You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Can I parallelize just one large layer across gpus and keep all other layers the same and work with distributed data parallel?
For example, a regular resnet18 model, which is small enough but the number of classes is 1 million, how do I parallelize just the classification layer?
The text was updated successfully, but these errors were encountered:
Can I parallelize just one large layer across gpus and keep all other layers the same and work with distributed data parallel?
For example, a regular resnet18 model, which is small enough but the number of classes is 1 million, how do I parallelize just the classification layer?
The text was updated successfully, but these errors were encountered: