You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm not sure whether this distributed training codes are correct. I tried training with 8xGPU and got lower accuracy than training with. My setting is:
I think training with batch-size=4N on 1 GPU is exactly equivalent to training with batch-size=N on 4 GPU ( when syncBN is on )
What's the correct way to train on multi-GPU ?
The text was updated successfully, but these errors were encountered:
Hi @Cathy0908, how far your results different from the paper. I have trained on Cars196 with one Nvidia A100, and all the results is lower than in the paper. The lower the K value in recall@K, the further the gap.
some codes in
train.py
is like:I'm not sure whether this distributed training codes are correct. I tried training with 8xGPU and got lower accuracy than training with. My setting is:
Train 1 GPU:
Train 8 GPU:
I think training with batch-size=4N on 1 GPU is exactly equivalent to training with batch-size=N on 4 GPU ( when syncBN is on )
What's the correct way to train on multi-GPU ?
The text was updated successfully, but these errors were encountered: