Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

why only 49-batches for weights? #17

Open
777Adonis opened this issue Sep 10, 2019 · 1 comment
Open

why only 49-batches for weights? #17

777Adonis opened this issue Sep 10, 2019 · 1 comment

Comments

@777Adonis
Copy link

Thank you for your contribution, but your pre-training weights file has only 49 batches. I want to know if you have used the learning rate attenuation strategy?

@elvisyjlin
Copy link
Owner

You're welcome. Sorry for my late reply. I was busy in the past 3 months.

You probably meant weights.49.pth, which is the model weights at the 49th epoch. We apply the same learning rate attenuation strategy as the code by the original author. The model is updated with base learning rate in the first 100 epochs whilst the learning rate is halved in the last 100 epochs.

However, the whole training is too long. It took about 2 weeks to train under one setting on my machine. For the sack of time, we simply train each model for 49 epochs. Please do the training yourself to get the 200-epoched model. Yet, believe me, the performance is actually good enough with 49 epochs.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants