Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why use fixstep lr #8

Open
liu6381810 opened this issue Aug 12, 2020 · 3 comments
Open

Why use fixstep lr #8

liu6381810 opened this issue Aug 12, 2020 · 3 comments
Labels
question Further information is requested

Comments

@liu6381810
Copy link

Thanks for your work!
I have 2 questions:
In the paper, the author use cos lr but why fixsLR used in your code?
And now W3A2 acc has reported but I want to know whether you train W4A4 resnet18 reaching the acc in the paper?

@zhutmost
Copy link
Owner

Ahhh, I cannot obtain better acc with cos decay, and I used step decay but fixed decay.
And the config.yaml is just a template. Please refer the configuration file in the example folder.

I have tried W4A4, but its accuracy is less than the authors' one.

@zhutmost zhutmost added the question Further information is requested label Aug 12, 2020
@liu6381810
Copy link
Author

Thanks for your reply,
Another quesiton I want to know is have you ever verified the effect of grad_scale
In my early quantization experiment(not lsq) weight lr is often much smaller than scale lr
but with grad scale it's approximate decay the lr for scale, it's confused for me

@zhutmost
Copy link
Owner

Sorry not yet.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants