Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

dtheta update issue? transforming dtheta for getting theta? #20

Open
mamunir opened this issue May 21, 2019 · 1 comment
Open

dtheta update issue? transforming dtheta for getting theta? #20

mamunir opened this issue May 21, 2019 · 1 comment

Comments

@mamunir
Copy link

mamunir commented May 21, 2019

updating dtheta gives higher than 1 in the parameters, even more than 1000? is it normal? I do not think so. Do I miss some point?

I re-implement in python
I send [dV value, U value, and dU value for every single coordinate respectively, x-y normalized coordinates] to backprop function.

@daerduoCarey your help in this regard is appreciated.

@daerduoCarey
Copy link
Owner

Training such a layer requires some careful tuning to control the gradients. Without your details of training, I don't think I can help. A practical good thing is to try adjust your learning rate to be smaller enough first to see if you still observe this gradient overflow. Or, you can try to make the lr for the stn to be smaller than the rest backbone.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants