You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This exercise is also bit misleading since it gives you the impression that we got out of local minima because we did zeroth-order, when in fact, we could have achieved the same effect by doing first order while injecting stochasticity.
I'd love to reimplement this problem with what we've learned with various gradient estimators over the past year.
The text was updated successfully, but these errors were encountered:
The stochastic approximation has
x <- x - eta * [ l(x + w) - l(x) ] w
but the true gradient should be
x <- x - eta * [ l(x + w) - l(x) ] w / sigma^2
so we seem to be off at a scale.
This exercise is also bit misleading since it gives you the impression that we got out of local minima because we did zeroth-order, when in fact, we could have achieved the same effect by doing first order while injecting stochasticity.
I'd love to reimplement this problem with what we've learned with various gradient estimators over the past year.
The text was updated successfully, but these errors were encountered: