-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
gradient estimation #38
Comments
The basic problem we face is computing the gradients and its gradient: The general Monte Carlo estimator of the stochastic objective function is: There are 2 general families of methods to develop gradient estimators for the stochastic gradient
These families each have different strengths and weaknesses due to their construction. There are 3 major classes of estimators that fall into these 2 families considered in the review paper. Score function estimator This estimator belongs to the class of derivatives of measure. It's named after the score function, the gradient of the log probability with respect to parameters: $$ This final formulation as an expectation motivates the straightforward Monte Carlo estimator: Pathwise estimator As the name suggests, this one belongs to the family of derivatives of paths. The pathwise estimator is so called because we can usually separate the sampling of the measure Which leads to $$ The Monte Carlo estimator of the gradient is again straightforward once we get the expectation: This one is nice when we can apply it, because after our draw from the base distribution we push Measure-valued estimator This also belongs to the family of derivatives of measure. It depends on some relatively new results on weak derivatives, which are derivatives of a density with respect to a single parameter depending on a specific decomposition. Where the triple The estimator is defined for each index of $$ In the final line we see that we get the estimator as a difference in expectations, so we can formulate the Monte Carlo estimator straightforwardly. |
What is StochasticAD.jl doing?
Can we line up notation?
The construction is given below, without much care for rigor. Let's saw we already do our "primal" draw This seems to be what's implemented in the code, see first part and second part |
@adolgert we should pay attention to this: https://github.com/arnauqb/diff_abms_tutorial/tree/main |
For me, I need to read https://arxiv.org/abs/1906.10652 as background and then familiarize myself with https://github.com/gaurav-arya/StochasticAD.jl/tree/main and https://arxiv.org/abs/2210.08572
The text was updated successfully, but these errors were encountered: