You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Is it possible to share the exact hyperparameters used for the ablations on the Rainbow environment? I am trying to recreate these results using the smaller, default Transformer size (3 layers, 128 dim, 8 heads). However, I find that most problems are solved at exactly 64 nodes expanded. (Interestingly there also seems to be a jump in Figure 4 from the paper at 64 nodes expanded.)
Here are what my current results look like:
Thanks so much!
The text was updated successfully, but these errors were encountered:
I just checked my wandb logs, the hyperparameters you used are indeed correct. There is a small quirk in the way the number of nodes is calculated across different methods. Particularly, for rejection sampling it isn't clear if, say, the 20th generation of parallel generation of 64 programs should be counted as 20 expanded nodes, or 64 since that compute was spent.
I forget how we did these between different methods, let me take a look into the codebase and re-run these ablations again to see if maybe there is a bug with quantization.
Very nice work!
Is it possible to share the exact hyperparameters used for the ablations on the Rainbow environment? I am trying to recreate these results using the smaller, default Transformer size (3 layers, 128 dim, 8 heads). However, I find that most problems are solved at exactly 64 nodes expanded. (Interestingly there also seems to be a jump in Figure 4 from the paper at 64 nodes expanded.)
Here are what my current results look like:
Thanks so much!
The text was updated successfully, but these errors were encountered: