-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pretraining Datasets #3
Comments
Sorry for the late reply, we use the Viznet tables for pre-training the column encoder which can be found in this page: https://github.com/megagonlabs/sato/tree/master/table_data |
Thanks a lot for your sourcing the code and your answer @jw-megagon. I was wondering did you train the model on all tables of VisNet (80000) or you used the multi-column sets only? Moreover, could you please provide the hyperparameters (--batch_size, --lr --lm, --n_epochs , --max_len , --size, --projector, --augment_op, --sample_meth, --table_order) you used in the training process? Thanks in advance! |
To obtain the training data for Viznet, I saved all the tables from the folders within the |
I have obtained the Thanks in advance! : ) |
Thank you for open-sourcing the code! I didn't find descriptions about pretraining datasets in the paper. Was Starmie pertained on benchmark datasets?
The text was updated successfully, but these errors were encountered: