Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Hyperparameter Settings and Distillation Rounds in CIFAR-10 Knowledge Distillation #30

Open
ConquerMyHer opened this issue Dec 17, 2024 · 0 comments

Comments

@ConquerMyHer
Copy link

A respectful salute to your work.

I am currently attempting to reproduce the CIFAR-10 knowledge distillation results in your paper, which are based on inversion and adversarial losses.

Could you please provide more details on the following points?

  1. For the distillation setting, should I use the same inversion hyperparameters as those listed under the CIFAR-10 directory? If not, could you kindly provide the corresponding hyperparameters?

  2. As I understand, your distillation process is iterative. In other words, you first generate a batch of synthetic samples, then train the student model on these samples, and repeat this cycle multiple times so that the adversarial loss can take effect. Could you confirm this understanding? Also, could you please provide the number of synthetic samples generated in each distillation round, as well as the total number of rounds you conducted?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant