Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About parameter number and in_dim #10

Open
cheryyunl opened this issue Mar 18, 2024 · 7 comments
Open

About parameter number and in_dim #10

cheryyunl opened this issue Mar 18, 2024 · 7 comments

Comments

@cheryyunl
Copy link

Hi, I find that in the experiments for entire networks, the parameter number is very large (more than 50,000 or 100,000).
How to use the encoder to train the parameter embedding? If the in_dim = parameter number?
Would you like to provide the training details (especially the autoencoder) in the entire network generation?

Thank you for your reply!!!

@cheryyunl cheryyunl changed the title bAout parameter number and in_dim About parameter number and in_dim Mar 18, 2024
@1zeryu
Copy link
Collaborator

1zeryu commented Mar 19, 2024

Thanks for your feedback.
For layer parameter, you should change system.ae_model.in_dim = parameter num.
You should change the hyperparameter about fold_rate or layer num for autoencoder, the details about autoencoder in https://github.com/NUS-HPC-AI-Lab/Neural-Network-Diffusion/blob/main/core/module/modules/encoder.py.

@cheryyunl
Copy link
Author

Thank you so much for your quick reply!
In the encoder.py for the entire network parameter generator, if in_dim = 50, 000, will there be a lot of pressure on the memory and training time required for training? Is there any dimensionality reduction operation?

@1zeryu
Copy link
Collaborator

1zeryu commented Mar 19, 2024

Your problems are not accidental. We also find it tricky to deal with large model parameters

The information that can be provided is that Latent_AE_cnn_big under core/module/modules/autoencoder.py is the autoencoder model for large model parameters. But it also takes a long time to train, and our future task is to solve large-scale parameter generation.

Thanks for your feedback again.

@cheryyunl
Copy link
Author

Thank you!!! I will try to use Latent_AE_cnn_big for training autoencoder.
An additional question is for the entire network parameters, the paper also use 200 checkpoints for parameter encoder training? For Latent_AE_cnn_big, the number of samples looks small. Do you try to use more checkpoints?

Thanks again!!

@1zeryu
Copy link
Collaborator

1zeryu commented Mar 25, 2024

Sorry to hear that you are having such difficulty. But I never try more checkpoints.
We will update the p-diff for easier large parameter generation in the future.

@SKDDJ
Copy link

SKDDJ commented Mar 25, 2024

Hi @1zeryu , would you mind adding more details about your code in the README? I am having trouble understanding the hierarchical structure and how to make modifications.

@1zeryu
Copy link
Collaborator

1zeryu commented Mar 26, 2024

Sure, I will do it soon.
Thanks for your feedback.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants