Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

降低训练显存占用可能性 #5

Open
codexq123 opened this issue Jan 9, 2025 · 1 comment
Open

降低训练显存占用可能性 #5

codexq123 opened this issue Jan 9, 2025 · 1 comment

Comments

@codexq123
Copy link

有可能利用fp8训练,将显存占用降低到24g吗

@Passenger12138
Copy link
Owner

If your hardware supports the FP8 format, you can certainly use FP8 for training and adopt mixed precision, where the large model operates in BF16 and the LoRA module in FP8, to reduce memory usage. However, it's important to note that the LoRA module itself has relatively few parameters, so using FP8 may not yield significant benefits in this case.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants