Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

关于并行计算 #17

Open
penglo opened this issue Jun 13, 2024 · 1 comment
Open

关于并行计算 #17

penglo opened this issue Jun 13, 2024 · 1 comment

Comments

@penglo
Copy link

penglo commented Jun 13, 2024

你好,恭喜您取得如此优秀的成果,我在复现你的代码的时候也是使用了8块3090,但是还是跑不起来您所使用的batchsize ,我研究了一下您使用的多卡训练的代码采用的是数据并行的方式,但是大多数我们用来解决多卡运行显存不足时,通常使用的是模型并行的方式,想请教一下您是否做过这方面的部署呢,如果有,希望您开源一下代码,或者我们进行一些交流,我的邮箱是[email protected]
Hello,Congratulations on achieving such excellent results. While reproducing your code, I also used 8 RTX 3090 GPUs, but I still couldn't manage to run the batch size you used. After some research, I noticed that your multi-GPU training code uses data parallelism, while most of us typically use model parallelism to solve the issue of insufficient GPU memory. I would like to ask if you have deployed model parallelism in your setup. If so, could you please share your code, or perhaps we could have some discussions on this matter? My email is [email protected] regards.

@MasterIzumi
Copy link
Collaborator

@penglo Hi, 本文中的网络比较轻量,应该不会遇到一块卡显存不够存模型而需要模型并行的情况。我们仅提供了DDP数据并行的实现,关于模型并行并没有进行尝试。

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants