-
Notifications
You must be signed in to change notification settings - Fork 4.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
hidden_dim constraint in transformer cuda kernel #491
Comments
Hi @szhengac Thanks for pointing this out. We are currently working on supporting arbitrary dimensions. There will be a code update soon to add this feature. Please stay tuned! :-) Thanks. |
Hi @szhengac Sorry for the delay! Could you please try this and let me know if it works for your training environment? Thank you, |
@RezaYazdaniAminabadi. Thanks for the good work! I will try it when I get some free time. |
Hi @RezaYazdaniAminabadi, we recently came across another issue of using DeepSpeed Transformer kernel as shown in #589 |
I found that there is constraint on the dimensionality when we use the transformer cuda kernel: https://github.com/microsoft/DeepSpeed/blob/d720fdb6857f4b71d922ca1e8efbe5271b5fb7c2/csrc/transformer/normalize_kernels.cu#L232-L250
I wonder what is the reason behind it? Is there any plan to support arbitrary dimensionality? Or, If I want to use
hidden_dim=4096 or 8192
, what do I need to do to make it work? Thanks.The text was updated successfully, but these errors were encountered: