Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch size beyond 16 is throwing an error #1177

Closed
venkycreator opened this issue Aug 1, 2024 · 2 comments
Closed

Batch size beyond 16 is throwing an error #1177

venkycreator opened this issue Aug 1, 2024 · 2 comments

Comments

@venkycreator
Copy link

Earlier I used to do inference using the past code till batch size of 64 , but don't know now it is not even running till batch size of 16 what things should i need to change in order to run as like earlier

@regisss
Copy link
Collaborator

regisss commented Aug 1, 2024

Please give more information:

  • what code are you talking about?
  • the commands you ran?
  • what are the versions of the library you used?

@venkycreator
Copy link
Author

venkycreator commented Aug 2, 2024

Earlier there is a code (for text generation run_generation.py) which doesn't have flash attention and also there is a argument called attn_softmax_bf16 , i exactly dont know what is the reason that i was unable to run batch size beyond 16

The commands that were there in the readme were only executed

And below attached file is the error that i got while running
error.txt

@regisss regisss closed this as completed Dec 16, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants