Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add sequence packing for lm-eval-harness #850

Merged
merged 21 commits into from
Jan 8, 2025
Merged

Add sequence packing for lm-eval-harness #850

merged 21 commits into from
Jan 8, 2025

Conversation

dlwh
Copy link
Member

@dlwh dlwh commented Jan 7, 2025

Speedup was less than I hoped, but I think that can be improved with better packing strategies (and also removing some batching overhead. the bottleneck is now data loading??)

@dlwh dlwh changed the title Packing Add sequence packing for lm-eval-harness Jan 7, 2025
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

was this intended to be part of the PR? Just checking

Copy link
Member Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

nah, but i don't care enough to take it out :-)

@nikil-ravi
Copy link
Contributor

nikil-ravi commented Jan 8, 2025

Curious whether we know the rough ratio of padding:non padding tokens with this packing strategy?

@dlwh
Copy link
Member Author

dlwh commented Jan 8, 2025

It's about 90% real tokens (as opposed to ~90% padding before)

@dlwh dlwh merged commit 93a8aa9 into main Jan 8, 2025
8 checks passed
@dlwh dlwh deleted the packing branch January 8, 2025 18:32
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants