Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ads parallel MLPerf accuracy eval. #1059

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open

Conversation

patemotter
Copy link
Collaborator

@patemotter patemotter commented Nov 22, 2024

For MLPerf accuracy runs we need to compute the ROUGE scores and compare against expected outputs.

This change adds a parallelized the code, which provides a speedup of 20x (10m -> 0.5m) when evaluating the 25k MLPerf dataset for Llama2-70B. Since this doesn't change the underlying computation of the scores we get the same results as the serial version.

Original Serial Results
{'rouge1': 44.3842, 'rouge2': 21.8723, 'rougeL': 28.5027, 'rougeLsum': 41.947, 'gen_len': 28135648, 'gen_num': 24576, 'gen_tok_len': 7171079, 'tokens_per_sample': 291.8}

real    10m4.329s
user    10m8.792s
sys     0m8.158s


Parallel Results
{'rouge1': 44.3842, 'rouge2': 21.8723, 'rougeL': 28.5027, 'rougeLsum': 41.947, 'gen_len': 28135648, 'gen_num': 24576, 'gen_tok_len': 7171079, 'tokens_per_sample': 291.8}

real    0m29.519s
user    13m2.545s
sys     2m43.586s

@patemotter patemotter changed the title Parallelizes MLPerf accuracy eval. Ads parallel MLPerf accuracy eval. Nov 22, 2024
@vipannalla
Copy link
Collaborator

Can add a way (env_car or option in llama_offline_run.sh) to use this new and faster eval file?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants