Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

add example for Llama3 vllm server #404

Merged
merged 2 commits into from
Nov 11, 2024

Conversation

cameron-chen
Copy link
Contributor

Hello, I am creating this PR to share the example of evaluating by local model using API call (vllm server).

I find this approach can be quite useful when:

  • server the local annotator on the cluster so that other nodes only need to call API for evaluating.
  • the user wants to use "weighted"-style annotator similar with weighted_alpaca_eval_gpt4_turbo.

Please let me know if this looks good. I am happy to add more detailed instructions.


To use the API, add the client config local_configs.yaml and activate it:

default:
    - api_key: "token-abc123"
      base_url: "http://localhost:8000/v1"

@YannDubs
Copy link
Collaborator

LGTM, can you add a README.md in src/alpaca_eval/evaluators_configs/weighted_alpaca_eval_vllm_llama3_70b/ that explains the setup and goal? thanks!

@cameron-chen
Copy link
Contributor Author

LGTM, can you add a README.md in src/alpaca_eval/evaluators_configs/weighted_alpaca_eval_vllm_llama3_70b/ that explains the setup and goal? thanks!

Hi @YannDubs , thanks for looking into the PR! I have added README to provide more info regarding setup.

@cameron-chen
Copy link
Contributor Author

Hi @YannDubs , I wanted to check in on the PR after adding the README as you suggested. If there’s anything more I can clarify or adjust, please let me know. Thanks again for your feedback so far!

@YannDubs
Copy link
Collaborator

Look great, thanks @cameron-chen
sorry for the delay on that one

@YannDubs YannDubs closed this Nov 11, 2024
@YannDubs YannDubs reopened this Nov 11, 2024
@YannDubs YannDubs merged commit 0b4af76 into tatsu-lab:main Nov 11, 2024
2 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants