Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support Yi & StableLM models, change default maximum length of generated tokens for smooth chat. #57

Merged
merged 3 commits into from
Jul 15, 2024

Conversation

guoqingbao
Copy link
Collaborator

Key changes in this PR:

  1. Support Yi & StableLM models
  2. The maximum length of generated tokens changed to 1/5 of max_seq_len by default, this will enalbe smooth chat with minimal breaks. Users are allowed to set max_gen_tokens for the model throught parameters.

Tested cases:

cargo run --release -- --port 2000 --weight-path /home/stablelm-zephyr-3b/ stable-lm --repeat-last-n 32
cargo run --release -- --port 2000 --weight-path /home/yi-6b/ yi --repeat-last-n 32

@guoqingbao guoqingbao merged commit 0be4121 into master Jul 15, 2024
5 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant