Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

B1565 #16

Merged
merged 10 commits into from
Nov 26, 2023
Merged

B1565 #16

merged 10 commits into from
Nov 26, 2023

Conversation

Nexesenex
Copy link
Owner

No description provided.

jammm and others added 10 commits November 24, 2023 09:52
* Update README.md to use PATH for Windows ROCm

* Update README.md

* Update README.md
llama_token_eos(const struct llama_model *) is currently getting struct llama_context type variable context as a parameter.
* ggml-cuda : support stablelm rope

* remove unused freq_base kernel parameter

* add n_dims parameter to llm_build_k_shift, default to n_rot via overload

* llama : fix llm_build_k_shift args

---------

Co-authored-by: Georgi Gerganov <[email protected]>
* Add openai-compatible POST /v1/chat/completions API endpoint to server example

* fix code style

* Update server README.md

* Improve server README.md

* Fix server.cpp code style according to review

* server : some style changes

* server : indentation

* server : enable special tokens during tokenization by default

* server : minor code style

* server : change random string generator

* straightforward /v1/models endpoint

---------

Co-authored-by: kir-gadjello <[email protected]>
Co-authored-by: Tobi Lütke <[email protected]>
* reserve space for codepoints

* improvement for the appended 0
* Use mmap in torch load, prefer .bin files when loading

* Revert .bin > .safetensors preference
@Nexesenex Nexesenex merged commit 24e8996 into Nexesenex:master_experimental Nov 26, 2023
31 of 36 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

8 participants