You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
print(llm.create_chat_completion([{'role': 'user', 'content': 'What is the capital of France? Replay using a JSON: {"answer": "YOUR_ANSWER"}!'}], response_format={ "type": "json_object" }))
Due to:
libc++abi: terminating due to uncaught exception of type std::out_of_range: vector
[1] 25897 abort /Users/shakedz/bitbucket/achilles/.venv/bin/python
But without the response_format:
print(llm.create_chat_completion([{'role': 'user', 'content': 'What is the capital of France? Replay using a JSON: {"answer": "YOUR_ANSWER"}!'}]))
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Expected Behavior
Please provide a detailed written description of what you were trying to do, and what you expected
llama-cpp-python
to do.Current Behavior
Using
response_format={ "type": "json_object" }
crashesEnvironment and Context
I'm using:
Which were downloaded form here:
https://huggingface.co/lmstudio-community/Meta-Llama-3.1-8B-Instruct-GGUF/tree/main
This fails:
Due to:
But without the response_format:
It works:
System
CMAKE_ARGS="-DGGML_METAL=on" pip install llama-cpp-python
The text was updated successfully, but these errors were encountered: