-
Notifications
You must be signed in to change notification settings - Fork 115
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add a recipe that works with function calling models #562
Comments
Depends on #601 |
#601 merged, can we close this? |
@rhatdan no. This issue is related to the need to have something similar to the chatbot/rag recipe. So we can add it to ai-lab extension. |
To create a recipe providing function calling, we have the following requirements:
I opened abetlen/llama-cpp-python#1636 which should be fixed abetlen/llama-cpp-python#1637 |
I made some experiments locally with the amazing Instructor library and here are some demos I made using llama-cpp-python 0.83 (any version above crashes) Instructor libraryInformation extractionGiven an unstructured text, return a python model. demo-instructor-1.mp4A basic pydantic model provided, the instructor will translate the llm response to the model provided, with validation etc. Single text classificationdemo-instructor-single-classification.mp4Given an enum and a text, got the enum value as result. Allowing basic single text classification. Multi text classificationdemo-multi-classification.mp4We can request multiple values as result, allowing to get a multi-classification. Function callingThe function calling need a bit of manual change, as we need abetlen/llama-cpp-python#1509 applied, and some adjustment to be able to uses the demo-function-calling.mp4Thoses example are took from the llama-cpp-python repository2 Footnotes |
Fixes containers#562 Signed-off-by: Jeff MAURY <[email protected]>
In podman desktop ai lab we are planning to add function calling models support -> containers/podman-desktop-extension-ai-lab#1116
We would need a recipe + a list of recommended models to work with
The text was updated successfully, but these errors were encountered: