From 2e64b0ab3a10f5b454df149c1041c9fe3289affc Mon Sep 17 00:00:00 2001 From: Kasper Marstal Date: Thu, 21 Nov 2024 22:28:46 +0100 Subject: [PATCH] feat: Update Llamafile version, add larger Llamafile models --- src/Cellm/appsettings.Local.Anthropic.json | 2 +- src/Cellm/appsettings.Local.Llamafile.GPU.json | 4 +++- src/Cellm/appsettings.Local.Llamafile.json | 6 ++++-- src/Cellm/appsettings.Local.Mistral.json | 2 +- src/Cellm/appsettings.Local.OpenAi.json | 2 +- 5 files changed, 10 insertions(+), 6 deletions(-) diff --git a/src/Cellm/appsettings.Local.Anthropic.json b/src/Cellm/appsettings.Local.Anthropic.json index 646b425..e8f2ea5 100644 --- a/src/Cellm/appsettings.Local.Anthropic.json +++ b/src/Cellm/appsettings.Local.Anthropic.json @@ -1,7 +1,7 @@ { "AnthropicConfiguration": { "DefaultModel": "claude-3-5-sonnet-20241022", - "ApiKey": "YOUR_ANTHROPIC_APIKEY" + "ApiKey": "YOUR_ANTHROPIC_API_KEY" }, "CellmConfiguration": { "DefaultProvider": "Anthropic", diff --git a/src/Cellm/appsettings.Local.Llamafile.GPU.json b/src/Cellm/appsettings.Local.Llamafile.GPU.json index c5e4de6..2c84d7a 100644 --- a/src/Cellm/appsettings.Local.Llamafile.GPU.json +++ b/src/Cellm/appsettings.Local.Llamafile.GPU.json @@ -5,8 +5,10 @@ "DefaultModel": "gemma-2-2b", "Models": { "gemma-2-2b": "https://huggingface.co/bartowski/gemma-2-2b-it-GGUF/resolve/main/gemma-2-2b-it-Q6_K.gguf", + "gemma-2-9b": "https://huggingface.co/bartowski/gemma-2-9b-it-GGUF/resolve/main/gemma-2-9b-it-Q4_K_L.gguf", "llama-3.2-1b": "https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF/resolve/main/Llama-3.2-1B-Instruct-Q6_K_L.gguf", - "llama-3.2-3b": "https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q4_K_L.gguf" + "llama-3.2-3b": "https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q4_K_L.gguf", + "ministral-8b": "https://huggingface.co/bartowski/Ministral-8B-Instruct-2410-GGUF/resolve/main/Ministral-8B-Instruct-2410-Q5_K_L.gguf" }, "GPU": true, "GpuLayers": 999 diff --git a/src/Cellm/appsettings.Local.Llamafile.json b/src/Cellm/appsettings.Local.Llamafile.json index bb50b19..c6474fb 100644 --- a/src/Cellm/appsettings.Local.Llamafile.json +++ b/src/Cellm/appsettings.Local.Llamafile.json @@ -1,12 +1,14 @@ { "LlamafileConfiguration": { - "LlamafileUrl": "https://github.com/Mozilla-Ocho/llamafile/releases/download/0.8.13/llamafile-0.8.13", + "LlamafileUrl": "https://github.com/Mozilla-Ocho/llamafile/releases/download/0.8.16/llamafile-0.8.16", "BaseAddress": "http://127.0.0.1", "DefaultModel": "gemma-2-2b", "Models": { "gemma-2-2b": "https://huggingface.co/bartowski/gemma-2-2b-it-GGUF/resolve/main/gemma-2-2b-it-Q6_K.gguf", + "gemma-2-9b": "https://huggingface.co/bartowski/gemma-2-9b-it-GGUF/resolve/main/gemma-2-9b-it-Q4_K_L.gguf", "llama-3.2-1b": "https://huggingface.co/bartowski/Llama-3.2-1B-Instruct-GGUF/resolve/main/Llama-3.2-1B-Instruct-Q6_K_L.gguf", - "llama-3.2-3b": "https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q4_K_L.gguf" + "llama-3.2-3b": "https://huggingface.co/bartowski/Llama-3.2-3B-Instruct-GGUF/resolve/main/Llama-3.2-3B-Instruct-Q4_K_L.gguf", + "ministral-8b": "https://huggingface.co/bartowski/Ministral-8B-Instruct-2410-GGUF/resolve/main/Ministral-8B-Instruct-2410-Q5_K_L.gguf" } }, "CellmConfiguration": { diff --git a/src/Cellm/appsettings.Local.Mistral.json b/src/Cellm/appsettings.Local.Mistral.json index f00ceb0..4a64398 100644 --- a/src/Cellm/appsettings.Local.Mistral.json +++ b/src/Cellm/appsettings.Local.Mistral.json @@ -2,7 +2,7 @@ "OpenAiConfiguration": { "BaseAddress": "https://api.mistral.ai", "DefaultModel": "mistral-small-latest", - "ApiKey": "YOUR_MISTRAL_APIKEY" + "ApiKey": "YOUR_MISTRAL_API_KEY" }, "CellmConfiguration": { "DefaultProvider": "OpenAI", diff --git a/src/Cellm/appsettings.Local.OpenAi.json b/src/Cellm/appsettings.Local.OpenAi.json index 295db65..de345fc 100644 --- a/src/Cellm/appsettings.Local.OpenAi.json +++ b/src/Cellm/appsettings.Local.OpenAi.json @@ -1,7 +1,7 @@ { "OpenAiConfiguration": { "DefaultModel": "gpt-4o-mini", - "ApiKey": "YOUR_OPENAI_APIKEY" + "ApiKey": "YOUR_OPENAI_API_KEY" }, "CellmConfiguration": { "DefaultProvider": "OpenAI",