diff --git a/package.json b/package.json index 62839969..02b88b9a 100644 --- a/package.json +++ b/package.json @@ -239,7 +239,8 @@ "ollama", "llamacpp", "lmstudio", - "oobabooga" + "oobabooga", + "other" ], "default": "ollama", "description": "The API provider to use (sets the paths and port automatically to defaults)." diff --git a/src/common/types.ts b/src/common/types.ts index 79dac069..ebd857d5 100644 --- a/src/common/types.ts +++ b/src/common/types.ts @@ -147,7 +147,8 @@ export const ApiProviders = { OllamaWebUi: 'ollamawebui', LlamaCpp: 'llamacpp', LMStudio: 'lmstudio', - Oobabooga: 'oobabooga' + Oobabooga: 'oobabooga', + Other: 'other' } as const export interface OllamaModel { diff --git a/src/extension/model-options.ts b/src/extension/model-options.ts index c221cbcf..e1aac855 100644 --- a/src/extension/model-options.ts +++ b/src/extension/model-options.ts @@ -38,10 +38,10 @@ export function createStreamRequestBody( temperature: options.temperature, n_predict: options.numPredictChat } + case ApiProviders.Other: default: return { model: options.model, - prompt, stream: true, max_tokens: options.numPredictChat, messages: options.messages,