From 45e84cea0cc3efbb55bc814caff7a9e696256ee9 Mon Sep 17 00:00:00 2001 From: rjmacarthy Date: Wed, 6 Nov 2024 09:37:11 +0000 Subject: [PATCH] 3.18.4 --- package-lock.json | 4 ++-- package.json | 7 +++++-- 2 files changed, 7 insertions(+), 4 deletions(-) diff --git a/package-lock.json b/package-lock.json index bf2623ce..80a828ef 100644 --- a/package-lock.json +++ b/package-lock.json @@ -1,12 +1,12 @@ { "name": "twinny", - "version": "3.18.3", + "version": "3.18.4", "lockfileVersion": 3, "requires": true, "packages": { "": { "name": "twinny", - "version": "3.18.3", + "version": "3.18.4", "cpu": [ "x64", "arm64" diff --git a/package.json b/package.json index b39f5467..c8b357a0 100644 --- a/package.json +++ b/package.json @@ -2,7 +2,7 @@ "name": "twinny", "displayName": "twinny - AI Code Completion and Chat", "description": "Locally hosted AI code completion plugin for vscode", - "version": "3.18.3", + "version": "3.18.4", "icon": "assets/icon.png", "keywords": [ "code-inference", @@ -369,7 +369,10 @@ }, "twinny.ollamaApiPort": { "order": 15, - "type": ["number", "null"], + "type": [ + "number", + "null" + ], "default": 11434, "markdownDescription": "Sets the port number for the Ollama API. The default is `11434`, but you can change it if your Ollama instance uses a different port.", "required": false