From 3d147c14a14c6bc8f5d97fcad707270bf90d78a2 Mon Sep 17 00:00:00 2001 From: Shroominic Date: Sat, 9 Dec 2023 16:28:32 -0800 Subject: [PATCH] =?UTF-8?q?=F0=9F=94=A7=20Adjust=20model=20parameters?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- src/funcchain/_llms.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/src/funcchain/_llms.py b/src/funcchain/_llms.py index 8390d6e..2d256eb 100644 --- a/src/funcchain/_llms.py +++ b/src/funcchain/_llms.py @@ -36,7 +36,7 @@ class _LlamaCppCommon(BaseLanguageModel): lora_path: Optional[str] = None """The path to the Llama LoRA. If None, no LoRa is loaded.""" - n_ctx: int = Field(512, alias="n_ctx") + n_ctx: int = Field(4096, alias="n_ctx") """Token context window.""" n_parts: int = Field(-1, alias="n_parts") @@ -72,7 +72,7 @@ class _LlamaCppCommon(BaseLanguageModel): suffix: Optional[str] = Field(None) """A suffix to append to the generated text. If None, no suffix is appended.""" - max_tokens: Optional[int] = 256 + max_tokens: Optional[int] = 512 """The maximum number of tokens to generate.""" temperature: Optional[float] = 0.8 @@ -128,7 +128,7 @@ class _LlamaCppCommon(BaseLanguageModel): emojis. At most one of grammar_path and grammar should be passed in. """ - verbose: bool = True + verbose: bool = False """Print verbose output to stderr.""" @root_validator()