Skip to content

Commit

Permalink
copy
Browse files Browse the repository at this point in the history
  • Loading branch information
paul-gauthier committed Nov 24, 2024
1 parent 939d7ea commit 0c59d32
Showing 1 changed file with 10 additions and 4 deletions.
14 changes: 10 additions & 4 deletions aider/website/docs/llms/ollama.md
Original file line number Diff line number Diff line change
Expand Up @@ -44,6 +44,13 @@ setx OLLAMA_API_KEY <api-key> # Windows, restart shell after setx

[Ollama uses a 2k context window by default](https://github.com/ollama/ollama/blob/main/docs/faq.md#how-can-i-specify-the-context-window-size),
which is very small for working with aider.
Unlike most other LLM servers, Ollama does not throw an error if you submit
a request that exceeds the context window.
Instead, it just silently truncates the request by discarding the "oldest" messages
in the chat to make it fit within the context window.

All of the Ollama results above were collected with at least an 8k context window, which
is large enough to attempt all the coding problems in the benchmark.

You can set the Ollama server's context window with a
[`.aider.model.settings.yml` file](https://aider.chat/docs/config/adv-model-settings.html#model-settings)
Expand All @@ -52,14 +59,13 @@ like this:
```
- name: aider/extra_params
extra_params:
num_ctx: 65536
num_ctx: 8192
```

That uses the special model name `aider/extra_params` to set it for *all* models. You should probably use a specific model name like:

```
- name: ollama_chat/qwen2.5-coder:32b-instruct-fp16
- name: ollama/qwen2.5-coder:32b-instruct-fp16
extra_params:
num_ctx: 65536
num_ctx: 8192
```

0 comments on commit 0c59d32

Please sign in to comment.