ollama: Add Qwen3 and Gemma3 (default to 16K context) (#29580)

If you have the VRAM you can increase the context by adding this to your
settings.json:

```json
  "language_models": {
    "ollama": {
      "available_models": [
        { "max_tokens": 65536, "name": "qwen3", "display_name": "Qwen3-64k" }
      ]
    }
  },
```

Release Notes:

- ollama: Add support for Qwen3. Defaults to 16K token context. See:
[Assistant Configuration
Docs](https://zed.dev/docs/assistant/configuration#ollama-context) to
increase.
This commit is contained in:
Peter Tripp 2025-04-29 01:44:28 +00:00 committed by GitHub
parent 2cc5a0de26
commit 4dc8ce8cf7
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -84,7 +84,9 @@ fn get_max_tokens(name: &str) -> usize {
"mistral" | "codestral" | "mixstral" | "llava" | "qwen2" | "qwen2.5-coder"
| "dolphin-mixtral" => 32768,
"llama3.1" | "llama3.2" | "llama3.3" | "phi3" | "phi3.5" | "phi4" | "command-r"
| "deepseek-coder-v2" | "deepseek-v3" | "deepseek-r1" | "yi-coder" => 128000,
| "qwen3" | "gemma3" | "deepseek-coder-v2" | "deepseek-v3" | "deepseek-r1" | "yi-coder" => {
128000
}
_ => DEFAULT_TOKENS,
}
.clamp(1, MAXIMUM_TOKENS)