Add ollama phi4 context size defaults (#23036)
Add `phi4` maximum context length (128K). By default this clamps to `16384` but if you have enough video memory you can set it higher or connect to a non-local machine via settings: ```json "language_models": { "ollama": { "api_url": "http://localhost:11434", "available_models": [ { "name": "phi4", "display_name": "Phi4 64K", "max_tokens": 65536 } ] } } ``` Release Notes: - Improve support for Phi4 with ollama.
This commit is contained in:
parent
0e4a619c9f
commit
67525cca71
1 changed files with 2 additions and 2 deletions
|
@ -83,8 +83,8 @@ fn get_max_tokens(name: &str) -> usize {
|
|||
"codellama" | "starcoder2" => 16384,
|
||||
"mistral" | "codestral" | "mixstral" | "llava" | "qwen2" | "qwen2.5-coder"
|
||||
| "dolphin-mixtral" => 32768,
|
||||
"llama3.1" | "phi3" | "phi3.5" | "command-r" | "deepseek-coder-v2" | "yi-coder"
|
||||
| "llama3.2" => 128000,
|
||||
"llama3.1" | "phi3" | "phi3.5" | "phi4" | "command-r" | "deepseek-coder-v2"
|
||||
| "yi-coder" | "llama3.2" => 128000,
|
||||
_ => DEFAULT_TOKENS,
|
||||
}
|
||||
.clamp(1, MAXIMUM_TOKENS)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue