mirror of
https://github.com/ollama/ollama.git
synced 2025-12-05 19:16:53 -06:00
llm: Enable flash attention for mistral3 by default
This commit is contained in:
@@ -831,6 +831,7 @@ func (f GGML) FlashAttention() bool {
|
||||
return slices.Contains([]string{
|
||||
"gemma3",
|
||||
"gptoss", "gpt-oss",
|
||||
"mistral3",
|
||||
"qwen3", "qwen3moe",
|
||||
"qwen3vl", "qwen3vlmoe",
|
||||
}, f.KV().String("general.architecture"))
|
||||
|
||||
Reference in New Issue
Block a user