[PR #15050] [MERGED] ggml: force flash attention off for grok #77285

Closed
opened 2026-05-05 09:57:10 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/15050
Author: @rick-github
Created: 3/25/2026
Status: Merged
Merged: 3/25/2026
Merged by: @jessegross

Base: mainHead: grok


📝 Commits (1)

  • 97d09e4 ggml: force flash attention off for grok

📊 Changes

1 file changed (+1 additions, -1 deletions)

View changed files

📝 fs/ggml/ggml.go (+1 -1)

📄 Description

By default, ollama supports FA for grok, but llama.cpp disables it. If KV cache quantization has been set this causes the runner to crash.

Fixes: #15043


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/15050 **Author:** [@rick-github](https://github.com/rick-github) **Created:** 3/25/2026 **Status:** ✅ Merged **Merged:** 3/25/2026 **Merged by:** [@jessegross](https://github.com/jessegross) **Base:** `main` ← **Head:** `grok` --- ### 📝 Commits (1) - [`97d09e4`](https://github.com/ollama/ollama/commit/97d09e406009905bec12bf8b73a5caea1af46100) ggml: force flash attention off for grok ### 📊 Changes **1 file changed** (+1 additions, -1 deletions) <details> <summary>View changed files</summary> 📝 `fs/ggml/ggml.go` (+1 -1) </details> ### 📄 Description By default, ollama supports FA for grok, but llama.cpp [disables it](https://github.com/ollama/ollama/blob/4fda69809a3fecf73d0f71657ef50f8f7e8f43f7/llama/llama.cpp/src/llama-context.cpp#L2418). If KV cache quantization has been set this causes the runner to crash. Fixes: #15043 --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-05-05 09:57:10 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#77285