[PR #2018] [MERGED] fix: cache prompt causes kv cache to fill and not return after some time #10760

Closed
opened 2026-04-12 23:09:55 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/2018
Author: @BruceMacD
Created: 1/16/2024
Status: Merged
Merged: 1/16/2024
Merged by: @BruceMacD

Base: mainHead: brucemacd/cache-prompt-stuck


📝 Commits (1)

📊 Changes

1 file changed (+0 additions, -1 deletions)

View changed files

📝 llm/dyn_ext_server.go (+0 -1)

📄 Description

  • prompt cache causes inferance to hang after some time

This is a temporary fix to mitigate #1994 if I can't fix the root cause before the next release.


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/2018 **Author:** [@BruceMacD](https://github.com/BruceMacD) **Created:** 1/16/2024 **Status:** ✅ Merged **Merged:** 1/16/2024 **Merged by:** [@BruceMacD](https://github.com/BruceMacD) **Base:** `main` ← **Head:** `brucemacd/cache-prompt-stuck` --- ### 📝 Commits (1) - [`743493a`](https://github.com/ollama/ollama/commit/743493a0c5d7dd8b14e58fb106758eb39c1749e6) do not cache prompt ### 📊 Changes **1 file changed** (+0 additions, -1 deletions) <details> <summary>View changed files</summary> 📝 `llm/dyn_ext_server.go` (+0 -1) </details> ### 📄 Description - prompt cache causes inferance to hang after some time This is a temporary fix to mitigate #1994 if I can't fix the root cause before the next release. --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-12 23:09:55 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#10760