[PR #825] [MERGED] relay CUDA errors to the client #15617

Closed
opened 2026-04-16 05:03:22 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/825
Author: @BruceMacD
Created: 10/17/2023
Status: Merged
Merged: 10/18/2023
Merged by: @BruceMacD

Base: mainHead: brucemacd/relay-cuda-errors


📝 Commits (1)

  • cf0a525 relay CUDA errors to the client

📊 Changes

1 file changed (+35 additions, -12 deletions)

View changed files

📝 llm/llama.go (+35 -12)

📄 Description

When the llama.cpp runner failed with CUDA error the error message was not relayed to the client. Instead the client would only see an EOF error. Update the llama.cpp subprocess log monitor to capture CUDA errors and relay them to the client.

Before:

Error: error reading llm response: unexpected EOF

After:

Error: llama runner exited, you may not have enough available memory to run this model

or the actual error is available


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/825 **Author:** [@BruceMacD](https://github.com/BruceMacD) **Created:** 10/17/2023 **Status:** ✅ Merged **Merged:** 10/18/2023 **Merged by:** [@BruceMacD](https://github.com/BruceMacD) **Base:** `main` ← **Head:** `brucemacd/relay-cuda-errors` --- ### 📝 Commits (1) - [`cf0a525`](https://github.com/ollama/ollama/commit/cf0a525aa1a4534b2e520589f2897d2c82f3df40) relay CUDA errors to the client ### 📊 Changes **1 file changed** (+35 additions, -12 deletions) <details> <summary>View changed files</summary> 📝 `llm/llama.go` (+35 -12) </details> ### 📄 Description When the llama.cpp runner failed with CUDA error the error message was not relayed to the client. Instead the client would only see an EOF error. Update the llama.cpp subprocess log monitor to capture CUDA errors and relay them to the client. Before: ``` Error: error reading llm response: unexpected EOF ``` After: ``` Error: llama runner exited, you may not have enough available memory to run this model ``` or the actual error is available --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-16 05:03:22 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#15617