[GH-ISSUE #15290] I'm having an issue when using the Ollama cloud model, especially when generating large outputs. #71843

Open
opened 2026-05-05 02:41:20 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @Hevoon on GitHub (Apr 3, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/15290

What is the issue?

I’d like to subscribe to Ollama Pro, but when I use the free Ollama cloud models, I encounter an error when the function call outputs a large amount of content:

{"error":"Post v1/chat/completions?ts=1775218167": unexpected EOF"}

When using streaming, I get an error where chunk.finishReason = other.

If I directly use /llama.com/v1 with the ckimi-k2.5:cloud model (or any cloud model), normal conversations work fine. However, when using function calls to generate a slightly larger HTML page (such as a homepage), the process gets interrupted and throws an error:

SocketError: other side closed, code UND_ERR_SOCKET.

Will upgrading to Pro resolve this issue? Or is my usage approach incorrect? Could someone experienced help clarify this?

Relevant log output


OS

No response

GPU

No response

CPU

No response

Ollama version

No response

Originally created by @Hevoon on GitHub (Apr 3, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/15290 ### What is the issue? I’d like to subscribe to Ollama Pro, but when I use the free Ollama cloud models, I encounter an error when the function call outputs a large amount of content: {"error":"Post v1/chat/completions?ts=1775218167": unexpected EOF"} When using streaming, I get an error where chunk.finishReason = other. If I directly use /llama.com/v1 with the ckimi-k2.5:cloud model (or any cloud model), normal conversations work fine. However, when using function calls to generate a slightly larger HTML page (such as a homepage), the process gets interrupted and throws an error: SocketError: other side closed, code UND_ERR_SOCKET. Will upgrading to Pro resolve this issue? Or is my usage approach incorrect? Could someone experienced help clarify this? ### Relevant log output ```shell ``` ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
GiteaMirror added the bug label 2026-05-05 02:41:20 -05:00
Author
Owner

@imaboku commented on GitHub (Apr 8, 2026):

I recently started using Pro to blend cloud models with my local models and it has been working great with kimi-k2.5:cloud except the last couple of days I have started to see this same issue. I tried glm-5.1:cloud this evening and had the same issues. Im running claude code with some fairly complex coding tasks. Earlier this week it was working fine and just started give more of these errors today:
level=WARN source=cloud_proxy.go:257 msg="cloud proxy response copy failed" path=/v1/messages upstream_path=/v1/messages status=200 request_context_canceled=false request_context_err=<nil> error="unexpected EOF"

<!-- gh-comment-id:4203911722 --> @imaboku commented on GitHub (Apr 8, 2026): I recently started using Pro to blend cloud models with my local models and it has been working great with kimi-k2.5:cloud except the last couple of days I have started to see this same issue. I tried glm-5.1:cloud this evening and had the same issues. Im running claude code with some fairly complex coding tasks. Earlier this week it was working fine and just started give more of these errors today: `level=WARN source=cloud_proxy.go:257 msg="cloud proxy response copy failed" path=/v1/messages upstream_path=/v1/messages status=200 request_context_canceled=false request_context_err=<nil> error="unexpected EOF"`
Author
Owner

@Hevoon commented on GitHub (Apr 8, 2026):

did you activate Ollama Pro? I'm wondering if it's because I'm using the free version.

<!-- gh-comment-id:4205765972 --> @Hevoon commented on GitHub (Apr 8, 2026): did you activate Ollama Pro? I'm wondering if it's because I'm using the free version.
Author
Owner

@imaboku commented on GitHub (Apr 9, 2026):

Yes, using Pro access, same issue.

<!-- gh-comment-id:4210600254 --> @imaboku commented on GitHub (Apr 9, 2026): Yes, using Pro access, same issue.
Author
Owner

@PureBlissAK commented on GitHub (Apr 18, 2026):

🤖 Automated Triage & Analysis Report

Issue: #15290
Analyzed: 2026-04-18T18:22:41.417788

Analysis

  • Type: unknown
  • Severity: medium
  • Components: unknown

Implementation Plan

  • Effort: medium
  • Steps:

This issue has been triaged and marked for implementation.

<!-- gh-comment-id:4274310446 --> @PureBlissAK commented on GitHub (Apr 18, 2026): <!-- ollama-issue-orchestrator:v1 issue:15290 --> ## 🤖 Automated Triage & Analysis Report **Issue**: #15290 **Analyzed**: 2026-04-18T18:22:41.417788 ### Analysis - **Type**: unknown - **Severity**: medium - **Components**: unknown ### Implementation Plan - **Effort**: medium - **Steps**: *This issue has been triaged and marked for implementation.*
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#71843