[GH-ISSUE #15006] Go back and solve the problem #9643

Open
opened 2026-04-12 22:32:12 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @djfjjcjcjcjfjfjlsnfjcj-creator on GitHub (Mar 22, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/15006

What is the issue?

Hello,

I am facing an issue while using Ollama.

When I send a long code (around 4000–10000 lines), the response gets cut off and does not complete. Sometimes it also stops suddenly or shows errors like incomplete output.

This makes it difficult to work with large code files, as the model does not finish generating the full response.

Here are my details:

  • Device: Android phone
  • د
  • Model: deepseek-v3.1:671b-cloud
  • Issue: Output gets truncated / stops منتصف الرد

I have tried sending smaller parts, but I need to process large code in one request.

Can you please help me fix this issue or suggest a solution?

Thank you.

Before the update, it was working perfectly and there were no problems; after the update, problems started to appear..

Relevant log output


OS

Windows

GPU

No response

CPU

No response

Ollama version

No response

Originally created by @djfjjcjcjcjfjfjlsnfjcj-creator on GitHub (Mar 22, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/15006 ### What is the issue? Hello, I am facing an issue while using Ollama. When I send a long code (around 4000–10000 lines), the response gets cut off and does not complete. Sometimes it also stops suddenly or shows errors like incomplete output. This makes it difficult to work with large code files, as the model does not finish generating the full response. Here are my details: - Device: Android phone - د - Model: deepseek-v3.1:671b-cloud - Issue: Output gets truncated / stops منتصف الرد I have tried sending smaller parts, but I need to process large code in one request. Can you please help me fix this issue or suggest a solution? Thank you. Before the update, it was working perfectly and there were no problems; after the update, problems started to appear.. ### Relevant log output ```shell ``` ### OS Windows ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
GiteaMirror added the cloudbug labels 2026-04-12 22:32:12 -05:00
Author
Owner

@mirraman commented on GitHub (Mar 23, 2026):

This is most likely due to token or output limits rather than a bug. When you send very large inputs like 4k–10k lines, the model may hit its maximum context or output size, which causes the response to be truncated or stop midway. It’s possible the new version has stricter limits or timeouts.

<!-- gh-comment-id:4113858630 --> @mirraman commented on GitHub (Mar 23, 2026): This is most likely due to token or output limits rather than a bug. When you send very large inputs like 4k–10k lines, the model may hit its maximum context or output size, which causes the response to be truncated or stop midway. It’s possible the new version has stricter limits or timeouts.
Author
Owner

@mirraman commented on GitHub (Mar 23, 2026):

Try by breaking the code into smaller chunks or compress the code and process it in parts, or ask the model to analyze specific sections instead of the entire file at once.

<!-- gh-comment-id:4113870841 --> @mirraman commented on GitHub (Mar 23, 2026): Try by breaking the code into smaller chunks or compress the code and process it in parts, or ask the model to analyze specific sections instead of the entire file at once.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#9643