[GH-ISSUE #5537] deepseek code v2 inference downgrade after a few inference #65494

Open
opened 2026-05-03 21:28:41 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @kidoln on GitHub (Jul 8, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/5537

What is the issue?

I am using ollama for deepseed coder model make fill-in-middel inference, and it was working fine initially. However, after more than 50 consecutive inferences, the model output suddenly became abnormal, producing nonsensical content.

abnormal output like this.

#generategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerate

In these calls, the call parameters remain consistent.

data = {
                "model": "deepseek-coder-v2:latest",              
                "prompt": prompt,
                "raw": True,
                "stream": False,
                "keep_alive": -1,
                "seed": 40,
                "options": {
                    "num_predict": num_p,
                    "top_k": 1,
                    "top_p": 0.0,
                    "temperature": 0.0,
                    "repeat_penalty": 1.1,
                }
}

OS

Linux

GPU

Nvidia

CPU

Intel

Ollama version

0.1.48

Originally created by @kidoln on GitHub (Jul 8, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/5537 ### What is the issue? I am using ollama for deepseed coder model make fill-in-middel inference, and it was working fine initially. However, after more than 50 consecutive inferences, the model output suddenly became abnormal, producing nonsensical content. abnormal output like this. #generategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerategenerate In these calls, the call parameters remain consistent. ``` data = { "model": "deepseek-coder-v2:latest", "prompt": prompt, "raw": True, "stream": False, "keep_alive": -1, "seed": 40, "options": { "num_predict": num_p, "top_k": 1, "top_p": 0.0, "temperature": 0.0, "repeat_penalty": 1.1, } } ``` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.48
GiteaMirror added the bug label 2026-05-03 21:28:41 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#65494