[GH-ISSUE #5418] DeepSeek-Coder-V2 (Lite) spouts GGGs #3391

Closed
opened 2026-04-12 14:01:31 -05:00 by GiteaMirror · 9 comments
Owner

Originally created by @lorenzodimauro97 on GitHub (Jul 1, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/5418

What is the issue?

Eventually when using deepseek-coder-v2:16b-lite-instruct-q8_0 and Open Web UI (but also other means like for example continuedev) the model will stop working and spout GGGGGGGGGGGGGGGGGGGGGGGGGGGGGGG regardless of the input:

image

The only fix is to force stop and restart the model, which is bothersome enough to be worth issuing a bug

OS

Windows

GPU

AMD

CPU

AMD

Ollama version

0.1.48

Originally created by @lorenzodimauro97 on GitHub (Jul 1, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/5418 ### What is the issue? Eventually when using deepseek-coder-v2:16b-lite-instruct-q8_0 and Open Web UI (but also other means like for example continuedev) the model will stop working and spout GGGGGGGGGGGGGGGGGGGGGGGGGGGGGGG regardless of the input: ![image](https://github.com/ollama/ollama/assets/50343905/b90c39ae-9269-4f7f-b3a5-dda65186a208) The only fix is to force stop and restart the model, which is bothersome enough to be worth issuing a bug ### OS Windows ### GPU AMD ### CPU AMD ### Ollama version 0.1.48
GiteaMirror added the bug label 2026-04-12 14:01:31 -05:00
Author
Owner

@igorschlum commented on GitHub (Jul 1, 2024):

Hi @lorenzodimauro97, with how many queries do the GGG appear?

<!-- gh-comment-id:2201233682 --> @igorschlum commented on GitHub (Jul 1, 2024): Hi @lorenzodimauro97, with how many queries do the GGG appear?
Author
Owner

@lorenzodimauro97 commented on GitHub (Jul 1, 2024):

It's not consistent, they can appear as soon as 2 queries or as long as about 30+, furthermore, there seems to be no errors on the console log

<!-- gh-comment-id:2201274364 --> @lorenzodimauro97 commented on GitHub (Jul 1, 2024): It's not consistent, they can appear as soon as 2 queries or as long as about 30+, furthermore, there seems to be no errors on the console log
Author
Owner

@igorschlum commented on GitHub (Jul 1, 2024):

BTW this model is the only one exept ChatGPT that can answer correctly this prompt:
how many days between febuary 2nd 2021 and may 23rd, 2024?
I will try to ask him more questions. I'm on a macbook with 32GB or RAM.

<!-- gh-comment-id:2201294854 --> @igorschlum commented on GitHub (Jul 1, 2024): BTW this model is the only one exept ChatGPT that can answer correctly this prompt: how many days between febuary 2nd 2021 and may 23rd, 2024? I will try to ask him more questions. I'm on a macbook with 32GB or RAM.
Author
Owner

@jmorganca commented on GitHub (Jul 1, 2024):

Hi there, would it be possible to re-pull the model? This should fix things. Let me know if it continues though

<!-- gh-comment-id:2201329772 --> @jmorganca commented on GitHub (Jul 1, 2024): Hi there, would it be possible to re-pull the model? This should fix things. Let me know if it continues though
Author
Owner

@igorschlum commented on GitHub (Jul 1, 2024):

@jmorganca I played with this model since then and had no problem on MacOS.

<!-- gh-comment-id:2201446381 --> @igorschlum commented on GitHub (Jul 1, 2024): @jmorganca I played with this model since then and had no problem on MacOS.
Author
Owner

@lorenzodimauro97 commented on GitHub (Jul 2, 2024):

Hi there, would it be possible to re-pull the model? This should fix things. Let me know if it continues though

I did re-pull the model but problem still occurs:
image

<!-- gh-comment-id:2202664753 --> @lorenzodimauro97 commented on GitHub (Jul 2, 2024): > Hi there, would it be possible to re-pull the model? This should fix things. Let me know if it continues though I did re-pull the model but problem still occurs: ![image](https://github.com/ollama/ollama/assets/50343905/971dd3dd-15af-48e6-a7b7-d1826960fbf4)
Author
Owner

@igorschlum commented on GitHub (Jul 2, 2024):

I played quite a lot with this Model on MacOS without any problem. How much memory do you have? Can you write a simple Python script that can reproduce the issue so other users of Ollama could try to help on this issue?

<!-- gh-comment-id:2203125288 --> @igorschlum commented on GitHub (Jul 2, 2024): I played quite a lot with this Model on MacOS without any problem. How much memory do you have? Can you write a simple Python script that can reproduce the issue so other users of Ollama could try to help on this issue?
Author
Owner

@lorenzodimauro97 commented on GitHub (Jul 2, 2024):

I played quite a lot with this Model on MacOS without any problem. How much memory do you have? Can you write a simple Python script that can reproduce the issue so other users of Ollama could try to help on this issue?

I have 48 GB DDR4 at 3800 MHz, a Ryzen 5800X and an AMD 6800XT. I am using Windows 11.

I do not incur in memory saturation with the model, I stay at about 50-60% RAM usage, going 80-90 with heavy coding, but still, should be enough.

I do not have the means to reproduce it because it does not happen consistently. It can happen as often as after 3 queries or even after half a day's worth of coding, but still I find it hard to reproduce.

What I can say though is that it's the only model that has this issue, none of my other models seem to fail as hard and as frequently as this one does.

<!-- gh-comment-id:2203725611 --> @lorenzodimauro97 commented on GitHub (Jul 2, 2024): > I played quite a lot with this Model on MacOS without any problem. How much memory do you have? Can you write a simple Python script that can reproduce the issue so other users of Ollama could try to help on this issue? I have 48 GB DDR4 at 3800 MHz, a Ryzen 5800X and an AMD 6800XT. I am using Windows 11. I do not incur in memory saturation with the model, I stay at about 50-60% RAM usage, going 80-90 with heavy coding, but still, should be enough. I do not have the means to reproduce it because it does not happen consistently. It can happen as often as after 3 queries or even after half a day's worth of coding, but still I find it hard to reproduce. What I can say though is that it's the only model that has this issue, none of my other models seem to fail as hard and as frequently as this one does.
Author
Owner

@igorschlum commented on GitHub (Jul 2, 2024):

Thank you. Your graphics card has 16 GB of VRAM, but the model requires 17 GB of RAM, so it cannot fully fit in VRAM. Could you try testing a smaller model, such as ollama run deepseek-coder-v2:latest, and see if the issue with GGGGG still occurs? Then, test the deepseek-coder-v2:16b-lite-instruct-q5_0 model, which is 11 GB, to see if the problem can be reproduced.

<!-- gh-comment-id:2203994621 --> @igorschlum commented on GitHub (Jul 2, 2024): Thank you. Your graphics card has 16 GB of VRAM, but the model requires 17 GB of RAM, so it cannot fully fit in VRAM. Could you try testing a smaller model, such as ollama run deepseek-coder-v2:latest, and see if the issue with GGGGG still occurs? Then, test the deepseek-coder-v2:16b-lite-instruct-q5_0 model, which is 11 GB, to see if the problem can be reproduced.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#3391