[GH-ISSUE #14719] deepseek-r1:70b model now takes 69 GB of RAM to run a 43 GB model? #35276

Closed
opened 2026-04-22 19:40:13 -05:00 by GiteaMirror · 3 comments
Owner

Originally created by @alpha754293 on GitHub (Mar 8, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/14719

What is the issue?

I used to be able to run the deepseek-r1:70b model with my pair of 3090s which just fit within the total 48 GB of VRAM.

Now, when I tried to launch it with ollama run deepseek-r1:70b, it gives me this error instead:

Error: 500 Internal Server Error: model requires more system memory (69.0 GiB) than is available (65.6 GiB)

What changed? Why is it asking for more RAM than is needed to run the model where/when it used to work before?

Relevant log output

Error: 500 Internal Server Error: model requires more system memory (69.0 GiB) than is available (65.6 GiB)

OS

Linux

GPU

Nvidia

CPU

Intel

Ollama version

0.17.7

Originally created by @alpha754293 on GitHub (Mar 8, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/14719 ### What is the issue? I used to be able to run the `deepseek-r1:70b` model with my pair of 3090s which _just_ fit within the total 48 GB of VRAM. Now, when I tried to launch it with `ollama run deepseek-r1:70b`, it gives me this error instead: `Error: 500 Internal Server Error: model requires more system memory (69.0 GiB) than is available (65.6 GiB)` What changed? Why is it asking for more RAM than is needed to run the model where/when it used to work before? ### Relevant log output ```shell Error: 500 Internal Server Error: model requires more system memory (69.0 GiB) than is available (65.6 GiB) ``` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.17.7
GiteaMirror added the bug label 2026-04-22 19:40:13 -05:00
Author
Owner

@rick-github commented on GitHub (Mar 8, 2026):

Perhaps #14116. Server logs will aid in debugging.

<!-- gh-comment-id:4019679788 --> @rick-github commented on GitHub (Mar 8, 2026): Perhaps #14116. [Server logs](https://docs.ollama.com/troubleshooting) will aid in debugging.
Author
Owner

@alpha754293 commented on GitHub (Mar 8, 2026):

Perhaps #14116. Server logs will aid in debugging.

My apologies. Here you go.

log.txt

And here is the docker compose YAML file that I use (inside dockge) to start it, but it seems to be mostly different to how I am starting ollama.

compose.yaml

<!-- gh-comment-id:4019713908 --> @alpha754293 commented on GitHub (Mar 8, 2026): > Perhaps [#14116](https://github.com/ollama/ollama/issues/14116). [Server logs](https://docs.ollama.com/troubleshooting) will aid in debugging. My apologies. Here you go. [log.txt](https://github.com/user-attachments/files/25826699/log.txt) And here is the docker compose YAML file that I use (inside dockge) to start it, but it seems to be mostly different to _how_ I am starting ollama. [compose.yaml](https://github.com/user-attachments/files/25826716/compose.yaml)
Author
Owner

@alpha754293 commented on GitHub (Mar 8, 2026):

I think that I might have figured it out. It would appear that as of v0.15.5, there was a change to the context window defaults that for systems with a total of 48 GB of VRAM (I presume), that it set the default context window size to 32768.

Thanks.

<!-- gh-comment-id:4019751352 --> @alpha754293 commented on GitHub (Mar 8, 2026): I think that I might have figured it out. It would appear that as of v0.15.5, there was a change to the context window defaults that for systems with a total of 48 GB of VRAM (I presume), that it set the default context window size to 32768. Thanks.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#35276