[GH-ISSUE #4478] Ram not releasing #64837

Closed
opened 2026-05-03 18:55:38 -05:00 by GiteaMirror · 3 comments
Owner

Originally created by @Stampsm on GitHub (May 16, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/4478

Originally assigned to: @dhiltgen on GitHub.

What is the issue?

I seem to have an issue with ram not releasing after model unloads. I waited for the new version with the PS command and verified that the model was unloaded after the default 5 minute timeout and a large duffer time after that just to be safe, but my ram usage was still showing ~60GB on this fresh install of Ubuntu the exact same as it was before the model unloaded. When I try to again load a large model after a while when it's loading my system will crash from what seems like an out of memory issues as this proxmox VM image only has ~120GB allocated to it. Nothing else is running on this image as it was a fresh install just for Ollama and openwebui interface. Is there any suggestions or info needed?

OS

Linux

GPU

Other

CPU

Intel

Ollama version

0.1.38

Originally created by @Stampsm on GitHub (May 16, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/4478 Originally assigned to: @dhiltgen on GitHub. ### What is the issue? I seem to have an issue with ram not releasing after model unloads. I waited for the new version with the PS command and verified that the model was unloaded after the default 5 minute timeout and a large duffer time after that just to be safe, but my ram usage was still showing ~60GB on this fresh install of Ubuntu the exact same as it was before the model unloaded. When I try to again load a large model after a while when it's loading my system will crash from what seems like an out of memory issues as this proxmox VM image only has ~120GB allocated to it. Nothing else is running on this image as it was a fresh install just for Ollama and openwebui interface. Is there any suggestions or info needed? ### OS Linux ### GPU Other ### CPU Intel ### Ollama version 0.1.38
GiteaMirror added the bug label 2026-05-03 18:55:38 -05:00
Author
Owner

@Stampsm commented on GitHub (May 17, 2024):

I just had this again. When I load a larger model and it shows unloaded a few hours later then try to load a different large model my whole system soon crashes. Is there a log someplace I can review?

<!-- gh-comment-id:2118040422 --> @Stampsm commented on GitHub (May 17, 2024): I just had this again. When I load a larger model and it shows unloaded a few hours later then try to load a different large model my whole system soon crashes. Is there a log someplace I can review?
Author
Owner

@dhiltgen commented on GitHub (May 18, 2024):

Can you share a little more data? How are you measuring RAM usage before/during/after loading models? Could you run something like ps -eo pid,ppid,rss,args | grep ollama demonstrating what you're seeing?

<!-- gh-comment-id:2118870490 --> @dhiltgen commented on GitHub (May 18, 2024): Can you share a little more data? How are you measuring RAM usage before/during/after loading models? Could you run something like `ps -eo pid,ppid,rss,args | grep ollama` demonstrating what you're seeing?
Author
Owner

@dhiltgen commented on GitHub (May 31, 2024):

If you're still having trouble, please share more info about your measurements and I'll re-open the issue. You can find our troubleshooting docs here https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md

<!-- gh-comment-id:2143022000 --> @dhiltgen commented on GitHub (May 31, 2024): If you're still having trouble, please share more info about your measurements and I'll re-open the issue. You can find our troubleshooting docs here https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#64837