[GH-ISSUE #7013] Option to Override a Model's Memory Requirements #66499

Closed
opened 2026-05-04 06:51:57 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @dabockster on GitHub (Sep 28, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/7013

Originally assigned to: @dhiltgen on GitHub.

I was trying to load the 70b Llama3 model and Ollama says I need 33.6 GB of 30.5 GB ram. I believe this is a safety thing Meta put into the model, so I want to have the ability to override this and attempt to run it on lower amounts of memory. I know this will likely dip into swap/page file space, possibly even causing kernel panics and BSODs. But I still want to make the attempt.

Computer, disable holodeck safety protocols. Authorization Picard-4-7-Alpha-Tango.

Originally created by @dabockster on GitHub (Sep 28, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/7013 Originally assigned to: @dhiltgen on GitHub. I was trying to load the 70b Llama3 model and Ollama says I need 33.6 GB of 30.5 GB ram. I believe this is a safety thing Meta put into the model, so I want to have the ability to override this and attempt to run it on lower amounts of memory. I know this will likely dip into swap/page file space, possibly even causing kernel panics and BSODs. But I still want to make the attempt. _Computer, disable holodeck safety protocols. Authorization Picard-4-7-Alpha-Tango._
GiteaMirror added the questionwindows labels 2026-05-04 06:51:58 -05:00
Author
Owner

@dhiltgen commented on GitHub (Sep 28, 2024):

We take swap space into consideration. If you increase the size of your swap file it should try to load it, but expect extremely slow performance if you're paging in/out of swap to fit the model.

If that doesn't work, please share your server log so we can see more details on why it wouldn't load the model and I'll reopen the issue.

<!-- gh-comment-id:2381015703 --> @dhiltgen commented on GitHub (Sep 28, 2024): We take swap space into consideration. If you increase the size of your swap file it should try to load it, but expect extremely slow performance if you're paging in/out of swap to fit the model. If that doesn't work, please share your server log so we can see more details on why it wouldn't load the model and I'll reopen the issue.
Author
Owner

@ndrsvh commented on GitHub (Jan 9, 2025):

We take swap space into consideration. If you increase the size of your swap file it should try to load it, but expect extremely slow performance if you're paging in/out of swap to fit the model.

If that doesn't work, please share your server log so we can see more details on why it wouldn't load the model and I'll reopen the issue.

Hello! Could you clarify, I have multiple swap files on different drives, which file will be used by ollama? Do I understand correctly that the one that will be determined by the operating system itself?

<!-- gh-comment-id:2580325393 --> @ndrsvh commented on GitHub (Jan 9, 2025): > We take swap space into consideration. If you increase the size of your swap file it should try to load it, but expect extremely slow performance if you're paging in/out of swap to fit the model. > > If that doesn't work, please share your server log so we can see more details on why it wouldn't load the model and I'll reopen the issue. Hello! Could you clarify, I have multiple swap files on different drives, which file will be used by ollama? Do I understand correctly that the one that will be determined by the operating system itself?
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#66499