[GH-ISSUE #467] Running a 70B Model with 16GB RAM: Possible Strategies? #46727

Closed
opened 2026-04-27 23:43:57 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @OguzcanOzdemir on GitHub (Sep 3, 2023).
Original GitHub issue: https://github.com/ollama/ollama/issues/467

Hello,

I'm currently working with a system that has 16GB of RAM, and I'm interested in running a 70B model for my project. I understand that according to the GitHub repository's documentation, a 70B model typically requires 32GB of RAM.

However, due to my system limitations, I'm looking for guidance on potential strategies or alternative methods to run a 70B model efficiently with 16GB of RAM.

Are there any techniques, optimizations, or workarounds that I can explore to make this possible? I would greatly appreciate any advice or suggestions on how to approach this challenge and still achieve acceptable performance.

Thank you for your assistance and insights.

Originally created by @OguzcanOzdemir on GitHub (Sep 3, 2023). Original GitHub issue: https://github.com/ollama/ollama/issues/467 Hello, I'm currently working with a system that has 16GB of RAM, and I'm interested in running a 70B model for my project. I understand that according to the GitHub repository's documentation, a 70B model typically requires 32GB of RAM. However, due to my system limitations, I'm looking for guidance on potential strategies or alternative methods to run a 70B model efficiently with 16GB of RAM. Are there any techniques, optimizations, or workarounds that I can explore to make this possible? I would greatly appreciate any advice or suggestions on how to approach this challenge and still achieve acceptable performance. Thank you for your assistance and insights.
GiteaMirror added the question label 2026-04-27 23:43:57 -05:00
Author
Owner

@ddkwork commented on GitHub (Sep 4, 2023):

are you tested 70b?

<!-- gh-comment-id:1705442700 --> @ddkwork commented on GitHub (Sep 4, 2023): are you tested 70b?
Author
Owner

@OguzcanOzdemir commented on GitHub (Sep 4, 2023):

are you tested 70b?

Yes I did. I got this message "Error: Post "http://127.0.0.1:11434/api/generate": EOF"

<!-- gh-comment-id:1705657340 --> @OguzcanOzdemir commented on GitHub (Sep 4, 2023): > are you tested 70b? Yes I did. I got this message "Error: Post "http://127.0.0.1:11434/api/generate": EOF"
Author
Owner

@technovangelist commented on GitHub (Sep 5, 2023):

You got that message because you don't have enough memory. Unfortunately there isn't anything that can be done about that. You really have two strategies available. Use a cloud instance with enough memory, or upgrade to a machine with 64GB of memory. Anything below 64 and you are likely to see that kind of message. Why do you think you need a 70B model? For most cases, smaller models will suffice. Perhaps try some smaller models that will fit on your hardware. I will close this issue, since there is nothing that can be done on our side.

<!-- gh-comment-id:1706908419 --> @technovangelist commented on GitHub (Sep 5, 2023): You got that message because you don't have enough memory. Unfortunately there isn't anything that can be done about that. You really have two strategies available. Use a cloud instance with enough memory, or upgrade to a machine with 64GB of memory. Anything below 64 and you are likely to see that kind of message. Why do you think you need a 70B model? For most cases, smaller models will suffice. Perhaps try some smaller models that will fit on your hardware. I will close this issue, since there is nothing that can be done on our side.
Author
Owner

@OguzcanOzdemir commented on GitHub (Sep 5, 2023):

You got that message because you don't have enough memory. Unfortunately there isn't anything that can be done about that. You really have two strategies available. Use a cloud instance with enough memory, or upgrade to a machine with 64GB of memory. Anything below 64 and you are likely to see that kind of message. Why do you think you need a 70B model? For most cases, smaller models will suffice. Perhaps try some smaller models that will fit on your hardware. I will close this issue, since there is nothing that can be done on our side.

For some difficult questions, I thought the 70b model would provide more creative answers. Thanks for the answer. I will continue to use the 13b model.

<!-- gh-comment-id:1707218831 --> @OguzcanOzdemir commented on GitHub (Sep 5, 2023): > You got that message because you don't have enough memory. Unfortunately there isn't anything that can be done about that. You really have two strategies available. Use a cloud instance with enough memory, or upgrade to a machine with 64GB of memory. Anything below 64 and you are likely to see that kind of message. Why do you think you need a 70B model? For most cases, smaller models will suffice. Perhaps try some smaller models that will fit on your hardware. I will close this issue, since there is nothing that can be done on our side. For some difficult questions, I thought the 70b model would provide more creative answers. Thanks for the answer. I will continue to use the 13b model.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#46727