[GH-ISSUE #6806] slow #30056

Closed
opened 2026-04-22 09:29:11 -05:00 by GiteaMirror · 9 comments
Owner

Originally created by @ayttop on GitHub (Sep 14, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/6806

What is the issue?

The program is very slow
3.10 ollama

OS

Windows

GPU

Intel

CPU

Intel

Ollama version

3.10

Originally created by @ayttop on GitHub (Sep 14, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/6806 ### What is the issue? The program is very slow 3.10 ollama ### OS Windows ### GPU Intel ### CPU Intel ### Ollama version 3.10
GiteaMirror added the needs more infobug labels 2026-04-22 09:29:11 -05:00
Author
Owner

@remco-pc commented on GitHub (Sep 14, 2024):

your hardware ?

<!-- gh-comment-id:2351156052 --> @remco-pc commented on GitHub (Sep 14, 2024): your hardware ?
Author
Owner

@cgimenez commented on GitHub (Sep 15, 2024):

And what model are you using ? I’ve pushed my Mac M2 (64 GB RAM) to its limits just trying to use LLaMA 3.1:70B." - ollama went hard to sigkill

[the doc is clear about the RAM needed depending on the model you wanna use]

<!-- gh-comment-id:2351583031 --> @cgimenez commented on GitHub (Sep 15, 2024): And what model are you using ? I’ve pushed my Mac M2 (64 GB RAM) to its limits just trying to use LLaMA 3.1:70B." - ollama went hard to sigkill [the doc is clear about the RAM needed depending on the model you wanna use]
Author
Owner

@pdevine commented on GitHub (Sep 15, 2024):

You can use ollama ps to figure out if the model is loaded into CPU or GPU.

<!-- gh-comment-id:2351633493 --> @pdevine commented on GitHub (Sep 15, 2024): You can use `ollama ps` to figure out if the model is loaded into CPU or GPU.
Author
Owner

@tcreek commented on GitHub (Sep 15, 2024):

That fact they list intel GPU is very telling on why it is very slow. It is probably using the CPU.

<!-- gh-comment-id:2351658294 --> @tcreek commented on GitHub (Sep 15, 2024): That fact they list intel GPU is very telling on why it is very slow. It is probably using the CPU.
Author
Owner

@ayttop commented on GitHub (Sep 15, 2024):

lama3.1 8q
no viga
Unusually slow, I deleted the first one and reinstalled it, and the same problem.

<!-- gh-comment-id:2351683456 --> @ayttop commented on GitHub (Sep 15, 2024): lama3.1 8q no viga Unusually slow, I deleted the first one and reinstalled it, and the same problem.
Author
Owner

@ayttop commented on GitHub (Sep 15, 2024):

i7
64ram

<!-- gh-comment-id:2351684829 --> @ayttop commented on GitHub (Sep 15, 2024): i7 64ram
Author
Owner

@tcreek commented on GitHub (Sep 15, 2024):

unless you have a modern nVidia or AMD GPU, it will continue to be slow

<!-- gh-comment-id:2351686133 --> @tcreek commented on GitHub (Sep 15, 2024): unless you have a modern nVidia or AMD GPU, it will continue to be slow
Author
Owner

@remco-pc commented on GitHub (Sep 15, 2024):

and what do you think is slow, can you think faster ?, my llama3.1:70B model is slow, like 12 minutes for a response, but i find it rather fast enough and workable, i can give it tasks in the future to solve

<!-- gh-comment-id:2351783173 --> @remco-pc commented on GitHub (Sep 15, 2024): and what do you think is slow, can you think faster ?, my llama3.1:70B model is slow, like 12 minutes for a response, but i find it rather fast enough and workable, i can give it tasks in the future to solve
Author
Owner

@pdevine commented on GitHub (Sep 15, 2024):

Hi @ayttop , sorry this is going slow for you. It's because you're using CPU inference (which you can tell with ollama ps). I'm not certain what your GPU is, but the Intel GPUs unfortunately aren't supported (yet). There are a number of issues already around supporting that, so I'm going to go ahead and close this issue (like #1590) so I'll go ahead and close this issue.

<!-- gh-comment-id:2351791354 --> @pdevine commented on GitHub (Sep 15, 2024): Hi @ayttop , sorry this is going slow for you. It's because you're using CPU inference (which you can tell with `ollama ps`). I'm not certain what your GPU is, but the Intel GPUs unfortunately aren't supported (yet). There are a number of issues already around supporting that, so I'm going to go ahead and close this issue (like #1590) so I'll go ahead and close this issue.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#30056