[GH-ISSUE #13765] Randomly uses CPU even when GPU available until I restart #71081

Closed
opened 2026-05-04 23:57:16 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @CooperCorona on GitHub (Jan 17, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/13765

What is the issue?

Occasionally, ollama will refuse to use my GPU. nvidia-smi shows I have plenty of available VRAM. But ollama run will load the model entirely on the CPU, with ollama ps showing 100% CPU usage. This happens seemingly randomly. I'll go days using ollama just fine, then the next time I run it, all CPU.

I'm running 0.13.5 via OpenWebUI's Ollama Docker image, which is at version v0.7.2, on Ubuntu 22.04.5 LTS. I notice this via OpenWebUI, which fails with 500: model requires more system memory (18.9 GiB) than is available (13.1 GiB) (because it's using CPU, not GPU). I have OpenWebUI to refuse to use the CPU, because before, it would silently use the CPU and just generate at 10% speed. This has occurred throughout OpenWebUI and Ollama versions.

Let me know if you need more information or would like me to do any debugging. I solve this just by restarting the Docker container, but it's quite annoying.

Relevant log output


OS

Docker

GPU

Nvidia

CPU

AMD

Ollama version

0.13.5

Originally created by @CooperCorona on GitHub (Jan 17, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/13765 ### What is the issue? Occasionally, ollama will refuse to use my GPU. `nvidia-smi` shows I have plenty of available VRAM. But `ollama run` will load the model entirely on the CPU, with `ollama ps` showing 100% CPU usage. This happens seemingly randomly. I'll go days using ollama just fine, then the next time I run it, all CPU. I'm running 0.13.5 via OpenWebUI's Ollama Docker image, which is at version v0.7.2, on Ubuntu 22.04.5 LTS. I notice this via OpenWebUI, which fails with `500: model requires more system memory (18.9 GiB) than is available (13.1 GiB)` (because it's using CPU, not GPU). I have OpenWebUI to refuse to use the CPU, because before, it would silently use the CPU and just generate at 10% speed. This has occurred throughout OpenWebUI and Ollama versions. Let me know if you need more information or would like me to do any debugging. I solve this just by restarting the Docker container, but it's quite annoying. ### Relevant log output ```shell ``` ### OS Docker ### GPU Nvidia ### CPU AMD ### Ollama version 0.13.5
GiteaMirror added the bugnvidia labels 2026-05-04 23:57:17 -05:00
Author
Owner

@mchiang0610 commented on GitHub (Jan 18, 2026):

Thank you for reporting this. May I ask if this issue exists on the latest Ollama version? v0.14.2

I'd love to know which GPU you have too.

<!-- gh-comment-id:3764942138 --> @mchiang0610 commented on GitHub (Jan 18, 2026): Thank you for reporting this. May I ask if this issue exists on the latest Ollama version? v0.14.2 I'd love to know which GPU you have too.
Author
Owner
<!-- gh-comment-id:3765078195 --> @rick-github commented on GitHub (Jan 18, 2026): https://github.com/ollama/ollama/blob/main/docs/troubleshooting.mdx#linux-docker
Author
Owner

@thomas-meier85 commented on GitHub (Jan 18, 2026):

Try adding

{
    "runtimes": {
        "nvidia": {
            "args": [],
            "path": "nvidia-container-runtime"
        }
    },
        "exec-opts": ["native.cgroupdriver=cgroupfs"]
}

to etc/docker/daemon.json

Restart your machine and check if the error still happens.

<!-- gh-comment-id:3765151503 --> @thomas-meier85 commented on GitHub (Jan 18, 2026): Try adding ``` { "runtimes": { "nvidia": { "args": [], "path": "nvidia-container-runtime" } }, "exec-opts": ["native.cgroupdriver=cgroupfs"] } ``` to etc/docker/daemon.json Restart your machine and check if the error still happens.
Author
Owner

@CooperCorona commented on GitHub (Mar 15, 2026):

sorry, completely missed all the responses here. This just happened again on Ollama v0.17.2. My GPU is an EVGA RTX 3090 and Nvidia FE 3090.

I'll try thomas' suggestion and see if it happens again. Since this bug is stale, I'll only reopen it if it occurs yet again.

<!-- gh-comment-id:4063564479 --> @CooperCorona commented on GitHub (Mar 15, 2026): sorry, completely missed all the responses here. This just happened again on Ollama v0.17.2. My GPU is an EVGA RTX 3090 and Nvidia FE 3090. I'll try thomas' suggestion and see if it happens again. Since this bug is stale, I'll only reopen it if it occurs yet again.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#71081