[GH-ISSUE #8437] run phi4 report timed out waiting for llama runner to start #31185

Closed
opened 2026-04-22 11:24:56 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @Jeffhop on GitHub (Jan 15, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/8437

What is the issue?

I used the docker ollama:latest image, entered the container, and got an error when running ollama run phi4 inside the container.

Image

Image

OS

Windows

GPU

Nvidia

CPU

Intel

Ollama version

docker ollama:latest

Originally created by @Jeffhop on GitHub (Jan 15, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/8437 ### What is the issue? I used the docker ollama:latest image, entered the container, and got an error when running ollama run phi4 inside the container. ![Image](https://github.com/user-attachments/assets/829319a5-f704-4bf5-83d6-cdabcbf01df8) ![Image](https://github.com/user-attachments/assets/41201578-ebf0-4b4d-bcc0-35d20e99f751) ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version docker ollama:latest
GiteaMirror added the bug label 2026-04-22 11:24:56 -05:00
Author
Owner

@rick-github commented on GitHub (Jan 15, 2025):

You can increase the load timeout with OLLAMA_LOAD_TIMEOUT. Perhaps access to the GPU is a bit slower inside a container.

<!-- gh-comment-id:2591967651 --> @rick-github commented on GitHub (Jan 15, 2025): You can increase the load timeout with [`OLLAMA_LOAD_TIMEOUT`](https://github.com/ollama/ollama/blob/2539f2dbf99ec1b8f44ece884bf2c8678fca3127/envconfig/config.go#L243). Perhaps access to the GPU is a bit slower inside a container.
Author
Owner

@Jeffhop commented on GitHub (Jan 15, 2025):

You can increase the load timeout with OLLAMA_LOAD_TIMEOUT. Perhaps access to the GPU is a bit slower inside a container.

How can I set it up? export OLLAMA_LOAD_TIMEOUT =xxx,like this?

<!-- gh-comment-id:2591979431 --> @Jeffhop commented on GitHub (Jan 15, 2025): > You can increase the load timeout with [`OLLAMA_LOAD_TIMEOUT`](https://github.com/ollama/ollama/blob/2539f2dbf99ec1b8f44ece884bf2c8678fca3127/envconfig/config.go#L243). Perhaps access to the GPU is a bit slower inside a container. How can I set it up? export OLLAMA_LOAD_TIMEOUT =xxx,like this?
Author
Owner

@rick-github commented on GitHub (Jan 15, 2025):

services:
  ollama:
    environment:
      - OLLAMA_LOAD_TIMEOUT=30m
<!-- gh-comment-id:2591987594 --> @rick-github commented on GitHub (Jan 15, 2025): ```yaml services: ollama: environment: - OLLAMA_LOAD_TIMEOUT=30m ```
Author
Owner

@Jeffhop commented on GitHub (Jan 15, 2025):

services:
ollama:
environment:
- OLLAMA_LOAD_TIMEOUT=30m
thanks,it works.

<!-- gh-comment-id:2592049969 --> @Jeffhop commented on GitHub (Jan 15, 2025): > services: > ollama: > environment: > - OLLAMA_LOAD_TIMEOUT=30m thanks,it works.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#31185