[GH-ISSUE #8470] ollama._types.ResponseError: timed out waiting for llama runner to start - progress 0.00 - #51962

Closed
opened 2026-04-28 21:24:15 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @legendier on GitHub (Jan 17, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/8470

The process of loading large models into GPU memory is very slow.
Then an error will occur:
ollama._types.ResponseError: timed out waiting for llama runner to start - progress 0.00 -

Previously used normally, but recently the large model has been unable to load successfully.
Why is this?

Originally created by @legendier on GitHub (Jan 17, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/8470 The process of loading large models into GPU memory is very slow. Then an error will occur: ”**ollama._types.ResponseError: timed out waiting for llama runner to start - progress 0.00 -** “ Previously used normally, but recently the large model has been unable to load successfully. Why is this?
Author
Owner

@rick-github commented on GitHub (Jan 17, 2025):

Ollama has a timeout for model loads, you can adjust this with OLLAMA_LOAD_TIMEOUT, eg OLLAMA_LOAD_TIMEOUT=30m.

<!-- gh-comment-id:2598335652 --> @rick-github commented on GitHub (Jan 17, 2025): Ollama has a timeout for model loads, you can adjust this with [`OLLAMA_LOAD_TIMEOUT`](https://github.com/ollama/ollama/blob/021817e59ace5e351b35b2e6881f83a09f038546/envconfig/config.go#L243), eg `OLLAMA_LOAD_TIMEOUT=30m`.
Author
Owner

@legendier commented on GitHub (Jan 18, 2025):

Ollama has a timeout for model loads, you can adjust this with OLLAMA_LOAD_TIMEOUT, eg OLLAMA_LOAD_TIMEOUT=30m.

Thank you for your reply.
I successfully loaded the large model using the command "OLLAMA_Load_TIMEOUT=30m bin/ollama serve"

<!-- gh-comment-id:2599535659 --> @legendier commented on GitHub (Jan 18, 2025): > Ollama has a timeout for model loads, you can adjust this with [`OLLAMA_LOAD_TIMEOUT`](https://github.com/ollama/ollama/blob/021817e59ace5e351b35b2e6881f83a09f038546/envconfig/config.go#L243), eg `OLLAMA_LOAD_TIMEOUT=30m`. Thank you for your reply. I successfully loaded the large model using the command "**OLLAMA_Load_TIMEOUT=30m bin/ollama serve**"
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#51962