[GH-ISSUE #5432] level=ERROR source=sched.go:388 msg="error loading llama server" error="llama runner process no longer running: -1 " #49912

Closed
opened 2026-04-28 13:23:30 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @popav4 on GitHub (Jul 2, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/5432

What is the issue?

Macbook Air M1
Run with Docker:
docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama
docker exec -it ollama ollama run codestral:22b
Error:

level=ERROR source=sched.go:388 msg="error loading llama server" error="llama runner process no longer running: -1 "
docker.log

OS

Docker

GPU

Other

CPU

Apple

Ollama version

0.1.48

Originally created by @popav4 on GitHub (Jul 2, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/5432 ### What is the issue? Macbook Air M1 Run with Docker: `docker run -d -v ollama:/root/.ollama -p 11434:11434 --name ollama ollama/ollama` `docker exec -it ollama ollama run codestral:22b` Error: > level=ERROR source=sched.go:388 msg="error loading llama server" error="llama runner process no longer running: -1 " [docker.log](https://github.com/user-attachments/files/16067480/docker.log) ### OS Docker ### GPU Other ### CPU Apple ### Ollama version 0.1.48
GiteaMirror added the bug label 2026-04-28 13:23:30 -05:00
Author
Owner

@rick-github commented on GitHub (Jul 2, 2024):

Logs say memory.available="[7.3 GiB]" and memory.required.full="13.1 GiB" so I would guess that your hardware can't load the model.

<!-- gh-comment-id:2203925995 --> @rick-github commented on GitHub (Jul 2, 2024): Logs say `memory.available="[7.3 GiB]"` and `memory.required.full="13.1 GiB"` so I would guess that your hardware can't load the model.
Author
Owner

@dhiltgen commented on GitHub (Jul 2, 2024):

This is a dup of #4955 where we don't currently detect that the first model we try to load wont fit in available system memory when running in CPU mode. Once that's fixed, we'll give a better error message instead of crashing when we fail to allocate enough memory. @popav4 you can increase the memory size in Docker for the container VM and then you'll be able to load this model.

<!-- gh-comment-id:2204322263 --> @dhiltgen commented on GitHub (Jul 2, 2024): This is a dup of #4955 where we don't currently detect that the first model we try to load wont fit in available system memory when running in CPU mode. Once that's fixed, we'll give a better error message instead of crashing when we fail to allocate enough memory. @popav4 you can increase the memory size in Docker for the container VM and then you'll be able to load this model.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#49912