[GH-ISSUE #4601] Error: llama runner process has terminated: signal: segmentation fault #2889

Closed
opened 2026-04-12 13:14:30 -05:00 by GiteaMirror · 8 comments
Owner

Originally created by @guiniao on GitHub (May 24, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/4601

What is the issue?

ollama run codellama:34b

error occurred:

pulling manifest
pulling f36b668ebcd3... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 19 GB
pulling 2e0493f67d0c... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 59 B
pulling c60122cb2728... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 132 B
pulling d5981b4f8e77... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 382 B
verifying sha256 digest
writing manifest
removing any unused layers
success

Error: llama runner process has terminated: signal: segmentation fault

codellama:70b and codellama:13b successful

OS

Linux

GPU

Nvidia

CPU

No response

Ollama version

0.1.38

Originally created by @guiniao on GitHub (May 24, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/4601 ### What is the issue? ollama run codellama:34b error occurred: pulling manifest pulling f36b668ebcd3... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 19 GB pulling 2e0493f67d0c... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 59 B pulling c60122cb2728... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 132 B pulling d5981b4f8e77... 100% ▕████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████▏ 382 B verifying sha256 digest writing manifest removing any unused layers success Error: llama runner process has terminated: signal: segmentation fault codellama:70b and codellama:13b successful ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version 0.1.38
GiteaMirror added the bug label 2026-04-12 13:14:30 -05:00
Author
Owner

@jmorganca commented on GitHub (May 24, 2024):

Sorry about this! Closing for #codellama3b

<!-- gh-comment-id:2130496804 --> @jmorganca commented on GitHub (May 24, 2024): Sorry about this! Closing for [#codellama3b](https://github.com/ollama/ollama/issues/4333)
Author
Owner

@Mozisword commented on GitHub (May 31, 2024):

C:\Users\OMEN>ollama run ChatLaw-13B-Q8:latest Error: llama runner process has terminated: exit status 0xc0000409,But running other models works fine

<!-- gh-comment-id:2141074897 --> @Mozisword commented on GitHub (May 31, 2024): C:\Users\OMEN>ollama run ChatLaw-13B-Q8:latest Error: llama runner process has terminated: exit status 0xc0000409,But running other models works fine
Author
Owner

@mhgrove commented on GitHub (Jun 7, 2024):

I am seeing this as well, ubuntu 23.10. started happening after i updated ollama from 0.1.33 to 0.1.41. Every codellama:34b model i've tried (34b, 34b-code, 34b-instruct-34b-python, 34b-q8_0) core dumps when i attempt to use it. 13b works, 70b works. none of the 34s work.

<!-- gh-comment-id:2153682032 --> @mhgrove commented on GitHub (Jun 7, 2024): I am seeing this as well, ubuntu 23.10. started happening after i updated ollama from 0.1.33 to 0.1.41. Every codellama:34b model i've tried (34b, 34b-code, 34b-instruct-34b-python, 34b-q8_0) core dumps when i attempt to use it. 13b works, 70b works. none of the 34s work.
Author
Owner

@francesco-carrella commented on GitHub (Jun 7, 2024):

Experiencing the exact same with codellama:34b model

<!-- gh-comment-id:2155272411 --> @francesco-carrella commented on GitHub (Jun 7, 2024): Experiencing the exact same with codellama:34b model
Author
Owner

@DigitalDevelooper commented on GitHub (Jun 9, 2024):

Same with 34b models and 13b python model, other models work fine

<!-- gh-comment-id:2156507792 --> @DigitalDevelooper commented on GitHub (Jun 9, 2024): Same with 34b models and 13b python model, other models work fine
Author
Owner

@kim-gtek commented on GitHub (Jun 28, 2024):

Same issue, any fix?
error="llama runner process has terminated: signal: segmentation fault "
ollama 34B

<!-- gh-comment-id:2196396236 --> @kim-gtek commented on GitHub (Jun 28, 2024): Same issue, any fix? error="llama runner process has terminated: signal: segmentation fault " ollama 34B
Author
Owner

@Enissay commented on GitHub (Jul 1, 2024):

Same here!

╰─ ollama run codellama:34b --verbose
Error: llama runner process has terminated: signal: segmentation fault (core dumped)
<!-- gh-comment-id:2200572405 --> @Enissay commented on GitHub (Jul 1, 2024): Same here! ```bash ╰─ ollama run codellama:34b --verbose Error: llama runner process has terminated: signal: segmentation fault (core dumped) ```
Author
Owner

@melhamamsy commented on GitHub (Aug 27, 2024):

Same:

curl -X POST http://localhost:11434/api/embeddings -H "Content-Type: application/json" -d "{\"model\": \"locusai/multi-qa-minilm-l6-cos-v1\", \"prompt\": \"How old are you?\" }"

{"error":"llama runner process has terminated: signal: segmentation fault (core dumped)"}

Ollama Image tag: https://hub.docker.com/layers/ollama/ollama/latest/images/sha256-e9b65ea2e50399543346b72cb4f000523a95952105a616bb1a073a9c1e6e1690?context=explore

Ollama Image Definition:

#docker-compose.yml
...
  ollama:
      build:
        context: .
        dockerfile: Dockerfile.ollama
      image: lex_fridman_podcast_ollama:latest
      container_name: ollama
      environment:
        - CHAT_MODEL=${CHAT_MODEL}
        - EMBED_MODEL=${EMBED_MODEL}
      volumes:
        - ollama_data:/root/.ollama
      ports:
        - "11434:11434"
      healthcheck:
        test: ["CMD-SHELL", "curl -s http://localhost:11434 || exit 1"]
        interval: 30s
        timeout: 10s
        retries: 5
        start_period: 30s
...
# Dockerfile.ollama
# Use the official Ollama image as the base
FROM ollama/ollama

# Install curl using apt-get instead of apt
RUN apt-get update && apt-get install -y curl && curl --version

free -h

total used free shared buff/cache available
15Gi 5.6Gi 1.6Gi 170Mi 8.4Gi 9.4Gi (Mem)
5.7Gi 0.0Ki 5.7Gi (Swap)

docker stats ollama

CONTAINER ID NAME CPU % MEM USAGE / LIMIT MEM % NET I/O BLOCK I/O PIDS
59ffa256fdd6 ollama 4.18% 48.98MiB / 15.53GiB 0.31% 93.6MB / 1.33MB 0B / 1.33GB 18

cat /etc/os-release

PRETTY_NAME="Ubuntu 22.04.3 LTS"
NAME="Ubuntu"
VERSION_ID="22.04"
VERSION="22.04.3 LTS (Jammy Jellyfish)"
VERSION_CODENAME=jammy
ID=ubuntu
ID_LIKE=debian
HOME_URL="https://www.ubuntu.com/"
SUPPORT_URL="https://help.ubuntu.com/"
BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/"
PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy"
UBUNTU_CODENAME=jammy

<!-- gh-comment-id:2312659729 --> @melhamamsy commented on GitHub (Aug 27, 2024): Same: ``` curl -X POST http://localhost:11434/api/embeddings -H "Content-Type: application/json" -d "{\"model\": \"locusai/multi-qa-minilm-l6-cos-v1\", \"prompt\": \"How old are you?\" }" ``` > {"error":"llama runner process has terminated: signal: segmentation fault (core dumped)"} **Ollama Image tag**: https://hub.docker.com/layers/ollama/ollama/latest/images/sha256-e9b65ea2e50399543346b72cb4f000523a95952105a616bb1a073a9c1e6e1690?context=explore **Ollama Image Definition:** ``` #docker-compose.yml ... ollama: build: context: . dockerfile: Dockerfile.ollama image: lex_fridman_podcast_ollama:latest container_name: ollama environment: - CHAT_MODEL=${CHAT_MODEL} - EMBED_MODEL=${EMBED_MODEL} volumes: - ollama_data:/root/.ollama ports: - "11434:11434" healthcheck: test: ["CMD-SHELL", "curl -s http://localhost:11434 || exit 1"] interval: 30s timeout: 10s retries: 5 start_period: 30s ... ``` ``` # Dockerfile.ollama # Use the official Ollama image as the base FROM ollama/ollama # Install curl using apt-get instead of apt RUN apt-get update && apt-get install -y curl && curl --version ``` ```free -h``` > total used free shared buff/cache available > 15Gi 5.6Gi 1.6Gi 170Mi 8.4Gi 9.4Gi (Mem) > 5.7Gi 0.0Ki 5.7Gi (Swap) `docker stats ollama` > CONTAINER ID NAME CPU % MEM USAGE / LIMIT MEM % NET I/O BLOCK I/O PIDS > 59ffa256fdd6 ollama 4.18% 48.98MiB / 15.53GiB 0.31% 93.6MB / 1.33MB 0B / 1.33GB 18 `cat /etc/os-release ` > PRETTY_NAME="Ubuntu 22.04.3 LTS" > NAME="Ubuntu" > VERSION_ID="22.04" > VERSION="22.04.3 LTS (Jammy Jellyfish)" > VERSION_CODENAME=jammy > ID=ubuntu > ID_LIKE=debian > HOME_URL="https://www.ubuntu.com/" > SUPPORT_URL="https://help.ubuntu.com/" > BUG_REPORT_URL="https://bugs.launchpad.net/ubuntu/" > PRIVACY_POLICY_URL="https://www.ubuntu.com/legal/terms-and-policies/privacy-policy" > UBUNTU_CODENAME=jammy
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#2889