[GH-ISSUE #13724] image runner failed #34758

Closed
opened 2026-04-22 18:35:05 -05:00 by GiteaMirror · 10 comments
Owner

Originally created by @wOvAN on GitHub (Jan 15, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/13724

What is the issue?

running x/z-image-turbo:bf16

Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127)

Relevant log output

llama  | time=2026-01-15T07:22:48.331Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax"
ollama  | time=2026-01-15T07:22:48.331Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/bin/ollama runner --ollama-engine --port 37559"
ollama  | time=2026-01-15T07:22:49.817Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/z-image-turbo:bf16 port=39333
ollama  | time=2026-01-15T07:22:50.021Z level=WARN source=server.go:141 msg=image-runner msg="/usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory"
ollama  | time=2026-01-15T07:22:50.021Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=201

OS

Docker

GPU

Nvidia

CPU

Intel, AMD

Ollama version

0.14.1

Originally created by @wOvAN on GitHub (Jan 15, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/13724 ### What is the issue? running x/z-image-turbo:bf16 Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127) ### Relevant log output ```shell llama | time=2026-01-15T07:22:48.331Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax" ollama | time=2026-01-15T07:22:48.331Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/bin/ollama runner --ollama-engine --port 37559" ollama | time=2026-01-15T07:22:49.817Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/z-image-turbo:bf16 port=39333 ollama | time=2026-01-15T07:22:50.021Z level=WARN source=server.go:141 msg=image-runner msg="/usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory" ollama | time=2026-01-15T07:22:50.021Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=201 ``` ### OS Docker ### GPU Nvidia ### CPU Intel, AMD ### Ollama version 0.14.1
GiteaMirror added the bug label 2026-04-22 18:35:05 -05:00
Author
Owner

@mchiang0610 commented on GitHub (Jan 16, 2026):

Thanks for reporting this. May I ask more about your system?

  • Operating system docker is running
  • system specs, including the GPU
  • have you tried this on different computers as you listed Intel and AMD for CPUs?
<!-- gh-comment-id:3757899926 --> @mchiang0610 commented on GitHub (Jan 16, 2026): Thanks for reporting this. May I ask more about your system? - Operating system docker is running - system specs, including the GPU - have you tried this on different computers as you listed Intel and AMD for CPUs?
Author
Owner

@YuenSzeHong commented on GitHub (Jan 16, 2026):

Ubunut 24.04, WSL 2

docker exec -it ollama bash
root@10724800ea30:/# ollama -v
ollama version is 0.14.2-rc1
Fri Jan 16 14:31:42 2026
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 560.35.02              Driver Version: 560.94         CUDA Version: 12.6     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce GTX 1650 ...    On  |   00000000:02:00.0 Off |                  N/A |
| N/A   49C    P8              2W /   37W |     166MiB /   4096MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI        PID   Type   Process name                              GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|  No running processes found                                                             |
+-----------------------------------------------------------------------------------------+

Docker side

time=2026-01-16T06:24:34.875Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax"

time=2026-01-16T06:24:34.893Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/z-image-turbo port=42857

time=2026-01-16T06:24:34.959Z level=WARN source=server.go:141 msg=image-runner msg="/usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory"

time=2026-01-16T06:24:34.959Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=123

Container Side

# ollama run --experimental x/z-image-turbo
>>> anime girl under cherry blossom

Error: 500 Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127)
<!-- gh-comment-id:3758389754 --> @YuenSzeHong commented on GitHub (Jan 16, 2026): Ubunut 24.04, WSL 2 ``` docker exec -it ollama bash root@10724800ea30:/# ollama -v ollama version is 0.14.2-rc1 ``` ``` Fri Jan 16 14:31:42 2026 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 560.35.02 Driver Version: 560.94 CUDA Version: 12.6 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce GTX 1650 ... On | 00000000:02:00.0 Off | N/A | | N/A 49C P8 2W / 37W | 166MiB / 4096MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+ ``` Docker side ``` time=2026-01-16T06:24:34.875Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax" time=2026-01-16T06:24:34.893Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/z-image-turbo port=42857 time=2026-01-16T06:24:34.959Z level=WARN source=server.go:141 msg=image-runner msg="/usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory" time=2026-01-16T06:24:34.959Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=123 ``` Container Side ```console # ollama run --experimental x/z-image-turbo >>> anime girl under cherry blossom Error: 500 Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127) ```
Author
Owner

@rick-github commented on GitHub (Jan 16, 2026):

The missing library in the docker image can be installed with:

docker exec -it ollama bash -c 'apt update && apt install -y libquadmath0'
<!-- gh-comment-id:3758585433 --> @rick-github commented on GitHub (Jan 16, 2026): The missing library in the docker image can be installed with: ``` docker exec -it ollama bash -c 'apt update && apt install -y libquadmath0' ```
Author
Owner

@alerque commented on GitHub (Jan 16, 2026):

Ubunut 24.04, WSL 2

The release notes specify that only macOS is supported for image generation models. Is this even expected to work on Linux?

<!-- gh-comment-id:3758594817 --> @alerque commented on GitHub (Jan 16, 2026): > Ubunut 24.04, WSL 2 The release notes specify that only macOS is supported for image generation models. Is this even expected to work on Linux?
Author
Owner

@rick-github commented on GitHub (Jan 16, 2026):

Release notes say

Experimental image generation models are available for macOS and Linux (CUDA) in Ollama:
<!-- gh-comment-id:3758621992 --> @rick-github commented on GitHub (Jan 16, 2026): Release notes say ``` Experimental image generation models are available for macOS and Linux (CUDA) in Ollama: ```
Author
Owner

@YuenSzeHong commented on GitHub (Jan 16, 2026):

Ubunut 24.04, WSL 2

The release notes specify that only macOS is supported for image generation models. Is this even expected to work on Linux?

Image

am i blind or you are blind? Linux (CUDA) is explisitly mentioned

<!-- gh-comment-id:3758684104 --> @YuenSzeHong commented on GitHub (Jan 16, 2026): > > Ubunut 24.04, WSL 2 > > The release notes specify that only macOS is supported for image generation models. Is this even expected to work on Linux? <img width="2160" height="1119" alt="Image" src="https://github.com/user-attachments/assets/9e75392f-a874-4e96-92f3-db1f18d75b44" /> am i blind or you are blind? `Linux (CUDA)` is explisitly mentioned
Author
Owner

@wOvAN commented on GitHub (Jan 16, 2026):

0.14.2 - linux + docker + nvidia, all the same

500 Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127)

<!-- gh-comment-id:3760427306 --> @wOvAN commented on GitHub (Jan 16, 2026): 0.14.2 - linux + docker + nvidia, all the same 500 Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127)
Author
Owner

@alerque commented on GitHub (Jan 16, 2026):

am i blind or you are blind? Linux (CUDA) is explisitly mentioned

Neither. I'm pretty sure the release notes got edited.

Also the README for the x/z-image-turbo also still says otherwise:

Image

So should the z-image-turbo be expected to work on Linux or not? Even with the current release notes it isn't clear whether that's something I should bother poking at or just wait until it's at least supposed to work before I give it a go.

<!-- gh-comment-id:3760569101 --> @alerque commented on GitHub (Jan 16, 2026): > am i blind or you are blind? `Linux (CUDA)` is explisitly mentioned Neither. I'm pretty sure the release notes got edited. Also the README for the [x/z-image-turbo](https://ollama.com/x/z-image-turbo) also still says otherwise: <img width="771" height="604" alt="Image" src="https://github.com/user-attachments/assets/9205fa5c-2dd9-4ab1-b0ec-c0a3d7062f62" /> So should the z-image-turbo be expected to work on Linux or not? Even with the current release notes it isn't clear whether that's something I should bother poking at or just wait until it's at least *supposed* to work before I give it a go.
Author
Owner

@YuenSzeHong commented on GitHub (Jan 17, 2026):

am i blind or you are blind? Linux (CUDA) is explisitly mentioned

Neither. I'm pretty sure the release notes got edited.

Also the README for the x/z-image-turbo also still says otherwise:

Image

So should the z-image-turbo be expected to work on Linux or not? Even with the current release notes it isn't clear whether that's something I should bother poking at or just wait until it's at least supposed to work before I give it a go.

so did you refer to the model doc or release note, the release note still have linux on it tho

<!-- gh-comment-id:3762332311 --> @YuenSzeHong commented on GitHub (Jan 17, 2026): > > am i blind or you are blind? `Linux (CUDA)` is explisitly mentioned > > Neither. I'm pretty sure the release notes got edited. > > Also the README for the [x/z-image-turbo](https://ollama.com/x/z-image-turbo) also still says otherwise: > > <img width="771" height="604" alt="Image" src="https://github.com/user-attachments/assets/9205fa5c-2dd9-4ab1-b0ec-c0a3d7062f62" /> > > So should the z-image-turbo be expected to work on Linux or not? Even with the current release notes it isn't clear whether that's something I should bother poking at or just wait until it's at least *supposed* to work before I give it a go. so did you refer to the model doc or release note, the release note still have linux on it tho
Author
Owner

@jnjnnjzch commented on GitHub (Jan 20, 2026):

Same issue

System: Ubuntu 22.04 + ollama docker 0.14.2
GPU: Nvidia RTX 4090 + CUDA 13.0 + Driver 580.95.05
CPU: Intel(R) Xeon(R) w7-3455

The Error reports as:

Error: 500 Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127)

I've tried docker exec -it ollama bash -c 'apt update && apt install -y libquadmath0' as @rick-github said (thanks by the way)

The missing library in the docker image can be installed with:

docker exec -it ollama bash -c 'apt update && apt install -y libquadmath0'

Here is the new error:

Error: 500 Internal Server Error: image runner failed: 2026/01/20 07:59:02 runner.go:65: INFO starting image runner model=x/z-image-turbo port=39315 (exit: exit status 255)

and another model gives the same error report:

Error: 500 Internal Server Error: image runner failed: 2026/01/20 08:00:02 runner.go:65: INFO starting image runner model=x/flux2-klein port=46651 (exit: exit status 255)

Hope to get any help, thanks!

Edit:
here is the logs

time=2026-01-20T08:15:34.428Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax"
time=2026-01-20T08:15:34.435Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/z-image-turbo port=36805
time=2026-01-20T08:15:34.621Z level=WARN source=server.go:141 msg=image-runner msg="2026/01/20 08:15:34 runner.go:65: INFO starting image runner model=x/z-image-turbo port=36805"
time=2026-01-20T08:15:34.626Z level=INFO source=server.go:134 msg=image-runner msg="Loading Z-Image model from manifest: x/z-image-turbo..."
time=2026-01-20T08:15:34.940Z level=INFO source=server.go:134 msg=image-runner msg="  Loading tokenizer... ✓"
time=2026-01-20T08:15:37.009Z level=INFO source=server.go:134 msg=image-runner msg="  Loading text encoder... ✓"
time=2026-01-20T08:15:37.536Z level=INFO source=server.go:134 msg=image-runner msg="  (11.3 GB, peak 11.3 GB)"
time=2026-01-20T08:15:41.305Z level=INFO source=server.go:134 msg=image-runner msg="  Loading transformer... ✓"
time=2026-01-20T08:15:43.206Z level=INFO source=server.go:134 msg=image-runner msg="MLX error: cudaMallocAsync(&data, size, stream) failed: out of memory at /go/src/github.com/ollama/ollama/build/_deps/mlx-c-src/mlx/c/transforms.cpp:73"
time=2026-01-20T08:15:43.430Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=929
[GIN] 2026/01/20 - 08:15:48 | 500 |  14.01500212s |       127.0.0.1 | POST     "/api/generate"

Edit: This model always runs out of memory, even I load ollama docker with 3 GPUs --gpus '"device=3,2,1"' . As my GPU has 24GB graphic memory, and the model size is 12GB, it is quite possible to be a bug.

For another model

time=2026-01-20T08:15:01.151Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax"
time=2026-01-20T08:15:01.155Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/flux2-klein port=41357
time=2026-01-20T08:15:01.341Z level=WARN source=server.go:141 msg=image-runner msg="2026/01/20 08:15:01 runner.go:65: INFO starting image runner model=x/flux2-klein port=41357"
time=2026-01-20T08:15:01.345Z level=INFO source=server.go:134 msg=image-runner msg="Loading Z-Image model from manifest: x/flux2-klein..."
time=2026-01-20T08:15:01.639Z level=INFO source=server.go:134 msg=image-runner msg="  Loading tokenizer... ✓"
time=2026-01-20T08:15:02.874Z level=INFO source=server.go:134 msg=image-runner msg="  Loading text encoder... MLX error: [dequantize] Shape of scales and biases does not match the matrix given the quantization parameters. Provided matrix of shape (4096,320) and scales/biases of shape (4096,80) with group_size=32 and bits=8. at /go/src/github.com/ollama/ollama/build/_deps/mlx-c-src/mlx/c/ops.cpp:1085"
time=2026-01-20T08:15:03.012Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=876
<!-- gh-comment-id:3771565881 --> @jnjnnjzch commented on GitHub (Jan 20, 2026): Same issue ``` System: Ubuntu 22.04 + ollama docker 0.14.2 GPU: Nvidia RTX 4090 + CUDA 13.0 + Driver 580.95.05 CPU: Intel(R) Xeon(R) w7-3455 ``` The Error reports as: ```bash Error: 500 Internal Server Error: image runner failed: /usr/bin/ollama-mlx: error while loading shared libraries: libquadmath.so.0: cannot open shared object file: No such file or directory (exit: exit status 127) ``` I've tried `docker exec -it ollama bash -c 'apt update && apt install -y libquadmath0'` as @rick-github said (thanks by the way) > The missing library in the docker image can be installed with: > > ``` > docker exec -it ollama bash -c 'apt update && apt install -y libquadmath0' > ``` Here is the new error: ``` Error: 500 Internal Server Error: image runner failed: 2026/01/20 07:59:02 runner.go:65: INFO starting image runner model=x/z-image-turbo port=39315 (exit: exit status 255) ``` and another model gives the same error report: ``` Error: 500 Internal Server Error: image runner failed: 2026/01/20 08:00:02 runner.go:65: INFO starting image runner model=x/flux2-klein port=46651 (exit: exit status 255) ``` Hope to get any help, thanks! Edit: here is the logs ```bash time=2026-01-20T08:15:34.428Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax" time=2026-01-20T08:15:34.435Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/z-image-turbo port=36805 time=2026-01-20T08:15:34.621Z level=WARN source=server.go:141 msg=image-runner msg="2026/01/20 08:15:34 runner.go:65: INFO starting image runner model=x/z-image-turbo port=36805" time=2026-01-20T08:15:34.626Z level=INFO source=server.go:134 msg=image-runner msg="Loading Z-Image model from manifest: x/z-image-turbo..." time=2026-01-20T08:15:34.940Z level=INFO source=server.go:134 msg=image-runner msg=" Loading tokenizer... ✓" time=2026-01-20T08:15:37.009Z level=INFO source=server.go:134 msg=image-runner msg=" Loading text encoder... ✓" time=2026-01-20T08:15:37.536Z level=INFO source=server.go:134 msg=image-runner msg=" (11.3 GB, peak 11.3 GB)" time=2026-01-20T08:15:41.305Z level=INFO source=server.go:134 msg=image-runner msg=" Loading transformer... ✓" time=2026-01-20T08:15:43.206Z level=INFO source=server.go:134 msg=image-runner msg="MLX error: cudaMallocAsync(&data, size, stream) failed: out of memory at /go/src/github.com/ollama/ollama/build/_deps/mlx-c-src/mlx/c/transforms.cpp:73" time=2026-01-20T08:15:43.430Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=929 [GIN] 2026/01/20 - 08:15:48 | 500 | 14.01500212s | 127.0.0.1 | POST "/api/generate" ``` Edit: This model always runs out of memory, even I load ollama docker with 3 GPUs `--gpus '"device=3,2,1"' `. As my GPU has 24GB graphic memory, and the model size is 12GB, it is quite possible to be a bug. For another model ```bash time=2026-01-20T08:15:01.151Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax" time=2026-01-20T08:15:01.155Z level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/usr/bin/ollama-mlx model=x/flux2-klein port=41357 time=2026-01-20T08:15:01.341Z level=WARN source=server.go:141 msg=image-runner msg="2026/01/20 08:15:01 runner.go:65: INFO starting image runner model=x/flux2-klein port=41357" time=2026-01-20T08:15:01.345Z level=INFO source=server.go:134 msg=image-runner msg="Loading Z-Image model from manifest: x/flux2-klein..." time=2026-01-20T08:15:01.639Z level=INFO source=server.go:134 msg=image-runner msg=" Loading tokenizer... ✓" time=2026-01-20T08:15:02.874Z level=INFO source=server.go:134 msg=image-runner msg=" Loading text encoder... MLX error: [dequantize] Shape of scales and biases does not match the matrix given the quantization parameters. Provided matrix of shape (4096,320) and scales/biases of shape (4096,80) with group_size=32 and bits=8. at /go/src/github.com/ollama/ollama/build/_deps/mlx-c-src/mlx/c/ops.cpp:1085" time=2026-01-20T08:15:03.012Z level=INFO source=server.go:320 msg="stopping image runner subprocess" pid=876 ```
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#34758