[GH-ISSUE #13780] Removing image generation model unexpectedly generates an image first #34790

Open
opened 2026-04-22 18:38:40 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @mchiang0610 on GitHub (Jan 19, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/13780

What is the issue?

ollama v0.14.3 (macOS Tahoe 26.2; Apple M4 Max)

ollama@ollamas-computer ~ % ollama serve
time=2026-01-19T14:53:56.396-05:00 level=INFO source=routes.go:1614 msg="server config" env="map[HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/Users/ollama/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false http_proxy: https_proxy: no_proxy:]"
time=2026-01-19T14:53:56.437-05:00 level=INFO source=images.go:499 msg="total blobs: 2907"
time=2026-01-19T14:53:56.444-05:00 level=INFO source=images.go:506 msg="total unused blobs removed: 0"
time=2026-01-19T14:53:56.445-05:00 level=INFO source=routes.go:1667 msg="Listening on 127.0.0.1:11434 (version 0.14.2)"
time=2026-01-19T14:53:56.445-05:00 level=INFO source=runner.go:67 msg="discovering available GPUs..."
time=2026-01-19T14:53:56.446-05:00 level=INFO source=server.go:429 msg="starting runner" cmd="/Applications/Ollama.app/Contents/Resources/ollama runner --ollama-engine --port 55398"
time=2026-01-19T14:53:56.493-05:00 level=INFO source=types.go:42 msg="inference compute" id=0 filter_id=0 library=Metal compute=0.0 name=Metal description="Apple M4 Max" libdirs="" driver=0.0 pci_id="" type=discrete total="107.5 GiB" available="107.5 GiB"
[GIN] 2026/01/19 - 14:54:07 | 200 |      26.333µs |       127.0.0.1 | HEAD     "/"
[GIN] 2026/01/19 - 14:54:07 | 200 |    35.80675ms |       127.0.0.1 | POST     "/api/show"
time=2026-01-19T14:54:07.399-05:00 level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/Applications/Ollama.app/Contents/Resources/ollama-mlx model=x/z-image-turbo:latest port=55439
time=2026-01-19T14:54:07.443-05:00 level=WARN source=server.go:141 msg=image-runner msg="2026/01/19 14:54:07 runner.go:65: INFO starting image runner model=x/z-image-turbo:latest port=55439"
time=2026-01-19T14:54:07.446-05:00 level=INFO source=server.go:134 msg=image-runner msg="Loading Z-Image model from manifest: x/z-image-turbo:latest..."
time=2026-01-19T14:54:07.634-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading tokenizer... ✓"
time=2026-01-19T14:54:08.105-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading text encoder... ✓"
time=2026-01-19T14:54:08.105-05:00 level=INFO source=server.go:134 msg=image-runner msg="  (4.5 GB, peak 4.5 GB)"
time=2026-01-19T14:54:08.819-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading transformer... ✓"
time=2026-01-19T14:54:08.819-05:00 level=INFO source=server.go:134 msg=image-runner msg="  (11.7 GB, peak 11.7 GB)"
time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading conv_in... ✓"
time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading mid block... ✓"
time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading up blocks... ✓ [4 blocks]"
time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading conv_norm_out... ✓"
time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loading conv_out... ✓"
time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg="  (11.9 GB, peak 11.9 GB)"
time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Loaded in 1.41s (11.9 GB VRAM)"
time=2026-01-19T14:54:08.854-05:00 level=WARN source=server.go:141 msg=image-runner msg="2026/01/19 14:54:08 runner.go:109: INFO image runner listening addr=127.0.0.1:55439"
time=2026-01-19T14:54:08.902-05:00 level=INFO source=server.go:225 msg="image runner is ready" port=55439
time=2026-01-19T14:54:09.305-05:00 level=INFO source=server.go:134 msg=image-runner msg="  TeaCache enabled: threshold=0.15"
time=2026-01-19T14:54:19.412-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 1/9: t=1.0000 (10.11s) [11.9 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:19.412-05:00 level=INFO source=server.go:134 msg=image-runner msg="    [TeaCache: reusing cached output]"
time=2026-01-19T14:54:19.413-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 2/9: t=0.9619 (0.00s) [11.9 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:19.413-05:00 level=INFO source=server.go:134 msg=image-runner msg="    [TeaCache: reusing cached output]"
time=2026-01-19T14:54:19.413-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 3/9: t=0.9170 (0.00s) [11.9 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:26.839-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 4/9: t=0.8633 (7.43s) [12.1 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:26.839-05:00 level=INFO source=server.go:134 msg=image-runner msg="    [TeaCache: reusing cached output]"
time=2026-01-19T14:54:26.839-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 5/9: t=0.7979 (0.00s) [11.9 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:33.813-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 6/9: t=0.7164 (6.97s) [12.0 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:33.813-05:00 level=INFO source=server.go:134 msg=image-runner msg="    [TeaCache: reusing cached output]"
time=2026-01-19T14:54:33.814-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 7/9: t=0.6123 (0.00s) [11.9 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:41.082-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 8/9: t=0.4743 (7.27s) [11.9 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:49.473-05:00 level=INFO source=server.go:134 msg=image-runner msg="  Step 9/9: t=0.2830 (8.39s) [12.1 GB active, 13.5 GB peak]"
time=2026-01-19T14:54:49.473-05:00 level=INFO source=server.go:134 msg=image-runner msg="  TeaCache stats: 4 hits, 5 misses (44.4% cache rate)"
time=2026-01-19T14:54:50.696-05:00 level=INFO source=server.go:134 msg=image-runner msg="Generated in 41.79s (9 steps)"
[GIN] 2026/01/19 - 14:54:50 | 200 | 43.504865458s |       127.0.0.1 | POST     "/api/generate"
[GIN] 2026/01/19 - 14:54:51 | 200 |  544.721292ms |       127.0.0.1 | DELETE   "/api/delete"

Ollama version

v0.14.3

Originally created by @mchiang0610 on GitHub (Jan 19, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/13780 ### What is the issue? ollama v0.14.3 (macOS Tahoe 26.2; Apple M4 Max) ``` ollama@ollamas-computer ~ % ollama serve time=2026-01-19T14:53:56.396-05:00 level=INFO source=routes.go:1614 msg="server config" env="map[HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/Users/ollama/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false http_proxy: https_proxy: no_proxy:]" time=2026-01-19T14:53:56.437-05:00 level=INFO source=images.go:499 msg="total blobs: 2907" time=2026-01-19T14:53:56.444-05:00 level=INFO source=images.go:506 msg="total unused blobs removed: 0" time=2026-01-19T14:53:56.445-05:00 level=INFO source=routes.go:1667 msg="Listening on 127.0.0.1:11434 (version 0.14.2)" time=2026-01-19T14:53:56.445-05:00 level=INFO source=runner.go:67 msg="discovering available GPUs..." time=2026-01-19T14:53:56.446-05:00 level=INFO source=server.go:429 msg="starting runner" cmd="/Applications/Ollama.app/Contents/Resources/ollama runner --ollama-engine --port 55398" time=2026-01-19T14:53:56.493-05:00 level=INFO source=types.go:42 msg="inference compute" id=0 filter_id=0 library=Metal compute=0.0 name=Metal description="Apple M4 Max" libdirs="" driver=0.0 pci_id="" type=discrete total="107.5 GiB" available="107.5 GiB" [GIN] 2026/01/19 - 14:54:07 | 200 | 26.333µs | 127.0.0.1 | HEAD "/" [GIN] 2026/01/19 - 14:54:07 | 200 | 35.80675ms | 127.0.0.1 | POST "/api/show" time=2026-01-19T14:54:07.399-05:00 level=INFO source=server.go:149 msg="starting ollama-mlx image runner subprocess" exe=/Applications/Ollama.app/Contents/Resources/ollama-mlx model=x/z-image-turbo:latest port=55439 time=2026-01-19T14:54:07.443-05:00 level=WARN source=server.go:141 msg=image-runner msg="2026/01/19 14:54:07 runner.go:65: INFO starting image runner model=x/z-image-turbo:latest port=55439" time=2026-01-19T14:54:07.446-05:00 level=INFO source=server.go:134 msg=image-runner msg="Loading Z-Image model from manifest: x/z-image-turbo:latest..." time=2026-01-19T14:54:07.634-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading tokenizer... ✓" time=2026-01-19T14:54:08.105-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading text encoder... ✓" time=2026-01-19T14:54:08.105-05:00 level=INFO source=server.go:134 msg=image-runner msg=" (4.5 GB, peak 4.5 GB)" time=2026-01-19T14:54:08.819-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading transformer... ✓" time=2026-01-19T14:54:08.819-05:00 level=INFO source=server.go:134 msg=image-runner msg=" (11.7 GB, peak 11.7 GB)" time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading conv_in... ✓" time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading mid block... ✓" time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading up blocks... ✓ [4 blocks]" time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading conv_norm_out... ✓" time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loading conv_out... ✓" time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg=" (11.9 GB, peak 11.9 GB)" time=2026-01-19T14:54:08.854-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Loaded in 1.41s (11.9 GB VRAM)" time=2026-01-19T14:54:08.854-05:00 level=WARN source=server.go:141 msg=image-runner msg="2026/01/19 14:54:08 runner.go:109: INFO image runner listening addr=127.0.0.1:55439" time=2026-01-19T14:54:08.902-05:00 level=INFO source=server.go:225 msg="image runner is ready" port=55439 time=2026-01-19T14:54:09.305-05:00 level=INFO source=server.go:134 msg=image-runner msg=" TeaCache enabled: threshold=0.15" time=2026-01-19T14:54:19.412-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 1/9: t=1.0000 (10.11s) [11.9 GB active, 13.5 GB peak]" time=2026-01-19T14:54:19.412-05:00 level=INFO source=server.go:134 msg=image-runner msg=" [TeaCache: reusing cached output]" time=2026-01-19T14:54:19.413-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 2/9: t=0.9619 (0.00s) [11.9 GB active, 13.5 GB peak]" time=2026-01-19T14:54:19.413-05:00 level=INFO source=server.go:134 msg=image-runner msg=" [TeaCache: reusing cached output]" time=2026-01-19T14:54:19.413-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 3/9: t=0.9170 (0.00s) [11.9 GB active, 13.5 GB peak]" time=2026-01-19T14:54:26.839-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 4/9: t=0.8633 (7.43s) [12.1 GB active, 13.5 GB peak]" time=2026-01-19T14:54:26.839-05:00 level=INFO source=server.go:134 msg=image-runner msg=" [TeaCache: reusing cached output]" time=2026-01-19T14:54:26.839-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 5/9: t=0.7979 (0.00s) [11.9 GB active, 13.5 GB peak]" time=2026-01-19T14:54:33.813-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 6/9: t=0.7164 (6.97s) [12.0 GB active, 13.5 GB peak]" time=2026-01-19T14:54:33.813-05:00 level=INFO source=server.go:134 msg=image-runner msg=" [TeaCache: reusing cached output]" time=2026-01-19T14:54:33.814-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 7/9: t=0.6123 (0.00s) [11.9 GB active, 13.5 GB peak]" time=2026-01-19T14:54:41.082-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 8/9: t=0.4743 (7.27s) [11.9 GB active, 13.5 GB peak]" time=2026-01-19T14:54:49.473-05:00 level=INFO source=server.go:134 msg=image-runner msg=" Step 9/9: t=0.2830 (8.39s) [12.1 GB active, 13.5 GB peak]" time=2026-01-19T14:54:49.473-05:00 level=INFO source=server.go:134 msg=image-runner msg=" TeaCache stats: 4 hits, 5 misses (44.4% cache rate)" time=2026-01-19T14:54:50.696-05:00 level=INFO source=server.go:134 msg=image-runner msg="Generated in 41.79s (9 steps)" [GIN] 2026/01/19 - 14:54:50 | 200 | 43.504865458s | 127.0.0.1 | POST "/api/generate" [GIN] 2026/01/19 - 14:54:51 | 200 | 544.721292ms | 127.0.0.1 | DELETE "/api/delete" ``` ### Ollama version v0.14.3
GiteaMirror added the bug label 2026-04-22 18:38:40 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#34790