[GH-ISSUE #9839] Error: Post "http://127.0.0.1:11434/api/show": dial tcp 127.0.0.1:11434: connectex: No connection could be made because the target machine actively refused it. #32200

Closed
opened 2026-04-22 13:14:52 -05:00 by GiteaMirror · 3 comments
Owner

Originally created by @yisheng926 on GitHub (Mar 18, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/9839

What is the issue?

The same issue occurs when downloading any model

Relevant log output

2025/03/18 09:08:33 routes.go:1259: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\Users\\yisheng\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-03-18T09:08:33.565+08:00 level=INFO source=images.go:757 msg="total blobs: 0"
time=2025-03-18T09:08:33.565+08:00 level=INFO source=images.go:764 msg="total unused blobs removed: 0"
time=2025-03-18T09:08:33.566+08:00 level=INFO source=routes.go:1310 msg="Listening on 127.0.0.1:11434 (version 0.5.4)"
time=2025-03-18T09:08:33.566+08:00 level=INFO source=routes.go:1339 msg="Dynamic LLM libraries" runners="[cpu cpu_avx cpu_avx2 cuda_v11_avx cuda_v12_avx rocm_avx]"
time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu.go:226 msg="looking for compatible GPUs"
time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1
time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu_windows.go:183 msg="efficiency cores detected" maxEfficiencyClass=1
time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=8 threads=20
time=2025-03-18T09:08:33.583+08:00 level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered"
time=2025-03-18T09:08:33.583+08:00 level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="15.6 GiB" available="10.1 GiB"
[GIN] 2025/03/18 - 09:08:52 | 200 |       501.3µs |       127.0.0.1 | HEAD     "/"
[GIN] 2025/03/18 - 09:08:52 | 404 |       501.1µs |       127.0.0.1 | POST     "/api/show"
panic: runtime error: index out of range [0] with length 0

goroutine 42 [running]:
github.com/ollama/ollama/server.(*blobDownload).Prepare(0xc0002b2b60, {0x7ff6bd283df0, 0xc00073ac80}, 0xc000144240, 0xc0003f45c0)
	github.com/ollama/ollama/server/download.go:175 +0x539
github.com/ollama/ollama/server.downloadBlob({0x7ff6bd283df0, 0xc00073ac80}, {{{0x7ff6bd0c5e4d, 0x5}, {0x7ff6bd0daafc, 0x12}, {0x7ff6bd0ce874, 0x7}, {0xc0002cc680, 0xb}, ...}, ...})
	github.com/ollama/ollama/server/download.go:489 +0x4da
github.com/ollama/ollama/server.PullModel({0x7ff6bd283df0, 0xc00073ac80}, {0xc0002cc680, 0x10}, 0xc0003f45c0, 0xc00021a2d0)
	github.com/ollama/ollama/server/images.go:889 +0x771
github.com/ollama/ollama/server.(*Server).PullHandler.func1()
	github.com/ollama/ollama/server/routes.go:595 +0x197
created by github.com/ollama/ollama/server.(*Server).PullHandler in goroutine 52
	github.com/ollama/ollama/server/routes.go:582 +0x691

OS

Windows

GPU

No response

CPU

Intel

Ollama version

0.5.4

Originally created by @yisheng926 on GitHub (Mar 18, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/9839 ### What is the issue? The same issue occurs when downloading any model ### Relevant log output ```shell 2025/03/18 09:08:33 routes.go:1259: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\Users\\yisheng\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-03-18T09:08:33.565+08:00 level=INFO source=images.go:757 msg="total blobs: 0" time=2025-03-18T09:08:33.565+08:00 level=INFO source=images.go:764 msg="total unused blobs removed: 0" time=2025-03-18T09:08:33.566+08:00 level=INFO source=routes.go:1310 msg="Listening on 127.0.0.1:11434 (version 0.5.4)" time=2025-03-18T09:08:33.566+08:00 level=INFO source=routes.go:1339 msg="Dynamic LLM libraries" runners="[cpu cpu_avx cpu_avx2 cuda_v11_avx cuda_v12_avx rocm_avx]" time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu.go:226 msg="looking for compatible GPUs" time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1 time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu_windows.go:183 msg="efficiency cores detected" maxEfficiencyClass=1 time=2025-03-18T09:08:33.566+08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=8 threads=20 time=2025-03-18T09:08:33.583+08:00 level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered" time=2025-03-18T09:08:33.583+08:00 level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="15.6 GiB" available="10.1 GiB" [GIN] 2025/03/18 - 09:08:52 | 200 | 501.3µs | 127.0.0.1 | HEAD "/" [GIN] 2025/03/18 - 09:08:52 | 404 | 501.1µs | 127.0.0.1 | POST "/api/show" panic: runtime error: index out of range [0] with length 0 goroutine 42 [running]: github.com/ollama/ollama/server.(*blobDownload).Prepare(0xc0002b2b60, {0x7ff6bd283df0, 0xc00073ac80}, 0xc000144240, 0xc0003f45c0) github.com/ollama/ollama/server/download.go:175 +0x539 github.com/ollama/ollama/server.downloadBlob({0x7ff6bd283df0, 0xc00073ac80}, {{{0x7ff6bd0c5e4d, 0x5}, {0x7ff6bd0daafc, 0x12}, {0x7ff6bd0ce874, 0x7}, {0xc0002cc680, 0xb}, ...}, ...}) github.com/ollama/ollama/server/download.go:489 +0x4da github.com/ollama/ollama/server.PullModel({0x7ff6bd283df0, 0xc00073ac80}, {0xc0002cc680, 0x10}, 0xc0003f45c0, 0xc00021a2d0) github.com/ollama/ollama/server/images.go:889 +0x771 github.com/ollama/ollama/server.(*Server).PullHandler.func1() github.com/ollama/ollama/server/routes.go:595 +0x197 created by github.com/ollama/ollama/server.(*Server).PullHandler in goroutine 52 github.com/ollama/ollama/server/routes.go:582 +0x691 ``` ### OS Windows ### GPU _No response_ ### CPU Intel ### Ollama version 0.5.4
GiteaMirror added the bug label 2026-04-22 13:14:52 -05:00
Author
Owner

@rick-github commented on GitHub (Mar 18, 2025):

https://github.com/ollama/ollama/pull/8746

Upgrade to 0.5.8 or newer.

<!-- gh-comment-id:2731360029 --> @rick-github commented on GitHub (Mar 18, 2025): https://github.com/ollama/ollama/pull/8746 [Upgrade](https://github.com/ollama/ollama/blob/main/docs/faq.md#how-can-i-upgrade-ollama) to 0.5.8 or newer.
Author
Owner

@yisheng926 commented on GitHub (Mar 18, 2025):

#8746

Upgrade to 0.5.8 or newer.

After updating to version 0.6.1, a new issue occurred

Image

<!-- gh-comment-id:2731376585 --> @yisheng926 commented on GitHub (Mar 18, 2025): > [#8746](https://github.com/ollama/ollama/pull/8746) > > [Upgrade](https://github.com/ollama/ollama/blob/main/docs/faq.md#how-can-i-upgrade-ollama) to 0.5.8 or newer. After updating to version 0.6.1, a new issue occurred ![Image](https://github.com/user-attachments/assets/c3b797f2-2a7a-48da-b74c-d83e3ed6a02f)
Author
Owner

@yisheng926 commented on GitHub (Mar 18, 2025):

#8746

Upgrade to 0.5.8 or newer.

2025/03/18 09:46:40 routes.go:1230: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:2048 OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY:cpu OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\Users\yisheng\.ollama\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-03-18T09:46:40.265+08:00 level=INFO source=images.go:432 msg="total blobs: 0"
time=2025-03-18T09:46:40.265+08:00 level=INFO source=images.go:439 msg="total unused blobs removed: 0"
time=2025-03-18T09:46:40.265+08:00 level=INFO source=routes.go:1297 msg="Listening on 127.0.0.1:11434 (version 0.6.1)"
time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs"
time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1
time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu_windows.go:183 msg="efficiency cores detected" maxEfficiencyClass=1
time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=8 threads=20
time=2025-03-18T09:46:40.299+08:00 level=INFO source=gpu.go:377 msg="no compatible GPUs were discovered"
time=2025-03-18T09:46:40.299+08:00 level=INFO source=types.go:130 msg="inference compute" id=0 library=cpu variant="" compute="" driver=0.0 name="" total="15.6 GiB" available="8.1 GiB"
[GIN] 2025/03/18 - 09:47:08 | 200 | 0s | 127.0.0.1 | HEAD "/"
[GIN] 2025/03/18 - 09:47:17 | 200 | 509.9µs | 127.0.0.1 | GET "/api/version"
[GIN] 2025/03/18 - 09:47:27 | 200 | 0s | 127.0.0.1 | HEAD "/"
[GIN] 2025/03/18 - 09:47:27 | 404 | 515.8µs | 127.0.0.1 | POST "/api/show"
[GIN] 2025/03/18 - 09:47:32 | 200 | 5.7646472s | 127.0.0.1 | POST "/api/pull"

<!-- gh-comment-id:2731377554 --> @yisheng926 commented on GitHub (Mar 18, 2025): > [#8746](https://github.com/ollama/ollama/pull/8746) > > [Upgrade](https://github.com/ollama/ollama/blob/main/docs/faq.md#how-can-i-upgrade-ollama) to 0.5.8 or newer. 2025/03/18 09:46:40 routes.go:1230: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:2048 OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY:cpu OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\Users\\yisheng\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-03-18T09:46:40.265+08:00 level=INFO source=images.go:432 msg="total blobs: 0" time=2025-03-18T09:46:40.265+08:00 level=INFO source=images.go:439 msg="total unused blobs removed: 0" time=2025-03-18T09:46:40.265+08:00 level=INFO source=routes.go:1297 msg="Listening on 127.0.0.1:11434 (version 0.6.1)" time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1 time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu_windows.go:183 msg="efficiency cores detected" maxEfficiencyClass=1 time=2025-03-18T09:46:40.265+08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=8 threads=20 time=2025-03-18T09:46:40.299+08:00 level=INFO source=gpu.go:377 msg="no compatible GPUs were discovered" time=2025-03-18T09:46:40.299+08:00 level=INFO source=types.go:130 msg="inference compute" id=0 library=cpu variant="" compute="" driver=0.0 name="" total="15.6 GiB" available="8.1 GiB" [GIN] 2025/03/18 - 09:47:08 | 200 | 0s | 127.0.0.1 | HEAD "/" [GIN] 2025/03/18 - 09:47:17 | 200 | 509.9µs | 127.0.0.1 | GET "/api/version" [GIN] 2025/03/18 - 09:47:27 | 200 | 0s | 127.0.0.1 | HEAD "/" [GIN] 2025/03/18 - 09:47:27 | 404 | 515.8µs | 127.0.0.1 | POST "/api/show" [GIN] 2025/03/18 - 09:47:32 | 200 | 5.7646472s | 127.0.0.1 | POST "/api/pull"
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#32200