[GH-ISSUE #9522] ollama run, windows 11, rtx 4090 just hangs #31969

Closed
opened 2026-04-22 12:48:58 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @ldemailly on GitHub (Mar 5, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/9522

What is the issue?

Latest ollama trying a basic

ollama run llama3.2:3b

just hangs with no activity shown in task manager and just

[GIN] 2025/03/05 - 09:45:12 | 200 |            0s |       127.0.0.1 | HEAD     "/"

in the logs (model is downloaded/pulled/checked fine)

Relevant log output

2025/03/05 09:44:46 routes.go:1215: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:2048 OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\Users\\l\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-03-05T09:44:46.360-08:00 level=INFO source=images.go:432 msg="total blobs: 16"
time=2025-03-05T09:44:46.361-08:00 level=INFO source=images.go:439 msg="total unused blobs removed: 0"
time=2025-03-05T09:44:46.361-08:00 level=INFO source=routes.go:1277 msg="Listening on 127.0.0.1:11434 (version 0.5.13)"
time=2025-03-05T09:44:46.361-08:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs"
time=2025-03-05T09:44:46.372-08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1
time=2025-03-05T09:44:46.372-08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=12 efficiency=0 threads=24
time=2025-03-05T09:44:46.686-08:00 level=INFO source=gpu.go:319 msg="detected OS VRAM overhead" id=GPU-3e0719c6-c053-80ac-1ced-8c58d06f03d5 library=cuda compute=8.9 driver=12.8 name="NVIDIA GeForce RTX 4090" overhead="702.4 MiB"
time=2025-03-05T09:44:47.080-08:00 level=INFO source=amd_windows.go:127 msg="unsupported Radeon iGPU detected skipping" id=0 total="24.0 GiB"
time=2025-03-05T09:44:47.082-08:00 level=INFO source=types.go:130 msg="inference compute" id=GPU-3e0719c6-c053-80ac-1ced-8c58d06f03d5 library=cuda variant=v12 compute=8.9 driver=12.8 name="NVIDIA GeForce RTX 4090" total="24.0 GiB" available="22.5 GiB"
[GIN] 2025/03/05 - 09:45:12 | 200 |            0s |       127.0.0.1 | HEAD     "/"

OS

Windows

GPU

Nvidia

CPU

AMD

Ollama version

ollama version is 0.5.13

Originally created by @ldemailly on GitHub (Mar 5, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/9522 ### What is the issue? Latest ollama trying a basic ``` ollama run llama3.2:3b ``` just hangs with no activity shown in task manager and just ``` [GIN] 2025/03/05 - 09:45:12 | 200 | 0s | 127.0.0.1 | HEAD "/" ``` in the logs (model is downloaded/pulled/checked fine) ### Relevant log output ```shell 2025/03/05 09:44:46 routes.go:1215: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:2048 OLLAMA_DEBUG:false OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\Users\\l\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-03-05T09:44:46.360-08:00 level=INFO source=images.go:432 msg="total blobs: 16" time=2025-03-05T09:44:46.361-08:00 level=INFO source=images.go:439 msg="total unused blobs removed: 0" time=2025-03-05T09:44:46.361-08:00 level=INFO source=routes.go:1277 msg="Listening on 127.0.0.1:11434 (version 0.5.13)" time=2025-03-05T09:44:46.361-08:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-03-05T09:44:46.372-08:00 level=INFO source=gpu_windows.go:167 msg=packages count=1 time=2025-03-05T09:44:46.372-08:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=12 efficiency=0 threads=24 time=2025-03-05T09:44:46.686-08:00 level=INFO source=gpu.go:319 msg="detected OS VRAM overhead" id=GPU-3e0719c6-c053-80ac-1ced-8c58d06f03d5 library=cuda compute=8.9 driver=12.8 name="NVIDIA GeForce RTX 4090" overhead="702.4 MiB" time=2025-03-05T09:44:47.080-08:00 level=INFO source=amd_windows.go:127 msg="unsupported Radeon iGPU detected skipping" id=0 total="24.0 GiB" time=2025-03-05T09:44:47.082-08:00 level=INFO source=types.go:130 msg="inference compute" id=GPU-3e0719c6-c053-80ac-1ced-8c58d06f03d5 library=cuda variant=v12 compute=8.9 driver=12.8 name="NVIDIA GeForce RTX 4090" total="24.0 GiB" available="22.5 GiB" [GIN] 2025/03/05 - 09:45:12 | 200 | 0s | 127.0.0.1 | HEAD "/" ``` ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version ollama version is 0.5.13
GiteaMirror added the bug label 2026-04-22 12:48:58 -05:00
Author
Owner

@ldemailly commented on GitHub (Mar 5, 2025):

looks like it hangs only from git bash and works fine from cmd.exe or web-ui
(maybe an entry for the faq?)

<!-- gh-comment-id:2701759296 --> @ldemailly commented on GitHub (Mar 5, 2025): looks like it hangs only from git bash and works fine from cmd.exe or web-ui (maybe an entry for the faq?)
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#31969