[GH-ISSUE #12618] Ollama serve fails to detect Nvidia GPUs after updating to the latest version #54889

Closed
opened 2026-04-29 07:49:08 -05:00 by GiteaMirror · 52 comments
Owner

Originally created by @catclaw on GitHub (Oct 14, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/12618

Originally assigned to: @dhiltgen on GitHub.

What is the issue?

Hi!
I'm experiencing a strange error after updating Ollama to the latest version today.
I'm running Ollama on a Windows 11 Enterprise 25H2 machine - no Docker.
When I start Ollama using "ollama serve", it fails to detect my GPUs and falls back to CPU.

time=2025-10-14T21:18:01.270+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-14T21:18:02.309+02:00 level=INFO source=images.go:522 msg="total blobs: 145" time=2025-10-14T21:18:02.326+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-14T21:18:02.340+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)" time=2025-10-14T21:18:02.348+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-14T21:18:03.030+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="101.0 GiB" time=2025-10-14T21:18:03.030+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

When I check the older logs, I can see that it detected the GPUs before the update:
time=2025-09-17T13:55:55.012+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-09-17T13:55:55.303+02:00 level=INFO source=images.go:477 msg="total blobs: 100" time=2025-09-17T13:55:55.314+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-09-17T13:55:55.324+02:00 level=INFO source=routes.go:1385 msg="Listening on [::]:11434 (version 0.11.11)" time=2025-09-17T13:55:55.324+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:56.007+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="156.4 MiB" time=2025-09-17T13:55:56.212+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="79.7 MiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB"

My Nvidia driver are the latest and I have CUDA 13.0 installed.
I've tried to uninstall and reinstall Ollama, my Nvidia driver, CUDA, etc. - but nothing seems to work.
I also added CUDA_VISIBLE_DEVICES:0,1,2 to my system variables - but Ollama still refuses to detect my GPUs.

I've googled and searched, and tried every tutorial, but I'm stuck with Ollama using CPU instead of my GPUs.
I'd appreciate any help I can get.
Sorry for bothering and thank you in advance.

Relevant log output

time=2025-10-14T21:18:01.270+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-10-14T21:18:02.309+02:00 level=INFO source=images.go:522 msg="total blobs: 145"
time=2025-10-14T21:18:02.326+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-10-14T21:18:02.340+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)"
time=2025-10-14T21:18:02.348+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..."
time=2025-10-14T21:18:03.030+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="101.0 GiB"
time=2025-10-14T21:18:03.030+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

Vs:

time=2025-09-17T13:55:55.012+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-09-17T13:55:55.303+02:00 level=INFO source=images.go:477 msg="total blobs: 100"
time=2025-09-17T13:55:55.314+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0"
time=2025-09-17T13:55:55.324+02:00 level=INFO source=routes.go:1385 msg="Listening on [::]:11434 (version 0.11.11)"
time=2025-09-17T13:55:55.324+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs"
time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2
time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28
time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28
time=2025-09-17T13:55:56.007+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="156.4 MiB"
time=2025-09-17T13:55:56.212+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="79.7 MiB"
time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB"
time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB"
time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB"

OS

Windows

GPU

Nvidia

CPU

Intel

Ollama version

0.12.5

Originally created by @catclaw on GitHub (Oct 14, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/12618 Originally assigned to: @dhiltgen on GitHub. ### What is the issue? Hi! I'm experiencing a strange error after updating Ollama to the latest version today. I'm running Ollama on a Windows 11 Enterprise 25H2 machine - no Docker. When I start Ollama using "ollama serve", it fails to detect my GPUs and falls back to CPU. `time=2025-10-14T21:18:01.270+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-14T21:18:02.309+02:00 level=INFO source=images.go:522 msg="total blobs: 145" time=2025-10-14T21:18:02.326+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-14T21:18:02.340+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)" time=2025-10-14T21:18:02.348+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-14T21:18:03.030+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="101.0 GiB" time=2025-10-14T21:18:03.030+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"` When I check the older logs, I can see that it detected the GPUs before the update: `time=2025-09-17T13:55:55.012+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-09-17T13:55:55.303+02:00 level=INFO source=images.go:477 msg="total blobs: 100" time=2025-09-17T13:55:55.314+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-09-17T13:55:55.324+02:00 level=INFO source=routes.go:1385 msg="Listening on [::]:11434 (version 0.11.11)" time=2025-09-17T13:55:55.324+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:56.007+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="156.4 MiB" time=2025-09-17T13:55:56.212+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="79.7 MiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" ` My Nvidia driver are the latest and I have CUDA 13.0 installed. I've tried to uninstall and reinstall Ollama, my Nvidia driver, CUDA, etc. - but nothing seems to work. I also added CUDA_VISIBLE_DEVICES:0,1,2 to my system variables - but Ollama still refuses to detect my GPUs. I've googled and searched, and tried every tutorial, but I'm stuck with Ollama using CPU instead of my GPUs. I'd appreciate any help I can get. Sorry for bothering and thank you in advance. ### Relevant log output ```shell time=2025-10-14T21:18:01.270+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-14T21:18:02.309+02:00 level=INFO source=images.go:522 msg="total blobs: 145" time=2025-10-14T21:18:02.326+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-14T21:18:02.340+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)" time=2025-10-14T21:18:02.348+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-14T21:18:03.030+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="101.0 GiB" time=2025-10-14T21:18:03.030+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" Vs: time=2025-09-17T13:55:55.012+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-09-17T13:55:55.303+02:00 level=INFO source=images.go:477 msg="total blobs: 100" time=2025-09-17T13:55:55.314+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-09-17T13:55:55.324+02:00 level=INFO source=routes.go:1385 msg="Listening on [::]:11434 (version 0.11.11)" time=2025-09-17T13:55:55.324+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:56.007+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="156.4 MiB" time=2025-09-17T13:55:56.212+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="79.7 MiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" ``` ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.12.5
GiteaMirror added the bugnvidiawindows labels 2026-04-29 07:49:09 -05:00
Author
Owner

@dhiltgen commented on GitHub (Oct 14, 2025):

@catclaw could you quit Ollama in the system tray, and then run the following in a powershell terminal and share the logs?

$env:OLLAMA_DEBUG="2"
ollama serve 2>&1 | % ToString | tee-object serve.log

Once it reports "inference compute" you can ^C and send the log.

<!-- gh-comment-id:3403371205 --> @dhiltgen commented on GitHub (Oct 14, 2025): @catclaw could you quit Ollama in the system tray, and then run the following in a powershell terminal and share the logs? ```powershell $env:OLLAMA_DEBUG="2" ollama serve 2>&1 | % ToString | tee-object serve.log ``` Once it reports "inference compute" you can `^C` and send the log.
Author
Owner

@catclaw commented on GitHub (Oct 14, 2025):

@catclaw could you quit Ollama in the system tray, and then run the following in a powershell terminal and share the logs?

$env:OLLAMA_DEBUG="2"
ollama serve 2>&1 | % ToString | tee-object serve.log
Once it reports "inference compute" you can ^C and send the log.

Yes, of course! Thank you so much for taking your time to look into this. 🙏

serve.log

<!-- gh-comment-id:3403743697 --> @catclaw commented on GitHub (Oct 14, 2025): > [@catclaw](https://github.com/catclaw) could you quit Ollama in the system tray, and then run the following in a powershell terminal and share the logs? > > $env:OLLAMA_DEBUG="2" > ollama serve 2>&1 | % ToString | tee-object serve.log > Once it reports "inference compute" you can `^C` and send the log. Yes, of course! Thank you so much for taking your time to look into this. 🙏 [serve.log](https://github.com/user-attachments/files/22914022/serve.log)
Author
Owner

@dhiltgen commented on GitHub (Oct 14, 2025):

My suspicion is it's from CUDA_VISIBLE_DEVICES=0,1,2. Try unsetting that first and see if it discovers the GPUs, or use the UUIDs instead which you can gather from nvidia-smi -L

If that doesn't solve it, can you share the output of nvidia-smi?

<!-- gh-comment-id:3403817000 --> @dhiltgen commented on GitHub (Oct 14, 2025): My suspicion is it's from `CUDA_VISIBLE_DEVICES=0,1,2`. Try unsetting that first and see if it discovers the GPUs, or use the UUIDs instead which you can gather from `nvidia-smi -L` If that doesn't solve it, can you share the output of `nvidia-smi`?
Author
Owner

@catclaw commented on GitHub (Oct 15, 2025):

My suspicion is it's from CUDA_VISIBLE_DEVICES=0,1,2. Try unsetting that first and see if it discovers the GPUs, or use the UUIDs instead which you can gather from nvidia-smi -L

If that doesn't solve it, can you share the output of nvidia-smi?

I added CUDA_VISIBLE_DEVICES=0,1,2 to System Variables, hoping Ollama would detect the GPUs after updating from 0.11.11 to 0.12.5, but it didn't work - and neither does using GPU UUIDs...
Meanwhile, I don't have any problems creating three ComfyUI instances in SwarmUI, for example, to run 3 generations simultaneously.
I've been up all night trying to figure out the problem, with no luck... 😵‍💫

C:\Windows\System32>nvidia-smi
Wed Oct 15 04:20:35 2025
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 581.57                 Driver Version: 581.57         CUDA Version: 13.0     |
+-----------------------------------------+------------------------+----------------------+
| GPU  Name                  Driver-Model | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce RTX 5070      WDDM  |   00000000:03:00.0  On |                  N/A |
|100%   45C    P0             31W /  250W |    3199MiB /  12227MiB |      7%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+
|   1  NVIDIA GeForce RTX 3060      WDDM  |   00000000:04:00.0  On |                  N/A |
|100%   41C    P0             38W /  170W |     909MiB /  12288MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+
|   2  NVIDIA GeForce RTX 3060      WDDM  |   00000000:83:00.0  On |                  N/A |
|100%   42C    P0             38W /  170W |     946MiB /  12288MiB |      3%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI              PID   Type   Process name                        GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A            3936    C+G   ....0.3537.71\msedgewebview2.exe      N/A      |
|    0   N/A  N/A            5992    C+G   ...cw5n1h2txyewy\WidgetBoard.exe      N/A      |
|    0   N/A  N/A            6952    C+G   ...em32\ApplicationFrameHost.exe      N/A      |
|    0   N/A  N/A           12512    C+G   ...XT CAM Beta\NZXT CAM Beta.exe      N/A      |
|    0   N/A  N/A           12736    C+G   ...-dash\dash\bin\OculusDash.exe      N/A      |
|    0   N/A  N/A           14012    C+G   ...Chrome\Application\chrome.exe      N/A      |
|    0   N/A  N/A           14520    C+G   ....0.3537.71\msedgewebview2.exe      N/A      |
|    0   N/A  N/A           16656    C+G   ...indows\System32\ShellHost.exe      N/A      |
|    0   N/A  N/A           16812    C+G   ...lus-runtime\OVRServer_x64.exe      N/A      |
|    0   N/A  N/A           20436    C+G   ...4__cv1g1gvanyjgm\WhatsApp.exe      N/A      |
|    0   N/A  N/A           20800    C+G   C:\Windows\explorer.exe               N/A      |
|    0   N/A  N/A           21116    C+G   ...e Beta\Application\chrome.exe      N/A      |
|    0   N/A  N/A           23596    C+G   ...8bbwe\Microsoft.CmdPal.UI.exe      N/A      |
|    0   N/A  N/A           23632    C+G   ..._cw5n1h2txyewy\SearchHost.exe      N/A      |
|    0   N/A  N/A           23808    C+G   ...y\StartMenuExperienceHost.exe      N/A      |
|    0   N/A  N/A           24508    C+G   ...5n1h2txyewy\TextInputHost.exe      N/A      |
|    0   N/A  N/A           25064    C+G   ...crosoft shared\ink\TabTip.exe      N/A      |
|    0   N/A  N/A           28432    C+G   ...8bbwe\PhoneExperienceHost.exe      N/A      |
|    0   N/A  N/A           28728    C+G   ....0.3537.71\msedgewebview2.exe      N/A      |
|    0   N/A  N/A           30808    C+G   ...s\Windows Sidebar\sidebar.exe      N/A      |
|    0   N/A  N/A           31180    C+G   ...culus-client\OculusClient.exe      N/A      |
|    0   N/A  N/A           31248    C+G   ...Chrome\Application\chrome.exe      N/A      |
|    0   N/A  N/A           31284    C+G   ...yb3d8bbwe\Notepad\Notepad.exe      N/A      |
|    0   N/A  N/A           35440    C+G   ...ef.win7x64\steamwebhelper.exe      N/A      |
|    0   N/A  N/A           37764    C+G   ...ndows\System32\mmgaserver.exe      N/A      |
|    0   N/A  N/A           38496    C+G   ...crosoft OneDrive\OneDrive.exe      N/A      |
|    0   N/A  N/A           41316    C+G   ...er\Application\AVGBrowser.exe      N/A      |
|    0   N/A  N/A           43692    C+G   ...x40ttqa\iCloud\iCloudHome.exe      N/A      |
|    0   N/A  N/A           46780    C+G   ...ntrolPanel\SystemSettings.exe      N/A      |
|    0   N/A  N/A           51780    C+G   ...xyewy\ShellExperienceHost.exe      N/A      |
|    0   N/A  N/A           53112    C+G   ...2p2nqsd0c76g0\app\ChatGPT.exe      N/A      |
|    0   N/A  N/A           53732    C+G   ...8wekyb3d8bbwe\WebViewHost.exe      N/A      |
|    0   N/A  N/A           53928    C+G   ....0.3537.71\msedgewebview2.exe      N/A      |
|    0   N/A  N/A           54628    C+G   ...lare WARP\Cloudflare WARP.exe      N/A      |
|    0   N/A  N/A           55140    C+G   ....0.3537.71\msedgewebview2.exe      N/A      |
|    0   N/A  N/A           56636    C+G   ...\Imaging Edge Desktop\ied.exe      N/A      |
|    0   N/A  N/A           58360    C+G   ...ons\ScpToolkit\ScpTrayApp.exe      N/A      |
|    1   N/A  N/A           20800    C+G   C:\Windows\explorer.exe               N/A      |
|    1   N/A  N/A           54628    C+G   ...lare WARP\Cloudflare WARP.exe      N/A      |
|    1   N/A  N/A           56636    C+G   ...\Imaging Edge Desktop\ied.exe      N/A      |
|    1   N/A  N/A           58360    C+G   ...ons\ScpToolkit\ScpTrayApp.exe      N/A      |
|    2   N/A  N/A           20800    C+G   C:\Windows\explorer.exe               N/A      |
|    2   N/A  N/A           31284    C+G   ...yb3d8bbwe\Notepad\Notepad.exe      N/A      |
|    2   N/A  N/A           54628    C+G   ...lare WARP\Cloudflare WARP.exe      N/A      |
|    2   N/A  N/A           56636    C+G   ...\Imaging Edge Desktop\ied.exe      N/A      |
|    2   N/A  N/A           58360    C+G   ...ons\ScpToolkit\ScpTrayApp.exe      N/A      |
+-----------------------------------------------------------------------------------------+

C:\Windows\System32>nvidia-smi -L
GPU 0: NVIDIA GeForce RTX 5070 (UUID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8)
GPU 1: NVIDIA GeForce RTX 3060 (UUID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced)
GPU 2: NVIDIA GeForce RTX 3060 (UUID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c)`

After adding GPU UUID:

time=2025-10-15T04:28:22.873+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8\",\"GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced\",\"GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-10-15T04:28:22.910+02:00 level=INFO source=images.go:522 msg="total blobs: 145"
time=2025-10-15T04:28:22.927+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-10-15T04:28:22.942+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)"
time=2025-10-15T04:28:22.947+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..."
time=2025-10-15T04:28:23.636+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="96.8 GiB"
time=2025-10-15T04:28:23.636+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"
<!-- gh-comment-id:3404317185 --> @catclaw commented on GitHub (Oct 15, 2025): > My suspicion is it's from `CUDA_VISIBLE_DEVICES=0,1,2`. Try unsetting that first and see if it discovers the GPUs, or use the UUIDs instead which you can gather from `nvidia-smi -L` > > If that doesn't solve it, can you share the output of `nvidia-smi`? I added CUDA_VISIBLE_DEVICES=0,1,2 to System Variables, hoping Ollama would detect the GPUs after updating from 0.11.11 to 0.12.5, but it didn't work - and neither does using GPU UUIDs... Meanwhile, I don't have any problems creating three ComfyUI instances in SwarmUI, for example, to run 3 generations simultaneously. I've been up all night trying to figure out the problem, with no luck... 😵‍💫 ``` C:\Windows\System32>nvidia-smi Wed Oct 15 04:20:35 2025 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 581.57 Driver Version: 581.57 CUDA Version: 13.0 | +-----------------------------------------+------------------------+----------------------+ | GPU Name Driver-Model | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce RTX 5070 WDDM | 00000000:03:00.0 On | N/A | |100% 45C P0 31W / 250W | 3199MiB / 12227MiB | 7% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA GeForce RTX 3060 WDDM | 00000000:04:00.0 On | N/A | |100% 41C P0 38W / 170W | 909MiB / 12288MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 2 NVIDIA GeForce RTX 3060 WDDM | 00000000:83:00.0 On | N/A | |100% 42C P0 38W / 170W | 946MiB / 12288MiB | 3% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | 0 N/A N/A 3936 C+G ....0.3537.71\msedgewebview2.exe N/A | | 0 N/A N/A 5992 C+G ...cw5n1h2txyewy\WidgetBoard.exe N/A | | 0 N/A N/A 6952 C+G ...em32\ApplicationFrameHost.exe N/A | | 0 N/A N/A 12512 C+G ...XT CAM Beta\NZXT CAM Beta.exe N/A | | 0 N/A N/A 12736 C+G ...-dash\dash\bin\OculusDash.exe N/A | | 0 N/A N/A 14012 C+G ...Chrome\Application\chrome.exe N/A | | 0 N/A N/A 14520 C+G ....0.3537.71\msedgewebview2.exe N/A | | 0 N/A N/A 16656 C+G ...indows\System32\ShellHost.exe N/A | | 0 N/A N/A 16812 C+G ...lus-runtime\OVRServer_x64.exe N/A | | 0 N/A N/A 20436 C+G ...4__cv1g1gvanyjgm\WhatsApp.exe N/A | | 0 N/A N/A 20800 C+G C:\Windows\explorer.exe N/A | | 0 N/A N/A 21116 C+G ...e Beta\Application\chrome.exe N/A | | 0 N/A N/A 23596 C+G ...8bbwe\Microsoft.CmdPal.UI.exe N/A | | 0 N/A N/A 23632 C+G ..._cw5n1h2txyewy\SearchHost.exe N/A | | 0 N/A N/A 23808 C+G ...y\StartMenuExperienceHost.exe N/A | | 0 N/A N/A 24508 C+G ...5n1h2txyewy\TextInputHost.exe N/A | | 0 N/A N/A 25064 C+G ...crosoft shared\ink\TabTip.exe N/A | | 0 N/A N/A 28432 C+G ...8bbwe\PhoneExperienceHost.exe N/A | | 0 N/A N/A 28728 C+G ....0.3537.71\msedgewebview2.exe N/A | | 0 N/A N/A 30808 C+G ...s\Windows Sidebar\sidebar.exe N/A | | 0 N/A N/A 31180 C+G ...culus-client\OculusClient.exe N/A | | 0 N/A N/A 31248 C+G ...Chrome\Application\chrome.exe N/A | | 0 N/A N/A 31284 C+G ...yb3d8bbwe\Notepad\Notepad.exe N/A | | 0 N/A N/A 35440 C+G ...ef.win7x64\steamwebhelper.exe N/A | | 0 N/A N/A 37764 C+G ...ndows\System32\mmgaserver.exe N/A | | 0 N/A N/A 38496 C+G ...crosoft OneDrive\OneDrive.exe N/A | | 0 N/A N/A 41316 C+G ...er\Application\AVGBrowser.exe N/A | | 0 N/A N/A 43692 C+G ...x40ttqa\iCloud\iCloudHome.exe N/A | | 0 N/A N/A 46780 C+G ...ntrolPanel\SystemSettings.exe N/A | | 0 N/A N/A 51780 C+G ...xyewy\ShellExperienceHost.exe N/A | | 0 N/A N/A 53112 C+G ...2p2nqsd0c76g0\app\ChatGPT.exe N/A | | 0 N/A N/A 53732 C+G ...8wekyb3d8bbwe\WebViewHost.exe N/A | | 0 N/A N/A 53928 C+G ....0.3537.71\msedgewebview2.exe N/A | | 0 N/A N/A 54628 C+G ...lare WARP\Cloudflare WARP.exe N/A | | 0 N/A N/A 55140 C+G ....0.3537.71\msedgewebview2.exe N/A | | 0 N/A N/A 56636 C+G ...\Imaging Edge Desktop\ied.exe N/A | | 0 N/A N/A 58360 C+G ...ons\ScpToolkit\ScpTrayApp.exe N/A | | 1 N/A N/A 20800 C+G C:\Windows\explorer.exe N/A | | 1 N/A N/A 54628 C+G ...lare WARP\Cloudflare WARP.exe N/A | | 1 N/A N/A 56636 C+G ...\Imaging Edge Desktop\ied.exe N/A | | 1 N/A N/A 58360 C+G ...ons\ScpToolkit\ScpTrayApp.exe N/A | | 2 N/A N/A 20800 C+G C:\Windows\explorer.exe N/A | | 2 N/A N/A 31284 C+G ...yb3d8bbwe\Notepad\Notepad.exe N/A | | 2 N/A N/A 54628 C+G ...lare WARP\Cloudflare WARP.exe N/A | | 2 N/A N/A 56636 C+G ...\Imaging Edge Desktop\ied.exe N/A | | 2 N/A N/A 58360 C+G ...ons\ScpToolkit\ScpTrayApp.exe N/A | +-----------------------------------------------------------------------------------------+ C:\Windows\System32>nvidia-smi -L GPU 0: NVIDIA GeForce RTX 5070 (UUID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8) GPU 1: NVIDIA GeForce RTX 3060 (UUID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced) GPU 2: NVIDIA GeForce RTX 3060 (UUID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c)` ``` After adding GPU UUID: ``` time=2025-10-15T04:28:22.873+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8\",\"GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced\",\"GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-15T04:28:22.910+02:00 level=INFO source=images.go:522 msg="total blobs: 145" time=2025-10-15T04:28:22.927+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-15T04:28:22.942+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)" time=2025-10-15T04:28:22.947+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-15T04:28:23.636+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="96.8 GiB" time=2025-10-15T04:28:23.636+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ```
Author
Owner

@dhiltgen commented on GitHub (Oct 15, 2025):

Could you share the server log with OLLAMA_DEBUG="2" without CUDA_VISIBLE_DEVICES set? I'm hoping that may have a little more information on what's going wrong.

<!-- gh-comment-id:3407115117 --> @dhiltgen commented on GitHub (Oct 15, 2025): Could you share the server log with OLLAMA_DEBUG="2" without CUDA_VISIBLE_DEVICES set? I'm hoping that may have a little more information on what's going wrong.
Author
Owner

@catclaw commented on GitHub (Oct 15, 2025):

Could you share the server log with OLLAMA_DEBUG="2" without CUDA_VISIBLE_DEVICES set? I'm hoping that may have a little more information on what's going wrong.

Yes, of course! Thank you for taking the time to look into this.
I've uninstalled and reinstalled Ollama (without deleting all my models) - but it still falls back to using CPU instead of GPU.
(I've removed CUDA_VISIBLE_DEVICES and set OLLAMA_DEBUG=2)
Here is the output:

time=2025-10-16T01:32:50.208+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-10-16T01:32:50.248+02:00 level=INFO source=images.go:522 msg="total blobs: 145"
time=2025-10-16T01:32:50.268+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-10-16T01:32:50.284+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)"
time=2025-10-16T01:32:50.285+02:00 level=DEBUG source=sched.go:122 msg="starting llm scheduler"
time=2025-10-16T01:32:50.291+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..."
time=2025-10-16T01:32:50.291+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
time=2025-10-16T01:32:50.342+02:00 level=TRACE source=runner.go:491 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V10_2=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20800_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVCUDASAMPLES10_2_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVCUDASAMPLES_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVTOOLSEXT_PATH=C:\\Program Files\\NVIDIA Corporation\\NvToolsExt\\ NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_DEBUG=2 OLLAMA_MODELS=H:\\LLM OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52893"
time=2025-10-16T01:32:50.489+02:00 level=INFO source=runner.go:1316 msg="starting ollama engine"
time=2025-10-16T01:32:50.495+02:00 level=INFO source=runner.go:1352 msg="Server listening on 127.0.0.1:52893"
time=2025-10-16T01:32:50.503+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32
time=2025-10-16T01:32:50.504+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-16T01:32:50.506+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-16T01:32:50.507+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32
time=2025-10-16T01:32:50.509+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.file_type default=0
time=2025-10-16T01:32:50.510+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.name default=""
time=2025-10-16T01:32:50.511+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.description default=""
time=2025-10-16T01:32:50.514+02:00 level=INFO source=ggml.go:133 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-16T01:32:50.516+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-16T01:32:50.544+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-10-16T01:32:50.564+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-16T01:32:50.565+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0
time=2025-10-16T01:32:50.570+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-16T01:32:50.572+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-16T01:32:50.574+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.577+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.582+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.584+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.587+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-16T01:32:50.589+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-16T01:32:50.592+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-16T01:32:50.594+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-16T01:32:50.596+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.596+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-16T01:32:50.599+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0
time=2025-10-16T01:32:50.599+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-16T01:32:50.602+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-16T01:32:50.608+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-16T01:32:50.610+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-16T01:32:50.611+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-16T01:32:50.613+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-16T01:32:50.614+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-16T01:32:50.617+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-16T01:32:50.620+02:00 level=DEBUG source=runner.go:1291 msg="dummy model load took" duration=117.8269ms
time=2025-10-16T01:32:50.621+02:00 level=DEBUG source=runner.go:1296 msg="gathering device infos took" duration=0s
time=2025-10-16T01:32:50.624+02:00 level=TRACE source=runner.go:510 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[]
time=2025-10-16T01:32:50.627+02:00 level=DEBUG source=runner.go:414 msg="bootstrap discovery took" duration=335.0667ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
time=2025-10-16T01:32:50.632+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[]
time=2025-10-16T01:32:50.636+02:00 level=TRACE source=runner.go:491 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V10_2=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20800_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVCUDASAMPLES10_2_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVCUDASAMPLES_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVTOOLSEXT_PATH=C:\\Program Files\\NVIDIA Corporation\\NvToolsExt\\ NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_DEBUG=2 OLLAMA_MODELS=H:\\LLM OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52907"
time=2025-10-16T01:32:50.793+02:00 level=INFO source=runner.go:1316 msg="starting ollama engine"
time=2025-10-16T01:32:50.799+02:00 level=INFO source=runner.go:1352 msg="Server listening on 127.0.0.1:52907"
time=2025-10-16T01:32:50.809+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32
time=2025-10-16T01:32:50.811+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-16T01:32:50.812+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-16T01:32:50.813+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32
time=2025-10-16T01:32:50.814+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.file_type default=0
time=2025-10-16T01:32:50.815+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.name default=""
time=2025-10-16T01:32:50.817+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.description default=""
time=2025-10-16T01:32:50.820+02:00 level=INFO source=ggml.go:133 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-16T01:32:50.821+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-16T01:32:50.847+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm"
time=2025-10-16T01:32:50.886+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-16T01:32:50.888+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0
time=2025-10-16T01:32:50.893+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-16T01:32:50.895+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-16T01:32:50.897+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.900+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.907+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.909+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.912+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-16T01:32:50.918+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-16T01:32:50.921+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-16T01:32:50.924+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-16T01:32:50.927+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-16T01:32:50.932+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-16T01:32:50.935+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0
time=2025-10-16T01:32:50.941+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-16T01:32:50.942+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-16T01:32:50.946+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-16T01:32:50.952+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-16T01:32:50.954+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-16T01:32:50.957+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-16T01:32:50.963+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-16T01:32:50.966+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-16T01:32:50.969+02:00 level=DEBUG source=runner.go:1291 msg="dummy model load took" duration=161.0183ms
time=2025-10-16T01:32:50.973+02:00 level=DEBUG source=runner.go:1296 msg="gathering device infos took" duration=0s
time=2025-10-16T01:32:50.977+02:00 level=TRACE source=runner.go:510 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[]
time=2025-10-16T01:32:50.983+02:00 level=DEBUG source=runner.go:414 msg="bootstrap discovery took" duration=349.6206ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[]
time=2025-10-16T01:32:50.989+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[]
time=2025-10-16T01:32:50.996+02:00 level=TRACE source=runner.go:491 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V10_2=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20800_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVCUDASAMPLES10_2_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVCUDASAMPLES_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVTOOLSEXT_PATH=C:\\Program Files\\NVIDIA Corporation\\NvToolsExt\\ NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_DEBUG=2 OLLAMA_MODELS=H:\\LLM OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52922"
time=2025-10-16T01:32:51.195+02:00 level=INFO source=runner.go:1316 msg="starting ollama engine"
time=2025-10-16T01:32:51.202+02:00 level=INFO source=runner.go:1352 msg="Server listening on 127.0.0.1:52922"
time=2025-10-16T01:32:51.207+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32
time=2025-10-16T01:32:51.210+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-16T01:32:51.213+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-16T01:32:51.215+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32
time=2025-10-16T01:32:51.218+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.file_type default=0
time=2025-10-16T01:32:51.221+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.name default=""
time=2025-10-16T01:32:51.226+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.description default=""
time=2025-10-16T01:32:51.229+02:00 level=INFO source=ggml.go:133 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-16T01:32:51.231+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-16T01:32:51.259+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
time=2025-10-16T01:32:51.306+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-16T01:32:51.306+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0
time=2025-10-16T01:32:51.311+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-16T01:32:51.313+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-16T01:32:51.317+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-16T01:32:51.327+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-16T01:32:51.336+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-16T01:32:51.342+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-16T01:32:51.346+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-16T01:32:51.351+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-16T01:32:51.359+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-16T01:32:51.361+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-16T01:32:51.365+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-16T01:32:51.370+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-16T01:32:51.374+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0
time=2025-10-16T01:32:51.376+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-16T01:32:51.378+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-16T01:32:51.385+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-16T01:32:51.388+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-16T01:32:51.390+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-16T01:32:51.393+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-16T01:32:51.398+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-16T01:32:51.401+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-16T01:32:51.403+02:00 level=DEBUG source=runner.go:1291 msg="dummy model load took" duration=198.0063ms
time=2025-10-16T01:32:51.407+02:00 level=DEBUG source=runner.go:1296 msg="gathering device infos took" duration=0s
time=2025-10-16T01:32:51.410+02:00 level=TRACE source=runner.go:510 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[]
time=2025-10-16T01:32:51.412+02:00 level=DEBUG source=runner.go:414 msg="bootstrap discovery took" duration=419.2702ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[]
time=2025-10-16T01:32:51.416+02:00 level=DEBUG source=runner.go:117 msg="filtering out unsupported or overlapping GPU library combinations" count=0
time=2025-10-16T01:32:51.421+02:00 level=TRACE source=runner.go:164 msg="supported GPU library combinations" supported=map[]
time=2025-10-16T01:32:51.424+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=1.1385586s
time=2025-10-16T01:32:51.426+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="93.0 GiB"
time=2025-10-16T01:32:51.427+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"
<!-- gh-comment-id:3408643840 --> @catclaw commented on GitHub (Oct 15, 2025): > Could you share the server log with OLLAMA_DEBUG="2" without CUDA_VISIBLE_DEVICES set? I'm hoping that may have a little more information on what's going wrong. Yes, of course! Thank you for taking the time to look into this. I've uninstalled and reinstalled Ollama (without deleting all my models) - but it still falls back to using CPU instead of GPU. (I've removed CUDA_VISIBLE_DEVICES and set OLLAMA_DEBUG=2) Here is the output: ``` time=2025-10-16T01:32:50.208+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-16T01:32:50.248+02:00 level=INFO source=images.go:522 msg="total blobs: 145" time=2025-10-16T01:32:50.268+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-16T01:32:50.284+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)" time=2025-10-16T01:32:50.285+02:00 level=DEBUG source=sched.go:122 msg="starting llm scheduler" time=2025-10-16T01:32:50.291+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-16T01:32:50.291+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-16T01:32:50.342+02:00 level=TRACE source=runner.go:491 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V10_2=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20800_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVCUDASAMPLES10_2_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVCUDASAMPLES_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVTOOLSEXT_PATH=C:\\Program Files\\NVIDIA Corporation\\NvToolsExt\\ NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_DEBUG=2 OLLAMA_MODELS=H:\\LLM OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52893" time=2025-10-16T01:32:50.489+02:00 level=INFO source=runner.go:1316 msg="starting ollama engine" time=2025-10-16T01:32:50.495+02:00 level=INFO source=runner.go:1352 msg="Server listening on 127.0.0.1:52893" time=2025-10-16T01:32:50.503+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T01:32:50.504+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-16T01:32:50.506+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-16T01:32:50.507+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T01:32:50.509+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.file_type default=0 time=2025-10-16T01:32:50.510+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.name default="" time=2025-10-16T01:32:50.511+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.description default="" time=2025-10-16T01:32:50.514+02:00 level=INFO source=ggml.go:133 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-16T01:32:50.516+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-16T01:32:50.544+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T01:32:50.564+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-16T01:32:50.565+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0 time=2025-10-16T01:32:50.570+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-16T01:32:50.572+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-16T01:32:50.574+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-16T01:32:50.577+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-16T01:32:50.582+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-16T01:32:50.584+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-16T01:32:50.587+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-16T01:32:50.589+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-16T01:32:50.592+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-16T01:32:50.594+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-16T01:32:50.596+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-16T01:32:50.596+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-16T01:32:50.599+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0 time=2025-10-16T01:32:50.599+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-16T01:32:50.602+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-16T01:32:50.608+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-16T01:32:50.610+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-16T01:32:50.611+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-16T01:32:50.613+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-16T01:32:50.614+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-16T01:32:50.617+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-16T01:32:50.620+02:00 level=DEBUG source=runner.go:1291 msg="dummy model load took" duration=117.8269ms time=2025-10-16T01:32:50.621+02:00 level=DEBUG source=runner.go:1296 msg="gathering device infos took" duration=0s time=2025-10-16T01:32:50.624+02:00 level=TRACE source=runner.go:510 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[] time=2025-10-16T01:32:50.627+02:00 level=DEBUG source=runner.go:414 msg="bootstrap discovery took" duration=335.0667ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-16T01:32:50.632+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-16T01:32:50.636+02:00 level=TRACE source=runner.go:491 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V10_2=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20800_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVCUDASAMPLES10_2_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVCUDASAMPLES_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVTOOLSEXT_PATH=C:\\Program Files\\NVIDIA Corporation\\NvToolsExt\\ NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_DEBUG=2 OLLAMA_MODELS=H:\\LLM OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52907" time=2025-10-16T01:32:50.793+02:00 level=INFO source=runner.go:1316 msg="starting ollama engine" time=2025-10-16T01:32:50.799+02:00 level=INFO source=runner.go:1352 msg="Server listening on 127.0.0.1:52907" time=2025-10-16T01:32:50.809+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T01:32:50.811+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-16T01:32:50.812+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-16T01:32:50.813+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T01:32:50.814+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.file_type default=0 time=2025-10-16T01:32:50.815+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.name default="" time=2025-10-16T01:32:50.817+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.description default="" time=2025-10-16T01:32:50.820+02:00 level=INFO source=ggml.go:133 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-16T01:32:50.821+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-16T01:32:50.847+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-10-16T01:32:50.886+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-16T01:32:50.888+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0 time=2025-10-16T01:32:50.893+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-16T01:32:50.895+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-16T01:32:50.897+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-16T01:32:50.900+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-16T01:32:50.907+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-16T01:32:50.909+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-16T01:32:50.912+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-16T01:32:50.918+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-16T01:32:50.921+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-16T01:32:50.924+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-16T01:32:50.927+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-16T01:32:50.932+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-16T01:32:50.935+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0 time=2025-10-16T01:32:50.941+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-16T01:32:50.942+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-16T01:32:50.946+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-16T01:32:50.952+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-16T01:32:50.954+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-16T01:32:50.957+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-16T01:32:50.963+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-16T01:32:50.966+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-16T01:32:50.969+02:00 level=DEBUG source=runner.go:1291 msg="dummy model load took" duration=161.0183ms time=2025-10-16T01:32:50.973+02:00 level=DEBUG source=runner.go:1296 msg="gathering device infos took" duration=0s time=2025-10-16T01:32:50.977+02:00 level=TRACE source=runner.go:510 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-10-16T01:32:50.983+02:00 level=DEBUG source=runner.go:414 msg="bootstrap discovery took" duration=349.6206ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-16T01:32:50.989+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-16T01:32:50.996+02:00 level=TRACE source=runner.go:491 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V10_2=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20800_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVCUDASAMPLES10_2_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVCUDASAMPLES_ROOT=C:\\ProgramData\\NVIDIA Corporation\\CUDA Samples\\v10.2 NVTOOLSEXT_PATH=C:\\Program Files\\NVIDIA Corporation\\NvToolsExt\\ NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_DEBUG=2 OLLAMA_MODELS=H:\\LLM OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52922" time=2025-10-16T01:32:51.195+02:00 level=INFO source=runner.go:1316 msg="starting ollama engine" time=2025-10-16T01:32:51.202+02:00 level=INFO source=runner.go:1352 msg="Server listening on 127.0.0.1:52922" time=2025-10-16T01:32:51.207+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T01:32:51.210+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-16T01:32:51.213+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-16T01:32:51.215+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T01:32:51.218+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.file_type default=0 time=2025-10-16T01:32:51.221+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.name default="" time=2025-10-16T01:32:51.226+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=general.description default="" time=2025-10-16T01:32:51.229+02:00 level=INFO source=ggml.go:133 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-16T01:32:51.231+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-16T01:32:51.259+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-10-16T01:32:51.306+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-16T01:32:51.306+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0 time=2025-10-16T01:32:51.311+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-16T01:32:51.313+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-16T01:32:51.317+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-16T01:32:51.327+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-16T01:32:51.336+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-16T01:32:51.342+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-16T01:32:51.346+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-16T01:32:51.351+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-16T01:32:51.359+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-16T01:32:51.361+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-16T01:32:51.365+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-16T01:32:51.370+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-16T01:32:51.374+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.block_count default=0 time=2025-10-16T01:32:51.376+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-16T01:32:51.378+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-16T01:32:51.385+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-16T01:32:51.388+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-16T01:32:51.390+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-16T01:32:51.393+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-16T01:32:51.398+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-16T01:32:51.401+02:00 level=DEBUG source=ggml.go:275 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-16T01:32:51.403+02:00 level=DEBUG source=runner.go:1291 msg="dummy model load took" duration=198.0063ms time=2025-10-16T01:32:51.407+02:00 level=DEBUG source=runner.go:1296 msg="gathering device infos took" duration=0s time=2025-10-16T01:32:51.410+02:00 level=TRACE source=runner.go:510 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[] time=2025-10-16T01:32:51.412+02:00 level=DEBUG source=runner.go:414 msg="bootstrap discovery took" duration=419.2702ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-16T01:32:51.416+02:00 level=DEBUG source=runner.go:117 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-16T01:32:51.421+02:00 level=TRACE source=runner.go:164 msg="supported GPU library combinations" supported=map[] time=2025-10-16T01:32:51.424+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=1.1385586s time=2025-10-16T01:32:51.426+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="93.0 GiB" time=2025-10-16T01:32:51.427+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ```
Author
Owner

@catclaw commented on GitHub (Oct 16, 2025):

This seems to be a bug in the new 0.12.5 version.
I uninstalled Ollama 0.12.5 and installed version 0.11.11 - and it's detecting my GPUs just fine?

Debug is set to 1:

time=2025-10-16T03:11:23.895+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-16T03:11:23.933+02:00 level=INFO source=images.go:477 msg="total blobs: 145" time=2025-10-16T03:11:23.951+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-10-16T03:11:23.965+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" time=2025-10-16T03:11:23.966+02:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler" time=2025-10-16T03:11:23.966+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-10-16T03:11:23.967+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-10-16T03:11:23.968+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-10-16T03:11:23.968+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-10-16T03:11:23.969+02:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-10-16T03:11:23.971+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvml.dll time=2025-10-16T03:11:23.972+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvml.dll C:\\tools\\ruby34\\bin\\nvml.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvml.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvml.dll C:\\Python312\\Scripts\\nvml.dll C:\\Python312\\nvml.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvml.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvml.dll C:\\Windows\\system32\\nvml.dll C:\\Windows\\nvml.dll C:\\Windows\\System32\\Wbem\\nvml.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\Windows\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\WinMerge\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvml.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvml.dll C:\\ProgramData\\chocolatey\\bin\\nvml.dll H:\\Fooocus\\python_embeded\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvml.dll C:\\Program Files\\dotnet\\nvml.dll C:\\Program Files\\PuTTY\\nvml.dll C:\\Program Files\\Process Lasso\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvml.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvml.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvml.dll C:\\Program Files\\gsudo\\Current\\nvml.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvml.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll C:\\WINDOWS\\nvml.dll C:\\WINDOWS\\System32\\Wbem\\nvml.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\WINDOWS\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\CrowdSec\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvml.dll C:\\Program Files\\GitHub CLI\\nvml.dll C:\\Program Files\\Go\\bin\\nvml.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvml.dll C:\\Program Files\\nodejs\\nvml.dll C:\\Program Files\\CMake\\bin\\nvml.dll C:\\Program Files\\Calibre2\\nvml.dll C:\\Program Files\\PowerShell\\7\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvml.dll C:\\Program Files\\Git\\cmd\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T03:11:23.985+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll" time=2025-10-16T03:11:23.989+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T03:11:24.043+02:00 level=DEBUG source=gpu.go:111 msg="nvidia-ml loaded" library=C:\Windows\system32\nvml.dll time=2025-10-16T03:11:24.045+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvcuda.dll time=2025-10-16T03:11:24.048+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvcuda.dll C:\\tools\\ruby34\\bin\\nvcuda.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvcuda.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvcuda.dll C:\\Python312\\Scripts\\nvcuda.dll C:\\Python312\\nvcuda.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvcuda.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvcuda.dll C:\\Windows\\system32\\nvcuda.dll C:\\Windows\\nvcuda.dll C:\\Windows\\System32\\Wbem\\nvcuda.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\Windows\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\WinMerge\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvcuda.dll C:\\ProgramData\\chocolatey\\bin\\nvcuda.dll H:\\Fooocus\\python_embeded\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvcuda.dll C:\\Program Files\\dotnet\\nvcuda.dll C:\\Program Files\\PuTTY\\nvcuda.dll C:\\Program Files\\Process Lasso\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvcuda.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvcuda.dll C:\\Program Files\\gsudo\\Current\\nvcuda.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvcuda.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll C:\\WINDOWS\\nvcuda.dll C:\\WINDOWS\\System32\\Wbem\\nvcuda.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\WINDOWS\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\CrowdSec\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvcuda.dll C:\\Program Files\\GitHub CLI\\nvcuda.dll C:\\Program Files\\Go\\bin\\nvcuda.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvcuda.dll C:\\Program Files\\nodejs\\nvcuda.dll C:\\Program Files\\CMake\\bin\\nvcuda.dll C:\\Program Files\\Calibre2\\nvcuda.dll C:\\Program Files\\PowerShell\\7\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvcuda.dll C:\\Program Files\\Git\\cmd\\nvcuda.dll c:\\windows\\system*\\nvcuda.dll]" time=2025-10-16T03:11:24.064+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll" time=2025-10-16T03:11:24.069+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll]" initializing C:\Windows\system32\nvcuda.dll dlsym: cuInit - 00007FFDC058DE10 dlsym: cuDriverGetVersion - 00007FFDC058DEB0 dlsym: cuDeviceGetCount - 00007FFDC058E6A6 dlsym: cuDeviceGet - 00007FFDC058E6A0 dlsym: cuDeviceGetAttribute - 00007FFDC058E000 dlsym: cuDeviceGetUuid - 00007FFDC058E6B2 dlsym: cuDeviceGetName - 00007FFDC058E6AC dlsym: cuCtxCreate_v3 - 00007FFDC058E72A dlsym: cuMemGetInfo_v2 - 00007FFDC058E82C dlsym: cuCtxDestroy - 00007FFDC058E73C calling cuInit calling cuDriverGetVersion raw version 0x32c8 CUDA driver version: 13.0 calling cuDeviceGetCount device count 3 time=2025-10-16T03:11:24.133+02:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=3 library=C:\Windows\system32\nvcuda.dll [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA totalMem 12226mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA freeMem 10929mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] Compute Capability 12.0 [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA totalMem 12287mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA freeMem 11245mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] Compute Capability 8.6 time=2025-10-16T03:11:24.495+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="343.9 MiB" [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA totalMem 12287mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA freeMem 11245mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] Compute Capability 8.6 time=2025-10-16T03:11:24.669+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="274.2 MiB" time=2025-10-16T03:11:24.677+02:00 level=DEBUG source=amd_windows.go:34 msg="unable to load amdhip64_6.dll, please make sure to upgrade to the latest amd driver: The specified module could not be found." releasing cuda driver library releasing nvml library time=2025-10-16T03:11:24.684+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-10-16T03:11:24.686+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-10-16T03:11:24.691+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB"

<!-- gh-comment-id:3408793932 --> @catclaw commented on GitHub (Oct 16, 2025): This seems to be a bug in the new 0.12.5 version. I uninstalled Ollama 0.12.5 and installed version 0.11.11 - and it's detecting my GPUs just fine? Debug is set to 1: `time=2025-10-16T03:11:23.895+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-16T03:11:23.933+02:00 level=INFO source=images.go:477 msg="total blobs: 145" time=2025-10-16T03:11:23.951+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-10-16T03:11:23.965+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" time=2025-10-16T03:11:23.966+02:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler" time=2025-10-16T03:11:23.966+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-10-16T03:11:23.967+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-10-16T03:11:23.968+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-10-16T03:11:23.968+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-10-16T03:11:23.969+02:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-10-16T03:11:23.971+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvml.dll time=2025-10-16T03:11:23.972+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvml.dll C:\\tools\\ruby34\\bin\\nvml.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvml.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvml.dll C:\\Python312\\Scripts\\nvml.dll C:\\Python312\\nvml.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvml.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvml.dll C:\\Windows\\system32\\nvml.dll C:\\Windows\\nvml.dll C:\\Windows\\System32\\Wbem\\nvml.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\Windows\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\WinMerge\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvml.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvml.dll C:\\ProgramData\\chocolatey\\bin\\nvml.dll H:\\Fooocus\\python_embeded\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvml.dll C:\\Program Files\\dotnet\\nvml.dll C:\\Program Files\\PuTTY\\nvml.dll C:\\Program Files\\Process Lasso\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvml.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvml.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvml.dll C:\\Program Files\\gsudo\\Current\\nvml.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvml.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll C:\\WINDOWS\\nvml.dll C:\\WINDOWS\\System32\\Wbem\\nvml.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\WINDOWS\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\CrowdSec\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvml.dll C:\\Program Files\\GitHub CLI\\nvml.dll C:\\Program Files\\Go\\bin\\nvml.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvml.dll C:\\Program Files\\nodejs\\nvml.dll C:\\Program Files\\CMake\\bin\\nvml.dll C:\\Program Files\\Calibre2\\nvml.dll C:\\Program Files\\PowerShell\\7\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvml.dll C:\\Program Files\\Git\\cmd\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T03:11:23.985+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll" time=2025-10-16T03:11:23.989+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T03:11:24.043+02:00 level=DEBUG source=gpu.go:111 msg="nvidia-ml loaded" library=C:\Windows\system32\nvml.dll time=2025-10-16T03:11:24.045+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvcuda.dll time=2025-10-16T03:11:24.048+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvcuda.dll C:\\tools\\ruby34\\bin\\nvcuda.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvcuda.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvcuda.dll C:\\Python312\\Scripts\\nvcuda.dll C:\\Python312\\nvcuda.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvcuda.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvcuda.dll C:\\Windows\\system32\\nvcuda.dll C:\\Windows\\nvcuda.dll C:\\Windows\\System32\\Wbem\\nvcuda.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\Windows\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\WinMerge\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvcuda.dll C:\\ProgramData\\chocolatey\\bin\\nvcuda.dll H:\\Fooocus\\python_embeded\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvcuda.dll C:\\Program Files\\dotnet\\nvcuda.dll C:\\Program Files\\PuTTY\\nvcuda.dll C:\\Program Files\\Process Lasso\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvcuda.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvcuda.dll C:\\Program Files\\gsudo\\Current\\nvcuda.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvcuda.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll C:\\WINDOWS\\nvcuda.dll C:\\WINDOWS\\System32\\Wbem\\nvcuda.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\WINDOWS\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\CrowdSec\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvcuda.dll C:\\Program Files\\GitHub CLI\\nvcuda.dll C:\\Program Files\\Go\\bin\\nvcuda.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvcuda.dll C:\\Program Files\\nodejs\\nvcuda.dll C:\\Program Files\\CMake\\bin\\nvcuda.dll C:\\Program Files\\Calibre2\\nvcuda.dll C:\\Program Files\\PowerShell\\7\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvcuda.dll C:\\Program Files\\Git\\cmd\\nvcuda.dll c:\\windows\\system*\\nvcuda.dll]" time=2025-10-16T03:11:24.064+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll" time=2025-10-16T03:11:24.069+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll]" initializing C:\Windows\system32\nvcuda.dll dlsym: cuInit - 00007FFDC058DE10 dlsym: cuDriverGetVersion - 00007FFDC058DEB0 dlsym: cuDeviceGetCount - 00007FFDC058E6A6 dlsym: cuDeviceGet - 00007FFDC058E6A0 dlsym: cuDeviceGetAttribute - 00007FFDC058E000 dlsym: cuDeviceGetUuid - 00007FFDC058E6B2 dlsym: cuDeviceGetName - 00007FFDC058E6AC dlsym: cuCtxCreate_v3 - 00007FFDC058E72A dlsym: cuMemGetInfo_v2 - 00007FFDC058E82C dlsym: cuCtxDestroy - 00007FFDC058E73C calling cuInit calling cuDriverGetVersion raw version 0x32c8 CUDA driver version: 13.0 calling cuDeviceGetCount device count 3 time=2025-10-16T03:11:24.133+02:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=3 library=C:\Windows\system32\nvcuda.dll [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA totalMem 12226mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA freeMem 10929mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] Compute Capability 12.0 [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA totalMem 12287mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA freeMem 11245mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] Compute Capability 8.6 time=2025-10-16T03:11:24.495+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="343.9 MiB" [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA totalMem 12287mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA freeMem 11245mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] Compute Capability 8.6 time=2025-10-16T03:11:24.669+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="274.2 MiB" time=2025-10-16T03:11:24.677+02:00 level=DEBUG source=amd_windows.go:34 msg="unable to load amdhip64_6.dll, please make sure to upgrade to the latest amd driver: The specified module could not be found." releasing cuda driver library releasing nvml library time=2025-10-16T03:11:24.684+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-10-16T03:11:24.686+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-10-16T03:11:24.691+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB"`
Author
Owner

@catclaw commented on GitHub (Oct 16, 2025):

I can't figure out what the problem is?
I've been using 70B models in OpenWebUI using GPUs only (OLLAMA__SCHED_SPREAD:1) - but I'm back to square one!
Ollama keeps falling back to CPU no matter what model I try to use? 😵‍💫

time=2025-10-16T04:14:43.126+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-16T04:14:43.167+02:00 level=INFO source=images.go:477 msg="total blobs: 145" time=2025-10-16T04:14:43.186+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-10-16T04:14:43.202+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" time=2025-10-16T04:14:43.202+02:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler" time=2025-10-16T04:14:43.202+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-10-16T04:14:43.203+02:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-10-16T04:14:43.204+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvml.dll time=2025-10-16T04:14:43.205+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvml.dll C:\\tools\\ruby34\\bin\\nvml.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvml.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvml.dll C:\\Python312\\Scripts\\nvml.dll C:\\Python312\\nvml.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvml.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvml.dll C:\\Windows\\system32\\nvml.dll C:\\Windows\\nvml.dll C:\\Windows\\System32\\Wbem\\nvml.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\Windows\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\WinMerge\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvml.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvml.dll C:\\ProgramData\\chocolatey\\bin\\nvml.dll H:\\Fooocus\\python_embeded\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvml.dll C:\\Program Files\\dotnet\\nvml.dll C:\\Program Files\\PuTTY\\nvml.dll C:\\Program Files\\Process Lasso\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvml.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvml.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvml.dll C:\\Program Files\\gsudo\\Current\\nvml.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvml.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll C:\\WINDOWS\\nvml.dll C:\\WINDOWS\\System32\\Wbem\\nvml.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\WINDOWS\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\CrowdSec\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvml.dll C:\\Program Files\\GitHub CLI\\nvml.dll C:\\Program Files\\Go\\bin\\nvml.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvml.dll C:\\Program Files\\nodejs\\nvml.dll C:\\Program Files\\CMake\\bin\\nvml.dll C:\\Program Files\\Calibre2\\nvml.dll C:\\Program Files\\PowerShell\\7\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvml.dll C:\\Program Files\\Git\\cmd\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T04:14:43.219+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll" time=2025-10-16T04:14:43.223+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T04:14:43.279+02:00 level=DEBUG source=gpu.go:111 msg="nvidia-ml loaded" library=C:\Windows\system32\nvml.dll time=2025-10-16T04:14:43.280+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvcuda.dll time=2025-10-16T04:14:43.283+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvcuda.dll C:\\tools\\ruby34\\bin\\nvcuda.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvcuda.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvcuda.dll C:\\Python312\\Scripts\\nvcuda.dll C:\\Python312\\nvcuda.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvcuda.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvcuda.dll C:\\Windows\\system32\\nvcuda.dll C:\\Windows\\nvcuda.dll C:\\Windows\\System32\\Wbem\\nvcuda.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\Windows\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\WinMerge\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvcuda.dll C:\\ProgramData\\chocolatey\\bin\\nvcuda.dll H:\\Fooocus\\python_embeded\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvcuda.dll C:\\Program Files\\dotnet\\nvcuda.dll C:\\Program Files\\PuTTY\\nvcuda.dll C:\\Program Files\\Process Lasso\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvcuda.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvcuda.dll C:\\Program Files\\gsudo\\Current\\nvcuda.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvcuda.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll C:\\WINDOWS\\nvcuda.dll C:\\WINDOWS\\System32\\Wbem\\nvcuda.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\WINDOWS\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\CrowdSec\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvcuda.dll C:\\Program Files\\GitHub CLI\\nvcuda.dll C:\\Program Files\\Go\\bin\\nvcuda.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvcuda.dll C:\\Program Files\\nodejs\\nvcuda.dll C:\\Program Files\\CMake\\bin\\nvcuda.dll C:\\Program Files\\Calibre2\\nvcuda.dll C:\\Program Files\\PowerShell\\7\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvcuda.dll C:\\Program Files\\Git\\cmd\\nvcuda.dll c:\\windows\\system*\\nvcuda.dll]" time=2025-10-16T04:14:43.301+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll" time=2025-10-16T04:14:43.306+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll]" initializing C:\Windows\system32\nvcuda.dll dlsym: cuInit - 00007FFDC0B5DE10 dlsym: cuDriverGetVersion - 00007FFDC0B5DEB0 dlsym: cuDeviceGetCount - 00007FFDC0B5E6A6 dlsym: cuDeviceGet - 00007FFDC0B5E6A0 dlsym: cuDeviceGetAttribute - 00007FFDC0B5E000 dlsym: cuDeviceGetUuid - 00007FFDC0B5E6B2 dlsym: cuDeviceGetName - 00007FFDC0B5E6AC dlsym: cuCtxCreate_v3 - 00007FFDC0B5E72A dlsym: cuMemGetInfo_v2 - 00007FFDC0B5E82C dlsym: cuCtxDestroy - 00007FFDC0B5E73C calling cuInit calling cuDriverGetVersion raw version 0x32c8 CUDA driver version: 13.0 calling cuDeviceGetCount device count 3 time=2025-10-16T04:14:43.371+02:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=3 library=C:\Windows\system32\nvcuda.dll [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA totalMem 12226mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA freeMem 10929mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] Compute Capability 12.0 [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA totalMem 12287mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA freeMem 11245mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] Compute Capability 8.6 time=2025-10-16T04:14:43.745+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA totalMem 12287mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA freeMem 11245mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] Compute Capability 8.6 time=2025-10-16T04:14:43.916+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" time=2025-10-16T04:14:43.925+02:00 level=DEBUG source=amd_windows.go:34 msg="unable to load amdhip64_6.dll, please make sure to upgrade to the latest amd driver: The specified module could not be found." releasing cuda driver library releasing nvml library time=2025-10-16T04:14:43.933+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-10-16T04:14:43.935+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-10-16T04:14:43.936+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" [GIN] 2025/10/16 - 04:17:36 | 200 | 0s | 127.0.0.1 | HEAD "/" [GIN] 2025/10/16 - 04:17:36 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:19:06 | 200 | 31.3281ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:19:06 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:21:40 | 200 | 30.1901ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:21:40 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:21:43 | 200 | 32.5242ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:21:43 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:04 | 200 | 28.1053ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:22:05 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:53 | 200 | 29.7858ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:22:53 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:54 | 200 | 515.4µs | 127.0.0.1 | GET "/api/version" time=2025-10-16T04:34:17.695+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="90.5 GiB" before.free_swap="468.1 GiB" now.total="127.9 GiB" now.free="88.7 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:17.716+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="10.7 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:17.725+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:17.733+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:17.741+02:00 level=DEBUG source=sched.go:188 msg="updating default concurrency" OLLAMA_MAX_LOADED_MODELS=9 gpu_count=3 time=2025-10-16T04:34:17.770+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T04:34:17.781+02:00 level=DEBUG source=sched.go:208 msg="loading first model" model=H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 llama_model_loader: loaded meta data with 27 key-value pairs and 707 tensors from H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 32B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 32B llama_model_loader: - kv 5: qwen3.block_count u32 = 64 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 25600 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 64 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 257 tensors llama_model_loader: - type f16: 64 tensors llama_model_loader: - type q4_K: 353 tensors llama_model_loader: - type q6_K: 33 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.81 GiB (4.93 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 1 print_info: model type = ?B print_info: model params = 32.76 B print_info: general.name = Qwen3 32B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 llama_model_load: vocab only - skipping tensors time=2025-10-16T04:34:18.349+02:00 level=WARN source=server.go:171 msg="requested context size too large for model" num_ctx=131072 n_ctx_train=40960 time=2025-10-16T04:34:18.353+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="88.7 GiB" before.free_swap="463.8 GiB" now.total="127.9 GiB" now.free="88.6 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:18.372+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="8.8 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:18.379+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="10.9 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:18.387+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:18.447+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.451+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.462+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.464+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.469+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.473+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.476+02:00 level=INFO source=server.go:399 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --model H:\\LLM\\blobs\\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 --port 63675" time=2025-10-16T04:34:18.479+02:00 level=DEBUG source=server.go:400 msg=subprocess CUDA_PATH="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_PATH_V10_2="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2" CUDA_PATH_V12_8="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8" CUDA_PATH_V12_9="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9" CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=1 OLLAMA_MAX_LOADED_MODELS=9 OLLAMA_MODELS=H:\LLM OLLAMA_SCHED_SPREAD=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.526+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="88.6 GiB" before.free_swap="463.8 GiB" now.total="127.9 GiB" now.free="88.6 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:18.575+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="8.8 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:18.583+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="10.9 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:18.590+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:18.596+02:00 level=INFO source=server.go:504 msg="system memory" total="127.9 GiB" free="88.6 GiB" free_swap="463.8 GiB" time=2025-10-16T04:34:18.596+02:00 level=DEBUG source=memory.go:181 msg=evaluating library=cuda gpu_count=3 available="[11.0 GiB 10.9 GiB 8.8 GiB]" time=2025-10-16T04:34:18.601+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=qwen3.vision.block_count default=0 time=2025-10-16T04:34:18.606+02:00 level=DEBUG source=ggml.go:609 msg="default cache size estimate" "attention MiB"=10240 "attention bytes"=10737418240 "recurrent MiB"=0 "recurrent bytes"=0 time=2025-10-16T04:34:18.609+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.612+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="10.9 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.617+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="8.8 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.621+02:00 level=DEBUG source=memory.go:411 msg="insufficient VRAM to load any model layers" time=2025-10-16T04:34:18.625+02:00 level=DEBUG source=memory.go:181 msg=evaluating library=cuda gpu_count=3 available="[8.8 GiB 10.9 GiB 11.0 GiB]" time=2025-10-16T04:34:18.627+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=qwen3.vision.block_count default=0 time=2025-10-16T04:34:18.629+02:00 level=DEBUG source=ggml.go:609 msg="default cache size estimate" "attention MiB"=10240 "attention bytes"=10737418240 "recurrent MiB"=0 "recurrent bytes"=0 time=2025-10-16T04:34:18.631+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="8.8 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.634+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="10.9 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.641+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.645+02:00 level=DEBUG source=memory.go:411 msg="insufficient VRAM to load any model layers" time=2025-10-16T04:34:18.646+02:00 level=INFO source=server.go:544 msg=offload library=cuda layers.requested=-1 layers.model=65 layers.offload=0 layers.split=[] memory.available="[8.8 GiB 10.9 GiB 11.0 GiB]" memory.gpu_overhead="0 B" memory.required.full="28.4 GiB" memory.required.partial="0 B" memory.required.kv="10.0 GiB" memory.required.allocations="[0 B 0 B 0 B]" memory.weights.total="18.4 GiB" memory.weights.repeating="17.8 GiB" memory.weights.nonrepeating="608.6 MiB" memory.graph.full="13.3 GiB" memory.graph.partial="13.3 GiB" time=2025-10-16T04:34:18.693+02:00 level=INFO source=runner.go:864 msg="starting go runner" time=2025-10-16T04:34:18.694+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-16T04:34:18.750+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.766+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-16T04:34:18.769+02:00 level=INFO source=runner.go:900 msg="Server listening on 127.0.0.1:63675" time=2025-10-16T04:34:18.786+02:00 level=INFO source=runner.go:799 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:false KvSize:40960 KvCacheType: NumThreads:28 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-10-16T04:34:18.786+02:00 level=INFO source=server.go:1251 msg="waiting for llama runner to start responding" time=2025-10-16T04:34:18.794+02:00 level=INFO source=server.go:1285 msg="waiting for server to become available" status="llm server loading model" llama_model_loader: loaded meta data with 27 key-value pairs and 707 tensors from H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 32B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 32B llama_model_loader: - kv 5: qwen3.block_count u32 = 64 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 25600 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 64 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 257 tensors llama_model_loader: - type f16: 64 tensors llama_model_loader: - type q4_K: 353 tensors llama_model_loader: - type q6_K: 33 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.81 GiB (4.93 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 0 print_info: n_ctx_train = 40960 print_info: n_embd = 5120 print_info: n_layer = 64 print_info: n_head = 64 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: is_swa_any = 0 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 8 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-06 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 25600 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = -1 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 40960 print_info: rope_finetuned = unknown print_info: model type = 32B print_info: model params = 32.76 B print_info: general.name = Qwen3 32B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = false) load_tensors: layer 0 assigned to device CPU, is_swa = 0 load_tensors: layer 1 assigned to device CPU, is_swa = 0 load_tensors: layer 2 assigned to device CPU, is_swa = 0 load_tensors: layer 3 assigned to device CPU, is_swa = 0 load_tensors: layer 4 assigned to device CPU, is_swa = 0 load_tensors: layer 5 assigned to device CPU, is_swa = 0 load_tensors: layer 6 assigned to device CPU, is_swa = 0 load_tensors: layer 7 assigned to device CPU, is_swa = 0 load_tensors: layer 8 assigned to device CPU, is_swa = 0 load_tensors: layer 9 assigned to device CPU, is_swa = 0 load_tensors: layer 10 assigned to device CPU, is_swa = 0 load_tensors: layer 11 assigned to device CPU, is_swa = 0 load_tensors: layer 12 assigned to device CPU, is_swa = 0 load_tensors: layer 13 assigned to device CPU, is_swa = 0 load_tensors: layer 14 assigned to device CPU, is_swa = 0 load_tensors: layer 15 assigned to device CPU, is_swa = 0 load_tensors: layer 16 assigned to device CPU, is_swa = 0 load_tensors: layer 17 assigned to device CPU, is_swa = 0 load_tensors: layer 18 assigned to device CPU, is_swa = 0 load_tensors: layer 19 assigned to device CPU, is_swa = 0 load_tensors: layer 20 assigned to device CPU, is_swa = 0 load_tensors: layer 21 assigned to device CPU, is_swa = 0 load_tensors: layer 22 assigned to device CPU, is_swa = 0 load_tensors: layer 23 assigned to device CPU, is_swa = 0 load_tensors: layer 24 assigned to device CPU, is_swa = 0 load_tensors: layer 25 assigned to device CPU, is_swa = 0 load_tensors: layer 26 assigned to device CPU, is_swa = 0 load_tensors: layer 27 assigned to device CPU, is_swa = 0 load_tensors: layer 28 assigned to device CPU, is_swa = 0 load_tensors: layer 29 assigned to device CPU, is_swa = 0 load_tensors: layer 30 assigned to device CPU, is_swa = 0 load_tensors: layer 31 assigned to device CPU, is_swa = 0 load_tensors: layer 32 assigned to device CPU, is_swa = 0 load_tensors: layer 33 assigned to device CPU, is_swa = 0 load_tensors: layer 34 assigned to device CPU, is_swa = 0 load_tensors: layer 35 assigned to device CPU, is_swa = 0 load_tensors: layer 36 assigned to device CPU, is_swa = 0 load_tensors: layer 37 assigned to device CPU, is_swa = 0 load_tensors: layer 38 assigned to device CPU, is_swa = 0 load_tensors: layer 39 assigned to device CPU, is_swa = 0 load_tensors: layer 40 assigned to device CPU, is_swa = 0 load_tensors: layer 41 assigned to device CPU, is_swa = 0 load_tensors: layer 42 assigned to device CPU, is_swa = 0 load_tensors: layer 43 assigned to device CPU, is_swa = 0 load_tensors: layer 44 assigned to device CPU, is_swa = 0 load_tensors: layer 45 assigned to device CPU, is_swa = 0 load_tensors: layer 46 assigned to device CPU, is_swa = 0 load_tensors: layer 47 assigned to device CPU, is_swa = 0 load_tensors: layer 48 assigned to device CPU, is_swa = 0 load_tensors: layer 49 assigned to device CPU, is_swa = 0 load_tensors: layer 50 assigned to device CPU, is_swa = 0 load_tensors: layer 51 assigned to device CPU, is_swa = 0 load_tensors: layer 52 assigned to device CPU, is_swa = 0 load_tensors: layer 53 assigned to device CPU, is_swa = 0 load_tensors: layer 54 assigned to device CPU, is_swa = 0 load_tensors: layer 55 assigned to device CPU, is_swa = 0 load_tensors: layer 56 assigned to device CPU, is_swa = 0 load_tensors: layer 57 assigned to device CPU, is_swa = 0 load_tensors: layer 58 assigned to device CPU, is_swa = 0 load_tensors: layer 59 assigned to device CPU, is_swa = 0 load_tensors: layer 60 assigned to device CPU, is_swa = 0 load_tensors: layer 61 assigned to device CPU, is_swa = 0 load_tensors: layer 62 assigned to device CPU, is_swa = 0 load_tensors: layer 63 assigned to device CPU, is_swa = 0 load_tensors: layer 64 assigned to device CPU, is_swa = 0 load_tensors: CPU model buffer size = 19259.71 MiB load_all_data: no device found for buffer type CPU for async uploads time=2025-10-16T04:34:22.813+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.02" time=2025-10-16T04:34:27.840+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.05" time=2025-10-16T04:34:28.092+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.06" time=2025-10-16T04:34:28.847+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.06" time=2025-10-16T04:34:29.601+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.106+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.358+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.612+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:31.367+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:32.373+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:32.878+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:33.130+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:33.382+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:34.137+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:34.892+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.396+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.648+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.901+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:36.405+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.11" time=2025-10-16T04:34:37.161+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.11" time=2025-10-16T04:34:37.665+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:37.917+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:38.171+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:38.674+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:39.178+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:39.682+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:39.933+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:40.438+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.14" time=2025-10-16T04:34:40.942+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.14" time=2025-10-16T04:34:41.194+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.446+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.699+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.951+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:42.707+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.16" time=2025-10-16T04:34:42.959+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.16" time=2025-10-16T04:34:43.211+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:43.715+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:44.220+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:44.724+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:44.976+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:45.480+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:45.734+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:45.986+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:46.239+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:46.491+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:46.743+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:47.247+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:47.500+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:47.752+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:48.256+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:48.761+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.013+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.266+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.519+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.772+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.23" time=2025-10-16T04:34:50.276+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.23" time=2025-10-16T04:34:50.529+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:50.782+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:51.287+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:51.792+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.045+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.298+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.804+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.26" time=2025-10-16T04:34:53.309+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.26" time=2025-10-16T04:34:53.813+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.066+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.319+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.824+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.329+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.583+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.835+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.341+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.592+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.846+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:57.101+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:57.353+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:58.111+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:58.364+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:58.617+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:59.122+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:34:59.375+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:34:59.880+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:35:00.134+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:00.639+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:00.893+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:01.399+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:01.656+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:01.908+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:02.415+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.35" time=2025-10-16T04:35:02.921+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.35" time=2025-10-16T04:35:03.175+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:03.680+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:03.935+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:04.190+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.443+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.697+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.950+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:05.456+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:05.709+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:05.962+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:06.467+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.39" time=2025-10-16T04:35:06.972+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.39" time=2025-10-16T04:35:07.479+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:07.733+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:07.986+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:08.492+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:08.745+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:08.998+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:09.251+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:09.505+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.009+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.264+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.516+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:10.770+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:11.023+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:11.530+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.036+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.289+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.796+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.45" time=2025-10-16T04:35:13.049+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.45" time=2025-10-16T04:35:13.555+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:13.809+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:14.063+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:14.569+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 40960 llama_context: n_ctx_per_seq = 40960 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = 0 llama_context: kv_unified = false llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 set_abort_callback: call llama_context: CPU output buffer size = 0.60 MiB create_memory: n_ctx = 40960 (padded) llama_kv_cache_unified: layer 0: dev = CPU llama_kv_cache_unified: layer 1: dev = CPU llama_kv_cache_unified: layer 2: dev = CPU llama_kv_cache_unified: layer 3: dev = CPU llama_kv_cache_unified: layer 4: dev = CPU llama_kv_cache_unified: layer 5: dev = CPU llama_kv_cache_unified: layer 6: dev = CPU llama_kv_cache_unified: layer 7: dev = CPU llama_kv_cache_unified: layer 8: dev = CPU llama_kv_cache_unified: layer 9: dev = CPU llama_kv_cache_unified: layer 10: dev = CPU llama_kv_cache_unified: layer 11: dev = CPU llama_kv_cache_unified: layer 12: dev = CPU llama_kv_cache_unified: layer 13: dev = CPU llama_kv_cache_unified: layer 14: dev = CPU llama_kv_cache_unified: layer 15: dev = CPU llama_kv_cache_unified: layer 16: dev = CPU llama_kv_cache_unified: layer 17: dev = CPU llama_kv_cache_unified: layer 18: dev = CPU llama_kv_cache_unified: layer 19: dev = CPU llama_kv_cache_unified: layer 20: dev = CPU llama_kv_cache_unified: layer 21: dev = CPU llama_kv_cache_unified: layer 22: dev = CPU llama_kv_cache_unified: layer 23: dev = CPU llama_kv_cache_unified: layer 24: dev = CPU llama_kv_cache_unified: layer 25: dev = CPU llama_kv_cache_unified: layer 26: dev = CPU llama_kv_cache_unified: layer 27: dev = CPU llama_kv_cache_unified: layer 28: dev = CPU llama_kv_cache_unified: layer 29: dev = CPU llama_kv_cache_unified: layer 30: dev = CPU llama_kv_cache_unified: layer 31: dev = CPU llama_kv_cache_unified: layer 32: dev = CPU llama_kv_cache_unified: layer 33: dev = CPU llama_kv_cache_unified: layer 34: dev = CPU llama_kv_cache_unified: layer 35: dev = CPU llama_kv_cache_unified: layer 36: dev = CPU llama_kv_cache_unified: layer 37: dev = CPU llama_kv_cache_unified: layer 38: dev = CPU llama_kv_cache_unified: layer 39: dev = CPU llama_kv_cache_unified: layer 40: dev = CPU llama_kv_cache_unified: layer 41: dev = CPU llama_kv_cache_unified: layer 42: dev = CPU llama_kv_cache_unified: layer 43: dev = CPU llama_kv_cache_unified: layer 44: dev = CPU llama_kv_cache_unified: layer 45: dev = CPU llama_kv_cache_unified: layer 46: dev = CPU llama_kv_cache_unified: layer 47: dev = CPU llama_kv_cache_unified: layer 48: dev = CPU llama_kv_cache_unified: layer 49: dev = CPU llama_kv_cache_unified: layer 50: dev = CPU llama_kv_cache_unified: layer 51: dev = CPU llama_kv_cache_unified: layer 52: dev = CPU llama_kv_cache_unified: layer 53: dev = CPU llama_kv_cache_unified: layer 54: dev = CPU llama_kv_cache_unified: layer 55: dev = CPU llama_kv_cache_unified: layer 56: dev = CPU llama_kv_cache_unified: layer 57: dev = CPU llama_kv_cache_unified: layer 58: dev = CPU llama_kv_cache_unified: layer 59: dev = CPU llama_kv_cache_unified: layer 60: dev = CPU llama_kv_cache_unified: layer 61: dev = CPU llama_kv_cache_unified: layer 62: dev = CPU llama_kv_cache_unified: layer 63: dev = CPU llama_kv_cache_unified: CPU KV buffer size = 10240.00 MiB llama_kv_cache_unified: size = 10240.00 MiB ( 40960 cells, 64 layers, 1/1 seqs), K (f16): 5120.00 MiB, V (f16): 5120.00 MiB llama_context: enumerating backends llama_context: backend_ptrs.size() = 1 llama_context: max_nodes = 5656 llama_context: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 0 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 llama_context: CPU compute buffer size = 5256.01 MiB llama_context: graph nodes = 2502 llama_context: graph splits = 1 time=2025-10-16T04:36:51.998+02:00 level=INFO source=server.go:1289 msg="llama runner started in 153.52 seconds" time=2025-10-16T04:36:51.999+02:00 level=INFO source=sched.go:473 msg="loaded runners" count=1 time=2025-10-16T04:36:52.002+02:00 level=INFO source=server.go:1251 msg="waiting for llama runner to start responding" time=2025-10-16T04:36:52.002+02:00 level=INFO source=server.go:1289 msg="llama runner started in 153.53 seconds" time=2025-10-16T04:36:52.003+02:00 level=DEBUG source=sched.go:485 msg="finished setting up" runner.name=registry.ollama.ai/library/qwen3:32b runner.inference=cuda runner.devices=3 runner.size="28.4 GiB" runner.vram="0 B" runner.parallel=1 runner.pid=50652 runner.model=H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 runner.num_ctx=131072 time=2025-10-16T04:36:52.013+02:00 level=DEBUG source=server.go:1388 msg="completion request" images=0 prompt=4250 format="" time=2025-10-16T04:36:52.022+02:00 level=DEBUG source=cache.go:104 msg="loading cache slot" id=0 cache=0 prompt=891 used=0 remaining=891

<!-- gh-comment-id:3408975679 --> @catclaw commented on GitHub (Oct 16, 2025): I can't figure out what the problem is? I've been using 70B models in OpenWebUI using GPUs only (OLLAMA__SCHED_SPREAD:1) - but I'm back to square one! Ollama keeps falling back to CPU no matter what model I try to use? 😵‍💫 `time=2025-10-16T04:14:43.126+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-16T04:14:43.167+02:00 level=INFO source=images.go:477 msg="total blobs: 145" time=2025-10-16T04:14:43.186+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-10-16T04:14:43.202+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" time=2025-10-16T04:14:43.202+02:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler" time=2025-10-16T04:14:43.202+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-10-16T04:14:43.203+02:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-10-16T04:14:43.204+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvml.dll time=2025-10-16T04:14:43.205+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvml.dll C:\\tools\\ruby34\\bin\\nvml.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvml.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvml.dll C:\\Python312\\Scripts\\nvml.dll C:\\Python312\\nvml.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvml.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvml.dll C:\\Windows\\system32\\nvml.dll C:\\Windows\\nvml.dll C:\\Windows\\System32\\Wbem\\nvml.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\Windows\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\WinMerge\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvml.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvml.dll C:\\ProgramData\\chocolatey\\bin\\nvml.dll H:\\Fooocus\\python_embeded\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvml.dll C:\\Program Files\\dotnet\\nvml.dll C:\\Program Files\\PuTTY\\nvml.dll C:\\Program Files\\Process Lasso\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvml.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvml.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvml.dll C:\\Program Files\\gsudo\\Current\\nvml.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvml.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll C:\\WINDOWS\\nvml.dll C:\\WINDOWS\\System32\\Wbem\\nvml.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\WINDOWS\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\CrowdSec\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvml.dll C:\\Program Files\\GitHub CLI\\nvml.dll C:\\Program Files\\Go\\bin\\nvml.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvml.dll C:\\Program Files\\nodejs\\nvml.dll C:\\Program Files\\CMake\\bin\\nvml.dll C:\\Program Files\\Calibre2\\nvml.dll C:\\Program Files\\PowerShell\\7\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvml.dll C:\\Program Files\\Git\\cmd\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T04:14:43.219+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll" time=2025-10-16T04:14:43.223+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T04:14:43.279+02:00 level=DEBUG source=gpu.go:111 msg="nvidia-ml loaded" library=C:\Windows\system32\nvml.dll time=2025-10-16T04:14:43.280+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvcuda.dll time=2025-10-16T04:14:43.283+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvcuda.dll C:\\tools\\ruby34\\bin\\nvcuda.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvcuda.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvcuda.dll C:\\Python312\\Scripts\\nvcuda.dll C:\\Python312\\nvcuda.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvcuda.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvcuda.dll C:\\Windows\\system32\\nvcuda.dll C:\\Windows\\nvcuda.dll C:\\Windows\\System32\\Wbem\\nvcuda.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\Windows\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\WinMerge\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvcuda.dll C:\\ProgramData\\chocolatey\\bin\\nvcuda.dll H:\\Fooocus\\python_embeded\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvcuda.dll C:\\Program Files\\dotnet\\nvcuda.dll C:\\Program Files\\PuTTY\\nvcuda.dll C:\\Program Files\\Process Lasso\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvcuda.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvcuda.dll C:\\Program Files\\gsudo\\Current\\nvcuda.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvcuda.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll C:\\WINDOWS\\nvcuda.dll C:\\WINDOWS\\System32\\Wbem\\nvcuda.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\WINDOWS\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\CrowdSec\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvcuda.dll C:\\Program Files\\GitHub CLI\\nvcuda.dll C:\\Program Files\\Go\\bin\\nvcuda.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvcuda.dll C:\\Program Files\\nodejs\\nvcuda.dll C:\\Program Files\\CMake\\bin\\nvcuda.dll C:\\Program Files\\Calibre2\\nvcuda.dll C:\\Program Files\\PowerShell\\7\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvcuda.dll C:\\Program Files\\Git\\cmd\\nvcuda.dll c:\\windows\\system*\\nvcuda.dll]" time=2025-10-16T04:14:43.301+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll" time=2025-10-16T04:14:43.306+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll]" initializing C:\Windows\system32\nvcuda.dll dlsym: cuInit - 00007FFDC0B5DE10 dlsym: cuDriverGetVersion - 00007FFDC0B5DEB0 dlsym: cuDeviceGetCount - 00007FFDC0B5E6A6 dlsym: cuDeviceGet - 00007FFDC0B5E6A0 dlsym: cuDeviceGetAttribute - 00007FFDC0B5E000 dlsym: cuDeviceGetUuid - 00007FFDC0B5E6B2 dlsym: cuDeviceGetName - 00007FFDC0B5E6AC dlsym: cuCtxCreate_v3 - 00007FFDC0B5E72A dlsym: cuMemGetInfo_v2 - 00007FFDC0B5E82C dlsym: cuCtxDestroy - 00007FFDC0B5E73C calling cuInit calling cuDriverGetVersion raw version 0x32c8 CUDA driver version: 13.0 calling cuDeviceGetCount device count 3 time=2025-10-16T04:14:43.371+02:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=3 library=C:\Windows\system32\nvcuda.dll [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA totalMem 12226mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA freeMem 10929mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] Compute Capability 12.0 [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA totalMem 12287mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA freeMem 11245mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] Compute Capability 8.6 time=2025-10-16T04:14:43.745+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA totalMem 12287mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA freeMem 11245mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] Compute Capability 8.6 time=2025-10-16T04:14:43.916+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" time=2025-10-16T04:14:43.925+02:00 level=DEBUG source=amd_windows.go:34 msg="unable to load amdhip64_6.dll, please make sure to upgrade to the latest amd driver: The specified module could not be found." releasing cuda driver library releasing nvml library time=2025-10-16T04:14:43.933+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-10-16T04:14:43.935+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-10-16T04:14:43.936+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" [GIN] 2025/10/16 - 04:17:36 | 200 | 0s | 127.0.0.1 | HEAD "/" [GIN] 2025/10/16 - 04:17:36 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:19:06 | 200 | 31.3281ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:19:06 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:21:40 | 200 | 30.1901ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:21:40 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:21:43 | 200 | 32.5242ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:21:43 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:04 | 200 | 28.1053ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:22:05 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:53 | 200 | 29.7858ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:22:53 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:54 | 200 | 515.4µs | 127.0.0.1 | GET "/api/version" time=2025-10-16T04:34:17.695+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="90.5 GiB" before.free_swap="468.1 GiB" now.total="127.9 GiB" now.free="88.7 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:17.716+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="10.7 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:17.725+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:17.733+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:17.741+02:00 level=DEBUG source=sched.go:188 msg="updating default concurrency" OLLAMA_MAX_LOADED_MODELS=9 gpu_count=3 time=2025-10-16T04:34:17.770+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T04:34:17.781+02:00 level=DEBUG source=sched.go:208 msg="loading first model" model=H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 llama_model_loader: loaded meta data with 27 key-value pairs and 707 tensors from H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 32B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 32B llama_model_loader: - kv 5: qwen3.block_count u32 = 64 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 25600 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 64 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 257 tensors llama_model_loader: - type f16: 64 tensors llama_model_loader: - type q4_K: 353 tensors llama_model_loader: - type q6_K: 33 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.81 GiB (4.93 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 1 print_info: model type = ?B print_info: model params = 32.76 B print_info: general.name = Qwen3 32B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 llama_model_load: vocab only - skipping tensors time=2025-10-16T04:34:18.349+02:00 level=WARN source=server.go:171 msg="requested context size too large for model" num_ctx=131072 n_ctx_train=40960 time=2025-10-16T04:34:18.353+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="88.7 GiB" before.free_swap="463.8 GiB" now.total="127.9 GiB" now.free="88.6 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:18.372+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="8.8 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:18.379+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="10.9 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:18.387+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:18.447+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.451+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.462+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.464+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.469+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.473+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.476+02:00 level=INFO source=server.go:399 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --model H:\\LLM\\blobs\\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 --port 63675" time=2025-10-16T04:34:18.479+02:00 level=DEBUG source=server.go:400 msg=subprocess CUDA_PATH="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_PATH_V10_2="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2" CUDA_PATH_V12_8="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8" CUDA_PATH_V12_9="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9" CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=1 OLLAMA_MAX_LOADED_MODELS=9 OLLAMA_MODELS=H:\LLM OLLAMA_SCHED_SPREAD=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.526+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="88.6 GiB" before.free_swap="463.8 GiB" now.total="127.9 GiB" now.free="88.6 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:18.575+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="8.8 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:18.583+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="10.9 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:18.590+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:18.596+02:00 level=INFO source=server.go:504 msg="system memory" total="127.9 GiB" free="88.6 GiB" free_swap="463.8 GiB" time=2025-10-16T04:34:18.596+02:00 level=DEBUG source=memory.go:181 msg=evaluating library=cuda gpu_count=3 available="[11.0 GiB 10.9 GiB 8.8 GiB]" time=2025-10-16T04:34:18.601+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=qwen3.vision.block_count default=0 time=2025-10-16T04:34:18.606+02:00 level=DEBUG source=ggml.go:609 msg="default cache size estimate" "attention MiB"=10240 "attention bytes"=10737418240 "recurrent MiB"=0 "recurrent bytes"=0 time=2025-10-16T04:34:18.609+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.612+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="10.9 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.617+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="8.8 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.621+02:00 level=DEBUG source=memory.go:411 msg="insufficient VRAM to load any model layers" time=2025-10-16T04:34:18.625+02:00 level=DEBUG source=memory.go:181 msg=evaluating library=cuda gpu_count=3 available="[8.8 GiB 10.9 GiB 11.0 GiB]" time=2025-10-16T04:34:18.627+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=qwen3.vision.block_count default=0 time=2025-10-16T04:34:18.629+02:00 level=DEBUG source=ggml.go:609 msg="default cache size estimate" "attention MiB"=10240 "attention bytes"=10737418240 "recurrent MiB"=0 "recurrent bytes"=0 time=2025-10-16T04:34:18.631+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="8.8 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.634+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="10.9 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.641+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.645+02:00 level=DEBUG source=memory.go:411 msg="insufficient VRAM to load any model layers" time=2025-10-16T04:34:18.646+02:00 level=INFO source=server.go:544 msg=offload library=cuda layers.requested=-1 layers.model=65 layers.offload=0 layers.split=[] memory.available="[8.8 GiB 10.9 GiB 11.0 GiB]" memory.gpu_overhead="0 B" memory.required.full="28.4 GiB" memory.required.partial="0 B" memory.required.kv="10.0 GiB" memory.required.allocations="[0 B 0 B 0 B]" memory.weights.total="18.4 GiB" memory.weights.repeating="17.8 GiB" memory.weights.nonrepeating="608.6 MiB" memory.graph.full="13.3 GiB" memory.graph.partial="13.3 GiB" time=2025-10-16T04:34:18.693+02:00 level=INFO source=runner.go:864 msg="starting go runner" time=2025-10-16T04:34:18.694+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-16T04:34:18.750+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.766+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-16T04:34:18.769+02:00 level=INFO source=runner.go:900 msg="Server listening on 127.0.0.1:63675" time=2025-10-16T04:34:18.786+02:00 level=INFO source=runner.go:799 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:false KvSize:40960 KvCacheType: NumThreads:28 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-10-16T04:34:18.786+02:00 level=INFO source=server.go:1251 msg="waiting for llama runner to start responding" time=2025-10-16T04:34:18.794+02:00 level=INFO source=server.go:1285 msg="waiting for server to become available" status="llm server loading model" llama_model_loader: loaded meta data with 27 key-value pairs and 707 tensors from H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 32B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 32B llama_model_loader: - kv 5: qwen3.block_count u32 = 64 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 25600 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 64 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 257 tensors llama_model_loader: - type f16: 64 tensors llama_model_loader: - type q4_K: 353 tensors llama_model_loader: - type q6_K: 33 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.81 GiB (4.93 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 0 print_info: n_ctx_train = 40960 print_info: n_embd = 5120 print_info: n_layer = 64 print_info: n_head = 64 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: is_swa_any = 0 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 8 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-06 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 25600 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = -1 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 40960 print_info: rope_finetuned = unknown print_info: model type = 32B print_info: model params = 32.76 B print_info: general.name = Qwen3 32B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = false) load_tensors: layer 0 assigned to device CPU, is_swa = 0 load_tensors: layer 1 assigned to device CPU, is_swa = 0 load_tensors: layer 2 assigned to device CPU, is_swa = 0 load_tensors: layer 3 assigned to device CPU, is_swa = 0 load_tensors: layer 4 assigned to device CPU, is_swa = 0 load_tensors: layer 5 assigned to device CPU, is_swa = 0 load_tensors: layer 6 assigned to device CPU, is_swa = 0 load_tensors: layer 7 assigned to device CPU, is_swa = 0 load_tensors: layer 8 assigned to device CPU, is_swa = 0 load_tensors: layer 9 assigned to device CPU, is_swa = 0 load_tensors: layer 10 assigned to device CPU, is_swa = 0 load_tensors: layer 11 assigned to device CPU, is_swa = 0 load_tensors: layer 12 assigned to device CPU, is_swa = 0 load_tensors: layer 13 assigned to device CPU, is_swa = 0 load_tensors: layer 14 assigned to device CPU, is_swa = 0 load_tensors: layer 15 assigned to device CPU, is_swa = 0 load_tensors: layer 16 assigned to device CPU, is_swa = 0 load_tensors: layer 17 assigned to device CPU, is_swa = 0 load_tensors: layer 18 assigned to device CPU, is_swa = 0 load_tensors: layer 19 assigned to device CPU, is_swa = 0 load_tensors: layer 20 assigned to device CPU, is_swa = 0 load_tensors: layer 21 assigned to device CPU, is_swa = 0 load_tensors: layer 22 assigned to device CPU, is_swa = 0 load_tensors: layer 23 assigned to device CPU, is_swa = 0 load_tensors: layer 24 assigned to device CPU, is_swa = 0 load_tensors: layer 25 assigned to device CPU, is_swa = 0 load_tensors: layer 26 assigned to device CPU, is_swa = 0 load_tensors: layer 27 assigned to device CPU, is_swa = 0 load_tensors: layer 28 assigned to device CPU, is_swa = 0 load_tensors: layer 29 assigned to device CPU, is_swa = 0 load_tensors: layer 30 assigned to device CPU, is_swa = 0 load_tensors: layer 31 assigned to device CPU, is_swa = 0 load_tensors: layer 32 assigned to device CPU, is_swa = 0 load_tensors: layer 33 assigned to device CPU, is_swa = 0 load_tensors: layer 34 assigned to device CPU, is_swa = 0 load_tensors: layer 35 assigned to device CPU, is_swa = 0 load_tensors: layer 36 assigned to device CPU, is_swa = 0 load_tensors: layer 37 assigned to device CPU, is_swa = 0 load_tensors: layer 38 assigned to device CPU, is_swa = 0 load_tensors: layer 39 assigned to device CPU, is_swa = 0 load_tensors: layer 40 assigned to device CPU, is_swa = 0 load_tensors: layer 41 assigned to device CPU, is_swa = 0 load_tensors: layer 42 assigned to device CPU, is_swa = 0 load_tensors: layer 43 assigned to device CPU, is_swa = 0 load_tensors: layer 44 assigned to device CPU, is_swa = 0 load_tensors: layer 45 assigned to device CPU, is_swa = 0 load_tensors: layer 46 assigned to device CPU, is_swa = 0 load_tensors: layer 47 assigned to device CPU, is_swa = 0 load_tensors: layer 48 assigned to device CPU, is_swa = 0 load_tensors: layer 49 assigned to device CPU, is_swa = 0 load_tensors: layer 50 assigned to device CPU, is_swa = 0 load_tensors: layer 51 assigned to device CPU, is_swa = 0 load_tensors: layer 52 assigned to device CPU, is_swa = 0 load_tensors: layer 53 assigned to device CPU, is_swa = 0 load_tensors: layer 54 assigned to device CPU, is_swa = 0 load_tensors: layer 55 assigned to device CPU, is_swa = 0 load_tensors: layer 56 assigned to device CPU, is_swa = 0 load_tensors: layer 57 assigned to device CPU, is_swa = 0 load_tensors: layer 58 assigned to device CPU, is_swa = 0 load_tensors: layer 59 assigned to device CPU, is_swa = 0 load_tensors: layer 60 assigned to device CPU, is_swa = 0 load_tensors: layer 61 assigned to device CPU, is_swa = 0 load_tensors: layer 62 assigned to device CPU, is_swa = 0 load_tensors: layer 63 assigned to device CPU, is_swa = 0 load_tensors: layer 64 assigned to device CPU, is_swa = 0 load_tensors: CPU model buffer size = 19259.71 MiB load_all_data: no device found for buffer type CPU for async uploads time=2025-10-16T04:34:22.813+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.02" time=2025-10-16T04:34:27.840+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.05" time=2025-10-16T04:34:28.092+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.06" time=2025-10-16T04:34:28.847+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.06" time=2025-10-16T04:34:29.601+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.106+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.358+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.612+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:31.367+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:32.373+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:32.878+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:33.130+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:33.382+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:34.137+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:34.892+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.396+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.648+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.901+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:36.405+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.11" time=2025-10-16T04:34:37.161+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.11" time=2025-10-16T04:34:37.665+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:37.917+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:38.171+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:38.674+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:39.178+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:39.682+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:39.933+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:40.438+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.14" time=2025-10-16T04:34:40.942+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.14" time=2025-10-16T04:34:41.194+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.446+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.699+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.951+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:42.707+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.16" time=2025-10-16T04:34:42.959+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.16" time=2025-10-16T04:34:43.211+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:43.715+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:44.220+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:44.724+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:44.976+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:45.480+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:45.734+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:45.986+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:46.239+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:46.491+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:46.743+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:47.247+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:47.500+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:47.752+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:48.256+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:48.761+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.013+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.266+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.519+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.772+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.23" time=2025-10-16T04:34:50.276+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.23" time=2025-10-16T04:34:50.529+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:50.782+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:51.287+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:51.792+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.045+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.298+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.804+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.26" time=2025-10-16T04:34:53.309+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.26" time=2025-10-16T04:34:53.813+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.066+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.319+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.824+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.329+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.583+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.835+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.341+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.592+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.846+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:57.101+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:57.353+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:58.111+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:58.364+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:58.617+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:59.122+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:34:59.375+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:34:59.880+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:35:00.134+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:00.639+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:00.893+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:01.399+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:01.656+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:01.908+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:02.415+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.35" time=2025-10-16T04:35:02.921+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.35" time=2025-10-16T04:35:03.175+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:03.680+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:03.935+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:04.190+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.443+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.697+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.950+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:05.456+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:05.709+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:05.962+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:06.467+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.39" time=2025-10-16T04:35:06.972+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.39" time=2025-10-16T04:35:07.479+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:07.733+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:07.986+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:08.492+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:08.745+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:08.998+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:09.251+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:09.505+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.009+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.264+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.516+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:10.770+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:11.023+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:11.530+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.036+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.289+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.796+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.45" time=2025-10-16T04:35:13.049+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.45" time=2025-10-16T04:35:13.555+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:13.809+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:14.063+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:14.569+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 40960 llama_context: n_ctx_per_seq = 40960 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = 0 llama_context: kv_unified = false llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 set_abort_callback: call llama_context: CPU output buffer size = 0.60 MiB create_memory: n_ctx = 40960 (padded) llama_kv_cache_unified: layer 0: dev = CPU llama_kv_cache_unified: layer 1: dev = CPU llama_kv_cache_unified: layer 2: dev = CPU llama_kv_cache_unified: layer 3: dev = CPU llama_kv_cache_unified: layer 4: dev = CPU llama_kv_cache_unified: layer 5: dev = CPU llama_kv_cache_unified: layer 6: dev = CPU llama_kv_cache_unified: layer 7: dev = CPU llama_kv_cache_unified: layer 8: dev = CPU llama_kv_cache_unified: layer 9: dev = CPU llama_kv_cache_unified: layer 10: dev = CPU llama_kv_cache_unified: layer 11: dev = CPU llama_kv_cache_unified: layer 12: dev = CPU llama_kv_cache_unified: layer 13: dev = CPU llama_kv_cache_unified: layer 14: dev = CPU llama_kv_cache_unified: layer 15: dev = CPU llama_kv_cache_unified: layer 16: dev = CPU llama_kv_cache_unified: layer 17: dev = CPU llama_kv_cache_unified: layer 18: dev = CPU llama_kv_cache_unified: layer 19: dev = CPU llama_kv_cache_unified: layer 20: dev = CPU llama_kv_cache_unified: layer 21: dev = CPU llama_kv_cache_unified: layer 22: dev = CPU llama_kv_cache_unified: layer 23: dev = CPU llama_kv_cache_unified: layer 24: dev = CPU llama_kv_cache_unified: layer 25: dev = CPU llama_kv_cache_unified: layer 26: dev = CPU llama_kv_cache_unified: layer 27: dev = CPU llama_kv_cache_unified: layer 28: dev = CPU llama_kv_cache_unified: layer 29: dev = CPU llama_kv_cache_unified: layer 30: dev = CPU llama_kv_cache_unified: layer 31: dev = CPU llama_kv_cache_unified: layer 32: dev = CPU llama_kv_cache_unified: layer 33: dev = CPU llama_kv_cache_unified: layer 34: dev = CPU llama_kv_cache_unified: layer 35: dev = CPU llama_kv_cache_unified: layer 36: dev = CPU llama_kv_cache_unified: layer 37: dev = CPU llama_kv_cache_unified: layer 38: dev = CPU llama_kv_cache_unified: layer 39: dev = CPU llama_kv_cache_unified: layer 40: dev = CPU llama_kv_cache_unified: layer 41: dev = CPU llama_kv_cache_unified: layer 42: dev = CPU llama_kv_cache_unified: layer 43: dev = CPU llama_kv_cache_unified: layer 44: dev = CPU llama_kv_cache_unified: layer 45: dev = CPU llama_kv_cache_unified: layer 46: dev = CPU llama_kv_cache_unified: layer 47: dev = CPU llama_kv_cache_unified: layer 48: dev = CPU llama_kv_cache_unified: layer 49: dev = CPU llama_kv_cache_unified: layer 50: dev = CPU llama_kv_cache_unified: layer 51: dev = CPU llama_kv_cache_unified: layer 52: dev = CPU llama_kv_cache_unified: layer 53: dev = CPU llama_kv_cache_unified: layer 54: dev = CPU llama_kv_cache_unified: layer 55: dev = CPU llama_kv_cache_unified: layer 56: dev = CPU llama_kv_cache_unified: layer 57: dev = CPU llama_kv_cache_unified: layer 58: dev = CPU llama_kv_cache_unified: layer 59: dev = CPU llama_kv_cache_unified: layer 60: dev = CPU llama_kv_cache_unified: layer 61: dev = CPU llama_kv_cache_unified: layer 62: dev = CPU llama_kv_cache_unified: layer 63: dev = CPU llama_kv_cache_unified: CPU KV buffer size = 10240.00 MiB llama_kv_cache_unified: size = 10240.00 MiB ( 40960 cells, 64 layers, 1/1 seqs), K (f16): 5120.00 MiB, V (f16): 5120.00 MiB llama_context: enumerating backends llama_context: backend_ptrs.size() = 1 llama_context: max_nodes = 5656 llama_context: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 0 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 llama_context: CPU compute buffer size = 5256.01 MiB llama_context: graph nodes = 2502 llama_context: graph splits = 1 time=2025-10-16T04:36:51.998+02:00 level=INFO source=server.go:1289 msg="llama runner started in 153.52 seconds" time=2025-10-16T04:36:51.999+02:00 level=INFO source=sched.go:473 msg="loaded runners" count=1 time=2025-10-16T04:36:52.002+02:00 level=INFO source=server.go:1251 msg="waiting for llama runner to start responding" time=2025-10-16T04:36:52.002+02:00 level=INFO source=server.go:1289 msg="llama runner started in 153.53 seconds" time=2025-10-16T04:36:52.003+02:00 level=DEBUG source=sched.go:485 msg="finished setting up" runner.name=registry.ollama.ai/library/qwen3:32b runner.inference=cuda runner.devices=3 runner.size="28.4 GiB" runner.vram="0 B" runner.parallel=1 runner.pid=50652 runner.model=H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 runner.num_ctx=131072 time=2025-10-16T04:36:52.013+02:00 level=DEBUG source=server.go:1388 msg="completion request" images=0 prompt=4250 format="" time=2025-10-16T04:36:52.022+02:00 level=DEBUG source=cache.go:104 msg="loading cache slot" id=0 cache=0 prompt=891 used=0 remaining=891`
Author
Owner

@Digital-Monk commented on GitHub (Oct 16, 2025):

I'm just jumping on the bandwagon. Haven't been able to get GPU inference since "upgrading" to 12.5 from 12.3. I have uninstalled, purged, and reinstalled everything, and still no luck. I've tried with and without CUDA_VISIBLE_DEVICES. This happened right after I powered up this machine after removing one of my GPUs to put in a different server, so I have been assuming that something went sideways during that change. But I've seen a couple of other notes about 12.5 causing grief. Let me get my system back into an up-to-date software/driver state and then I'll grab logs and send them.

<!-- gh-comment-id:3412402082 --> @Digital-Monk commented on GitHub (Oct 16, 2025): I'm just jumping on the bandwagon. Haven't been able to get GPU inference since "upgrading" to 12.5 from 12.3. I have uninstalled, purged, and reinstalled everything, and still no luck. I've tried with and without CUDA_VISIBLE_DEVICES. This happened right after I powered up this machine after removing one of my GPUs to put in a different server, so I have been assuming that something went sideways during that change. But I've seen a couple of other notes about 12.5 causing grief. Let me get my system back into an up-to-date software/driver state and then I'll grab logs and send them.
Author
Owner

@Digital-Monk commented on GitHub (Oct 16, 2025):

serve.log

smi.log

<!-- gh-comment-id:3412521930 --> @Digital-Monk commented on GitHub (Oct 16, 2025): [serve.log](https://github.com/user-attachments/files/22957348/serve.log) [smi.log](https://github.com/user-attachments/files/22957358/smi.log)
Author
Owner

@dhiltgen commented on GitHub (Oct 16, 2025):

I haven't been able to reproduce on a Windows system with NVIDIA GPUs running the same driver version 581.57. The cuda_v13 library should be able to enumerate the GPUs, but for some reason it's failing silently. It seems like the libraries may not be getting loaded at all. Your log shows

time=2025-10-16T01:32:50.291+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
...
time=2025-10-16T01:32:50.516+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-16T01:32:50.544+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-10-16T01:32:50.564+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)

What I see on my test system is:

time=2025-10-16T12:01:58.042-07:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
...
time=2025-10-16T12:01:58.068-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama
load_backend: loaded CPU backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll
time=2025-10-16T12:01:58.081-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
  Device 0: Quadro RTX 5000, compute capability 7.5, VMM: yes, ID: GPU-87228d0f-4cdc-31b2-6fae-658d3c36bd2f
load_backend: loaded CUDA backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll
time=2025-10-16T12:01:58.168-07:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)

So it's not loading the CPU or GPU libraries properly.

Can you share what files you see in C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama and C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13

Maybe there's a dependency problem, or an AV system getting in the way?

On the dependency front, perhaps try simplifying your PATH down and see if that makes any difference? Something like

$env:PATH="C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\"
$env:OLLAMA_DEBUG="2"
ollama serve 2>&1 | % ToString | tee-object serve.log

You could also try to install the latest MSVC C++ runtime, although I don't think that should have any impact.
https://aka.ms/vs/17/release/vc_redist.x64.exe

<!-- gh-comment-id:3412524281 --> @dhiltgen commented on GitHub (Oct 16, 2025): I haven't been able to reproduce on a Windows system with NVIDIA GPUs running the same driver version 581.57. The cuda_v13 library should be able to enumerate the GPUs, but for some reason it's failing silently. It seems like the libraries may not be getting loaded at all. Your log shows ``` time=2025-10-16T01:32:50.291+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] ... time=2025-10-16T01:32:50.516+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-16T01:32:50.544+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T01:32:50.564+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) ``` What I see on my test system is: ``` time=2025-10-16T12:01:58.042-07:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] ... time=2025-10-16T12:01:58.068-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama load_backend: loaded CPU backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll time=2025-10-16T12:01:58.081-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: Quadro RTX 5000, compute capability 7.5, VMM: yes, ID: GPU-87228d0f-4cdc-31b2-6fae-658d3c36bd2f load_backend: loaded CUDA backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll time=2025-10-16T12:01:58.168-07:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) ``` So it's not loading the CPU or GPU libraries properly. Can you share what files you see in `C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama` and `C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13` Maybe there's a dependency problem, or an AV system getting in the way? On the dependency front, perhaps try simplifying your PATH down and see if that makes any difference? Something like ```powershell $env:PATH="C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\" $env:OLLAMA_DEBUG="2" ollama serve 2>&1 | % ToString | tee-object serve.log ``` You could also try to install the latest MSVC C++ runtime, although I don't think that should have any impact. https://aka.ms/vs/17/release/vc_redist.x64.exe
Author
Owner

@dhiltgen commented on GitHub (Oct 16, 2025):

@Digital-Monk what type of GPU do you have?

The error in your logs says:

time=2025-10-16T13:47:36.676-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v12
ggml_cuda_init: failed to initialize CUDA: forward compatibility was attempted on non supported HW
load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v12/libggml-cuda.so

Oops, I see now you included the SMI output too NVIDIA GeForce GTX 960

This should be an easy fix.

<!-- gh-comment-id:3412533012 --> @dhiltgen commented on GitHub (Oct 16, 2025): @Digital-Monk what type of GPU do you have? The error in your logs says: ``` time=2025-10-16T13:47:36.676-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v12 ggml_cuda_init: failed to initialize CUDA: forward compatibility was attempted on non supported HW load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v12/libggml-cuda.so ``` Oops, I see now you included the SMI output too `NVIDIA GeForce GTX 960` This should be an easy fix.
Author
Owner

@Digital-Monk commented on GitHub (Oct 16, 2025):

Yeah, ancient hardware. Still faster than the CPU, though :)

<!-- gh-comment-id:3412548419 --> @Digital-Monk commented on GitHub (Oct 16, 2025): Yeah, ancient hardware. Still faster than the CPU, though :)
Author
Owner

@dhiltgen commented on GitHub (Oct 16, 2025):

@Digital-Monk #12666 should get this working again. Sorry about that.

<!-- gh-comment-id:3412558903 --> @dhiltgen commented on GitHub (Oct 16, 2025): @Digital-Monk #12666 should get this working again. Sorry about that.
Author
Owner

@Digital-Monk commented on GitHub (Oct 17, 2025):

0.12.6 fixed it for me. Thank you!

<!-- gh-comment-id:3415539966 --> @Digital-Monk commented on GitHub (Oct 17, 2025): 0.12.6 fixed it for me. Thank you!
Author
Owner

@catclaw commented on GitHub (Oct 17, 2025):

I haven't been able to reproduce on a Windows system with NVIDIA GPUs running the same driver version 581.57. The cuda_v13 library should be able to enumerate the GPUs, but for some reason it's failing silently. It seems like the libraries may not be getting loaded at all. Your log shows

time=2025-10-16T01:32:50.291+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
...
time=2025-10-16T01:32:50.516+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-16T01:32:50.544+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-10-16T01:32:50.564+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)

What I see on my test system is:

time=2025-10-16T12:01:58.042-07:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
...
time=2025-10-16T12:01:58.068-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama
load_backend: loaded CPU backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll
time=2025-10-16T12:01:58.081-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
  Device 0: Quadro RTX 5000, compute capability 7.5, VMM: yes, ID: GPU-87228d0f-4cdc-31b2-6fae-658d3c36bd2f
load_backend: loaded CUDA backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll
time=2025-10-16T12:01:58.168-07:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)

So it's not loading the CPU or GPU libraries properly.

Can you share what files you see in C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama and C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13

Maybe there's a dependency problem, or an AV system getting in the way?

On the dependency front, perhaps try simplifying your PATH down and see if that makes any difference? Something like

$env:PATH="C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama"
$env:OLLAMA_DEBUG="2"
ollama serve 2>&1 | % ToString | tee-object serve.log
You could also try to install the latest MSVC C++ runtime, although I don't think that should have any impact. https://aka.ms/vs/17/release/vc_redist.x64.exe

Ollama still refuses to load even the smallest models to the GPU, and falls back to CPU, no matter what I do...
I've tried pretty much everything.
I'm going to try removing all Nvidia drivers and reinstalling them from scratch, hoping it will fix the problem.

I tried:
$env:PATH="C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama"
$env:OLLAMA_DEBUG="2"
ollama serve 2>&1 | % ToString | tee-object serve.log

Didn't help either - but I'll upload the log file.

 Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama

2025-10-16  03:09    <DIR>          .
2025-10-16  03:10    <DIR>          ..
2025-10-16  03:09    <DIR>          cuda_v12
2025-10-16  03:09    <DIR>          cuda_v13
2025-09-15  17:23           853 496 ggml-base.dll
2025-09-15  17:23         1 011 192 ggml-cpu-alderlake.dll
2025-09-15  17:23         1 014 264 ggml-cpu-haswell.dll
2025-09-15  17:23         1 227 256 ggml-cpu-icelake.dll
2025-09-15  17:23           909 816 ggml-cpu-sandybridge.dll
2025-09-15  17:23         1 223 672 ggml-cpu-skylakex.dll
2025-09-15  17:23           744 440 ggml-cpu-sse42.dll
2025-09-15  17:23           737 784 ggml-cpu-x64.dll
2025-09-15  17:23       550 436 856 ggml-hip.dll
2025-10-16  03:09    <DIR>          rocm
               9 File(s)    558 158 776 bytes
               5 Dir(s)  1 018 273 652 736 bytes free
 Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13

2025-10-16  03:09    <DIR>          .
2025-10-16  03:09    <DIR>          ..
2025-09-15  17:23        50 286 072 cublas64_13.dll
2025-09-15  17:23       477 894 648 cublasLt64_13.dll
2025-09-15  17:23       259 710 968 ggml-cuda.dll
               3 File(s)    787 891 688 bytes
               2 Dir(s)  1 018 268 430 336 bytes free

serve2.log

<!-- gh-comment-id:3416428288 --> @catclaw commented on GitHub (Oct 17, 2025): > I haven't been able to reproduce on a Windows system with NVIDIA GPUs running the same driver version 581.57. The cuda_v13 library should be able to enumerate the GPUs, but for some reason it's failing silently. It seems like the libraries may not be getting loaded at all. Your log shows > > ``` > time=2025-10-16T01:32:50.291+02:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] > ... > time=2025-10-16T01:32:50.516+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" > time=2025-10-16T01:32:50.544+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" > time=2025-10-16T01:32:50.564+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) > ``` > > What I see on my test system is: > > ``` > time=2025-10-16T12:01:58.042-07:00 level=DEBUG source=runner.go:411 msg="spawing runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\daniel\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] > ... > time=2025-10-16T12:01:58.068-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama > load_backend: loaded CPU backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll > time=2025-10-16T12:01:58.081-07:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13 > ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no > ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no > ggml_cuda_init: found 1 CUDA devices: > Device 0: Quadro RTX 5000, compute capability 7.5, VMM: yes, ID: GPU-87228d0f-4cdc-31b2-6fae-658d3c36bd2f > load_backend: loaded CUDA backend from C:\Users\daniel\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll > time=2025-10-16T12:01:58.168-07:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) > ``` > > So it's not loading the CPU or GPU libraries properly. > > Can you share what files you see in `C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama` and `C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13` > > Maybe there's a dependency problem, or an AV system getting in the way? > > On the dependency front, perhaps try simplifying your PATH down and see if that makes any difference? Something like > > $env:PATH="C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\" > $env:OLLAMA_DEBUG="2" > ollama serve 2>&1 | % ToString | tee-object serve.log > You could also try to install the latest MSVC C++ runtime, although I don't think that should have any impact. https://aka.ms/vs/17/release/vc_redist.x64.exe Ollama still refuses to load even the smallest models to the GPU, and falls back to CPU, no matter what I do... I've tried pretty much everything. I'm going to try removing all Nvidia drivers and reinstalling them from scratch, hoping it will fix the problem. I tried: $env:PATH="C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\" $env:OLLAMA_DEBUG="2" ollama serve 2>&1 | % ToString | tee-object serve.log Didn't help either - but I'll upload the log file. ``` Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama 2025-10-16 03:09 <DIR> . 2025-10-16 03:10 <DIR> .. 2025-10-16 03:09 <DIR> cuda_v12 2025-10-16 03:09 <DIR> cuda_v13 2025-09-15 17:23 853 496 ggml-base.dll 2025-09-15 17:23 1 011 192 ggml-cpu-alderlake.dll 2025-09-15 17:23 1 014 264 ggml-cpu-haswell.dll 2025-09-15 17:23 1 227 256 ggml-cpu-icelake.dll 2025-09-15 17:23 909 816 ggml-cpu-sandybridge.dll 2025-09-15 17:23 1 223 672 ggml-cpu-skylakex.dll 2025-09-15 17:23 744 440 ggml-cpu-sse42.dll 2025-09-15 17:23 737 784 ggml-cpu-x64.dll 2025-09-15 17:23 550 436 856 ggml-hip.dll 2025-10-16 03:09 <DIR> rocm 9 File(s) 558 158 776 bytes 5 Dir(s) 1 018 273 652 736 bytes free ``` ``` Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13 2025-10-16 03:09 <DIR> . 2025-10-16 03:09 <DIR> .. 2025-09-15 17:23 50 286 072 cublas64_13.dll 2025-09-15 17:23 477 894 648 cublasLt64_13.dll 2025-09-15 17:23 259 710 968 ggml-cuda.dll 3 File(s) 787 891 688 bytes 2 Dir(s) 1 018 268 430 336 bytes free ``` [serve2.log](https://github.com/user-attachments/files/22976589/serve2.log)
Author
Owner

@dhiltgen commented on GitHub (Oct 17, 2025):

@catclaw the server2.log you attached appears to be from version 0.11.11 not 0.12.6

time=2025-10-17T18:55:33.248+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)"
<!-- gh-comment-id:3416986447 --> @dhiltgen commented on GitHub (Oct 17, 2025): @catclaw the server2.log you attached appears to be from version 0.11.11 not 0.12.6 ``` time=2025-10-17T18:55:33.248+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" ```
Author
Owner

@catclaw commented on GitHub (Oct 17, 2025):

@catclaw the server2.log you attached appears to be from version 0.11.11 not 0.12.6

time=2025-10-17T18:55:33.248+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)"

Yes, because the latest version refuses to detect my GPUs at all...
This is the output 0.12.6:

time=2025-10-17T22:23:33.303+02:00 level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]"
time=2025-10-17T22:23:33.338+02:00 level=INFO source=images.go:522 msg="total blobs: 146"
time=2025-10-17T22:23:33.355+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-10-17T22:23:33.368+02:00 level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)"
time=2025-10-17T22:23:33.368+02:00 level=DEBUG source=sched.go:123 msg="starting llm scheduler"
time=2025-10-17T22:23:33.375+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..."
time=2025-10-17T22:23:33.375+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[]
time=2025-10-17T22:23:33.408+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62929"
time=2025-10-17T22:23:33.543+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-17T22:23:33.549+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62929"
time=2025-10-17T22:23:33.562+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.569+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-17T22:23:33.571+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-17T22:23:33.579+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.582+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-17T22:23:33.588+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-17T22:23:33.589+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-17T22:23:33.591+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-17T22:23:33.591+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-17T22:23:33.619+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
time=2025-10-17T22:23:33.674+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-17T22:23:33.675+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.684+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.685+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-17T22:23:33.691+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-17T22:23:33.692+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-17T22:23:33.696+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.700+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-17T22:23:33.702+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-17T22:23:33.703+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-17T22:23:33.707+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-17T22:23:33.708+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-17T22:23:33.714+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-17T22:23:33.716+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=160.457ms
time=2025-10-17T22:23:33.717+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s
time=2025-10-17T22:23:33.720+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[]
time=2025-10-17T22:23:33.723+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=348.2686ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[]
time=2025-10-17T22:23:33.724+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
time=2025-10-17T22:23:33.730+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62943"
time=2025-10-17T22:23:33.878+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-17T22:23:33.884+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62943"
time=2025-10-17T22:23:33.890+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.892+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-17T22:23:33.893+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-17T22:23:33.897+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-17T22:23:33.898+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-17T22:23:33.901+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-17T22:23:33.905+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-17T22:23:33.932+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-10-17T22:23:33.953+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-17T22:23:33.954+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.959+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.963+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.964+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.967+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.969+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-17T22:23:33.970+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-17T22:23:33.976+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.980+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-17T22:23:33.984+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-17T22:23:33.985+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-17T22:23:33.990+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-17T22:23:33.996+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=108.2169ms
time=2025-10-17T22:23:33.999+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s
time=2025-10-17T22:23:34.003+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[]
time=2025-10-17T22:23:34.008+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=280.104ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
time=2025-10-17T22:23:34.015+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[]
time=2025-10-17T22:23:34.019+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 50590"
time=2025-10-17T22:23:34.186+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-17T22:23:34.192+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:50590"
time=2025-10-17T22:23:34.195+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:34.199+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-17T22:23:34.202+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-17T22:23:34.203+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:34.205+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-17T22:23:34.206+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-17T22:23:34.208+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-17T22:23:34.209+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-17T22:23:34.211+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-17T22:23:34.239+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm"
time=2025-10-17T22:23:34.280+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-17T22:23:34.281+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:34.284+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-17T22:23:34.285+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-17T22:23:34.288+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.289+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.291+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.292+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.296+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-17T22:23:34.298+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-17T22:23:34.299+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-17T22:23:34.300+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-17T22:23:34.301+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.303+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-17T22:23:34.304+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:34.308+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-17T22:23:34.310+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-17T22:23:34.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-17T22:23:34.314+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-17T22:23:34.317+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-17T22:23:34.320+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-17T22:23:34.321+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-17T22:23:34.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=130.8744ms
time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s
time=2025-10-17T22:23:34.328+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[]
time=2025-10-17T22:23:34.331+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=315.5108ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[]
time=2025-10-17T22:23:34.334+02:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0
time=2025-10-17T22:23:34.337+02:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[]
time=2025-10-17T22:23:34.339+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=969.0867ms
time=2025-10-17T22:23:34.342+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="95.5 GiB"
time=2025-10-17T22:23:34.343+02:00 level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"
<!-- gh-comment-id:3417010511 --> @catclaw commented on GitHub (Oct 17, 2025): > [@catclaw](https://github.com/catclaw) the server2.log you attached appears to be from version 0.11.11 not 0.12.6 > > ``` > time=2025-10-17T18:55:33.248+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" > ``` Yes, because the latest version refuses to detect my GPUs at all... This is the output 0.12.6: ``` time=2025-10-17T22:23:33.303+02:00 level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-17T22:23:33.338+02:00 level=INFO source=images.go:522 msg="total blobs: 146" time=2025-10-17T22:23:33.355+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-17T22:23:33.368+02:00 level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-17T22:23:33.368+02:00 level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-17T22:23:33.375+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-17T22:23:33.375+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-17T22:23:33.408+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62929" time=2025-10-17T22:23:33.543+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-17T22:23:33.549+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62929" time=2025-10-17T22:23:33.562+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.569+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-17T22:23:33.571+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-17T22:23:33.579+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.582+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-17T22:23:33.588+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-17T22:23:33.589+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-17T22:23:33.591+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-17T22:23:33.591+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-17T22:23:33.619+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-10-17T22:23:33.674+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-17T22:23:33.675+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-17T22:23:33.684+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-17T22:23:33.685+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-17T22:23:33.691+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-17T22:23:33.692+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-17T22:23:33.696+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.700+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-17T22:23:33.702+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-17T22:23:33.703+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-17T22:23:33.707+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-17T22:23:33.708+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-17T22:23:33.714+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-17T22:23:33.716+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=160.457ms time=2025-10-17T22:23:33.717+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-17T22:23:33.720+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[] time=2025-10-17T22:23:33.723+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=348.2686ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-17T22:23:33.724+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-17T22:23:33.730+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62943" time=2025-10-17T22:23:33.878+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-17T22:23:33.884+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62943" time=2025-10-17T22:23:33.890+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.892+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-17T22:23:33.893+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-17T22:23:33.897+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-17T22:23:33.898+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-17T22:23:33.901+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-17T22:23:33.905+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-17T22:23:33.932+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-17T22:23:33.953+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-17T22:23:33.954+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.959+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-17T22:23:33.963+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-17T22:23:33.964+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-17T22:23:33.967+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-17T22:23:33.969+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-17T22:23:33.970+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-17T22:23:33.976+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-17T22:23:33.980+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-17T22:23:33.984+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-17T22:23:33.985+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-17T22:23:33.990+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-17T22:23:33.996+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=108.2169ms time=2025-10-17T22:23:33.999+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-17T22:23:34.003+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[] time=2025-10-17T22:23:34.008+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=280.104ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-17T22:23:34.015+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-17T22:23:34.019+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 50590" time=2025-10-17T22:23:34.186+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-17T22:23:34.192+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:50590" time=2025-10-17T22:23:34.195+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:34.199+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-17T22:23:34.202+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-17T22:23:34.203+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:34.205+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-17T22:23:34.206+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-17T22:23:34.208+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-17T22:23:34.209+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-17T22:23:34.211+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-17T22:23:34.239+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-10-17T22:23:34.280+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-17T22:23:34.281+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:34.284+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-17T22:23:34.285+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-17T22:23:34.288+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-17T22:23:34.289+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-17T22:23:34.291+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-17T22:23:34.292+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-17T22:23:34.296+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-17T22:23:34.298+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-17T22:23:34.299+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-17T22:23:34.300+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-17T22:23:34.301+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-17T22:23:34.303+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-17T22:23:34.304+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:34.308+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-17T22:23:34.310+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-17T22:23:34.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-17T22:23:34.314+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-17T22:23:34.317+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-17T22:23:34.320+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-17T22:23:34.321+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-17T22:23:34.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=130.8744ms time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-17T22:23:34.328+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-10-17T22:23:34.331+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=315.5108ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-17T22:23:34.334+02:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-17T22:23:34.337+02:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-17T22:23:34.339+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=969.0867ms time=2025-10-17T22:23:34.342+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="95.5 GiB" time=2025-10-17T22:23:34.343+02:00 level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ```
Author
Owner

@catclaw commented on GitHub (Oct 17, 2025):

I've used Nvidia CleanUpTool to remove everything and reinstalled CUDA 12.8, then 12.9, and 13.0.
I tried running Ollama (both version 0.11.11 and 0.12.6) between the installations - but;
0.11.11 detects my GPUs - but falls back to using CPU, no matter how small the model is.
0.12.6 doesn't detect any of my GPUs.

Here is the output from 0.12.6:

time=2025-10-17T22:23:33.303+02:00 level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]"
time=2025-10-17T22:23:33.338+02:00 level=INFO source=images.go:522 msg="total blobs: 146"
time=2025-10-17T22:23:33.355+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-10-17T22:23:33.368+02:00 level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)"
time=2025-10-17T22:23:33.368+02:00 level=DEBUG source=sched.go:123 msg="starting llm scheduler"
time=2025-10-17T22:23:33.375+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..."
time=2025-10-17T22:23:33.375+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[]
time=2025-10-17T22:23:33.408+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62929"
time=2025-10-17T22:23:33.543+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-17T22:23:33.549+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62929"
time=2025-10-17T22:23:33.562+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.569+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-17T22:23:33.571+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-17T22:23:33.579+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.582+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-17T22:23:33.588+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-17T22:23:33.589+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-17T22:23:33.591+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-17T22:23:33.591+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-17T22:23:33.619+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
time=2025-10-17T22:23:33.674+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-17T22:23:33.675+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.684+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.685+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-17T22:23:33.691+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-17T22:23:33.692+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-17T22:23:33.696+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.700+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-17T22:23:33.702+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-17T22:23:33.703+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-17T22:23:33.707+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-17T22:23:33.708+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-17T22:23:33.714+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-17T22:23:33.716+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=160.457ms
time=2025-10-17T22:23:33.717+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s
time=2025-10-17T22:23:33.720+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[]
time=2025-10-17T22:23:33.723+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=348.2686ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[]
time=2025-10-17T22:23:33.724+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
time=2025-10-17T22:23:33.730+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62943"
time=2025-10-17T22:23:33.878+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-17T22:23:33.884+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62943"
time=2025-10-17T22:23:33.890+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.892+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-17T22:23:33.893+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-17T22:23:33.897+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-17T22:23:33.898+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-17T22:23:33.901+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-17T22:23:33.905+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-17T22:23:33.932+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-10-17T22:23:33.953+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-17T22:23:33.954+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.959+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.963+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.964+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.967+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.969+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-17T22:23:33.970+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-17T22:23:33.976+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-17T22:23:33.980+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-17T22:23:33.984+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-17T22:23:33.985+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-17T22:23:33.990+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-17T22:23:33.996+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=108.2169ms
time=2025-10-17T22:23:33.999+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s
time=2025-10-17T22:23:34.003+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[]
time=2025-10-17T22:23:34.008+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=280.104ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[]
time=2025-10-17T22:23:34.015+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[]
time=2025-10-17T22:23:34.019+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 50590"
time=2025-10-17T22:23:34.186+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-17T22:23:34.192+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:50590"
time=2025-10-17T22:23:34.195+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:34.199+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-17T22:23:34.202+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-17T22:23:34.203+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-17T22:23:34.205+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-17T22:23:34.206+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-17T22:23:34.208+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-17T22:23:34.209+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-17T22:23:34.211+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-10-17T22:23:34.239+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm"
time=2025-10-17T22:23:34.280+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-10-17T22:23:34.281+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:34.284+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-17T22:23:34.285+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-17T22:23:34.288+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.289+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.291+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.292+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.296+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-17T22:23:34.298+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-17T22:23:34.299+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-17T22:23:34.300+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-17T22:23:34.301+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-17T22:23:34.303+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-17T22:23:34.304+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-17T22:23:34.308+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-17T22:23:34.310+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-17T22:23:34.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-17T22:23:34.314+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-17T22:23:34.317+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-17T22:23:34.320+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-17T22:23:34.321+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-17T22:23:34.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=130.8744ms
time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s
time=2025-10-17T22:23:34.328+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[]
time=2025-10-17T22:23:34.331+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=315.5108ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[]
time=2025-10-17T22:23:34.334+02:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0
time=2025-10-17T22:23:34.337+02:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[]
time=2025-10-17T22:23:34.339+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=969.0867ms
time=2025-10-17T22:23:34.342+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="95.5 GiB"
time=2025-10-17T22:23:34.343+02:00 level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"
<!-- gh-comment-id:3417042876 --> @catclaw commented on GitHub (Oct 17, 2025): I've used Nvidia CleanUpTool to remove everything and reinstalled CUDA 12.8, then 12.9, and 13.0. I tried running Ollama (both version 0.11.11 and 0.12.6) between the installations - but; 0.11.11 detects my GPUs - but falls back to using CPU, no matter how small the model is. 0.12.6 doesn't detect any of my GPUs. Here is the output from 0.12.6: ``` time=2025-10-17T22:23:33.303+02:00 level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-17T22:23:33.338+02:00 level=INFO source=images.go:522 msg="total blobs: 146" time=2025-10-17T22:23:33.355+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-17T22:23:33.368+02:00 level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-17T22:23:33.368+02:00 level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-17T22:23:33.375+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-17T22:23:33.375+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-17T22:23:33.408+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62929" time=2025-10-17T22:23:33.543+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-17T22:23:33.549+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62929" time=2025-10-17T22:23:33.562+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.569+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-17T22:23:33.571+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-17T22:23:33.579+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.582+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-17T22:23:33.588+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-17T22:23:33.589+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-17T22:23:33.591+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-17T22:23:33.591+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-17T22:23:33.619+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-10-17T22:23:33.674+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-17T22:23:33.675+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-17T22:23:33.679+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-17T22:23:33.682+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-17T22:23:33.684+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-17T22:23:33.685+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-17T22:23:33.689+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-17T22:23:33.691+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-17T22:23:33.692+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-17T22:23:33.694+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-17T22:23:33.696+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.700+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-17T22:23:33.702+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-17T22:23:33.703+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-17T22:23:33.705+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-17T22:23:33.707+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-17T22:23:33.708+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-17T22:23:33.714+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-17T22:23:33.716+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=160.457ms time=2025-10-17T22:23:33.717+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-17T22:23:33.720+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[] time=2025-10-17T22:23:33.723+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=348.2686ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-17T22:23:33.724+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-17T22:23:33.730+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62943" time=2025-10-17T22:23:33.878+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-17T22:23:33.884+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62943" time=2025-10-17T22:23:33.890+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.892+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-17T22:23:33.893+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:33.895+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-17T22:23:33.897+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-17T22:23:33.898+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-17T22:23:33.901+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-17T22:23:33.905+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-17T22:23:33.932+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-17T22:23:33.953+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-17T22:23:33.954+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.959+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-17T22:23:33.962+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-17T22:23:33.963+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-17T22:23:33.964+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-17T22:23:33.967+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-17T22:23:33.969+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-17T22:23:33.970+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-17T22:23:33.972+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-17T22:23:33.976+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-17T22:23:33.980+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:33.983+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-17T22:23:33.984+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-17T22:23:33.985+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-17T22:23:33.987+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-17T22:23:33.990+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-17T22:23:33.996+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-17T22:23:33.997+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=108.2169ms time=2025-10-17T22:23:33.999+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-17T22:23:34.003+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[] time=2025-10-17T22:23:34.008+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=280.104ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-17T22:23:34.015+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-17T22:23:34.019+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_20448_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 50590" time=2025-10-17T22:23:34.186+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-17T22:23:34.192+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:50590" time=2025-10-17T22:23:34.195+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:34.199+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-17T22:23:34.202+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-17T22:23:34.203+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-17T22:23:34.205+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-17T22:23:34.206+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-17T22:23:34.208+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-17T22:23:34.209+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-17T22:23:34.211+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-17T22:23:34.239+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-10-17T22:23:34.280+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-17T22:23:34.281+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:34.284+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-17T22:23:34.285+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-17T22:23:34.288+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-17T22:23:34.289+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-17T22:23:34.291+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-17T22:23:34.292+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-17T22:23:34.296+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-17T22:23:34.298+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-17T22:23:34.299+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-17T22:23:34.300+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-17T22:23:34.301+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-17T22:23:34.303+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-17T22:23:34.304+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-17T22:23:34.308+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-17T22:23:34.310+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-17T22:23:34.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-17T22:23:34.314+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-17T22:23:34.317+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-17T22:23:34.320+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-17T22:23:34.321+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-17T22:23:34.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=130.8744ms time=2025-10-17T22:23:34.325+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-17T22:23:34.328+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-10-17T22:23:34.331+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=315.5108ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-17T22:23:34.334+02:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-17T22:23:34.337+02:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-17T22:23:34.339+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=969.0867ms time=2025-10-17T22:23:34.342+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="95.5 GiB" time=2025-10-17T22:23:34.343+02:00 level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ```
Author
Owner

@dhiltgen commented on GitHub (Oct 17, 2025):

@catclaw the directory listings you shared don't seem to match the file sizes I see for 0.12.6 - perhaps that was from an older version?

I'm still not sure what's going wrong, but for some reason, the ggml-*.dll libraries aren't loading. I'll try to add some more debug logging in a future version which might help troubleshoot. Until then, things to potentially try:

  • uninstall/reinstall Ollama 0.12.6
  • In file explorer, get "properties" on any of these ggml-*.dll files and verify they're signed by Ollama.
  • Check your AV software to see if it is blocking anything
<!-- gh-comment-id:3417334501 --> @dhiltgen commented on GitHub (Oct 17, 2025): @catclaw the directory listings you shared don't seem to match the file sizes I see for 0.12.6 - perhaps that was from an older version? I'm still not sure what's going wrong, but for some reason, the `ggml-*.dll` libraries aren't loading. I'll try to add some more debug logging in a future version which might help troubleshoot. Until then, things to potentially try: - uninstall/reinstall Ollama 0.12.6 - In file explorer, get "properties" on any of these ggml-*.dll files and verify they're signed by Ollama. - Check your AV software to see if it is blocking anything
Author
Owner

@aminalshmu commented on GitHub (Oct 18, 2025):

Have been experiencing this same issue on Gentoo since around 0.12.5, this is from the just compiled git version:

$ ollama serve
time=2025-10-18T10:08:21.757-05:00 level=INFO source=routes.go:1515 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/home/XXX/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"
time=2025-10-18T10:08:21.758-05:00 level=INFO source=routes.go:1568 msg="Listening on 127.0.0.1:11434 (version 0.12.6-6-gbc1a818-dirty)"
time=2025-10-18T10:08:21.758-05:00 level=DEBUG source=sched.go:125 msg="starting llm scheduler"
time=2025-10-18T10:08:21.758-05:00 level=INFO source=runner.go:80 msg="discovering available GPUs..."
time=2025-10-18T10:08:21.758-05:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs=[]
time=2025-10-18T10:08:21.758-05:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash SESSION_MANAGER=local/HOSTNAME:@/tmp/.ICE-unix/25034,unix/HOSTNAME:/tmp/.ICE-unix/25034 COLORTERM=truecolor XDG_CONFIG_DIRS=/etc/xdg LESS=-R -M --shift 5 XDG_MENU_PREFIX=gnome- JDK_HOME=/etc/java-config-2/current-system-vm sparky=00:1c:25:e1:60:61 CONFIG_PROTECT_MASK=/etc/sandbox.d /etc/fonts/fonts.conf /etc/gentoo-release /etc/terminfo /etc/dconf /etc/ca-certificates.conf /etc/revdep-rebuild PKG_CONFIG_PATH=/opt/cuda/pkgconfig TERMCAP=SC|screen.xterm-256color|VT 100/ANSI X3.64 virtual terminal:DO=\\E[%dB:LE=\\E[%dD:RI=\\E[%dC:UP=\\E[%dA:bs:bt=\\E[Z:cd=\\E[J:ce=\\E[K:cl=\\E[H\\E[J:cm=\\E[%i%d;%dH:ct=\\E[3g:do=^J:nd=\\E[C:pt:rc=\\E8:rs=\\Ec:sc=\\E7:st=\\EH:up=\\EM:le=^H:bl=^G:cr=^M:it#8:ho=\\E[H:nw=\\EE:ta=^I:is=\\E)0:li#42:co#139:am:xn:xv:LP:sr=\\EM:al=\\E[L:AL=\\E[%dL:cs=\\E[%i%d;%dr:dl=\\E[M:DL=\\E[%dM:dc=\\E[P:DC=\\E[%dP:im=\\E[4h:ei=\\E[4l:mi:IC=\\E[%d@:ks=\\E[?1h\\E=:ke=\\E[?1l\\E>:vi=\\E[?25l:ve=\\E[34h\\E[?25h:vs=\\E[34l:ti=\\E[?1049h:te=\\E[?1049l:us=\\E[4m:ue=\\E[24m:so=\\E[3m:se=\\E[23m:mb=\\E[5m:md=\\E[1m:mh=\\E[2m:mr=\\E[7m:me=\\E[m:ms:Co#8:pa#64:AF=\\E[3%dm:AB=\\E[4%dm:op=\\E[39;49m:AX:vb=\\Eg:G0:as=\\E(0:ae=\\E(B:ac=\\140\\140aaffggjjkkllmmnnooppqqrrssttuuvvwwxxyyzz{{||}}~~..--++,,hhII00:po=\\E[5i:pf=\\E[4i:Z0=\\E[?3h:Z1=\\E[?3l:Km=\\E[<:k0=\\E[10~:k1=\\EOP:k2=\\EOQ:k3=\\EOR:k4=\\EOS:k5=\\E[15~:k6=\\E[17~:k7=\\E[18~:k8=\\E[19~:k9=\\E[20~:k;=\\E[21~:F1=\\E[23~:F2=\\E[24~:kB=\\E[Z:kh=\\E[1~:@1=\\E[1~:kH=\\E[4~:@7=\\E[4~:kN=\\E[6~:kP=\\E[5~:kI=\\E[2~:kD=\\E[3~:ku=\\EOA:kd=\\EOB:kr=\\EOC:kl=\\EOD:km: JAVA_HOME=/etc/java-config-2/current-system-vm WINDOW=0 XMODIFIERS=@im=ibus DESKTOP_SESSION=gnome EDITOR=/bin/nano GTK_MODULES=canberra-gtk-module PWD=/home/XXX CONFIG_PROTECT=/usr/share/gnupg/qualified.txt /usr/share/config XDG_SESSION_DESKTOP=gnome LOGNAME=XXX XDG_SESSION_TYPE=x11 MANPATH=/etc/java-config-2/current-system-vm/man:/usr/share/gcc-data/x86_64-pc-linux-gnu/15/man:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/man:/etc/java-config-2/current-system-vm/man/:/usr/local/share/man:/usr/share/man:/usr/lib/rust/man-bin-1.86.0:/usr/lib/rust/man-bin-1.88.0:/usr/lib/rust/man-bin-1.89.0:/usr/lib/rust/man-bin-1.90.0:/usr/lib/llvm/21/share/man:/usr/lib/llvm/20/share/man:/usr/lib/llvm/17/share/man:/etc/eselect/wine/share/man XAUTHORITY=/run/user/1000/gdm/Xauthority WINDOWPATH=8 MOTD_SHOWN=pam GDM_LANG=en_US.utf8 HOME=/home/XXX USERNAME=XXX LANG=en_US.utf8 LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.7z=01;31:*.ace=01;31:*.alz=01;31:*.apk=01;31:*.arc=01;31:*.arj=01;31:*.bz=01;31:*.bz2=01;31:*.cab=01;31:*.cpio=01;31:*.crate=01;31:*.deb=01;31:*.drpm=01;31:*.dwm=01;31:*.dz=01;31:*.ear=01;31:*.egg=01;31:*.esd=01;31:*.gz=01;31:*.jar=01;31:*.lha=01;31:*.lrz=01;31:*.lz=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.lzo=01;31:*.pyz=01;31:*.rar=01;31:*.rpm=01;31:*.rz=01;31:*.sar=01;31:*.swm=01;31:*.t7z=01;31:*.tar=01;31:*.taz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tgz=01;31:*.tlz=01;31:*.txz=01;31:*.tz=01;31:*.tzo=01;31:*.tzst=01;31:*.udeb=01;31:*.war=01;31:*.whl=01;31:*.wim=01;31:*.xz=01;31:*.z=01;31:*.zip=01;31:*.zoo=01;31:*.zst=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.jxl=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.cfg=00;32:*.conf=00;32:*.diff=00;32:*.doc=00;32:*.ini=00;32:*.log=00;32:*.patch=00;32:*.pdf=00;32:*.ps=00;32:*.tex=00;32:*.txt=00;32:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: XDG_CURRENT_DESKTOP=GNOME FLTK_DOCDIR=/usr/share/doc/fltk-1.3.8/html VTE_VERSION=8004 GNOME_TERMINAL_SCREEN=/org/gnome/Terminal/screen/d1303d53_9095_4f23_a7fc_bf458373742b INFOPATH=/usr/share/gcc-data/x86_64-pc-linux-gnu/15/info:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/info:/usr/share/autoconf-2.72/info:/usr/share/autoconf-2.13/info:/usr/share/automake-1.18.1/info:/usr/share/info MOZ_GMP_PATH=/usr/lib64/nsbrowser/plugins/gmp-gmpopenh264/system-installed JAVAC=/etc/java-config-2/current-system-vm/bin/javac XDG_SESSION_CLASS=user TERM=screen.xterm-256color LESSOPEN=|lesspipe %s USER=XXX GNOME_TERMINAL_SERVICE=:1.69 MANPAGER=manpager NPM_CONFIG_GLOBALCONFIG=/etc/npm/npmrc DISPLAY=:1 SHLVL=2 PAGER=/usr/bin/less LEX=flex QT_IM_MODULE=ibus OLLAMA_DEBUG=2 XSESSION=Xfce4 XDG_RUNTIME_DIR=/run/user/1000 GCC_SPECS= GSETTINGS_BACKEND=dconf XCURSOR_THEME= XDG_DATA_DIRS=/usr/local/share:/usr/share:/etc/eselect/wine/share:/usr/share/gdm PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/opt/bin:/usr/lib/llvm/21/bin:/usr/lib/llvm/20/bin:/usr/lib/llvm/17/bin:/etc/eselect/wine/bin:/opt/cuda/bin STY=25642.pts-0.HOSTNAME GDMSESSION=gnome DBUS_SESSION_BUS_ADDRESS=unix:path=/tmp/dbus-C44hIT5Fs3,guid=783a1deeb9952b0e471df78368f22f07 LV2_PATH=/usr/lib64/lv2 MAIL=/var/mail/XXX SHELL_SETS_TITLE=0 LADSPA_PATH=/usr/lib64/ladspa _=/usr/bin/ollama LD_LIBRARY_PATH=/usr/lib64/ollama:/usr/lib64/ollama OLLAMA_LIBRARY_PATH=/usr/lib64/ollama]" cmd="/usr/bin/ollama runner --ollama-engine --port 45645"
time=2025-10-18T10:08:21.762-05:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-18T10:08:21.762-05:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:45645"
time=2025-10-18T10:08:21.768-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-18T10:08:21.769-05:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/lib64/ollama
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
  Device 0: NVIDIA GeForce RTX 5070 Ti, compute capability 12.0, VMM: yes, ID: GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d
load_backend: loaded CUDA backend from /usr/lib64/ollama/libggml-cuda.so
ggml_vulkan: Found 1 Vulkan devices:
ggml_vulkan: 0 = NVIDIA GeForce RTX 5070 Ti (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2
load_backend: loaded Vulkan backend from /usr/lib64/ollama/libggml-vulkan.so
load_backend: loaded CPU backend from /usr/lib64/ollama/libggml-cpu-icelake.so
time=2025-10-18T10:08:21.904-05:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX_VNNI=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.AVX512_BF16=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=520 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(gcc)
time=2025-10-18T10:08:21.953-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
ggml_backend_vk_get_device_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=195.166182ms
ggml_backend_cuda_device_get_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528
ggml_backend_vk_get_device_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528
ggml_backend_vk_get_device_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528
time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=26.265872ms
time=2025-10-18T10:08:21.990-05:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] devices="[{DeviceID:{ID:GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 Ti FilteredID: Integrated:false PCIID:01:00.0 TotalMemory:17094934528 FreeMemory:16288514048 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[/usr/lib64/ollama]} {DeviceID:{ID:0 Library:Vulkan} Name:Vulkan0 Description:NVIDIA GeForce RTX 5070 Ti FilteredID:0 Integrated:false PCIID:01:00.0 TotalMemory:17094934528 FreeMemory:16288514048 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[/usr/lib64/ollama]}]"
time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=232.449707ms OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs=[]
time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=2
time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:130 msg="verifying GPU is supported" library=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=12.0 pci_id=01:00.0
time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:130 msg="verifying GPU is supported" library=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=0.0 pci_id=01:00.0
time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]"
time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]"
time=2025-10-18T10:08:21.991-05:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash SESSION_MANAGER=local/HOSTNAME:@/tmp/.ICE-unix/25034,unix/HOSTNAME:/tmp/.ICE-unix/25034 COLORTERM=truecolor XDG_CONFIG_DIRS=/etc/xdg LESS=-R -M --shift 5 XDG_MENU_PREFIX=gnome- JDK_HOME=/etc/java-config-2/current-system-vm sparky=00:1c:25:e1:60:61 CONFIG_PROTECT_MASK=/etc/sandbox.d /etc/fonts/fonts.conf /etc/gentoo-release /etc/terminfo /etc/dconf /etc/ca-certificates.conf /etc/revdep-rebuild PKG_CONFIG_PATH=/opt/cuda/pkgconfig TERMCAP=SC|screen.xterm-256color|VT 100/ANSI X3.64 virtual terminal:DO=\\E[%dB:LE=\\E[%dD:RI=\\E[%dC:UP=\\E[%dA:bs:bt=\\E[Z:cd=\\E[J:ce=\\E[K:cl=\\E[H\\E[J:cm=\\E[%i%d;%dH:ct=\\E[3g:do=^J:nd=\\E[C:pt:rc=\\E8:rs=\\Ec:sc=\\E7:st=\\EH:up=\\EM:le=^H:bl=^G:cr=^M:it#8:ho=\\E[H:nw=\\EE:ta=^I:is=\\E)0:li#42:co#139:am:xn:xv:LP:sr=\\EM:al=\\E[L:AL=\\E[%dL:cs=\\E[%i%d;%dr:dl=\\E[M:DL=\\E[%dM:dc=\\E[P:DC=\\E[%dP:im=\\E[4h:ei=\\E[4l:mi:IC=\\E[%d@:ks=\\E[?1h\\E=:ke=\\E[?1l\\E>:vi=\\E[?25l:ve=\\E[34h\\E[?25h:vs=\\E[34l:ti=\\E[?1049h:te=\\E[?1049l:us=\\E[4m:ue=\\E[24m:so=\\E[3m:se=\\E[23m:mb=\\E[5m:md=\\E[1m:mh=\\E[2m:mr=\\E[7m:me=\\E[m:ms:Co#8:pa#64:AF=\\E[3%dm:AB=\\E[4%dm:op=\\E[39;49m:AX:vb=\\Eg:G0:as=\\E(0:ae=\\E(B:ac=\\140\\140aaffggjjkkllmmnnooppqqrrssttuuvvwwxxyyzz{{||}}~~..--++,,hhII00:po=\\E[5i:pf=\\E[4i:Z0=\\E[?3h:Z1=\\E[?3l:Km=\\E[<:k0=\\E[10~:k1=\\EOP:k2=\\EOQ:k3=\\EOR:k4=\\EOS:k5=\\E[15~:k6=\\E[17~:k7=\\E[18~:k8=\\E[19~:k9=\\E[20~:k;=\\E[21~:F1=\\E[23~:F2=\\E[24~:kB=\\E[Z:kh=\\E[1~:@1=\\E[1~:kH=\\E[4~:@7=\\E[4~:kN=\\E[6~:kP=\\E[5~:kI=\\E[2~:kD=\\E[3~:ku=\\EOA:kd=\\EOB:kr=\\EOC:kl=\\EOD:km: JAVA_HOME=/etc/java-config-2/current-system-vm WINDOW=0 XMODIFIERS=@im=ibus DESKTOP_SESSION=gnome EDITOR=/bin/nano GTK_MODULES=canberra-gtk-module PWD=/home/XXX CONFIG_PROTECT=/usr/share/gnupg/qualified.txt /usr/share/config XDG_SESSION_DESKTOP=gnome LOGNAME=XXX XDG_SESSION_TYPE=x11 MANPATH=/etc/java-config-2/current-system-vm/man:/usr/share/gcc-data/x86_64-pc-linux-gnu/15/man:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/man:/etc/java-config-2/current-system-vm/man/:/usr/local/share/man:/usr/share/man:/usr/lib/rust/man-bin-1.86.0:/usr/lib/rust/man-bin-1.88.0:/usr/lib/rust/man-bin-1.89.0:/usr/lib/rust/man-bin-1.90.0:/usr/lib/llvm/21/share/man:/usr/lib/llvm/20/share/man:/usr/lib/llvm/17/share/man:/etc/eselect/wine/share/man XAUTHORITY=/run/user/1000/gdm/Xauthority WINDOWPATH=8 MOTD_SHOWN=pam GDM_LANG=en_US.utf8 HOME=/home/XXX USERNAME=XXX LANG=en_US.utf8 LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.7z=01;31:*.ace=01;31:*.alz=01;31:*.apk=01;31:*.arc=01;31:*.arj=01;31:*.bz=01;31:*.bz2=01;31:*.cab=01;31:*.cpio=01;31:*.crate=01;31:*.deb=01;31:*.drpm=01;31:*.dwm=01;31:*.dz=01;31:*.ear=01;31:*.egg=01;31:*.esd=01;31:*.gz=01;31:*.jar=01;31:*.lha=01;31:*.lrz=01;31:*.lz=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.lzo=01;31:*.pyz=01;31:*.rar=01;31:*.rpm=01;31:*.rz=01;31:*.sar=01;31:*.swm=01;31:*.t7z=01;31:*.tar=01;31:*.taz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tgz=01;31:*.tlz=01;31:*.txz=01;31:*.tz=01;31:*.tzo=01;31:*.tzst=01;31:*.udeb=01;31:*.war=01;31:*.whl=01;31:*.wim=01;31:*.xz=01;31:*.z=01;31:*.zip=01;31:*.zoo=01;31:*.zst=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.jxl=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.cfg=00;32:*.conf=00;32:*.diff=00;32:*.doc=00;32:*.ini=00;32:*.log=00;32:*.patch=00;32:*.pdf=00;32:*.ps=00;32:*.tex=00;32:*.txt=00;32:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: XDG_CURRENT_DESKTOP=GNOME FLTK_DOCDIR=/usr/share/doc/fltk-1.3.8/html VTE_VERSION=8004 GNOME_TERMINAL_SCREEN=/org/gnome/Terminal/screen/d1303d53_9095_4f23_a7fc_bf458373742b INFOPATH=/usr/share/gcc-data/x86_64-pc-linux-gnu/15/info:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/info:/usr/share/autoconf-2.72/info:/usr/share/autoconf-2.13/info:/usr/share/automake-1.18.1/info:/usr/share/info MOZ_GMP_PATH=/usr/lib64/nsbrowser/plugins/gmp-gmpopenh264/system-installed JAVAC=/etc/java-config-2/current-system-vm/bin/javac XDG_SESSION_CLASS=user TERM=screen.xterm-256color LESSOPEN=|lesspipe %s USER=XXX GNOME_TERMINAL_SERVICE=:1.69 MANPAGER=manpager NPM_CONFIG_GLOBALCONFIG=/etc/npm/npmrc DISPLAY=:1 SHLVL=2 PAGER=/usr/bin/less LEX=flex QT_IM_MODULE=ibus OLLAMA_DEBUG=2 XSESSION=Xfce4 XDG_RUNTIME_DIR=/run/user/1000 GCC_SPECS= GSETTINGS_BACKEND=dconf XCURSOR_THEME= XDG_DATA_DIRS=/usr/local/share:/usr/share:/etc/eselect/wine/share:/usr/share/gdm PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/opt/bin:/usr/lib/llvm/21/bin:/usr/lib/llvm/20/bin:/usr/lib/llvm/17/bin:/etc/eselect/wine/bin:/opt/cuda/bin STY=25642.pts-0.HOSTNAME GDMSESSION=gnome DBUS_SESSION_BUS_ADDRESS=unix:path=/tmp/dbus-C44hIT5Fs3,guid=783a1deeb9952b0e471df78368f22f07 LV2_PATH=/usr/lib64/lv2 MAIL=/var/mail/XXX SHELL_SETS_TITLE=0 LADSPA_PATH=/usr/lib64/ladspa _=/usr/bin/ollama LD_LIBRARY_PATH=/usr/lib64/ollama:/usr/lib64/ollama OLLAMA_LIBRARY_PATH=/usr/lib64/ollama GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]" cmd="/usr/bin/ollama runner --ollama-engine --port 35787"
time=2025-10-18T10:08:21.991-05:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash SESSION_MANAGER=local/HOSTNAME:@/tmp/.ICE-unix/25034,unix/HOSTNAME:/tmp/.ICE-unix/25034 COLORTERM=truecolor XDG_CONFIG_DIRS=/etc/xdg LESS=-R -M --shift 5 XDG_MENU_PREFIX=gnome- JDK_HOME=/etc/java-config-2/current-system-vm sparky=00:1c:25:e1:60:61 CONFIG_PROTECT_MASK=/etc/sandbox.d /etc/fonts/fonts.conf /etc/gentoo-release /etc/terminfo /etc/dconf /etc/ca-certificates.conf /etc/revdep-rebuild PKG_CONFIG_PATH=/opt/cuda/pkgconfig TERMCAP=SC|screen.xterm-256color|VT 100/ANSI X3.64 virtual terminal:DO=\\E[%dB:LE=\\E[%dD:RI=\\E[%dC:UP=\\E[%dA:bs:bt=\\E[Z:cd=\\E[J:ce=\\E[K:cl=\\E[H\\E[J:cm=\\E[%i%d;%dH:ct=\\E[3g:do=^J:nd=\\E[C:pt:rc=\\E8:rs=\\Ec:sc=\\E7:st=\\EH:up=\\EM:le=^H:bl=^G:cr=^M:it#8:ho=\\E[H:nw=\\EE:ta=^I:is=\\E)0:li#42:co#139:am:xn:xv:LP:sr=\\EM:al=\\E[L:AL=\\E[%dL:cs=\\E[%i%d;%dr:dl=\\E[M:DL=\\E[%dM:dc=\\E[P:DC=\\E[%dP:im=\\E[4h:ei=\\E[4l:mi:IC=\\E[%d@:ks=\\E[?1h\\E=:ke=\\E[?1l\\E>:vi=\\E[?25l:ve=\\E[34h\\E[?25h:vs=\\E[34l:ti=\\E[?1049h:te=\\E[?1049l:us=\\E[4m:ue=\\E[24m:so=\\E[3m:se=\\E[23m:mb=\\E[5m:md=\\E[1m:mh=\\E[2m:mr=\\E[7m:me=\\E[m:ms:Co#8:pa#64:AF=\\E[3%dm:AB=\\E[4%dm:op=\\E[39;49m:AX:vb=\\Eg:G0:as=\\E(0:ae=\\E(B:ac=\\140\\140aaffggjjkkllmmnnooppqqrrssttuuvvwwxxyyzz{{||}}~~..--++,,hhII00:po=\\E[5i:pf=\\E[4i:Z0=\\E[?3h:Z1=\\E[?3l:Km=\\E[<:k0=\\E[10~:k1=\\EOP:k2=\\EOQ:k3=\\EOR:k4=\\EOS:k5=\\E[15~:k6=\\E[17~:k7=\\E[18~:k8=\\E[19~:k9=\\E[20~:k;=\\E[21~:F1=\\E[23~:F2=\\E[24~:kB=\\E[Z:kh=\\E[1~:@1=\\E[1~:kH=\\E[4~:@7=\\E[4~:kN=\\E[6~:kP=\\E[5~:kI=\\E[2~:kD=\\E[3~:ku=\\EOA:kd=\\EOB:kr=\\EOC:kl=\\EOD:km: JAVA_HOME=/etc/java-config-2/current-system-vm WINDOW=0 XMODIFIERS=@im=ibus DESKTOP_SESSION=gnome EDITOR=/bin/nano GTK_MODULES=canberra-gtk-module PWD=/home/XXX CONFIG_PROTECT=/usr/share/gnupg/qualified.txt /usr/share/config XDG_SESSION_DESKTOP=gnome LOGNAME=XXX XDG_SESSION_TYPE=x11 MANPATH=/etc/java-config-2/current-system-vm/man:/usr/share/gcc-data/x86_64-pc-linux-gnu/15/man:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/man:/etc/java-config-2/current-system-vm/man/:/usr/local/share/man:/usr/share/man:/usr/lib/rust/man-bin-1.86.0:/usr/lib/rust/man-bin-1.88.0:/usr/lib/rust/man-bin-1.89.0:/usr/lib/rust/man-bin-1.90.0:/usr/lib/llvm/21/share/man:/usr/lib/llvm/20/share/man:/usr/lib/llvm/17/share/man:/etc/eselect/wine/share/man XAUTHORITY=/run/user/1000/gdm/Xauthority WINDOWPATH=8 MOTD_SHOWN=pam GDM_LANG=en_US.utf8 HOME=/home/XXX USERNAME=XXX LANG=en_US.utf8 LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.7z=01;31:*.ace=01;31:*.alz=01;31:*.apk=01;31:*.arc=01;31:*.arj=01;31:*.bz=01;31:*.bz2=01;31:*.cab=01;31:*.cpio=01;31:*.crate=01;31:*.deb=01;31:*.drpm=01;31:*.dwm=01;31:*.dz=01;31:*.ear=01;31:*.egg=01;31:*.esd=01;31:*.gz=01;31:*.jar=01;31:*.lha=01;31:*.lrz=01;31:*.lz=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.lzo=01;31:*.pyz=01;31:*.rar=01;31:*.rpm=01;31:*.rz=01;31:*.sar=01;31:*.swm=01;31:*.t7z=01;31:*.tar=01;31:*.taz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tgz=01;31:*.tlz=01;31:*.txz=01;31:*.tz=01;31:*.tzo=01;31:*.tzst=01;31:*.udeb=01;31:*.war=01;31:*.whl=01;31:*.wim=01;31:*.xz=01;31:*.z=01;31:*.zip=01;31:*.zoo=01;31:*.zst=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.jxl=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.cfg=00;32:*.conf=00;32:*.diff=00;32:*.doc=00;32:*.ini=00;32:*.log=00;32:*.patch=00;32:*.pdf=00;32:*.ps=00;32:*.tex=00;32:*.txt=00;32:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: XDG_CURRENT_DESKTOP=GNOME FLTK_DOCDIR=/usr/share/doc/fltk-1.3.8/html VTE_VERSION=8004 GNOME_TERMINAL_SCREEN=/org/gnome/Terminal/screen/d1303d53_9095_4f23_a7fc_bf458373742b INFOPATH=/usr/share/gcc-data/x86_64-pc-linux-gnu/15/info:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/info:/usr/share/autoconf-2.72/info:/usr/share/autoconf-2.13/info:/usr/share/automake-1.18.1/info:/usr/share/info MOZ_GMP_PATH=/usr/lib64/nsbrowser/plugins/gmp-gmpopenh264/system-installed JAVAC=/etc/java-config-2/current-system-vm/bin/javac XDG_SESSION_CLASS=user TERM=screen.xterm-256color LESSOPEN=|lesspipe %s USER=XXX GNOME_TERMINAL_SERVICE=:1.69 MANPAGER=manpager NPM_CONFIG_GLOBALCONFIG=/etc/npm/npmrc DISPLAY=:1 SHLVL=2 PAGER=/usr/bin/less LEX=flex QT_IM_MODULE=ibus OLLAMA_DEBUG=2 XSESSION=Xfce4 XDG_RUNTIME_DIR=/run/user/1000 GCC_SPECS= GSETTINGS_BACKEND=dconf XCURSOR_THEME= XDG_DATA_DIRS=/usr/local/share:/usr/share:/etc/eselect/wine/share:/usr/share/gdm PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/opt/bin:/usr/lib/llvm/21/bin:/usr/lib/llvm/20/bin:/usr/lib/llvm/17/bin:/etc/eselect/wine/bin:/opt/cuda/bin STY=25642.pts-0.HOSTNAME GDMSESSION=gnome DBUS_SESSION_BUS_ADDRESS=unix:path=/tmp/dbus-C44hIT5Fs3,guid=783a1deeb9952b0e471df78368f22f07 LV2_PATH=/usr/lib64/lv2 MAIL=/var/mail/XXX SHELL_SETS_TITLE=0 LADSPA_PATH=/usr/lib64/ladspa _=/usr/bin/ollama LD_LIBRARY_PATH=/usr/lib64/ollama:/usr/lib64/ollama OLLAMA_LIBRARY_PATH=/usr/lib64/ollama GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]" cmd="/usr/bin/ollama runner --ollama-engine --port 32957"
time=2025-10-18T10:08:21.996-05:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-18T10:08:21.996-05:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:32957"
time=2025-10-18T10:08:21.996-05:00 level=INFO source=runner.go:1332 msg="starting ollama engine"
time=2025-10-18T10:08:21.997-05:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:35787"
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-10-18T10:08:22.002-05:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-18T10:08:22.002-05:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/lib64/ollama
time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/lib64/ollama
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
/var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu:329: GGML_ASSERT(ggml_cuda_has_arch(info.devices[id].cc) && "ggml was not compiled with support for this arch") failed
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
/var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu:329: GGML_ASSERT(ggml_cuda_has_arch(info.devices[id].cc) && "ggml was not compiled with support for this arch") failed
/usr/lib64/ollama/libggml-base.so(+0x147a5) [0x7f718c46f7a5]
/usr/lib64/ollama/libggml-base.so(ggml_print_backtrace+0x203) [0x7f718c46fae3]
/usr/lib64/ollama/libggml-base.so(ggml_abort+0x147) [0x7f718c46fca7]
/usr/lib64/ollama/libggml-cuda.so(+0xe9f5b) [0x7f71824e9f5b]
/usr/lib64/ollama/libggml-cuda.so(_Z14ggml_cuda_infov+0x5d) [0x7f71824eacdd]
/usr/lib64/ollama/libggml-cuda.so(ggml_backend_cuda_reg+0x95) [0x7f71824ebe65]
/usr/bin/ollama() [0x11db25a]
/usr/bin/ollama() [0x11d9344]
/usr/bin/ollama() [0x11da9e5]
/usr/bin/ollama() [0x4ae3a4]
/usr/lib64/ollama/libggml-base.so(+0x147a5) [0x7fb0111ad7a5]
/usr/lib64/ollama/libggml-base.so(ggml_print_backtrace+0x203) [0x7fb0111adae3]
/usr/lib64/ollama/libggml-base.so(ggml_abort+0x147) [0x7fb0111adca7]
/usr/lib64/ollama/libggml-cuda.so(+0xe9f5b) [0x7faffece9f5b]
/usr/lib64/ollama/libggml-cuda.so(_Z14ggml_cuda_infov+0x5d) [0x7faffeceacdd]
/usr/lib64/ollama/libggml-cuda.so(ggml_backend_cuda_reg+0x95) [0x7faffecebe65]
/usr/bin/ollama() [0x11db25a]
/usr/bin/ollama() [0x11d9344]
/usr/bin/ollama() [0x11da9e5]
/usr/bin/ollama() [0x4ae3a4]
SIGABRT: abort
PC=0x7f71d5ad289c m=0 sigcode=18446744073709551610
signal arrived during cgo execution

goroutine 35 gp=0xc000588540 m=0 mp=0x21d9220 [syscall]:
runtime.cgocall(0x11ad7b0, 0xc000058838)
        /usr/lib/go/src/runtime/cgocall.go:167 +0x4b fp=0xc000058810 sp=0xc0000587d8 pc=0x4a30eb
github.com/ollama/ollama/ml/backend/ggml/ggml/src._Cfunc_ggml_backend_load_all_from_path(0xda168a0)
        _cgo_gotypes.go:195 +0x3a fp=0xc000058838 sp=0xc000058810 pc=0x857a5a
github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1.1({0xc00003e134, 0x11})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go:97 +0xf5 fp=0xc0000588d0 sp=0xc000058838 pc=0x857515
github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1()
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go:98 +0x545 fp=0xc000058b48 sp=0xc0000588d0 pc=0x857365
github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func2.1()
        /usr/lib/go/src/sync/oncefunc.go:33 +0x4b fp=0xc000058b80 sp=0xc000058b48 pc=0x856d6b
sync.(*Once).doSlow(0x2283fc0?, 0x0?)
        /usr/lib/go/src/sync/once.go:78 +0xac fp=0xc000058bd8 sp=0xc000058b80 pc=0x4b8dec
sync.(*Once).Do(0xc000058c78?, 0xc?)
        /usr/lib/go/src/sync/once.go:69 +0x19 fp=0xc000058bf8 sp=0xc000058bd8 pc=0x4b8d19
github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func2()
        /usr/lib/go/src/sync/oncefunc.go:22 +0x39 fp=0xc000058c30 sp=0xc000058bf8 pc=0x856cf9
github.com/ollama/ollama/ml/backend/ggml.init.func1()
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go:46 +0x1e fp=0xc000058cb8 sp=0xc000058c30 pc=0x8e319e
github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2.1()
        /usr/lib/go/src/sync/oncefunc.go:33 +0x4b fp=0xc000058cf0 sp=0xc000058cb8 pc=0x8e30cb
sync.(*Once).doSlow(0x8000?, 0x100000182ec00?)
        /usr/lib/go/src/sync/once.go:78 +0xac fp=0xc000058d48 sp=0xc000058cf0 pc=0x4b8dec
sync.(*Once).Do(0x22845d0?, 0x4b8ea0?)
        /usr/lib/go/src/sync/once.go:69 +0x19 fp=0xc000058d68 sp=0xc000058d48 pc=0x4b8d19
github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2()
        /usr/lib/go/src/sync/oncefunc.go:22 +0x39 fp=0xc000058da0 sp=0xc000058d68 pc=0x8e3059
github.com/ollama/ollama/ml/backend/ggml.New({0xc0003b01b0, 0x13}, {0x0, 0x6, {0xc0004f1a00, 0x1, 0x1}, 0x0})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go:145 +0x124 fp=0xc000059678 sp=0xc000058da0 pc=0x8eb564
github.com/ollama/ollama/ml.NewBackend({0xc0003b01b0, 0x13}, {0x0, 0x6, {0xc0004f1a00, 0x1, 0x1}, 0x0})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend.go:92 +0x9c fp=0xc0000596c8 sp=0xc000059678 pc=0x85941c
github.com/ollama/ollama/model.New({0xc0003b01b0?, 0xc0005f3978?}, {0x0, 0x6, {0xc0004f1a00, 0x1, 0x1}, 0x0})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/model/model.go:106 +0x66 fp=0xc000059788 sp=0xc0000596c8 pc=0x900346
github.com/ollama/ollama/runner/ollamarunner.(*Server).info(0xc00022b0e0, {0x1836750, 0xc0004125a0}, 0xc0005b0c20?)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1302 +0x533 fp=0xc000059ab8 sp=0xc000059788 pc=0x9bf833
github.com/ollama/ollama/runner/ollamarunner.(*Server).info-fm({0x1836750?, 0xc0004125a0?}, 0xc0005f3b30?)
        <autogenerated>:1 +0x36 fp=0xc000059ae8 sp=0xc000059ab8 pc=0x9c0d16
net/http.HandlerFunc.ServeHTTP(0xc000482000?, {0x1836750?, 0xc0004125a0?}, 0x7a1196?)
        /usr/lib/go/src/net/http/server.go:2322 +0x29 fp=0xc000059b10 sp=0xc000059ae8 pc=0x7a8d49
net/http.(*ServeMux).ServeHTTP(0x4a2fb9?, {0x1836750, 0xc0004125a0}, 0xc0005a9a40)
        /usr/lib/go/src/net/http/server.go:2861 +0x1c7 fp=0xc000059b60 sp=0xc000059b10 pc=0x7aac27
net/http.serverHandler.ServeHTTP({0xc0000aa000?}, {0x1836750?, 0xc0004125a0?}, 0x1?)
        /usr/lib/go/src/net/http/server.go:3340 +0x8e fp=0xc000059b90 sp=0xc000059b60 pc=0x7c850e
net/http.(*conn).serve(0xc0000f1c20, {0x1838b08, 0xc0005c12c0})
        /usr/lib/go/src/net/http/server.go:2109 +0x665 fp=0xc000059fb8 sp=0xc000059b90 pc=0x7a6e45
net/http.(*Server).Serve.gowrap3()
        /usr/lib/go/src/net/http/server.go:3493 +0x28 fp=0xc000059fe0 sp=0xc000059fb8 pc=0x7acb08
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000059fe8 sp=0xc000059fe0 pc=0x4ae701
created by net/http.(*Server).Serve in goroutine 1
        /usr/lib/go/src/net/http/server.go:3493 +0x485

goroutine 1 gp=0xc000002380 m=nil [IO wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc0004c3790 sp=0xc0004c3770 pc=0x4a656e
runtime.netpollblock(0xc0004c37e0?, 0x43b5e6?, 0x0?)
        /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc0004c37c8 sp=0xc0004c3790 pc=0x46a317
internal/poll.runtime_pollWait(0x7f71d58b5400, 0x72)
        /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc0004c37e8 sp=0xc0004c37c8 pc=0x4a5745
internal/poll.(*pollDesc).wait(0xc00059f280?, 0x9004461b7?, 0x0)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc0004c3810 sp=0xc0004c37e8 pc=0x52e627
internal/poll.(*pollDesc).waitRead(...)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89
internal/poll.(*FD).Accept(0xc00059f280)
        /usr/lib/go/src/internal/poll/fd_unix.go:613 +0x28c fp=0xc0004c38b8 sp=0xc0004c3810 pc=0x533a4c
net.(*netFD).accept(0xc00059f280)
        /usr/lib/go/src/net/fd_unix.go:161 +0x29 fp=0xc0004c3970 sp=0xc0004c38b8 pc=0x59dfa9
net.(*TCPListener).accept(0xc0004f1900)
        /usr/lib/go/src/net/tcpsock_posix.go:159 +0x1b fp=0xc0004c39c0 sp=0xc0004c3970 pc=0x5b369b
net.(*TCPListener).Accept(0xc0004f1900)
        /usr/lib/go/src/net/tcpsock.go:380 +0x30 fp=0xc0004c39f0 sp=0xc0004c39c0 pc=0x5b2530
net/http.(*onceCloseListener).Accept(0xc0000f1c20?)
        <autogenerated>:1 +0x24 fp=0xc0004c3a08 sp=0xc0004c39f0 pc=0x7d4d24
net/http.(*Server).Serve(0xc0005c2e00, {0x1836570, 0xc0004f1900})
        /usr/lib/go/src/net/http/server.go:3463 +0x30c fp=0xc0004c3b38 sp=0xc0004c3a08 pc=0x7ac70c
github.com/ollama/ollama/runner/ollamarunner.Execute({0xc000036080, 0x2, 0x2})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1368 +0x954 fp=0xc0004c3d08 sp=0xc0004c3b38 pc=0x9c0734
github.com/ollama/ollama/runner.Execute({0xc000036060?, 0x0?, 0x0?})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/runner.go:20 +0xc9 fp=0xc0004c3d30 sp=0xc0004c3d08 pc=0x9c1029
github.com/ollama/ollama/cmd.NewCLI.func2(0xc0005c2b00?, {0x1657910?, 0x4?, 0x1657914?})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/cmd/cmd.go:1769 +0x45 fp=0xc0004c3d58 sp=0xc0004c3d30 pc=0x11341c5
github.com/spf13/cobra.(*Command).execute(0xc0005c5b08, {0xc0005c0d50, 0x3, 0x3})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:940 +0x88a fp=0xc0004c3e78 sp=0xc0004c3d58 pc=0x61772a
github.com/spf13/cobra.(*Command).ExecuteC(0xc00048ef08)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:1068 +0x398 fp=0xc0004c3f30 sp=0xc0004c3e78 pc=0x617f58
github.com/spf13/cobra.(*Command).Execute(...)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:992
github.com/spf13/cobra.(*Command).ExecuteContext(...)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:985
main.main()
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/main.go:12 +0x4d fp=0xc0004c3f50 sp=0xc0004c3f30 pc=0x1134c8d
runtime.main()
        /usr/lib/go/src/runtime/proc.go:285 +0x29d fp=0xc0004c3fe0 sp=0xc0004c3f50 pc=0x471b5d
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0004c3fe8 sp=0xc0004c3fe0 pc=0x4ae701

goroutine 2 gp=0xc000002e00 m=nil [force gc (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007cfa8 sp=0xc00007cf88 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.forcegchelper()
        /usr/lib/go/src/runtime/proc.go:373 +0xb3 fp=0xc00007cfe0 sp=0xc00007cfa8 pc=0x471e93
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007cfe8 sp=0xc00007cfe0 pc=0x4ae701
created by runtime.init.7 in goroutine 1
        /usr/lib/go/src/runtime/proc.go:361 +0x1a

goroutine 3 gp=0xc000003340 m=nil [GC sweep wait]:
runtime.gopark(0x1?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007d780 sp=0xc00007d760 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.bgsweep(0xc0000a8000)
        /usr/lib/go/src/runtime/mgcsweep.go:323 +0xdf fp=0xc00007d7c8 sp=0xc00007d780 pc=0x45bc1f
runtime.gcenable.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:212 +0x25 fp=0xc00007d7e0 sp=0xc00007d7c8 pc=0x44fc05
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007d7e8 sp=0xc00007d7e0 pc=0x4ae701
created by runtime.gcenable in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:212 +0x66

goroutine 4 gp=0xc000003500 m=nil [GC scavenge wait]:
runtime.gopark(0x10000?, 0x1822d88?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007df78 sp=0xc00007df58 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.(*scavengerState).park(0x21d62a0)
        /usr/lib/go/src/runtime/mgcscavenge.go:425 +0x49 fp=0xc00007dfa8 sp=0xc00007df78 pc=0x459689
runtime.bgscavenge(0xc0000a8000)
        /usr/lib/go/src/runtime/mgcscavenge.go:658 +0x59 fp=0xc00007dfc8 sp=0xc00007dfa8 pc=0x459c39
runtime.gcenable.gowrap2()
        /usr/lib/go/src/runtime/mgc.go:213 +0x25 fp=0xc00007dfe0 sp=0xc00007dfc8 pc=0x44fba5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007dfe8 sp=0xc00007dfe0 pc=0x4ae701
created by runtime.gcenable in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:213 +0xa5

goroutine 5 gp=0xc000003dc0 m=nil [finalizer wait]:
runtime.gopark(0x480e35?, 0x447565?, 0xb8?, 0x1?, 0xc000002380?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007c620 sp=0xc00007c600 pc=0x4a656e
runtime.runFinalizers()
        /usr/lib/go/src/runtime/mfinal.go:210 +0x107 fp=0xc00007c7e0 sp=0xc00007c620 pc=0x44eb07
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007c7e8 sp=0xc00007c7e0 pc=0x4ae701
created by runtime.createfing in goroutine 1
        /usr/lib/go/src/runtime/mfinal.go:172 +0x3d

goroutine 6 gp=0xc0001d28c0 m=nil [cleanup wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007e768 sp=0xc00007e748 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.(*cleanupQueue).dequeue(0x21d6c00)
        /usr/lib/go/src/runtime/mcleanup.go:439 +0xc5 fp=0xc00007e7a0 sp=0xc00007e768 pc=0x44bd25
runtime.runCleanups()
        /usr/lib/go/src/runtime/mcleanup.go:635 +0x45 fp=0xc00007e7e0 sp=0xc00007e7a0 pc=0x44c3e5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007e7e8 sp=0xc00007e7e0 pc=0x4ae701
created by runtime.(*cleanupQueue).createGs in goroutine 1
        /usr/lib/go/src/runtime/mcleanup.go:589 +0xa5

goroutine 7 gp=0xc0001d3180 m=nil [GC worker (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007ef38 sp=0xc00007ef18 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007efc8 sp=0xc00007ef38 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007efe0 sp=0xc00007efc8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007efe8 sp=0xc00007efe0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 18 gp=0xc000504000 m=nil [GC worker (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000078738 sp=0xc000078718 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0000787c8 sp=0xc000078738 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0000787e0 sp=0xc0000787c8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0000787e8 sp=0xc0000787e0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 34 gp=0xc000102380 m=nil [GC worker (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000118738 sp=0xc000118718 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001187c8 sp=0xc000118738 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001187e0 sp=0xc0001187c8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001187e8 sp=0xc0001187e0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 8 gp=0xc0001d36c0 m=nil [GC worker (idle)]:
runtime.gopark(0x59364534a969?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007f738 sp=0xc00007f718 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007f7c8 sp=0xc00007f738 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007f7e0 sp=0xc00007f7c8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007f7e8 sp=0xc00007f7e0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 9 gp=0xc0001d3880 m=nil [GC worker (idle)]:
runtime.gopark(0x59364534a991?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007ff38 sp=0xc00007ff18 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007ffc8 sp=0xc00007ff38 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007ffe0 sp=0xc00007ffc8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007ffe8 sp=0xc00007ffe0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 10 gp=0xc0001d3a40 m=nil [GC worker (idle)]:
runtime.gopark(0x593645401bdb?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000114738 sp=0xc000114718 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001147c8 sp=0xc000114738 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001147e0 sp=0xc0001147c8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001147e8 sp=0xc0001147e0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 11 gp=0xc000588380 m=nil [sync.WaitGroup.Wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0xc0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00008ca90 sp=0xc00008ca70 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.semacquire1(0xc00022b198, 0x0, 0x1, 0x0, 0x19)
        /usr/lib/go/src/runtime/sema.go:192 +0x21d fp=0xc00008caf8 sp=0xc00008ca90 pc=0x48589d
sync.runtime_SemacquireWaitGroup(0x0?, 0x0?)
        /usr/lib/go/src/runtime/sema.go:114 +0x2e fp=0xc00008cb30 sp=0xc00008caf8 pc=0x4a806e
sync.(*WaitGroup).Wait(0xc00022b190)
        /usr/lib/go/src/sync/waitgroup.go:206 +0x85 fp=0xc00008cb58 sp=0xc00008cb30 pc=0x4ba365
github.com/ollama/ollama/runner/ollamarunner.(*Server).run(0xc00022b0e0, {0x1838b40, 0xc0005dc000})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:415 +0x45 fp=0xc00008cfb8 sp=0xc00008cb58 pc=0x9b7605
github.com/ollama/ollama/runner/ollamarunner.Execute.gowrap1()
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x28 fp=0xc00008cfe0 sp=0xc00008cfb8 pc=0x9c09a8
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00008cfe8 sp=0xc00008cfe0 pc=0x4ae701
created by github.com/ollama/ollama/runner/ollamarunner.Execute in goroutine 1
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x4c9

goroutine 12 gp=0xc0005041c0 m=nil [IO wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0xb?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000117dd8 sp=0xc000117db8 pc=0x4a656e
runtime.netpollblock(0x4cab98?, 0x43b5e6?, 0x0?)
        /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc000117e10 sp=0xc000117dd8 pc=0x46a317
internal/poll.runtime_pollWait(0x7f71d58b5200, 0x72)
        /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc000117e30 sp=0xc000117e10 pc=0x4a5745
internal/poll.(*pollDesc).wait(0xc00059f300?, 0xc0000aa021?, 0x0)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc000117e58 sp=0xc000117e30 pc=0x52e627
internal/poll.(*pollDesc).waitRead(...)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89
internal/poll.(*FD).Read(0xc00059f300, {0xc0000aa021, 0x1, 0x1})
        /usr/lib/go/src/internal/poll/fd_unix.go:165 +0x279 fp=0xc000117ef0 sp=0xc000117e58 pc=0x52f919
net.(*netFD).Read(0xc00059f300, {0xc0000aa021?, 0x0?SIGABRT: abort
PC=0x7fb05b8a489c m=5 sigcode=18446744073709551610
signal arrived during cgo execution

goroutine 38 gp=0xc0001d5a40 m=5 mp=0xc000083808 [syscall]:
runtime.cgocall(0x11ad7b0, 0xc000058838)
        , /usr/lib/go/src/runtime/cgocall.go:167 +0x00x4b? fp=}0xc000058810)
 sp=    0xc0000587d8/usr/lib/go/src/net/fd_posix.go pc=:0x4a30eb
68 +0x25 fp=0xc000117f38 sp=github.com/ollama/ollama/ml/backend/ggml/ggml/src._Cfunc_ggml_backend_load_all_from_path0xc000117ef0( pc=0x7fb008000be00x59c105)

        _cgo_gotypes.go:net.(*conn).Read195( +0x3a0xc00011e000 fp=, 0xc000058838{ sp=0xc0000aa0210xc000058810? pc=, 0x857a5a0x0
?, github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1.10x0(?{}0xc00003e134)
,       0x11/usr/lib/go/src/net/net.go}:)
196      +/var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go0x45: fp=970xc000117f80 + sp=0xf50xc000117f38 fp= pc=0xc0000588d00x5aa105 sp=
0xc000058838 pc=0x857515net/http.(*connReader).backgroundRead
(0xc0000aa000github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1)
(       )
/usr/lib/go/src/net/http/server.go      :/var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go702: +980x33 + fp=0x5450xc000117fc8 fp= sp=0xc000058b480xc000117f80 sp= pc=0xc0000588d00x7a12f3 pc=
0x857365
net/http.(*connReader).startBackgroundRead.gowrap2(github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func2.1)
(       )
/usr/lib/go/src/net/http/server.go      :/usr/lib/go/src/sync/oncefunc.go698: +330x25 + fp=0x4b0xc000117fe0 fp= sp=0xc000058b800xc000117fc8 sp= pc=0xc000058b480x7a1225 pc=
0x856d6b
runtime.goexit(sync.(*Once).doSlow{(}0x2283fc0)
?       , /usr/lib/go/src/runtime/asm_amd64.s0x0:?1693)
 +      0x1/usr/lib/go/src/sync/once.go fp=:0xc000117fe878 sp= +0xc000117fe00xac pc= fp=0x4ae7010xc000058bd8
 sp=created by 0xc000058b80net/http.(*connReader).startBackgroundRead pc= in goroutine 0x4b8dec35

sync.(*Once).Do (/usr/lib/go/src/net/http/server.go0xc000058c78:?698,  +0xc0xb6?
)

        rax    /usr/lib/go/src/sync/once.go0x0:
69rbx     +0x2c9d0x19
 fp=rcx    0xc000058bf80x7f71d5ad289c sp=
0xc000058bd8rdx     pc=0x60x4b8d19

rdi    github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func20x2c9d(
)
rsi     0x2c9d/usr/lib/go/src/sync/oncefunc.go
:rbp    220x7f7182965286 +
0x39rsp     fp=0x7ffc3726e0d00xc000058c30
 sp=r8     0xc000058bf80x0 pc=
0x856cf9r9     
0x0
github.com/ollama/ollama/ml/backend/ggml.init.func1r10    (0x0)

        r11    /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go0x246:
46r12     +0x1490x1e
 fp=r13    0xc000058cb80x7f718291eff0 sp=
0xc000058c30r14     pc=0x60x8e319e

r15    github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2.10x7ffc3726f7c0(
)
rip     0x7f71d5ad289c/usr/lib/go/src/sync/oncefunc.go
:rflags 330x246 +
0x4bcs      fp=0x330xc000058cf0
 sp=fs     0xc000058cb80x0 pc=
0x8e30cbgs     
0x0sync.(*Once).doSlow
(0x8000?, 0x100000182ec00?)
        /usr/lib/go/src/sync/once.go:78 +0xac fp=0xc000058d48 sp=0xc000058cf0 pc=0x4b8dec
sync.(*Once).Do(0x22845d0?, 0x4b8ea0?)
        /usr/lib/go/src/sync/once.go:69 +0x19 fp=0xc000058d68 sp=0xc000058d48 pc=0x4b8d19
github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2()
        /usr/lib/go/src/sync/oncefunc.go:22 +0x39 fp=0xc000058da0 sp=0xc000058d68 pc=0x8e3059
github.com/ollama/ollama/ml/backend/ggml.New({0xc0003b0900, 0x13}, {0x0, 0x6, {0xc000688940, 0x1, 0x1}, 0x0})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go:145 +0x124 fp=0xc000059678 sp=0xc000058da0 pc=0x8eb564
github.com/ollama/ollama/ml.NewBackend({0xc0003b0900, 0x13}, {0x0, 0x6, {0xc000688940, 0x1, 0x1}, 0x0})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend.go:92 +0x9c fp=0xc0000596c8 sp=0xc000059678 pc=0x85941c
github.com/ollama/ollama/model.New({0xc0003b0900?, 0xc000165978?}, {0x0, 0x6, {0xc000688940, 0x1, 0x1}, 0x0})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/model/model.go:106 +0x66 fp=0xc000059788 sp=0xc0000596c8 pc=0x900346
github.com/ollama/ollama/runner/ollamarunner.(*Server).info(0xc00022af00, {0x1836750, 0xc0005332c0}, 0xc000503140?)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1302 +0x533 fp=0xc000059ab8 sp=0xc000059788 pc=0x9bf833
github.com/ollama/ollama/runner/ollamarunner.(*Server).info-fm({0x1836750?, 0xc0005332c0?}, 0xc000165b30?)
        <autogenerated>:1 +0x36 fp=0xc000059ae8 sp=0xc000059ab8 pc=0x9c0d16
net/http.HandlerFunc.ServeHTTP(0xc000543140?, {0x1836750?, 0xc0005332c0?}, 0x7a1196?)
        /usr/lib/go/src/net/http/server.go:2322 +0x29 fp=0xc000059b10 sp=0xc000059ae8 pc=0x7a8d49
net/http.(*ServeMux).ServeHTTP(0x4a2fb9?, {0x1836750, 0xc0005332c0}, 0xc000379540)
        /usr/lib/go/src/net/http/server.go:2861 +0x1c7 fp=0xc000059b60 sp=0xc000059b10 pc=0x7aac27
net/http.serverHandler.ServeHTTP({0xc000688780?}, {0x1836750?, 0xc0005332c0?}, 0x1?)
        /usr/lib/go/src/net/http/server.go:3340 +0x8e fp=0xc000059b90 sp=0xc000059b60 pc=0x7c850e
net/http.(*conn).serve(0xc00012a3f0, {0x1838b08, 0xc000216570})
        /usr/lib/go/src/net/http/server.go:2109 +0x665 fp=0xc000059fb8 sp=0xc000059b90 pc=0x7a6e45
net/http.(*Server).Serve.gowrap3()
        /usr/lib/go/src/net/http/server.go:3493 +0x28 fp=0xc000059fe0 sp=0xc000059fb8 pc=0x7acb08
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000059fe8 sp=0xc000059fe0 pc=0x4ae701
created by net/http.(*Server).Serve in goroutine 1
        /usr/lib/go/src/net/http/server.go:3493 +0x485

goroutine 1 gp=0xc000002380 m=nil [IO wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc0004af790 sp=0xc0004af770 pc=0x4a656e
runtime.netpollblock(0xc0004af7e0?, 0x43b5e6?, 0x0?)
        /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc0004af7c8 sp=0xc0004af790 pc=0x46a317
internal/poll.runtime_pollWait(0x7fb05b677400, 0x72)
        /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc0004af7e8 sp=0xc0004af7c8 pc=0x4a5745
internal/poll.(*pollDesc).wait(0xc000121600?, 0x900000036?, 0x0)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc0004af810 sp=0xc0004af7e8 pc=0x52e627
internal/poll.(*pollDesc).waitRead(...)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89
internal/poll.(*FD).Accept(0xc000121600)
        /usr/lib/go/src/internal/poll/fd_unix.go:613 +0x28c fp=0xc0004af8b8 sp=0xc0004af810 pc=0x533a4c
net.(*netFD).accept(0xc000121600)
        /usr/lib/go/src/net/fd_unix.go:161 +0x29 fp=0xc0004af970 sp=0xc0004af8b8 pc=0x59dfa9
net.(*TCPListener).accept(0xc000688740)
        /usr/lib/go/src/net/tcpsock_posix.go:159 +0x1b fp=0xc0004af9c0 sp=0xc0004af970 pc=0x5b369b
net.(*TCPListener).Accept(0xc000688740)
        /usr/lib/go/src/net/tcpsock.go:380 +0x30 fp=0xc0004af9f0 sp=0xc0004af9c0 pc=0x5b2530
net/http.(*onceCloseListener).Accept(0xc00012a3f0?)
        <autogenerated>:1 +0x24 fp=0xc0004afa08 sp=0xc0004af9f0 pc=0x7d4d24
net/http.(*Server).Serve(0xc0001f5700, {0x1836570, 0xc000688740})
        /usr/lib/go/src/net/http/server.go:3463 +0x30c fp=0xc0004afb38 sp=0xc0004afa08 pc=0x7ac70c
github.com/ollama/ollama/runner/ollamarunner.Execute({0xc000036080, 0x2, 0x2})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1368 +0x954 fp=0xc0004afd08 sp=0xc0004afb38 pc=0x9c0734
github.com/ollama/ollama/runner.Execute({0xc000036060?, 0x0?, 0x0?})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/runner.go:20 +0xc9 fp=0xc0004afd30 sp=0xc0004afd08 pc=0x9c1029
github.com/ollama/ollama/cmd.NewCLI.func2(0xc0001f5400?, {0x1657910?, 0x4?, 0x1657914?})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/cmd/cmd.go:1769 +0x45 fp=0xc0004afd58 sp=0xc0004afd30 pc=0x11341c5
github.com/spf13/cobra.(*Command).execute(0xc00069fb08, {0xc000216000, 0x3, 0x3})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:940 +0x88a fp=0xc0004afe78 sp=0xc0004afd58 pc=0x61772a
github.com/spf13/cobra.(*Command).ExecuteC(0xc00049ac08)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:1068 +0x398 fp=0xc0004aff30 sp=0xc0004afe78 pc=0x617f58
github.com/spf13/cobra.(*Command).Execute(...)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:992
github.com/spf13/cobra.(*Command).ExecuteContext(...)
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:985
main.main()
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/main.go:12 +0x4d fp=0xc0004aff50 sp=0xc0004aff30 pc=0x1134c8d
runtime.main()
        /usr/lib/go/src/runtime/proc.go:285 +0x29d fp=0xc0004affe0 sp=0xc0004aff50 pc=0x471b5d
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0004affe8 sp=0xc0004affe0 pc=0x4ae701

goroutine 2 gp=0xc000002e00 m=nil [force gc (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007cfa8 sp=0xc00007cf88 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.forcegchelper()
        /usr/lib/go/src/runtime/proc.go:373 +0xb3 fp=0xc00007cfe0 sp=0xc00007cfa8 pc=0x471e93
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007cfe8 sp=0xc00007cfe0 pc=0x4ae701
created by runtime.init.7 in goroutine 1
        /usr/lib/go/src/runtime/proc.go:361 +0x1a

goroutine 3 gp=0xc000003340 m=nil [GC sweep wait]:
runtime.gopark(0x1?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007d780 sp=0xc00007d760 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.bgsweep(0xc0000a8000)
        /usr/lib/go/src/runtime/mgcsweep.go:323 +0xdf fp=0xc00007d7c8 sp=0xc00007d780 pc=0x45bc1f
runtime.gcenable.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:212 +0x25 fp=0xc00007d7e0 sp=0xc00007d7c8 pc=0x44fc05
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007d7e8 sp=0xc00007d7e0 pc=0x4ae701
created by runtime.gcenable in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:212 +0x66

goroutine 4 gp=0xc000003500 m=nil [GC scavenge wait]:
runtime.gopark(0x10000?, 0x1822d88?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007df78 sp=0xc00007df58 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.(*scavengerState).park(0x21d62a0)
        /usr/lib/go/src/runtime/mgcscavenge.go:425 +0x49 fp=0xc00007dfa8 sp=0xc00007df78 pc=0x459689
runtime.bgscavenge(0xc0000a8000)
        /usr/lib/go/src/runtime/mgcscavenge.go:658 +0x59 fp=0xc00007dfc8 sp=0xc00007dfa8 pc=0x459c39
runtime.gcenable.gowrap2()
        /usr/lib/go/src/runtime/mgc.go:213 +0x25 fp=0xc00007dfe0 sp=0xc00007dfc8 pc=0x44fba5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007dfe8 sp=0xc00007dfe0 pc=0x4ae701
created by runtime.gcenable in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:213 +0xa5

goroutine 5 gp=0xc000003dc0 m=nil [finalizer wait]:
runtime.gopark(0x480e35?, 0x447565?, 0xb8?, 0x1?, 0xc000002380?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007c620 sp=0xc00007c600 pc=0x4a656e
runtime.runFinalizers()
        /usr/lib/go/src/runtime/mfinal.go:210 +0x107 fp=0xc00007c7e0 sp=0xc00007c620 pc=0x44eb07
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007c7e8 sp=0xc00007c7e0 pc=0x4ae701
created by runtime.createfing in goroutine 1
        /usr/lib/go/src/runtime/mfinal.go:172 +0x3d

goroutine 6 gp=0xc0001d48c0 m=nil [cleanup wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007e768 sp=0xc00007e748 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.(*cleanupQueue).dequeue(0x21d6c00)
        /usr/lib/go/src/runtime/mcleanup.go:439 +0xc5 fp=0xc00007e7a0 sp=0xc00007e768 pc=0x44bd25
runtime.runCleanups()
        /usr/lib/go/src/runtime/mcleanup.go:635 +0x45 fp=0xc00007e7e0 sp=0xc00007e7a0 pc=0x44c3e5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007e7e8 sp=0xc00007e7e0 pc=0x4ae701
created by runtime.(*cleanupQueue).createGs in goroutine 1
        /usr/lib/go/src/runtime/mcleanup.go:589 +0xa5

goroutine 7 gp=0xc0001d5340 m=nil [GC worker (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007ef38 sp=0xc00007ef18 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007efc8 sp=0xc00007ef38 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007efe0 sp=0xc00007efc8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007efe8 sp=0xc00007efe0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 18 gp=0xc000504000 m=nil [GC worker (idle)]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000078738 sp=0xc000078718 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0000787c8 sp=0xc000078738 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0000787e0 sp=0xc0000787c8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0000787e8 sp=0xc0000787e0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 34 gp=0xc000102380 m=nil [GC worker (idle)]:
runtime.gopark(0x5936452d22e8?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000118738 sp=0xc000118718 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001187c8 sp=0xc000118738 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001187e0 sp=0xc0001187c8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001187e8 sp=0xc0001187e0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 35 gp=0xc000102540 m=nil [GC worker (idle)]:
runtime.gopark(0x5936452d2bee?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000118f38 sp=0xc000118f18 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc000118fc8 sp=0xc000118f38 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc000118fe0 sp=0xc000118fc8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000118fe8 sp=0xc000118fe0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 36 gp=0xc000102700 m=nil [GC worker (idle)]:
runtime.gopark(0x5936452d22d4?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000119738 sp=0xc000119718 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001197c8 sp=0xc000119738 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001197e0 sp=0xc0001197c8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001197e8 sp=0xc0001197e0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 8 gp=0xc0001d5500 m=nil [GC worker (idle)]:
runtime.gopark(0x5936452d37d8?, 0x0?, 0x0?, 0x0?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00008ff38 sp=0xc00008ff18 pc=0x4a656e
runtime.gcBgMarkWorker(0xc0000b76c0)
        /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00008ffc8 sp=0xc00008ff38 pc=0x4522eb
runtime.gcBgMarkStartWorkers.gowrap1()
        /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00008ffe0 sp=0xc00008ffc8 pc=0x4521c5
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00008ffe8 sp=0xc00008ffe0 pc=0x4ae701
created by runtime.gcBgMarkStartWorkers in goroutine 1
        /usr/lib/go/src/runtime/mgc.go:1373 +0x105

goroutine 37 gp=0xc0001d5880 m=nil [sync.WaitGroup.Wait]:
runtime.gopark(0x0?, 0x0?, 0xe0?, 0xc1?, 0x0?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00008ea90 sp=0xc00008ea70 pc=0x4a656e
runtime.goparkunlock(...)
        /usr/lib/go/src/runtime/proc.go:466
runtime.semacquire1(0xc00022afb8, 0x0, 0x1, 0x0, 0x19)
        /usr/lib/go/src/runtime/sema.go:192 +0x21d fp=0xc00008eaf8 sp=0xc00008ea90 pc=0x48589d
sync.runtime_SemacquireWaitGroup(0x0?, 0x0?)
        /usr/lib/go/src/runtime/sema.go:114 +0x2e fp=0xc00008eb30 sp=0xc00008eaf8 pc=0x4a806e
sync.(*WaitGroup).Wait(0xc00022afb0)
        /usr/lib/go/src/sync/waitgroup.go:206 +0x85 fp=0xc00008eb58 sp=0xc00008eb30 pc=0x4ba365
github.com/ollama/ollama/runner/ollamarunner.(*Server).run(0xc00022af00, {0x1838b40, 0xc00068c230})
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:415 +0x45 fp=0xc00008efb8 sp=0xc00008eb58 pc=0x9b7605
github.com/ollama/ollama/runner/ollamarunner.Execute.gowrap1()
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x28 fp=0xc00008efe0 sp=0xc00008efb8 pc=0x9c09a8
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00008efe8 sp=0xc00008efe0 pc=0x4ae701
created by github.com/ollama/ollama/runner/ollamarunner.Execute in goroutine 1
        /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x4c9

goroutine 39 gp=0xc0001d5c00 m=nil [IO wait]:
runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0xb?)
        /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000078dd8 sp=0xc000078db8 pc=0x4a656e
runtime.netpollblock(0x4cab98?, 0x43b5e6?, 0x0?)
        /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc000078e10 sp=0xc000078dd8 pc=0x46a317
internal/poll.runtime_pollWait(0x7fb05b677200, 0x72)
        /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc000078e30 sp=0xc000078e10 pc=0x4a5745
internal/poll.(*pollDesc).wait(0xc000121680?, 0xc0006887a1?, 0x0)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc000078e58 sp=0xc000078e30 pc=0x52e627
internal/poll.(*pollDesc).waitRead(...)
        /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89
internal/poll.(*FD).Read(0xc000121680, {0xc0006887a1, 0x1, 0x1})
        /usr/lib/go/src/internal/poll/fd_unix.go:165 +0x279 fp=0xc000078ef0 sp=0xc000078e58 pc=0x52f919
net.(*netFD).Read(0xc000121680, {0xc0006887a1?, 0x0?, 0x0?})
        /usr/lib/go/src/net/fd_posix.go:68 +0x25 fp=0xc000078f38 sp=0xc000078ef0 pc=0x59c105
net.(*conn).Read(0xc0000806e8, {0xc0006887a1?, 0x0?, 0x0?})
        /usr/lib/go/src/net/net.go:196 +0x45 fp=0xc000078f80 sp=0xc000078f38 pc=0x5aa105
net/http.(*connReader).backgroundRead(0xc000688780)
        /usr/lib/go/src/net/http/server.go:702 +0x33 fp=0xc000078fc8 sp=0xc000078f80 pc=0x7a12f3
net/http.(*connReader).startBackgroundRead.gowrap2()
        /usr/lib/go/src/net/http/server.go:698 +0x25 fp=0xc000078fe0 sp=0xc000078fc8 pc=0x7a1225
runtime.goexit({})
        /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000078fe8 sp=0xc000078fe0 pc=0x4ae701
created by net/http.(*connReader).startBackgroundRead in goroutine 38
        /usr/lib/go/src/net/http/server.go:698 +0xb6

rax    0x0
rbx    0x2ca3
rcx    0x7fb05b8a489c
rdx    0x6
rdi    0x2c9e
rsi    0x2ca3
rbp    0x7fafff165286
rsp    0x7fb013b9b660
r8     0x0
r9     0x0
r10    0x0
r11    0x246
r12    0x149
r13    0x7fafff11eff0
r14    0x6
r15    0x7fb013b9cd50
rip    0x7fb05b8a489c
rflags 0x246
cs     0x33
fs     0x0
gs     0x0
time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:543 msg="runner exited" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]" code=2
time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] devices=[]
time=2025-10-18T10:08:22.061-05:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=70.942693ms OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]"
time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:543 msg="runner exited" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]" code=2
time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] devices=[]
time=2025-10-18T10:08:22.061-05:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=71.139921ms OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]"
time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[]
time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:182 msg="removing unsupported or overlapping GPU combination" libDir=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=12.0 pci_id=01:00.0
time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:182 msg="removing unsupported or overlapping GPU combination" libDir=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=0.0 pci_id=01:00.0
time=2025-10-18T10:08:22.061-05:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=303.766267ms
time=2025-10-18T10:08:22.061-05:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="123.4 GiB" available="120.1 GiB"
time=2025-10-18T10:08:22.061-05:00 level=INFO source=routes.go:1609 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

However version 0.12.3 from guru overlay (same use flags, blas+cuda) it detects the GPU without issue:

$ ollama serve
time=2025-10-18T09:44:05.937-05:00 level=INFO source=routes.go:1475 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/home/XXX/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"
time=2025-10-18T09:44:05.938-05:00 level=INFO source=routes.go:1528 msg="Listening on 127.0.0.1:11434 (version 0.12.3)"
time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler"
time=2025-10-18T09:44:05.938-05:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs"
time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA"
time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=gpu.go:520 msg="Searching for GPU library" name=libcuda.so*
time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=gpu.go:544 msg="gpu library search" globs="[/usr/lib64/ollama/libcuda.so* /home/XXX/libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]"
time=2025-10-18T09:44:05.942-05:00 level=DEBUG source=gpu.go:577 msg="discovered GPU libraries" paths="[/usr/lib/libcuda.so.580.95.05 /usr/lib64/libcuda.so.580.95.05]"
initializing /usr/lib/libcuda.so.580.95.05
library /usr/lib/libcuda.so.580.95.05 load err: /usr/lib/libcuda.so.580.95.05: wrong ELF class: ELFCLASS32
time=2025-10-18T09:44:05.942-05:00 level=DEBUG source=gpu.go:628 msg="skipping 32bit library" library=/usr/lib/libcuda.so.580.95.05
initializing /usr/lib64/libcuda.so.580.95.05
dlsym: cuInit - 0x7fb131d05d00
dlsym: cuDriverGetVersion - 0x7fb131d05dc0
dlsym: cuDeviceGetCount - 0x7fb131d05f40
dlsym: cuDeviceGet - 0x7fb131d05e80
dlsym: cuDeviceGetAttribute - 0x7fb131d28f20
dlsym: cuDeviceGetUuid - 0x7fb131d7bd10
dlsym: cuDeviceGetName - 0x7fb131d06000
dlsym: cuCtxCreate_v3 - 0x7fb131d79750
dlsym: cuMemGetInfo_v2 - 0x7fb131d2d460
dlsym: cuCtxDestroy - 0x7fb131d7b650
calling cuInit
calling cuDriverGetVersion
raw version 0x32c8
CUDA driver version: 13.0
calling cuDeviceGetCount
device count 1
time=2025-10-18T09:44:05.963-05:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=1 library=/usr/lib64/libcuda.so.580.95.05
[GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d] CUDA totalMem 15806mb
[GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d] CUDA freeMem 15347mb
[GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d] Compute Capability 12.0
time=2025-10-18T09:44:06.081-05:00 level=WARN source=amd_linux.go:61 msg="ollama recommends running the https://www.amd.com/en/support/download/linux-drivers.html" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory"
time=2025-10-18T09:44:06.081-05:00 level=INFO source=amd_linux.go:406 msg="no compatible amdgpu devices detected"
releasing cuda driver library
time=2025-10-18T09:44:06.081-05:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070 Ti" total="15.4 GiB" available="15.0 GiB"
time=2025-10-18T09:44:06.081-05:00 level=INFO source=routes.go:1569 msg="entering low vram mode" "total vram"="15.4 GiB" threshold="20.0 GiB"

Edited to include debug level 2 log. Thanks for your help with this! Let me know if I can provide any more info.

<!-- gh-comment-id:3418474125 --> @aminalshmu commented on GitHub (Oct 18, 2025): Have been experiencing this same issue on Gentoo since around 0.12.5, this is from the just compiled git version: ``` $ ollama serve time=2025-10-18T10:08:21.757-05:00 level=INFO source=routes.go:1515 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/home/XXX/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-10-18T10:08:21.758-05:00 level=INFO source=routes.go:1568 msg="Listening on 127.0.0.1:11434 (version 0.12.6-6-gbc1a818-dirty)" time=2025-10-18T10:08:21.758-05:00 level=DEBUG source=sched.go:125 msg="starting llm scheduler" time=2025-10-18T10:08:21.758-05:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-18T10:08:21.758-05:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs=[] time=2025-10-18T10:08:21.758-05:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash SESSION_MANAGER=local/HOSTNAME:@/tmp/.ICE-unix/25034,unix/HOSTNAME:/tmp/.ICE-unix/25034 COLORTERM=truecolor XDG_CONFIG_DIRS=/etc/xdg LESS=-R -M --shift 5 XDG_MENU_PREFIX=gnome- JDK_HOME=/etc/java-config-2/current-system-vm sparky=00:1c:25:e1:60:61 CONFIG_PROTECT_MASK=/etc/sandbox.d /etc/fonts/fonts.conf /etc/gentoo-release /etc/terminfo /etc/dconf /etc/ca-certificates.conf /etc/revdep-rebuild PKG_CONFIG_PATH=/opt/cuda/pkgconfig TERMCAP=SC|screen.xterm-256color|VT 100/ANSI X3.64 virtual terminal:DO=\\E[%dB:LE=\\E[%dD:RI=\\E[%dC:UP=\\E[%dA:bs:bt=\\E[Z:cd=\\E[J:ce=\\E[K:cl=\\E[H\\E[J:cm=\\E[%i%d;%dH:ct=\\E[3g:do=^J:nd=\\E[C:pt:rc=\\E8:rs=\\Ec:sc=\\E7:st=\\EH:up=\\EM:le=^H:bl=^G:cr=^M:it#8:ho=\\E[H:nw=\\EE:ta=^I:is=\\E)0:li#42:co#139:am:xn:xv:LP:sr=\\EM:al=\\E[L:AL=\\E[%dL:cs=\\E[%i%d;%dr:dl=\\E[M:DL=\\E[%dM:dc=\\E[P:DC=\\E[%dP:im=\\E[4h:ei=\\E[4l:mi:IC=\\E[%d@:ks=\\E[?1h\\E=:ke=\\E[?1l\\E>:vi=\\E[?25l:ve=\\E[34h\\E[?25h:vs=\\E[34l:ti=\\E[?1049h:te=\\E[?1049l:us=\\E[4m:ue=\\E[24m:so=\\E[3m:se=\\E[23m:mb=\\E[5m:md=\\E[1m:mh=\\E[2m:mr=\\E[7m:me=\\E[m:ms:Co#8:pa#64:AF=\\E[3%dm:AB=\\E[4%dm:op=\\E[39;49m:AX:vb=\\Eg:G0:as=\\E(0:ae=\\E(B:ac=\\140\\140aaffggjjkkllmmnnooppqqrrssttuuvvwwxxyyzz{{||}}~~..--++,,hhII00:po=\\E[5i:pf=\\E[4i:Z0=\\E[?3h:Z1=\\E[?3l:Km=\\E[<:k0=\\E[10~:k1=\\EOP:k2=\\EOQ:k3=\\EOR:k4=\\EOS:k5=\\E[15~:k6=\\E[17~:k7=\\E[18~:k8=\\E[19~:k9=\\E[20~:k;=\\E[21~:F1=\\E[23~:F2=\\E[24~:kB=\\E[Z:kh=\\E[1~:@1=\\E[1~:kH=\\E[4~:@7=\\E[4~:kN=\\E[6~:kP=\\E[5~:kI=\\E[2~:kD=\\E[3~:ku=\\EOA:kd=\\EOB:kr=\\EOC:kl=\\EOD:km: JAVA_HOME=/etc/java-config-2/current-system-vm WINDOW=0 XMODIFIERS=@im=ibus DESKTOP_SESSION=gnome EDITOR=/bin/nano GTK_MODULES=canberra-gtk-module PWD=/home/XXX CONFIG_PROTECT=/usr/share/gnupg/qualified.txt /usr/share/config XDG_SESSION_DESKTOP=gnome LOGNAME=XXX XDG_SESSION_TYPE=x11 MANPATH=/etc/java-config-2/current-system-vm/man:/usr/share/gcc-data/x86_64-pc-linux-gnu/15/man:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/man:/etc/java-config-2/current-system-vm/man/:/usr/local/share/man:/usr/share/man:/usr/lib/rust/man-bin-1.86.0:/usr/lib/rust/man-bin-1.88.0:/usr/lib/rust/man-bin-1.89.0:/usr/lib/rust/man-bin-1.90.0:/usr/lib/llvm/21/share/man:/usr/lib/llvm/20/share/man:/usr/lib/llvm/17/share/man:/etc/eselect/wine/share/man XAUTHORITY=/run/user/1000/gdm/Xauthority WINDOWPATH=8 MOTD_SHOWN=pam GDM_LANG=en_US.utf8 HOME=/home/XXX USERNAME=XXX LANG=en_US.utf8 LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.7z=01;31:*.ace=01;31:*.alz=01;31:*.apk=01;31:*.arc=01;31:*.arj=01;31:*.bz=01;31:*.bz2=01;31:*.cab=01;31:*.cpio=01;31:*.crate=01;31:*.deb=01;31:*.drpm=01;31:*.dwm=01;31:*.dz=01;31:*.ear=01;31:*.egg=01;31:*.esd=01;31:*.gz=01;31:*.jar=01;31:*.lha=01;31:*.lrz=01;31:*.lz=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.lzo=01;31:*.pyz=01;31:*.rar=01;31:*.rpm=01;31:*.rz=01;31:*.sar=01;31:*.swm=01;31:*.t7z=01;31:*.tar=01;31:*.taz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tgz=01;31:*.tlz=01;31:*.txz=01;31:*.tz=01;31:*.tzo=01;31:*.tzst=01;31:*.udeb=01;31:*.war=01;31:*.whl=01;31:*.wim=01;31:*.xz=01;31:*.z=01;31:*.zip=01;31:*.zoo=01;31:*.zst=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.jxl=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.cfg=00;32:*.conf=00;32:*.diff=00;32:*.doc=00;32:*.ini=00;32:*.log=00;32:*.patch=00;32:*.pdf=00;32:*.ps=00;32:*.tex=00;32:*.txt=00;32:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: XDG_CURRENT_DESKTOP=GNOME FLTK_DOCDIR=/usr/share/doc/fltk-1.3.8/html VTE_VERSION=8004 GNOME_TERMINAL_SCREEN=/org/gnome/Terminal/screen/d1303d53_9095_4f23_a7fc_bf458373742b INFOPATH=/usr/share/gcc-data/x86_64-pc-linux-gnu/15/info:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/info:/usr/share/autoconf-2.72/info:/usr/share/autoconf-2.13/info:/usr/share/automake-1.18.1/info:/usr/share/info MOZ_GMP_PATH=/usr/lib64/nsbrowser/plugins/gmp-gmpopenh264/system-installed JAVAC=/etc/java-config-2/current-system-vm/bin/javac XDG_SESSION_CLASS=user TERM=screen.xterm-256color LESSOPEN=|lesspipe %s USER=XXX GNOME_TERMINAL_SERVICE=:1.69 MANPAGER=manpager NPM_CONFIG_GLOBALCONFIG=/etc/npm/npmrc DISPLAY=:1 SHLVL=2 PAGER=/usr/bin/less LEX=flex QT_IM_MODULE=ibus OLLAMA_DEBUG=2 XSESSION=Xfce4 XDG_RUNTIME_DIR=/run/user/1000 GCC_SPECS= GSETTINGS_BACKEND=dconf XCURSOR_THEME= XDG_DATA_DIRS=/usr/local/share:/usr/share:/etc/eselect/wine/share:/usr/share/gdm PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/opt/bin:/usr/lib/llvm/21/bin:/usr/lib/llvm/20/bin:/usr/lib/llvm/17/bin:/etc/eselect/wine/bin:/opt/cuda/bin STY=25642.pts-0.HOSTNAME GDMSESSION=gnome DBUS_SESSION_BUS_ADDRESS=unix:path=/tmp/dbus-C44hIT5Fs3,guid=783a1deeb9952b0e471df78368f22f07 LV2_PATH=/usr/lib64/lv2 MAIL=/var/mail/XXX SHELL_SETS_TITLE=0 LADSPA_PATH=/usr/lib64/ladspa _=/usr/bin/ollama LD_LIBRARY_PATH=/usr/lib64/ollama:/usr/lib64/ollama OLLAMA_LIBRARY_PATH=/usr/lib64/ollama]" cmd="/usr/bin/ollama runner --ollama-engine --port 45645" time=2025-10-18T10:08:21.762-05:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-18T10:08:21.762-05:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:45645" time=2025-10-18T10:08:21.768-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-18T10:08:21.769-05:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-18T10:08:21.769-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/lib64/ollama ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 5070 Ti, compute capability 12.0, VMM: yes, ID: GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d load_backend: loaded CUDA backend from /usr/lib64/ollama/libggml-cuda.so ggml_vulkan: Found 1 Vulkan devices: ggml_vulkan: 0 = NVIDIA GeForce RTX 5070 Ti (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2 load_backend: loaded Vulkan backend from /usr/lib64/ollama/libggml-vulkan.so load_backend: loaded CPU backend from /usr/lib64/ollama/libggml-cpu-icelake.so time=2025-10-18T10:08:21.904-05:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX_VNNI=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.AVX512_BF16=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=520 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(gcc) time=2025-10-18T10:08:21.953-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 ggml_backend_vk_get_device_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528 time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-18T10:08:21.963-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-18T10:08:21.964-05:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=195.166182ms ggml_backend_cuda_device_get_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528 ggml_backend_vk_get_device_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528 ggml_backend_vk_get_device_memory utilizing NVML memory reporting free: 16288514048 total: 17094934528 time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=26.265872ms time=2025-10-18T10:08:21.990-05:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] devices="[{DeviceID:{ID:GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 Ti FilteredID: Integrated:false PCIID:01:00.0 TotalMemory:17094934528 FreeMemory:16288514048 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[/usr/lib64/ollama]} {DeviceID:{ID:0 Library:Vulkan} Name:Vulkan0 Description:NVIDIA GeForce RTX 5070 Ti FilteredID:0 Integrated:false PCIID:01:00.0 TotalMemory:17094934528 FreeMemory:16288514048 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[/usr/lib64/ollama]}]" time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=232.449707ms OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs=[] time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=2 time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:130 msg="verifying GPU is supported" library=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=12.0 pci_id=01:00.0 time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:130 msg="verifying GPU is supported" library=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=0.0 pci_id=01:00.0 time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]" time=2025-10-18T10:08:21.990-05:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]" time=2025-10-18T10:08:21.991-05:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash SESSION_MANAGER=local/HOSTNAME:@/tmp/.ICE-unix/25034,unix/HOSTNAME:/tmp/.ICE-unix/25034 COLORTERM=truecolor XDG_CONFIG_DIRS=/etc/xdg LESS=-R -M --shift 5 XDG_MENU_PREFIX=gnome- JDK_HOME=/etc/java-config-2/current-system-vm sparky=00:1c:25:e1:60:61 CONFIG_PROTECT_MASK=/etc/sandbox.d /etc/fonts/fonts.conf /etc/gentoo-release /etc/terminfo /etc/dconf /etc/ca-certificates.conf /etc/revdep-rebuild PKG_CONFIG_PATH=/opt/cuda/pkgconfig TERMCAP=SC|screen.xterm-256color|VT 100/ANSI X3.64 virtual terminal:DO=\\E[%dB:LE=\\E[%dD:RI=\\E[%dC:UP=\\E[%dA:bs:bt=\\E[Z:cd=\\E[J:ce=\\E[K:cl=\\E[H\\E[J:cm=\\E[%i%d;%dH:ct=\\E[3g:do=^J:nd=\\E[C:pt:rc=\\E8:rs=\\Ec:sc=\\E7:st=\\EH:up=\\EM:le=^H:bl=^G:cr=^M:it#8:ho=\\E[H:nw=\\EE:ta=^I:is=\\E)0:li#42:co#139:am:xn:xv:LP:sr=\\EM:al=\\E[L:AL=\\E[%dL:cs=\\E[%i%d;%dr:dl=\\E[M:DL=\\E[%dM:dc=\\E[P:DC=\\E[%dP:im=\\E[4h:ei=\\E[4l:mi:IC=\\E[%d@:ks=\\E[?1h\\E=:ke=\\E[?1l\\E>:vi=\\E[?25l:ve=\\E[34h\\E[?25h:vs=\\E[34l:ti=\\E[?1049h:te=\\E[?1049l:us=\\E[4m:ue=\\E[24m:so=\\E[3m:se=\\E[23m:mb=\\E[5m:md=\\E[1m:mh=\\E[2m:mr=\\E[7m:me=\\E[m:ms:Co#8:pa#64:AF=\\E[3%dm:AB=\\E[4%dm:op=\\E[39;49m:AX:vb=\\Eg:G0:as=\\E(0:ae=\\E(B:ac=\\140\\140aaffggjjkkllmmnnooppqqrrssttuuvvwwxxyyzz{{||}}~~..--++,,hhII00:po=\\E[5i:pf=\\E[4i:Z0=\\E[?3h:Z1=\\E[?3l:Km=\\E[<:k0=\\E[10~:k1=\\EOP:k2=\\EOQ:k3=\\EOR:k4=\\EOS:k5=\\E[15~:k6=\\E[17~:k7=\\E[18~:k8=\\E[19~:k9=\\E[20~:k;=\\E[21~:F1=\\E[23~:F2=\\E[24~:kB=\\E[Z:kh=\\E[1~:@1=\\E[1~:kH=\\E[4~:@7=\\E[4~:kN=\\E[6~:kP=\\E[5~:kI=\\E[2~:kD=\\E[3~:ku=\\EOA:kd=\\EOB:kr=\\EOC:kl=\\EOD:km: JAVA_HOME=/etc/java-config-2/current-system-vm WINDOW=0 XMODIFIERS=@im=ibus DESKTOP_SESSION=gnome EDITOR=/bin/nano GTK_MODULES=canberra-gtk-module PWD=/home/XXX CONFIG_PROTECT=/usr/share/gnupg/qualified.txt /usr/share/config XDG_SESSION_DESKTOP=gnome LOGNAME=XXX XDG_SESSION_TYPE=x11 MANPATH=/etc/java-config-2/current-system-vm/man:/usr/share/gcc-data/x86_64-pc-linux-gnu/15/man:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/man:/etc/java-config-2/current-system-vm/man/:/usr/local/share/man:/usr/share/man:/usr/lib/rust/man-bin-1.86.0:/usr/lib/rust/man-bin-1.88.0:/usr/lib/rust/man-bin-1.89.0:/usr/lib/rust/man-bin-1.90.0:/usr/lib/llvm/21/share/man:/usr/lib/llvm/20/share/man:/usr/lib/llvm/17/share/man:/etc/eselect/wine/share/man XAUTHORITY=/run/user/1000/gdm/Xauthority WINDOWPATH=8 MOTD_SHOWN=pam GDM_LANG=en_US.utf8 HOME=/home/XXX USERNAME=XXX LANG=en_US.utf8 LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.7z=01;31:*.ace=01;31:*.alz=01;31:*.apk=01;31:*.arc=01;31:*.arj=01;31:*.bz=01;31:*.bz2=01;31:*.cab=01;31:*.cpio=01;31:*.crate=01;31:*.deb=01;31:*.drpm=01;31:*.dwm=01;31:*.dz=01;31:*.ear=01;31:*.egg=01;31:*.esd=01;31:*.gz=01;31:*.jar=01;31:*.lha=01;31:*.lrz=01;31:*.lz=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.lzo=01;31:*.pyz=01;31:*.rar=01;31:*.rpm=01;31:*.rz=01;31:*.sar=01;31:*.swm=01;31:*.t7z=01;31:*.tar=01;31:*.taz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tgz=01;31:*.tlz=01;31:*.txz=01;31:*.tz=01;31:*.tzo=01;31:*.tzst=01;31:*.udeb=01;31:*.war=01;31:*.whl=01;31:*.wim=01;31:*.xz=01;31:*.z=01;31:*.zip=01;31:*.zoo=01;31:*.zst=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.jxl=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.cfg=00;32:*.conf=00;32:*.diff=00;32:*.doc=00;32:*.ini=00;32:*.log=00;32:*.patch=00;32:*.pdf=00;32:*.ps=00;32:*.tex=00;32:*.txt=00;32:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: XDG_CURRENT_DESKTOP=GNOME FLTK_DOCDIR=/usr/share/doc/fltk-1.3.8/html VTE_VERSION=8004 GNOME_TERMINAL_SCREEN=/org/gnome/Terminal/screen/d1303d53_9095_4f23_a7fc_bf458373742b INFOPATH=/usr/share/gcc-data/x86_64-pc-linux-gnu/15/info:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/info:/usr/share/autoconf-2.72/info:/usr/share/autoconf-2.13/info:/usr/share/automake-1.18.1/info:/usr/share/info MOZ_GMP_PATH=/usr/lib64/nsbrowser/plugins/gmp-gmpopenh264/system-installed JAVAC=/etc/java-config-2/current-system-vm/bin/javac XDG_SESSION_CLASS=user TERM=screen.xterm-256color LESSOPEN=|lesspipe %s USER=XXX GNOME_TERMINAL_SERVICE=:1.69 MANPAGER=manpager NPM_CONFIG_GLOBALCONFIG=/etc/npm/npmrc DISPLAY=:1 SHLVL=2 PAGER=/usr/bin/less LEX=flex QT_IM_MODULE=ibus OLLAMA_DEBUG=2 XSESSION=Xfce4 XDG_RUNTIME_DIR=/run/user/1000 GCC_SPECS= GSETTINGS_BACKEND=dconf XCURSOR_THEME= XDG_DATA_DIRS=/usr/local/share:/usr/share:/etc/eselect/wine/share:/usr/share/gdm PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/opt/bin:/usr/lib/llvm/21/bin:/usr/lib/llvm/20/bin:/usr/lib/llvm/17/bin:/etc/eselect/wine/bin:/opt/cuda/bin STY=25642.pts-0.HOSTNAME GDMSESSION=gnome DBUS_SESSION_BUS_ADDRESS=unix:path=/tmp/dbus-C44hIT5Fs3,guid=783a1deeb9952b0e471df78368f22f07 LV2_PATH=/usr/lib64/lv2 MAIL=/var/mail/XXX SHELL_SETS_TITLE=0 LADSPA_PATH=/usr/lib64/ladspa _=/usr/bin/ollama LD_LIBRARY_PATH=/usr/lib64/ollama:/usr/lib64/ollama OLLAMA_LIBRARY_PATH=/usr/lib64/ollama GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]" cmd="/usr/bin/ollama runner --ollama-engine --port 35787" time=2025-10-18T10:08:21.991-05:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash SESSION_MANAGER=local/HOSTNAME:@/tmp/.ICE-unix/25034,unix/HOSTNAME:/tmp/.ICE-unix/25034 COLORTERM=truecolor XDG_CONFIG_DIRS=/etc/xdg LESS=-R -M --shift 5 XDG_MENU_PREFIX=gnome- JDK_HOME=/etc/java-config-2/current-system-vm sparky=00:1c:25:e1:60:61 CONFIG_PROTECT_MASK=/etc/sandbox.d /etc/fonts/fonts.conf /etc/gentoo-release /etc/terminfo /etc/dconf /etc/ca-certificates.conf /etc/revdep-rebuild PKG_CONFIG_PATH=/opt/cuda/pkgconfig TERMCAP=SC|screen.xterm-256color|VT 100/ANSI X3.64 virtual terminal:DO=\\E[%dB:LE=\\E[%dD:RI=\\E[%dC:UP=\\E[%dA:bs:bt=\\E[Z:cd=\\E[J:ce=\\E[K:cl=\\E[H\\E[J:cm=\\E[%i%d;%dH:ct=\\E[3g:do=^J:nd=\\E[C:pt:rc=\\E8:rs=\\Ec:sc=\\E7:st=\\EH:up=\\EM:le=^H:bl=^G:cr=^M:it#8:ho=\\E[H:nw=\\EE:ta=^I:is=\\E)0:li#42:co#139:am:xn:xv:LP:sr=\\EM:al=\\E[L:AL=\\E[%dL:cs=\\E[%i%d;%dr:dl=\\E[M:DL=\\E[%dM:dc=\\E[P:DC=\\E[%dP:im=\\E[4h:ei=\\E[4l:mi:IC=\\E[%d@:ks=\\E[?1h\\E=:ke=\\E[?1l\\E>:vi=\\E[?25l:ve=\\E[34h\\E[?25h:vs=\\E[34l:ti=\\E[?1049h:te=\\E[?1049l:us=\\E[4m:ue=\\E[24m:so=\\E[3m:se=\\E[23m:mb=\\E[5m:md=\\E[1m:mh=\\E[2m:mr=\\E[7m:me=\\E[m:ms:Co#8:pa#64:AF=\\E[3%dm:AB=\\E[4%dm:op=\\E[39;49m:AX:vb=\\Eg:G0:as=\\E(0:ae=\\E(B:ac=\\140\\140aaffggjjkkllmmnnooppqqrrssttuuvvwwxxyyzz{{||}}~~..--++,,hhII00:po=\\E[5i:pf=\\E[4i:Z0=\\E[?3h:Z1=\\E[?3l:Km=\\E[<:k0=\\E[10~:k1=\\EOP:k2=\\EOQ:k3=\\EOR:k4=\\EOS:k5=\\E[15~:k6=\\E[17~:k7=\\E[18~:k8=\\E[19~:k9=\\E[20~:k;=\\E[21~:F1=\\E[23~:F2=\\E[24~:kB=\\E[Z:kh=\\E[1~:@1=\\E[1~:kH=\\E[4~:@7=\\E[4~:kN=\\E[6~:kP=\\E[5~:kI=\\E[2~:kD=\\E[3~:ku=\\EOA:kd=\\EOB:kr=\\EOC:kl=\\EOD:km: JAVA_HOME=/etc/java-config-2/current-system-vm WINDOW=0 XMODIFIERS=@im=ibus DESKTOP_SESSION=gnome EDITOR=/bin/nano GTK_MODULES=canberra-gtk-module PWD=/home/XXX CONFIG_PROTECT=/usr/share/gnupg/qualified.txt /usr/share/config XDG_SESSION_DESKTOP=gnome LOGNAME=XXX XDG_SESSION_TYPE=x11 MANPATH=/etc/java-config-2/current-system-vm/man:/usr/share/gcc-data/x86_64-pc-linux-gnu/15/man:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/man:/etc/java-config-2/current-system-vm/man/:/usr/local/share/man:/usr/share/man:/usr/lib/rust/man-bin-1.86.0:/usr/lib/rust/man-bin-1.88.0:/usr/lib/rust/man-bin-1.89.0:/usr/lib/rust/man-bin-1.90.0:/usr/lib/llvm/21/share/man:/usr/lib/llvm/20/share/man:/usr/lib/llvm/17/share/man:/etc/eselect/wine/share/man XAUTHORITY=/run/user/1000/gdm/Xauthority WINDOWPATH=8 MOTD_SHOWN=pam GDM_LANG=en_US.utf8 HOME=/home/XXX USERNAME=XXX LANG=en_US.utf8 LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=01;05;37;41:mi=01;05;37;41:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.7z=01;31:*.ace=01;31:*.alz=01;31:*.apk=01;31:*.arc=01;31:*.arj=01;31:*.bz=01;31:*.bz2=01;31:*.cab=01;31:*.cpio=01;31:*.crate=01;31:*.deb=01;31:*.drpm=01;31:*.dwm=01;31:*.dz=01;31:*.ear=01;31:*.egg=01;31:*.esd=01;31:*.gz=01;31:*.jar=01;31:*.lha=01;31:*.lrz=01;31:*.lz=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.lzo=01;31:*.pyz=01;31:*.rar=01;31:*.rpm=01;31:*.rz=01;31:*.sar=01;31:*.swm=01;31:*.t7z=01;31:*.tar=01;31:*.taz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tgz=01;31:*.tlz=01;31:*.txz=01;31:*.tz=01;31:*.tzo=01;31:*.tzst=01;31:*.udeb=01;31:*.war=01;31:*.whl=01;31:*.wim=01;31:*.xz=01;31:*.z=01;31:*.zip=01;31:*.zoo=01;31:*.zst=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.jxl=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.cfg=00;32:*.conf=00;32:*.diff=00;32:*.doc=00;32:*.ini=00;32:*.log=00;32:*.patch=00;32:*.pdf=00;32:*.ps=00;32:*.tex=00;32:*.txt=00;32:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: XDG_CURRENT_DESKTOP=GNOME FLTK_DOCDIR=/usr/share/doc/fltk-1.3.8/html VTE_VERSION=8004 GNOME_TERMINAL_SCREEN=/org/gnome/Terminal/screen/d1303d53_9095_4f23_a7fc_bf458373742b INFOPATH=/usr/share/gcc-data/x86_64-pc-linux-gnu/15/info:/usr/share/binutils-data/x86_64-pc-linux-gnu/2.44/info:/usr/share/autoconf-2.72/info:/usr/share/autoconf-2.13/info:/usr/share/automake-1.18.1/info:/usr/share/info MOZ_GMP_PATH=/usr/lib64/nsbrowser/plugins/gmp-gmpopenh264/system-installed JAVAC=/etc/java-config-2/current-system-vm/bin/javac XDG_SESSION_CLASS=user TERM=screen.xterm-256color LESSOPEN=|lesspipe %s USER=XXX GNOME_TERMINAL_SERVICE=:1.69 MANPAGER=manpager NPM_CONFIG_GLOBALCONFIG=/etc/npm/npmrc DISPLAY=:1 SHLVL=2 PAGER=/usr/bin/less LEX=flex QT_IM_MODULE=ibus OLLAMA_DEBUG=2 XSESSION=Xfce4 XDG_RUNTIME_DIR=/run/user/1000 GCC_SPECS= GSETTINGS_BACKEND=dconf XCURSOR_THEME= XDG_DATA_DIRS=/usr/local/share:/usr/share:/etc/eselect/wine/share:/usr/share/gdm PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/opt/bin:/usr/lib/llvm/21/bin:/usr/lib/llvm/20/bin:/usr/lib/llvm/17/bin:/etc/eselect/wine/bin:/opt/cuda/bin STY=25642.pts-0.HOSTNAME GDMSESSION=gnome DBUS_SESSION_BUS_ADDRESS=unix:path=/tmp/dbus-C44hIT5Fs3,guid=783a1deeb9952b0e471df78368f22f07 LV2_PATH=/usr/lib64/lv2 MAIL=/var/mail/XXX SHELL_SETS_TITLE=0 LADSPA_PATH=/usr/lib64/ladspa _=/usr/bin/ollama LD_LIBRARY_PATH=/usr/lib64/ollama:/usr/lib64/ollama OLLAMA_LIBRARY_PATH=/usr/lib64/ollama GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]" cmd="/usr/bin/ollama runner --ollama-engine --port 32957" time=2025-10-18T10:08:21.996-05:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-18T10:08:21.996-05:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:32957" time=2025-10-18T10:08:21.996-05:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-18T10:08:21.997-05:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:35787" time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-18T10:08:22.002-05:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-18T10:08:22.002-05:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/lib64/ollama time=2025-10-18T10:08:22.002-05:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/lib64/ollama ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu:329: GGML_ASSERT(ggml_cuda_has_arch(info.devices[id].cc) && "ggml was not compiled with support for this arch") failed ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu:329: GGML_ASSERT(ggml_cuda_has_arch(info.devices[id].cc) && "ggml was not compiled with support for this arch") failed /usr/lib64/ollama/libggml-base.so(+0x147a5) [0x7f718c46f7a5] /usr/lib64/ollama/libggml-base.so(ggml_print_backtrace+0x203) [0x7f718c46fae3] /usr/lib64/ollama/libggml-base.so(ggml_abort+0x147) [0x7f718c46fca7] /usr/lib64/ollama/libggml-cuda.so(+0xe9f5b) [0x7f71824e9f5b] /usr/lib64/ollama/libggml-cuda.so(_Z14ggml_cuda_infov+0x5d) [0x7f71824eacdd] /usr/lib64/ollama/libggml-cuda.so(ggml_backend_cuda_reg+0x95) [0x7f71824ebe65] /usr/bin/ollama() [0x11db25a] /usr/bin/ollama() [0x11d9344] /usr/bin/ollama() [0x11da9e5] /usr/bin/ollama() [0x4ae3a4] /usr/lib64/ollama/libggml-base.so(+0x147a5) [0x7fb0111ad7a5] /usr/lib64/ollama/libggml-base.so(ggml_print_backtrace+0x203) [0x7fb0111adae3] /usr/lib64/ollama/libggml-base.so(ggml_abort+0x147) [0x7fb0111adca7] /usr/lib64/ollama/libggml-cuda.so(+0xe9f5b) [0x7faffece9f5b] /usr/lib64/ollama/libggml-cuda.so(_Z14ggml_cuda_infov+0x5d) [0x7faffeceacdd] /usr/lib64/ollama/libggml-cuda.so(ggml_backend_cuda_reg+0x95) [0x7faffecebe65] /usr/bin/ollama() [0x11db25a] /usr/bin/ollama() [0x11d9344] /usr/bin/ollama() [0x11da9e5] /usr/bin/ollama() [0x4ae3a4] SIGABRT: abort PC=0x7f71d5ad289c m=0 sigcode=18446744073709551610 signal arrived during cgo execution goroutine 35 gp=0xc000588540 m=0 mp=0x21d9220 [syscall]: runtime.cgocall(0x11ad7b0, 0xc000058838) /usr/lib/go/src/runtime/cgocall.go:167 +0x4b fp=0xc000058810 sp=0xc0000587d8 pc=0x4a30eb github.com/ollama/ollama/ml/backend/ggml/ggml/src._Cfunc_ggml_backend_load_all_from_path(0xda168a0) _cgo_gotypes.go:195 +0x3a fp=0xc000058838 sp=0xc000058810 pc=0x857a5a github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1.1({0xc00003e134, 0x11}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go:97 +0xf5 fp=0xc0000588d0 sp=0xc000058838 pc=0x857515 github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1() /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go:98 +0x545 fp=0xc000058b48 sp=0xc0000588d0 pc=0x857365 github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func2.1() /usr/lib/go/src/sync/oncefunc.go:33 +0x4b fp=0xc000058b80 sp=0xc000058b48 pc=0x856d6b sync.(*Once).doSlow(0x2283fc0?, 0x0?) /usr/lib/go/src/sync/once.go:78 +0xac fp=0xc000058bd8 sp=0xc000058b80 pc=0x4b8dec sync.(*Once).Do(0xc000058c78?, 0xc?) /usr/lib/go/src/sync/once.go:69 +0x19 fp=0xc000058bf8 sp=0xc000058bd8 pc=0x4b8d19 github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func2() /usr/lib/go/src/sync/oncefunc.go:22 +0x39 fp=0xc000058c30 sp=0xc000058bf8 pc=0x856cf9 github.com/ollama/ollama/ml/backend/ggml.init.func1() /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go:46 +0x1e fp=0xc000058cb8 sp=0xc000058c30 pc=0x8e319e github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2.1() /usr/lib/go/src/sync/oncefunc.go:33 +0x4b fp=0xc000058cf0 sp=0xc000058cb8 pc=0x8e30cb sync.(*Once).doSlow(0x8000?, 0x100000182ec00?) /usr/lib/go/src/sync/once.go:78 +0xac fp=0xc000058d48 sp=0xc000058cf0 pc=0x4b8dec sync.(*Once).Do(0x22845d0?, 0x4b8ea0?) /usr/lib/go/src/sync/once.go:69 +0x19 fp=0xc000058d68 sp=0xc000058d48 pc=0x4b8d19 github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2() /usr/lib/go/src/sync/oncefunc.go:22 +0x39 fp=0xc000058da0 sp=0xc000058d68 pc=0x8e3059 github.com/ollama/ollama/ml/backend/ggml.New({0xc0003b01b0, 0x13}, {0x0, 0x6, {0xc0004f1a00, 0x1, 0x1}, 0x0}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go:145 +0x124 fp=0xc000059678 sp=0xc000058da0 pc=0x8eb564 github.com/ollama/ollama/ml.NewBackend({0xc0003b01b0, 0x13}, {0x0, 0x6, {0xc0004f1a00, 0x1, 0x1}, 0x0}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend.go:92 +0x9c fp=0xc0000596c8 sp=0xc000059678 pc=0x85941c github.com/ollama/ollama/model.New({0xc0003b01b0?, 0xc0005f3978?}, {0x0, 0x6, {0xc0004f1a00, 0x1, 0x1}, 0x0}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/model/model.go:106 +0x66 fp=0xc000059788 sp=0xc0000596c8 pc=0x900346 github.com/ollama/ollama/runner/ollamarunner.(*Server).info(0xc00022b0e0, {0x1836750, 0xc0004125a0}, 0xc0005b0c20?) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1302 +0x533 fp=0xc000059ab8 sp=0xc000059788 pc=0x9bf833 github.com/ollama/ollama/runner/ollamarunner.(*Server).info-fm({0x1836750?, 0xc0004125a0?}, 0xc0005f3b30?) <autogenerated>:1 +0x36 fp=0xc000059ae8 sp=0xc000059ab8 pc=0x9c0d16 net/http.HandlerFunc.ServeHTTP(0xc000482000?, {0x1836750?, 0xc0004125a0?}, 0x7a1196?) /usr/lib/go/src/net/http/server.go:2322 +0x29 fp=0xc000059b10 sp=0xc000059ae8 pc=0x7a8d49 net/http.(*ServeMux).ServeHTTP(0x4a2fb9?, {0x1836750, 0xc0004125a0}, 0xc0005a9a40) /usr/lib/go/src/net/http/server.go:2861 +0x1c7 fp=0xc000059b60 sp=0xc000059b10 pc=0x7aac27 net/http.serverHandler.ServeHTTP({0xc0000aa000?}, {0x1836750?, 0xc0004125a0?}, 0x1?) /usr/lib/go/src/net/http/server.go:3340 +0x8e fp=0xc000059b90 sp=0xc000059b60 pc=0x7c850e net/http.(*conn).serve(0xc0000f1c20, {0x1838b08, 0xc0005c12c0}) /usr/lib/go/src/net/http/server.go:2109 +0x665 fp=0xc000059fb8 sp=0xc000059b90 pc=0x7a6e45 net/http.(*Server).Serve.gowrap3() /usr/lib/go/src/net/http/server.go:3493 +0x28 fp=0xc000059fe0 sp=0xc000059fb8 pc=0x7acb08 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000059fe8 sp=0xc000059fe0 pc=0x4ae701 created by net/http.(*Server).Serve in goroutine 1 /usr/lib/go/src/net/http/server.go:3493 +0x485 goroutine 1 gp=0xc000002380 m=nil [IO wait]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc0004c3790 sp=0xc0004c3770 pc=0x4a656e runtime.netpollblock(0xc0004c37e0?, 0x43b5e6?, 0x0?) /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc0004c37c8 sp=0xc0004c3790 pc=0x46a317 internal/poll.runtime_pollWait(0x7f71d58b5400, 0x72) /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc0004c37e8 sp=0xc0004c37c8 pc=0x4a5745 internal/poll.(*pollDesc).wait(0xc00059f280?, 0x9004461b7?, 0x0) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc0004c3810 sp=0xc0004c37e8 pc=0x52e627 internal/poll.(*pollDesc).waitRead(...) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Accept(0xc00059f280) /usr/lib/go/src/internal/poll/fd_unix.go:613 +0x28c fp=0xc0004c38b8 sp=0xc0004c3810 pc=0x533a4c net.(*netFD).accept(0xc00059f280) /usr/lib/go/src/net/fd_unix.go:161 +0x29 fp=0xc0004c3970 sp=0xc0004c38b8 pc=0x59dfa9 net.(*TCPListener).accept(0xc0004f1900) /usr/lib/go/src/net/tcpsock_posix.go:159 +0x1b fp=0xc0004c39c0 sp=0xc0004c3970 pc=0x5b369b net.(*TCPListener).Accept(0xc0004f1900) /usr/lib/go/src/net/tcpsock.go:380 +0x30 fp=0xc0004c39f0 sp=0xc0004c39c0 pc=0x5b2530 net/http.(*onceCloseListener).Accept(0xc0000f1c20?) <autogenerated>:1 +0x24 fp=0xc0004c3a08 sp=0xc0004c39f0 pc=0x7d4d24 net/http.(*Server).Serve(0xc0005c2e00, {0x1836570, 0xc0004f1900}) /usr/lib/go/src/net/http/server.go:3463 +0x30c fp=0xc0004c3b38 sp=0xc0004c3a08 pc=0x7ac70c github.com/ollama/ollama/runner/ollamarunner.Execute({0xc000036080, 0x2, 0x2}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1368 +0x954 fp=0xc0004c3d08 sp=0xc0004c3b38 pc=0x9c0734 github.com/ollama/ollama/runner.Execute({0xc000036060?, 0x0?, 0x0?}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/runner.go:20 +0xc9 fp=0xc0004c3d30 sp=0xc0004c3d08 pc=0x9c1029 github.com/ollama/ollama/cmd.NewCLI.func2(0xc0005c2b00?, {0x1657910?, 0x4?, 0x1657914?}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/cmd/cmd.go:1769 +0x45 fp=0xc0004c3d58 sp=0xc0004c3d30 pc=0x11341c5 github.com/spf13/cobra.(*Command).execute(0xc0005c5b08, {0xc0005c0d50, 0x3, 0x3}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:940 +0x88a fp=0xc0004c3e78 sp=0xc0004c3d58 pc=0x61772a github.com/spf13/cobra.(*Command).ExecuteC(0xc00048ef08) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:1068 +0x398 fp=0xc0004c3f30 sp=0xc0004c3e78 pc=0x617f58 github.com/spf13/cobra.(*Command).Execute(...) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:992 github.com/spf13/cobra.(*Command).ExecuteContext(...) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:985 main.main() /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/main.go:12 +0x4d fp=0xc0004c3f50 sp=0xc0004c3f30 pc=0x1134c8d runtime.main() /usr/lib/go/src/runtime/proc.go:285 +0x29d fp=0xc0004c3fe0 sp=0xc0004c3f50 pc=0x471b5d runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0004c3fe8 sp=0xc0004c3fe0 pc=0x4ae701 goroutine 2 gp=0xc000002e00 m=nil [force gc (idle)]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007cfa8 sp=0xc00007cf88 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.forcegchelper() /usr/lib/go/src/runtime/proc.go:373 +0xb3 fp=0xc00007cfe0 sp=0xc00007cfa8 pc=0x471e93 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007cfe8 sp=0xc00007cfe0 pc=0x4ae701 created by runtime.init.7 in goroutine 1 /usr/lib/go/src/runtime/proc.go:361 +0x1a goroutine 3 gp=0xc000003340 m=nil [GC sweep wait]: runtime.gopark(0x1?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007d780 sp=0xc00007d760 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.bgsweep(0xc0000a8000) /usr/lib/go/src/runtime/mgcsweep.go:323 +0xdf fp=0xc00007d7c8 sp=0xc00007d780 pc=0x45bc1f runtime.gcenable.gowrap1() /usr/lib/go/src/runtime/mgc.go:212 +0x25 fp=0xc00007d7e0 sp=0xc00007d7c8 pc=0x44fc05 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007d7e8 sp=0xc00007d7e0 pc=0x4ae701 created by runtime.gcenable in goroutine 1 /usr/lib/go/src/runtime/mgc.go:212 +0x66 goroutine 4 gp=0xc000003500 m=nil [GC scavenge wait]: runtime.gopark(0x10000?, 0x1822d88?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007df78 sp=0xc00007df58 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.(*scavengerState).park(0x21d62a0) /usr/lib/go/src/runtime/mgcscavenge.go:425 +0x49 fp=0xc00007dfa8 sp=0xc00007df78 pc=0x459689 runtime.bgscavenge(0xc0000a8000) /usr/lib/go/src/runtime/mgcscavenge.go:658 +0x59 fp=0xc00007dfc8 sp=0xc00007dfa8 pc=0x459c39 runtime.gcenable.gowrap2() /usr/lib/go/src/runtime/mgc.go:213 +0x25 fp=0xc00007dfe0 sp=0xc00007dfc8 pc=0x44fba5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007dfe8 sp=0xc00007dfe0 pc=0x4ae701 created by runtime.gcenable in goroutine 1 /usr/lib/go/src/runtime/mgc.go:213 +0xa5 goroutine 5 gp=0xc000003dc0 m=nil [finalizer wait]: runtime.gopark(0x480e35?, 0x447565?, 0xb8?, 0x1?, 0xc000002380?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007c620 sp=0xc00007c600 pc=0x4a656e runtime.runFinalizers() /usr/lib/go/src/runtime/mfinal.go:210 +0x107 fp=0xc00007c7e0 sp=0xc00007c620 pc=0x44eb07 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007c7e8 sp=0xc00007c7e0 pc=0x4ae701 created by runtime.createfing in goroutine 1 /usr/lib/go/src/runtime/mfinal.go:172 +0x3d goroutine 6 gp=0xc0001d28c0 m=nil [cleanup wait]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007e768 sp=0xc00007e748 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.(*cleanupQueue).dequeue(0x21d6c00) /usr/lib/go/src/runtime/mcleanup.go:439 +0xc5 fp=0xc00007e7a0 sp=0xc00007e768 pc=0x44bd25 runtime.runCleanups() /usr/lib/go/src/runtime/mcleanup.go:635 +0x45 fp=0xc00007e7e0 sp=0xc00007e7a0 pc=0x44c3e5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007e7e8 sp=0xc00007e7e0 pc=0x4ae701 created by runtime.(*cleanupQueue).createGs in goroutine 1 /usr/lib/go/src/runtime/mcleanup.go:589 +0xa5 goroutine 7 gp=0xc0001d3180 m=nil [GC worker (idle)]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007ef38 sp=0xc00007ef18 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007efc8 sp=0xc00007ef38 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007efe0 sp=0xc00007efc8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007efe8 sp=0xc00007efe0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 18 gp=0xc000504000 m=nil [GC worker (idle)]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000078738 sp=0xc000078718 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0000787c8 sp=0xc000078738 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0000787e0 sp=0xc0000787c8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0000787e8 sp=0xc0000787e0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 34 gp=0xc000102380 m=nil [GC worker (idle)]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000118738 sp=0xc000118718 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001187c8 sp=0xc000118738 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001187e0 sp=0xc0001187c8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001187e8 sp=0xc0001187e0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 8 gp=0xc0001d36c0 m=nil [GC worker (idle)]: runtime.gopark(0x59364534a969?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007f738 sp=0xc00007f718 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007f7c8 sp=0xc00007f738 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007f7e0 sp=0xc00007f7c8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007f7e8 sp=0xc00007f7e0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 9 gp=0xc0001d3880 m=nil [GC worker (idle)]: runtime.gopark(0x59364534a991?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007ff38 sp=0xc00007ff18 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007ffc8 sp=0xc00007ff38 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007ffe0 sp=0xc00007ffc8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007ffe8 sp=0xc00007ffe0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 10 gp=0xc0001d3a40 m=nil [GC worker (idle)]: runtime.gopark(0x593645401bdb?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000114738 sp=0xc000114718 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001147c8 sp=0xc000114738 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001147e0 sp=0xc0001147c8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001147e8 sp=0xc0001147e0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 11 gp=0xc000588380 m=nil [sync.WaitGroup.Wait]: runtime.gopark(0x0?, 0x0?, 0x0?, 0xc0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00008ca90 sp=0xc00008ca70 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.semacquire1(0xc00022b198, 0x0, 0x1, 0x0, 0x19) /usr/lib/go/src/runtime/sema.go:192 +0x21d fp=0xc00008caf8 sp=0xc00008ca90 pc=0x48589d sync.runtime_SemacquireWaitGroup(0x0?, 0x0?) /usr/lib/go/src/runtime/sema.go:114 +0x2e fp=0xc00008cb30 sp=0xc00008caf8 pc=0x4a806e sync.(*WaitGroup).Wait(0xc00022b190) /usr/lib/go/src/sync/waitgroup.go:206 +0x85 fp=0xc00008cb58 sp=0xc00008cb30 pc=0x4ba365 github.com/ollama/ollama/runner/ollamarunner.(*Server).run(0xc00022b0e0, {0x1838b40, 0xc0005dc000}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:415 +0x45 fp=0xc00008cfb8 sp=0xc00008cb58 pc=0x9b7605 github.com/ollama/ollama/runner/ollamarunner.Execute.gowrap1() /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x28 fp=0xc00008cfe0 sp=0xc00008cfb8 pc=0x9c09a8 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00008cfe8 sp=0xc00008cfe0 pc=0x4ae701 created by github.com/ollama/ollama/runner/ollamarunner.Execute in goroutine 1 /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x4c9 goroutine 12 gp=0xc0005041c0 m=nil [IO wait]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0xb?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000117dd8 sp=0xc000117db8 pc=0x4a656e runtime.netpollblock(0x4cab98?, 0x43b5e6?, 0x0?) /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc000117e10 sp=0xc000117dd8 pc=0x46a317 internal/poll.runtime_pollWait(0x7f71d58b5200, 0x72) /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc000117e30 sp=0xc000117e10 pc=0x4a5745 internal/poll.(*pollDesc).wait(0xc00059f300?, 0xc0000aa021?, 0x0) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc000117e58 sp=0xc000117e30 pc=0x52e627 internal/poll.(*pollDesc).waitRead(...) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0xc00059f300, {0xc0000aa021, 0x1, 0x1}) /usr/lib/go/src/internal/poll/fd_unix.go:165 +0x279 fp=0xc000117ef0 sp=0xc000117e58 pc=0x52f919 net.(*netFD).Read(0xc00059f300, {0xc0000aa021?, 0x0?SIGABRT: abort PC=0x7fb05b8a489c m=5 sigcode=18446744073709551610 signal arrived during cgo execution goroutine 38 gp=0xc0001d5a40 m=5 mp=0xc000083808 [syscall]: runtime.cgocall(0x11ad7b0, 0xc000058838) , /usr/lib/go/src/runtime/cgocall.go:167 +0x00x4b? fp=}0xc000058810) sp= 0xc0000587d8/usr/lib/go/src/net/fd_posix.go pc=:0x4a30eb 68 +0x25 fp=0xc000117f38 sp=github.com/ollama/ollama/ml/backend/ggml/ggml/src._Cfunc_ggml_backend_load_all_from_path0xc000117ef0( pc=0x7fb008000be00x59c105) _cgo_gotypes.go:net.(*conn).Read195( +0x3a0xc00011e000 fp=, 0xc000058838{ sp=0xc0000aa0210xc000058810? pc=, 0x857a5a0x0 ?, github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1.10x0(?{}0xc00003e134) , 0x11/usr/lib/go/src/net/net.go}:) 196 +/var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go0x45: fp=970xc000117f80 + sp=0xf50xc000117f38 fp= pc=0xc0000588d00x5aa105 sp= 0xc000058838 pc=0x857515net/http.(*connReader).backgroundRead (0xc0000aa000github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.func1) ( ) /usr/lib/go/src/net/http/server.go :/var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml/src/ggml.go702: +980x33 + fp=0x5450xc000117fc8 fp= sp=0xc000058b480xc000117f80 sp= pc=0xc0000588d00x7a12f3 pc= 0x857365 net/http.(*connReader).startBackgroundRead.gowrap2(github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func2.1) ( ) /usr/lib/go/src/net/http/server.go :/usr/lib/go/src/sync/oncefunc.go698: +330x25 + fp=0x4b0xc000117fe0 fp= sp=0xc000058b800xc000117fc8 sp= pc=0xc000058b480x7a1225 pc= 0x856d6b runtime.goexit(sync.(*Once).doSlow{(}0x2283fc0) ? , /usr/lib/go/src/runtime/asm_amd64.s0x0:?1693) + 0x1/usr/lib/go/src/sync/once.go fp=:0xc000117fe878 sp= +0xc000117fe00xac pc= fp=0x4ae7010xc000058bd8 sp=created by 0xc000058b80net/http.(*connReader).startBackgroundRead pc= in goroutine 0x4b8dec35 sync.(*Once).Do (/usr/lib/go/src/net/http/server.go0xc000058c78:?698, +0xc0xb6? ) rax /usr/lib/go/src/sync/once.go0x0: 69rbx +0x2c9d0x19 fp=rcx 0xc000058bf80x7f71d5ad289c sp= 0xc000058bd8rdx pc=0x60x4b8d19 rdi github.com/ollama/ollama/ml/backend/ggml/ggml/src.init.OnceFunc.func20x2c9d( ) rsi 0x2c9d/usr/lib/go/src/sync/oncefunc.go :rbp 220x7f7182965286 + 0x39rsp fp=0x7ffc3726e0d00xc000058c30 sp=r8 0xc000058bf80x0 pc= 0x856cf9r9 0x0 github.com/ollama/ollama/ml/backend/ggml.init.func1r10 (0x0) r11 /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go0x246: 46r12 +0x1490x1e fp=r13 0xc000058cb80x7f718291eff0 sp= 0xc000058c30r14 pc=0x60x8e319e r15 github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2.10x7ffc3726f7c0( ) rip 0x7f71d5ad289c/usr/lib/go/src/sync/oncefunc.go :rflags 330x246 + 0x4bcs fp=0x330xc000058cf0 sp=fs 0xc000058cb80x0 pc= 0x8e30cbgs 0x0sync.(*Once).doSlow (0x8000?, 0x100000182ec00?) /usr/lib/go/src/sync/once.go:78 +0xac fp=0xc000058d48 sp=0xc000058cf0 pc=0x4b8dec sync.(*Once).Do(0x22845d0?, 0x4b8ea0?) /usr/lib/go/src/sync/once.go:69 +0x19 fp=0xc000058d68 sp=0xc000058d48 pc=0x4b8d19 github.com/ollama/ollama/ml/backend/ggml.init.OnceFunc.func2() /usr/lib/go/src/sync/oncefunc.go:22 +0x39 fp=0xc000058da0 sp=0xc000058d68 pc=0x8e3059 github.com/ollama/ollama/ml/backend/ggml.New({0xc0003b0900, 0x13}, {0x0, 0x6, {0xc000688940, 0x1, 0x1}, 0x0}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend/ggml/ggml.go:145 +0x124 fp=0xc000059678 sp=0xc000058da0 pc=0x8eb564 github.com/ollama/ollama/ml.NewBackend({0xc0003b0900, 0x13}, {0x0, 0x6, {0xc000688940, 0x1, 0x1}, 0x0}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/ml/backend.go:92 +0x9c fp=0xc0000596c8 sp=0xc000059678 pc=0x85941c github.com/ollama/ollama/model.New({0xc0003b0900?, 0xc000165978?}, {0x0, 0x6, {0xc000688940, 0x1, 0x1}, 0x0}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/model/model.go:106 +0x66 fp=0xc000059788 sp=0xc0000596c8 pc=0x900346 github.com/ollama/ollama/runner/ollamarunner.(*Server).info(0xc00022af00, {0x1836750, 0xc0005332c0}, 0xc000503140?) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1302 +0x533 fp=0xc000059ab8 sp=0xc000059788 pc=0x9bf833 github.com/ollama/ollama/runner/ollamarunner.(*Server).info-fm({0x1836750?, 0xc0005332c0?}, 0xc000165b30?) <autogenerated>:1 +0x36 fp=0xc000059ae8 sp=0xc000059ab8 pc=0x9c0d16 net/http.HandlerFunc.ServeHTTP(0xc000543140?, {0x1836750?, 0xc0005332c0?}, 0x7a1196?) /usr/lib/go/src/net/http/server.go:2322 +0x29 fp=0xc000059b10 sp=0xc000059ae8 pc=0x7a8d49 net/http.(*ServeMux).ServeHTTP(0x4a2fb9?, {0x1836750, 0xc0005332c0}, 0xc000379540) /usr/lib/go/src/net/http/server.go:2861 +0x1c7 fp=0xc000059b60 sp=0xc000059b10 pc=0x7aac27 net/http.serverHandler.ServeHTTP({0xc000688780?}, {0x1836750?, 0xc0005332c0?}, 0x1?) /usr/lib/go/src/net/http/server.go:3340 +0x8e fp=0xc000059b90 sp=0xc000059b60 pc=0x7c850e net/http.(*conn).serve(0xc00012a3f0, {0x1838b08, 0xc000216570}) /usr/lib/go/src/net/http/server.go:2109 +0x665 fp=0xc000059fb8 sp=0xc000059b90 pc=0x7a6e45 net/http.(*Server).Serve.gowrap3() /usr/lib/go/src/net/http/server.go:3493 +0x28 fp=0xc000059fe0 sp=0xc000059fb8 pc=0x7acb08 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000059fe8 sp=0xc000059fe0 pc=0x4ae701 created by net/http.(*Server).Serve in goroutine 1 /usr/lib/go/src/net/http/server.go:3493 +0x485 goroutine 1 gp=0xc000002380 m=nil [IO wait]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc0004af790 sp=0xc0004af770 pc=0x4a656e runtime.netpollblock(0xc0004af7e0?, 0x43b5e6?, 0x0?) /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc0004af7c8 sp=0xc0004af790 pc=0x46a317 internal/poll.runtime_pollWait(0x7fb05b677400, 0x72) /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc0004af7e8 sp=0xc0004af7c8 pc=0x4a5745 internal/poll.(*pollDesc).wait(0xc000121600?, 0x900000036?, 0x0) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc0004af810 sp=0xc0004af7e8 pc=0x52e627 internal/poll.(*pollDesc).waitRead(...) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Accept(0xc000121600) /usr/lib/go/src/internal/poll/fd_unix.go:613 +0x28c fp=0xc0004af8b8 sp=0xc0004af810 pc=0x533a4c net.(*netFD).accept(0xc000121600) /usr/lib/go/src/net/fd_unix.go:161 +0x29 fp=0xc0004af970 sp=0xc0004af8b8 pc=0x59dfa9 net.(*TCPListener).accept(0xc000688740) /usr/lib/go/src/net/tcpsock_posix.go:159 +0x1b fp=0xc0004af9c0 sp=0xc0004af970 pc=0x5b369b net.(*TCPListener).Accept(0xc000688740) /usr/lib/go/src/net/tcpsock.go:380 +0x30 fp=0xc0004af9f0 sp=0xc0004af9c0 pc=0x5b2530 net/http.(*onceCloseListener).Accept(0xc00012a3f0?) <autogenerated>:1 +0x24 fp=0xc0004afa08 sp=0xc0004af9f0 pc=0x7d4d24 net/http.(*Server).Serve(0xc0001f5700, {0x1836570, 0xc000688740}) /usr/lib/go/src/net/http/server.go:3463 +0x30c fp=0xc0004afb38 sp=0xc0004afa08 pc=0x7ac70c github.com/ollama/ollama/runner/ollamarunner.Execute({0xc000036080, 0x2, 0x2}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1368 +0x954 fp=0xc0004afd08 sp=0xc0004afb38 pc=0x9c0734 github.com/ollama/ollama/runner.Execute({0xc000036060?, 0x0?, 0x0?}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/runner.go:20 +0xc9 fp=0xc0004afd30 sp=0xc0004afd08 pc=0x9c1029 github.com/ollama/ollama/cmd.NewCLI.func2(0xc0001f5400?, {0x1657910?, 0x4?, 0x1657914?}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/cmd/cmd.go:1769 +0x45 fp=0xc0004afd58 sp=0xc0004afd30 pc=0x11341c5 github.com/spf13/cobra.(*Command).execute(0xc00069fb08, {0xc000216000, 0x3, 0x3}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:940 +0x88a fp=0xc0004afe78 sp=0xc0004afd58 pc=0x61772a github.com/spf13/cobra.(*Command).ExecuteC(0xc00049ac08) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:1068 +0x398 fp=0xc0004aff30 sp=0xc0004afe78 pc=0x617f58 github.com/spf13/cobra.(*Command).Execute(...) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:992 github.com/spf13/cobra.(*Command).ExecuteContext(...) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/vendor/github.com/spf13/cobra/command.go:985 main.main() /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/main.go:12 +0x4d fp=0xc0004aff50 sp=0xc0004aff30 pc=0x1134c8d runtime.main() /usr/lib/go/src/runtime/proc.go:285 +0x29d fp=0xc0004affe0 sp=0xc0004aff50 pc=0x471b5d runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0004affe8 sp=0xc0004affe0 pc=0x4ae701 goroutine 2 gp=0xc000002e00 m=nil [force gc (idle)]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007cfa8 sp=0xc00007cf88 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.forcegchelper() /usr/lib/go/src/runtime/proc.go:373 +0xb3 fp=0xc00007cfe0 sp=0xc00007cfa8 pc=0x471e93 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007cfe8 sp=0xc00007cfe0 pc=0x4ae701 created by runtime.init.7 in goroutine 1 /usr/lib/go/src/runtime/proc.go:361 +0x1a goroutine 3 gp=0xc000003340 m=nil [GC sweep wait]: runtime.gopark(0x1?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007d780 sp=0xc00007d760 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.bgsweep(0xc0000a8000) /usr/lib/go/src/runtime/mgcsweep.go:323 +0xdf fp=0xc00007d7c8 sp=0xc00007d780 pc=0x45bc1f runtime.gcenable.gowrap1() /usr/lib/go/src/runtime/mgc.go:212 +0x25 fp=0xc00007d7e0 sp=0xc00007d7c8 pc=0x44fc05 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007d7e8 sp=0xc00007d7e0 pc=0x4ae701 created by runtime.gcenable in goroutine 1 /usr/lib/go/src/runtime/mgc.go:212 +0x66 goroutine 4 gp=0xc000003500 m=nil [GC scavenge wait]: runtime.gopark(0x10000?, 0x1822d88?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007df78 sp=0xc00007df58 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.(*scavengerState).park(0x21d62a0) /usr/lib/go/src/runtime/mgcscavenge.go:425 +0x49 fp=0xc00007dfa8 sp=0xc00007df78 pc=0x459689 runtime.bgscavenge(0xc0000a8000) /usr/lib/go/src/runtime/mgcscavenge.go:658 +0x59 fp=0xc00007dfc8 sp=0xc00007dfa8 pc=0x459c39 runtime.gcenable.gowrap2() /usr/lib/go/src/runtime/mgc.go:213 +0x25 fp=0xc00007dfe0 sp=0xc00007dfc8 pc=0x44fba5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007dfe8 sp=0xc00007dfe0 pc=0x4ae701 created by runtime.gcenable in goroutine 1 /usr/lib/go/src/runtime/mgc.go:213 +0xa5 goroutine 5 gp=0xc000003dc0 m=nil [finalizer wait]: runtime.gopark(0x480e35?, 0x447565?, 0xb8?, 0x1?, 0xc000002380?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007c620 sp=0xc00007c600 pc=0x4a656e runtime.runFinalizers() /usr/lib/go/src/runtime/mfinal.go:210 +0x107 fp=0xc00007c7e0 sp=0xc00007c620 pc=0x44eb07 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007c7e8 sp=0xc00007c7e0 pc=0x4ae701 created by runtime.createfing in goroutine 1 /usr/lib/go/src/runtime/mfinal.go:172 +0x3d goroutine 6 gp=0xc0001d48c0 m=nil [cleanup wait]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007e768 sp=0xc00007e748 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.(*cleanupQueue).dequeue(0x21d6c00) /usr/lib/go/src/runtime/mcleanup.go:439 +0xc5 fp=0xc00007e7a0 sp=0xc00007e768 pc=0x44bd25 runtime.runCleanups() /usr/lib/go/src/runtime/mcleanup.go:635 +0x45 fp=0xc00007e7e0 sp=0xc00007e7a0 pc=0x44c3e5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007e7e8 sp=0xc00007e7e0 pc=0x4ae701 created by runtime.(*cleanupQueue).createGs in goroutine 1 /usr/lib/go/src/runtime/mcleanup.go:589 +0xa5 goroutine 7 gp=0xc0001d5340 m=nil [GC worker (idle)]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00007ef38 sp=0xc00007ef18 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00007efc8 sp=0xc00007ef38 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00007efe0 sp=0xc00007efc8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00007efe8 sp=0xc00007efe0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 18 gp=0xc000504000 m=nil [GC worker (idle)]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000078738 sp=0xc000078718 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0000787c8 sp=0xc000078738 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0000787e0 sp=0xc0000787c8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0000787e8 sp=0xc0000787e0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 34 gp=0xc000102380 m=nil [GC worker (idle)]: runtime.gopark(0x5936452d22e8?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000118738 sp=0xc000118718 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001187c8 sp=0xc000118738 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001187e0 sp=0xc0001187c8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001187e8 sp=0xc0001187e0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 35 gp=0xc000102540 m=nil [GC worker (idle)]: runtime.gopark(0x5936452d2bee?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000118f38 sp=0xc000118f18 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc000118fc8 sp=0xc000118f38 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc000118fe0 sp=0xc000118fc8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000118fe8 sp=0xc000118fe0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 36 gp=0xc000102700 m=nil [GC worker (idle)]: runtime.gopark(0x5936452d22d4?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000119738 sp=0xc000119718 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc0001197c8 sp=0xc000119738 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc0001197e0 sp=0xc0001197c8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc0001197e8 sp=0xc0001197e0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 8 gp=0xc0001d5500 m=nil [GC worker (idle)]: runtime.gopark(0x5936452d37d8?, 0x0?, 0x0?, 0x0?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00008ff38 sp=0xc00008ff18 pc=0x4a656e runtime.gcBgMarkWorker(0xc0000b76c0) /usr/lib/go/src/runtime/mgc.go:1463 +0xeb fp=0xc00008ffc8 sp=0xc00008ff38 pc=0x4522eb runtime.gcBgMarkStartWorkers.gowrap1() /usr/lib/go/src/runtime/mgc.go:1373 +0x25 fp=0xc00008ffe0 sp=0xc00008ffc8 pc=0x4521c5 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00008ffe8 sp=0xc00008ffe0 pc=0x4ae701 created by runtime.gcBgMarkStartWorkers in goroutine 1 /usr/lib/go/src/runtime/mgc.go:1373 +0x105 goroutine 37 gp=0xc0001d5880 m=nil [sync.WaitGroup.Wait]: runtime.gopark(0x0?, 0x0?, 0xe0?, 0xc1?, 0x0?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc00008ea90 sp=0xc00008ea70 pc=0x4a656e runtime.goparkunlock(...) /usr/lib/go/src/runtime/proc.go:466 runtime.semacquire1(0xc00022afb8, 0x0, 0x1, 0x0, 0x19) /usr/lib/go/src/runtime/sema.go:192 +0x21d fp=0xc00008eaf8 sp=0xc00008ea90 pc=0x48589d sync.runtime_SemacquireWaitGroup(0x0?, 0x0?) /usr/lib/go/src/runtime/sema.go:114 +0x2e fp=0xc00008eb30 sp=0xc00008eaf8 pc=0x4a806e sync.(*WaitGroup).Wait(0xc00022afb0) /usr/lib/go/src/sync/waitgroup.go:206 +0x85 fp=0xc00008eb58 sp=0xc00008eb30 pc=0x4ba365 github.com/ollama/ollama/runner/ollamarunner.(*Server).run(0xc00022af00, {0x1838b40, 0xc00068c230}) /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:415 +0x45 fp=0xc00008efb8 sp=0xc00008eb58 pc=0x9b7605 github.com/ollama/ollama/runner/ollamarunner.Execute.gowrap1() /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x28 fp=0xc00008efe0 sp=0xc00008efb8 pc=0x9c09a8 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc00008efe8 sp=0xc00008efe0 pc=0x4ae701 created by github.com/ollama/ollama/runner/ollamarunner.Execute in goroutine 1 /var/tmp/portage/sci-ml/ollama-9999/work/ollama-9999/runner/ollamarunner/runner.go:1345 +0x4c9 goroutine 39 gp=0xc0001d5c00 m=nil [IO wait]: runtime.gopark(0x0?, 0x0?, 0x0?, 0x0?, 0xb?) /usr/lib/go/src/runtime/proc.go:460 +0xce fp=0xc000078dd8 sp=0xc000078db8 pc=0x4a656e runtime.netpollblock(0x4cab98?, 0x43b5e6?, 0x0?) /usr/lib/go/src/runtime/netpoll.go:575 +0xf7 fp=0xc000078e10 sp=0xc000078dd8 pc=0x46a317 internal/poll.runtime_pollWait(0x7fb05b677200, 0x72) /usr/lib/go/src/runtime/netpoll.go:351 +0x85 fp=0xc000078e30 sp=0xc000078e10 pc=0x4a5745 internal/poll.(*pollDesc).wait(0xc000121680?, 0xc0006887a1?, 0x0) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:84 +0x27 fp=0xc000078e58 sp=0xc000078e30 pc=0x52e627 internal/poll.(*pollDesc).waitRead(...) /usr/lib/go/src/internal/poll/fd_poll_runtime.go:89 internal/poll.(*FD).Read(0xc000121680, {0xc0006887a1, 0x1, 0x1}) /usr/lib/go/src/internal/poll/fd_unix.go:165 +0x279 fp=0xc000078ef0 sp=0xc000078e58 pc=0x52f919 net.(*netFD).Read(0xc000121680, {0xc0006887a1?, 0x0?, 0x0?}) /usr/lib/go/src/net/fd_posix.go:68 +0x25 fp=0xc000078f38 sp=0xc000078ef0 pc=0x59c105 net.(*conn).Read(0xc0000806e8, {0xc0006887a1?, 0x0?, 0x0?}) /usr/lib/go/src/net/net.go:196 +0x45 fp=0xc000078f80 sp=0xc000078f38 pc=0x5aa105 net/http.(*connReader).backgroundRead(0xc000688780) /usr/lib/go/src/net/http/server.go:702 +0x33 fp=0xc000078fc8 sp=0xc000078f80 pc=0x7a12f3 net/http.(*connReader).startBackgroundRead.gowrap2() /usr/lib/go/src/net/http/server.go:698 +0x25 fp=0xc000078fe0 sp=0xc000078fc8 pc=0x7a1225 runtime.goexit({}) /usr/lib/go/src/runtime/asm_amd64.s:1693 +0x1 fp=0xc000078fe8 sp=0xc000078fe0 pc=0x4ae701 created by net/http.(*connReader).startBackgroundRead in goroutine 38 /usr/lib/go/src/net/http/server.go:698 +0xb6 rax 0x0 rbx 0x2ca3 rcx 0x7fb05b8a489c rdx 0x6 rdi 0x2c9e rsi 0x2ca3 rbp 0x7fafff165286 rsp 0x7fb013b9b660 r8 0x0 r9 0x0 r10 0x0 r11 0x246 r12 0x149 r13 0x7fafff11eff0 r14 0x6 r15 0x7fb013b9cd50 rip 0x7fb05b8a489c rflags 0x246 cs 0x33 fs 0x0 gs 0x0 time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:543 msg="runner exited" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]" code=2 time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] devices=[] time=2025-10-18T10:08:22.061-05:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=70.942693ms OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 CUDA_VISIBLE_DEVICES=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d]" time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:543 msg="runner exited" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]" code=2 time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] devices=[] time=2025-10-18T10:08:22.061-05:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=71.139921ms OLLAMA_LIBRARY_PATH=[/usr/lib64/ollama] extra_envs="[GGML_CUDA_INIT=1 GGML_VK_VISIBLE_DEVICES=0]" time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:182 msg="removing unsupported or overlapping GPU combination" libDir=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=12.0 pci_id=01:00.0 time=2025-10-18T10:08:22.061-05:00 level=TRACE source=runner.go:182 msg="removing unsupported or overlapping GPU combination" libDir=/usr/lib64/ollama description="NVIDIA GeForce RTX 5070 Ti" compute=0.0 pci_id=01:00.0 time=2025-10-18T10:08:22.061-05:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=303.766267ms time=2025-10-18T10:08:22.061-05:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="123.4 GiB" available="120.1 GiB" time=2025-10-18T10:08:22.061-05:00 level=INFO source=routes.go:1609 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ``` However version 0.12.3 from guru overlay (same use flags, blas+cuda) it detects the GPU without issue: ``` $ ollama serve time=2025-10-18T09:44:05.937-05:00 level=INFO source=routes.go:1475 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/home/XXX/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-10-18T09:44:05.938-05:00 level=INFO source=routes.go:1528 msg="Listening on 127.0.0.1:11434 (version 0.12.3)" time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler" time=2025-10-18T09:44:05.938-05:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=gpu.go:520 msg="Searching for GPU library" name=libcuda.so* time=2025-10-18T09:44:05.938-05:00 level=DEBUG source=gpu.go:544 msg="gpu library search" globs="[/usr/lib64/ollama/libcuda.so* /home/XXX/libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]" time=2025-10-18T09:44:05.942-05:00 level=DEBUG source=gpu.go:577 msg="discovered GPU libraries" paths="[/usr/lib/libcuda.so.580.95.05 /usr/lib64/libcuda.so.580.95.05]" initializing /usr/lib/libcuda.so.580.95.05 library /usr/lib/libcuda.so.580.95.05 load err: /usr/lib/libcuda.so.580.95.05: wrong ELF class: ELFCLASS32 time=2025-10-18T09:44:05.942-05:00 level=DEBUG source=gpu.go:628 msg="skipping 32bit library" library=/usr/lib/libcuda.so.580.95.05 initializing /usr/lib64/libcuda.so.580.95.05 dlsym: cuInit - 0x7fb131d05d00 dlsym: cuDriverGetVersion - 0x7fb131d05dc0 dlsym: cuDeviceGetCount - 0x7fb131d05f40 dlsym: cuDeviceGet - 0x7fb131d05e80 dlsym: cuDeviceGetAttribute - 0x7fb131d28f20 dlsym: cuDeviceGetUuid - 0x7fb131d7bd10 dlsym: cuDeviceGetName - 0x7fb131d06000 dlsym: cuCtxCreate_v3 - 0x7fb131d79750 dlsym: cuMemGetInfo_v2 - 0x7fb131d2d460 dlsym: cuCtxDestroy - 0x7fb131d7b650 calling cuInit calling cuDriverGetVersion raw version 0x32c8 CUDA driver version: 13.0 calling cuDeviceGetCount device count 1 time=2025-10-18T09:44:05.963-05:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=1 library=/usr/lib64/libcuda.so.580.95.05 [GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d] CUDA totalMem 15806mb [GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d] CUDA freeMem 15347mb [GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d] Compute Capability 12.0 time=2025-10-18T09:44:06.081-05:00 level=WARN source=amd_linux.go:61 msg="ollama recommends running the https://www.amd.com/en/support/download/linux-drivers.html" error="amdgpu version file missing: /sys/module/amdgpu/version stat /sys/module/amdgpu/version: no such file or directory" time=2025-10-18T09:44:06.081-05:00 level=INFO source=amd_linux.go:406 msg="no compatible amdgpu devices detected" releasing cuda driver library time=2025-10-18T09:44:06.081-05:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-91a8a7fb-08c1-2a27-9a63-8143ff8da46d library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070 Ti" total="15.4 GiB" available="15.0 GiB" time=2025-10-18T09:44:06.081-05:00 level=INFO source=routes.go:1569 msg="entering low vram mode" "total vram"="15.4 GiB" threshold="20.0 GiB" ``` Edited to include debug level 2 log. Thanks for your help with this! Let me know if I can provide any more info.
Author
Owner

@dhiltgen commented on GitHub (Oct 18, 2025):

@aminalshmu do the official binaries correctly discover your GPU? It looks like your build didn't include the compute capability for your GPU. How did you build the native components? What CUDA version do you have installed?

<!-- gh-comment-id:3418730137 --> @dhiltgen commented on GitHub (Oct 18, 2025): @aminalshmu do the official binaries correctly discover your GPU? It looks like your build didn't include the compute capability for your GPU. How did you build the native components? What CUDA version do you have installed?
Author
Owner

@aminalshmu commented on GitHub (Oct 18, 2025):

Yes, official binary package detects GPU. I have been using the guru gentoo overlay, the -9999 package fetches current repo at build time:

https://github.com/gentoo/guru/tree/master/sci-ml/ollama

It was working with the live builds until about a week ago, around 0.12.5 I believe, and the guru ebuild has not changed in 2 weeks, seemed to work fine with source versions around 0.12.4. I have CUDA 13.0 and NVidia driver 580.95.05 FWIW.

I'm thinking this is a completely separate issue from catclaw as the Windows binary does not detect his GPUs but the 0.12.6 amd64 package works fine for me on Linux, just not building from source. Will use the binary for now but would like to figure out why the source version isn't working.

<!-- gh-comment-id:3418747266 --> @aminalshmu commented on GitHub (Oct 18, 2025): Yes, official binary package detects GPU. I have been using the guru gentoo overlay, the -9999 package fetches current repo at build time: https://github.com/gentoo/guru/tree/master/sci-ml/ollama It was working with the live builds until about a week ago, around 0.12.5 I believe, and the guru ebuild has not changed in 2 weeks, seemed to work fine with source versions around 0.12.4. I have CUDA 13.0 and NVidia driver 580.95.05 FWIW. I'm thinking this is a completely separate issue from catclaw as the Windows binary does not detect his GPUs but the 0.12.6 amd64 package works fine for me on Linux, just not building from source. Will use the binary for now but would like to figure out why the source version isn't working.
Author
Owner

@miversen33 commented on GitHub (Oct 19, 2025):

Disregard the below. In my case it was a stupid cgroup oversight with nvidia-uvm.


I am noticing identical behavior with Ollama 0.12.5 and 0.12.6 on linux.

I am unsure if my issue is the same, but the result is the same. Ollama seems to refuse to talk to my 2 P40 GPUs. Been pulling my hair out for the past few days trying to debug it but from everything I can tell, the issue is simply that ollama doesn't like the GPUs.

In my case I installed with the installation script found here and when I run it, I notice that ollama doesn't have runners for anything pre cuda_12. For my GPUs, they run cuda_v11 and it seems that while Ollama can see the GPUS, it doesn't know how to talk to them as there is no runner for them

Github is fighting me and won't let me upload the log as a file so here is my log, maybe it will help shed some light on stuff? time=2025-10-19T14:59:21.474Z level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-10-19T14:59:21.474Z level=INFO source=images.go:522 msg="total blobs: 0" time=2025-10-19T14:59:21.474Z level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-19T14:59:21.474Z level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-19T14:59:21.475Z level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-19T14:59:21.475Z level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-19T14:59:21.475Z level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" extra_envs=[] time=2025-10-19T14:59:21.476Z level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash CGO_LDFLAGS=-L/opt/ollama-source/build/lib/ollama -L/usr/local/cuda-11.8/lib64 PWD=/usr/local/lib/ollama LOGNAME=root XDG_SESSION_TYPE=tty CXX=/usr/bin/g++-11 HOME=/root LANG=C LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: SSH_CONNECTION=10.40.1.52 34792 10.40.1.20 22 CUDAHOSTCXX=/usr/bin/g++-11 LESSCLOSE=/usr/bin/lesspipe %s %s XDG_SESSION_CLASS=user TERM=tmux-256color LESSOPEN=| /usr/bin/lesspipe %s USER=root SHLVL=1 XDG_SESSION_ID=3329 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v12:/opt/ollama-source/build/lib/ollama:/usr/local/cuda-11.8/lib64:/usr/local/cuda-11.8/lib64: XDG_RUNTIME_DIR=/run/user/0 SSH_CLIENT=10.40.1.52 34792 22 CUDA_HOME=/usr/local/cuda-11.8 PATH=/usr/local/cuda-11.8/bin:/usr/local/cuda-11.8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin CC=/usr/bin/gcc-11 CGO_ENABLED=1 DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/0/bus SSH_TTY=/dev/pts/3 OLDPWD=/root _=/usr/local/bin/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v12]" cmd="/usr/local/bin/ollama runner --ollama-engine --port 34211" time=2025-10-19T14:59:21.495Z level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-19T14:59:21.497Z level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:34211" time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.498Z level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-19T14:59:21.498Z level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-19T14:59:21.498Z level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu-haswell.so time=2025-10-19T14:59:21.520Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v12 ggml_cuda_init: failed to initialize CUDA: unknown error load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v12/libggml-cuda.so time=2025-10-19T14:59:21.626Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-19T14:59:21.627Z level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=129.726648ms time=2025-10-19T14:59:21.627Z level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=810ns time=2025-10-19T14:59:21.628Z level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" devices=[] time=2025-10-19T14:59:21.628Z level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=153.310479ms OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" extra_envs=[] time=2025-10-19T14:59:21.628Z level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" extra_envs=[] time=2025-10-19T14:59:21.629Z level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash CGO_LDFLAGS=-L/opt/ollama-source/build/lib/ollama -L/usr/local/cuda-11.8/lib64 PWD=/usr/local/lib/ollama LOGNAME=root XDG_SESSION_TYPE=tty CXX=/usr/bin/g++-11 HOME=/root LANG=C LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: SSH_CONNECTION=10.40.1.52 34792 10.40.1.20 22 CUDAHOSTCXX=/usr/bin/g++-11 LESSCLOSE=/usr/bin/lesspipe %s %s XDG_SESSION_CLASS=user TERM=tmux-256color LESSOPEN=| /usr/bin/lesspipe %s USER=root SHLVL=1 XDG_SESSION_ID=3329 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v13:/opt/ollama-source/build/lib/ollama:/usr/local/cuda-11.8/lib64:/usr/local/cuda-11.8/lib64: XDG_RUNTIME_DIR=/run/user/0 SSH_CLIENT=10.40.1.52 34792 22 CUDA_HOME=/usr/local/cuda-11.8 PATH=/usr/local/cuda-11.8/bin:/usr/local/cuda-11.8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin CC=/usr/bin/gcc-11 CGO_ENABLED=1 DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/0/bus SSH_TTY=/dev/pts/3 OLDPWD=/root _=/usr/local/bin/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v13]" cmd="/usr/local/bin/ollama runner --ollama-engine --port 41121" time=2025-10-19T14:59:21.678Z level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-19T14:59:21.678Z level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:41121" time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.680Z level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-19T14:59:21.680Z level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-19T14:59:21.680Z level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu-haswell.so time=2025-10-19T14:59:21.701Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v13 ggml_cuda_init: failed to initialize CUDA: unknown error load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v13/libggml-cuda.so time=2025-10-19T14:59:21.818Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-10-19T14:59:21.818Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-19T14:59:21.819Z level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=139.677341ms time=2025-10-19T14:59:21.819Z level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=850ns time=2025-10-19T14:59:21.820Z level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" devices=[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=191.574864ms OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" extra_envs=[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-19T14:59:21.820Z level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=345.517504ms time=2025-10-19T14:59:21.821Z level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="8.0 GiB" available="7.6 GiB" time=2025-10-19T14:59:21.821Z level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

Nvidia details

  • cuda version 11.8 (per nvcc)
  • nvidia version: 580.82.07 (per nvidia-smi)

Card details:

  • 2 Nvidia Tesla P40s

Os details: Ubuntu 24.04 (in an LXC)

<!-- gh-comment-id:3419734453 --> @miversen33 commented on GitHub (Oct 19, 2025): Disregard the below. In my case it was a stupid cgroup oversight with `nvidia-uvm`. --- I am noticing identical behavior with Ollama 0.12.5 and 0.12.6 on linux. I am unsure if my issue is the same, but the result is the same. Ollama seems to refuse to talk to my 2 P40 GPUs. Been pulling my hair out for the past few days trying to debug it but from everything I can tell, the issue is simply that ollama doesn't like the GPUs. In my case I installed with the installation script found [here](https://github.com/ollama/ollama/blob/main/scripts/install.sh) and when I run it, I notice that ollama doesn't have runners for anything pre cuda_12. For my GPUs, they run cuda_v11 and it seems that while Ollama _can_ see the GPUS, it doesn't know how to talk to them as there is no runner for them <details> <summary>Github is fighting me and won't let me upload the log as a file so here is my log, maybe it will help shed some light on stuff?</summary> time=2025-10-19T14:59:21.474Z level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-10-19T14:59:21.474Z level=INFO source=images.go:522 msg="total blobs: 0" time=2025-10-19T14:59:21.474Z level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-19T14:59:21.474Z level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-19T14:59:21.475Z level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-19T14:59:21.475Z level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-19T14:59:21.475Z level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" extra_envs=[] time=2025-10-19T14:59:21.476Z level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash CGO_LDFLAGS=-L/opt/ollama-source/build/lib/ollama -L/usr/local/cuda-11.8/lib64 PWD=/usr/local/lib/ollama LOGNAME=root XDG_SESSION_TYPE=tty CXX=/usr/bin/g++-11 HOME=/root LANG=C LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: SSH_CONNECTION=10.40.1.52 34792 10.40.1.20 22 CUDAHOSTCXX=/usr/bin/g++-11 LESSCLOSE=/usr/bin/lesspipe %s %s XDG_SESSION_CLASS=user TERM=tmux-256color LESSOPEN=| /usr/bin/lesspipe %s USER=root SHLVL=1 XDG_SESSION_ID=3329 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v12:/opt/ollama-source/build/lib/ollama:/usr/local/cuda-11.8/lib64:/usr/local/cuda-11.8/lib64: XDG_RUNTIME_DIR=/run/user/0 SSH_CLIENT=10.40.1.52 34792 22 CUDA_HOME=/usr/local/cuda-11.8 PATH=/usr/local/cuda-11.8/bin:/usr/local/cuda-11.8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin CC=/usr/bin/gcc-11 CGO_ENABLED=1 DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/0/bus SSH_TTY=/dev/pts/3 OLDPWD=/root _=/usr/local/bin/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v12]" cmd="/usr/local/bin/ollama runner --ollama-engine --port 34211" time=2025-10-19T14:59:21.495Z level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-19T14:59:21.497Z level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:34211" time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.498Z level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-19T14:59:21.498Z level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-19T14:59:21.498Z level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu-haswell.so time=2025-10-19T14:59:21.520Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v12 ggml_cuda_init: failed to initialize CUDA: unknown error load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v12/libggml-cuda.so time=2025-10-19T14:59:21.626Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-19T14:59:21.627Z level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=129.726648ms time=2025-10-19T14:59:21.627Z level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=810ns time=2025-10-19T14:59:21.628Z level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" devices=[] time=2025-10-19T14:59:21.628Z level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=153.310479ms OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" extra_envs=[] time=2025-10-19T14:59:21.628Z level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" extra_envs=[] time=2025-10-19T14:59:21.629Z level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash CGO_LDFLAGS=-L/opt/ollama-source/build/lib/ollama -L/usr/local/cuda-11.8/lib64 PWD=/usr/local/lib/ollama LOGNAME=root XDG_SESSION_TYPE=tty CXX=/usr/bin/g++-11 HOME=/root LANG=C LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: SSH_CONNECTION=10.40.1.52 34792 10.40.1.20 22 CUDAHOSTCXX=/usr/bin/g++-11 LESSCLOSE=/usr/bin/lesspipe %s %s XDG_SESSION_CLASS=user TERM=tmux-256color LESSOPEN=| /usr/bin/lesspipe %s USER=root SHLVL=1 XDG_SESSION_ID=3329 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v13:/opt/ollama-source/build/lib/ollama:/usr/local/cuda-11.8/lib64:/usr/local/cuda-11.8/lib64: XDG_RUNTIME_DIR=/run/user/0 SSH_CLIENT=10.40.1.52 34792 22 CUDA_HOME=/usr/local/cuda-11.8 PATH=/usr/local/cuda-11.8/bin:/usr/local/cuda-11.8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin CC=/usr/bin/gcc-11 CGO_ENABLED=1 DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/0/bus SSH_TTY=/dev/pts/3 OLDPWD=/root _=/usr/local/bin/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v13]" cmd="/usr/local/bin/ollama runner --ollama-engine --port 41121" time=2025-10-19T14:59:21.678Z level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-19T14:59:21.678Z level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:41121" time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.680Z level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-19T14:59:21.680Z level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-19T14:59:21.680Z level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu-haswell.so time=2025-10-19T14:59:21.701Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v13 ggml_cuda_init: failed to initialize CUDA: unknown error load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v13/libggml-cuda.so time=2025-10-19T14:59:21.818Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-10-19T14:59:21.818Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-19T14:59:21.819Z level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=139.677341ms time=2025-10-19T14:59:21.819Z level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=850ns time=2025-10-19T14:59:21.820Z level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" devices=[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=191.574864ms OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" extra_envs=[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-19T14:59:21.820Z level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=345.517504ms time=2025-10-19T14:59:21.821Z level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="8.0 GiB" available="7.6 GiB" time=2025-10-19T14:59:21.821Z level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" </details> Nvidia details - cuda version 11.8 (per nvcc) - nvidia version: 580.82.07 (per nvidia-smi) Card details: - 2 Nvidia Tesla P40s Os details: Ubuntu 24.04 (in an LXC)
Author
Owner

@dhiltgen commented on GitHub (Oct 20, 2025):

@catclaw the logging update is merged, so when 0.12.7 ships (later this week) please give it a try with OLLAMA_DEBUG=2 and hopefully we'll get a little more detail on why none of the libraries are loading properly on your system.

<!-- gh-comment-id:3422785571 --> @dhiltgen commented on GitHub (Oct 20, 2025): @catclaw the logging update is merged, so when 0.12.7 ships (later this week) please give it a try with OLLAMA_DEBUG=2 and hopefully we'll get a little more detail on why none of the libraries are loading properly on your system.
Author
Owner

@comicchang commented on GitHub (Oct 21, 2025):

I am experiencing the same issue.

I am on Arch Linux, using the following packages:

  • cuda 13.0.2-1
  • ollama 0.12.6-1
  • ollama-cuda 0.12.6-1

Here is my nvidia-smi output:

$ nvidia-smi
Tue Oct 21 23:38:38 2025       
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 580.95.05              Driver Version: 580.95.05      CUDA Version: 13.0     |
+-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA GeForce GTX 1080 Ti     Off |   00000000:01:00.0 Off |                  N/A |
| 29%   27C    P8             10W /  250W |      48MiB /  11264MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+
|   1  NVIDIA GeForce GTX 1080 Ti     Off |   00000000:02:00.0 Off |                  N/A |
| 23%   29C    P8              8W /  250W |       7MiB /  11264MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes:                                                                              |
|  GPU   GI   CI              PID   Type   Process name                        GPU Memory |
|        ID   ID                                                               Usage      |
|=========================================================================================|
|    0   N/A  N/A            2223      G   /usr/lib/Xorg                            23MiB |
|    0   N/A  N/A            2243      G   /usr/bin/sddm-greeter-qt6                21MiB |
|    1   N/A  N/A            2223      G   /usr/lib/Xorg                             4MiB |
+-----------------------------------------------------------------------------------------+

And here is my OLLAMA_DEBUG=2 ollama serve output:

debug.txt

<!-- gh-comment-id:3427319616 --> @comicchang commented on GitHub (Oct 21, 2025): I am experiencing the same issue. I am on Arch Linux, using the following packages: * `cuda 13.0.2-1` * `ollama 0.12.6-1` * `ollama-cuda 0.12.6-1` Here is my `nvidia-smi` output: ``` $ nvidia-smi Tue Oct 21 23:38:38 2025 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 580.95.05 Driver Version: 580.95.05 CUDA Version: 13.0 | +-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA GeForce GTX 1080 Ti Off | 00000000:01:00.0 Off | N/A | | 29% 27C P8 10W / 250W | 48MiB / 11264MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA GeForce GTX 1080 Ti Off | 00000000:02:00.0 Off | N/A | | 23% 29C P8 8W / 250W | 7MiB / 11264MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | 0 N/A N/A 2223 G /usr/lib/Xorg 23MiB | | 0 N/A N/A 2243 G /usr/bin/sddm-greeter-qt6 21MiB | | 1 N/A N/A 2223 G /usr/lib/Xorg 4MiB | +-----------------------------------------------------------------------------------------+ ``` And here is my `OLLAMA_DEBUG=2 ollama serve` output: [debug.txt](https://github.com/user-attachments/files/23024472/debug.txt)
Author
Owner

@fernvenue commented on GitHub (Oct 21, 2025):

Same here:

~# nvidia-smi    
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 550.163.01             Driver Version: 550.163.01     CUDA Version: 12.4     |
|-----------------------------------------+------------------------+----------------------+
| GPU  Name                 Persistence-M | Bus-Id          Disp.A | Volatile Uncorr. ECC |
| Fan  Temp   Perf          Pwr:Usage/Cap |           Memory-Usage | GPU-Util  Compute M. |
|                                         |                        |               MIG M. |
|=========================================+========================+======================|
|   0  NVIDIA RTX 5000 Ada Gene...    On  |   00000000:01:00.0 Off |                  Off |
| 30%   26C    P8              9W /  250W |   26792MiB /  32760MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+
|   1  NVIDIA RTX 5000 Ada Gene...    On  |   00000000:02:00.0 Off |                  Off |
| 30%   30C    P8              9W /  250W |       5MiB /  32760MiB |      0%      Default |
|                                         |                        |                  N/A |
+-----------------------------------------+------------------------+----------------------+
<!-- gh-comment-id:3428265153 --> @fernvenue commented on GitHub (Oct 21, 2025): Same here: ``` ~# nvidia-smi +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 550.163.01 Driver Version: 550.163.01 CUDA Version: 12.4 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA RTX 5000 Ada Gene... On | 00000000:01:00.0 Off | Off | | 30% 26C P8 9W / 250W | 26792MiB / 32760MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA RTX 5000 Ada Gene... On | 00000000:02:00.0 Off | Off | | 30% 30C P8 9W / 250W | 5MiB / 32760MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ ```
Author
Owner

@dhiltgen commented on GitHub (Oct 21, 2025):

@comicchang it looks like the Arch packages aren't being built correctly. Until that's resolved by those maintainers, please try to use the Ollama official binaries and it should detect your GPU. If not, please share an updated log.

<!-- gh-comment-id:3428412317 --> @dhiltgen commented on GitHub (Oct 21, 2025): @comicchang it looks like the Arch packages aren't being built correctly. Until that's resolved by those maintainers, please try to use the Ollama [official binaries](https://github.com/ollama/ollama/blob/main/docs/linux.md) and it should detect your GPU. If not, please share an updated log.
Author
Owner

@fernvenue commented on GitHub (Oct 21, 2025):

Hi @dhiltgen, thx for your explain, just tried the latest official build, and seems it can now detect GPU correctly, thx again :)

<!-- gh-comment-id:3428463124 --> @fernvenue commented on GitHub (Oct 21, 2025): Hi @dhiltgen, thx for your explain, just tried the latest official build, and seems it can now detect GPU correctly, thx again :)
Author
Owner

@comicchang commented on GitHub (Oct 22, 2025):

@comicchang it looks like the Arch packages aren't being built correctly. Until that's resolved by those maintainers, please try to use the Ollama official binaries and it should detect your GPU. If not, please share an updated log.

Thanks, I checked the ollama PKGBUILD file and found that support for compute capability 6.1 (sm_61) was indeed removed. I will try to build the package myself.

<!-- gh-comment-id:3430123901 --> @comicchang commented on GitHub (Oct 22, 2025): > [@comicchang](https://github.com/comicchang) it looks like the Arch packages aren't being built correctly. Until that's resolved by those maintainers, please try to use the Ollama [official binaries](https://github.com/ollama/ollama/blob/main/docs/linux.md) and it should detect your GPU. If not, please share an updated log. Thanks, I checked the ollama [PKGBUILD file](https://gitlab.archlinux.org/archlinux/packaging/packages/ollama/-/blob/45e731d84a6edab405f0b1c50e0ff3b1234fd887/PKGBUILD) and found that support for compute capability 6.1 (sm_61) was indeed removed. I will try to build the package myself.
Author
Owner

@catclaw commented on GitHub (Oct 25, 2025):

@catclaw the logging update is merged, so when 0.12.7 ships (later this week) please give it a try with OLLAMA_DEBUG=2 and hopefully we'll get a little more detail on why none of the libraries are loading properly on your system.

Will do! Thank you so much!
I've made a lot of changes to my system, but the problem still persists.

This is the current output from "ollama serve" in a console:

time=2025-10-25T20:18:21.577+02:00 level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-25T20:18:21.619+02:00 level=INFO source=images.go:522 msg="total blobs: 146" time=2025-10-25T20:18:21.636+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-25T20:18:21.651+02:00 level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-25T20:18:21.652+02:00 level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-25T20:18:21.658+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-25T20:18:21.658+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-25T20:18:21.709+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62569" time=2025-10-25T20:18:21.870+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:21.877+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62569" time=2025-10-25T20:18:21.883+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:21.885+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:21.887+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:21.888+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:21.888+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:21.890+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:21.893+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:21.896+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:21.898+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:21.927+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-10-25T20:18:21.962+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:21.964+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:21.967+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:21.969+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:21.971+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:21.973+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:21.974+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:21.977+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:21.979+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:21.979+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:21.981+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:21.982+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:21.984+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:21.988+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:21.990+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:21.992+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:21.992+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:21.995+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:21.996+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:21.998+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.001+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.002+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.003+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.005+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=122.7994ms time=2025-10-25T20:18:22.005+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.009+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-10-25T20:18:22.012+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=353.0328ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-25T20:18:22.015+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-25T20:18:22.023+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62584" time=2025-10-25T20:18:22.194+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:22.199+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62584" time=2025-10-25T20:18:22.211+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.213+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:22.214+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:22.215+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.216+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:22.219+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:22.220+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:22.221+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:22.223+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:22.252+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-10-25T20:18:22.301+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:22.302+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.304+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:22.306+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:22.307+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:22.308+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:22.310+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:22.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:22.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:22.315+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:22.317+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:22.318+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:22.319+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:22.320+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:22.321+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:22.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:22.329+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:22.329+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:22.331+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.332+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.333+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.333+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.335+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=124.9775ms time=2025-10-25T20:18:22.339+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.342+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[] time=2025-10-25T20:18:22.344+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=325.1639ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-25T20:18:22.345+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-25T20:18:22.349+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62602" time=2025-10-25T20:18:22.526+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:22.532+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62602" time=2025-10-25T20:18:22.544+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.545+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:22.546+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:22.547+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.549+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:22.552+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:22.553+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:22.554+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:22.555+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:22.584+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-25T20:18:22.607+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:22.607+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.610+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:22.612+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:22.612+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:22.614+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:22.615+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:22.616+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:22.617+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:22.621+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:22.623+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:22.623+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:22.625+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:22.625+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:22.628+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.628+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:22.629+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:22.634+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:22.635+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:22.636+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.637+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.638+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.639+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.640+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=97.5824ms time=2025-10-25T20:18:22.642+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.649+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[] time=2025-10-25T20:18:22.651+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=304.9441ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-25T20:18:22.652+02:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-25T20:18:22.653+02:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-25T20:18:22.654+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=1.0015121s time=2025-10-25T20:18:22.655+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="85.5 GiB" time=2025-10-25T20:18:22.659+02:00 level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

This is the list of all files of 0.12.6 installation:

` Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama

2025-10-25 20:14

.
2025-10-22 02:44 ..
2025-10-16 23:00 7 502 app.ico
2025-10-25 20:14 lib
2025-10-16 23:09 13 897 208 ollama app.exe
2025-10-16 23:09 32 979 448 ollama.exe
2025-10-25 20:14 446 008 unins000.dat
2025-10-25 20:12 3 706 360 unins000.exe
2025-10-25 20:14 25 291 unins000.msg
6 File(s) 51 061 817 bytes

Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib

2025-10-25 20:14

.
2025-10-25 20:14 ..
2025-10-25 20:13 ollama
2025-10-25 20:14 2 138 Ollama.lnk
1 File(s) 2 138 bytes

Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama

2025-10-25 20:13

.
2025-10-25 20:14 ..
2025-10-25 20:13 cuda_v12
2025-10-25 20:13 cuda_v13
2025-10-16 23:09 902 136 ggml-base.dll
2025-10-16 23:09 1 067 512 ggml-cpu-alderlake.dll
2025-10-16 23:09 1 070 584 ggml-cpu-haswell.dll
2025-10-16 23:09 1 296 888 ggml-cpu-icelake.dll
2025-10-16 23:09 955 896 ggml-cpu-sandybridge.dll
2025-10-16 23:09 1 295 352 ggml-cpu-skylakex.dll
2025-10-16 23:09 794 616 ggml-cpu-sse42.dll
2025-10-16 23:09 783 352 ggml-cpu-x64.dll
2025-10-25 20:13 rocm
8 File(s) 8 166 336 bytes

Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12

2025-10-25 20:13

.
2025-10-25 20:13 ..
2025-10-16 23:09 113 720 824 cublas64_12.dll
2025-10-16 23:09 692 449 784 cublasLt64_12.dll
2025-10-16 23:09 582 136 cudart64_12.dll
2025-10-16 23:09 1 569 384 952 ggml-cuda.dll
4 File(s) 2 376 137 696 bytes

Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13

2025-10-25 20:13

.
2025-10-25 20:13 ..
2025-10-16 23:09 50 286 072 cublas64_13.dll
2025-10-16 23:09 477 894 648 cublasLt64_13.dll
2025-10-16 23:09 293 704 696 ggml-cuda.dll
3 File(s) 821 885 416 bytes

Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm

2025-10-25 20:13

.
2025-10-25 20:13 ..
2025-10-16 23:09 23 660 536 amdhip64_6.dll
2025-10-16 23:09 111 860 216 amd_comgr_2.dll
2025-10-16 23:09 426 622 968 ggml-hip.dll
2025-10-16 23:09 908 792 hipblas.dll
2025-10-25 20:13 rocblas
2025-10-16 23:09 416 692 728 rocblas.dll
5 File(s) 979 745 240 bytes

Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\rocblas

2025-10-25 20:13

.
2025-10-25 20:13 ..
2025-10-25 20:14 library
0 File(s) 0 bytes

Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\rocblas\library

2025-10-25 20:14

.
2025-10-25 20:13 ..
2025-10-16 23:03 248 288 Kernels.so-000-gfx1030.hsaco
2025-10-16 23:03 262 120 Kernels.so-000-gfx1100.hsaco
2025-10-16 23:03 262 376 Kernels.so-000-gfx1101.hsaco
2025-10-16 23:03 262 120 Kernels.so-000-gfx1102.hsaco
2025-10-16 23:03 260 840 Kernels.so-000-gfx1151.hsaco
2025-10-16 23:03 23 026 TensileLibrary_lazy_gfx1030.dat
2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1100.dat
2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1101.dat
2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1102.dat
2025-10-16 23:03 17 653 TensileLibrary_lazy_gfx1151.dat
2025-10-16 23:03 43 504 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 161 304 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 170 784 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 37 230 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 134 320 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 142 264 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 43 848 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 161 048 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 170 016 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 169 760 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 169 760 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 37 398 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 133 808 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 141 752 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 56 121 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 474 768 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 500 632 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 16 798 632 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 549 603 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 17 275 880 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 583 354 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 11 487 568 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 244 064 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 53 167 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 437 856 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 472 168 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 16 199 336 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 645 565 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 6 410 600 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 309 377 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 8 035 592 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 215 028 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 39 542 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 189 456 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 199 448 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 199 448 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 198 936 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 198 424 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 15 258 976 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 590 496 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 12 111 384 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 477 855 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 516 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 317 608 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 53 294 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 434 272 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 468 328 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 467 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 468 328 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 467 048 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 11 708 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 576 526 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 6 024 760 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 382 857 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 4 275 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 145 331 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 54 356 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 449 936 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 465 048 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 465 048 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 466 072 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 464 536 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 32 004 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 240 096 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 252 392 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 17 086 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 81 496 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 85 344 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 28 317 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 195 440 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 204 152 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 298 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 286 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 326 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 49 024 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 49 024 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 48 768 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 48 768 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 43 888 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 200 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 43 880 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 45 936 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 094 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 20 008 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 8 802 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 33 864 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 110 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 20 008 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 110 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 19 752 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 27 664 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 186 632 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 200 720 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 5 667 184 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 320 632 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 12 455 016 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 518 011 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 16 953 096 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 626 281 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 9 981 064 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 257 641 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 20 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 90 288 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 97 720 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 10 899 056 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 606 844 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 171 584 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 568 803 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 5 949 184 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 336 152 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 9 584 152 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 322 795 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 20 980 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 104 888 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 112 832 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 113 088 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 112 832 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 113 088 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 5 846 216 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 346 808 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 478 496 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 507 834 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 11 279 360 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 504 971 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 7 762 144 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 240 050 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 20 976 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 89 776 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 96 696 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 96 952 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 96 952 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 96 696 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 8 805 984 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 433 493 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 7 842 920 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 526 415 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 4 909 552 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 375 873 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 3 627 104 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 164 717 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 74 464 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 380 904 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 395 504 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 9 250 488 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 624 107 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 13 129 136 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 540 665 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 14 238 888 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 523 364 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 4 892 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 139 468 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 52 394 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 196 904 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 206 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 206 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 207 920 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 206 384 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 13 248 936 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 159 612 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 984 424 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 562 937 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 5 833 984 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 329 670 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 12 036 552 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 391 906 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 45 640 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 108 496 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 112 600 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 1 538 176 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 143 490 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 11 003 984 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 537 678 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 9 958 440 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 473 572 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 8 283 312 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 251 527 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 45 664 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 164 056 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 171 488 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 171 488 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 172 768 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 171 232 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 11 981 432 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 060 470 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 8 804 720 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 563 577 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 4 893 264 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 376 677 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 5 387 384 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 226 414 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 46 618 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 340 912 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 349 624 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 349 624 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 348 856 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 348 856 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 28 091 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 201 336 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 211 072 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 20 519 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 102 576 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 106 936 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 31 766 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 236 000 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 244 968 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 244 712 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 246 504 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 244 712 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 70 739 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 491 840 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 539 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 539 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 539 976 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 538 952 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 3 253 936 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 318 729 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 22 255 688 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 679 573 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 15 519 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 571 645 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 17 097 848 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 327 608 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 56 959 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 334 288 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 364 760 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 364 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 365 528 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 364 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 2 191 064 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 197 238 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 17 182 176 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 698 938 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 7 241 096 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 413 453 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 15 757 872 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 397 205 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 67 638 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 463 320 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 509 664 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 509 664 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 510 432 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 509 408 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 2 486 384 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 234 821 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 16 680 632 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 633 900 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 5 576 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 337 236 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 14 921 928 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 375 463 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 65 854 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 411 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 451 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 2 302 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 234 701 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 926 528 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 710 893 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 3 804 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 387 494 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 7 114 904 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 318 061 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 53 296 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 15 770 432 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 263 969 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 15 770 432 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 1 263 969 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 976 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 016 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 224 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 110 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 34 856 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 16 283 560 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 447 637 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 16 283 560 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 1 447 637 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 984 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 017 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 232 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 34 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 8 104 512 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 563 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 8 104 512 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 563 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 984 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 016 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 232 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 080 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 47 416 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 47 416 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 47 672 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 47 160 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 17 987 376 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 423 043 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 17 987 376 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 1 423 043 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 992 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 017 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 240 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 330 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 326 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 358 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 760 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 47 984 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 47 976 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 115 291 TensileManifest.txt
487 File(s) 788 770 641 bytes

 Total Files Listed:
         514 File(s)  5 025 769 284 bytes`

Meanwhile, I'm using LM Studio on the same port as Ollama (Not at the same time) to make things easier, and linked my existing models using Gollama -L, and it's working without any issues, besides I feel that I don't have the same capabilities, and LM Studio feels slower, and I get random "Out of memory" errors on 70B models...
Can't wait till 0.12.7 release

Output from LM Studio server:

[2025-10-25 20:35:43][DEBUG][LM Studio] GPU Configuration: Strategy: evenly Priority: [] Disabled GPUs: [] Limit weight offload to dedicated GPU Memory: OFF Offload KV Cache to GPU: ON [2025-10-25 20:35:43][DEBUG][LM Studio] Live GPU memory info (source 'LMS Core'): GPU 0: NVIDIA GeForce RTX 5070 (Used: 4.12 GB, Total: 12.82 GB, Free: 8.70 GB) GPU 1: NVIDIA GeForce RTX 3060 (Used: 970.50 MB, Total: 12.88 GB, Free: 11.91 GB) GPU 2: NVIDIA GeForce RTX 3060 (Used: 998.65 MB, Total: 12.88 GB, Free: 11.89 GB) [2025-10-25 20:35:43][DEBUG][LM Studio] Model load size estimate with raw num offload layers 'max' and context length '131072': Model: 18.40 GB Context: 46.55 GB Total: 64.96 GB [2025-10-25 20:35:43][DEBUG][LM Studio] Strict GPU VRAM cap is OFF: GPU offload layers will not be checked for adjustment [LM Studio] Resolved GPU config options: Num Offload Layers: max Num CPU Expert Layers: 0 Main GPU: 0 Tensor Split: [0] Disabled GPUs: [] [2025-10-25 20:35:43][DEBUG] ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes [2025-10-25 20:35:43][DEBUG] CUDA : ARCHS = 750,800,890,900,1000,1200 | USE_GRAPHS = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | BMI2 = 1 | LLAMAFILE = 1 | OPENMP = 1 | REPACK = 1 | [2025-10-25 20:35:43][DEBUG] llama_model_load_from_file_impl: using device CUDA0 (NVIDIA GeForce RTX 5070) (0000:03:00.0) - 10929 MiB free llama_model_load_from_file_impl: using device CUDA1 (NVIDIA GeForce RTX 3060) (0000:04:00.0) - 11245 MiB free llama_model_load_from_file_impl: using device CUDA2 (NVIDIA GeForce RTX 3060) (0000:83:00.0) - 11245 MiB free [2025-10-25 20:35:44][DEBUG] llama_model_loader: loaded meta data with 44 key-value pairs and 771 tensors from C:\Users\Iman Mirbioki\.lmstudio\models\aratan\DeepSeek-R1-32B-Uncensored\DeepSeek-R1-32B-Uncensored.gguf (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen2 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = DeepSeek R1 Distill Qwen 32B Uncensored llama_model_loader: - kv 3: general.finetune str = Uncensored llama_model_loader: - kv 4: general.basename str = DeepSeek-R1-Distill-Qwen llama_model_loader: - kv 5: general.size_label str = 32B llama_model_loader: - kv 6: general.license str = mit llama_model_loader: - kv 7: general.base_model.count u32 = 1 llama_model_loader: - kv 8: general.base_model.0.name str = DeepSeek R1 Distill Qwen 32B llama_model_loader: - kv 9: general.base_model.0.organization str = Deepseek Ai llama_model_loader: - kv 10: general.base_model.0.repo_url str = https://huggingface.co/deepseek-ai/De... llama_model_loader: - kv 11: general.dataset.count u32 = 1 llama_model_loader: - kv 12: general.dataset.0.name str = Uncensor llama_model_loader: - kv 13: general.dataset.0.organization str = Guilherme34 llama_model_loader: - kv 14: general.dataset.0.repo_url str = https://huggingface.co/Guilherme34/un... llama_model_loader: - kv 15: general.tags arr[str,1] = ["generated_from_trainer"] llama_model_loader: - kv 16: qwen2.block_count u32 = 64 llama_model_loader: - kv 17: qwen2.context_length u32 = 131072 llama_model_loader: - kv 18: qwen2.embedding_length u32 = 5120 llama_model_loader: - kv 19: qwen2.feed_forward_length u32 = 27648 llama_model_loader: - kv 20: qwen2.attention.head_count u32 = 40 llama_model_loader: - kv 21: qwen2.attention.head_count_kv u32 = 8 llama_model_loader: - kv 22: qwen2.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 23: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 24: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 25: tokenizer.ggml.pre str = deepseek-r1-qwen [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 26: tokenizer.ggml.tokens arr[str,151665] = ["!", "\"", "#", "$", "%", "&", "'", ... [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 27: tokenizer.ggml.token_type arr[i32,151665] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 28: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 29: tokenizer.ggml.bos_token_id u32 = 151646 llama_model_loader: - kv 30: tokenizer.ggml.eos_token_id u32 = 151643 llama_model_loader: - kv 31: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 32: tokenizer.ggml.add_bos_token bool = true llama_model_loader: - kv 33: tokenizer.ggml.add_eos_token bool = false llama_model_loader: - kv 34: tokenizer.chat_template str = {% if not add_generation_prompt is de... llama_model_loader: - kv 35: general.quantization_version u32 = 2 llama_model_loader: - kv 36: general.file_type u32 = 30 llama_model_loader: - kv 37: general.url str = https://huggingface.co/mradermacher/D... llama_model_loader: - kv 38: mradermacher.quantize_version str = 2 llama_model_loader: - kv 39: mradermacher.quantized_by str = mradermacher llama_model_loader: - kv 40: mradermacher.quantized_at str = 2025-01-26T04:57:59+01:00 llama_model_loader: - kv 41: mradermacher.quantized_on str = marco llama_model_loader: - kv 42: general.source.url str = https://huggingface.co/nicoboss/DeepS... llama_model_loader: - kv 43: mradermacher.convert_type str = hf llama_model_loader: - type f32: 321 tensors llama_model_loader: - type q5_K: 72 tensors llama_model_loader: - type q6_K: 1 tensors llama_model_loader: - type iq4_xs: 377 tensors print_info: file format = GGUF V3 (latest) print_info: file type = IQ4_XS - 4.25 bpw print_info: file size = 16.63 GiB (4.36 BPW) [2025-10-25 20:35:44][DEBUG] load: special_eos_id is not in special_eog_ids - the tokenizer config may be incorrect load: printing all EOG tokens: load: - 151643 ('<|end▁of▁sentence|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') [2025-10-25 20:35:44][DEBUG] load: special tokens cache size = 22 [2025-10-25 20:35:44][DEBUG] load: token to piece cache size = 0.9310 MB print_info: arch = qwen2 print_info: vocab_only = 0 print_info: n_ctx_train = 131072 print_info: n_embd = 5120 print_info: n_layer = 64 print_info: n_head = 40 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: is_swa_any = 0 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 5 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-05 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 [2025-10-25 20:35:44][DEBUG] print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 27648 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = -1 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 131072 print_info: rope_finetuned = unknown print_info: model type = 32B print_info: model params = 32.76 B print_info: general.name = DeepSeek R1 Distill Qwen 32B Uncensored print_info: vocab type = BPE print_info: n_vocab = 151665 print_info: n_merges = 151387 print_info: BOS token = 151646 '<|begin▁of▁sentence|>' print_info: EOS token = 151643 '<|end▁of▁sentence|>' print_info: EOT token = 151643 '<|end▁of▁sentence|>' print_info: PAD token = 151643 '<|end▁of▁sentence|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|end▁of▁sentence|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = true) [2025-10-25 20:37:03][DEBUG] load_tensors: offloading 64 repeating layers to GPU load_tensors: offloading output layer to GPU load_tensors: offloaded 65/65 layers to GPU load_tensors: CUDA2 model buffer size = 5565.08 MiB load_tensors: CUDA1 model buffer size = 5453.34 MiB load_tensors: CUDA0 model buffer size = 5622.09 MiB load_tensors: CPU_Mapped model buffer size = 393.42 MiB [2025-10-25 20:37:19][DEBUG] llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 131072 llama_context: n_ctx_per_seq = 131072 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = disabled llama_context: kv_unified = false llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 [2025-10-25 20:37:19][DEBUG] llama_context: CUDA_Host output buffer size = 0.58 MiB

I've been going through all my logs, and it seems like the problem started when Win11 update 25H2 was released - but it's just a guess.

<!-- gh-comment-id:3447276333 --> @catclaw commented on GitHub (Oct 25, 2025): > [@catclaw](https://github.com/catclaw) the logging update is merged, so when 0.12.7 ships (later this week) please give it a try with OLLAMA_DEBUG=2 and hopefully we'll get a little more detail on why none of the libraries are loading properly on your system. Will do! Thank you so much! I've made a lot of changes to my system, but the problem still persists. This is the current output from "ollama serve" in a console: `time=2025-10-25T20:18:21.577+02:00 level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-25T20:18:21.619+02:00 level=INFO source=images.go:522 msg="total blobs: 146" time=2025-10-25T20:18:21.636+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-25T20:18:21.651+02:00 level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-25T20:18:21.652+02:00 level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-25T20:18:21.658+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-25T20:18:21.658+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-25T20:18:21.709+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62569" time=2025-10-25T20:18:21.870+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:21.877+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62569" time=2025-10-25T20:18:21.883+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:21.885+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:21.887+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:21.888+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:21.888+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:21.890+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:21.893+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:21.896+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:21.898+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:21.927+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-10-25T20:18:21.962+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:21.964+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:21.967+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:21.969+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:21.971+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:21.973+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:21.974+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:21.977+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:21.979+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:21.979+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:21.981+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:21.982+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:21.984+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:21.988+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:21.990+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:21.992+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:21.992+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:21.995+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:21.996+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:21.998+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.001+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.002+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.003+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.005+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=122.7994ms time=2025-10-25T20:18:22.005+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.009+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-10-25T20:18:22.012+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=353.0328ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-25T20:18:22.015+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-25T20:18:22.023+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62584" time=2025-10-25T20:18:22.194+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:22.199+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62584" time=2025-10-25T20:18:22.211+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.213+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:22.214+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:22.215+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.216+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:22.219+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:22.220+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:22.221+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:22.223+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:22.252+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-10-25T20:18:22.301+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:22.302+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.304+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:22.306+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:22.307+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:22.308+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:22.310+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:22.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:22.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:22.315+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:22.317+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:22.318+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:22.319+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:22.320+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:22.321+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:22.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:22.329+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:22.329+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:22.331+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.332+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.333+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.333+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.335+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=124.9775ms time=2025-10-25T20:18:22.339+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.342+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[] time=2025-10-25T20:18:22.344+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=325.1639ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-25T20:18:22.345+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-25T20:18:22.349+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62602" time=2025-10-25T20:18:22.526+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:22.532+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62602" time=2025-10-25T20:18:22.544+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.545+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:22.546+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:22.547+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.549+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:22.552+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:22.553+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:22.554+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:22.555+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:22.584+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-25T20:18:22.607+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:22.607+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.610+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:22.612+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:22.612+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:22.614+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:22.615+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:22.616+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:22.617+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:22.621+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:22.623+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:22.623+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:22.625+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:22.625+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:22.628+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.628+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:22.629+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:22.634+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:22.635+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:22.636+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.637+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.638+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.639+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.640+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=97.5824ms time=2025-10-25T20:18:22.642+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.649+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[] time=2025-10-25T20:18:22.651+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=304.9441ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-25T20:18:22.652+02:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-25T20:18:22.653+02:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-25T20:18:22.654+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=1.0015121s time=2025-10-25T20:18:22.655+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="85.5 GiB" time=2025-10-25T20:18:22.659+02:00 level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"` This is the list of all files of 0.12.6 installation: ` Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama 2025-10-25 20:14 <DIR> . 2025-10-22 02:44 <DIR> .. 2025-10-16 23:00 7 502 app.ico 2025-10-25 20:14 <DIR> lib 2025-10-16 23:09 13 897 208 ollama app.exe 2025-10-16 23:09 32 979 448 ollama.exe 2025-10-25 20:14 446 008 unins000.dat 2025-10-25 20:12 3 706 360 unins000.exe 2025-10-25 20:14 25 291 unins000.msg 6 File(s) 51 061 817 bytes Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib 2025-10-25 20:14 <DIR> . 2025-10-25 20:14 <DIR> .. 2025-10-25 20:13 <DIR> ollama 2025-10-25 20:14 2 138 Ollama.lnk 1 File(s) 2 138 bytes Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama 2025-10-25 20:13 <DIR> . 2025-10-25 20:14 <DIR> .. 2025-10-25 20:13 <DIR> cuda_v12 2025-10-25 20:13 <DIR> cuda_v13 2025-10-16 23:09 902 136 ggml-base.dll 2025-10-16 23:09 1 067 512 ggml-cpu-alderlake.dll 2025-10-16 23:09 1 070 584 ggml-cpu-haswell.dll 2025-10-16 23:09 1 296 888 ggml-cpu-icelake.dll 2025-10-16 23:09 955 896 ggml-cpu-sandybridge.dll 2025-10-16 23:09 1 295 352 ggml-cpu-skylakex.dll 2025-10-16 23:09 794 616 ggml-cpu-sse42.dll 2025-10-16 23:09 783 352 ggml-cpu-x64.dll 2025-10-25 20:13 <DIR> rocm 8 File(s) 8 166 336 bytes Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12 2025-10-25 20:13 <DIR> . 2025-10-25 20:13 <DIR> .. 2025-10-16 23:09 113 720 824 cublas64_12.dll 2025-10-16 23:09 692 449 784 cublasLt64_12.dll 2025-10-16 23:09 582 136 cudart64_12.dll 2025-10-16 23:09 1 569 384 952 ggml-cuda.dll 4 File(s) 2 376 137 696 bytes Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13 2025-10-25 20:13 <DIR> . 2025-10-25 20:13 <DIR> .. 2025-10-16 23:09 50 286 072 cublas64_13.dll 2025-10-16 23:09 477 894 648 cublasLt64_13.dll 2025-10-16 23:09 293 704 696 ggml-cuda.dll 3 File(s) 821 885 416 bytes Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm 2025-10-25 20:13 <DIR> . 2025-10-25 20:13 <DIR> .. 2025-10-16 23:09 23 660 536 amdhip64_6.dll 2025-10-16 23:09 111 860 216 amd_comgr_2.dll 2025-10-16 23:09 426 622 968 ggml-hip.dll 2025-10-16 23:09 908 792 hipblas.dll 2025-10-25 20:13 <DIR> rocblas 2025-10-16 23:09 416 692 728 rocblas.dll 5 File(s) 979 745 240 bytes Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\rocblas 2025-10-25 20:13 <DIR> . 2025-10-25 20:13 <DIR> .. 2025-10-25 20:14 <DIR> library 0 File(s) 0 bytes Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\rocblas\library 2025-10-25 20:14 <DIR> . 2025-10-25 20:13 <DIR> .. 2025-10-16 23:03 248 288 Kernels.so-000-gfx1030.hsaco 2025-10-16 23:03 262 120 Kernels.so-000-gfx1100.hsaco 2025-10-16 23:03 262 376 Kernels.so-000-gfx1101.hsaco 2025-10-16 23:03 262 120 Kernels.so-000-gfx1102.hsaco 2025-10-16 23:03 260 840 Kernels.so-000-gfx1151.hsaco 2025-10-16 23:03 23 026 TensileLibrary_lazy_gfx1030.dat 2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1100.dat 2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1101.dat 2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1102.dat 2025-10-16 23:03 17 653 TensileLibrary_lazy_gfx1151.dat 2025-10-16 23:03 43 504 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 161 304 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 170 784 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 37 230 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 134 320 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 142 264 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 43 848 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 161 048 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 170 016 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 169 760 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 169 760 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 37 398 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 133 808 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 141 752 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 56 121 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 474 768 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 500 632 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 16 798 632 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 549 603 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 17 275 880 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 583 354 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 11 487 568 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 244 064 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 53 167 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 437 856 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 472 168 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 16 199 336 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 645 565 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 6 410 600 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 309 377 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 8 035 592 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 215 028 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 39 542 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 189 456 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 199 448 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 199 448 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 198 936 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 198 424 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 15 258 976 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 590 496 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 12 111 384 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 477 855 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 13 516 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 317 608 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 53 294 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 434 272 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 468 328 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 467 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 468 328 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 467 048 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 11 708 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 576 526 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 6 024 760 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 382 857 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 4 275 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 145 331 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 54 356 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 449 936 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 465 048 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 465 048 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 466 072 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 464 536 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 32 004 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 240 096 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 252 392 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 17 086 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 81 496 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 85 344 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 28 317 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 195 440 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 204 152 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 298 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback.dat 2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 286 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 326 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback.dat 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 49 024 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 49 024 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 48 768 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 48 768 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 43 888 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 200 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback.dat 2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 43 880 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 45 936 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 094 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 20 008 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 8 802 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 33 864 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 110 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 20 008 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 110 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 19 752 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 27 664 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 186 632 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 200 720 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 5 667 184 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 320 632 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 12 455 016 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 518 011 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 16 953 096 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 626 281 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 9 981 064 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 257 641 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 20 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 90 288 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 97 720 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 10 899 056 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 606 844 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 10 171 584 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 568 803 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 5 949 184 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 336 152 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 9 584 152 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 322 795 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 20 980 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 104 888 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 112 832 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 113 088 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 112 832 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 113 088 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 5 846 216 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 346 808 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 10 478 496 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 507 834 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 11 279 360 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 504 971 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 7 762 144 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 240 050 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 20 976 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 89 776 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 96 696 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 96 952 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 96 952 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 96 696 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 8 805 984 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 433 493 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 7 842 920 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 526 415 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 4 909 552 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 375 873 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 3 627 104 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 164 717 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 74 464 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 380 904 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 395 504 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 9 250 488 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 624 107 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 13 129 136 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 540 665 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 14 238 888 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 523 364 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 4 892 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 139 468 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 52 394 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 196 904 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 206 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 206 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 207 920 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 206 384 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 13 248 936 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 1 159 612 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 10 984 424 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 562 937 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 5 833 984 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 329 670 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 12 036 552 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 391 906 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 45 640 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 108 496 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 112 600 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 1 538 176 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 143 490 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 11 003 984 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 537 678 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 9 958 440 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 473 572 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 8 283 312 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 251 527 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 45 664 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 164 056 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 171 488 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 171 488 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 172 768 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 171 232 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 11 981 432 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 1 060 470 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 8 804 720 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 563 577 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 4 893 264 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 376 677 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 5 387 384 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 226 414 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 46 618 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 340 912 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 349 624 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 349 624 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 348 856 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 348 856 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 28 091 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 201 336 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 211 072 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 20 519 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 102 576 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 106 936 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 31 766 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 236 000 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 244 968 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 244 712 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 246 504 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 244 712 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 70 739 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 491 840 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 539 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 539 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 539 976 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 538 952 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 3 253 936 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 318 729 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 22 255 688 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 679 573 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 15 519 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 571 645 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 17 097 848 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 327 608 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 56 959 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 334 288 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 364 760 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 364 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 365 528 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 364 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 2 191 064 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 197 238 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 17 182 176 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 698 938 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 7 241 096 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 413 453 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 15 757 872 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 397 205 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 67 638 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 463 320 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 509 664 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 509 664 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 510 432 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 509 408 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 2 486 384 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 234 821 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 16 680 632 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 633 900 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 5 576 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 337 236 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 14 921 928 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 375 463 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 65 854 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 411 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 451 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 2 302 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 234 701 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 10 926 528 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 710 893 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 3 804 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 387 494 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 7 114 904 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 318 061 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 53 296 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 15 770 432 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 1 263 969 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 15 770 432 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 1 263 969 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 27 976 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 2 016 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 13 224 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 7 110 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 34 856 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 16 283 560 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 1 447 637 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 16 283 560 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 1 447 637 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 27 984 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 2 017 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 13 232 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 34 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 8 104 512 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 563 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 8 104 512 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 563 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 27 984 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 2 016 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 13 232 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 44 080 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 47 416 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 47 416 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 47 672 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 47 160 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 17 987 376 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co 2025-10-16 23:03 1 423 043 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat 2025-10-16 23:03 17 987 376 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co 2025-10-16 23:03 1 423 043 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat 2025-10-16 23:03 27 992 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co 2025-10-16 23:03 2 017 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat 2025-10-16 23:03 13 240 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co 2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat 2025-10-16 23:03 7 330 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback.dat 2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 326 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 358 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback.dat 2025-10-16 23:03 48 760 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 47 984 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback.dat 2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat 2025-10-16 23:03 47 976 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco 2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco 2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco 2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco 2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco 2025-10-16 23:03 115 291 TensileManifest.txt 487 File(s) 788 770 641 bytes Total Files Listed: 514 File(s) 5 025 769 284 bytes` Meanwhile, I'm using LM Studio on the same port as Ollama (Not at the same time) to make things easier, and linked my existing models using Gollama -L, and it's working without any issues, besides I feel that I don't have the same capabilities, and LM Studio feels slower, and I get random "Out of memory" errors on 70B models... Can't wait till 0.12.7 release Output from LM Studio server: `[2025-10-25 20:35:43][DEBUG][LM Studio] GPU Configuration: Strategy: evenly Priority: [] Disabled GPUs: [] Limit weight offload to dedicated GPU Memory: OFF Offload KV Cache to GPU: ON [2025-10-25 20:35:43][DEBUG][LM Studio] Live GPU memory info (source 'LMS Core'): GPU 0: NVIDIA GeForce RTX 5070 (Used: 4.12 GB, Total: 12.82 GB, Free: 8.70 GB) GPU 1: NVIDIA GeForce RTX 3060 (Used: 970.50 MB, Total: 12.88 GB, Free: 11.91 GB) GPU 2: NVIDIA GeForce RTX 3060 (Used: 998.65 MB, Total: 12.88 GB, Free: 11.89 GB) [2025-10-25 20:35:43][DEBUG][LM Studio] Model load size estimate with raw num offload layers 'max' and context length '131072': Model: 18.40 GB Context: 46.55 GB Total: 64.96 GB [2025-10-25 20:35:43][DEBUG][LM Studio] Strict GPU VRAM cap is OFF: GPU offload layers will not be checked for adjustment [LM Studio] Resolved GPU config options: Num Offload Layers: max Num CPU Expert Layers: 0 Main GPU: 0 Tensor Split: [0] Disabled GPUs: [] [2025-10-25 20:35:43][DEBUG] ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes [2025-10-25 20:35:43][DEBUG] CUDA : ARCHS = 750,800,890,900,1000,1200 | USE_GRAPHS = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | BMI2 = 1 | LLAMAFILE = 1 | OPENMP = 1 | REPACK = 1 | [2025-10-25 20:35:43][DEBUG] llama_model_load_from_file_impl: using device CUDA0 (NVIDIA GeForce RTX 5070) (0000:03:00.0) - 10929 MiB free llama_model_load_from_file_impl: using device CUDA1 (NVIDIA GeForce RTX 3060) (0000:04:00.0) - 11245 MiB free llama_model_load_from_file_impl: using device CUDA2 (NVIDIA GeForce RTX 3060) (0000:83:00.0) - 11245 MiB free [2025-10-25 20:35:44][DEBUG] llama_model_loader: loaded meta data with 44 key-value pairs and 771 tensors from C:\Users\Iman Mirbioki\.lmstudio\models\aratan\DeepSeek-R1-32B-Uncensored\DeepSeek-R1-32B-Uncensored.gguf (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen2 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = DeepSeek R1 Distill Qwen 32B Uncensored llama_model_loader: - kv 3: general.finetune str = Uncensored llama_model_loader: - kv 4: general.basename str = DeepSeek-R1-Distill-Qwen llama_model_loader: - kv 5: general.size_label str = 32B llama_model_loader: - kv 6: general.license str = mit llama_model_loader: - kv 7: general.base_model.count u32 = 1 llama_model_loader: - kv 8: general.base_model.0.name str = DeepSeek R1 Distill Qwen 32B llama_model_loader: - kv 9: general.base_model.0.organization str = Deepseek Ai llama_model_loader: - kv 10: general.base_model.0.repo_url str = https://huggingface.co/deepseek-ai/De... llama_model_loader: - kv 11: general.dataset.count u32 = 1 llama_model_loader: - kv 12: general.dataset.0.name str = Uncensor llama_model_loader: - kv 13: general.dataset.0.organization str = Guilherme34 llama_model_loader: - kv 14: general.dataset.0.repo_url str = https://huggingface.co/Guilherme34/un... llama_model_loader: - kv 15: general.tags arr[str,1] = ["generated_from_trainer"] llama_model_loader: - kv 16: qwen2.block_count u32 = 64 llama_model_loader: - kv 17: qwen2.context_length u32 = 131072 llama_model_loader: - kv 18: qwen2.embedding_length u32 = 5120 llama_model_loader: - kv 19: qwen2.feed_forward_length u32 = 27648 llama_model_loader: - kv 20: qwen2.attention.head_count u32 = 40 llama_model_loader: - kv 21: qwen2.attention.head_count_kv u32 = 8 llama_model_loader: - kv 22: qwen2.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 23: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 24: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 25: tokenizer.ggml.pre str = deepseek-r1-qwen [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 26: tokenizer.ggml.tokens arr[str,151665] = ["!", "\"", "#", "$", "%", "&", "'", ... [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 27: tokenizer.ggml.token_type arr[i32,151665] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 28: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 29: tokenizer.ggml.bos_token_id u32 = 151646 llama_model_loader: - kv 30: tokenizer.ggml.eos_token_id u32 = 151643 llama_model_loader: - kv 31: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 32: tokenizer.ggml.add_bos_token bool = true llama_model_loader: - kv 33: tokenizer.ggml.add_eos_token bool = false llama_model_loader: - kv 34: tokenizer.chat_template str = {% if not add_generation_prompt is de... llama_model_loader: - kv 35: general.quantization_version u32 = 2 llama_model_loader: - kv 36: general.file_type u32 = 30 llama_model_loader: - kv 37: general.url str = https://huggingface.co/mradermacher/D... llama_model_loader: - kv 38: mradermacher.quantize_version str = 2 llama_model_loader: - kv 39: mradermacher.quantized_by str = mradermacher llama_model_loader: - kv 40: mradermacher.quantized_at str = 2025-01-26T04:57:59+01:00 llama_model_loader: - kv 41: mradermacher.quantized_on str = marco llama_model_loader: - kv 42: general.source.url str = https://huggingface.co/nicoboss/DeepS... llama_model_loader: - kv 43: mradermacher.convert_type str = hf llama_model_loader: - type f32: 321 tensors llama_model_loader: - type q5_K: 72 tensors llama_model_loader: - type q6_K: 1 tensors llama_model_loader: - type iq4_xs: 377 tensors print_info: file format = GGUF V3 (latest) print_info: file type = IQ4_XS - 4.25 bpw print_info: file size = 16.63 GiB (4.36 BPW) [2025-10-25 20:35:44][DEBUG] load: special_eos_id is not in special_eog_ids - the tokenizer config may be incorrect load: printing all EOG tokens: load: - 151643 ('<|end▁of▁sentence|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') [2025-10-25 20:35:44][DEBUG] load: special tokens cache size = 22 [2025-10-25 20:35:44][DEBUG] load: token to piece cache size = 0.9310 MB print_info: arch = qwen2 print_info: vocab_only = 0 print_info: n_ctx_train = 131072 print_info: n_embd = 5120 print_info: n_layer = 64 print_info: n_head = 40 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: is_swa_any = 0 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 5 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-05 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 [2025-10-25 20:35:44][DEBUG] print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 27648 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = -1 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 131072 print_info: rope_finetuned = unknown print_info: model type = 32B print_info: model params = 32.76 B print_info: general.name = DeepSeek R1 Distill Qwen 32B Uncensored print_info: vocab type = BPE print_info: n_vocab = 151665 print_info: n_merges = 151387 print_info: BOS token = 151646 '<|begin▁of▁sentence|>' print_info: EOS token = 151643 '<|end▁of▁sentence|>' print_info: EOT token = 151643 '<|end▁of▁sentence|>' print_info: PAD token = 151643 '<|end▁of▁sentence|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|end▁of▁sentence|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = true) [2025-10-25 20:37:03][DEBUG] load_tensors: offloading 64 repeating layers to GPU load_tensors: offloading output layer to GPU load_tensors: offloaded 65/65 layers to GPU load_tensors: CUDA2 model buffer size = 5565.08 MiB load_tensors: CUDA1 model buffer size = 5453.34 MiB load_tensors: CUDA0 model buffer size = 5622.09 MiB load_tensors: CPU_Mapped model buffer size = 393.42 MiB [2025-10-25 20:37:19][DEBUG] llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 131072 llama_context: n_ctx_per_seq = 131072 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = disabled llama_context: kv_unified = false llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 [2025-10-25 20:37:19][DEBUG] llama_context: CUDA_Host output buffer size = 0.58 MiB ` I've been going through all my logs, and it seems like the problem started when Win11 update 25H2 was released - but it's just a guess.
Author
Owner

@pisarevel commented on GitHub (Oct 31, 2025):

I had the same problem.
While searching for a solution, I found a table of CUDA requirements for graphics card driver versions. I realized that my driver was outdated for the current CUDA version.
After updating the driver, everything begin to work as expected.
You can find the table here: https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html

<!-- gh-comment-id:3474825230 --> @pisarevel commented on GitHub (Oct 31, 2025): I had the same problem. While searching for a solution, I found a table of CUDA requirements for graphics card driver versions. I realized that my driver was outdated for the current CUDA version. After updating the driver, everything begin to work as expected. You can find the table here: https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html
Author
Owner

@catclaw commented on GitHub (Nov 3, 2025):

I've updated to the latest CUDA for Windows 11, my drivers match, too, MSVC C++ runtime is the latest (I've tried "repair", too) but after upgrading to the latest Ollama version (0.12.9), it still refuses to detect my GPUs.
Version 0.11.11 detects the GPUs, but falls back to the CPUs when trying to load a model.
The log is included. (For ver. 0.12.9)
I've tried several other tools like LM Studio, and they detect the GPUs with no problems?
Please help - I really need to get Ollama up and working... 🙏🙏🙏

serve.log

<!-- gh-comment-id:3482712495 --> @catclaw commented on GitHub (Nov 3, 2025): I've updated to the latest CUDA for Windows 11, my drivers match, too, MSVC C++ runtime is the latest (I've tried "repair", too) but after upgrading to the latest Ollama version (0.12.9), it still refuses to detect my GPUs. Version 0.11.11 detects the GPUs, but falls back to the CPUs when trying to load a model. The log is included. (For ver. 0.12.9) I've tried several other tools like LM Studio, and they detect the GPUs with no problems? Please help - I really need to get Ollama up and working... 🙏🙏🙏 [serve.log](https://github.com/user-attachments/files/23318478/serve.log)
Author
Owner

@dhiltgen commented on GitHub (Nov 4, 2025):

@catclaw with the updated logging code we have a little more detail, but unfortunately still not an exact root cause on what's causing this problem on your system. None of the dynamic libraries will load properly

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
time=2025-11-03T22:32:16.738+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException

For a missing dependency, I would expect the error to say "The specified module could not be found" so the fact you're seeing "procedure" seems to imply the way we're linking our libraries is causing an incompatibility with something on your system.

Do you have any other software on your system that's doing preload/injection? Any unusual AV software?

Maybe we'll be able to get a little more insight if you can scan the libraries on your system? You might try https://github.com/lucasg/Dependencies/ which has a nice GUI that can show dependencies, and in this case, maybe it will help us understand what is causing all the DLLs listed above to not load?

<!-- gh-comment-id:3487259161 --> @dhiltgen commented on GitHub (Nov 4, 2025): @catclaw with the updated logging code we have a little more detail, but unfortunately still not an exact root cause on what's causing this problem on your system. None of the dynamic libraries will load properly ``` dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found. System.Management.Automation.RemoteException dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found. System.Management.Automation.RemoteException dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found. System.Management.Automation.RemoteException dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found. System.Management.Automation.RemoteException dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found. System.Management.Automation.RemoteException dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found. System.Management.Automation.RemoteException dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found. System.Management.Automation.RemoteException time=2025-11-03T22:32:16.738+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found. System.Management.Automation.RemoteException ``` For a missing dependency, I would expect the error to say "The specified module could not be found" so the fact you're seeing "procedure" seems to imply the way we're linking our libraries is causing an incompatibility with something on your system. Do you have any other software on your system that's doing preload/injection? Any unusual AV software? Maybe we'll be able to get a little more insight if you can scan the libraries on your system? You might try https://github.com/lucasg/Dependencies/ which has a nice GUI that can show dependencies, and in this case, maybe it will help us understand what is causing all the DLLs listed above to not load?
Author
Owner

@catclaw commented on GitHub (Nov 4, 2025):

@catclaw with the updated logging code we have a little more detail, but unfortunately still not an exact root cause on what's causing this problem on your system. None of the dynamic libraries will load properly

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
time=2025-11-03T22:32:16.738+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException

For a missing dependency, I would expect the error to say "The specified module could not be found" so the fact you're seeing "procedure" seems to imply the way we're linking our libraries is causing an incompatibility with something on your system.

Do you have any other software on your system that's doing preload/injection? Any unusual AV software?

Maybe we'll be able to get a little more insight if you can scan the libraries on your system? You might try https://github.com/lucasg/Dependencies/ which has a nice GUI that can show dependencies, and in this case, maybe it will help us understand what is causing all the DLLs listed above to not load?

Will do! Thank you so much!
I don't have anything unusual on my system. I've been using AVG AV for years (paid version).
I'll try Dependencies (I'm familiar with it), and see what I can find.
I'm also going to try uninstalling AVG Antivirus and see if that helps. (It's impossible to stop the AV services fully.)
However, I don't think the problem is with the AV.
Ollama 0.11.8 and 0.11.11 detect my GPUs, but they fall back to the CPU when I try to load a model. (No matter the size.)
I'm also going to try rebooting with all but MS services and see if it helps.
Meanwhile, I've added the serve.log for both 0.11.8 (which detects my GPUs) and 0.12.9 (which doesn't).
It's strange, as LM Studio works in a similar way, but it detects my GPUs and loads the models to VRAM. (I used Gollama -L to link my 35+ models to LM Studio.)
This all happened overnight after I updated Windows 11 (Enterprise) from 24H2 to 25H2, which I became aware of quickly, as I use OllamaVision and MagicPrompt in SwarmUI to rewrite and optimize my prompts. I noticed how my system became almost unresponsive and CPU usage went sky-high.
I'll be back with more information - and thank you so much again for taking your time to look at this.
I'm really grateful.

serve_0.12_9.log
serve_0_11_8.log

<!-- gh-comment-id:3488205082 --> @catclaw commented on GitHub (Nov 4, 2025): > [@catclaw](https://github.com/catclaw) with the updated logging code we have a little more detail, but unfortunately still not an exact root cause on what's causing this problem on your system. None of the dynamic libraries will load properly > > ``` > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > time=2025-11-03T22:32:16.738+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > ``` > > For a missing dependency, I would expect the error to say "The specified module could not be found" so the fact you're seeing "procedure" seems to imply the way we're linking our libraries is causing an incompatibility with something on your system. > > Do you have any other software on your system that's doing preload/injection? Any unusual AV software? > > Maybe we'll be able to get a little more insight if you can scan the libraries on your system? You might try https://github.com/lucasg/Dependencies/ which has a nice GUI that can show dependencies, and in this case, maybe it will help us understand what is causing all the DLLs listed above to not load? Will do! Thank you so much! I don't have anything unusual on my system. I've been using AVG AV for years (paid version). I'll try Dependencies (I'm familiar with it), and see what I can find. I'm also going to try uninstalling AVG Antivirus and see if that helps. (It's impossible to stop the AV services fully.) However, I don't think the problem is with the AV. Ollama 0.11.8 and 0.11.11 detect my GPUs, but they fall back to the CPU when I try to load a model. (No matter the size.) I'm also going to try rebooting with all but MS services and see if it helps. Meanwhile, I've added the serve.log for both 0.11.8 (which detects my GPUs) and 0.12.9 (which doesn't). It's strange, as LM Studio works in a similar way, but it detects my GPUs and loads the models to VRAM. (I used Gollama -L to link my 35+ models to LM Studio.) This all happened overnight after I updated Windows 11 (Enterprise) from 24H2 to 25H2, which I became aware of quickly, as I use OllamaVision and MagicPrompt in SwarmUI to rewrite and optimize my prompts. I noticed how my system became almost unresponsive and CPU usage went sky-high. I'll be back with more information - and thank you so much again for taking your time to look at this. I'm really grateful. [serve_0.12_9.log](https://github.com/user-attachments/files/23346327/serve_0.12_9.log) [serve_0_11_8.log](https://github.com/user-attachments/files/23346326/serve_0_11_8.log)
Author
Owner

@catclaw commented on GitHub (Nov 5, 2025):

@catclaw with the updated logging code we have a little more detail, but unfortunately still not an exact root cause on what's causing this problem on your system. None of the dynamic libraries will load properly

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException
time=2025-11-03T22:32:16.738+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found.
System.Management.Automation.RemoteException

For a missing dependency, I would expect the error to say "The specified module could not be found" so the fact you're seeing "procedure" seems to imply the way we're linking our libraries is causing an incompatibility with something on your system.

Do you have any other software on your system that's doing preload/injection? Any unusual AV software?

Maybe we'll be able to get a little more insight if you can scan the libraries on your system? You might try https://github.com/lucasg/Dependencies/ which has a nice GUI that can show dependencies, and in this case, maybe it will help us understand what is causing all the DLLs listed above to not load?

This is what I found out:

[NOT_FOUND] ext-ms-win-security-chambers-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-oobe-query-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-core-app-package-registration-l1-1-0.dll :
[NOT_FOUND] ext-ms-win32-subsystem-query-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-com-suspendresiliency-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-appmodel-deployment-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-appmodel-usercontext-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-gaming-devicefamily-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-ui-viewmanagement-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-2.dll :
[NOT_FOUND] ext-ms-win-gdi-edgegdi-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-appmodel-viewscalefactor-l1-1-0.dll :
[NOT_FOUND] ext-ms-onecore-defaultdiscovery-l1-1-0.dll :
[NOT_FOUND] wpaxholder.dll :
[NOT_FOUND] ext-ms-windowscore-deviceinfo-l1-1-0.dll :
[NOT_FOUND] ext-ms-mf-pal-l2-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-input-l1-1-1.dll :
[NOT_FOUND] api-ms-win-coreui-secruntime-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-ntuser-private-l1-1-1.dll :
[NOT_FOUND] ext-ms-win-media-codecpack-mounting-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-core-winrt-remote-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-com-apartmentrestriction-l1-1-0.dll :
[NOT_FOUND] PdmUtilities.dll :
[NOT_FOUND] HvsiFileTrust.dll :
[NOT_FOUND] ext-ms-win-audiocore-pal-l1-2-0.dll :
[NOT_FOUND] ext-ms-win-audiocore-spatial-l1-1-0.dll :
[NOT_FOUND] UpdateAPI.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-1.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-0.dll :
[NOT_FOUND] AzureAttestManager.dll :
[NOT_FOUND] AzureAttestNormal.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-1-2.dll :
[NOT_FOUND] ext-ms-win-networking-wlanstorage-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-1-3.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-2-0.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-2-1.dll :
[NOT_FOUND] ext-ms-onecore-orientation-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-internal-l1-1-0.dll :
[NOT_FOUND] ext-ms-onecore-shellchromeapi-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-3.dll :
[NOT_FOUND] ext-ms-win-core-xbrm-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-core-xbrm-l1-1-1.dll :
[NOT_FOUND] ext-ms-win-ntdsa-activedirectoryserver-l1-1-0.dll :

After reinstalling the correct Visual C++ Redistributable package (once again - I had to uninstall the previous version), and running sfc /scannow, dism.exe /online /cleanup-image /scanhealth, dism.exe /online /cleanup-image /restorehealth, and dism.exe /online /cleanup-image /startcomponentcleanup - the DLL-files were still missing.

If you look at the end of apisets, some APIs call non-existing DLLs. ([ ]).

[-] Api Sets Map :
api-ms-onecoreuap-print-render-l1-1-0 -> [ printrenderapihost.dll ]
api-ms-win-appmodel-advertisingid-l1-1-0 -> [ kernel.appcore.dll ]
api-ms-win-appmodel-identity-l1-2-0 -> [ kernel.appcore.dll ]
api-ms-win-appmodel-lifecyclepolicy-l1-1-0 -> [ rmclient.dll ]
api-ms-win-appmodel-runtime-internal-l1-1-11 -> [ kernel.appcore.dll ]
api-ms-win-appmodel-runtime-l1-1-7 -> [ kernel.appcore.dll ]
api-ms-win-appmodel-state-l1-1-2 -> [ kernel.appcore.dll ]
api-ms-win-appmodel-state-l1-2-0 -> [ kernel.appcore.dll ]
api-ms-win-appmodel-unlock-l1-1-0 -> [ kernel.appcore.dll ]
api-ms-win-audiocore-spatial-config-l1-1-0 -> [ windows.media.devices.dll ]
api-ms-win-base-bootconfig-l1-1-0 -> [ advapi32.dll ]
api-ms-win-base-util-l1-1-0 -> [ advapi32.dll ]
api-ms-win-composition-redirection-l1-1-0 -> [ dwmredir.dll ]
api-ms-win-composition-windowmanager-l1-1-0 -> [ udwm.dll ]
api-ms-win-containers-cmclient-l1-1-1 -> [ cmclient.dll ]
api-ms-win-containers-cmclient-l1-2-0 -> [ cmclient.dll ]
api-ms-win-containers-cmclient-l1-3-0 -> [ cmclient.dll ]
api-ms-win-containers-cmclient-l1-4-0 -> [ cmclient.dll ]
api-ms-win-containers-cmclient-l1-5-3 -> [ cmclient.dll ]
api-ms-win-containers-cmdiagclient-l1-1-2 -> [ cmclient.dll ]
api-ms-win-containers-cmservicingclient-l1-1-1 -> [ cmclient.dll ]
api-ms-win-containers-cmservicingclient-l1-2-2 -> [ cmclient.dll ]
api-ms-win-core-apiquery-l1-1-2 -> [ ntdll.dll ]
api-ms-win-core-apiquery-l2-1-1 -> [ kernelbase.dll ]
api-ms-win-core-appcompat-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-appinit-l1-1-0 -> [ kernel32.dll,kernelbase.dll ]
api-ms-win-core-atoms-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-backgroundtask-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-bicltapi-l1-1-6 -> [ bi.dll ]
api-ms-win-core-biplmapi-l1-1-5 -> [ twinapi.appcore.dll ]
api-ms-win-core-biplmapi-l1-2-0 -> [ twinapi.appcore.dll ]
api-ms-win-core-biptcltapi-l1-1-7 -> [ twinapi.appcore.dll ]
api-ms-win-core-calendar-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-com-l1-1-3 -> [ combase.dll ]
api-ms-win-core-com-l2-1-1 -> [ coml2.dll ]
api-ms-win-core-com-midlproxystub-l1-1-0 -> [ combase.dll ]
api-ms-win-core-com-private-l1-1-1 -> [ combase.dll ]
api-ms-win-core-com-private-l1-2-0 -> [ combase.dll ]
api-ms-win-core-com-private-l1-3-1 -> [ combase.dll ]
api-ms-win-core-comm-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-commandlinetoargv-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-console-ansi-l2-1-0 -> [ kernel32.dll ]
api-ms-win-core-console-internal-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-console-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-console-l1-2-2 -> [ kernelbase.dll ]
api-ms-win-core-console-l2-1-0 -> [ kernelbase.dll ]
api-ms-win-core-console-l2-2-0 -> [ kernelbase.dll ]
api-ms-win-core-console-l3-1-0 -> [ kernelbase.dll ]
api-ms-win-core-console-l3-2-0 -> [ kernelbase.dll ]
api-ms-win-core-crt-l1-1-0 -> [ ntdll.dll ]
api-ms-win-core-crt-l2-1-0 -> [ kernelbase.dll ]
api-ms-win-core-datetime-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-debug-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-debug-minidump-l1-1-0 -> [ dbgcore.dll ]
api-ms-win-core-delayload-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-enclave-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-errorhandling-l1-1-3 -> [ kernelbase.dll ]
api-ms-win-core-featurestaging-l1-1-1 -> [ shcore.dll ]
api-ms-win-core-featuretoggles-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-fibers-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-fibers-l2-1-1 -> [ kernelbase.dll ]
api-ms-win-core-file-ansi-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-file-ansi-l2-1-0 -> [ kernel32.dll ]
api-ms-win-core-file-fromapp-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-file-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-file-l1-2-5 -> [ kernelbase.dll ]
api-ms-win-core-file-l2-1-4 -> [ kernelbase.dll ]
api-ms-win-core-firmware-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-guard-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-handle-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-heap-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-heap-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-heap-l2-1-0 -> [ kernelbase.dll ]
api-ms-win-core-heap-obsolete-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-interlocked-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-interlocked-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-io-l1-1-1 -> [ kernel32.dll,kernelbase.dll ]
api-ms-win-core-ioring-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-job-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-job-l2-1-1 -> [ kernel32.dll ]
api-ms-win-core-kernel32-legacy-ansi-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-kernel32-legacy-l1-1-6 -> [ kernel32.dll ]
api-ms-win-core-kernel32-private-l1-1-2 -> [ kernel32.dll ]
api-ms-win-core-kernel32-private-l1-2-0 -> [ kernel32.dll ]
api-ms-win-core-largeinteger-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-libraryloader-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-libraryloader-l1-2-3 -> [ kernelbase.dll ]
api-ms-win-core-libraryloader-l2-1-0 -> [ kernelbase.dll ]
api-ms-win-core-libraryloader-private-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-localization-ansi-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-localization-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-localization-l1-2-4 -> [ kernelbase.dll ]
api-ms-win-core-localization-l2-1-0 -> [ kernelbase.dll ]
api-ms-win-core-localization-obsolete-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-localization-obsolete-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-localization-obsolete-l1-3-0 -> [ kernelbase.dll ]
api-ms-win-core-localization-private-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-localregistry-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-marshal-l1-1-0 -> [ combase.dll ]
api-ms-win-core-memory-l1-1-9 -> [ kernelbase.dll ]
api-ms-win-core-misc-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-multipleproviderrouter-l1-1-0 -> [ mpr.dll ]
api-ms-win-core-namedpipe-ansi-l1-1-1 -> [ kernel32.dll ]
api-ms-win-core-namedpipe-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-namedpipe-l1-2-2 -> [ kernelbase.dll ]
api-ms-win-core-namespace-ansi-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-namespace-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-normalization-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-path-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-pcw-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-perfcounters-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-perfcounters-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-privateprofile-l1-1-1 -> [ kernel32.dll ]
api-ms-win-core-processenvironment-ansi-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-processenvironment-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-processenvironment-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-processsecurity-l1-1-0 -> [ kernel32.dll,kernelbase.dll ]
api-ms-win-core-processsnapshot-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-processthreads-l1-1-8 -> [ kernel32.dll,kernelbase.dll ]
api-ms-win-core-processtopology-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-processtopology-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-processtopology-obsolete-l1-1-1 -> [ kernel32.dll ]
api-ms-win-core-processtopology-private-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-profile-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-psapi-ansi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-psapi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-psapi-obsolete-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-psapiansi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-psm-app-l1-1-0 -> [ twinapi.appcore.dll ]
api-ms-win-core-psm-appnotify-l1-1-1 -> [ twinapi.appcore.dll ]
api-ms-win-core-psm-info-l1-1-1 -> [ appsruprov.dll ]
api-ms-win-core-psm-key-l1-1-3 -> [ kernelbase.dll ]
api-ms-win-core-psm-plm-l1-1-3 -> [ twinapi.appcore.dll ]
api-ms-win-core-psm-plm-l1-2-0 -> [ twinapi.appcore.dll ]
api-ms-win-core-psm-plm-l1-3-0 -> [ twinapi.appcore.dll ]
api-ms-win-core-psm-rtimer-l1-1-1 -> [ twinapi.appcore.dll ]
api-ms-win-core-psm-tc-l1-1-1 -> [ twinapi.appcore.dll ]
api-ms-win-core-quirks-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-realtime-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-registry-fromapp-l1-1-0 -> [ reguwpapi.dll ]
api-ms-win-core-registry-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-registry-l2-1-0 -> [ advapi32.dll ]
api-ms-win-core-registry-l2-2-0 -> [ advapi32.dll ]
api-ms-win-core-registry-l2-3-0 -> [ advapi32.dll ]
api-ms-win-core-registry-private-l1-1-0 -> [ advapi32.dll ]
api-ms-win-core-registryuserspecific-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-rtlsupport-l1-1-1 -> [ ntdll.dll ]
api-ms-win-core-rtlsupport-l1-2-2 -> [ ntdll.dll ]
api-ms-win-core-shlwapi-legacy-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-shlwapi-obsolete-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-shlwapi-obsolete-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-shutdown-ansi-l1-1-0 -> [ advapi32.dll ]
api-ms-win-core-shutdown-l1-1-1 -> [ advapi32.dll ]
api-ms-win-core-sidebyside-ansi-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-sidebyside-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-slapi-l1-1-0 -> [ clipc.dll ]
api-ms-win-core-state-helpers-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-string-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-string-l2-1-1 -> [ kernelbase.dll ]
api-ms-win-core-string-obsolete-l1-1-1 -> [ kernel32.dll ]
api-ms-win-core-stringansi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-stringloader-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-synch-ansi-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-synch-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-synch-l1-2-1 -> [ kernelbase.dll ]
api-ms-win-core-sysinfo-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-sysinfo-l1-2-7 -> [ kernelbase.dll ]
api-ms-win-core-sysinfo-l2-1-0 -> [ advapi32.dll ]
api-ms-win-core-systemtopology-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-core-textinput-client-l1-1-1 -> [ textinputframework.dll ]
api-ms-win-core-textinput-client-l1-2-0 -> [ textinputframework.dll ]
api-ms-win-core-threadpool-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-threadpool-l1-2-0 -> [ kernelbase.dll ]
api-ms-win-core-threadpool-legacy-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-threadpool-private-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-timezone-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-timezone-private-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-toolhelp-l1-1-1 -> [ kernel32.dll ]
api-ms-win-core-ums-l1-1-0 -> [ kernel32.dll ]
api-ms-win-core-url-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-util-l1-1-1 -> [ kernel32.dll,kernelbase.dll ]
api-ms-win-core-version-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-version-private-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-versionansi-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-core-windowsceip-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-core-windowserrorreporting-l1-1-3 -> [ kernelbase.dll ]
api-ms-win-core-winrt-error-l1-1-1 -> [ combase.dll ]
api-ms-win-core-winrt-errorprivate-l1-1-1 -> [ combase.dll ]
api-ms-win-core-winrt-l1-1-0 -> [ combase.dll ]
api-ms-win-core-winrt-propertysetprivate-l1-1-1 -> [ wintypes.dll ]
api-ms-win-core-winrt-registration-l1-1-0 -> [ combase.dll ]
api-ms-win-core-winrt-robuffer-l1-1-0 -> [ wintypes.dll ]
api-ms-win-core-winrt-roparameterizediid-l1-1-0 -> [ combase.dll ]
api-ms-win-core-winrt-string-l1-1-1 -> [ combase.dll ]
api-ms-win-core-wow64-l1-1-3 -> [ kernelbase.dll ]
api-ms-win-core-xstate-l1-1-3 -> [ ntdll.dll ]
api-ms-win-core-xstate-l2-1-2 -> [ kernelbase.dll ]
api-ms-win-coremessaging-host-l1-1-0 -> [  ]
api-ms-win-coreui-secruntime-l1-1-0 -> [  ]
api-ms-win-crt-conio-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-convert-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-environment-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-filesystem-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-heap-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-locale-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-math-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-multibyte-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-private-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-process-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-runtime-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-stdio-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-string-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-time-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-crt-utility-l1-1-0 -> [ ucrtbase.dll ]
api-ms-win-deprecated-apis-advapi-l1-1-0 -> [  ]
api-ms-win-deprecated-apis-legacy-l1-1-0 -> [  ]
api-ms-win-deprecated-apis-legacy-l1-2-0 -> [  ]
api-ms-win-deprecated-apis-obsolete-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-devices-config-l1-1-2 -> [ cfgmgr32.dll ]
api-ms-win-devices-query-l1-1-1 -> [ cfgmgr32.dll ]
api-ms-win-devices-swdevice-l1-1-1 -> [ cfgmgr32.dll ]
api-ms-win-downlevel-advapi32-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-downlevel-advapi32-l2-1-0 -> [ sechost.dll ]
api-ms-win-downlevel-advapi32-l3-1-0 -> [ ntmarta.dll ]
api-ms-win-downlevel-advapi32-l4-1-0 -> [ advapi32.dll ]
api-ms-win-downlevel-kernel32-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-downlevel-kernel32-l2-1-0 -> [ kernel32.dll ]
api-ms-win-downlevel-normaliz-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-downlevel-ole32-l1-1-0 -> [ combase.dll ]
api-ms-win-downlevel-shell32-l1-1-0 -> [ shcore.dll ]
api-ms-win-downlevel-shlwapi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-downlevel-shlwapi-l2-1-0 -> [ shcore.dll ]
api-ms-win-downlevel-user32-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-downlevel-version-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-dwmapi-l1-1-0 -> [ dwmapi.dll ]
api-ms-win-dx-d3dkmt-l1-1-8 -> [ gdi32.dll ]
api-ms-win-eventing-classicprovider-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-eventing-consumer-l1-1-2 -> [ sechost.dll ]
api-ms-win-eventing-controller-l1-1-1 -> [ sechost.dll ]
api-ms-win-eventing-legacy-l1-1-0 -> [ advapi32.dll ]
api-ms-win-eventing-obsolete-l1-1-0 -> [ sechost.dll ]
api-ms-win-eventing-provider-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-eventing-tdh-l1-1-2 -> [ tdh.dll ]
api-ms-win-eventlog-legacy-l1-1-0 -> [ advapi32.dll ]
api-ms-win-gaming-deviceinformation-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-gaming-expandedresources-l1-1-0 -> [ gamemode.dll ]
api-ms-win-gaming-experience-l1-1-0 -> [ gamemode.dll ]
api-ms-win-gaming-tcui-l1-1-4 -> [ gamingtcui.dll ]
api-ms-win-gdi-dpiinfo-l1-1-0 -> [ gdi32.dll ]
api-ms-win-gdi-internal-uap-l1-1-0 -> [ gdi32full.dll ]
api-ms-win-ham-apphistory-l1-1-0 -> [ rmclient.dll ]
api-ms-win-ham-hamplm-l1-1-0 -> [ rmclient.dll ]
api-ms-win-http-time-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-input-ie-interactioncontext-l1-1-0 -> [  ]
api-ms-win-legacy-shlwapi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-mm-joystick-l1-1-0 -> [ winmm.dll ]
api-ms-win-mm-mci-l1-1-0 -> [ winmm.dll ]
api-ms-win-mm-misc-l1-1-1 -> [ winmmbase.dll ]
api-ms-win-mm-misc-l2-1-0 -> [ winmm.dll ]
api-ms-win-mm-mme-l1-1-0 -> [ winmmbase.dll ]
api-ms-win-mm-playsound-l1-1-0 -> [ winmm.dll ]
api-ms-win-mm-time-l1-1-0 -> [ kernel32.dll ]
api-ms-win-net-isolation-l1-1-1 -> [ firewallapi.dll ]
api-ms-win-networking-interfacecontexts-l1-1-0 -> [ ondemandconnroutehelper.dll ]
api-ms-win-ngc-serialization-l1-1-1 -> [ ngckeyenum.dll ]
api-ms-win-ntuser-ie-message-l1-1-0 -> [ user32.dll ]
api-ms-win-ntuser-ie-window-l1-1-0 -> [ user32.dll ]
api-ms-win-ntuser-ie-wmpointer-l1-1-0 -> [ user32.dll ]
api-ms-win-ntuser-rectangle-l1-1-0 -> [ user32.dll ]
api-ms-win-ntuser-sysparams-l1-1-0 -> [ user32.dll ]
api-ms-win-obsolete-localization-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-obsolete-psapi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-obsolete-shlwapi-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-ole32-ie-l1-1-0 -> [ ole32.dll ]
api-ms-win-oobe-notification-l1-1-0 -> [ kernel32.dll ]
api-ms-win-perf-legacy-l1-1-0 -> [ advapi32.dll ]
api-ms-win-power-base-l1-1-0 -> [ powrprof.dll ]
api-ms-win-power-limitsmanagement-l1-1-0 -> [ powrprof.dll ]
api-ms-win-power-setting-l1-1-1 -> [ powrprof.dll ]
api-ms-win-privacy-coreprivacysettingsstore-l1-1-0 -> [ coreprivacysettingsstore.dll ]
api-ms-win-ro-typeresolution-l1-1-1 -> [ wintypes.dll ]
api-ms-win-rtcore-minuser-private-l1-1-1 -> [  ]
api-ms-win-rtcore-navigation-l1-1-0 -> [  ]
api-ms-win-rtcore-ntuser-clipboard-l1-1-0 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-draw-l1-1-0 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-powermanagement-l1-1-0 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-private-l1-1-11 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-shell-l1-1-0 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-synch-l1-1-0 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-window-l1-1-0 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-winevent-l1-1-0 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-wmpointer-l1-1-3 -> [ user32.dll ]
api-ms-win-rtcore-ntuser-wmpointer-l1-2-0 -> [ user32.dll ]
api-ms-win-rtcore-ole32-clipboard-l1-1-1 -> [ ole32.dll ]
api-ms-win-rtcore-session-l1-1-0 -> [  ]
api-ms-win-security-accesshlpr-l1-1-0 -> [ sechost.dll ]
api-ms-win-security-activedirectoryclient-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-security-appcontainer-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-security-audit-l1-1-1 -> [ sechost.dll ]
api-ms-win-security-base-ansi-l1-1-0 -> [ advapi32.dll ]
api-ms-win-security-base-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-security-base-l1-2-2 -> [ kernelbase.dll ]
api-ms-win-security-base-private-l1-1-2 -> [ kernelbase.dll ]
api-ms-win-security-capability-l1-1-1 -> [ sechost.dll ]
api-ms-win-security-cpwl-l1-1-0 -> [ advapi32.dll ]
api-ms-win-security-credentials-l1-1-0 -> [ sechost.dll ]
api-ms-win-security-credentials-l2-1-1 -> [ sechost.dll ]
api-ms-win-security-cryptoapi-l1-1-0 -> [ cryptsp.dll ]
api-ms-win-security-grouppolicy-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-security-isolatedcontainer-l1-1-1 -> [ shcore.dll ]
api-ms-win-security-isolationapi-l1-1-0 -> [ sechost.dll ]
api-ms-win-security-isolationapi-l1-2-0 -> [ sechost.dll ]
api-ms-win-security-isolationpolicy-l1-1-0 -> [ sechost.dll ]
api-ms-win-security-isolationpolicy-l1-2-0 -> [ sechost.dll ]
api-ms-win-security-licenseprotection-l1-1-0 -> [ licenseprotection.dll ]
api-ms-win-security-logon-l1-1-1 -> [ advapi32.dll ]
api-ms-win-security-lsalookup-ansi-l2-1-0 -> [ advapi32.dll ]
api-ms-win-security-lsalookup-l1-1-2 -> [ sechost.dll ]
api-ms-win-security-lsalookup-l2-1-1 -> [ advapi32.dll ]
api-ms-win-security-lsapolicy-l1-1-2 -> [ sechost.dll ]
api-ms-win-security-provider-ansi-l1-1-0 -> [ advapi32.dll ]
api-ms-win-security-provider-l1-1-0 -> [ ntmarta.dll ]
api-ms-win-security-sddl-ansi-l1-1-0 -> [ advapi32.dll ]
api-ms-win-security-sddl-l1-1-0 -> [ sechost.dll ]
api-ms-win-security-sddl-private-l1-1-0 -> [ sechost.dll ]
api-ms-win-security-sddlparsecond-l1-1-1 -> [ sechost.dll ]
api-ms-win-security-systemfunctions-l1-1-0 -> [ advapi32.dll ]
api-ms-win-security-trustee-l1-1-2 -> [ advapi32.dll ]
api-ms-win-service-core-ansi-l1-1-1 -> [ advapi32.dll ]
api-ms-win-service-core-l1-1-5 -> [ sechost.dll ]
api-ms-win-service-legacy-l1-1-0 -> [ advapi32.dll ]
api-ms-win-service-management-l1-1-0 -> [ sechost.dll ]
api-ms-win-service-management-l2-1-0 -> [ sechost.dll ]
api-ms-win-service-private-l1-1-5 -> [ sechost.dll ]
api-ms-win-service-private-l1-2-2 -> [ sechost.dll ]
api-ms-win-service-winsvc-l1-1-0 -> [ sechost.dll ]
api-ms-win-service-winsvc-l1-2-0 -> [ sechost.dll ]
api-ms-win-shcore-comhelpers-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-obsolete-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-path-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-registry-l1-1-1 -> [ shcore.dll ]
api-ms-win-shcore-scaling-l1-1-2 -> [ shcore.dll ]
api-ms-win-shcore-stream-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-stream-winrt-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-sysinfo-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-taskpool-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-thread-l1-1-0 -> [ shcore.dll ]
api-ms-win-shcore-unicodeansi-l1-1-0 -> [ shcore.dll ]
api-ms-win-shell-associations-l1-1-3 -> [ windows.storage.dll ]
api-ms-win-shell-changenotify-l1-1-1 -> [ windows.storage.dll ]
api-ms-win-shell-dataobject-l1-1-1 -> [ windows.storage.dll ]
api-ms-win-shell-namespace-l1-1-1 -> [ windows.storage.dll ]
api-ms-win-shell-shdirectory-l1-1-0 -> [ shcore.dll ]
api-ms-win-shell-shell32legacy-shdirectory-l1-1-0 -> [  ]
api-ms-win-shell-shellcom-l1-1-0 -> [ kernelbase.dll ]
api-ms-win-shell-shellfolders-l1-1-1 -> [ windows.storage.dll ]
api-ms-win-shlwapi-ie-l1-1-0 -> [ shlwapi.dll ]
api-ms-win-shlwapi-winrt-storage-l1-1-1 -> [ shlwapi.dll ]
api-ms-win-stateseparation-helpers-l1-1-1 -> [ kernelbase.dll ]
api-ms-win-storage-exports-external-l1-1-2 -> [ windows.storage.dll ]
api-ms-win-storage-exports-internal-l1-1-0 -> [ windows.storage.dll ]
api-ms-win-storage-reserve-l1-1-0 -> [ storageusage.dll ]
api-ms-win-winrt-search-folder-l1-1-1 -> [ windows.storage.search.dll ]
api-ms-win-wsl-api-l1-1-0 -> [ wslapi.dll ]
ext-ms-mf-pal-l2-1-1 -> [  ]
ext-ms-net-eap-sim-l1-1-0 -> [ eapsimextdesktop.dll ]
ext-ms-net-vpn-soh-l1-1-0 -> [ vpnsohdesktop.dll ]
ext-ms-onecore-appchromeapi-l1-1-0 -> [  ]
ext-ms-onecore-appdefaults-l1-1-0 -> [ windows.storage.dll ]
ext-ms-onecore-appmodel-deployment-internal-l1-1-2 -> [ appxdeploymentclient.dll ]
ext-ms-onecore-appmodel-emclient-l1-1-0 -> [  ]
ext-ms-onecore-appmodel-emsvcs-l1-1-0 -> [  ]
ext-ms-onecore-appmodel-pacmanclient-l1-1-0 -> [  ]
ext-ms-onecore-appmodel-staterepository-appextension-l1-1-0 -> [ windows.staterepositoryclient.dll ]
ext-ms-onecore-appmodel-staterepository-cache-l1-1-5 -> [ windows.staterepositorycore.dll ]
ext-ms-onecore-appmodel-staterepository-internal-l1-1-7 -> [ windows.staterepositoryclient.dll ]
ext-ms-onecore-appmodel-staterepository-pkgextension-l1-1-0 -> [ windows.staterepositoryclient.dll ]
ext-ms-onecore-appmodel-tdlmigration-l1-1-1 -> [ tdlmigration.dll ]
ext-ms-onecore-comp-dwmmonitor-l1-1-0 -> [  ]
ext-ms-onecore-dcomp-l1-1-0 -> [ dcomp.dll ]
ext-ms-onecore-defaultdiscovery-l1-1-0 -> [  ]
ext-ms-onecore-hcap-svf-l1-1-0 -> [  ]
ext-ms-onecore-hlink-l1-1-0 -> [ hlink.dll ]
ext-ms-onecore-hnetcfg-l1-1-0 -> [ hnetcfgclient.dll ]
ext-ms-onecore-ipnathlp-l1-1-0 -> [ ipnathlpclient.dll ]
ext-ms-onecore-mpc-input-l1-1-0 -> [  ]
ext-ms-onecore-orientation-l1-1-0 -> [  ]
ext-ms-onecore-service-devicedirectory-claims-l1-1-0 -> [ ddcclaimsapi.dll ]
ext-ms-onecore-shellchromeapi-l1-1-2 -> [  ]
ext-ms-onecore-shellremindersapi-l1-1-0 -> [  ]
ext-ms-onecore-shlwapi-l1-1-0 -> [ shlwapi.dll ]
ext-ms-onecore-spectrumsyncclient-l1-1-0 -> [  ]
ext-ms-win-accel-api-km-l1-1-0 -> [ winaccel.sys ]
ext-ms-win-adsi-activeds-l1-1-0 -> [ activeds.dll ]
ext-ms-win-advapi32-auth-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-encryptedfile-l1-1-1 -> [ advapi32.dll ]
ext-ms-win-advapi32-eventlog-ansi-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-eventlog-l1-1-2 -> [ advapi32.dll ]
ext-ms-win-advapi32-hwprof-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-idletask-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-lsa-l1-1-4 -> [ advapi32.dll ]
ext-ms-win-advapi32-msi-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-npusername-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-ntmarta-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-psm-app-l1-1-0 -> [ twinapi.appcore.dll ]
ext-ms-win-advapi32-registry-l1-1-1 -> [ advapi32.dll ]
ext-ms-win-advapi32-safer-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-advapi32-shutdown-l1-1-0 -> [ advapi32.dll ]
ext-ms-win-appcompat-aeinv-l1-1-1 -> [ aeinv.dll ]
ext-ms-win-appcompat-aepic-l1-1-0 -> [ aepic.dll ]
ext-ms-win-appcompat-apphelp-l1-1-2 -> [ apphelp.dll ]
ext-ms-win-appcompat-pcacli-l1-1-0 -> [ pcacli.dll ]
ext-ms-win-appmodel-activation-l1-1-2 -> [ activationmanager.dll ]
ext-ms-win-appmodel-appcontainerpath-l1-1-0 -> [  ]
ext-ms-win-appmodel-appexecutionalias-l1-1-5 -> [ apisethost.appexecutionalias.dll ]
ext-ms-win-appmodel-datasharingservice-extensions-l1-1-0 -> [  ]
ext-ms-win-appmodel-daxcore-l1-1-3 -> [ daxexec.dll ]
ext-ms-win-appmodel-deployment-l1-1-1 -> [  ]
ext-ms-win-appmodel-deploymentvolumes-l1-1-1 -> [  ]
ext-ms-win-appmodel-opc-l1-1-0 -> [ opcservices.dll ]
ext-ms-win-appmodel-registrycompatibility-l1-1-0 -> [ appxdeploymentextensions.desktop.dll ]
ext-ms-win-appmodel-restrictedappcontainer-internal-l1-1-0 -> [ kernel.appcore.dll ]
ext-ms-win-appmodel-shellexecute-l1-1-0 -> [ windows.storage.dll ]
ext-ms-win-appmodel-state-ext-l1-2-0 -> [ kernel.appcore.dll ]
ext-ms-win-appmodel-usercontext-l1-1-0 -> [  ]
ext-ms-win-appmodel-viewscalefactor-l1-1-0 -> [  ]
ext-ms-win-appxdeploymentclient-appxdeploy-l1-1-1 -> [ appxdeploymentclient.dll ]
ext-ms-win-appxdeploymentclient-appxdeployonecore-l1-1-1 -> [ appxdeploymentclient.dll ]
ext-ms-win-audio-spatial-systemsound-l1-1-0 -> [  ]
ext-ms-win-audiocore-coreaudiopolicymanager-l1-1-0 -> [ coreaudiopolicymanagerext.dll ]
ext-ms-win-audiocore-pal-l1-2-0 -> [  ]
ext-ms-win-audiocore-policymanager-l1-1-1 -> [  ]
ext-ms-win-audiocore-spatial-l1-1-0 -> [  ]
ext-ms-win-authz-claimpolicies-l1-1-0 -> [ authz.dll ]
ext-ms-win-authz-context-l1-1-0 -> [ authz.dll ]
ext-ms-win-authz-remote-l1-1-0 -> [ logoncli.dll ]
ext-ms-win-base-psapi-l1-1-0 -> [ psapi.dll ]
ext-ms-win-base-rstrtmgr-l1-1-0 -> [ rstrtmgr.dll ]
ext-ms-win-biometrics-winbio-core-l1-1-7 -> [ winbio.dll ]
ext-ms-win-biometrics-winbio-l1-1-0 -> [ winbioext.dll ]
ext-ms-win-biometrics-winbio-l1-2-0 -> [ winbioext.dll ]
ext-ms-win-biometrics-winbio-l1-3-0 -> [ winbioext.dll ]
ext-ms-win-bluetooth-apis-internal-l1-1-0 -> [ bluetoothapis.dll ]
ext-ms-win-bluetooth-apis-l1-1-0 -> [ bluetoothapis.dll ]
ext-ms-win-bluetooth-apis-private-l1-1-0 -> [ bluetoothapis.dll ]
ext-ms-win-branding-winbrand-l1-1-2 -> [ winbrand.dll ]
ext-ms-win-branding-winbrand-l1-2-0 -> [ winbrand.dll ]
ext-ms-win-capabilityaccessmanager-storage-l1-1-0 -> [ capabilityaccessmanager.desktop.storage.dll ]
ext-ms-win-casting-device-l1-1-0 -> [  ]
ext-ms-win-casting-lockscreen-l1-1-0 -> [ miracastreceiverext.dll ]
ext-ms-win-casting-receiver-l1-1-1 -> [  ]
ext-ms-win-casting-shell-l1-1-0 -> [ castingshellext.dll ]
ext-ms-win-ci-management-l1-1-3 -> [ manageci.dll ]
ext-ms-win-ci-xbox-l1-1-0 -> [  ]
ext-ms-win-cloudap-tbal-l1-1-0 -> [  ]
ext-ms-win-clouddomainjoin-usermanagement-l1-1-0 -> [  ]
ext-ms-win-cluster-clusapi-l1-1-6 -> [ clusapi.dll ]
ext-ms-win-cluster-resutils-l1-1-3 -> [ resutils.dll ]
ext-ms-win-cmd-util-l1-1-0 -> [ cmdext.dll ]
ext-ms-win-cng-rng-l1-1-1 -> [ bcryptprimitives.dll ]
ext-ms-win-com-apartmentrestriction-l1-1-0 -> [  ]
ext-ms-win-com-clbcatq-l1-1-0 -> [ clbcatq.dll ]
ext-ms-win-com-coml2-l1-1-1 -> [ coml2.dll ]
ext-ms-win-com-ole32-l1-1-5 -> [ ole32.dll ]
ext-ms-win-com-ole32-l1-2-0 -> [ ole32.dll ]
ext-ms-win-com-ole32-l1-3-0 -> [ ole32.dll ]
ext-ms-win-com-ole32-l1-4-0 -> [ ole32.dll ]
ext-ms-win-com-psmregister-l1-1-0 -> [ kernel.appcore.dll ]
ext-ms-win-com-psmregister-l1-2-2 -> [ kernel.appcore.dll ]
ext-ms-win-com-psmregister-l1-3-1 -> [ kernel.appcore.dll ]
ext-ms-win-com-sta-l1-1-0 -> [ ole32.dll ]
ext-ms-win-com-suspendresiliency-l1-1-0 -> [  ]
ext-ms-win-composition-ghost-l1-1-0 -> [ dwmghost.dll ]
ext-ms-win-composition-holographic-l1-1-0 -> [  ]
ext-ms-win-composition-init-l1-1-0 -> [ dwminit.dll ]
ext-ms-win-compositor-hosting-l1-1-1 -> [ ism.dll ]
ext-ms-win-compositor-hosting-l1-2-1 -> [ ism.dll ]
ext-ms-win-compositor-hosting-l1-3-0 -> [ ism.dll ]
ext-ms-win-connectionattribution-api-l1-1-0 -> [ connectionattributionapi.dll ]
ext-ms-win-containers-policymanagercli-l1-1-1 -> [  ]
ext-ms-win-core-app-package-registration-l1-1-1 -> [  ]
ext-ms-win-core-app-package-volume-l1-1-0 -> [  ]
ext-ms-win-core-container-init-l1-1-0 -> [  ]
ext-ms-win-core-dhcp6client-l1-1-0 -> [  ]
ext-ms-win-core-game-streaming-l1-1-0 -> [ gamestreamingext.dll ]
ext-ms-win-core-iuri-l1-1-0 -> [ urlmon.dll ]
ext-ms-win-core-licensemanager-l1-1-2 -> [  ]
ext-ms-win-core-marshal-l2-1-0 -> [ ole32.dll ]
ext-ms-win-core-pkeyhelper-l1-1-0 -> [ pkeyhelper.dll ]
ext-ms-win-core-psm-bi-l1-1-0 -> [ bisrv.dll ]
ext-ms-win-core-psm-bi-l1-2-0 -> [ bisrv.dll ]
ext-ms-win-core-psm-extendedresourcemode-l1-1-0 -> [  ]
ext-ms-win-core-psm-service-l1-1-6 -> [ psmserviceexthost.dll ]
ext-ms-win-core-resourcemanager-l1-1-0 -> [ rmclient.dll ]
ext-ms-win-core-resourcemanager-l1-2-1 -> [ rmclient.dll ]
ext-ms-win-core-resourcepolicy-l1-1-2 -> [ resourcepolicyclient.dll ]
ext-ms-win-core-resourcepolicyserver-l1-1-1 -> [ resourcepolicyserver.dll ]
ext-ms-win-core-stateseparationext-l1-1-0 -> [  ]
ext-ms-win-core-storelicensing-l1-1-0 -> [ licensemanagerapi.dll ]
ext-ms-win-core-storelicensing-l1-2-1 -> [ licensemanagerapi.dll ]
ext-ms-win-core-symbolicnames-l1-1-0 -> [ tdhres.dll ]
ext-ms-win-core-win32k-base-export-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-baseinit-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-common-export-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-common-input-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-common-inputrim-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-common-user-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-dcomp-l1-1-3 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-ddccigdi-l1-1-1 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-dxgdi-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-dxgk-internal-l1-1-0 -> [ dxgkrnl.sys ]
ext-ms-win-core-win32k-dxgk-l1-1-0 -> [ dxgkrnl.sys ]
ext-ms-win-core-win32k-flipmgr-l1-1-1 -> [ dxgkrnl.sys ]
ext-ms-win-core-win32k-full-export-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-full-float-export-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-full-floatobj-export-l1-1-0 -> [  ]
ext-ms-win-core-win32k-fulldcompbase-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-fulldwm-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-fullgdi-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-fulluser-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-fulluser64-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-core-win32k-fulluserbase-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-gdi-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-input-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-inputmit-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-inputrim-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-mindwm-l1-1-0 -> [  ]
ext-ms-win-core-win32k-mininput-l1-1-0 -> [  ]
ext-ms-win-core-win32k-mininputmit-l1-1-0 -> [  ]
ext-ms-win-core-win32k-mininputmitbase-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-minuser-l1-1-0 -> [  ]
ext-ms-win-core-win32k-opmgdi-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-surfmgr-l1-1-1 -> [ dxgkrnl.sys ]
ext-ms-win-core-win32k-tokenmgr-l1-1-0 -> [ dxgkrnl.sys ]
ext-ms-win-core-win32k-user-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-userdisplay-l1-1-0 -> [ win32kbase.sys ]
ext-ms-win-core-win32k-userinit-l1-1-0 -> [ win32k.sys ]
ext-ms-win-core-winrt-remote-l1-1-0 -> [  ]
ext-ms-win-core-winsrv-l1-1-0 -> [ winsrvext.dll ]
ext-ms-win-core-winsrv-min-l1-1-0 -> [  ]
ext-ms-win-core-xbrm-l1-1-1 -> [  ]
ext-ms-win-coreui-l1-1-0 -> [  ]
ext-ms-win-coreui-navshutdown-l1-1-0 -> [ navshutdown.dll ]
ext-ms-win-crypto-xbox-l1-1-0 -> [  ]
ext-ms-win-deployment-productenumerator-l1-1-0 -> [ productenumerator.dll ]
ext-ms-win-desktopappx-l1-1-7 -> [ daxexec.dll ]
ext-ms-win-desktopappx-l1-2-2 -> [ daxexec.dll ]
ext-ms-win-devmgmt-dm-l1-1-3 -> [ dmapisetextimpldesktop.dll ]
ext-ms-win-devmgmt-policy-l1-1-3 -> [ policymanager.dll ]
ext-ms-win-direct2d-desktop-l1-1-0 -> [ direct2ddesktop.dll ]
ext-ms-win-domainjoin-netjoin-l1-1-0 -> [ netjoin.dll ]
ext-ms-win-dot3-grouppolicy-l1-1-0 -> [ dot3gpclnt.dll ]
ext-ms-win-driver-recovery-l1-1-0 -> [ drvsetup.dll ]
ext-ms-win-driver-setup-l1-1-0 -> [ drvsetup.dll ]
ext-ms-win-driver-setup-wu-l1-1-1 -> [ drvsetup.dll ]
ext-ms-win-drvinst-desktop-l1-1-0 -> [ newdev.dll ]
ext-ms-win-dwmapi-ext-l1-1-2 -> [ dwmapi.dll ]
ext-ms-win-dwmapidxgi-ext-l1-1-1 -> [ dwmapi.dll ]
ext-ms-win-dx-d3d9-l1-1-0 -> [ d3d9.dll ]
ext-ms-win-dx-d3dkmt-dxcore-l1-1-5 -> [ dxcore.dll ]
ext-ms-win-dx-d3dkmt-gdi-l1-1-0 -> [ gdi32.dll ]
ext-ms-win-dx-ddraw-l1-1-0 -> [ ddraw.dll ]
ext-ms-win-dx-dinput8-l1-1-0 -> [ dinput8.dll ]
ext-ms-win-dx-dxdbhelper-l1-1-4 -> [ directxdatabasehelper.dll ]
ext-ms-win-dxcore-internal-l1-1-0 -> [ dxcore.dll ]
ext-ms-win-dxcore-l1-1-0 -> [ dxcore.dll ]
ext-ms-win-edputil-policy-l1-1-2 -> [ edputil.dll ]
ext-ms-win-els-elscore-l1-1-0 -> [ elscore.dll ]
ext-ms-win-eventing-pdh-l1-1-3 -> [ pdh.dll ]
ext-ms-win-eventing-rundown-l1-1-0 -> [ etwrundown.dll ]
ext-ms-win-eventing-tdh-ext-l1-1-0 -> [ tdh.dll ]
ext-ms-win-eventing-tdh-priv-l1-1-0 -> [ tdh.dll ]
ext-ms-win-eventing-wdi-l1-1-0 -> [ wdi.dll ]
ext-ms-win-familysafety-childaccount-l1-1-0 -> [ familysafetyext.dll ]
ext-ms-win-feclient-encryptedfile-l1-1-3 -> [ feclient.dll ]
ext-ms-win-firewallapi-webproxy-l1-1-1 -> [ firewallapi.dll ]
ext-ms-win-font-fontgroups-l1-1-0 -> [ fontgroupsoverride.dll ]
ext-ms-win-font-setup-l1-1-0 -> [ muifontsetup.dll ]
ext-ms-win-fs-clfs-l1-1-0 -> [ clfs.sys ]
ext-ms-win-fs-cscapi-l1-1-1 -> [ cscapi.dll ]
ext-ms-win-fs-vssapi-l1-1-0 -> [ vssapi.dll ]
ext-ms-win-fsutilext-ifsutil-l1-1-0 -> [ fsutilext.dll ]
ext-ms-win-fsutilext-ulib-l1-1-0 -> [ fsutilext.dll ]
ext-ms-win-fveapi-query-l1-1-0 -> [ fveapi.dll ]
ext-ms-win-gaming-devicefamily-l1-1-0 -> [  ]
ext-ms-win-gaming-gamechatoverlay-l1-1-0 -> [ gamechatoverlayext.dll ]
ext-ms-win-gaming-xblgamesave-l1-1-0 -> [ xblgamesaveext.dll ]
ext-ms-win-gaming-xinput-l1-1-0 -> [ xinputuap.dll ]
ext-ms-win-gdi-clipping-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-gdi-dc-create-l1-1-2 -> [ gdi32full.dll ]
ext-ms-win-gdi-dc-l1-2-1 -> [ gdi32full.dll ]
ext-ms-win-gdi-devcaps-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-gdi-draw-l1-1-3 -> [ gdi32full.dll ]
ext-ms-win-gdi-edgegdi-l1-1-0 -> [  ]
ext-ms-win-gdi-font-l1-1-3 -> [ gdi32full.dll ]
ext-ms-win-gdi-gdiplus-l1-1-0 -> [ gdiplus.dll ]
ext-ms-win-gdi-internal-desktop-l1-1-6 -> [ gdi32full.dll ]
ext-ms-win-gdi-internal-uap-init-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-gdi-metafile-l1-1-2 -> [ gdi32full.dll ]
ext-ms-win-gdi-path-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-gdi-print-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-gdi-private-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-gdi-render-l1-1-0 -> [ gdi32.dll ]
ext-ms-win-gdi-rgn-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-gdi-wcs-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-globalization-input-l1-1-3 -> [ globinputhost.dll ]
ext-ms-win-gpapi-grouppolicy-l1-1-0 -> [ gpapi.dll ]
ext-ms-win-gpsvc-grouppolicy-l1-1-0 -> [ gpsvc.dll ]
ext-ms-win-gui-dui70-l1-1-0 -> [ dui70.dll ]
ext-ms-win-gui-ieui-l1-1-0 -> [ ieui.dll ]
ext-ms-win-gui-uxinit-l1-1-1 -> [ uxinit.dll ]
ext-ms-win-hcihealth-utils-l1-1-0 -> [  ]
ext-ms-win-hostactivitymanager-bi-ham-ext-l1-1-0 -> [ psmserviceexthost.dll ]
ext-ms-win-hostactivitymanager-ham-private-ext-l1-1-0 -> [ psmserviceexthost.dll ]
ext-ms-win-hostactivitymanager-hostidstore-l1-1-1 -> [ rmclient.dll ]
ext-ms-win-hyperv-compute-l1-1-1 -> [  ]
ext-ms-win-hyperv-compute-l1-2-5 -> [ computecore.dll ]
ext-ms-win-hyperv-compute-legacy-l1-1-0 -> [ vmcompute.dll ]
ext-ms-win-hyperv-computenetwork-l1-1-1 -> [ computenetwork.dll ]
ext-ms-win-hyperv-computestorage-l1-1-2 -> [ computestorage.dll ]
ext-ms-win-hyperv-devicevirtualization-l1-1-1 -> [ vmdevicehost.dll ]
ext-ms-win-hyperv-devicevirtualization-l1-2-2 -> [ vmdevicehost.dll ]
ext-ms-win-hyperv-hgs-l1-1-0 -> [ vmhgs.dll ]
ext-ms-win-hyperv-hvemulation-l1-1-0 -> [ winhvemulation.dll ]
ext-ms-win-hyperv-hvplatform-l1-1-5 -> [ winhvplatform.dll ]
ext-ms-win-ie-textinput-l1-1-0 -> [  ]
ext-ms-win-imm-l1-1-3 -> [ imm32.dll ]
ext-ms-win-kernel32-appcompat-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-datetime-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-elevation-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-errorhandling-l1-1-0 -> [ kernel32.dll,faultrep.dll ]
ext-ms-win-kernel32-file-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-localization-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-package-current-l1-1-0 -> [ kernel.appcore.dll ]
ext-ms-win-kernel32-package-l1-1-2 -> [ kernel.appcore.dll ]
ext-ms-win-kernel32-process-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-quirks-l1-1-1 -> [ kernel32.dll ]
ext-ms-win-kernel32-registry-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-sidebyside-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-transacted-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-updateresource-l1-1-0 -> [ kernel32.dll ]
ext-ms-win-kernel32-windowserrorreporting-l1-1-1 -> [ kernel32.dll ]
ext-ms-win-kernelbase-processthread-l1-1-3 -> [ kernel32.dll ]
ext-ms-win-kernelbase-processthread-l1-2-0 -> [ kernel32.dll ]
ext-ms-win-kioskmode-config-l1-1-0 -> [  ]
ext-ms-win-laps-l1-1-1 -> [ laps.dll ]
ext-ms-win-lighting-lamparray-l1-1-1 -> [ lamparray.dll ]
ext-ms-win-mapi-mapi32-l1-1-0 -> [ mapistub.dll ]
ext-ms-win-media-avi-l1-1-0 -> [ avifil32.dll ]
ext-ms-win-media-codecpack-mounting-l1-1-0 -> [  ]
ext-ms-win-mf-devicespecific-l1-1-0 -> [  ]
ext-ms-win-mf-vfw-l1-1-0 -> [ mfvfw.dll ]
ext-ms-win-mininput-cursorhost-l1-1-0 -> [ inputhost.dll ]
ext-ms-win-mininput-extensions-l1-1-0 -> [  ]
ext-ms-win-mininput-inputhost-l1-1-1 -> [ inputhost.dll ]
ext-ms-win-mininput-inputhost-l1-2-1 -> [ inputhost.dll ]
ext-ms-win-mininput-inputhost-l1-3-0 -> [ inputhost.dll ]
ext-ms-win-mininput-inputhost-l1-4-0 -> [ inputhost.dll ]
ext-ms-win-mininput-systeminputhost-l1-1-0 -> [ ism.dll ]
ext-ms-win-mininput-systeminputhost-l1-2-0 -> [ ism.dll ]
ext-ms-win-mm-io-l1-1-0 -> [ winmmbase.dll ]
ext-ms-win-mm-msacm-l1-1-0 -> [ msacm32.dll ]
ext-ms-win-mm-pehelper-l1-1-0 -> [ mf.dll ]
ext-ms-win-mm-wmvcore-l1-1-0 -> [ wmvcore.dll ]
ext-ms-win-moderncore-win32k-base-ntgdi-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-moderncore-win32k-base-ntuser-l1-1-0 -> [ win32kfull.sys ]
ext-ms-win-moderncore-win32k-base-sysentry-l1-1-0 -> [ win32k.sys ]
ext-ms-win-mpr-multipleproviderrouter-l1-1-0 -> [ mprext.dll ]
ext-ms-win-mrmcorer-resmanager-l1-1-0 -> [ mrmcorer.dll ]
ext-ms-win-msa-device-l1-1-0 -> [  ]
ext-ms-win-msa-ui-l1-1-0 -> [ msauserext.dll ]
ext-ms-win-msa-user-l1-1-1 -> [ msauserext.dll ]
ext-ms-win-msi-misc-l1-1-0 -> [ msi.dll ]
ext-ms-win-msiltcfg-msi-l1-1-0 -> [ msiltcfg.dll ]
ext-ms-win-msimg-draw-l1-1-0 -> [ msimg32.dll ]
ext-ms-win-net-cmvpn-l1-1-0 -> [ cmintegrator.dll ]
ext-ms-win-net-httpproxyext-l1-1-0 -> [ httpprxc.dll ]
ext-ms-win-net-isoext-l1-1-0 -> [ firewallapi.dll ]
ext-ms-win-net-netbios-l1-1-0 -> [ netbios.dll ]
ext-ms-win-net-netshell-l1-1-0 -> [ netshell.dll ]
ext-ms-win-net-nfdapi-l1-1-1 -> [ ndfapi.dll ]
ext-ms-win-net-vpn-l1-1-0 -> [  ]
ext-ms-win-netio-l1-1-0 -> [ netio.sys ]
ext-ms-win-netprovision-netprovfw-l1-1-0 -> [ netprovfw.dll ]
ext-ms-win-networking-iphlpsvc-l1-1-0 -> [  ]
ext-ms-win-networking-mpssvc-l1-1-0 -> [  ]
ext-ms-win-networking-ncsiuserprobe-l1-1-0 -> [  ]
ext-ms-win-networking-radiomonitor-l1-1-0 -> [ windows.devices.radios.dll ]
ext-ms-win-networking-teredo-l1-1-0 -> [ windows.networking.connectivity.dll ]
ext-ms-win-networking-wcmapi-l1-1-1 -> [ wcmapi.dll ]
ext-ms-win-networking-winipsec-l1-1-0 -> [ winipsec.dll ]
ext-ms-win-networking-wlanapi-l1-1-0 -> [ wlanapi.dll ]
ext-ms-win-networking-wlanstorage-l1-1-0 -> [  ]
ext-ms-win-networking-xblconnectivity-l1-1-0 -> [  ]
ext-ms-win-newdev-config-l1-1-2 -> [ newdev.dll ]
ext-ms-win-nfc-semgr-l1-1-0 -> [ semgrsvc.dll ]
ext-ms-win-ntdsa-activedirectoryasyncthreadqueue-l1-1-0 -> [  ]
ext-ms-win-ntdsa-activedirectoryserver-l1-1-2 -> [  ]
ext-ms-win-ntdsapi-activedirectoryclient-l1-1-1 -> [ ntdsapi.dll ]
ext-ms-win-ntos-clipsp-l1-1-0 -> [ clipsp.sys ]
ext-ms-win-ntos-dg-l1-1-0 -> [  ]
ext-ms-win-ntos-globmerger-l1-1-0 -> [ globmerger.sys ]
ext-ms-win-ntos-kcminitcfg-l1-1-0 -> [ cmimcext.sys ]
ext-ms-win-ntos-ksecurity-l1-1-1 -> [  ]
ext-ms-win-ntos-ksr-l1-1-5 -> [  ]
ext-ms-win-ntos-processparameters-l1-1-0 -> [  ]
ext-ms-win-ntos-runlevels-l1-1-0 -> [  ]
ext-ms-win-ntos-stateseparation-l1-1-0 -> [  ]
ext-ms-win-ntos-tm-l1-1-0 -> [ tm.sys ]
ext-ms-win-ntos-trace-l1-1-0 -> [  ]
ext-ms-win-ntos-ucode-l1-1-0 -> [ ntosext.sys ]
ext-ms-win-ntos-vail-l1-1-0 -> [  ]
ext-ms-win-ntos-vmsvc-l1-1-0 -> [ vmsvcext.sys ]
ext-ms-win-ntos-werkernel-l1-1-1 -> [ werkernel.sys ]
ext-ms-win-ntos-win32k-l1-1-0 -> [ win32k.sys ]
ext-ms-win-ntuser-caret-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-chartranslation-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-dc-access-ext-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-dde-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-dialogbox-l1-1-3 -> [ user32.dll ]
ext-ms-win-ntuser-draw-l1-1-2 -> [ user32.dll ]
ext-ms-win-ntuser-gui-l1-1-1 -> [ user32.dll ]
ext-ms-win-ntuser-gui-l1-2-0 -> [ user32.dll ]
ext-ms-win-ntuser-gui-l1-3-1 -> [ user32.dll ]
ext-ms-win-ntuser-keyboard-ansi-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-keyboard-l1-1-1 -> [ user32.dll ]
ext-ms-win-ntuser-keyboard-l1-2-0 -> [ user32.dll ]
ext-ms-win-ntuser-keyboard-l1-3-2 -> [ user32.dll ]
ext-ms-win-ntuser-menu-l1-1-3 -> [ user32.dll ]
ext-ms-win-ntuser-message-l1-1-3 -> [ user32.dll ]
ext-ms-win-ntuser-misc-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-misc-l1-2-0 -> [ user32.dll ]
ext-ms-win-ntuser-misc-l1-3-0 -> [ user32.dll ]
ext-ms-win-ntuser-misc-l1-5-1 -> [ user32.dll ]
ext-ms-win-ntuser-misc-l1-6-0 -> [ user32.dll ]
ext-ms-win-ntuser-misc-l1-7-0 -> [ user32.dll ]
ext-ms-win-ntuser-mit-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-mouse-l1-1-1 -> [ user32.dll ]
ext-ms-win-ntuser-powermanagement-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-private-l1-1-1 -> [ user32.dll ]
ext-ms-win-ntuser-private-l1-2-0 -> [ user32.dll ]
ext-ms-win-ntuser-private-l1-3-3 -> [ user32.dll ]
ext-ms-win-ntuser-private-l1-4-0 -> [ user32.dll ]
ext-ms-win-ntuser-private-l1-5-0 -> [ user32.dll ]
ext-ms-win-ntuser-private-l1-6-3 -> [ user32.dll ]
ext-ms-win-ntuser-rawinput-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-rawinput-l1-2-0 -> [ user32.dll ]
ext-ms-win-ntuser-rectangle-ext-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-rim-l1-1-2 -> [ user32.dll ]
ext-ms-win-ntuser-rim-l1-2-1 -> [ user32.dll ]
ext-ms-win-ntuser-rotationmanager-l1-1-2 -> [ user32.dll ]
ext-ms-win-ntuser-server-l1-1-1 -> [ user32.dll ]
ext-ms-win-ntuser-string-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-synch-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-sysparams-ext-l1-1-1 -> [ user32.dll ]
ext-ms-win-ntuser-touch-hittest-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-uicontext-ext-l1-1-0 -> [ user32.dll ]
ext-ms-win-ntuser-window-l1-1-6 -> [ user32.dll ]
ext-ms-win-ntuser-windowclass-l1-1-2 -> [ user32.dll ]
ext-ms-win-ntuser-windowstation-ansi-l1-1-1 -> [ user32.dll ]
ext-ms-win-ntuser-windowstation-l1-1-2 -> [ user32.dll ]
ext-ms-win-odbc-odbc32-l1-1-0 -> [ odbc32.dll ]
ext-ms-win-ole32-bindctx-l1-1-0 -> [ ole32.dll ]
ext-ms-win-ole32-ie-ext-l1-1-0 -> [ ole32.dll ]
ext-ms-win-ole32-oleautomation-l1-1-0 -> [ ole32.dll ]
ext-ms-win-oleacc-l1-1-2 -> [ oleacc.dll ]
ext-ms-win-onecore-shutdown-l1-1-0 -> [ twinapi.appcore.dll ]
ext-ms-win-oobe-query-l1-1-0 -> [  ]
ext-ms-win-packagevirtualizationcontext-l1-1-0 -> [ daxexec.dll ]
ext-ms-win-parentalcontrols-setup-l1-1-0 -> [ wpcapi.dll ]
ext-ms-win-perception-device-l1-1-1 -> [  ]
ext-ms-win-pinenrollment-enrollment-l1-1-2 -> [ pinenrollmenthelper.dll ]
ext-ms-win-printer-prntvpt-l1-1-2 -> [ prntvpt.dll ]
ext-ms-win-printer-winspool-core-l1-1-0 -> [ winspool.drv ]
ext-ms-win-printer-winspool-l1-1-4 -> [ winspool.drv ]
ext-ms-win-printer-winspool-l1-2-0 -> [ winspool.drv ]
ext-ms-win-profile-extender-l1-1-0 -> [ userenv.dll ]
ext-ms-win-profile-load-l1-1-0 -> [  ]
ext-ms-win-profile-profsvc-l1-1-0 -> [ profsvcext.dll ]
ext-ms-win-profile-userenv-l1-1-1 -> [ profext.dll ]
ext-ms-win-provisioning-options-l1-1-0 -> [  ]
ext-ms-win-provisioning-platform-l1-1-2 -> [ provplatformdesktop.dll ]
ext-ms-win-ras-rasapi32-l1-1-2 -> [ rasapi32.dll ]
ext-ms-win-ras-rasdlg-l1-1-0 -> [ rasdlg.dll ]
ext-ms-win-ras-rasman-l1-1-0 -> [ rasman.dll ]
ext-ms-win-ras-tapi32-l1-1-1 -> [ tapi32.dll ]
ext-ms-win-raschapext-eap-l1-1-0 -> [ raschapext.dll ]
ext-ms-win-rastlsext-eap-l1-1-0 -> [ rastlsext.dll ]
ext-ms-win-rdr-davhlpr-l1-1-0 -> [ davhlpr.dll ]
ext-ms-win-reinfo-query-l1-1-0 -> [ reinfo.dll ]
ext-ms-win-remotewipe-platform-l1-1-0 -> [  ]
ext-ms-win-resourcemanager-activitycoordinator-l1-1-1 -> [ rmclient.dll ]
ext-ms-win-resourcemanager-crm-l1-1-0 -> [ rmclient.dll ]
ext-ms-win-resourcemanager-crm-l1-2-0 -> [ rmclient.dll ]
ext-ms-win-resourcemanager-crm-private-ext-l1-1-0 -> [ psmserviceexthost.dll ]
ext-ms-win-resourcemanager-gamemode-l1-1-0 -> [ rmclient.dll ]
ext-ms-win-resourcemanager-gamemode-l1-2-1 -> [ rmclient.dll ]
ext-ms-win-resourcemanager-limits-l1-1-0 -> [ rmclient.dll ]
ext-ms-win-resources-deployment-l1-1-0 -> [ mrmdeploy.dll ]
ext-ms-win-resources-languageoverlay-l1-1-7 -> [ languageoverlayutil.dll ]
ext-ms-win-ro-typeresolution-l1-1-1 -> [ wintypes.dll ]
ext-ms-win-rometadata-dispenser-l1-1-0 -> [ rometadata.dll ]
ext-ms-win-rpc-firewallportuse-l1-1-0 -> [ rpcrtremote.dll ]
ext-ms-win-rpc-ssl-l1-1-0 -> [ rpcrtremote.dll ]
ext-ms-win-rtcore-gdi-devcaps-l1-1-1 -> [ gdi32.dll ]
ext-ms-win-rtcore-gdi-object-l1-1-0 -> [ gdi32.dll ]
ext-ms-win-rtcore-gdi-rgn-l1-1-1 -> [ gdi32.dll ]
ext-ms-win-rtcore-minuser-display-l1-1-0 -> [  ]
ext-ms-win-rtcore-minuser-host-l1-1-0 -> [  ]
ext-ms-win-rtcore-minuser-input-l1-1-4 -> [  ]
ext-ms-win-rtcore-minuser-internal-l1-1-0 -> [  ]
ext-ms-win-rtcore-minuser-private-ext-l1-1-3 -> [  ]
ext-ms-win-rtcore-ntuser-controllernavigation-l1-1-2 -> [ inputhost.dll ]
ext-ms-win-rtcore-ntuser-cursor-l1-1-1 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-dc-access-l1-1-1 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-dialogbox-l1-1-0 -> [  ]
ext-ms-win-rtcore-ntuser-dpi-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-dpi-l1-2-2 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-draw-l1-1-0 -> [  ]
ext-ms-win-rtcore-ntuser-gui-l1-1-1 -> [  ]
ext-ms-win-rtcore-ntuser-iam-l1-1-2 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-inputintercept-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-integration-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-keyboard-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-message-ansi-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-message-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-mininit-l1-1-0 -> [  ]
ext-ms-win-rtcore-ntuser-misc-l1-1-0 -> [  ]
ext-ms-win-rtcore-ntuser-mouse-l1-1-0 -> [  ]
ext-ms-win-rtcore-ntuser-powermanagement-l1-1-0 -> [  ]
ext-ms-win-rtcore-ntuser-private-l1-1-1 -> [  ]
ext-ms-win-rtcore-ntuser-rawinput-l1-1-1 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-rawinput-l1-2-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-synch-ext-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-syscolors-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-sysparams-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-usersecurity-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-window-ansi-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-window-ext-l1-1-1 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-window-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-winevent-ext-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-winstamin-l1-1-0 -> [  ]
ext-ms-win-rtcore-ntuser-wmpointer-l1-1-0 -> [ user32.dll ]
ext-ms-win-rtcore-ntuser-wmpointermin-l1-1-0 -> [  ]
ext-ms-win-rtcore-ole32-dragdrop-l1-1-0 -> [ ole32.dll ]
ext-ms-win-rtcore-ole32-misc-l1-1-0 -> [ ole32.dll ]
ext-ms-win-rtcore-webview-l1-1-0 -> [  ]
ext-ms-win-samsrv-accountstore-l1-1-1 -> [ samsrv.dll ]
ext-ms-win-scesrv-server-l1-1-0 -> [ scesrv.dll ]
ext-ms-win-search-folder-l1-1-1 -> [ searchfolder.dll ]
ext-ms-win-search-lifetimemanager-l1-1-0 -> [ mssrch.dll ]
ext-ms-win-secur32-translatename-l1-1-0 -> [ secur32.dll ]
ext-ms-win-security-appinfoext-l1-1-0 -> [ appinfoext.dll ]
ext-ms-win-security-authbrokerui-l1-1-0 -> [ authbrokerui.dll ]
ext-ms-win-security-authz-helper-l1-1-0 -> [  ]
ext-ms-win-security-capauthz-l1-1-1 -> [ capauthz.dll ]
ext-ms-win-security-catalog-database-l1-1-0 -> [ cryptcatsvc.dll ]
ext-ms-win-security-certpoleng-l1-1-0 -> [ certpoleng.dll ]
ext-ms-win-security-cfl-l1-1-2 -> [ cflapi.dll ]
ext-ms-win-security-chambers-l1-1-1 -> [  ]
ext-ms-win-security-credui-internal-l1-1-1 -> [ wincredui.dll ]
ext-ms-win-security-credui-l1-1-1 -> [ credui.dll ]
ext-ms-win-security-crosscontainerauthhelper-l1-1-0 -> [  ]
ext-ms-win-security-cryptui-l1-1-1 -> [ cryptui.dll ]
ext-ms-win-security-developerunlock-l1-1-0 -> [  ]
ext-ms-win-security-deviceid-l1-1-0 -> [  ]
ext-ms-win-security-efs-l1-1-1 -> [ efsext.dll ]
ext-ms-win-security-efswrt-l1-1-4 -> [ efswrt.dll ]
ext-ms-win-security-kerberos-l1-1-0 -> [ kerberos.dll ]
ext-ms-win-security-lsaadt-l1-1-0 -> [ lsaadt.dll ]
ext-ms-win-security-lsaadtpriv-l1-1-0 -> [ lsaadt.dll ]
ext-ms-win-security-lsaauditrpc-l1-1-0 -> [ lsaadt.dll ]
ext-ms-win-security-ngc-local-l1-1-0 -> [ ngclocal.dll ]
ext-ms-win-security-shutdownext-l1-1-0 -> [ shutdownext.dll ]
ext-ms-win-security-slc-l1-1-0 -> [ slc.dll ]
ext-ms-win-security-srp-l1-1-1 -> [ srpapi.dll ]
ext-ms-win-security-tokenbrokerui-l1-1-0 -> [ tokenbrokerui.dll ]
ext-ms-win-security-vaultcds-l1-1-0 -> [ vaultcds.dll ]
ext-ms-win-security-vaultcds-l1-2-0 -> [ vaultcds.dll ]
ext-ms-win-security-vaultcli-l1-1-1 -> [ vaultcli.dll ]
ext-ms-win-security-winscard-l1-1-1 -> [ winscard.dll ]
ext-ms-win-sensors-core-private-l1-1-8 -> [ sensorsnativeapi.dll ]
ext-ms-win-sensors-utilities-private-l1-1-5 -> [ sensorsutilsv2.dll ]
ext-ms-win-servicing-uapi-l1-1-2 -> [ servicinguapi.dll ]
ext-ms-win-session-candidateaccountmgr-l1-1-0 -> [ camext.dll ]
ext-ms-win-session-userinit-l1-1-0 -> [ userinitext.dll ]
ext-ms-win-session-usermgr-l1-1-0 -> [ usermgrcli.dll ]
ext-ms-win-session-usermgr-l1-2-1 -> [ usermgrcli.dll ]
ext-ms-win-session-usertoken-l1-1-0 -> [ wtsapi32.dll ]
ext-ms-win-session-wininit-l1-1-1 -> [ wininitext.dll ]
ext-ms-win-session-wininit-l1-2-0 -> [ wininitext.dll ]
ext-ms-win-session-winlogon-l1-1-2 -> [ winlogonext.dll ]
ext-ms-win-session-winlogon-notify-l1-1-0 -> [  ]
ext-ms-win-session-winsta-l1-1-6 -> [ winsta.dll ]
ext-ms-win-session-wtsapi32-l1-1-2 -> [ wtsapi32.dll ]
ext-ms-win-setupapi-classinstallers-l1-1-2 -> [ setupapi.dll ]
ext-ms-win-setupapi-devobj-l1-1-0 -> [  ]
ext-ms-win-setupapi-inf-l1-1-1 -> [ setupapi.dll ]
ext-ms-win-setupapi-logging-l1-1-0 -> [ setupapi.dll ]
ext-ms-win-shell-aclui-l1-1-0 -> [ aclui.dll ]
ext-ms-win-shell-browsersettingsync-l1-1-0 -> [  ]
ext-ms-win-shell-comctl32-da-l1-1-0 -> [ comctl32.dll ]
ext-ms-win-shell-comctl32-init-l1-1-1 -> [ comctl32.dll ]
ext-ms-win-shell-comctl32-l1-1-0 -> [ comctl32.dll ]
ext-ms-win-shell-comctl32-window-l1-1-0 -> [ comctl32.dll ]
ext-ms-win-shell-comdlg32-l1-1-1 -> [ comdlg32.dll ]
ext-ms-win-shell-directory-l1-1-0 -> [ windows.storage.dll ]
ext-ms-win-shell-efsadu-l1-1-0 -> [ efsadu.dll ]
ext-ms-win-shell-embeddedmode-l1-1-0 -> [ embeddedmodesvcapi.dll ]
ext-ms-win-shell-exports-internal-l1-1-1 -> [ shell32.dll ]
ext-ms-win-shell-fileplaceholder-l1-1-0 -> [ windows.fileexplorer.common.dll ]
ext-ms-win-shell-ntshrui-l1-1-0 -> [ ntshrui.dll ]
ext-ms-win-shell-propsys-l1-1-1 -> [ propsys.dll ]
ext-ms-win-shell-shdocvw-l1-1-0 -> [ shdocvw.dll ]
ext-ms-win-shell-shell32-l1-2-3 -> [ shell32.dll ]
ext-ms-win-shell-shell32-l1-3-0 -> [ shell32.dll ]
ext-ms-win-shell-shell32-l1-4-0 -> [ shell32.dll ]
ext-ms-win-shell-shell32-l1-5-0 -> [ shell32.dll ]
ext-ms-win-shell-shlwapi-l1-1-2 -> [ shlwapi.dll ]
ext-ms-win-shell-shlwapi-l1-2-1 -> [ shlwapi.dll ]
ext-ms-win-shell32-shellcom-l1-1-0 -> [ windows.storage.dll ]
ext-ms-win-shell32-shellfolders-l1-1-1 -> [ windows.storage.dll ]
ext-ms-win-shell32-shellfolders-l1-2-1 -> [ windows.storage.dll ]
ext-ms-win-smbshare-browser-l1-1-0 -> [ browser.dll ]
ext-ms-win-smbshare-browserclient-l1-1-0 -> [ browcli.dll ]
ext-ms-win-smbshare-sscore-l1-1-0 -> [ sscoreext.dll ]
ext-ms-win-spinf-inf-l1-1-0 -> [ spinf.dll ]
ext-ms-win-storage-hbaapi-l1-1-1 -> [ hbaapi.dll ]
ext-ms-win-storage-iscsidsc-l1-1-0 -> [ iscsidsc.dll ]
ext-ms-win-storage-sense-l1-1-0 -> [ storageusage.dll ]
ext-ms-win-storage-sense-l1-2-5 -> [ storageusage.dll ]
ext-ms-win-sxs-oleautomation-l1-1-0 -> [ sxs.dll ]
ext-ms-win-sysmain-pfapi-l1-1-1 -> [ pfclient.dll ]
ext-ms-win-sysmain-pfsapi-l1-1-0 -> [ pfclient.dll ]
ext-ms-win-sysmain-plmapi-l1-1-1 -> [ pfclient.dll ]
ext-ms-win-sysmain-sfltapi-l1-1-1 -> [  ]
ext-ms-win-system-metrics-override-l1-1-1 -> [  ]
ext-ms-win-teapext-eap-l1-1-0 -> [ eapteapext.dll ]
ext-ms-win-test-sys1-l1-1-0 -> [  ]
ext-ms-win-test-sys2-l1-1-0 -> [  ]
ext-ms-win-tsf-inputsetting-l1-1-0 -> [ input.dll ]
ext-ms-win-tsf-msctf-l1-1-4 -> [ msctf.dll ]
ext-ms-win-ttlsext-eap-l1-1-0 -> [ ttlsext.dll ]
ext-ms-win-ui-viewmanagement-l1-1-0 -> [  ]
ext-ms-win-uiacore-l1-1-3 -> [ uiautomationcore.dll ]
ext-ms-win-umpoext-umpo-l1-1-0 -> [ umpoext.dll ]
ext-ms-win-usp10-l1-1-0 -> [ gdi32full.dll ]
ext-ms-win-uwf-servicing-apis-l1-1-1 -> [ uwfservicingapi.dll ]
ext-ms-win-uxtheme-themes-l1-1-3 -> [ uxtheme.dll ]
ext-ms-win-vmbus-hvsocket-l1-1-0 -> [ hvsocket.sys ]
ext-ms-win-webtokenrequest-win32-l1-1-0 -> [  ]
ext-ms-win-wer-reporting-l1-1-3 -> [ wer.dll ]
ext-ms-win-wer-ui-l1-1-1 -> [ werui.dll ]
ext-ms-win-wer-wct-l1-1-0 -> [ wer.dll ]
ext-ms-win-wer-xbox-l1-1-3 -> [  ]
ext-ms-win-wer-xbox-l1-2-1 -> [  ]
ext-ms-win-wevtapi-eventlog-l1-1-3 -> [ wevtapi.dll ]
ext-ms-win-windowing-external-l1-1-0 -> [ windows.ui.dll ]
ext-ms-win-windowing-internal-l1-1-0 -> [  ]
ext-ms-win-winlogon-mincreds-l1-1-0 -> [  ]
ext-ms-win-winrt-device-access-l1-1-0 -> [ deviceaccess.dll ]
ext-ms-win-winrt-storage-l1-1-0 -> [ windows.storage.dll ]
ext-ms-win-winrt-storage-l1-2-3 -> [ windows.storage.dll ]
ext-ms-win-winrt-storage-win32broker-l1-1-0 -> [ windows.storage.onecore.dll ]
ext-ms-win-wlan-grouppolicy-l1-1-0 -> [ wlgpclnt.dll ]
ext-ms-win-wlan-onexui-l1-1-0 -> [ onexui.dll ]
ext-ms-win-wlan-scard-l1-1-0 -> [ winscard.dll ]
ext-ms-win-wpc-webfilter-l1-1-0 -> [ wpcwebfilter.dll ]
ext-ms-win-wpn-phoneext-l1-1-0 -> [  ]
ext-ms-win-wrp-sfc-l1-1-0 -> [ sfc.dll ]
ext-ms-win-wsclient-devlicense-l1-1-1 -> [ wsclient.dll ]
ext-ms-win-wwaext-misc-l1-1-0 -> [ wwaext.dll ]
ext-ms-win-wwaext-module-l1-1-0 -> [ wwaext.dll ]
ext-ms-win-wwan-wwapi-l1-1-3 -> [ wwapi.dll ]
ext-ms-win-xaml-controls-l1-1-0 -> [ windows.ui.xaml.phone.dll ]
ext-ms-win-xaml-pal-l1-1-0 -> [  ]
ext-ms-win-xaudio-platform-l1-1-0 -> [  ]
ext-ms-win-xblauth-console-l1-1-0 -> [  ]
ext-ms-win-xboxlive-xboxnetapisvc-l1-1-0 -> [  ]
ext-ms-win32-subsystem-query-l1-1-0 -> [  ]
ext-ms-windowscore-deviceinfo-l1-1-0 -> [  ]

I've included the dependencies chain in a log file.

log.txt

<!-- gh-comment-id:3488823712 --> @catclaw commented on GitHub (Nov 5, 2025): > [@catclaw](https://github.com/catclaw) with the updated logging code we have a little more detail, but unfortunately still not an exact root cause on what's causing this problem on your system. None of the dynamic libraries will load properly > > ``` > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > time=2025-11-03T22:32:16.738+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" > dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found. > System.Management.Automation.RemoteException > ``` > > For a missing dependency, I would expect the error to say "The specified module could not be found" so the fact you're seeing "procedure" seems to imply the way we're linking our libraries is causing an incompatibility with something on your system. > > Do you have any other software on your system that's doing preload/injection? Any unusual AV software? > > Maybe we'll be able to get a little more insight if you can scan the libraries on your system? You might try https://github.com/lucasg/Dependencies/ which has a nice GUI that can show dependencies, and in this case, maybe it will help us understand what is causing all the DLLs listed above to not load? This is what I found out: ``` [NOT_FOUND] ext-ms-win-security-chambers-l1-1-0.dll : [NOT_FOUND] ext-ms-win-oobe-query-l1-1-0.dll : [NOT_FOUND] ext-ms-win-core-app-package-registration-l1-1-0.dll : [NOT_FOUND] ext-ms-win32-subsystem-query-l1-1-0.dll : [NOT_FOUND] ext-ms-win-com-suspendresiliency-l1-1-0.dll : [NOT_FOUND] ext-ms-win-appmodel-deployment-l1-1-0.dll : [NOT_FOUND] ext-ms-win-appmodel-usercontext-l1-1-0.dll : [NOT_FOUND] ext-ms-win-gaming-devicefamily-l1-1-0.dll : [NOT_FOUND] ext-ms-win-ui-viewmanagement-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-2.dll : [NOT_FOUND] ext-ms-win-gdi-edgegdi-l1-1-0.dll : [NOT_FOUND] ext-ms-win-appmodel-viewscalefactor-l1-1-0.dll : [NOT_FOUND] ext-ms-onecore-defaultdiscovery-l1-1-0.dll : [NOT_FOUND] wpaxholder.dll : [NOT_FOUND] ext-ms-windowscore-deviceinfo-l1-1-0.dll : [NOT_FOUND] ext-ms-mf-pal-l2-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-input-l1-1-1.dll : [NOT_FOUND] api-ms-win-coreui-secruntime-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-ntuser-private-l1-1-1.dll : [NOT_FOUND] ext-ms-win-media-codecpack-mounting-l1-1-0.dll : [NOT_FOUND] ext-ms-win-core-winrt-remote-l1-1-0.dll : [NOT_FOUND] ext-ms-win-com-apartmentrestriction-l1-1-0.dll : [NOT_FOUND] PdmUtilities.dll : [NOT_FOUND] HvsiFileTrust.dll : [NOT_FOUND] ext-ms-win-audiocore-pal-l1-2-0.dll : [NOT_FOUND] ext-ms-win-audiocore-spatial-l1-1-0.dll : [NOT_FOUND] UpdateAPI.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-1.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-0.dll : [NOT_FOUND] AzureAttestManager.dll : [NOT_FOUND] AzureAttestNormal.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-1-2.dll : [NOT_FOUND] ext-ms-win-networking-wlanstorage-l1-1-0.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-1-3.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-1-0.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-2-0.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-2-1.dll : [NOT_FOUND] ext-ms-onecore-orientation-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-internal-l1-1-0.dll : [NOT_FOUND] ext-ms-onecore-shellchromeapi-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-3.dll : [NOT_FOUND] ext-ms-win-core-xbrm-l1-1-0.dll : [NOT_FOUND] ext-ms-win-core-xbrm-l1-1-1.dll : [NOT_FOUND] ext-ms-win-ntdsa-activedirectoryserver-l1-1-0.dll : ``` After reinstalling the correct Visual C++ Redistributable package (once again - I had to uninstall the previous version), and running sfc /scannow, dism.exe /online /cleanup-image /scanhealth, dism.exe /online /cleanup-image /restorehealth, and dism.exe /online /cleanup-image /startcomponentcleanup - the DLL-files were still missing. If you look at the end of apisets, some APIs call non-existing DLLs. ([ ]). ``` [-] Api Sets Map : api-ms-onecoreuap-print-render-l1-1-0 -> [ printrenderapihost.dll ] api-ms-win-appmodel-advertisingid-l1-1-0 -> [ kernel.appcore.dll ] api-ms-win-appmodel-identity-l1-2-0 -> [ kernel.appcore.dll ] api-ms-win-appmodel-lifecyclepolicy-l1-1-0 -> [ rmclient.dll ] api-ms-win-appmodel-runtime-internal-l1-1-11 -> [ kernel.appcore.dll ] api-ms-win-appmodel-runtime-l1-1-7 -> [ kernel.appcore.dll ] api-ms-win-appmodel-state-l1-1-2 -> [ kernel.appcore.dll ] api-ms-win-appmodel-state-l1-2-0 -> [ kernel.appcore.dll ] api-ms-win-appmodel-unlock-l1-1-0 -> [ kernel.appcore.dll ] api-ms-win-audiocore-spatial-config-l1-1-0 -> [ windows.media.devices.dll ] api-ms-win-base-bootconfig-l1-1-0 -> [ advapi32.dll ] api-ms-win-base-util-l1-1-0 -> [ advapi32.dll ] api-ms-win-composition-redirection-l1-1-0 -> [ dwmredir.dll ] api-ms-win-composition-windowmanager-l1-1-0 -> [ udwm.dll ] api-ms-win-containers-cmclient-l1-1-1 -> [ cmclient.dll ] api-ms-win-containers-cmclient-l1-2-0 -> [ cmclient.dll ] api-ms-win-containers-cmclient-l1-3-0 -> [ cmclient.dll ] api-ms-win-containers-cmclient-l1-4-0 -> [ cmclient.dll ] api-ms-win-containers-cmclient-l1-5-3 -> [ cmclient.dll ] api-ms-win-containers-cmdiagclient-l1-1-2 -> [ cmclient.dll ] api-ms-win-containers-cmservicingclient-l1-1-1 -> [ cmclient.dll ] api-ms-win-containers-cmservicingclient-l1-2-2 -> [ cmclient.dll ] api-ms-win-core-apiquery-l1-1-2 -> [ ntdll.dll ] api-ms-win-core-apiquery-l2-1-1 -> [ kernelbase.dll ] api-ms-win-core-appcompat-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-appinit-l1-1-0 -> [ kernel32.dll,kernelbase.dll ] api-ms-win-core-atoms-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-backgroundtask-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-bicltapi-l1-1-6 -> [ bi.dll ] api-ms-win-core-biplmapi-l1-1-5 -> [ twinapi.appcore.dll ] api-ms-win-core-biplmapi-l1-2-0 -> [ twinapi.appcore.dll ] api-ms-win-core-biptcltapi-l1-1-7 -> [ twinapi.appcore.dll ] api-ms-win-core-calendar-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-com-l1-1-3 -> [ combase.dll ] api-ms-win-core-com-l2-1-1 -> [ coml2.dll ] api-ms-win-core-com-midlproxystub-l1-1-0 -> [ combase.dll ] api-ms-win-core-com-private-l1-1-1 -> [ combase.dll ] api-ms-win-core-com-private-l1-2-0 -> [ combase.dll ] api-ms-win-core-com-private-l1-3-1 -> [ combase.dll ] api-ms-win-core-comm-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-commandlinetoargv-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-console-ansi-l2-1-0 -> [ kernel32.dll ] api-ms-win-core-console-internal-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-console-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-console-l1-2-2 -> [ kernelbase.dll ] api-ms-win-core-console-l2-1-0 -> [ kernelbase.dll ] api-ms-win-core-console-l2-2-0 -> [ kernelbase.dll ] api-ms-win-core-console-l3-1-0 -> [ kernelbase.dll ] api-ms-win-core-console-l3-2-0 -> [ kernelbase.dll ] api-ms-win-core-crt-l1-1-0 -> [ ntdll.dll ] api-ms-win-core-crt-l2-1-0 -> [ kernelbase.dll ] api-ms-win-core-datetime-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-debug-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-debug-minidump-l1-1-0 -> [ dbgcore.dll ] api-ms-win-core-delayload-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-enclave-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-errorhandling-l1-1-3 -> [ kernelbase.dll ] api-ms-win-core-featurestaging-l1-1-1 -> [ shcore.dll ] api-ms-win-core-featuretoggles-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-fibers-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-fibers-l2-1-1 -> [ kernelbase.dll ] api-ms-win-core-file-ansi-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-file-ansi-l2-1-0 -> [ kernel32.dll ] api-ms-win-core-file-fromapp-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-file-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-file-l1-2-5 -> [ kernelbase.dll ] api-ms-win-core-file-l2-1-4 -> [ kernelbase.dll ] api-ms-win-core-firmware-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-guard-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-handle-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-heap-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-heap-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-heap-l2-1-0 -> [ kernelbase.dll ] api-ms-win-core-heap-obsolete-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-interlocked-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-interlocked-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-io-l1-1-1 -> [ kernel32.dll,kernelbase.dll ] api-ms-win-core-ioring-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-job-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-job-l2-1-1 -> [ kernel32.dll ] api-ms-win-core-kernel32-legacy-ansi-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-kernel32-legacy-l1-1-6 -> [ kernel32.dll ] api-ms-win-core-kernel32-private-l1-1-2 -> [ kernel32.dll ] api-ms-win-core-kernel32-private-l1-2-0 -> [ kernel32.dll ] api-ms-win-core-largeinteger-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-libraryloader-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-libraryloader-l1-2-3 -> [ kernelbase.dll ] api-ms-win-core-libraryloader-l2-1-0 -> [ kernelbase.dll ] api-ms-win-core-libraryloader-private-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-localization-ansi-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-localization-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-localization-l1-2-4 -> [ kernelbase.dll ] api-ms-win-core-localization-l2-1-0 -> [ kernelbase.dll ] api-ms-win-core-localization-obsolete-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-localization-obsolete-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-localization-obsolete-l1-3-0 -> [ kernelbase.dll ] api-ms-win-core-localization-private-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-localregistry-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-marshal-l1-1-0 -> [ combase.dll ] api-ms-win-core-memory-l1-1-9 -> [ kernelbase.dll ] api-ms-win-core-misc-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-multipleproviderrouter-l1-1-0 -> [ mpr.dll ] api-ms-win-core-namedpipe-ansi-l1-1-1 -> [ kernel32.dll ] api-ms-win-core-namedpipe-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-namedpipe-l1-2-2 -> [ kernelbase.dll ] api-ms-win-core-namespace-ansi-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-namespace-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-normalization-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-path-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-pcw-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-perfcounters-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-perfcounters-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-privateprofile-l1-1-1 -> [ kernel32.dll ] api-ms-win-core-processenvironment-ansi-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-processenvironment-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-processenvironment-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-processsecurity-l1-1-0 -> [ kernel32.dll,kernelbase.dll ] api-ms-win-core-processsnapshot-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-processthreads-l1-1-8 -> [ kernel32.dll,kernelbase.dll ] api-ms-win-core-processtopology-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-processtopology-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-processtopology-obsolete-l1-1-1 -> [ kernel32.dll ] api-ms-win-core-processtopology-private-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-profile-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-psapi-ansi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-psapi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-psapi-obsolete-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-psapiansi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-psm-app-l1-1-0 -> [ twinapi.appcore.dll ] api-ms-win-core-psm-appnotify-l1-1-1 -> [ twinapi.appcore.dll ] api-ms-win-core-psm-info-l1-1-1 -> [ appsruprov.dll ] api-ms-win-core-psm-key-l1-1-3 -> [ kernelbase.dll ] api-ms-win-core-psm-plm-l1-1-3 -> [ twinapi.appcore.dll ] api-ms-win-core-psm-plm-l1-2-0 -> [ twinapi.appcore.dll ] api-ms-win-core-psm-plm-l1-3-0 -> [ twinapi.appcore.dll ] api-ms-win-core-psm-rtimer-l1-1-1 -> [ twinapi.appcore.dll ] api-ms-win-core-psm-tc-l1-1-1 -> [ twinapi.appcore.dll ] api-ms-win-core-quirks-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-realtime-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-registry-fromapp-l1-1-0 -> [ reguwpapi.dll ] api-ms-win-core-registry-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-registry-l2-1-0 -> [ advapi32.dll ] api-ms-win-core-registry-l2-2-0 -> [ advapi32.dll ] api-ms-win-core-registry-l2-3-0 -> [ advapi32.dll ] api-ms-win-core-registry-private-l1-1-0 -> [ advapi32.dll ] api-ms-win-core-registryuserspecific-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-rtlsupport-l1-1-1 -> [ ntdll.dll ] api-ms-win-core-rtlsupport-l1-2-2 -> [ ntdll.dll ] api-ms-win-core-shlwapi-legacy-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-shlwapi-obsolete-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-shlwapi-obsolete-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-shutdown-ansi-l1-1-0 -> [ advapi32.dll ] api-ms-win-core-shutdown-l1-1-1 -> [ advapi32.dll ] api-ms-win-core-sidebyside-ansi-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-sidebyside-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-slapi-l1-1-0 -> [ clipc.dll ] api-ms-win-core-state-helpers-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-string-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-string-l2-1-1 -> [ kernelbase.dll ] api-ms-win-core-string-obsolete-l1-1-1 -> [ kernel32.dll ] api-ms-win-core-stringansi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-stringloader-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-synch-ansi-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-synch-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-synch-l1-2-1 -> [ kernelbase.dll ] api-ms-win-core-sysinfo-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-sysinfo-l1-2-7 -> [ kernelbase.dll ] api-ms-win-core-sysinfo-l2-1-0 -> [ advapi32.dll ] api-ms-win-core-systemtopology-l1-1-2 -> [ kernelbase.dll ] api-ms-win-core-textinput-client-l1-1-1 -> [ textinputframework.dll ] api-ms-win-core-textinput-client-l1-2-0 -> [ textinputframework.dll ] api-ms-win-core-threadpool-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-threadpool-l1-2-0 -> [ kernelbase.dll ] api-ms-win-core-threadpool-legacy-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-threadpool-private-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-timezone-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-timezone-private-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-toolhelp-l1-1-1 -> [ kernel32.dll ] api-ms-win-core-ums-l1-1-0 -> [ kernel32.dll ] api-ms-win-core-url-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-util-l1-1-1 -> [ kernel32.dll,kernelbase.dll ] api-ms-win-core-version-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-version-private-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-versionansi-l1-1-1 -> [ kernelbase.dll ] api-ms-win-core-windowsceip-l1-1-0 -> [ kernelbase.dll ] api-ms-win-core-windowserrorreporting-l1-1-3 -> [ kernelbase.dll ] api-ms-win-core-winrt-error-l1-1-1 -> [ combase.dll ] api-ms-win-core-winrt-errorprivate-l1-1-1 -> [ combase.dll ] api-ms-win-core-winrt-l1-1-0 -> [ combase.dll ] api-ms-win-core-winrt-propertysetprivate-l1-1-1 -> [ wintypes.dll ] api-ms-win-core-winrt-registration-l1-1-0 -> [ combase.dll ] api-ms-win-core-winrt-robuffer-l1-1-0 -> [ wintypes.dll ] api-ms-win-core-winrt-roparameterizediid-l1-1-0 -> [ combase.dll ] api-ms-win-core-winrt-string-l1-1-1 -> [ combase.dll ] api-ms-win-core-wow64-l1-1-3 -> [ kernelbase.dll ] api-ms-win-core-xstate-l1-1-3 -> [ ntdll.dll ] api-ms-win-core-xstate-l2-1-2 -> [ kernelbase.dll ] api-ms-win-coremessaging-host-l1-1-0 -> [ ] api-ms-win-coreui-secruntime-l1-1-0 -> [ ] api-ms-win-crt-conio-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-convert-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-environment-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-filesystem-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-heap-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-locale-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-math-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-multibyte-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-private-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-process-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-runtime-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-stdio-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-string-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-time-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-crt-utility-l1-1-0 -> [ ucrtbase.dll ] api-ms-win-deprecated-apis-advapi-l1-1-0 -> [ ] api-ms-win-deprecated-apis-legacy-l1-1-0 -> [ ] api-ms-win-deprecated-apis-legacy-l1-2-0 -> [ ] api-ms-win-deprecated-apis-obsolete-l1-1-0 -> [ kernelbase.dll ] api-ms-win-devices-config-l1-1-2 -> [ cfgmgr32.dll ] api-ms-win-devices-query-l1-1-1 -> [ cfgmgr32.dll ] api-ms-win-devices-swdevice-l1-1-1 -> [ cfgmgr32.dll ] api-ms-win-downlevel-advapi32-l1-1-0 -> [ kernelbase.dll ] api-ms-win-downlevel-advapi32-l2-1-0 -> [ sechost.dll ] api-ms-win-downlevel-advapi32-l3-1-0 -> [ ntmarta.dll ] api-ms-win-downlevel-advapi32-l4-1-0 -> [ advapi32.dll ] api-ms-win-downlevel-kernel32-l1-1-0 -> [ kernelbase.dll ] api-ms-win-downlevel-kernel32-l2-1-0 -> [ kernel32.dll ] api-ms-win-downlevel-normaliz-l1-1-0 -> [ kernelbase.dll ] api-ms-win-downlevel-ole32-l1-1-0 -> [ combase.dll ] api-ms-win-downlevel-shell32-l1-1-0 -> [ shcore.dll ] api-ms-win-downlevel-shlwapi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-downlevel-shlwapi-l2-1-0 -> [ shcore.dll ] api-ms-win-downlevel-user32-l1-1-0 -> [ kernelbase.dll ] api-ms-win-downlevel-version-l1-1-0 -> [ kernelbase.dll ] api-ms-win-dwmapi-l1-1-0 -> [ dwmapi.dll ] api-ms-win-dx-d3dkmt-l1-1-8 -> [ gdi32.dll ] api-ms-win-eventing-classicprovider-l1-1-0 -> [ kernelbase.dll ] api-ms-win-eventing-consumer-l1-1-2 -> [ sechost.dll ] api-ms-win-eventing-controller-l1-1-1 -> [ sechost.dll ] api-ms-win-eventing-legacy-l1-1-0 -> [ advapi32.dll ] api-ms-win-eventing-obsolete-l1-1-0 -> [ sechost.dll ] api-ms-win-eventing-provider-l1-1-0 -> [ kernelbase.dll ] api-ms-win-eventing-tdh-l1-1-2 -> [ tdh.dll ] api-ms-win-eventlog-legacy-l1-1-0 -> [ advapi32.dll ] api-ms-win-gaming-deviceinformation-l1-1-0 -> [ kernelbase.dll ] api-ms-win-gaming-expandedresources-l1-1-0 -> [ gamemode.dll ] api-ms-win-gaming-experience-l1-1-0 -> [ gamemode.dll ] api-ms-win-gaming-tcui-l1-1-4 -> [ gamingtcui.dll ] api-ms-win-gdi-dpiinfo-l1-1-0 -> [ gdi32.dll ] api-ms-win-gdi-internal-uap-l1-1-0 -> [ gdi32full.dll ] api-ms-win-ham-apphistory-l1-1-0 -> [ rmclient.dll ] api-ms-win-ham-hamplm-l1-1-0 -> [ rmclient.dll ] api-ms-win-http-time-l1-1-0 -> [ kernelbase.dll ] api-ms-win-input-ie-interactioncontext-l1-1-0 -> [ ] api-ms-win-legacy-shlwapi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-mm-joystick-l1-1-0 -> [ winmm.dll ] api-ms-win-mm-mci-l1-1-0 -> [ winmm.dll ] api-ms-win-mm-misc-l1-1-1 -> [ winmmbase.dll ] api-ms-win-mm-misc-l2-1-0 -> [ winmm.dll ] api-ms-win-mm-mme-l1-1-0 -> [ winmmbase.dll ] api-ms-win-mm-playsound-l1-1-0 -> [ winmm.dll ] api-ms-win-mm-time-l1-1-0 -> [ kernel32.dll ] api-ms-win-net-isolation-l1-1-1 -> [ firewallapi.dll ] api-ms-win-networking-interfacecontexts-l1-1-0 -> [ ondemandconnroutehelper.dll ] api-ms-win-ngc-serialization-l1-1-1 -> [ ngckeyenum.dll ] api-ms-win-ntuser-ie-message-l1-1-0 -> [ user32.dll ] api-ms-win-ntuser-ie-window-l1-1-0 -> [ user32.dll ] api-ms-win-ntuser-ie-wmpointer-l1-1-0 -> [ user32.dll ] api-ms-win-ntuser-rectangle-l1-1-0 -> [ user32.dll ] api-ms-win-ntuser-sysparams-l1-1-0 -> [ user32.dll ] api-ms-win-obsolete-localization-l1-1-0 -> [ kernelbase.dll ] api-ms-win-obsolete-psapi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-obsolete-shlwapi-l1-1-0 -> [ kernelbase.dll ] api-ms-win-ole32-ie-l1-1-0 -> [ ole32.dll ] api-ms-win-oobe-notification-l1-1-0 -> [ kernel32.dll ] api-ms-win-perf-legacy-l1-1-0 -> [ advapi32.dll ] api-ms-win-power-base-l1-1-0 -> [ powrprof.dll ] api-ms-win-power-limitsmanagement-l1-1-0 -> [ powrprof.dll ] api-ms-win-power-setting-l1-1-1 -> [ powrprof.dll ] api-ms-win-privacy-coreprivacysettingsstore-l1-1-0 -> [ coreprivacysettingsstore.dll ] api-ms-win-ro-typeresolution-l1-1-1 -> [ wintypes.dll ] api-ms-win-rtcore-minuser-private-l1-1-1 -> [ ] api-ms-win-rtcore-navigation-l1-1-0 -> [ ] api-ms-win-rtcore-ntuser-clipboard-l1-1-0 -> [ user32.dll ] api-ms-win-rtcore-ntuser-draw-l1-1-0 -> [ user32.dll ] api-ms-win-rtcore-ntuser-powermanagement-l1-1-0 -> [ user32.dll ] api-ms-win-rtcore-ntuser-private-l1-1-11 -> [ user32.dll ] api-ms-win-rtcore-ntuser-shell-l1-1-0 -> [ user32.dll ] api-ms-win-rtcore-ntuser-synch-l1-1-0 -> [ user32.dll ] api-ms-win-rtcore-ntuser-window-l1-1-0 -> [ user32.dll ] api-ms-win-rtcore-ntuser-winevent-l1-1-0 -> [ user32.dll ] api-ms-win-rtcore-ntuser-wmpointer-l1-1-3 -> [ user32.dll ] api-ms-win-rtcore-ntuser-wmpointer-l1-2-0 -> [ user32.dll ] api-ms-win-rtcore-ole32-clipboard-l1-1-1 -> [ ole32.dll ] api-ms-win-rtcore-session-l1-1-0 -> [ ] api-ms-win-security-accesshlpr-l1-1-0 -> [ sechost.dll ] api-ms-win-security-activedirectoryclient-l1-1-1 -> [ kernelbase.dll ] api-ms-win-security-appcontainer-l1-1-0 -> [ kernelbase.dll ] api-ms-win-security-audit-l1-1-1 -> [ sechost.dll ] api-ms-win-security-base-ansi-l1-1-0 -> [ advapi32.dll ] api-ms-win-security-base-l1-1-1 -> [ kernelbase.dll ] api-ms-win-security-base-l1-2-2 -> [ kernelbase.dll ] api-ms-win-security-base-private-l1-1-2 -> [ kernelbase.dll ] api-ms-win-security-capability-l1-1-1 -> [ sechost.dll ] api-ms-win-security-cpwl-l1-1-0 -> [ advapi32.dll ] api-ms-win-security-credentials-l1-1-0 -> [ sechost.dll ] api-ms-win-security-credentials-l2-1-1 -> [ sechost.dll ] api-ms-win-security-cryptoapi-l1-1-0 -> [ cryptsp.dll ] api-ms-win-security-grouppolicy-l1-1-0 -> [ kernelbase.dll ] api-ms-win-security-isolatedcontainer-l1-1-1 -> [ shcore.dll ] api-ms-win-security-isolationapi-l1-1-0 -> [ sechost.dll ] api-ms-win-security-isolationapi-l1-2-0 -> [ sechost.dll ] api-ms-win-security-isolationpolicy-l1-1-0 -> [ sechost.dll ] api-ms-win-security-isolationpolicy-l1-2-0 -> [ sechost.dll ] api-ms-win-security-licenseprotection-l1-1-0 -> [ licenseprotection.dll ] api-ms-win-security-logon-l1-1-1 -> [ advapi32.dll ] api-ms-win-security-lsalookup-ansi-l2-1-0 -> [ advapi32.dll ] api-ms-win-security-lsalookup-l1-1-2 -> [ sechost.dll ] api-ms-win-security-lsalookup-l2-1-1 -> [ advapi32.dll ] api-ms-win-security-lsapolicy-l1-1-2 -> [ sechost.dll ] api-ms-win-security-provider-ansi-l1-1-0 -> [ advapi32.dll ] api-ms-win-security-provider-l1-1-0 -> [ ntmarta.dll ] api-ms-win-security-sddl-ansi-l1-1-0 -> [ advapi32.dll ] api-ms-win-security-sddl-l1-1-0 -> [ sechost.dll ] api-ms-win-security-sddl-private-l1-1-0 -> [ sechost.dll ] api-ms-win-security-sddlparsecond-l1-1-1 -> [ sechost.dll ] api-ms-win-security-systemfunctions-l1-1-0 -> [ advapi32.dll ] api-ms-win-security-trustee-l1-1-2 -> [ advapi32.dll ] api-ms-win-service-core-ansi-l1-1-1 -> [ advapi32.dll ] api-ms-win-service-core-l1-1-5 -> [ sechost.dll ] api-ms-win-service-legacy-l1-1-0 -> [ advapi32.dll ] api-ms-win-service-management-l1-1-0 -> [ sechost.dll ] api-ms-win-service-management-l2-1-0 -> [ sechost.dll ] api-ms-win-service-private-l1-1-5 -> [ sechost.dll ] api-ms-win-service-private-l1-2-2 -> [ sechost.dll ] api-ms-win-service-winsvc-l1-1-0 -> [ sechost.dll ] api-ms-win-service-winsvc-l1-2-0 -> [ sechost.dll ] api-ms-win-shcore-comhelpers-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-obsolete-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-path-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-registry-l1-1-1 -> [ shcore.dll ] api-ms-win-shcore-scaling-l1-1-2 -> [ shcore.dll ] api-ms-win-shcore-stream-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-stream-winrt-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-sysinfo-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-taskpool-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-thread-l1-1-0 -> [ shcore.dll ] api-ms-win-shcore-unicodeansi-l1-1-0 -> [ shcore.dll ] api-ms-win-shell-associations-l1-1-3 -> [ windows.storage.dll ] api-ms-win-shell-changenotify-l1-1-1 -> [ windows.storage.dll ] api-ms-win-shell-dataobject-l1-1-1 -> [ windows.storage.dll ] api-ms-win-shell-namespace-l1-1-1 -> [ windows.storage.dll ] api-ms-win-shell-shdirectory-l1-1-0 -> [ shcore.dll ] api-ms-win-shell-shell32legacy-shdirectory-l1-1-0 -> [ ] api-ms-win-shell-shellcom-l1-1-0 -> [ kernelbase.dll ] api-ms-win-shell-shellfolders-l1-1-1 -> [ windows.storage.dll ] api-ms-win-shlwapi-ie-l1-1-0 -> [ shlwapi.dll ] api-ms-win-shlwapi-winrt-storage-l1-1-1 -> [ shlwapi.dll ] api-ms-win-stateseparation-helpers-l1-1-1 -> [ kernelbase.dll ] api-ms-win-storage-exports-external-l1-1-2 -> [ windows.storage.dll ] api-ms-win-storage-exports-internal-l1-1-0 -> [ windows.storage.dll ] api-ms-win-storage-reserve-l1-1-0 -> [ storageusage.dll ] api-ms-win-winrt-search-folder-l1-1-1 -> [ windows.storage.search.dll ] api-ms-win-wsl-api-l1-1-0 -> [ wslapi.dll ] ext-ms-mf-pal-l2-1-1 -> [ ] ext-ms-net-eap-sim-l1-1-0 -> [ eapsimextdesktop.dll ] ext-ms-net-vpn-soh-l1-1-0 -> [ vpnsohdesktop.dll ] ext-ms-onecore-appchromeapi-l1-1-0 -> [ ] ext-ms-onecore-appdefaults-l1-1-0 -> [ windows.storage.dll ] ext-ms-onecore-appmodel-deployment-internal-l1-1-2 -> [ appxdeploymentclient.dll ] ext-ms-onecore-appmodel-emclient-l1-1-0 -> [ ] ext-ms-onecore-appmodel-emsvcs-l1-1-0 -> [ ] ext-ms-onecore-appmodel-pacmanclient-l1-1-0 -> [ ] ext-ms-onecore-appmodel-staterepository-appextension-l1-1-0 -> [ windows.staterepositoryclient.dll ] ext-ms-onecore-appmodel-staterepository-cache-l1-1-5 -> [ windows.staterepositorycore.dll ] ext-ms-onecore-appmodel-staterepository-internal-l1-1-7 -> [ windows.staterepositoryclient.dll ] ext-ms-onecore-appmodel-staterepository-pkgextension-l1-1-0 -> [ windows.staterepositoryclient.dll ] ext-ms-onecore-appmodel-tdlmigration-l1-1-1 -> [ tdlmigration.dll ] ext-ms-onecore-comp-dwmmonitor-l1-1-0 -> [ ] ext-ms-onecore-dcomp-l1-1-0 -> [ dcomp.dll ] ext-ms-onecore-defaultdiscovery-l1-1-0 -> [ ] ext-ms-onecore-hcap-svf-l1-1-0 -> [ ] ext-ms-onecore-hlink-l1-1-0 -> [ hlink.dll ] ext-ms-onecore-hnetcfg-l1-1-0 -> [ hnetcfgclient.dll ] ext-ms-onecore-ipnathlp-l1-1-0 -> [ ipnathlpclient.dll ] ext-ms-onecore-mpc-input-l1-1-0 -> [ ] ext-ms-onecore-orientation-l1-1-0 -> [ ] ext-ms-onecore-service-devicedirectory-claims-l1-1-0 -> [ ddcclaimsapi.dll ] ext-ms-onecore-shellchromeapi-l1-1-2 -> [ ] ext-ms-onecore-shellremindersapi-l1-1-0 -> [ ] ext-ms-onecore-shlwapi-l1-1-0 -> [ shlwapi.dll ] ext-ms-onecore-spectrumsyncclient-l1-1-0 -> [ ] ext-ms-win-accel-api-km-l1-1-0 -> [ winaccel.sys ] ext-ms-win-adsi-activeds-l1-1-0 -> [ activeds.dll ] ext-ms-win-advapi32-auth-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-encryptedfile-l1-1-1 -> [ advapi32.dll ] ext-ms-win-advapi32-eventlog-ansi-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-eventlog-l1-1-2 -> [ advapi32.dll ] ext-ms-win-advapi32-hwprof-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-idletask-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-lsa-l1-1-4 -> [ advapi32.dll ] ext-ms-win-advapi32-msi-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-npusername-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-ntmarta-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-psm-app-l1-1-0 -> [ twinapi.appcore.dll ] ext-ms-win-advapi32-registry-l1-1-1 -> [ advapi32.dll ] ext-ms-win-advapi32-safer-l1-1-0 -> [ advapi32.dll ] ext-ms-win-advapi32-shutdown-l1-1-0 -> [ advapi32.dll ] ext-ms-win-appcompat-aeinv-l1-1-1 -> [ aeinv.dll ] ext-ms-win-appcompat-aepic-l1-1-0 -> [ aepic.dll ] ext-ms-win-appcompat-apphelp-l1-1-2 -> [ apphelp.dll ] ext-ms-win-appcompat-pcacli-l1-1-0 -> [ pcacli.dll ] ext-ms-win-appmodel-activation-l1-1-2 -> [ activationmanager.dll ] ext-ms-win-appmodel-appcontainerpath-l1-1-0 -> [ ] ext-ms-win-appmodel-appexecutionalias-l1-1-5 -> [ apisethost.appexecutionalias.dll ] ext-ms-win-appmodel-datasharingservice-extensions-l1-1-0 -> [ ] ext-ms-win-appmodel-daxcore-l1-1-3 -> [ daxexec.dll ] ext-ms-win-appmodel-deployment-l1-1-1 -> [ ] ext-ms-win-appmodel-deploymentvolumes-l1-1-1 -> [ ] ext-ms-win-appmodel-opc-l1-1-0 -> [ opcservices.dll ] ext-ms-win-appmodel-registrycompatibility-l1-1-0 -> [ appxdeploymentextensions.desktop.dll ] ext-ms-win-appmodel-restrictedappcontainer-internal-l1-1-0 -> [ kernel.appcore.dll ] ext-ms-win-appmodel-shellexecute-l1-1-0 -> [ windows.storage.dll ] ext-ms-win-appmodel-state-ext-l1-2-0 -> [ kernel.appcore.dll ] ext-ms-win-appmodel-usercontext-l1-1-0 -> [ ] ext-ms-win-appmodel-viewscalefactor-l1-1-0 -> [ ] ext-ms-win-appxdeploymentclient-appxdeploy-l1-1-1 -> [ appxdeploymentclient.dll ] ext-ms-win-appxdeploymentclient-appxdeployonecore-l1-1-1 -> [ appxdeploymentclient.dll ] ext-ms-win-audio-spatial-systemsound-l1-1-0 -> [ ] ext-ms-win-audiocore-coreaudiopolicymanager-l1-1-0 -> [ coreaudiopolicymanagerext.dll ] ext-ms-win-audiocore-pal-l1-2-0 -> [ ] ext-ms-win-audiocore-policymanager-l1-1-1 -> [ ] ext-ms-win-audiocore-spatial-l1-1-0 -> [ ] ext-ms-win-authz-claimpolicies-l1-1-0 -> [ authz.dll ] ext-ms-win-authz-context-l1-1-0 -> [ authz.dll ] ext-ms-win-authz-remote-l1-1-0 -> [ logoncli.dll ] ext-ms-win-base-psapi-l1-1-0 -> [ psapi.dll ] ext-ms-win-base-rstrtmgr-l1-1-0 -> [ rstrtmgr.dll ] ext-ms-win-biometrics-winbio-core-l1-1-7 -> [ winbio.dll ] ext-ms-win-biometrics-winbio-l1-1-0 -> [ winbioext.dll ] ext-ms-win-biometrics-winbio-l1-2-0 -> [ winbioext.dll ] ext-ms-win-biometrics-winbio-l1-3-0 -> [ winbioext.dll ] ext-ms-win-bluetooth-apis-internal-l1-1-0 -> [ bluetoothapis.dll ] ext-ms-win-bluetooth-apis-l1-1-0 -> [ bluetoothapis.dll ] ext-ms-win-bluetooth-apis-private-l1-1-0 -> [ bluetoothapis.dll ] ext-ms-win-branding-winbrand-l1-1-2 -> [ winbrand.dll ] ext-ms-win-branding-winbrand-l1-2-0 -> [ winbrand.dll ] ext-ms-win-capabilityaccessmanager-storage-l1-1-0 -> [ capabilityaccessmanager.desktop.storage.dll ] ext-ms-win-casting-device-l1-1-0 -> [ ] ext-ms-win-casting-lockscreen-l1-1-0 -> [ miracastreceiverext.dll ] ext-ms-win-casting-receiver-l1-1-1 -> [ ] ext-ms-win-casting-shell-l1-1-0 -> [ castingshellext.dll ] ext-ms-win-ci-management-l1-1-3 -> [ manageci.dll ] ext-ms-win-ci-xbox-l1-1-0 -> [ ] ext-ms-win-cloudap-tbal-l1-1-0 -> [ ] ext-ms-win-clouddomainjoin-usermanagement-l1-1-0 -> [ ] ext-ms-win-cluster-clusapi-l1-1-6 -> [ clusapi.dll ] ext-ms-win-cluster-resutils-l1-1-3 -> [ resutils.dll ] ext-ms-win-cmd-util-l1-1-0 -> [ cmdext.dll ] ext-ms-win-cng-rng-l1-1-1 -> [ bcryptprimitives.dll ] ext-ms-win-com-apartmentrestriction-l1-1-0 -> [ ] ext-ms-win-com-clbcatq-l1-1-0 -> [ clbcatq.dll ] ext-ms-win-com-coml2-l1-1-1 -> [ coml2.dll ] ext-ms-win-com-ole32-l1-1-5 -> [ ole32.dll ] ext-ms-win-com-ole32-l1-2-0 -> [ ole32.dll ] ext-ms-win-com-ole32-l1-3-0 -> [ ole32.dll ] ext-ms-win-com-ole32-l1-4-0 -> [ ole32.dll ] ext-ms-win-com-psmregister-l1-1-0 -> [ kernel.appcore.dll ] ext-ms-win-com-psmregister-l1-2-2 -> [ kernel.appcore.dll ] ext-ms-win-com-psmregister-l1-3-1 -> [ kernel.appcore.dll ] ext-ms-win-com-sta-l1-1-0 -> [ ole32.dll ] ext-ms-win-com-suspendresiliency-l1-1-0 -> [ ] ext-ms-win-composition-ghost-l1-1-0 -> [ dwmghost.dll ] ext-ms-win-composition-holographic-l1-1-0 -> [ ] ext-ms-win-composition-init-l1-1-0 -> [ dwminit.dll ] ext-ms-win-compositor-hosting-l1-1-1 -> [ ism.dll ] ext-ms-win-compositor-hosting-l1-2-1 -> [ ism.dll ] ext-ms-win-compositor-hosting-l1-3-0 -> [ ism.dll ] ext-ms-win-connectionattribution-api-l1-1-0 -> [ connectionattributionapi.dll ] ext-ms-win-containers-policymanagercli-l1-1-1 -> [ ] ext-ms-win-core-app-package-registration-l1-1-1 -> [ ] ext-ms-win-core-app-package-volume-l1-1-0 -> [ ] ext-ms-win-core-container-init-l1-1-0 -> [ ] ext-ms-win-core-dhcp6client-l1-1-0 -> [ ] ext-ms-win-core-game-streaming-l1-1-0 -> [ gamestreamingext.dll ] ext-ms-win-core-iuri-l1-1-0 -> [ urlmon.dll ] ext-ms-win-core-licensemanager-l1-1-2 -> [ ] ext-ms-win-core-marshal-l2-1-0 -> [ ole32.dll ] ext-ms-win-core-pkeyhelper-l1-1-0 -> [ pkeyhelper.dll ] ext-ms-win-core-psm-bi-l1-1-0 -> [ bisrv.dll ] ext-ms-win-core-psm-bi-l1-2-0 -> [ bisrv.dll ] ext-ms-win-core-psm-extendedresourcemode-l1-1-0 -> [ ] ext-ms-win-core-psm-service-l1-1-6 -> [ psmserviceexthost.dll ] ext-ms-win-core-resourcemanager-l1-1-0 -> [ rmclient.dll ] ext-ms-win-core-resourcemanager-l1-2-1 -> [ rmclient.dll ] ext-ms-win-core-resourcepolicy-l1-1-2 -> [ resourcepolicyclient.dll ] ext-ms-win-core-resourcepolicyserver-l1-1-1 -> [ resourcepolicyserver.dll ] ext-ms-win-core-stateseparationext-l1-1-0 -> [ ] ext-ms-win-core-storelicensing-l1-1-0 -> [ licensemanagerapi.dll ] ext-ms-win-core-storelicensing-l1-2-1 -> [ licensemanagerapi.dll ] ext-ms-win-core-symbolicnames-l1-1-0 -> [ tdhres.dll ] ext-ms-win-core-win32k-base-export-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-baseinit-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-common-export-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-common-input-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-common-inputrim-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-common-user-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-dcomp-l1-1-3 -> [ win32kbase.sys ] ext-ms-win-core-win32k-ddccigdi-l1-1-1 -> [ win32kfull.sys ] ext-ms-win-core-win32k-dxgdi-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-dxgk-internal-l1-1-0 -> [ dxgkrnl.sys ] ext-ms-win-core-win32k-dxgk-l1-1-0 -> [ dxgkrnl.sys ] ext-ms-win-core-win32k-flipmgr-l1-1-1 -> [ dxgkrnl.sys ] ext-ms-win-core-win32k-full-export-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-full-float-export-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-full-floatobj-export-l1-1-0 -> [ ] ext-ms-win-core-win32k-fulldcompbase-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-fulldwm-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-fullgdi-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-fulluser-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-fulluser64-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-core-win32k-fulluserbase-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-gdi-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-input-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-inputmit-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-inputrim-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-mindwm-l1-1-0 -> [ ] ext-ms-win-core-win32k-mininput-l1-1-0 -> [ ] ext-ms-win-core-win32k-mininputmit-l1-1-0 -> [ ] ext-ms-win-core-win32k-mininputmitbase-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-minuser-l1-1-0 -> [ ] ext-ms-win-core-win32k-opmgdi-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-surfmgr-l1-1-1 -> [ dxgkrnl.sys ] ext-ms-win-core-win32k-tokenmgr-l1-1-0 -> [ dxgkrnl.sys ] ext-ms-win-core-win32k-user-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-userdisplay-l1-1-0 -> [ win32kbase.sys ] ext-ms-win-core-win32k-userinit-l1-1-0 -> [ win32k.sys ] ext-ms-win-core-winrt-remote-l1-1-0 -> [ ] ext-ms-win-core-winsrv-l1-1-0 -> [ winsrvext.dll ] ext-ms-win-core-winsrv-min-l1-1-0 -> [ ] ext-ms-win-core-xbrm-l1-1-1 -> [ ] ext-ms-win-coreui-l1-1-0 -> [ ] ext-ms-win-coreui-navshutdown-l1-1-0 -> [ navshutdown.dll ] ext-ms-win-crypto-xbox-l1-1-0 -> [ ] ext-ms-win-deployment-productenumerator-l1-1-0 -> [ productenumerator.dll ] ext-ms-win-desktopappx-l1-1-7 -> [ daxexec.dll ] ext-ms-win-desktopappx-l1-2-2 -> [ daxexec.dll ] ext-ms-win-devmgmt-dm-l1-1-3 -> [ dmapisetextimpldesktop.dll ] ext-ms-win-devmgmt-policy-l1-1-3 -> [ policymanager.dll ] ext-ms-win-direct2d-desktop-l1-1-0 -> [ direct2ddesktop.dll ] ext-ms-win-domainjoin-netjoin-l1-1-0 -> [ netjoin.dll ] ext-ms-win-dot3-grouppolicy-l1-1-0 -> [ dot3gpclnt.dll ] ext-ms-win-driver-recovery-l1-1-0 -> [ drvsetup.dll ] ext-ms-win-driver-setup-l1-1-0 -> [ drvsetup.dll ] ext-ms-win-driver-setup-wu-l1-1-1 -> [ drvsetup.dll ] ext-ms-win-drvinst-desktop-l1-1-0 -> [ newdev.dll ] ext-ms-win-dwmapi-ext-l1-1-2 -> [ dwmapi.dll ] ext-ms-win-dwmapidxgi-ext-l1-1-1 -> [ dwmapi.dll ] ext-ms-win-dx-d3d9-l1-1-0 -> [ d3d9.dll ] ext-ms-win-dx-d3dkmt-dxcore-l1-1-5 -> [ dxcore.dll ] ext-ms-win-dx-d3dkmt-gdi-l1-1-0 -> [ gdi32.dll ] ext-ms-win-dx-ddraw-l1-1-0 -> [ ddraw.dll ] ext-ms-win-dx-dinput8-l1-1-0 -> [ dinput8.dll ] ext-ms-win-dx-dxdbhelper-l1-1-4 -> [ directxdatabasehelper.dll ] ext-ms-win-dxcore-internal-l1-1-0 -> [ dxcore.dll ] ext-ms-win-dxcore-l1-1-0 -> [ dxcore.dll ] ext-ms-win-edputil-policy-l1-1-2 -> [ edputil.dll ] ext-ms-win-els-elscore-l1-1-0 -> [ elscore.dll ] ext-ms-win-eventing-pdh-l1-1-3 -> [ pdh.dll ] ext-ms-win-eventing-rundown-l1-1-0 -> [ etwrundown.dll ] ext-ms-win-eventing-tdh-ext-l1-1-0 -> [ tdh.dll ] ext-ms-win-eventing-tdh-priv-l1-1-0 -> [ tdh.dll ] ext-ms-win-eventing-wdi-l1-1-0 -> [ wdi.dll ] ext-ms-win-familysafety-childaccount-l1-1-0 -> [ familysafetyext.dll ] ext-ms-win-feclient-encryptedfile-l1-1-3 -> [ feclient.dll ] ext-ms-win-firewallapi-webproxy-l1-1-1 -> [ firewallapi.dll ] ext-ms-win-font-fontgroups-l1-1-0 -> [ fontgroupsoverride.dll ] ext-ms-win-font-setup-l1-1-0 -> [ muifontsetup.dll ] ext-ms-win-fs-clfs-l1-1-0 -> [ clfs.sys ] ext-ms-win-fs-cscapi-l1-1-1 -> [ cscapi.dll ] ext-ms-win-fs-vssapi-l1-1-0 -> [ vssapi.dll ] ext-ms-win-fsutilext-ifsutil-l1-1-0 -> [ fsutilext.dll ] ext-ms-win-fsutilext-ulib-l1-1-0 -> [ fsutilext.dll ] ext-ms-win-fveapi-query-l1-1-0 -> [ fveapi.dll ] ext-ms-win-gaming-devicefamily-l1-1-0 -> [ ] ext-ms-win-gaming-gamechatoverlay-l1-1-0 -> [ gamechatoverlayext.dll ] ext-ms-win-gaming-xblgamesave-l1-1-0 -> [ xblgamesaveext.dll ] ext-ms-win-gaming-xinput-l1-1-0 -> [ xinputuap.dll ] ext-ms-win-gdi-clipping-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-gdi-dc-create-l1-1-2 -> [ gdi32full.dll ] ext-ms-win-gdi-dc-l1-2-1 -> [ gdi32full.dll ] ext-ms-win-gdi-devcaps-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-gdi-draw-l1-1-3 -> [ gdi32full.dll ] ext-ms-win-gdi-edgegdi-l1-1-0 -> [ ] ext-ms-win-gdi-font-l1-1-3 -> [ gdi32full.dll ] ext-ms-win-gdi-gdiplus-l1-1-0 -> [ gdiplus.dll ] ext-ms-win-gdi-internal-desktop-l1-1-6 -> [ gdi32full.dll ] ext-ms-win-gdi-internal-uap-init-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-gdi-metafile-l1-1-2 -> [ gdi32full.dll ] ext-ms-win-gdi-path-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-gdi-print-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-gdi-private-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-gdi-render-l1-1-0 -> [ gdi32.dll ] ext-ms-win-gdi-rgn-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-gdi-wcs-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-globalization-input-l1-1-3 -> [ globinputhost.dll ] ext-ms-win-gpapi-grouppolicy-l1-1-0 -> [ gpapi.dll ] ext-ms-win-gpsvc-grouppolicy-l1-1-0 -> [ gpsvc.dll ] ext-ms-win-gui-dui70-l1-1-0 -> [ dui70.dll ] ext-ms-win-gui-ieui-l1-1-0 -> [ ieui.dll ] ext-ms-win-gui-uxinit-l1-1-1 -> [ uxinit.dll ] ext-ms-win-hcihealth-utils-l1-1-0 -> [ ] ext-ms-win-hostactivitymanager-bi-ham-ext-l1-1-0 -> [ psmserviceexthost.dll ] ext-ms-win-hostactivitymanager-ham-private-ext-l1-1-0 -> [ psmserviceexthost.dll ] ext-ms-win-hostactivitymanager-hostidstore-l1-1-1 -> [ rmclient.dll ] ext-ms-win-hyperv-compute-l1-1-1 -> [ ] ext-ms-win-hyperv-compute-l1-2-5 -> [ computecore.dll ] ext-ms-win-hyperv-compute-legacy-l1-1-0 -> [ vmcompute.dll ] ext-ms-win-hyperv-computenetwork-l1-1-1 -> [ computenetwork.dll ] ext-ms-win-hyperv-computestorage-l1-1-2 -> [ computestorage.dll ] ext-ms-win-hyperv-devicevirtualization-l1-1-1 -> [ vmdevicehost.dll ] ext-ms-win-hyperv-devicevirtualization-l1-2-2 -> [ vmdevicehost.dll ] ext-ms-win-hyperv-hgs-l1-1-0 -> [ vmhgs.dll ] ext-ms-win-hyperv-hvemulation-l1-1-0 -> [ winhvemulation.dll ] ext-ms-win-hyperv-hvplatform-l1-1-5 -> [ winhvplatform.dll ] ext-ms-win-ie-textinput-l1-1-0 -> [ ] ext-ms-win-imm-l1-1-3 -> [ imm32.dll ] ext-ms-win-kernel32-appcompat-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-datetime-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-elevation-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-errorhandling-l1-1-0 -> [ kernel32.dll,faultrep.dll ] ext-ms-win-kernel32-file-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-localization-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-package-current-l1-1-0 -> [ kernel.appcore.dll ] ext-ms-win-kernel32-package-l1-1-2 -> [ kernel.appcore.dll ] ext-ms-win-kernel32-process-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-quirks-l1-1-1 -> [ kernel32.dll ] ext-ms-win-kernel32-registry-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-sidebyside-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-transacted-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-updateresource-l1-1-0 -> [ kernel32.dll ] ext-ms-win-kernel32-windowserrorreporting-l1-1-1 -> [ kernel32.dll ] ext-ms-win-kernelbase-processthread-l1-1-3 -> [ kernel32.dll ] ext-ms-win-kernelbase-processthread-l1-2-0 -> [ kernel32.dll ] ext-ms-win-kioskmode-config-l1-1-0 -> [ ] ext-ms-win-laps-l1-1-1 -> [ laps.dll ] ext-ms-win-lighting-lamparray-l1-1-1 -> [ lamparray.dll ] ext-ms-win-mapi-mapi32-l1-1-0 -> [ mapistub.dll ] ext-ms-win-media-avi-l1-1-0 -> [ avifil32.dll ] ext-ms-win-media-codecpack-mounting-l1-1-0 -> [ ] ext-ms-win-mf-devicespecific-l1-1-0 -> [ ] ext-ms-win-mf-vfw-l1-1-0 -> [ mfvfw.dll ] ext-ms-win-mininput-cursorhost-l1-1-0 -> [ inputhost.dll ] ext-ms-win-mininput-extensions-l1-1-0 -> [ ] ext-ms-win-mininput-inputhost-l1-1-1 -> [ inputhost.dll ] ext-ms-win-mininput-inputhost-l1-2-1 -> [ inputhost.dll ] ext-ms-win-mininput-inputhost-l1-3-0 -> [ inputhost.dll ] ext-ms-win-mininput-inputhost-l1-4-0 -> [ inputhost.dll ] ext-ms-win-mininput-systeminputhost-l1-1-0 -> [ ism.dll ] ext-ms-win-mininput-systeminputhost-l1-2-0 -> [ ism.dll ] ext-ms-win-mm-io-l1-1-0 -> [ winmmbase.dll ] ext-ms-win-mm-msacm-l1-1-0 -> [ msacm32.dll ] ext-ms-win-mm-pehelper-l1-1-0 -> [ mf.dll ] ext-ms-win-mm-wmvcore-l1-1-0 -> [ wmvcore.dll ] ext-ms-win-moderncore-win32k-base-ntgdi-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-moderncore-win32k-base-ntuser-l1-1-0 -> [ win32kfull.sys ] ext-ms-win-moderncore-win32k-base-sysentry-l1-1-0 -> [ win32k.sys ] ext-ms-win-mpr-multipleproviderrouter-l1-1-0 -> [ mprext.dll ] ext-ms-win-mrmcorer-resmanager-l1-1-0 -> [ mrmcorer.dll ] ext-ms-win-msa-device-l1-1-0 -> [ ] ext-ms-win-msa-ui-l1-1-0 -> [ msauserext.dll ] ext-ms-win-msa-user-l1-1-1 -> [ msauserext.dll ] ext-ms-win-msi-misc-l1-1-0 -> [ msi.dll ] ext-ms-win-msiltcfg-msi-l1-1-0 -> [ msiltcfg.dll ] ext-ms-win-msimg-draw-l1-1-0 -> [ msimg32.dll ] ext-ms-win-net-cmvpn-l1-1-0 -> [ cmintegrator.dll ] ext-ms-win-net-httpproxyext-l1-1-0 -> [ httpprxc.dll ] ext-ms-win-net-isoext-l1-1-0 -> [ firewallapi.dll ] ext-ms-win-net-netbios-l1-1-0 -> [ netbios.dll ] ext-ms-win-net-netshell-l1-1-0 -> [ netshell.dll ] ext-ms-win-net-nfdapi-l1-1-1 -> [ ndfapi.dll ] ext-ms-win-net-vpn-l1-1-0 -> [ ] ext-ms-win-netio-l1-1-0 -> [ netio.sys ] ext-ms-win-netprovision-netprovfw-l1-1-0 -> [ netprovfw.dll ] ext-ms-win-networking-iphlpsvc-l1-1-0 -> [ ] ext-ms-win-networking-mpssvc-l1-1-0 -> [ ] ext-ms-win-networking-ncsiuserprobe-l1-1-0 -> [ ] ext-ms-win-networking-radiomonitor-l1-1-0 -> [ windows.devices.radios.dll ] ext-ms-win-networking-teredo-l1-1-0 -> [ windows.networking.connectivity.dll ] ext-ms-win-networking-wcmapi-l1-1-1 -> [ wcmapi.dll ] ext-ms-win-networking-winipsec-l1-1-0 -> [ winipsec.dll ] ext-ms-win-networking-wlanapi-l1-1-0 -> [ wlanapi.dll ] ext-ms-win-networking-wlanstorage-l1-1-0 -> [ ] ext-ms-win-networking-xblconnectivity-l1-1-0 -> [ ] ext-ms-win-newdev-config-l1-1-2 -> [ newdev.dll ] ext-ms-win-nfc-semgr-l1-1-0 -> [ semgrsvc.dll ] ext-ms-win-ntdsa-activedirectoryasyncthreadqueue-l1-1-0 -> [ ] ext-ms-win-ntdsa-activedirectoryserver-l1-1-2 -> [ ] ext-ms-win-ntdsapi-activedirectoryclient-l1-1-1 -> [ ntdsapi.dll ] ext-ms-win-ntos-clipsp-l1-1-0 -> [ clipsp.sys ] ext-ms-win-ntos-dg-l1-1-0 -> [ ] ext-ms-win-ntos-globmerger-l1-1-0 -> [ globmerger.sys ] ext-ms-win-ntos-kcminitcfg-l1-1-0 -> [ cmimcext.sys ] ext-ms-win-ntos-ksecurity-l1-1-1 -> [ ] ext-ms-win-ntos-ksr-l1-1-5 -> [ ] ext-ms-win-ntos-processparameters-l1-1-0 -> [ ] ext-ms-win-ntos-runlevels-l1-1-0 -> [ ] ext-ms-win-ntos-stateseparation-l1-1-0 -> [ ] ext-ms-win-ntos-tm-l1-1-0 -> [ tm.sys ] ext-ms-win-ntos-trace-l1-1-0 -> [ ] ext-ms-win-ntos-ucode-l1-1-0 -> [ ntosext.sys ] ext-ms-win-ntos-vail-l1-1-0 -> [ ] ext-ms-win-ntos-vmsvc-l1-1-0 -> [ vmsvcext.sys ] ext-ms-win-ntos-werkernel-l1-1-1 -> [ werkernel.sys ] ext-ms-win-ntos-win32k-l1-1-0 -> [ win32k.sys ] ext-ms-win-ntuser-caret-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-chartranslation-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-dc-access-ext-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-dde-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-dialogbox-l1-1-3 -> [ user32.dll ] ext-ms-win-ntuser-draw-l1-1-2 -> [ user32.dll ] ext-ms-win-ntuser-gui-l1-1-1 -> [ user32.dll ] ext-ms-win-ntuser-gui-l1-2-0 -> [ user32.dll ] ext-ms-win-ntuser-gui-l1-3-1 -> [ user32.dll ] ext-ms-win-ntuser-keyboard-ansi-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-keyboard-l1-1-1 -> [ user32.dll ] ext-ms-win-ntuser-keyboard-l1-2-0 -> [ user32.dll ] ext-ms-win-ntuser-keyboard-l1-3-2 -> [ user32.dll ] ext-ms-win-ntuser-menu-l1-1-3 -> [ user32.dll ] ext-ms-win-ntuser-message-l1-1-3 -> [ user32.dll ] ext-ms-win-ntuser-misc-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-misc-l1-2-0 -> [ user32.dll ] ext-ms-win-ntuser-misc-l1-3-0 -> [ user32.dll ] ext-ms-win-ntuser-misc-l1-5-1 -> [ user32.dll ] ext-ms-win-ntuser-misc-l1-6-0 -> [ user32.dll ] ext-ms-win-ntuser-misc-l1-7-0 -> [ user32.dll ] ext-ms-win-ntuser-mit-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-mouse-l1-1-1 -> [ user32.dll ] ext-ms-win-ntuser-powermanagement-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-private-l1-1-1 -> [ user32.dll ] ext-ms-win-ntuser-private-l1-2-0 -> [ user32.dll ] ext-ms-win-ntuser-private-l1-3-3 -> [ user32.dll ] ext-ms-win-ntuser-private-l1-4-0 -> [ user32.dll ] ext-ms-win-ntuser-private-l1-5-0 -> [ user32.dll ] ext-ms-win-ntuser-private-l1-6-3 -> [ user32.dll ] ext-ms-win-ntuser-rawinput-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-rawinput-l1-2-0 -> [ user32.dll ] ext-ms-win-ntuser-rectangle-ext-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-rim-l1-1-2 -> [ user32.dll ] ext-ms-win-ntuser-rim-l1-2-1 -> [ user32.dll ] ext-ms-win-ntuser-rotationmanager-l1-1-2 -> [ user32.dll ] ext-ms-win-ntuser-server-l1-1-1 -> [ user32.dll ] ext-ms-win-ntuser-string-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-synch-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-sysparams-ext-l1-1-1 -> [ user32.dll ] ext-ms-win-ntuser-touch-hittest-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-uicontext-ext-l1-1-0 -> [ user32.dll ] ext-ms-win-ntuser-window-l1-1-6 -> [ user32.dll ] ext-ms-win-ntuser-windowclass-l1-1-2 -> [ user32.dll ] ext-ms-win-ntuser-windowstation-ansi-l1-1-1 -> [ user32.dll ] ext-ms-win-ntuser-windowstation-l1-1-2 -> [ user32.dll ] ext-ms-win-odbc-odbc32-l1-1-0 -> [ odbc32.dll ] ext-ms-win-ole32-bindctx-l1-1-0 -> [ ole32.dll ] ext-ms-win-ole32-ie-ext-l1-1-0 -> [ ole32.dll ] ext-ms-win-ole32-oleautomation-l1-1-0 -> [ ole32.dll ] ext-ms-win-oleacc-l1-1-2 -> [ oleacc.dll ] ext-ms-win-onecore-shutdown-l1-1-0 -> [ twinapi.appcore.dll ] ext-ms-win-oobe-query-l1-1-0 -> [ ] ext-ms-win-packagevirtualizationcontext-l1-1-0 -> [ daxexec.dll ] ext-ms-win-parentalcontrols-setup-l1-1-0 -> [ wpcapi.dll ] ext-ms-win-perception-device-l1-1-1 -> [ ] ext-ms-win-pinenrollment-enrollment-l1-1-2 -> [ pinenrollmenthelper.dll ] ext-ms-win-printer-prntvpt-l1-1-2 -> [ prntvpt.dll ] ext-ms-win-printer-winspool-core-l1-1-0 -> [ winspool.drv ] ext-ms-win-printer-winspool-l1-1-4 -> [ winspool.drv ] ext-ms-win-printer-winspool-l1-2-0 -> [ winspool.drv ] ext-ms-win-profile-extender-l1-1-0 -> [ userenv.dll ] ext-ms-win-profile-load-l1-1-0 -> [ ] ext-ms-win-profile-profsvc-l1-1-0 -> [ profsvcext.dll ] ext-ms-win-profile-userenv-l1-1-1 -> [ profext.dll ] ext-ms-win-provisioning-options-l1-1-0 -> [ ] ext-ms-win-provisioning-platform-l1-1-2 -> [ provplatformdesktop.dll ] ext-ms-win-ras-rasapi32-l1-1-2 -> [ rasapi32.dll ] ext-ms-win-ras-rasdlg-l1-1-0 -> [ rasdlg.dll ] ext-ms-win-ras-rasman-l1-1-0 -> [ rasman.dll ] ext-ms-win-ras-tapi32-l1-1-1 -> [ tapi32.dll ] ext-ms-win-raschapext-eap-l1-1-0 -> [ raschapext.dll ] ext-ms-win-rastlsext-eap-l1-1-0 -> [ rastlsext.dll ] ext-ms-win-rdr-davhlpr-l1-1-0 -> [ davhlpr.dll ] ext-ms-win-reinfo-query-l1-1-0 -> [ reinfo.dll ] ext-ms-win-remotewipe-platform-l1-1-0 -> [ ] ext-ms-win-resourcemanager-activitycoordinator-l1-1-1 -> [ rmclient.dll ] ext-ms-win-resourcemanager-crm-l1-1-0 -> [ rmclient.dll ] ext-ms-win-resourcemanager-crm-l1-2-0 -> [ rmclient.dll ] ext-ms-win-resourcemanager-crm-private-ext-l1-1-0 -> [ psmserviceexthost.dll ] ext-ms-win-resourcemanager-gamemode-l1-1-0 -> [ rmclient.dll ] ext-ms-win-resourcemanager-gamemode-l1-2-1 -> [ rmclient.dll ] ext-ms-win-resourcemanager-limits-l1-1-0 -> [ rmclient.dll ] ext-ms-win-resources-deployment-l1-1-0 -> [ mrmdeploy.dll ] ext-ms-win-resources-languageoverlay-l1-1-7 -> [ languageoverlayutil.dll ] ext-ms-win-ro-typeresolution-l1-1-1 -> [ wintypes.dll ] ext-ms-win-rometadata-dispenser-l1-1-0 -> [ rometadata.dll ] ext-ms-win-rpc-firewallportuse-l1-1-0 -> [ rpcrtremote.dll ] ext-ms-win-rpc-ssl-l1-1-0 -> [ rpcrtremote.dll ] ext-ms-win-rtcore-gdi-devcaps-l1-1-1 -> [ gdi32.dll ] ext-ms-win-rtcore-gdi-object-l1-1-0 -> [ gdi32.dll ] ext-ms-win-rtcore-gdi-rgn-l1-1-1 -> [ gdi32.dll ] ext-ms-win-rtcore-minuser-display-l1-1-0 -> [ ] ext-ms-win-rtcore-minuser-host-l1-1-0 -> [ ] ext-ms-win-rtcore-minuser-input-l1-1-4 -> [ ] ext-ms-win-rtcore-minuser-internal-l1-1-0 -> [ ] ext-ms-win-rtcore-minuser-private-ext-l1-1-3 -> [ ] ext-ms-win-rtcore-ntuser-controllernavigation-l1-1-2 -> [ inputhost.dll ] ext-ms-win-rtcore-ntuser-cursor-l1-1-1 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-dc-access-l1-1-1 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-dialogbox-l1-1-0 -> [ ] ext-ms-win-rtcore-ntuser-dpi-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-dpi-l1-2-2 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-draw-l1-1-0 -> [ ] ext-ms-win-rtcore-ntuser-gui-l1-1-1 -> [ ] ext-ms-win-rtcore-ntuser-iam-l1-1-2 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-inputintercept-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-integration-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-keyboard-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-message-ansi-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-message-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-mininit-l1-1-0 -> [ ] ext-ms-win-rtcore-ntuser-misc-l1-1-0 -> [ ] ext-ms-win-rtcore-ntuser-mouse-l1-1-0 -> [ ] ext-ms-win-rtcore-ntuser-powermanagement-l1-1-0 -> [ ] ext-ms-win-rtcore-ntuser-private-l1-1-1 -> [ ] ext-ms-win-rtcore-ntuser-rawinput-l1-1-1 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-rawinput-l1-2-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-synch-ext-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-syscolors-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-sysparams-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-usersecurity-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-window-ansi-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-window-ext-l1-1-1 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-window-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-winevent-ext-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-winstamin-l1-1-0 -> [ ] ext-ms-win-rtcore-ntuser-wmpointer-l1-1-0 -> [ user32.dll ] ext-ms-win-rtcore-ntuser-wmpointermin-l1-1-0 -> [ ] ext-ms-win-rtcore-ole32-dragdrop-l1-1-0 -> [ ole32.dll ] ext-ms-win-rtcore-ole32-misc-l1-1-0 -> [ ole32.dll ] ext-ms-win-rtcore-webview-l1-1-0 -> [ ] ext-ms-win-samsrv-accountstore-l1-1-1 -> [ samsrv.dll ] ext-ms-win-scesrv-server-l1-1-0 -> [ scesrv.dll ] ext-ms-win-search-folder-l1-1-1 -> [ searchfolder.dll ] ext-ms-win-search-lifetimemanager-l1-1-0 -> [ mssrch.dll ] ext-ms-win-secur32-translatename-l1-1-0 -> [ secur32.dll ] ext-ms-win-security-appinfoext-l1-1-0 -> [ appinfoext.dll ] ext-ms-win-security-authbrokerui-l1-1-0 -> [ authbrokerui.dll ] ext-ms-win-security-authz-helper-l1-1-0 -> [ ] ext-ms-win-security-capauthz-l1-1-1 -> [ capauthz.dll ] ext-ms-win-security-catalog-database-l1-1-0 -> [ cryptcatsvc.dll ] ext-ms-win-security-certpoleng-l1-1-0 -> [ certpoleng.dll ] ext-ms-win-security-cfl-l1-1-2 -> [ cflapi.dll ] ext-ms-win-security-chambers-l1-1-1 -> [ ] ext-ms-win-security-credui-internal-l1-1-1 -> [ wincredui.dll ] ext-ms-win-security-credui-l1-1-1 -> [ credui.dll ] ext-ms-win-security-crosscontainerauthhelper-l1-1-0 -> [ ] ext-ms-win-security-cryptui-l1-1-1 -> [ cryptui.dll ] ext-ms-win-security-developerunlock-l1-1-0 -> [ ] ext-ms-win-security-deviceid-l1-1-0 -> [ ] ext-ms-win-security-efs-l1-1-1 -> [ efsext.dll ] ext-ms-win-security-efswrt-l1-1-4 -> [ efswrt.dll ] ext-ms-win-security-kerberos-l1-1-0 -> [ kerberos.dll ] ext-ms-win-security-lsaadt-l1-1-0 -> [ lsaadt.dll ] ext-ms-win-security-lsaadtpriv-l1-1-0 -> [ lsaadt.dll ] ext-ms-win-security-lsaauditrpc-l1-1-0 -> [ lsaadt.dll ] ext-ms-win-security-ngc-local-l1-1-0 -> [ ngclocal.dll ] ext-ms-win-security-shutdownext-l1-1-0 -> [ shutdownext.dll ] ext-ms-win-security-slc-l1-1-0 -> [ slc.dll ] ext-ms-win-security-srp-l1-1-1 -> [ srpapi.dll ] ext-ms-win-security-tokenbrokerui-l1-1-0 -> [ tokenbrokerui.dll ] ext-ms-win-security-vaultcds-l1-1-0 -> [ vaultcds.dll ] ext-ms-win-security-vaultcds-l1-2-0 -> [ vaultcds.dll ] ext-ms-win-security-vaultcli-l1-1-1 -> [ vaultcli.dll ] ext-ms-win-security-winscard-l1-1-1 -> [ winscard.dll ] ext-ms-win-sensors-core-private-l1-1-8 -> [ sensorsnativeapi.dll ] ext-ms-win-sensors-utilities-private-l1-1-5 -> [ sensorsutilsv2.dll ] ext-ms-win-servicing-uapi-l1-1-2 -> [ servicinguapi.dll ] ext-ms-win-session-candidateaccountmgr-l1-1-0 -> [ camext.dll ] ext-ms-win-session-userinit-l1-1-0 -> [ userinitext.dll ] ext-ms-win-session-usermgr-l1-1-0 -> [ usermgrcli.dll ] ext-ms-win-session-usermgr-l1-2-1 -> [ usermgrcli.dll ] ext-ms-win-session-usertoken-l1-1-0 -> [ wtsapi32.dll ] ext-ms-win-session-wininit-l1-1-1 -> [ wininitext.dll ] ext-ms-win-session-wininit-l1-2-0 -> [ wininitext.dll ] ext-ms-win-session-winlogon-l1-1-2 -> [ winlogonext.dll ] ext-ms-win-session-winlogon-notify-l1-1-0 -> [ ] ext-ms-win-session-winsta-l1-1-6 -> [ winsta.dll ] ext-ms-win-session-wtsapi32-l1-1-2 -> [ wtsapi32.dll ] ext-ms-win-setupapi-classinstallers-l1-1-2 -> [ setupapi.dll ] ext-ms-win-setupapi-devobj-l1-1-0 -> [ ] ext-ms-win-setupapi-inf-l1-1-1 -> [ setupapi.dll ] ext-ms-win-setupapi-logging-l1-1-0 -> [ setupapi.dll ] ext-ms-win-shell-aclui-l1-1-0 -> [ aclui.dll ] ext-ms-win-shell-browsersettingsync-l1-1-0 -> [ ] ext-ms-win-shell-comctl32-da-l1-1-0 -> [ comctl32.dll ] ext-ms-win-shell-comctl32-init-l1-1-1 -> [ comctl32.dll ] ext-ms-win-shell-comctl32-l1-1-0 -> [ comctl32.dll ] ext-ms-win-shell-comctl32-window-l1-1-0 -> [ comctl32.dll ] ext-ms-win-shell-comdlg32-l1-1-1 -> [ comdlg32.dll ] ext-ms-win-shell-directory-l1-1-0 -> [ windows.storage.dll ] ext-ms-win-shell-efsadu-l1-1-0 -> [ efsadu.dll ] ext-ms-win-shell-embeddedmode-l1-1-0 -> [ embeddedmodesvcapi.dll ] ext-ms-win-shell-exports-internal-l1-1-1 -> [ shell32.dll ] ext-ms-win-shell-fileplaceholder-l1-1-0 -> [ windows.fileexplorer.common.dll ] ext-ms-win-shell-ntshrui-l1-1-0 -> [ ntshrui.dll ] ext-ms-win-shell-propsys-l1-1-1 -> [ propsys.dll ] ext-ms-win-shell-shdocvw-l1-1-0 -> [ shdocvw.dll ] ext-ms-win-shell-shell32-l1-2-3 -> [ shell32.dll ] ext-ms-win-shell-shell32-l1-3-0 -> [ shell32.dll ] ext-ms-win-shell-shell32-l1-4-0 -> [ shell32.dll ] ext-ms-win-shell-shell32-l1-5-0 -> [ shell32.dll ] ext-ms-win-shell-shlwapi-l1-1-2 -> [ shlwapi.dll ] ext-ms-win-shell-shlwapi-l1-2-1 -> [ shlwapi.dll ] ext-ms-win-shell32-shellcom-l1-1-0 -> [ windows.storage.dll ] ext-ms-win-shell32-shellfolders-l1-1-1 -> [ windows.storage.dll ] ext-ms-win-shell32-shellfolders-l1-2-1 -> [ windows.storage.dll ] ext-ms-win-smbshare-browser-l1-1-0 -> [ browser.dll ] ext-ms-win-smbshare-browserclient-l1-1-0 -> [ browcli.dll ] ext-ms-win-smbshare-sscore-l1-1-0 -> [ sscoreext.dll ] ext-ms-win-spinf-inf-l1-1-0 -> [ spinf.dll ] ext-ms-win-storage-hbaapi-l1-1-1 -> [ hbaapi.dll ] ext-ms-win-storage-iscsidsc-l1-1-0 -> [ iscsidsc.dll ] ext-ms-win-storage-sense-l1-1-0 -> [ storageusage.dll ] ext-ms-win-storage-sense-l1-2-5 -> [ storageusage.dll ] ext-ms-win-sxs-oleautomation-l1-1-0 -> [ sxs.dll ] ext-ms-win-sysmain-pfapi-l1-1-1 -> [ pfclient.dll ] ext-ms-win-sysmain-pfsapi-l1-1-0 -> [ pfclient.dll ] ext-ms-win-sysmain-plmapi-l1-1-1 -> [ pfclient.dll ] ext-ms-win-sysmain-sfltapi-l1-1-1 -> [ ] ext-ms-win-system-metrics-override-l1-1-1 -> [ ] ext-ms-win-teapext-eap-l1-1-0 -> [ eapteapext.dll ] ext-ms-win-test-sys1-l1-1-0 -> [ ] ext-ms-win-test-sys2-l1-1-0 -> [ ] ext-ms-win-tsf-inputsetting-l1-1-0 -> [ input.dll ] ext-ms-win-tsf-msctf-l1-1-4 -> [ msctf.dll ] ext-ms-win-ttlsext-eap-l1-1-0 -> [ ttlsext.dll ] ext-ms-win-ui-viewmanagement-l1-1-0 -> [ ] ext-ms-win-uiacore-l1-1-3 -> [ uiautomationcore.dll ] ext-ms-win-umpoext-umpo-l1-1-0 -> [ umpoext.dll ] ext-ms-win-usp10-l1-1-0 -> [ gdi32full.dll ] ext-ms-win-uwf-servicing-apis-l1-1-1 -> [ uwfservicingapi.dll ] ext-ms-win-uxtheme-themes-l1-1-3 -> [ uxtheme.dll ] ext-ms-win-vmbus-hvsocket-l1-1-0 -> [ hvsocket.sys ] ext-ms-win-webtokenrequest-win32-l1-1-0 -> [ ] ext-ms-win-wer-reporting-l1-1-3 -> [ wer.dll ] ext-ms-win-wer-ui-l1-1-1 -> [ werui.dll ] ext-ms-win-wer-wct-l1-1-0 -> [ wer.dll ] ext-ms-win-wer-xbox-l1-1-3 -> [ ] ext-ms-win-wer-xbox-l1-2-1 -> [ ] ext-ms-win-wevtapi-eventlog-l1-1-3 -> [ wevtapi.dll ] ext-ms-win-windowing-external-l1-1-0 -> [ windows.ui.dll ] ext-ms-win-windowing-internal-l1-1-0 -> [ ] ext-ms-win-winlogon-mincreds-l1-1-0 -> [ ] ext-ms-win-winrt-device-access-l1-1-0 -> [ deviceaccess.dll ] ext-ms-win-winrt-storage-l1-1-0 -> [ windows.storage.dll ] ext-ms-win-winrt-storage-l1-2-3 -> [ windows.storage.dll ] ext-ms-win-winrt-storage-win32broker-l1-1-0 -> [ windows.storage.onecore.dll ] ext-ms-win-wlan-grouppolicy-l1-1-0 -> [ wlgpclnt.dll ] ext-ms-win-wlan-onexui-l1-1-0 -> [ onexui.dll ] ext-ms-win-wlan-scard-l1-1-0 -> [ winscard.dll ] ext-ms-win-wpc-webfilter-l1-1-0 -> [ wpcwebfilter.dll ] ext-ms-win-wpn-phoneext-l1-1-0 -> [ ] ext-ms-win-wrp-sfc-l1-1-0 -> [ sfc.dll ] ext-ms-win-wsclient-devlicense-l1-1-1 -> [ wsclient.dll ] ext-ms-win-wwaext-misc-l1-1-0 -> [ wwaext.dll ] ext-ms-win-wwaext-module-l1-1-0 -> [ wwaext.dll ] ext-ms-win-wwan-wwapi-l1-1-3 -> [ wwapi.dll ] ext-ms-win-xaml-controls-l1-1-0 -> [ windows.ui.xaml.phone.dll ] ext-ms-win-xaml-pal-l1-1-0 -> [ ] ext-ms-win-xaudio-platform-l1-1-0 -> [ ] ext-ms-win-xblauth-console-l1-1-0 -> [ ] ext-ms-win-xboxlive-xboxnetapisvc-l1-1-0 -> [ ] ext-ms-win32-subsystem-query-l1-1-0 -> [ ] ext-ms-windowscore-deviceinfo-l1-1-0 -> [ ] ``` I've included the dependencies chain in a log file. [log.txt](https://github.com/user-attachments/files/23348751/log.txt)
Author
Owner

@catclaw commented on GitHub (Nov 5, 2025):

I can't explain this - but I started Windows 11 Sandbox and installed Ollama, quit the "ollama app.exe" (in the taskbar), and restarted it in cmd with ollama serve...
...And it detected all my 3 GPUs, no errors, no hassle...
The problem is that as soon as I close the Sandbox, everything will disappear, and the strange thing is that the Sandbox is a "copy" of my existing Windows 11 installation, running in a VM. (With GPU passthrough.) 😳😳😳😳

I feel.... lost!

Image
<!-- gh-comment-id:3491803126 --> @catclaw commented on GitHub (Nov 5, 2025): I can't explain this - but I started Windows 11 Sandbox and installed Ollama, quit the "ollama app.exe" (in the taskbar), and restarted it in cmd with ollama serve... ...And it detected all my 3 GPUs, no errors, no hassle... The problem is that as soon as I close the Sandbox, everything will disappear, and the strange thing is that the Sandbox is a "copy" of my existing Windows 11 installation, running in a VM. (With GPU passthrough.) 😳😳😳😳 I feel.... lost! <img width="2561" height="1545" alt="Image" src="https://github.com/user-attachments/assets/06aeaf44-ef95-4b44-aaf8-d2db2f3ac50e" />
Author
Owner

@dhiltgen commented on GitHub (Nov 5, 2025):

@catclaw one point of clarification - we dynamically load the ggml-*.dll files at runtime, so they won't show up as dependencies of main ollama.exe binary. You'll want to scan any of those files mentioned above with The specified procedure could not be found errors.

<!-- gh-comment-id:3492833461 --> @dhiltgen commented on GitHub (Nov 5, 2025): @catclaw one point of clarification - we dynamically load the `ggml-*.dll` files at runtime, so they won't show up as dependencies of main ollama.exe binary. You'll want to scan any of those files mentioned above with `The specified procedure could not be found` errors.
Author
Owner

@catclaw commented on GitHub (Nov 7, 2025):

Sorry, guys and gals...
I've installed the latest Ollama version, but it still refuses to detect my GPUs.
I've uninstalled CUDA 12.9 and 12.8 and only kept 13.0.88
Visual C++ Red. x64 2015-2022 is installed (actually ALL of them, down to 2005)
I have tried every solution I can think of, and Dependencies is not giving me any clues besides these files, which dependencies.exe -modules shows as missing in CLI and are marked as orange in the GUI (missing):

[NOT_FOUND] ext-ms-win-security-chambers-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-oobe-query-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-core-app-package-registration-l1-1-0.dll :
[NOT_FOUND] ext-ms-win32-subsystem-query-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-com-suspendresiliency-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-appmodel-deployment-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-appmodel-usercontext-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-gaming-devicefamily-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-ui-viewmanagement-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-2.dll :
[NOT_FOUND] ext-ms-win-gdi-edgegdi-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-appmodel-viewscalefactor-l1-1-0.dll :
[NOT_FOUND] ext-ms-onecore-defaultdiscovery-l1-1-0.dll :
[NOT_FOUND] wpaxholder.dll :
[NOT_FOUND] ext-ms-windowscore-deviceinfo-l1-1-0.dll :
[NOT_FOUND] ext-ms-mf-pal-l2-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-input-l1-1-1.dll :
[NOT_FOUND] api-ms-win-coreui-secruntime-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-ntuser-private-l1-1-1.dll :
[NOT_FOUND] ext-ms-win-media-codecpack-mounting-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-core-winrt-remote-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-com-apartmentrestriction-l1-1-0.dll :
[NOT_FOUND] PdmUtilities.dll :
[NOT_FOUND] HvsiFileTrust.dll :
[NOT_FOUND] ext-ms-win-audiocore-pal-l1-2-0.dll :
[NOT_FOUND] ext-ms-win-audiocore-spatial-l1-1-0.dll :
[NOT_FOUND] UpdateAPI.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-1.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-0.dll :
[NOT_FOUND] AzureAttestManager.dll :
[NOT_FOUND] AzureAttestNormal.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-1-2.dll :
[NOT_FOUND] ext-ms-win-networking-wlanstorage-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-1-3.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-2-0.dll :
[NOT_FOUND] ext-ms-win-wer-xbox-l1-2-1.dll :
[NOT_FOUND] ext-ms-onecore-orientation-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-internal-l1-1-0.dll :
[NOT_FOUND] ext-ms-onecore-shellchromeapi-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-3.dll :
[NOT_FOUND] ext-ms-win-core-xbrm-l1-1-0.dll :
[NOT_FOUND] ext-ms-win-core-xbrm-l1-1-1.dll :
[NOT_FOUND] ext-ms-win-ntdsa-activedirectoryserver-l1-1-0.dll :

(Which also the command line version shows as not found.)
When I install and run Ollama in the Sandbox, it works just fine, but I noticed something important - I thought the Sandbox VM was a copy of my existing Windows 11 installation - which it is not. Winver reports 24H2, while I have 25H2 on my computer. (An no way to go back, without having to reinstall several hundreds programs - some very large and time-consuming like the Adobe family.

This is the debug output from Ollama in Sandbox: (Debug = 2)

time=2025-11-07T21:56:11.714+01:00 level=INFO source=routes.go:1525 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\Users\\WDAGUtilityAccount\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]"
time=2025-11-07T21:56:11.714+01:00 level=INFO source=images.go:522 msg="total blobs: 0"
time=2025-11-07T21:56:11.714+01:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-11-07T21:56:11.714+01:00 level=INFO source=routes.go:1578 msg="Listening on 127.0.0.1:11434 (version 0.12.10)"
time=2025-11-07T21:56:11.714+01:00 level=DEBUG source=sched.go:120 msg="starting llm scheduler"
time=2025-11-07T21:56:11.719+01:00 level=INFO source=runner.go:67 msg="discovering available GPUs..."
time=2025-11-07T21:56:11.719+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extraEnvs=map[]
time=2025-11-07T21:56:11.743+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 65165"
time=2025-11-07T21:56:11.744+01:00 level=DEBUG source=server.go:401 msg=subprocess OLLAMA_DEBUG=2 PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Users\WDAGUtilityAccount\AppData\Local\Microsoft\WindowsApps;;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama OLLAMA_LIBRARY_PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12
time=2025-11-07T21:56:11.827+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine"
time=2025-11-07T21:56:11.830+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:65165"
time=2025-11-07T21:56:11.839+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string
time=2025-11-07T21:56:11.839+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string
time=2025-11-07T21:56:11.839+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:56:11.840+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-11-07T21:56:11.841+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama
load_backend: loaded CPU backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-07T21:56:11.880+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 3 CUDA devices:
  Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8
  Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced
  Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c
load_backend: loaded CUDA backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll
time=2025-11-07T21:56:12.938+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)
time=2025-11-07T21:56:12.938+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=1.1014927s
ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8102727680 total: 12820938752
ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11958046720 total: 12884901888
ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11895758848 total: 12884901888
time=2025-11-07T21:56:13.908+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=967.9884ms
time=2025-11-07T21:56:13.910+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:8102727680 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11958046720 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11895758848 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]}]"
time=2025-11-07T21:56:13.911+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=2.1922203s OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=map[]
time=2025-11-07T21:56:13.911+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[]
time=2025-11-07T21:56:13.915+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 65175"
time=2025-11-07T21:56:13.915+01:00 level=DEBUG source=server.go:401 msg=subprocess OLLAMA_DEBUG=2 PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Users\WDAGUtilityAccount\AppData\Local\Microsoft\WindowsApps;;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama OLLAMA_LIBRARY_PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13
time=2025-11-07T21:56:14.001+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine"
time=2025-11-07T21:56:14.004+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:65175"
time=2025-11-07T21:56:14.010+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string
time=2025-11-07T21:56:14.010+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string
time=2025-11-07T21:56:14.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:56:14.011+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:56:14.011+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-11-07T21:56:14.012+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-11-07T21:56:14.013+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-11-07T21:56:14.013+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-07T21:56:14.013+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama
load_backend: loaded CPU backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-07T21:56:14.047+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 3 CUDA devices:
  Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8
  Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced
  Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c
load_backend: loaded CUDA backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll
time=2025-11-07T21:56:14.969+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)
time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=961.571ms
ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8097550336 total: 12820938752
ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11958046720 total: 12884901888
ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11895758848 total: 12884901888
time=2025-11-07T21:56:15.759+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=787.6786ms
time=2025-11-07T21:56:15.765+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:8097550336 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11958046720 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11895758848 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]}]"
time=2025-11-07T21:56:15.766+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=1.8546423s OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[]
time=2025-11-07T21:56:15.766+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extraEnvs=map[]
time=2025-11-07T21:56:15.770+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 65185"
time=2025-11-07T21:56:15.770+01:00 level=DEBUG source=server.go:401 msg=subprocess OLLAMA_DEBUG=2 PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Users\WDAGUtilityAccount\AppData\Local\Microsoft\WindowsApps;;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama OLLAMA_LIBRARY_PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm
time=2025-11-07T21:56:15.871+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine"
time=2025-11-07T21:56:15.874+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:65185"
time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string
time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string
time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-11-07T21:56:15.880+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-07T21:56:15.881+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama
load_backend: loaded CPU backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-07T21:56:15.917+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm
ggml_cuda_init: failed to initialize ROCm: no ROCm-capable device is detected
load_backend: loaded ROCm backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm\ggml-hip.dll
time=2025-11-07T21:56:15.991+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(clang)
time=2025-11-07T21:56:15.992+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=116.0823ms
time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s
time=2025-11-07T21:56:15.994+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[]
time=2025-11-07T21:56:15.996+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=229.7144ms OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=map[]
time=2025-11-07T21:56:15.996+01:00 level=DEBUG source=runner.go:113 msg="evluating which if any devices to filter out" initial_count=6
time=2025-11-07T21:56:15.996+01:00 level=TRACE source=runner.go:153 msg="supported GPU library combinations before filtering" supported=map[]
time=2025-11-07T21:56:15.996+01:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=4.2818412s
time=2025-11-07T21:56:15.996+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced filter_id="" library=CUDA compute=8.6 name=CUDA1 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v12 driver=13.0 pci_id=0000:04:00.0 type=discrete total="12.0 GiB" available="11.1 GiB"
time=2025-11-07T21:56:15.996+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c filter_id="" library=CUDA compute=8.6 name=CUDA2 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v12 driver=13.0 pci_id=0000:83:00.0 type=discrete total="12.0 GiB" available="11.1 GiB"
time=2025-11-07T21:56:15.996+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 filter_id="" library=CUDA compute=12.0 name=CUDA0 description="NVIDIA GeForce RTX 5070" libdirs=ollama,cuda_v12 driver=13.0 pci_id=0000:03:00.0 type=discrete total="11.9 GiB" available="7.5 GiB"

It detects my GPUs and loads models to VRAM.
This is the serve.log:
serve.log

However, when I install and run Ollama on my computer (outside Sandbox):
This is the output: (Debug = 2)

time=2025-11-07T21:24:59.289+01:00 level=INFO source=routes.go:1525 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:3 OLLAMA_MODELS:H:\\LMM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:3 OLLAMA_ORIGINS:[*\"; ollama serve http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]"
time=2025-11-07T21:24:59.295+01:00 level=INFO source=images.go:522 msg="total blobs: 0"
time=2025-11-07T21:24:59.299+01:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-11-07T21:24:59.302+01:00 level=INFO source=routes.go:1578 msg="Listening on 127.0.0.1:11434 (version 0.12.10)"
time=2025-11-07T21:24:59.302+01:00 level=DEBUG source=sched.go:120 msg="starting llm scheduler"
time=2025-11-07T21:24:59.308+01:00 level=INFO source=runner.go:67 msg="discovering available GPUs..."
time=2025-11-07T21:24:59.308+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[]
time=2025-11-07T21:24:59.340+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52942"
time=2025-11-07T21:24:59.340+01:00 level=DEBUG source=server.go:401 msg=subprocess CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LMM OLLAMA_NUM_PARALLEL=3 OLLAMA_ORIGINS="\"*\"; ollama serve" OLLAMA_SCHED_SPREAD=1 OLLAMA_TMPDIR=H:\LLM\temp PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm\\rocblas\\library" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-11-07T21:24:59.493+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine"
time=2025-11-07T21:24:59.497+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:52942"
time=2025-11-07T21:24:59.502+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string
time=2025-11-07T21:24:59.502+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string
time=2025-11-07T21:24:59.502+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-11-07T21:24:59.503+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found.

time=2025-11-07T21:24:59.528+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll: The specified procedure could not be found.

time=2025-11-07T21:24:59.550+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-11-07T21:24:59.550+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=51.6783ms
time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s
time=2025-11-07T21:24:59.554+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[]
time=2025-11-07T21:24:59.578+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=269.7244ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[]
time=2025-11-07T21:24:59.579+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extraEnvs=map[]
time=2025-11-07T21:24:59.583+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52958"
time=2025-11-07T21:24:59.588+01:00 level=DEBUG source=server.go:401 msg=subprocess CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LMM OLLAMA_NUM_PARALLEL=3 OLLAMA_ORIGINS="\"*\"; ollama serve" OLLAMA_SCHED_SPREAD=1 OLLAMA_TMPDIR=H:\LLM\temp PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm\\rocblas\\library" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm"
time=2025-11-07T21:24:59.750+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine"
time=2025-11-07T21:24:59.754+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:52958"
time=2025-11-07T21:24:59.766+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string
time=2025-11-07T21:24:59.766+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string
time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-11-07T21:24:59.767+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found.

time=2025-11-07T21:24:59.790+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\ggml-hip.dll: The specified procedure could not be found.

time=2025-11-07T21:24:59.824+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=61.7733ms
time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s
time=2025-11-07T21:24:59.828+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[]
time=2025-11-07T21:24:59.842+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=263.4474ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=map[]
time=2025-11-07T21:24:59.844+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extraEnvs=map[]
time=2025-11-07T21:24:59.849+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52703"
time=2025-11-07T21:24:59.851+01:00 level=DEBUG source=server.go:401 msg=subprocess CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LMM OLLAMA_NUM_PARALLEL=3 OLLAMA_ORIGINS="\"*\"; ollama serve" OLLAMA_SCHED_SPREAD=1 OLLAMA_TMPDIR=H:\LLM\temp PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm\\rocblas\\library" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
time=2025-11-07T21:24:59.994+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine"
time=2025-11-07T21:24:59.998+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:52703"
time=2025-11-07T21:25:00.009+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string
time=2025-11-07T21:25:00.009+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string
time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-11-07T21:25:00.010+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found.

dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found.

time=2025-11-07T21:25:00.034+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found.

time=2025-11-07T21:25:00.077+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang)
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=70.9579ms
time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s
time=2025-11-07T21:25:00.081+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[]
time=2025-11-07T21:25:00.095+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=251.0928ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=map[]
time=2025-11-07T21:25:00.096+01:00 level=DEBUG source=runner.go:113 msg="evluating which if any devices to filter out" initial_count=0
time=2025-11-07T21:25:00.097+01:00 level=TRACE source=runner.go:153 msg="supported GPU library combinations before filtering" supported=map[]
time=2025-11-07T21:25:00.098+01:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=794.929ms
time=2025-11-07T21:25:00.100+01:00 level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="101.3 GiB"
time=2025-11-07T21:25:00.100+01:00 level=INFO source=routes.go:1619 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

And this is the serve.log:
serve.log

I've tried everything I could think of - I've added Ollama LIB folders to PATH, I've copied the DLL files reported missing with Dependencies.exe/DependenciesGUI.exe from other x64 software to the LIB folder...

I don't know what to do anymore? 😖😭😭😭

<!-- gh-comment-id:3505115197 --> @catclaw commented on GitHub (Nov 7, 2025): Sorry, guys and gals... I've installed the latest Ollama version, but it still refuses to detect my GPUs. I've uninstalled CUDA 12.9 and 12.8 and only kept 13.0.88 Visual C++ Red. x64 2015-2022 is installed (actually ALL of them, down to 2005) I have tried every solution I can think of, and Dependencies is not giving me any clues besides these files, which dependencies.exe -modules shows as missing in CLI and are marked as orange in the GUI (missing): ``` [NOT_FOUND] ext-ms-win-security-chambers-l1-1-0.dll : [NOT_FOUND] ext-ms-win-oobe-query-l1-1-0.dll : [NOT_FOUND] ext-ms-win-core-app-package-registration-l1-1-0.dll : [NOT_FOUND] ext-ms-win32-subsystem-query-l1-1-0.dll : [NOT_FOUND] ext-ms-win-com-suspendresiliency-l1-1-0.dll : [NOT_FOUND] ext-ms-win-appmodel-deployment-l1-1-0.dll : [NOT_FOUND] ext-ms-win-appmodel-usercontext-l1-1-0.dll : [NOT_FOUND] ext-ms-win-gaming-devicefamily-l1-1-0.dll : [NOT_FOUND] ext-ms-win-ui-viewmanagement-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-2.dll : [NOT_FOUND] ext-ms-win-gdi-edgegdi-l1-1-0.dll : [NOT_FOUND] ext-ms-win-appmodel-viewscalefactor-l1-1-0.dll : [NOT_FOUND] ext-ms-onecore-defaultdiscovery-l1-1-0.dll : [NOT_FOUND] wpaxholder.dll : [NOT_FOUND] ext-ms-windowscore-deviceinfo-l1-1-0.dll : [NOT_FOUND] ext-ms-mf-pal-l2-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-input-l1-1-1.dll : [NOT_FOUND] api-ms-win-coreui-secruntime-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-ntuser-private-l1-1-1.dll : [NOT_FOUND] ext-ms-win-media-codecpack-mounting-l1-1-0.dll : [NOT_FOUND] ext-ms-win-core-winrt-remote-l1-1-0.dll : [NOT_FOUND] ext-ms-win-com-apartmentrestriction-l1-1-0.dll : [NOT_FOUND] PdmUtilities.dll : [NOT_FOUND] HvsiFileTrust.dll : [NOT_FOUND] ext-ms-win-audiocore-pal-l1-2-0.dll : [NOT_FOUND] ext-ms-win-audiocore-spatial-l1-1-0.dll : [NOT_FOUND] UpdateAPI.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-1.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-0.dll : [NOT_FOUND] AzureAttestManager.dll : [NOT_FOUND] AzureAttestNormal.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-1-2.dll : [NOT_FOUND] ext-ms-win-networking-wlanstorage-l1-1-0.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-1-3.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-1-0.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-2-0.dll : [NOT_FOUND] ext-ms-win-wer-xbox-l1-2-1.dll : [NOT_FOUND] ext-ms-onecore-orientation-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-internal-l1-1-0.dll : [NOT_FOUND] ext-ms-onecore-shellchromeapi-l1-1-0.dll : [NOT_FOUND] ext-ms-win-rtcore-minuser-private-ext-l1-1-3.dll : [NOT_FOUND] ext-ms-win-core-xbrm-l1-1-0.dll : [NOT_FOUND] ext-ms-win-core-xbrm-l1-1-1.dll : [NOT_FOUND] ext-ms-win-ntdsa-activedirectoryserver-l1-1-0.dll : ``` (Which also the command line version shows as not found.) When I install and run Ollama in the Sandbox, it works just fine, but I noticed something important - I thought the Sandbox VM was a copy of my existing Windows 11 installation - which it is not. Winver reports 24H2, while I have 25H2 on my computer. (An no way to go back, without having to reinstall several hundreds programs - some very large and time-consuming like the Adobe family. This is the debug output from Ollama in Sandbox: (Debug = 2) ``` time=2025-11-07T21:56:11.714+01:00 level=INFO source=routes.go:1525 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:C:\\Users\\WDAGUtilityAccount\\.ollama\\models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-11-07T21:56:11.714+01:00 level=INFO source=images.go:522 msg="total blobs: 0" time=2025-11-07T21:56:11.714+01:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-11-07T21:56:11.714+01:00 level=INFO source=routes.go:1578 msg="Listening on 127.0.0.1:11434 (version 0.12.10)" time=2025-11-07T21:56:11.714+01:00 level=DEBUG source=sched.go:120 msg="starting llm scheduler" time=2025-11-07T21:56:11.719+01:00 level=INFO source=runner.go:67 msg="discovering available GPUs..." time=2025-11-07T21:56:11.719+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extraEnvs=map[] time=2025-11-07T21:56:11.743+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 65165" time=2025-11-07T21:56:11.744+01:00 level=DEBUG source=server.go:401 msg=subprocess OLLAMA_DEBUG=2 PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Users\WDAGUtilityAccount\AppData\Local\Microsoft\WindowsApps;;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama OLLAMA_LIBRARY_PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12 time=2025-11-07T21:56:11.827+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine" time=2025-11-07T21:56:11.830+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:65165" time=2025-11-07T21:56:11.839+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string time=2025-11-07T21:56:11.839+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string time=2025-11-07T21:56:11.839+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:56:11.840+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-11-07T21:56:11.841+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-07T21:56:11.841+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama load_backend: loaded CPU backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-07T21:56:11.880+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c load_backend: loaded CUDA backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll time=2025-11-07T21:56:12.938+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) time=2025-11-07T21:56:12.938+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-07T21:56:12.940+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=1.1014927s ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8102727680 total: 12820938752 ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11958046720 total: 12884901888 ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11895758848 total: 12884901888 time=2025-11-07T21:56:13.908+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=967.9884ms time=2025-11-07T21:56:13.910+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:8102727680 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11958046720 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11895758848 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]}]" time=2025-11-07T21:56:13.911+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=2.1922203s OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=map[] time=2025-11-07T21:56:13.911+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[] time=2025-11-07T21:56:13.915+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 65175" time=2025-11-07T21:56:13.915+01:00 level=DEBUG source=server.go:401 msg=subprocess OLLAMA_DEBUG=2 PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Users\WDAGUtilityAccount\AppData\Local\Microsoft\WindowsApps;;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama OLLAMA_LIBRARY_PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13 time=2025-11-07T21:56:14.001+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine" time=2025-11-07T21:56:14.004+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:65175" time=2025-11-07T21:56:14.010+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string time=2025-11-07T21:56:14.010+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string time=2025-11-07T21:56:14.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:56:14.011+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:56:14.011+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-11-07T21:56:14.012+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-11-07T21:56:14.013+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-11-07T21:56:14.013+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-07T21:56:14.013+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama load_backend: loaded CPU backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-07T21:56:14.047+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c load_backend: loaded CUDA backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll time=2025-11-07T21:56:14.969+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-07T21:56:14.970+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-07T21:56:14.971+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=961.571ms ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8097550336 total: 12820938752 ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11958046720 total: 12884901888 ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11895758848 total: 12884901888 time=2025-11-07T21:56:15.759+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=787.6786ms time=2025-11-07T21:56:15.765+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:8097550336 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11958046720 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11895758848 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]}]" time=2025-11-07T21:56:15.766+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=1.8546423s OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[] time=2025-11-07T21:56:15.766+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extraEnvs=map[] time=2025-11-07T21:56:15.770+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 65185" time=2025-11-07T21:56:15.770+01:00 level=DEBUG source=server.go:401 msg=subprocess OLLAMA_DEBUG=2 PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;C:\Users\WDAGUtilityAccount\AppData\Local\Microsoft\WindowsApps;;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama OLLAMA_LIBRARY_PATH=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama;C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm time=2025-11-07T21:56:15.871+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine" time=2025-11-07T21:56:15.874+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:65185" time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-11-07T21:56:15.880+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-11-07T21:56:15.880+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-07T21:56:15.881+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama load_backend: loaded CPU backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-07T21:56:15.917+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm ggml_cuda_init: failed to initialize ROCm: no ROCm-capable device is detected load_backend: loaded ROCm backend from C:\Users\WDAGUtilityAccount\AppData\Local\Programs\Ollama\lib\ollama\rocm\ggml-hip.dll time=2025-11-07T21:56:15.991+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(clang) time=2025-11-07T21:56:15.992+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-07T21:56:15.993+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=116.0823ms time=2025-11-07T21:56:15.994+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s time=2025-11-07T21:56:15.994+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-11-07T21:56:15.996+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=229.7144ms OLLAMA_LIBRARY_PATH="[C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\WDAGUtilityAccount\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=map[] time=2025-11-07T21:56:15.996+01:00 level=DEBUG source=runner.go:113 msg="evluating which if any devices to filter out" initial_count=6 time=2025-11-07T21:56:15.996+01:00 level=TRACE source=runner.go:153 msg="supported GPU library combinations before filtering" supported=map[] time=2025-11-07T21:56:15.996+01:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=4.2818412s time=2025-11-07T21:56:15.996+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced filter_id="" library=CUDA compute=8.6 name=CUDA1 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v12 driver=13.0 pci_id=0000:04:00.0 type=discrete total="12.0 GiB" available="11.1 GiB" time=2025-11-07T21:56:15.996+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c filter_id="" library=CUDA compute=8.6 name=CUDA2 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v12 driver=13.0 pci_id=0000:83:00.0 type=discrete total="12.0 GiB" available="11.1 GiB" time=2025-11-07T21:56:15.996+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 filter_id="" library=CUDA compute=12.0 name=CUDA0 description="NVIDIA GeForce RTX 5070" libdirs=ollama,cuda_v12 driver=13.0 pci_id=0000:03:00.0 type=discrete total="11.9 GiB" available="7.5 GiB" ``` It detects my GPUs and loads models to VRAM. This is the serve.log: [serve.log](https://github.com/user-attachments/files/23425756/serve.log) However, when I install and run Ollama on my computer (outside Sandbox): This is the output: (Debug = 2) ``` time=2025-11-07T21:24:59.289+01:00 level=INFO source=routes.go:1525 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:3 OLLAMA_MODELS:H:\\LMM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:3 OLLAMA_ORIGINS:[*\"; ollama serve http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-11-07T21:24:59.295+01:00 level=INFO source=images.go:522 msg="total blobs: 0" time=2025-11-07T21:24:59.299+01:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-11-07T21:24:59.302+01:00 level=INFO source=routes.go:1578 msg="Listening on 127.0.0.1:11434 (version 0.12.10)" time=2025-11-07T21:24:59.302+01:00 level=DEBUG source=sched.go:120 msg="starting llm scheduler" time=2025-11-07T21:24:59.308+01:00 level=INFO source=runner.go:67 msg="discovering available GPUs..." time=2025-11-07T21:24:59.308+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[] time=2025-11-07T21:24:59.340+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52942" time=2025-11-07T21:24:59.340+01:00 level=DEBUG source=server.go:401 msg=subprocess CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LMM OLLAMA_NUM_PARALLEL=3 OLLAMA_ORIGINS="\"*\"; ollama serve" OLLAMA_SCHED_SPREAD=1 OLLAMA_TMPDIR=H:\LLM\temp PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm\\rocblas\\library" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-11-07T21:24:59.493+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine" time=2025-11-07T21:24:59.497+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:52942" time=2025-11-07T21:24:59.502+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string time=2025-11-07T21:24:59.502+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string time=2025-11-07T21:24:59.502+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-11-07T21:24:59.503+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-07T21:24:59.503+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found. time=2025-11-07T21:24:59.528+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll: The specified procedure could not be found. time=2025-11-07T21:24:59.550+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-11-07T21:24:59.550+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-07T21:24:59.551+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=51.6783ms time=2025-11-07T21:24:59.552+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s time=2025-11-07T21:24:59.554+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[] time=2025-11-07T21:24:59.578+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=269.7244ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[] time=2025-11-07T21:24:59.579+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extraEnvs=map[] time=2025-11-07T21:24:59.583+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52958" time=2025-11-07T21:24:59.588+01:00 level=DEBUG source=server.go:401 msg=subprocess CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LMM OLLAMA_NUM_PARALLEL=3 OLLAMA_ORIGINS="\"*\"; ollama serve" OLLAMA_SCHED_SPREAD=1 OLLAMA_TMPDIR=H:\LLM\temp PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm\\rocblas\\library" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-11-07T21:24:59.750+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine" time=2025-11-07T21:24:59.754+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:52958" time=2025-11-07T21:24:59.766+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string time=2025-11-07T21:24:59.766+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-11-07T21:24:59.767+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-07T21:24:59.767+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found. time=2025-11-07T21:24:59.790+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\ggml-hip.dll: The specified procedure could not be found. time=2025-11-07T21:24:59.824+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-07T21:24:59.825+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=61.7733ms time=2025-11-07T21:24:59.826+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s time=2025-11-07T21:24:59.828+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-11-07T21:24:59.842+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=263.4474ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=map[] time=2025-11-07T21:24:59.844+01:00 level=TRACE source=runner.go:418 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extraEnvs=map[] time=2025-11-07T21:24:59.849+01:00 level=INFO source=server.go:400 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 52703" time=2025-11-07T21:24:59.851+01:00 level=DEBUG source=server.go:401 msg=subprocess CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LMM OLLAMA_NUM_PARALLEL=3 OLLAMA_ORIGINS="\"*\"; ollama serve" OLLAMA_SCHED_SPREAD=1 OLLAMA_TMPDIR=H:\LLM\temp PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm\\rocblas\\library" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-11-07T21:24:59.994+01:00 level=INFO source=runner.go:1349 msg="starting ollama engine" time=2025-11-07T21:24:59.998+01:00 level=INFO source=runner.go:1384 msg="Server listening on 127.0.0.1:52703" time=2025-11-07T21:25:00.009+01:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string time=2025-11-07T21:25:00.009+01:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-11-07T21:25:00.010+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-07T21:25:00.010+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-alderlake.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-icelake.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sandybridge.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-skylakex.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-sse42.dll: The specified procedure could not be found. dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-x64.dll: The specified procedure could not be found. time=2025-11-07T21:25:00.034+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" dl_load_library unable to load library C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll: The specified procedure could not be found. time=2025-11-07T21:25:00.077+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-07T21:25:00.078+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=runner.go:1324 msg="dummy model load took" duration=70.9579ms time=2025-11-07T21:25:00.079+01:00 level=DEBUG source=runner.go:1329 msg="gathering device infos took" duration=0s time=2025-11-07T21:25:00.081+01:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[] time=2025-11-07T21:25:00.095+01:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=251.0928ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=map[] time=2025-11-07T21:25:00.096+01:00 level=DEBUG source=runner.go:113 msg="evluating which if any devices to filter out" initial_count=0 time=2025-11-07T21:25:00.097+01:00 level=TRACE source=runner.go:153 msg="supported GPU library combinations before filtering" supported=map[] time=2025-11-07T21:25:00.098+01:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=794.929ms time=2025-11-07T21:25:00.100+01:00 level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="101.3 GiB" time=2025-11-07T21:25:00.100+01:00 level=INFO source=routes.go:1619 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ``` And this is the serve.log: [serve.log](https://github.com/user-attachments/files/23425798/serve.log) I've tried everything I could think of - I've added Ollama LIB folders to PATH, I've copied the DLL files reported missing with Dependencies.exe/DependenciesGUI.exe from other x64 software to the LIB folder... I don't know what to do anymore? 😖😭😭😭
Author
Owner

@Anyeos commented on GitHub (Nov 12, 2025):

Update:

I missed to build the "build" subdirectory part. I issued cmake -B build and cmake --build build and now it is working.
Maybe @catclaw you are forgeting the libraries folder or you are putting them on another place. It must be in "./lib/ollama" path (as indicated by the documentation). That will provide you GPU and CPU acceleration.

My original post:

Same happens to me and I build it from source. The old version detect my GPU automatically.
And I am under Ubuntu 24.04
It can be a bug in source code or some dependency that directly is skipped.
I use go run . serve

I tried go build again but nothing changed.

time=2025-11-12T03:49:56.979-03:00 level=INFO source=runner.go:1398 msg="starting ollama engine"
time=2025-11-12T03:49:56.979-03:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:42279"
time=2025-11-12T03:49:56.980-03:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string
time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string
time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32
time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0
time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default=""
time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default=""
time=2025-11-12T03:49:56.981-03:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/mnt/InteligenciaArtificial/LenguajeNatural/Ollama/ollama/build/lib/ollama
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
  Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes, ID: GPU-7c0e39f9-9204-ef46-3c23-511371e6c901
time=2025-11-12T03:49:57.159-03:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc)
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=178.878976ms
time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=540ns
time=2025-11-12T03:49:57.160-03:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/mnt/InteligenciaArtificial/LenguajeNatural/Ollama/ollama/build/lib/ollama] devices=[]
time=2025-11-12T03:49:57.160-03:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=190.600957ms OLLAMA_LIBRARY_PATH=[/mnt/InteligenciaArtificial/LenguajeNatural/Ollama/ollama/build/lib/ollama] extra_envs=map[]
time=2025-11-12T03:49:57.160-03:00 level=DEBUG source=runner.go:113 msg="evluating which if any devices to filter out" initial_count=0
time=2025-11-12T03:49:57.160-03:00 level=TRACE source=runner.go:153 msg="supported GPU library combinations before filtering" supported=map[]
time=2025-11-12T03:49:57.160-03:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=190.792858ms
time=2025-11-12T03:49:57.160-03:00 level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="47.0 GiB" available="38.6 GiB"
time=2025-11-12T03:49:57.160-03:00 level=INFO source=routes.go:1638 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

Call my attention that it says "total vram"="0 B" <- Why? What is happening there?

<!-- gh-comment-id:3520327051 --> @Anyeos commented on GitHub (Nov 12, 2025): ### Update: I missed to build the "build" subdirectory part. I issued cmake -B build and cmake --build build and now it is working. Maybe @catclaw you are forgeting the libraries folder or you are putting them on another place. It must be in "./lib/ollama" path (as indicated by the documentation). That will provide you GPU and CPU acceleration. ### My original post: Same happens to me and I build it from source. The old version detect my GPU automatically. And I am under Ubuntu 24.04 It can be a bug in source code or some dependency that directly is skipped. I use go run . serve I tried go build again but nothing changed. ``` time=2025-11-12T03:49:56.979-03:00 level=INFO source=runner.go:1398 msg="starting ollama engine" time=2025-11-12T03:49:56.979-03:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:42279" time=2025-11-12T03:49:56.980-03:00 level=DEBUG source=gguf.go:590 msg=general.architecture type=string time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=gguf.go:590 msg=tokenizer.ggml.model type=string time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-11-12T03:49:56.981-03:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-12T03:49:56.981-03:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/mnt/InteligenciaArtificial/LenguajeNatural/Ollama/ollama/build/lib/ollama ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes, ID: GPU-7c0e39f9-9204-ef46-3c23-511371e6c901 time=2025-11-12T03:49:57.159-03:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc) time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=178.878976ms time=2025-11-12T03:49:57.159-03:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=540ns time=2025-11-12T03:49:57.160-03:00 level=TRACE source=runner.go:445 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/mnt/InteligenciaArtificial/LenguajeNatural/Ollama/ollama/build/lib/ollama] devices=[] time=2025-11-12T03:49:57.160-03:00 level=DEBUG source=runner.go:415 msg="bootstrap discovery took" duration=190.600957ms OLLAMA_LIBRARY_PATH=[/mnt/InteligenciaArtificial/LenguajeNatural/Ollama/ollama/build/lib/ollama] extra_envs=map[] time=2025-11-12T03:49:57.160-03:00 level=DEBUG source=runner.go:113 msg="evluating which if any devices to filter out" initial_count=0 time=2025-11-12T03:49:57.160-03:00 level=TRACE source=runner.go:153 msg="supported GPU library combinations before filtering" supported=map[] time=2025-11-12T03:49:57.160-03:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=190.792858ms time=2025-11-12T03:49:57.160-03:00 level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="47.0 GiB" available="38.6 GiB" time=2025-11-12T03:49:57.160-03:00 level=INFO source=routes.go:1638 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ``` Call my attention that it says "total vram"="0 B" <- Why? What is happening there?
Author
Owner

@VSOL3008 commented on GitHub (Nov 12, 2025):

I had the same problem,
for some reason Ollama 0.12.4 have some bug with the environment variable OLLAMA_LLM_LIBRARY=CUDA, basically its causing an error preventing the detection of the multiple GPUs. But this variable was working in older versions of Ollama, now in new version it probably have different behavior.
So what helped for me, was just removing this OLLAMA_LLM_LIBRARY variable.
I hope it helps.

<!-- gh-comment-id:3520649969 --> @VSOL3008 commented on GitHub (Nov 12, 2025): I had the same problem, for some reason Ollama 0.12.4 have some bug with the environment variable OLLAMA_LLM_LIBRARY=CUDA, basically its causing an error preventing the detection of the multiple GPUs. But this variable was working in older versions of Ollama, now in new version it probably have different behavior. **So what helped for me, was just removing this OLLAMA_LLM_LIBRARY variable.** I hope it helps.
Author
Owner

@dhiltgen commented on GitHub (Nov 12, 2025):

@catclaw so far, I haven't seen anyone else report the error The specified procedure could not be found causing all of the libraries not to load. Given you have development tools installed on your system, perhaps you could try to build from source and see if the problem disappears? My best guess at this point is your system DLLs are somehow different than the standard ones we link against in our CI.

<!-- gh-comment-id:3522663840 --> @dhiltgen commented on GitHub (Nov 12, 2025): @catclaw so far, I haven't seen anyone else report the error `The specified procedure could not be found` causing all of the libraries not to load. Given you have development tools installed on your system, perhaps you could try to build from source and see if the problem disappears? My best guess at this point is your system DLLs are somehow different than the standard ones we link against in our CI.
Author
Owner

@catclaw commented on GitHub (Nov 13, 2025):

I've created a clean install of Windows 11 (Enterprise) ver. 24H2, on a new computer, and moved my GPUs over to the new motherboard.
Ollama worked fine - till I updated to 25H2, and the same problem occurred once again.
I've been diving into some MS technical data, and Nvidia developers forum - I see "CUDA context issue introduced in the 25H2 driver layer" repeating everywhere.
I'm not entirely sure what it means, but it seems like MS has made some low-layer changes to how Windows handles GPU/CUDA drivers.
I'm currently running Ollama (latest version) on WSL2 (Ubuntu), with the following in my .bashrc
export OLLAMA_MODELS=/mnt/h/LLM
export CUDA_VISIBLE_DEVICES=0,1,2 #3 is a PNY Nvidia Tesla A100 80GB, which I use in ComfyUI, while using Ollama for writing my prompts.
export OLLAMA_SCHED_SPREAD=1
export OLLAMA_NUM_PARALLEL=3
export OLLAMA_MAX_QUEUE=3
export OLLAMA_NO_MMAP=1
export OLLAMA_LOAD_TIMEOUT=10m
export OLLAMA_DEBUG=2

However, I can't use localhost on Windows anymore to reach Ollama, and have to type my LAN IP.
If anyone is going to create a similar setup, make sure to include OLLAMA_NO_MMAP=1 as ext4 fs is faster than NTFS, and without it, you might run into trouble loading larger models.

That said...
I hope you find the solution to this problem, as WSL uses a lot of my system resources and RAM, slowing down video generation for example.

<!-- gh-comment-id:3530015529 --> @catclaw commented on GitHub (Nov 13, 2025): I've created a clean install of Windows 11 (Enterprise) ver. 24H2, on a new computer, and moved my GPUs over to the new motherboard. Ollama worked fine - till I updated to 25H2, and the same problem occurred once again. I've been diving into some MS technical data, and Nvidia developers forum - I see "CUDA context issue introduced in the 25H2 driver layer" repeating everywhere. I'm not entirely sure what it means, but it seems like MS has made some low-layer changes to how Windows handles GPU/CUDA drivers. I'm currently running Ollama (latest version) on WSL2 (Ubuntu), with the following in my .bashrc export OLLAMA_MODELS=/mnt/h/LLM export CUDA_VISIBLE_DEVICES=0,1,2 #3 is a PNY Nvidia Tesla A100 80GB, which I use in ComfyUI, while using Ollama for writing my prompts. export OLLAMA_SCHED_SPREAD=1 export OLLAMA_NUM_PARALLEL=3 export OLLAMA_MAX_QUEUE=3 export OLLAMA_NO_MMAP=1 export OLLAMA_LOAD_TIMEOUT=10m export OLLAMA_DEBUG=2 However, I can't use localhost on Windows anymore to reach Ollama, and have to type my LAN IP. If anyone is going to create a similar setup, make sure to include OLLAMA_NO_MMAP=1 as ext4 fs is faster than NTFS, and without it, you might run into trouble loading larger models. That said... I hope you find the solution to this problem, as WSL uses a lot of my system resources and RAM, slowing down video generation for example.
Author
Owner

@dhiltgen commented on GitHub (Nov 13, 2025):

At least that narrows it down a bit.

I don't think it's related to CUDA, since the logs you shared showed even the CPU libraries failed to load. It appears there's something about the way we're building the windows binaries which makes them incompatible with Windows Enterprise 25H2. We have Windows Pro 25H2 test systems that all seem to work OK. In our CI, we build on Windows Server github runners, and the compiler used for the CPU libraries is C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/Llvm/x64/bin/clang.exe which I would have thought would produce compatible binaries.

I still think it would be interesting to see if locally built binaries work on your system if you have a chance to try that.

<!-- gh-comment-id:3530136201 --> @dhiltgen commented on GitHub (Nov 13, 2025): At least that narrows it down a bit. I don't think it's related to CUDA, since the logs you shared showed even the CPU libraries failed to load. It appears there's something about the way we're building the windows binaries which makes them incompatible with Windows Enterprise 25H2. We have Windows Pro 25H2 test systems that all seem to work OK. In our CI, we build on Windows Server github runners, and the compiler used for the CPU libraries is `C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/Llvm/x64/bin/clang.exe` which I would have thought would produce compatible binaries. I still think it would be interesting to see if locally built binaries work on your system if you have a chance to try that.
Author
Owner

@dhiltgen commented on GitHub (Nov 14, 2025):

I just spun up a Windows 11 Enterprise 25H2 VM on Azure, and it is able to load the libraries without error. @catclaw could there be some other piece of software intervening, like AV software on your system? If you have 3rd party AV software, could you try disabling that temporarily and see if the libraries load correctly? On my test VM, I see MsMpEng.exe chewing up a ton of CPU cycles as we're doing our initial bootstrapping, but it appears to be letting the libraries load. All the DLLs look correctly signed by our signing key.

If you are in a corporate environment and can't disable the AV software, can you check with your IT to see if there are logs showing it is blocking the libraries from loading?

<!-- gh-comment-id:3534727005 --> @dhiltgen commented on GitHub (Nov 14, 2025): I just spun up a Windows 11 Enterprise 25H2 VM on Azure, and it is able to load the libraries without error. @catclaw could there be some other piece of software intervening, like AV software on your system? If you have 3rd party AV software, could you try disabling that temporarily and see if the libraries load correctly? On my test VM, I see MsMpEng.exe chewing up a ton of CPU cycles as we're doing our initial bootstrapping, but it appears to be letting the libraries load. All the DLLs look correctly signed by our signing key. If you are in a corporate environment and can't disable the AV software, can you check with your IT to see if there are logs showing it is blocking the libraries from loading?
Author
Owner

@catclaw commented on GitHub (Nov 15, 2025):

I just spun up a Windows 11 Enterprise 25H2 VM on Azure, and it is able to load the libraries without error. @catclaw could there be some other piece of software intervening, like AV software on your system? If you have 3rd party AV software, could you try disabling that temporarily and see if the libraries load correctly? On my test VM, I see MsMpEng.exe chewing up a ton of CPU cycles as we're doing our initial bootstrapping, but it appears to be letting the libraries load. All the DLLs look correctly signed by our signing key.

If you are in a corporate environment and can't disable the AV software, can you check with your IT to see if there are logs showing it is blocking the libraries from loading?

Hi!
Thank you for taking the time to test it on a VM. I really appreciate it.
Perhaps the updates have something to do with the libs not being loaded?
My firmware was also updated from 2006 to 2009.
Image

Image

No, I'm not in a corporate environment - it's my own personal computer. However, I had some trouble using PNY Nvidia Tesla A100 80GB on the Pro-edition (I could either use my 3 RTX GPUs or A100 - not all 4 at the same time), so I contacted both PNY and Microsoft support. They both suggested that I use a server edition of Windows, or the Enterprise edition, which I did, and now I can use all 4 GPUs at the same time. (I never got an explanation for why it would matter?)

Yes, I installed a bare Windows 11 Enterprise edition, without any AV software, or anything else.
The Ollama setup EXE file installed vc_redist.x64.exe, but nothing else.
Despite this, Ollama fails to detect my GPUs.
However, I installed Windows 11 Enterprise 24H2 (with all updates, except 25H2) - and Ollama worked just fine; it detected my GPUs and loaded models to VRAM.
(Yes, MsMpEng.exe spins up pretty good during the initial loading of a model, but it's a part of Windows Security - so I guess it's screening for something.)
As I mentioned earlier, some people are reporting "CUDA context issue introduced in the 25H2 driver layer" related to software that uses CUDA 13.0.x and Win11 25H2 - but I have no idea what it's about?

I'm going to try WinDbg and step through the libs loading and see if I can find anything.

<!-- gh-comment-id:3536825671 --> @catclaw commented on GitHub (Nov 15, 2025): > I just spun up a Windows 11 Enterprise 25H2 VM on Azure, and it is able to load the libraries without error. [@catclaw](https://github.com/catclaw) could there be some other piece of software intervening, like AV software on your system? If you have 3rd party AV software, could you try disabling that temporarily and see if the libraries load correctly? On my test VM, I see MsMpEng.exe chewing up a ton of CPU cycles as we're doing our initial bootstrapping, but it appears to be letting the libraries load. All the DLLs look correctly signed by our signing key. > > If you are in a corporate environment and can't disable the AV software, can you check with your IT to see if there are logs showing it is blocking the libraries from loading? Hi! Thank you for taking the time to test it on a VM. I really appreciate it. Perhaps the updates have something to do with the libs not being loaded? My firmware was also updated from 2006 to 2009. <img width="847" height="758" alt="Image" src="https://github.com/user-attachments/assets/fab8f9c2-d396-4b31-8d4e-db295438201b" /> <img width="847" height="309" alt="Image" src="https://github.com/user-attachments/assets/25dd4d71-d0fa-4b18-a847-6f23b882fbbf" /> No, I'm not in a corporate environment - it's my own personal computer. However, I had some trouble using PNY Nvidia Tesla A100 80GB on the Pro-edition (I could either use my 3 RTX GPUs or A100 - not all 4 at the same time), so I contacted both PNY and Microsoft support. They both suggested that I use a server edition of Windows, or the Enterprise edition, which I did, and now I can use all 4 GPUs at the same time. (I never got an explanation for why it would matter?) Yes, I installed a bare Windows 11 Enterprise edition, without any AV software, or anything else. The Ollama setup EXE file installed vc_redist.x64.exe, but nothing else. Despite this, Ollama fails to detect my GPUs. However, I installed Windows 11 Enterprise 24H2 (with all updates, except 25H2) - and Ollama worked just fine; it detected my GPUs and loaded models to VRAM. (Yes, MsMpEng.exe spins up pretty good during the initial loading of a model, but it's a part of Windows Security - so I guess it's screening for something.) As I mentioned earlier, some people are reporting "CUDA context issue introduced in the 25H2 driver layer" related to software that uses CUDA 13.0.x and Win11 25H2 - but I have no idea what it's about? I'm going to try WinDbg and step through the libs loading and see if I can find anything.
Author
Owner

@dhiltgen commented on GitHub (Nov 15, 2025):

I'm going to try WinDbg and step through the libs loading and see if I can find anything.

I was hoping I could repro and then I was going to try gflags -i ollama.exe +sls and try running it under a debugger, which apparently will show more detail on LoadLibrary failures. One wrinkle to be aware of is we use subprocesses, so you'll have to get the debugger to follow child processes. (The initial ollama serve spawns another copy of ollama which is what calls LoadLibrary). You can look at the logs to see the commands it's running and try running the runner subprocess manually.

<!-- gh-comment-id:3536901522 --> @dhiltgen commented on GitHub (Nov 15, 2025): > I'm going to try WinDbg and step through the libs loading and see if I can find anything. I was hoping I could repro and then I was going to try `gflags -i ollama.exe +sls` and try running it under a debugger, which apparently will show more detail on LoadLibrary failures. One wrinkle to be aware of is we use subprocesses, so you'll have to get the debugger to follow child processes. (The initial `ollama serve` spawns another copy of `ollama` which is what calls LoadLibrary). You can look at the logs to see the commands it's running and try running the runner subprocess manually.
Author
Owner

@dhiltgen commented on GitHub (Nov 17, 2025):

@catclaw based on this comment a possible explanation is another version of ggml-base.dll someplace on your system that's conflicting. Can you look around to see if you have another copy of this anywhere in the PATH? Try removing it and see if that resolves the problem.

<!-- gh-comment-id:3543718664 --> @dhiltgen commented on GitHub (Nov 17, 2025): @catclaw based on [this comment](https://github.com/ollama/ollama/issues/12672#issuecomment-3537970076) a possible explanation is another version of `ggml-base.dll` someplace on your system that's conflicting. Can you look around to see if you have another copy of this anywhere in the PATH? Try removing it and see if that resolves the problem.
Author
Owner

@catclaw commented on GitHub (Nov 22, 2025):

Well... The new 0.13.0 version finally detects my GPUs, but still loads models into RAM and uses the CPU. (~70-80% - please se attached image at the end of this post.)
(No, there are no other ggml-base.dll in PATH; ComfyUI, SwarmUI, Whisper, etc. are all in their own folders outside the path.)
Also, when I run ollama serve, it seems like my Nvidia drivers crash and restart, as all my 4 monitors go black, and the same thing happens as when I restart the drivers manually by pressing CTRL+SHIFT+WIN+B.

time=2025-11-22T20:56:39.151+01:00 level=INFO source=routes.go:1544 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:3 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:3 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true OLLAMA_VULKAN:true ROCR_VISIBLE_DEVICES:]"
time=2025-11-22T20:56:39.431+01:00 level=INFO source=images.go:522 msg="total blobs: 146"
time=2025-11-22T20:56:39.449+01:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0"
time=2025-11-22T20:56:39.463+01:00 level=INFO source=routes.go:1597 msg="Listening on 127.0.0.1:11434 (version 0.13.0)"
time=2025-11-22T20:56:39.464+01:00 level=DEBUG source=sched.go:120 msg="starting llm scheduler"
time=2025-11-22T20:56:39.471+01:00 level=INFO source=runner.go:67 msg="discovering available GPUs..."
time=2025-11-22T20:56:39.471+01:00 level=WARN source=runner.go:470 msg="user overrode visible devices" CUDA_VISIBLE_DEVICES=0,1,2
time=2025-11-22T20:56:39.472+01:00 level=WARN source=runner.go:474 msg="if GPUs are not correctly discovered, unset and try again"
time=2025-11-22T20:56:39.472+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[]
time=2025-11-22T20:56:39.520+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 58193"
time=2025-11-22T20:56:39.521+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-11-22T20:56:39.718+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine"
time=2025-11-22T20:56:39.722+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:58193"
time=2025-11-22T20:56:39.734+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string
time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string
time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0
time=2025-11-22T20:56:39.736+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default=""
time=2025-11-22T20:56:39.736+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default=""
time=2025-11-22T20:56:39.737+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-22T20:56:39.737+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-22T20:56:39.788+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 3 CUDA devices:
  Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8
  Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced
  Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c
load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll
time=2025-11-22T20:56:40.050+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)
time=2025-11-22T20:56:40.051+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=320.1729ms
ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 9020006400 total: 12820938752
ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11452166144 total: 12884901888
ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11689582592 total: 12884901888
time=2025-11-22T20:56:40.458+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=405.3281ms
time=2025-11-22T20:56:40.461+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:9020006400 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11452166144 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11689582592 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]}]"
time=2025-11-22T20:56:40.466+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=994.0386ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[]
time=2025-11-22T20:56:40.485+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extraEnvs=map[]
time=2025-11-22T20:56:40.491+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 57842"
time=2025-11-22T20:56:40.493+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm"
time=2025-11-22T20:56:40.704+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine"
time=2025-11-22T20:56:40.708+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:57842"
time=2025-11-22T20:56:40.714+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string
time=2025-11-22T20:56:40.714+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string
time=2025-11-22T20:56:40.715+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:40.715+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0
time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default=""
time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default=""
time=2025-11-22T20:56:40.716+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-22T20:56:40.764+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm"
ggml_cuda_init: failed to initialize ROCm: no ROCm-capable device is detected
load_backend: loaded ROCm backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\ggml-hip.dll
time=2025-11-22T20:56:40.834+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(clang)
time=2025-11-22T20:56:40.834+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=122.4886ms
time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=0s
time=2025-11-22T20:56:40.838+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[]
time=2025-11-22T20:56:40.859+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=374.0745ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=map[]
time=2025-11-22T20:56:40.879+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]" extraEnvs=map[]
time=2025-11-22T20:56:40.886+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 58909"
time=2025-11-22T20:56:40.901+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan"
time=2025-11-22T20:56:41.103+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine"
time=2025-11-22T20:56:41.107+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:58909"
time=2025-11-22T20:56:41.116+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string
time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string
time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0
time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default=""
time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default=""
time=2025-11-22T20:56:41.117+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-22T20:56:41.118+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-22T20:56:41.165+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan"
ggml_vulkan: Found 3 Vulkan devices:
ggml_vulkan: 0 = NVIDIA GeForce RTX 5070 (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2
ggml_vulkan: 1 = NVIDIA GeForce RTX 3060 (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2
ggml_vulkan: 2 = NVIDIA GeForce RTX 3060 (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2
load_backend: loaded Vulkan backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\vulkan\ggml-vulkan.dll
time=2025-11-22T20:56:41.749+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(clang)
time=2025-11-22T20:56:42.164+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
ggml_backend_vk_get_device_memory called: uuid 7c7493fb-894e-ffa3-0d90-00b83a5731e8
ggml_backend_vk_get_device_memory called: luid 0x00000000000814dc
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 5070) with LUID 0x00000000000814dc detected. Dedicated Total: 12430868480.00 bytes (11.58 GB), Dedicated Usage: 3487313920.00 bytes (3.25 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 8943554560 total: 12430868480
ggml_backend_vk_get_device_memory called: uuid 67bf118a-52b6-0f51-c4ad-c561cdfbb20c
ggml_backend_vk_get_device_memory called: luid 0x0000000000084879
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000084879 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 917102592.00 bytes (0.85 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11786395648 total: 12703498240
ggml_backend_vk_get_device_memory called: uuid e2200bae-da5e-037d-9fc2-d0d8c6081ced
ggml_backend_vk_get_device_memory called: luid 0x0000000000083770
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000083770 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 1154629632.00 bytes (1.08 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11548868608 total: 12703498240
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=2.2566731s
ggml_backend_vk_get_device_memory called: uuid 7c7493fb-894e-ffa3-0d90-00b83a5731e8
ggml_backend_vk_get_device_memory called: luid 0x00000000000814dc
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 5070) with LUID 0x00000000000814dc detected. Dedicated Total: 12430868480.00 bytes (11.58 GB), Dedicated Usage: 3439132672.00 bytes (3.20 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 8991735808 total: 12430868480
ggml_backend_vk_get_device_memory called: uuid 7c7493fb-894e-ffa3-0d90-00b83a5731e8
ggml_backend_vk_get_device_memory called: luid 0x00000000000814dc
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 5070) with LUID 0x00000000000814dc detected. Dedicated Total: 12430868480.00 bytes (11.58 GB), Dedicated Usage: 3439132672.00 bytes (3.20 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 8991735808 total: 12430868480
ggml_backend_vk_get_device_memory called: uuid 67bf118a-52b6-0f51-c4ad-c561cdfbb20c
ggml_backend_vk_get_device_memory called: luid 0x0000000000084879
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000084879 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 917102592.00 bytes (0.85 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11786395648 total: 12703498240
ggml_backend_vk_get_device_memory called: uuid 67bf118a-52b6-0f51-c4ad-c561cdfbb20c
ggml_backend_vk_get_device_memory called: luid 0x0000000000084879
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000084879 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 917102592.00 bytes (0.85 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11786395648 total: 12703498240
ggml_backend_vk_get_device_memory called: uuid e2200bae-da5e-037d-9fc2-d0d8c6081ced
ggml_backend_vk_get_device_memory called: luid 0x0000000000083770
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000083770 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 1154629632.00 bytes (1.08 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11548868608 total: 12703498240
ggml_backend_vk_get_device_memory called: uuid e2200bae-da5e-037d-9fc2-d0d8c6081ced
ggml_backend_vk_get_device_memory called: luid 0x0000000000083770
ggml_dxgi_pdh_init called
DXGI + PDH Initialized. Getting GPU free memory info
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB
[DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB
[DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB
Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000083770 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 1154629632.00 bytes (1.08 GB)
ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11548868608 total: 12703498240
time=2025-11-22T20:56:44.143+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=771.8209ms
time=2025-11-22T20:56:44.145+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]" devices="[{DeviceID:{ID:7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:Vulkan} Name:Vulkan0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12430868480 FreeMemory:8991735808 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]} {DeviceID:{ID:67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:Vulkan} Name:Vulkan1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12703498240 FreeMemory:11786395648 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]} {DeviceID:{ID:e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:Vulkan} Name:Vulkan2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12703498240 FreeMemory:11548868608 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]}]"
time=2025-11-22T20:56:44.153+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=3.2735283s OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]" extra_envs=map[]
time=2025-11-22T20:56:44.242+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extraEnvs=map[]
time=2025-11-22T20:56:44.253+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 57435"
time=2025-11-22T20:56:44.254+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
time=2025-11-22T20:56:44.443+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine"
time=2025-11-22T20:56:44.448+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:57435"
time=2025-11-22T20:56:44.455+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string
time=2025-11-22T20:56:44.455+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string
time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0
time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default=""
time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default=""
time=2025-11-22T20:56:44.456+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-22T20:56:44.502+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12"
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 3 CUDA devices:
  Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8
  Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced
  Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c
load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll
time=2025-11-22T20:56:44.734+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=281.9024ms
ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 9015746560 total: 12820938752
ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11452166144 total: 12884901888
ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11689582592 total: 12884901888
time=2025-11-22T20:56:45.197+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=461.3986ms
time=2025-11-22T20:56:45.199+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:9015746560 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11452166144 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11689582592 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]}]"
time=2025-11-22T20:56:45.206+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=964.5419ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=map[]
time=2025-11-22T20:56:45.229+01:00 level=DEBUG source=runner.go:120 msg="evaluating which, if any, devices to filter out" initial_count=9
time=2025-11-22T20:56:45.231+01:00 level=TRACE source=runner.go:160 msg="supported GPU library combinations before filtering" supported=map[]
time=2025-11-22T20:56:45.232+01:00 level=DEBUG source=runner.go:213 msg="dropping duplicate device" id=7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=Vulkan compute=0.0 name=Vulkan0 description="NVIDIA GeForce RTX 5070" libdirs="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama,C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" driver=0.0 pci_id=0000:03:00.0 type=discrete total="11.6 GiB" available="8.4 GiB"
time=2025-11-22T20:56:45.234+01:00 level=DEBUG source=runner.go:213 msg="dropping duplicate device" id=e2200bae-da5e-037d-9fc2-d0d8c6081ced library=Vulkan compute=0.0 name=Vulkan2 description="NVIDIA GeForce RTX 3060" libdirs="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama,C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" driver=0.0 pci_id=0000:04:00.0 type=discrete total="11.8 GiB" available="10.8 GiB"
time=2025-11-22T20:56:45.235+01:00 level=DEBUG source=runner.go:213 msg="dropping duplicate device" id=67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=Vulkan compute=0.0 name=Vulkan1 description="NVIDIA GeForce RTX 3060" libdirs="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama,C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" driver=0.0 pci_id=0000:83:00.0 type=discrete total="11.8 GiB" available="11.0 GiB"
time=2025-11-22T20:56:45.242+01:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=5.7781411s
time=2025-11-22T20:56:45.244+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c filter_id="" library=CUDA compute=8.6 name=CUDA2 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v13 driver=13.0 pci_id=0000:83:00.0 type=discrete total="12.0 GiB" available="10.9 GiB"
time=2025-11-22T20:56:45.245+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced filter_id="" library=CUDA compute=8.6 name=CUDA1 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v13 driver=13.0 pci_id=0000:04:00.0 type=discrete total="12.0 GiB" available="10.7 GiB"
time=2025-11-22T20:56:45.246+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 filter_id="" library=CUDA compute=12.0 name=CUDA0 description="NVIDIA GeForce RTX 5070" libdirs=ollama,cuda_v13 driver=13.0 pci_id=0000:03:00.0 type=discrete total="11.9 GiB" available="8.4 GiB"
[GIN] 2025/11/22 - 20:56:57 | 200 |            0s |       127.0.0.1 | HEAD     "/"
[GIN] 2025/11/22 - 20:56:57 | 200 |     44.7027ms |       127.0.0.1 | GET      "/api/tags"
[GIN] 2025/11/22 - 21:15:48 | 200 |     28.5907ms |       127.0.0.1 | GET      "/api/tags"
[GIN] 2025/11/22 - 21:15:49 | 200 |            0s |       127.0.0.1 | GET      "/api/ps"
[GIN] 2025/11/22 - 21:15:53 | 200 |     26.5895ms |       127.0.0.1 | GET      "/api/tags"
[GIN] 2025/11/22 - 21:15:53 | 200 |            0s |       127.0.0.1 | GET      "/api/ps"
time=2025-11-22T21:16:51.500+01:00 level=TRACE source=sched.go:146 msg="processing incoming request" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9
time=2025-11-22T21:16:51.501+01:00 level=TRACE source=sched.go:179 msg="refreshing GPU list" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9
time=2025-11-22T21:16:51.505+01:00 level=DEBUG source=runner.go:250 msg="refreshing free memory"
time=2025-11-22T21:16:51.506+01:00 level=DEBUG source=runner.go:314 msg="unable to refresh all GPUs with existing runners, performing bootstrap discovery"
time=2025-11-22T21:16:51.507+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[]
time=2025-11-22T21:16:51.547+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 56402"
time=2025-11-22T21:16:51.549+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-11-22T21:16:51.766+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine"
time=2025-11-22T21:16:51.771+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:56402"
time=2025-11-22T21:16:51.777+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string
time=2025-11-22T21:16:51.778+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string
time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0
time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default=""
time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default=""
time=2025-11-22T21:16:51.779+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-22T21:16:51.830+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 3 CUDA devices:
  Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8
  Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced
  Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c
load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll
time=2025-11-22T21:16:52.124+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)
time=2025-11-22T21:16:52.124+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default=""
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0
time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
time=2025-11-22T21:16:52.126+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000
time=2025-11-22T21:16:52.126+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1
time=2025-11-22T21:16:52.126+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=349.5975ms
ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8893997056 total: 12820938752
ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11387678720 total: 12884901888
ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11551105024 total: 12884901888
time=2025-11-22T21:16:52.530+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=403.6765ms
time=2025-11-22T21:16:52.533+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:8893997056 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11387678720 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11551105024 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]}]"
time=2025-11-22T21:16:52.540+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=1.0331979s OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[]
time=2025-11-22T21:16:52.563+01:00 level=DEBUG source=runner.go:40 msg="overall device VRAM discovery took" duration=1.0581286s
time=2025-11-22T21:16:52.565+01:00 level=TRACE source=sched.go:182 msg="refreshing system information" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9
time=2025-11-22T21:16:52.566+01:00 level=TRACE source=gpu.go:22 msg="performing CPU discovery"
time=2025-11-22T21:16:52.569+01:00 level=INFO source=cpu_windows.go:148 msg=packages count=2
time=2025-11-22T21:16:52.571+01:00 level=INFO source=cpu_windows.go:195 msg="" package=0 cores=14 efficiency=0 threads=28
time=2025-11-22T21:16:52.573+01:00 level=INFO source=cpu_windows.go:195 msg="" package=1 cores=14 efficiency=0 threads=28
time=2025-11-22T21:16:52.574+01:00 level=TRACE source=gpu.go:25 msg="CPU discovery completed" duration=4.9149ms
time=2025-11-22T21:16:52.575+01:00 level=DEBUG source=sched.go:194 msg="updating default concurrency" OLLAMA_MAX_LOADED_MODELS=9 gpu_count=3
time=2025-11-22T21:16:52.577+01:00 level=TRACE source=sched.go:198 msg="loading model metadata" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9
time=2025-11-22T21:16:52.604+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32
time=2025-11-22T21:16:52.617+01:00 level=TRACE source=sched.go:206 msg="updating free space" gpu_count=3 model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9
time=2025-11-22T21:16:52.637+01:00 level=DEBUG source=sched.go:211 msg="loading first model" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9
llama_model_loader: loaded meta data with 34 key-value pairs and 771 tensors from H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv   0:                       general.architecture str              = qwen2
llama_model_loader: - kv   1:                               general.type str              = model
llama_model_loader: - kv   2:                               general.name str              = Qwen2.5 Coder 32B Instruct
llama_model_loader: - kv   3:                           general.finetune str              = Instruct
llama_model_loader: - kv   4:                           general.basename str              = Qwen2.5-Coder
llama_model_loader: - kv   5:                         general.size_label str              = 32B
llama_model_loader: - kv   6:                            general.license str              = apache-2.0
llama_model_loader: - kv   7:                       general.license.link str              = https://huggingface.co/Qwen/Qwen2.5-C...
llama_model_loader: - kv   8:                   general.base_model.count u32              = 1
llama_model_loader: - kv   9:                  general.base_model.0.name str              = Qwen2.5 Coder 32B
llama_model_loader: - kv  10:          general.base_model.0.organization str              = Qwen
llama_model_loader: - kv  11:              general.base_model.0.repo_url str              = https://huggingface.co/Qwen/Qwen2.5-C...
llama_model_loader: - kv  12:                               general.tags arr[str,6]       = ["code", "codeqwen", "chat", "qwen", ...
llama_model_loader: - kv  13:                          general.languages arr[str,1]       = ["en"]
llama_model_loader: - kv  14:                          qwen2.block_count u32              = 64
llama_model_loader: - kv  15:                       qwen2.context_length u32              = 32768
llama_model_loader: - kv  16:                     qwen2.embedding_length u32              = 5120
llama_model_loader: - kv  17:                  qwen2.feed_forward_length u32              = 27648
llama_model_loader: - kv  18:                 qwen2.attention.head_count u32              = 40
llama_model_loader: - kv  19:              qwen2.attention.head_count_kv u32              = 8
llama_model_loader: - kv  20:                       qwen2.rope.freq_base f32              = 1000000.000000
llama_model_loader: - kv  21:     qwen2.attention.layer_norm_rms_epsilon f32              = 0.000001
llama_model_loader: - kv  22:                          general.file_type u32              = 15
llama_model_loader: - kv  23:                       tokenizer.ggml.model str              = gpt2
llama_model_loader: - kv  24:                         tokenizer.ggml.pre str              = qwen2
llama_model_loader: - kv  25:                      tokenizer.ggml.tokens arr[str,152064]  = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv  26:                  tokenizer.ggml.token_type arr[i32,152064]  = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv  27:                      tokenizer.ggml.merges arr[str,151387]  = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",...
llama_model_loader: - kv  28:                tokenizer.ggml.eos_token_id u32              = 151645
llama_model_loader: - kv  29:            tokenizer.ggml.padding_token_id u32              = 151643
llama_model_loader: - kv  30:                tokenizer.ggml.bos_token_id u32              = 151643
llama_model_loader: - kv  31:               tokenizer.ggml.add_bos_token bool             = false
llama_model_loader: - kv  32:                    tokenizer.chat_template str              = {%- if tools %}\n    {{- '<|im_start|>...
llama_model_loader: - kv  33:               general.quantization_version u32              = 2
llama_model_loader: - type  f32:  321 tensors
llama_model_loader: - type q4_K:  385 tensors
llama_model_loader: - type q6_K:   65 tensors
print_info: file format = GGUF V3 (latest)
print_info: file type   = Q4_K - Medium
print_info: file size   = 18.48 GiB (4.85 BPW)
init_tokenizer: initializing tokenizer for type 2
load: control token: 151659 '<|fim_prefix|>' is not marked as EOG
load: control token: 151656 '<|video_pad|>' is not marked as EOG
load: control token: 151655 '<|image_pad|>' is not marked as EOG
load: control token: 151653 '<|vision_end|>' is not marked as EOG
load: control token: 151652 '<|vision_start|>' is not marked as EOG
load: control token: 151651 '<|quad_end|>' is not marked as EOG
load: control token: 151649 '<|box_end|>' is not marked as EOG
load: control token: 151648 '<|box_start|>' is not marked as EOG
load: control token: 151646 '<|object_ref_start|>' is not marked as EOG
load: control token: 151644 '<|im_start|>' is not marked as EOG
load: control token: 151661 '<|fim_suffix|>' is not marked as EOG
load: control token: 151647 '<|object_ref_end|>' is not marked as EOG
load: control token: 151660 '<|fim_middle|>' is not marked as EOG
load: control token: 151654 '<|vision_pad|>' is not marked as EOG
load: control token: 151650 '<|quad_start|>' is not marked as EOG
load: printing all EOG tokens:
load:   - 151643 ('<|endoftext|>')
load:   - 151645 ('<|im_end|>')
load:   - 151662 ('<|fim_pad|>')
load:   - 151663 ('<|repo_name|>')
load:   - 151664 ('<|file_sep|>')
load: special tokens cache size = 22
load: token to piece cache size = 0.9310 MB
print_info: arch             = qwen2
print_info: vocab_only       = 1
print_info: model type       = ?B
print_info: model params     = 32.76 B
print_info: general.name     = Qwen2.5 Coder 32B Instruct
print_info: vocab type       = BPE
print_info: n_vocab          = 152064
print_info: n_merges         = 151387
print_info: BOS token        = 151643 '<|endoftext|>'
print_info: EOS token        = 151645 '<|im_end|>'
print_info: EOT token        = 151645 '<|im_end|>'
print_info: PAD token        = 151643 '<|endoftext|>'
print_info: LF token         = 198 'Ċ'
print_info: FIM PRE token    = 151659 '<|fim_prefix|>'
print_info: FIM SUF token    = 151661 '<|fim_suffix|>'
print_info: FIM MID token    = 151660 '<|fim_middle|>'
print_info: FIM PAD token    = 151662 '<|fim_pad|>'
print_info: FIM REP token    = 151663 '<|repo_name|>'
print_info: FIM SEP token    = 151664 '<|file_sep|>'
print_info: EOG token        = 151643 '<|endoftext|>'
print_info: EOG token        = 151645 '<|im_end|>'
print_info: EOG token        = 151662 '<|fim_pad|>'
print_info: EOG token        = 151663 '<|repo_name|>'
print_info: EOG token        = 151664 '<|file_sep|>'
print_info: max token length = 256
llama_model_load: vocab only - skipping tensors
time=2025-11-22T21:16:53.186+01:00 level=WARN source=server.go:167 msg="requested context size too large for model" num_ctx=131072 n_ctx_train=32768
time=2025-11-22T21:16:53.191+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --model H:\\LLM\\blobs\\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 --port 58133"
time=2025-11-22T21:16:53.196+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-11-22T21:16:53.257+01:00 level=INFO source=sched.go:443 msg="system memory" total="127.9 GiB" free="82.3 GiB" free_swap="453.9 GiB"
time=2025-11-22T21:16:53.261+01:00 level=INFO source=sched.go:450 msg="gpu memory" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=CUDA available="7.8 GiB" free="8.3 GiB" minimum="457.0 MiB" overhead="0 B"
time=2025-11-22T21:16:53.268+01:00 level=INFO source=sched.go:450 msg="gpu memory" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=CUDA available="10.2 GiB" free="10.6 GiB" minimum="457.0 MiB" overhead="0 B"
time=2025-11-22T21:16:53.276+01:00 level=INFO source=sched.go:450 msg="gpu memory" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=CUDA available="10.3 GiB" free="10.8 GiB" minimum="457.0 MiB" overhead="0 B"
time=2025-11-22T21:16:53.282+01:00 level=INFO source=server.go:459 msg="loading model" "model layers"=65 requested=-1
time=2025-11-22T21:16:53.287+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=qwen2.attention.key_length default=128
time=2025-11-22T21:16:53.292+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=qwen2.attention.value_length default=128
time=2025-11-22T21:16:53.296+01:00 level=DEBUG source=ggml.go:613 msg="default cache size estimate" "attention MiB"=24576 "attention bytes"=25769803776 "recurrent MiB"=0 "recurrent bytes"=0
time=2025-11-22T21:16:53.321+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=0 size="681.7 MiB"
time=2025-11-22T21:16:53.327+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=1 size="681.7 MiB"
time=2025-11-22T21:16:53.335+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=2 size="681.7 MiB"
time=2025-11-22T21:16:53.342+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=3 size="681.7 MiB"
time=2025-11-22T21:16:53.352+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=4 size="681.7 MiB"
time=2025-11-22T21:16:53.353+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=5 size="681.7 MiB"
time=2025-11-22T21:16:53.356+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=6 size="681.7 MiB"
time=2025-11-22T21:16:53.366+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=7 size="681.7 MiB"
time=2025-11-22T21:16:53.368+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=8 size="645.6 MiB"
time=2025-11-22T21:16:53.370+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=9 size="646.9 MiB"
time=2025-11-22T21:16:53.373+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=10 size="645.6 MiB"
time=2025-11-22T21:16:53.381+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=11 size="646.9 MiB"
time=2025-11-22T21:16:53.384+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=12 size="680.4 MiB"
time=2025-11-22T21:16:53.385+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=13 size="680.4 MiB"
time=2025-11-22T21:16:53.386+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=14 size="645.6 MiB"
time=2025-11-22T21:16:53.390+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=15 size="645.6 MiB"
time=2025-11-22T21:16:53.396+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=16 size="681.7 MiB"
time=2025-11-22T21:16:53.399+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=17 size="645.6 MiB"
time=2025-11-22T21:16:53.400+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=18 size="645.6 MiB"
time=2025-11-22T21:16:53.402+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=19 size="681.7 MiB"
time=2025-11-22T21:16:53.405+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=20 size="645.6 MiB"
time=2025-11-22T21:16:53.411+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=21 size="645.6 MiB"
time=2025-11-22T21:16:53.413+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=22 size="681.7 MiB"
time=2025-11-22T21:16:53.414+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=23 size="645.6 MiB"
time=2025-11-22T21:16:53.416+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=24 size="645.6 MiB"
time=2025-11-22T21:16:53.420+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=25 size="681.7 MiB"
time=2025-11-22T21:16:53.423+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=26 size="645.6 MiB"
time=2025-11-22T21:16:53.428+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=27 size="645.6 MiB"
time=2025-11-22T21:16:53.430+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=28 size="681.7 MiB"
time=2025-11-22T21:16:53.431+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=29 size="645.6 MiB"
time=2025-11-22T21:16:53.438+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=30 size="645.6 MiB"
time=2025-11-22T21:16:53.440+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=31 size="681.7 MiB"
time=2025-11-22T21:16:53.443+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=32 size="645.6 MiB"
time=2025-11-22T21:16:53.444+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=33 size="645.6 MiB"
time=2025-11-22T21:16:53.446+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=34 size="681.7 MiB"
time=2025-11-22T21:16:53.453+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=35 size="645.6 MiB"
time=2025-11-22T21:16:53.455+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=36 size="645.6 MiB"
time=2025-11-22T21:16:53.453+01:00 level=INFO source=runner.go:963 msg="starting go runner"
time=2025-11-22T21:16:53.455+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama"
time=2025-11-22T21:16:53.462+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=37 size="681.7 MiB"
time=2025-11-22T21:16:53.474+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=38 size="645.6 MiB"
time=2025-11-22T21:16:53.479+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=39 size="645.6 MiB"
time=2025-11-22T21:16:53.480+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=40 size="681.7 MiB"
time=2025-11-22T21:16:53.484+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=41 size="645.6 MiB"
time=2025-11-22T21:16:53.486+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=42 size="645.6 MiB"
time=2025-11-22T21:16:53.491+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=43 size="681.7 MiB"
time=2025-11-22T21:16:53.496+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=44 size="645.6 MiB"
time=2025-11-22T21:16:53.500+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=45 size="645.6 MiB"
time=2025-11-22T21:16:53.504+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=46 size="681.7 MiB"
time=2025-11-22T21:16:53.506+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=47 size="645.6 MiB"
time=2025-11-22T21:16:53.509+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=48 size="645.6 MiB"
time=2025-11-22T21:16:53.514+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=49 size="681.7 MiB"
time=2025-11-22T21:16:53.516+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=50 size="645.6 MiB"
time=2025-11-22T21:16:53.520+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=51 size="645.6 MiB"
time=2025-11-22T21:16:53.521+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=52 size="681.7 MiB"
time=2025-11-22T21:16:53.530+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=53 size="645.6 MiB"
time=2025-11-22T21:16:53.532+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=54 size="645.6 MiB"
time=2025-11-22T21:16:53.533+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=55 size="681.7 MiB"
time=2025-11-22T21:16:53.536+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=56 size="681.7 MiB"
load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll
time=2025-11-22T21:16:53.504+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13"
time=2025-11-22T21:16:53.538+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=57 size="681.7 MiB"
time=2025-11-22T21:16:53.557+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=58 size="681.7 MiB"
time=2025-11-22T21:16:53.559+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=59 size="681.7 MiB"
time=2025-11-22T21:16:53.561+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=60 size="681.7 MiB"
time=2025-11-22T21:16:53.563+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=61 size="681.7 MiB"
time=2025-11-22T21:16:53.567+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=62 size="681.7 MiB"
time=2025-11-22T21:16:53.572+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=63 size="681.7 MiB"
time=2025-11-22T21:16:53.573+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=64 size="609.1 MiB"
time=2025-11-22T21:16:53.577+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=CUDA "available layer vram"="9.6 GiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="0 B"
time=2025-11-22T21:16:53.581+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=CUDA "available layer vram"="9.5 GiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="0 B"
time=2025-11-22T21:16:53.594+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=CUDA "available layer vram"="7.2 GiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="0 B"
time=2025-11-22T21:16:53.597+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=0 size="681.7 MiB"
time=2025-11-22T21:16:53.602+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=1 size="681.7 MiB"
time=2025-11-22T21:16:53.605+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=2 size="681.7 MiB"
time=2025-11-22T21:16:53.608+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=3 size="681.7 MiB"
time=2025-11-22T21:16:53.612+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=4 size="681.7 MiB"
time=2025-11-22T21:16:53.615+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=5 size="681.7 MiB"
time=2025-11-22T21:16:53.616+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=6 size="681.7 MiB"
time=2025-11-22T21:16:53.618+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=7 size="681.7 MiB"
time=2025-11-22T21:16:53.621+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=8 size="645.6 MiB"
time=2025-11-22T21:16:53.625+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=9 size="646.9 MiB"
time=2025-11-22T21:16:53.626+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=10 size="645.6 MiB"
time=2025-11-22T21:16:53.628+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=11 size="646.9 MiB"
time=2025-11-22T21:16:53.629+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=12 size="680.4 MiB"
time=2025-11-22T21:16:53.630+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=13 size="680.4 MiB"
time=2025-11-22T21:16:53.633+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=14 size="645.6 MiB"
time=2025-11-22T21:16:53.637+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=15 size="645.6 MiB"
time=2025-11-22T21:16:53.638+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=16 size="681.7 MiB"
time=2025-11-22T21:16:53.638+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=17 size="645.6 MiB"
time=2025-11-22T21:16:53.640+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=18 size="645.6 MiB"
time=2025-11-22T21:16:53.641+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=19 size="681.7 MiB"
time=2025-11-22T21:16:53.643+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=20 size="645.6 MiB"
time=2025-11-22T21:16:53.644+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=21 size="645.6 MiB"
time=2025-11-22T21:16:53.648+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=22 size="681.7 MiB"
time=2025-11-22T21:16:53.651+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=23 size="645.6 MiB"
time=2025-11-22T21:16:53.653+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=24 size="645.6 MiB"
time=2025-11-22T21:16:53.655+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=25 size="681.7 MiB"
time=2025-11-22T21:16:53.659+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=26 size="645.6 MiB"
time=2025-11-22T21:16:53.660+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=27 size="645.6 MiB"
time=2025-11-22T21:16:53.662+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=28 size="681.7 MiB"
time=2025-11-22T21:16:53.663+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=29 size="645.6 MiB"
time=2025-11-22T21:16:53.665+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=30 size="645.6 MiB"
time=2025-11-22T21:16:53.666+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=31 size="681.7 MiB"
time=2025-11-22T21:16:53.669+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=32 size="645.6 MiB"
time=2025-11-22T21:16:53.670+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=33 size="645.6 MiB"
time=2025-11-22T21:16:53.672+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=34 size="681.7 MiB"
time=2025-11-22T21:16:53.673+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=35 size="645.6 MiB"
time=2025-11-22T21:16:53.675+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=36 size="645.6 MiB"
time=2025-11-22T21:16:53.676+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=37 size="681.7 MiB"
time=2025-11-22T21:16:53.678+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=38 size="645.6 MiB"
time=2025-11-22T21:16:53.681+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=39 size="645.6 MiB"
time=2025-11-22T21:16:53.682+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=40 size="681.7 MiB"
time=2025-11-22T21:16:53.684+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=41 size="645.6 MiB"
time=2025-11-22T21:16:53.685+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=42 size="645.6 MiB"
time=2025-11-22T21:16:53.687+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=43 size="681.7 MiB"
time=2025-11-22T21:16:53.689+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=44 size="645.6 MiB"
time=2025-11-22T21:16:53.694+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=45 size="645.6 MiB"
time=2025-11-22T21:16:53.695+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=46 size="681.7 MiB"
time=2025-11-22T21:16:53.697+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=47 size="645.6 MiB"
time=2025-11-22T21:16:53.699+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=48 size="645.6 MiB"
time=2025-11-22T21:16:53.700+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=49 size="681.7 MiB"
time=2025-11-22T21:16:53.704+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=50 size="645.6 MiB"
time=2025-11-22T21:16:53.705+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=51 size="645.6 MiB"
time=2025-11-22T21:16:53.707+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=52 size="681.7 MiB"
time=2025-11-22T21:16:53.709+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=53 size="645.6 MiB"
time=2025-11-22T21:16:53.711+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=54 size="645.6 MiB"
time=2025-11-22T21:16:53.715+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=55 size="681.7 MiB"
time=2025-11-22T21:16:53.716+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=56 size="681.7 MiB"
time=2025-11-22T21:16:53.717+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=57 size="681.7 MiB"
time=2025-11-22T21:16:53.720+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=58 size="681.7 MiB"
time=2025-11-22T21:16:53.720+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=59 size="681.7 MiB"
time=2025-11-22T21:16:53.723+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=60 size="681.7 MiB"
time=2025-11-22T21:16:53.726+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=61 size="681.7 MiB"
time=2025-11-22T21:16:53.728+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=62 size="681.7 MiB"
time=2025-11-22T21:16:53.729+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=63 size="681.7 MiB"
time=2025-11-22T21:16:53.731+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=64 size="609.1 MiB"
time=2025-11-22T21:16:53.733+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=CUDA "available layer vram"="65.2 MiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="9.6 GiB"
time=2025-11-22T21:16:53.737+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=CUDA "available layer vram"="0 B" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="9.6 GiB"
time=2025-11-22T21:16:53.742+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=CUDA "available layer vram"="0 B" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="9.6 GiB"
time=2025-11-22T21:16:53.746+01:00 level=DEBUG source=server.go:1002 msg="insufficient VRAM to load any model layers"
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 3 CUDA devices:
  Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8
  Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced
  Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c
load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll
time=2025-11-22T21:16:53.752+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)
time=2025-11-22T21:16:53.755+01:00 level=INFO source=runner.go:999 msg="Server listening on 127.0.0.1:58133"
time=2025-11-22T21:16:53.747+01:00 level=DEBUG source=server.go:614 msg=memory estimate.CPU.Weights="[312184832 312184832 312184832 312184832 312184832 312184832 312184832 312184832 274337792 275689472 274337792 275689472 310833152 310833152 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 312184832 312184832 312184832 312184832 312184832 312184832 312184832 312184832 638689280]" estimate.CPU.Cache="[402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 0]" estimate.CUDA0.ID=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 estimate.CUDA0.Graph=10288650240 estimate.CUDA1.ID=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced estimate.CUDA1.Graph=10288650240 estimate.CUDA2.ID=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c estimate.CUDA2.Graph=10288650240
time=2025-11-22T21:16:53.787+01:00 level=INFO source=device.go:245 msg="model weights" device=CPU size="18.1 GiB"
time=2025-11-22T21:16:53.795+01:00 level=INFO source=device.go:256 msg="kv cache" device=CPU size="24.0 GiB"
time=2025-11-22T21:16:53.796+01:00 level=INFO source=device.go:262 msg="compute graph" device=CUDA0 size="9.6 GiB"
time=2025-11-22T21:16:53.798+01:00 level=INFO source=device.go:262 msg="compute graph" device=CUDA1 size="9.6 GiB"
time=2025-11-22T21:16:53.799+01:00 level=INFO source=device.go:262 msg="compute graph" device=CUDA2 size="9.6 GiB"
time=2025-11-22T21:16:53.801+01:00 level=INFO source=device.go:272 msg="total memory" size="70.8 GiB"
time=2025-11-22T21:16:53.807+01:00 level=INFO source=runner.go:893 msg=load request="{Operation:commit LoraPath:[] Parallel:3 BatchSize:512 FlashAttention:false KvSize:98304 KvCacheType: NumThreads:28 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2025-11-22T21:16:53.808+01:00 level=INFO source=server.go:1294 msg="waiting for llama runner to start responding"
time=2025-11-22T21:16:53.813+01:00 level=INFO source=server.go:1328 msg="waiting for server to become available" status="llm server loading model"
ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8894390272 total: 12820938752
llama_model_load_from_file_impl: using device CUDA0 (NVIDIA GeForce RTX 5070) (0000:03:00.0) - 8482 MiB free
ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11387678720 total: 12884901888
llama_model_load_from_file_impl: using device CUDA1 (NVIDIA GeForce RTX 3060) (0000:04:00.0) - 10860 MiB free
ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11551105024 total: 12884901888
llama_model_load_from_file_impl: using device CUDA2 (NVIDIA GeForce RTX 3060) (0000:83:00.0) - 11015 MiB free
llama_model_loader: loaded meta data with 34 key-value pairs and 771 tensors from H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv   0:                       general.architecture str              = qwen2
llama_model_loader: - kv   1:                               general.type str              = model
llama_model_loader: - kv   2:                               general.name str              = Qwen2.5 Coder 32B Instruct
llama_model_loader: - kv   3:                           general.finetune str              = Instruct
llama_model_loader: - kv   4:                           general.basename str              = Qwen2.5-Coder
llama_model_loader: - kv   5:                         general.size_label str              = 32B
llama_model_loader: - kv   6:                            general.license str              = apache-2.0
llama_model_loader: - kv   7:                       general.license.link str              = https://huggingface.co/Qwen/Qwen2.5-C...
llama_model_loader: - kv   8:                   general.base_model.count u32              = 1
llama_model_loader: - kv   9:                  general.base_model.0.name str              = Qwen2.5 Coder 32B
llama_model_loader: - kv  10:          general.base_model.0.organization str              = Qwen
llama_model_loader: - kv  11:              general.base_model.0.repo_url str              = https://huggingface.co/Qwen/Qwen2.5-C...
llama_model_loader: - kv  12:                               general.tags arr[str,6]       = ["code", "codeqwen", "chat", "qwen", ...
llama_model_loader: - kv  13:                          general.languages arr[str,1]       = ["en"]
llama_model_loader: - kv  14:                          qwen2.block_count u32              = 64
llama_model_loader: - kv  15:                       qwen2.context_length u32              = 32768
llama_model_loader: - kv  16:                     qwen2.embedding_length u32              = 5120
llama_model_loader: - kv  17:                  qwen2.feed_forward_length u32              = 27648
llama_model_loader: - kv  18:                 qwen2.attention.head_count u32              = 40
llama_model_loader: - kv  19:              qwen2.attention.head_count_kv u32              = 8
llama_model_loader: - kv  20:                       qwen2.rope.freq_base f32              = 1000000.000000
llama_model_loader: - kv  21:     qwen2.attention.layer_norm_rms_epsilon f32              = 0.000001
llama_model_loader: - kv  22:                          general.file_type u32              = 15
llama_model_loader: - kv  23:                       tokenizer.ggml.model str              = gpt2
llama_model_loader: - kv  24:                         tokenizer.ggml.pre str              = qwen2
llama_model_loader: - kv  25:                      tokenizer.ggml.tokens arr[str,152064]  = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv  26:                  tokenizer.ggml.token_type arr[i32,152064]  = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv  27:                      tokenizer.ggml.merges arr[str,151387]  = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",...
llama_model_loader: - kv  28:                tokenizer.ggml.eos_token_id u32              = 151645
llama_model_loader: - kv  29:            tokenizer.ggml.padding_token_id u32              = 151643
llama_model_loader: - kv  30:                tokenizer.ggml.bos_token_id u32              = 151643
llama_model_loader: - kv  31:               tokenizer.ggml.add_bos_token bool             = false
llama_model_loader: - kv  32:                    tokenizer.chat_template str              = {%- if tools %}\n    {{- '<|im_start|>...
llama_model_loader: - kv  33:               general.quantization_version u32              = 2
llama_model_loader: - type  f32:  321 tensors
llama_model_loader: - type q4_K:  385 tensors
llama_model_loader: - type q6_K:   65 tensors
print_info: file format = GGUF V3 (latest)
print_info: file type   = Q4_K - Medium
print_info: file size   = 18.48 GiB (4.85 BPW)
init_tokenizer: initializing tokenizer for type 2
load: control token: 151659 '<|fim_prefix|>' is not marked as EOG
load: control token: 151656 '<|video_pad|>' is not marked as EOG
load: control token: 151655 '<|image_pad|>' is not marked as EOG
load: control token: 151653 '<|vision_end|>' is not marked as EOG
load: control token: 151652 '<|vision_start|>' is not marked as EOG
load: control token: 151651 '<|quad_end|>' is not marked as EOG
load: control token: 151649 '<|box_end|>' is not marked as EOG
load: control token: 151648 '<|box_start|>' is not marked as EOG
load: control token: 151646 '<|object_ref_start|>' is not marked as EOG
load: control token: 151644 '<|im_start|>' is not marked as EOG
load: control token: 151661 '<|fim_suffix|>' is not marked as EOG
load: control token: 151647 '<|object_ref_end|>' is not marked as EOG
load: control token: 151660 '<|fim_middle|>' is not marked as EOG
load: control token: 151654 '<|vision_pad|>' is not marked as EOG
load: control token: 151650 '<|quad_start|>' is not marked as EOG
load: printing all EOG tokens:
load:   - 151643 ('<|endoftext|>')
load:   - 151645 ('<|im_end|>')
load:   - 151662 ('<|fim_pad|>')
load:   - 151663 ('<|repo_name|>')
load:   - 151664 ('<|file_sep|>')
load: special tokens cache size = 22
load: token to piece cache size = 0.9310 MB
print_info: arch             = qwen2
print_info: vocab_only       = 0
print_info: n_ctx_train      = 32768
print_info: n_embd           = 5120
print_info: n_layer          = 64
print_info: n_head           = 40
print_info: n_head_kv        = 8
print_info: n_rot            = 128
print_info: n_swa            = 0
print_info: is_swa_any       = 0
print_info: n_embd_head_k    = 128
print_info: n_embd_head_v    = 128
print_info: n_gqa            = 5
print_info: n_embd_k_gqa     = 1024
print_info: n_embd_v_gqa     = 1024
print_info: f_norm_eps       = 0.0e+00
print_info: f_norm_rms_eps   = 1.0e-06
print_info: f_clamp_kqv      = 0.0e+00
print_info: f_max_alibi_bias = 0.0e+00
print_info: f_logit_scale    = 0.0e+00
print_info: f_attn_scale     = 0.0e+00
print_info: n_ff             = 27648
print_info: n_expert         = 0
print_info: n_expert_used    = 0
print_info: causal attn      = 1
print_info: pooling type     = -1
print_info: rope type        = 2
print_info: rope scaling     = linear
print_info: freq_base_train  = 1000000.0
print_info: freq_scale_train = 1
print_info: n_ctx_orig_yarn  = 32768
print_info: rope_finetuned   = unknown
print_info: model type       = 32B
print_info: model params     = 32.76 B
print_info: general.name     = Qwen2.5 Coder 32B Instruct
print_info: vocab type       = BPE
print_info: n_vocab          = 152064
print_info: n_merges         = 151387
print_info: BOS token        = 151643 '<|endoftext|>'
print_info: EOS token        = 151645 '<|im_end|>'
print_info: EOT token        = 151645 '<|im_end|>'
print_info: PAD token        = 151643 '<|endoftext|>'
print_info: LF token         = 198 'Ċ'
print_info: FIM PRE token    = 151659 '<|fim_prefix|>'
print_info: FIM SUF token    = 151661 '<|fim_suffix|>'
print_info: FIM MID token    = 151660 '<|fim_middle|>'
print_info: FIM PAD token    = 151662 '<|fim_pad|>'
print_info: FIM REP token    = 151663 '<|repo_name|>'
print_info: FIM SEP token    = 151664 '<|file_sep|>'
print_info: EOG token        = 151643 '<|endoftext|>'
print_info: EOG token        = 151645 '<|im_end|>'
print_info: EOG token        = 151662 '<|fim_pad|>'
print_info: EOG token        = 151663 '<|repo_name|>'
print_info: EOG token        = 151664 '<|file_sep|>'
print_info: max token length = 256
load_tensors: loading model tensors, this can take a while... (mmap = false)
ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8888201216 total: 12820938752
ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11387678720 total: 12884901888
ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11551105024 total: 12884901888
load_tensors: layer   0 assigned to device CPU, is_swa = 0
load_tensors: layer   1 assigned to device CPU, is_swa = 0
load_tensors: layer   2 assigned to device CPU, is_swa = 0
load_tensors: layer   3 assigned to device CPU, is_swa = 0
load_tensors: layer   4 assigned to device CPU, is_swa = 0
load_tensors: layer   5 assigned to device CPU, is_swa = 0
load_tensors: layer   6 assigned to device CPU, is_swa = 0
load_tensors: layer   7 assigned to device CPU, is_swa = 0
load_tensors: layer   8 assigned to device CPU, is_swa = 0
load_tensors: layer   9 assigned to device CPU, is_swa = 0
load_tensors: layer  10 assigned to device CPU, is_swa = 0
load_tensors: layer  11 assigned to device CPU, is_swa = 0
load_tensors: layer  12 assigned to device CPU, is_swa = 0
load_tensors: layer  13 assigned to device CPU, is_swa = 0
load_tensors: layer  14 assigned to device CPU, is_swa = 0
load_tensors: layer  15 assigned to device CPU, is_swa = 0
load_tensors: layer  16 assigned to device CPU, is_swa = 0
load_tensors: layer  17 assigned to device CPU, is_swa = 0
load_tensors: layer  18 assigned to device CPU, is_swa = 0
load_tensors: layer  19 assigned to device CPU, is_swa = 0
load_tensors: layer  20 assigned to device CPU, is_swa = 0
load_tensors: layer  21 assigned to device CPU, is_swa = 0
load_tensors: layer  22 assigned to device CPU, is_swa = 0
load_tensors: layer  23 assigned to device CPU, is_swa = 0
load_tensors: layer  24 assigned to device CPU, is_swa = 0
load_tensors: layer  25 assigned to device CPU, is_swa = 0
load_tensors: layer  26 assigned to device CPU, is_swa = 0
load_tensors: layer  27 assigned to device CPU, is_swa = 0
load_tensors: layer  28 assigned to device CPU, is_swa = 0
load_tensors: layer  29 assigned to device CPU, is_swa = 0
load_tensors: layer  30 assigned to device CPU, is_swa = 0
load_tensors: layer  31 assigned to device CPU, is_swa = 0
load_tensors: layer  32 assigned to device CPU, is_swa = 0
load_tensors: layer  33 assigned to device CPU, is_swa = 0
load_tensors: layer  34 assigned to device CPU, is_swa = 0
load_tensors: layer  35 assigned to device CPU, is_swa = 0
load_tensors: layer  36 assigned to device CPU, is_swa = 0
load_tensors: layer  37 assigned to device CPU, is_swa = 0
load_tensors: layer  38 assigned to device CPU, is_swa = 0
load_tensors: layer  39 assigned to device CPU, is_swa = 0
load_tensors: layer  40 assigned to device CPU, is_swa = 0
load_tensors: layer  41 assigned to device CPU, is_swa = 0
load_tensors: layer  42 assigned to device CPU, is_swa = 0
load_tensors: layer  43 assigned to device CPU, is_swa = 0
load_tensors: layer  44 assigned to device CPU, is_swa = 0
load_tensors: layer  45 assigned to device CPU, is_swa = 0
load_tensors: layer  46 assigned to device CPU, is_swa = 0
load_tensors: layer  47 assigned to device CPU, is_swa = 0
load_tensors: layer  48 assigned to device CPU, is_swa = 0
load_tensors: layer  49 assigned to device CPU, is_swa = 0
load_tensors: layer  50 assigned to device CPU, is_swa = 0
load_tensors: layer  51 assigned to device CPU, is_swa = 0
load_tensors: layer  52 assigned to device CPU, is_swa = 0
load_tensors: layer  53 assigned to device CPU, is_swa = 0
load_tensors: layer  54 assigned to device CPU, is_swa = 0
load_tensors: layer  55 assigned to device CPU, is_swa = 0
load_tensors: layer  56 assigned to device CPU, is_swa = 0
load_tensors: layer  57 assigned to device CPU, is_swa = 0
load_tensors: layer  58 assigned to device CPU, is_swa = 0
load_tensors: layer  59 assigned to device CPU, is_swa = 0
load_tensors: layer  60 assigned to device CPU, is_swa = 0
load_tensors: layer  61 assigned to device CPU, is_swa = 0
load_tensors: layer  62 assigned to device CPU, is_swa = 0
load_tensors: layer  63 assigned to device CPU, is_swa = 0
load_tensors: layer  64 assigned to device CPU, is_swa = 0
create_tensor: loading tensor token_embd.weight
create_tensor: loading tensor output_norm.weight
create_tensor: loading tensor output.weight
create_tensor: loading tensor blk.0.attn_norm.weight
create_tensor: loading tensor blk.0.attn_q.weight
create_tensor: loading tensor blk.0.attn_k.weight
create_tensor: loading tensor blk.0.attn_v.weight
create_tensor: loading tensor blk.0.attn_output.weight
create_tensor: loading tensor blk.0.attn_q.bias
create_tensor: loading tensor blk.0.attn_k.bias
create_tensor: loading tensor blk.0.attn_v.bias
create_tensor: loading tensor blk.0.ffn_norm.weight
create_tensor: loading tensor blk.0.ffn_gate.weight
create_tensor: loading tensor blk.0.ffn_down.weight
create_tensor: loading tensor blk.0.ffn_up.weight
create_tensor: loading tensor blk.1.attn_norm.weight
create_tensor: loading tensor blk.1.attn_q.weight
create_tensor: loading tensor blk.1.attn_k.weight
create_tensor: loading tensor blk.1.attn_v.weight
create_tensor: loading tensor blk.1.attn_output.weight
create_tensor: loading tensor blk.1.attn_q.bias
create_tensor: loading tensor blk.1.attn_k.bias
create_tensor: loading tensor blk.1.attn_v.bias
create_tensor: loading tensor blk.1.ffn_norm.weight
create_tensor: loading tensor blk.1.ffn_gate.weight
create_tensor: loading tensor blk.1.ffn_down.weight
create_tensor: loading tensor blk.1.ffn_up.weight
create_tensor: loading tensor blk.2.attn_norm.weight
create_tensor: loading tensor blk.2.attn_q.weight
create_tensor: loading tensor blk.2.attn_k.weight
create_tensor: loading tensor blk.2.attn_v.weight
create_tensor: loading tensor blk.2.attn_output.weight
create_tensor: loading tensor blk.2.attn_q.bias
create_tensor: loading tensor blk.2.attn_k.bias
create_tensor: loading tensor blk.2.attn_v.bias
create_tensor: loading tensor blk.2.ffn_norm.weight
create_tensor: loading tensor blk.2.ffn_gate.weight
create_tensor: loading tensor blk.2.ffn_down.weight
create_tensor: loading tensor blk.2.ffn_up.weight
create_tensor: loading tensor blk.3.attn_norm.weight
create_tensor: loading tensor blk.3.attn_q.weight
create_tensor: loading tensor blk.3.attn_k.weight
create_tensor: loading tensor blk.3.attn_v.weight
create_tensor: loading tensor blk.3.attn_output.weight
create_tensor: loading tensor blk.3.attn_q.bias
create_tensor: loading tensor blk.3.attn_k.bias
create_tensor: loading tensor blk.3.attn_v.bias
create_tensor: loading tensor blk.3.ffn_norm.weight
create_tensor: loading tensor blk.3.ffn_gate.weight
create_tensor: loading tensor blk.3.ffn_down.weight
create_tensor: loading tensor blk.3.ffn_up.weight
create_tensor: loading tensor blk.4.attn_norm.weight
create_tensor: loading tensor blk.4.attn_q.weight
create_tensor: loading tensor blk.4.attn_k.weight
create_tensor: loading tensor blk.4.attn_v.weight
create_tensor: loading tensor blk.4.attn_output.weight
create_tensor: loading tensor blk.4.attn_q.bias
create_tensor: loading tensor blk.4.attn_k.bias
create_tensor: loading tensor blk.4.attn_v.bias
create_tensor: loading tensor blk.4.ffn_norm.weight
create_tensor: loading tensor blk.4.ffn_gate.weight
create_tensor: loading tensor blk.4.ffn_down.weight
create_tensor: loading tensor blk.4.ffn_up.weight
create_tensor: loading tensor blk.5.attn_norm.weight
create_tensor: loading tensor blk.5.attn_q.weight
create_tensor: loading tensor blk.5.attn_k.weight
create_tensor: loading tensor blk.5.attn_v.weight
create_tensor: loading tensor blk.5.attn_output.weight
create_tensor: loading tensor blk.5.attn_q.bias
create_tensor: loading tensor blk.5.attn_k.bias
create_tensor: loading tensor blk.5.attn_v.bias
create_tensor: loading tensor blk.5.ffn_norm.weight
create_tensor: loading tensor blk.5.ffn_gate.weight
create_tensor: loading tensor blk.5.ffn_down.weight
create_tensor: loading tensor blk.5.ffn_up.weight
create_tensor: loading tensor blk.6.attn_norm.weight
create_tensor: loading tensor blk.6.attn_q.weight
create_tensor: loading tensor blk.6.attn_k.weight
create_tensor: loading tensor blk.6.attn_v.weight
create_tensor: loading tensor blk.6.attn_output.weight
create_tensor: loading tensor blk.6.attn_q.bias
create_tensor: loading tensor blk.6.attn_k.bias
create_tensor: loading tensor blk.6.attn_v.bias
create_tensor: loading tensor blk.6.ffn_norm.weight
create_tensor: loading tensor blk.6.ffn_gate.weight
create_tensor: loading tensor blk.6.ffn_down.weight
create_tensor: loading tensor blk.6.ffn_up.weight
create_tensor: loading tensor blk.7.attn_norm.weight
create_tensor: loading tensor blk.7.attn_q.weight
create_tensor: loading tensor blk.7.attn_k.weight
create_tensor: loading tensor blk.7.attn_v.weight
create_tensor: loading tensor blk.7.attn_output.weight
create_tensor: loading tensor blk.7.attn_q.bias
create_tensor: loading tensor blk.7.attn_k.bias
create_tensor: loading tensor blk.7.attn_v.bias
create_tensor: loading tensor blk.7.ffn_norm.weight
create_tensor: loading tensor blk.7.ffn_gate.weight
create_tensor: loading tensor blk.7.ffn_down.weight
create_tensor: loading tensor blk.7.ffn_up.weight
create_tensor: loading tensor blk.8.attn_norm.weight
create_tensor: loading tensor blk.8.attn_q.weight
create_tensor: loading tensor blk.8.attn_k.weight
create_tensor: loading tensor blk.8.attn_v.weight
create_tensor: loading tensor blk.8.attn_output.weight
create_tensor: loading tensor blk.8.attn_q.bias
create_tensor: loading tensor blk.8.attn_k.bias
create_tensor: loading tensor blk.8.attn_v.bias
create_tensor: loading tensor blk.8.ffn_norm.weight
create_tensor: loading tensor blk.8.ffn_gate.weight
create_tensor: loading tensor blk.8.ffn_down.weight
create_tensor: loading tensor blk.8.ffn_up.weight
create_tensor: loading tensor blk.9.attn_norm.weight
create_tensor: loading tensor blk.9.attn_q.weight
create_tensor: loading tensor blk.9.attn_k.weight
create_tensor: loading tensor blk.9.attn_v.weight
create_tensor: loading tensor blk.9.attn_output.weight
create_tensor: loading tensor blk.9.attn_q.bias
create_tensor: loading tensor blk.9.attn_k.bias
create_tensor: loading tensor blk.9.attn_v.bias
create_tensor: loading tensor blk.9.ffn_norm.weight
create_tensor: loading tensor blk.9.ffn_gate.weight
create_tensor: loading tensor blk.9.ffn_down.weight
create_tensor: loading tensor blk.9.ffn_up.weight
create_tensor: loading tensor blk.10.attn_norm.weight
create_tensor: loading tensor blk.10.attn_q.weight
create_tensor: loading tensor blk.10.attn_k.weight
create_tensor: loading tensor blk.10.attn_v.weight
create_tensor: loading tensor blk.10.attn_output.weight
create_tensor: loading tensor blk.10.attn_q.bias
create_tensor: loading tensor blk.10.attn_k.bias
create_tensor: loading tensor blk.10.attn_v.bias
create_tensor: loading tensor blk.10.ffn_norm.weight
create_tensor: loading tensor blk.10.ffn_gate.weight
create_tensor: loading tensor blk.10.ffn_down.weight
create_tensor: loading tensor blk.10.ffn_up.weight
create_tensor: loading tensor blk.11.attn_norm.weight
create_tensor: loading tensor blk.11.attn_q.weight
create_tensor: loading tensor blk.11.attn_k.weight
create_tensor: loading tensor blk.11.attn_v.weight
create_tensor: loading tensor blk.11.attn_output.weight
create_tensor: loading tensor blk.11.attn_q.bias
create_tensor: loading tensor blk.11.attn_k.bias
create_tensor: loading tensor blk.11.attn_v.bias
create_tensor: loading tensor blk.11.ffn_norm.weight
create_tensor: loading tensor blk.11.ffn_gate.weight
create_tensor: loading tensor blk.11.ffn_down.weight
create_tensor: loading tensor blk.11.ffn_up.weight
create_tensor: loading tensor blk.12.attn_norm.weight
create_tensor: loading tensor blk.12.attn_q.weight
create_tensor: loading tensor blk.12.attn_k.weight
create_tensor: loading tensor blk.12.attn_v.weight
create_tensor: loading tensor blk.12.attn_output.weight
create_tensor: loading tensor blk.12.attn_q.bias
create_tensor: loading tensor blk.12.attn_k.bias
create_tensor: loading tensor blk.12.attn_v.bias
create_tensor: loading tensor blk.12.ffn_norm.weight
create_tensor: loading tensor blk.12.ffn_gate.weight
create_tensor: loading tensor blk.12.ffn_down.weight
create_tensor: loading tensor blk.12.ffn_up.weight
create_tensor: loading tensor blk.13.attn_norm.weight
create_tensor: loading tensor blk.13.attn_q.weight
create_tensor: loading tensor blk.13.attn_k.weight
create_tensor: loading tensor blk.13.attn_v.weight
create_tensor: loading tensor blk.13.attn_output.weight
create_tensor: loading tensor blk.13.attn_q.bias
create_tensor: loading tensor blk.13.attn_k.bias
create_tensor: loading tensor blk.13.attn_v.bias
create_tensor: loading tensor blk.13.ffn_norm.weight
create_tensor: loading tensor blk.13.ffn_gate.weight
create_tensor: loading tensor blk.13.ffn_down.weight
create_tensor: loading tensor blk.13.ffn_up.weight
create_tensor: loading tensor blk.14.attn_norm.weight
create_tensor: loading tensor blk.14.attn_q.weight
create_tensor: loading tensor blk.14.attn_k.weight
create_tensor: loading tensor blk.14.attn_v.weight
create_tensor: loading tensor blk.14.attn_output.weight
create_tensor: loading tensor blk.14.attn_q.bias
create_tensor: loading tensor blk.14.attn_k.bias
create_tensor: loading tensor blk.14.attn_v.bias
create_tensor: loading tensor blk.14.ffn_norm.weight
create_tensor: loading tensor blk.14.ffn_gate.weight
create_tensor: loading tensor blk.14.ffn_down.weight
create_tensor: loading tensor blk.14.ffn_up.weight
create_tensor: loading tensor blk.15.attn_norm.weight
create_tensor: loading tensor blk.15.attn_q.weight
create_tensor: loading tensor blk.15.attn_k.weight
create_tensor: loading tensor blk.15.attn_v.weight
create_tensor: loading tensor blk.15.attn_output.weight
create_tensor: loading tensor blk.15.attn_q.bias
create_tensor: loading tensor blk.15.attn_k.bias
create_tensor: loading tensor blk.15.attn_v.bias
create_tensor: loading tensor blk.15.ffn_norm.weight
create_tensor: loading tensor blk.15.ffn_gate.weight
create_tensor: loading tensor blk.15.ffn_down.weight
create_tensor: loading tensor blk.15.ffn_up.weight
create_tensor: loading tensor blk.16.attn_norm.weight
create_tensor: loading tensor blk.16.attn_q.weight
create_tensor: loading tensor blk.16.attn_k.weight
create_tensor: loading tensor blk.16.attn_v.weight
create_tensor: loading tensor blk.16.attn_output.weight
create_tensor: loading tensor blk.16.attn_q.bias
create_tensor: loading tensor blk.16.attn_k.bias
create_tensor: loading tensor blk.16.attn_v.bias
create_tensor: loading tensor blk.16.ffn_norm.weight
create_tensor: loading tensor blk.16.ffn_gate.weight
create_tensor: loading tensor blk.16.ffn_down.weight
create_tensor: loading tensor blk.16.ffn_up.weight
create_tensor: loading tensor blk.17.attn_norm.weight
create_tensor: loading tensor blk.17.attn_q.weight
create_tensor: loading tensor blk.17.attn_k.weight
create_tensor: loading tensor blk.17.attn_v.weight
create_tensor: loading tensor blk.17.attn_output.weight
create_tensor: loading tensor blk.17.attn_q.bias
create_tensor: loading tensor blk.17.attn_k.bias
create_tensor: loading tensor blk.17.attn_v.bias
create_tensor: loading tensor blk.17.ffn_norm.weight
create_tensor: loading tensor blk.17.ffn_gate.weight
create_tensor: loading tensor blk.17.ffn_down.weight
create_tensor: loading tensor blk.17.ffn_up.weight
create_tensor: loading tensor blk.18.attn_norm.weight
create_tensor: loading tensor blk.18.attn_q.weight
create_tensor: loading tensor blk.18.attn_k.weight
create_tensor: loading tensor blk.18.attn_v.weight
create_tensor: loading tensor blk.18.attn_output.weight
create_tensor: loading tensor blk.18.attn_q.bias
create_tensor: loading tensor blk.18.attn_k.bias
create_tensor: loading tensor blk.18.attn_v.bias
create_tensor: loading tensor blk.18.ffn_norm.weight
create_tensor: loading tensor blk.18.ffn_gate.weight
create_tensor: loading tensor blk.18.ffn_down.weight
create_tensor: loading tensor blk.18.ffn_up.weight
create_tensor: loading tensor blk.19.attn_norm.weight
create_tensor: loading tensor blk.19.attn_q.weight
create_tensor: loading tensor blk.19.attn_k.weight
create_tensor: loading tensor blk.19.attn_v.weight
create_tensor: loading tensor blk.19.attn_output.weight
create_tensor: loading tensor blk.19.attn_q.bias
create_tensor: loading tensor blk.19.attn_k.bias
create_tensor: loading tensor blk.19.attn_v.bias
create_tensor: loading tensor blk.19.ffn_norm.weight
create_tensor: loading tensor blk.19.ffn_gate.weight
create_tensor: loading tensor blk.19.ffn_down.weight
create_tensor: loading tensor blk.19.ffn_up.weight
create_tensor: loading tensor blk.20.attn_norm.weight
create_tensor: loading tensor blk.20.attn_q.weight
create_tensor: loading tensor blk.20.attn_k.weight
create_tensor: loading tensor blk.20.attn_v.weight
create_tensor: loading tensor blk.20.attn_output.weight
create_tensor: loading tensor blk.20.attn_q.bias
create_tensor: loading tensor blk.20.attn_k.bias
create_tensor: loading tensor blk.20.attn_v.bias
create_tensor: loading tensor blk.20.ffn_norm.weight
create_tensor: loading tensor blk.20.ffn_gate.weight
create_tensor: loading tensor blk.20.ffn_down.weight
create_tensor: loading tensor blk.20.ffn_up.weight
create_tensor: loading tensor blk.21.attn_norm.weight
create_tensor: loading tensor blk.21.attn_q.weight
create_tensor: loading tensor blk.21.attn_k.weight
create_tensor: loading tensor blk.21.attn_v.weight
create_tensor: loading tensor blk.21.attn_output.weight
create_tensor: loading tensor blk.21.attn_q.bias
create_tensor: loading tensor blk.21.attn_k.bias
create_tensor: loading tensor blk.21.attn_v.bias
create_tensor: loading tensor blk.21.ffn_norm.weight
create_tensor: loading tensor blk.21.ffn_gate.weight
create_tensor: loading tensor blk.21.ffn_down.weight
create_tensor: loading tensor blk.21.ffn_up.weight
create_tensor: loading tensor blk.22.attn_norm.weight
create_tensor: loading tensor blk.22.attn_q.weight
create_tensor: loading tensor blk.22.attn_k.weight
create_tensor: loading tensor blk.22.attn_v.weight
create_tensor: loading tensor blk.22.attn_output.weight
create_tensor: loading tensor blk.22.attn_q.bias
create_tensor: loading tensor blk.22.attn_k.bias
create_tensor: loading tensor blk.22.attn_v.bias
create_tensor: loading tensor blk.22.ffn_norm.weight
create_tensor: loading tensor blk.22.ffn_gate.weight
create_tensor: loading tensor blk.22.ffn_down.weight
create_tensor: loading tensor blk.22.ffn_up.weight
create_tensor: loading tensor blk.23.attn_norm.weight
create_tensor: loading tensor blk.23.attn_q.weight
create_tensor: loading tensor blk.23.attn_k.weight
create_tensor: loading tensor blk.23.attn_v.weight
create_tensor: loading tensor blk.23.attn_output.weight
create_tensor: loading tensor blk.23.attn_q.bias
create_tensor: loading tensor blk.23.attn_k.bias
create_tensor: loading tensor blk.23.attn_v.bias
create_tensor: loading tensor blk.23.ffn_norm.weight
create_tensor: loading tensor blk.23.ffn_gate.weight
create_tensor: loading tensor blk.23.ffn_down.weight
create_tensor: loading tensor blk.23.ffn_up.weight
create_tensor: loading tensor blk.24.attn_norm.weight
create_tensor: loading tensor blk.24.attn_q.weight
create_tensor: loading tensor blk.24.attn_k.weight
create_tensor: loading tensor blk.24.attn_v.weight
create_tensor: loading tensor blk.24.attn_output.weight
create_tensor: loading tensor blk.24.attn_q.bias
create_tensor: loading tensor blk.24.attn_k.bias
create_tensor: loading tensor blk.24.attn_v.bias
create_tensor: loading tensor blk.24.ffn_norm.weight
create_tensor: loading tensor blk.24.ffn_gate.weight
create_tensor: loading tensor blk.24.ffn_down.weight
create_tensor: loading tensor blk.24.ffn_up.weight
create_tensor: loading tensor blk.25.attn_norm.weight
create_tensor: loading tensor blk.25.attn_q.weight
create_tensor: loading tensor blk.25.attn_k.weight
create_tensor: loading tensor blk.25.attn_v.weight
create_tensor: loading tensor blk.25.attn_output.weight
create_tensor: loading tensor blk.25.attn_q.bias
create_tensor: loading tensor blk.25.attn_k.bias
create_tensor: loading tensor blk.25.attn_v.bias
create_tensor: loading tensor blk.25.ffn_norm.weight
create_tensor: loading tensor blk.25.ffn_gate.weight
create_tensor: loading tensor blk.25.ffn_down.weight
create_tensor: loading tensor blk.25.ffn_up.weight
create_tensor: loading tensor blk.26.attn_norm.weight
create_tensor: loading tensor blk.26.attn_q.weight
create_tensor: loading tensor blk.26.attn_k.weight
create_tensor: loading tensor blk.26.attn_v.weight
create_tensor: loading tensor blk.26.attn_output.weight
create_tensor: loading tensor blk.26.attn_q.bias
create_tensor: loading tensor blk.26.attn_k.bias
create_tensor: loading tensor blk.26.attn_v.bias
create_tensor: loading tensor blk.26.ffn_norm.weight
create_tensor: loading tensor blk.26.ffn_gate.weight
create_tensor: loading tensor blk.26.ffn_down.weight
create_tensor: loading tensor blk.26.ffn_up.weight
create_tensor: loading tensor blk.27.attn_norm.weight
create_tensor: loading tensor blk.27.attn_q.weight
create_tensor: loading tensor blk.27.attn_k.weight
create_tensor: loading tensor blk.27.attn_v.weight
create_tensor: loading tensor blk.27.attn_output.weight
create_tensor: loading tensor blk.27.attn_q.bias
create_tensor: loading tensor blk.27.attn_k.bias
create_tensor: loading tensor blk.27.attn_v.bias
create_tensor: loading tensor blk.27.ffn_norm.weight
create_tensor: loading tensor blk.27.ffn_gate.weight
create_tensor: loading tensor blk.27.ffn_down.weight
create_tensor: loading tensor blk.27.ffn_up.weight
create_tensor: loading tensor blk.28.attn_norm.weight
create_tensor: loading tensor blk.28.attn_q.weight
create_tensor: loading tensor blk.28.attn_k.weight
create_tensor: loading tensor blk.28.attn_v.weight
create_tensor: loading tensor blk.28.attn_output.weight
create_tensor: loading tensor blk.28.attn_q.bias
create_tensor: loading tensor blk.28.attn_k.bias
create_tensor: loading tensor blk.28.attn_v.bias
create_tensor: loading tensor blk.28.ffn_norm.weight
create_tensor: loading tensor blk.28.ffn_gate.weight
create_tensor: loading tensor blk.28.ffn_down.weight
create_tensor: loading tensor blk.28.ffn_up.weight
create_tensor: loading tensor blk.29.attn_norm.weight
create_tensor: loading tensor blk.29.attn_q.weight
create_tensor: loading tensor blk.29.attn_k.weight
create_tensor: loading tensor blk.29.attn_v.weight
create_tensor: loading tensor blk.29.attn_output.weight
create_tensor: loading tensor blk.29.attn_q.bias
create_tensor: loading tensor blk.29.attn_k.bias
create_tensor: loading tensor blk.29.attn_v.bias
create_tensor: loading tensor blk.29.ffn_norm.weight
create_tensor: loading tensor blk.29.ffn_gate.weight
create_tensor: loading tensor blk.29.ffn_down.weight
create_tensor: loading tensor blk.29.ffn_up.weight
create_tensor: loading tensor blk.30.attn_norm.weight
create_tensor: loading tensor blk.30.attn_q.weight
create_tensor: loading tensor blk.30.attn_k.weight
create_tensor: loading tensor blk.30.attn_v.weight
create_tensor: loading tensor blk.30.attn_output.weight
create_tensor: loading tensor blk.30.attn_q.bias
create_tensor: loading tensor blk.30.attn_k.bias
create_tensor: loading tensor blk.30.attn_v.bias
create_tensor: loading tensor blk.30.ffn_norm.weight
create_tensor: loading tensor blk.30.ffn_gate.weight
create_tensor: loading tensor blk.30.ffn_down.weight
create_tensor: loading tensor blk.30.ffn_up.weight
create_tensor: loading tensor blk.31.attn_norm.weight
create_tensor: loading tensor blk.31.attn_q.weight
create_tensor: loading tensor blk.31.attn_k.weight
create_tensor: loading tensor blk.31.attn_v.weight
create_tensor: loading tensor blk.31.attn_output.weight
create_tensor: loading tensor blk.31.attn_q.bias
create_tensor: loading tensor blk.31.attn_k.bias
create_tensor: loading tensor blk.31.attn_v.bias
create_tensor: loading tensor blk.31.ffn_norm.weight
create_tensor: loading tensor blk.31.ffn_gate.weight
create_tensor: loading tensor blk.31.ffn_down.weight
create_tensor: loading tensor blk.31.ffn_up.weight
create_tensor: loading tensor blk.32.attn_norm.weight
create_tensor: loading tensor blk.32.attn_q.weight
create_tensor: loading tensor blk.32.attn_k.weight
create_tensor: loading tensor blk.32.attn_v.weight
create_tensor: loading tensor blk.32.attn_output.weight
create_tensor: loading tensor blk.32.attn_q.bias
create_tensor: loading tensor blk.32.attn_k.bias
create_tensor: loading tensor blk.32.attn_v.bias
create_tensor: loading tensor blk.32.ffn_norm.weight
create_tensor: loading tensor blk.32.ffn_gate.weight
create_tensor: loading tensor blk.32.ffn_down.weight
create_tensor: loading tensor blk.32.ffn_up.weight
create_tensor: loading tensor blk.33.attn_norm.weight
create_tensor: loading tensor blk.33.attn_q.weight
create_tensor: loading tensor blk.33.attn_k.weight
create_tensor: loading tensor blk.33.attn_v.weight
create_tensor: loading tensor blk.33.attn_output.weight
create_tensor: loading tensor blk.33.attn_q.bias
create_tensor: loading tensor blk.33.attn_k.bias
create_tensor: loading tensor blk.33.attn_v.bias
create_tensor: loading tensor blk.33.ffn_norm.weight
create_tensor: loading tensor blk.33.ffn_gate.weight
create_tensor: loading tensor blk.33.ffn_down.weight
create_tensor: loading tensor blk.33.ffn_up.weight
create_tensor: loading tensor blk.34.attn_norm.weight
create_tensor: loading tensor blk.34.attn_q.weight
create_tensor: loading tensor blk.34.attn_k.weight
create_tensor: loading tensor blk.34.attn_v.weight
create_tensor: loading tensor blk.34.attn_output.weight
create_tensor: loading tensor blk.34.attn_q.bias
create_tensor: loading tensor blk.34.attn_k.bias
create_tensor: loading tensor blk.34.attn_v.bias
create_tensor: loading tensor blk.34.ffn_norm.weight
create_tensor: loading tensor blk.34.ffn_gate.weight
create_tensor: loading tensor blk.34.ffn_down.weight
create_tensor: loading tensor blk.34.ffn_up.weight
create_tensor: loading tensor blk.35.attn_norm.weight
create_tensor: loading tensor blk.35.attn_q.weight
create_tensor: loading tensor blk.35.attn_k.weight
create_tensor: loading tensor blk.35.attn_v.weight
create_tensor: loading tensor blk.35.attn_output.weight
create_tensor: loading tensor blk.35.attn_q.bias
create_tensor: loading tensor blk.35.attn_k.bias
create_tensor: loading tensor blk.35.attn_v.bias
create_tensor: loading tensor blk.35.ffn_norm.weight
create_tensor: loading tensor blk.35.ffn_gate.weight
create_tensor: loading tensor blk.35.ffn_down.weight
create_tensor: loading tensor blk.35.ffn_up.weight
create_tensor: loading tensor blk.36.attn_norm.weight
create_tensor: loading tensor blk.36.attn_q.weight
create_tensor: loading tensor blk.36.attn_k.weight
create_tensor: loading tensor blk.36.attn_v.weight
create_tensor: loading tensor blk.36.attn_output.weight
create_tensor: loading tensor blk.36.attn_q.bias
create_tensor: loading tensor blk.36.attn_k.bias
create_tensor: loading tensor blk.36.attn_v.bias
create_tensor: loading tensor blk.36.ffn_norm.weight
create_tensor: loading tensor blk.36.ffn_gate.weight
create_tensor: loading tensor blk.36.ffn_down.weight
create_tensor: loading tensor blk.36.ffn_up.weight
create_tensor: loading tensor blk.37.attn_norm.weight
create_tensor: loading tensor blk.37.attn_q.weight
create_tensor: loading tensor blk.37.attn_k.weight
create_tensor: loading tensor blk.37.attn_v.weight
create_tensor: loading tensor blk.37.attn_output.weight
create_tensor: loading tensor blk.37.attn_q.bias
create_tensor: loading tensor blk.37.attn_k.bias
create_tensor: loading tensor blk.37.attn_v.bias
create_tensor: loading tensor blk.37.ffn_norm.weight
create_tensor: loading tensor blk.37.ffn_gate.weight
create_tensor: loading tensor blk.37.ffn_down.weight
create_tensor: loading tensor blk.37.ffn_up.weight
create_tensor: loading tensor blk.38.attn_norm.weight
create_tensor: loading tensor blk.38.attn_q.weight
create_tensor: loading tensor blk.38.attn_k.weight
create_tensor: loading tensor blk.38.attn_v.weight
create_tensor: loading tensor blk.38.attn_output.weight
create_tensor: loading tensor blk.38.attn_q.bias
create_tensor: loading tensor blk.38.attn_k.bias
create_tensor: loading tensor blk.38.attn_v.bias
create_tensor: loading tensor blk.38.ffn_norm.weight
create_tensor: loading tensor blk.38.ffn_gate.weight
create_tensor: loading tensor blk.38.ffn_down.weight
create_tensor: loading tensor blk.38.ffn_up.weight
create_tensor: loading tensor blk.39.attn_norm.weight
create_tensor: loading tensor blk.39.attn_q.weight
create_tensor: loading tensor blk.39.attn_k.weight
create_tensor: loading tensor blk.39.attn_v.weight
create_tensor: loading tensor blk.39.attn_output.weight
create_tensor: loading tensor blk.39.attn_q.bias
create_tensor: loading tensor blk.39.attn_k.bias
create_tensor: loading tensor blk.39.attn_v.bias
create_tensor: loading tensor blk.39.ffn_norm.weight
create_tensor: loading tensor blk.39.ffn_gate.weight
create_tensor: loading tensor blk.39.ffn_down.weight
create_tensor: loading tensor blk.39.ffn_up.weight
create_tensor: loading tensor blk.40.attn_norm.weight
create_tensor: loading tensor blk.40.attn_q.weight
create_tensor: loading tensor blk.40.attn_k.weight
create_tensor: loading tensor blk.40.attn_v.weight
create_tensor: loading tensor blk.40.attn_output.weight
create_tensor: loading tensor blk.40.attn_q.bias
create_tensor: loading tensor blk.40.attn_k.bias
create_tensor: loading tensor blk.40.attn_v.bias
create_tensor: loading tensor blk.40.ffn_norm.weight
create_tensor: loading tensor blk.40.ffn_gate.weight
create_tensor: loading tensor blk.40.ffn_down.weight
create_tensor: loading tensor blk.40.ffn_up.weight
create_tensor: loading tensor blk.41.attn_norm.weight
create_tensor: loading tensor blk.41.attn_q.weight
create_tensor: loading tensor blk.41.attn_k.weight
create_tensor: loading tensor blk.41.attn_v.weight
create_tensor: loading tensor blk.41.attn_output.weight
create_tensor: loading tensor blk.41.attn_q.bias
create_tensor: loading tensor blk.41.attn_k.bias
create_tensor: loading tensor blk.41.attn_v.bias
create_tensor: loading tensor blk.41.ffn_norm.weight
create_tensor: loading tensor blk.41.ffn_gate.weight
create_tensor: loading tensor blk.41.ffn_down.weight
create_tensor: loading tensor blk.41.ffn_up.weight
create_tensor: loading tensor blk.42.attn_norm.weight
create_tensor: loading tensor blk.42.attn_q.weight
create_tensor: loading tensor blk.42.attn_k.weight
create_tensor: loading tensor blk.42.attn_v.weight
create_tensor: loading tensor blk.42.attn_output.weight
create_tensor: loading tensor blk.42.attn_q.bias
create_tensor: loading tensor blk.42.attn_k.bias
create_tensor: loading tensor blk.42.attn_v.bias
create_tensor: loading tensor blk.42.ffn_norm.weight
create_tensor: loading tensor blk.42.ffn_gate.weight
create_tensor: loading tensor blk.42.ffn_down.weight
create_tensor: loading tensor blk.42.ffn_up.weight
create_tensor: loading tensor blk.43.attn_norm.weight
create_tensor: loading tensor blk.43.attn_q.weight
create_tensor: loading tensor blk.43.attn_k.weight
create_tensor: loading tensor blk.43.attn_v.weight
create_tensor: loading tensor blk.43.attn_output.weight
create_tensor: loading tensor blk.43.attn_q.bias
create_tensor: loading tensor blk.43.attn_k.bias
create_tensor: loading tensor blk.43.attn_v.bias
create_tensor: loading tensor blk.43.ffn_norm.weight
create_tensor: loading tensor blk.43.ffn_gate.weight
create_tensor: loading tensor blk.43.ffn_down.weight
create_tensor: loading tensor blk.43.ffn_up.weight
create_tensor: loading tensor blk.44.attn_norm.weight
create_tensor: loading tensor blk.44.attn_q.weight
create_tensor: loading tensor blk.44.attn_k.weight
create_tensor: loading tensor blk.44.attn_v.weight
create_tensor: loading tensor blk.44.attn_output.weight
create_tensor: loading tensor blk.44.attn_q.bias
create_tensor: loading tensor blk.44.attn_k.bias
create_tensor: loading tensor blk.44.attn_v.bias
create_tensor: loading tensor blk.44.ffn_norm.weight
create_tensor: loading tensor blk.44.ffn_gate.weight
create_tensor: loading tensor blk.44.ffn_down.weight
create_tensor: loading tensor blk.44.ffn_up.weight
create_tensor: loading tensor blk.45.attn_norm.weight
create_tensor: loading tensor blk.45.attn_q.weight
create_tensor: loading tensor blk.45.attn_k.weight
create_tensor: loading tensor blk.45.attn_v.weight
create_tensor: loading tensor blk.45.attn_output.weight
create_tensor: loading tensor blk.45.attn_q.bias
create_tensor: loading tensor blk.45.attn_k.bias
create_tensor: loading tensor blk.45.attn_v.bias
create_tensor: loading tensor blk.45.ffn_norm.weight
create_tensor: loading tensor blk.45.ffn_gate.weight
create_tensor: loading tensor blk.45.ffn_down.weight
create_tensor: loading tensor blk.45.ffn_up.weight
create_tensor: loading tensor blk.46.attn_norm.weight
create_tensor: loading tensor blk.46.attn_q.weight
create_tensor: loading tensor blk.46.attn_k.weight
create_tensor: loading tensor blk.46.attn_v.weight
create_tensor: loading tensor blk.46.attn_output.weight
create_tensor: loading tensor blk.46.attn_q.bias
create_tensor: loading tensor blk.46.attn_k.bias
create_tensor: loading tensor blk.46.attn_v.bias
create_tensor: loading tensor blk.46.ffn_norm.weight
create_tensor: loading tensor blk.46.ffn_gate.weight
create_tensor: loading tensor blk.46.ffn_down.weight
create_tensor: loading tensor blk.46.ffn_up.weight
create_tensor: loading tensor blk.47.attn_norm.weight
create_tensor: loading tensor blk.47.attn_q.weight
create_tensor: loading tensor blk.47.attn_k.weight
create_tensor: loading tensor blk.47.attn_v.weight
create_tensor: loading tensor blk.47.attn_output.weight
create_tensor: loading tensor blk.47.attn_q.bias
create_tensor: loading tensor blk.47.attn_k.bias
create_tensor: loading tensor blk.47.attn_v.bias
create_tensor: loading tensor blk.47.ffn_norm.weight
create_tensor: loading tensor blk.47.ffn_gate.weight
create_tensor: loading tensor blk.47.ffn_down.weight
create_tensor: loading tensor blk.47.ffn_up.weight
create_tensor: loading tensor blk.48.attn_norm.weight
create_tensor: loading tensor blk.48.attn_q.weight
create_tensor: loading tensor blk.48.attn_k.weight
create_tensor: loading tensor blk.48.attn_v.weight
create_tensor: loading tensor blk.48.attn_output.weight
create_tensor: loading tensor blk.48.attn_q.bias
create_tensor: loading tensor blk.48.attn_k.bias
create_tensor: loading tensor blk.48.attn_v.bias
create_tensor: loading tensor blk.48.ffn_norm.weight
create_tensor: loading tensor blk.48.ffn_gate.weight
create_tensor: loading tensor blk.48.ffn_down.weight
create_tensor: loading tensor blk.48.ffn_up.weight
create_tensor: loading tensor blk.49.attn_norm.weight
create_tensor: loading tensor blk.49.attn_q.weight
create_tensor: loading tensor blk.49.attn_k.weight
create_tensor: loading tensor blk.49.attn_v.weight
create_tensor: loading tensor blk.49.attn_output.weight
create_tensor: loading tensor blk.49.attn_q.bias
create_tensor: loading tensor blk.49.attn_k.bias
create_tensor: loading tensor blk.49.attn_v.bias
create_tensor: loading tensor blk.49.ffn_norm.weight
create_tensor: loading tensor blk.49.ffn_gate.weight
create_tensor: loading tensor blk.49.ffn_down.weight
create_tensor: loading tensor blk.49.ffn_up.weight
create_tensor: loading tensor blk.50.attn_norm.weight
create_tensor: loading tensor blk.50.attn_q.weight
create_tensor: loading tensor blk.50.attn_k.weight
create_tensor: loading tensor blk.50.attn_v.weight
create_tensor: loading tensor blk.50.attn_output.weight
create_tensor: loading tensor blk.50.attn_q.bias
create_tensor: loading tensor blk.50.attn_k.bias
create_tensor: loading tensor blk.50.attn_v.bias
create_tensor: loading tensor blk.50.ffn_norm.weight
create_tensor: loading tensor blk.50.ffn_gate.weight
create_tensor: loading tensor blk.50.ffn_down.weight
create_tensor: loading tensor blk.50.ffn_up.weight
create_tensor: loading tensor blk.51.attn_norm.weight
create_tensor: loading tensor blk.51.attn_q.weight
create_tensor: loading tensor blk.51.attn_k.weight
create_tensor: loading tensor blk.51.attn_v.weight
create_tensor: loading tensor blk.51.attn_output.weight
create_tensor: loading tensor blk.51.attn_q.bias
create_tensor: loading tensor blk.51.attn_k.bias
create_tensor: loading tensor blk.51.attn_v.bias
create_tensor: loading tensor blk.51.ffn_norm.weight
create_tensor: loading tensor blk.51.ffn_gate.weight
create_tensor: loading tensor blk.51.ffn_down.weight
create_tensor: loading tensor blk.51.ffn_up.weight
create_tensor: loading tensor blk.52.attn_norm.weight
create_tensor: loading tensor blk.52.attn_q.weight
create_tensor: loading tensor blk.52.attn_k.weight
create_tensor: loading tensor blk.52.attn_v.weight
create_tensor: loading tensor blk.52.attn_output.weight
create_tensor: loading tensor blk.52.attn_q.bias
create_tensor: loading tensor blk.52.attn_k.bias
create_tensor: loading tensor blk.52.attn_v.bias
create_tensor: loading tensor blk.52.ffn_norm.weight
create_tensor: loading tensor blk.52.ffn_gate.weight
create_tensor: loading tensor blk.52.ffn_down.weight
create_tensor: loading tensor blk.52.ffn_up.weight
create_tensor: loading tensor blk.53.attn_norm.weight
create_tensor: loading tensor blk.53.attn_q.weight
create_tensor: loading tensor blk.53.attn_k.weight
create_tensor: loading tensor blk.53.attn_v.weight
create_tensor: loading tensor blk.53.attn_output.weight
create_tensor: loading tensor blk.53.attn_q.bias
create_tensor: loading tensor blk.53.attn_k.bias
create_tensor: loading tensor blk.53.attn_v.bias
create_tensor: loading tensor blk.53.ffn_norm.weight
create_tensor: loading tensor blk.53.ffn_gate.weight
create_tensor: loading tensor blk.53.ffn_down.weight
create_tensor: loading tensor blk.53.ffn_up.weight
create_tensor: loading tensor blk.54.attn_norm.weight
create_tensor: loading tensor blk.54.attn_q.weight
create_tensor: loading tensor blk.54.attn_k.weight
create_tensor: loading tensor blk.54.attn_v.weight
create_tensor: loading tensor blk.54.attn_output.weight
create_tensor: loading tensor blk.54.attn_q.bias
create_tensor: loading tensor blk.54.attn_k.bias
create_tensor: loading tensor blk.54.attn_v.bias
create_tensor: loading tensor blk.54.ffn_norm.weight
create_tensor: loading tensor blk.54.ffn_gate.weight
create_tensor: loading tensor blk.54.ffn_down.weight
create_tensor: loading tensor blk.54.ffn_up.weight
create_tensor: loading tensor blk.55.attn_norm.weight
create_tensor: loading tensor blk.55.attn_q.weight
create_tensor: loading tensor blk.55.attn_k.weight
create_tensor: loading tensor blk.55.attn_v.weight
create_tensor: loading tensor blk.55.attn_output.weight
create_tensor: loading tensor blk.55.attn_q.bias
create_tensor: loading tensor blk.55.attn_k.bias
create_tensor: loading tensor blk.55.attn_v.bias
create_tensor: loading tensor blk.55.ffn_norm.weight
create_tensor: loading tensor blk.55.ffn_gate.weight
create_tensor: loading tensor blk.55.ffn_down.weight
create_tensor: loading tensor blk.55.ffn_up.weight
create_tensor: loading tensor blk.56.attn_norm.weight
create_tensor: loading tensor blk.56.attn_q.weight
create_tensor: loading tensor blk.56.attn_k.weight
create_tensor: loading tensor blk.56.attn_v.weight
create_tensor: loading tensor blk.56.attn_output.weight
create_tensor: loading tensor blk.56.attn_q.bias
create_tensor: loading tensor blk.56.attn_k.bias
create_tensor: loading tensor blk.56.attn_v.bias
create_tensor: loading tensor blk.56.ffn_norm.weight
create_tensor: loading tensor blk.56.ffn_gate.weight
create_tensor: loading tensor blk.56.ffn_down.weight
create_tensor: loading tensor blk.56.ffn_up.weight
create_tensor: loading tensor blk.57.attn_norm.weight
create_tensor: loading tensor blk.57.attn_q.weight
create_tensor: loading tensor blk.57.attn_k.weight
create_tensor: loading tensor blk.57.attn_v.weight
create_tensor: loading tensor blk.57.attn_output.weight
create_tensor: loading tensor blk.57.attn_q.bias
create_tensor: loading tensor blk.57.attn_k.bias
create_tensor: loading tensor blk.57.attn_v.bias
create_tensor: loading tensor blk.57.ffn_norm.weight
create_tensor: loading tensor blk.57.ffn_gate.weight
create_tensor: loading tensor blk.57.ffn_down.weight
create_tensor: loading tensor blk.57.ffn_up.weight
create_tensor: loading tensor blk.58.attn_norm.weight
create_tensor: loading tensor blk.58.attn_q.weight
create_tensor: loading tensor blk.58.attn_k.weight
create_tensor: loading tensor blk.58.attn_v.weight
create_tensor: loading tensor blk.58.attn_output.weight
create_tensor: loading tensor blk.58.attn_q.bias
create_tensor: loading tensor blk.58.attn_k.bias
create_tensor: loading tensor blk.58.attn_v.bias
create_tensor: loading tensor blk.58.ffn_norm.weight
create_tensor: loading tensor blk.58.ffn_gate.weight
create_tensor: loading tensor blk.58.ffn_down.weight
create_tensor: loading tensor blk.58.ffn_up.weight
create_tensor: loading tensor blk.59.attn_norm.weight
create_tensor: loading tensor blk.59.attn_q.weight
create_tensor: loading tensor blk.59.attn_k.weight
create_tensor: loading tensor blk.59.attn_v.weight
create_tensor: loading tensor blk.59.attn_output.weight
create_tensor: loading tensor blk.59.attn_q.bias
create_tensor: loading tensor blk.59.attn_k.bias
create_tensor: loading tensor blk.59.attn_v.bias
create_tensor: loading tensor blk.59.ffn_norm.weight
create_tensor: loading tensor blk.59.ffn_gate.weight
create_tensor: loading tensor blk.59.ffn_down.weight
create_tensor: loading tensor blk.59.ffn_up.weight
create_tensor: loading tensor blk.60.attn_norm.weight
create_tensor: loading tensor blk.60.attn_q.weight
create_tensor: loading tensor blk.60.attn_k.weight
create_tensor: loading tensor blk.60.attn_v.weight
create_tensor: loading tensor blk.60.attn_output.weight
create_tensor: loading tensor blk.60.attn_q.bias
create_tensor: loading tensor blk.60.attn_k.bias
create_tensor: loading tensor blk.60.attn_v.bias
create_tensor: loading tensor blk.60.ffn_norm.weight
create_tensor: loading tensor blk.60.ffn_gate.weight
create_tensor: loading tensor blk.60.ffn_down.weight
create_tensor: loading tensor blk.60.ffn_up.weight
create_tensor: loading tensor blk.61.attn_norm.weight
create_tensor: loading tensor blk.61.attn_q.weight
create_tensor: loading tensor blk.61.attn_k.weight
create_tensor: loading tensor blk.61.attn_v.weight
create_tensor: loading tensor blk.61.attn_output.weight
create_tensor: loading tensor blk.61.attn_q.bias
create_tensor: loading tensor blk.61.attn_k.bias
create_tensor: loading tensor blk.61.attn_v.bias
create_tensor: loading tensor blk.61.ffn_norm.weight
create_tensor: loading tensor blk.61.ffn_gate.weight
create_tensor: loading tensor blk.61.ffn_down.weight
create_tensor: loading tensor blk.61.ffn_up.weight
create_tensor: loading tensor blk.62.attn_norm.weight
create_tensor: loading tensor blk.62.attn_q.weight
create_tensor: loading tensor blk.62.attn_k.weight
create_tensor: loading tensor blk.62.attn_v.weight
create_tensor: loading tensor blk.62.attn_output.weight
create_tensor: loading tensor blk.62.attn_q.bias
create_tensor: loading tensor blk.62.attn_k.bias
create_tensor: loading tensor blk.62.attn_v.bias
create_tensor: loading tensor blk.62.ffn_norm.weight
create_tensor: loading tensor blk.62.ffn_gate.weight
create_tensor: loading tensor blk.62.ffn_down.weight
create_tensor: loading tensor blk.62.ffn_up.weight
create_tensor: loading tensor blk.63.attn_norm.weight
create_tensor: loading tensor blk.63.attn_q.weight
create_tensor: loading tensor blk.63.attn_k.weight
create_tensor: loading tensor blk.63.attn_v.weight
create_tensor: loading tensor blk.63.attn_output.weight
create_tensor: loading tensor blk.63.attn_q.bias
create_tensor: loading tensor blk.63.attn_k.bias
create_tensor: loading tensor blk.63.attn_v.bias
create_tensor: loading tensor blk.63.ffn_norm.weight
create_tensor: loading tensor blk.63.ffn_gate.weight
create_tensor: loading tensor blk.63.ffn_down.weight
create_tensor: loading tensor blk.63.ffn_up.weight
load_tensors: tensor 'token_embd.weight' (q4_K) (and 0 others) cannot be used with preferred buffer type CUDA_Host, using CPU instead
load_tensors: offloading 0 repeating layers to GPU
load_tensors: offloaded 0/65 layers to GPU
load_tensors:    CUDA_Host model buffer size = 18508.35 MiB
load_tensors:          CPU model buffer size =   417.66 MiB
load_all_data: buffer type CUDA_Host is not the default buffer type for device CUDA0 for async uploads
time=2025-11-22T21:16:59.342+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.00"
time=2025-11-22T21:17:04.372+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.03"
time=2025-11-22T21:17:04.623+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.03"
time=2025-11-22T21:17:04.876+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.03"
time=2025-11-22T21:17:05.380+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.04"
time=2025-11-22T21:17:06.136+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.04"
time=2025-11-22T21:17:06.639+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.05"
time=2025-11-22T21:17:06.891+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.05"
time=2025-11-22T21:17:07.647+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.05"
time=2025-11-22T21:17:08.402+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.06"
time=2025-11-22T21:17:08.906+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.06"
time=2025-11-22T21:17:09.158+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.07"
time=2025-11-22T21:17:09.914+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.07"
time=2025-11-22T21:17:10.669+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08"
time=2025-11-22T21:17:11.272+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08"
time=2025-11-22T21:17:11.524+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08"
time=2025-11-22T21:17:11.776+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08"
time=2025-11-22T21:17:12.280+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.09"
time=2025-11-22T21:17:13.036+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.09"
time=2025-11-22T21:17:13.540+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10"
time=2025-11-22T21:17:13.793+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10"
time=2025-11-22T21:17:14.045+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10"
time=2025-11-22T21:17:14.550+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10"
time=2025-11-22T21:17:15.306+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.11"
time=2025-11-22T21:17:15.811+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.11"
time=2025-11-22T21:17:16.064+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.11"
time=2025-11-22T21:17:16.820+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.12"
time=2025-11-22T21:17:17.577+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.12"
time=2025-11-22T21:17:18.081+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.13"
time=2025-11-22T21:17:18.335+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.13"
time=2025-11-22T21:17:19.092+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.13"
time=2025-11-22T21:17:19.849+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.14"
time=2025-11-22T21:17:20.353+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.14"
time=2025-11-22T21:17:20.607+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.14"
time=2025-11-22T21:17:21.123+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.15"
time=2025-11-22T21:17:22.134+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.15"
time=2025-11-22T21:17:22.641+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.16"
time=2025-11-22T21:17:22.894+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.16"
time=2025-11-22T21:17:23.400+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.16"
time=2025-11-22T21:17:46.032+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.32"
time=2025-11-22T21:17:46.286+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.33"
time=2025-11-22T21:17:46.790+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.33"
time=2025-11-22T21:17:47.044+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.33"
time=2025-11-22T21:17:47.549+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.34"
time=2025-11-22T21:17:48.307+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.34"
time=2025-11-22T21:17:48.812+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.35"
time=2025-11-22T21:17:49.065+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.35"
time=2025-11-22T21:17:49.823+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.35"
time=2025-11-22T21:17:50.327+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36"
time=2025-11-22T21:17:50.833+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36"
time=2025-11-22T21:17:51.086+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36"
time=2025-11-22T21:17:51.339+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36"
time=2025-11-22T21:17:51.845+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.37"
time=2025-11-22T21:17:52.855+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.37"
time=2025-11-22T21:17:53.360+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.37"
time=2025-11-22T21:17:53.613+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38"
time=2025-11-22T21:17:53.866+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38"
time=2025-11-22T21:17:54.372+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38"
time=2025-11-22T21:17:54.879+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38"
time=2025-11-22T21:17:55.637+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.39"
time=2025-11-22T21:17:55.890+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.39"
time=2025-11-22T21:17:56.395+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.39"
time=2025-11-22T21:17:56.901+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40"
time=2025-11-22T21:17:57.406+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40"
time=2025-11-22T21:17:57.659+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40"
time=2025-11-22T21:17:57.912+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40"
time=2025-11-22T21:17:58.418+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.41"
time=2025-11-22T21:17:59.175+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.41"
time=2025-11-22T21:17:59.682+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.42"
time=2025-11-22T21:17:59.935+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.42"
time=2025-11-22T21:18:00.441+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.42"
time=2025-11-22T21:18:01.200+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.43"
time=2025-11-22T21:18:01.705+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.43"
time=2025-11-22T21:18:01.958+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.43"
time=2025-11-22T21:18:02.464+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.44"
time=2025-11-22T21:18:02.969+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.44"
time=2025-11-22T21:18:03.475+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.45"
time=2025-11-22T21:18:03.728+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.45"
time=2025-11-22T21:18:04.487+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.45"
time=2025-11-22T21:18:05.246+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.46"
time=2025-11-22T21:18:05.752+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.46"
time=2025-11-22T21:18:06.006+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.46"
time=2025-11-22T21:18:06.513+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.47"
time=2025-11-22T21:18:07.019+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.47"
time=2025-11-22T21:18:07.525+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.48"
time=2025-11-22T21:18:07.778+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.48"
time=2025-11-22T21:18:08.538+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.48"
time=2025-11-22T21:18:09.044+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.49"
time=2025-11-22T21:18:09.550+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.49"
time=2025-11-22T21:18:09.803+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.49"
time=2025-11-22T21:18:10.309+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.50"
time=2025-11-22T21:18:11.321+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.50"
time=2025-11-22T21:18:11.828+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51"
time=2025-11-22T21:18:12.082+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51"
time=2025-11-22T21:18:12.588+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51"
time=2025-11-22T21:18:13.095+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51"
time=2025-11-22T21:18:13.856+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.52"
time=2025-11-22T21:18:14.110+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.52"
time=2025-11-22T21:18:14.617+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.52"
time=2025-11-22T21:18:15.124+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53"
time=2025-11-22T21:18:15.631+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53"
time=2025-11-22T21:18:15.885+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53"
time=2025-11-22T21:18:16.139+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53"
time=2025-11-22T21:18:16.645+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.54"
time=2025-11-22T21:18:17.404+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.54"
time=2025-11-22T21:18:17.911+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.55"
time=2025-11-22T21:18:18.165+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.55"
time=2025-11-22T21:18:18.672+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.55"
time=2025-11-22T21:18:19.433+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.56"
time=2025-11-22T21:18:19.940+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.56"
time=2025-11-22T21:18:20.195+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.56"
time=2025-11-22T21:18:20.702+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.57"
time=2025-11-22T21:18:21.209+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.57"
time=2025-11-22T21:18:21.968+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.58"
time=2025-11-22T21:18:22.222+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.58"
time=2025-11-22T21:18:22.983+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.58"
time=2025-11-22T21:18:23.744+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.59"
time=2025-11-22T21:18:24.504+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.59"
time=2025-11-22T21:18:24.759+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.59"
time=2025-11-22T21:18:25.520+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.60"
time=2025-11-22T21:18:26.027+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.60"
time=2025-11-22T21:18:26.788+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.61"
time=2025-11-22T21:18:27.042+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.61"
time=2025-11-22T21:18:27.549+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.61"
time=2025-11-22T21:18:28.057+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62"
time=2025-11-22T21:18:28.564+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62"
time=2025-11-22T21:18:28.818+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62"
time=2025-11-22T21:18:29.073+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62"
time=2025-11-22T21:18:29.580+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.63"
time=2025-11-22T21:18:30.341+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.63"
time=2025-11-22T21:18:30.850+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.64"
time=2025-11-22T21:18:31.104+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.64"
time=2025-11-22T21:18:31.611+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.64"
time=2025-11-22T21:18:32.371+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65"
time=2025-11-22T21:18:32.879+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65"
time=2025-11-22T21:18:33.133+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65"
time=2025-11-22T21:18:33.641+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65"
time=2025-11-22T21:18:34.148+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.66"
time=2025-11-22T21:18:34.657+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.66"
time=2025-11-22T21:18:34.912+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.66"
time=2025-11-22T21:18:35.420+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.67"
time=2025-11-22T21:18:36.434+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.67"
time=2025-11-22T21:18:36.942+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68"
time=2025-11-22T21:18:37.196+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68"
time=2025-11-22T21:18:37.450+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68"
time=2025-11-22T21:18:37.958+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68"
time=2025-11-22T21:18:38.466+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.69"
time=2025-11-22T21:18:38.973+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.69"
time=2025-11-22T21:18:39.228+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.69"
time=2025-11-22T21:18:39.735+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.70"
time=2025-11-22T21:18:40.499+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.70"
time=2025-11-22T21:18:41.007+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.71"
time=2025-11-22T21:18:41.261+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.71"
time=2025-11-22T21:18:41.771+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.71"
time=2025-11-22T21:18:42.533+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72"
time=2025-11-22T21:18:43.042+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72"
time=2025-11-22T21:18:43.298+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72"
time=2025-11-22T21:18:43.552+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72"
time=2025-11-22T21:18:44.060+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.73"
time=2025-11-22T21:18:44.567+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.73"
time=2025-11-22T21:18:45.076+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.74"
time=2025-11-22T21:18:45.330+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.74"
time=2025-11-22T21:18:45.838+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.74"
time=2025-11-22T21:18:46.601+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.75"
time=2025-11-22T21:18:47.110+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.75"
time=2025-11-22T21:18:47.365+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.75"
time=2025-11-22T21:18:48.127+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.76"
time=2025-11-22T21:18:48.890+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.76"
time=2025-11-22T21:18:49.399+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.77"
time=2025-11-22T21:18:49.653+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.77"
time=2025-11-22T21:18:50.415+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.77"
time=2025-11-22T21:18:50.924+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78"
time=2025-11-22T21:18:51.433+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78"
time=2025-11-22T21:18:51.688+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78"
time=2025-11-22T21:18:51.943+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78"
time=2025-11-22T21:18:52.451+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78"
time=2025-11-22T21:18:52.960+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.79"
time=2025-11-22T21:18:53.723+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.79"
time=2025-11-22T21:18:53.978+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.79"
time=2025-11-22T21:18:54.486+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.80"
time=2025-11-22T21:18:55.503+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.80"
time=2025-11-22T21:18:56.012+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.81"
time=2025-11-22T21:18:56.267+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.81"
time=2025-11-22T21:18:57.030+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.81"
time=2025-11-22T21:18:57.540+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.82"
time=2025-11-22T21:18:58.303+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.82"
time=2025-11-22T21:18:58.559+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.82"
time=2025-11-22T21:18:59.069+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.83"
time=2025-11-22T21:18:59.578+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.83"
time=2025-11-22T21:19:00.088+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.84"
time=2025-11-22T21:19:00.342+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.84"
time=2025-11-22T21:19:00.852+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.84"
time=2025-11-22T21:19:01.616+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.85"
time=2025-11-22T21:19:02.127+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.85"
time=2025-11-22T21:19:02.382+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.85"
time=2025-11-22T21:19:03.146+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.86"
time=2025-11-22T21:19:03.911+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.86"
time=2025-11-22T21:19:04.421+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.87"
time=2025-11-22T21:19:04.677+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.87"
time=2025-11-22T21:19:05.186+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.87"
time=2025-11-22T21:19:05.950+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.88"
time=2025-11-22T21:19:06.459+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.88"
time=2025-11-22T21:19:06.715+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.88"
time=2025-11-22T21:19:06.970+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.89"
time=2025-11-22T21:19:07.481+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.89"
time=2025-11-22T21:19:08.246+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.90"
time=2025-11-22T21:19:09.012+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.90"
time=2025-11-22T21:19:09.267+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.90"
time=2025-11-22T21:19:09.775+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.91"
time=2025-11-22T21:19:10.795+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.91"
time=2025-11-22T21:19:11.305+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.92"
time=2025-11-22T21:19:11.560+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.92"
time=2025-11-22T21:19:12.071+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.92"
time=2025-11-22T21:19:12.836+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.93"
time=2025-11-22T21:19:13.602+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.93"
time=2025-11-22T21:19:13.857+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.93"
time=2025-11-22T21:19:14.368+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.94"
time=2025-11-22T21:19:15.133+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.94"
time=2025-11-22T21:19:15.643+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.95"
time=2025-11-22T21:19:15.898+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.95"
time=2025-11-22T21:19:16.410+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.95"
time=2025-11-22T21:19:17.175+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96"
time=2025-11-22T21:19:17.686+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96"
time=2025-11-22T21:19:17.941+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96"
time=2025-11-22T21:19:18.197+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96"
time=2025-11-22T21:19:18.708+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.97"
time=2025-11-22T21:19:19.473+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.97"
load_all_data: no device found for buffer type CPU for async uploads
time=2025-11-22T21:19:19.983+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.98"
llama_context: constructing llama_context
llama_context: n_seq_max     = 3
llama_context: n_ctx         = 98304
llama_context: n_ctx_per_seq = 32768
llama_context: n_batch       = 1536
llama_context: n_ubatch      = 512
llama_context: causal_attn   = 1
llama_context: flash_attn    = disabled
llama_context: kv_unified    = false
llama_context: freq_base     = 1000000.0
llama_context: freq_scale    = 1
set_abort_callback: call
llama_context:        CPU  output buffer size =     1.80 MiB
create_memory: n_ctx = 98304 (padded)
llama_kv_cache: layer   0: dev = CPU
llama_kv_cache: layer   1: dev = CPU
llama_kv_cache: layer   2: dev = CPU
llama_kv_cache: layer   3: dev = CPU
llama_kv_cache: layer   4: dev = CPU
llama_kv_cache: layer   5: dev = CPU
llama_kv_cache: layer   6: dev = CPU
llama_kv_cache: layer   7: dev = CPU
llama_kv_cache: layer   8: dev = CPU
llama_kv_cache: layer   9: dev = CPU
llama_kv_cache: layer  10: dev = CPU
llama_kv_cache: layer  11: dev = CPU
llama_kv_cache: layer  12: dev = CPU
llama_kv_cache: layer  13: dev = CPU
llama_kv_cache: layer  14: dev = CPU
llama_kv_cache: layer  15: dev = CPU
llama_kv_cache: layer  16: dev = CPU
llama_kv_cache: layer  17: dev = CPU
llama_kv_cache: layer  18: dev = CPU
llama_kv_cache: layer  19: dev = CPU
llama_kv_cache: layer  20: dev = CPU
llama_kv_cache: layer  21: dev = CPU
llama_kv_cache: layer  22: dev = CPU
llama_kv_cache: layer  23: dev = CPU
llama_kv_cache: layer  24: dev = CPU
llama_kv_cache: layer  25: dev = CPU
llama_kv_cache: layer  26: dev = CPU
llama_kv_cache: layer  27: dev = CPU
llama_kv_cache: layer  28: dev = CPU
llama_kv_cache: layer  29: dev = CPU
llama_kv_cache: layer  30: dev = CPU
llama_kv_cache: layer  31: dev = CPU
llama_kv_cache: layer  32: dev = CPU
llama_kv_cache: layer  33: dev = CPU
llama_kv_cache: layer  34: dev = CPU
llama_kv_cache: layer  35: dev = CPU
llama_kv_cache: layer  36: dev = CPU
llama_kv_cache: layer  37: dev = CPU
llama_kv_cache: layer  38: dev = CPU
llama_kv_cache: layer  39: dev = CPU
llama_kv_cache: layer  40: dev = CPU
llama_kv_cache: layer  41: dev = CPU
llama_kv_cache: layer  42: dev = CPU
llama_kv_cache: layer  43: dev = CPU
llama_kv_cache: layer  44: dev = CPU
llama_kv_cache: layer  45: dev = CPU
llama_kv_cache: layer  46: dev = CPU
llama_kv_cache: layer  47: dev = CPU
llama_kv_cache: layer  48: dev = CPU
llama_kv_cache: layer  49: dev = CPU
llama_kv_cache: layer  50: dev = CPU
llama_kv_cache: layer  51: dev = CPU
llama_kv_cache: layer  52: dev = CPU
llama_kv_cache: layer  53: dev = CPU
llama_kv_cache: layer  54: dev = CPU
llama_kv_cache: layer  55: dev = CPU
llama_kv_cache: layer  56: dev = CPU
llama_kv_cache: layer  57: dev = CPU
llama_kv_cache: layer  58: dev = CPU
llama_kv_cache: layer  59: dev = CPU
llama_kv_cache: layer  60: dev = CPU
llama_kv_cache: layer  61: dev = CPU
llama_kv_cache: layer  62: dev = CPU
llama_kv_cache: layer  63: dev = CPU
llama_kv_cache:        CPU KV buffer size = 24576.00 MiB
time=2025-11-22T21:19:23.555+01:00 level=DEBUG source=server.go:1338 msg="model load progress 1.00"
time=2025-11-22T21:19:23.812+01:00 level=DEBUG source=server.go:1341 msg="model load completed, waiting for server to become available" status="llm server loading model"
llama_kv_cache: size = 24576.00 MiB ( 32768 cells,  64 layers,  3/3 seqs), K (f16): 12288.00 MiB, V (f16): 12288.00 MiB
llama_context: enumerating backends
llama_context: backend_ptrs.size() = 4
llama_context: max_nodes = 6168
llama_context: reserving full memory module
llama_context: worst-case: n_tokens = 512, n_seqs = 3, n_outputs = 3
graph_reserve: reserving a graph for ubatch with n_tokens =  512, n_seqs =  3, n_outputs =  512
graph_reserve: making n_tokens a multiple of n_seqs - n_tokens = 513, n_seqs = 3, n_outputs = 512
graph_reserve: reserving a graph for ubatch with n_tokens =    3, n_seqs =  3, n_outputs =    3
graph_reserve: reserving a graph for ubatch with n_tokens =  512, n_seqs =  3, n_outputs =  512
graph_reserve: making n_tokens a multiple of n_seqs - n_tokens = 513, n_seqs = 3, n_outputs = 512
llama_context:      CUDA0 compute buffer size =  3051.08 MiB
llama_context:  CUDA_Host compute buffer size =    86.04 MiB
llama_context: graph nodes  = 2566
llama_context: graph splits = 900 (with bs=512), 1 (with bs=1)
time=2025-11-22T21:21:15.211+01:00 level=INFO source=server.go:1332 msg="llama runner started in 261.95 seconds"
time=2025-11-22T21:21:15.213+01:00 level=INFO source=sched.go:517 msg="loaded runners" count=1
time=2025-11-22T21:21:15.215+01:00 level=INFO source=server.go:1294 msg="waiting for llama runner to start responding"
time=2025-11-22T21:21:15.223+01:00 level=INFO source=server.go:1332 msg="llama runner started in 261.97 seconds"
time=2025-11-22T21:21:15.224+01:00 level=DEBUG source=sched.go:529 msg="finished setting up" runner.name=registry.ollama.ai/library/qwen2.5-coder:32b runner.size="70.8 GiB" runner.vram="28.7 GiB" runner.parallel=3 runner.pid=77624 runner.model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 runner.num_ctx=131072
time=2025-11-22T21:21:15.229+01:00 level=DEBUG source=server.go:1465 msg="completion request" images=0 prompt=1586 format=""
Image
<!-- gh-comment-id:3567044996 --> @catclaw commented on GitHub (Nov 22, 2025): Well... The new 0.13.0 version finally detects my GPUs, but still loads models into RAM and uses the CPU. (~70-80% - please se attached image at the end of this post.) (No, there are no other ggml-base.dll in PATH; ComfyUI, SwarmUI, Whisper, etc. are all in their own folders outside the path.) Also, when I run ollama serve, it seems like my Nvidia drivers crash and restart, as all my 4 monitors go black, and the same thing happens as when I restart the drivers manually by pressing CTRL+SHIFT+WIN+B. ``` time=2025-11-22T20:56:39.151+01:00 level=INFO source=routes.go:1544 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:3 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:3 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true OLLAMA_VULKAN:true ROCR_VISIBLE_DEVICES:]" time=2025-11-22T20:56:39.431+01:00 level=INFO source=images.go:522 msg="total blobs: 146" time=2025-11-22T20:56:39.449+01:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-11-22T20:56:39.463+01:00 level=INFO source=routes.go:1597 msg="Listening on 127.0.0.1:11434 (version 0.13.0)" time=2025-11-22T20:56:39.464+01:00 level=DEBUG source=sched.go:120 msg="starting llm scheduler" time=2025-11-22T20:56:39.471+01:00 level=INFO source=runner.go:67 msg="discovering available GPUs..." time=2025-11-22T20:56:39.471+01:00 level=WARN source=runner.go:470 msg="user overrode visible devices" CUDA_VISIBLE_DEVICES=0,1,2 time=2025-11-22T20:56:39.472+01:00 level=WARN source=runner.go:474 msg="if GPUs are not correctly discovered, unset and try again" time=2025-11-22T20:56:39.472+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[] time=2025-11-22T20:56:39.520+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 58193" time=2025-11-22T20:56:39.521+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-11-22T20:56:39.718+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine" time=2025-11-22T20:56:39.722+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:58193" time=2025-11-22T20:56:39.734+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:39.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0 time=2025-11-22T20:56:39.736+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default="" time=2025-11-22T20:56:39.736+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default="" time=2025-11-22T20:56:39.737+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-22T20:56:39.737+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-22T20:56:39.788+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll time=2025-11-22T20:56:40.050+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) time=2025-11-22T20:56:40.051+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-22T20:56:40.052+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=320.1729ms ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 9020006400 total: 12820938752 ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11452166144 total: 12884901888 ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11689582592 total: 12884901888 time=2025-11-22T20:56:40.458+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=405.3281ms time=2025-11-22T20:56:40.461+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:9020006400 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11452166144 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11689582592 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]}]" time=2025-11-22T20:56:40.466+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=994.0386ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[] time=2025-11-22T20:56:40.485+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extraEnvs=map[] time=2025-11-22T20:56:40.491+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 57842" time=2025-11-22T20:56:40.493+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-11-22T20:56:40.704+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine" time=2025-11-22T20:56:40.708+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:57842" time=2025-11-22T20:56:40.714+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string time=2025-11-22T20:56:40.714+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string time=2025-11-22T20:56:40.715+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:40.715+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0 time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default="" time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default="" time=2025-11-22T20:56:40.716+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-22T20:56:40.716+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-22T20:56:40.764+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" ggml_cuda_init: failed to initialize ROCm: no ROCm-capable device is detected load_backend: loaded ROCm backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\ggml-hip.dll time=2025-11-22T20:56:40.834+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(clang) time=2025-11-22T20:56:40.834+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=122.4886ms time=2025-11-22T20:56:40.835+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=0s time=2025-11-22T20:56:40.838+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-11-22T20:56:40.859+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=374.0745ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=map[] time=2025-11-22T20:56:40.879+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]" extraEnvs=map[] time=2025-11-22T20:56:40.886+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 58909" time=2025-11-22T20:56:40.901+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" time=2025-11-22T20:56:41.103+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine" time=2025-11-22T20:56:41.107+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:58909" time=2025-11-22T20:56:41.116+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0 time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default="" time=2025-11-22T20:56:41.117+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default="" time=2025-11-22T20:56:41.117+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-22T20:56:41.118+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-22T20:56:41.165+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" ggml_vulkan: Found 3 Vulkan devices: ggml_vulkan: 0 = NVIDIA GeForce RTX 5070 (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2 ggml_vulkan: 1 = NVIDIA GeForce RTX 3060 (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2 ggml_vulkan: 2 = NVIDIA GeForce RTX 3060 (NVIDIA) | uma: 0 | fp16: 1 | bf16: 1 | warp size: 32 | shared memory: 49152 | int dot: 1 | matrix cores: NV_coopmat2 load_backend: loaded Vulkan backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\vulkan\ggml-vulkan.dll time=2025-11-22T20:56:41.749+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(clang) time=2025-11-22T20:56:42.164+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 ggml_backend_vk_get_device_memory called: uuid 7c7493fb-894e-ffa3-0d90-00b83a5731e8 ggml_backend_vk_get_device_memory called: luid 0x00000000000814dc ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 5070) with LUID 0x00000000000814dc detected. Dedicated Total: 12430868480.00 bytes (11.58 GB), Dedicated Usage: 3487313920.00 bytes (3.25 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 8943554560 total: 12430868480 ggml_backend_vk_get_device_memory called: uuid 67bf118a-52b6-0f51-c4ad-c561cdfbb20c ggml_backend_vk_get_device_memory called: luid 0x0000000000084879 ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000084879 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 917102592.00 bytes (0.85 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11786395648 total: 12703498240 ggml_backend_vk_get_device_memory called: uuid e2200bae-da5e-037d-9fc2-d0d8c6081ced ggml_backend_vk_get_device_memory called: luid 0x0000000000083770 ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000083770 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 1154629632.00 bytes (1.08 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11548868608 total: 12703498240 time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-22T20:56:43.370+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-22T20:56:43.371+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=2.2566731s ggml_backend_vk_get_device_memory called: uuid 7c7493fb-894e-ffa3-0d90-00b83a5731e8 ggml_backend_vk_get_device_memory called: luid 0x00000000000814dc ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 5070) with LUID 0x00000000000814dc detected. Dedicated Total: 12430868480.00 bytes (11.58 GB), Dedicated Usage: 3439132672.00 bytes (3.20 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 8991735808 total: 12430868480 ggml_backend_vk_get_device_memory called: uuid 7c7493fb-894e-ffa3-0d90-00b83a5731e8 ggml_backend_vk_get_device_memory called: luid 0x00000000000814dc ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 5070) with LUID 0x00000000000814dc detected. Dedicated Total: 12430868480.00 bytes (11.58 GB), Dedicated Usage: 3439132672.00 bytes (3.20 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 8991735808 total: 12430868480 ggml_backend_vk_get_device_memory called: uuid 67bf118a-52b6-0f51-c4ad-c561cdfbb20c ggml_backend_vk_get_device_memory called: luid 0x0000000000084879 ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000084879 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 917102592.00 bytes (0.85 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11786395648 total: 12703498240 ggml_backend_vk_get_device_memory called: uuid 67bf118a-52b6-0f51-c4ad-c561cdfbb20c ggml_backend_vk_get_device_memory called: luid 0x0000000000084879 ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000084879 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 917102592.00 bytes (0.85 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11786395648 total: 12703498240 ggml_backend_vk_get_device_memory called: uuid e2200bae-da5e-037d-9fc2-d0d8c6081ced ggml_backend_vk_get_device_memory called: luid 0x0000000000083770 ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000083770 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 1154629632.00 bytes (1.08 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11548868608 total: 12703498240 ggml_backend_vk_get_device_memory called: uuid e2200bae-da5e-037d-9fc2-d0d8c6081ced ggml_backend_vk_get_device_memory called: luid 0x0000000000083770 ggml_dxgi_pdh_init called DXGI + PDH Initialized. Getting GPU free memory info [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000814DC, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000084879, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 3060, LUID: 0x0000000000083770, Dedicated: 11.83 GB, Shared: 63.95 GB [DXGI] Adapter Description: NVIDIA GeForce RTX 5070, LUID: 0x00000000000913FB, Dedicated: 11.58 GB, Shared: 63.95 GB [DXGI] Adapter Description: Microsoft Basic Render Driver, LUID: 0x000000000008371F, Dedicated: 0.00 GB, Shared: 63.95 GB Discrete GPU (NVIDIA GeForce RTX 3060) with LUID 0x0000000000083770 detected. Dedicated Total: 12703498240.00 bytes (11.83 GB), Dedicated Usage: 1154629632.00 bytes (1.08 GB) ggml_backend_vk_get_device_memory utilizing DXGI + PDH memory reporting free: 11548868608 total: 12703498240 time=2025-11-22T20:56:44.143+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=771.8209ms time=2025-11-22T20:56:44.145+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]" devices="[{DeviceID:{ID:7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:Vulkan} Name:Vulkan0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12430868480 FreeMemory:8991735808 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]} {DeviceID:{ID:67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:Vulkan} Name:Vulkan1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12703498240 FreeMemory:11786395648 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]} {DeviceID:{ID:e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:Vulkan} Name:Vulkan2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12703498240 FreeMemory:11548868608 ComputeMajor:0 ComputeMinor:0 DriverMajor:0 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]}]" time=2025-11-22T20:56:44.153+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=3.2735283s OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan]" extra_envs=map[] time=2025-11-22T20:56:44.242+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extraEnvs=map[] time=2025-11-22T20:56:44.253+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 57435" time=2025-11-22T20:56:44.254+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-11-22T20:56:44.443+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine" time=2025-11-22T20:56:44.448+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:57435" time=2025-11-22T20:56:44.455+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string time=2025-11-22T20:56:44.455+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0 time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default="" time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default="" time=2025-11-22T20:56:44.456+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-22T20:56:44.456+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-22T20:56:44.502+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12\ggml-cuda.dll time=2025-11-22T20:56:44.734+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=500,520,600,610,700,750,800,860,890,900,1200 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-22T20:56:44.735+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=281.9024ms ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 9015746560 total: 12820938752 ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11452166144 total: 12884901888 ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11689582592 total: 12884901888 time=2025-11-22T20:56:45.197+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=461.3986ms time=2025-11-22T20:56:45.199+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:9015746560 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11452166144 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11689582592 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]}]" time=2025-11-22T20:56:45.206+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=964.5419ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=map[] time=2025-11-22T20:56:45.229+01:00 level=DEBUG source=runner.go:120 msg="evaluating which, if any, devices to filter out" initial_count=9 time=2025-11-22T20:56:45.231+01:00 level=TRACE source=runner.go:160 msg="supported GPU library combinations before filtering" supported=map[] time=2025-11-22T20:56:45.232+01:00 level=DEBUG source=runner.go:213 msg="dropping duplicate device" id=7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=Vulkan compute=0.0 name=Vulkan0 description="NVIDIA GeForce RTX 5070" libdirs="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama,C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" driver=0.0 pci_id=0000:03:00.0 type=discrete total="11.6 GiB" available="8.4 GiB" time=2025-11-22T20:56:45.234+01:00 level=DEBUG source=runner.go:213 msg="dropping duplicate device" id=e2200bae-da5e-037d-9fc2-d0d8c6081ced library=Vulkan compute=0.0 name=Vulkan2 description="NVIDIA GeForce RTX 3060" libdirs="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama,C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" driver=0.0 pci_id=0000:04:00.0 type=discrete total="11.8 GiB" available="10.8 GiB" time=2025-11-22T20:56:45.235+01:00 level=DEBUG source=runner.go:213 msg="dropping duplicate device" id=67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=Vulkan compute=0.0 name=Vulkan1 description="NVIDIA GeForce RTX 3060" libdirs="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama,C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\vulkan" driver=0.0 pci_id=0000:83:00.0 type=discrete total="11.8 GiB" available="11.0 GiB" time=2025-11-22T20:56:45.242+01:00 level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=5.7781411s time=2025-11-22T20:56:45.244+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c filter_id="" library=CUDA compute=8.6 name=CUDA2 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v13 driver=13.0 pci_id=0000:83:00.0 type=discrete total="12.0 GiB" available="10.9 GiB" time=2025-11-22T20:56:45.245+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced filter_id="" library=CUDA compute=8.6 name=CUDA1 description="NVIDIA GeForce RTX 3060" libdirs=ollama,cuda_v13 driver=13.0 pci_id=0000:04:00.0 type=discrete total="12.0 GiB" available="10.7 GiB" time=2025-11-22T20:56:45.246+01:00 level=INFO source=types.go:42 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 filter_id="" library=CUDA compute=12.0 name=CUDA0 description="NVIDIA GeForce RTX 5070" libdirs=ollama,cuda_v13 driver=13.0 pci_id=0000:03:00.0 type=discrete total="11.9 GiB" available="8.4 GiB" [GIN] 2025/11/22 - 20:56:57 | 200 | 0s | 127.0.0.1 | HEAD "/" [GIN] 2025/11/22 - 20:56:57 | 200 | 44.7027ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/11/22 - 21:15:48 | 200 | 28.5907ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/11/22 - 21:15:49 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/11/22 - 21:15:53 | 200 | 26.5895ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/11/22 - 21:15:53 | 200 | 0s | 127.0.0.1 | GET "/api/ps" time=2025-11-22T21:16:51.500+01:00 level=TRACE source=sched.go:146 msg="processing incoming request" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 time=2025-11-22T21:16:51.501+01:00 level=TRACE source=sched.go:179 msg="refreshing GPU list" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 time=2025-11-22T21:16:51.505+01:00 level=DEBUG source=runner.go:250 msg="refreshing free memory" time=2025-11-22T21:16:51.506+01:00 level=DEBUG source=runner.go:314 msg="unable to refresh all GPUs with existing runners, performing bootstrap discovery" time=2025-11-22T21:16:51.507+01:00 level=TRACE source=runner.go:425 msg="starting runner for device discovery" libDirs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extraEnvs=map[] time=2025-11-22T21:16:51.547+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 56402" time=2025-11-22T21:16:51.549+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-11-22T21:16:51.766+01:00 level=INFO source=runner.go:1398 msg="starting ollama engine" time=2025-11-22T21:16:51.771+01:00 level=INFO source=runner.go:1433 msg="Server listening on 127.0.0.1:56402" time=2025-11-22T21:16:51.777+01:00 level=DEBUG source=gguf.go:589 msg=general.architecture type=string time=2025-11-22T21:16:51.778+01:00 level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.file_type default=0 time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.name default="" time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.description default="" time=2025-11-22T21:16:51.779+01:00 level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-11-22T21:16:51.779+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-22T21:16:51.830+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll time=2025-11-22T21:16:52.124+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) time=2025-11-22T21:16:52.124+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.pooling_type default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.expert_count default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.block_count default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.embedding_length default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-11-22T21:16:52.125+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-11-22T21:16:52.126+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-11-22T21:16:52.126+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-11-22T21:16:52.126+01:00 level=DEBUG source=runner.go:1373 msg="dummy model load took" duration=349.5975ms ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8893997056 total: 12820938752 ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11387678720 total: 12884901888 ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11551105024 total: 12884901888 time=2025-11-22T21:16:52.530+01:00 level=DEBUG source=runner.go:1378 msg="gathering device infos took" duration=403.6765ms time=2025-11-22T21:16:52.533+01:00 level=TRACE source=runner.go:452 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices="[{DeviceID:{ID:GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Library:CUDA} Name:CUDA0 Description:NVIDIA GeForce RTX 5070 FilterID: Integrated:false PCIID:0000:03:00.0 TotalMemory:12820938752 FreeMemory:8893997056 ComputeMajor:12 ComputeMinor:0 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Library:CUDA} Name:CUDA1 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:04:00.0 TotalMemory:12884901888 FreeMemory:11387678720 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]} {DeviceID:{ID:GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c Library:CUDA} Name:CUDA2 Description:NVIDIA GeForce RTX 3060 FilterID: Integrated:false PCIID:0000:83:00.0 TotalMemory:12884901888 FreeMemory:11551105024 ComputeMajor:8 ComputeMinor:6 DriverMajor:13 DriverMinor:0 LibraryPath:[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]}]" time=2025-11-22T21:16:52.540+01:00 level=DEBUG source=runner.go:422 msg="bootstrap discovery took" duration=1.0331979s OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=map[] time=2025-11-22T21:16:52.563+01:00 level=DEBUG source=runner.go:40 msg="overall device VRAM discovery took" duration=1.0581286s time=2025-11-22T21:16:52.565+01:00 level=TRACE source=sched.go:182 msg="refreshing system information" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 time=2025-11-22T21:16:52.566+01:00 level=TRACE source=gpu.go:22 msg="performing CPU discovery" time=2025-11-22T21:16:52.569+01:00 level=INFO source=cpu_windows.go:148 msg=packages count=2 time=2025-11-22T21:16:52.571+01:00 level=INFO source=cpu_windows.go:195 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-11-22T21:16:52.573+01:00 level=INFO source=cpu_windows.go:195 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-11-22T21:16:52.574+01:00 level=TRACE source=gpu.go:25 msg="CPU discovery completed" duration=4.9149ms time=2025-11-22T21:16:52.575+01:00 level=DEBUG source=sched.go:194 msg="updating default concurrency" OLLAMA_MAX_LOADED_MODELS=9 gpu_count=3 time=2025-11-22T21:16:52.577+01:00 level=TRACE source=sched.go:198 msg="loading model metadata" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 time=2025-11-22T21:16:52.604+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=general.alignment default=32 time=2025-11-22T21:16:52.617+01:00 level=TRACE source=sched.go:206 msg="updating free space" gpu_count=3 model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 time=2025-11-22T21:16:52.637+01:00 level=DEBUG source=sched.go:211 msg="loading first model" model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 llama_model_loader: loaded meta data with 34 key-value pairs and 771 tensors from H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen2 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen2.5 Coder 32B Instruct llama_model_loader: - kv 3: general.finetune str = Instruct llama_model_loader: - kv 4: general.basename str = Qwen2.5-Coder llama_model_loader: - kv 5: general.size_label str = 32B llama_model_loader: - kv 6: general.license str = apache-2.0 llama_model_loader: - kv 7: general.license.link str = https://huggingface.co/Qwen/Qwen2.5-C... llama_model_loader: - kv 8: general.base_model.count u32 = 1 llama_model_loader: - kv 9: general.base_model.0.name str = Qwen2.5 Coder 32B llama_model_loader: - kv 10: general.base_model.0.organization str = Qwen llama_model_loader: - kv 11: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen2.5-C... llama_model_loader: - kv 12: general.tags arr[str,6] = ["code", "codeqwen", "chat", "qwen", ... llama_model_loader: - kv 13: general.languages arr[str,1] = ["en"] llama_model_loader: - kv 14: qwen2.block_count u32 = 64 llama_model_loader: - kv 15: qwen2.context_length u32 = 32768 llama_model_loader: - kv 16: qwen2.embedding_length u32 = 5120 llama_model_loader: - kv 17: qwen2.feed_forward_length u32 = 27648 llama_model_loader: - kv 18: qwen2.attention.head_count u32 = 40 llama_model_loader: - kv 19: qwen2.attention.head_count_kv u32 = 8 llama_model_loader: - kv 20: qwen2.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 21: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 22: general.file_type u32 = 15 llama_model_loader: - kv 23: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 24: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 25: tokenizer.ggml.tokens arr[str,152064] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 26: tokenizer.ggml.token_type arr[i32,152064] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 27: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 28: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 29: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 30: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 31: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 32: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 33: general.quantization_version u32 = 2 llama_model_loader: - type f32: 321 tensors llama_model_loader: - type q4_K: 385 tensors llama_model_loader: - type q6_K: 65 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.48 GiB (4.85 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 22 load: token to piece cache size = 0.9310 MB print_info: arch = qwen2 print_info: vocab_only = 1 print_info: model type = ?B print_info: model params = 32.76 B print_info: general.name = Qwen2.5 Coder 32B Instruct print_info: vocab type = BPE print_info: n_vocab = 152064 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 llama_model_load: vocab only - skipping tensors time=2025-11-22T21:16:53.186+01:00 level=WARN source=server.go:167 msg="requested context size too large for model" num_ctx=131072 n_ctx_train=32768 time=2025-11-22T21:16:53.191+01:00 level=INFO source=server.go:392 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --model H:\\LLM\\blobs\\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 --port 58133" time=2025-11-22T21:16:53.196+01:00 level=DEBUG source=server.go:393 msg=subprocess CUDA_VISIBLE_DEVICES=0,1,2 OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MAX_QUEUE=3 OLLAMA_MODELS=H:\LLM OLLAMA_NUM_PARALLEL=3 OLLAMA_SCHED_SPREAD=1 OLLAMA_VULKAN=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Program Files\\Calibre2\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\dotnet\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\Program Files\\Microsoft SQL Server\\170\\Tools\\Binn\\;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-11-22T21:16:53.257+01:00 level=INFO source=sched.go:443 msg="system memory" total="127.9 GiB" free="82.3 GiB" free_swap="453.9 GiB" time=2025-11-22T21:16:53.261+01:00 level=INFO source=sched.go:450 msg="gpu memory" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=CUDA available="7.8 GiB" free="8.3 GiB" minimum="457.0 MiB" overhead="0 B" time=2025-11-22T21:16:53.268+01:00 level=INFO source=sched.go:450 msg="gpu memory" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=CUDA available="10.2 GiB" free="10.6 GiB" minimum="457.0 MiB" overhead="0 B" time=2025-11-22T21:16:53.276+01:00 level=INFO source=sched.go:450 msg="gpu memory" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=CUDA available="10.3 GiB" free="10.8 GiB" minimum="457.0 MiB" overhead="0 B" time=2025-11-22T21:16:53.282+01:00 level=INFO source=server.go:459 msg="loading model" "model layers"=65 requested=-1 time=2025-11-22T21:16:53.287+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=qwen2.attention.key_length default=128 time=2025-11-22T21:16:53.292+01:00 level=DEBUG source=ggml.go:278 msg="key with type not found" key=qwen2.attention.value_length default=128 time=2025-11-22T21:16:53.296+01:00 level=DEBUG source=ggml.go:613 msg="default cache size estimate" "attention MiB"=24576 "attention bytes"=25769803776 "recurrent MiB"=0 "recurrent bytes"=0 time=2025-11-22T21:16:53.321+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=0 size="681.7 MiB" time=2025-11-22T21:16:53.327+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=1 size="681.7 MiB" time=2025-11-22T21:16:53.335+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=2 size="681.7 MiB" time=2025-11-22T21:16:53.342+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=3 size="681.7 MiB" time=2025-11-22T21:16:53.352+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=4 size="681.7 MiB" time=2025-11-22T21:16:53.353+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=5 size="681.7 MiB" time=2025-11-22T21:16:53.356+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=6 size="681.7 MiB" time=2025-11-22T21:16:53.366+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=7 size="681.7 MiB" time=2025-11-22T21:16:53.368+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=8 size="645.6 MiB" time=2025-11-22T21:16:53.370+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=9 size="646.9 MiB" time=2025-11-22T21:16:53.373+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=10 size="645.6 MiB" time=2025-11-22T21:16:53.381+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=11 size="646.9 MiB" time=2025-11-22T21:16:53.384+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=12 size="680.4 MiB" time=2025-11-22T21:16:53.385+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=13 size="680.4 MiB" time=2025-11-22T21:16:53.386+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=14 size="645.6 MiB" time=2025-11-22T21:16:53.390+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=15 size="645.6 MiB" time=2025-11-22T21:16:53.396+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=16 size="681.7 MiB" time=2025-11-22T21:16:53.399+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=17 size="645.6 MiB" time=2025-11-22T21:16:53.400+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=18 size="645.6 MiB" time=2025-11-22T21:16:53.402+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=19 size="681.7 MiB" time=2025-11-22T21:16:53.405+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=20 size="645.6 MiB" time=2025-11-22T21:16:53.411+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=21 size="645.6 MiB" time=2025-11-22T21:16:53.413+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=22 size="681.7 MiB" time=2025-11-22T21:16:53.414+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=23 size="645.6 MiB" time=2025-11-22T21:16:53.416+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=24 size="645.6 MiB" time=2025-11-22T21:16:53.420+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=25 size="681.7 MiB" time=2025-11-22T21:16:53.423+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=26 size="645.6 MiB" time=2025-11-22T21:16:53.428+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=27 size="645.6 MiB" time=2025-11-22T21:16:53.430+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=28 size="681.7 MiB" time=2025-11-22T21:16:53.431+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=29 size="645.6 MiB" time=2025-11-22T21:16:53.438+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=30 size="645.6 MiB" time=2025-11-22T21:16:53.440+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=31 size="681.7 MiB" time=2025-11-22T21:16:53.443+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=32 size="645.6 MiB" time=2025-11-22T21:16:53.444+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=33 size="645.6 MiB" time=2025-11-22T21:16:53.446+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=34 size="681.7 MiB" time=2025-11-22T21:16:53.453+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=35 size="645.6 MiB" time=2025-11-22T21:16:53.455+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=36 size="645.6 MiB" time=2025-11-22T21:16:53.453+01:00 level=INFO source=runner.go:963 msg="starting go runner" time=2025-11-22T21:16:53.455+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-11-22T21:16:53.462+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=37 size="681.7 MiB" time=2025-11-22T21:16:53.474+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=38 size="645.6 MiB" time=2025-11-22T21:16:53.479+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=39 size="645.6 MiB" time=2025-11-22T21:16:53.480+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=40 size="681.7 MiB" time=2025-11-22T21:16:53.484+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=41 size="645.6 MiB" time=2025-11-22T21:16:53.486+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=42 size="645.6 MiB" time=2025-11-22T21:16:53.491+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=43 size="681.7 MiB" time=2025-11-22T21:16:53.496+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=44 size="645.6 MiB" time=2025-11-22T21:16:53.500+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=45 size="645.6 MiB" time=2025-11-22T21:16:53.504+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=46 size="681.7 MiB" time=2025-11-22T21:16:53.506+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=47 size="645.6 MiB" time=2025-11-22T21:16:53.509+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=48 size="645.6 MiB" time=2025-11-22T21:16:53.514+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=49 size="681.7 MiB" time=2025-11-22T21:16:53.516+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=50 size="645.6 MiB" time=2025-11-22T21:16:53.520+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=51 size="645.6 MiB" time=2025-11-22T21:16:53.521+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=52 size="681.7 MiB" time=2025-11-22T21:16:53.530+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=53 size="645.6 MiB" time=2025-11-22T21:16:53.532+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=54 size="645.6 MiB" time=2025-11-22T21:16:53.533+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=55 size="681.7 MiB" time=2025-11-22T21:16:53.536+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=56 size="681.7 MiB" load_backend: loaded CPU backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\ggml-cpu-haswell.dll time=2025-11-22T21:16:53.504+01:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-11-22T21:16:53.538+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=57 size="681.7 MiB" time=2025-11-22T21:16:53.557+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=58 size="681.7 MiB" time=2025-11-22T21:16:53.559+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=59 size="681.7 MiB" time=2025-11-22T21:16:53.561+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=60 size="681.7 MiB" time=2025-11-22T21:16:53.563+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=61 size="681.7 MiB" time=2025-11-22T21:16:53.567+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=62 size="681.7 MiB" time=2025-11-22T21:16:53.572+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=63 size="681.7 MiB" time=2025-11-22T21:16:53.573+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=64 size="609.1 MiB" time=2025-11-22T21:16:53.577+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=CUDA "available layer vram"="9.6 GiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="0 B" time=2025-11-22T21:16:53.581+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=CUDA "available layer vram"="9.5 GiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="0 B" time=2025-11-22T21:16:53.594+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=CUDA "available layer vram"="7.2 GiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="0 B" time=2025-11-22T21:16:53.597+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=0 size="681.7 MiB" time=2025-11-22T21:16:53.602+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=1 size="681.7 MiB" time=2025-11-22T21:16:53.605+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=2 size="681.7 MiB" time=2025-11-22T21:16:53.608+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=3 size="681.7 MiB" time=2025-11-22T21:16:53.612+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=4 size="681.7 MiB" time=2025-11-22T21:16:53.615+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=5 size="681.7 MiB" time=2025-11-22T21:16:53.616+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=6 size="681.7 MiB" time=2025-11-22T21:16:53.618+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=7 size="681.7 MiB" time=2025-11-22T21:16:53.621+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=8 size="645.6 MiB" time=2025-11-22T21:16:53.625+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=9 size="646.9 MiB" time=2025-11-22T21:16:53.626+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=10 size="645.6 MiB" time=2025-11-22T21:16:53.628+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=11 size="646.9 MiB" time=2025-11-22T21:16:53.629+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=12 size="680.4 MiB" time=2025-11-22T21:16:53.630+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=13 size="680.4 MiB" time=2025-11-22T21:16:53.633+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=14 size="645.6 MiB" time=2025-11-22T21:16:53.637+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=15 size="645.6 MiB" time=2025-11-22T21:16:53.638+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=16 size="681.7 MiB" time=2025-11-22T21:16:53.638+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=17 size="645.6 MiB" time=2025-11-22T21:16:53.640+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=18 size="645.6 MiB" time=2025-11-22T21:16:53.641+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=19 size="681.7 MiB" time=2025-11-22T21:16:53.643+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=20 size="645.6 MiB" time=2025-11-22T21:16:53.644+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=21 size="645.6 MiB" time=2025-11-22T21:16:53.648+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=22 size="681.7 MiB" time=2025-11-22T21:16:53.651+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=23 size="645.6 MiB" time=2025-11-22T21:16:53.653+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=24 size="645.6 MiB" time=2025-11-22T21:16:53.655+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=25 size="681.7 MiB" time=2025-11-22T21:16:53.659+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=26 size="645.6 MiB" time=2025-11-22T21:16:53.660+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=27 size="645.6 MiB" time=2025-11-22T21:16:53.662+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=28 size="681.7 MiB" time=2025-11-22T21:16:53.663+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=29 size="645.6 MiB" time=2025-11-22T21:16:53.665+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=30 size="645.6 MiB" time=2025-11-22T21:16:53.666+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=31 size="681.7 MiB" time=2025-11-22T21:16:53.669+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=32 size="645.6 MiB" time=2025-11-22T21:16:53.670+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=33 size="645.6 MiB" time=2025-11-22T21:16:53.672+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=34 size="681.7 MiB" time=2025-11-22T21:16:53.673+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=35 size="645.6 MiB" time=2025-11-22T21:16:53.675+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=36 size="645.6 MiB" time=2025-11-22T21:16:53.676+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=37 size="681.7 MiB" time=2025-11-22T21:16:53.678+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=38 size="645.6 MiB" time=2025-11-22T21:16:53.681+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=39 size="645.6 MiB" time=2025-11-22T21:16:53.682+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=40 size="681.7 MiB" time=2025-11-22T21:16:53.684+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=41 size="645.6 MiB" time=2025-11-22T21:16:53.685+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=42 size="645.6 MiB" time=2025-11-22T21:16:53.687+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=43 size="681.7 MiB" time=2025-11-22T21:16:53.689+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=44 size="645.6 MiB" time=2025-11-22T21:16:53.694+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=45 size="645.6 MiB" time=2025-11-22T21:16:53.695+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=46 size="681.7 MiB" time=2025-11-22T21:16:53.697+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=47 size="645.6 MiB" time=2025-11-22T21:16:53.699+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=48 size="645.6 MiB" time=2025-11-22T21:16:53.700+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=49 size="681.7 MiB" time=2025-11-22T21:16:53.704+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=50 size="645.6 MiB" time=2025-11-22T21:16:53.705+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=51 size="645.6 MiB" time=2025-11-22T21:16:53.707+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=52 size="681.7 MiB" time=2025-11-22T21:16:53.709+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=53 size="645.6 MiB" time=2025-11-22T21:16:53.711+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=54 size="645.6 MiB" time=2025-11-22T21:16:53.715+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=55 size="681.7 MiB" time=2025-11-22T21:16:53.716+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=56 size="681.7 MiB" time=2025-11-22T21:16:53.717+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=57 size="681.7 MiB" time=2025-11-22T21:16:53.720+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=58 size="681.7 MiB" time=2025-11-22T21:16:53.720+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=59 size="681.7 MiB" time=2025-11-22T21:16:53.723+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=60 size="681.7 MiB" time=2025-11-22T21:16:53.726+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=61 size="681.7 MiB" time=2025-11-22T21:16:53.728+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=62 size="681.7 MiB" time=2025-11-22T21:16:53.729+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=63 size="681.7 MiB" time=2025-11-22T21:16:53.731+01:00 level=TRACE source=server.go:896 msg="layer to assign" layer=64 size="609.1 MiB" time=2025-11-22T21:16:53.733+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=CUDA "available layer vram"="65.2 MiB" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="9.6 GiB" time=2025-11-22T21:16:53.737+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=CUDA "available layer vram"="0 B" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="9.6 GiB" time=2025-11-22T21:16:53.742+01:00 level=DEBUG source=server.go:921 msg="available gpu" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=CUDA "available layer vram"="0 B" backoff=0.00 minimum="457.0 MiB" overhead="0 B" graph="9.6 GiB" time=2025-11-22T21:16:53.746+01:00 level=DEBUG source=server.go:1002 msg="insufficient VRAM to load any model layers" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes, ID: GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes, ID: GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c load_backend: loaded CUDA backend from C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13\ggml-cuda.dll time=2025-11-22T21:16:53.752+01:00 level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 CUDA.1.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.1.USE_GRAPHS=1 CUDA.1.PEER_MAX_BATCH_SIZE=128 CUDA.2.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.2.USE_GRAPHS=1 CUDA.2.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) time=2025-11-22T21:16:53.755+01:00 level=INFO source=runner.go:999 msg="Server listening on 127.0.0.1:58133" time=2025-11-22T21:16:53.747+01:00 level=DEBUG source=server.go:614 msg=memory estimate.CPU.Weights="[312184832 312184832 312184832 312184832 312184832 312184832 312184832 312184832 274337792 275689472 274337792 275689472 310833152 310833152 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 274337792 274337792 312184832 312184832 312184832 312184832 312184832 312184832 312184832 312184832 312184832 638689280]" estimate.CPU.Cache="[402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 402653184 0]" estimate.CUDA0.ID=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 estimate.CUDA0.Graph=10288650240 estimate.CUDA1.ID=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced estimate.CUDA1.Graph=10288650240 estimate.CUDA2.ID=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c estimate.CUDA2.Graph=10288650240 time=2025-11-22T21:16:53.787+01:00 level=INFO source=device.go:245 msg="model weights" device=CPU size="18.1 GiB" time=2025-11-22T21:16:53.795+01:00 level=INFO source=device.go:256 msg="kv cache" device=CPU size="24.0 GiB" time=2025-11-22T21:16:53.796+01:00 level=INFO source=device.go:262 msg="compute graph" device=CUDA0 size="9.6 GiB" time=2025-11-22T21:16:53.798+01:00 level=INFO source=device.go:262 msg="compute graph" device=CUDA1 size="9.6 GiB" time=2025-11-22T21:16:53.799+01:00 level=INFO source=device.go:262 msg="compute graph" device=CUDA2 size="9.6 GiB" time=2025-11-22T21:16:53.801+01:00 level=INFO source=device.go:272 msg="total memory" size="70.8 GiB" time=2025-11-22T21:16:53.807+01:00 level=INFO source=runner.go:893 msg=load request="{Operation:commit LoraPath:[] Parallel:3 BatchSize:512 FlashAttention:false KvSize:98304 KvCacheType: NumThreads:28 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-11-22T21:16:53.808+01:00 level=INFO source=server.go:1294 msg="waiting for llama runner to start responding" time=2025-11-22T21:16:53.813+01:00 level=INFO source=server.go:1328 msg="waiting for server to become available" status="llm server loading model" ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8894390272 total: 12820938752 llama_model_load_from_file_impl: using device CUDA0 (NVIDIA GeForce RTX 5070) (0000:03:00.0) - 8482 MiB free ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11387678720 total: 12884901888 llama_model_load_from_file_impl: using device CUDA1 (NVIDIA GeForce RTX 3060) (0000:04:00.0) - 10860 MiB free ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11551105024 total: 12884901888 llama_model_load_from_file_impl: using device CUDA2 (NVIDIA GeForce RTX 3060) (0000:83:00.0) - 11015 MiB free llama_model_loader: loaded meta data with 34 key-value pairs and 771 tensors from H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen2 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen2.5 Coder 32B Instruct llama_model_loader: - kv 3: general.finetune str = Instruct llama_model_loader: - kv 4: general.basename str = Qwen2.5-Coder llama_model_loader: - kv 5: general.size_label str = 32B llama_model_loader: - kv 6: general.license str = apache-2.0 llama_model_loader: - kv 7: general.license.link str = https://huggingface.co/Qwen/Qwen2.5-C... llama_model_loader: - kv 8: general.base_model.count u32 = 1 llama_model_loader: - kv 9: general.base_model.0.name str = Qwen2.5 Coder 32B llama_model_loader: - kv 10: general.base_model.0.organization str = Qwen llama_model_loader: - kv 11: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen2.5-C... llama_model_loader: - kv 12: general.tags arr[str,6] = ["code", "codeqwen", "chat", "qwen", ... llama_model_loader: - kv 13: general.languages arr[str,1] = ["en"] llama_model_loader: - kv 14: qwen2.block_count u32 = 64 llama_model_loader: - kv 15: qwen2.context_length u32 = 32768 llama_model_loader: - kv 16: qwen2.embedding_length u32 = 5120 llama_model_loader: - kv 17: qwen2.feed_forward_length u32 = 27648 llama_model_loader: - kv 18: qwen2.attention.head_count u32 = 40 llama_model_loader: - kv 19: qwen2.attention.head_count_kv u32 = 8 llama_model_loader: - kv 20: qwen2.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 21: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 22: general.file_type u32 = 15 llama_model_loader: - kv 23: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 24: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 25: tokenizer.ggml.tokens arr[str,152064] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 26: tokenizer.ggml.token_type arr[i32,152064] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 27: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 28: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 29: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 30: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 31: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 32: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 33: general.quantization_version u32 = 2 llama_model_loader: - type f32: 321 tensors llama_model_loader: - type q4_K: 385 tensors llama_model_loader: - type q6_K: 65 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.48 GiB (4.85 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 22 load: token to piece cache size = 0.9310 MB print_info: arch = qwen2 print_info: vocab_only = 0 print_info: n_ctx_train = 32768 print_info: n_embd = 5120 print_info: n_layer = 64 print_info: n_head = 40 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: is_swa_any = 0 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 5 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-06 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 27648 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = -1 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 32768 print_info: rope_finetuned = unknown print_info: model type = 32B print_info: model params = 32.76 B print_info: general.name = Qwen2.5 Coder 32B Instruct print_info: vocab type = BPE print_info: n_vocab = 152064 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = false) ggml_backend_cuda_device_get_memory device GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 utilizing NVML memory reporting free: 8888201216 total: 12820938752 ggml_backend_cuda_device_get_memory device GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced utilizing NVML memory reporting free: 11387678720 total: 12884901888 ggml_backend_cuda_device_get_memory device GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c utilizing NVML memory reporting free: 11551105024 total: 12884901888 load_tensors: layer 0 assigned to device CPU, is_swa = 0 load_tensors: layer 1 assigned to device CPU, is_swa = 0 load_tensors: layer 2 assigned to device CPU, is_swa = 0 load_tensors: layer 3 assigned to device CPU, is_swa = 0 load_tensors: layer 4 assigned to device CPU, is_swa = 0 load_tensors: layer 5 assigned to device CPU, is_swa = 0 load_tensors: layer 6 assigned to device CPU, is_swa = 0 load_tensors: layer 7 assigned to device CPU, is_swa = 0 load_tensors: layer 8 assigned to device CPU, is_swa = 0 load_tensors: layer 9 assigned to device CPU, is_swa = 0 load_tensors: layer 10 assigned to device CPU, is_swa = 0 load_tensors: layer 11 assigned to device CPU, is_swa = 0 load_tensors: layer 12 assigned to device CPU, is_swa = 0 load_tensors: layer 13 assigned to device CPU, is_swa = 0 load_tensors: layer 14 assigned to device CPU, is_swa = 0 load_tensors: layer 15 assigned to device CPU, is_swa = 0 load_tensors: layer 16 assigned to device CPU, is_swa = 0 load_tensors: layer 17 assigned to device CPU, is_swa = 0 load_tensors: layer 18 assigned to device CPU, is_swa = 0 load_tensors: layer 19 assigned to device CPU, is_swa = 0 load_tensors: layer 20 assigned to device CPU, is_swa = 0 load_tensors: layer 21 assigned to device CPU, is_swa = 0 load_tensors: layer 22 assigned to device CPU, is_swa = 0 load_tensors: layer 23 assigned to device CPU, is_swa = 0 load_tensors: layer 24 assigned to device CPU, is_swa = 0 load_tensors: layer 25 assigned to device CPU, is_swa = 0 load_tensors: layer 26 assigned to device CPU, is_swa = 0 load_tensors: layer 27 assigned to device CPU, is_swa = 0 load_tensors: layer 28 assigned to device CPU, is_swa = 0 load_tensors: layer 29 assigned to device CPU, is_swa = 0 load_tensors: layer 30 assigned to device CPU, is_swa = 0 load_tensors: layer 31 assigned to device CPU, is_swa = 0 load_tensors: layer 32 assigned to device CPU, is_swa = 0 load_tensors: layer 33 assigned to device CPU, is_swa = 0 load_tensors: layer 34 assigned to device CPU, is_swa = 0 load_tensors: layer 35 assigned to device CPU, is_swa = 0 load_tensors: layer 36 assigned to device CPU, is_swa = 0 load_tensors: layer 37 assigned to device CPU, is_swa = 0 load_tensors: layer 38 assigned to device CPU, is_swa = 0 load_tensors: layer 39 assigned to device CPU, is_swa = 0 load_tensors: layer 40 assigned to device CPU, is_swa = 0 load_tensors: layer 41 assigned to device CPU, is_swa = 0 load_tensors: layer 42 assigned to device CPU, is_swa = 0 load_tensors: layer 43 assigned to device CPU, is_swa = 0 load_tensors: layer 44 assigned to device CPU, is_swa = 0 load_tensors: layer 45 assigned to device CPU, is_swa = 0 load_tensors: layer 46 assigned to device CPU, is_swa = 0 load_tensors: layer 47 assigned to device CPU, is_swa = 0 load_tensors: layer 48 assigned to device CPU, is_swa = 0 load_tensors: layer 49 assigned to device CPU, is_swa = 0 load_tensors: layer 50 assigned to device CPU, is_swa = 0 load_tensors: layer 51 assigned to device CPU, is_swa = 0 load_tensors: layer 52 assigned to device CPU, is_swa = 0 load_tensors: layer 53 assigned to device CPU, is_swa = 0 load_tensors: layer 54 assigned to device CPU, is_swa = 0 load_tensors: layer 55 assigned to device CPU, is_swa = 0 load_tensors: layer 56 assigned to device CPU, is_swa = 0 load_tensors: layer 57 assigned to device CPU, is_swa = 0 load_tensors: layer 58 assigned to device CPU, is_swa = 0 load_tensors: layer 59 assigned to device CPU, is_swa = 0 load_tensors: layer 60 assigned to device CPU, is_swa = 0 load_tensors: layer 61 assigned to device CPU, is_swa = 0 load_tensors: layer 62 assigned to device CPU, is_swa = 0 load_tensors: layer 63 assigned to device CPU, is_swa = 0 load_tensors: layer 64 assigned to device CPU, is_swa = 0 create_tensor: loading tensor token_embd.weight create_tensor: loading tensor output_norm.weight create_tensor: loading tensor output.weight create_tensor: loading tensor blk.0.attn_norm.weight create_tensor: loading tensor blk.0.attn_q.weight create_tensor: loading tensor blk.0.attn_k.weight create_tensor: loading tensor blk.0.attn_v.weight create_tensor: loading tensor blk.0.attn_output.weight create_tensor: loading tensor blk.0.attn_q.bias create_tensor: loading tensor blk.0.attn_k.bias create_tensor: loading tensor blk.0.attn_v.bias create_tensor: loading tensor blk.0.ffn_norm.weight create_tensor: loading tensor blk.0.ffn_gate.weight create_tensor: loading tensor blk.0.ffn_down.weight create_tensor: loading tensor blk.0.ffn_up.weight create_tensor: loading tensor blk.1.attn_norm.weight create_tensor: loading tensor blk.1.attn_q.weight create_tensor: loading tensor blk.1.attn_k.weight create_tensor: loading tensor blk.1.attn_v.weight create_tensor: loading tensor blk.1.attn_output.weight create_tensor: loading tensor blk.1.attn_q.bias create_tensor: loading tensor blk.1.attn_k.bias create_tensor: loading tensor blk.1.attn_v.bias create_tensor: loading tensor blk.1.ffn_norm.weight create_tensor: loading tensor blk.1.ffn_gate.weight create_tensor: loading tensor blk.1.ffn_down.weight create_tensor: loading tensor blk.1.ffn_up.weight create_tensor: loading tensor blk.2.attn_norm.weight create_tensor: loading tensor blk.2.attn_q.weight create_tensor: loading tensor blk.2.attn_k.weight create_tensor: loading tensor blk.2.attn_v.weight create_tensor: loading tensor blk.2.attn_output.weight create_tensor: loading tensor blk.2.attn_q.bias create_tensor: loading tensor blk.2.attn_k.bias create_tensor: loading tensor blk.2.attn_v.bias create_tensor: loading tensor blk.2.ffn_norm.weight create_tensor: loading tensor blk.2.ffn_gate.weight create_tensor: loading tensor blk.2.ffn_down.weight create_tensor: loading tensor blk.2.ffn_up.weight create_tensor: loading tensor blk.3.attn_norm.weight create_tensor: loading tensor blk.3.attn_q.weight create_tensor: loading tensor blk.3.attn_k.weight create_tensor: loading tensor blk.3.attn_v.weight create_tensor: loading tensor blk.3.attn_output.weight create_tensor: loading tensor blk.3.attn_q.bias create_tensor: loading tensor blk.3.attn_k.bias create_tensor: loading tensor blk.3.attn_v.bias create_tensor: loading tensor blk.3.ffn_norm.weight create_tensor: loading tensor blk.3.ffn_gate.weight create_tensor: loading tensor blk.3.ffn_down.weight create_tensor: loading tensor blk.3.ffn_up.weight create_tensor: loading tensor blk.4.attn_norm.weight create_tensor: loading tensor blk.4.attn_q.weight create_tensor: loading tensor blk.4.attn_k.weight create_tensor: loading tensor blk.4.attn_v.weight create_tensor: loading tensor blk.4.attn_output.weight create_tensor: loading tensor blk.4.attn_q.bias create_tensor: loading tensor blk.4.attn_k.bias create_tensor: loading tensor blk.4.attn_v.bias create_tensor: loading tensor blk.4.ffn_norm.weight create_tensor: loading tensor blk.4.ffn_gate.weight create_tensor: loading tensor blk.4.ffn_down.weight create_tensor: loading tensor blk.4.ffn_up.weight create_tensor: loading tensor blk.5.attn_norm.weight create_tensor: loading tensor blk.5.attn_q.weight create_tensor: loading tensor blk.5.attn_k.weight create_tensor: loading tensor blk.5.attn_v.weight create_tensor: loading tensor blk.5.attn_output.weight create_tensor: loading tensor blk.5.attn_q.bias create_tensor: loading tensor blk.5.attn_k.bias create_tensor: loading tensor blk.5.attn_v.bias create_tensor: loading tensor blk.5.ffn_norm.weight create_tensor: loading tensor blk.5.ffn_gate.weight create_tensor: loading tensor blk.5.ffn_down.weight create_tensor: loading tensor blk.5.ffn_up.weight create_tensor: loading tensor blk.6.attn_norm.weight create_tensor: loading tensor blk.6.attn_q.weight create_tensor: loading tensor blk.6.attn_k.weight create_tensor: loading tensor blk.6.attn_v.weight create_tensor: loading tensor blk.6.attn_output.weight create_tensor: loading tensor blk.6.attn_q.bias create_tensor: loading tensor blk.6.attn_k.bias create_tensor: loading tensor blk.6.attn_v.bias create_tensor: loading tensor blk.6.ffn_norm.weight create_tensor: loading tensor blk.6.ffn_gate.weight create_tensor: loading tensor blk.6.ffn_down.weight create_tensor: loading tensor blk.6.ffn_up.weight create_tensor: loading tensor blk.7.attn_norm.weight create_tensor: loading tensor blk.7.attn_q.weight create_tensor: loading tensor blk.7.attn_k.weight create_tensor: loading tensor blk.7.attn_v.weight create_tensor: loading tensor blk.7.attn_output.weight create_tensor: loading tensor blk.7.attn_q.bias create_tensor: loading tensor blk.7.attn_k.bias create_tensor: loading tensor blk.7.attn_v.bias create_tensor: loading tensor blk.7.ffn_norm.weight create_tensor: loading tensor blk.7.ffn_gate.weight create_tensor: loading tensor blk.7.ffn_down.weight create_tensor: loading tensor blk.7.ffn_up.weight create_tensor: loading tensor blk.8.attn_norm.weight create_tensor: loading tensor blk.8.attn_q.weight create_tensor: loading tensor blk.8.attn_k.weight create_tensor: loading tensor blk.8.attn_v.weight create_tensor: loading tensor blk.8.attn_output.weight create_tensor: loading tensor blk.8.attn_q.bias create_tensor: loading tensor blk.8.attn_k.bias create_tensor: loading tensor blk.8.attn_v.bias create_tensor: loading tensor blk.8.ffn_norm.weight create_tensor: loading tensor blk.8.ffn_gate.weight create_tensor: loading tensor blk.8.ffn_down.weight create_tensor: loading tensor blk.8.ffn_up.weight create_tensor: loading tensor blk.9.attn_norm.weight create_tensor: loading tensor blk.9.attn_q.weight create_tensor: loading tensor blk.9.attn_k.weight create_tensor: loading tensor blk.9.attn_v.weight create_tensor: loading tensor blk.9.attn_output.weight create_tensor: loading tensor blk.9.attn_q.bias create_tensor: loading tensor blk.9.attn_k.bias create_tensor: loading tensor blk.9.attn_v.bias create_tensor: loading tensor blk.9.ffn_norm.weight create_tensor: loading tensor blk.9.ffn_gate.weight create_tensor: loading tensor blk.9.ffn_down.weight create_tensor: loading tensor blk.9.ffn_up.weight create_tensor: loading tensor blk.10.attn_norm.weight create_tensor: loading tensor blk.10.attn_q.weight create_tensor: loading tensor blk.10.attn_k.weight create_tensor: loading tensor blk.10.attn_v.weight create_tensor: loading tensor blk.10.attn_output.weight create_tensor: loading tensor blk.10.attn_q.bias create_tensor: loading tensor blk.10.attn_k.bias create_tensor: loading tensor blk.10.attn_v.bias create_tensor: loading tensor blk.10.ffn_norm.weight create_tensor: loading tensor blk.10.ffn_gate.weight create_tensor: loading tensor blk.10.ffn_down.weight create_tensor: loading tensor blk.10.ffn_up.weight create_tensor: loading tensor blk.11.attn_norm.weight create_tensor: loading tensor blk.11.attn_q.weight create_tensor: loading tensor blk.11.attn_k.weight create_tensor: loading tensor blk.11.attn_v.weight create_tensor: loading tensor blk.11.attn_output.weight create_tensor: loading tensor blk.11.attn_q.bias create_tensor: loading tensor blk.11.attn_k.bias create_tensor: loading tensor blk.11.attn_v.bias create_tensor: loading tensor blk.11.ffn_norm.weight create_tensor: loading tensor blk.11.ffn_gate.weight create_tensor: loading tensor blk.11.ffn_down.weight create_tensor: loading tensor blk.11.ffn_up.weight create_tensor: loading tensor blk.12.attn_norm.weight create_tensor: loading tensor blk.12.attn_q.weight create_tensor: loading tensor blk.12.attn_k.weight create_tensor: loading tensor blk.12.attn_v.weight create_tensor: loading tensor blk.12.attn_output.weight create_tensor: loading tensor blk.12.attn_q.bias create_tensor: loading tensor blk.12.attn_k.bias create_tensor: loading tensor blk.12.attn_v.bias create_tensor: loading tensor blk.12.ffn_norm.weight create_tensor: loading tensor blk.12.ffn_gate.weight create_tensor: loading tensor blk.12.ffn_down.weight create_tensor: loading tensor blk.12.ffn_up.weight create_tensor: loading tensor blk.13.attn_norm.weight create_tensor: loading tensor blk.13.attn_q.weight create_tensor: loading tensor blk.13.attn_k.weight create_tensor: loading tensor blk.13.attn_v.weight create_tensor: loading tensor blk.13.attn_output.weight create_tensor: loading tensor blk.13.attn_q.bias create_tensor: loading tensor blk.13.attn_k.bias create_tensor: loading tensor blk.13.attn_v.bias create_tensor: loading tensor blk.13.ffn_norm.weight create_tensor: loading tensor blk.13.ffn_gate.weight create_tensor: loading tensor blk.13.ffn_down.weight create_tensor: loading tensor blk.13.ffn_up.weight create_tensor: loading tensor blk.14.attn_norm.weight create_tensor: loading tensor blk.14.attn_q.weight create_tensor: loading tensor blk.14.attn_k.weight create_tensor: loading tensor blk.14.attn_v.weight create_tensor: loading tensor blk.14.attn_output.weight create_tensor: loading tensor blk.14.attn_q.bias create_tensor: loading tensor blk.14.attn_k.bias create_tensor: loading tensor blk.14.attn_v.bias create_tensor: loading tensor blk.14.ffn_norm.weight create_tensor: loading tensor blk.14.ffn_gate.weight create_tensor: loading tensor blk.14.ffn_down.weight create_tensor: loading tensor blk.14.ffn_up.weight create_tensor: loading tensor blk.15.attn_norm.weight create_tensor: loading tensor blk.15.attn_q.weight create_tensor: loading tensor blk.15.attn_k.weight create_tensor: loading tensor blk.15.attn_v.weight create_tensor: loading tensor blk.15.attn_output.weight create_tensor: loading tensor blk.15.attn_q.bias create_tensor: loading tensor blk.15.attn_k.bias create_tensor: loading tensor blk.15.attn_v.bias create_tensor: loading tensor blk.15.ffn_norm.weight create_tensor: loading tensor blk.15.ffn_gate.weight create_tensor: loading tensor blk.15.ffn_down.weight create_tensor: loading tensor blk.15.ffn_up.weight create_tensor: loading tensor blk.16.attn_norm.weight create_tensor: loading tensor blk.16.attn_q.weight create_tensor: loading tensor blk.16.attn_k.weight create_tensor: loading tensor blk.16.attn_v.weight create_tensor: loading tensor blk.16.attn_output.weight create_tensor: loading tensor blk.16.attn_q.bias create_tensor: loading tensor blk.16.attn_k.bias create_tensor: loading tensor blk.16.attn_v.bias create_tensor: loading tensor blk.16.ffn_norm.weight create_tensor: loading tensor blk.16.ffn_gate.weight create_tensor: loading tensor blk.16.ffn_down.weight create_tensor: loading tensor blk.16.ffn_up.weight create_tensor: loading tensor blk.17.attn_norm.weight create_tensor: loading tensor blk.17.attn_q.weight create_tensor: loading tensor blk.17.attn_k.weight create_tensor: loading tensor blk.17.attn_v.weight create_tensor: loading tensor blk.17.attn_output.weight create_tensor: loading tensor blk.17.attn_q.bias create_tensor: loading tensor blk.17.attn_k.bias create_tensor: loading tensor blk.17.attn_v.bias create_tensor: loading tensor blk.17.ffn_norm.weight create_tensor: loading tensor blk.17.ffn_gate.weight create_tensor: loading tensor blk.17.ffn_down.weight create_tensor: loading tensor blk.17.ffn_up.weight create_tensor: loading tensor blk.18.attn_norm.weight create_tensor: loading tensor blk.18.attn_q.weight create_tensor: loading tensor blk.18.attn_k.weight create_tensor: loading tensor blk.18.attn_v.weight create_tensor: loading tensor blk.18.attn_output.weight create_tensor: loading tensor blk.18.attn_q.bias create_tensor: loading tensor blk.18.attn_k.bias create_tensor: loading tensor blk.18.attn_v.bias create_tensor: loading tensor blk.18.ffn_norm.weight create_tensor: loading tensor blk.18.ffn_gate.weight create_tensor: loading tensor blk.18.ffn_down.weight create_tensor: loading tensor blk.18.ffn_up.weight create_tensor: loading tensor blk.19.attn_norm.weight create_tensor: loading tensor blk.19.attn_q.weight create_tensor: loading tensor blk.19.attn_k.weight create_tensor: loading tensor blk.19.attn_v.weight create_tensor: loading tensor blk.19.attn_output.weight create_tensor: loading tensor blk.19.attn_q.bias create_tensor: loading tensor blk.19.attn_k.bias create_tensor: loading tensor blk.19.attn_v.bias create_tensor: loading tensor blk.19.ffn_norm.weight create_tensor: loading tensor blk.19.ffn_gate.weight create_tensor: loading tensor blk.19.ffn_down.weight create_tensor: loading tensor blk.19.ffn_up.weight create_tensor: loading tensor blk.20.attn_norm.weight create_tensor: loading tensor blk.20.attn_q.weight create_tensor: loading tensor blk.20.attn_k.weight create_tensor: loading tensor blk.20.attn_v.weight create_tensor: loading tensor blk.20.attn_output.weight create_tensor: loading tensor blk.20.attn_q.bias create_tensor: loading tensor blk.20.attn_k.bias create_tensor: loading tensor blk.20.attn_v.bias create_tensor: loading tensor blk.20.ffn_norm.weight create_tensor: loading tensor blk.20.ffn_gate.weight create_tensor: loading tensor blk.20.ffn_down.weight create_tensor: loading tensor blk.20.ffn_up.weight create_tensor: loading tensor blk.21.attn_norm.weight create_tensor: loading tensor blk.21.attn_q.weight create_tensor: loading tensor blk.21.attn_k.weight create_tensor: loading tensor blk.21.attn_v.weight create_tensor: loading tensor blk.21.attn_output.weight create_tensor: loading tensor blk.21.attn_q.bias create_tensor: loading tensor blk.21.attn_k.bias create_tensor: loading tensor blk.21.attn_v.bias create_tensor: loading tensor blk.21.ffn_norm.weight create_tensor: loading tensor blk.21.ffn_gate.weight create_tensor: loading tensor blk.21.ffn_down.weight create_tensor: loading tensor blk.21.ffn_up.weight create_tensor: loading tensor blk.22.attn_norm.weight create_tensor: loading tensor blk.22.attn_q.weight create_tensor: loading tensor blk.22.attn_k.weight create_tensor: loading tensor blk.22.attn_v.weight create_tensor: loading tensor blk.22.attn_output.weight create_tensor: loading tensor blk.22.attn_q.bias create_tensor: loading tensor blk.22.attn_k.bias create_tensor: loading tensor blk.22.attn_v.bias create_tensor: loading tensor blk.22.ffn_norm.weight create_tensor: loading tensor blk.22.ffn_gate.weight create_tensor: loading tensor blk.22.ffn_down.weight create_tensor: loading tensor blk.22.ffn_up.weight create_tensor: loading tensor blk.23.attn_norm.weight create_tensor: loading tensor blk.23.attn_q.weight create_tensor: loading tensor blk.23.attn_k.weight create_tensor: loading tensor blk.23.attn_v.weight create_tensor: loading tensor blk.23.attn_output.weight create_tensor: loading tensor blk.23.attn_q.bias create_tensor: loading tensor blk.23.attn_k.bias create_tensor: loading tensor blk.23.attn_v.bias create_tensor: loading tensor blk.23.ffn_norm.weight create_tensor: loading tensor blk.23.ffn_gate.weight create_tensor: loading tensor blk.23.ffn_down.weight create_tensor: loading tensor blk.23.ffn_up.weight create_tensor: loading tensor blk.24.attn_norm.weight create_tensor: loading tensor blk.24.attn_q.weight create_tensor: loading tensor blk.24.attn_k.weight create_tensor: loading tensor blk.24.attn_v.weight create_tensor: loading tensor blk.24.attn_output.weight create_tensor: loading tensor blk.24.attn_q.bias create_tensor: loading tensor blk.24.attn_k.bias create_tensor: loading tensor blk.24.attn_v.bias create_tensor: loading tensor blk.24.ffn_norm.weight create_tensor: loading tensor blk.24.ffn_gate.weight create_tensor: loading tensor blk.24.ffn_down.weight create_tensor: loading tensor blk.24.ffn_up.weight create_tensor: loading tensor blk.25.attn_norm.weight create_tensor: loading tensor blk.25.attn_q.weight create_tensor: loading tensor blk.25.attn_k.weight create_tensor: loading tensor blk.25.attn_v.weight create_tensor: loading tensor blk.25.attn_output.weight create_tensor: loading tensor blk.25.attn_q.bias create_tensor: loading tensor blk.25.attn_k.bias create_tensor: loading tensor blk.25.attn_v.bias create_tensor: loading tensor blk.25.ffn_norm.weight create_tensor: loading tensor blk.25.ffn_gate.weight create_tensor: loading tensor blk.25.ffn_down.weight create_tensor: loading tensor blk.25.ffn_up.weight create_tensor: loading tensor blk.26.attn_norm.weight create_tensor: loading tensor blk.26.attn_q.weight create_tensor: loading tensor blk.26.attn_k.weight create_tensor: loading tensor blk.26.attn_v.weight create_tensor: loading tensor blk.26.attn_output.weight create_tensor: loading tensor blk.26.attn_q.bias create_tensor: loading tensor blk.26.attn_k.bias create_tensor: loading tensor blk.26.attn_v.bias create_tensor: loading tensor blk.26.ffn_norm.weight create_tensor: loading tensor blk.26.ffn_gate.weight create_tensor: loading tensor blk.26.ffn_down.weight create_tensor: loading tensor blk.26.ffn_up.weight create_tensor: loading tensor blk.27.attn_norm.weight create_tensor: loading tensor blk.27.attn_q.weight create_tensor: loading tensor blk.27.attn_k.weight create_tensor: loading tensor blk.27.attn_v.weight create_tensor: loading tensor blk.27.attn_output.weight create_tensor: loading tensor blk.27.attn_q.bias create_tensor: loading tensor blk.27.attn_k.bias create_tensor: loading tensor blk.27.attn_v.bias create_tensor: loading tensor blk.27.ffn_norm.weight create_tensor: loading tensor blk.27.ffn_gate.weight create_tensor: loading tensor blk.27.ffn_down.weight create_tensor: loading tensor blk.27.ffn_up.weight create_tensor: loading tensor blk.28.attn_norm.weight create_tensor: loading tensor blk.28.attn_q.weight create_tensor: loading tensor blk.28.attn_k.weight create_tensor: loading tensor blk.28.attn_v.weight create_tensor: loading tensor blk.28.attn_output.weight create_tensor: loading tensor blk.28.attn_q.bias create_tensor: loading tensor blk.28.attn_k.bias create_tensor: loading tensor blk.28.attn_v.bias create_tensor: loading tensor blk.28.ffn_norm.weight create_tensor: loading tensor blk.28.ffn_gate.weight create_tensor: loading tensor blk.28.ffn_down.weight create_tensor: loading tensor blk.28.ffn_up.weight create_tensor: loading tensor blk.29.attn_norm.weight create_tensor: loading tensor blk.29.attn_q.weight create_tensor: loading tensor blk.29.attn_k.weight create_tensor: loading tensor blk.29.attn_v.weight create_tensor: loading tensor blk.29.attn_output.weight create_tensor: loading tensor blk.29.attn_q.bias create_tensor: loading tensor blk.29.attn_k.bias create_tensor: loading tensor blk.29.attn_v.bias create_tensor: loading tensor blk.29.ffn_norm.weight create_tensor: loading tensor blk.29.ffn_gate.weight create_tensor: loading tensor blk.29.ffn_down.weight create_tensor: loading tensor blk.29.ffn_up.weight create_tensor: loading tensor blk.30.attn_norm.weight create_tensor: loading tensor blk.30.attn_q.weight create_tensor: loading tensor blk.30.attn_k.weight create_tensor: loading tensor blk.30.attn_v.weight create_tensor: loading tensor blk.30.attn_output.weight create_tensor: loading tensor blk.30.attn_q.bias create_tensor: loading tensor blk.30.attn_k.bias create_tensor: loading tensor blk.30.attn_v.bias create_tensor: loading tensor blk.30.ffn_norm.weight create_tensor: loading tensor blk.30.ffn_gate.weight create_tensor: loading tensor blk.30.ffn_down.weight create_tensor: loading tensor blk.30.ffn_up.weight create_tensor: loading tensor blk.31.attn_norm.weight create_tensor: loading tensor blk.31.attn_q.weight create_tensor: loading tensor blk.31.attn_k.weight create_tensor: loading tensor blk.31.attn_v.weight create_tensor: loading tensor blk.31.attn_output.weight create_tensor: loading tensor blk.31.attn_q.bias create_tensor: loading tensor blk.31.attn_k.bias create_tensor: loading tensor blk.31.attn_v.bias create_tensor: loading tensor blk.31.ffn_norm.weight create_tensor: loading tensor blk.31.ffn_gate.weight create_tensor: loading tensor blk.31.ffn_down.weight create_tensor: loading tensor blk.31.ffn_up.weight create_tensor: loading tensor blk.32.attn_norm.weight create_tensor: loading tensor blk.32.attn_q.weight create_tensor: loading tensor blk.32.attn_k.weight create_tensor: loading tensor blk.32.attn_v.weight create_tensor: loading tensor blk.32.attn_output.weight create_tensor: loading tensor blk.32.attn_q.bias create_tensor: loading tensor blk.32.attn_k.bias create_tensor: loading tensor blk.32.attn_v.bias create_tensor: loading tensor blk.32.ffn_norm.weight create_tensor: loading tensor blk.32.ffn_gate.weight create_tensor: loading tensor blk.32.ffn_down.weight create_tensor: loading tensor blk.32.ffn_up.weight create_tensor: loading tensor blk.33.attn_norm.weight create_tensor: loading tensor blk.33.attn_q.weight create_tensor: loading tensor blk.33.attn_k.weight create_tensor: loading tensor blk.33.attn_v.weight create_tensor: loading tensor blk.33.attn_output.weight create_tensor: loading tensor blk.33.attn_q.bias create_tensor: loading tensor blk.33.attn_k.bias create_tensor: loading tensor blk.33.attn_v.bias create_tensor: loading tensor blk.33.ffn_norm.weight create_tensor: loading tensor blk.33.ffn_gate.weight create_tensor: loading tensor blk.33.ffn_down.weight create_tensor: loading tensor blk.33.ffn_up.weight create_tensor: loading tensor blk.34.attn_norm.weight create_tensor: loading tensor blk.34.attn_q.weight create_tensor: loading tensor blk.34.attn_k.weight create_tensor: loading tensor blk.34.attn_v.weight create_tensor: loading tensor blk.34.attn_output.weight create_tensor: loading tensor blk.34.attn_q.bias create_tensor: loading tensor blk.34.attn_k.bias create_tensor: loading tensor blk.34.attn_v.bias create_tensor: loading tensor blk.34.ffn_norm.weight create_tensor: loading tensor blk.34.ffn_gate.weight create_tensor: loading tensor blk.34.ffn_down.weight create_tensor: loading tensor blk.34.ffn_up.weight create_tensor: loading tensor blk.35.attn_norm.weight create_tensor: loading tensor blk.35.attn_q.weight create_tensor: loading tensor blk.35.attn_k.weight create_tensor: loading tensor blk.35.attn_v.weight create_tensor: loading tensor blk.35.attn_output.weight create_tensor: loading tensor blk.35.attn_q.bias create_tensor: loading tensor blk.35.attn_k.bias create_tensor: loading tensor blk.35.attn_v.bias create_tensor: loading tensor blk.35.ffn_norm.weight create_tensor: loading tensor blk.35.ffn_gate.weight create_tensor: loading tensor blk.35.ffn_down.weight create_tensor: loading tensor blk.35.ffn_up.weight create_tensor: loading tensor blk.36.attn_norm.weight create_tensor: loading tensor blk.36.attn_q.weight create_tensor: loading tensor blk.36.attn_k.weight create_tensor: loading tensor blk.36.attn_v.weight create_tensor: loading tensor blk.36.attn_output.weight create_tensor: loading tensor blk.36.attn_q.bias create_tensor: loading tensor blk.36.attn_k.bias create_tensor: loading tensor blk.36.attn_v.bias create_tensor: loading tensor blk.36.ffn_norm.weight create_tensor: loading tensor blk.36.ffn_gate.weight create_tensor: loading tensor blk.36.ffn_down.weight create_tensor: loading tensor blk.36.ffn_up.weight create_tensor: loading tensor blk.37.attn_norm.weight create_tensor: loading tensor blk.37.attn_q.weight create_tensor: loading tensor blk.37.attn_k.weight create_tensor: loading tensor blk.37.attn_v.weight create_tensor: loading tensor blk.37.attn_output.weight create_tensor: loading tensor blk.37.attn_q.bias create_tensor: loading tensor blk.37.attn_k.bias create_tensor: loading tensor blk.37.attn_v.bias create_tensor: loading tensor blk.37.ffn_norm.weight create_tensor: loading tensor blk.37.ffn_gate.weight create_tensor: loading tensor blk.37.ffn_down.weight create_tensor: loading tensor blk.37.ffn_up.weight create_tensor: loading tensor blk.38.attn_norm.weight create_tensor: loading tensor blk.38.attn_q.weight create_tensor: loading tensor blk.38.attn_k.weight create_tensor: loading tensor blk.38.attn_v.weight create_tensor: loading tensor blk.38.attn_output.weight create_tensor: loading tensor blk.38.attn_q.bias create_tensor: loading tensor blk.38.attn_k.bias create_tensor: loading tensor blk.38.attn_v.bias create_tensor: loading tensor blk.38.ffn_norm.weight create_tensor: loading tensor blk.38.ffn_gate.weight create_tensor: loading tensor blk.38.ffn_down.weight create_tensor: loading tensor blk.38.ffn_up.weight create_tensor: loading tensor blk.39.attn_norm.weight create_tensor: loading tensor blk.39.attn_q.weight create_tensor: loading tensor blk.39.attn_k.weight create_tensor: loading tensor blk.39.attn_v.weight create_tensor: loading tensor blk.39.attn_output.weight create_tensor: loading tensor blk.39.attn_q.bias create_tensor: loading tensor blk.39.attn_k.bias create_tensor: loading tensor blk.39.attn_v.bias create_tensor: loading tensor blk.39.ffn_norm.weight create_tensor: loading tensor blk.39.ffn_gate.weight create_tensor: loading tensor blk.39.ffn_down.weight create_tensor: loading tensor blk.39.ffn_up.weight create_tensor: loading tensor blk.40.attn_norm.weight create_tensor: loading tensor blk.40.attn_q.weight create_tensor: loading tensor blk.40.attn_k.weight create_tensor: loading tensor blk.40.attn_v.weight create_tensor: loading tensor blk.40.attn_output.weight create_tensor: loading tensor blk.40.attn_q.bias create_tensor: loading tensor blk.40.attn_k.bias create_tensor: loading tensor blk.40.attn_v.bias create_tensor: loading tensor blk.40.ffn_norm.weight create_tensor: loading tensor blk.40.ffn_gate.weight create_tensor: loading tensor blk.40.ffn_down.weight create_tensor: loading tensor blk.40.ffn_up.weight create_tensor: loading tensor blk.41.attn_norm.weight create_tensor: loading tensor blk.41.attn_q.weight create_tensor: loading tensor blk.41.attn_k.weight create_tensor: loading tensor blk.41.attn_v.weight create_tensor: loading tensor blk.41.attn_output.weight create_tensor: loading tensor blk.41.attn_q.bias create_tensor: loading tensor blk.41.attn_k.bias create_tensor: loading tensor blk.41.attn_v.bias create_tensor: loading tensor blk.41.ffn_norm.weight create_tensor: loading tensor blk.41.ffn_gate.weight create_tensor: loading tensor blk.41.ffn_down.weight create_tensor: loading tensor blk.41.ffn_up.weight create_tensor: loading tensor blk.42.attn_norm.weight create_tensor: loading tensor blk.42.attn_q.weight create_tensor: loading tensor blk.42.attn_k.weight create_tensor: loading tensor blk.42.attn_v.weight create_tensor: loading tensor blk.42.attn_output.weight create_tensor: loading tensor blk.42.attn_q.bias create_tensor: loading tensor blk.42.attn_k.bias create_tensor: loading tensor blk.42.attn_v.bias create_tensor: loading tensor blk.42.ffn_norm.weight create_tensor: loading tensor blk.42.ffn_gate.weight create_tensor: loading tensor blk.42.ffn_down.weight create_tensor: loading tensor blk.42.ffn_up.weight create_tensor: loading tensor blk.43.attn_norm.weight create_tensor: loading tensor blk.43.attn_q.weight create_tensor: loading tensor blk.43.attn_k.weight create_tensor: loading tensor blk.43.attn_v.weight create_tensor: loading tensor blk.43.attn_output.weight create_tensor: loading tensor blk.43.attn_q.bias create_tensor: loading tensor blk.43.attn_k.bias create_tensor: loading tensor blk.43.attn_v.bias create_tensor: loading tensor blk.43.ffn_norm.weight create_tensor: loading tensor blk.43.ffn_gate.weight create_tensor: loading tensor blk.43.ffn_down.weight create_tensor: loading tensor blk.43.ffn_up.weight create_tensor: loading tensor blk.44.attn_norm.weight create_tensor: loading tensor blk.44.attn_q.weight create_tensor: loading tensor blk.44.attn_k.weight create_tensor: loading tensor blk.44.attn_v.weight create_tensor: loading tensor blk.44.attn_output.weight create_tensor: loading tensor blk.44.attn_q.bias create_tensor: loading tensor blk.44.attn_k.bias create_tensor: loading tensor blk.44.attn_v.bias create_tensor: loading tensor blk.44.ffn_norm.weight create_tensor: loading tensor blk.44.ffn_gate.weight create_tensor: loading tensor blk.44.ffn_down.weight create_tensor: loading tensor blk.44.ffn_up.weight create_tensor: loading tensor blk.45.attn_norm.weight create_tensor: loading tensor blk.45.attn_q.weight create_tensor: loading tensor blk.45.attn_k.weight create_tensor: loading tensor blk.45.attn_v.weight create_tensor: loading tensor blk.45.attn_output.weight create_tensor: loading tensor blk.45.attn_q.bias create_tensor: loading tensor blk.45.attn_k.bias create_tensor: loading tensor blk.45.attn_v.bias create_tensor: loading tensor blk.45.ffn_norm.weight create_tensor: loading tensor blk.45.ffn_gate.weight create_tensor: loading tensor blk.45.ffn_down.weight create_tensor: loading tensor blk.45.ffn_up.weight create_tensor: loading tensor blk.46.attn_norm.weight create_tensor: loading tensor blk.46.attn_q.weight create_tensor: loading tensor blk.46.attn_k.weight create_tensor: loading tensor blk.46.attn_v.weight create_tensor: loading tensor blk.46.attn_output.weight create_tensor: loading tensor blk.46.attn_q.bias create_tensor: loading tensor blk.46.attn_k.bias create_tensor: loading tensor blk.46.attn_v.bias create_tensor: loading tensor blk.46.ffn_norm.weight create_tensor: loading tensor blk.46.ffn_gate.weight create_tensor: loading tensor blk.46.ffn_down.weight create_tensor: loading tensor blk.46.ffn_up.weight create_tensor: loading tensor blk.47.attn_norm.weight create_tensor: loading tensor blk.47.attn_q.weight create_tensor: loading tensor blk.47.attn_k.weight create_tensor: loading tensor blk.47.attn_v.weight create_tensor: loading tensor blk.47.attn_output.weight create_tensor: loading tensor blk.47.attn_q.bias create_tensor: loading tensor blk.47.attn_k.bias create_tensor: loading tensor blk.47.attn_v.bias create_tensor: loading tensor blk.47.ffn_norm.weight create_tensor: loading tensor blk.47.ffn_gate.weight create_tensor: loading tensor blk.47.ffn_down.weight create_tensor: loading tensor blk.47.ffn_up.weight create_tensor: loading tensor blk.48.attn_norm.weight create_tensor: loading tensor blk.48.attn_q.weight create_tensor: loading tensor blk.48.attn_k.weight create_tensor: loading tensor blk.48.attn_v.weight create_tensor: loading tensor blk.48.attn_output.weight create_tensor: loading tensor blk.48.attn_q.bias create_tensor: loading tensor blk.48.attn_k.bias create_tensor: loading tensor blk.48.attn_v.bias create_tensor: loading tensor blk.48.ffn_norm.weight create_tensor: loading tensor blk.48.ffn_gate.weight create_tensor: loading tensor blk.48.ffn_down.weight create_tensor: loading tensor blk.48.ffn_up.weight create_tensor: loading tensor blk.49.attn_norm.weight create_tensor: loading tensor blk.49.attn_q.weight create_tensor: loading tensor blk.49.attn_k.weight create_tensor: loading tensor blk.49.attn_v.weight create_tensor: loading tensor blk.49.attn_output.weight create_tensor: loading tensor blk.49.attn_q.bias create_tensor: loading tensor blk.49.attn_k.bias create_tensor: loading tensor blk.49.attn_v.bias create_tensor: loading tensor blk.49.ffn_norm.weight create_tensor: loading tensor blk.49.ffn_gate.weight create_tensor: loading tensor blk.49.ffn_down.weight create_tensor: loading tensor blk.49.ffn_up.weight create_tensor: loading tensor blk.50.attn_norm.weight create_tensor: loading tensor blk.50.attn_q.weight create_tensor: loading tensor blk.50.attn_k.weight create_tensor: loading tensor blk.50.attn_v.weight create_tensor: loading tensor blk.50.attn_output.weight create_tensor: loading tensor blk.50.attn_q.bias create_tensor: loading tensor blk.50.attn_k.bias create_tensor: loading tensor blk.50.attn_v.bias create_tensor: loading tensor blk.50.ffn_norm.weight create_tensor: loading tensor blk.50.ffn_gate.weight create_tensor: loading tensor blk.50.ffn_down.weight create_tensor: loading tensor blk.50.ffn_up.weight create_tensor: loading tensor blk.51.attn_norm.weight create_tensor: loading tensor blk.51.attn_q.weight create_tensor: loading tensor blk.51.attn_k.weight create_tensor: loading tensor blk.51.attn_v.weight create_tensor: loading tensor blk.51.attn_output.weight create_tensor: loading tensor blk.51.attn_q.bias create_tensor: loading tensor blk.51.attn_k.bias create_tensor: loading tensor blk.51.attn_v.bias create_tensor: loading tensor blk.51.ffn_norm.weight create_tensor: loading tensor blk.51.ffn_gate.weight create_tensor: loading tensor blk.51.ffn_down.weight create_tensor: loading tensor blk.51.ffn_up.weight create_tensor: loading tensor blk.52.attn_norm.weight create_tensor: loading tensor blk.52.attn_q.weight create_tensor: loading tensor blk.52.attn_k.weight create_tensor: loading tensor blk.52.attn_v.weight create_tensor: loading tensor blk.52.attn_output.weight create_tensor: loading tensor blk.52.attn_q.bias create_tensor: loading tensor blk.52.attn_k.bias create_tensor: loading tensor blk.52.attn_v.bias create_tensor: loading tensor blk.52.ffn_norm.weight create_tensor: loading tensor blk.52.ffn_gate.weight create_tensor: loading tensor blk.52.ffn_down.weight create_tensor: loading tensor blk.52.ffn_up.weight create_tensor: loading tensor blk.53.attn_norm.weight create_tensor: loading tensor blk.53.attn_q.weight create_tensor: loading tensor blk.53.attn_k.weight create_tensor: loading tensor blk.53.attn_v.weight create_tensor: loading tensor blk.53.attn_output.weight create_tensor: loading tensor blk.53.attn_q.bias create_tensor: loading tensor blk.53.attn_k.bias create_tensor: loading tensor blk.53.attn_v.bias create_tensor: loading tensor blk.53.ffn_norm.weight create_tensor: loading tensor blk.53.ffn_gate.weight create_tensor: loading tensor blk.53.ffn_down.weight create_tensor: loading tensor blk.53.ffn_up.weight create_tensor: loading tensor blk.54.attn_norm.weight create_tensor: loading tensor blk.54.attn_q.weight create_tensor: loading tensor blk.54.attn_k.weight create_tensor: loading tensor blk.54.attn_v.weight create_tensor: loading tensor blk.54.attn_output.weight create_tensor: loading tensor blk.54.attn_q.bias create_tensor: loading tensor blk.54.attn_k.bias create_tensor: loading tensor blk.54.attn_v.bias create_tensor: loading tensor blk.54.ffn_norm.weight create_tensor: loading tensor blk.54.ffn_gate.weight create_tensor: loading tensor blk.54.ffn_down.weight create_tensor: loading tensor blk.54.ffn_up.weight create_tensor: loading tensor blk.55.attn_norm.weight create_tensor: loading tensor blk.55.attn_q.weight create_tensor: loading tensor blk.55.attn_k.weight create_tensor: loading tensor blk.55.attn_v.weight create_tensor: loading tensor blk.55.attn_output.weight create_tensor: loading tensor blk.55.attn_q.bias create_tensor: loading tensor blk.55.attn_k.bias create_tensor: loading tensor blk.55.attn_v.bias create_tensor: loading tensor blk.55.ffn_norm.weight create_tensor: loading tensor blk.55.ffn_gate.weight create_tensor: loading tensor blk.55.ffn_down.weight create_tensor: loading tensor blk.55.ffn_up.weight create_tensor: loading tensor blk.56.attn_norm.weight create_tensor: loading tensor blk.56.attn_q.weight create_tensor: loading tensor blk.56.attn_k.weight create_tensor: loading tensor blk.56.attn_v.weight create_tensor: loading tensor blk.56.attn_output.weight create_tensor: loading tensor blk.56.attn_q.bias create_tensor: loading tensor blk.56.attn_k.bias create_tensor: loading tensor blk.56.attn_v.bias create_tensor: loading tensor blk.56.ffn_norm.weight create_tensor: loading tensor blk.56.ffn_gate.weight create_tensor: loading tensor blk.56.ffn_down.weight create_tensor: loading tensor blk.56.ffn_up.weight create_tensor: loading tensor blk.57.attn_norm.weight create_tensor: loading tensor blk.57.attn_q.weight create_tensor: loading tensor blk.57.attn_k.weight create_tensor: loading tensor blk.57.attn_v.weight create_tensor: loading tensor blk.57.attn_output.weight create_tensor: loading tensor blk.57.attn_q.bias create_tensor: loading tensor blk.57.attn_k.bias create_tensor: loading tensor blk.57.attn_v.bias create_tensor: loading tensor blk.57.ffn_norm.weight create_tensor: loading tensor blk.57.ffn_gate.weight create_tensor: loading tensor blk.57.ffn_down.weight create_tensor: loading tensor blk.57.ffn_up.weight create_tensor: loading tensor blk.58.attn_norm.weight create_tensor: loading tensor blk.58.attn_q.weight create_tensor: loading tensor blk.58.attn_k.weight create_tensor: loading tensor blk.58.attn_v.weight create_tensor: loading tensor blk.58.attn_output.weight create_tensor: loading tensor blk.58.attn_q.bias create_tensor: loading tensor blk.58.attn_k.bias create_tensor: loading tensor blk.58.attn_v.bias create_tensor: loading tensor blk.58.ffn_norm.weight create_tensor: loading tensor blk.58.ffn_gate.weight create_tensor: loading tensor blk.58.ffn_down.weight create_tensor: loading tensor blk.58.ffn_up.weight create_tensor: loading tensor blk.59.attn_norm.weight create_tensor: loading tensor blk.59.attn_q.weight create_tensor: loading tensor blk.59.attn_k.weight create_tensor: loading tensor blk.59.attn_v.weight create_tensor: loading tensor blk.59.attn_output.weight create_tensor: loading tensor blk.59.attn_q.bias create_tensor: loading tensor blk.59.attn_k.bias create_tensor: loading tensor blk.59.attn_v.bias create_tensor: loading tensor blk.59.ffn_norm.weight create_tensor: loading tensor blk.59.ffn_gate.weight create_tensor: loading tensor blk.59.ffn_down.weight create_tensor: loading tensor blk.59.ffn_up.weight create_tensor: loading tensor blk.60.attn_norm.weight create_tensor: loading tensor blk.60.attn_q.weight create_tensor: loading tensor blk.60.attn_k.weight create_tensor: loading tensor blk.60.attn_v.weight create_tensor: loading tensor blk.60.attn_output.weight create_tensor: loading tensor blk.60.attn_q.bias create_tensor: loading tensor blk.60.attn_k.bias create_tensor: loading tensor blk.60.attn_v.bias create_tensor: loading tensor blk.60.ffn_norm.weight create_tensor: loading tensor blk.60.ffn_gate.weight create_tensor: loading tensor blk.60.ffn_down.weight create_tensor: loading tensor blk.60.ffn_up.weight create_tensor: loading tensor blk.61.attn_norm.weight create_tensor: loading tensor blk.61.attn_q.weight create_tensor: loading tensor blk.61.attn_k.weight create_tensor: loading tensor blk.61.attn_v.weight create_tensor: loading tensor blk.61.attn_output.weight create_tensor: loading tensor blk.61.attn_q.bias create_tensor: loading tensor blk.61.attn_k.bias create_tensor: loading tensor blk.61.attn_v.bias create_tensor: loading tensor blk.61.ffn_norm.weight create_tensor: loading tensor blk.61.ffn_gate.weight create_tensor: loading tensor blk.61.ffn_down.weight create_tensor: loading tensor blk.61.ffn_up.weight create_tensor: loading tensor blk.62.attn_norm.weight create_tensor: loading tensor blk.62.attn_q.weight create_tensor: loading tensor blk.62.attn_k.weight create_tensor: loading tensor blk.62.attn_v.weight create_tensor: loading tensor blk.62.attn_output.weight create_tensor: loading tensor blk.62.attn_q.bias create_tensor: loading tensor blk.62.attn_k.bias create_tensor: loading tensor blk.62.attn_v.bias create_tensor: loading tensor blk.62.ffn_norm.weight create_tensor: loading tensor blk.62.ffn_gate.weight create_tensor: loading tensor blk.62.ffn_down.weight create_tensor: loading tensor blk.62.ffn_up.weight create_tensor: loading tensor blk.63.attn_norm.weight create_tensor: loading tensor blk.63.attn_q.weight create_tensor: loading tensor blk.63.attn_k.weight create_tensor: loading tensor blk.63.attn_v.weight create_tensor: loading tensor blk.63.attn_output.weight create_tensor: loading tensor blk.63.attn_q.bias create_tensor: loading tensor blk.63.attn_k.bias create_tensor: loading tensor blk.63.attn_v.bias create_tensor: loading tensor blk.63.ffn_norm.weight create_tensor: loading tensor blk.63.ffn_gate.weight create_tensor: loading tensor blk.63.ffn_down.weight create_tensor: loading tensor blk.63.ffn_up.weight load_tensors: tensor 'token_embd.weight' (q4_K) (and 0 others) cannot be used with preferred buffer type CUDA_Host, using CPU instead load_tensors: offloading 0 repeating layers to GPU load_tensors: offloaded 0/65 layers to GPU load_tensors: CUDA_Host model buffer size = 18508.35 MiB load_tensors: CPU model buffer size = 417.66 MiB load_all_data: buffer type CUDA_Host is not the default buffer type for device CUDA0 for async uploads time=2025-11-22T21:16:59.342+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.00" time=2025-11-22T21:17:04.372+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.03" time=2025-11-22T21:17:04.623+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.03" time=2025-11-22T21:17:04.876+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.03" time=2025-11-22T21:17:05.380+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.04" time=2025-11-22T21:17:06.136+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.04" time=2025-11-22T21:17:06.639+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.05" time=2025-11-22T21:17:06.891+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.05" time=2025-11-22T21:17:07.647+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.05" time=2025-11-22T21:17:08.402+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.06" time=2025-11-22T21:17:08.906+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.06" time=2025-11-22T21:17:09.158+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.07" time=2025-11-22T21:17:09.914+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.07" time=2025-11-22T21:17:10.669+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08" time=2025-11-22T21:17:11.272+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08" time=2025-11-22T21:17:11.524+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08" time=2025-11-22T21:17:11.776+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.08" time=2025-11-22T21:17:12.280+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.09" time=2025-11-22T21:17:13.036+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.09" time=2025-11-22T21:17:13.540+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10" time=2025-11-22T21:17:13.793+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10" time=2025-11-22T21:17:14.045+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10" time=2025-11-22T21:17:14.550+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.10" time=2025-11-22T21:17:15.306+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.11" time=2025-11-22T21:17:15.811+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.11" time=2025-11-22T21:17:16.064+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.11" time=2025-11-22T21:17:16.820+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.12" time=2025-11-22T21:17:17.577+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.12" time=2025-11-22T21:17:18.081+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.13" time=2025-11-22T21:17:18.335+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.13" time=2025-11-22T21:17:19.092+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.13" time=2025-11-22T21:17:19.849+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.14" time=2025-11-22T21:17:20.353+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.14" time=2025-11-22T21:17:20.607+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.14" time=2025-11-22T21:17:21.123+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.15" time=2025-11-22T21:17:22.134+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.15" time=2025-11-22T21:17:22.641+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.16" time=2025-11-22T21:17:22.894+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.16" time=2025-11-22T21:17:23.400+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.16" time=2025-11-22T21:17:46.032+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.32" time=2025-11-22T21:17:46.286+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.33" time=2025-11-22T21:17:46.790+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.33" time=2025-11-22T21:17:47.044+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.33" time=2025-11-22T21:17:47.549+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.34" time=2025-11-22T21:17:48.307+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.34" time=2025-11-22T21:17:48.812+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.35" time=2025-11-22T21:17:49.065+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.35" time=2025-11-22T21:17:49.823+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.35" time=2025-11-22T21:17:50.327+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36" time=2025-11-22T21:17:50.833+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36" time=2025-11-22T21:17:51.086+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36" time=2025-11-22T21:17:51.339+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.36" time=2025-11-22T21:17:51.845+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.37" time=2025-11-22T21:17:52.855+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.37" time=2025-11-22T21:17:53.360+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.37" time=2025-11-22T21:17:53.613+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38" time=2025-11-22T21:17:53.866+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38" time=2025-11-22T21:17:54.372+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38" time=2025-11-22T21:17:54.879+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.38" time=2025-11-22T21:17:55.637+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.39" time=2025-11-22T21:17:55.890+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.39" time=2025-11-22T21:17:56.395+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.39" time=2025-11-22T21:17:56.901+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40" time=2025-11-22T21:17:57.406+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40" time=2025-11-22T21:17:57.659+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40" time=2025-11-22T21:17:57.912+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.40" time=2025-11-22T21:17:58.418+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.41" time=2025-11-22T21:17:59.175+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.41" time=2025-11-22T21:17:59.682+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.42" time=2025-11-22T21:17:59.935+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.42" time=2025-11-22T21:18:00.441+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.42" time=2025-11-22T21:18:01.200+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.43" time=2025-11-22T21:18:01.705+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.43" time=2025-11-22T21:18:01.958+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.43" time=2025-11-22T21:18:02.464+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.44" time=2025-11-22T21:18:02.969+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.44" time=2025-11-22T21:18:03.475+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.45" time=2025-11-22T21:18:03.728+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.45" time=2025-11-22T21:18:04.487+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.45" time=2025-11-22T21:18:05.246+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.46" time=2025-11-22T21:18:05.752+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.46" time=2025-11-22T21:18:06.006+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.46" time=2025-11-22T21:18:06.513+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.47" time=2025-11-22T21:18:07.019+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.47" time=2025-11-22T21:18:07.525+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.48" time=2025-11-22T21:18:07.778+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.48" time=2025-11-22T21:18:08.538+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.48" time=2025-11-22T21:18:09.044+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.49" time=2025-11-22T21:18:09.550+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.49" time=2025-11-22T21:18:09.803+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.49" time=2025-11-22T21:18:10.309+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.50" time=2025-11-22T21:18:11.321+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.50" time=2025-11-22T21:18:11.828+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51" time=2025-11-22T21:18:12.082+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51" time=2025-11-22T21:18:12.588+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51" time=2025-11-22T21:18:13.095+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.51" time=2025-11-22T21:18:13.856+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.52" time=2025-11-22T21:18:14.110+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.52" time=2025-11-22T21:18:14.617+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.52" time=2025-11-22T21:18:15.124+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53" time=2025-11-22T21:18:15.631+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53" time=2025-11-22T21:18:15.885+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53" time=2025-11-22T21:18:16.139+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.53" time=2025-11-22T21:18:16.645+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.54" time=2025-11-22T21:18:17.404+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.54" time=2025-11-22T21:18:17.911+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.55" time=2025-11-22T21:18:18.165+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.55" time=2025-11-22T21:18:18.672+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.55" time=2025-11-22T21:18:19.433+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.56" time=2025-11-22T21:18:19.940+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.56" time=2025-11-22T21:18:20.195+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.56" time=2025-11-22T21:18:20.702+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.57" time=2025-11-22T21:18:21.209+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.57" time=2025-11-22T21:18:21.968+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.58" time=2025-11-22T21:18:22.222+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.58" time=2025-11-22T21:18:22.983+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.58" time=2025-11-22T21:18:23.744+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.59" time=2025-11-22T21:18:24.504+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.59" time=2025-11-22T21:18:24.759+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.59" time=2025-11-22T21:18:25.520+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.60" time=2025-11-22T21:18:26.027+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.60" time=2025-11-22T21:18:26.788+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.61" time=2025-11-22T21:18:27.042+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.61" time=2025-11-22T21:18:27.549+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.61" time=2025-11-22T21:18:28.057+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62" time=2025-11-22T21:18:28.564+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62" time=2025-11-22T21:18:28.818+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62" time=2025-11-22T21:18:29.073+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.62" time=2025-11-22T21:18:29.580+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.63" time=2025-11-22T21:18:30.341+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.63" time=2025-11-22T21:18:30.850+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.64" time=2025-11-22T21:18:31.104+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.64" time=2025-11-22T21:18:31.611+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.64" time=2025-11-22T21:18:32.371+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65" time=2025-11-22T21:18:32.879+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65" time=2025-11-22T21:18:33.133+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65" time=2025-11-22T21:18:33.641+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.65" time=2025-11-22T21:18:34.148+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.66" time=2025-11-22T21:18:34.657+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.66" time=2025-11-22T21:18:34.912+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.66" time=2025-11-22T21:18:35.420+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.67" time=2025-11-22T21:18:36.434+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.67" time=2025-11-22T21:18:36.942+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68" time=2025-11-22T21:18:37.196+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68" time=2025-11-22T21:18:37.450+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68" time=2025-11-22T21:18:37.958+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.68" time=2025-11-22T21:18:38.466+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.69" time=2025-11-22T21:18:38.973+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.69" time=2025-11-22T21:18:39.228+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.69" time=2025-11-22T21:18:39.735+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.70" time=2025-11-22T21:18:40.499+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.70" time=2025-11-22T21:18:41.007+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.71" time=2025-11-22T21:18:41.261+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.71" time=2025-11-22T21:18:41.771+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.71" time=2025-11-22T21:18:42.533+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72" time=2025-11-22T21:18:43.042+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72" time=2025-11-22T21:18:43.298+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72" time=2025-11-22T21:18:43.552+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.72" time=2025-11-22T21:18:44.060+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.73" time=2025-11-22T21:18:44.567+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.73" time=2025-11-22T21:18:45.076+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.74" time=2025-11-22T21:18:45.330+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.74" time=2025-11-22T21:18:45.838+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.74" time=2025-11-22T21:18:46.601+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.75" time=2025-11-22T21:18:47.110+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.75" time=2025-11-22T21:18:47.365+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.75" time=2025-11-22T21:18:48.127+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.76" time=2025-11-22T21:18:48.890+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.76" time=2025-11-22T21:18:49.399+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.77" time=2025-11-22T21:18:49.653+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.77" time=2025-11-22T21:18:50.415+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.77" time=2025-11-22T21:18:50.924+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78" time=2025-11-22T21:18:51.433+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78" time=2025-11-22T21:18:51.688+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78" time=2025-11-22T21:18:51.943+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78" time=2025-11-22T21:18:52.451+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.78" time=2025-11-22T21:18:52.960+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.79" time=2025-11-22T21:18:53.723+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.79" time=2025-11-22T21:18:53.978+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.79" time=2025-11-22T21:18:54.486+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.80" time=2025-11-22T21:18:55.503+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.80" time=2025-11-22T21:18:56.012+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.81" time=2025-11-22T21:18:56.267+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.81" time=2025-11-22T21:18:57.030+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.81" time=2025-11-22T21:18:57.540+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.82" time=2025-11-22T21:18:58.303+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.82" time=2025-11-22T21:18:58.559+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.82" time=2025-11-22T21:18:59.069+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.83" time=2025-11-22T21:18:59.578+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.83" time=2025-11-22T21:19:00.088+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.84" time=2025-11-22T21:19:00.342+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.84" time=2025-11-22T21:19:00.852+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.84" time=2025-11-22T21:19:01.616+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.85" time=2025-11-22T21:19:02.127+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.85" time=2025-11-22T21:19:02.382+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.85" time=2025-11-22T21:19:03.146+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.86" time=2025-11-22T21:19:03.911+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.86" time=2025-11-22T21:19:04.421+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.87" time=2025-11-22T21:19:04.677+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.87" time=2025-11-22T21:19:05.186+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.87" time=2025-11-22T21:19:05.950+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.88" time=2025-11-22T21:19:06.459+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.88" time=2025-11-22T21:19:06.715+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.88" time=2025-11-22T21:19:06.970+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.89" time=2025-11-22T21:19:07.481+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.89" time=2025-11-22T21:19:08.246+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.90" time=2025-11-22T21:19:09.012+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.90" time=2025-11-22T21:19:09.267+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.90" time=2025-11-22T21:19:09.775+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.91" time=2025-11-22T21:19:10.795+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.91" time=2025-11-22T21:19:11.305+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.92" time=2025-11-22T21:19:11.560+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.92" time=2025-11-22T21:19:12.071+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.92" time=2025-11-22T21:19:12.836+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.93" time=2025-11-22T21:19:13.602+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.93" time=2025-11-22T21:19:13.857+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.93" time=2025-11-22T21:19:14.368+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.94" time=2025-11-22T21:19:15.133+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.94" time=2025-11-22T21:19:15.643+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.95" time=2025-11-22T21:19:15.898+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.95" time=2025-11-22T21:19:16.410+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.95" time=2025-11-22T21:19:17.175+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96" time=2025-11-22T21:19:17.686+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96" time=2025-11-22T21:19:17.941+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96" time=2025-11-22T21:19:18.197+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.96" time=2025-11-22T21:19:18.708+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.97" time=2025-11-22T21:19:19.473+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.97" load_all_data: no device found for buffer type CPU for async uploads time=2025-11-22T21:19:19.983+01:00 level=DEBUG source=server.go:1338 msg="model load progress 0.98" llama_context: constructing llama_context llama_context: n_seq_max = 3 llama_context: n_ctx = 98304 llama_context: n_ctx_per_seq = 32768 llama_context: n_batch = 1536 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = disabled llama_context: kv_unified = false llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 set_abort_callback: call llama_context: CPU output buffer size = 1.80 MiB create_memory: n_ctx = 98304 (padded) llama_kv_cache: layer 0: dev = CPU llama_kv_cache: layer 1: dev = CPU llama_kv_cache: layer 2: dev = CPU llama_kv_cache: layer 3: dev = CPU llama_kv_cache: layer 4: dev = CPU llama_kv_cache: layer 5: dev = CPU llama_kv_cache: layer 6: dev = CPU llama_kv_cache: layer 7: dev = CPU llama_kv_cache: layer 8: dev = CPU llama_kv_cache: layer 9: dev = CPU llama_kv_cache: layer 10: dev = CPU llama_kv_cache: layer 11: dev = CPU llama_kv_cache: layer 12: dev = CPU llama_kv_cache: layer 13: dev = CPU llama_kv_cache: layer 14: dev = CPU llama_kv_cache: layer 15: dev = CPU llama_kv_cache: layer 16: dev = CPU llama_kv_cache: layer 17: dev = CPU llama_kv_cache: layer 18: dev = CPU llama_kv_cache: layer 19: dev = CPU llama_kv_cache: layer 20: dev = CPU llama_kv_cache: layer 21: dev = CPU llama_kv_cache: layer 22: dev = CPU llama_kv_cache: layer 23: dev = CPU llama_kv_cache: layer 24: dev = CPU llama_kv_cache: layer 25: dev = CPU llama_kv_cache: layer 26: dev = CPU llama_kv_cache: layer 27: dev = CPU llama_kv_cache: layer 28: dev = CPU llama_kv_cache: layer 29: dev = CPU llama_kv_cache: layer 30: dev = CPU llama_kv_cache: layer 31: dev = CPU llama_kv_cache: layer 32: dev = CPU llama_kv_cache: layer 33: dev = CPU llama_kv_cache: layer 34: dev = CPU llama_kv_cache: layer 35: dev = CPU llama_kv_cache: layer 36: dev = CPU llama_kv_cache: layer 37: dev = CPU llama_kv_cache: layer 38: dev = CPU llama_kv_cache: layer 39: dev = CPU llama_kv_cache: layer 40: dev = CPU llama_kv_cache: layer 41: dev = CPU llama_kv_cache: layer 42: dev = CPU llama_kv_cache: layer 43: dev = CPU llama_kv_cache: layer 44: dev = CPU llama_kv_cache: layer 45: dev = CPU llama_kv_cache: layer 46: dev = CPU llama_kv_cache: layer 47: dev = CPU llama_kv_cache: layer 48: dev = CPU llama_kv_cache: layer 49: dev = CPU llama_kv_cache: layer 50: dev = CPU llama_kv_cache: layer 51: dev = CPU llama_kv_cache: layer 52: dev = CPU llama_kv_cache: layer 53: dev = CPU llama_kv_cache: layer 54: dev = CPU llama_kv_cache: layer 55: dev = CPU llama_kv_cache: layer 56: dev = CPU llama_kv_cache: layer 57: dev = CPU llama_kv_cache: layer 58: dev = CPU llama_kv_cache: layer 59: dev = CPU llama_kv_cache: layer 60: dev = CPU llama_kv_cache: layer 61: dev = CPU llama_kv_cache: layer 62: dev = CPU llama_kv_cache: layer 63: dev = CPU llama_kv_cache: CPU KV buffer size = 24576.00 MiB time=2025-11-22T21:19:23.555+01:00 level=DEBUG source=server.go:1338 msg="model load progress 1.00" time=2025-11-22T21:19:23.812+01:00 level=DEBUG source=server.go:1341 msg="model load completed, waiting for server to become available" status="llm server loading model" llama_kv_cache: size = 24576.00 MiB ( 32768 cells, 64 layers, 3/3 seqs), K (f16): 12288.00 MiB, V (f16): 12288.00 MiB llama_context: enumerating backends llama_context: backend_ptrs.size() = 4 llama_context: max_nodes = 6168 llama_context: reserving full memory module llama_context: worst-case: n_tokens = 512, n_seqs = 3, n_outputs = 3 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 3, n_outputs = 512 graph_reserve: making n_tokens a multiple of n_seqs - n_tokens = 513, n_seqs = 3, n_outputs = 512 graph_reserve: reserving a graph for ubatch with n_tokens = 3, n_seqs = 3, n_outputs = 3 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 3, n_outputs = 512 graph_reserve: making n_tokens a multiple of n_seqs - n_tokens = 513, n_seqs = 3, n_outputs = 512 llama_context: CUDA0 compute buffer size = 3051.08 MiB llama_context: CUDA_Host compute buffer size = 86.04 MiB llama_context: graph nodes = 2566 llama_context: graph splits = 900 (with bs=512), 1 (with bs=1) time=2025-11-22T21:21:15.211+01:00 level=INFO source=server.go:1332 msg="llama runner started in 261.95 seconds" time=2025-11-22T21:21:15.213+01:00 level=INFO source=sched.go:517 msg="loaded runners" count=1 time=2025-11-22T21:21:15.215+01:00 level=INFO source=server.go:1294 msg="waiting for llama runner to start responding" time=2025-11-22T21:21:15.223+01:00 level=INFO source=server.go:1332 msg="llama runner started in 261.97 seconds" time=2025-11-22T21:21:15.224+01:00 level=DEBUG source=sched.go:529 msg="finished setting up" runner.name=registry.ollama.ai/library/qwen2.5-coder:32b runner.size="70.8 GiB" runner.vram="28.7 GiB" runner.parallel=3 runner.pid=77624 runner.model=H:\LLM\blobs\sha256-ac3d1ba8aa77755dab3806d9024e9c385ea0d5b412d6bdf9157f8a4a7e9fc0d9 runner.num_ctx=131072 time=2025-11-22T21:21:15.229+01:00 level=DEBUG source=server.go:1465 msg="completion request" images=0 prompt=1586 format="" ``` <img width="913" height="634" alt="Image" src="https://github.com/user-attachments/assets/2c38e212-7428-44a0-89b1-e8ee908a0e3f" />
Author
Owner

@catclaw commented on GitHub (Nov 23, 2025):

I've finally fixed it!!
I copied ggml-base.dll from Ollama installation folder to C:\Windows and now all the models are loading to VRAM of my GPUs!

<!-- gh-comment-id:3567445779 --> @catclaw commented on GitHub (Nov 23, 2025): **I've finally fixed it!!** I copied `ggml-base.dll` from Ollama installation folder to `C:\Windows` and now all the models are loading to VRAM of my GPUs!
Author
Owner

@dhiltgen commented on GitHub (Dec 1, 2025):

@catclaw that "fix" will likely break on the next update. You should remove ggml-base.dll from any system-wide locations so that the version that is found is within the Ollama install directory, as that file changes on every release, and is not binary compatible between releases.

<!-- gh-comment-id:3598543532 --> @dhiltgen commented on GitHub (Dec 1, 2025): @catclaw that "fix" will likely break on the next update. You should remove ggml-base.dll from any system-wide locations so that the version that is found is within the Ollama install directory, as that file changes on every release, and is not binary compatible between releases.
Author
Owner

@mwprado commented on GitHub (Dec 15, 2025):

I Use fedora 43, cuda 12 and tesla P4 the last after 0.12.5, it stops GPU use.

<!-- gh-comment-id:3652664573 --> @mwprado commented on GitHub (Dec 15, 2025): I Use fedora 43, cuda 12 and tesla P4 the last after 0.12.5, it stops GPU use.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#54889