[GH-ISSUE #14083] ollama version 0.15.4 cant't discover GPU #9195

Closed
opened 2026-04-12 22:02:39 -05:00 by GiteaMirror · 7 comments
Owner

Originally created by @qiaojian1989 on GitHub (Feb 5, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/14083

What is the issue?

root@localai:~# nvidia-smi
Thu Feb 5 02:19:13 2026
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 590.48.01 Driver Version: 590.48.01 CUDA Version: 13.1 |
+-----------------------------------------+------------------------+----------------------+
| GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 Tesla T4 On | 00000000:03:00.0 Off | 0 |
| N/A 40C P8 14W / 70W | 0MiB / 15360MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
| 1 Tesla T4 On | 00000000:13:00.0 Off | Off |
| N/A 44C P8 13W / 70W | 0MiB / 16384MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+

+-----------------------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=========================================================================================|
| No running processes found |
+-----------------------------------------------------------------------------------------+
root@localai:# systemctl restart ollama.service
root@localai:
# systemctl status ollama.service
● ollama.service - Ollama Service
Loaded: loaded (/etc/systemd/system/ollama.service; enabled; vendor preset: enabled)
Active: active (running) since Thu 2026-02-05 02:19:25 UTC; 5s ago
Main PID: 69621 (ollama)
Tasks: 12 (limit: 76914)
Memory: 11.6M
CPU: 86ms
CGroup: /system.slice/ollama.service
└─69621 /usr/local/bin/ollama serve

Feb 05 02:19:25 localai systemd[1]: Started Ollama Service.
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.064Z level=INFO source=routes.go:1631 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES>
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.069Z level=INFO source=images.go:473 msg="total blobs: 53"
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.071Z level=INFO source=images.go:480 msg="total unused blobs removed: 0"
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.072Z level=INFO source=routes.go:1684 msg="Listening on [::]:11434 (version 0.15.4)"
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.073Z level=INFO source=runner.go:67 msg="discovering available GPUs..."
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.073Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --p>
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.110Z level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cp>
Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.110Z level=INFO source=routes.go:1725 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"

Relevant log output


OS

Linux

GPU

Nvidia

CPU

Intel

Ollama version

0.15.4

Originally created by @qiaojian1989 on GitHub (Feb 5, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/14083 ### What is the issue? root@localai:~# nvidia-smi Thu Feb 5 02:19:13 2026 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 590.48.01 Driver Version: 590.48.01 CUDA Version: 13.1 | +-----------------------------------------+------------------------+----------------------+ | GPU Name Persistence-M | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 Tesla T4 On | 00000000:03:00.0 Off | 0 | | N/A 40C P8 14W / 70W | 0MiB / 15360MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 Tesla T4 On | 00000000:13:00.0 Off | Off | | N/A 44C P8 13W / 70W | 0MiB / 16384MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ +-----------------------------------------------------------------------------------------+ | Processes: | | GPU GI CI PID Type Process name GPU Memory | | ID ID Usage | |=========================================================================================| | No running processes found | +-----------------------------------------------------------------------------------------+ root@localai:~# systemctl restart ollama.service root@localai:~# systemctl status ollama.service ● ollama.service - Ollama Service Loaded: loaded (/etc/systemd/system/ollama.service; enabled; vendor preset: enabled) Active: active (running) since Thu 2026-02-05 02:19:25 UTC; 5s ago Main PID: 69621 (ollama) Tasks: 12 (limit: 76914) Memory: 11.6M CPU: 86ms CGroup: /system.slice/ollama.service └─69621 /usr/local/bin/ollama serve Feb 05 02:19:25 localai systemd[1]: Started Ollama Service. Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.064Z level=INFO source=routes.go:1631 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES> Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.069Z level=INFO source=images.go:473 msg="total blobs: 53" Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.071Z level=INFO source=images.go:480 msg="total unused blobs removed: 0" Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.072Z level=INFO source=routes.go:1684 msg="Listening on [::]:11434 (version 0.15.4)" Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.073Z level=INFO source=runner.go:67 msg="discovering available GPUs..." Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.073Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --p> Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.110Z level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cp> Feb 05 02:19:25 localai ollama[69621]: time=2026-02-05T02:19:25.110Z level=INFO source=routes.go:1725 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB" ### Relevant log output ```shell ``` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.15.4
GiteaMirror added the bug label 2026-04-12 22:02:39 -05:00
Author
Owner

@rick-github commented on GitHub (Feb 5, 2026):

Set OLLAMA_DEBUG=2 in the server environment, restart the server, and post the output of:

journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p'
<!-- gh-comment-id:3850732200 --> @rick-github commented on GitHub (Feb 5, 2026): Set `OLLAMA_DEBUG=2` in the server environment, restart the server, and post the output of: ``` journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p' ```
Author
Owner

@qiaojian1989 commented on GitHub (Feb 5, 2026):

root@localai:~# journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p'
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:03.989Z level=INFO source=routes.go:1631 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/usr/share/ollama/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false OLLAMA_VULKAN:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:473 msg="total blobs: 53"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:480 msg="total unused blobs removed: 0"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=INFO source=routes.go:1684 msg="Listening on [::]:11434 (version 0.15.4)"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=DEBUG source=sched.go:121 msg="starting llm scheduler"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=INFO source=runner.go:67 msg="discovering available GPUs..."
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=TRACE source=runner.go:440 msg="starting runner for device discovery" libDirs=[/usr/local/lib/ollama] extraEnvs=map[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 33703"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=DEBUG source=server.go:430 msg=subprocess PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/snap/bin OLLAMA_HOST=0.0.0.0:11434 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.534Z level=INFO source=runner.go:1405 msg="starting ollama engine"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.535Z level=INFO source=runner.go:1440 msg="Server listening on 127.0.0.1:33703"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=general.architecture type=string
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.file_type default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.name default=""
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.description default=""
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.546Z level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc)
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.pooling_type default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.expert_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.pre default=""
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.embedding_length default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count_kv default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.key_length default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.dimension_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.freq_base default=100000
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.scaling.factor default=1
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1380 msg="dummy model load took" duration=58.248689ms
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1385 msg="gathering device infos took" duration=998ns
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:467 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] devices=[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:437 msg="bootstrap discovery took" duration=252.252062ms OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] extra_envs=map[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:124 msg="evaluating which, if any, devices to filter out" initial_count=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:174 msg="supported GPU library combinations before filtering" supported=map[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=253.189035ms
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="62.8 GiB" available="61.6 GiB"
<!-- gh-comment-id:3850757858 --> @qiaojian1989 commented on GitHub (Feb 5, 2026): ``` root@localai:~# journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p' Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:03.989Z level=INFO source=routes.go:1631 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/usr/share/ollama/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false OLLAMA_VULKAN:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:473 msg="total blobs: 53" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:480 msg="total unused blobs removed: 0" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=INFO source=routes.go:1684 msg="Listening on [::]:11434 (version 0.15.4)" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=DEBUG source=sched.go:121 msg="starting llm scheduler" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=INFO source=runner.go:67 msg="discovering available GPUs..." Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=TRACE source=runner.go:440 msg="starting runner for device discovery" libDirs=[/usr/local/lib/ollama] extraEnvs=map[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 33703" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=DEBUG source=server.go:430 msg=subprocess PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/snap/bin OLLAMA_HOST=0.0.0.0:11434 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.534Z level=INFO source=runner.go:1405 msg="starting ollama engine" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.535Z level=INFO source=runner.go:1440 msg="Server listening on 127.0.0.1:33703" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=general.architecture type=string Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.file_type default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.name default="" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.description default="" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.546Z level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc) Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.pooling_type default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.expert_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.pre default="" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.embedding_length default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count_kv default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.key_length default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.dimension_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.freq_base default=100000 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.scaling.factor default=1 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1380 msg="dummy model load took" duration=58.248689ms Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1385 msg="gathering device infos took" duration=998ns Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:467 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] devices=[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:437 msg="bootstrap discovery took" duration=252.252062ms OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] extra_envs=map[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:124 msg="evaluating which, if any, devices to filter out" initial_count=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:174 msg="supported GPU library combinations before filtering" supported=map[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=253.189035ms Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="62.8 GiB" available="61.6 GiB" ```
Author
Owner

@qiaojian1989 commented on GitHub (Feb 5, 2026):

Set OLLAMA_DEBUG=2 in the server environment, restart the server, and post the output of:

journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p'

root@localai:~# journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p'
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:03.989Z level=INFO source=routes.go:1631 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/usr/share/ollama/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false OLLAMA_VULKAN:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:473 msg="total blobs: 53"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:480 msg="total unused blobs removed: 0"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=INFO source=routes.go:1684 msg="Listening on [::]:11434 (version 0.15.4)"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=DEBUG source=sched.go:121 msg="starting llm scheduler"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=INFO source=runner.go:67 msg="discovering available GPUs..."
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=TRACE source=runner.go:440 msg="starting runner for device discovery" libDirs=[/usr/local/lib/ollama] extraEnvs=map[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 33703"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=DEBUG source=server.go:430 msg=subprocess PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/snap/bin OLLAMA_HOST=0.0.0.0:11434 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.534Z level=INFO source=runner.go:1405 msg="starting ollama engine"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.535Z level=INFO source=runner.go:1440 msg="Server listening on 127.0.0.1:33703"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=general.architecture type=string
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.file_type default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.name default=""
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.description default=""
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.546Z level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc)
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.pooling_type default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.expert_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}"
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.pre default=""
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.embedding_length default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count_kv default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.key_length default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.dimension_count default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.freq_base default=100000
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.scaling.factor default=1
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1380 msg="dummy model load took" duration=58.248689ms
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1385 msg="gathering device infos took" duration=998ns
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:467 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] devices=[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:437 msg="bootstrap discovery took" duration=252.252062ms OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] extra_envs=map[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:124 msg="evaluating which, if any, devices to filter out" initial_count=0
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:174 msg="supported GPU library combinations before filtering" supported=map[]
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=253.189035ms
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="62.8 GiB" available="61.6 GiB"

<!-- gh-comment-id:3850760293 --> @qiaojian1989 commented on GitHub (Feb 5, 2026): > Set `OLLAMA_DEBUG=2` in the server environment, restart the server, and post the output of: > > ``` > journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p' > ``` root@localai:~# journalctl -u ollama --no-pager --since "$(systemctl show ollama --property=ActiveEnterTimestamp --value)" | sed -ne '/server config/,/inference compute/p' Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:03.989Z level=INFO source=routes.go:1631 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/usr/share/ollama/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false OLLAMA_VULKAN:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:473 msg="total blobs: 53" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.342Z level=INFO source=images.go:480 msg="total unused blobs removed: 0" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=INFO source=routes.go:1684 msg="Listening on [::]:11434 (version 0.15.4)" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.343Z level=DEBUG source=sched.go:121 msg="starting llm scheduler" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=INFO source=runner.go:67 msg="discovering available GPUs..." Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.344Z level=TRACE source=runner.go:440 msg="starting runner for device discovery" libDirs=[/usr/local/lib/ollama] extraEnvs=map[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 33703" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.425Z level=DEBUG source=server.go:430 msg=subprocess PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/snap/bin OLLAMA_HOST=0.0.0.0:11434 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.534Z level=INFO source=runner.go:1405 msg="starting ollama engine" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.535Z level=INFO source=runner.go:1440 msg="Server listening on 127.0.0.1:33703" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=general.architecture type=string Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=gguf.go:589 msg=tokenizer.ggml.model type=string Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.alignment default=32 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.file_type default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.name default="" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=general.description default="" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=INFO source=ggml.go:136 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.546Z level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc) Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.pooling_type default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.expert_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.590Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=tokenizer.ggml.pre default="" Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.block_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.embedding_length default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.head_count_kv default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.595Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.key_length default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.dimension_count default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.freq_base default=100000 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=ggml.go:298 msg="key with type not found" key=llama.rope.scaling.factor default=1 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1380 msg="dummy model load took" duration=58.248689ms Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:1385 msg="gathering device infos took" duration=998ns Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:467 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] devices=[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:437 msg="bootstrap discovery took" duration=252.252062ms OLLAMA_LIBRARY_PATH=[/usr/local/lib/ollama] extra_envs=map[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:124 msg="evaluating which, if any, devices to filter out" initial_count=0 Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=TRACE source=runner.go:174 msg="supported GPU library combinations before filtering" supported=map[] Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=DEBUG source=runner.go:40 msg="GPU bootstrap discovery took" duration=253.189035ms Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.596Z level=INFO source=types.go:60 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="62.8 GiB" available="61.6 GiB"
Author
Owner

@rick-github commented on GitHub (Feb 5, 2026):

Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama
Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.546Z level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc)

No backends found. What's the output of

find /usr/local/lib/ollama
<!-- gh-comment-id:3850764076 --> @rick-github commented on GitHub (Feb 5, 2026): ``` Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.538Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama Feb 05 02:38:04 localai ollama[1063]: time=2026-02-05T02:38:04.546Z level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(gcc) ``` No backends found. What's the output of ``` find /usr/local/lib/ollama ```
Author
Owner

@qiaojian1989 commented on GitHub (Feb 5, 2026):

oh, I found out the problem. Thank you very much, guy.

<!-- gh-comment-id:3850844547 --> @qiaojian1989 commented on GitHub (Feb 5, 2026): oh, I found out the problem. Thank you very much, guy.
Author
Owner

@rick-github commented on GitHub (Feb 5, 2026):

What was the problem?

<!-- gh-comment-id:3850866411 --> @rick-github commented on GitHub (Feb 5, 2026): What was the problem?
Author
Owner

@qiaojian1989 commented on GitHub (Feb 6, 2026):

I found the directory /usr/local/lib/ollama was empty, maybe I was updating ollama making some mistake.

<!-- gh-comment-id:3857294623 --> @qiaojian1989 commented on GitHub (Feb 6, 2026): I found the directory /usr/local/lib/ollama was empty, maybe I was updating ollama making some mistake.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#9195