[GH-ISSUE #8791] Unable to load cudart library in docker or docker compose #67765

Open
opened 2026-05-04 11:37:52 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @nonodev96 on GitHub (Feb 3, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/8791

What is the issue?

2025/02/03 16:40:03 routes.go:1187: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:true OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"
time=2025-02-03T16:40:03.807Z level=INFO source=images.go:432 msg="total blobs: 5"
time=2025-02-03T16:40:03.807Z level=INFO source=images.go:439 msg="total unused blobs removed: 0"
time=2025-02-03T16:40:03.807Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.6-0-g2539f2d-dirty)"
time=2025-02-03T16:40:03.807Z level=DEBUG source=common.go:80 msg="runners located" dir=/usr/lib/ollama/runners
time=2025-02-03T16:40:03.807Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx/ollama_llama_server
time=2025-02-03T16:40:03.808Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx2/ollama_llama_server
time=2025-02-03T16:40:03.808Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cuda_v11_avx/ollama_llama_server
time=2025-02-03T16:40:03.808Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cuda_v12_avx/ollama_llama_server
time=2025-02-03T16:40:03.808Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cuda_v11_avx cuda_v12_avx cpu cpu_avx cpu_avx2]"
time=2025-02-03T16:40:03.808Z level=DEBUG source=routes.go:1268 msg="Override detection logic by setting OLLAMA_LLM_LIBRARY"
time=2025-02-03T16:40:03.808Z level=DEBUG source=sched.go:105 msg="starting llm scheduler"
time=2025-02-03T16:40:03.808Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs"
time=2025-02-03T16:40:03.808Z level=DEBUG source=gpu.go:99 msg="searching for GPU discovery libraries for NVIDIA"
time=2025-02-03T16:40:03.808Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcuda.so*
time=2025-02-03T16:40:03.808Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcuda.so* /usr/lib/ollama/libcuda.so* /usr/local/nvidia/lib/libcuda.so* /usr/local/nvidia/lib64/libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]"
time=2025-02-03T16:40:03.809Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths="[/usr/lib/x86_64-linux-gnu/libcuda.so.1 /usr/lib/wsl/drivers/nv_dispi.inf_amd64_ab3196e1830c9b6c/libcuda.so.1.1]"
initializing /usr/lib/x86_64-linux-gnu/libcuda.so.1
dlsym: cuInit - 0x7f500c220ec0
dlsym: cuDriverGetVersion - 0x7f500c220e80
dlsym: cuDeviceGetCount - 0x7f500c220efd
dlsym: cuDeviceGet - 0x7f500c220ef7
dlsym: cuDeviceGetAttribute - 0x7f500c220de0
dlsym: cuDeviceGetUuid - 0x7f500c220f09
dlsym: cuDeviceGetName - 0x7f500c220f03
[GIN-debug] [WARNING] Creating an Engine instance with the Logger and Recovery middleware already attached.
[GIN-debug] [WARNING] Running in "debug" mode. Switch to "release" mode in production.
 - using env:   export GIN_MODE=release
 - using code:  gin.SetMode(gin.ReleaseMode)
[GIN-debug] POST   /api/pull                 --> github.com/ollama/ollama/server.(*Server).PullHandler-fm (5 handlers)
[GIN-debug] POST   /api/generate             --> github.com/ollama/ollama/server.(*Server).GenerateHandler-fm (5 handlers)
[GIN-debug] POST   /api/chat                 --> github.com/ollama/ollama/server.(*Server).ChatHandler-fm (5 handlers)
[GIN-debug] POST   /api/embed                --> github.com/ollama/ollama/server.(*Server).EmbedHandler-fm (5 handlers)
[GIN-debug] POST   /api/embeddings           --> github.com/ollama/ollama/server.(*Server).EmbeddingsHandler-fm (5 handlers)
[GIN-debug] POST   /api/create               --> github.com/ollama/ollama/server.(*Server).CreateHandler-fm (5 handlers)
[GIN-debug] POST   /api/push                 --> github.com/ollama/ollama/server.(*Server).PushHandler-fm (5 handlers)
[GIN-debug] POST   /api/copy                 --> github.com/ollama/ollama/server.(*Server).CopyHandler-fm (5 handlers)
[GIN-debug] DELETE /api/delete               --> github.com/ollama/ollama/server.(*Server).DeleteHandler-fm (5 handlers)
[GIN-debug] POST   /api/show                 --> github.com/ollama/ollama/server.(*Server).ShowHandler-fm (5 handlers)
[GIN-debug] POST   /api/blobs/:digest        --> github.com/ollama/ollama/server.(*Server).CreateBlobHandler-fm (5 handlers)
[GIN-debug] HEAD   /api/blobs/:digest        --> github.com/ollama/ollama/server.(*Server).HeadBlobHandler-fm (5 handlers)
[GIN-debug] GET    /api/ps                   --> github.com/ollama/ollama/server.(*Server).PsHandler-fm (5 handlers)
[GIN-debug] POST   /v1/chat/completions      --> github.com/ollama/ollama/server.(*Server).ChatHandler-fm (6 handlers)
[GIN-debug] POST   /v1/completions           --> github.com/ollama/ollama/server.(*Server).GenerateHandler-fm (6 handlers)
[GIN-debug] POST   /v1/embeddings            --> github.com/ollama/ollama/server.(*Server).EmbedHandler-fm (6 handlers)
[GIN-debug] GET    /v1/models                --> github.com/ollama/ollama/server.(*Server).ListHandler-fm (6 handlers)
[GIN-debug] GET    /v1/models/:model         --> github.com/ollama/ollama/server.(*Server).ShowHandler-fm (6 handlers)
[GIN-debug] GET    /                         --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func1 (5 handlers)
[GIN-debug] GET    /api/tags                 --> github.com/ollama/ollama/server.(*Server).ListHandler-fm (5 handlers)
[GIN-debug] GET    /api/version              --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func2 (5 handlers)
[GIN-debug] HEAD   /                         --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func1 (5 handlers)
[GIN-debug] HEAD   /api/tags                 --> github.com/ollama/ollama/server.(*Server).ListHandler-fm (5 handlers)
[GIN-debug] HEAD   /api/version              --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func2 (5 handlers)
dlsym: cuCtxCreate_v3 - 0x7f500c220f7b
dlsym: cuMemGetInfo_v2 - 0x7f500c22107d
dlsym: cuCtxDestroy - 0x7f500c220f8d
calling cuInit
cuInit err: 500
time=2025-02-03T16:40:03.884Z level=INFO source=gpu.go:630 msg="Unable to load cudart library /usr/lib/x86_64-linux-gnu/libcuda.so.1: cuda driver library init failure: 500"
initializing /usr/lib/wsl/drivers/nv_dispi.inf_amd64_ab3196e1830c9b6c/libcuda.so.1.1
dlsym: cuInit - 0x7f4fe640e220
dlsym: cuDriverGetVersion - 0x7f4fe640e1f0
dlsym: cuDeviceGetCount - 0x7f4fe640e190
dlsym: cuDeviceGet - 0x7f4fe640e1c0
dlsym: cuDeviceGetAttribute - 0x7f4fe640e040
dlsym: cuDeviceGetUuid - 0x7f4fe640e130
dlsym: cuDeviceGetName - 0x7f4fe640e160
dlsym: cuCtxCreate_v3 - 0x7f4fe640dd70
dlsym: cuMemGetInfo_v2 - 0x7f4fe640d230
dlsym: cuCtxDestroy - 0x7f4fe64087c0
calling cuInit
cuInit err: 500
time=2025-02-03T16:40:03.885Z level=INFO source=gpu.go:630 msg="Unable to load cudart library /usr/lib/wsl/drivers/nv_dispi.inf_amd64_ab3196e1830c9b6c/libcuda.so.1.1: cuda driver library init failure: 500"
time=2025-02-03T16:40:03.885Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcudart.so*
time=2025-02-03T16:40:03.885Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/local/nvidia/lib/libcudart.so* /usr/local/nvidia/lib64/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/local/cuda/lib64/libcudart.so* /usr/lib/x86_64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/x86_64-linux-gnu/libcudart.so* /usr/lib/wsl/lib/libcudart.so* /usr/lib/wsl/drivers/*/libcudart.so* /opt/cuda/lib64/libcudart.so* /usr/local/cuda*/targets/aarch64-linux/lib/libcudart.so* /usr/lib/aarch64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/aarch64-linux-gnu/libcudart.so* /usr/local/cuda/lib*/libcudart.so* /usr/lib*/libcudart.so* /usr/local/lib*/libcudart.so*]"
time=2025-02-03T16:40:03.885Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths="[/usr/lib/ollama/libcudart.so.11.3.109 /usr/lib/ollama/libcudart.so.12.4.127]"
cudaSetDevice err: 500
time=2025-02-03T16:40:03.887Z level=DEBUG source=gpu.go:592 msg="Unable to load cudart library /usr/lib/ollama/libcudart.so.11.3.109: cudart init failure: 500"
cudaSetDevice err: 500
time=2025-02-03T16:40:03.891Z level=DEBUG source=gpu.go:592 msg="Unable to load cudart library /usr/lib/ollama/libcudart.so.12.4.127: cudart init failure: 500"
time=2025-02-03T16:40:03.891Z level=DEBUG source=amd_linux.go:421 msg="amdgpu driver not detected /sys/module/amdgpu"
time=2025-02-03T16:40:03.891Z level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered"
time=2025-02-03T16:40:03.891Z level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="15.6 GiB" available="14.1 GiB"

In local computer works with SDK NVIDIA, nvcc --version cuda_12.4.r12.4.

NVIDIA RTX 4070

OS

Windows, Docker

GPU

Nvidia

CPU

Intel

Ollama version

0.5.7 and 0.5.6

Originally created by @nonodev96 on GitHub (Feb 3, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/8791 ### What is the issue? ```ansi 2025/02/03 16:40:03 routes.go:1187: INFO server config env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_DEBUG:true OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-02-03T16:40:03.807Z level=INFO source=images.go:432 msg="total blobs: 5" time=2025-02-03T16:40:03.807Z level=INFO source=images.go:439 msg="total unused blobs removed: 0" time=2025-02-03T16:40:03.807Z level=INFO source=routes.go:1238 msg="Listening on [::]:11434 (version 0.5.6-0-g2539f2d-dirty)" time=2025-02-03T16:40:03.807Z level=DEBUG source=common.go:80 msg="runners located" dir=/usr/lib/ollama/runners time=2025-02-03T16:40:03.807Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx/ollama_llama_server time=2025-02-03T16:40:03.808Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cpu_avx2/ollama_llama_server time=2025-02-03T16:40:03.808Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cuda_v11_avx/ollama_llama_server time=2025-02-03T16:40:03.808Z level=DEBUG source=common.go:124 msg="availableServers : found" file=/usr/lib/ollama/runners/cuda_v12_avx/ollama_llama_server time=2025-02-03T16:40:03.808Z level=INFO source=routes.go:1267 msg="Dynamic LLM libraries" runners="[cuda_v11_avx cuda_v12_avx cpu cpu_avx cpu_avx2]" time=2025-02-03T16:40:03.808Z level=DEBUG source=routes.go:1268 msg="Override detection logic by setting OLLAMA_LLM_LIBRARY" time=2025-02-03T16:40:03.808Z level=DEBUG source=sched.go:105 msg="starting llm scheduler" time=2025-02-03T16:40:03.808Z level=INFO source=gpu.go:226 msg="looking for compatible GPUs" time=2025-02-03T16:40:03.808Z level=DEBUG source=gpu.go:99 msg="searching for GPU discovery libraries for NVIDIA" time=2025-02-03T16:40:03.808Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcuda.so* time=2025-02-03T16:40:03.808Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcuda.so* /usr/lib/ollama/libcuda.so* /usr/local/nvidia/lib/libcuda.so* /usr/local/nvidia/lib64/libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]" time=2025-02-03T16:40:03.809Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths="[/usr/lib/x86_64-linux-gnu/libcuda.so.1 /usr/lib/wsl/drivers/nv_dispi.inf_amd64_ab3196e1830c9b6c/libcuda.so.1.1]" initializing /usr/lib/x86_64-linux-gnu/libcuda.so.1 dlsym: cuInit - 0x7f500c220ec0 dlsym: cuDriverGetVersion - 0x7f500c220e80 dlsym: cuDeviceGetCount - 0x7f500c220efd dlsym: cuDeviceGet - 0x7f500c220ef7 dlsym: cuDeviceGetAttribute - 0x7f500c220de0 dlsym: cuDeviceGetUuid - 0x7f500c220f09 dlsym: cuDeviceGetName - 0x7f500c220f03 [GIN-debug] [WARNING] Creating an Engine instance with the Logger and Recovery middleware already attached. [GIN-debug] [WARNING] Running in "debug" mode. Switch to "release" mode in production. - using env: export GIN_MODE=release - using code: gin.SetMode(gin.ReleaseMode) [GIN-debug] POST /api/pull --> github.com/ollama/ollama/server.(*Server).PullHandler-fm (5 handlers) [GIN-debug] POST /api/generate --> github.com/ollama/ollama/server.(*Server).GenerateHandler-fm (5 handlers) [GIN-debug] POST /api/chat --> github.com/ollama/ollama/server.(*Server).ChatHandler-fm (5 handlers) [GIN-debug] POST /api/embed --> github.com/ollama/ollama/server.(*Server).EmbedHandler-fm (5 handlers) [GIN-debug] POST /api/embeddings --> github.com/ollama/ollama/server.(*Server).EmbeddingsHandler-fm (5 handlers) [GIN-debug] POST /api/create --> github.com/ollama/ollama/server.(*Server).CreateHandler-fm (5 handlers) [GIN-debug] POST /api/push --> github.com/ollama/ollama/server.(*Server).PushHandler-fm (5 handlers) [GIN-debug] POST /api/copy --> github.com/ollama/ollama/server.(*Server).CopyHandler-fm (5 handlers) [GIN-debug] DELETE /api/delete --> github.com/ollama/ollama/server.(*Server).DeleteHandler-fm (5 handlers) [GIN-debug] POST /api/show --> github.com/ollama/ollama/server.(*Server).ShowHandler-fm (5 handlers) [GIN-debug] POST /api/blobs/:digest --> github.com/ollama/ollama/server.(*Server).CreateBlobHandler-fm (5 handlers) [GIN-debug] HEAD /api/blobs/:digest --> github.com/ollama/ollama/server.(*Server).HeadBlobHandler-fm (5 handlers) [GIN-debug] GET /api/ps --> github.com/ollama/ollama/server.(*Server).PsHandler-fm (5 handlers) [GIN-debug] POST /v1/chat/completions --> github.com/ollama/ollama/server.(*Server).ChatHandler-fm (6 handlers) [GIN-debug] POST /v1/completions --> github.com/ollama/ollama/server.(*Server).GenerateHandler-fm (6 handlers) [GIN-debug] POST /v1/embeddings --> github.com/ollama/ollama/server.(*Server).EmbedHandler-fm (6 handlers) [GIN-debug] GET /v1/models --> github.com/ollama/ollama/server.(*Server).ListHandler-fm (6 handlers) [GIN-debug] GET /v1/models/:model --> github.com/ollama/ollama/server.(*Server).ShowHandler-fm (6 handlers) [GIN-debug] GET / --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func1 (5 handlers) [GIN-debug] GET /api/tags --> github.com/ollama/ollama/server.(*Server).ListHandler-fm (5 handlers) [GIN-debug] GET /api/version --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func2 (5 handlers) [GIN-debug] HEAD / --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func1 (5 handlers) [GIN-debug] HEAD /api/tags --> github.com/ollama/ollama/server.(*Server).ListHandler-fm (5 handlers) [GIN-debug] HEAD /api/version --> github.com/ollama/ollama/server.(*Server).GenerateRoutes.func2 (5 handlers) dlsym: cuCtxCreate_v3 - 0x7f500c220f7b dlsym: cuMemGetInfo_v2 - 0x7f500c22107d dlsym: cuCtxDestroy - 0x7f500c220f8d calling cuInit cuInit err: 500 time=2025-02-03T16:40:03.884Z level=INFO source=gpu.go:630 msg="Unable to load cudart library /usr/lib/x86_64-linux-gnu/libcuda.so.1: cuda driver library init failure: 500" initializing /usr/lib/wsl/drivers/nv_dispi.inf_amd64_ab3196e1830c9b6c/libcuda.so.1.1 dlsym: cuInit - 0x7f4fe640e220 dlsym: cuDriverGetVersion - 0x7f4fe640e1f0 dlsym: cuDeviceGetCount - 0x7f4fe640e190 dlsym: cuDeviceGet - 0x7f4fe640e1c0 dlsym: cuDeviceGetAttribute - 0x7f4fe640e040 dlsym: cuDeviceGetUuid - 0x7f4fe640e130 dlsym: cuDeviceGetName - 0x7f4fe640e160 dlsym: cuCtxCreate_v3 - 0x7f4fe640dd70 dlsym: cuMemGetInfo_v2 - 0x7f4fe640d230 dlsym: cuCtxDestroy - 0x7f4fe64087c0 calling cuInit cuInit err: 500 time=2025-02-03T16:40:03.885Z level=INFO source=gpu.go:630 msg="Unable to load cudart library /usr/lib/wsl/drivers/nv_dispi.inf_amd64_ab3196e1830c9b6c/libcuda.so.1.1: cuda driver library init failure: 500" time=2025-02-03T16:40:03.885Z level=DEBUG source=gpu.go:517 msg="Searching for GPU library" name=libcudart.so* time=2025-02-03T16:40:03.885Z level=DEBUG source=gpu.go:543 msg="gpu library search" globs="[/usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/local/nvidia/lib/libcudart.so* /usr/local/nvidia/lib64/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/lib/ollama/libcudart.so* /usr/local/cuda/lib64/libcudart.so* /usr/lib/x86_64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/x86_64-linux-gnu/libcudart.so* /usr/lib/wsl/lib/libcudart.so* /usr/lib/wsl/drivers/*/libcudart.so* /opt/cuda/lib64/libcudart.so* /usr/local/cuda*/targets/aarch64-linux/lib/libcudart.so* /usr/lib/aarch64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/aarch64-linux-gnu/libcudart.so* /usr/local/cuda/lib*/libcudart.so* /usr/lib*/libcudart.so* /usr/local/lib*/libcudart.so*]" time=2025-02-03T16:40:03.885Z level=DEBUG source=gpu.go:576 msg="discovered GPU libraries" paths="[/usr/lib/ollama/libcudart.so.11.3.109 /usr/lib/ollama/libcudart.so.12.4.127]" cudaSetDevice err: 500 time=2025-02-03T16:40:03.887Z level=DEBUG source=gpu.go:592 msg="Unable to load cudart library /usr/lib/ollama/libcudart.so.11.3.109: cudart init failure: 500" cudaSetDevice err: 500 time=2025-02-03T16:40:03.891Z level=DEBUG source=gpu.go:592 msg="Unable to load cudart library /usr/lib/ollama/libcudart.so.12.4.127: cudart init failure: 500" time=2025-02-03T16:40:03.891Z level=DEBUG source=amd_linux.go:421 msg="amdgpu driver not detected /sys/module/amdgpu" time=2025-02-03T16:40:03.891Z level=INFO source=gpu.go:392 msg="no compatible GPUs were discovered" time=2025-02-03T16:40:03.891Z level=INFO source=types.go:131 msg="inference compute" id=0 library=cpu variant=avx2 compute="" driver=0.0 name="" total="15.6 GiB" available="14.1 GiB" ``` In local computer works with SDK NVIDIA, `nvcc --version` `cuda_12.4.r12.4`. NVIDIA RTX 4070 ### OS Windows, Docker ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.7 and 0.5.6
GiteaMirror added the bugwslwindowsdocker labels 2026-05-04 11:37:54 -05:00
Author
Owner

@rick-github commented on GitHub (Feb 3, 2025):

error 500 is CUDA_ERROR_NOT_FOUND:
CUDA_ERROR_NOT_FOUND = 500

  • This indicates that a named symbol was not found. Examples of symbols are global/constant variable names, driver function names, texture names, and surface names.

The code is calling cuInit which has entry point (dlsym: cuInit - 0x7f500c220ec0) so presumably something else later in the call stack is missing. Do you have an up to date version of the CUDA toolkit?

<!-- gh-comment-id:2631642751 --> @rick-github commented on GitHub (Feb 3, 2025): error 500 is [`CUDA_ERROR_NOT_FOUND`](https://docs.nvidia.com/cuda/cuda-driver-api/group__CUDA__TYPES.html#group__CUDA__TYPES_1ggc6c391505e117393cc2558fff6bfc2e990696c86fcee1f536a1ec7d25867feeb:~:text=CUDA_ERROR_NOT_FOUND%20%3D%20500,and%20surface%20names.): CUDA_ERROR_NOT_FOUND = 500 - This indicates that a named symbol was not found. Examples of symbols are global/constant variable names, driver function names, texture names, and surface names. The code is calling `cuInit` which has entry point (`dlsym: cuInit - 0x7f500c220ec0`) so presumably something else later in the call stack is missing. Do you have an up to date version of the [CUDA toolkit](https://github.com/ollama/ollama/blob/main/docs/docker.md#nvidia-gpu)?
Author
Owner

@nonodev96 commented on GitHub (Feb 4, 2025):

HI @rick-github

I have version cuda_12.4.r12.4 and RTX 4070.
I had to downgrade from version 12.6 previously, it has been weeks ago, and I have had no problems with other projects.

I upgraded to the latest version and I get the same error.

<!-- gh-comment-id:2635266160 --> @nonodev96 commented on GitHub (Feb 4, 2025): HI @rick-github I have version cuda_12.4.r12.4 and RTX 4070. I had to downgrade from version 12.6 previously, it has been weeks ago, and I have had no problems with other projects. I upgraded to the latest version and I get the same error.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#67765