[GH-ISSUE #10781] AMD GPU not being detected in docker ROCm windows 11 RX 7800 XT #53594

Open
opened 2026-04-29 04:03:40 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @UpcraftLP on GitHub (May 20, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/10781

What is the issue?

using the docker rocm images (both 0.7.0 and 0.6.8), ollama fails to detect my AMD GPU.


OS: Windows 11 (24H2, build 26100.4061)
Docker version: 28.1.1, backend WSL2
GPU: AMD Radeon RX 7800 XT
Driver version: Radeon PRO, version 24.20.52
ROCm version: 6.2


Note:

the AMD documentation seems outdated, I had to add --privileged flag to the docker command or it would refuse to accept /dev/kfd.


docker command used:

docker run --privileged -d --device /dev/kfd --device /dev/dri -v ollama:/root/.ollama -p 11434:11434 --name ollama -e OLLAMA_DEBUG=1 ollama/ollama:0.7.0-rocm


hipinfo output:

device#                           0
Name:                             AMD  Radeon RX 7800 XT
pciBusID:                         3
pciDeviceID:                      0
pciDomainID:                      0
multiProcessorCount:              30
maxThreadsPerMultiProcessor:      2048
isMultiGpuBoard:                  0
clockRate:                        2254 Mhz
memoryClockRate:                  1219 Mhz
memoryBusWidth:                   0
totalGlobalMem:                   15.98 GB
totalConstMem:                    2147483647
sharedMemPerBlock:                64.00 KB
canMapHostMemory:                 1
regsPerBlock:                     0
warpSize:                         32
l2CacheSize:                      4194304
computeMode:                      0
maxThreadsPerBlock:               1024
maxThreadsDim.x:                  1024
maxThreadsDim.y:                  1024
maxThreadsDim.z:                  1024
maxGridSize.x:                    2147483647
maxGridSize.y:                    65536
maxGridSize.z:                    65536
major:                            11
minor:                            0
concurrentKernels:                1
cooperativeLaunch:                0
cooperativeMultiDeviceLaunch:     0
isIntegrated:                     0
maxTexture1D:                     16384
maxTexture2D.width:               16384
maxTexture2D.height:              16384
maxTexture3D.width:               2048
maxTexture3D.height:              2048
maxTexture3D.depth:               2048
hostNativeAtomicSupported:        1
isLargeBar:                       0
asicRevision:                     0
maxSharedMemoryPerMultiProcessor: 64.00 KB
clockInstructionRate:             1000.00 Mhz
arch.hasGlobalInt32Atomics:       1
arch.hasGlobalFloatAtomicExch:    1
arch.hasSharedInt32Atomics:       1
arch.hasSharedFloatAtomicExch:    1
arch.hasFloatAtomicAdd:           1
arch.hasGlobalInt64Atomics:       1
arch.hasSharedInt64Atomics:       1
arch.hasDoubles:                  1
arch.hasWarpVote:                 1
arch.hasWarpBallot:               1
arch.hasWarpShuffle:              1
arch.hasFunnelShift:              0
arch.hasThreadFenceSystem:        1
arch.hasSyncThreadsExt:           0
arch.hasSurfaceFuncs:             0
arch.has3dGrid:                   1
arch.hasDynamicParallelism:       0
gcnArchName:                      gfx1101
peers:
non-peers:                        device#0

memInfo.total:                    15.98 GB
memInfo.free:                     15.84 GB (99%)

Relevant log output

time=2025-05-20T18:28:25.260Z level=INFO source=routes.go:1205 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost⁠ http://localhost:*⁠ https://localhost:*⁠ http://127.0.0.1⁠ https://127.0.0.1⁠ http://127.0.0.1:*⁠ https://127.0.0.1:*⁠ http://0.0.0.0⁠ https://0.0.0.0⁠ http://0.0.0.0:*⁠ https://0.0.0.0:*⁠ app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"

time=2025-05-20T18:28:25.260Z level=INFO source=images.go:463 msg="total blobs: 5"

time=2025-05-20T18:28:25.260Z level=INFO source=images.go:470 msg="total unused blobs removed: 0"

time=2025-05-20T18:28:25.260Z level=INFO source=routes.go:1258 msg="Listening on [::]:11434 (version 0.7.0)"

time=2025-05-20T18:28:25.261Z level=DEBUG source=sched.go:108 msg="starting llm scheduler"

time=2025-05-20T18:28:25.261Z level=INFO source=gpu.go:217 msg="looking for compatible GPUs"

time=2025-05-20T18:28:25.261Z level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA"

time=2025-05-20T18:28:25.261Z level=DEBUG source=gpu.go:501 msg="Searching for GPU library" name=libcuda.so*

time=2025-05-20T18:28:25.261Z level=DEBUG source=gpu.go:525 msg="gpu library search" globs="[/usr/lib/ollama/libcuda.so* /usr/local/nvidia/lib/libcuda.so* /usr/local/nvidia/lib64/libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]"

time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:558 msg="discovered GPU libraries" paths=[]

time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:501 msg="Searching for GPU library" name=libcudart.so*

time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:525 msg="gpu library search" globs="[/usr/lib/ollama/libcudart.so* /usr/local/nvidia/lib/libcudart.so* /usr/local/nvidia/lib64/libcudart.so* /usr/lib/ollama/cuda_v*/libcudart.so* /usr/local/cuda/lib64/libcudart.so* /usr/lib/x86_64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/x86_64-linux-gnu/libcudart.so* /usr/lib/wsl/lib/libcudart.so* /usr/lib/wsl/drivers/*/libcudart.so* /opt/cuda/lib64/libcudart.so* /usr/local/cuda*/targets/aarch64-linux/lib/libcudart.so* /usr/lib/aarch64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/aarch64-linux-gnu/libcudart.so* /usr/local/cuda/lib*/libcudart.so* /usr/lib*/libcudart.so* /usr/local/lib*/libcudart.so*]"

time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:558 msg="discovered GPU libraries" paths=[]

time=2025-05-20T18:28:25.262Z level=DEBUG source=amd_linux.go:419 msg="amdgpu driver not detected /sys/module/amdgpu"

time=2025-05-20T18:28:25.262Z level=INFO source=gpu.go:377 msg="no compatible GPUs were discovered"

time=2025-05-20T18:28:25.262Z level=INFO source=types.go:130 msg="inference compute" id=0 library=cpu variant="" compute="" driver=0.0 name="" total="31.2 GiB" available="29.8 GiB"

OS

Windows, Docker

GPU

AMD

CPU

AMD

Ollama version

0.7.0

Originally created by @UpcraftLP on GitHub (May 20, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/10781 ### What is the issue? using the docker `rocm` images (both 0.7.0 and 0.6.8), ollama fails to detect my AMD GPU. --- OS: Windows 11 (24H2, build 26100.4061) Docker version: 28.1.1, backend WSL2 GPU: AMD Radeon RX 7800 XT Driver version: Radeon PRO, version 24.20.52 ROCm version: 6.2 --- ### Note: the AMD documentation seems outdated, I had to add `--privileged` flag to the docker command or it would refuse to accept `/dev/kfd`. --- ### docker command used: `docker run --privileged -d --device /dev/kfd --device /dev/dri -v ollama:/root/.ollama -p 11434:11434 --name ollama -e OLLAMA_DEBUG=1 ollama/ollama:0.7.0-rocm` --- ### `hipinfo` output: ``` device# 0 Name: AMD Radeon RX 7800 XT pciBusID: 3 pciDeviceID: 0 pciDomainID: 0 multiProcessorCount: 30 maxThreadsPerMultiProcessor: 2048 isMultiGpuBoard: 0 clockRate: 2254 Mhz memoryClockRate: 1219 Mhz memoryBusWidth: 0 totalGlobalMem: 15.98 GB totalConstMem: 2147483647 sharedMemPerBlock: 64.00 KB canMapHostMemory: 1 regsPerBlock: 0 warpSize: 32 l2CacheSize: 4194304 computeMode: 0 maxThreadsPerBlock: 1024 maxThreadsDim.x: 1024 maxThreadsDim.y: 1024 maxThreadsDim.z: 1024 maxGridSize.x: 2147483647 maxGridSize.y: 65536 maxGridSize.z: 65536 major: 11 minor: 0 concurrentKernels: 1 cooperativeLaunch: 0 cooperativeMultiDeviceLaunch: 0 isIntegrated: 0 maxTexture1D: 16384 maxTexture2D.width: 16384 maxTexture2D.height: 16384 maxTexture3D.width: 2048 maxTexture3D.height: 2048 maxTexture3D.depth: 2048 hostNativeAtomicSupported: 1 isLargeBar: 0 asicRevision: 0 maxSharedMemoryPerMultiProcessor: 64.00 KB clockInstructionRate: 1000.00 Mhz arch.hasGlobalInt32Atomics: 1 arch.hasGlobalFloatAtomicExch: 1 arch.hasSharedInt32Atomics: 1 arch.hasSharedFloatAtomicExch: 1 arch.hasFloatAtomicAdd: 1 arch.hasGlobalInt64Atomics: 1 arch.hasSharedInt64Atomics: 1 arch.hasDoubles: 1 arch.hasWarpVote: 1 arch.hasWarpBallot: 1 arch.hasWarpShuffle: 1 arch.hasFunnelShift: 0 arch.hasThreadFenceSystem: 1 arch.hasSyncThreadsExt: 0 arch.hasSurfaceFuncs: 0 arch.has3dGrid: 1 arch.hasDynamicParallelism: 0 gcnArchName: gfx1101 peers: non-peers: device#0 memInfo.total: 15.98 GB memInfo.free: 15.84 GB (99%) ``` --- ### Relevant log output ```shell time=2025-05-20T18:28:25.260Z level=INFO source=routes.go:1205 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:0 OLLAMA_ORIGINS:[http://localhost https://localhost⁠ http://localhost:*⁠ https://localhost:*⁠ http://127.0.0.1⁠ https://127.0.0.1⁠ http://127.0.0.1:*⁠ https://127.0.0.1:*⁠ http://0.0.0.0⁠ https://0.0.0.0⁠ http://0.0.0.0:*⁠ https://0.0.0.0:*⁠ app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-05-20T18:28:25.260Z level=INFO source=images.go:463 msg="total blobs: 5" time=2025-05-20T18:28:25.260Z level=INFO source=images.go:470 msg="total unused blobs removed: 0" time=2025-05-20T18:28:25.260Z level=INFO source=routes.go:1258 msg="Listening on [::]:11434 (version 0.7.0)" time=2025-05-20T18:28:25.261Z level=DEBUG source=sched.go:108 msg="starting llm scheduler" time=2025-05-20T18:28:25.261Z level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-05-20T18:28:25.261Z level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-05-20T18:28:25.261Z level=DEBUG source=gpu.go:501 msg="Searching for GPU library" name=libcuda.so* time=2025-05-20T18:28:25.261Z level=DEBUG source=gpu.go:525 msg="gpu library search" globs="[/usr/lib/ollama/libcuda.so* /usr/local/nvidia/lib/libcuda.so* /usr/local/nvidia/lib64/libcuda.so* /usr/local/cuda*/targets/*/lib/libcuda.so* /usr/lib/*-linux-gnu/nvidia/current/libcuda.so* /usr/lib/*-linux-gnu/libcuda.so* /usr/lib/wsl/lib/libcuda.so* /usr/lib/wsl/drivers/*/libcuda.so* /opt/cuda/lib*/libcuda.so* /usr/local/cuda/lib*/libcuda.so* /usr/lib*/libcuda.so* /usr/local/lib*/libcuda.so*]" time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:558 msg="discovered GPU libraries" paths=[] time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:501 msg="Searching for GPU library" name=libcudart.so* time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:525 msg="gpu library search" globs="[/usr/lib/ollama/libcudart.so* /usr/local/nvidia/lib/libcudart.so* /usr/local/nvidia/lib64/libcudart.so* /usr/lib/ollama/cuda_v*/libcudart.so* /usr/local/cuda/lib64/libcudart.so* /usr/lib/x86_64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/x86_64-linux-gnu/libcudart.so* /usr/lib/wsl/lib/libcudart.so* /usr/lib/wsl/drivers/*/libcudart.so* /opt/cuda/lib64/libcudart.so* /usr/local/cuda*/targets/aarch64-linux/lib/libcudart.so* /usr/lib/aarch64-linux-gnu/nvidia/current/libcudart.so* /usr/lib/aarch64-linux-gnu/libcudart.so* /usr/local/cuda/lib*/libcudart.so* /usr/lib*/libcudart.so* /usr/local/lib*/libcudart.so*]" time=2025-05-20T18:28:25.262Z level=DEBUG source=gpu.go:558 msg="discovered GPU libraries" paths=[] time=2025-05-20T18:28:25.262Z level=DEBUG source=amd_linux.go:419 msg="amdgpu driver not detected /sys/module/amdgpu" time=2025-05-20T18:28:25.262Z level=INFO source=gpu.go:377 msg="no compatible GPUs were discovered" time=2025-05-20T18:28:25.262Z level=INFO source=types.go:130 msg="inference compute" id=0 library=cpu variant="" compute="" driver=0.0 name="" total="31.2 GiB" available="29.8 GiB" ``` ### OS Windows, Docker ### GPU AMD ### CPU AMD ### Ollama version 0.7.0
GiteaMirror added the bug label 2026-04-29 04:03:40 -05:00
Author
Owner

@rick-github commented on GitHub (May 20, 2025):

Docker documentation only mentions support for Nvidia GPUs on Windows.

<!-- gh-comment-id:2895476136 --> @rick-github commented on GitHub (May 20, 2025): [Docker documentation](https://docs.docker.com/desktop/features/gpu/) only mentions support for Nvidia GPUs on Windows.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#53594