Does not support Intel integrated graphics, please add support #8604

Open
opened 2025-11-12 14:46:58 -06:00 by GiteaMirror · 2 comments
Owner

Originally created by @fly23423 on GitHub (Nov 5, 2025).

What is the issue?

Does not support Intel integrated graphics, please add support

Relevant log output

time=2025-11-05T10:38:43.973Z level=INFO source=server.go:653 msg="loading model" "model layers"=29 requested=-1
time=2025-11-05T10:38:43.973Z level=INFO source=server.go:658 msg="system memory" total="15.4 GiB" free="10.6 GiB" free_swap="28.1 GiB"
time=2025-11-05T10:38:43.974Z level=INFO source=runner.go:1222 msg=load request="{Operation:fit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2025-11-05T10:38:44.018Z level=INFO source=ggml.go:136 msg="" architecture=qwen3 file_type=Q4_K_M name="Qwen3 1.7B" description="" num_tensors=311 num_key_values=28
load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-sse42.so
time=2025-11-05T10:38:44.030Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc)
time=2025-11-05T10:38:44.056Z level=INFO source=runner.go:1222 msg=load request="{Operation:alloc LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2025-11-05T10:38:44.341Z level=INFO source=runner.go:1222 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:482 msg="offloading 0 repeating layers to GPU"
time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:486 msg="offloading output layer to CPU"
time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:494 msg="offloaded 0/29 layers to GPU"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:217 msg="model weights" device=CPU size="1.3 GiB"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:228 msg="kv cache" device=CPU size="448.0 MiB"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:239 msg="compute graph" device=CPU size="52.0 MiB"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:244 msg="total memory" size="1.7 GiB"

OS

Docker

GPU

Intel

CPU

Intel

Ollama version

0.12.9

Originally created by @fly23423 on GitHub (Nov 5, 2025). ### What is the issue? Does not support Intel integrated graphics, please add support ### Relevant log output ```shell time=2025-11-05T10:38:43.973Z level=INFO source=server.go:653 msg="loading model" "model layers"=29 requested=-1 time=2025-11-05T10:38:43.973Z level=INFO source=server.go:658 msg="system memory" total="15.4 GiB" free="10.6 GiB" free_swap="28.1 GiB" time=2025-11-05T10:38:43.974Z level=INFO source=runner.go:1222 msg=load request="{Operation:fit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-11-05T10:38:44.018Z level=INFO source=ggml.go:136 msg="" architecture=qwen3 file_type=Q4_K_M name="Qwen3 1.7B" description="" num_tensors=311 num_key_values=28 load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-sse42.so time=2025-11-05T10:38:44.030Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-11-05T10:38:44.056Z level=INFO source=runner.go:1222 msg=load request="{Operation:alloc LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-11-05T10:38:44.341Z level=INFO source=runner.go:1222 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:482 msg="offloading 0 repeating layers to GPU" time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:486 msg="offloading output layer to CPU" time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:494 msg="offloaded 0/29 layers to GPU" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:217 msg="model weights" device=CPU size="1.3 GiB" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:228 msg="kv cache" device=CPU size="448.0 MiB" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:239 msg="compute graph" device=CPU size="52.0 MiB" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:244 msg="total memory" size="1.7 GiB" ``` ### OS Docker ### GPU Intel ### CPU Intel ### Ollama version 0.12.9
GiteaMirror added the bug label 2025-11-12 14:46:58 -06:00
Author
Owner

@rick-github commented on GitHub (Nov 5, 2025):

Which Intel GPU do you have? It might be supported by the up-coming Vulkan backend.

@rick-github commented on GitHub (Nov 5, 2025): Which Intel GPU do you have? It might be supported by the up-coming Vulkan backend.
Author
Owner

@fly23423 commented on GitHub (Nov 5, 2025):

Intel n5105

@fly23423 commented on GitHub (Nov 5, 2025): Intel n5105
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama-ollama#8604