[GH-ISSUE #12967] Does not support Intel integrated graphics, please add support #34356

Open
opened 2026-04-22 17:49:46 -05:00 by GiteaMirror · 3 comments
Owner

Originally created by @fly23423 on GitHub (Nov 5, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/12967

What is the issue?

Does not support Intel integrated graphics, please add support

Relevant log output

time=2025-11-05T10:38:43.973Z level=INFO source=server.go:653 msg="loading model" "model layers"=29 requested=-1
time=2025-11-05T10:38:43.973Z level=INFO source=server.go:658 msg="system memory" total="15.4 GiB" free="10.6 GiB" free_swap="28.1 GiB"
time=2025-11-05T10:38:43.974Z level=INFO source=runner.go:1222 msg=load request="{Operation:fit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2025-11-05T10:38:44.018Z level=INFO source=ggml.go:136 msg="" architecture=qwen3 file_type=Q4_K_M name="Qwen3 1.7B" description="" num_tensors=311 num_key_values=28
load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-sse42.so
time=2025-11-05T10:38:44.030Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc)
time=2025-11-05T10:38:44.056Z level=INFO source=runner.go:1222 msg=load request="{Operation:alloc LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2025-11-05T10:38:44.341Z level=INFO source=runner.go:1222 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:482 msg="offloading 0 repeating layers to GPU"
time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:486 msg="offloading output layer to CPU"
time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:494 msg="offloaded 0/29 layers to GPU"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:217 msg="model weights" device=CPU size="1.3 GiB"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:228 msg="kv cache" device=CPU size="448.0 MiB"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:239 msg="compute graph" device=CPU size="52.0 MiB"
time=2025-11-05T10:38:44.341Z level=INFO source=device.go:244 msg="total memory" size="1.7 GiB"

OS

Docker

GPU

Intel

CPU

Intel

Ollama version

0.12.9

Originally created by @fly23423 on GitHub (Nov 5, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/12967 ### What is the issue? Does not support Intel integrated graphics, please add support ### Relevant log output ```shell time=2025-11-05T10:38:43.973Z level=INFO source=server.go:653 msg="loading model" "model layers"=29 requested=-1 time=2025-11-05T10:38:43.973Z level=INFO source=server.go:658 msg="system memory" total="15.4 GiB" free="10.6 GiB" free_swap="28.1 GiB" time=2025-11-05T10:38:43.974Z level=INFO source=runner.go:1222 msg=load request="{Operation:fit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-11-05T10:38:44.018Z level=INFO source=ggml.go:136 msg="" architecture=qwen3 file_type=Q4_K_M name="Qwen3 1.7B" description="" num_tensors=311 num_key_values=28 load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-sse42.so time=2025-11-05T10:38:44.030Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-11-05T10:38:44.056Z level=INFO source=runner.go:1222 msg=load request="{Operation:alloc LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-11-05T10:38:44.341Z level=INFO source=runner.go:1222 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:true KvSize:4096 KvCacheType: NumThreads:4 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:482 msg="offloading 0 repeating layers to GPU" time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:486 msg="offloading output layer to CPU" time=2025-11-05T10:38:44.341Z level=INFO source=ggml.go:494 msg="offloaded 0/29 layers to GPU" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:217 msg="model weights" device=CPU size="1.3 GiB" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:228 msg="kv cache" device=CPU size="448.0 MiB" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:239 msg="compute graph" device=CPU size="52.0 MiB" time=2025-11-05T10:38:44.341Z level=INFO source=device.go:244 msg="total memory" size="1.7 GiB" ``` ### OS Docker ### GPU Intel ### CPU Intel ### Ollama version 0.12.9
GiteaMirror added the intelbug labels 2026-04-22 17:49:46 -05:00
Author
Owner

@rick-github commented on GitHub (Nov 5, 2025):

Which Intel GPU do you have? It might be supported by the up-coming Vulkan backend.

<!-- gh-comment-id:3491025777 --> @rick-github commented on GitHub (Nov 5, 2025): Which Intel GPU do you have? It might be supported by the up-coming Vulkan backend.
Author
Owner

@fly23423 commented on GitHub (Nov 5, 2025):

Intel n5105

<!-- gh-comment-id:3493623192 --> @fly23423 commented on GitHub (Nov 5, 2025): Intel n5105
Author
Owner

@dhiltgen commented on GitHub (Nov 14, 2025):

In 0.12.11 Vulkan is now included in the official binaries, but still experimental. To enable, set OLLAMA_VULKAN=1 for the server. https://github.com/ollama/ollama/blob/main/docs/faq.mdx#how-do-i-configure-ollama-server

<!-- gh-comment-id:3530328275 --> @dhiltgen commented on GitHub (Nov 14, 2025): In 0.12.11 Vulkan is now included in the official binaries, but still experimental. To enable, set OLLAMA_VULKAN=1 for the server. https://github.com/ollama/ollama/blob/main/docs/faq.mdx#how-do-i-configure-ollama-server
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#34356