[GH-ISSUE #5652] A path to GPU support for Ollama in a VM/container on Apple Silicon #50037

Open
opened 2026-04-28 13:55:06 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @easp on GitHub (Jul 12, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/5652

I came across a short discussion in the llama.cpp repo about using a GPU on Apple Silicon from within a vm/container. Performance isn't as good as bare metal, but it's a significant improvement over CPU-only inference.

The meat of it is this work from March which provides a virtualized Vulkan interface inside the VM which is then passed-through to a vulkan-to-metal layer on the host MacOS.

For this to be relevant to Ollama I think it would require this proof of concept work with podman to be implemented in Docker Desktop for Mac, but I'm noting it here because it shows a way forward.

Originally created by @easp on GitHub (Jul 12, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/5652 I came across a short discussion in the llama.cpp repo about using a [GPU on Apple Silicon from within a vm/container](https://github.com/ggerganov/llama.cpp/discussions/8042). Performance isn't as good as bare metal, but it's a significant improvement over CPU-only inference. The meat of it is [this work](https://sinrega.org/2024-03-06-enabling-containers-gpu-macos/) from March which provides a virtualized Vulkan interface inside the VM which is then passed-through to a vulkan-to-metal layer on the host MacOS. For this to be relevant to Ollama I think it would require this proof of concept work with podman to be implemented in Docker Desktop for Mac, but I'm noting it here because it shows a way forward.
GiteaMirror added the feature request label 2026-04-28 13:55:06 -05:00
Author
Owner

@hybra commented on GitHub (Jul 15, 2024):

Have a look:
https://github.com/ggerganov/llama.cpp/discussions/8042#discussioncomment-10044780

<!-- gh-comment-id:2229190932 --> @hybra commented on GitHub (Jul 15, 2024): Have a look: https://github.com/ggerganov/llama.cpp/discussions/8042#discussioncomment-10044780
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#50037