[GH-ISSUE #4373] Feature request: support for OpenCL #64767

Open
opened 2026-05-03 18:45:04 -05:00 by GiteaMirror · 7 comments
Owner

Originally created by @alnoses on GitHub (May 12, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/4373

Allow the possibility of running the models on those GPUs that do not support, have stopped supporting or have never had CUDA or ROCm support such as Intel (dedicated and integrated) or integrated AMD GPUs or old models such as Polaris, the offer Qualcomm ARM...

Originally created by @alnoses on GitHub (May 12, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/4373 Allow the possibility of running the models on those GPUs that do not support, have stopped supporting or have never had CUDA or ROCm support such as Intel (dedicated and integrated) or integrated AMD GPUs or old models such as Polaris, the offer Qualcomm ARM...
GiteaMirror added the feature request label 2026-05-03 18:45:04 -05:00
Author
Owner

@vt-alt commented on GitHub (Jul 7, 2024):

I wanted to suggest building with CLBlast support but it seems CLBlast support is recently removed from llama.cpp as "Superseded by Vulkan" https://github.com/ggerganov/llama.cpp/pull/7735

<!-- gh-comment-id:2212615845 --> @vt-alt commented on GitHub (Jul 7, 2024): I wanted to suggest building with CLBlast support but it seems CLBlast support is recently removed from `llama.cpp` as "Superseded by Vulkan" https://github.com/ggerganov/llama.cpp/pull/7735
Author
Owner

@serialscriptr commented on GitHub (Sep 8, 2024):

OpenCL support would provide better performance on SBCs such as the Orange Pi 5 or any other device that uses the rk3588. The Mali drivers have very capable opencl support but no vulken drivers.

<!-- gh-comment-id:2336861440 --> @serialscriptr commented on GitHub (Sep 8, 2024): OpenCL support would provide better performance on SBCs such as the Orange Pi 5 or any other device that uses the rk3588. The Mali drivers have very capable opencl support but no vulken drivers.
Author
Owner

@vpereira commented on GitHub (Feb 3, 2025):

with openCL we would be able to run small models (< 4 GB) straight from the GPUs.. it's interesting for instance for those running deepseek-r1:1.5b as local code completion tool..

<!-- gh-comment-id:2630685339 --> @vpereira commented on GitHub (Feb 3, 2025): with openCL we would be able to run small models (< 4 GB) straight from the GPUs.. it's interesting for instance for those running `deepseek-r1:1.5b` as local code completion tool..
Author
Owner
<!-- gh-comment-id:2630733675 --> @pjaaskel commented on GitHub (Feb 3, 2025): https://www.qualcomm.com/developer/blog/2024/11/introducing-new-opn-cl-gpu-backend-llama-cpp-for-qualcomm-adreno-gpu
Author
Owner

@fish4terrisa-MSDSM commented on GitHub (Apr 7, 2025):

Any updates? Seems that OpenCL support is brought back to llama.cpp and is no longer tied with clblast now.

<!-- gh-comment-id:2782004347 --> @fish4terrisa-MSDSM commented on GitHub (Apr 7, 2025): Any updates? Seems that OpenCL support is brought back to llama.cpp and is no longer tied with clblast now.
Author
Owner

@dan-and commented on GitHub (May 6, 2025):

Also PoCL (https://github.com/pocl/pocl/blob/release_7_0/doc/sphinx/source/notes_7_0.rst ) would allow a single interface to a heterogenic set of backend devices

<!-- gh-comment-id:2854424333 --> @dan-and commented on GitHub (May 6, 2025): Also PoCL (https://github.com/pocl/pocl/blob/release_7_0/doc/sphinx/source/notes_7_0.rst ) would allow a single interface to a heterogenic set of backend devices
Author
Owner

@vpereira commented on GitHub (May 6, 2025):

Also PoCL (https://github.com/pocl/pocl/blob/release_7_0/doc/sphinx/source/notes_7_0.rst ) would allow a single interface to a heterogenic set of backend devices

it however doesn't support AMD GPUs, AFAICS...

<!-- gh-comment-id:2854862371 --> @vpereira commented on GitHub (May 6, 2025): > Also PoCL (https://github.com/pocl/pocl/blob/release_7_0/doc/sphinx/source/notes_7_0.rst ) would allow a single interface to a heterogenic set of backend devices it however doesn't support AMD GPUs, AFAICS...
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#64767