[GH-ISSUE #13050] Ollama 0.12.10 fails to find CUDA compiler (fixed with work around) #70706

Closed
opened 2026-05-04 22:38:44 -05:00 by GiteaMirror · 7 comments
Owner

Originally created by @abcbarryn on GitHub (Nov 11, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/13050

What is the issue?

-- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF
-- CMAKE_SYSTEM_PROCESSOR: x86_64
-- GGML_SYSTEM_ARCH: x86
-- Including CPU backend
-- x86 detected
-- Adding CPU backend variant ggml-cpu-x64:
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX
-- x86 detected
-- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2
-- x86 detected
-- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512
-- x86 detected
-- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI
-- x86 detected
-- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI
-- Vulkan found
-- GL_KHR_cooperative_matrix supported by glslc
-- GL_NV_cooperative_matrix2 supported by glslc
-- GL_EXT_integer_dot_product supported by glslc
-- GL_EXT_bfloat16 supported by glslc
-- Configuring done (0.3s)
-- Generating done (0.1s)
-- Build files have been written to: /usr/src/ollama-0.12.10/build

Relevant log output

Partial output of build of 0.12.9 on the same system for comparison...
-- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF
-- CMAKE_SYSTEM_PROCESSOR: x86_64
-- GGML_SYSTEM_ARCH: x86
-- Including CPU backend
-- x86 detected
-- Adding CPU backend variant ggml-cpu-x64:
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX
-- x86 detected
-- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2
-- x86 detected
-- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512
-- x86 detected
-- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI
-- x86 detected
-- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI
-- CUDA Toolkit found
-- Using CUDA architectures: native
-- Vulkan found
-- GL_KHR_cooperative_matrix supported by glslc
-- GL_NV_cooperative_matrix2 supported by glslc
-- GL_EXT_integer_dot_product supported by glslc
-- GL_EXT_bfloat16 supported by glslc
-- Configuring done (0.4s)
-- Generating done (0.1s)
-- Build files have been written to: /usr/src/ollama-0.12.9/build

Note that the lines:
-- CUDA Toolkit found
-- Using CUDA architectures: native
are missing from the 0.12.10 build output.

OS

Linux

GPU

Nvidia

CPU

Intel

Ollama version

0.12.10

Originally created by @abcbarryn on GitHub (Nov 11, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/13050 ### What is the issue? -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- GGML_SYSTEM_ARCH: x86 -- Including CPU backend -- x86 detected -- Adding CPU backend variant ggml-cpu-x64: -- x86 detected -- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42 -- x86 detected -- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX -- x86 detected -- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2 -- x86 detected -- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512 -- x86 detected -- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI -- x86 detected -- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI -- Vulkan found -- GL_KHR_cooperative_matrix supported by glslc -- GL_NV_cooperative_matrix2 supported by glslc -- GL_EXT_integer_dot_product supported by glslc -- GL_EXT_bfloat16 supported by glslc -- Configuring done (0.3s) -- Generating done (0.1s) -- Build files have been written to: /usr/src/ollama-0.12.10/build ### Relevant log output ```shell Partial output of build of 0.12.9 on the same system for comparison... -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- GGML_SYSTEM_ARCH: x86 -- Including CPU backend -- x86 detected -- Adding CPU backend variant ggml-cpu-x64: -- x86 detected -- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42 -- x86 detected -- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX -- x86 detected -- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2 -- x86 detected -- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512 -- x86 detected -- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI -- x86 detected -- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI -- CUDA Toolkit found -- Using CUDA architectures: native -- Vulkan found -- GL_KHR_cooperative_matrix supported by glslc -- GL_NV_cooperative_matrix2 supported by glslc -- GL_EXT_integer_dot_product supported by glslc -- GL_EXT_bfloat16 supported by glslc -- Configuring done (0.4s) -- Generating done (0.1s) -- Build files have been written to: /usr/src/ollama-0.12.9/build Note that the lines: -- CUDA Toolkit found -- Using CUDA architectures: native are missing from the 0.12.10 build output. ``` ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.12.10
GiteaMirror added the buildbug labels 2026-05-04 22:38:45 -05:00
Author
Owner

@rick-github commented on GitHub (Nov 11, 2025):

$ mkdir 0.12.10 && cd 0.12.10 && git clone https://github.com/ollama/ollama -b v0.12.10 . && cmake -B build
Cloning into '.'...
remote: Enumerating objects: 41175, done.
remote: Counting objects: 100% (2/2), done.
remote: Compressing objects: 100% (2/2), done.
remote: Total 41175 (delta 0), reused 0 (delta 0), pack-reused 41173 (from 3)
Receiving objects: 100% (41175/41175), 65.12 MiB | 29.88 MiB/s, done.
Resolving deltas: 100% (25869/25869), done.
Note: switching to '80d34260ea16e76c9ef0d014a86cc130421855f1'.

You are in 'detached HEAD' state. You can look around, make experimental
changes and commit them, and you can discard any commits you make in this
state without impacting any branches by switching back to a branch.

If you want to create a new branch to retain commits you create, you may
do so (now or later) by using -c with the switch command. Example:

  git switch -c <new-branch-name>

Or undo this operation with:

  git switch -

Turn off this advice by setting config variable advice.detachedHead to false

-- The C compiler identification is GNU 13.2.0
-- The CXX compiler identification is GNU 13.2.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /usr/bin/cc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success
-- Found Threads: TRUE  
-- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF
-- CMAKE_SYSTEM_PROCESSOR: x86_64
-- GGML_SYSTEM_ARCH: x86
-- Including CPU backend
-- x86 detected
-- Adding CPU backend variant ggml-cpu-x64:  
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX
-- x86 detected
-- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2
-- x86 detected
-- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512
-- x86 detected
-- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI
-- x86 detected
-- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI
-- Looking for a CUDA compiler
-- Looking for a CUDA compiler - /usr/bin/nvcc
-- Found CUDAToolkit: /usr/include (found version "12.0.140") 
-- CUDA Toolkit found
-- Using CUDA architectures: native
-- The CUDA compiler identification is NVIDIA 12.0.140
-- Detecting CUDA compiler ABI info
-- Detecting CUDA compiler ABI info - done
-- Check for working CUDA compiler: /usr/bin/nvcc - skipped
-- Detecting CUDA compile features
-- Detecting CUDA compile features - done
-- Looking for a HIP compiler
-- Looking for a HIP compiler - NOTFOUND
-- Found Vulkan: /home/rick/ollama/1.4.328.1/x86_64/lib/libvulkan.so (found version "1.4.328") found components: glslc glslangValidator 
-- Vulkan found
-- GL_KHR_cooperative_matrix supported by glslc
-- GL_NV_cooperative_matrix2 supported by glslc
-- GL_EXT_integer_dot_product supported by glslc
-- GL_EXT_bfloat16 supported by glslc
-- Configuring done (6.9s)
-- Generating done (0.0s)
-- Build files have been written to: /home/rick/ollama/0.12.10/build
<!-- gh-comment-id:3517536688 --> @rick-github commented on GitHub (Nov 11, 2025): ```console $ mkdir 0.12.10 && cd 0.12.10 && git clone https://github.com/ollama/ollama -b v0.12.10 . && cmake -B build Cloning into '.'... remote: Enumerating objects: 41175, done. remote: Counting objects: 100% (2/2), done. remote: Compressing objects: 100% (2/2), done. remote: Total 41175 (delta 0), reused 0 (delta 0), pack-reused 41173 (from 3) Receiving objects: 100% (41175/41175), 65.12 MiB | 29.88 MiB/s, done. Resolving deltas: 100% (25869/25869), done. Note: switching to '80d34260ea16e76c9ef0d014a86cc130421855f1'. You are in 'detached HEAD' state. You can look around, make experimental changes and commit them, and you can discard any commits you make in this state without impacting any branches by switching back to a branch. If you want to create a new branch to retain commits you create, you may do so (now or later) by using -c with the switch command. Example: git switch -c <new-branch-name> Or undo this operation with: git switch - Turn off this advice by setting config variable advice.detachedHead to false -- The C compiler identification is GNU 13.2.0 -- The CXX compiler identification is GNU 13.2.0 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success -- Found Threads: TRUE -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- GGML_SYSTEM_ARCH: x86 -- Including CPU backend -- x86 detected -- Adding CPU backend variant ggml-cpu-x64: -- x86 detected -- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42 -- x86 detected -- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX -- x86 detected -- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2 -- x86 detected -- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512 -- x86 detected -- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI -- x86 detected -- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI -- Looking for a CUDA compiler -- Looking for a CUDA compiler - /usr/bin/nvcc -- Found CUDAToolkit: /usr/include (found version "12.0.140") -- CUDA Toolkit found -- Using CUDA architectures: native -- The CUDA compiler identification is NVIDIA 12.0.140 -- Detecting CUDA compiler ABI info -- Detecting CUDA compiler ABI info - done -- Check for working CUDA compiler: /usr/bin/nvcc - skipped -- Detecting CUDA compile features -- Detecting CUDA compile features - done -- Looking for a HIP compiler -- Looking for a HIP compiler - NOTFOUND -- Found Vulkan: /home/rick/ollama/1.4.328.1/x86_64/lib/libvulkan.so (found version "1.4.328") found components: glslc glslangValidator -- Vulkan found -- GL_KHR_cooperative_matrix supported by glslc -- GL_NV_cooperative_matrix2 supported by glslc -- GL_EXT_integer_dot_product supported by glslc -- GL_EXT_bfloat16 supported by glslc -- Configuring done (6.9s) -- Generating done (0.0s) -- Build files have been written to: /home/rick/ollama/0.12.10/build ```
Author
Owner

@abcbarryn commented on GitHub (Nov 12, 2025):

Interesting, I downloaded 0.12.10 from the release page and while 0.12.9 detected CUDA, 0.12.10 on my system downloaded from the release page did not.

I re downloaded it using git clone and got this output...

-- The C compiler identification is GNU 14.3.0
-- The CXX compiler identification is GNU 14.3.0
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /usr/bin/cc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /usr/bin/c++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD
-- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success
-- Found Threads: TRUE
-- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF
-- CMAKE_SYSTEM_PROCESSOR: x86_64
-- GGML_SYSTEM_ARCH: x86
-- Including CPU backend
-- x86 detected
-- Adding CPU backend variant ggml-cpu-x64:
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42
-- x86 detected
-- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX
-- x86 detected
-- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2
-- x86 detected
-- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512
-- x86 detected
-- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI
-- x86 detected
-- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI
-- Looking for a CUDA compiler
-- Looking for a CUDA compiler - NOTFOUND
-- Looking for a HIP compiler
-- Looking for a HIP compiler - NOTFOUND
-- Found Vulkan: /usr/lib64/libvulkan.so (found version "1.3.275") found components: glslc missing components: glslangValidator
-- Vulkan found
-- GL_KHR_cooperative_matrix supported by glslc
-- GL_NV_cooperative_matrix2 supported by glslc
-- GL_EXT_integer_dot_product supported by glslc
-- GL_EXT_bfloat16 supported by glslc
-- Configuring done (1.2s)
-- Generating done (0.1s)
-- Build files have been written to: /usr/src/ollama-0.12.10/build
<!-- gh-comment-id:3519814244 --> @abcbarryn commented on GitHub (Nov 12, 2025): Interesting, I downloaded 0.12.10 from the release page and while 0.12.9 detected CUDA, 0.12.10 on my system downloaded from the release page did not. I re downloaded it using git clone and got this output... ``` -- The C compiler identification is GNU 14.3.0 -- The CXX compiler identification is GNU 14.3.0 -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /usr/bin/cc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /usr/bin/c++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Performing Test CMAKE_HAVE_LIBC_PTHREAD -- Performing Test CMAKE_HAVE_LIBC_PTHREAD - Success -- Found Threads: TRUE -- Warning: ccache not found - consider installing it for faster compilation or disable this warning with GGML_CCACHE=OFF -- CMAKE_SYSTEM_PROCESSOR: x86_64 -- GGML_SYSTEM_ARCH: x86 -- Including CPU backend -- x86 detected -- Adding CPU backend variant ggml-cpu-x64: -- x86 detected -- Adding CPU backend variant ggml-cpu-sse42: -msse4.2 GGML_SSE42 -- x86 detected -- Adding CPU backend variant ggml-cpu-sandybridge: -msse4.2;-mavx GGML_SSE42;GGML_AVX -- x86 detected -- Adding CPU backend variant ggml-cpu-haswell: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2 GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2 -- x86 detected -- Adding CPU backend variant ggml-cpu-skylakex: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512 -- x86 detected -- Adding CPU backend variant ggml-cpu-icelake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavx512f;-mavx512cd;-mavx512vl;-mavx512dq;-mavx512bw;-mavx512vbmi;-mavx512vnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX512;GGML_AVX512_VBMI;GGML_AVX512_VNNI -- x86 detected -- Adding CPU backend variant ggml-cpu-alderlake: -msse4.2;-mf16c;-mfma;-mbmi2;-mavx;-mavx2;-mavxvnni GGML_SSE42;GGML_F16C;GGML_FMA;GGML_BMI2;GGML_AVX;GGML_AVX2;GGML_AVX_VNNI -- Looking for a CUDA compiler -- Looking for a CUDA compiler - NOTFOUND -- Looking for a HIP compiler -- Looking for a HIP compiler - NOTFOUND -- Found Vulkan: /usr/lib64/libvulkan.so (found version "1.3.275") found components: glslc missing components: glslangValidator -- Vulkan found -- GL_KHR_cooperative_matrix supported by glslc -- GL_NV_cooperative_matrix2 supported by glslc -- GL_EXT_integer_dot_product supported by glslc -- GL_EXT_bfloat16 supported by glslc -- Configuring done (1.2s) -- Generating done (0.1s) -- Build files have been written to: /usr/src/ollama-0.12.10/build ```
Author
Owner

@abcbarryn commented on GitHub (Nov 12, 2025):

It looks like maybe adding: export INCLUDES="-I /usr/local/cuda/include"
got it working but this was not need on 0.12.9...?

<!-- gh-comment-id:3519910846 --> @abcbarryn commented on GitHub (Nov 12, 2025): It looks like maybe adding: export INCLUDES="-I /usr/local/cuda/include" got it working but this was not need on 0.12.9...?
Author
Owner

@YonTracks commented on GitHub (Nov 14, 2025):

Howdy! Not sure if this helps, but I ran into similar problems and it turned out to be my CUDA + CMake setup.

When CUDA 12.8 → 13 rolled out (somewhere around this commit: 2aba569a2a
), my builds started failing because I still had old Ollama + old CUDA remnants on my system. I basically had to start fresh.

What fixed it for me on Windows:

Clean up CUDA installs
Make sure only one CUDA version is installed.
I previously had multiple: CUDA_PATH, CUDA_PATH_V12, and leftover PATH entries like
C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v13.0\bin (in both user + system PATH).
Having multiple toolkits/env vars completely corrupted the build.
I uninstalled all CUDA toolkits and manually removed the stray env vars + PATH entries before reinstalling the correct version.

Use VS Code’s CMake Tools extension to configure the build
Since recent CMake updates, running

cmake -B build
cmake --build build --config Release

directly from PowerShell or terminal or any CLI stopped producing correct builds for me (needing like ming or similar).
Using the CMake Tools extension in VS Code fixed everything — it configures the project properly, compiles much faster, and provides clean logs. Auto-build also works great.

Hopefully this helps anyone hitting strange Windows/CUDA/CMake issues.
good luck.

<!-- gh-comment-id:3531426253 --> @YonTracks commented on GitHub (Nov 14, 2025): Howdy! Not sure if this helps, but I ran into similar problems and it turned out to be my CUDA + CMake setup. When CUDA 12.8 → 13 rolled out (somewhere around this commit: https://github.com/ollama/ollama/commit/2aba569a2a593f56651ded7f5011480ece70c80f ), my builds started failing because I still had old Ollama + old CUDA remnants on my system. I basically had to start fresh. What fixed it for me on Windows: Clean up CUDA installs Make sure only one CUDA version is installed. I previously had multiple: CUDA_PATH, CUDA_PATH_V12, and leftover PATH entries like C:\Program Files\NVIDIA GPU Computing Toolkit\CUDA\v13.0\bin (in both user + system PATH). Having multiple toolkits/env vars completely corrupted the build. I uninstalled all CUDA toolkits and manually removed the stray env vars + PATH entries before reinstalling the correct version. Use VS Code’s CMake Tools extension to configure the build Since recent CMake updates, running cmake -B build cmake --build build --config Release directly from PowerShell or terminal or any CLI stopped producing correct builds for me (needing like ming or similar). Using the CMake Tools extension in VS Code fixed everything — it configures the project properly, compiles much faster, and provides clean logs. Auto-build also works great. Hopefully this helps anyone hitting strange Windows/CUDA/CMake issues. good luck.
Author
Owner

@YonTracks commented on GitHub (Nov 14, 2025):

then after a successful build, I can then run powershell env:PKG_VERSION = "0.12.10 and env:VERSION = "0.12.10" and then build the OllamaSetup.exe installer via powershell -ExecutionPolicy Bypass -File .\scripts\build_windows.ps1 for the custom build.

Hopefully this helps, not hinders.
If there's anything I'm doing wrong or could do a better way let me know cheers, but this works flawless for me, cuda, no cuda, cuda 13 or whatever and ROCm.

good luck

<!-- gh-comment-id:3531479845 --> @YonTracks commented on GitHub (Nov 14, 2025): then after a successful build, I can then run powershell `env:PKG_VERSION = "0.12.10` and `env:VERSION = "0.12.10"` and then build the OllamaSetup.exe installer via `powershell -ExecutionPolicy Bypass -File .\scripts\build_windows.ps1` for the custom build. Hopefully this helps, not hinders. If there's anything I'm doing wrong or could do a better way let me know cheers, but this works flawless for me, cuda, no cuda, cuda 13 or whatever and ROCm. good luck
Author
Owner

@abcbarryn commented on GitHub (Nov 14, 2025):

export INCLUDES="-I /usr/local/cuda/include"
got it working. Apparently the 0.12.10 CMakeLists.txt file could not find the CUDA include directory, but 0.12.9 did.

<!-- gh-comment-id:3532311338 --> @abcbarryn commented on GitHub (Nov 14, 2025): export INCLUDES="-I /usr/local/cuda/include" got it working. Apparently the 0.12.10 CMakeLists.txt file could not find the CUDA include directory, but 0.12.9 did.
Author
Owner

@abcbarryn commented on GitHub (Nov 16, 2025):

export INCLUDES="-I /usr/local/cuda/include" caused issue #12872, I got Ollama to build properly using these settings:

export CMAKE_CUDA_COMPILER=/usr/local/cuda/bin/nvcc
export PATH="/usr/local/cuda/nvvm/bin:/usr/local/cuda/bin:/usr/local/cuda/lib64:/sbin:/usr/sbin:/usr/bin:/bin"
export LD_LIBRARY_PATH="/lib64:/usr/lib64:/usr/local/cuda/lib64:/usr/local/cuda/targets/x86_64-linux/lib"
cmake -B build -DCMAKE_CUDA_COMPILER="$CMAKE_CUDA_COMPILER"
cmake --build build

I had to add the -DCMAKE_CUDA_COMPILER="$CMAKE_CUDA_COMPILER" to the cmake command and set the path for cmake to properly find the CUDA compiler. I removed export INCLUDES="-I /usr/local/cuda/include" which was causing issue #12872
Using CUDA 12.9 with my GPU, it grumbles: nvcc warning : Support for offline compilation for architectures prior to '<compute/sm/lto>_75' will be removed in a future release (Use -Wno-deprecated-gpu-targets to suppress warning). but it compiles.

<!-- gh-comment-id:3538512696 --> @abcbarryn commented on GitHub (Nov 16, 2025): `export INCLUDES="-I /usr/local/cuda/include"` caused issue #12872, I got Ollama to build properly using these settings: ``` export CMAKE_CUDA_COMPILER=/usr/local/cuda/bin/nvcc export PATH="/usr/local/cuda/nvvm/bin:/usr/local/cuda/bin:/usr/local/cuda/lib64:/sbin:/usr/sbin:/usr/bin:/bin" export LD_LIBRARY_PATH="/lib64:/usr/lib64:/usr/local/cuda/lib64:/usr/local/cuda/targets/x86_64-linux/lib" cmake -B build -DCMAKE_CUDA_COMPILER="$CMAKE_CUDA_COMPILER" cmake --build build ``` I had to add the -DCMAKE_CUDA_COMPILER="$CMAKE_CUDA_COMPILER" to the cmake command and set the path for cmake to properly find the CUDA compiler. I removed `export INCLUDES="-I /usr/local/cuda/include"` which was causing issue #12872 Using CUDA 12.9 with my GPU, it grumbles: `nvcc warning : Support for offline compilation for architectures prior to '<compute/sm/lto>_75' will be removed in a future release (Use -Wno-deprecated-gpu-targets to suppress warning).` but it compiles.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#70706