-
released this
2025-10-16 15:07:41 -05:00 | 185 commits to main since this release📅 Originally published on GitHub: Wed, 15 Oct 2025 23:02:31 GMT
🏷️ Git tag created: Thu, 16 Oct 2025 20:07:41 GMTWhat's Changed
- Ollama's app now supports searching when running DeepSeek-V3.1, Qwen3 and other models that support tool calling.
- Flash attention is now enabled by default for Gemma 3, improving performance and memory utilization
- Fixed issue where Ollama would hang while generating responses
- Fixed issue where
qwen3-coderwould act in raw mode when using/api/generateorollama run qwen3-coder <prompt> - Fixed
qwen3-embeddingproviding invalid results - Ollama will now evict models correctly when
num_gpuis set - Fixed issue where
tool_indexwith a value of0would not be sent to the model
Experimental Vulkan Support
Experimental support for Vulkan is now available when you build locally from source. This will enable additional GPUs from AMD, and Intel which are not currently supported by Ollama. To build locally, install the Vulkan SDK and set VULKAN_SDK in your environment, then follow the developer instructions. In a future release, Vulkan support will be included in the binary release as well. Please file issues if you run into any problems.
New Contributors
- @yajianggroup made their first contribution in https://github.com/ollama/ollama/pull/12377
- @inforithmics made their first contribution in https://github.com/ollama/ollama/pull/11835
- @sbhavani made their first contribution in https://github.com/ollama/ollama/pull/12619
Full Changelog: https://github.com/ollama/ollama/compare/v0.12.5...v0.12.6
Downloads
mirror of
https://github.com/ollama/ollama.git
synced 2025-12-05 18:46:22 -06:00