Logo
Explore Help
Sign In
github-starred/ollama-ollama
2
0
Fork 0
You've already forked ollama-ollama
mirror of https://github.com/ollama/ollama.git synced 2025-12-05 18:46:22 -06:00
Code Issues 2.2k Packages Projects Releases 100 Wiki Activity

100 Releases 358 Tags

RSS Feed
  • v0.9.0 5f57b0ef42
    Compare

    Stable

    GiteaMirror released this 2025-05-28 21:38:52 -05:00 | 535 commits to main since this release

    📅 Originally published on GitHub: Thu, 29 May 2025 05:41:01 GMT
    🏷️ Git tag created: Thu, 29 May 2025 02:38:52 GMT

    ollama thinking

    New models

    • DeepSeek-R1-2508: DeepSeek-R1 has received a minor version upgrade to DeepSeek-R1-0528 for the 8 billion parameter distilled model and the full 671 billion parameter model. In this update, DeepSeek R1 has significantly improved its reasoning and inference capabilities.

    Thinking

    Ollama now has the ability to enable or disable thinking. This gives users the flexibility to choose the model’s thinking behavior for different applications and use cases.

    When thinking is enabled, the output will separate the model’s thinking from the model’s output. When thinking is disabled, the model will not think and directly output the content.

    Models that support thinking:

    • DeepSeek R1
    • Qwen 3
    • more will be added under thinking models.

    When running a model that supports thinking, Ollama will now display the model's thoughts:

    % ollama run deepseek-r1
    >>> How many Rs are in strawberry
    Thinking...
    First, I need to understand what the question is asking. It's asking how many letters 'R' are present in the word "strawberry."
    
    Next, I'll examine each letter in the word individually.
    
    I'll start from the beginning and count every occurrence of the letter 'R.'
    
    After reviewing all the letters, I determine that there are three instances where the letter 'R' appears in the word "strawberry."
    ...done thinking.
    
    There are three **Rs** in the word **"strawberry"**.
    

    In Ollama's API, a model's thinking is now returned as a separate thinking field for easy parsing:

    {
      "message": {
        "role": "assistant",
        "thinking": "First, I need to understand what the question is asking. It's asking how many letters 'R' are present in the word "strawberry...",
        "content": "There are **3** instances of the letter **R** in the word **"strawberry."**"
      }
    }
    

    Turning thinking on and off

    In the API, thinking can be enabled by passing "think": true and disabled by passing "think": false

    curl http://localhost:11434/api/chat -d '{
      "model": "deepseek-r1",
      "messages": [
        {
          "role": "user",
          "content": "Why is the sky blue?"
        },
      ],
      "think": true
    }'
    

    In Ollama's CLI, use /set think and /set nothink to enable and disable thinking.

    What's Changed

    • Add thinking support to Ollama

    Full Changelog: https://github.com/ollama/ollama/compare/v0.8.0...v0.9.0

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      22 MiB
      2025-11-12 02:24:15 -06:00
    • Ollama-darwin.zip
      177 MiB
      2025-11-12 02:24:30 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 02:26:26 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 02:29:22 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      434 MiB
      2025-11-12 02:30:03 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      345 MiB
      2025-11-12 02:30:33 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 02:33:00 -06:00
    • OllamaSetup.exe
      980 MiB
      2025-11-12 02:37:43 -06:00
    • ollama-windows-amd64-rocm.zip
      423 MiB
      2025-11-12 02:33:40 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 02:36:01 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 02:36:04 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 02:37:44 -06:00
  • v0.8.0 aa25aff10d
    Compare

    Stable

    GiteaMirror released this 2025-05-27 18:50:57 -05:00 | 536 commits to main since this release

    📅 Originally published on GitHub: Tue, 27 May 2025 19:55:15 GMT
    🏷️ Git tag created: Tue, 27 May 2025 23:50:57 GMT

    image of ollama with a bigger and better tool

    What's Changed

    • Ollama will now stream responses with tool calls blog post
    • Logs will now include better memory estimate debug information when running models in Ollama's engine.

    New Contributors

    • @hellotunamayo made their first contribution in https://github.com/ollama/ollama/pull/10790

    Full Changelog: https://github.com/ollama/ollama/compare/v0.7.1...v0.8.0

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      22 MiB
      2025-11-12 02:10:59 -06:00
    • Ollama-darwin.zip
      177 MiB
      2025-11-12 02:11:15 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 02:13:07 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 02:15:28 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      434 MiB
      2025-11-12 02:16:08 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      345 MiB
      2025-11-12 02:16:39 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 02:19:35 -06:00
    • OllamaSetup.exe
      981 MiB
      2025-11-12 02:24:10 -06:00
    • ollama-windows-amd64-rocm.zip
      423 MiB
      2025-11-12 02:20:16 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 02:22:41 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 02:22:43 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 02:24:10 -06:00
  • v0.7.1 884d26093c
    Compare

    Stable

    GiteaMirror released this 2025-05-22 20:53:31 -05:00 | 548 commits to main since this release

    📅 Originally published on GitHub: Wed, 21 May 2025 22:17:35 GMT
    🏷️ Git tag created: Fri, 23 May 2025 01:53:31 GMT

    What's Changed

    • Improved model memory management to allocate sufficient memory to prevent crashes when running multimodal models in certain situations
    • Enhanced memory estimation for models to prevent unintended memory offloading
    • ollama show will now show ... when data is truncated
    • Fixed crash that would occur with qwen2.5vl
    • Fixed crash on Nvidia's CUDA for llama3.2-vision
    • Support for Alibaba's Qwen 3 and Qwen 2 architectures in Ollama's new multimodal engine

    New Contributors

    • @ronxldwilson made their first contribution in https://github.com/ollama/ollama/pull/10763
    • @DarkCaster made their first contribution in https://github.com/ollama/ollama/pull/10779

    Full Changelog: https://github.com/ollama/ollama/compare/v0.7.0...v0.7.1

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      22 MiB
      2025-11-12 01:56:53 -06:00
    • Ollama-darwin.zip
      177 MiB
      2025-11-12 01:57:10 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 01:59:35 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 02:02:02 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      434 MiB
      2025-11-12 02:02:56 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      345 MiB
      2025-11-12 02:03:29 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 02:05:59 -06:00
    • OllamaSetup.exe
      980 MiB
      2025-11-12 02:10:54 -06:00
    • ollama-windows-amd64-rocm.zip
      423 MiB
      2025-11-12 02:06:36 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 02:09:23 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 02:09:26 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 02:10:55 -06:00
  • v0.7.0 bd68d3ae50
    Compare

    Stable

    GiteaMirror released this 2025-05-14 18:42:30 -05:00 | 580 commits to main since this release

    📅 Originally published on GitHub: Tue, 13 May 2025 00:10:33 GMT
    🏷️ Git tag created: Wed, 14 May 2025 23:42:30 GMT

    multimodality-uncompressed

    Ollama now supports multimodal models via Ollama’s new engine, starting with new vision multimodal models:

    • Meta Llama 4
    • Google Gemma 3
    • Qwen 2.5 VL
    • Mistral Small 3.1
    • and more vision models.

    What's Changed

    • Ollama now supports providing WebP images as input to multimodal models
    • Fixed issue where a blank terminal window would appear when runnings models on Windows
    • Fixed error that would occur when running llama4 on NVIDIA GPUs
    • Reduced log level of key not found message
    • Ollama will now correct remove quotes from image paths when sending images as input with ollama run
    • Improved performance of importing safetensors models via ollama create
    • Improved prompt processing speeds of Qwen3 MoE on macOS
    • Fixed issue where providing large JSON schemas in structured output requests would result in an error
    • Ollama's API will now return code 405 instead of 404 for methods that are not allowed
    • Fixed issue where ollama processes would continue to run after a model was unloaded

    New Contributors

    • @ashokgelal made their first contribution in https://github.com/ollama/ollama/pull/8668
    • @Aharon-Bensadoun made their first contribution in https://github.com/ollama/ollama/pull/9719
    • @HardCodeDev777 made their first contribution in https://github.com/ollama/ollama/pull/10664

    Full Changelog: https://github.com/ollama/ollama/compare/v0.6.8...v0.7.0

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      22 MiB
      2025-11-12 01:43:21 -06:00
    • Ollama-darwin.zip
      177 MiB
      2025-11-12 01:43:37 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 01:45:31 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 01:48:14 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      434 MiB
      2025-11-12 01:49:10 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      345 MiB
      2025-11-12 01:49:42 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 01:52:14 -06:00
    • OllamaSetup.exe
      980 MiB
      2025-11-12 01:56:49 -06:00
    • ollama-windows-amd64-rocm.zip
      423 MiB
      2025-11-12 01:52:52 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 01:55:18 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 01:55:20 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 01:56:49 -06:00
  • v0.6.8 6a74bba7e7
    Compare

    Stable

    GiteaMirror released this 2025-05-03 15:11:48 -05:00 | 626 commits to main since this release

    📅 Originally published on GitHub: Sat, 03 May 2025 22:56:44 GMT
    🏷️ Git tag created: Sat, 03 May 2025 20:11:48 GMT

    What's Changed

    • Performance improvements for Qwen 3 MoE models (30b-a3b and 235b-a22b) on NVIDIA and AMD GPUs
    • Fixed GGML_ASSERT(tensor->op == GGML_OP_UNARY) failed issue caused by conflicting installations
    • Fixed a memory leak that occurred when providing images as input
    • ollama show will now correctly label older vision models such as llava
    • Reduced out of memory errors by improving worst-case memory estimations
    • Fix issue that resulted in a context canceled error

    New Contributors

    • @AliAhmedNada made their first contribution in https://github.com/ollama/ollama/pull/10522
    • @HarshNevse made their first contribution in https://github.com/ollama/ollama/pull/10465

    Full Changelog: https://github.com/ollama/ollama/compare/v0.6.7...v0.6.8

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      22 MiB
      2025-11-12 01:29:52 -06:00
    • Ollama-darwin.zip
      176 MiB
      2025-11-12 01:30:09 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 01:32:05 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 01:34:26 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      434 MiB
      2025-11-12 01:35:04 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      345 MiB
      2025-11-12 01:35:36 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 01:38:15 -06:00
    • OllamaSetup.exe
      979 MiB
      2025-11-12 01:43:17 -06:00
    • ollama-windows-amd64-rocm.zip
      422 MiB
      2025-11-12 01:39:06 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 01:41:50 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 01:41:53 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 01:43:17 -06:00
  • v0.6.7 a7835c6716
    Compare

    Stable

    GiteaMirror released this 2025-04-30 19:59:31 -05:00 | 638 commits to main since this release

    📅 Originally published on GitHub: Sat, 26 Apr 2025 09:16:02 GMT
    🏷️ Git tag created: Thu, 01 May 2025 00:59:31 GMT

    New models

    • Qwen 3: Qwen3 is the latest generation of large language models in Qwen series, offering a comprehensive suite of dense and mixture-of-experts (MoE) models.
    • Phi 4 reasoning and Phi-4-mini-reasoning: New state-of-the-art reasoning models from Microsoft
    • Llama 4: state-of-the-art multi-modal models from Meta

    What's Changed

    • Add support for Meta's Llama 4 multimodal models
    • Add support for Microsoft's Phi 4 reasoning models, and Phi 4 mini reasoning model
    • Increased default context window to 4096 tokens
    • Fixed issue where image paths would not be recognized with ~ when being provided to ollama run
    • Improved output quality when using JSON mode in certain scenarios
    • Fixed tensor->op == GGML_OP_UNARY errors when running a model due to conflicting inference libraries
    • Fixed issue where model would be stuck in the Stopping... state

    New Contributors

    • @greengrass821 made their first contribution in https://github.com/ollama/ollama/pull/10339
    • @richardshiue made their first contribution in https://github.com/ollama/ollama/pull/10335
    • @aduermael made their first contribution in https://github.com/ollama/ollama/pull/10386

    Full Changelog: https://github.com/ollama/ollama/compare/v0.6.6...v0.6.7

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      22 MiB
      2025-11-12 01:15:56 -06:00
    • Ollama-darwin.zip
      176 MiB
      2025-11-12 01:16:12 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 01:18:36 -06:00
    • ollama-linux-amd64.tgz
      1.5 GiB
      2025-11-12 01:21:15 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      424 MiB
      2025-11-12 01:22:00 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      340 MiB
      2025-11-12 01:22:37 -06:00
    • ollama-linux-arm64.tgz
      1.5 GiB
      2025-11-12 01:24:52 -06:00
    • OllamaSetup.exe
      954 MiB
      2025-11-12 01:29:47 -06:00
    • ollama-windows-amd64-rocm.zip
      382 MiB
      2025-11-12 01:25:29 -06:00
    • ollama-windows-amd64.zip
      1.5 GiB
      2025-11-12 01:28:00 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 01:28:02 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 01:29:48 -06:00
  • v0.6.6 88738b357b
    Compare

    Stable

    GiteaMirror released this 2025-04-18 20:13:05 -05:00 | 679 commits to main since this release

    📅 Originally published on GitHub: Thu, 17 Apr 2025 04:34:58 GMT
    🏷️ Git tag created: Sat, 19 Apr 2025 01:13:05 GMT

    New models

    • IBM Granite 3.3: 2B and 8B models with 128K context length that have been fine-tuned for improved reasoning and instruction-following capabilities.
    • DeepCoder: a fully open-Source 14B coder model at O3-mini level, with a 1.5B version also available.

    What's Changed

    • New, faster model downloading: OLLAMA_EXPERIMENT=client2 ollama serve will run Ollama using a new downloader with improved performance and reliability when running ollama pull. Please share feedback here!
    • Fixed memory leak issues when running Gemma 3, Mistral Small 3.1 and other models on Ollama
    • Improved performance of ollama create when importing models from Safetensors
    • Ollama will now allow tool function parameters with either a single type or an array of types by @rozgo
    • Fixed certain out of memory issues from not reserving enough memory at startup
    • Fix nondeterministic model unload order by @IreGaddr
    • Include the items and $defs fields to properly handle array types in the API by @sheffler
    • OpenAI-Beta headers are now included in the CORS safelist by @drifkin
    • Fixed issue where model tensor data would be corrupted when importing models from Safetensors

    New Contributors

    • @drifkin made their first contribution in https://github.com/ollama/ollama/pull/10169
    • @rozgo made their first contribution in https://github.com/ollama/ollama/pull/9434
    • @qwerty108109 made their first contribution in https://github.com/ollama/ollama/pull/10168
    • @IreGaddr made their first contribution in https://github.com/ollama/ollama/pull/10185
    • @sheffler made their first contribution in https://github.com/ollama/ollama/pull/10091

    Full Changelog: https://github.com/ollama/ollama/compare/v0.6.5...v0.6.6

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      22 MiB
      2025-11-12 01:03:10 -06:00
    • Ollama-darwin.zip
      176 MiB
      2025-11-12 01:03:27 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 01:05:14 -06:00
    • ollama-linux-amd64.tgz
      1.5 GiB
      2025-11-12 01:07:37 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      424 MiB
      2025-11-12 01:08:25 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      340 MiB
      2025-11-12 01:09:02 -06:00
    • ollama-linux-arm64.tgz
      1.5 GiB
      2025-11-12 01:11:31 -06:00
    • OllamaSetup.exe
      954 MiB
      2025-11-12 01:15:52 -06:00
    • ollama-windows-amd64-rocm.zip
      382 MiB
      2025-11-12 01:12:09 -06:00
    • ollama-windows-amd64.zip
      1.5 GiB
      2025-11-12 01:14:24 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 01:14:27 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 01:15:52 -06:00
  • v0.6.5 0f3f9e353d
    Compare

    Stable

    GiteaMirror released this 2025-04-04 19:04:24 -05:00 | 720 commits to main since this release

    📅 Originally published on GitHub: Sun, 06 Apr 2025 00:15:39 GMT
    🏷️ Git tag created: Sat, 05 Apr 2025 00:04:24 GMT

    New models

    • Mistral Small 3.1: the best performing vision model in its weight class.

    What's Changed

    • Support for Mistral Small 3.1
    • Improved model loading times for Gemma 3 on network-backed filesystems such as Google Cloud Storage FUSE

    New Contributors

    • @danhipke made their first contribution in https://github.com/ollama/ollama/pull/10133

    Full Changelog: https://github.com/ollama/ollama/compare/v0.6.4...v0.6.5

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      21 MiB
      2025-11-12 00:49:33 -06:00
    • Ollama-darwin.zip
      175 MiB
      2025-11-12 00:49:54 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 00:51:59 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 00:54:22 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      451 MiB
      2025-11-12 00:55:03 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      343 MiB
      2025-11-12 00:55:36 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 00:58:24 -06:00
    • OllamaSetup.exe
      1003 MiB
      2025-11-12 01:03:05 -06:00
    • ollama-windows-amd64-rocm.zip
      385 MiB
      2025-11-12 00:59:05 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 01:01:38 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 01:01:40 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 01:03:06 -06:00
  • v0.6.4 b51e0f397c
    Compare

    Stable

    GiteaMirror released this 2025-04-02 15:22:56 -05:00 | 725 commits to main since this release

    📅 Originally published on GitHub: Wed, 02 Apr 2025 22:14:24 GMT
    🏷️ Git tag created: Wed, 02 Apr 2025 20:22:56 GMT

    What's Changed

    • /api/show will now include model capabilities such as vision
    • Fixed certain out-of-memory errors that would occur with parallel requests with Gemma 3
    • Gemma 3 will now properly understand and output certain multilingual characters
    • Fixed context shifting issues with models using the DeepSeek architecture
    • Fixed issues with 0.6.3 where Gemma 3's output quality would worsen after 512 or 1024 tokens
    • Added AMD RDNA4 support on Linux

    New Contributors

    • @saman-amd made their first contribution in https://github.com/ollama/ollama/pull/9878
    • @leandroBorgesFerreira made their first contribution in https://github.com/ollama/ollama/pull/10042
    • @Abyss-c0re made their first contribution in https://github.com/ollama/ollama/pull/9955
    • @uggrock made their first contribution in https://github.com/ollama/ollama/pull/9983
    • @IsAurora6 made their first contribution in https://github.com/ollama/ollama/pull/10057

    Full Changelog: https://github.com/ollama/ollama/compare/v0.6.3...v0.6.4

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      21 MiB
      2025-11-12 00:34:22 -06:00
    • Ollama-darwin.zip
      175 MiB
      2025-11-12 00:34:39 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 00:36:38 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 00:40:09 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      451 MiB
      2025-11-12 00:40:50 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      343 MiB
      2025-11-12 00:41:34 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 00:44:32 -06:00
    • OllamaSetup.exe
      1003 MiB
      2025-11-12 00:49:28 -06:00
    • ollama-windows-amd64-rocm.zip
      385 MiB
      2025-11-12 00:45:08 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 00:47:42 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 00:47:46 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 00:49:29 -06:00
  • v0.6.3 e5d84fb90b
    Compare

    Stable

    GiteaMirror released this 2025-03-26 15:39:01 -05:00 | 743 commits to main since this release

    📅 Originally published on GitHub: Sat, 22 Mar 2025 02:56:08 GMT
    🏷️ Git tag created: Wed, 26 Mar 2025 20:39:01 GMT

    What's Changed

    • New sliding window attention optimizations for Gemma 3, improving inference speed and memory allocation for long context windows.
    • Improved loading speed of Gemma 3
    • ollama create will now return the name of unsupported architectures
    • Fixed error talloc->buffer_id >= 0 when running a model
    • Fixed (int)sched->hash_set.size >= graph->n_nodes + graph->n_leafs error when running a model
    • ollama create will now automatically select the right template when importing Gemma 3 from safetensors
    • ollama show -v will now correctly render boolean values as true or false

    New Contributors

    • @rylativity made their first contribution in https://github.com/ollama/ollama/pull/9874

    Full Changelog: https://github.com/ollama/ollama/compare/v0.6.2...v0.6.3

    Downloads
    • Source Code (ZIP)
    • Source Code (TAR.GZ)
    • ollama-darwin.tgz
      21 MiB
      2025-11-12 00:19:00 -06:00
    • Ollama-darwin.zip
      175 MiB
      2025-11-12 00:19:23 -06:00
    • ollama-linux-amd64-rocm.tgz
      1.2 GiB
      2025-11-12 00:21:25 -06:00
    • ollama-linux-amd64.tgz
      1.6 GiB
      2025-11-12 00:24:22 -06:00
    • ollama-linux-arm64-jetpack5.tgz
      451 MiB
      2025-11-12 00:25:04 -06:00
    • ollama-linux-arm64-jetpack6.tgz
      343 MiB
      2025-11-12 00:25:40 -06:00
    • ollama-linux-arm64.tgz
      1.6 GiB
      2025-11-12 00:29:05 -06:00
    • OllamaSetup.exe
      1001 MiB
      2025-11-12 00:34:17 -06:00
    • ollama-windows-amd64-rocm.zip
      368 MiB
      2025-11-12 00:29:49 -06:00
    • ollama-windows-amd64.zip
      1.6 GiB
      2025-11-12 00:32:28 -06:00
    • ollama-windows-arm64.zip
      20 MiB
      2025-11-12 00:32:30 -06:00
    • sha256sum.txt
      1018 B
      2025-11-12 00:34:18 -06:00
First Previous ... 2 3 4 5 6 ... Next Last
Powered by Gitea Version: 1.24.6 Page: 77ms Template: 20ms
English
Bahasa Indonesia Deutsch English Español Français Gaeilge Italiano Latviešu Magyar nyelv Nederlands Polski Português de Portugal Português do Brasil Suomi Svenska Türkçe Čeština Ελληνικά Български Русский Українська فارسی മലയാളം 日本語 简体中文 繁體中文(台灣) 繁體中文(香港) 한국어
Licenses API