[PR #15213] [CLOSED] Feature/bonsai q1 0 support #77373

Closed
opened 2026-05-05 10:03:03 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/15213
Author: @nareshnavinash
Created: 4/2/2026
Status: Closed

Base: mainHead: feature/bonsai-q1-0-support


📝 Commits (2)

  • d5078b4 feat: add Bonsai Q1_0 1-bit quantization support
  • bdcec13 docs: update README for Bonsai Q1_0 fork

📊 Changes

38 files changed (+3861 additions, -55 deletions)

View changed files

📝 README.md (+64 -51)
📝 fs/ggml/ggml.go (+8 -1)
📝 fs/ggml/type.go (+22 -0)
📝 fs/gguf/tensor.go (+20 -1)
📝 llama/llama.cpp/include/llama.h (+2 -0)
📝 llama/llama.cpp/src/llama-model-loader.cpp (+4 -0)
📝 llama/llama.cpp/src/llama-quant.cpp (+2 -0)
llama/patches/0035-ggml-add-Q1_0-and-Q1_0_g128-1-bit-quantization-suppo.patch (+2009 -0)
📝 ml/backend/ggml/ggml/include/ggml.h (+5 -1)
📝 ml/backend/ggml/ggml/src/ggml-common.h (+21 -0)
📝 ml/backend/ggml/ggml/src/ggml-cpu/arch/arm/quants.c (+140 -0)
📝 ml/backend/ggml/ggml/src/ggml-cpu/arch/x86/quants.c (+159 -0)
📝 ml/backend/ggml/ggml/src/ggml-cpu/ggml-cpu.c (+12 -0)
📝 ml/backend/ggml/ggml/src/ggml-cpu/ops.cpp (+4 -0)
📝 ml/backend/ggml/ggml/src/ggml-cpu/quants.c (+101 -0)
📝 ml/backend/ggml/ggml/src/ggml-cpu/quants.h (+6 -0)
📝 ml/backend/ggml/ggml/src/ggml-cuda/common.cuh (+14 -0)
📝 ml/backend/ggml/ggml/src/ggml-cuda/convert.cu (+20 -0)
📝 ml/backend/ggml/ggml/src/ggml-cuda/dequantize.cuh (+46 -0)
📝 ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu (+4 -0)

...and 18 more files

📄 Description

No description provided


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/15213 **Author:** [@nareshnavinash](https://github.com/nareshnavinash) **Created:** 4/2/2026 **Status:** ❌ Closed **Base:** `main` ← **Head:** `feature/bonsai-q1-0-support` --- ### 📝 Commits (2) - [`d5078b4`](https://github.com/ollama/ollama/commit/d5078b4a18ef7d13f744a56632d6124cc1ac77e9) feat: add Bonsai Q1_0 1-bit quantization support - [`bdcec13`](https://github.com/ollama/ollama/commit/bdcec130865b5c9fbfb870ec0edd27333fa2c751) docs: update README for Bonsai Q1_0 fork ### 📊 Changes **38 files changed** (+3861 additions, -55 deletions) <details> <summary>View changed files</summary> 📝 `README.md` (+64 -51) 📝 `fs/ggml/ggml.go` (+8 -1) 📝 `fs/ggml/type.go` (+22 -0) 📝 `fs/gguf/tensor.go` (+20 -1) 📝 `llama/llama.cpp/include/llama.h` (+2 -0) 📝 `llama/llama.cpp/src/llama-model-loader.cpp` (+4 -0) 📝 `llama/llama.cpp/src/llama-quant.cpp` (+2 -0) ➕ `llama/patches/0035-ggml-add-Q1_0-and-Q1_0_g128-1-bit-quantization-suppo.patch` (+2009 -0) 📝 `ml/backend/ggml/ggml/include/ggml.h` (+5 -1) 📝 `ml/backend/ggml/ggml/src/ggml-common.h` (+21 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cpu/arch/arm/quants.c` (+140 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cpu/arch/x86/quants.c` (+159 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cpu/ggml-cpu.c` (+12 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cpu/ops.cpp` (+4 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cpu/quants.c` (+101 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cpu/quants.h` (+6 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cuda/common.cuh` (+14 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cuda/convert.cu` (+20 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cuda/dequantize.cuh` (+46 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu` (+4 -0) _...and 18 more files_ </details> ### 📄 Description _No description provided_ --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-05-05 10:03:03 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#77373