[PR #1497] [CLOSED] patches: Clean up llama.cpp patches, update submodules to latest upstream #15873

Closed
opened 2026-04-16 05:10:49 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/1497
Author: @tmc
Created: 12/13/2023
Status: Closed

Base: mainHead: patch-cleanup


📝 Commits (2)

  • 379877f patches: Clean up llama.cpp patches, update submodules to latest upstream
  • fd74756 llama.cpp: Update both to 948ff137ec37f1ec74c02905917fa0afc9b97514

📊 Changes

10 files changed (+1 additions, -257 deletions)

View changed files

📝 llm/llama.cpp/generate_darwin_amd64.go (+0 -4)
📝 llm/llama.cpp/generate_darwin_arm64.go (+0 -4)
📝 llm/llama.cpp/generate_linux.go (+0 -3)
📝 llm/llama.cpp/generate_windows.go (+0 -2)
📝 llm/llama.cpp/ggml (+1 -1)
llm/llama.cpp/patches/0001-add-detokenize-endpoint.patch (+0 -51)
llm/llama.cpp/patches/0002-34B-model-support.patch (+0 -89)
llm/llama.cpp/patches/0003-metal-fix-synchronization-in-new-matrix-multiplicati.patch (+0 -30)
llm/llama.cpp/patches/0004-metal-add-missing-barriers-for-mul-mat-2699.patch (+0 -41)
llm/llama.cpp/patches/0005-ggml-support-CUDA-s-half-type-for-aarch64-1455-2670.patch (+0 -32)

📄 Description

This updates llama.cpp submodules to latest (fecac4) and removes the patches that have landed in llama.cpp already (most of them).

Since the "ggml" source tree doesn't appear to need to be patched anymore it seems like we can reduce down to one submodule here but I did not perform that refactor for simplicity, let me know if you'd like me to tackle that.


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/1497 **Author:** [@tmc](https://github.com/tmc) **Created:** 12/13/2023 **Status:** ❌ Closed **Base:** `main` ← **Head:** `patch-cleanup` --- ### 📝 Commits (2) - [`379877f`](https://github.com/ollama/ollama/commit/379877f916a8973f153d30751d144a6b13e32ef8) patches: Clean up llama.cpp patches, update submodules to latest upstream - [`fd74756`](https://github.com/ollama/ollama/commit/fd747567f52544aaae3f85292ea43dd843f24fe9) llama.cpp: Update both to 948ff137ec37f1ec74c02905917fa0afc9b97514 ### 📊 Changes **10 files changed** (+1 additions, -257 deletions) <details> <summary>View changed files</summary> 📝 `llm/llama.cpp/generate_darwin_amd64.go` (+0 -4) 📝 `llm/llama.cpp/generate_darwin_arm64.go` (+0 -4) 📝 `llm/llama.cpp/generate_linux.go` (+0 -3) 📝 `llm/llama.cpp/generate_windows.go` (+0 -2) 📝 `llm/llama.cpp/ggml` (+1 -1) ➖ `llm/llama.cpp/patches/0001-add-detokenize-endpoint.patch` (+0 -51) ➖ `llm/llama.cpp/patches/0002-34B-model-support.patch` (+0 -89) ➖ `llm/llama.cpp/patches/0003-metal-fix-synchronization-in-new-matrix-multiplicati.patch` (+0 -30) ➖ `llm/llama.cpp/patches/0004-metal-add-missing-barriers-for-mul-mat-2699.patch` (+0 -41) ➖ `llm/llama.cpp/patches/0005-ggml-support-CUDA-s-half-type-for-aarch64-1455-2670.patch` (+0 -32) </details> ### 📄 Description This updates llama.cpp submodules to latest (fecac4) and removes the patches that have landed in llama.cpp already (most of them). Since the "ggml" source tree doesn't appear to need to be patched anymore it seems like we can reduce down to one submodule here but I did not perform that refactor for simplicity, let me know if you'd like me to tackle that. --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-16 05:10:49 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#15873