[PR #8122] [CLOSED] build: streamline build #12633

Closed
opened 2026-04-13 00:05:20 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/8122
Author: @dhiltgen
Created: 12/16/2024
Status: Closed

Base: mainHead: streamline


📝 Commits (3)

  • ee2bfb2 build: streamline build
  • 1680409 build: clean up patches on create
  • af61353 ci: fix runner trigger

📊 Changes

15 files changed (+6 additions, -69 deletions)

View changed files

📝 .github/workflows/test.yaml (+1 -1)
📝 Dockerfile (+0 -2)
📝 llama/ggml-cuda/ggml-cuda.cu (+0 -2)
📝 llama/patches/0006-blas.patch (+0 -0)
llama/patches/0006-conditional-fattn.patch (+0 -25)
📝 llama/patches/0007-add-mllama-support.patch (+0 -0)
📝 llama/patches/0008-add-unpad-operator.patch (+2 -2)
📝 llama/patches/0009-fix-deepseek-deseret-regex.patch (+0 -0)
📝 llama/patches/0010-relative-include-paths.patch (+0 -0)
📝 llama/patches/0011-Maintain-ordering-for-rules-for-grammar.patch (+0 -0)
📝 llama/patches/0012-fix-missing-arg-in-static-assert-on-windows.patch (+0 -0)
📝 make/Makefile.sync (+1 -0)
📝 make/gpu.make (+2 -16)
📝 scripts/env.sh (+0 -1)
scripts/fast.sh (+0 -20)

📄 Description

This wiring was intended to make a faster developer build by disabling flash attention but the added complexity and friction on updates makes this less useful


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/8122 **Author:** [@dhiltgen](https://github.com/dhiltgen) **Created:** 12/16/2024 **Status:** ❌ Closed **Base:** `main` ← **Head:** `streamline` --- ### 📝 Commits (3) - [`ee2bfb2`](https://github.com/ollama/ollama/commit/ee2bfb2e4302bb99b12c437bc4bd3e1efe776887) build: streamline build - [`1680409`](https://github.com/ollama/ollama/commit/1680409fe909404cc2d10511c08dec4da3de8f49) build: clean up patches on create - [`af61353`](https://github.com/ollama/ollama/commit/af61353af23e6a368390477300860cb7fce93203) ci: fix runner trigger ### 📊 Changes **15 files changed** (+6 additions, -69 deletions) <details> <summary>View changed files</summary> 📝 `.github/workflows/test.yaml` (+1 -1) 📝 `Dockerfile` (+0 -2) 📝 `llama/ggml-cuda/ggml-cuda.cu` (+0 -2) 📝 `llama/patches/0006-blas.patch` (+0 -0) ➖ `llama/patches/0006-conditional-fattn.patch` (+0 -25) 📝 `llama/patches/0007-add-mllama-support.patch` (+0 -0) 📝 `llama/patches/0008-add-unpad-operator.patch` (+2 -2) 📝 `llama/patches/0009-fix-deepseek-deseret-regex.patch` (+0 -0) 📝 `llama/patches/0010-relative-include-paths.patch` (+0 -0) 📝 `llama/patches/0011-Maintain-ordering-for-rules-for-grammar.patch` (+0 -0) 📝 `llama/patches/0012-fix-missing-arg-in-static-assert-on-windows.patch` (+0 -0) 📝 `make/Makefile.sync` (+1 -0) 📝 `make/gpu.make` (+2 -16) 📝 `scripts/env.sh` (+0 -1) ➖ `scripts/fast.sh` (+0 -20) </details> ### 📄 Description This wiring was intended to make a faster developer build by disabling flash attention but the added complexity and friction on updates makes this less useful --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-13 00:05:20 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#12633