[PR #13196] [MERGED] amd: use GTT on iGPUs on linux #60813

Closed
opened 2026-04-29 15:56:17 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/13196
Author: @dhiltgen
Created: 11/21/2025
Status: Merged
Merged: 12/23/2025
Merged by: @dhiltgen

Base: mainHead: amd_gtt


📝 Commits (1)

  • 6c72bbe amd: use GTT on iGPUs on linux

📊 Changes

9 files changed (+83 additions, -25 deletions)

View changed files

📝 llama/patches/0024-GPU-discovery-enhancements.patch (+44 -15)
📝 llama/patches/0028-Add-memory-detection-using-DXGI-PDH.patch (+3 -3)
📝 llama/patches/0029-ggml-cuda-skip-large-batches.patch (+1 -1)
📝 llama/patches/0030-fix-bakllava-regression.patch (+0 -0)
📝 llama/patches/0031-win-exit-instead-of-abort.patch (+0 -0)
📝 ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu (+1 -1)
📝 ml/backend/ggml/ggml/src/ggml-impl.h (+1 -1)
📝 ml/backend/ggml/ggml/src/ggml-vulkan/ggml-vulkan.cpp (+1 -1)
📝 ml/backend/ggml/ggml/src/mem_hip.cpp (+32 -3)

📄 Description

On Linux, look at the GTT memory information for iGPUs.

I've only been able to verify this on a single system so far, and it seems to work correctly. Draft until we can verify on more systems to ensure we don't inadvertently report too much available memory and start causing OOM crashes or gibberish responses.

Inspired by:

Fixes:


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/13196 **Author:** [@dhiltgen](https://github.com/dhiltgen) **Created:** 11/21/2025 **Status:** ✅ Merged **Merged:** 12/23/2025 **Merged by:** [@dhiltgen](https://github.com/dhiltgen) **Base:** `main` ← **Head:** `amd_gtt` --- ### 📝 Commits (1) - [`6c72bbe`](https://github.com/ollama/ollama/commit/6c72bbe47df7a36cade08fd0b16b80a41ea42559) amd: use GTT on iGPUs on linux ### 📊 Changes **9 files changed** (+83 additions, -25 deletions) <details> <summary>View changed files</summary> 📝 `llama/patches/0024-GPU-discovery-enhancements.patch` (+44 -15) 📝 `llama/patches/0028-Add-memory-detection-using-DXGI-PDH.patch` (+3 -3) 📝 `llama/patches/0029-ggml-cuda-skip-large-batches.patch` (+1 -1) 📝 `llama/patches/0030-fix-bakllava-regression.patch` (+0 -0) 📝 `llama/patches/0031-win-exit-instead-of-abort.patch` (+0 -0) 📝 `ml/backend/ggml/ggml/src/ggml-cuda/ggml-cuda.cu` (+1 -1) 📝 `ml/backend/ggml/ggml/src/ggml-impl.h` (+1 -1) 📝 `ml/backend/ggml/ggml/src/ggml-vulkan/ggml-vulkan.cpp` (+1 -1) 📝 `ml/backend/ggml/ggml/src/mem_hip.cpp` (+32 -3) </details> ### 📄 Description On Linux, look at the GTT memory information for iGPUs. I've only been able to verify this on a single system so far, and it seems to work correctly. Draft until we can verify on more systems to ensure we don't inadvertently report too much available memory and start causing OOM crashes or gibberish responses. Inspired by: - Closes #6282 - Closes #5426 Fixes: - Fixes #2637 - Fixes #4392 - Fixes #5471 - Fixes #6362 - Fixes #12062 - Fixes #12342 - Fixes #12411 - Fixes #13107 - Fixes #13173 --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-29 15:56:17 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#60813