[PR #4266] [CLOSED] Support forced spreading for multi GPU #58128

Closed
opened 2026-04-29 12:56:54 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/4266
Author: @dhiltgen
Created: 5/8/2024
Status: Closed

Base: mainHead: concurrency


📝 Commits (1)

  • 96ec776 Support forced spreading for multi GPU

📊 Changes

2 files changed (+18 additions, -4 deletions)

View changed files

📝 server/envconfig/config.go (+12 -0)
📝 server/sched.go (+6 -4)

📄 Description

Our default behavior today is to try to fit into a single GPU if possible. Some users would prefer the old behavior of always spreading across multiple GPUs even if the model can fit into one. This exposes that tunable behavior.

Fixes #4198


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/4266 **Author:** [@dhiltgen](https://github.com/dhiltgen) **Created:** 5/8/2024 **Status:** ❌ Closed **Base:** `main` ← **Head:** `concurrency` --- ### 📝 Commits (1) - [`96ec776`](https://github.com/ollama/ollama/commit/96ec776349ff77e2a5e8be8c9ec125055dd89db2) Support forced spreading for multi GPU ### 📊 Changes **2 files changed** (+18 additions, -4 deletions) <details> <summary>View changed files</summary> 📝 `server/envconfig/config.go` (+12 -0) 📝 `server/sched.go` (+6 -4) </details> ### 📄 Description Our default behavior today is to try to fit into a single GPU if possible. Some users would prefer the old behavior of always spreading across multiple GPUs even if the model can fit into one. This exposes that tunable behavior. Fixes #4198 --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-29 12:56:54 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#58128