[PR #10844] discover/gpu.go: Add Support for Distributed Inferencing (continued) #13386

Open
opened 2026-04-13 00:25:42 -05:00 by GiteaMirror · 0 comments
Owner

Original Pull Request: https://github.com/ollama/ollama/pull/10844

State: open
Merged: No


This PR builds on top of the work done by @ecyht2 on #6729, following issue #4643.
It aims to add RPC support to Ollama based on llama.cpp RPC mechanism to allow distributed inference across multiple devices.

This PR has been tested and confirmed working on MacOS (fixing a race condition in distributed inference). best performance can be achieved by connecting the devices using Thunderbolt 4.

This PR also adds the ollama rpc command that allows running the RPC server on the other computer.

**Original Pull Request:** https://github.com/ollama/ollama/pull/10844 **State:** open **Merged:** No --- This PR builds on top of the work done by @ecyht2 on #6729, following issue #4643. It aims to add RPC support to Ollama based on llama.cpp RPC mechanism to allow distributed inference across multiple devices. This PR has been tested and confirmed working on MacOS (fixing a race condition in distributed inference). best performance can be achieved by connecting the devices using Thunderbolt 4. This PR also adds the `ollama rpc` command that allows running the RPC server on the other computer.
GiteaMirror added the pull-request label 2026-04-13 00:25:42 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#13386