[GH-ISSUE #2800] Make Ollama serve requests in parallel #27454

Closed
opened 2026-04-22 04:48:39 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @pankajkumar229 on GitHub (Feb 28, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/2800

It seems that requests are served sequentially. It would help if they can be served in parallel.

Originally created by @pankajkumar229 on GitHub (Feb 28, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/2800 It seems that requests are served sequentially. It would help if they can be served in parallel.
Author
Owner

@kevintanhongann commented on GitHub (Feb 29, 2024):

streaming support.

<!-- gh-comment-id:1970277217 --> @kevintanhongann commented on GitHub (Feb 29, 2024): streaming support.
Author
Owner

@pdevine commented on GitHub (Mar 1, 2024):

Closing as a dupe of #358

<!-- gh-comment-id:1972288637 --> @pdevine commented on GitHub (Mar 1, 2024): Closing as a dupe of #358
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#27454