[PR #11125] Increase default context size to 8k #13448

Closed
opened 2026-04-13 00:27:36 -05:00 by GiteaMirror · 0 comments
Owner

Original Pull Request: https://github.com/ollama/ollama/pull/11125

State: closed
Merged: No


Adjust the default parallism down to 1 so this has minimal impacts on memory usage. Note for small VRAM GPUs this may cause fewer layers to load if we already reduced the parallelism down to 1. This may also impact some users of mllama which already reduced num parallel to 1.

Draft until I can test on various VRAM sizes to confirm minimal impact.

**Original Pull Request:** https://github.com/ollama/ollama/pull/11125 **State:** closed **Merged:** No --- Adjust the default parallism down to 1 so this has minimal impacts on memory usage. Note for small VRAM GPUs this may cause fewer layers to load if we already reduced the parallelism down to 1. This may also impact some users of mllama which already reduced num parallel to 1. Draft until I can test on various VRAM sizes to confirm minimal impact.
GiteaMirror added the pull-request label 2026-04-13 00:27:36 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#13448