mirror of
https://github.com/open-webui/open-webui.git
synced 2026-03-09 23:35:09 -05:00
feat: Max Context Length Mode #5640
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @Notbici on GitHub (Jun 25, 2025).
Check Existing Issues
Problem Description
When I download a model, its always set to 2048 or something close to that in num_ctx.
I've got to go in and change it, I know majority of the userbase are on consumer GPUs so this is a good compromise for them, I don't know where to stop this behavior.
The other problem is if you set num_ctx PER chat, then the second message triggers the chat window title generation which uses a different context length, this causes the model to unload making the 3rd message slow before it stays. Could be a separate bug.
Desired Solution you'd like
Have a mode, like a admin setting or mode that simply always uses max context length.
With a large machine with a lot of vram, there's rarely a model that I have problems loading, its more the contrary where I might turn it down on specific models.
Alternatives Considered
just editing every model I have in the Admin -> Models tab and setting num_ctx there.
Additional Context
No response
@tjbck commented on GitHub (Jun 26, 2025):
Like you mentioned you can configure from the model editor level or from Ollama.