mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-06 10:58:17 -05:00
[GH-ISSUE #5600] num_predict slider max value too small. #14046
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @JamesClarke7283 on GitHub (Sep 22, 2024).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/5600
Bug Report
Installation Method
Docker
Environment
Open WebUI Version: v0.3.23
Ollama (if applicable): 0.3.10
Operating System: ArchLinux
Browser (if applicable): Zen 1.0.1-a.2 (Fork of Firefox)
Confirmation:
Expected Behavior:
Able to set
num_predictto at most65536tokens.Actual Behavior:
I can only move the slider to
16000Description
Bug Summary:
There are now models that support higher
num_predictamounts, like65536Reproduction Details
Steps to Reproduce:
Go to the Chat Window or Model Editor, and in advanced parameters, try and change
num_predictto a number higher than16000, it wont do it.Logs and Screenshots
N/A
Additional Information
This is the highest number i have seen(where the num_predict could fill the whole context window):
https://openrouter.ai/models/meta-llama/llama-3.1-405b-instruct/providers
This is the more likely case:
https://openrouter.ai/models/openai/o1-mini/providers
Soon, seeing a max output token limit of
65536will be commonplace.Even standardly used models, go slightly over the
16000limit, like16384:https://platform.openai.com/docs/models/gpt-4o
Conclusion
I think a upper limit of
65536or128000would be good(the former, being conservative and meeting most models of today, the later, covering the edge cases like llama3.1, and also future models).