mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-06 19:08:59 -05:00
[GH-ISSUE #11284] issue: Cannot correctly handle the think tag of qwq model #54837
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @qhjqhj on GitHub (Mar 6, 2025).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/11284
Check Existing Issues
Installation Method
Git Clone
Open WebUI Version
0.5.20
Ollama Version (if applicable)
0.5.13
Operating System
windows 11
Browser (if applicable)
No response
Confirmation
README.md.Expected Behavior
Separate thinking from formal output
Actual Behavior
The content of thinking and the content of formal output cannot be distinguished, and the formal content cannot be output
Steps to Reproduce
When running the qwq:latest model of Ollama on the backend, the content in the tag is output as normal content, and part of the code output seems to be lost
Logs & Screenshots
Additional Information
No response
@Eidenz commented on GitHub (Mar 6, 2025):
Adding to this, it doesn't even seem to have its normal output in the thinking tag.
On my end, everything in the thinking tag seems to indeed be the thought process, but once it's done it outputs no answer, and openwebui acts as if the model was still thinking.
@mihkhub commented on GitHub (Mar 6, 2025):
QwQ-32b thinking tag pair "
@EntropyYue commented on GitHub (Mar 6, 2025):
It seems that the model's thinking output has been stopped before it has finished, please try increasing
nun_ctxandnum_predict@Duxon commented on GitHub (Mar 6, 2025):
How to increase
num_ctx(to 131072) in the Open-WebUI settings? I can't find that seeting.@EntropyYue commented on GitHub (Mar 6, 2025):
If you are using en-US, it will display as
Context Length@qhjqhj commented on GitHub (Mar 6, 2025):
I tried it several times, and the total number of tokens generated was different. It seems that the issue is not caused by the maximum length limit.
@bjj commented on GitHub (Mar 6, 2025):
See #11259
@EntropyYue commented on GitHub (Mar 6, 2025):
Please provide screenshots of other situations, which will help us better solve the problem