mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-05 18:38:17 -05:00
issue: Non-closing think section with latest openAI API backend with reasoning_parser #5618
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @jingyibo123 on GitHub (Jun 23, 2025).
Check Existing Issues
Installation Method
Git Clone
Open WebUI Version
0.6.15
Ollama Version (if applicable)
No response
Operating System
Ubuntu 20.04
Browser (if applicable)
No response
Confirmation
README.md.Expected Behavior
Expected behavior: show
reasoning_contentin the frontend folded think-section, and when reasoning_content is over and content starts, finish the think-section and show content outside..Actual Behavior
When working on think-enabled models such as qwen3 on latest openAI server deployed with lmdeploy, which returns first
reasoning_contentthencontentwithout</think>clouse, the open-webui frontend shows a ever-lasting thinking section including both the reasoning_content and the actual content.lmdeploy's API's behavior ( stream response):
vllm's API's behavior ( stream response):
without reasoning enabled:
Open-webuiworks OK when not enabling reasoning or enabled with vllm, NOT with deploy
Steps to Reproduce
Deploy mode with lmdeploy
Logs & Screenshots
No error in logs.
Additional Information
No response
@tjbck commented on GitHub (Jun 24, 2025):
This is a model level issue.
@jingyibo123 commented on GitHub (Jun 25, 2025):
@tjbck I'd like to know what is OpenWebUI's expected streaming response format for a think-enabled openAI-compatible chat API?