[GH-ISSUE #16289] issue: Laggy Reasoning Content Formatting #56514

Closed
opened 2026-05-05 19:34:26 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @yuchenxie4645 on GitHub (Aug 5, 2025).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/16289

For reasoning models like Qwen3, DeepSeek, and other models from API, long reasoning for hard problems like AIME and USACO problems results in slow formatting, and lag. Lag causes token streaming to be slowed down dramatically within UI. I was thinking maybe a reasoning stripper/summarizer that google and openai uses for their reasoning models? How do we implement it?

Originally created by @yuchenxie4645 on GitHub (Aug 5, 2025). Original GitHub issue: https://github.com/open-webui/open-webui/issues/16289 For reasoning models like Qwen3, DeepSeek, and other models from API, long reasoning for hard problems like AIME and USACO problems results in slow formatting, and lag. Lag causes token streaming to be slowed down dramatically within UI. I was thinking maybe a reasoning stripper/summarizer that google and openai uses for their reasoning models? How do we implement it?
GiteaMirror added the bug label 2026-05-05 19:34:26 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#56514