mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-08 04:16:03 -05:00
[GH-ISSUE #24144] issue: LLM response constantly gets "stuck" on most models for most queries with tools #58875
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @peterwwillis on GitHub (Apr 26, 2026).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/24144
Check Existing Issues
Installation Method
Docker
Open WebUI Version
v0.9.2
Ollama Version (if applicable)
No response
Operating System
Ubuntu 24.04
Browser (if applicable)
Firefox 150
Confirmation
README.md.Expected Behavior
Tools run, AI continues to output tokens to completion. (This doesn't happen... see below)
Actual Behavior
I use different LLM providers: OpenCode Go, OpenRouter, OpenAI, etc. I also use various models, mainly latest MiniMax, Qwen, GLM, Kimi, and some GPT.
Web search is Searxng running in a local Docker.
I'll have a chat open and enable web search and code execution (and I make sure native tool calling is selected).
Randomly, probably 25% of the time, if the model runs a web search or code execution, the model just stops outputting tokens after running those tools. I'll then send "did you get stuck?" or something and suddenly it's replying again, but of course it skips the previous tool call info.
This has been going on for months. There's no debugging information for me to tell what is going wrong. Basically OpenWebUI is randomly broken a quarter of the time.
This doesn't happen if I don't select the tools.
Steps to Reproduce
Logs & Screenshots
Conversations so far have been private. If you can provide me with a way to generate some kind of useful log, I'll make more attempts and gather the logs.
Additional Information
This has been going on for 2 months. I really want to find an alternative to OpenWebUI so I can actually use the functionality.
@belugaming commented on GitHub (Apr 26, 2026):
The same problem, especially when calling the web search tool multiple times, the web page will freeze, even the Mac Studio M1 ultra 64G RAM will have a serious lag in the LLM search of 100 websites.
@silenceroom commented on GitHub (Apr 28, 2026):
had exactly same problem here. I was guessing it might related to the networking issue or something.
It seems that the model just gets stuck after execute the web_search for a round or tying to call a complicated skill to do something. Some modesl if you click the "continue" button it'll keep runnnig, but most of the times it just repeat what it had done and gets stuck again.
@Classic298 commented on GitHub (May 2, 2026):
I have never experienced this and I also use OpenAI and OpenRouter and Kimi K2.5 on native tool calling. Have used it all. I never experienced this.
And there's no reason for why this should be happening. What's your precise setup? What's your aiohttp timeout if you configured any? What's your network connectivity to open webui but also from open webui the Provider look like?
How can i reproduce this?
Are you maybe just silently running out of context?