mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-07 03:18:23 -05:00
[GH-ISSUE #3296] chat request timesout after 60s, and context lenght not detected automatically #51880
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @amida47 on GitHub (Jun 19, 2024).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/3296
Bug Report
Description
Bug Summary:
on OpenWebui when I try long context, the chat request timeout after on 1 min
Steps to Reproduce:
create a GitHub codespace with 4-cores
install ollama through
curl -fsSL https://ollama.com/install.sh | shrun
ollama serverun
ollama run phi3:14b-medium-128k-instruct-q4_0run
docker run -d --network=host -v open-webui:/app/backend/data -e OLLAMA_BASE_URL=http://127.0.0.1:11434 --name open-webui --restart always ghcr.io/open-webui/open-webui:maingo to openwebui
test connection to ollama by running simple prompt
upload long .tex document and ask to summarize
the request to /chat timesout after 60 s
Expected Behavior:
start streaming as normal
Actual Behavior:
the request to /chat timesout after 60 s
Environment
Open WebUI Version: latest as of 06/19/2024
Ollama (if applicable): ollama version is 0.1.44
Operating System: [e.g., Windows 10, macOS Big Sur, Ubuntu 20.04]
Browser (if applicable): [e.g., Chrome 100.0, Firefox 98.0]
Reproduction Details
Confirmation:
Logs and Screenshots
https://github.com/open-webui/open-webui/assets/49212088/b0b98f79-cd2a-4f7a-b0a3-c750ee5d384a
another example, it started streaming but once the request time reaches 60s it failed

the logs for this screenshot , you can see the context lenght ollama used 2048 while phi model has 128k in context so openwebui didn't read this context