[Feature] Continue inference while switching between chats. #715

Closed
opened 2025-11-11 14:29:40 -06:00 by GiteaMirror · 0 comments
Owner

Originally created by @imkebe on GitHub (Apr 24, 2024).

While locally hosted LLM's are not as fast as cloud ones, and there is an option to provide multiple independent hosting endpoints I would like to have two or more (might be able to configure it) active chats. Currently while I send a query I need to stay in the context of the conversation to finish it. When I switch to another one i lost the current context of the ongoing one. When it finishes it often mess up the conversation. There should be an pending icon in the left chat list if the query was send and the inference is ongoing.

Originally created by @imkebe on GitHub (Apr 24, 2024). While locally hosted LLM's are not as fast as cloud ones, and there is an option to provide multiple independent hosting endpoints I would like to have two or more (might be able to configure it) active chats. Currently while I send a query I need to stay in the context of the conversation to finish it. When I switch to another one i lost the current context of the ongoing one. When it finishes it often mess up the conversation. There should be an pending icon in the left chat list if the query was send and the inference is ongoing.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#715