mirror of
https://github.com/open-webui/open-webui.git
synced 2026-03-12 01:54:38 -05:00
BUG No completion from Ollama-served model proxied behind LiteLLM #751
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @corticalstack on GitHub (Apr 28, 2024).
Bug Report
Description
Configured a proxied LiteLLM model in open-webui. Cannot get response completion from the model, UI visualisation just shows a blining awaiting response.
Note I do get compeltion from said model exposed via LiteLLM when inferring LiteLLM directly, with the example curl below:
When configuring the ollama/llama3 model in open-webui LiteLLM settings, I have tried both
http://192.168.1.12:4000andhttp://192.168.1.12:4000/chat/completionsas api base urls.Also note I do get completioned from other LLM endpoints proxied behind LiteLLM sich as OpenAI GPT-3.5-Turbo and Groq Llama3-70B.
Thanks for any help.
Environment
Open WebUI Version: [e.g., 0.1.121]
**Ollama 0.1.32
**LiteLLM 1.35.29 OAS 3.1
**Ubuntu 22.04.3 LTS
Operating System: [e.g., Windows 10, macOS Big Sur, Ubuntu 20.04]
Browser (if applicable): [e.g., Chrome 100.0, Firefox 98.0]
@justinh-rahb commented on GitHub (Apr 28, 2024):
Base URLs should be that of your Ollama server, not the LiteLLM proxy:
@corticalstack commented on GitHub (Apr 28, 2024):
That's defeating my objective, which is to have EVERY LLM made available for seelction behind a proxy, for purposes of load balancing, capturing token consumption etc.
If I wanted open-webui to have direct access to ollama runner, I'd just configure the OLLAMA_BASE_URL.