mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-06 10:58:17 -05:00
[GH-ISSUE #22861] issue: OWUI Ollama proxy does not support anthropic compatibility endpoints #58495
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @Umutayb on GitHub (Mar 19, 2026).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/22861
Check Existing Issues
Installation Method
Docker
Open WebUI Version
0.8.10
Ollama Version (if applicable)
0.18.2
Operating System
MacOS, Ubuntu
Browser (if applicable)
No response
Confirmation
README.md.Expected Behavior
The ollama proxy api spec(request validations) should match the description @ https://docs.ollama.com/api/anthropic-compatibility
curl -X POST "http://localhost:11435/v1/messages" \ -H "Content-Type: application/json" \ -d '{ "model": "qwen3.5:35b", "max_tokens": 1024, "messages": [ { "role": "user", "content": "Hello!" } ] }' {"id":"msg_f639a3a98b1fb905507e9704","type":"message","role":"assistant","model":"qwen3.5:35b","content":[{"type":"thinking","thinking":"Thinking Process:\n\n1. **Analyze the Input:**\n * Input: \"Hello!\"\n * Intent: Greeting.\n * Tone: Friendly, casual.\n * Context: Start of a conversation.\n\n2. **Determine the appropriate response:**\n * Acknowledge the greeting.\n * Offer assistance.\n * Maintain a friendly and helpful tone.\n * Keep it concise but open-ended.\n\n3. **Draft potential responses:**\n * *Option 1:* Hello! How can I help you today?\n * *Option 2:* Hi there! Hope you're having a great day. What's on your mind?\n * *Option 3:* Hello! Nice to meet you. How can I assist you?\n * *Option 4:* Hey! How's it going?\n\n4. **Select the best option:** Option 1 or 2 is usually best for an AI assistant. Let's go with a friendly greeting + offer of help.\n\n5. **Refine the response:** \"Hello! How can I help you today?\" is clear and direct. Let's add a bit more warmth. \"Hello! It's nice to meet you. How can I assist you today?\"\n\n6. **Final Polish:** Keep it simple. \"Hello! How can I help you today?\" is standard and effective. Or \"Hi there! How's your day going?\" Let's combine friendliness with utility.\n\n *Selected:* \"Hello! How can I help you today?\" (Simple, effective).\n *Alternative:* \"Hi! It's great to hear from you. What can I do for you today?\"\n\n Let's go with a warm, friendly greeting that invites further interaction.\n\n7. **Final Output Generation:** \"Hello! How can I help you today?\" (Or similar). I'll make it slightly more engaging.\n\n \"Hello! It's great to hear from you. How can I assist you today?\"\n\n Actually, keeping it punchy is good. \"Hello! How are you doing today? Is there something specific you'd like to chat about or get help with?\"\n\n Let's settle on: \"Hello! How can I help you today?\" (Safe, standard).\n Wait, I can be a bit more conversational.\n \"Hello! 👋 How's your day going? Is there anything I can help you with?\"\n\n Okay, final decision: Friendly greeting + offer of help.\n\n *Draft:* \"Hello! How can I assist you today?\"\n\n *Revised:* \"Hello! It's nice to meet you. How can I help you today?\"\n\n Let's go with a balanced approach.\n\n *Final Choice:* \"Hello! How can I help you today?\" (Simple and direct).\n Actually, adding an emoji or a bit of warmth is nice. \"Hello! 👋 How can I help you today?\"\n\n Let's just output the text.\n\n *Self-Correction during thought:* The user just said \"Hello!\". I should respond in kind.\n\n Response: \"Hello! How can I help you today?\"\n\n Wait, I should check if there are any specific constraints. None.\n\n Okay, ready to output.cw"},{"type":"text","text":"Hello! 👋 How can I help you today?"}],"stop_reason":"end_turn","usage":{"input_tokens":12,"output_tokens":723}}Actual Behavior
api responds with 405 method not allowed
curl -X POST "http://localhost:3000/ollama/v1/messages" \ -H "Content-Type: application/json" \ -d '{ "model": "qwen3.5:35b", "max_tokens": 1024, "messages": [ { "role": "user", "content": "Hello!" } ] }' {"detail":"Method Not Allowed"}Steps to Reproduce
Start OWUI via docker.
use /ollama proxy to call ollama api for
Anthropic compatibilityas described in https://docs.ollama.com/api/anthropic-compatibility.Logs & Screenshots
2026-03-19 16:57:52.250 | INFO | uvicorn.protocols.http.httptools_impl:send:483 - *********** - "POST /ollama/v1/messages HTTP/1.1" 405Additional Information
No response
@tjbck commented on GitHub (Mar 20, 2026):
Addressed in dev.
@Umutayb commented on GitHub (Mar 21, 2026):
Would be nice to know the commit id to track it @tjbck
@Umutayb commented on GitHub (Mar 21, 2026):
for future reference:
f23296b22d