[GH-ISSUE #14728] /v1/chat/completions not working with qwen3.5 multimodal model #35284

Closed
opened 2026-04-22 19:40:54 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @sweihub on GitHub (Mar 9, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/14728

What is the issue?

Example:
https://docs.ollama.com/api/openai-compatibility#/v1/chat/completions-with-vision-example

Not working with qwen3.5:27b which is a multimodal LLM.

[2026-03-09 16:03:19.776] [DEBUG] [ureq::stream:395] connecting to localhost:11434 at 127.0.0.1:11434
[2026-03-09 16:03:19.776] [DEBUG] [ureq::stream:202] created stream: Stream(TcpStream { addr: 127.0.0.1:57784, peer: 127.0.0.1:11434, fd: 3 })
[2026-03-09 16:03:19.776] [DEBUG] [ureq::unit:261] sending request POST http://localhost:11434/v1/chat/completions
[2026-03-09 16:03:19.776] [DEBUG] [ureq::unit:480] writing prelude: POST /v1/chat/completions HTTP/1.1
Host: localhost:11434
User-Agent: ureq/2.12.1
Accept: */*
Content-Type: application/json
accept-encoding: gzip
Content-Length: 3783642
[2026-03-09 16:03:45.965] [DEBUG] [ureq::response:396] Body entirely buffered (length: 206)
[2026-03-09 16:03:45.965] [DEBUG] [ureq::pool:130] adding stream to pool: http|localhost|11434 -> Stream(TcpStream { addr: 127.0.0.1:57784, peer: 127.0.0.1:11434, fd: 3 })
[2026-03-09 16:03:45.965] [DEBUG] [ureq::unit:314] response 500 to POST http://localhost:11434/v1/chat/completions
[2026-03-09 16:03:45.965] [DEBUG] [ureq::stream:322] dropping stream: Stream(TcpStream { addr: 127.0.0.1:57784, peer: 127.0.0.1:11434, fd: 3 })
Error: Failed to OCR page 1

Caused by:
    0: Failed to connect to VML server at http://localhost:11434/v1
    1: http://localhost:11434/v1/chat/completions: status code 500

Relevant log output


OS

Ubuntu 22.04.4 LTS

GPU

4 x A100

CPU

No response

Ollama version

ollama version is 0.17.4

Originally created by @sweihub on GitHub (Mar 9, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/14728 ### What is the issue? Example: https://docs.ollama.com/api/openai-compatibility#/v1/chat/completions-with-vision-example Not working with `qwen3.5:27b` which is a multimodal LLM. ``` [2026-03-09 16:03:19.776] [DEBUG] [ureq::stream:395] connecting to localhost:11434 at 127.0.0.1:11434 [2026-03-09 16:03:19.776] [DEBUG] [ureq::stream:202] created stream: Stream(TcpStream { addr: 127.0.0.1:57784, peer: 127.0.0.1:11434, fd: 3 }) [2026-03-09 16:03:19.776] [DEBUG] [ureq::unit:261] sending request POST http://localhost:11434/v1/chat/completions [2026-03-09 16:03:19.776] [DEBUG] [ureq::unit:480] writing prelude: POST /v1/chat/completions HTTP/1.1 Host: localhost:11434 User-Agent: ureq/2.12.1 Accept: */* Content-Type: application/json accept-encoding: gzip Content-Length: 3783642 [2026-03-09 16:03:45.965] [DEBUG] [ureq::response:396] Body entirely buffered (length: 206) [2026-03-09 16:03:45.965] [DEBUG] [ureq::pool:130] adding stream to pool: http|localhost|11434 -> Stream(TcpStream { addr: 127.0.0.1:57784, peer: 127.0.0.1:11434, fd: 3 }) [2026-03-09 16:03:45.965] [DEBUG] [ureq::unit:314] response 500 to POST http://localhost:11434/v1/chat/completions [2026-03-09 16:03:45.965] [DEBUG] [ureq::stream:322] dropping stream: Stream(TcpStream { addr: 127.0.0.1:57784, peer: 127.0.0.1:11434, fd: 3 }) Error: Failed to OCR page 1 Caused by: 0: Failed to connect to VML server at http://localhost:11434/v1 1: http://localhost:11434/v1/chat/completions: status code 500 ``` ### Relevant log output ```shell ``` ### OS Ubuntu 22.04.4 LTS ### GPU 4 x A100 ### CPU _No response_ ### Ollama version ollama version is 0.17.4
GiteaMirror added the bug label 2026-04-22 19:40:54 -05:00
Author
Owner

@rick-github commented on GitHub (Mar 9, 2026):

Server logs will aid in debugging.

<!-- gh-comment-id:4021905702 --> @rick-github commented on GitHub (Mar 9, 2026): [Server logs](https://docs.ollama.com/troubleshooting) will aid in debugging.
Author
Owner

@sweihub commented on GitHub (Mar 9, 2026):

HTTP 500 means my ollama server is too busy, I tested again, it worked, so close as not-a-bug!

<!-- gh-comment-id:4022322844 --> @sweihub commented on GitHub (Mar 9, 2026): HTTP 500 means my ollama server is too busy, I tested again, it worked, so close as not-a-bug!
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#35284