mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-08 04:16:03 -05:00
[GH-ISSUE #23256] issue: Reranker: raw prompt deprecation warning with vLLM ≥ 0.18 (InputProcessor raw prompts removed) #58599
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @danialkhatib on GitHub (Mar 31, 2026).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/23256
Check Existing Issues
Installation Method
Docker
Open WebUI Version
v0.8.12
Ollama Version (if applicable)
No response
Operating System
Red Hat Enterprise Linux release 9.6
Browser (if applicable)
No response
Confirmation
README.md.Expected Behavior
Bug Report: Reranker sends raw prompts to vLLM
InputProcessor(deprecated in v0.18)Labels:
bug,rag / reranker,vllmEnvironment
Expected Behavior
Rerank requests from Open WebUI should be formatted in a way that is compatible with vLLM's
InputProcessor. Prompts should be passed through vLLM'sRenderer.render_cmpl()orRenderer.render_chat()pipeline before being submitted to the/v1/rerankendpoint, producing no deprecation warnings and ensuring forward compatibility with vLLM ≥ 0.18.Actual Behavior
On every rerank request, vLLM 0.18.1 logs the following deprecation warning:
Open WebUI is sending raw prompt strings directly to
/v1/rerankinstead of using the rendered prompt format that vLLM now expects. Requests currently return200 OK, but this is likely relying on a fallback that may be fully removed in a future patch, which would break reranking silently or with an unhandled error.Relevant vLLM Logs
Additional Context
The fix needs to be applied in Open WebUI's reranker HTTP client code — specifically wherever it constructs the request body for
POST /v1/rerank. The prompt strings need to be passed through vLLM'sRendererbefore being submitted, or the payload format needs to match what vLLM'sInputProcessornow expects as a non-raw-prompt input.vLLM issue reference: https://github.com/vllm-project/vllm/blob/main/vllm/v1/engine/input_processor.py
Steps to Reproduce
Steps to Reproduce
--runner poolingand modelzeroentropy/zerank-2on vLLM 0.18.1/v1/rerankInputProcessordeprecation warning in vLLM logs on every requestLogs & Screenshots
Additional Information
Vllm Config