mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-06 10:58:17 -05:00
[PR #22556] [CLOSED] fix: use max_completion_tokens for newer OpenAI models #26755
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
📋 Pull Request Information
Original PR: https://github.com/open-webui/open-webui/pull/22556
Author: @alvinttang
Created: 3/11/2026
Status: ❌ Closed
Base:
dev← Head:fix/max-completion-tokens📝 Commits (10+)
fe6783cMerge pull request #19030 from open-webui/devfc05e0aMerge pull request #19405 from open-webui/deve3faec6Merge pull request #19416 from open-webui/dev9899293Merge pull request #19448 from open-webui/dev140605eMerge pull request #19462 from open-webui/dev6f1486fMerge pull request #19466 from open-webui/devd95f533Merge pull request #19729 from open-webui/deva7271530.6.43 (#20093)6adde20Merge pull request #20394 from open-webui/devf9b0534Merge pull request #20522 from open-webui/dev📊 Changes
1 file changed (+42 additions, -20 deletions)
View changed files
📝
backend/open_webui/routers/openai.py(+42 -20)📄 Description
Summary
gpt-4.1,gpt-4.1-mini,gpt-4.1-nano,gpt-5,gpt-5.2) reject the deprecatedmax_tokensparameter and requiremax_completion_tokensinstead, causing 400 errors during title/emoji/task generation.is_openai_new_api_model()to match all models that use the newer API conventions (o-series + gpt-4.1+ + gpt-5+), while keepingis_openai_reasoning_model()for reasoning-specific behavior (system→developer role conversion, temperature restrictions)._extract_model_name()helper to handle provider-prefixed model names (e.g.openai/gpt-5.2from LiteLLM proxies).Fixes #22524
Changes
backend/open_webui/routers/openai.py:_extract_model_name(): strips provider prefixes (openai/gpt-5→gpt-5)is_openai_new_api_model(): matcheso1,o3,o4,gpt-4.1,gpt-5familiesopenai_reasoning_model_handler→openai_new_api_model_handler: appliesmax_completion_tokensconversion to all newer models, but only applies system→developer role conversion for reasoning modelsconvert_to_azure_payloadto useis_openai_new_api_modelfor token param conversion, while keeping temperature restriction for reasoning models onlygenerate_chat_completionto use the new broader model checkTest plan
gpt-5.2as task model (no 400 error)gpt-4.1-minias task modelgpt-4o(usesmax_tokens)o3,o4-mini) still get system→developer role conversionmax_tokensopenai/gpt-5.2) works correctlyContributor License Agreement
🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.