mirror of
https://github.com/open-webui/open-webui.git
synced 2026-05-07 03:18:23 -05:00
[GH-ISSUE #24341] feat: thinking_budget_tokens for llama.cpp #58938
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @alkeryn on GitHub (May 4, 2026).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/24341
Check Existing Issues
Verify Feature Scope
Problem Description
with llama.cpp it's possible to disable thinking through the api using : thinking_budget_tokens: 0, there may also be other methods that work with it.
anyway, currently there is no convenient builtin way to disable thinking for llama.cpp
it can be done per chat by adding "thinking_budget_tokens: 0" as a custom field in the chat settings, but there is no way to do it globaly.
Desired Solution you'd like
either add an option for llama.cpp in settings or allow custom fields not only in the chat settings but globaly.
Alternatives Considered
No response
Additional Context
No response
@owui-terminator[bot] commented on GitHub (May 4, 2026):
🔍 Similar Issues Found
I found some existing issues that might be related. Please check if any of these are duplicates or contain helpful solutions:
#23703 issue: Notes feature not compatible with llama.cpp, enable_thinking is always injected?
by TomTheWise ·
bug#8688 feat: built-in cpu-only llama cpp integration
by tjbck ·
enhancement,help wanted#17428 issue: Support Think Parsing with llama.cpp + GPT-OSS
by AbdullahMPrograms ·
bug#17350 issue: Llama.cpp server timing metrics not parsed correctly
by ITankForCAD ·
bug#16251 issue: When using llama.cpp as backend, pressing stop doesn't stop token generation
by OracleToes ·
bug💡 If this is a duplicate, consider closing it and adding details to the existing issue.
This comment was generated automatically. React with 👍 if helpful, 👎 if not.
@Classic298 commented on GitHub (May 4, 2026):
Filter?
@alkeryn commented on GitHub (May 4, 2026):
@Classic298 filter?
@Classic298 commented on GitHub (May 4, 2026):
this can be easily done with a filter, did you look into it? https://docs.openwebui.com/features/extensibility/plugin/functions/filter
@alkeryn commented on GitHub (May 4, 2026):
@Classic298 fair enough, though i do feel like it'd make sense as a builtin feature, we can already add custom args in the current chat, why couldn't we add them globaly in here?
@Classic298 commented on GitHub (May 4, 2026):
@alkeryn you can, scroll further down - you can add custom parameters. Open WebUI cannot had dozens and dozens more parameters for model or inference dependent engines, that's why you can add custom paramaters at the very end as well
@alkeryn commented on GitHub (May 4, 2026):
@Classic298 you can only for a single chat ie :
but the "add new custom setting" doesn't show up in the global config here :
@Classic298 commented on GitHub (May 4, 2026):
@alkeryn go to admin panel > models and add it to the model if your desire there in the advanced paramaters
@alkeryn commented on GitHub (May 4, 2026):
@Classic298 sure, but per model is annoying, would be nice to have a global setting or at least per connection / provider.
@Classic298 commented on GitHub (May 4, 2026):
@alkeryn admin panel > settings > models > top right: settings - this should also do it
@alkeryn commented on GitHub (May 4, 2026):
@Classic298 i don't see an option for it, weird.

@Classic298 commented on GitHub (May 4, 2026):
hmmmmmm open a feature request specifically for adding custom param IN THIS MODAL that you here opened so that you can add the same custom param to all models pls thx