[GH-ISSUE #21024] issue: Open WebUI does not distinguish the difference between image editing and image creation prompts. #58025

Closed
opened 2026-05-05 22:12:00 -05:00 by GiteaMirror · 20 comments
Owner

Originally created by @MattariOnline on GitHub (Jan 29, 2026).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/21024

Check Existing Issues

  • I have searched for any existing and/or related issues.
  • I have searched for any existing and/or related discussions.
  • I have also searched in the CLOSED issues AND CLOSED discussions and found no related items (your issue might already be addressed on the development branch!).
  • I am using the latest version of Open WebUI.

Installation Method

Docker

Open WebUI Version

v0.7.2

Ollama Version (if applicable)

0.15.1

Operating System

Windows 11

Browser (if applicable)

No response

Confirmation

  • I have read and followed all instructions in README.md.
  • I am using the latest version of both Open WebUI and Ollama.
  • I have included the browser console logs.
  • I have included the Docker container logs.
  • I have provided every relevant configuration, setting, and environment variable used in my setup.
  • I have clearly listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc).
  • I have documented step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation. My steps:
  • Start with the initial platform/version/OS and dependencies used,
  • Specify exact install/launch/configure commands,
  • List URLs visited, user input (incl. example values/emails/passwords if needed),
  • Describe all options and toggles enabled or changed,
  • Include any files or environmental changes,
  • Identify the expected and actual result at each stage,
  • Ensure any reasonably skilled user can follow and hit the same issue.

Expected Behavior

When requesting a new image, the image creation model (in this case, Z-image-Turbo) should be used.
When requesting changes or edits to an image, the image edit model should be used (such as Flux2-Klein).

Actual Behavior

When requesting image generation in a new conversation, it uses the image creation model.
For all subsequent image generation requests, even if requesting to "create a new picture" or "generate new picture," only the image edit model is used. There does not appear to be any manual override (ie. tool or button) to force the use of the image creation model over the edit model, aside from outright disabling the image edit model in the Admin settings.

Steps to Reproduce

The Token-Walkthrough for Environment Setup:

  1. Install Windows 11 Pro fresh on a computer.
  2. Install Ollama for Windows, ComfyUI Desktop, Docker Desktop, and docker pull the open-webui image.
  3. Install any desired LLM models (ie. Qwen 0.5b, Llama 3.2 3b q4, gpt-oss:20b; all of these were tested with this issue).
  4. Add and setup workflows in ComfyUI for Z-Image-Turbo and Flux.2 Klein 4B Distilled.
  5. Export as API JSON files, upload to Open WebUI under Admin Panel > Settings > Images after enabling Image Generation and Image Edit.image_flux2_klein_image_edit_4b_distilled_prompt-separation.json
    image_z_image_turbo.json
  6. Setup nodes and save.
    Image
  7. Go to Workspace, New Model, choose a model (see any above), enable the Capabilities for Vision, File Upload, File Context, Image Generation, Builtin Tools, and enable Default Features: Image Generation.
    The Important Part:
  8. Start a new conversation with your image model.
  9. Ask it to create an image, then observe if the creation or edit model was used; it should use the image creation model.
  10. Ask it to edit that image, or to make a new one, and observe it uses the image edit model now.
  11. Proceed to ask for as many new images as you'd like, each of which will be made with the edit model.
  12. Create a new conversation and ask to generate a picture; it will use the image creation model instead of the edit model.
  13. Ask for any changes or any other new images and, once again, it will only use the edit model henceforth for that conversation.

Example prompts used to generate a picture, edit it, and request an entirely new one:

  • Create a house on a hill.
  • Remake that picture in anime style.
  • Create a new picture: A small dragon sitting in a teacup, with the market streets of a fantasy town in the background. Anime style, no realism.

Logs & Screenshots

(No applicable browser console logs.)

Open WebUI Screenshot:
Image
I made three prompts: create an image, edit that image, then create a new image.
It would be expected that it uses the creation model (Z-image-Turbo), the edit model (Flux2 Klein), then the creation model again.

ComfyUI Job Queue (newest at top):
Image
As seen by the job queue, the creation model is used, then the edit model, then the edit model is subsequently used for that conversation. It will only revert back to the creation model for a new chat.

Additional Information

No response

Originally created by @MattariOnline on GitHub (Jan 29, 2026). Original GitHub issue: https://github.com/open-webui/open-webui/issues/21024 ### Check Existing Issues - [x] I have searched for any existing and/or related issues. - [x] I have searched for any existing and/or related discussions. - [x] I have also searched in the CLOSED issues AND CLOSED discussions and found no related items (your issue might already be addressed on the development branch!). - [x] I am using the latest version of Open WebUI. ### Installation Method Docker ### Open WebUI Version v0.7.2 ### Ollama Version (if applicable) 0.15.1 ### Operating System Windows 11 ### Browser (if applicable) _No response_ ### Confirmation - [x] I have read and followed all instructions in `README.md`. - [x] I am using the latest version of **both** Open WebUI and Ollama. - [x] I have included the browser console logs. - [x] I have included the Docker container logs. - [x] I have **provided every relevant configuration, setting, and environment variable used in my setup.** - [x] I have clearly **listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup** (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc). - [x] I have documented **step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation**. My steps: - Start with the initial platform/version/OS and dependencies used, - Specify exact install/launch/configure commands, - List URLs visited, user input (incl. example values/emails/passwords if needed), - Describe all options and toggles enabled or changed, - Include any files or environmental changes, - Identify the expected and actual result at each stage, - Ensure any reasonably skilled user can follow and hit the same issue. ### Expected Behavior When requesting a new image, the image creation model (in this case, Z-image-Turbo) should be used. When requesting changes or edits to an image, the image edit model should be used (such as Flux2-Klein). ### Actual Behavior When requesting image generation in a new conversation, it uses the image creation model. For all subsequent image generation requests, even if requesting to "create a new picture" or "generate new picture," only the image edit model is used. There does not appear to be any manual override (ie. tool or button) to force the use of the image creation model over the edit model, aside from outright disabling the image edit model in the Admin settings. ### Steps to Reproduce **The Token-Walkthrough for Environment Setup:** 1. Install Windows 11 Pro fresh on a computer. 2. Install Ollama for Windows, ComfyUI Desktop, Docker Desktop, and docker pull the open-webui image. 3. Install any desired LLM models (ie. Qwen 0.5b, Llama 3.2 3b q4, gpt-oss:20b; all of these were tested with this issue). 4. Add and setup workflows in ComfyUI for Z-Image-Turbo and Flux.2 Klein 4B Distilled. 5. Export as API JSON files, upload to Open WebUI under Admin Panel > Settings > Images after enabling Image Generation and Image Edit.[image_flux2_klein_image_edit_4b_distilled_prompt-separation.json](https://github.com/user-attachments/files/24926729/image_flux2_klein_image_edit_4b_distilled_prompt-separation.json) [image_z_image_turbo.json](https://github.com/user-attachments/files/24926728/image_z_image_turbo.json) 6. Setup nodes and save. <img width="426" height="890" alt="Image" src="https://github.com/user-attachments/assets/d10d5595-b50e-4fba-afff-3ddfe0f8210d" /> 7. Go to Workspace, New Model, choose a model (see any above), enable the Capabilities for Vision, File Upload, File Context, Image Generation, Builtin Tools, and enable Default Features: Image Generation. **The Important Part:** 8. Start a new conversation with your image model. 9. Ask it to create an image, then observe if the creation or edit model was used; it should use the image creation model. 10. Ask it to edit that image, or to make a new one, and observe it uses the image edit model now. 11. Proceed to ask for as many new images as you'd like, each of which will be made with the edit model. 12. Create a new conversation and ask to generate a picture; it will use the image creation model instead of the edit model. 13. Ask for any changes or any other new images and, once again, it will only use the edit model henceforth for that conversation. Example prompts used to generate a picture, edit it, and request an entirely new one: - Create a house on a hill. - Remake that picture in anime style. - Create a new picture: A small dragon sitting in a teacup, with the market streets of a fantasy town in the background. Anime style, no realism. ### Logs & Screenshots (No applicable browser console logs.) **Open WebUI Screenshot:** <img width="712" height="1211" alt="Image" src="https://github.com/user-attachments/assets/5b4d6890-7b15-43b7-b8f5-424843ce215a" /> I made three prompts: create an image, edit that image, then create a new image. It would be expected that it uses the creation model (Z-image-Turbo), the edit model (Flux2 Klein), then the creation model again. **ComfyUI Job Queue (newest at top):** <img width="375" height="340" alt="Image" src="https://github.com/user-attachments/assets/49a33439-1374-4a73-96f5-a7e856925262" /> As seen by the job queue, the creation model is used, then the edit model, then the edit model is subsequently used for that conversation. It will only revert back to the creation model for a new chat. ### Additional Information _No response_
GiteaMirror added the bug label 2026-05-05 22:12:00 -05:00
Author
Owner

@owui-terminator[bot] commented on GitHub (Jan 29, 2026):

🔍 Similar Issues Found

I found some existing issues that might be related to this one. Please check if any of these are duplicates or contain helpful solutions:

  1. #20237 issue: Image editing stopped working
    by tomasloksa • Dec 29, 2025 • bug, confirmed issue

  2. #20754 issue: New upload of an image does not work with image editing
    by iChristGit • Jan 17, 2026 • bug

  3. #19339 issue: Multiple input images for image edit and generation are passed as separate instances instead of as a list
    by eml-henn • Nov 21, 2025 • bug

  4. #19187 issue: Image generation menu gone.
    by calebrio02 • Nov 14, 2025 • bug

  5. #20091 issue: image is regarded as binary in temp chat
    by funnycups • Dec 22, 2025 • bug

Show 5 more related issues
  1. #19750 issue: Image Edit does not distinguish between User and Assistant images
    by eml-henn • Dec 04, 2025 • bug, confirmed issue

  2. #18726 issue: web search and image generation
    by allmazz • Oct 29, 2025 • bug

  3. #18995 issue: image generation and edition doesn’t work on temporary chats
    by futureshield • Nov 06, 2025 • bug

  4. #19825 Image Generation and Web Search trigger on every message
    by bcnation • Dec 08, 2025 • bug

  5. #19987 issue: There is a lack of visual consistency between the home page and the chat interface.
    by i-iooi-i • Dec 16, 2025 • bug


💡 Tips:

  • If this is a duplicate, please consider closing this issue and adding any additional details to the existing one
  • If you found a solution in any of these issues, please share it here to help others

This comment was generated automatically by a bot. Please react with a 👍 if this comment was helpful, or a 👎 if it was not.

<!-- gh-comment-id:3815320763 --> @owui-terminator[bot] commented on GitHub (Jan 29, 2026): 🔍 **Similar Issues Found** I found some existing issues that might be related to this one. Please check if any of these are duplicates or contain helpful solutions: 1. [#20237](https://github.com/open-webui/open-webui/issues/20237) **issue: Image editing stopped working** *by tomasloksa • Dec 29, 2025 • `bug`, `confirmed issue`* 2. [#20754](https://github.com/open-webui/open-webui/issues/20754) **issue: New upload of an image does not work with image editing** *by iChristGit • Jan 17, 2026 • `bug`* 3. [#19339](https://github.com/open-webui/open-webui/issues/19339) **issue: Multiple input images for image edit and generation are passed as separate instances instead of as a list** *by eml-henn • Nov 21, 2025 • `bug`* 4. [#19187](https://github.com/open-webui/open-webui/issues/19187) **issue: Image generation menu gone.** *by calebrio02 • Nov 14, 2025 • `bug`* 5. [#20091](https://github.com/open-webui/open-webui/issues/20091) **issue: image is regarded as binary in temp chat** *by funnycups • Dec 22, 2025 • `bug`* <details> <summary>Show 5 more related issues</summary> 6. [#19750](https://github.com/open-webui/open-webui/issues/19750) **issue: Image Edit does not distinguish between User and Assistant images** *by eml-henn • Dec 04, 2025 • `bug`, `confirmed issue`* 7. [#18726](https://github.com/open-webui/open-webui/issues/18726) **issue: web search and image generation** *by allmazz • Oct 29, 2025 • `bug`* 8. [#18995](https://github.com/open-webui/open-webui/issues/18995) **issue: image generation and edition doesn’t work on temporary chats** *by futureshield • Nov 06, 2025 • `bug`* 9. [#19825](https://github.com/open-webui/open-webui/issues/19825) **Image Generation and Web Search trigger on every message** *by bcnation • Dec 08, 2025 • `bug`* 10. [#19987](https://github.com/open-webui/open-webui/issues/19987) **issue: There is a lack of visual consistency between the home page and the chat interface.** *by i-iooi-i • Dec 16, 2025 • `bug`* </details> --- 💡 **Tips:** - If this is a duplicate, please consider closing this issue and adding any additional details to the existing one - If you found a solution in any of these issues, please share it here to help others *This comment was generated automatically by a bot.* Please react with a 👍 if this comment was helpful, or a 👎 if it was not.
Author
Owner

@Classic298 commented on GitHub (Jan 29, 2026):

Very closely related

https://github.com/open-webui/open-webui/issues/20754

<!-- gh-comment-id:3815756762 --> @Classic298 commented on GitHub (Jan 29, 2026): Very closely related https://github.com/open-webui/open-webui/issues/20754
Author
Owner

@MattariOnline commented on GitHub (Jan 31, 2026):

Very closely related

#20754

Potentially, however my primary issue is that Open WebUI does not understand when it needs to utilize the image generation tool rather than the image editing tool.

That said, I wouldn't rule out that it could be related to it fixating on editing the initial image; I'll be sure to watch that issue and test accordingly when a potential resolution is pushed for it. If it affects this, I'll be sure to note the observation and close if resolved.

Thanks for mentioning it!

<!-- gh-comment-id:3827588663 --> @MattariOnline commented on GitHub (Jan 31, 2026): > Very closely related > > [#20754](https://github.com/open-webui/open-webui/issues/20754) Potentially, however my primary issue is that Open WebUI does not understand when it needs to utilize the image generation tool rather than the image editing tool. That said, I wouldn't rule out that it could be related to it fixating on editing the initial image; I'll be sure to watch that issue and test accordingly when a potential resolution is pushed for it. If it affects this, I'll be sure to note the observation and close if resolved. Thanks for mentioning it!
Author
Owner

@gitFox117 commented on GitHub (Feb 3, 2026):

Image

When editing for the second time on the same page, this prompt appears—do you encounter this issue on your side?

""{"status": "success", "message": "The edited image has been successfully generated and is already visible to the user in the chat. You do not need to display or embed the image again - just acknowledge that it has been created.", "images": [{"url": "/api/v1/files/9f2e7b3a-4c1d-4e8f-b5a6-1d8c9e7f2a3b/content"}]}""

<!-- gh-comment-id:3837945843 --> @gitFox117 commented on GitHub (Feb 3, 2026): <img width="843" height="464" alt="Image" src="https://github.com/user-attachments/assets/a76619a0-d211-4496-b52d-5e7756f3bcc3" /> When editing for the second time on the same page, this prompt appears—do you encounter this issue on your side? "&quot;{&quot;status&quot;: &quot;success&quot;, &quot;message&quot;: &quot;The edited image has been successfully generated and is already visible to the user in the chat. You do not need to display or embed the image again - just acknowledge that it has been created.&quot;, &quot;images&quot;: [{&quot;url&quot;: &quot;/api/v1/files/9f2e7b3a-4c1d-4e8f-b5a6-1d8c9e7f2a3b/content&quot;}]}&quot;"
Author
Owner

@MattariOnline commented on GitHub (Feb 3, 2026):

When editing for the second time on the same page, this prompt appears—do you encounter this issue on your side?

""{"status": "success", "message": "The edited image has been successfully generated and is already visible to the user in the chat. You do not need to display or embed the image again - just acknowledge that it has been created.", "images": [{"url": "/api/v1/files/9f2e7b3a-4c1d-4e8f-b5a6-1d8c9e7f2a3b/content"}]}""

I haven't noticed anything like that, though honestly I reduce response times by removing the LLM response entirely.

(To be clear for anyone else looking into this issue as a whole; I have tested with several models with all default settings. When I'm not trying to troubleshoot image creation/editing, I use the zero-response setup for fast output, so this has nothing to do with the issue I'm reporting.)

<!-- gh-comment-id:3838122490 --> @MattariOnline commented on GitHub (Feb 3, 2026): > When editing for the second time on the same page, this prompt appears—do you encounter this issue on your side? > > ""{"status": "success", "message": "The edited image has been successfully generated and is already visible to the user in the chat. You do not need to display or embed the image again - just acknowledge that it has been created.", "images": [{"url": "/api/v1/files/9f2e7b3a-4c1d-4e8f-b5a6-1d8c9e7f2a3b/content"}]}"" I haven't noticed anything like that, though honestly I reduce response times by removing the LLM response entirely. (To be clear for anyone else looking into this issue as a whole; I have tested with several models with all default settings. When I'm _not_ trying to troubleshoot image creation/editing, I use the zero-response setup for fast output, so this has nothing to do with the issue I'm reporting.)
Author
Owner

@MattariOnline commented on GitHub (Feb 14, 2026):

Issue #20754 was reportedly fixed and closed with v0.8.0 about 20 hours ago, as of writing.

Tested the latest v0.8.1 and can confirm my issue still persists.

Explicitly asking it to "Create a new image of.." as suggested by the Open WebUI documentation does not trigger the use of the image creation API; it only continues to use the image editing API.

<!-- gh-comment-id:3902165871 --> @MattariOnline commented on GitHub (Feb 14, 2026): Issue #20754 was reportedly fixed and closed with v0.8.0 about 20 hours ago, as of writing. Tested the latest v0.8.1 and can confirm my issue still persists. Explicitly asking it to "Create a new image of.." as suggested by the Open WebUI documentation does _not_ trigger the use of the image creation API; it only continues to use the image editing API.
Author
Owner

@Classic298 commented on GitHub (Feb 14, 2026):

I can now say as of 0.8.0 and 0.8.1 the model can handle user input for edits, assistant input for edits AND assitant+user input for edits just fine if using the default built in image edit function with for example an OpenAI or Gemini model.

OpenRouter direct integrations does not work as well. There, sometimes the model does not use it's own previously generated image . OpenRouter specific quirk it seems

<!-- gh-comment-id:3902182026 --> @Classic298 commented on GitHub (Feb 14, 2026): I can now say as of 0.8.0 and 0.8.1 the model can handle user input for edits, assistant input for edits AND assitant+user input for edits just fine if using the default built in image edit function with for example an OpenAI or Gemini model. OpenRouter direct integrations does not work as well. There, sometimes the model does not use it's own previously generated image . OpenRouter specific quirk it seems
Author
Owner

@MattariOnline commented on GitHub (Feb 14, 2026):

I can now say as of 0.8.0 and 0.8.1 the model can handle user input for edits, assistant input for edits AND assitant+user input for edits just fine if using the default built in image edit function with for example an OpenAI or Gemini model.

OpenRouter direct integrations does not work as well. There, sometimes the model does not use it's own previously generated image . OpenRouter specific quirk it seems

Just in case the original post wasn't clear, it seems like you're focusing on edits, but the issue I'm reporting is it isn't possible to go from editing back to image creation within the same chat.
Editing works fine, but I'm concerned with returning to new/fresh image creation.

<!-- gh-comment-id:3902198822 --> @MattariOnline commented on GitHub (Feb 14, 2026): > I can now say as of 0.8.0 and 0.8.1 the model can handle user input for edits, assistant input for edits AND assitant+user input for edits just fine if using the default built in image edit function with for example an OpenAI or Gemini model. > > OpenRouter direct integrations does not work as well. There, sometimes the model does not use it's own previously generated image . OpenRouter specific quirk it seems Just in case the original post wasn't clear, it seems like you're focusing on edits, but the issue I'm reporting is _it isn't possible to go from editing back to image creation within the same chat._ Editing works fine, but I'm concerned with returning to new/fresh image creation.
Author
Owner

@Arjenlodder commented on GitHub (Mar 3, 2026):

Same issue here! Going back to creating a new image, results in editting the previous image.

<!-- gh-comment-id:3993425817 --> @Arjenlodder commented on GitHub (Mar 3, 2026): Same issue here! Going back to creating a **new** image, results in editting the previous image.
Author
Owner

@pfn commented on GitHub (Mar 9, 2026):

use native tool calling. this problem goes away.

<!-- gh-comment-id:4027174923 --> @pfn commented on GitHub (Mar 9, 2026): use native tool calling. this problem goes away.
Author
Owner

@MattariOnline commented on GitHub (Mar 9, 2026):

use native tool calling. this problem goes away.

That's not really a solution; at best, it's a work-around for non-local models or some cases with very large models.
Even the OWUI docs advise using things like GPT-5, Claude 4.5 Sonnet, etc. It does suggest large models such as Qwen 3 32B or Llama 3.3 70B could work, but advises results can vary significantly.

If it isn't possible to build in basic tool functionality to discern the difference between "Create an image" and "Edit an image," then there should at least be a button/toggle to facilitate the difference, such as "Edit previous image."

<!-- gh-comment-id:4027444099 --> @MattariOnline commented on GitHub (Mar 9, 2026): > use native tool calling. this problem goes away. That's not really a solution; at best, it's a work-around for non-local models or some cases with very large models. Even the OWUI docs advise using things like GPT-5, Claude 4.5 Sonnet, etc. It does suggest large models such as Qwen 3 32B or Llama 3.3 70B could work, but advises results can vary significantly. If it isn't possible to build in basic tool functionality to discern the difference between "Create an image" and "Edit an image," then there should at least be a button/toggle to facilitate the difference, such as "Edit previous image."
Author
Owner

@pfn commented on GitHub (Mar 9, 2026):

Native tool calling is the solution. default tool calling mode is a mistake, imo.

Clearly, this is the behavior you want:

Image
<!-- gh-comment-id:4027553072 --> @pfn commented on GitHub (Mar 9, 2026): Native tool calling *is* the solution. default tool calling mode is a mistake, imo. Clearly, this is the behavior you want: <img width="968" height="1965" alt="Image" src="https://github.com/user-attachments/assets/f79abea6-171e-4202-8f8e-32d31fcd4d97" />
Author
Owner

@MattariOnline commented on GitHub (Mar 10, 2026):

Native tool calling is the solution. default tool calling mode is a mistake, imo.

Well once figuring out how to force it to work, it's not bad for a solution, but definitely limits which models can work with it.
Thankfully, GPT-OSS:20b knows how to call both creation and editing, so that'll suit my needs. I still think a toggle for default (non-native) mode to "Edit previous image" or not would be a solid addition, especially if limited by VRAM. Thankfully Strix Halo doesn't care and will happily run both, but my old 24Gb GPU rig would've been touchier, and not all even have the benefit of a 24Gb dedicated GPU for an AI server.

<!-- gh-comment-id:4027727724 --> @MattariOnline commented on GitHub (Mar 10, 2026): > Native tool calling _is_ the solution. default tool calling mode is a mistake, imo. Well once figuring out how to force it to work, it's not bad for a solution, but definitely limits which models can work with it. Thankfully, GPT-OSS:20b knows how to call both creation and editing, so that'll suit my needs. I still think a toggle for default (non-native) mode to "Edit previous image" or not would be a solid addition, especially if limited by VRAM. Thankfully Strix Halo doesn't care and will happily run both, but my old 24Gb GPU rig would've been touchier, and not all even have the benefit of a 24Gb dedicated GPU for an AI server.
Author
Owner

@pfn commented on GitHub (Mar 10, 2026):

Just about no modern model worth using won't support native tool calling.

<!-- gh-comment-id:4027891990 --> @pfn commented on GitHub (Mar 10, 2026): Just about no modern model worth using won't support native tool calling.
Author
Owner

@MattariOnline commented on GitHub (Mar 10, 2026):

Just about no modern model worth using won't support native tool calling.

Subjective and definitely not true, as I've tested several decent models which didn't support it, but thankfully GPT-OSS:20b is supported which satisfies my needs.

<!-- gh-comment-id:4028098150 --> @MattariOnline commented on GitHub (Mar 10, 2026): > Just about no modern model worth using won't support native tool calling. Subjective and definitely not true, as I've tested several decent models which didn't support it, but thankfully GPT-OSS:20b is supported which satisfies my needs.
Author
Owner

@pfn commented on GitHub (Mar 10, 2026):

Subjective and definitely not true, as I've tested several decent models which didn't support it

considering that you didn't bother naming them, they're probably not worth using.

<!-- gh-comment-id:4028304346 --> @pfn commented on GitHub (Mar 10, 2026): > Subjective and definitely not true, as I've tested several decent models which didn't support it considering that you didn't bother naming them, they're probably not worth using.
Author
Owner

@MattariOnline commented on GitHub (Mar 10, 2026):

Subjective and definitely not true, as I've tested several decent models which didn't support it

considering that you didn't bother naming them, they're probably not worth using.

A bit rude for github, much? Well it was several Llama 3.1/3.2 variants, Granite 4 iirc, some Qwen models, but I didn't have exact specifics and I'm currently clocking a high fever with a flu so I didn't want to do all the proper testing again so as to accurately reference them all.

Feel free to test which models do/don't work and report back; I'll concede to anything that works. Otherwise I'm gonna go back to fighting my flu thanks.

<!-- gh-comment-id:4028469541 --> @MattariOnline commented on GitHub (Mar 10, 2026): > > Subjective and definitely not true, as I've tested several decent models which didn't support it > > considering that you didn't bother naming them, they're probably not worth using. A bit rude for github, much? Well it was several Llama 3.1/3.2 variants, Granite 4 iirc, some Qwen models, but I didn't have exact specifics and I'm currently clocking a high fever with a flu so I didn't want to do all the proper testing __again__ so as to accurately reference them all. Feel free to test which models do/don't work and report back; I'll concede to anything that works. Otherwise I'm gonna go back to fighting my flu thanks.
Author
Owner

@pfn commented on GitHub (Mar 10, 2026):

out of the recent models that I've tried:

yes:

  • GLM 4.7 (and Flash)
  • GLM 4.6V
  • GLM 4.5 Air
  • Qwen3 & 3.5 (all varieties, VL, Coder, Next)
  • gpt-oss-120b
  • Nemotron
  • Step 3.5
  • MiniMax M2.5

no:

  • llama 3.3

take care of your flu

<!-- gh-comment-id:4028575229 --> @pfn commented on GitHub (Mar 10, 2026): out of the recent models that I've tried: yes: * GLM 4.7 (and Flash) * GLM 4.6V * GLM 4.5 Air * Qwen3 & 3.5 (all varieties, VL, Coder, Next) * gpt-oss-120b * Nemotron * Step 3.5 * MiniMax M2.5 no: * llama 3.3 take care of your flu
Author
Owner

@MattariOnline commented on GitHub (Mar 10, 2026):

out of the recent models that I've tried:

yes:

  • GLM 4.7 (and Flash)
  • GLM 4.6V
  • GLM 4.5 Air
  • Qwen3 & 3.5 (all varieties, VL, Coder, Next)
  • gpt-oss-120b
  • Nemotron
  • Step 3.5
  • MiniMax M2.5

no:

  • llama 3.3

take care of your flu

Ty sir, the fever finally broke just barely. Also weird that Llama 3.3 is a dud. Tried Claude 3.5 32b iirc, but weirdly Claude A3O 30b (or something like that) starts thinking..weirdly.

<!-- gh-comment-id:4030623536 --> @MattariOnline commented on GitHub (Mar 10, 2026): > out of the recent models that I've tried: > > yes: > > * GLM 4.7 (and Flash) > * GLM 4.6V > * GLM 4.5 Air > * Qwen3 & 3.5 (all varieties, VL, Coder, Next) > * gpt-oss-120b > * Nemotron > * Step 3.5 > * MiniMax M2.5 > > no: > > * llama 3.3 > > take care of your flu Ty sir, the fever finally broke just barely. Also weird that Llama 3.3 is a dud. Tried Claude 3.5 32b iirc, but weirdly Claude A3O 30b (or something like that) starts thinking..weirdly.
Author
Owner

@Classic298 commented on GitHub (Apr 14, 2026):

native tool calling should be used, default tool calling is legacy and not reliable in this case and cannot fully detect all cases since its automated detection and cannot distinguish based on what the user wants. native tool calling can because the user decides.

Similar question was posted to discussion a month ago and was closed with this reason. Documentation has since long been updated to notify default tool calling as unrecommended and legacy. Native should be used

<!-- gh-comment-id:4244398406 --> @Classic298 commented on GitHub (Apr 14, 2026): native tool calling should be used, default tool calling is legacy and not reliable in this case and cannot fully detect all cases since its automated detection and cannot distinguish based on what the user wants. native tool calling can because the user decides. Similar question was posted to discussion a month ago and was closed with this reason. Documentation has since long been updated to notify default tool calling as unrecommended and legacy. Native should be used
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#58025