issue: Chat no longer works after an update #4497

Closed
opened 2025-11-11 15:55:27 -06:00 by GiteaMirror · 0 comments
Owner

Originally created by @mistrjirka on GitHub (Mar 19, 2025).

Check Existing Issues

  • I have searched the existing issues and discussions.
  • I am using the latest version of Open WebUI.

Installation Method

Git Clone

Open WebUI Version

v0.5.20

Ollama Version (if applicable)

ollama version is 0.6.1

Operating System

Arch linux

Browser (if applicable)

Firefox

Confirmation

  • I have read and followed all instructions in README.md.
  • I am using the latest version of both Open WebUI and Ollama.
  • I have included the browser console logs.
  • I have included the Docker container logs.
  • I have listed steps to reproduce the bug in detail.

Expected Behavior

When I send message to AI I expect it to respond.

Actual Behavior

Ai never responds just acts as if it was loading.

Image

Steps to Reproduce

Update to latest version of openwebui and ollama and try to use the openwebui.

Logs & Screenshots

I looked through the logs and I cannot find anything that would suggest something broken. The Website is infinitely waiting on post request "new"
This is the request that is infinitely loading as cURL:
curl 'http://10.154.224.4:8080/api/v1/chats/new' -X POST -H 'User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:136.0) Gecko/20100101 Firefox/136.0' -H 'Accept: application/json' -H 'Accept-Language: cs,sk;q=0.8,en-US;q=0.5,en;q=0.3' -H 'Accept-Encoding: gzip, deflate' -H 'Referer: http://10.154.224.4:8080/' -H 'Content-Type: application/json' -H 'authorization: Bearer eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpZCI6IjBmOWMyZWQ0LTBiYjEtNDU4Ny04NjJiLWEzZjNmZDg1NDE1NyJ9.fcNihgrjUO8QZhHYPAvv9VJ0OlM5Csf95BGvpoc83Uc' -H 'Origin: http://10.154.224.4:8080' -H 'Connection: keep-alive' -H 'Cookie: token=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpZCI6IjBmOWMyZWQ0LTBiYjEtNDU4Ny04NjJiLWEzZjNmZDg1NDE1NyJ9.fcNihgrjUO8QZhHYPAvv9VJ0OlM5Csf95BGvpoc83Uc' -H 'Priority: u=0' -H 'Pragma: no-cache' -H 'Cache-Control: no-cache' --data-raw $'{"chat":{"id":"","title":"Nov\xfd chat","models":["deepseek-r1:1.5b"],"params":{},"history":{"messages":{"7d5009b8-b990-41d0-a75e-2f406db93e8d":{"id":"7d5009b8-b990-41d0-a75e-2f406db93e8d","parentId":null,"childrenIds":[],"role":"user","content":"Jak se vede. Um\xed\u0161 \u010desky?","timestamp":1742412270,"models":["deepseek-r1:1.5b"]}},"currentId":"7d5009b8-b990-41d0-a75e-2f406db93e8d"},"messages":[{"id":"7d5009b8-b990-41d0-a75e-2f406db93e8d","parentId":null,"childrenIds":[],"role":"user","content":"Jak se vede. Um\xed\u0161 \u010desky?","timestamp":1742412270,"models":["deepseek-r1:1.5b"]}],"tags":[],"timestamp":1742412270120}}'

open-webui.log

ollama.log

console-export-2025-3-19_20-41-48.txt

Additional Information

I tested that ollama works when I do ollama run.
I am currently tunneled to the machine using wireguard vpn and I was not sure if I should use the url in the api calls of the vpn ip or the local ip of the machine. I tried both and it shows for both that the connection to Ollama is working (when I click on verify).
I am running ollama locally and open-webui in docker. The system worked before udpating ollama and openwebui to the latest version
Bit disturbing is that I do not see any error in the logs

Originally created by @mistrjirka on GitHub (Mar 19, 2025). ### Check Existing Issues - [x] I have searched the existing issues and discussions. - [x] I am using the latest version of Open WebUI. ### Installation Method Git Clone ### Open WebUI Version v0.5.20 ### Ollama Version (if applicable) ollama version is 0.6.1 ### Operating System Arch linux ### Browser (if applicable) Firefox ### Confirmation - [x] I have read and followed all instructions in `README.md`. - [x] I am using the latest version of **both** Open WebUI and Ollama. - [x] I have included the browser console logs. - [x] I have included the Docker container logs. - [x] I have listed steps to reproduce the bug in detail. ### Expected Behavior When I send message to AI I expect it to respond. ### Actual Behavior Ai never responds just acts as if it was loading. ![Image](https://github.com/user-attachments/assets/8c90dc88-324b-4036-a76a-e0d3e0fc5523) ### Steps to Reproduce Update to latest version of openwebui and ollama and try to use the openwebui. ### Logs & Screenshots I looked through the logs and I cannot find anything that would suggest something broken. The Website is infinitely waiting on post request "new" This is the request that is infinitely loading as cURL: ```curl 'http://10.154.224.4:8080/api/v1/chats/new' -X POST -H 'User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:136.0) Gecko/20100101 Firefox/136.0' -H 'Accept: application/json' -H 'Accept-Language: cs,sk;q=0.8,en-US;q=0.5,en;q=0.3' -H 'Accept-Encoding: gzip, deflate' -H 'Referer: http://10.154.224.4:8080/' -H 'Content-Type: application/json' -H 'authorization: Bearer eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpZCI6IjBmOWMyZWQ0LTBiYjEtNDU4Ny04NjJiLWEzZjNmZDg1NDE1NyJ9.fcNihgrjUO8QZhHYPAvv9VJ0OlM5Csf95BGvpoc83Uc' -H 'Origin: http://10.154.224.4:8080' -H 'Connection: keep-alive' -H 'Cookie: token=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpZCI6IjBmOWMyZWQ0LTBiYjEtNDU4Ny04NjJiLWEzZjNmZDg1NDE1NyJ9.fcNihgrjUO8QZhHYPAvv9VJ0OlM5Csf95BGvpoc83Uc' -H 'Priority: u=0' -H 'Pragma: no-cache' -H 'Cache-Control: no-cache' --data-raw $'{"chat":{"id":"","title":"Nov\xfd chat","models":["deepseek-r1:1.5b"],"params":{},"history":{"messages":{"7d5009b8-b990-41d0-a75e-2f406db93e8d":{"id":"7d5009b8-b990-41d0-a75e-2f406db93e8d","parentId":null,"childrenIds":[],"role":"user","content":"Jak se vede. Um\xed\u0161 \u010desky?","timestamp":1742412270,"models":["deepseek-r1:1.5b"]}},"currentId":"7d5009b8-b990-41d0-a75e-2f406db93e8d"},"messages":[{"id":"7d5009b8-b990-41d0-a75e-2f406db93e8d","parentId":null,"childrenIds":[],"role":"user","content":"Jak se vede. Um\xed\u0161 \u010desky?","timestamp":1742412270,"models":["deepseek-r1:1.5b"]}],"tags":[],"timestamp":1742412270120}}'``` [open-webui.log](https://github.com/user-attachments/files/19349918/open-webui.log) [ollama.log](https://github.com/user-attachments/files/19349982/ollama.log) [console-export-2025-3-19_20-41-48.txt](https://github.com/user-attachments/files/19350025/console-export-2025-3-19_20-41-48.txt) ### Additional Information I tested that ollama works when I do ollama run. I am currently tunneled to the machine using wireguard vpn and I was not sure if I should use the url in the api calls of the vpn ip or the local ip of the machine. I tried both and it shows for both that the connection to Ollama is working (when I click on verify). I am running ollama locally and open-webui in docker. The system worked before udpating ollama and openwebui to the latest version Bit disturbing is that I do not see any error in the logs
GiteaMirror added the bug label 2025-11-11 15:55:27 -06:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#4497