[GH-ISSUE #1599] Large amount of RAG data possibly causing crash consistently #28091

Closed
opened 2026-04-25 02:49:38 -05:00 by GiteaMirror · 6 comments
Owner

Originally created by @yesilzeytin on GitHub (Apr 18, 2024).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/1599

Bug Report

Description

Bug Summary:
I have been working with Open WebUI for a few days now using several different GGUF models on a weak hardware. With deepseek-coder:1B, it has been going fine and smoothly with and without RAG. My documents until now were consisting of specifications and books converted from pdf to txt. However, when I tried to include 400 different code files with small sizes and tried to use their tag to provide as RAG data, the model just shuts down and gives the error: "Uh-oh! There was an issue connecting to Ollama." After that, I lose the Open WebUI itself. Therefore, there may be a problem with RAG implementation for large file counts. In Docker logs, I observe an error message related to "too many open files", so the problem most likely is due to that.

Steps to Reproduce:

  • Run the script "run-compose.sh" from the Terminal as normal.
  • New chat. Select your model and ask anything, you get a reply.
  • New chat. Select your model, give some RAG data of a few documents totaling more than 10 MB, you get a reply.
  • New chat. Select your model, give a RAG tag consisting of 400 small documents totaling only 3 MB, you get the error "Uh-oh! There was an issue connecting to Ollama." in chat after waiting a while.
  • New chat. Select your model and write anything. No response and no attempt at all, like Open WebUI is stalled.
  • Refresh the localhost:3000 for checking Open WebUI to see that it loads only a white blank screen.
  • Rerun the script "run-compose.sh" from the Terminal to restart.
  • Open WebUI works and the models are responsive again until the same thing is tried.

Expected Behavior:
Open WebUI shouldn't have suddenly crash with many RAG data files, especially when it proves to be capable of working with larger data sizes.

Actual Behavior:
Open WebUI crashes and needs to be restarted to be able to work again.

Environment

  • Operating System: Ubuntu 22.04.3 LTS on WSL instance on Windows 10
  • Browser (if applicable): Version 123.0.6312.124 (Official Build) (64-bit)
  • Hardware: Intel 1165G7 @2.80 GHz, 16 GB DDR4 RAM, no GPU.

Reproduction Details

Confirmation:

  • [+] I have read and followed all the instructions provided in the README.md.
  • [+] I am on the latest version of both Open WebUI and Ollama.
  • [+] I have included the browser console logs.
  • [+] I have included the Docker container logs.

Logs and Screenshots

Browser Console Logs:
localhost.log

Docker Container Logs:
docker.log

Screenshots (if applicable):
chrome_dULVBqKX8f

chrome_oqpwsq6GdS

chrome_umb7Vla5kZ

Installation Method

Docker

Additional Information

Firstly, using this framework has been a blast so far and thanks a lot for creating and sharing such a wonderful thing. I am aware of that RAG feature is in Alpha stage and may have errors, but wanted to report it anyway thinking it could potentially help you further improve this already-great tool.

Note

If the bug report is incomplete or does not follow the provided instructions, it may not be addressed. Please ensure that you have followed the steps outlined in the README.md and troubleshooting.md documents, and provide all necessary information for us to reproduce and address the issue. Thank you!

Originally created by @yesilzeytin on GitHub (Apr 18, 2024). Original GitHub issue: https://github.com/open-webui/open-webui/issues/1599 # Bug Report ## Description **Bug Summary:** I have been working with Open WebUI for a few days now using several different GGUF models on a weak hardware. With deepseek-coder:1B, it has been going fine and smoothly with and without RAG. My documents until now were consisting of specifications and books converted from pdf to txt. However, when I tried to include 400 different code files with small sizes and tried to use their tag to provide as RAG data, the model just shuts down and gives the error: **"Uh-oh! There was an issue connecting to Ollama."** After that, I lose the Open WebUI itself. Therefore, there may be a problem with RAG implementation for large file counts. In Docker logs, I observe an error message related to **"too many open files"**, so the problem most likely is due to that. **Steps to Reproduce:** - Run the script "run-compose.sh" from the Terminal as normal. - New chat. Select your model and ask anything, you get a reply. - New chat. Select your model, give some RAG data of a few documents totaling more than 10 MB, you get a reply. - New chat. Select your model, give a RAG tag consisting of 400 small documents totaling only 3 MB, you get the error "Uh-oh! There was an issue connecting to Ollama." in chat after waiting a while. - New chat. Select your model and write anything. No response and no attempt at all, like Open WebUI is stalled. - Refresh the localhost:3000 for checking Open WebUI to see that it loads only a white blank screen. - Rerun the script "run-compose.sh" from the Terminal to restart. - Open WebUI works and the models are responsive again until the same thing is tried. **Expected Behavior:** Open WebUI shouldn't have suddenly crash with many RAG data files, especially when it proves to be capable of working with larger data sizes. **Actual Behavior:** Open WebUI crashes and needs to be restarted to be able to work again. ## Environment - **Operating System:** Ubuntu 22.04.3 LTS on WSL instance on Windows 10 - **Browser (if applicable):** Version 123.0.6312.124 (Official Build) (64-bit) - **Hardware:** Intel 1165G7 @2.80 GHz, 16 GB DDR4 RAM, no GPU. ## Reproduction Details **Confirmation:** - [+] I have read and followed all the instructions provided in the README.md. - [+] I am on the latest version of both Open WebUI and Ollama. - [+] I have included the browser console logs. - [+] I have included the Docker container logs. ## Logs and Screenshots **Browser Console Logs:** [localhost.log](https://github.com/open-webui/open-webui/files/15024424/localhost.log) **Docker Container Logs:** [docker.log](https://github.com/open-webui/open-webui/files/15024427/docker.log) **Screenshots (if applicable):** ![chrome_dULVBqKX8f](https://github.com/open-webui/open-webui/assets/29541720/d406daf0-27e2-4371-bc39-2f4abd63f860) ![chrome_oqpwsq6GdS](https://github.com/open-webui/open-webui/assets/29541720/8ea9c37c-c2d9-4917-b9fc-5ddf31038a08) ![chrome_umb7Vla5kZ](https://github.com/open-webui/open-webui/assets/29541720/d9eb0478-6ead-4be3-8e2c-274ee95d33e7) ## Installation Method Docker ## Additional Information Firstly, using this framework has been a blast so far and thanks a lot for creating and sharing such a wonderful thing. I am aware of that RAG feature is in Alpha stage and may have errors, but wanted to report it anyway thinking it could potentially help you further improve this already-great tool. ## Note If the bug report is incomplete or does not follow the provided instructions, it may not be addressed. Please ensure that you have followed the steps outlined in the README.md and troubleshooting.md documents, and provide all necessary information for us to reproduce and address the issue. Thank you!
GiteaMirror added the enhancementhelp wanted labels 2026-04-25 02:49:38 -05:00
Author
Owner

@justinh-rahb commented on GitHub (Apr 18, 2024):

"Too many open files" would be an operating system imposed limit. There's kernel parameters that can be tuned to increase it, running ulimit -n 2048 in the container shell would temporarily double the default limit so you could test it.

<!-- gh-comment-id:2063721241 --> @justinh-rahb commented on GitHub (Apr 18, 2024): "Too many open files" would be an operating system imposed limit. There's kernel parameters that can be tuned to increase it, running `ulimit -n 2048` in the container shell would temporarily double the default limit so you could test it.
Author
Owner

@yesilzeytin commented on GitHub (Apr 18, 2024):

I wasn't expecting such a quick response. Thanks a lot. Unfortunately, the method you proposed still cause the same error no matter what file limit I set. What you say sounds quite logical but I wonder if there could be something else considering this setup also fails:
image

<!-- gh-comment-id:2063817765 --> @yesilzeytin commented on GitHub (Apr 18, 2024): I wasn't expecting such a quick response. Thanks a lot. Unfortunately, the method you proposed still cause the same error no matter what file limit I set. What you say sounds quite logical but I wonder if there could be something else considering this setup also fails: ![image](https://github.com/open-webui/open-webui/assets/29541720/f187c8e9-070b-417f-beec-57bc34344097)
Author
Owner

@justinh-rahb commented on GitHub (Apr 18, 2024):

Will investigate further @yesilzeytin. It's also possible that the host machine itself needs this tweak applied as well.

<!-- gh-comment-id:2063855477 --> @justinh-rahb commented on GitHub (Apr 18, 2024): Will investigate further @yesilzeytin. It's also possible that the host machine itself needs this tweak applied as well.
Author
Owner

@jarekmor commented on GitHub (Apr 19, 2024):

I have got the same problem.
I am using OI with the speakleash/Bielik-7B-Instruct-v0.1 model with 4K context window. The model is based on Mistral architecture and was fine-tuned on Polish data set. The problem i got is that after every 6-7 passages of conversation I got error mentioned above ("Uh-oh! "). My guess is that the cause is the short context window which can not handle properly the conversation history but I can be wrong. I am using it with my documents embedded into the vectorstore.
unlimit = unlimited on both docker container and my host machine.

<!-- gh-comment-id:2066497685 --> @jarekmor commented on GitHub (Apr 19, 2024): I have got the same problem. I am using OI with the speakleash/Bielik-7B-Instruct-v0.1 model with 4K context window. The model is based on Mistral architecture and was fine-tuned on Polish data set. The problem i got is that after every 6-7 passages of conversation I got error mentioned above ("Uh-oh! "). My guess is that the cause is the short context window which can not handle properly the conversation history but I can be wrong. I am using it with my documents embedded into the vectorstore. unlimit = unlimited on both docker container and my host machine.
Author
Owner

@justinh-rahb commented on GitHub (Apr 19, 2024):

The problem i got is that after every 6-7 passages of conversation I got error mentioned above ("Uh-oh! ").

@jarekmor I've seen that you've posted this on our Discord. I don't believe it to be related to the issue in this thread, it's not the same error. If you'd like you should open a new issue for that matter if you aren't getting a response on Discord.

<!-- gh-comment-id:2066531298 --> @justinh-rahb commented on GitHub (Apr 19, 2024): > The problem i got is that after every 6-7 passages of conversation I got error mentioned above ("Uh-oh! "). @jarekmor I've seen that you've posted this on our Discord. I don't believe it to be related to the issue in this thread, it's not the same error. If you'd like you should open a new issue for that matter if you aren't getting a response on Discord.
Author
Owner

@tjbck commented on GitHub (May 7, 2024):

Closing in favour of #2044, Let's continue our discussion there!

<!-- gh-comment-id:2098984667 --> @tjbck commented on GitHub (May 7, 2024): Closing in favour of #2044, Let's continue our discussion there!
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#28091