issue: Document Upload on Large PDFs Results in Gateway Timeout #5773

Closed
opened 2025-11-11 16:33:19 -06:00 by GiteaMirror · 7 comments
Owner

Originally created by @kns-12 on GitHub (Jul 16, 2025).

Check Existing Issues

  • I have searched the existing issues and discussions.
  • I am using the latest version of Open WebUI.

Installation Method

Git Clone

Open WebUI Version

0.6.16

Ollama Version (if applicable)

No response

Operating System

Official Helm-Chart => Kubernetes/Docker

Browser (if applicable)

137.0.7151.122

Confirmation

  • I have read and followed all instructions in README.md.
  • I am using the latest version of both Open WebUI and Ollama.
  • I have included the browser console logs.
  • I have included the Docker container logs.
  • I have provided every relevant configuration, setting, and environment variable used in my setup.
  • I have clearly listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc).
  • I have documented step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation. My steps:
  • Start with the initial platform/version/OS and dependencies used,
  • Specify exact install/launch/configure commands,
  • List URLs visited, user input (incl. example values/emails/passwords if needed),
  • Describe all options and toggles enabled or changed,
  • Include any files or environmental changes,
  • Identify the expected and actual result at each stage,
  • Ensure any reasonably skilled user can follow and hit the same issue.

Expected Behavior

Uploading a PDF with 150-200 pages should be embedded fine without any timeout occuring.

Actual Behavior

The uploading of the PDF file results in a 504 Gateway Timeout after 30s.

Image

Steps to Reproduce

Go into any chat with your model of choice.
In my case its Llama 3.3 70B (Which doesnt really matter since the embedding is done by another model)

Upload my attached example pdf filled with lorem ipsum and 600 pages.
lorem_ipsum_600_pages.pdf
It has 600 pages because of the small size and complexity the timeout doesn't occur on 200-300 pages.

As soon as you drag & drop the file into the chat the embedding and chunking starts.

After 30 seconds the Gateway Timeout should occur.

Logs & Screenshots

These are my settings for the rag.
Image

Additional Information

The current embedding model is BAAI/bge-m3.
Also I am using apache Tika as a helm-chart in the latest available version: 2.9.0-full
Chunk Size is 1500 and overlap is 100.

I don't think any of that matters, it's just the fact that the embedding exceeds the Gateway Timeout of 30 seconds.
Unfortunately theres lots of PDFs with a certain size that will break the timeout.
In my case its lots of technical documentations that have hundreds of pages.
Searching them via RAG would be a great help.

A possible solution would be an environment variable so everyone can set the timeout to their liking, since it's pretty specific for everyone

Originally created by @kns-12 on GitHub (Jul 16, 2025). ### Check Existing Issues - [x] I have searched the existing issues and discussions. - [x] I am using the latest version of Open WebUI. ### Installation Method Git Clone ### Open WebUI Version 0.6.16 ### Ollama Version (if applicable) _No response_ ### Operating System Official Helm-Chart => Kubernetes/Docker ### Browser (if applicable) 137.0.7151.122 ### Confirmation - [x] I have read and followed all instructions in `README.md`. - [x] I am using the latest version of **both** Open WebUI and Ollama. - [x] I have included the browser console logs. - [x] I have included the Docker container logs. - [x] I have **provided every relevant configuration, setting, and environment variable used in my setup.** - [x] I have clearly **listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup** (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc). - [x] I have documented **step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation**. My steps: - Start with the initial platform/version/OS and dependencies used, - Specify exact install/launch/configure commands, - List URLs visited, user input (incl. example values/emails/passwords if needed), - Describe all options and toggles enabled or changed, - Include any files or environmental changes, - Identify the expected and actual result at each stage, - Ensure any reasonably skilled user can follow and hit the same issue. ### Expected Behavior Uploading a PDF with 150-200 pages should be embedded fine without any timeout occuring. ### Actual Behavior The uploading of the PDF file results in a 504 Gateway Timeout after 30s. <img width="525" height="145" alt="Image" src="https://github.com/user-attachments/assets/eef4259a-02a5-40db-9ec5-bf854756ba26" /> ### Steps to Reproduce Go into any chat with your model of choice. In my case its Llama 3.3 70B (Which doesnt really matter since the embedding is done by another model) Upload my attached example pdf filled with lorem ipsum and 600 pages. [lorem_ipsum_600_pages.pdf](https://github.com/user-attachments/files/21248338/lorem_ipsum_600_pages.pdf) It has 600 pages because of the small size and complexity the timeout doesn't occur on 200-300 pages. As soon as you drag & drop the file into the chat the embedding and chunking starts. After 30 seconds the Gateway Timeout should occur. ### Logs & Screenshots These are my settings for the rag. <img width="2110" height="390" alt="Image" src="https://github.com/user-attachments/assets/276a4660-ae18-4a8e-950d-af4053d9d518" /> ### Additional Information The current embedding model is BAAI/bge-m3. Also I am using apache Tika as a helm-chart in the latest available version: 2.9.0-full Chunk Size is 1500 and overlap is 100. I don't think any of that matters, it's just the fact that the embedding exceeds the Gateway Timeout of 30 seconds. Unfortunately theres lots of PDFs with a certain size that will break the timeout. In my case its lots of technical documentations that have hundreds of pages. Searching them via RAG would be a great help. A possible solution would be an environment variable so everyone can set the timeout to their liking, since it's pretty specific for everyone
GiteaMirror added the bug label 2025-11-11 16:33:19 -06:00
Author
Owner

@tjbck commented on GitHub (Jul 16, 2025):

#15023

@tjbck commented on GitHub (Jul 16, 2025): #15023
Author
Owner

@rgaricano commented on GitHub (Jul 16, 2025):

working fine in my endpoint (4 core+16GB+GPU 12GB)
500 pages pdf
Image

@rgaricano commented on GitHub (Jul 16, 2025): working fine in my endpoint (4 core+16GB+GPU 12GB) 500 pages pdf ![Image](https://github.com/user-attachments/assets/b9d83bb2-cb3b-4e24-bd18-84c0416d7015)
Author
Owner

@rgaricano commented on GitHub (Jul 16, 2025):

with yours

Image

Image

@rgaricano commented on GitHub (Jul 16, 2025): with yours ![Image](https://github.com/user-attachments/assets/c30978e0-2ca6-42e2-844b-a5b34d45f863) ![Image](https://github.com/user-attachments/assets/e84aaf0d-b59d-4fc8-ad70-392b7cd03705)
Author
Owner

@kns-12 commented on GitHub (Jul 16, 2025):

Oh okay that's nice, thank you for testing it!
Would you mind sharing how long the POST call to /api/v1/files/ took to be processed?
As mentioned I am receiving a Gateway Timeout after 30s each time.
Smaller documents that do not exceed that duration work fine.
Image

Are you using Apache Tika?

@kns-12 commented on GitHub (Jul 16, 2025): Oh okay that's nice, thank you for testing it! Would you mind sharing how long the POST call to /api/v1/files/ took to be processed? As mentioned I am receiving a Gateway Timeout after 30s each time. Smaller documents that do not exceed that duration work fine. <img width="378" height="239" alt="Image" src="https://github.com/user-attachments/assets/38be70a8-72b3-4f8e-ada3-6e4ee0e2c6a0" /> Are you using Apache Tika?
Author
Owner

@rgaricano commented on GitHub (Jul 16, 2025):

in my end 15,02sg,
No, I'm using default (but I have changed the default pyPDFLoader by pyMuPDFLoader, change only relevant for OCR related)

Image

@rgaricano commented on GitHub (Jul 16, 2025): in my end 15,02sg, No, I'm using default (but I have changed the default pyPDFLoader by pyMuPDFLoader, change only relevant for OCR related) ![Image](https://github.com/user-attachments/assets/14ff1a9c-1a7e-4d7e-9303-35f2c0ec3ff1)
Author
Owner

@rgaricano commented on GitHub (Jul 16, 2025):

with tika, 25sg

Image

Image

@rgaricano commented on GitHub (Jul 16, 2025): with tika, 25sg ![Image](https://github.com/user-attachments/assets/5814b98a-93cc-463c-beb4-b6abeb8815f2) ![Image](https://github.com/user-attachments/assets/dfdd94dc-b429-4bf5-ba20-bdfb02ff01bb)
Author
Owner

@kns-12 commented on GitHub (Jul 18, 2025):

@rgaricano I got another embedding model to work using ollama "nomic-embed-text". The "BAAI/bge-m3" was apparently just to slow for the 30 seconds Gateway Timeout. Still it would be good to be able to change this timeout.

@kns-12 commented on GitHub (Jul 18, 2025): @rgaricano I got another embedding model to work using ollama "nomic-embed-text". The "BAAI/bge-m3" was apparently just to slow for the 30 seconds Gateway Timeout. Still it would be good to be able to change this timeout.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#5773