[GH-ISSUE #16313] issue: Voice mode seems to be broken on desktop web browsers #33391

Closed
opened 2026-04-25 07:17:58 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @shadowshadow725 on GitHub (Aug 6, 2025).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/16313

Check Existing Issues

  • I have searched the existing issues and discussions.
  • I am using the latest version of Open WebUI.

Installation Method

Docker

Open WebUI Version

0.6.18

Ollama Version (if applicable)

latest

Operating System

Ubuntu 24.04 lts

Browser (if applicable)

chrome and edge desktop

Confirmation

  • I have read and followed all instructions in README.md.
  • I am using the latest version of both Open WebUI and Ollama.
  • I have included the browser console logs.
  • I have included the Docker container logs.
  • I have provided every relevant configuration, setting, and environment variable used in my setup.
  • I have clearly listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc).
  • I have documented step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation. My steps:
  • Start with the initial platform/version/OS and dependencies used,
  • Specify exact install/launch/configure commands,
  • List URLs visited, user input (incl. example values/emails/passwords if needed),
  • Describe all options and toggles enabled or changed,
  • Include any files or environmental changes,
  • Identify the expected and actual result at each stage,
  • Ensure any reasonably skilled user can follow and hit the same issue.

Expected Behavior

When I talk the white circle that pops up when I click voice mode will change in size and the model will respond to me.

Actual Behavior

The white circle indicator shows up but doesn't respond to my speech and the model does nothing.

Image

Steps to Reproduce

Running
ubuntu 24.04 lts
rtx 3060 12g
Docker version 28.3.2, build 578ccf6
The following command is used to run the containers

docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama --restart unless-stopped ollama/ollama
docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v open-webui:/app/backend/data --name open-webui --restart unless-stopped ghcr.io/open-webui/open-webui:main

I've made sure that I have a mic connected to my computer and the browser is using the correct mic and the mic is working. When I click voice mode it doesn't seem to detect that I'm speaking. I've tried using voice mode from an android phone using chrome and the model responds to me as expected and the speech indicator works as expected.

Logs & Screenshots

Image

some logs from the ollama container

time=2025-08-06T03:32:41.415Z level=INFO source=sched.go:786 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e gpu=GPU-8876e5e1-cf0e-d479-b427-0fe7fcb38333 parallel=1 available=12301828096 required="10.0 GiB" time=2025-08-06T03:32:41.539Z level=INFO source=server.go:135 msg="system memory" total="31.2 GiB" free="27.0 GiB" free_swap="1.4 GiB" time=2025-08-06T03:32:41.539Z level=INFO source=server.go:175 msg=offload library=cuda layers.requested=-1 layers.model=41 layers.offload=41 layers.split="" memory.available="[11.5 GiB]" memory.gpu_overhead="0 B" memory.required.full="10.0 GiB" memory.required.partial="10.0 GiB" memory.required.kv="640.0 MiB" memory.required.allocations="[10.0 GiB]" memory.weights.total="8.2 GiB" memory.weights.repeating="7.6 GiB" memory.weights.nonrepeating="608.6 MiB" memory.graph.full="533.3 MiB" memory.graph.partial="533.3 MiB" llama_model_loader: loaded meta data with 27 key-value pairs and 443 tensors from /root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 14B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 14B llama_model_loader: - kv 5: qwen3.block_count u32 = 40 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 17408 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 40 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 161 tensors llama_model_loader: - type f16: 40 tensors llama_model_loader: - type q4_K: 221 tensors llama_model_loader: - type q6_K: 21 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 8.63 GiB (5.02 BPW) load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 1 print_info: model type = ?B print_info: model params = 14.77 B print_info: general.name = Qwen3 14B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 llama_model_load: vocab only - skipping tensors time=2025-08-06T03:32:41.677Z level=INFO source=server.go:438 msg="starting llama server" cmd="/usr/bin/ollama runner --model /root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e --ctx-size 4096 --batch-size 512 --n-gpu-layers 41 --threads 8 --parallel 1 --port 45447" time=2025-08-06T03:32:41.677Z level=INFO source=sched.go:481 msg="loaded runners" count=1 time=2025-08-06T03:32:41.677Z level=INFO source=server.go:598 msg="waiting for llama runner to start responding" time=2025-08-06T03:32:41.678Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server not responding" time=2025-08-06T03:32:41.690Z level=INFO source=runner.go:815 msg="starting go runner" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes load_backend: loaded CUDA backend from /usr/lib/ollama/libggml-cuda.so load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-icelake.so time=2025-08-06T03:32:41.749Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,600,610,700,750,800,860,870,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(gcc) time=2025-08-06T03:32:41.749Z level=INFO source=runner.go:874 msg="Server listening on 127.0.0.1:45447" llama_model_load_from_file_impl: using device CUDA0 (NVIDIA GeForce RTX 3060) - 11731 MiB free llama_model_loader: loaded meta data with 27 key-value pairs and 443 tensors from /root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 14B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 14B llama_model_loader: - kv 5: qwen3.block_count u32 = 40 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 17408 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 40 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 161 tensors llama_model_loader: - type f16: 40 tensors llama_model_loader: - type q4_K: 221 tensors llama_model_loader: - type q6_K: 21 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 8.63 GiB (5.02 BPW) time=2025-08-06T03:32:41.929Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server loading model" load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 0 print_info: n_ctx_train = 40960 print_info: n_embd = 5120 print_info: n_layer = 40 print_info: n_head = 40 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: n_swa_pattern = 1 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 5 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-06 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 17408 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = 0 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 40960 print_info: rope_finetuned = unknown print_info: ssm_d_conv = 0 print_info: ssm_d_inner = 0 print_info: ssm_d_state = 0 print_info: ssm_dt_rank = 0 print_info: ssm_dt_b_c_rms = 0 print_info: model type = 14B print_info: model params = 14.77 B print_info: general.name = Qwen3 14B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = true) load_tensors: offloading 40 repeating layers to GPU load_tensors: offloading output layer to GPU load_tensors: offloaded 41/41 layers to GPU load_tensors: CUDA0 model buffer size = 8423.47 MiB load_tensors: CPU_Mapped model buffer size = 417.30 MiB llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 4096 llama_context: n_ctx_per_seq = 4096 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = 0 llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 llama_context: n_ctx_per_seq (4096) < n_ctx_train (40960) -- the full capacity of the model will not be utilized llama_context: CUDA_Host output buffer size = 0.60 MiB llama_kv_cache_unified: kv_size = 4096, type_k = 'f16', type_v = 'f16', n_layer = 40, can_shift = 1, padding = 32 llama_kv_cache_unified: CUDA0 KV buffer size = 640.00 MiB llama_kv_cache_unified: KV self size = 640.00 MiB, K (f16): 320.00 MiB, V (f16): 320.00 MiB llama_context: CUDA0 compute buffer size = 368.00 MiB llama_context: CUDA_Host compute buffer size = 18.01 MiB llama_context: graph nodes = 1526 llama_context: graph splits = 2 time=2025-08-06T03:32:47.200Z level=INFO source=server.go:637 msg="llama runner started in 5.52 seconds" [GIN] 2025/08/06 - 03:32:52 | 200 | 10.98706337s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:33:26 | 200 | 18.320083604s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:33:53 | 200 | 13.387027252s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:35:03 | 200 | 1.594868544s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:35:57 | 200 | 5.442233ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:35:57 | 200 | 63.13µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:36:03 | 200 | 43.825µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:36:08 | 200 | 48.027µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:36:17 | 200 | 7.44382486s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:36:34 | 200 | 16.825114491s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:36:46 | 200 | 2.246994ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:36:46 | 200 | 71.171µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:37:01 | 200 | 26.755593194s | 172.17.0.1 | POST "/api/chat" time=2025-08-06T03:37:03.245Z level=INFO source=sched.go:546 msg="updated VRAM based on existing loaded models" gpu=GPU-8876e5e1-cf0e-d479-b427-0fe7fcb38333 library=cuda total="11.6 GiB" available="1.6 GiB" [GIN] 2025/08/06 - 03:37:20 | 200 | 2.591057ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:20 | 200 | 48.328µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:37:21 | 200 | 38.474µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:37:21 | 200 | 2.641725ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:21 | 200 | 41.419µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:37:24 | 200 | 23.389278513s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:25 | 200 | 2.369078ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:25 | 200 | 21.353µs | 172.17.0.1 | GET "/api/ps" time=2025-08-06T03:37:25.243Z level=INFO source=sched.go:786 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-38e8dcc30df4eb0e29eaf5c74ba6ce3f2cd66badad50768fc14362acfb8b8cb6 gpu=GPU-8876e5e1-cf0e-d479-b427-0fe7fcb38333 parallel=1 available=12301828096 required="5.4 GiB" time=2025-08-06T03:37:25.366Z level=INFO source=server.go:135 msg="system memory" total="31.2 GiB" free="26.8 GiB" free_swap="1.2 GiB" time=2025-08-06T03:37:25.367Z level=INFO source=server.go:175 msg=offload library=cuda layers.requested=-1 layers.model=36 layers.offload=36 layers.split="" memory.available="[11.5 GiB]" memory.gpu_overhead="0 B" memory.required.full="5.4 GiB" memory.required.partial="5.4 GiB" memory.required.kv="280.0 MiB" memory.required.allocations="[5.4 GiB]" memory.weights.total="2.6 GiB" memory.weights.repeating="2.2 GiB" memory.weights.nonrepeating="420.4 MiB" memory.graph.full="2.0 GiB" memory.graph.partial="3.7 GiB" time=2025-08-06T03:37:25.421Z level=INFO source=server.go:438 msg="starting llama server" cmd="/usr/bin/ollama runner --ollama-engine --model /root/.ollama/models/blobs/sha256-38e8dcc30df4eb0e29eaf5c74ba6ce3f2cd66badad50768fc14362acfb8b8cb6 --ctx-size 4096 --batch-size 512 --n-gpu-layers 36 --threads 8 --parallel 1 --port 39807" time=2025-08-06T03:37:25.421Z level=INFO source=sched.go:481 msg="loaded runners" count=1 time=2025-08-06T03:37:25.421Z level=INFO source=server.go:598 msg="waiting for llama runner to start responding" time=2025-08-06T03:37:25.421Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server not responding" time=2025-08-06T03:37:25.431Z level=INFO source=runner.go:925 msg="starting ollama engine" time=2025-08-06T03:37:25.431Z level=INFO source=runner.go:983 msg="Server listening on 127.0.0.1:39807" time=2025-08-06T03:37:25.494Z level=INFO source=ggml.go:92 msg="" architecture=gemma3n file_type=Q4_K_M name="" description="" num_tensors=847 num_key_values=40 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes load_backend: loaded CUDA backend from /usr/lib/ollama/libggml-cuda.so load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-icelake.so time=2025-08-06T03:37:25.539Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,600,610,700,750,800,860,870,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(gcc) time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:365 msg="offloading 35 repeating layers to GPU" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:371 msg="offloading output layer to GPU" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:376 msg="offloaded 36/36 layers to GPU" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:379 msg="model weights" buffer=CPU size="420.4 MiB" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:379 msg="model weights" buffer=CUDA0 size="7.0 GiB" time=2025-08-06T03:37:25.627Z level=INFO source=ggml.go:668 msg="compute graph" backend=CUDA0 buffer_type=CUDA0 size="135.5 MiB" time=2025-08-06T03:37:25.627Z level=INFO source=ggml.go:668 msg="compute graph" backend=CPU buffer_type=CPU size="4.0 MiB" time=2025-08-06T03:37:25.672Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server loading model" time=2025-08-06T03:37:28.181Z level=INFO source=server.go:637 msg="llama runner started in 2.76 seconds" [GIN] 2025/08/06 - 03:37:29 | 200 | 26.550326175s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:30 | 200 | 1.394084314s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:31 | 200 | 654.35957ms | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:32 | 200 | 1.072350544s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:49 | 200 | 3.118832ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:49 | 200 | 39.63µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:39:36 | 200 | 2.815778ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:39:36 | 200 | 32.844µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:40:27 | 200 | 48.372µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:43:45 | 200 | 3.171724ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:43:45 | 200 | 32.527µs | 172.17.0.1 | GET "/api/ps"

some logs from the openwebui containter

2025-08-06 03:37:20.635 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:20.635 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:20.638 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/pinned HTTP/1.1" 200 - {} 2025-08-06 03:37:20.651 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:20.651 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:20.660 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:21.024 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:37:21.031 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/configs/banners HTTP/1.1" 200 - {} 2025-08-06 03:37:21.037 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/tools/ HTTP/1.1" 200 - {} 2025-08-06 03:37:21.059 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /ollama/api/version HTTP/1.1" 200 - {} 2025-08-06 03:37:21.094 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/user/settings HTTP/1.1" 200 - {} 2025-08-06 03:37:21.209 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/version/updates HTTP/1.1" 200 - {} 2025-08-06 03:37:21.476 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:21.476 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:21.808 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:37:23.677 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:37:25.154 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:25.155 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:25.383 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:37:28.374 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/chat/completions HTTP/1.1" 200 - {} 2025-08-06 03:37:28.497 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:28.522 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:29.488 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/v1/audio/speech HTTP/1.1" 500 - {} Exception in ASGI application Traceback (most recent call last): File "/usr/local/lib/python3.11/site-packages/uvicorn/protocols/http/httptools_impl.py", line 409, in run_asgi result = await app( # type: ignore[func-returns-value] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/uvicorn/middleware/proxy_headers.py", line 60, in __call__ return await self.app(scope, receive, send) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/fastapi/applications.py", line 1054, in __call__ await super().__call__(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/applications.py", line 112, in __call__ await self.middleware_stack(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/errors.py", line 187, in __call__ raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/errors.py", line 165, in __call__ await self.app(scope, receive, _send) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/cors.py", line 93, in __call__ await self.simple_response(scope, receive, send, request_headers=headers) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/cors.py", line 144, in simple_response await self.app(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1162, in inspect_websocket return await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1141, in check_url response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1127, in commit_session_after_request response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/utils/security_headers.py", line 11, in dispatch response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1113, in dispatch response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette_compress/__init__.py", line 92, in __call__ return await self._zstd(scope, receive, send) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette_compress/_zstd_legacy.py", line 100, in __call__ await self.app(scope, receive, wrapper) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/exceptions.py", line 62, in __call__ await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app raise exc File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app await app(scope, receive, sender) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 715, in __call__ await self.middleware_stack(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 735, in app await route.handle(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 288, in handle await self.app(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 76, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app raise exc File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app await app(scope, receive, sender) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 73, in app response = await f(request) ^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/fastapi/routing.py", line 301, in app raw_response = await run_endpoint_function( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/fastapi/routing.py", line 212, in run_endpoint_function return await dependant.call(**values) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/routers/audio.py", line 517, in speech load_speech_pipeline(request) File "/app/backend/open_webui/routers/audio.py", line 303, in load_speech_pipeline request.app.state.speech_speaker_embeddings_dataset = load_dataset( ^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 1392, in load_dataset builder_instance = load_dataset_builder( ^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 1132, in load_dataset_builder dataset_module = dataset_module_factory( ^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 1031, in dataset_module_factory raise e1 from None File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 989, in dataset_module_factory raise RuntimeError(f"Dataset scripts are no longer supported, but found {filename}") RuntimeError: Dataset scripts are no longer supported, but found cmu-arctic-xvectors.py 2025-08-06 03:37:29.496 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /audio/notification.mp3 HTTP/1.1" 206 - {} 2025-08-06 03:37:29.496 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/chat/completed HTTP/1.1" 200 - {} 2025-08-06 03:37:29.517 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /static/apple-touch-icon.png HTTP/1.1" 200 - {} 2025-08-06 03:37:29.525 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/v1/chats/0f79969d-8c97-49cf-88e9-6511e03be0dc HTTP/1.1" 200 - {} 2025-08-06 03:37:29.543 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:29.565 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:31.494 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:31.502 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:31.547 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:31.585 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:32.578 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/all/tags HTTP/1.1" 200 - {} 2025-08-06 03:37:32.681 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/0f79969d-8c97-49cf-88e9-6511e03be0dc HTTP/1.1" 200 - {} 2025-08-06 03:37:32.706 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/all/tags HTTP/1.1" 200 - {} 2025-08-06 03:37:49.732 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:49.733 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:49.984 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:38:08.921 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /manifest.json HTTP/1.1" 200 - {} 2025-08-06 03:38:08.921 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/0.BEiiGd97.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.921 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/purify.DQwrSZtH.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Collapsible.DZAVxvaL.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Modal.CyLKLEmt.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Skeleton.Cr25vAl_.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/ConfirmDialog.kaSBQ3kP.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.924 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Messages.Bbuu_qma.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.925 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/MapSelector.CIGW-MKW.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.926 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Drawer.Bz8t_RYw.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.926 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Chat.CWA0G-PE.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.926 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/2.BXeiR8Q0.css HTTP/1.1" 304 - {} 2025-08-06 03:38:20.264 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:38:23.683 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:39:20.972 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:39:23.684 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:39:36.388 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:39:36.388 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:39:36.486 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/0f79969d-8c97-49cf-88e9-6511e03be0dc HTTP/1.1" 200 - {} 2025-08-06 03:39:36.510 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2a5699a7-9cc3-4745-9f4d-7b408f9879dc HTTP/1.1" 200 - {} 2025-08-06 03:39:36.517 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/10cfaedf-d0db-414e-90a6-1297c4a1a67a HTTP/1.1" 200 - {} 2025-08-06 03:39:36.548 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e7b1a31a-3226-46de-8e70-90a46a1b1e62 HTTP/1.1" 200 - {} 2025-08-06 03:39:36.556 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/541d0f9a-0a0a-4e8f-be1a-804c3189519f HTTP/1.1" 200 - {} 2025-08-06 03:39:36.576 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e05b6db7-1b3c-4e06-b622-6aa375565b9d HTTP/1.1" 200 - {} 2025-08-06 03:39:36.598 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2d5d7d91-54d6-44e4-a7f5-9f8fb1975a44 HTTP/1.1" 200 - {} 2025-08-06 03:39:36.716 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/291b1336-1016-441e-bbfc-4c6dc5e82046 HTTP/1.1" 200 - {} 2025-08-06 03:39:36.839 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:39:37.039 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/0b5a3a3d-0581-497d-b287-9806e3ee2dc0 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.053 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e1bac2cf-887e-473e-abdb-dd515ba10161 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.056 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/77cf441b-f231-4c73-a4a8-4254c6d5e002 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.056 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/09f66bae-2d38-43ec-b39a-a2586972b726 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.234 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/291a2e20-84eb-438a-9f67-e81a9de76c90 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.241 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/16ddce7c-f97b-4ba4-a028-d4503eaa6935 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.241 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/08fa0b98-f121-4673-843e-06069fde3dbb HTTP/1.1" 200 - {} 2025-08-06 03:39:37.242 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/f2fc367c-4f70-4085-9254-1f2a185171cb HTTP/1.1" 200 - {} 2025-08-06 03:39:37.242 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1e9ca672-5dca-499f-8b2b-3c5b7d45c66b HTTP/1.1" 200 - {} 2025-08-06 03:39:37.270 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/c61ad61d-0aa7-475e-8614-da28d088fe57 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.270 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/8c40c89b-06e9-48bd-81e6-c4051d630e14 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.270 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/9a784cd1-e81f-411a-8cd7-d750ca3831c2 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.273 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/5c15e756-7963-4cd6-95a7-1fd03e45b00c HTTP/1.1" 200 - {} 2025-08-06 03:39:37.274 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/08ea4579-4d0d-449a-bec2-6f7d27bbff74 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.274 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/ff7f8aa5-fd0a-4856-a7b6-a08c96509767 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.303 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/bd9c328f-abd4-4b25-b8d8-348bcfd38306 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.304 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/5baf6686-1ab4-46b4-9b15-4ca653e6bd97 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.304 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/f7178b16-5b4f-47c7-82dd-b129ded7c5df HTTP/1.1" 200 - {} 2025-08-06 03:39:37.304 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/8f37e3e9-a360-49df-a4c4-76b8fef2a0e4 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.309 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/69d7e923-cd9f-445b-a9f4-5dbdd87452d1 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.310 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/4dc97d6b-6522-44d7-a385-e0b6f1b43ad4 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.335 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/975c1a64-356a-4444-a6e0-db8cfbfad5c1 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/f01d1aba-2723-4aec-8e40-6acbab8b53b3 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/331da5b4-9211-4cab-98fc-5033cc25b585 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e8cff7db-4d78-40bd-b5ba-418a7899608e HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/9c501207-9049-48cd-97a8-78f9a701ee67 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/b6ec8a47-52fe-4067-8a10-1707fe13328d HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2000cd81-6fdf-4306-adb2-db00e771997d HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/9d5c76fa-cd37-4e5f-a0ce-25ce956874f5 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/19.D5tH4moG.js HTTP/1.1" 200 - {} 2025-08-06 03:39:37.370 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/17131d5c-e521-498b-bf7b-70cd5764bf04 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.377 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/89c333b2-6d31-4040-9f1d-68116372c5f3 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1969a05b-3362-4fe0-b0eb-4fabef4ad838 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2775eab7-5937-409b-a50a-ab34e877757f HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e6acdce3-ff67-41d1-9f39-b1ea1d40849a HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/8105e1ca-28d0-47ed-8daa-546e687e5ee0 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.443 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/caa37dc6-c410-4934-a260-1368cb4240e6 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.443 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/bad3f560-e73e-4cbf-8ae0-614e910ac555 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.444 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/c87a7b7c-c2cf-41e7-8ca5-e4f5efcbf1f7 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.448 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/16b06372-d991-4511-a68a-a6846e278693 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.450 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/0fe5cbc8-93af-43a2-821e-ad867ed61cf7 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.450 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/26d2e83d-7d8d-4251-a24d-4ce15c4dcc52 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.450 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/a0350689-d3cc-4cb4-afa3-339061fc5392 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.458 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/dc49270b-f60a-47c7-a266-ee4a635bf986 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.459 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/?page=2 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.459 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/02587867-f7b4-402f-a303-581f45beecec HTTP/1.1" 200 - {} 2025-08-06 03:39:37.459 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/22a29d68-c6d9-417c-9e67-312fe658b3b7 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.460 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/02ddc96e-b97f-4bd2-a0ce-cbabb0ebc088 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.460 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/7424616b-2d3c-44b4-830d-e230548b45ad HTTP/1.1" 200 - {} 2025-08-06 03:39:37.464 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1ba04b95-d3de-4dbf-b90e-0de68e3684cd HTTP/1.1" 200 - {} 2025-08-06 03:39:37.476 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:39:38.005 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:39:38.025 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:39:38.465 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1a144301-ac99-48c6-8408-faf79f1d9545 HTTP/1.1" 200 - {} 2025-08-06 03:39:38.489 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:39:39.446 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/3.DQTEdc7F.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.446 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/9.OrSiIAXf.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.475 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/29.wEbTgpRj.css HTTP/1.1" 200 - {} 2025-08-06 03:39:39.475 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/Ck0cvSaj.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.476 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/17.DB2DlSzW.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.478 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/DQtrEHa0.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.478 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/V4NzNuWG.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.479 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/TZ3cQ_RY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.498 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:39.500 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/18.DPUDtlSY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.504 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:39.523 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:39.529 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:41.013 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/C83syoBY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.014 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/CICvqL9d.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.014 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/DbZwnr2B.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.016 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/CFy5Jpyl.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.017 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/15.BhTxcmdY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.017 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/Dg4-BTkH.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.017 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/CeDBf-bV.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.275 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/webhook HTTP/1.1" 200 - {} 2025-08-06 03:39:41.277 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap/server HTTP/1.1" 200 - {} 2025-08-06 03:39:41.278 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config HTTP/1.1" 200 - {} 2025-08-06 03:39:41.281 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/16.DqmmaYsb.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.291 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap HTTP/1.1" 200 - {} 2025-08-06 03:39:41.309 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/webhook HTTP/1.1" 200 - {} 2025-08-06 03:39:41.310 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap/server HTTP/1.1" 200 - {} 2025-08-06 03:39:41.312 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config HTTP/1.1" 200 - {} 2025-08-06 03:39:41.322 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap HTTP/1.1" 200 - {} 2025-08-06 03:39:41.498 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/version/updates HTTP/1.1" 200 - {} 2025-08-06 03:39:41.604 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/version/updates HTTP/1.1" 200 - {} 2025-08-06 03:40:21.205 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:40:23.687 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:40:26.605 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/10.C41L3GeI.js HTTP/1.1" 200 - {} 2025-08-06 03:40:26.606 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/Cfz980HS.js HTTP/1.1" 200 - {} 2025-08-06 03:40:26.607 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/HKxhIwoV.js HTTP/1.1" 200 - {} 2025-08-06 03:40:27.346 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /ollama/api/version HTTP/1.1" 200 - {} 2025-08-06 03:40:27.388 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/user/settings HTTP/1.1" 200 - {} 2025-08-06 03:40:28.695 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:41:21.977 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:41:23.687 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:42:22.982 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:42:23.715 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:43:23.692 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:43:23.971 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:43:45.819 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:43:45.819 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:43:46.031 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:44:23.690 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:44:23.983 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:45:23.694 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:45:24.981 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:46:23.697 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:46:25.971 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:47:23.697 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:47:25.976 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:48:23.697 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:48:25.991 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:49:23.703 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:49:26.980 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:50:23.704 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:50:26.981 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {}

browser logs (pretty much nothing)

[tiptap warn]: Duplicate extension names found: ['codeBlock', 'bulletList', 'listItem', 'listKeymap', 'orderedList']. This can lead to issues.

Additional Information

No response

Originally created by @shadowshadow725 on GitHub (Aug 6, 2025). Original GitHub issue: https://github.com/open-webui/open-webui/issues/16313 ### Check Existing Issues - [x] I have searched the existing issues and discussions. - [x] I am using the latest version of Open WebUI. ### Installation Method Docker ### Open WebUI Version 0.6.18 ### Ollama Version (if applicable) latest ### Operating System Ubuntu 24.04 lts ### Browser (if applicable) chrome and edge desktop ### Confirmation - [x] I have read and followed all instructions in `README.md`. - [x] I am using the latest version of **both** Open WebUI and Ollama. - [x] I have included the browser console logs. - [x] I have included the Docker container logs. - [x] I have **provided every relevant configuration, setting, and environment variable used in my setup.** - [x] I have clearly **listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup** (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc). - [x] I have documented **step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation**. My steps: - Start with the initial platform/version/OS and dependencies used, - Specify exact install/launch/configure commands, - List URLs visited, user input (incl. example values/emails/passwords if needed), - Describe all options and toggles enabled or changed, - Include any files or environmental changes, - Identify the expected and actual result at each stage, - Ensure any reasonably skilled user can follow and hit the same issue. ### Expected Behavior When I talk the white circle that pops up when I click voice mode will change in size and the model will respond to me. ### Actual Behavior The white circle indicator shows up but doesn't respond to my speech and the model does nothing. <img width="1631" height="980" alt="Image" src="https://github.com/user-attachments/assets/012c51d5-86a9-4ce8-8519-5f9ceec6f73f" /> ### Steps to Reproduce Running ubuntu 24.04 lts rtx 3060 12g Docker version 28.3.2, build 578ccf6 The following command is used to run the containers docker run -d --gpus=all -v ollama:/root/.ollama -p 11434:11434 --name ollama --restart unless-stopped ollama/ollama docker run -d -p 3000:8080 --add-host=host.docker.internal:host-gateway -v open-webui:/app/backend/data --name open-webui --restart unless-stopped ghcr.io/open-webui/open-webui:main I've made sure that I have a mic connected to my computer and the browser is using the correct mic and the mic is working. When I click voice mode it doesn't seem to detect that I'm speaking. I've tried using voice mode from an android phone using chrome and the model responds to me as expected and the speech indicator works as expected. ### Logs & Screenshots <img width="1640" height="992" alt="Image" src="https://github.com/user-attachments/assets/1a330380-bfb2-4a58-8bf8-a661a1c40645" /> some logs from the ollama container `time=2025-08-06T03:32:41.415Z level=INFO source=sched.go:786 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e gpu=GPU-8876e5e1-cf0e-d479-b427-0fe7fcb38333 parallel=1 available=12301828096 required="10.0 GiB" time=2025-08-06T03:32:41.539Z level=INFO source=server.go:135 msg="system memory" total="31.2 GiB" free="27.0 GiB" free_swap="1.4 GiB" time=2025-08-06T03:32:41.539Z level=INFO source=server.go:175 msg=offload library=cuda layers.requested=-1 layers.model=41 layers.offload=41 layers.split="" memory.available="[11.5 GiB]" memory.gpu_overhead="0 B" memory.required.full="10.0 GiB" memory.required.partial="10.0 GiB" memory.required.kv="640.0 MiB" memory.required.allocations="[10.0 GiB]" memory.weights.total="8.2 GiB" memory.weights.repeating="7.6 GiB" memory.weights.nonrepeating="608.6 MiB" memory.graph.full="533.3 MiB" memory.graph.partial="533.3 MiB" llama_model_loader: loaded meta data with 27 key-value pairs and 443 tensors from /root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 14B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 14B llama_model_loader: - kv 5: qwen3.block_count u32 = 40 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 17408 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 40 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 161 tensors llama_model_loader: - type f16: 40 tensors llama_model_loader: - type q4_K: 221 tensors llama_model_loader: - type q6_K: 21 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 8.63 GiB (5.02 BPW) load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 1 print_info: model type = ?B print_info: model params = 14.77 B print_info: general.name = Qwen3 14B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 llama_model_load: vocab only - skipping tensors time=2025-08-06T03:32:41.677Z level=INFO source=server.go:438 msg="starting llama server" cmd="/usr/bin/ollama runner --model /root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e --ctx-size 4096 --batch-size 512 --n-gpu-layers 41 --threads 8 --parallel 1 --port 45447" time=2025-08-06T03:32:41.677Z level=INFO source=sched.go:481 msg="loaded runners" count=1 time=2025-08-06T03:32:41.677Z level=INFO source=server.go:598 msg="waiting for llama runner to start responding" time=2025-08-06T03:32:41.678Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server not responding" time=2025-08-06T03:32:41.690Z level=INFO source=runner.go:815 msg="starting go runner" ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes load_backend: loaded CUDA backend from /usr/lib/ollama/libggml-cuda.so load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-icelake.so time=2025-08-06T03:32:41.749Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,600,610,700,750,800,860,870,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(gcc) time=2025-08-06T03:32:41.749Z level=INFO source=runner.go:874 msg="Server listening on 127.0.0.1:45447" llama_model_load_from_file_impl: using device CUDA0 (NVIDIA GeForce RTX 3060) - 11731 MiB free llama_model_loader: loaded meta data with 27 key-value pairs and 443 tensors from /root/.ollama/models/blobs/sha256-a8cc1361f3145dc01f6d77c6c82c9116b9ffe3c97b34716fe20418455876c40e (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 14B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 14B llama_model_loader: - kv 5: qwen3.block_count u32 = 40 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 17408 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 40 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 161 tensors llama_model_loader: - type f16: 40 tensors llama_model_loader: - type q4_K: 221 tensors llama_model_loader: - type q6_K: 21 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 8.63 GiB (5.02 BPW) time=2025-08-06T03:32:41.929Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server loading model" load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 0 print_info: n_ctx_train = 40960 print_info: n_embd = 5120 print_info: n_layer = 40 print_info: n_head = 40 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: n_swa_pattern = 1 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 5 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-06 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 17408 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = 0 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 40960 print_info: rope_finetuned = unknown print_info: ssm_d_conv = 0 print_info: ssm_d_inner = 0 print_info: ssm_d_state = 0 print_info: ssm_dt_rank = 0 print_info: ssm_dt_b_c_rms = 0 print_info: model type = 14B print_info: model params = 14.77 B print_info: general.name = Qwen3 14B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = true) load_tensors: offloading 40 repeating layers to GPU load_tensors: offloading output layer to GPU load_tensors: offloaded 41/41 layers to GPU load_tensors: CUDA0 model buffer size = 8423.47 MiB load_tensors: CPU_Mapped model buffer size = 417.30 MiB llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 4096 llama_context: n_ctx_per_seq = 4096 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = 0 llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 llama_context: n_ctx_per_seq (4096) < n_ctx_train (40960) -- the full capacity of the model will not be utilized llama_context: CUDA_Host output buffer size = 0.60 MiB llama_kv_cache_unified: kv_size = 4096, type_k = 'f16', type_v = 'f16', n_layer = 40, can_shift = 1, padding = 32 llama_kv_cache_unified: CUDA0 KV buffer size = 640.00 MiB llama_kv_cache_unified: KV self size = 640.00 MiB, K (f16): 320.00 MiB, V (f16): 320.00 MiB llama_context: CUDA0 compute buffer size = 368.00 MiB llama_context: CUDA_Host compute buffer size = 18.01 MiB llama_context: graph nodes = 1526 llama_context: graph splits = 2 time=2025-08-06T03:32:47.200Z level=INFO source=server.go:637 msg="llama runner started in 5.52 seconds" [GIN] 2025/08/06 - 03:32:52 | 200 | 10.98706337s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:33:26 | 200 | 18.320083604s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:33:53 | 200 | 13.387027252s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:35:03 | 200 | 1.594868544s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:35:57 | 200 | 5.442233ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:35:57 | 200 | 63.13µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:36:03 | 200 | 43.825µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:36:08 | 200 | 48.027µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:36:17 | 200 | 7.44382486s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:36:34 | 200 | 16.825114491s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:36:46 | 200 | 2.246994ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:36:46 | 200 | 71.171µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:37:01 | 200 | 26.755593194s | 172.17.0.1 | POST "/api/chat" time=2025-08-06T03:37:03.245Z level=INFO source=sched.go:546 msg="updated VRAM based on existing loaded models" gpu=GPU-8876e5e1-cf0e-d479-b427-0fe7fcb38333 library=cuda total="11.6 GiB" available="1.6 GiB" [GIN] 2025/08/06 - 03:37:20 | 200 | 2.591057ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:20 | 200 | 48.328µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:37:21 | 200 | 38.474µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:37:21 | 200 | 2.641725ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:21 | 200 | 41.419µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:37:24 | 200 | 23.389278513s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:25 | 200 | 2.369078ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:25 | 200 | 21.353µs | 172.17.0.1 | GET "/api/ps" time=2025-08-06T03:37:25.243Z level=INFO source=sched.go:786 msg="new model will fit in available VRAM in single GPU, loading" model=/root/.ollama/models/blobs/sha256-38e8dcc30df4eb0e29eaf5c74ba6ce3f2cd66badad50768fc14362acfb8b8cb6 gpu=GPU-8876e5e1-cf0e-d479-b427-0fe7fcb38333 parallel=1 available=12301828096 required="5.4 GiB" time=2025-08-06T03:37:25.366Z level=INFO source=server.go:135 msg="system memory" total="31.2 GiB" free="26.8 GiB" free_swap="1.2 GiB" time=2025-08-06T03:37:25.367Z level=INFO source=server.go:175 msg=offload library=cuda layers.requested=-1 layers.model=36 layers.offload=36 layers.split="" memory.available="[11.5 GiB]" memory.gpu_overhead="0 B" memory.required.full="5.4 GiB" memory.required.partial="5.4 GiB" memory.required.kv="280.0 MiB" memory.required.allocations="[5.4 GiB]" memory.weights.total="2.6 GiB" memory.weights.repeating="2.2 GiB" memory.weights.nonrepeating="420.4 MiB" memory.graph.full="2.0 GiB" memory.graph.partial="3.7 GiB" time=2025-08-06T03:37:25.421Z level=INFO source=server.go:438 msg="starting llama server" cmd="/usr/bin/ollama runner --ollama-engine --model /root/.ollama/models/blobs/sha256-38e8dcc30df4eb0e29eaf5c74ba6ce3f2cd66badad50768fc14362acfb8b8cb6 --ctx-size 4096 --batch-size 512 --n-gpu-layers 36 --threads 8 --parallel 1 --port 39807" time=2025-08-06T03:37:25.421Z level=INFO source=sched.go:481 msg="loaded runners" count=1 time=2025-08-06T03:37:25.421Z level=INFO source=server.go:598 msg="waiting for llama runner to start responding" time=2025-08-06T03:37:25.421Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server not responding" time=2025-08-06T03:37:25.431Z level=INFO source=runner.go:925 msg="starting ollama engine" time=2025-08-06T03:37:25.431Z level=INFO source=runner.go:983 msg="Server listening on 127.0.0.1:39807" time=2025-08-06T03:37:25.494Z level=INFO source=ggml.go:92 msg="" architecture=gemma3n file_type=Q4_K_M name="" description="" num_tensors=847 num_key_values=40 ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes load_backend: loaded CUDA backend from /usr/lib/ollama/libggml-cuda.so load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-icelake.so time=2025-08-06T03:37:25.539Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=500,600,610,700,750,800,860,870,890,900,1200 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(gcc) time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:365 msg="offloading 35 repeating layers to GPU" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:371 msg="offloading output layer to GPU" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:376 msg="offloaded 36/36 layers to GPU" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:379 msg="model weights" buffer=CPU size="420.4 MiB" time=2025-08-06T03:37:25.620Z level=INFO source=ggml.go:379 msg="model weights" buffer=CUDA0 size="7.0 GiB" time=2025-08-06T03:37:25.627Z level=INFO source=ggml.go:668 msg="compute graph" backend=CUDA0 buffer_type=CUDA0 size="135.5 MiB" time=2025-08-06T03:37:25.627Z level=INFO source=ggml.go:668 msg="compute graph" backend=CPU buffer_type=CPU size="4.0 MiB" time=2025-08-06T03:37:25.672Z level=INFO source=server.go:632 msg="waiting for server to become available" status="llm server loading model" time=2025-08-06T03:37:28.181Z level=INFO source=server.go:637 msg="llama runner started in 2.76 seconds" [GIN] 2025/08/06 - 03:37:29 | 200 | 26.550326175s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:30 | 200 | 1.394084314s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:31 | 200 | 654.35957ms | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:32 | 200 | 1.072350544s | 172.17.0.1 | POST "/api/chat" [GIN] 2025/08/06 - 03:37:49 | 200 | 3.118832ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:37:49 | 200 | 39.63µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:39:36 | 200 | 2.815778ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:39:36 | 200 | 32.844µs | 172.17.0.1 | GET "/api/ps" [GIN] 2025/08/06 - 03:40:27 | 200 | 48.372µs | 172.17.0.1 | GET "/api/version" [GIN] 2025/08/06 - 03:43:45 | 200 | 3.171724ms | 172.17.0.1 | GET "/api/tags" [GIN] 2025/08/06 - 03:43:45 | 200 | 32.527µs | 172.17.0.1 | GET "/api/ps"` some logs from the openwebui containter ` 2025-08-06 03:37:20.635 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:20.635 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:20.638 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/pinned HTTP/1.1" 200 - {} 2025-08-06 03:37:20.651 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:20.651 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:20.660 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:21.024 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:37:21.031 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/configs/banners HTTP/1.1" 200 - {} 2025-08-06 03:37:21.037 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/tools/ HTTP/1.1" 200 - {} 2025-08-06 03:37:21.059 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /ollama/api/version HTTP/1.1" 200 - {} 2025-08-06 03:37:21.094 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/user/settings HTTP/1.1" 200 - {} 2025-08-06 03:37:21.209 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/version/updates HTTP/1.1" 200 - {} 2025-08-06 03:37:21.476 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:21.476 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:21.808 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:37:23.677 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:37:25.154 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:25.155 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:25.383 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:37:28.374 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/chat/completions HTTP/1.1" 200 - {} 2025-08-06 03:37:28.497 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:28.522 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:29.488 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/v1/audio/speech HTTP/1.1" 500 - {} Exception in ASGI application Traceback (most recent call last): File "/usr/local/lib/python3.11/site-packages/uvicorn/protocols/http/httptools_impl.py", line 409, in run_asgi result = await app( # type: ignore[func-returns-value] ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/uvicorn/middleware/proxy_headers.py", line 60, in __call__ return await self.app(scope, receive, send) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/fastapi/applications.py", line 1054, in __call__ await super().__call__(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/applications.py", line 112, in __call__ await self.middleware_stack(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/errors.py", line 187, in __call__ raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/errors.py", line 165, in __call__ await self.app(scope, receive, _send) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/cors.py", line 93, in __call__ await self.simple_response(scope, receive, send, request_headers=headers) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/cors.py", line 144, in simple_response await self.app(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1162, in inspect_websocket return await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1141, in check_url response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1127, in commit_session_after_request response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/utils/security_headers.py", line 11, in dispatch response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 177, in __call__ with recv_stream, send_stream, collapse_excgroups(): File "/usr/local/lib/python3.11/contextlib.py", line 158, in __exit__ self.gen.throw(typ, value, traceback) File "/usr/local/lib/python3.11/site-packages/starlette/_utils.py", line 82, in collapse_excgroups raise exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 179, in __call__ response = await self.dispatch_func(request, call_next) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/main.py", line 1113, in dispatch response = await call_next(request) ^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 154, in call_next raise app_exc File "/usr/local/lib/python3.11/site-packages/starlette/middleware/base.py", line 141, in coro await self.app(scope, receive_or_disconnect, send_no_error) File "/usr/local/lib/python3.11/site-packages/starlette_compress/__init__.py", line 92, in __call__ return await self._zstd(scope, receive, send) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/starlette_compress/_zstd_legacy.py", line 100, in __call__ await self.app(scope, receive, wrapper) File "/usr/local/lib/python3.11/site-packages/starlette/middleware/exceptions.py", line 62, in __call__ await wrap_app_handling_exceptions(self.app, conn)(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app raise exc File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app await app(scope, receive, sender) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 715, in __call__ await self.middleware_stack(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 735, in app await route.handle(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 288, in handle await self.app(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 76, in app await wrap_app_handling_exceptions(app, request)(scope, receive, send) File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 53, in wrapped_app raise exc File "/usr/local/lib/python3.11/site-packages/starlette/_exception_handler.py", line 42, in wrapped_app await app(scope, receive, sender) File "/usr/local/lib/python3.11/site-packages/starlette/routing.py", line 73, in app response = await f(request) ^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/fastapi/routing.py", line 301, in app raw_response = await run_endpoint_function( ^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/fastapi/routing.py", line 212, in run_endpoint_function return await dependant.call(**values) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/app/backend/open_webui/routers/audio.py", line 517, in speech load_speech_pipeline(request) File "/app/backend/open_webui/routers/audio.py", line 303, in load_speech_pipeline request.app.state.speech_speaker_embeddings_dataset = load_dataset( ^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 1392, in load_dataset builder_instance = load_dataset_builder( ^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 1132, in load_dataset_builder dataset_module = dataset_module_factory( ^^^^^^^^^^^^^^^^^^^^^^^ File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 1031, in dataset_module_factory raise e1 from None File "/usr/local/lib/python3.11/site-packages/datasets/load.py", line 989, in dataset_module_factory raise RuntimeError(f"Dataset scripts are no longer supported, but found {filename}") RuntimeError: Dataset scripts are no longer supported, but found cmu-arctic-xvectors.py 2025-08-06 03:37:29.496 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /audio/notification.mp3 HTTP/1.1" 206 - {} 2025-08-06 03:37:29.496 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/chat/completed HTTP/1.1" 200 - {} 2025-08-06 03:37:29.517 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /static/apple-touch-icon.png HTTP/1.1" 200 - {} 2025-08-06 03:37:29.525 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "POST /api/v1/chats/0f79969d-8c97-49cf-88e9-6511e03be0dc HTTP/1.1" 200 - {} 2025-08-06 03:37:29.543 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:29.565 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:31.494 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:31.502 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:31.547 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/?page=1 HTTP/1.1" 200 - {} 2025-08-06 03:37:31.585 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:37:32.578 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/all/tags HTTP/1.1" 200 - {} 2025-08-06 03:37:32.681 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/0f79969d-8c97-49cf-88e9-6511e03be0dc HTTP/1.1" 200 - {} 2025-08-06 03:37:32.706 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /api/v1/chats/all/tags HTTP/1.1" 200 - {} 2025-08-06 03:37:49.732 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:37:49.733 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:37:49.984 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:38:08.921 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /manifest.json HTTP/1.1" 200 - {} 2025-08-06 03:38:08.921 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/0.BEiiGd97.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.921 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/purify.DQwrSZtH.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Collapsible.DZAVxvaL.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Modal.CyLKLEmt.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Skeleton.Cr25vAl_.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.922 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/ConfirmDialog.kaSBQ3kP.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.924 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Messages.Bbuu_qma.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.925 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/MapSelector.CIGW-MKW.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.926 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Drawer.Bz8t_RYw.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.926 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/Chat.CWA0G-PE.css HTTP/1.1" 304 - {} 2025-08-06 03:38:08.926 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/2.BXeiR8Q0.css HTTP/1.1" 304 - {} 2025-08-06 03:38:20.264 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:38:23.683 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:39:20.972 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:39:23.684 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:39:36.388 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:39:36.388 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:39:36.486 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/0f79969d-8c97-49cf-88e9-6511e03be0dc HTTP/1.1" 200 - {} 2025-08-06 03:39:36.510 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2a5699a7-9cc3-4745-9f4d-7b408f9879dc HTTP/1.1" 200 - {} 2025-08-06 03:39:36.517 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/10cfaedf-d0db-414e-90a6-1297c4a1a67a HTTP/1.1" 200 - {} 2025-08-06 03:39:36.548 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e7b1a31a-3226-46de-8e70-90a46a1b1e62 HTTP/1.1" 200 - {} 2025-08-06 03:39:36.556 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/541d0f9a-0a0a-4e8f-be1a-804c3189519f HTTP/1.1" 200 - {} 2025-08-06 03:39:36.576 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e05b6db7-1b3c-4e06-b622-6aa375565b9d HTTP/1.1" 200 - {} 2025-08-06 03:39:36.598 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2d5d7d91-54d6-44e4-a7f5-9f8fb1975a44 HTTP/1.1" 200 - {} 2025-08-06 03:39:36.716 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/291b1336-1016-441e-bbfc-4c6dc5e82046 HTTP/1.1" 200 - {} 2025-08-06 03:39:36.839 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:39:37.039 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/0b5a3a3d-0581-497d-b287-9806e3ee2dc0 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.053 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e1bac2cf-887e-473e-abdb-dd515ba10161 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.056 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/77cf441b-f231-4c73-a4a8-4254c6d5e002 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.056 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/09f66bae-2d38-43ec-b39a-a2586972b726 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.234 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/291a2e20-84eb-438a-9f67-e81a9de76c90 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.241 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/16ddce7c-f97b-4ba4-a028-d4503eaa6935 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.241 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/08fa0b98-f121-4673-843e-06069fde3dbb HTTP/1.1" 200 - {} 2025-08-06 03:39:37.242 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/f2fc367c-4f70-4085-9254-1f2a185171cb HTTP/1.1" 200 - {} 2025-08-06 03:39:37.242 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1e9ca672-5dca-499f-8b2b-3c5b7d45c66b HTTP/1.1" 200 - {} 2025-08-06 03:39:37.270 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/c61ad61d-0aa7-475e-8614-da28d088fe57 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.270 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/8c40c89b-06e9-48bd-81e6-c4051d630e14 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.270 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/9a784cd1-e81f-411a-8cd7-d750ca3831c2 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.273 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/5c15e756-7963-4cd6-95a7-1fd03e45b00c HTTP/1.1" 200 - {} 2025-08-06 03:39:37.274 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/08ea4579-4d0d-449a-bec2-6f7d27bbff74 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.274 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/ff7f8aa5-fd0a-4856-a7b6-a08c96509767 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.303 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/bd9c328f-abd4-4b25-b8d8-348bcfd38306 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.304 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/5baf6686-1ab4-46b4-9b15-4ca653e6bd97 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.304 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/f7178b16-5b4f-47c7-82dd-b129ded7c5df HTTP/1.1" 200 - {} 2025-08-06 03:39:37.304 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/8f37e3e9-a360-49df-a4c4-76b8fef2a0e4 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.309 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/69d7e923-cd9f-445b-a9f4-5dbdd87452d1 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.310 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/4dc97d6b-6522-44d7-a385-e0b6f1b43ad4 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.335 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/975c1a64-356a-4444-a6e0-db8cfbfad5c1 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/f01d1aba-2723-4aec-8e40-6acbab8b53b3 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/331da5b4-9211-4cab-98fc-5033cc25b585 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e8cff7db-4d78-40bd-b5ba-418a7899608e HTTP/1.1" 200 - {} 2025-08-06 03:39:37.336 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/9c501207-9049-48cd-97a8-78f9a701ee67 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/b6ec8a47-52fe-4067-8a10-1707fe13328d HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2000cd81-6fdf-4306-adb2-db00e771997d HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/9d5c76fa-cd37-4e5f-a0ce-25ce956874f5 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.369 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/19.D5tH4moG.js HTTP/1.1" 200 - {} 2025-08-06 03:39:37.370 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/17131d5c-e521-498b-bf7b-70cd5764bf04 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.377 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/89c333b2-6d31-4040-9f1d-68116372c5f3 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1969a05b-3362-4fe0-b0eb-4fabef4ad838 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/2775eab7-5937-409b-a50a-ab34e877757f HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/e6acdce3-ff67-41d1-9f39-b1ea1d40849a HTTP/1.1" 200 - {} 2025-08-06 03:39:37.389 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/8105e1ca-28d0-47ed-8daa-546e687e5ee0 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.443 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/caa37dc6-c410-4934-a260-1368cb4240e6 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.443 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/bad3f560-e73e-4cbf-8ae0-614e910ac555 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.444 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/c87a7b7c-c2cf-41e7-8ca5-e4f5efcbf1f7 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.448 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/16b06372-d991-4511-a68a-a6846e278693 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.450 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/0fe5cbc8-93af-43a2-821e-ad867ed61cf7 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.450 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/26d2e83d-7d8d-4251-a24d-4ce15c4dcc52 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.450 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/a0350689-d3cc-4cb4-afa3-339061fc5392 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.458 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/dc49270b-f60a-47c7-a266-ee4a635bf986 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.459 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/?page=2 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.459 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/02587867-f7b4-402f-a303-581f45beecec HTTP/1.1" 200 - {} 2025-08-06 03:39:37.459 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/22a29d68-c6d9-417c-9e67-312fe658b3b7 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.460 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/02ddc96e-b97f-4bd2-a0ce-cbabb0ebc088 HTTP/1.1" 200 - {} 2025-08-06 03:39:37.460 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/7424616b-2d3c-44b4-830d-e230548b45ad HTTP/1.1" 200 - {} 2025-08-06 03:39:37.464 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1ba04b95-d3de-4dbf-b90e-0de68e3684cd HTTP/1.1" 200 - {} 2025-08-06 03:39:37.476 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/folders/ HTTP/1.1" 200 - {} 2025-08-06 03:39:38.005 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:39:38.025 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:39:38.465 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/chats/1a144301-ac99-48c6-8408-faf79f1d9545 HTTP/1.1" 200 - {} 2025-08-06 03:39:38.489 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:39:39.446 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/3.DQTEdc7F.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.446 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/9.OrSiIAXf.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.475 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/assets/29.wEbTgpRj.css HTTP/1.1" 200 - {} 2025-08-06 03:39:39.475 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/Ck0cvSaj.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.476 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/17.DB2DlSzW.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.478 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/DQtrEHa0.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.478 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/V4NzNuWG.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.479 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/TZ3cQ_RY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.498 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:39.500 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/18.DPUDtlSY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:39.504 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:39.523 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:39.529 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/?page=1&order_by=created_at&direction=asc HTTP/1.1" 200 - {} 2025-08-06 03:39:41.013 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/C83syoBY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.014 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/CICvqL9d.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.014 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/DbZwnr2B.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.016 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/CFy5Jpyl.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.017 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/15.BhTxcmdY.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.017 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/Dg4-BTkH.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.017 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/CeDBf-bV.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.275 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/webhook HTTP/1.1" 200 - {} 2025-08-06 03:39:41.277 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap/server HTTP/1.1" 200 - {} 2025-08-06 03:39:41.278 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config HTTP/1.1" 200 - {} 2025-08-06 03:39:41.281 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/16.DqmmaYsb.js HTTP/1.1" 200 - {} 2025-08-06 03:39:41.291 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap HTTP/1.1" 200 - {} 2025-08-06 03:39:41.309 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/webhook HTTP/1.1" 200 - {} 2025-08-06 03:39:41.310 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap/server HTTP/1.1" 200 - {} 2025-08-06 03:39:41.312 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config HTTP/1.1" 200 - {} 2025-08-06 03:39:41.322 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/auths/admin/config/ldap HTTP/1.1" 200 - {} 2025-08-06 03:39:41.498 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/version/updates HTTP/1.1" 200 - {} 2025-08-06 03:39:41.604 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/version/updates HTTP/1.1" 200 - {} 2025-08-06 03:40:21.205 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:40:23.687 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:40:26.605 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/nodes/10.C41L3GeI.js HTTP/1.1" 200 - {} 2025-08-06 03:40:26.606 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/Cfz980HS.js HTTP/1.1" 200 - {} 2025-08-06 03:40:26.607 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/immutable/chunks/HKxhIwoV.js HTTP/1.1" 200 - {} 2025-08-06 03:40:27.346 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /ollama/api/version HTTP/1.1" 200 - {} 2025-08-06 03:40:27.388 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/v1/users/user/settings HTTP/1.1" 200 - {} 2025-08-06 03:40:28.695 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/usage HTTP/1.1" 200 - {} 2025-08-06 03:41:21.977 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:41:23.687 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:42:22.982 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:42:23.715 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:43:23.692 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:43:23.971 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:43:45.819 | INFO | open_webui.routers.openai:get_all_models:392 - get_all_models() - {} 2025-08-06 03:43:45.819 | INFO | open_webui.routers.ollama:get_all_models:344 - get_all_models() - {} 2025-08-06 03:43:46.031 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /api/models HTTP/1.1" 200 - {} 2025-08-06 03:44:23.690 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:44:23.983 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:45:23.694 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:45:24.981 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:46:23.697 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:46:25.971 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:47:23.697 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:47:25.976 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:48:23.697 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:48:25.991 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:49:23.703 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:49:26.980 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:50:23.704 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.125:0 - "GET /_app/version.json HTTP/1.1" 200 - {} 2025-08-06 03:50:26.981 | INFO | uvicorn.protocols.http.httptools_impl:send:476 - 10.0.0.126:0 - "GET /_app/version.json HTTP/1.1" 200 - {} ` browser logs (pretty much nothing) `[tiptap warn]: Duplicate extension names found: ['codeBlock', 'bulletList', 'listItem', 'listKeymap', 'orderedList']. This can lead to issues.` ### Additional Information _No response_
GiteaMirror added the bug label 2026-04-25 07:17:58 -05:00
Author
Owner

@tjbck commented on GitHub (Aug 6, 2025):

Unable to reproduce from our end, keep us updated!

<!-- gh-comment-id:3158235231 --> @tjbck commented on GitHub (Aug 6, 2025): Unable to reproduce from our end, keep us updated!
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#33391