[GH-ISSUE #20994] issue: ollama version in open-webui:ollama docker container is stuck at 0.13.5 #34883

Closed
opened 2026-04-25 09:03:48 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @wyattearp on GitHub (Jan 28, 2026).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/20994

Check Existing Issues

  • I have searched for any existing and/or related issues.
  • I have searched for any existing and/or related discussions.
  • I have also searched in the CLOSED issues AND CLOSED discussions and found no related items (your issue might already be addressed on the development branch!).
  • I am using the latest version of Open WebUI.

Installation Method

Git Clone

Open WebUI Version

docker pull ghcr.io/open-webui/open-webui:ollama (latest)

Ollama Version (if applicable)

0.13.5

Operating System

Ubuntu 24.04 (arm64)

Browser (if applicable)

Ubuntu 24.04 (arm64)

Confirmation

  • I have read and followed all instructions in README.md.
  • I am using the latest version of both Open WebUI and Ollama.
  • I have included the browser console logs.
  • I have included the Docker container logs.
  • I have provided every relevant configuration, setting, and environment variable used in my setup.
  • I have clearly listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc).
  • I have documented step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation. My steps:
  • Start with the initial platform/version/OS and dependencies used,
  • Specify exact install/launch/configure commands,
  • List URLs visited, user input (incl. example values/emails/passwords if needed),
  • Describe all options and toggles enabled or changed,
  • Include any files or environmental changes,
  • Identify the expected and actual result at each stage,
  • Ensure any reasonably skilled user can follow and hit the same issue.

Expected Behavior

Ollama version is at the latest available (0.14 or greater)

Actual Behavior

Ollama is stuck at version 0.13.5

Steps to Reproduce

  1. docker pull ghcr.io/open-webui/open-webui:ollama
  2. docker stop open-webui
  3. docker container rm open-webui
  4. docker run -d -p 12000:8080 -p 11434:11434 --gpus=all -v open-webui:/app/backend/data -v open-webui-ollama:/root/.ollama -e "OLLAMA_HOST=0.0.0.0" -e "OLLAMA_BASE_URL=http://127.0.0.1:11434" --name open-webui ghcr.io/open-webui/open-webui:ollama
  5. docker exec -it open-webui /bin/bash
  6. ollama --version
wyatt@spark-6f4b:~$ cat openwebui-ollama-startup.sh 
#!/bin/bash
docker pull ghcr.io/open-webui/open-webui:ollama
docker stop open-webui 
docker container rm open-webui
docker run -d -p 12000:8080 -p 11434:11434 --gpus=all   -v open-webui:/app/backend/data   -v open-webui-ollama:/root/.ollama   -e "OLLAMA_HOST=0.0.0.0"   -e "OLLAMA_BASE_URL=http://127.0.0.1:11434"   --name open-webui ghcr.io/open-webui/open-webui:ollama
wyatt@spark-6f4b:~$ cat /etc/lsb-release 
DISTRIB_ID=Ubuntu
DISTRIB_RELEASE=24.04
DISTRIB_CODENAME=noble
DISTRIB_DESCRIPTION="Ubuntu 24.04.3 LTS"
wyatt@spark-6f4b:~$ docker exec -it open-webui /bin/bash
root@6f4aa5d849bc:/app/backend# ollama --version
ollama version is 0.13.5

Logs & Screenshots

Loading WEBUI_SECRET_KEY from file, not provided as an environment variable.
Generating WEBUI_SECRET_KEY
Loading WEBUI_SECRET_KEY from .webui_secret_key
USE_OLLAMA is set to true, starting ollama serve.
time=2026-01-28T05:21:01.570Z level=INFO source=routes.go:1554 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false OLLAMA_VULKAN:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]"
time=2026-01-28T05:21:01.571Z level=INFO source=images.go:493 msg="total blobs: 23"
time=2026-01-28T05:21:01.571Z level=INFO source=images.go:500 msg="total unused blobs removed: 0"
time=2026-01-28T05:21:01.571Z level=INFO source=routes.go:1607 msg="Listening on [::]:11434 (version 0.13.5)"
time=2026-01-28T05:21:01.572Z level=INFO source=runner.go:67 msg="discovering available GPUs..."
time=2026-01-28T05:21:01.572Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 44219"
time=2026-01-28T05:21:01.945Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 44527"
time=2026-01-28T05:21:02.239Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 39193"
time=2026-01-28T05:21:02.240Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 36283"
time=2026-01-28T05:21:02.653Z level=INFO source=types.go:42 msg="inference compute" id=GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f filter_id="" library=CUDA compute=12.1 name=CUDA0 description="NVIDIA GB10" libdirs=ollama,cuda_v13 driver=13.0 pci_id=000f:01:00.0 type=iGPU total="119.7 GiB" available="115.3 GiB"
INFO  [alembic.runtime.migration] Context impl SQLiteImpl.
INFO  [alembic.runtime.migration] Will assume non-transactional DDL.
WARNI [open_webui.env] 

WARNING: CORS_ALLOW_ORIGIN IS SET TO '*' - NOT RECOMMENDED FOR PRODUCTION DEPLOYMENTS.

WARNI [langchain_community.utils.user_agent] USER_AGENT environment variable not set, consider setting it to identify your requests.

 ██████╗ ██████╗ ███████╗███╗   ██╗    ██╗    ██╗███████╗██████╗ ██╗   ██╗██╗
██╔═══██╗██╔══██╗██╔════╝████╗  ██║    ██║    ██║██╔════╝██╔══██╗██║   ██║██║
██║   ██║██████╔╝█████╗  ██╔██╗ ██║    ██║ █╗ ██║█████╗  ██████╔╝██║   ██║██║
██║   ██║██╔═══╝ ██╔══╝  ██║╚██╗██║    ██║███╗██║██╔══╝  ██╔══██╗██║   ██║██║
╚██████╔╝██║     ███████╗██║ ╚████║    ╚███╔███╔╝███████╗██████╔╝╚██████╔╝██║
 ╚═════╝ ╚═╝     ╚══════╝╚═╝  ╚═══╝     ╚══╝╚══╝ ╚══════╝╚═════╝  ╚═════╝ ╚═╝


v0.7.2 - building the best AI user interface.

https://github.com/open-webui/open-webui

Fetching 30 files: 100%|██████████| 30/30 [00:00<00:00, 24432.84it/s]
INFO:     Started server process [1]
INFO:     Waiting for application startup.
2026-01-28 05:21:09.819 | INFO     | open_webui.utils.logger:start_logger:166 - GLOBAL_LOG_LEVEL: INFO
2026-01-28 05:21:09.819 | INFO     | open_webui.main:lifespan:604 - Installing external dependencies of functions and tools...
2026-01-28 05:21:09.823 | INFO     | open_webui.utils.plugin:install_frontmatter_requirements:286 - No requirements found in frontmatter.
time=2026-01-28T05:21:31.456Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 38609"
time=2026-01-28T05:21:31.819Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax"
time=2026-01-28T05:21:31.965Z level=INFO source=server.go:245 msg="enabling flash attention"
time=2026-01-28T05:21:31.965Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --model /root/.ollama/models/blobs/sha256-e7b273f9636059a689e3ddcab3716e4f65abe0143ac978e46673ad0e52d09efb --port 36259"
time=2026-01-28T05:21:31.965Z level=INFO source=sched.go:443 msg="system memory" total="119.7 GiB" free="118.8 GiB" free_swap="16.0 GiB"
time=2026-01-28T05:21:31.965Z level=INFO source=sched.go:450 msg="gpu memory" id=GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f library=CUDA available="114.4 GiB" free="114.8 GiB" minimum="457.0 MiB" overhead="0 B"
time=2026-01-28T05:21:31.965Z level=INFO source=server.go:746 msg="loading model" "model layers"=25 requested=-1
time=2026-01-28T05:21:31.973Z level=INFO source=runner.go:1405 msg="starting ollama engine"
time=2026-01-28T05:21:31.973Z level=INFO source=runner.go:1440 msg="Server listening on 127.0.0.1:36259"
time=2026-01-28T05:21:31.977Z level=INFO source=runner.go:1278 msg=load request="{Operation:fit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:Enabled KvSize:8192 KvCacheType: NumThreads:20 GPULayers:25[ID:GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f Layers:25(0..24)] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2026-01-28T05:21:32.015Z level=INFO source=ggml.go:136 msg="" architecture=gptoss file_type=MXFP4 name="" description="" num_tensors=459 num_key_values=32
load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu.so
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
  Device 0: NVIDIA GB10, compute capability 12.1, VMM: yes, ID: GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f
load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v13/libggml-cuda.so
time=2026-01-28T05:21:32.298Z level=INFO source=ggml.go:104 msg=system CPU.0.NEON=1 CPU.0.ARM_FMA=1 CPU.0.LLAMAFILE=1 CPU.1.NEON=1 CPU.1.ARM_FMA=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang)
time=2026-01-28T05:21:49.277Z level=INFO source=runner.go:1278 msg=load request="{Operation:alloc LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:Enabled KvSize:8192 KvCacheType: NumThreads:20 GPULayers:25[ID:GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f Layers:25(0..24)] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2026-01-28T05:21:49.718Z level=INFO source=runner.go:1278 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:Enabled KvSize:8192 KvCacheType: NumThreads:20 GPULayers:25[ID:GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f Layers:25(0..24)] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}"
time=2026-01-28T05:21:49.718Z level=INFO source=ggml.go:482 msg="offloading 24 repeating layers to GPU"
time=2026-01-28T05:21:49.718Z level=INFO source=ggml.go:489 msg="offloading output layer to GPU"
time=2026-01-28T05:21:49.718Z level=INFO source=ggml.go:494 msg="offloaded 25/25 layers to GPU"
time=2026-01-28T05:21:49.718Z level=INFO source=device.go:240 msg="model weights" device=CUDA0 size="11.8 GiB"
time=2026-01-28T05:21:49.718Z level=INFO source=device.go:245 msg="model weights" device=CPU size="1.1 GiB"
time=2026-01-28T05:21:49.718Z level=INFO source=device.go:251 msg="kv cache" device=CUDA0 size="300.0 MiB"
time=2026-01-28T05:21:49.718Z level=INFO source=device.go:262 msg="compute graph" device=CUDA0 size="124.6 MiB"
time=2026-01-28T05:21:49.718Z level=INFO source=device.go:267 msg="compute graph" device=CPU size="5.6 MiB"
time=2026-01-28T05:21:49.718Z level=INFO source=device.go:272 msg="total memory" size="13.3 GiB"
time=2026-01-28T05:21:49.718Z level=INFO source=sched.go:517 msg="loaded runners" count=1
time=2026-01-28T05:21:49.718Z level=INFO source=server.go:1338 msg="waiting for llama runner to start responding"
time=2026-01-28T05:21:49.721Z level=INFO source=server.go:1372 msg="waiting for server to become available" status="llm server loading model"
time=2026-01-28T05:21:50.473Z level=INFO source=server.go:1376 msg="llama runner started in 18.51 seconds"
[GIN] 2026/01/28 - 05:21:56 | 200 | 25.489044258s | 192.168.100.252 | POST     "/v1/chat/completions"
[GIN] 2026/01/28 - 05:22:35 | 404 |         4.8µs | 192.168.100.252 | POST     "/v1/messages/count_tokens?beta=true"
[GIN] 2026/01/28 - 05:22:35 | 404 |       1.376µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:22:38 | 404 |       4.304µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:22:38 | 404 |       4.208µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:22:38 | 404 |       3.856µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:23:02 | 404 |       4.032µs | 192.168.100.252 | POST     "/v1/messages/count_tokens?beta=true"
[GIN] 2026/01/28 - 05:23:02 | 404 |       2.016µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:23:48 | 404 |           4µs | 192.168.100.252 | POST     "/v1/messages/count_tokens?beta=true"
[GIN] 2026/01/28 - 05:23:48 | 404 |       1.184µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:23:50 | 404 |       4.032µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:23:50 | 404 |       3.664µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:23:50 | 404 |       3.312µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:24:18 | 404 |        3.36µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:24:18 | 404 |       3.936µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:24:18 | 404 |       3.664µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:24:19 | 404 |       4.304µs | 192.168.100.252 | POST     "/v1/messages/count_tokens?beta=true"
[GIN] 2026/01/28 - 05:24:19 | 404 |       3.664µs | 192.168.100.252 | POST     "/v1/messages?beta=true"
[GIN] 2026/01/28 - 05:24:28 | 404 |       3.952µs | 192.168.100.252 | POST     "/v1/messages"
[GIN] 2026/01/28 - 05:27:12 | 200 |      44.733µs | 192.168.100.252 | GET      "/api/version"
[GIN] 2026/01/28 - 05:27:31 | 200 |      34.302µs | 192.168.100.252 | GET      "/api/version"
[GIN] 2026/01/28 - 05:31:06 | 200 |      31.823µs |       127.0.0.1 | GET      "/api/version"
[GIN] 2026/01/28 - 05:31:14 | 200 |      27.407µs |       127.0.0.1 | HEAD     "/"
[GIN] 2026/01/28 - 05:31:14 | 200 |     802.931µs |       127.0.0.1 | GET      "/api/tags"
[GIN] 2026/01/28 - 12:16:35 | 200 |      33.265µs |       127.0.0.1 | GET      "/api/version"

Additional Information

Talking with @Kemon in the discord revealed there is a package dependency that is holding back ollama:


Is should have had the latest! Tested ollama and dev-ollama images, both has the 0.13.5 version.
Looks like Ollama requires another package that has stopped the installation of the newest version.
Workaround until it will be fixed.

Enter docker container
docker exec -it open-webui bash

Update and install required package
apt update
apt install zstd

Install/Update Ollama:
curl -fsSL https://ollama.com/install.sh | sh

Restart the container:
kill 1 (or just restart it normally)

When doing this ollama is updated:

root@6f4aa5d849bc:/app/backend# curl -fsSL https://ollama.com/install.sh | sh
>>> Cleaning up old version at /usr/local/lib/ollama
>>> Installing ollama to /usr/local
>>> Downloading ollama-linux-arm64.tar.zst
######################################################################## 100.0%
WARNING: Unable to detect NVIDIA/AMD GPU. Install lspci or lshw to automatically detect and install GPU dependencies.
>>> The Ollama API is now available at 127.0.0.1:11434.
>>> Install complete. Run "ollama" from the command line.
root@6f4aa5d849bc:/app/backend# ollama version
Error: unknown command "version" for "ollama"
root@6f4aa5d849bc:/app/backend# ollama version
Error: unknown command "version" for "ollama"
root@6f4aa5d849bc:/app/backend# kill 1
root@6f4aa5d849bc:/app/backend# wyatt@spark-6f4b:~$ docker restart open-webui 
open-webui
wyatt@spark-6f4b:~$ docker exec -it open-webui /bin/bash
root@6f4aa5d849bc:/app/backend# ollama --version
ollama version is 0.15.2
root@6f4aa5d849bc:/app/backend# 

NOTE: I'm not sure if this is because the :ollama tag is no longer being used / supported OR if it's really an issue with the build tag.

Originally created by @wyattearp on GitHub (Jan 28, 2026). Original GitHub issue: https://github.com/open-webui/open-webui/issues/20994 ### Check Existing Issues - [x] I have searched for any existing and/or related issues. - [x] I have searched for any existing and/or related discussions. - [x] I have also searched in the CLOSED issues AND CLOSED discussions and found no related items (your issue might already be addressed on the development branch!). - [x] I am using the latest version of Open WebUI. ### Installation Method Git Clone ### Open WebUI Version docker pull ghcr.io/open-webui/open-webui:ollama (latest) ### Ollama Version (if applicable) 0.13.5 ### Operating System Ubuntu 24.04 (arm64) ### Browser (if applicable) Ubuntu 24.04 (arm64) ### Confirmation - [x] I have read and followed all instructions in `README.md`. - [x] I am using the latest version of **both** Open WebUI and Ollama. - [x] I have included the browser console logs. - [x] I have included the Docker container logs. - [x] I have **provided every relevant configuration, setting, and environment variable used in my setup.** - [x] I have clearly **listed every relevant configuration, custom setting, environment variable, and command-line option that influences my setup** (such as Docker Compose overrides, .env values, browser settings, authentication configurations, etc). - [x] I have documented **step-by-step reproduction instructions that are precise, sequential, and leave nothing to interpretation**. My steps: - Start with the initial platform/version/OS and dependencies used, - Specify exact install/launch/configure commands, - List URLs visited, user input (incl. example values/emails/passwords if needed), - Describe all options and toggles enabled or changed, - Include any files or environmental changes, - Identify the expected and actual result at each stage, - Ensure any reasonably skilled user can follow and hit the same issue. ### Expected Behavior Ollama version is at the latest available (0.14 or greater) ### Actual Behavior Ollama is stuck at version 0.13.5 ### Steps to Reproduce 1. docker pull ghcr.io/open-webui/open-webui:ollama 2. docker stop open-webui 3. docker container rm open-webui 4. docker run -d -p 12000:8080 -p 11434:11434 --gpus=all -v open-webui:/app/backend/data -v open-webui-ollama:/root/.ollama -e "OLLAMA_HOST=0.0.0.0" -e "OLLAMA_BASE_URL=http://127.0.0.1:11434" --name open-webui ghcr.io/open-webui/open-webui:ollama 5. docker exec -it open-webui /bin/bash 6. ollama --version ``` wyatt@spark-6f4b:~$ cat openwebui-ollama-startup.sh #!/bin/bash docker pull ghcr.io/open-webui/open-webui:ollama docker stop open-webui docker container rm open-webui docker run -d -p 12000:8080 -p 11434:11434 --gpus=all -v open-webui:/app/backend/data -v open-webui-ollama:/root/.ollama -e "OLLAMA_HOST=0.0.0.0" -e "OLLAMA_BASE_URL=http://127.0.0.1:11434" --name open-webui ghcr.io/open-webui/open-webui:ollama wyatt@spark-6f4b:~$ cat /etc/lsb-release DISTRIB_ID=Ubuntu DISTRIB_RELEASE=24.04 DISTRIB_CODENAME=noble DISTRIB_DESCRIPTION="Ubuntu 24.04.3 LTS" wyatt@spark-6f4b:~$ docker exec -it open-webui /bin/bash root@6f4aa5d849bc:/app/backend# ollama --version ollama version is 0.13.5 ``` ### Logs & Screenshots ``` Loading WEBUI_SECRET_KEY from file, not provided as an environment variable. Generating WEBUI_SECRET_KEY Loading WEBUI_SECRET_KEY from .webui_secret_key USE_OLLAMA is set to true, starting ollama serve. time=2026-01-28T05:21:01.570Z level=INFO source=routes.go:1554 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false OLLAMA_VULKAN:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2026-01-28T05:21:01.571Z level=INFO source=images.go:493 msg="total blobs: 23" time=2026-01-28T05:21:01.571Z level=INFO source=images.go:500 msg="total unused blobs removed: 0" time=2026-01-28T05:21:01.571Z level=INFO source=routes.go:1607 msg="Listening on [::]:11434 (version 0.13.5)" time=2026-01-28T05:21:01.572Z level=INFO source=runner.go:67 msg="discovering available GPUs..." time=2026-01-28T05:21:01.572Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 44219" time=2026-01-28T05:21:01.945Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 44527" time=2026-01-28T05:21:02.239Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 39193" time=2026-01-28T05:21:02.240Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 36283" time=2026-01-28T05:21:02.653Z level=INFO source=types.go:42 msg="inference compute" id=GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f filter_id="" library=CUDA compute=12.1 name=CUDA0 description="NVIDIA GB10" libdirs=ollama,cuda_v13 driver=13.0 pci_id=000f:01:00.0 type=iGPU total="119.7 GiB" available="115.3 GiB" INFO [alembic.runtime.migration] Context impl SQLiteImpl. INFO [alembic.runtime.migration] Will assume non-transactional DDL. WARNI [open_webui.env] WARNING: CORS_ALLOW_ORIGIN IS SET TO '*' - NOT RECOMMENDED FOR PRODUCTION DEPLOYMENTS. WARNI [langchain_community.utils.user_agent] USER_AGENT environment variable not set, consider setting it to identify your requests. ██████╗ ██████╗ ███████╗███╗ ██╗ ██╗ ██╗███████╗██████╗ ██╗ ██╗██╗ ██╔═══██╗██╔══██╗██╔════╝████╗ ██║ ██║ ██║██╔════╝██╔══██╗██║ ██║██║ ██║ ██║██████╔╝█████╗ ██╔██╗ ██║ ██║ █╗ ██║█████╗ ██████╔╝██║ ██║██║ ██║ ██║██╔═══╝ ██╔══╝ ██║╚██╗██║ ██║███╗██║██╔══╝ ██╔══██╗██║ ██║██║ ╚██████╔╝██║ ███████╗██║ ╚████║ ╚███╔███╔╝███████╗██████╔╝╚██████╔╝██║ ╚═════╝ ╚═╝ ╚══════╝╚═╝ ╚═══╝ ╚══╝╚══╝ ╚══════╝╚═════╝ ╚═════╝ ╚═╝ v0.7.2 - building the best AI user interface. https://github.com/open-webui/open-webui Fetching 30 files: 100%|██████████| 30/30 [00:00<00:00, 24432.84it/s] INFO: Started server process [1] INFO: Waiting for application startup. 2026-01-28 05:21:09.819 | INFO | open_webui.utils.logger:start_logger:166 - GLOBAL_LOG_LEVEL: INFO 2026-01-28 05:21:09.819 | INFO | open_webui.main:lifespan:604 - Installing external dependencies of functions and tools... 2026-01-28 05:21:09.823 | INFO | open_webui.utils.plugin:install_frontmatter_requirements:286 - No requirements found in frontmatter. time=2026-01-28T05:21:31.456Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --port 38609" time=2026-01-28T05:21:31.819Z level=WARN source=cpu_linux.go:130 msg="failed to parse CPU allowed micro secs" error="strconv.ParseInt: parsing \"max\": invalid syntax" time=2026-01-28T05:21:31.965Z level=INFO source=server.go:245 msg="enabling flash attention" time=2026-01-28T05:21:31.965Z level=INFO source=server.go:429 msg="starting runner" cmd="/usr/local/bin/ollama runner --ollama-engine --model /root/.ollama/models/blobs/sha256-e7b273f9636059a689e3ddcab3716e4f65abe0143ac978e46673ad0e52d09efb --port 36259" time=2026-01-28T05:21:31.965Z level=INFO source=sched.go:443 msg="system memory" total="119.7 GiB" free="118.8 GiB" free_swap="16.0 GiB" time=2026-01-28T05:21:31.965Z level=INFO source=sched.go:450 msg="gpu memory" id=GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f library=CUDA available="114.4 GiB" free="114.8 GiB" minimum="457.0 MiB" overhead="0 B" time=2026-01-28T05:21:31.965Z level=INFO source=server.go:746 msg="loading model" "model layers"=25 requested=-1 time=2026-01-28T05:21:31.973Z level=INFO source=runner.go:1405 msg="starting ollama engine" time=2026-01-28T05:21:31.973Z level=INFO source=runner.go:1440 msg="Server listening on 127.0.0.1:36259" time=2026-01-28T05:21:31.977Z level=INFO source=runner.go:1278 msg=load request="{Operation:fit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:Enabled KvSize:8192 KvCacheType: NumThreads:20 GPULayers:25[ID:GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f Layers:25(0..24)] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2026-01-28T05:21:32.015Z level=INFO source=ggml.go:136 msg="" architecture=gptoss file_type=MXFP4 name="" description="" num_tensors=459 num_key_values=32 load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu.so ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 1 CUDA devices: Device 0: NVIDIA GB10, compute capability 12.1, VMM: yes, ID: GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v13/libggml-cuda.so time=2026-01-28T05:21:32.298Z level=INFO source=ggml.go:104 msg=system CPU.0.NEON=1 CPU.0.ARM_FMA=1 CPU.0.LLAMAFILE=1 CPU.1.NEON=1 CPU.1.ARM_FMA=1 CPU.1.LLAMAFILE=1 CUDA.0.ARCHS=750,800,860,870,890,900,1000,1030,1100,1200,1210 CUDA.0.USE_GRAPHS=1 CUDA.0.PEER_MAX_BATCH_SIZE=128 compiler=cgo(clang) time=2026-01-28T05:21:49.277Z level=INFO source=runner.go:1278 msg=load request="{Operation:alloc LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:Enabled KvSize:8192 KvCacheType: NumThreads:20 GPULayers:25[ID:GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f Layers:25(0..24)] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2026-01-28T05:21:49.718Z level=INFO source=runner.go:1278 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:Enabled KvSize:8192 KvCacheType: NumThreads:20 GPULayers:25[ID:GPU-3c00ea44-acfe-fd2a-855b-c3190f90fc4f Layers:25(0..24)] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2026-01-28T05:21:49.718Z level=INFO source=ggml.go:482 msg="offloading 24 repeating layers to GPU" time=2026-01-28T05:21:49.718Z level=INFO source=ggml.go:489 msg="offloading output layer to GPU" time=2026-01-28T05:21:49.718Z level=INFO source=ggml.go:494 msg="offloaded 25/25 layers to GPU" time=2026-01-28T05:21:49.718Z level=INFO source=device.go:240 msg="model weights" device=CUDA0 size="11.8 GiB" time=2026-01-28T05:21:49.718Z level=INFO source=device.go:245 msg="model weights" device=CPU size="1.1 GiB" time=2026-01-28T05:21:49.718Z level=INFO source=device.go:251 msg="kv cache" device=CUDA0 size="300.0 MiB" time=2026-01-28T05:21:49.718Z level=INFO source=device.go:262 msg="compute graph" device=CUDA0 size="124.6 MiB" time=2026-01-28T05:21:49.718Z level=INFO source=device.go:267 msg="compute graph" device=CPU size="5.6 MiB" time=2026-01-28T05:21:49.718Z level=INFO source=device.go:272 msg="total memory" size="13.3 GiB" time=2026-01-28T05:21:49.718Z level=INFO source=sched.go:517 msg="loaded runners" count=1 time=2026-01-28T05:21:49.718Z level=INFO source=server.go:1338 msg="waiting for llama runner to start responding" time=2026-01-28T05:21:49.721Z level=INFO source=server.go:1372 msg="waiting for server to become available" status="llm server loading model" time=2026-01-28T05:21:50.473Z level=INFO source=server.go:1376 msg="llama runner started in 18.51 seconds" [GIN] 2026/01/28 - 05:21:56 | 200 | 25.489044258s | 192.168.100.252 | POST "/v1/chat/completions" [GIN] 2026/01/28 - 05:22:35 | 404 | 4.8µs | 192.168.100.252 | POST "/v1/messages/count_tokens?beta=true" [GIN] 2026/01/28 - 05:22:35 | 404 | 1.376µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:22:38 | 404 | 4.304µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:22:38 | 404 | 4.208µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:22:38 | 404 | 3.856µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:23:02 | 404 | 4.032µs | 192.168.100.252 | POST "/v1/messages/count_tokens?beta=true" [GIN] 2026/01/28 - 05:23:02 | 404 | 2.016µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:23:48 | 404 | 4µs | 192.168.100.252 | POST "/v1/messages/count_tokens?beta=true" [GIN] 2026/01/28 - 05:23:48 | 404 | 1.184µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:23:50 | 404 | 4.032µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:23:50 | 404 | 3.664µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:23:50 | 404 | 3.312µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:24:18 | 404 | 3.36µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:24:18 | 404 | 3.936µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:24:18 | 404 | 3.664µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:24:19 | 404 | 4.304µs | 192.168.100.252 | POST "/v1/messages/count_tokens?beta=true" [GIN] 2026/01/28 - 05:24:19 | 404 | 3.664µs | 192.168.100.252 | POST "/v1/messages?beta=true" [GIN] 2026/01/28 - 05:24:28 | 404 | 3.952µs | 192.168.100.252 | POST "/v1/messages" [GIN] 2026/01/28 - 05:27:12 | 200 | 44.733µs | 192.168.100.252 | GET "/api/version" [GIN] 2026/01/28 - 05:27:31 | 200 | 34.302µs | 192.168.100.252 | GET "/api/version" [GIN] 2026/01/28 - 05:31:06 | 200 | 31.823µs | 127.0.0.1 | GET "/api/version" [GIN] 2026/01/28 - 05:31:14 | 200 | 27.407µs | 127.0.0.1 | HEAD "/" [GIN] 2026/01/28 - 05:31:14 | 200 | 802.931µs | 127.0.0.1 | GET "/api/tags" [GIN] 2026/01/28 - 12:16:35 | 200 | 33.265µs | 127.0.0.1 | GET "/api/version" ``` ### Additional Information Talking with @Kemon in the discord revealed there is a package dependency that is holding back ollama: ``` Is should have had the latest! Tested ollama and dev-ollama images, both has the 0.13.5 version. Looks like Ollama requires another package that has stopped the installation of the newest version. Workaround until it will be fixed. Enter docker container docker exec -it open-webui bash Update and install required package apt update apt install zstd Install/Update Ollama: curl -fsSL https://ollama.com/install.sh | sh Restart the container: kill 1 (or just restart it normally) ``` When doing this ollama is updated: ``` root@6f4aa5d849bc:/app/backend# curl -fsSL https://ollama.com/install.sh | sh >>> Cleaning up old version at /usr/local/lib/ollama >>> Installing ollama to /usr/local >>> Downloading ollama-linux-arm64.tar.zst ######################################################################## 100.0% WARNING: Unable to detect NVIDIA/AMD GPU. Install lspci or lshw to automatically detect and install GPU dependencies. >>> The Ollama API is now available at 127.0.0.1:11434. >>> Install complete. Run "ollama" from the command line. root@6f4aa5d849bc:/app/backend# ollama version Error: unknown command "version" for "ollama" root@6f4aa5d849bc:/app/backend# ollama version Error: unknown command "version" for "ollama" root@6f4aa5d849bc:/app/backend# kill 1 root@6f4aa5d849bc:/app/backend# wyatt@spark-6f4b:~$ docker restart open-webui open-webui wyatt@spark-6f4b:~$ docker exec -it open-webui /bin/bash root@6f4aa5d849bc:/app/backend# ollama --version ollama version is 0.15.2 root@6f4aa5d849bc:/app/backend# ``` **NOTE**: I'm not sure if this is because the :ollama tag is no longer being used / supported OR if it's really an issue with the build tag.
GiteaMirror added the bug label 2026-04-25 09:03:48 -05:00
Author
Owner

@owui-terminator[bot] commented on GitHub (Jan 28, 2026):

🔍 Similar Issues Found

I found some existing issues that might be related to this one. Please check if any of these are duplicates or contain helpful solutions:

  1. #20901 issue: Upgrade to 0.7.2 breaks API access
    by huornlmj • Jan 23, 2026 • bug

  2. #20327 issue: Unable to use any Open WebUI version newer than 0.6.25 due to hybrid search performance
    by galvanoid • Jan 02, 2026 • bug

  3. #16496 issue: Stale ollama version in official docker images
    by robmurrer • Aug 11, 2025 • bug

  4. #18145 issue: 0.6.33 regression
    by Ark-Levy • Oct 08, 2025 • bug

  5. #14529 issue: Open WebUI does not work on versions after version 0.6.7
    by OpenSoftware-World • May 30, 2025 • bug

Show 5 more related issues
  1. #19755 issue: Error: 404, message='Not Found', url='http://ollama:11434/api/embed'
    by stevewillett • Dec 04, 2025 • bug

  2. #19376 issue: open-webui docker compose service does not resolve ollama service name
    by SorenPoulsen • Nov 21, 2025 • bug

  3. #19417 issue: v0.6.37 SQL Error
    by AKHYP • Nov 24, 2025 • bug

  4. #19496 issue: 500 internal server error appears in v0.6.40
    by cloudtuotuo • Nov 26, 2025 • bug

  5. #18084 issue: DEV docker image broken
    by by-lin • Oct 06, 2025 • bug


💡 Tips:

  • If this is a duplicate, please consider closing this issue and adding any additional details to the existing one
  • If you found a solution in any of these issues, please share it here to help others

This comment was generated automatically by a bot. Please react with a 👍 if this comment was helpful, or a 👎 if it was not.

<!-- gh-comment-id:3811007227 --> @owui-terminator[bot] commented on GitHub (Jan 28, 2026): 🔍 **Similar Issues Found** I found some existing issues that might be related to this one. Please check if any of these are duplicates or contain helpful solutions: 1. [#20901](https://github.com/open-webui/open-webui/issues/20901) **issue: Upgrade to 0.7.2 breaks API access** *by huornlmj • Jan 23, 2026 • `bug`* 2. [#20327](https://github.com/open-webui/open-webui/issues/20327) **issue: Unable to use any Open WebUI version newer than 0.6.25 due to hybrid search performance** *by galvanoid • Jan 02, 2026 • `bug`* 3. [#16496](https://github.com/open-webui/open-webui/issues/16496) **issue: Stale ollama version in official docker images** *by robmurrer • Aug 11, 2025 • `bug`* 4. [#18145](https://github.com/open-webui/open-webui/issues/18145) **issue: 0.6.33 regression** *by Ark-Levy • Oct 08, 2025 • `bug`* 5. [#14529](https://github.com/open-webui/open-webui/issues/14529) **issue: Open WebUI does not work on versions after version 0.6.7** *by OpenSoftware-World • May 30, 2025 • `bug`* <details> <summary>Show 5 more related issues</summary> 6. [#19755](https://github.com/open-webui/open-webui/issues/19755) **issue: Error: 404, message='Not Found', url='http://ollama:11434/api/embed'** *by stevewillett • Dec 04, 2025 • `bug`* 7. [#19376](https://github.com/open-webui/open-webui/issues/19376) **issue: open-webui docker compose service does not resolve ollama service name** *by SorenPoulsen • Nov 21, 2025 • `bug`* 8. [#19417](https://github.com/open-webui/open-webui/issues/19417) **issue: v0.6.37 SQL Error** *by AKHYP • Nov 24, 2025 • `bug`* 9. [#19496](https://github.com/open-webui/open-webui/issues/19496) **issue: 500 internal server error appears in v0.6.40** *by cloudtuotuo • Nov 26, 2025 • `bug`* 10. [#18084](https://github.com/open-webui/open-webui/issues/18084) **issue: DEV docker image broken** *by by-lin • Oct 06, 2025 • `bug`* </details> --- 💡 **Tips:** - If this is a duplicate, please consider closing this issue and adding any additional details to the existing one - If you found a solution in any of these issues, please share it here to help others *This comment was generated automatically by a bot.* Please react with a 👍 if this comment was helpful, or a 👎 if it was not.
Author
Owner

@tjbck commented on GitHub (Jan 29, 2026):

This requires version update.

<!-- gh-comment-id:3818702073 --> @tjbck commented on GitHub (Jan 29, 2026): This requires version update.
Author
Owner

@wyattearp commented on GitHub (Jan 30, 2026):

Note that the 0.15.2 version of ollama is running into some other issues with claude code: https://github.com/ollama/ollama/issues/13949

<!-- gh-comment-id:3821122545 --> @wyattearp commented on GitHub (Jan 30, 2026): Note that the 0.15.2 version of ollama is running into some other issues with claude code: https://github.com/ollama/ollama/issues/13949
Author
Owner

@wyattearp commented on GitHub (Jan 30, 2026):

@tjbck - the reason the USE_OLLAMA=true doesn't build anymore is because it's missing the zstd package - will you open this back up and I'll submit a PR to fix it.

<!-- gh-comment-id:3821340737 --> @wyattearp commented on GitHub (Jan 30, 2026): @tjbck - the reason the `USE_OLLAMA=true` doesn't build anymore is because it's missing the `zstd` package - will you open this back up and I'll submit a PR to fix it.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#34883