[GH-ISSUE #11014] Ollama v0.9 use higher VRAM than v0.68 #7261

Closed
opened 2026-04-12 19:18:36 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @indogood1 on GitHub (Jun 8, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/11014

What is the issue?

ver 9.0
D:>ollama ps
NAME ID SIZE PROCESSOR UNTIL
gemma3:12b 6fd036cefda5 14 GB 19%/81% CPU/GPU Forever
ver 0.6.8
D:>ollama ps
NAME ID SIZE PROCESSOR UNTIL
gemma3:12b 6fd036cefda5 14 GB 100% GPU Forever

D:>nvidia-smi
Sun Jun 8 11:29:27 2025
+-----------------------------------------------------------------------------------------+
| NVIDIA-SMI 572.42 Driver Version: 572.42 CUDA Version: 12.8 |
|-----------------------------------------+------------------------+----------------------+
| GPU Name Driver-Model | Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|=========================================+========================+======================|
| 0 NVIDIA CMP 70HX WDDM | 00000000:10:00.0 Off | N/A |
| 30% 39C P8 10W / 220W | 0MiB / 8192MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+
| 1 NVIDIA P104-100 WDDM | 00000000:2E:00.0 Off | N/A |
| 53% 41C P8 6W / 90W | 0MiB / 8192MiB | 0% Default |
| | | N/A |
+-----------------------------------------+------------------------+----------------------+

server-v0.9.log

server-v0.68.log

Relevant log output


OS

Windows

GPU

Nvidia

CPU

Intel, AMD

Ollama version

0.68 and 0.9

Originally created by @indogood1 on GitHub (Jun 8, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/11014 ### What is the issue? ver 9.0 D:\>ollama ps NAME ID SIZE PROCESSOR UNTIL gemma3:12b 6fd036cefda5 14 GB 19%/81% CPU/GPU Forever ver 0.6.8 D:\>ollama ps NAME ID SIZE PROCESSOR UNTIL gemma3:12b 6fd036cefda5 14 GB 100% GPU Forever D:\>nvidia-smi Sun Jun 8 11:29:27 2025 +-----------------------------------------------------------------------------------------+ | NVIDIA-SMI 572.42 Driver Version: 572.42 CUDA Version: 12.8 | |-----------------------------------------+------------------------+----------------------+ | GPU Name Driver-Model | Bus-Id Disp.A | Volatile Uncorr. ECC | | Fan Temp Perf Pwr:Usage/Cap | Memory-Usage | GPU-Util Compute M. | | | | MIG M. | |=========================================+========================+======================| | 0 NVIDIA CMP 70HX WDDM | 00000000:10:00.0 Off | N/A | | 30% 39C P8 10W / 220W | 0MiB / 8192MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ | 1 NVIDIA P104-100 WDDM | 00000000:2E:00.0 Off | N/A | | 53% 41C P8 6W / 90W | 0MiB / 8192MiB | 0% Default | | | | N/A | +-----------------------------------------+------------------------+----------------------+ [server-v0.9.log](https://github.com/user-attachments/files/20642823/server-v0.9.log) [server-v0.68.log](https://github.com/user-attachments/files/20642825/server-v0.68.log) ### Relevant log output ```shell ``` ### OS Windows ### GPU Nvidia ### CPU Intel, AMD ### Ollama version 0.68 and 0.9
GiteaMirror added the bug label 2026-04-12 19:18:36 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#7261