[GH-ISSUE #7566] Having trouble with vram using priority #51330

Closed
opened 2026-04-28 19:30:19 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @morika546 on GitHub (Nov 8, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/7566

What is the issue?

My GPU is 6800xt with 16 GB VRAM,and my RAM is 64 GB.When I running the model,It always using so many shared VRAM while there is still a lot of free dedicated VRAM.For example, when I run a 13 b sized model,it needs 12GB VRAM,and only 4GB Runs by dedicated VRAM,the resist 8GB runs by shared VRAM,That makes even the processor is 100% GPU,the model still runs in a very very slow speed.
I have checked the serve.log,and doubt if this makes the problem:

time=2024-11-06T22:18:35.707+08:00 level=INFO source=sched.go:185 msg="one or more GPUs detected that are unable to accurately report free memory - disabling default concurrency

I am new to ollama,and have been suffering from the bug for two weeks,I will appreciate it if anyone can help me.Here is the full serve.log:

2_server.log

OS

Windows

GPU

AMD

CPU

AMD

Ollama version

0.4.0

Originally created by @morika546 on GitHub (Nov 8, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/7566 ### What is the issue? My GPU is 6800xt with 16 GB VRAM,and my RAM is 64 GB.When I running the model,It always using so many shared VRAM while there is still a lot of free dedicated VRAM.For example, when I run a 13 b sized model,it needs 12GB VRAM,and only 4GB Runs by dedicated VRAM,the resist 8GB runs by shared VRAM,That makes even the processor is 100% GPU,the model still runs in a very very slow speed. I have checked the serve.log,and doubt if this makes the problem: time=2024-11-06T22:18:35.707+08:00 level=INFO source=sched.go:185 msg="one or more GPUs detected that are unable to accurately report free memory - disabling default concurrency I am new to ollama,and have been suffering from the bug for two weeks,I will appreciate it if anyone can help me.Here is the full serve.log: [2_server.log](https://github.com/user-attachments/files/17672177/2_server.log) ### OS Windows ### GPU AMD ### CPU AMD ### Ollama version 0.4.0
GiteaMirror added the bug label 2026-04-28 19:30:19 -05:00
Author
Owner

@morika546 commented on GitHub (Nov 8, 2024):

solved by reinstall Redon gpu driver 24.8.1

<!-- gh-comment-id:2464207995 --> @morika546 commented on GitHub (Nov 8, 2024): solved by reinstall Redon gpu driver 24.8.1
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#51330