[GH-ISSUE #5477] How to setting N GPU usage? #65461

Closed
opened 2026-05-03 21:22:18 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @CaoYunzhou on GitHub (Jul 4, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/5477

What is the issue?

why use one GPU
I have four gpu device
image

OS

Linux

GPU

Nvidia

CPU

Intel

Ollama version

0.1.48

Originally created by @CaoYunzhou on GitHub (Jul 4, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/5477 ### What is the issue? why use one GPU I have four gpu device ![image](https://github.com/ollama/ollama/assets/28099773/775bb006-e7de-48cf-b001-98a49ff66595) ### OS Linux ### GPU Nvidia ### CPU Intel ### Ollama version 0.1.48
GiteaMirror added the bug label 2026-05-03 21:22:19 -05:00
Author
Owner

@CaoYunzhou commented on GitHub (Jul 4, 2024):

0   N/A  N/A     65952      C   ...unners/cuda_v11/ollama_llama_server      39122MiB
maybe only use number 0 deviced
<!-- gh-comment-id:2208195810 --> @CaoYunzhou commented on GitHub (Jul 4, 2024): 0 N/A N/A 65952 C ...unners/cuda_v11/ollama_llama_server 39122MiB maybe only use number 0 deviced
Author
Owner

@CaoYunzhou commented on GitHub (Jul 4, 2024):

I reslove find in issue
[Unit]
Description=Ollama Service
After=network-online.target

[Service]
ExecStart=/usr/local/bin/ollama serve
#Environment="CUDA_VISIBLE_DEVICES=32a47599-188e-dd99-4336-152ad66edb31,9e751506-fbbf-dc41-2475-60f78c4b87b8,db7c29cf-1f42-bbf7-04a4-c81b7eaa27b5,9bc1c429-31c1-8c38-7075-0de3e1992bd8"
#Environment="CUDA_VISIBLE_DEVICES=GPU-32a47599-188e-dd99-4336-152ad66edb31,GPU-9e751506-fbbf-dc41-2475-60f78c4b87b8,GPU-db7c29cf-1f42-bbf7-04a4-c81b7eaa27b5,GPU-9bc1c429-31c1-8c38-7075-0de3e1992bd8"
Environment="CUDA_VISIBLE_DEVICES=0,1,2,3"
Environment="OLLAMA_MAX_LOADED_MODELS=9"
Environment="OLLAMA_NUM_PARALLEL=100"
Environment="OLLAMA_HOST=0.0.0.0:11434"

User=ollama
Group=ollama
Restart=always
RestartSec=3
Environment="PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin"

[Install]
WantedBy=default.target

#-----------
need this env is working
Environment="CUDA_VISIBLE_DEVICES=0,1,2,3"
Environment="OLLAMA_MAX_LOADED_MODELS=9"
Environment="OLLAMA_NUM_PARALLEL=100"
Environment="OLLAMA_HOST=0.0.0.0:11434"

<!-- gh-comment-id:2208291659 --> @CaoYunzhou commented on GitHub (Jul 4, 2024): I reslove find in issue [Unit] Description=Ollama Service After=network-online.target [Service] ExecStart=/usr/local/bin/ollama serve #Environment="CUDA_VISIBLE_DEVICES=32a47599-188e-dd99-4336-152ad66edb31,9e751506-fbbf-dc41-2475-60f78c4b87b8,db7c29cf-1f42-bbf7-04a4-c81b7eaa27b5,9bc1c429-31c1-8c38-7075-0de3e1992bd8" #Environment="CUDA_VISIBLE_DEVICES=GPU-32a47599-188e-dd99-4336-152ad66edb31,GPU-9e751506-fbbf-dc41-2475-60f78c4b87b8,GPU-db7c29cf-1f42-bbf7-04a4-c81b7eaa27b5,GPU-9bc1c429-31c1-8c38-7075-0de3e1992bd8" Environment="CUDA_VISIBLE_DEVICES=0,1,2,3" Environment="OLLAMA_MAX_LOADED_MODELS=9" Environment="OLLAMA_NUM_PARALLEL=100" Environment="OLLAMA_HOST=0.0.0.0:11434" User=ollama Group=ollama Restart=always RestartSec=3 Environment="PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin" [Install] WantedBy=default.target #----------- need this env is working Environment="CUDA_VISIBLE_DEVICES=0,1,2,3" Environment="OLLAMA_MAX_LOADED_MODELS=9" Environment="OLLAMA_NUM_PARALLEL=100" Environment="OLLAMA_HOST=0.0.0.0:11434"
Author
Owner

@haier-1314 commented on GitHub (Jul 15, 2024):

I have the same problem, please tell me how to solve it

<!-- gh-comment-id:2227597649 --> @haier-1314 commented on GitHub (Jul 15, 2024): I have the same problem, please tell me how to solve it
Author
Owner

@CaoYunzhou commented on GitHub (Jul 16, 2024):

need this env is working
Environment="CUDA_VISIBLE_DEVICES=0,1,2,3"
Environment="OLLAMA_MAX_LOADED_MODELS=9"
Environment="OLLAMA_NUM_PARALLEL=100"
Environment="OLLAMA_HOST=0.0.0.0:11434"

in your system setting change this ollama env service
need this env is working
Environment="CUDA_VISIBLE_DEVICES=0,1,2,3"
Environment="OLLAMA_MAX_LOADED_MODELS=9"
Environment="OLLAMA_NUM_PARALLEL=100"
Environment="OLLAMA_HOST=0.0.0.0:11434"

<!-- gh-comment-id:2230584577 --> @CaoYunzhou commented on GitHub (Jul 16, 2024): > need this env is working > Environment="CUDA_VISIBLE_DEVICES=0,1,2,3" > Environment="OLLAMA_MAX_LOADED_MODELS=9" > Environment="OLLAMA_NUM_PARALLEL=100" > Environment="OLLAMA_HOST=0.0.0.0:11434" in your system setting change this ollama env service need this env is working Environment="CUDA_VISIBLE_DEVICES=0,1,2,3" Environment="OLLAMA_MAX_LOADED_MODELS=9" Environment="OLLAMA_NUM_PARALLEL=100" Environment="OLLAMA_HOST=0.0.0.0:11434"
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#65461