[GH-ISSUE #2742] How to improve ollama performance #63688

Closed
opened 2026-05-03 14:40:52 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @gautam-fairpe on GitHub (Feb 25, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/2742

Originally assigned to: @dhiltgen on GitHub.

current model params :
FROM llama2:13b-chat
PARAMETER temperature 0.2
PARAMETER num_ctx 4096
PARAMETER num_thread 16
PARAMETER use_mmap False

System config :
Ram 108 GB
T4 graphics card 16 gb
Screenshot from 2024-02-25 17-57-04

Also hardly any ram is being used.
Using ollama python bindings to get the result but due to some params issue not getting the result as expected. What am i missing here ?

Originally created by @gautam-fairpe on GitHub (Feb 25, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/2742 Originally assigned to: @dhiltgen on GitHub. current model params : FROM llama2:13b-chat PARAMETER temperature 0.2 PARAMETER num_ctx 4096 PARAMETER num_thread 16 PARAMETER use_mmap False System config : Ram 108 GB T4 graphics card 16 gb ![Screenshot from 2024-02-25 17-57-04](https://github.com/ollama/ollama/assets/127822235/24854715-93b3-4732-9b6f-bd1a373a9417) Also hardly any ram is being used. Using ollama python bindings to get the result but due to some params issue not getting the result as expected. What am i missing here ?
GiteaMirror added the performancenvidia labels 2026-05-03 14:40:53 -05:00
Author
Owner

@easp commented on GitHub (Feb 25, 2024):

Look in the log file to see what it says about GPU detection and model layers being offloaded to GPU

https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues

<!-- gh-comment-id:1963006601 --> @easp commented on GitHub (Feb 25, 2024): Look in the log file to see what it says about GPU detection and model layers being offloaded to GPU https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues
Author
Owner

@dhiltgen commented on GitHub (Mar 11, 2024):

If you're still having performance problems, please share your server log so we can better understand what's going wrong.

(also please make sure to upgrade to the latest version in case a bug fix resolves the problem)

<!-- gh-comment-id:1989466317 --> @dhiltgen commented on GitHub (Mar 11, 2024): If you're still having performance problems, please share your server log so we can better understand what's going wrong. (also please make sure to upgrade to the latest version in case a bug fix resolves the problem)
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#63688