[PR #14263] docs: add GPU troubleshooting and VRAM sizing guidance #45841

Open
opened 2026-04-25 01:27:42 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/14263
Author: @akuligowski9
Created: 2/14/2026
Status: 🔄 Open

Base: mainHead: docs/gpu-troubleshooting-vram-guide


📝 Commits (1)

  • 5bd633e docs: add GPU troubleshooting and VRAM sizing guidance

📊 Changes

1 file changed (+50 additions, -0 deletions)

View changed files

📝 docs/troubleshooting.mdx (+50 -0)

📄 Description

Summary

  • Adds "Model Running Slower Than Expected" section to docs/troubleshooting.mdx
  • Covers how to check GPU utilization with ollama ps
  • Explains common reasons for CPU fallback (model too large, context too high, other models loaded)
  • Includes approximate VRAM requirements table for 4-bit quantized models (1B through 70B)
  • Lists actionable steps when a model doesn't fit in VRAM

Context

Resolves #14260

The existing troubleshooting docs cover GPU discovery failures but not the far more common scenario where the GPU is detected but VRAM is insufficient. This fills the gap where most confused users land. Related issues: #4809 #9774 #8144 #14257 #6864 #4996

Test plan

  • Documentation renders correctly
  • VRAM estimates are reasonable for listed model sizes at 4-bit quantization

🤖 Generated with Claude Code


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/14263 **Author:** [@akuligowski9](https://github.com/akuligowski9) **Created:** 2/14/2026 **Status:** 🔄 Open **Base:** `main` ← **Head:** `docs/gpu-troubleshooting-vram-guide` --- ### 📝 Commits (1) - [`5bd633e`](https://github.com/ollama/ollama/commit/5bd633e25a78a238daff349bf49333016e98e13f) docs: add GPU troubleshooting and VRAM sizing guidance ### 📊 Changes **1 file changed** (+50 additions, -0 deletions) <details> <summary>View changed files</summary> 📝 `docs/troubleshooting.mdx` (+50 -0) </details> ### 📄 Description ## Summary - Adds "Model Running Slower Than Expected" section to `docs/troubleshooting.mdx` - Covers how to check GPU utilization with `ollama ps` - Explains common reasons for CPU fallback (model too large, context too high, other models loaded) - Includes approximate VRAM requirements table for 4-bit quantized models (1B through 70B) - Lists actionable steps when a model doesn't fit in VRAM ## Context Resolves #14260 The existing troubleshooting docs cover GPU discovery failures but not the far more common scenario where the GPU is detected but VRAM is insufficient. This fills the gap where most confused users land. Related issues: #4809 #9774 #8144 #14257 #6864 #4996 ## Test plan - [ ] Documentation renders correctly - [ ] VRAM estimates are reasonable for listed model sizes at 4-bit quantization 🤖 Generated with [Claude Code](https://claude.com/claude-code) --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-25 01:27:42 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#45841