mirror of
https://github.com/ollama/ollama.git
synced 2026-03-11 17:34:04 -05:00
Replace binary low VRAM mode with tiered VRAM thresholds that set default context lengths for all models: - < 24 GiB VRAM: 4,096 context - 24-48 GiB VRAM: 32,768 context - >= 48 GiB VRAM: 262,144 context
3.0 KiB
3.0 KiB