update context length format.

This commit is contained in:
Maternion
2026-02-10 10:13:04 +05:30
committed by Jesse Gross
parent c980e19995
commit 2dbb000908

View File

@@ -6,9 +6,9 @@ Context length is the maximum number of tokens that the model has access to in m
<Note>
Ollama defaults to the following context lengths based on VRAM:
- < 24 GiB VRAM: 4,096 context
- 24-48 GiB VRAM: 32,768 context
- &gt;= 48 GiB VRAM: 262,144 context
- < 24 GiB VRAM: 4k context
- 24-48 GiB VRAM: 32k context
- &gt;= 48 GiB VRAM: 256k context
</Note>
Tasks which require large context like web search, agents, and coding tools should be set to at least 64000 tokens.