[GH-ISSUE #8277] mistral-nemo - context window 1024000? #5295

Open
opened 2026-04-12 16:28:41 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @mjaniec2013 on GitHub (Dec 31, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/8277

What is the issue?

model_name='mistral-nemo'
ollama.show(model_name)['modelinfo']

{'general.architecture': 'llama',
'general.basename': 'Mistral-Nemo',
'general.file_type': 2,
'general.finetune': 'Instruct',
'general.languages': ['en', 'fr', 'de', 'es', 'it', 'pt', 'ru', 'zh', 'ja'],
'general.license': 'apache-2.0',
'general.parameter_count': 12247782400,
'general.quantization_version': 2,
'general.size_label': '12B',
'general.type': 'model',
'general.version': '2407',
'llama.attention.head_count': 32,
'llama.attention.head_count_kv': 8,
'llama.attention.key_length': 128,
'llama.attention.layer_norm_rms_epsilon': 1e-05,
'llama.attention.value_length': 128,
'llama.block_count': 40,
'llama.context_length': 1024000, <<<<<
'llama.embedding_length': 5120,
'llama.feed_forward_length': 14336,
'llama.rope.dimension_count': 128,
'llama.rope.freq_base': 1000000,
'llama.vocab_size': 131072,
'tokenizer.ggml.add_bos_token': True,
'tokenizer.ggml.add_eos_token': False,
'tokenizer.ggml.add_space_prefix': False,
'tokenizer.ggml.bos_token_id': 1,
'tokenizer.ggml.eos_token_id': 2,
'tokenizer.ggml.merges': None,
'tokenizer.ggml.model': 'gpt2',
'tokenizer.ggml.pre': 'tekken',
'tokenizer.ggml.token_type': None,
'tokenizer.ggml.tokens': None,
'tokenizer.ggml.unknown_token_id': 0}

Shouldn't it be 128k instead of 1M+?

OS

Windows

GPU

Nvidia

CPU

Intel

Ollama version

0.5.4

Originally created by @mjaniec2013 on GitHub (Dec 31, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/8277 ### What is the issue? model_name='mistral-nemo' ollama.show(model_name)['modelinfo'] {'general.architecture': 'llama', 'general.basename': 'Mistral-Nemo', 'general.file_type': 2, 'general.finetune': 'Instruct', 'general.languages': ['en', 'fr', 'de', 'es', 'it', 'pt', 'ru', 'zh', 'ja'], 'general.license': 'apache-2.0', 'general.parameter_count': 12247782400, 'general.quantization_version': 2, 'general.size_label': '12B', 'general.type': 'model', 'general.version': '2407', 'llama.attention.head_count': 32, 'llama.attention.head_count_kv': 8, 'llama.attention.key_length': 128, 'llama.attention.layer_norm_rms_epsilon': 1e-05, 'llama.attention.value_length': 128, 'llama.block_count': 40, 'llama.context_length': 1024000, <<<<< 'llama.embedding_length': 5120, 'llama.feed_forward_length': 14336, 'llama.rope.dimension_count': 128, 'llama.rope.freq_base': 1000000, 'llama.vocab_size': 131072, 'tokenizer.ggml.add_bos_token': True, 'tokenizer.ggml.add_eos_token': False, 'tokenizer.ggml.add_space_prefix': False, 'tokenizer.ggml.bos_token_id': 1, 'tokenizer.ggml.eos_token_id': 2, 'tokenizer.ggml.merges': None, 'tokenizer.ggml.model': 'gpt2', 'tokenizer.ggml.pre': 'tekken', 'tokenizer.ggml.token_type': None, 'tokenizer.ggml.tokens': None, 'tokenizer.ggml.unknown_token_id': 0} Shouldn't it be 128k instead of 1M+? ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version 0.5.4
GiteaMirror added the bug label 2026-04-12 16:28:41 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#5295