[GH-ISSUE #13796] Incorrect general.name metadata in GGUF files for llama3.3:70b variants #34798

Open
opened 2026-04-22 18:39:40 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @kenokall on GitHub (Jan 20, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/13796

What is the issue?

The GGUF files for the llama3.3:70b model (across multiple quantizations) have an incorrect general.name field in their metadata. The field states llama3.1 instead of the expected llama3.3.

Evidence:

  • Log output from loading the model shows the wrong name.
  • Inspecting the GGUF files linked from the manifest confirms the mismatch.
  • The llama3.3:8b variants correctly show llama3.3 in the same field.
  • The license file in the manifest correctly references Llama 3.3, suggesting this is a metadata error in the GGUF packing.

Impact: This may cause confusion in the logs or UIs that rely on the general.name field for identification.

Relevant log output

llama_model_loader: loaded meta data with 36 key-value pairs and 724 tensors from /root/.ollama/models/blobs/sha256-4824460d29f2058aaf6e1118a63a7a197a09bed509f0e7d4e2efb1ee273b447d (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
llama_model_loader: - kv   0:                       general.architecture str              = llama
llama_model_loader: - kv   1:                               general.type str              = model
llama_model_loader: - kv   2:                               general.name str              = Llama 3.1 70B Instruct 2024 12
llama_model_loader: - kv   3:                            general.version str              = 2024-12
llama_model_loader: - kv   4:                           general.finetune str              = Instruct
llama_model_loader: - kv   5:                           general.basename str              = Llama-3.1
llama_model_loader: - kv   6:                         general.size_label str              = 70B
llama_model_loader: - kv   7:                            general.license str              = llama3.1
llama_model_loader: - kv   8:                   general.base_model.count u32              = 1
llama_model_loader: - kv   9:                  general.base_model.0.name str              = Llama 3.1 70B
llama_model_loader: - kv  10:          general.base_model.0.organization str              = Meta Llama
llama_model_loader: - kv  11:              general.base_model.0.repo_url str              = https://huggingface.co/meta-llama/Lla...
llama_model_loader: - kv  12:                               general.tags arr[str,5]       = ["facebook", "meta", "pytorch", "llam...
llama_model_loader: - kv  13:                          general.languages arr[str,7]       = ["fr", "it", "pt", "hi", "es", "th", ...
llama_model_loader: - kv  14:                          llama.block_count u32              = 80
llama_model_loader: - kv  15:                       llama.context_length u32              = 131072
llama_model_loader: - kv  16:                     llama.embedding_length u32              = 8192
llama_model_loader: - kv  17:                  llama.feed_forward_length u32              = 28672
llama_model_loader: - kv  18:                 llama.attention.head_count u32              = 64
llama_model_loader: - kv  19:              llama.attention.head_count_kv u32              = 8
llama_model_loader: - kv  20:                       llama.rope.freq_base f32              = 500000.000000
llama_model_loader: - kv  21:     llama.attention.layer_norm_rms_epsilon f32              = 0.000010
llama_model_loader: - kv  22:                 llama.attention.key_length u32              = 128
llama_model_loader: - kv  23:               llama.attention.value_length u32              = 128
llama_model_loader: - kv  24:                          general.file_type u32              = 15
llama_model_loader: - kv  25:                           llama.vocab_size u32              = 128256
llama_model_loader: - kv  26:                 llama.rope.dimension_count u32              = 128
llama_model_loader: - kv  27:                       tokenizer.ggml.model str              = gpt2
llama_model_loader: - kv  28:                         tokenizer.ggml.pre str              = llama-bpe
llama_model_loader: - kv  29:                      tokenizer.ggml.tokens arr[str,128256]  = ["!", "\"", "#", "$", "%", "&", "'", ...
llama_model_loader: - kv  30:                  tokenizer.ggml.token_type arr[i32,128256]  = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ...
llama_model_loader: - kv  31:                      tokenizer.ggml.merges arr[str,280147]  = ["Ġ Ġ", "Ġ ĠĠĠ", "ĠĠ ĠĠ", "...
llama_model_loader: - kv  32:                tokenizer.ggml.bos_token_id u32              = 128000
llama_model_loader: - kv  33:                tokenizer.ggml.eos_token_id u32              = 128009
llama_model_loader: - kv  34:                    tokenizer.chat_template str              = {{- bos_token }}\n{%- if custom_tools ...
llama_model_loader: - kv  35:               general.quantization_version u32              = 2
llama_model_loader: - type  f32:  162 tensors
llama_model_loader: - type q4_K:  441 tensors
llama_model_loader: - type q5_K:   40 tensors
llama_model_loader: - type q6_K:   81 tensors
print_info: file format = GGUF V3 (latest)
print_info: file type   = Q4_K - Medium
print_info: file size   = 39.59 GiB (4.82 BPW) 
load: printing all EOG tokens:
load:   - 128001 ('<|end_of_text|>')
load:   - 128008 ('<|eom_id|>')
load:   - 128009 ('<|eot_id|>')
load: special tokens cache size = 256
load: token to piece cache size = 0.7999 MB
print_info: arch             = llama
print_info: vocab_only       = 0
print_info: no_alloc         = 0
print_info: n_ctx_train      = 131072
print_info: n_embd           = 8192





GGUF:

GGUF^C^@^@^@Ô^B^@^@^@^@^@^@$^@^@^@^@^@^@^@^T^@^@^@^@^@^@^@general.architecture^H^@^@^@^E^@^@^@^@^@^@^@llama^L^@^@^@^@^@^@^@general.type^H^@^@^@^E^@^@^@^@^@^@^@model^L^@^@^@^@^@^@^@general.name^H^@^@^@^^^@^@^@^@^@^@^@Llama 3.1 70B Instruct 2024 12^O^@^@^@^@^@^@^@general.version^H^@^@^@^G^@^@^@^@^@^@^@2024-12^P^@^@^@^@^@^@^@general.finetune^H^@^@^@^H^@^@^@^@^@^@^@Instruct^P^@^@^@^@^@^@^@general.basename^H^@^@^@    ^@^@^@^@^@^@^@Llama-3.1^R^@^@^@^@^@^@^@general.size_label^H^@^@^@^C^@^@^@^@^@^@^@70B^O^@^@^@^@^@^@^@general.license^H^@^@^@^H^@^@^@^@^@^@^@llama3.1^X^@^@^@^@^@^@^@general.base_model.count^D^@^@^@^A^@^@^@^Y^@^@^@^@^@^@^@general.base_model.0.name^H^@^@^@^M^@^@^@^@^@^@^@Llama 3.1 70B!^@^@^@^@^@^@^@general.base_model.0.organization^H^@^@^@
^@^@^@^@^@^@^@Meta Llama^]^@^@^@^@^@^@^@general.base_model.0.repo_url^H^@^@^@/

OS

Linux

GPU

Nvidia

CPU

Other

Ollama version

0.14.2

Originally created by @kenokall on GitHub (Jan 20, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/13796 ### What is the issue? The GGUF files for the llama3.3:70b model (across multiple quantizations) have an incorrect general.name field in their metadata. The field states llama3.1 instead of the expected llama3.3. Evidence: - Log output from loading the model shows the wrong name. - Inspecting the GGUF files linked from the manifest confirms the mismatch. - The llama3.3:8b variants correctly show llama3.3 in the same field. - The license file in the manifest correctly references Llama 3.3, suggesting this is a metadata error in the GGUF packing. Impact: This may cause confusion in the logs or UIs that rely on the general.name field for identification. ### Relevant log output ```shell llama_model_loader: loaded meta data with 36 key-value pairs and 724 tensors from /root/.ollama/models/blobs/sha256-4824460d29f2058aaf6e1118a63a7a197a09bed509f0e7d4e2efb1ee273b447d (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = llama llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Llama 3.1 70B Instruct 2024 12 llama_model_loader: - kv 3: general.version str = 2024-12 llama_model_loader: - kv 4: general.finetune str = Instruct llama_model_loader: - kv 5: general.basename str = Llama-3.1 llama_model_loader: - kv 6: general.size_label str = 70B llama_model_loader: - kv 7: general.license str = llama3.1 llama_model_loader: - kv 8: general.base_model.count u32 = 1 llama_model_loader: - kv 9: general.base_model.0.name str = Llama 3.1 70B llama_model_loader: - kv 10: general.base_model.0.organization str = Meta Llama llama_model_loader: - kv 11: general.base_model.0.repo_url str = https://huggingface.co/meta-llama/Lla... llama_model_loader: - kv 12: general.tags arr[str,5] = ["facebook", "meta", "pytorch", "llam... llama_model_loader: - kv 13: general.languages arr[str,7] = ["fr", "it", "pt", "hi", "es", "th", ... llama_model_loader: - kv 14: llama.block_count u32 = 80 llama_model_loader: - kv 15: llama.context_length u32 = 131072 llama_model_loader: - kv 16: llama.embedding_length u32 = 8192 llama_model_loader: - kv 17: llama.feed_forward_length u32 = 28672 llama_model_loader: - kv 18: llama.attention.head_count u32 = 64 llama_model_loader: - kv 19: llama.attention.head_count_kv u32 = 8 llama_model_loader: - kv 20: llama.rope.freq_base f32 = 500000.000000 llama_model_loader: - kv 21: llama.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 22: llama.attention.key_length u32 = 128 llama_model_loader: - kv 23: llama.attention.value_length u32 = 128 llama_model_loader: - kv 24: general.file_type u32 = 15 llama_model_loader: - kv 25: llama.vocab_size u32 = 128256 llama_model_loader: - kv 26: llama.rope.dimension_count u32 = 128 llama_model_loader: - kv 27: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 28: tokenizer.ggml.pre str = llama-bpe llama_model_loader: - kv 29: tokenizer.ggml.tokens arr[str,128256] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 30: tokenizer.ggml.token_type arr[i32,128256] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 31: tokenizer.ggml.merges arr[str,280147] = ["Ġ Ġ", "Ġ ĠĠĠ", "ĠĠ ĠĠ", "... llama_model_loader: - kv 32: tokenizer.ggml.bos_token_id u32 = 128000 llama_model_loader: - kv 33: tokenizer.ggml.eos_token_id u32 = 128009 llama_model_loader: - kv 34: tokenizer.chat_template str = {{- bos_token }}\n{%- if custom_tools ... llama_model_loader: - kv 35: general.quantization_version u32 = 2 llama_model_loader: - type f32: 162 tensors llama_model_loader: - type q4_K: 441 tensors llama_model_loader: - type q5_K: 40 tensors llama_model_loader: - type q6_K: 81 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 39.59 GiB (4.82 BPW) load: printing all EOG tokens: load: - 128001 ('<|end_of_text|>') load: - 128008 ('<|eom_id|>') load: - 128009 ('<|eot_id|>') load: special tokens cache size = 256 load: token to piece cache size = 0.7999 MB print_info: arch = llama print_info: vocab_only = 0 print_info: no_alloc = 0 print_info: n_ctx_train = 131072 print_info: n_embd = 8192 GGUF: GGUF^C^@^@^@Ô^B^@^@^@^@^@^@$^@^@^@^@^@^@^@^T^@^@^@^@^@^@^@general.architecture^H^@^@^@^E^@^@^@^@^@^@^@llama^L^@^@^@^@^@^@^@general.type^H^@^@^@^E^@^@^@^@^@^@^@model^L^@^@^@^@^@^@^@general.name^H^@^@^@^^^@^@^@^@^@^@^@Llama 3.1 70B Instruct 2024 12^O^@^@^@^@^@^@^@general.version^H^@^@^@^G^@^@^@^@^@^@^@2024-12^P^@^@^@^@^@^@^@general.finetune^H^@^@^@^H^@^@^@^@^@^@^@Instruct^P^@^@^@^@^@^@^@general.basename^H^@^@^@ ^@^@^@^@^@^@^@Llama-3.1^R^@^@^@^@^@^@^@general.size_label^H^@^@^@^C^@^@^@^@^@^@^@70B^O^@^@^@^@^@^@^@general.license^H^@^@^@^H^@^@^@^@^@^@^@llama3.1^X^@^@^@^@^@^@^@general.base_model.count^D^@^@^@^A^@^@^@^Y^@^@^@^@^@^@^@general.base_model.0.name^H^@^@^@^M^@^@^@^@^@^@^@Llama 3.1 70B!^@^@^@^@^@^@^@general.base_model.0.organization^H^@^@^@ ^@^@^@^@^@^@^@Meta Llama^]^@^@^@^@^@^@^@general.base_model.0.repo_url^H^@^@^@/ ``` ### OS Linux ### GPU Nvidia ### CPU Other ### Ollama version 0.14.2
GiteaMirror added the bug label 2026-04-22 18:39:40 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#34798