[GH-ISSUE #10208] ollama show --verbose reporting wrong infromation #6696

Closed
opened 2026-04-12 18:25:57 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @pt13762104 on GitHub (Apr 10, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/10208

What is the issue?

I downloaded llama3.2:3b, but ollama show --verbose isn't working right. It's displaying Q3_K_M for Q4_K and Q4_K_M for Q6_K.
I have confirmed that the model is the same one on ollama.com.

Relevant log output

Tensors
    rope_freqs.weight            F32       [64]             
    token_embd.weight            Q4_K_S    [3072 128256]    
    blk.0.attn_norm.weight       F32       [3072]           
    blk.0.ffn_down.weight        Q4_K_S    [8192 3072]      
    blk.0.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.0.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.0.ffn_norm.weight        F32       [3072]           
    blk.0.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.0.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.0.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.0.attn_v.weight          Q4_K_S    [3072 1024]      
    blk.1.attn_norm.weight       F32       [3072]           
    blk.1.ffn_down.weight        Q4_K_S    [8192 3072]      
    blk.1.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.1.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.1.ffn_norm.weight        F32       [3072]           
    blk.1.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.1.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.1.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.1.attn_v.weight          Q4_K_S    [3072 1024]      
    blk.10.attn_norm.weight      F32       [3072]           
    blk.10.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.10.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.10.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.10.ffn_norm.weight       F32       [3072]           
    blk.10.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.10.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.10.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.10.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.11.attn_norm.weight      F32       [3072]           
    blk.11.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.11.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.11.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.11.ffn_norm.weight       F32       [3072]           
    blk.11.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.11.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.11.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.11.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.12.attn_norm.weight      F32       [3072]           
    blk.12.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.12.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.12.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.12.ffn_norm.weight       F32       [3072]           
    blk.12.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.12.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.12.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.12.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.13.attn_norm.weight      F32       [3072]           
    blk.13.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.13.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.13.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.13.ffn_norm.weight       F32       [3072]           
    blk.13.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.13.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.13.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.13.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.14.attn_norm.weight      F32       [3072]           
    blk.14.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.14.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.14.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.14.ffn_norm.weight       F32       [3072]           
    blk.14.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.14.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.14.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.14.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.15.attn_norm.weight      F32       [3072]           
    blk.15.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.15.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.15.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.15.ffn_norm.weight       F32       [3072]           
    blk.15.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.15.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.15.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.15.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.16.attn_norm.weight      F32       [3072]           
    blk.16.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.16.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.16.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.16.ffn_norm.weight       F32       [3072]           
    blk.16.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.16.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.16.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.16.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.17.attn_norm.weight      F32       [3072]           
    blk.17.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.17.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.17.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.17.ffn_norm.weight       F32       [3072]           
    blk.17.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.17.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.17.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.17.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.18.attn_norm.weight      F32       [3072]           
    blk.18.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.18.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.18.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.18.ffn_norm.weight       F32       [3072]           
    blk.18.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.18.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.18.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.18.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.19.attn_norm.weight      F32       [3072]           
    blk.19.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.19.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.19.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.19.ffn_norm.weight       F32       [3072]           
    blk.19.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.19.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.19.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.19.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.2.attn_norm.weight       F32       [3072]           
    blk.2.ffn_down.weight        Q3_K_M    [8192 3072]      
    blk.2.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.2.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.2.ffn_norm.weight        F32       [3072]           
    blk.2.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.2.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.2.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.2.attn_v.weight          Q3_K_M    [3072 1024]      
    blk.20.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.20.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.20.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.20.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.20.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.20.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.3.attn_norm.weight       F32       [3072]           
    blk.3.ffn_down.weight        Q3_K_M    [8192 3072]      
    blk.3.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.3.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.3.ffn_norm.weight        F32       [3072]           
    blk.3.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.3.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.3.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.3.attn_v.weight          Q4_K_S    [3072 1024]      
    blk.4.attn_norm.weight       F32       [3072]           
    blk.4.ffn_down.weight        Q4_K_S    [8192 3072]      
    blk.4.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.4.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.4.ffn_norm.weight        F32       [3072]           
    blk.4.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.4.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.4.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.4.attn_v.weight          Q3_K_M    [3072 1024]      
    blk.5.attn_norm.weight       F32       [3072]           
    blk.5.ffn_down.weight        Q3_K_M    [8192 3072]      
    blk.5.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.5.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.5.ffn_norm.weight        F32       [3072]           
    blk.5.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.5.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.5.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.5.attn_v.weight          Q3_K_M    [3072 1024]      
    blk.6.attn_norm.weight       F32       [3072]           
    blk.6.ffn_down.weight        Q3_K_M    [8192 3072]      
    blk.6.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.6.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.6.ffn_norm.weight        F32       [3072]           
    blk.6.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.6.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.6.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.6.attn_v.weight          Q4_K_S    [3072 1024]      
    blk.7.attn_norm.weight       F32       [3072]           
    blk.7.ffn_down.weight        Q4_K_S    [8192 3072]      
    blk.7.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.7.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.7.ffn_norm.weight        F32       [3072]           
    blk.7.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.7.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.7.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.7.attn_v.weight          Q3_K_M    [3072 1024]      
    blk.8.attn_norm.weight       F32       [3072]           
    blk.8.ffn_down.weight        Q3_K_M    [8192 3072]      
    blk.8.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.8.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.8.ffn_norm.weight        F32       [3072]           
    blk.8.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.8.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.8.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.8.attn_v.weight          Q3_K_M    [3072 1024]      
    blk.9.attn_norm.weight       F32       [3072]           
    blk.9.ffn_down.weight        Q3_K_M    [8192 3072]      
    blk.9.ffn_gate.weight        Q3_K_M    [3072 8192]      
    blk.9.ffn_up.weight          Q3_K_M    [3072 8192]      
    blk.9.ffn_norm.weight        F32       [3072]           
    blk.9.attn_k.weight          Q3_K_M    [3072 1024]      
    blk.9.attn_output.weight     Q3_K_M    [3072 3072]      
    blk.9.attn_q.weight          Q3_K_M    [3072 3072]      
    blk.9.attn_v.weight          Q4_K_S    [3072 1024]      
    blk.20.attn_norm.weight      F32       [3072]           
    blk.20.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.20.ffn_norm.weight       F32       [3072]           
    blk.21.attn_norm.weight      F32       [3072]           
    blk.21.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.21.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.21.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.21.ffn_norm.weight       F32       [3072]           
    blk.21.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.21.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.21.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.21.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.22.attn_norm.weight      F32       [3072]           
    blk.22.ffn_down.weight       Q3_K_M    [8192 3072]      
    blk.22.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.22.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.22.ffn_norm.weight       F32       [3072]           
    blk.22.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.22.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.22.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.22.attn_v.weight         Q3_K_M    [3072 1024]      
    blk.23.attn_norm.weight      F32       [3072]           
    blk.23.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.23.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.23.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.23.ffn_norm.weight       F32       [3072]           
    blk.23.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.23.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.23.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.23.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.24.attn_norm.weight      F32       [3072]           
    blk.24.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.24.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.24.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.24.ffn_norm.weight       F32       [3072]           
    blk.24.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.24.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.24.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.24.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.25.attn_norm.weight      F32       [3072]           
    blk.25.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.25.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.25.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.25.ffn_norm.weight       F32       [3072]           
    blk.25.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.25.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.25.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.25.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.26.attn_norm.weight      F32       [3072]           
    blk.26.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.26.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.26.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.26.ffn_norm.weight       F32       [3072]           
    blk.26.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.26.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.26.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.26.attn_v.weight         Q4_K_S    [3072 1024]      
    blk.27.attn_norm.weight      F32       [3072]           
    blk.27.ffn_down.weight       Q4_K_S    [8192 3072]      
    blk.27.ffn_gate.weight       Q3_K_M    [3072 8192]      
    blk.27.ffn_up.weight         Q3_K_M    [3072 8192]      
    blk.27.ffn_norm.weight       F32       [3072]           
    blk.27.attn_k.weight         Q3_K_M    [3072 1024]      
    blk.27.attn_output.weight    Q3_K_M    [3072 3072]      
    blk.27.attn_q.weight         Q3_K_M    [3072 3072]      
    blk.27.attn_v.weight         Q4_K_S    [3072 1024]      
    output_norm.weight           F32       [3072]

OS

No response

GPU

No response

CPU

No response

Ollama version

No response

Originally created by @pt13762104 on GitHub (Apr 10, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/10208 ### What is the issue? I downloaded `llama3.2:3b`, but `ollama show --verbose` isn't working right. It's displaying `Q3_K_M` for `Q4_K` and `Q4_K_M` for `Q6_K`. I have confirmed that the model is the same one on ollama.com. ### Relevant log output ```shell Tensors rope_freqs.weight F32 [64] token_embd.weight Q4_K_S [3072 128256] blk.0.attn_norm.weight F32 [3072] blk.0.ffn_down.weight Q4_K_S [8192 3072] blk.0.ffn_gate.weight Q3_K_M [3072 8192] blk.0.ffn_up.weight Q3_K_M [3072 8192] blk.0.ffn_norm.weight F32 [3072] blk.0.attn_k.weight Q3_K_M [3072 1024] blk.0.attn_output.weight Q3_K_M [3072 3072] blk.0.attn_q.weight Q3_K_M [3072 3072] blk.0.attn_v.weight Q4_K_S [3072 1024] blk.1.attn_norm.weight F32 [3072] blk.1.ffn_down.weight Q4_K_S [8192 3072] blk.1.ffn_gate.weight Q3_K_M [3072 8192] blk.1.ffn_up.weight Q3_K_M [3072 8192] blk.1.ffn_norm.weight F32 [3072] blk.1.attn_k.weight Q3_K_M [3072 1024] blk.1.attn_output.weight Q3_K_M [3072 3072] blk.1.attn_q.weight Q3_K_M [3072 3072] blk.1.attn_v.weight Q4_K_S [3072 1024] blk.10.attn_norm.weight F32 [3072] blk.10.ffn_down.weight Q4_K_S [8192 3072] blk.10.ffn_gate.weight Q3_K_M [3072 8192] blk.10.ffn_up.weight Q3_K_M [3072 8192] blk.10.ffn_norm.weight F32 [3072] blk.10.attn_k.weight Q3_K_M [3072 1024] blk.10.attn_output.weight Q3_K_M [3072 3072] blk.10.attn_q.weight Q3_K_M [3072 3072] blk.10.attn_v.weight Q4_K_S [3072 1024] blk.11.attn_norm.weight F32 [3072] blk.11.ffn_down.weight Q3_K_M [8192 3072] blk.11.ffn_gate.weight Q3_K_M [3072 8192] blk.11.ffn_up.weight Q3_K_M [3072 8192] blk.11.ffn_norm.weight F32 [3072] blk.11.attn_k.weight Q3_K_M [3072 1024] blk.11.attn_output.weight Q3_K_M [3072 3072] blk.11.attn_q.weight Q3_K_M [3072 3072] blk.11.attn_v.weight Q3_K_M [3072 1024] blk.12.attn_norm.weight F32 [3072] blk.12.ffn_down.weight Q3_K_M [8192 3072] blk.12.ffn_gate.weight Q3_K_M [3072 8192] blk.12.ffn_up.weight Q3_K_M [3072 8192] blk.12.ffn_norm.weight F32 [3072] blk.12.attn_k.weight Q3_K_M [3072 1024] blk.12.attn_output.weight Q3_K_M [3072 3072] blk.12.attn_q.weight Q3_K_M [3072 3072] blk.12.attn_v.weight Q3_K_M [3072 1024] blk.13.attn_norm.weight F32 [3072] blk.13.ffn_down.weight Q4_K_S [8192 3072] blk.13.ffn_gate.weight Q3_K_M [3072 8192] blk.13.ffn_up.weight Q3_K_M [3072 8192] blk.13.ffn_norm.weight F32 [3072] blk.13.attn_k.weight Q3_K_M [3072 1024] blk.13.attn_output.weight Q3_K_M [3072 3072] blk.13.attn_q.weight Q3_K_M [3072 3072] blk.13.attn_v.weight Q4_K_S [3072 1024] blk.14.attn_norm.weight F32 [3072] blk.14.ffn_down.weight Q3_K_M [8192 3072] blk.14.ffn_gate.weight Q3_K_M [3072 8192] blk.14.ffn_up.weight Q3_K_M [3072 8192] blk.14.ffn_norm.weight F32 [3072] blk.14.attn_k.weight Q3_K_M [3072 1024] blk.14.attn_output.weight Q3_K_M [3072 3072] blk.14.attn_q.weight Q3_K_M [3072 3072] blk.14.attn_v.weight Q3_K_M [3072 1024] blk.15.attn_norm.weight F32 [3072] blk.15.ffn_down.weight Q3_K_M [8192 3072] blk.15.ffn_gate.weight Q3_K_M [3072 8192] blk.15.ffn_up.weight Q3_K_M [3072 8192] blk.15.ffn_norm.weight F32 [3072] blk.15.attn_k.weight Q3_K_M [3072 1024] blk.15.attn_output.weight Q3_K_M [3072 3072] blk.15.attn_q.weight Q3_K_M [3072 3072] blk.15.attn_v.weight Q3_K_M [3072 1024] blk.16.attn_norm.weight F32 [3072] blk.16.ffn_down.weight Q4_K_S [8192 3072] blk.16.ffn_gate.weight Q3_K_M [3072 8192] blk.16.ffn_up.weight Q3_K_M [3072 8192] blk.16.ffn_norm.weight F32 [3072] blk.16.attn_k.weight Q3_K_M [3072 1024] blk.16.attn_output.weight Q3_K_M [3072 3072] blk.16.attn_q.weight Q3_K_M [3072 3072] blk.16.attn_v.weight Q4_K_S [3072 1024] blk.17.attn_norm.weight F32 [3072] blk.17.ffn_down.weight Q3_K_M [8192 3072] blk.17.ffn_gate.weight Q3_K_M [3072 8192] blk.17.ffn_up.weight Q3_K_M [3072 8192] blk.17.ffn_norm.weight F32 [3072] blk.17.attn_k.weight Q3_K_M [3072 1024] blk.17.attn_output.weight Q3_K_M [3072 3072] blk.17.attn_q.weight Q3_K_M [3072 3072] blk.17.attn_v.weight Q3_K_M [3072 1024] blk.18.attn_norm.weight F32 [3072] blk.18.ffn_down.weight Q3_K_M [8192 3072] blk.18.ffn_gate.weight Q3_K_M [3072 8192] blk.18.ffn_up.weight Q3_K_M [3072 8192] blk.18.ffn_norm.weight F32 [3072] blk.18.attn_k.weight Q3_K_M [3072 1024] blk.18.attn_output.weight Q3_K_M [3072 3072] blk.18.attn_q.weight Q3_K_M [3072 3072] blk.18.attn_v.weight Q3_K_M [3072 1024] blk.19.attn_norm.weight F32 [3072] blk.19.ffn_down.weight Q4_K_S [8192 3072] blk.19.ffn_gate.weight Q3_K_M [3072 8192] blk.19.ffn_up.weight Q3_K_M [3072 8192] blk.19.ffn_norm.weight F32 [3072] blk.19.attn_k.weight Q3_K_M [3072 1024] blk.19.attn_output.weight Q3_K_M [3072 3072] blk.19.attn_q.weight Q3_K_M [3072 3072] blk.19.attn_v.weight Q4_K_S [3072 1024] blk.2.attn_norm.weight F32 [3072] blk.2.ffn_down.weight Q3_K_M [8192 3072] blk.2.ffn_gate.weight Q3_K_M [3072 8192] blk.2.ffn_up.weight Q3_K_M [3072 8192] blk.2.ffn_norm.weight F32 [3072] blk.2.attn_k.weight Q3_K_M [3072 1024] blk.2.attn_output.weight Q3_K_M [3072 3072] blk.2.attn_q.weight Q3_K_M [3072 3072] blk.2.attn_v.weight Q3_K_M [3072 1024] blk.20.ffn_gate.weight Q3_K_M [3072 8192] blk.20.ffn_up.weight Q3_K_M [3072 8192] blk.20.attn_k.weight Q3_K_M [3072 1024] blk.20.attn_output.weight Q3_K_M [3072 3072] blk.20.attn_q.weight Q3_K_M [3072 3072] blk.20.attn_v.weight Q3_K_M [3072 1024] blk.3.attn_norm.weight F32 [3072] blk.3.ffn_down.weight Q3_K_M [8192 3072] blk.3.ffn_gate.weight Q3_K_M [3072 8192] blk.3.ffn_up.weight Q3_K_M [3072 8192] blk.3.ffn_norm.weight F32 [3072] blk.3.attn_k.weight Q3_K_M [3072 1024] blk.3.attn_output.weight Q3_K_M [3072 3072] blk.3.attn_q.weight Q3_K_M [3072 3072] blk.3.attn_v.weight Q4_K_S [3072 1024] blk.4.attn_norm.weight F32 [3072] blk.4.ffn_down.weight Q4_K_S [8192 3072] blk.4.ffn_gate.weight Q3_K_M [3072 8192] blk.4.ffn_up.weight Q3_K_M [3072 8192] blk.4.ffn_norm.weight F32 [3072] blk.4.attn_k.weight Q3_K_M [3072 1024] blk.4.attn_output.weight Q3_K_M [3072 3072] blk.4.attn_q.weight Q3_K_M [3072 3072] blk.4.attn_v.weight Q3_K_M [3072 1024] blk.5.attn_norm.weight F32 [3072] blk.5.ffn_down.weight Q3_K_M [8192 3072] blk.5.ffn_gate.weight Q3_K_M [3072 8192] blk.5.ffn_up.weight Q3_K_M [3072 8192] blk.5.ffn_norm.weight F32 [3072] blk.5.attn_k.weight Q3_K_M [3072 1024] blk.5.attn_output.weight Q3_K_M [3072 3072] blk.5.attn_q.weight Q3_K_M [3072 3072] blk.5.attn_v.weight Q3_K_M [3072 1024] blk.6.attn_norm.weight F32 [3072] blk.6.ffn_down.weight Q3_K_M [8192 3072] blk.6.ffn_gate.weight Q3_K_M [3072 8192] blk.6.ffn_up.weight Q3_K_M [3072 8192] blk.6.ffn_norm.weight F32 [3072] blk.6.attn_k.weight Q3_K_M [3072 1024] blk.6.attn_output.weight Q3_K_M [3072 3072] blk.6.attn_q.weight Q3_K_M [3072 3072] blk.6.attn_v.weight Q4_K_S [3072 1024] blk.7.attn_norm.weight F32 [3072] blk.7.ffn_down.weight Q4_K_S [8192 3072] blk.7.ffn_gate.weight Q3_K_M [3072 8192] blk.7.ffn_up.weight Q3_K_M [3072 8192] blk.7.ffn_norm.weight F32 [3072] blk.7.attn_k.weight Q3_K_M [3072 1024] blk.7.attn_output.weight Q3_K_M [3072 3072] blk.7.attn_q.weight Q3_K_M [3072 3072] blk.7.attn_v.weight Q3_K_M [3072 1024] blk.8.attn_norm.weight F32 [3072] blk.8.ffn_down.weight Q3_K_M [8192 3072] blk.8.ffn_gate.weight Q3_K_M [3072 8192] blk.8.ffn_up.weight Q3_K_M [3072 8192] blk.8.ffn_norm.weight F32 [3072] blk.8.attn_k.weight Q3_K_M [3072 1024] blk.8.attn_output.weight Q3_K_M [3072 3072] blk.8.attn_q.weight Q3_K_M [3072 3072] blk.8.attn_v.weight Q3_K_M [3072 1024] blk.9.attn_norm.weight F32 [3072] blk.9.ffn_down.weight Q3_K_M [8192 3072] blk.9.ffn_gate.weight Q3_K_M [3072 8192] blk.9.ffn_up.weight Q3_K_M [3072 8192] blk.9.ffn_norm.weight F32 [3072] blk.9.attn_k.weight Q3_K_M [3072 1024] blk.9.attn_output.weight Q3_K_M [3072 3072] blk.9.attn_q.weight Q3_K_M [3072 3072] blk.9.attn_v.weight Q4_K_S [3072 1024] blk.20.attn_norm.weight F32 [3072] blk.20.ffn_down.weight Q4_K_S [8192 3072] blk.20.ffn_norm.weight F32 [3072] blk.21.attn_norm.weight F32 [3072] blk.21.ffn_down.weight Q3_K_M [8192 3072] blk.21.ffn_gate.weight Q3_K_M [3072 8192] blk.21.ffn_up.weight Q3_K_M [3072 8192] blk.21.ffn_norm.weight F32 [3072] blk.21.attn_k.weight Q3_K_M [3072 1024] blk.21.attn_output.weight Q3_K_M [3072 3072] blk.21.attn_q.weight Q3_K_M [3072 3072] blk.21.attn_v.weight Q3_K_M [3072 1024] blk.22.attn_norm.weight F32 [3072] blk.22.ffn_down.weight Q3_K_M [8192 3072] blk.22.ffn_gate.weight Q3_K_M [3072 8192] blk.22.ffn_up.weight Q3_K_M [3072 8192] blk.22.ffn_norm.weight F32 [3072] blk.22.attn_k.weight Q3_K_M [3072 1024] blk.22.attn_output.weight Q3_K_M [3072 3072] blk.22.attn_q.weight Q3_K_M [3072 3072] blk.22.attn_v.weight Q3_K_M [3072 1024] blk.23.attn_norm.weight F32 [3072] blk.23.ffn_down.weight Q4_K_S [8192 3072] blk.23.ffn_gate.weight Q3_K_M [3072 8192] blk.23.ffn_up.weight Q3_K_M [3072 8192] blk.23.ffn_norm.weight F32 [3072] blk.23.attn_k.weight Q3_K_M [3072 1024] blk.23.attn_output.weight Q3_K_M [3072 3072] blk.23.attn_q.weight Q3_K_M [3072 3072] blk.23.attn_v.weight Q4_K_S [3072 1024] blk.24.attn_norm.weight F32 [3072] blk.24.ffn_down.weight Q4_K_S [8192 3072] blk.24.ffn_gate.weight Q3_K_M [3072 8192] blk.24.ffn_up.weight Q3_K_M [3072 8192] blk.24.ffn_norm.weight F32 [3072] blk.24.attn_k.weight Q3_K_M [3072 1024] blk.24.attn_output.weight Q3_K_M [3072 3072] blk.24.attn_q.weight Q3_K_M [3072 3072] blk.24.attn_v.weight Q4_K_S [3072 1024] blk.25.attn_norm.weight F32 [3072] blk.25.ffn_down.weight Q4_K_S [8192 3072] blk.25.ffn_gate.weight Q3_K_M [3072 8192] blk.25.ffn_up.weight Q3_K_M [3072 8192] blk.25.ffn_norm.weight F32 [3072] blk.25.attn_k.weight Q3_K_M [3072 1024] blk.25.attn_output.weight Q3_K_M [3072 3072] blk.25.attn_q.weight Q3_K_M [3072 3072] blk.25.attn_v.weight Q4_K_S [3072 1024] blk.26.attn_norm.weight F32 [3072] blk.26.ffn_down.weight Q4_K_S [8192 3072] blk.26.ffn_gate.weight Q3_K_M [3072 8192] blk.26.ffn_up.weight Q3_K_M [3072 8192] blk.26.ffn_norm.weight F32 [3072] blk.26.attn_k.weight Q3_K_M [3072 1024] blk.26.attn_output.weight Q3_K_M [3072 3072] blk.26.attn_q.weight Q3_K_M [3072 3072] blk.26.attn_v.weight Q4_K_S [3072 1024] blk.27.attn_norm.weight F32 [3072] blk.27.ffn_down.weight Q4_K_S [8192 3072] blk.27.ffn_gate.weight Q3_K_M [3072 8192] blk.27.ffn_up.weight Q3_K_M [3072 8192] blk.27.ffn_norm.weight F32 [3072] blk.27.attn_k.weight Q3_K_M [3072 1024] blk.27.attn_output.weight Q3_K_M [3072 3072] blk.27.attn_q.weight Q3_K_M [3072 3072] blk.27.attn_v.weight Q4_K_S [3072 1024] output_norm.weight F32 [3072] ``` ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
GiteaMirror added the bug label 2026-04-12 18:25:57 -05:00
Author
Owner

@rick-github commented on GitHub (Feb 10, 2026):

$ ollama show -v llama3.2:3b-instruct-q4_K_M | grep blk.0
    blk.0.attn_norm.weight       F32     [3072]           
    blk.0.ffn_down.weight        Q6_K    [8192 3072]      
    blk.0.ffn_gate.weight        Q4_K    [3072 8192]      
    blk.0.ffn_up.weight          Q4_K    [3072 8192]      
    blk.0.ffn_norm.weight        F32     [3072]           
    blk.0.attn_k.weight          Q4_K    [3072 1024]      
    blk.0.attn_output.weight     Q4_K    [3072 3072]      
    blk.0.attn_q.weight          Q4_K    [3072 3072]      
    blk.0.attn_v.weight          Q6_K    [3072 1024]      
<!-- gh-comment-id:3874716538 --> @rick-github commented on GitHub (Feb 10, 2026): ```console $ ollama show -v llama3.2:3b-instruct-q4_K_M | grep blk.0 blk.0.attn_norm.weight F32 [3072] blk.0.ffn_down.weight Q6_K [8192 3072] blk.0.ffn_gate.weight Q4_K [3072 8192] blk.0.ffn_up.weight Q4_K [3072 8192] blk.0.ffn_norm.weight F32 [3072] blk.0.attn_k.weight Q4_K [3072 1024] blk.0.attn_output.weight Q4_K [3072 3072] blk.0.attn_q.weight Q4_K [3072 3072] blk.0.attn_v.weight Q6_K [3072 1024] ```
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#6696