[GH-ISSUE #5531] ollama create --quantize does not show proper error if quantizing an unsupported model architecture #3454

Closed
opened 2026-04-12 14:08:00 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @jmorganca on GitHub (Jul 7, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/5531

Originally assigned to: @joshyan1 on GitHub.

What is the issue?

When quantizing a model with ollama create -q, unknown model architecture errors are not properly bubbled up. E.g. from the logs:

llama_model_loader: - type  f32:  121 tensors
llama_model_loader: - type  f16:  162 tensors
llama_model_quantize: failed to quantize: unknown model architecture: 'chatglm'

User error:

% ollama create -f Modelfile -q q4_0 test
transferring model data 
quantizing F16 model to Q4_0 
Error: llama_model_quantize: 1

OS

No response

GPU

No response

CPU

No response

Ollama version

No response

Originally created by @jmorganca on GitHub (Jul 7, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/5531 Originally assigned to: @joshyan1 on GitHub. ### What is the issue? When quantizing a model with `ollama create -q`, unknown model architecture errors are not properly bubbled up. E.g. from the logs: ``` llama_model_loader: - type f32: 121 tensors llama_model_loader: - type f16: 162 tensors llama_model_quantize: failed to quantize: unknown model architecture: 'chatglm' ``` User error: ``` % ollama create -f Modelfile -q q4_0 test transferring model data quantizing F16 model to Q4_0 Error: llama_model_quantize: 1 ``` ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
GiteaMirror added the bug label 2026-04-12 14:08:00 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#3454