[GH-ISSUE #9682] Ollama 0.6.0: unknown model architecture: 'gemma3' #32078

Closed
opened 2026-04-22 12:59:22 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @chigkim on GitHub (Mar 12, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/9682

Originally assigned to: @pdevine on GitHub.

What is the issue?

Are more quants for gemma3 coming to Ollama library? There seem to be only q4_K_M and fp16.
I tried to quantize to q8_0, and I got this error.

modelfile:
FROM gemma3:27b-it-fp16

% ollama create gemma3:27-it-q8_0 --quantize q8_0 -f gemma3.modelfile

Relevant log output

gathering model components 
quantizing F16 model to Q8_0 
Error: llama_model_quantize: 1

Log:
llama_model_quantize: failed to quantize: unknown model architecture: 'gemma3'

OS

macOS

GPU

Apple

CPU

Apple

Ollama version

0.6.0

Originally created by @chigkim on GitHub (Mar 12, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/9682 Originally assigned to: @pdevine on GitHub. ### What is the issue? Are more quants for gemma3 coming to Ollama library? There seem to be only q4_K_M and fp16. I tried to quantize to q8_0, and I got this error. modelfile: FROM gemma3:27b-it-fp16 % ollama create gemma3:27-it-q8_0 --quantize q8_0 -f gemma3.modelfile ### Relevant log output ```shell gathering model components quantizing F16 model to Q8_0 Error: llama_model_quantize: 1 Log: llama_model_quantize: failed to quantize: unknown model architecture: 'gemma3' ``` ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version 0.6.0
GiteaMirror added the bug label 2026-04-22 12:59:22 -05:00
Author
Owner

@Hioness commented on GitHub (Mar 12, 2025):

I've had the same issue with the 1b version.

> ollama create gemma3:1b-Q8_0 -f .\modelfile.md -q Q8_0
gathering model components
quantizing F16 model to Q8_0
Error: llama_model_quantize: 1
<!-- gh-comment-id:2718318769 --> @Hioness commented on GitHub (Mar 12, 2025): I've had the same issue with the 1b version. ```` > ollama create gemma3:1b-Q8_0 -f .\modelfile.md -q Q8_0 gathering model components quantizing F16 model to Q8_0 Error: llama_model_quantize: 1 ````
Author
Owner

@pdevine commented on GitHub (Mar 12, 2025):

I have a fix for this, but it's not quite ready yet.

<!-- gh-comment-id:2718841166 --> @pdevine commented on GitHub (Mar 12, 2025): I have a fix for this, but it's not quite ready yet.
Author
Owner

@lowlyocean commented on GitHub (Mar 14, 2025):

I have a fix for this, but it's not quite ready yet.

Is this something we can test before the next official ollama release version?

<!-- gh-comment-id:2724975917 --> @lowlyocean commented on GitHub (Mar 14, 2025): > I have a fix for this, but it's not quite ready yet. Is this something we can test before the next official ollama release version?
Author
Owner

@pdevine commented on GitHub (Mar 15, 2025):

Fixed w/ #9776

This should work in Ollama 0.6.2 (didn't make it work 0.6.1 unfortunately) or you can build from main.

<!-- gh-comment-id:2726071432 --> @pdevine commented on GitHub (Mar 15, 2025): Fixed w/ #9776 This should work in Ollama `0.6.2` (didn't make it work `0.6.1` unfortunately) or you can build from `main`.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#32078