[GH-ISSUE #15436] MLX model support for gemma 4 #9865

Open
opened 2026-04-12 22:43:42 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @toughcoding on GitHub (Apr 8, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/15436

Gemma 26B in mlx not supported yet

hf download mlx-community/gemma-4-26b-a4b-it-mxfp4

cd ~/.cache/huggingface/hub/models--mlx-community--gemma-4-26b-a4b-it-mxfp4
cp -rL snapshots/4d7ba9cbe85182e0b6882dfe72fc2fc6ea2b4afb final
echo "FROM final" > Modelfile

ollama create --experimental gemma-4-26b-a4b-it-mxfp4
importing safetensors model 
importing safetensors model 
importing model-00001-of-00003.safetensors (444 tensors) 
importing model-00002-of-00003.safetensors (452 tensors) 
importing model-00003-of-00003.safetensors (594 tensors) 
importing config config.json 
importing config generation_config.json 
importing config processor_config.json 
importing config tokenizer.json 
importing config tokenizer_config.json 
writing manifest for gemma-4-26b-a4b-it-mxfp4 
successfully imported gemma-4-26b-a4b-it-mxfp4 with 1048 layers

rm -Rf final

ollama list

NAME                                                     ID              SIZE     MODIFIED           

gemma-4-26b-a4b-it-mxfp4:latest                          2d1b607d3ef8    14 GB    About a minute ago

ollama --verbose run gemma-4-26b-a4b-it-mxfp4:latest 

Error: 500 Internal Server Error: mlx runner failed: Error: unsupported architecture: Gemma4ForConditionalGeneration (exit: exit status 1)
Originally created by @toughcoding on GitHub (Apr 8, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/15436 Gemma 26B in mlx not supported yet ``` hf download mlx-community/gemma-4-26b-a4b-it-mxfp4 cd ~/.cache/huggingface/hub/models--mlx-community--gemma-4-26b-a4b-it-mxfp4 cp -rL snapshots/4d7ba9cbe85182e0b6882dfe72fc2fc6ea2b4afb final echo "FROM final" > Modelfile ollama create --experimental gemma-4-26b-a4b-it-mxfp4 importing safetensors model  importing safetensors model  importing model-00001-of-00003.safetensors (444 tensors)  importing model-00002-of-00003.safetensors (452 tensors)  importing model-00003-of-00003.safetensors (594 tensors)  importing config config.json  importing config generation_config.json  importing config processor_config.json  importing config tokenizer.json  importing config tokenizer_config.json  writing manifest for gemma-4-26b-a4b-it-mxfp4  successfully imported gemma-4-26b-a4b-it-mxfp4 with 1048 layers rm -Rf final ollama list NAME                                                     ID              SIZE     MODIFIED            gemma-4-26b-a4b-it-mxfp4:latest                          2d1b607d3ef8    14 GB    About a minute ago ollama --verbose run gemma-4-26b-a4b-it-mxfp4:latest  Error: 500 Internal Server Error: mlx runner failed: Error: unsupported architecture: Gemma4ForConditionalGeneration (exit: exit status 1) ```
GiteaMirror added the mlxmodel labels 2026-04-12 22:43:42 -05:00
Author
Owner

@rick-github commented on GitHub (Apr 8, 2026):

https://github.com/ollama/ollama/pull/15244

<!-- gh-comment-id:4210123750 --> @rick-github commented on GitHub (Apr 8, 2026): https://github.com/ollama/ollama/pull/15244
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#9865