[GH-ISSUE #15436] MLX model support for gemma 4 #56378

Open
opened 2026-04-29 10:44:23 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @toughcoding on GitHub (Apr 8, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/15436

Gemma 26B in mlx not supported yet

hf download mlx-community/gemma-4-26b-a4b-it-mxfp4

cd ~/.cache/huggingface/hub/models--mlx-community--gemma-4-26b-a4b-it-mxfp4
cp -rL snapshots/4d7ba9cbe85182e0b6882dfe72fc2fc6ea2b4afb final
echo "FROM final" > Modelfile

ollama create --experimental gemma-4-26b-a4b-it-mxfp4
importing safetensors model 
importing safetensors model 
importing model-00001-of-00003.safetensors (444 tensors) 
importing model-00002-of-00003.safetensors (452 tensors) 
importing model-00003-of-00003.safetensors (594 tensors) 
importing config config.json 
importing config generation_config.json 
importing config processor_config.json 
importing config tokenizer.json 
importing config tokenizer_config.json 
writing manifest for gemma-4-26b-a4b-it-mxfp4 
successfully imported gemma-4-26b-a4b-it-mxfp4 with 1048 layers

rm -Rf final

ollama list

NAME                                                     ID              SIZE     MODIFIED           

gemma-4-26b-a4b-it-mxfp4:latest                          2d1b607d3ef8    14 GB    About a minute ago

ollama --verbose run gemma-4-26b-a4b-it-mxfp4:latest 

Error: 500 Internal Server Error: mlx runner failed: Error: unsupported architecture: Gemma4ForConditionalGeneration (exit: exit status 1)
Originally created by @toughcoding on GitHub (Apr 8, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/15436 Gemma 26B in mlx not supported yet ``` hf download mlx-community/gemma-4-26b-a4b-it-mxfp4 cd ~/.cache/huggingface/hub/models--mlx-community--gemma-4-26b-a4b-it-mxfp4 cp -rL snapshots/4d7ba9cbe85182e0b6882dfe72fc2fc6ea2b4afb final echo "FROM final" > Modelfile ollama create --experimental gemma-4-26b-a4b-it-mxfp4 importing safetensors model  importing safetensors model  importing model-00001-of-00003.safetensors (444 tensors)  importing model-00002-of-00003.safetensors (452 tensors)  importing model-00003-of-00003.safetensors (594 tensors)  importing config config.json  importing config generation_config.json  importing config processor_config.json  importing config tokenizer.json  importing config tokenizer_config.json  writing manifest for gemma-4-26b-a4b-it-mxfp4  successfully imported gemma-4-26b-a4b-it-mxfp4 with 1048 layers rm -Rf final ollama list NAME                                                     ID              SIZE     MODIFIED            gemma-4-26b-a4b-it-mxfp4:latest                          2d1b607d3ef8    14 GB    About a minute ago ollama --verbose run gemma-4-26b-a4b-it-mxfp4:latest  Error: 500 Internal Server Error: mlx runner failed: Error: unsupported architecture: Gemma4ForConditionalGeneration (exit: exit status 1) ```
GiteaMirror added the mlxmodel labels 2026-04-29 10:44:23 -05:00
Author
Owner

@rick-github commented on GitHub (Apr 8, 2026):

https://github.com/ollama/ollama/pull/15244

<!-- gh-comment-id:4210123750 --> @rick-github commented on GitHub (Apr 8, 2026): https://github.com/ollama/ollama/pull/15244
Author
Owner

@PureBlissAK commented on GitHub (Apr 18, 2026):

🤖 Automated Triage & Analysis Report

Issue: #15436
Analyzed: 2026-04-18T18:21:31.735209

Analysis

  • Type: unknown
  • Severity: medium
  • Components: unknown

Implementation Plan

  • Effort: medium
  • Steps:

This issue has been triaged and marked for implementation.

<!-- gh-comment-id:4274308477 --> @PureBlissAK commented on GitHub (Apr 18, 2026): <!-- ollama-issue-orchestrator:v1 issue:15436 --> ## 🤖 Automated Triage & Analysis Report **Issue**: #15436 **Analyzed**: 2026-04-18T18:21:31.735209 ### Analysis - **Type**: unknown - **Severity**: medium - **Components**: unknown ### Implementation Plan - **Effort**: medium - **Steps**: *This issue has been triaged and marked for implementation.*
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#56378