[GH-ISSUE #8548] ollama create --quantize q4_K_M not working #5516

Closed
opened 2026-04-12 16:46:05 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @hgKang02 on GitHub (Jan 23, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/8548

What is the issue?

Dear ollama authors,

I hope this issue finds you well. First of all, thank you for such a wonderful project. It helped me to play around with different agentic AI models and tasks.

However, the quantize command is not working for me when I try to quantize the llava-next model. I have created a Modelfile on the same folder where the safetensors are placed. So for instance, if the model's safetensors are placed at /user/agent-models/llava-next, inside this folder I have created the Modelfile with context
FROM .

Now when I run the ollama create --quantize q4_K_M mymodel I get

gathering model components
Error: no Modelfile or safetensors files found

Would this be because ollama does not support Llava architecture for quantize function? Or is there anything that I am doing wrong?

Thank you and it will be much appreciated if I can hear back about this problem.

Sincerely,

OS

Linux

GPU

Nvidia

CPU

No response

Ollama version

0.5.7

Originally created by @hgKang02 on GitHub (Jan 23, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/8548 ### What is the issue? Dear ollama authors, I hope this issue finds you well. First of all, thank you for such a wonderful project. It helped me to play around with different agentic AI models and tasks. However, the quantize command is not working for me when I try to quantize the llava-next model. I have created a Modelfile on the same folder where the safetensors are placed. So for instance, if the model's safetensors are placed at /user/agent-models/llava-next, inside this folder I have created the Modelfile with context `FROM .` Now when I run the `ollama create --quantize q4_K_M mymodel` I get **gathering model components Error: no Modelfile or safetensors files found** Would this be because ollama does not support Llava architecture for quantize function? Or is there anything that I am doing wrong? Thank you and it will be much appreciated if I can hear back about this problem. Sincerely, ### OS Linux ### GPU Nvidia ### CPU _No response_ ### Ollama version 0.5.7
GiteaMirror added the bug label 2026-04-12 16:46:05 -05:00
Author
Owner

@rick-github commented on GitHub (Jan 23, 2025):

I don't know if ollama will quantize a llava model, but I don't think that's the problem you have here. The model create feature underwent some surgery in 0.5.5 and still has some rough edges. Try rolling back to 0.5.4. You can create a model in 0.5.4 and it will work in 0.5.5+.

<!-- gh-comment-id:2610113059 --> @rick-github commented on GitHub (Jan 23, 2025): I don't know if ollama will quantize a llava model, but I don't think that's the problem you have here. The model create feature underwent some surgery in 0.5.5 and still has some rough edges. Try rolling back to 0.5.4. You can create a model in 0.5.4 and it will work in 0.5.5+.
Author
Owner

@hgKang02 commented on GitHub (Jan 24, 2025):

It works when I use 0.5.4. Thank you!

<!-- gh-comment-id:2613084249 --> @hgKang02 commented on GitHub (Jan 24, 2025): It works when I use 0.5.4. Thank you!
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#5516