[GH-ISSUE #7914] how to use Modefile build customer model when gguf file have sevral sub files? #5065

Closed
opened 2026-04-12 16:09:33 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @cqray1990 on GitHub (Dec 3, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/7914

What is the issue?

like this: qwen2.5-instruct-0001.gguf
qwen2.5-instruct-0002.gguf

OS

No response

GPU

No response

CPU

No response

Ollama version

No response

Originally created by @cqray1990 on GitHub (Dec 3, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/7914 ### What is the issue? like this: qwen2.5-instruct-0001.gguf qwen2.5-instruct-0002.gguf ### OS _No response_ ### GPU _No response_ ### CPU _No response_ ### Ollama version _No response_
GiteaMirror added the bug label 2026-04-12 16:09:33 -05:00
Author
Owner

@rick-github commented on GitHub (Dec 3, 2024):

Merge them into a single file. There are serveral projects on github that have a program to do that.

If you have docker installed, it's quite convenient to pull the llama.cpp docker image (warning, 4GB): docker pull ghcr.io/ggerganov/llama.cpp:full. It has a bunch of tools for dealing with GGUF files (splitting, merging, quantizing, etc)

$ mv qwen2.5-instruct-0001.gguf qwen2.5-instruct-00001-of-00002.gguf
$ mv qwen2.5-instruct-0002.gguf qwen2.5-instruct-00002-of-00002.gguf
$ docker run --rm -it -v .:/workdir --workdir /workdir --entrypoint /app/llama-gguf-split ghcr.io/ggerganov/llama.cpp:full --merge qwen2.5-instruct-00001-of-00002.gguf qwen2.5-instruct.gguf
<!-- gh-comment-id:2514348036 --> @rick-github commented on GitHub (Dec 3, 2024): Merge them into a single file. There are serveral projects on github that have a program to do that. If you have docker installed, it's quite convenient to pull the [llama.cpp](https://github.com/ggerganov/llama.cpp) docker image (warning, 4GB): `docker pull ghcr.io/ggerganov/llama.cpp:full`. It has a bunch of tools for dealing with GGUF files (splitting, merging, quantizing, etc) ```console $ mv qwen2.5-instruct-0001.gguf qwen2.5-instruct-00001-of-00002.gguf $ mv qwen2.5-instruct-0002.gguf qwen2.5-instruct-00002-of-00002.gguf $ docker run --rm -it -v .:/workdir --workdir /workdir --entrypoint /app/llama-gguf-split ghcr.io/ggerganov/llama.cpp:full --merge qwen2.5-instruct-00001-of-00002.gguf qwen2.5-instruct.gguf ```
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#5065