[GH-ISSUE #15447] Error: 400 (empty body) when pulling hf.co GGUF models after successful download #9872

Open
opened 2026-04-12 22:44:00 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @comdevx on GitHub (Apr 9, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/15447

What is the issue?

When pulling a GGUF model from HuggingFace (hf.co/...), the download completes successfully but Ollama returns Error: 400: (empty body) and fails to register the model. No manifest is created and the model does not appear in ollama list.

Reproduction steps:

ollama run hf.co/arsovskidev/Gemma-4-E4B-Claude-4.6-Opus-Reasoning-Distilled:Q8_0

Output:

pulling manifest
pulling fe4c25ea2f4e: 100% ▕████████████████████████████████████▏ 8.0 GB
Error: 400:

Also reproduced with :Q4_K_M variant (5.3 GB).

GGUF file details

Inspecting the downloaded blob confirms it is a valid GGUF v3 file:

  • general.architecture = gemma4
  • general.quantized_by = Unsloth
  • 720 tensors, 42 KV pairs

Workaround

Manually creating the model from the already-downloaded blob works:

cat > Modelfile << 'MODELFILE'
FROM /path/to/.ollama/models/blobs/sha256-<hash>

TEMPLATE """<bos>{{ if .System }}<start_of_turn>user
{{ .System }}<end_of_turn>
{{ end }}{{ range .Messages }}{{ if eq .Role "user" }}<start_of_turn>user
{{ .Content }}<end_of_turn>
<start_of_turn>model
{{ else if eq .Role "assistant" }}{{ .Content }}<end_of_turn>
{{ end }}{{ end }}<start_of_turn>model
"""

PARAMETER stop "<end_of_turn>"
PARAMETER stop "<eos>"
MODELFILE

ollama create my-model -f Modelfile

Expected behavior

Model should be registered and runnable after a successful download, same as ollama create from a local blob.

Relevant log output

Server log only shows 200 POST /api/pull — no error details logged server-side for the 400 response.

OS

macOS Darwin 25.3.0 (Apple Silicon)

GPU

Apple Silicon (Metal)

CPU

Apple Silicon

Ollama version

ollama version is 0.20.4
Originally created by @comdevx on GitHub (Apr 9, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/15447 ### What is the issue? When pulling a GGUF model from HuggingFace (`hf.co/...`), the download completes successfully but Ollama returns `Error: 400:` (empty body) and fails to register the model. No manifest is created and the model does not appear in `ollama list`. **Reproduction steps:** ``` ollama run hf.co/arsovskidev/Gemma-4-E4B-Claude-4.6-Opus-Reasoning-Distilled:Q8_0 ``` Output: ``` pulling manifest pulling fe4c25ea2f4e: 100% ▕████████████████████████████████████▏ 8.0 GB Error: 400: ``` Also reproduced with `:Q4_K_M` variant (5.3 GB). ### GGUF file details Inspecting the downloaded blob confirms it is a valid GGUF v3 file: - `general.architecture = gemma4` - `general.quantized_by = Unsloth` - 720 tensors, 42 KV pairs ### Workaround Manually creating the model from the already-downloaded blob works: ```bash cat > Modelfile << 'MODELFILE' FROM /path/to/.ollama/models/blobs/sha256-<hash> TEMPLATE """<bos>{{ if .System }}<start_of_turn>user {{ .System }}<end_of_turn> {{ end }}{{ range .Messages }}{{ if eq .Role "user" }}<start_of_turn>user {{ .Content }}<end_of_turn> <start_of_turn>model {{ else if eq .Role "assistant" }}{{ .Content }}<end_of_turn> {{ end }}{{ end }}<start_of_turn>model """ PARAMETER stop "<end_of_turn>" PARAMETER stop "<eos>" MODELFILE ollama create my-model -f Modelfile ``` ### Expected behavior Model should be registered and runnable after a successful download, same as `ollama create` from a local blob. ### Relevant log output Server log only shows `200 POST /api/pull` — no error details logged server-side for the 400 response. ### OS macOS Darwin 25.3.0 (Apple Silicon) ### GPU Apple Silicon (Metal) ### CPU Apple Silicon ### Ollama version ``` ollama version is 0.20.4 ```
GiteaMirror added the networking label 2026-04-12 22:44:00 -05:00
Author
Owner

@rick-github commented on GitHub (Apr 9, 2026):

It fails because fetching the config blob specified in the manifest pulled from HF returns an error:

  "error": "The specified tag is not available in the repository. Please use another tag or \"latest\""
<!-- gh-comment-id:4213134585 --> @rick-github commented on GitHub (Apr 9, 2026): It fails because fetching the config blob specified in the manifest pulled from HF returns an error: ``` "error": "The specified tag is not available in the repository. Please use another tag or \"latest\"" ```
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#9872