[GH-ISSUE #15447] Error: 400 (empty body) when pulling hf.co GGUF models after successful download #56385

Open
opened 2026-04-29 10:44:52 -05:00 by GiteaMirror · 3 comments
Owner

Originally created by @comdevx on GitHub (Apr 9, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/15447

What is the issue?

When pulling a GGUF model from HuggingFace (hf.co/...), the download completes successfully but Ollama returns Error: 400: (empty body) and fails to register the model. No manifest is created and the model does not appear in ollama list.

Reproduction steps:

ollama run hf.co/arsovskidev/Gemma-4-E4B-Claude-4.6-Opus-Reasoning-Distilled:Q8_0

Output:

pulling manifest
pulling fe4c25ea2f4e: 100% ▕████████████████████████████████████▏ 8.0 GB
Error: 400:

Also reproduced with :Q4_K_M variant (5.3 GB).

GGUF file details

Inspecting the downloaded blob confirms it is a valid GGUF v3 file:

  • general.architecture = gemma4
  • general.quantized_by = Unsloth
  • 720 tensors, 42 KV pairs

Workaround

Manually creating the model from the already-downloaded blob works:

cat > Modelfile << 'MODELFILE'
FROM /path/to/.ollama/models/blobs/sha256-<hash>

TEMPLATE """<bos>{{ if .System }}<start_of_turn>user
{{ .System }}<end_of_turn>
{{ end }}{{ range .Messages }}{{ if eq .Role "user" }}<start_of_turn>user
{{ .Content }}<end_of_turn>
<start_of_turn>model
{{ else if eq .Role "assistant" }}{{ .Content }}<end_of_turn>
{{ end }}{{ end }}<start_of_turn>model
"""

PARAMETER stop "<end_of_turn>"
PARAMETER stop "<eos>"
MODELFILE

ollama create my-model -f Modelfile

Expected behavior

Model should be registered and runnable after a successful download, same as ollama create from a local blob.

Relevant log output

Server log only shows 200 POST /api/pull — no error details logged server-side for the 400 response.

OS

macOS Darwin 25.3.0 (Apple Silicon)

GPU

Apple Silicon (Metal)

CPU

Apple Silicon

Ollama version

ollama version is 0.20.4
Originally created by @comdevx on GitHub (Apr 9, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/15447 ### What is the issue? When pulling a GGUF model from HuggingFace (`hf.co/...`), the download completes successfully but Ollama returns `Error: 400:` (empty body) and fails to register the model. No manifest is created and the model does not appear in `ollama list`. **Reproduction steps:** ``` ollama run hf.co/arsovskidev/Gemma-4-E4B-Claude-4.6-Opus-Reasoning-Distilled:Q8_0 ``` Output: ``` pulling manifest pulling fe4c25ea2f4e: 100% ▕████████████████████████████████████▏ 8.0 GB Error: 400: ``` Also reproduced with `:Q4_K_M` variant (5.3 GB). ### GGUF file details Inspecting the downloaded blob confirms it is a valid GGUF v3 file: - `general.architecture = gemma4` - `general.quantized_by = Unsloth` - 720 tensors, 42 KV pairs ### Workaround Manually creating the model from the already-downloaded blob works: ```bash cat > Modelfile << 'MODELFILE' FROM /path/to/.ollama/models/blobs/sha256-<hash> TEMPLATE """<bos>{{ if .System }}<start_of_turn>user {{ .System }}<end_of_turn> {{ end }}{{ range .Messages }}{{ if eq .Role "user" }}<start_of_turn>user {{ .Content }}<end_of_turn> <start_of_turn>model {{ else if eq .Role "assistant" }}{{ .Content }}<end_of_turn> {{ end }}{{ end }}<start_of_turn>model """ PARAMETER stop "<end_of_turn>" PARAMETER stop "<eos>" MODELFILE ollama create my-model -f Modelfile ``` ### Expected behavior Model should be registered and runnable after a successful download, same as `ollama create` from a local blob. ### Relevant log output Server log only shows `200 POST /api/pull` — no error details logged server-side for the 400 response. ### OS macOS Darwin 25.3.0 (Apple Silicon) ### GPU Apple Silicon (Metal) ### CPU Apple Silicon ### Ollama version ``` ollama version is 0.20.4 ```
GiteaMirror added the networking label 2026-04-29 10:44:52 -05:00
Author
Owner

@rick-github commented on GitHub (Apr 9, 2026):

It fails because fetching the config blob specified in the manifest pulled from HF returns an error:

  "error": "The specified tag is not available in the repository. Please use another tag or \"latest\""
<!-- gh-comment-id:4213134585 --> @rick-github commented on GitHub (Apr 9, 2026): It fails because fetching the config blob specified in the manifest pulled from HF returns an error: ``` "error": "The specified tag is not available in the repository. Please use another tag or \"latest\"" ```
Author
Owner

@mcd1992 commented on GitHub (Apr 15, 2026):

Getting the same issue and your workaround does work. Thank you.

It fails because fetching the config blob specified in the manifest pulled from HF returns an error:

  "error": "The specified tag is not available in the repository. Please use another tag or \"latest\""

The error happens after it spends time actually downloading the model parts. And it happens even if you explicitly give a tag that exists like latest or BF16.

<!-- gh-comment-id:4256264212 --> @mcd1992 commented on GitHub (Apr 15, 2026): Getting the same issue and your workaround does work. Thank you. > It fails because fetching the config blob specified in the manifest pulled from HF returns an error: > > ``` > "error": "The specified tag is not available in the repository. Please use another tag or \"latest\"" > ``` The error happens after it spends time actually downloading the model parts. And it happens even if you explicitly give a tag that exists like `latest` or `BF16`.
Author
Owner

@PureBlissAK commented on GitHub (Apr 18, 2026):

🤖 Automated Triage & Analysis Report

Issue: #15447
Analyzed: 2026-04-18T18:21:30.083083

Analysis

  • Type: unknown
  • Severity: medium
  • Components: unknown

Implementation Plan

  • Effort: medium
  • Steps:

This issue has been triaged and marked for implementation.

<!-- gh-comment-id:4274308404 --> @PureBlissAK commented on GitHub (Apr 18, 2026): <!-- ollama-issue-orchestrator:v1 issue:15447 --> ## 🤖 Automated Triage & Analysis Report **Issue**: #15447 **Analyzed**: 2026-04-18T18:21:30.083083 ### Analysis - **Type**: unknown - **Severity**: medium - **Components**: unknown ### Implementation Plan - **Effort**: medium - **Steps**: *This issue has been triaged and marked for implementation.*
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#56385