[GH-ISSUE #8464] DeepSeek-v3 fails on M4 Max Pro 128GB with ollama v0.5.7: “signal: killed” error #31205

Closed
opened 2026-04-22 11:26:16 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @totorofly on GitHub (Jan 17, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/8464

What is the issue?

Description:
I’m running a fully specced M4 Max Pro with 128GB of unified memory. When I execute:

ollama run deepseek-v3

(using deepseek-v3 with ollama v0.5.7), I encounter this error:

Error: llama runner process has terminated: signal: killed

Image

I’m using Q4_K_M quantization, which should be 4-bit. Given 128GB of unified memory, it seems like there should be enough memory available. Why is this error occurring, and is there any workaround or additional configuration needed to prevent the process from being killed?

Thank you for your help!

OS

macOS

GPU

Apple

CPU

Apple

Ollama version

v0.5.7

Originally created by @totorofly on GitHub (Jan 17, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/8464 ### What is the issue? Description: I’m running a fully specced M4 Max Pro with 128GB of unified memory. When I execute: ```ollama run deepseek-v3``` (using [deepseek-v3](https://ollama.com/library/deepseek-v3) with ollama v0.5.7), I encounter this error: ```Error: llama runner process has terminated: signal: killed``` <img width="445" alt="Image" src="https://github.com/user-attachments/assets/18610684-f54b-4a61-bf70-20133f1a7895" /> I’m using Q4_K_M quantization, which should be 4-bit. Given 128GB of unified memory, it seems like there should be enough memory available. Why is this error occurring, and is there any workaround or additional configuration needed to prevent the process from being killed? Thank you for your help! ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version v0.5.7
GiteaMirror added the bug label 2026-04-22 11:26:16 -05:00
Author
Owner

@rick-github commented on GitHub (Jan 17, 2025):

The model is 404GB. Server logs will provide insight.

<!-- gh-comment-id:2597234647 --> @rick-github commented on GitHub (Jan 17, 2025): The model is 404GB. [Server logs](https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#how-to-troubleshoot-issues) will provide insight.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#31205