[GH-ISSUE #12266] Ollama serve crashes on windows #33915

Closed
opened 2026-04-22 17:05:54 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @wanshicheng on GitHub (Sep 12, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/12266

What is the issue?

ResponseError: do embedding request: Post "http://127.0.0.1:51742/embedding": read tcp 127.0.0.1:51769->127.0.0.1:51742: wsarecv: An existing connection was forcibly closed by the remote host. (status code: 500)

Relevant log output

time=2025-09-12T17:23:43.269+08:00 level=INFO source=server.go:1288 msg="llama runner started in 2.07 seconds"
time=2025-09-12T17:23:43.269+08:00 level=INFO source=sched.go:473 msg="loaded runners" count=1
time=2025-09-12T17:23:43.269+08:00 level=INFO source=server.go:1250 msg="waiting for llama runner to start responding"
time=2025-09-12T17:23:43.270+08:00 level=INFO source=server.go:1288 msg="llama runner started in 2.08 seconds"
C:/a/ollama/ollama/ml/backend/ggml/ggml/src/ggml-cpu/ops.cpp:5280: GGML_ASSERT(i01 >= 0 && i01 < ne01) failed
[GIN] 2025/09/12 - 17:23:43 | 500 |    4.1103598s |       127.0.0.1 | POST     "/api/embed"
time=2025-09-12T17:23:43.807+08:00 level=ERROR source=server.go:424 msg="llama runner terminated" error="exit status 0xc0000409"

OS

Windows

GPU

Intel, Nvidia

CPU

Intel

Ollama version

0.11.10

Originally created by @wanshicheng on GitHub (Sep 12, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/12266 ### What is the issue? ResponseError: do embedding request: Post "http://127.0.0.1:51742/embedding": read tcp 127.0.0.1:51769->127.0.0.1:51742: wsarecv: An existing connection was forcibly closed by the remote host. (status code: 500) ### Relevant log output ```shell time=2025-09-12T17:23:43.269+08:00 level=INFO source=server.go:1288 msg="llama runner started in 2.07 seconds" time=2025-09-12T17:23:43.269+08:00 level=INFO source=sched.go:473 msg="loaded runners" count=1 time=2025-09-12T17:23:43.269+08:00 level=INFO source=server.go:1250 msg="waiting for llama runner to start responding" time=2025-09-12T17:23:43.270+08:00 level=INFO source=server.go:1288 msg="llama runner started in 2.08 seconds" C:/a/ollama/ollama/ml/backend/ggml/ggml/src/ggml-cpu/ops.cpp:5280: GGML_ASSERT(i01 >= 0 && i01 < ne01) failed [GIN] 2025/09/12 - 17:23:43 | 500 | 4.1103598s | 127.0.0.1 | POST "/api/embed" time=2025-09-12T17:23:43.807+08:00 level=ERROR source=server.go:424 msg="llama runner terminated" error="exit status 0xc0000409" ``` ### OS Windows ### GPU Intel, Nvidia ### CPU Intel ### Ollama version 0.11.10
GiteaMirror added the bug label 2026-04-22 17:05:54 -05:00
Author
Owner

@rick-github commented on GitHub (Sep 12, 2025):

Model? Input? More log might help.

<!-- gh-comment-id:3285153623 --> @rick-github commented on GitHub (Sep 12, 2025): Model? Input? More log might help.
Author
Owner

@wanshicheng commented on GitHub (Sep 12, 2025):

Model? Input? More log might help.

Before upgrading to this version 0.11.10, my code was functioning properly.

The code is very simple:

from langchain_ollama import OllamaEmbeddings

embeddings = OllamaEmbeddings(model="ryanshillington/Qwen3-Embedding-0.6B:latest")
embeddings.embed_query("Hello, world!")

The full log:

server.log

<!-- gh-comment-id:3285449980 --> @wanshicheng commented on GitHub (Sep 12, 2025): > Model? Input? More log might help. Before upgrading to this version 0.11.10, my code was functioning properly. The code is very simple: ```python from langchain_ollama import OllamaEmbeddings embeddings = OllamaEmbeddings(model="ryanshillington/Qwen3-Embedding-0.6B:latest") embeddings.embed_query("Hello, world!") ``` The full log: [server.log](https://github.com/user-attachments/files/22299898/server.log)
Author
Owner

@rick-github commented on GitHub (Sep 12, 2025):

#12014

<!-- gh-comment-id:3285866550 --> @rick-github commented on GitHub (Sep 12, 2025): #12014
Author
Owner

@pdevine commented on GitHub (Sep 12, 2025):

Going to close as a dupe.

<!-- gh-comment-id:3286885558 --> @pdevine commented on GitHub (Sep 12, 2025): Going to close as a dupe.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#33915