[GH-ISSUE #12710] After updating to v0.12.6 mxbai-embed-large stopped working. #34196

Closed
opened 2026-04-22 17:35:25 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @Triedge-sys on GitHub (Oct 20, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/12710

Originally assigned to: @jmorganca on GitHub.

What is the issue?

I'm using Silly Tavern + Ollama for vector storage. Everything was working fine, but after updating to 0.12.6, mxbai-embed-large stopped working, reporting that the context limit had been exceeded in the logs. Okay, so I installed Ollama 0.12.4 and everything works fine.

Relevant log output

time=2025-10-20T23:42:56.406+03:00 level=INFO source=server.go:1310 msg="llama runner started in 6.34 seconds"
time=2025-10-20T23:42:56.406+03:00 level=INFO source=sched.go:482 msg="loaded runners" count=1
time=2025-10-20T23:42:56.406+03:00 level=INFO source=server.go:1272 msg="waiting for llama runner to start responding"
time=2025-10-20T23:42:56.407+03:00 level=INFO source=server.go:1310 msg="llama runner started in 6.34 seconds"
time=2025-10-20T23:42:56.412+03:00 level=INFO source=server.go:1635 msg="llm embedding error: Failed to create new sequence: the input length exceeds the context length"

OS

Windows

GPU

Nvidia

CPU

Intel

Ollama version

v0.12.6

Originally created by @Triedge-sys on GitHub (Oct 20, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/12710 Originally assigned to: @jmorganca on GitHub. ### What is the issue? I'm using Silly Tavern + Ollama for vector storage. Everything was working fine, but after updating to 0.12.6, mxbai-embed-large stopped working, reporting that the context limit had been exceeded in the logs. Okay, so I installed Ollama 0.12.4 and everything works fine. ### Relevant log output ```shell time=2025-10-20T23:42:56.406+03:00 level=INFO source=server.go:1310 msg="llama runner started in 6.34 seconds" time=2025-10-20T23:42:56.406+03:00 level=INFO source=sched.go:482 msg="loaded runners" count=1 time=2025-10-20T23:42:56.406+03:00 level=INFO source=server.go:1272 msg="waiting for llama runner to start responding" time=2025-10-20T23:42:56.407+03:00 level=INFO source=server.go:1310 msg="llama runner started in 6.34 seconds" time=2025-10-20T23:42:56.412+03:00 level=INFO source=server.go:1635 msg="llm embedding error: Failed to create new sequence: the input length exceeds the context length" ``` ### OS Windows ### GPU Nvidia ### CPU Intel ### Ollama version v0.12.6
GiteaMirror added the bug label 2026-04-22 17:35:26 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#34196