[GH-ISSUE #7153] Go LLM server slice bounds error and a nil pointer dereference, leading to a server crash. #4541

Closed
opened 2026-04-12 15:28:45 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @BruceMacD on GitHub (Oct 9, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/7153

Originally assigned to: @jessegross on GitHub.

What is the issue?

While running the new experimental Go LLM server for code completion I hit a nil pointer de-reference.

This was using the /generate endpoint via the code here:
c3cbbd6fbf/core/llm/llms/Ollama.ts (L266)

time=2024-10-09T10:55:03.285-07:00 level=DEBUG source=sched.go:466 msg="context for request finished"
time=2024-10-09T10:55:03.285-07:00 level=DEBUG source=sched.go:357 msg="after processing request finished event" modelPath=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a refCount=1
time=2024-10-09T10:55:03.302-07:00 level=DEBUG source=runner.go:454 msg="hit stop token" stop="[<|endoftext|> <|endoftext|> <|fim_prefix|> <|fim_middle|> <|fim_suffix|> <|fim_pad|> <|repo_name|> <|file_sep|> <|im_start|> <|im_end|> \n\n \r\n\r\n /src/ #- coding: utf-8 ```]"
panic: runtime error: slice bounds out of range [:340] with capacity 339

goroutine 21 [running]:
main.(*Server).processBatch(0x1400014e1b0, 0x140000afee8, 0x140000afe80)
	/Users/bruce/Development/ollama/llama/runner/runner.go:461 +0x880
main.(*Server).run(0x1400014e1b0, {0x10065bab0, 0x140001e4000})
	/Users/bruce/Development/ollama/llama/runner/runner.go:316 +0x20c
created by main.main in goroutine 1
	/Users/bruce/Development/ollama/llama/runner/runner.go:853 +0xb3c
time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:1094 msg="stopping llama server"
time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:1100 msg="waiting for llama server to exit"
time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:439 msg="llama runner terminated" error="exit status 2"
time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:1105 msg="llama server stopped"
[GIN] 2024/10/09 - 10:55:03 | 200 |  448.250333ms |       127.0.0.1 | POST     "/api/generate"
time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=sched.go:407 msg="context for request finished"
time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=sched.go:339 msg="runner with non-zero duration has gone idle, adding timer" modelPath=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a duration=30m0s
time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=sched.go:357 msg="after processing request finished event" modelPath=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a refCount=0
time=2024-10-09T10:55:03.392-07:00 level=DEBUG source=sched.go:575 msg="evaluating already loaded" model=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x2 addr=0xa8 pc=0x102b6a3c0]

goroutine 10 [running]:
github.com/ollama/ollama/llm.(*llmServer).getServerStatus(0x1400013b800, {0x103af89e0?, 0x14000594fc0?})
	/Users/bruce/Development/ollama/llm/server.go:511 +0x40
github.com/ollama/ollama/llm.(*llmServer).Ping(0x103aaa040?, {0x103af89e0?, 0x14000594fc0?})
	/Users/bruce/Development/ollama/llm/server.go:585 +0x28
github.com/ollama/ollama/server.(*runnerRef).needsReload(0x140004e4b40, {0x103af8970, 0x1400059b090}, 0x1400011fe00)
	/Users/bruce/Development/ollama/server/sched.go:604 +0x324
github.com/ollama/ollama/server.(*Scheduler).processPending(0x140000ca420, {0x103af8970, 0x1400059b090})
	/Users/bruce/Development/ollama/server/sched.go:147 +0x3a8
github.com/ollama/ollama/server.(*Scheduler).Run.func1()
	/Users/bruce/Development/ollama/server/sched.go:107 +0x28
created by github.com/ollama/ollama/server.(*Scheduler).Run in goroutine 1
	/Users/bruce/Development/ollama/server/sched.go:106 +0xc4

OS

macOS

GPU

Apple

CPU

Apple

Ollama version

main branch

Originally created by @BruceMacD on GitHub (Oct 9, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/7153 Originally assigned to: @jessegross on GitHub. ### What is the issue? While running the new experimental Go LLM server for code completion I hit a nil pointer de-reference. This was using the `/generate` endpoint via the code here: https://github.com/continuedev/continue/blob/c3cbbd6fbff99b2ae920bc7a5530b80d29d2e5a9/core/llm/llms/Ollama.ts#L266 ``` time=2024-10-09T10:55:03.285-07:00 level=DEBUG source=sched.go:466 msg="context for request finished" time=2024-10-09T10:55:03.285-07:00 level=DEBUG source=sched.go:357 msg="after processing request finished event" modelPath=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a refCount=1 time=2024-10-09T10:55:03.302-07:00 level=DEBUG source=runner.go:454 msg="hit stop token" stop="[<|endoftext|> <|endoftext|> <|fim_prefix|> <|fim_middle|> <|fim_suffix|> <|fim_pad|> <|repo_name|> <|file_sep|> <|im_start|> <|im_end|> \n\n \r\n\r\n /src/ #- coding: utf-8 ```]" panic: runtime error: slice bounds out of range [:340] with capacity 339 goroutine 21 [running]: main.(*Server).processBatch(0x1400014e1b0, 0x140000afee8, 0x140000afe80) /Users/bruce/Development/ollama/llama/runner/runner.go:461 +0x880 main.(*Server).run(0x1400014e1b0, {0x10065bab0, 0x140001e4000}) /Users/bruce/Development/ollama/llama/runner/runner.go:316 +0x20c created by main.main in goroutine 1 /Users/bruce/Development/ollama/llama/runner/runner.go:853 +0xb3c time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:1094 msg="stopping llama server" time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:1100 msg="waiting for llama server to exit" time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:439 msg="llama runner terminated" error="exit status 2" time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=server.go:1105 msg="llama server stopped" [GIN] 2024/10/09 - 10:55:03 | 200 | 448.250333ms | 127.0.0.1 | POST "/api/generate" time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=sched.go:407 msg="context for request finished" time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=sched.go:339 msg="runner with non-zero duration has gone idle, adding timer" modelPath=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a duration=30m0s time=2024-10-09T10:55:03.314-07:00 level=DEBUG source=sched.go:357 msg="after processing request finished event" modelPath=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a refCount=0 time=2024-10-09T10:55:03.392-07:00 level=DEBUG source=sched.go:575 msg="evaluating already loaded" model=/Users/bruce/.ollama/models/blobs/sha256-8636de72744c6d04ee8efeddd74207eab2a9722f64766c2796b626ec10da3a4a panic: runtime error: invalid memory address or nil pointer dereference [signal SIGSEGV: segmentation violation code=0x2 addr=0xa8 pc=0x102b6a3c0] goroutine 10 [running]: github.com/ollama/ollama/llm.(*llmServer).getServerStatus(0x1400013b800, {0x103af89e0?, 0x14000594fc0?}) /Users/bruce/Development/ollama/llm/server.go:511 +0x40 github.com/ollama/ollama/llm.(*llmServer).Ping(0x103aaa040?, {0x103af89e0?, 0x14000594fc0?}) /Users/bruce/Development/ollama/llm/server.go:585 +0x28 github.com/ollama/ollama/server.(*runnerRef).needsReload(0x140004e4b40, {0x103af8970, 0x1400059b090}, 0x1400011fe00) /Users/bruce/Development/ollama/server/sched.go:604 +0x324 github.com/ollama/ollama/server.(*Scheduler).processPending(0x140000ca420, {0x103af8970, 0x1400059b090}) /Users/bruce/Development/ollama/server/sched.go:147 +0x3a8 github.com/ollama/ollama/server.(*Scheduler).Run.func1() /Users/bruce/Development/ollama/server/sched.go:107 +0x28 created by github.com/ollama/ollama/server.(*Scheduler).Run in goroutine 1 /Users/bruce/Development/ollama/server/sched.go:106 +0xc4 ``` ### OS macOS ### GPU Apple ### CPU Apple ### Ollama version main branch
GiteaMirror added the bug label 2026-04-12 15:28:45 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#4541