mirror of
https://github.com/ollama/ollama.git
synced 2025-12-05 18:46:22 -06:00
Ollama Runner Fails with “Exit Status 2” and Random Non-Responsive Behavior on Windows #8587
Open
opened 2025-11-12 14:46:34 -06:00 by GiteaMirror
·
23 comments
No Branch/Tag Specified
main
parth/olmo3-thinking-renderer
drifkin/responses
jmorganca/write-int32
parth/render-template
jmorganca/intellect-3
parth/remove-prealloc-linter
nicole/olmo-model
jmorganca/cmd-eval
grace/deepseek-parser
nicole/nomic-embed-text-fix
jessegross/multi_chunk_reserve
grace/deepseek-renderer-parser
hoyyeva/proxy-ollama-serve
hoyyeva/model-capabilities-after-download
mxyng/ropeopts
mxyng/lint-2
nicole/nomic-embed-text-v2
hoyyeva/add-gemini-3-pro-preview
hoyyeva/load-model-list
hoyyeva/app-ollama-apis
mxyng/expand-path
mxyng/environ-2
hoyyeva/thinking-rendering
nicole/truncation
hoyyeva/deeplink-json-encoding
parth/improve-tool-calling-tests
hoyyeva/conversation
hoyyeva/assistant-edit-response
hoyyeva/thinking
origin/brucemacd/invalid-char-i-err
parth/improve-tool-calling
jmorganca/required-omitempty
grace/qwen3-vl-tests
mxyng/iter-client
parth/docs-readme
nicole/embed-test
pdevine/integration-benchstat
parth/remove-generate-cmd
parth/add-toolcall-id
mxyng/server-tests
jmorganca/glm-4.6
jmorganca/gin-h-compat
drifkin/stable-tool-args
pdevine/qwen3-more-thinking
parth/add-websearch-client
nicole/websearch_local
jmorganca/qwen3-coder-updates
grace/deepseek-v3-migration-tests
mxyng/fix-create
jmorganca/cloud-errors
pdevine/parser-tidy
revert-12233-parth/simplify-entrypoints-runner
parth/enable-so-gpt-oss
brucemacd/qwen3vl
jmorganca/readme-simplify
parth/gpt-oss-structured-outputs
revert-12039-jmorganca/tools-braces
mxyng/embeddings
mxyng/gguf
mxyng/benchmark
mxyng/types-null
parth/move-parsing
mxyng/gemma2
jmorganca/docs
mxyng/16-bit
mxyng/create-stdin
pdevine/authorizedkeys
mxyng/quant
parth/opt-in-error-context-window
brucemacd/cache-models
brucemacd/runner-completion
jmorganca/llama-update-6
brucemacd/benchmark-list
brucemacd/partial-read-caps
parth/deepseek-r1-tools
mxyng/omit-array
parth/tool-prefix-temp
brucemacd/runner-test
jmorganca/qwen25vl
brucemacd/model-forward-test-ext
parth/python-function-parsing
jmorganca/cuda-compression-none
drifkin/num-parallel
drifkin/chat-truncation-fix
jmorganca/sync
parth/python-tools-calling
drifkin/array-head-count
brucemacd/create-no-loop
parth/server-enable-content-stream-with-tools
qwen25omni
mxyng/v3
brucemacd/ropeconfig
jmorganca/silence-tokenizer
parth/sample-so-test
parth/sampling-structured-outputs
brucemacd/doc-go-engine
parth/constrained-sampling-json
jmorganca/mistral-wip
brucemacd/mistral-small-convert
parth/sample-unmarshal-json-for-params
brucemacd/jomorganca/mistral
pdevine/bfloat16
jmorganca/mistral
brucemacd/mistral
pdevine/logging
parth/sample-correctness-fix
parth/sample-fix-sorting
jmorgan/sample-fix-sorting-extras
jmorganca/temp-0-images
brucemacd/parallel-embed-models
brucemacd/shim-grammar
jmorganca/fix-gguf-error
bmizerany/nameswork
jmorganca/faster-releases
bmizerany/validatenames
brucemacd/err-no-vocab
brucemacd/rope-config
brucemacd/err-hint
brucemacd/qwen2_5
brucemacd/logprobs
brucemacd/new_runner_graph_bench
progress-flicker
brucemacd/forward-test
brucemacd/go_qwen2
pdevine/gemma2
jmorganca/add-missing-symlink-eval
mxyng/next-debug
parth/set-context-size-openai
brucemacd/next-bpe-bench
brucemacd/next-bpe-test
brucemacd/new_runner_e2e
brucemacd/new_runner_qwen2
pdevine/convert-cohere2
brucemacd/convert-cli
parth/log-probs
mxyng/next-mlx
mxyng/cmd-history
parth/templating
parth/tokenize-detokenize
brucemacd/check-key-register
bmizerany/grammar
jmorganca/vendor-081b29bd
mxyng/func-checks
jmorganca/fix-null-format
parth/fix-default-to-warn-json
jmorganca/qwen2vl
jmorganca/no-concat
parth/cmd-cleanup-SO
brucemacd/check-key-register-structured-err
parth/openai-stream-usage
parth/fix-referencing-so
stream-tools-stop
jmorganca/degin-1
brucemacd/install-path-clean
brucemacd/push-name-validation
brucemacd/browser-key-register
jmorganca/openai-fix-first-message
jmorganca/fix-proxy
jessegross/sample
parth/disallow-streaming-tools
dhiltgen/remove_submodule
jmorganca/ga
jmorganca/mllama
pdevine/newlines
pdevine/geems-2b
jmorganca/llama-bump
mxyng/modelname-7
mxyng/gin-slog
mxyng/modelname-6
jyan/convert-prog
jyan/quant5
paligemma-support
pdevine/import-docs
jmorganca/openai-context
jyan/paligemma
jyan/p2
jyan/palitest
bmizerany/embedspeedup
jmorganca/llama-vit
brucemacd/allow-ollama
royh/ep-methods
royh/whisper
mxyng/api-models
mxyng/fix-memory
jyan/q4_4/8
jyan/ollama-v
royh/stream-tools
roy-embed-parallel
bmizerany/hrm
revert-5963-revert-5924-mxyng/llama3.1-rope
royh/embed-viz
jyan/local2
jyan/auth
jyan/local
jyan/parse-temp
jmorganca/template-mistral
jyan/reord-g
royh-openai-suffixdocs
royh-imgembed
royh-embed-parallel
jyan/quant4
royh-precision
jyan/progress
pdevine/fix-template
jyan/quant3
pdevine/ggla
mxyng/update-registry-domain
jmorganca/ggml-static
mxyng/create-context
jyan/v0.146
mxyng/layers-from-files
build_dist
bmizerany/noseek
royh-ls
royh-name
timeout
mxyng/server-timestamp
bmizerany/nosillyggufslurps
royh-params
jmorganca/llama-cpp-7c26775
royh-openai-delete
royh-show-rigid
jmorganca/enable-fa
jmorganca/no-error-template
jyan/format
royh-testdelete
bmizerany/fastverify
language_support
pdevine/ps-glitches
brucemacd/tokenize
bruce/iq-quants
bmizerany/filepathwithcoloninhost
mxyng/split-bin
bmizerany/client-registry
jmorganca/if-none-match
native
jmorganca/native
jmorganca/batch-embeddings
jmorganca/initcmake
jmorganca/mm
pdevine/showggmlinfo
modenameenforcealphanum
bmizerany/modenameenforcealphanum
jmorganca/done-reason
jmorganca/llama-cpp-8960fe8
ollama.com
bmizerany/filepathnobuild
bmizerany/types/model/defaultfix
rmdisplaylong
nogogen
bmizerany/x
modelfile-readme
bmizerany/replacecolon
jmorganca/limit
jmorganca/execstack
jmorganca/replace-assets
mxyng/tune-concurrency
jmorganca/testing
whitespace-detection
jmorganca/options
upgrade-all
scratch
cuda-search
mattw/airenamer
mattw/allmodelsonhuggingface
mattw/quantcontext
mattw/whatneedstorun
brucemacd/llama-mem-calc
mattw/faq-context
mattw/communitylinks
mattw/noprune
mattw/python-functioncalling
rename
mxyng/install
pulse
remove-first
editor
mattw/selfqueryingretrieval
cgo
mattw/howtoquant
api
matt/streamingapi
format-config
mxyng/extra-args
shell
update-nous-hermes
cp-model
upload-progress
fix-unknown-model
fix-model-names
delete-fix
insecure-registry
ls
deletemodels
progressbar
readme-updates
license-layers
skip-list
list-models
modelpath
matt/examplemodelfiles
distribution
go-opts
v0.13.2-rc2
v0.13.2-rc1
v0.13.2-rc0
v0.13.1
v0.13.1-rc2
v0.13.1-rc1
v0.13.1-rc0
v0.13.0
v0.13.0-rc0
v0.12.11
v0.12.11-rc1
v0.12.11-rc0
v0.12.10
v0.12.10-rc1
v0.12.10-rc0
v0.12.9-rc0
v0.12.9
v0.12.8
v0.12.8-rc0
v0.12.7
v0.12.7-rc1
v0.12.7-rc0
v0.12.6
v0.12.6-rc1
v0.12.6-rc0
v0.12.5
v0.12.5-rc0
v0.12.4
v0.12.4-rc7
v0.12.4-rc6
v0.12.4-rc5
v0.12.4-rc4
v0.12.4-rc3
v0.12.4-rc2
v0.12.4-rc1
v0.12.4-rc0
v0.12.3
v0.12.2
v0.12.2-rc0
v0.12.1
v0.12.1-rc1
v0.12.1-rc2
v0.12.1-rc0
v0.12.0
v0.12.0-rc1
v0.12.0-rc0
v0.11.11
v0.11.11-rc3
v0.11.11-rc2
v0.11.11-rc1
v0.11.11-rc0
v0.11.10
v0.11.9
v0.11.9-rc0
v0.11.8
v0.11.8-rc0
v0.11.7-rc1
v0.11.7-rc0
v0.11.7
v0.11.6
v0.11.6-rc0
v0.11.5-rc4
v0.11.5-rc3
v0.11.5
v0.11.5-rc5
v0.11.5-rc2
v0.11.5-rc1
v0.11.5-rc0
v0.11.4
v0.11.4-rc0
v0.11.3
v0.11.3-rc0
v0.11.2
v0.11.1
v0.11.0
v0.10.1
v0.10.0
v0.10.0-rc4
v0.10.0-rc3
v0.10.0-rc2
v0.10.0-rc1
v0.10.0-rc0
v0.9.7-rc1
v0.9.7-rc0
v0.9.6
v0.9.6-rc0
v0.9.5
v0.9.4-rc5
v0.9.4-rc6
v0.9.4
v0.9.4-rc3
v0.9.4-rc4
v0.9.4-rc1
v0.9.4-rc2
v0.9.4-rc0
v0.9.3
v0.9.3-rc5
v0.9.4-citest0
v0.9.3-rc4
v0.9.3-rc3
v0.9.3-rc2
v0.9.3-rc1
v0.9.3-rc0
v0.9.2
v0.9.1
v0.9.1-rc1
v0.9.1-rc0
v0.9.0
v0.9.0-rc0
v0.8.0
v0.8.0-rc0
v0.7.1-rc2
v0.7.1
v0.7.1-rc1
v0.7.1-rc0
v0.7.0
v0.7.0-rc1
v0.7.0-rc0
v0.6.8
v0.6.8-rc0
v0.6.7
v0.6.7-rc2
v0.6.7-rc1
v0.6.7-rc0
v0.6.6
v0.6.6-rc2
v0.6.6-rc1
v0.6.6-rc0
v0.6.5-rc1
v0.6.5
v0.6.5-rc0
v0.6.4-rc0
v0.6.4
v0.6.3-rc1
v0.6.3
v0.6.3-rc0
v0.6.2
v0.6.2-rc0
v0.6.1
v0.6.1-rc0
v0.6.0-rc0
v0.6.0
v0.5.13
v0.5.13-rc6
v0.5.13-rc5
v0.5.13-rc4
v0.5.13-rc3
v0.5.13-rc2
v0.5.13-rc1
v0.5.13-rc0
v0.5.12
v0.5.12-rc1
v0.5.12-rc0
v0.5.11
v0.5.10
v0.5.9
v0.5.9-rc0
v0.5.8-rc13
v0.5.8
v0.5.8-rc12
v0.5.8-rc11
v0.5.8-rc10
v0.5.8-rc9
v0.5.8-rc8
v0.5.8-rc7
v0.5.8-rc6
v0.5.8-rc5
v0.5.8-rc4
v0.5.8-rc3
v0.5.8-rc2
v0.5.8-rc1
v0.5.8-rc0
v0.5.7
v0.5.6
v0.5.5
v0.5.5-rc0
v0.5.4
v0.5.3
v0.5.3-rc0
v0.5.2
v0.5.2-rc3
v0.5.2-rc2
v0.5.2-rc1
v0.5.2-rc0
v0.5.1
v0.5.0
v0.5.0-rc1
v0.4.8-rc0
v0.4.7
v0.4.6
v0.4.5
v0.4.4
v0.4.3
v0.4.3-rc0
v0.4.2
v0.4.2-rc1
v0.4.2-rc0
v0.4.1
v0.4.1-rc0
v0.4.0
v0.4.0-rc8
v0.4.0-rc7
v0.4.0-rc6
v0.4.0-rc5
v0.4.0-rc4
v0.4.0-rc3
v0.4.0-rc2
v0.4.0-rc1
v0.4.0-rc0
v0.4.0-ci3
v0.3.14
v0.3.14-rc0
v0.3.13
v0.3.12
v0.3.12-rc5
v0.3.12-rc4
v0.3.12-rc3
v0.3.12-rc2
v0.3.12-rc1
v0.3.11
v0.3.11-rc4
v0.3.11-rc3
v0.3.11-rc2
v0.3.11-rc1
v0.3.10
v0.3.10-rc1
v0.3.9
v0.3.8
v0.3.7
v0.3.7-rc6
v0.3.7-rc5
v0.3.7-rc4
v0.3.7-rc3
v0.3.7-rc2
v0.3.7-rc1
v0.3.6
v0.3.5
v0.3.4
v0.3.3
v0.3.2
v0.3.1
v0.3.0
v0.2.8
v0.2.8-rc2
v0.2.8-rc1
v0.2.7
v0.2.6
v0.2.5
v0.2.4
v0.2.3
v0.2.2
v0.2.2-rc2
v0.2.2-rc1
v0.2.1
v0.2.0
v0.1.49-rc14
v0.1.49-rc13
v0.1.49-rc12
v0.1.49-rc11
v0.1.49-rc10
v0.1.49-rc9
v0.1.49-rc8
v0.1.49-rc7
v0.1.49-rc6
v0.1.49-rc4
v0.1.49-rc5
v0.1.49-rc3
v0.1.49-rc2
v0.1.49-rc1
v0.1.48
v0.1.47
v0.1.46
v0.1.45-rc5
v0.1.45
v0.1.45-rc4
v0.1.45-rc3
v0.1.45-rc2
v0.1.45-rc1
v0.1.44
v0.1.43
v0.1.42
v0.1.41
v0.1.40
v0.1.40-rc1
v0.1.39
v0.1.39-rc2
v0.1.39-rc1
v0.1.38
v0.1.37
v0.1.36
v0.1.35
v0.1.35-rc1
v0.1.34
v0.1.34-rc1
v0.1.33
v0.1.33-rc7
v0.1.33-rc6
v0.1.33-rc5
v0.1.33-rc4
v0.1.33-rc3
v0.1.33-rc2
v0.1.33-rc1
v0.1.32
v0.1.32-rc2
v0.1.32-rc1
v0.1.31
v0.1.30
v0.1.29
v0.1.28
v0.1.27
v0.1.26
v0.1.25
v0.1.24
v0.1.23
v0.1.22
v0.1.21
v0.1.20
v0.1.19
v0.1.18
v0.1.17
v0.1.16
v0.1.15
v0.1.14
v0.1.13
v0.1.12
v0.1.11
v0.1.10
v0.1.9
v0.1.8
v0.1.7
v0.1.6
v0.1.5
v0.1.4
v0.1.3
v0.1.2
v0.1.1
v0.1.0
v0.0.21
v0.0.20
v0.0.19
v0.0.18
v0.0.17
v0.0.16
v0.0.15
v0.0.14
v0.0.13
v0.0.12
v0.0.11
v0.0.10
v0.0.9
v0.0.8
v0.0.7
v0.0.6
v0.0.5
v0.0.4
v0.0.3
v0.0.2
v0.0.1
Labels
Clear labels
amd
api
app
bug
build
cli
client2
cloud
compatibility
context-length
create
docker
documentation
embeddings
engine
feature request
feedback wanted
good first issue
gpt-oss
gpu
harmony
help wanted
install
integration
intel
js
linux
macos
memory
model
needs more info
networking
nvidia
ollama.com
performance
pull-request
Mirrored from GitHub Pull Request
python
question
registry
rendering
thinking
tools
top
windows
wsl
No Label
bug
Milestone
No items
No Milestone
Projects
Clear projects
No project
No Assignees
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: github-starred/ollama-ollama#8587
Reference in New Issue
Block a user
Blocking a user prevents them from interacting with repositories, such as opening or commenting on pull requests or issues. Learn more about blocking a user.
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @Anurag1940 on GitHub (Nov 4, 2025).
Ollama Runner fails intermittently on Windows when running models like llama3.2, gemma3:4b, and phi3:mini.
When executing a simple command such as:
ollama run llama3.2 "Hello"
it either terminates immediately with:
Error: 500 Internal Server Error: llama runner process has terminated: exit status 2
or hangs indefinitely without providing any output or visible error.
This happens both in GPU mode and CPU-only mode ($env:OLLAMA_NO_GPU=1).
Expected behavior: the model should initialize and respond normally without termination or hanging.
Actual behavior:
The process stops abruptly or becomes non-responsive.
Logs indicate “entering low VRAM mode” despite having sufficient system memory (~11.7 GiB total).
Restarting the Ollama daemon and re-pulling models did not resolve the issue.
Relevant log output:
time=2025-11-01T10:28:26.946+05:30 level=INFO source=routes.go:1577 msg="Listening on 127.0.0.1:11434 (version 0.12.8)"
time=2025-11-01T10:28:29.534+05:30 level=INFO source=routes.go:1618 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"
Error: 500 Internal Server Error: llama runner process has terminated: exit status 2
System details:
OS: Windows 11 (PowerShell environment)
Ollama version: 0.12.8
Installed models: llama3.2:latest, gemma3:4b, phi3:mini
System memory: 11.7 GiB total / 1.2 GiB available
Tested in both GPU and CPU-only configurations
Troubleshooting steps already performed:
Restarted Ollama service and system
Cleared cache and re-pulled models
Verified ports and memory allocation
Switched between GPU and CPU modes
Despite these steps, the runner process remains unstable and occasionally fails without any visible logs or output.
Requesting guidance on possible configuration adjustments, additional debug parameters, or diagnostic utilities to trace this behavior further.
server-1.log
server-2.log
server-3.log
server-4.log
OS
Windows
GPU
Intel
CPU
Intel
Ollama version
0.12.9
@rick-github commented on GitHub (Nov 4, 2025):
OLLAMA_NO_GPUis not an ollama configuration variable so has no effect. But the logs show ollama never successfully detects a GPU, so CPU is always used. The logs also don't show a model load or runner crash, so there's little information to go on. If you setOLLAMA_DEBUG=2and post the resulting logs it will be easier to make progress,@dhiltgen commented on GitHub (Nov 4, 2025):
12G of system memory, with only 2.7G available isn't going to be able to load very many models.
Intel GPUs are not officially supported yet, but Vulkan support is coming soon which will enable many Intel GPUs. However if your GPU is an iGPU, it may struggle to load models with so little available memory.
@Nantris commented on GitHub (Nov 9, 2025):
Upgraded from
0.12.3to0.12.10on Windows 11 and Ollama is entirely unusable now with the same error,Error: 500 Internal Server Error: llama runner process has terminated: exit status 2Nvidia GPU and plenty of RAM - everything worked fine on
0.12.3but I was getting empty responses fromgranite4:7b-a1b-hintermittently so I upgraded and now I can't run any model.I noticed the program is no longer
ollama.exebut insteadollama app.exeand I noticed it makes anollama app.exefolder in AppData. I wonder if any of this (the file extension in the folder name or the space in the executable name) are causing issues.@Nantris commented on GitHub (Nov 9, 2025):
Also tried setting
OLLAMA_DEBUG=1andOLLAMA_DEBUG=2but nothing prints besides that error, and nothing is logged to any file either.@Nantris commented on GitHub (Nov 9, 2025):
0.12.3is the last version I can useollama run granite4:7b-a1b-hwithout facing this error.When I tried running any model in
0.12.4in the GUI I got:400 Bad Request: registry.ollama.ai/library/granite4:7b-a1b-h does not support thinking
But in the CLI it's the message from above (
Server Error: llama runner process has terminated: exit status 2)@rick-github commented on GitHub (Nov 9, 2025):
ollama.exe is the CLI/server, ollama app.exe is the UI. You have to set
OLLAMA_DEBUGin the server environment for it to have any effect, and then check theserver.logfile in%LOCALAPPDATA%\Ollama.@Nantris commented on GitHub (Nov 9, 2025):
I didn't see
ollama.exerunning anymore when I usedollama runafter0.12.3. I doubt that's the issue, but maybe something to look into. I just spent 30 minutes installing various versions so I'm not inclined to do any more bisecting now that I'm back on0.12.3and it runs (albeit maybe with tool-calling bugs)@Nantris commented on GitHub (Nov 9, 2025):
Oh and to clarify the above, in case it's unclear, I was using the CLI/server so I can confirm 100% nothing gets logged whatsoever (at least in
0.12.10- I didn't test in any older version except0.12.3I know it works)@rick-github commented on GitHub (Nov 9, 2025):
Stop the ollama server by clicking on the systray icon and selecting "Quit Ollama". Open a CMD window and run the following:
Then open a second CMD windows and run:
What's the output in the first CMD window?
@Nantris commented on GitHub (Nov 9, 2025):
Thank you for your replies @rick-github. I did do exactly that. Respectfully, I think you're underestimating my technical proficiency. The output is as previously stated:
Server Error: llama runner process has terminated: exit status 2Unfortunately because it logs nothing anywhere that's all I can offer. Nothing in Windows Event Logs either. I installed with the
OllamaSetup.exeand I exited Windows Terminal between each new install.@rick-github commented on GitHub (Nov 9, 2025):
What command do you run that displays
Server Error: llama runner process has terminated: exit status 2?@Nantris commented on GitHub (Nov 9, 2025):
ollama run granite4:7b-a1b-hExchange any other model and the error is the same.
@rick-github commented on GitHub (Nov 9, 2025):
ollama run granite4:7b-a1b-hcannot emit that message without connecting to a server. In the first CMD window in my advice above, you should either have a failed server start, or a bunch of log lines. What is the content of the first CMD window?@Nantris commented on GitHub (Nov 9, 2025):
Thanks again for your reply and I apologize as I misread your message.
So yes that logs and that also resolves the problem. It seems in the past it was never necessary to run
ollama serveand usingollama runwould open the app in the system tray automatically. But I don't see anything in the release notes for0.12.4that suggests that that's expected.@rick-github commented on GitHub (Nov 9, 2025):
The purpose of starting the server from the CMD window was to increase the visibility of the logs to determine the cause of
Server Error: llama runner process has terminated: exit status 2. If this error is no longer occurring, it seems it was a transient issue. If it re-occurs, update this issue (or create a new issue) with the server log.@TigerGod commented on GitHub (Nov 9, 2025):
希望别再更新了,用上一个老版本就挺好的,更新后影响太大了。。。
@Nantris commented on GitHub (Nov 9, 2025):
@rick-github I feel you're overlooking the change in behavior, and perhaps I was not clear enough about it.
ollama run [model]"just works" in0.12.3. In0.12.4and beyond it produces the error unless you runollama servefirst. That seems to be the cause for this issue existing at all. It's definitely not transient and I didn't upload the log because the issue IS thatollama servenow needs to be run first, but when it is, it runs fine.If this is intended, it should be documented.
@rick-github commented on GitHub (Nov 9, 2025):
The behaviour hasn't changed. There was a bug in the 0.12.4 to 0.12.9 range that caused model loading to stall, perhaps that's what you experienced. If
ollama run(orollama listif you want to avoid a load stall) is run in a terminal window when the server is not running, the server will be started.If, in 0.12.4 and beyond, you do not manually start the ollama server by running
ollama servein a command window, and you runollama run granite4:7b-a1b-hand get anError: llama runner process has terminated: exit status 2message, then there must be a server running, either started as part of theStartup appsor by the autostart triggered by theruncommand. In that case, the server log will contain details about the runner crash.@Nantris commented on GitHub (Nov 10, 2025):
I'm in
0.12.10now and whether it was intentional or not, I can assure you the behavior changed on Windows in0.12.4.From your instructions, it sounds like the old behavior was unexpected but I don't know for sure. Was
ollama runever supposed to work without separately starting the server first? Because it did.I have run
ollama run [model]hundreds of times and it just works as stated, but as of0.12.4it no longer works and it errors as stated. It immediately starts working if you runollama servefirst and separately as you advised. If you do not, what happens instead is thatollama app.exeas well as twoollama.exeinstances and you get theError: llama runner process has terminated: exit status 2If you use the GUI which it starts when you run
ollama run [model], there it errors:500 Internal Server Error: llama runner process has terminated: exit status 2. - This also happens if you run it from the start menu. (The first time I installed0.12.10the GUI app was not starting which may make some of my earlier reports confusing to reconcile)As far as I can tell, the GUI app no longer ever works, but the CLI interface works fine if you run
ollama serve. That workaround doesn't work for the GUI app because it seems to end anyollama servethat's running, and trying to run it after the GUI app, sensibly, yields:Error: listen tcp 127.0.0.1:11434: bind: Only one usage of each socket address (protocol/network address/port) is normally permitted.Please let me know if there's any information you'd like me to try, investigate, or share to try to give you a more complete insight into this. I have it working for my development now and don't use the GUI, so it's fine with me this new way it works, but it is definitely new.
I apologize for the lengthy report but wanted to be complete.
@rick-github commented on GitHub (Nov 10, 2025):
The code that does server start if missing was committed early 2024 and wasn't modified during 0.12.*. AFAIK there have been no other reports of this failing and I am unable to reproduce this, so it seems like it's specific to your installation. In which case I propose moving the discussion of this to a new issue.
The reason for the termination of the runner process will be in the server log. If there is no information in the log it could be another problem with your installation. What's the output of
dir %LOCALAPPDATA%\Ollama?@Apil120 commented on GitHub (Nov 11, 2025):
Have you tried doing following the steps mentioned by @rick-github in this issue? I used taskkill to kill the ollama process and ran ollama list, which seems to have fixed the issue for me.
@YonTracks commented on GitHub (Nov 11, 2025):
did you uninstall the old before updating the new (old ollama was more forgiving with this), updating seems an issue on windows (I think app id or something) sometimes 2 ollama's, especially if the version name was modified like me, example
version 0.12.10-yontracks. good luck. I'm a bad communicator but check that. cheers. good luck.@YonTracks commented on GitHub (Nov 11, 2025):
I modified the iss script to do it (check for previous installs, remove and update accordingly). cheers. good luck.
heres the actual, app error when using
ollama runwithout running it first via the app.ollama serveworks fine (console load, with console streaming logs etc). and via the app it works fine, but UI loads.clarity: so, with no ollama processes running at all.
I run
ollama run llama3.1no
server.logas the server exits and rather silently?, tricky error messages anyway etc.app.log