mirror of
https://github.com/ollama/ollama.git
synced 2026-05-06 16:11:34 -05:00
[GH-ISSUE #12618] Ollama serve fails to detect Nvidia GPUs after updating to the latest version #8376
Closed
opened 2026-04-12 21:00:18 -05:00 by GiteaMirror
·
52 comments
No Branch/Tag Specified
main
dhiltgen/ci
parth-launch-plan-gating
hoyyeva/anthropic-reference-images-path
parth-anthropic-reference-images-path
brucemacd/download-before-remove
hoyyeva/editor-config-repair
parth-mlx-decode-checkpoints
parth-launch-codex-app
hoyyeva/fix-codex-model-metadata-warning
hoyyeva/qwen
parth/hide-claude-desktop-till-release
hoyyeva/opencode-image-modality
parth-add-claude-code-autoinstall
release_v0.22.0
pdevine/manifest-list
codex/fix-codex-model-metadata-warning
pdevine/addressable-manifest
brucemacd/launch-fetch-reccomended
jmorganca/llama-compat
launch-copilot-cli
hoyyeva/opencode-thinking
release_v0.20.7
parth-auto-save-backup
parth-test
jmorganca/gemma4-audio-replacements
fix-manifest-digest-on-pull
hoyyeva/vscode-improve
brucemacd/install-server-wait
parth/update-claude-docs
brucemac/start-ap-install
pdevine/mlx-update
pdevine/qwen35_vision
drifkin/api-show-fallback
mintlify/image-generation-1773352582
hoyyeva/server-context-length-local-config
jmorganca/faster-reptition-penalties
jmorganca/convert-nemotron
parth-pi-thinking
pdevine/sampling-penalties
jmorganca/fix-create-quantization-memory
dongchen/resumable_transfer_fix
pdevine/sampling-cache-error
jessegross/mlx-usage
hoyyeva/openclaw-config
hoyyeva/app-html
pdevine/qwen3next
brucemacd/sign-sh-install
brucemacd/tui-update
brucemacd/usage-api
jmorganca/launch-empty
fix-app-dist-embed
mxyng/mlx-compile
mxyng/mlx-quant
mxyng/mlx-glm4.7
mxyng/mlx
brucemacd/simplify-model-picker
jmorganca/qwen3-concurrent
fix-glm-4.7-flash-mla-config
drifkin/qwen3-coder-opening-tag
brucemacd/usage-cli
fix-cuda12-fattn-shmem
ollama-imagegen-docs
parth/fix-multiline-inputs
brucemacd/config-docs
mxyng/model-files
mxyng/simple-execute
fix-imagegen-ollama-models
mxyng/async-upload
jmorganca/lazy-no-dtype-changes
imagegen-auto-detect-create
parth/decrease-concurrent-download-hf
fix-mlx-quantize-init
jmorganca/x-cleanup
usage
imagegen-readme
jmorganca/glm-image
mlx-gpu-cd
jmorganca/imagegen-modelfile
parth/agent-skills
parth/agent-allowlist
parth/signed-in-offline
parth/agents
parth/fix-context-chopping
improve-cloud-flow
parth/add-models-websearch
parth/prompt-renderer-mcp
jmorganca/native-settings
jmorganca/download-stream-hash
jmorganca/client2-rebased
brucemacd/oai-chat-req-multipart
jessegross/multi_chunk_reserve
grace/additional-omit-empty
grace/mistral-3-large
mxyng/tokenizer2
mxyng/tokenizer
jessegross/flash
hoyyeva/windows-nacked-app
mxyng/cleanup-attention
grace/deepseek-parser
hoyyeva/remember-unsent-prompt
parth/add-lfs-pointer-error-conversion
parth/olmo2-test2
hoyyeva/ollama-launchagent-plist
nicole/olmo-model
parth/olmo-test
mxyng/remove-embedded
parth/render-template
jmorganca/intellect-3
parth/remove-prealloc-linter
jmorganca/cmd-eval
nicole/nomic-embed-text-fix
mxyng/lint-2
hoyyeva/add-gemini-3-pro-preview
hoyyeva/load-model-list
mxyng/expand-path
mxyng/environ-2
hoyyeva/deeplink-json-encoding
parth/improve-tool-calling-tests
hoyyeva/conversation
hoyyeva/assistant-edit-response
hoyyeva/thinking
origin/brucemacd/invalid-char-i-err
parth/improve-tool-calling
jmorganca/required-omitempty
grace/qwen3-vl-tests
mxyng/iter-client
parth/docs-readme
nicole/embed-test
pdevine/integration-benchstat
parth/remove-generate-cmd
parth/add-toolcall-id
mxyng/server-tests
jmorganca/glm-4.6
jmorganca/gin-h-compat
drifkin/stable-tool-args
pdevine/qwen3-more-thinking
parth/add-websearch-client
nicole/websearch_local
jmorganca/qwen3-coder-updates
grace/deepseek-v3-migration-tests
mxyng/fix-create
jmorganca/cloud-errors
pdevine/parser-tidy
revert-12233-parth/simplify-entrypoints-runner
parth/enable-so-gpt-oss
brucemacd/qwen3vl
jmorganca/readme-simplify
parth/gpt-oss-structured-outputs
revert-12039-jmorganca/tools-braces
mxyng/embeddings
mxyng/gguf
mxyng/benchmark
mxyng/types-null
parth/move-parsing
mxyng/gemma2
jmorganca/docs
mxyng/16-bit
mxyng/create-stdin
pdevine/authorizedkeys
mxyng/quant
parth/opt-in-error-context-window
brucemacd/cache-models
brucemacd/runner-completion
jmorganca/llama-update-6
brucemacd/benchmark-list
brucemacd/partial-read-caps
parth/deepseek-r1-tools
mxyng/omit-array
parth/tool-prefix-temp
brucemacd/runner-test
jmorganca/qwen25vl
brucemacd/model-forward-test-ext
parth/python-function-parsing
jmorganca/cuda-compression-none
drifkin/num-parallel
drifkin/chat-truncation-fix
jmorganca/sync
parth/python-tools-calling
drifkin/array-head-count
brucemacd/create-no-loop
parth/server-enable-content-stream-with-tools
qwen25omni
mxyng/v3
brucemacd/ropeconfig
jmorganca/silence-tokenizer
parth/sample-so-test
parth/sampling-structured-outputs
brucemacd/doc-go-engine
parth/constrained-sampling-json
jmorganca/mistral-wip
brucemacd/mistral-small-convert
parth/sample-unmarshal-json-for-params
brucemacd/jomorganca/mistral
pdevine/bfloat16
jmorganca/mistral
brucemacd/mistral
pdevine/logging
parth/sample-correctness-fix
parth/sample-fix-sorting
jmorgan/sample-fix-sorting-extras
jmorganca/temp-0-images
brucemacd/parallel-embed-models
brucemacd/shim-grammar
jmorganca/fix-gguf-error
bmizerany/nameswork
jmorganca/faster-releases
bmizerany/validatenames
brucemacd/err-no-vocab
brucemacd/rope-config
brucemacd/err-hint
brucemacd/qwen2_5
brucemacd/logprobs
brucemacd/new_runner_graph_bench
progress-flicker
brucemacd/forward-test
brucemacd/go_qwen2
pdevine/gemma2
jmorganca/add-missing-symlink-eval
mxyng/next-debug
parth/set-context-size-openai
brucemacd/next-bpe-bench
brucemacd/next-bpe-test
brucemacd/new_runner_e2e
brucemacd/new_runner_qwen2
pdevine/convert-cohere2
brucemacd/convert-cli
parth/log-probs
mxyng/next-mlx
mxyng/cmd-history
parth/templating
parth/tokenize-detokenize
brucemacd/check-key-register
bmizerany/grammar
jmorganca/vendor-081b29bd
mxyng/func-checks
jmorganca/fix-null-format
parth/fix-default-to-warn-json
jmorganca/qwen2vl
jmorganca/no-concat
parth/cmd-cleanup-SO
brucemacd/check-key-register-structured-err
parth/openai-stream-usage
parth/fix-referencing-so
stream-tools-stop
jmorganca/degin-1
brucemacd/install-path-clean
brucemacd/push-name-validation
brucemacd/browser-key-register
jmorganca/openai-fix-first-message
jmorganca/fix-proxy
jessegross/sample
parth/disallow-streaming-tools
dhiltgen/remove_submodule
jmorganca/ga
jmorganca/mllama
pdevine/newlines
pdevine/geems-2b
jmorganca/llama-bump
mxyng/modelname-7
mxyng/gin-slog
mxyng/modelname-6
jyan/convert-prog
jyan/quant5
paligemma-support
pdevine/import-docs
jmorganca/openai-context
jyan/paligemma
jyan/p2
jyan/palitest
bmizerany/embedspeedup
jmorganca/llama-vit
brucemacd/allow-ollama
royh/ep-methods
royh/whisper
mxyng/api-models
mxyng/fix-memory
jyan/q4_4/8
jyan/ollama-v
royh/stream-tools
roy-embed-parallel
bmizerany/hrm
revert-5963-revert-5924-mxyng/llama3.1-rope
royh/embed-viz
jyan/local2
jyan/auth
jyan/local
jyan/parse-temp
jmorganca/template-mistral
jyan/reord-g
royh-openai-suffixdocs
royh-imgembed
royh-embed-parallel
jyan/quant4
royh-precision
jyan/progress
pdevine/fix-template
jyan/quant3
pdevine/ggla
mxyng/update-registry-domain
jmorganca/ggml-static
mxyng/create-context
jyan/v0.146
mxyng/layers-from-files
build_dist
bmizerany/noseek
royh-ls
royh-name
timeout
mxyng/server-timestamp
bmizerany/nosillyggufslurps
royh-params
jmorganca/llama-cpp-7c26775
royh-openai-delete
royh-show-rigid
jmorganca/enable-fa
jmorganca/no-error-template
jyan/format
royh-testdelete
bmizerany/fastverify
language_support
pdevine/ps-glitches
brucemacd/tokenize
bruce/iq-quants
bmizerany/filepathwithcoloninhost
mxyng/split-bin
bmizerany/client-registry
jmorganca/if-none-match
native
jmorganca/native
jmorganca/batch-embeddings
jmorganca/initcmake
jmorganca/mm
pdevine/showggmlinfo
modenameenforcealphanum
bmizerany/modenameenforcealphanum
jmorganca/done-reason
jmorganca/llama-cpp-8960fe8
ollama.com
bmizerany/filepathnobuild
bmizerany/types/model/defaultfix
rmdisplaylong
nogogen
bmizerany/x
modelfile-readme
bmizerany/replacecolon
jmorganca/limit
jmorganca/execstack
jmorganca/replace-assets
mxyng/tune-concurrency
jmorganca/testing
whitespace-detection
jmorganca/options
upgrade-all
scratch
cuda-search
mattw/airenamer
mattw/allmodelsonhuggingface
mattw/quantcontext
mattw/whatneedstorun
brucemacd/llama-mem-calc
mattw/faq-context
mattw/communitylinks
mattw/noprune
mattw/python-functioncalling
rename
mxyng/install
pulse
remove-first
editor
mattw/selfqueryingretrieval
cgo
mattw/howtoquant
api
matt/streamingapi
format-config
mxyng/extra-args
shell
update-nous-hermes
cp-model
upload-progress
fix-unknown-model
fix-model-names
delete-fix
insecure-registry
ls
deletemodels
progressbar
readme-updates
license-layers
skip-list
list-models
modelpath
matt/examplemodelfiles
distribution
go-opts
v0.23.1
v0.23.1-rc0
v0.23.0
v0.23.0-rc0
v0.22.1
v0.22.1-rc1
v0.22.1-rc0
v0.22.0
v0.22.0-rc1
v0.21.3-rc0
v0.21.2-rc1
v0.21.2
v0.21.2-rc0
v0.21.1
v0.21.1-rc1
v0.21.1-rc0
v0.21.0
v0.21.0-rc1
v0.21.0-rc0
v0.20.8-rc0
v0.20.7
v0.20.7-rc1
v0.20.7-rc0
v0.20.6
v0.20.6-rc1
v0.20.6-rc0
v0.20.5
v0.20.5-rc2
v0.20.5-rc1
v0.20.5-rc0
v0.20.4
v0.20.4-rc2
v0.20.4-rc1
v0.20.4-rc0
v0.20.3
v0.20.3-rc0
v0.20.2
v0.20.1
v0.20.1-rc2
v0.20.1-rc1
v0.20.1-rc0
v0.20.0
v0.20.0-rc1
v0.20.0-rc0
v0.19.0
v0.19.0-rc2
v0.19.0-rc1
v0.19.0-rc0
v0.18.4-rc1
v0.18.4-rc0
v0.18.3
v0.18.3-rc2
v0.18.3-rc1
v0.18.3-rc0
v0.18.2
v0.18.2-rc1
v0.18.2-rc0
v0.18.1
v0.18.1-rc1
v0.18.1-rc0
v0.18.0
v0.18.0-rc2
v0.18.0-rc1
v0.18.0-rc0
v0.17.8-rc4
v0.17.8-rc3
v0.17.8-rc2
v0.17.8-rc1
v0.17.8-rc0
v0.17.7
v0.17.7-rc2
v0.17.7-rc1
v0.17.7-rc0
v0.17.6
v0.17.5
v0.17.4
v0.17.3
v0.17.2
v0.17.1
v0.17.1-rc2
v0.17.1-rc1
v0.17.1-rc0
v0.17.0
v0.17.0-rc2
v0.17.0-rc1
v0.17.0-rc0
v0.16.3
v0.16.3-rc2
v0.16.3-rc1
v0.16.3-rc0
v0.16.2
v0.16.2-rc0
v0.16.1
v0.16.0
v0.16.0-rc2
v0.16.0-rc0
v0.16.0-rc1
v0.15.6
v0.15.5
v0.15.5-rc5
v0.15.5-rc4
v0.15.5-rc3
v0.15.5-rc2
v0.15.5-rc1
v0.15.5-rc0
v0.15.4
v0.15.3
v0.15.2
v0.15.1
v0.15.1-rc1
v0.15.1-rc0
v0.15.0-rc6
v0.15.0
v0.15.0-rc5
v0.15.0-rc4
v0.15.0-rc3
v0.15.0-rc2
v0.15.0-rc1
v0.15.0-rc0
v0.14.3
v0.14.3-rc3
v0.14.3-rc2
v0.14.3-rc1
v0.14.3-rc0
v0.14.2
v0.14.2-rc1
v0.14.2-rc0
v0.14.1
v0.14.0-rc11
v0.14.0
v0.14.0-rc10
v0.14.0-rc9
v0.14.0-rc8
v0.14.0-rc7
v0.14.0-rc6
v0.14.0-rc5
v0.14.0-rc4
v0.14.0-rc3
v0.14.0-rc2
v0.14.0-rc1
v0.14.0-rc0
v0.13.5
v0.13.5-rc1
v0.13.5-rc0
v0.13.4-rc2
v0.13.4
v0.13.4-rc1
v0.13.4-rc0
v0.13.3
v0.13.3-rc1
v0.13.3-rc0
v0.13.2
v0.13.2-rc2
v0.13.2-rc1
v0.13.2-rc0
v0.13.1
v0.13.1-rc2
v0.13.1-rc1
v0.13.1-rc0
v0.13.0
v0.13.0-rc0
v0.12.11
v0.12.11-rc1
v0.12.11-rc0
v0.12.10
v0.12.10-rc1
v0.12.10-rc0
v0.12.9-rc0
v0.12.9
v0.12.8
v0.12.8-rc0
v0.12.7
v0.12.7-rc1
v0.12.7-rc0
v0.12.7-citest0
v0.12.6
v0.12.6-rc1
v0.12.6-rc0
v0.12.5
v0.12.5-rc0
v0.12.4
v0.12.4-rc7
v0.12.4-rc6
v0.12.4-rc5
v0.12.4-rc4
v0.12.4-rc3
v0.12.4-rc2
v0.12.4-rc1
v0.12.4-rc0
v0.12.3
v0.12.2
v0.12.2-rc0
v0.12.1
v0.12.1-rc1
v0.12.1-rc2
v0.12.1-rc0
v0.12.0
v0.12.0-rc1
v0.12.0-rc0
v0.11.11
v0.11.11-rc3
v0.11.11-rc2
v0.11.11-rc1
v0.11.11-rc0
v0.11.10
v0.11.9
v0.11.9-rc0
v0.11.8
v0.11.8-rc0
v0.11.7-rc1
v0.11.7-rc0
v0.11.7
v0.11.6
v0.11.6-rc0
v0.11.5-rc4
v0.11.5-rc3
v0.11.5
v0.11.5-rc5
v0.11.5-rc2
v0.11.5-rc1
v0.11.5-rc0
v0.11.4
v0.11.4-rc0
v0.11.3
v0.11.3-rc0
v0.11.2
v0.11.1
v0.11.0-rc0
v0.11.0-rc1
v0.11.0-rc2
v0.11.0
v0.10.2-int1
v0.10.1
v0.10.0
v0.10.0-rc4
v0.10.0-rc3
v0.10.0-rc2
v0.10.0-rc1
v0.10.0-rc0
v0.9.7-rc1
v0.9.7-rc0
v0.9.6
v0.9.6-rc0
v0.9.6-ci0
v0.9.5
v0.9.4-rc5
v0.9.4-rc6
v0.9.4
v0.9.4-rc3
v0.9.4-rc4
v0.9.4-rc1
v0.9.4-rc2
v0.9.4-rc0
v0.9.3
v0.9.3-rc5
v0.9.4-citest0
v0.9.3-rc4
v0.9.3-rc3
v0.9.3-rc2
v0.9.3-rc1
v0.9.3-rc0
v0.9.2
v0.9.1
v0.9.1-rc1
v0.9.1-rc0
v0.9.1-ci1
v0.9.1-ci0
v0.9.0
v0.9.0-rc0
v0.8.0
v0.8.0-rc0
v0.7.1-rc2
v0.7.1
v0.7.1-rc1
v0.7.1-rc0
v0.7.0
v0.7.0-rc1
v0.7.0-rc0
v0.6.9-rc0
v0.6.8
v0.6.8-rc0
v0.6.7
v0.6.7-rc2
v0.6.7-rc1
v0.6.7-rc0
v0.6.6
v0.6.6-rc2
v0.6.6-rc1
v0.6.6-rc0
v0.6.5-rc1
v0.6.5
v0.6.5-rc0
v0.6.4-rc0
v0.6.4
v0.6.3-rc1
v0.6.3
v0.6.3-rc0
v0.6.2
v0.6.2-rc0
v0.6.1
v0.6.1-rc0
v0.6.0-rc0
v0.6.0
v0.5.14-rc0
v0.5.13
v0.5.13-rc6
v0.5.13-rc5
v0.5.13-rc4
v0.5.13-rc3
v0.5.13-rc2
v0.5.13-rc1
v0.5.13-rc0
v0.5.12
v0.5.12-rc1
v0.5.12-rc0
v0.5.11
v0.5.10
v0.5.9
v0.5.9-rc0
v0.5.8-rc13
v0.5.8
v0.5.8-rc12
v0.5.8-rc11
v0.5.8-rc10
v0.5.8-rc9
v0.5.8-rc8
v0.5.8-rc7
v0.5.8-rc6
v0.5.8-rc5
v0.5.8-rc4
v0.5.8-rc3
v0.5.8-rc2
v0.5.8-rc1
v0.5.8-rc0
v0.5.7
v0.5.6
v0.5.5
v0.5.5-rc0
v0.5.4
v0.5.3
v0.5.3-rc0
v0.5.2
v0.5.2-rc3
v0.5.2-rc2
v0.5.2-rc1
v0.5.2-rc0
v0.5.1
v0.5.0
v0.5.0-rc1
v0.4.8-rc0
v0.4.7
v0.4.6
v0.4.5
v0.4.4
v0.4.3
v0.4.3-rc0
v0.4.2
v0.4.2-rc1
v0.4.2-rc0
v0.4.1
v0.4.1-rc0
v0.4.0
v0.4.0-rc8
v0.4.0-rc7
v0.4.0-rc6
v0.4.0-rc5
v0.4.0-rc4
v0.4.0-rc3
v0.4.0-rc2
v0.4.0-rc1
v0.4.0-rc0
v0.4.0-ci3
v0.3.14
v0.3.14-rc0
v0.3.13
v0.3.12
v0.3.12-rc5
v0.3.12-rc4
v0.3.12-rc3
v0.3.12-rc2
v0.3.12-rc1
v0.3.11
v0.3.11-rc4
v0.3.11-rc3
v0.3.11-rc2
v0.3.11-rc1
v0.3.10
v0.3.10-rc1
v0.3.9
v0.3.8
v0.3.7
v0.3.7-rc6
v0.3.7-rc5
v0.3.7-rc4
v0.3.7-rc3
v0.3.7-rc2
v0.3.7-rc1
v0.3.6
v0.3.5
v0.3.4
v0.3.3
v0.3.2
v0.3.1
v0.3.0
v0.2.8
v0.2.8-rc2
v0.2.8-rc1
v0.2.7
v0.2.6
v0.2.5
v0.2.4
v0.2.3
v0.2.2
v0.2.2-rc2
v0.2.2-rc1
v0.2.1
v0.2.0
v0.1.49-rc14
v0.1.49-rc13
v0.1.49-rc12
v0.1.49-rc11
v0.1.49-rc10
v0.1.49-rc9
v0.1.49-rc8
v0.1.49-rc7
v0.1.49-rc6
v0.1.49-rc4
v0.1.49-rc5
v0.1.49-rc3
v0.1.49-rc2
v0.1.49-rc1
v0.1.48
v0.1.47
v0.1.46
v0.1.45-rc5
v0.1.45
v0.1.45-rc4
v0.1.45-rc3
v0.1.45-rc2
v0.1.45-rc1
v0.1.44
v0.1.43
v0.1.42
v0.1.41
v0.1.40
v0.1.40-rc1
v0.1.39
v0.1.39-rc2
v0.1.39-rc1
v0.1.38
v0.1.37
v0.1.36
v0.1.35
v0.1.35-rc1
v0.1.34
v0.1.34-rc1
v0.1.33
v0.1.33-rc7
v0.1.33-rc6
v0.1.33-rc5
v0.1.33-rc4
v0.1.33-rc3
v0.1.33-rc2
v0.1.33-rc1
v0.1.32
v0.1.32-rc2
v0.1.32-rc1
v0.1.31
v0.1.30
v0.1.29
v0.1.28
v0.1.27
v0.1.26
v0.1.25
v0.1.24
v0.1.23
v0.1.22
v0.1.21
v0.1.20
v0.1.19
v0.1.18
v0.1.17
v0.1.16
v0.1.15
v0.1.14
v0.1.13
v0.1.12
v0.1.11
v0.1.10
v0.1.9
v0.1.8
v0.1.7
v0.1.6
v0.1.5
v0.1.4
v0.1.3
v0.1.2
v0.1.1
v0.1.0
v0.0.21
v0.0.20
v0.0.19
v0.0.18
v0.0.17
v0.0.16
v0.0.15
v0.0.14
v0.0.13
v0.0.12
v0.0.11
v0.0.10
v0.0.9
v0.0.8
v0.0.7
v0.0.6
v0.0.5
v0.0.4
v0.0.3
v0.0.2
v0.0.1
Labels
Clear labels
amd
api
app
bug
build
cli
cloud
compatibility
context-length
create
docker
documentation
embeddings
feature request
feedback wanted
good first issue
gpt-oss
gpu
harmony
help wanted
image
install
intel
js
launch
linux
macos
memory
mlx
model
needs more info
networking
nvidia
ollama.com
performance
pull-request
python
question
registry
rendering
thinking
tools
top
vulkan
windows
wsl
Mirrored from GitHub Pull Request
Milestone
No items
No Milestone
Projects
Clear projects
No project
No Assignees
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: github-starred/ollama#8376
Reference in New Issue
Block a user
Blocking a user prevents them from interacting with repositories, such as opening or commenting on pull requests or issues. Learn more about blocking a user.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @catclaw on GitHub (Oct 14, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/12618
Originally assigned to: @dhiltgen on GitHub.
What is the issue?
Hi!
I'm experiencing a strange error after updating Ollama to the latest version today.
I'm running Ollama on a Windows 11 Enterprise 25H2 machine - no Docker.
When I start Ollama using "ollama serve", it fails to detect my GPUs and falls back to CPU.
time=2025-10-14T21:18:01.270+02:00 level=INFO source=routes.go:1481 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-14T21:18:02.309+02:00 level=INFO source=images.go:522 msg="total blobs: 145" time=2025-10-14T21:18:02.326+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-14T21:18:02.340+02:00 level=INFO source=routes.go:1534 msg="Listening on 127.0.0.1:11434 (version 0.12.5)" time=2025-10-14T21:18:02.348+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-14T21:18:03.030+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="101.0 GiB" time=2025-10-14T21:18:03.030+02:00 level=INFO source=routes.go:1575 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"When I check the older logs, I can see that it detected the GPUs before the update:
time=2025-09-17T13:55:55.012+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:INFO OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://0.0.0.0:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-09-17T13:55:55.303+02:00 level=INFO source=images.go:477 msg="total blobs: 100" time=2025-09-17T13:55:55.314+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-09-17T13:55:55.324+02:00 level=INFO source=routes.go:1385 msg="Listening on [::]:11434 (version 0.11.11)" time=2025-09-17T13:55:55.324+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:55.325+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-09-17T13:55:56.007+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="156.4 MiB" time=2025-09-17T13:55:56.212+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="79.7 MiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-09-17T13:55:56.222+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB"My Nvidia driver are the latest and I have CUDA 13.0 installed.
I've tried to uninstall and reinstall Ollama, my Nvidia driver, CUDA, etc. - but nothing seems to work.
I also added CUDA_VISIBLE_DEVICES:0,1,2 to my system variables - but Ollama still refuses to detect my GPUs.
I've googled and searched, and tried every tutorial, but I'm stuck with Ollama using CPU instead of my GPUs.
I'd appreciate any help I can get.
Sorry for bothering and thank you in advance.
Relevant log output
OS
Windows
GPU
Nvidia
CPU
Intel
Ollama version
0.12.5
@dhiltgen commented on GitHub (Oct 14, 2025):
@catclaw could you quit Ollama in the system tray, and then run the following in a powershell terminal and share the logs?
Once it reports "inference compute" you can
^Cand send the log.@catclaw commented on GitHub (Oct 14, 2025):
Yes, of course! Thank you so much for taking your time to look into this. 🙏
serve.log
@dhiltgen commented on GitHub (Oct 14, 2025):
My suspicion is it's from
CUDA_VISIBLE_DEVICES=0,1,2. Try unsetting that first and see if it discovers the GPUs, or use the UUIDs instead which you can gather fromnvidia-smi -LIf that doesn't solve it, can you share the output of
nvidia-smi?@catclaw commented on GitHub (Oct 15, 2025):
I added CUDA_VISIBLE_DEVICES=0,1,2 to System Variables, hoping Ollama would detect the GPUs after updating from 0.11.11 to 0.12.5, but it didn't work - and neither does using GPU UUIDs...
Meanwhile, I don't have any problems creating three ComfyUI instances in SwarmUI, for example, to run 3 generations simultaneously.
I've been up all night trying to figure out the problem, with no luck... 😵💫
After adding GPU UUID:
@dhiltgen commented on GitHub (Oct 15, 2025):
Could you share the server log with OLLAMA_DEBUG="2" without CUDA_VISIBLE_DEVICES set? I'm hoping that may have a little more information on what's going wrong.
@catclaw commented on GitHub (Oct 15, 2025):
Yes, of course! Thank you for taking the time to look into this.
I've uninstalled and reinstalled Ollama (without deleting all my models) - but it still falls back to using CPU instead of GPU.
(I've removed CUDA_VISIBLE_DEVICES and set OLLAMA_DEBUG=2)
Here is the output:
@catclaw commented on GitHub (Oct 16, 2025):
This seems to be a bug in the new 0.12.5 version.
I uninstalled Ollama 0.12.5 and installed version 0.11.11 - and it's detecting my GPUs just fine?
Debug is set to 1:
time=2025-10-16T03:11:23.895+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES:]" time=2025-10-16T03:11:23.933+02:00 level=INFO source=images.go:477 msg="total blobs: 145" time=2025-10-16T03:11:23.951+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-10-16T03:11:23.965+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" time=2025-10-16T03:11:23.966+02:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler" time=2025-10-16T03:11:23.966+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-10-16T03:11:23.967+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-10-16T03:11:23.968+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-10-16T03:11:23.968+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-10-16T03:11:23.969+02:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-10-16T03:11:23.971+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvml.dll time=2025-10-16T03:11:23.972+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvml.dll C:\\tools\\ruby34\\bin\\nvml.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvml.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvml.dll C:\\Python312\\Scripts\\nvml.dll C:\\Python312\\nvml.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvml.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvml.dll C:\\Windows\\system32\\nvml.dll C:\\Windows\\nvml.dll C:\\Windows\\System32\\Wbem\\nvml.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\Windows\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\WinMerge\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvml.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvml.dll C:\\ProgramData\\chocolatey\\bin\\nvml.dll H:\\Fooocus\\python_embeded\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvml.dll C:\\Program Files\\dotnet\\nvml.dll C:\\Program Files\\PuTTY\\nvml.dll C:\\Program Files\\Process Lasso\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvml.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvml.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvml.dll C:\\Program Files\\gsudo\\Current\\nvml.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvml.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll C:\\WINDOWS\\nvml.dll C:\\WINDOWS\\System32\\Wbem\\nvml.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\WINDOWS\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\CrowdSec\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvml.dll C:\\Program Files\\GitHub CLI\\nvml.dll C:\\Program Files\\Go\\bin\\nvml.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvml.dll C:\\Program Files\\nodejs\\nvml.dll C:\\Program Files\\CMake\\bin\\nvml.dll C:\\Program Files\\Calibre2\\nvml.dll C:\\Program Files\\PowerShell\\7\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvml.dll C:\\Program Files\\Git\\cmd\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T03:11:23.985+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll" time=2025-10-16T03:11:23.989+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T03:11:24.043+02:00 level=DEBUG source=gpu.go:111 msg="nvidia-ml loaded" library=C:\Windows\system32\nvml.dll time=2025-10-16T03:11:24.045+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvcuda.dll time=2025-10-16T03:11:24.048+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvcuda.dll C:\\tools\\ruby34\\bin\\nvcuda.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvcuda.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvcuda.dll C:\\Python312\\Scripts\\nvcuda.dll C:\\Python312\\nvcuda.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvcuda.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvcuda.dll C:\\Windows\\system32\\nvcuda.dll C:\\Windows\\nvcuda.dll C:\\Windows\\System32\\Wbem\\nvcuda.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\Windows\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\WinMerge\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvcuda.dll C:\\ProgramData\\chocolatey\\bin\\nvcuda.dll H:\\Fooocus\\python_embeded\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvcuda.dll C:\\Program Files\\dotnet\\nvcuda.dll C:\\Program Files\\PuTTY\\nvcuda.dll C:\\Program Files\\Process Lasso\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvcuda.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvcuda.dll C:\\Program Files\\gsudo\\Current\\nvcuda.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvcuda.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll C:\\WINDOWS\\nvcuda.dll C:\\WINDOWS\\System32\\Wbem\\nvcuda.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\WINDOWS\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\CrowdSec\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvcuda.dll C:\\Program Files\\GitHub CLI\\nvcuda.dll C:\\Program Files\\Go\\bin\\nvcuda.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvcuda.dll C:\\Program Files\\nodejs\\nvcuda.dll C:\\Program Files\\CMake\\bin\\nvcuda.dll C:\\Program Files\\Calibre2\\nvcuda.dll C:\\Program Files\\PowerShell\\7\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvcuda.dll C:\\Program Files\\Git\\cmd\\nvcuda.dll c:\\windows\\system*\\nvcuda.dll]" time=2025-10-16T03:11:24.064+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll" time=2025-10-16T03:11:24.069+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll]" initializing C:\Windows\system32\nvcuda.dll dlsym: cuInit - 00007FFDC058DE10 dlsym: cuDriverGetVersion - 00007FFDC058DEB0 dlsym: cuDeviceGetCount - 00007FFDC058E6A6 dlsym: cuDeviceGet - 00007FFDC058E6A0 dlsym: cuDeviceGetAttribute - 00007FFDC058E000 dlsym: cuDeviceGetUuid - 00007FFDC058E6B2 dlsym: cuDeviceGetName - 00007FFDC058E6AC dlsym: cuCtxCreate_v3 - 00007FFDC058E72A dlsym: cuMemGetInfo_v2 - 00007FFDC058E82C dlsym: cuCtxDestroy - 00007FFDC058E73C calling cuInit calling cuDriverGetVersion raw version 0x32c8 CUDA driver version: 13.0 calling cuDeviceGetCount device count 3 time=2025-10-16T03:11:24.133+02:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=3 library=C:\Windows\system32\nvcuda.dll [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA totalMem 12226mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA freeMem 10929mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] Compute Capability 12.0 [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA totalMem 12287mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA freeMem 11245mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] Compute Capability 8.6 time=2025-10-16T03:11:24.495+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="343.9 MiB" [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA totalMem 12287mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA freeMem 11245mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] Compute Capability 8.6 time=2025-10-16T03:11:24.669+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="274.2 MiB" time=2025-10-16T03:11:24.677+02:00 level=DEBUG source=amd_windows.go:34 msg="unable to load amdhip64_6.dll, please make sure to upgrade to the latest amd driver: The specified module could not be found." releasing cuda driver library releasing nvml library time=2025-10-16T03:11:24.684+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-10-16T03:11:24.686+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-10-16T03:11:24.691+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB"@catclaw commented on GitHub (Oct 16, 2025):
I can't figure out what the problem is?
I've been using 70B models in OpenWebUI using GPUs only (OLLAMA__SCHED_SPREAD:1) - but I'm back to square one!
Ollama keeps falling back to CPU no matter what model I try to use? 😵💫
time=2025-10-16T04:14:43.126+02:00 level=INFO source=routes.go:1332 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-16T04:14:43.167+02:00 level=INFO source=images.go:477 msg="total blobs: 145" time=2025-10-16T04:14:43.186+02:00 level=INFO source=images.go:484 msg="total unused blobs removed: 0" time=2025-10-16T04:14:43.202+02:00 level=INFO source=routes.go:1385 msg="Listening on 127.0.0.1:11434 (version 0.11.11)" time=2025-10-16T04:14:43.202+02:00 level=DEBUG source=sched.go:121 msg="starting llm scheduler" time=2025-10-16T04:14:43.202+02:00 level=INFO source=gpu.go:217 msg="looking for compatible GPUs" time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:167 msg=packages count=2 time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:214 msg="" package=0 cores=14 efficiency=0 threads=28 time=2025-10-16T04:14:43.203+02:00 level=INFO source=gpu_windows.go:214 msg="" package=1 cores=14 efficiency=0 threads=28 time=2025-10-16T04:14:43.203+02:00 level=DEBUG source=gpu.go:98 msg="searching for GPU discovery libraries for NVIDIA" time=2025-10-16T04:14:43.204+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvml.dll time=2025-10-16T04:14:43.205+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvml.dll C:\\tools\\ruby34\\bin\\nvml.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvml.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvml.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvml.dll C:\\Python312\\Scripts\\nvml.dll C:\\Python312\\nvml.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvml.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvml.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvml.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvml.dll C:\\Windows\\system32\\nvml.dll C:\\Windows\\nvml.dll C:\\Windows\\System32\\Wbem\\nvml.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\Windows\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\WinMerge\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvml.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvml.dll C:\\ProgramData\\chocolatey\\bin\\nvml.dll H:\\Fooocus\\python_embeded\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvml.dll C:\\Program Files\\dotnet\\nvml.dll C:\\Program Files\\PuTTY\\nvml.dll C:\\Program Files\\Process Lasso\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvml.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvml.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvml.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvml.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvml.dll C:\\Program Files\\gsudo\\Current\\nvml.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvml.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll C:\\WINDOWS\\nvml.dll C:\\WINDOWS\\System32\\Wbem\\nvml.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvml.dll C:\\WINDOWS\\System32\\OpenSSH\\nvml.dll C:\\Program Files\\CrowdSec\\nvml.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvml.dll C:\\Program Files\\GitHub CLI\\nvml.dll C:\\Program Files\\Go\\bin\\nvml.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvml.dll C:\\Program Files\\nodejs\\nvml.dll C:\\Program Files\\CMake\\bin\\nvml.dll C:\\Program Files\\Calibre2\\nvml.dll C:\\Program Files\\PowerShell\\7\\nvml.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvml.dll C:\\Program Files\\Git\\cmd\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T04:14:43.219+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvml.dll" time=2025-10-16T04:14:43.223+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvml.dll C:\\WINDOWS\\system32\\nvml.dll c:\\Windows\\System32\\nvml.dll]" time=2025-10-16T04:14:43.279+02:00 level=DEBUG source=gpu.go:111 msg="nvidia-ml loaded" library=C:\Windows\system32\nvml.dll time=2025-10-16T04:14:43.280+02:00 level=DEBUG source=gpu.go:512 msg="Searching for GPU library" name=nvcuda.dll time=2025-10-16T04:14:43.283+02:00 level=DEBUG source=gpu.go:536 msg="gpu library search" globs="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\nvcuda.dll C:\\tools\\ruby34\\bin\\nvcuda.dll C:\\Program Files\\Oculus\\Support\\oculus-runtime\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin\\nvcuda.dll C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp\\nvcuda.dll C:\\VulkanSDK\\1.4.313.0\\Bin\\nvcuda.dll C:\\Python312\\Scripts\\nvcuda.dll C:\\Python312\\nvcuda.dll C:\\Program Files\\Zulu\\zulu-23\\bin\\nvcuda.dll C:\\Program Files\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath\\nvcuda.dll C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler\\nvcuda.dll C:\\Windows\\system32\\nvcuda.dll C:\\Windows\\nvcuda.dll C:\\Windows\\System32\\Wbem\\nvcuda.dll C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\Windows\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\WinMerge\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Microsoft VS Code\\bin\\nvcuda.dll C:\\ProgramData\\chocolatey\\bin\\nvcuda.dll H:\\Fooocus\\python_embeded\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts\\nvcuda.dll C:\\Program Files\\dotnet\\nvcuda.dll C:\\Program Files\\PuTTY\\nvcuda.dll C:\\Program Files\\Process Lasso\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts\\nvcuda.dll C:\\Users\\Iman Mirbioki\\miniconda3\\nvcuda.dll C:\\FPC\\3.2.2\\bin\\i386-Win32\\nvcuda.dll C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\nvcuda.dll C:\\Program Files\\Cloudflare\\Cloudflare WARP\\nvcuda.dll C:\\Program Files\\gsudo\\Current\\nvcuda.dll C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR\\nvcuda.dll C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll C:\\WINDOWS\\nvcuda.dll C:\\WINDOWS\\System32\\Wbem\\nvcuda.dll C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\nvcuda.dll C:\\WINDOWS\\System32\\OpenSSH\\nvcuda.dll C:\\Program Files\\CrowdSec\\nvcuda.dll C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\nvcuda.dll C:\\Program Files\\GitHub CLI\\nvcuda.dll C:\\Program Files\\Go\\bin\\nvcuda.dll C:\\Program Files\\gs\\gs10.05.1\\bin\\nvcuda.dll C:\\Program Files\\nodejs\\nvcuda.dll C:\\Program Files\\CMake\\bin\\nvcuda.dll C:\\Program Files\\Calibre2\\nvcuda.dll C:\\Program Files\\PowerShell\\7\\nvcuda.dll C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\nvcuda.dll C:\\Program Files\\Git\\cmd\\nvcuda.dll c:\\windows\\system*\\nvcuda.dll]" time=2025-10-16T04:14:43.301+02:00 level=DEBUG source=gpu.go:540 msg="skipping PhysX cuda library path" path="C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common\\nvcuda.dll" time=2025-10-16T04:14:43.306+02:00 level=DEBUG source=gpu.go:569 msg="discovered GPU libraries" paths="[C:\\Windows\\system32\\nvcuda.dll C:\\WINDOWS\\system32\\nvcuda.dll]" initializing C:\Windows\system32\nvcuda.dll dlsym: cuInit - 00007FFDC0B5DE10 dlsym: cuDriverGetVersion - 00007FFDC0B5DEB0 dlsym: cuDeviceGetCount - 00007FFDC0B5E6A6 dlsym: cuDeviceGet - 00007FFDC0B5E6A0 dlsym: cuDeviceGetAttribute - 00007FFDC0B5E000 dlsym: cuDeviceGetUuid - 00007FFDC0B5E6B2 dlsym: cuDeviceGetName - 00007FFDC0B5E6AC dlsym: cuCtxCreate_v3 - 00007FFDC0B5E72A dlsym: cuMemGetInfo_v2 - 00007FFDC0B5E82C dlsym: cuCtxDestroy - 00007FFDC0B5E73C calling cuInit calling cuDriverGetVersion raw version 0x32c8 CUDA driver version: 13.0 calling cuDeviceGetCount device count 3 time=2025-10-16T04:14:43.371+02:00 level=DEBUG source=gpu.go:125 msg="detected GPUs" count=3 library=C:\Windows\system32\nvcuda.dll [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA totalMem 12226mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] CUDA freeMem 10929mb [GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8] Compute Capability 12.0 [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA totalMem 12287mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] CUDA freeMem 11245mb [GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced] Compute Capability 8.6 time=2025-10-16T04:14:43.745+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA totalMem 12287mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] CUDA freeMem 11245mb [GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c] Compute Capability 8.6 time=2025-10-16T04:14:43.916+02:00 level=INFO source=gpu.go:321 msg="detected OS VRAM overhead" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" time=2025-10-16T04:14:43.925+02:00 level=DEBUG source=amd_windows.go:34 msg="unable to load amdhip64_6.dll, please make sure to upgrade to the latest amd driver: The specified module could not be found." releasing cuda driver library releasing nvml library time=2025-10-16T04:14:43.933+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="10.7 GiB" time=2025-10-16T04:14:43.935+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" time=2025-10-16T04:14:43.936+02:00 level=INFO source=types.go:131 msg="inference compute" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" [GIN] 2025/10/16 - 04:17:36 | 200 | 0s | 127.0.0.1 | HEAD "/" [GIN] 2025/10/16 - 04:17:36 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:19:06 | 200 | 31.3281ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:19:06 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:21:40 | 200 | 30.1901ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:21:40 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:21:43 | 200 | 32.5242ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:21:43 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:04 | 200 | 28.1053ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:22:05 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:53 | 200 | 29.7858ms | 127.0.0.1 | GET "/api/tags" [GIN] 2025/10/16 - 04:22:53 | 200 | 0s | 127.0.0.1 | GET "/api/ps" [GIN] 2025/10/16 - 04:22:54 | 200 | 515.4µs | 127.0.0.1 | GET "/api/version" time=2025-10-16T04:34:17.695+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="90.5 GiB" before.free_swap="468.1 GiB" now.total="127.9 GiB" now.free="88.7 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:17.716+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="10.7 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:17.725+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:17.733+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:17.741+02:00 level=DEBUG source=sched.go:188 msg="updating default concurrency" OLLAMA_MAX_LOADED_MODELS=9 gpu_count=3 time=2025-10-16T04:34:17.770+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=general.alignment default=32 time=2025-10-16T04:34:17.781+02:00 level=DEBUG source=sched.go:208 msg="loading first model" model=H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 llama_model_loader: loaded meta data with 27 key-value pairs and 707 tensors from H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 32B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 32B llama_model_loader: - kv 5: qwen3.block_count u32 = 64 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 25600 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 64 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 257 tensors llama_model_loader: - type f16: 64 tensors llama_model_loader: - type q4_K: 353 tensors llama_model_loader: - type q6_K: 33 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.81 GiB (4.93 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 1 print_info: model type = ?B print_info: model params = 32.76 B print_info: general.name = Qwen3 32B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 llama_model_load: vocab only - skipping tensors time=2025-10-16T04:34:18.349+02:00 level=WARN source=server.go:171 msg="requested context size too large for model" num_ctx=131072 n_ctx_train=40960 time=2025-10-16T04:34:18.353+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="88.7 GiB" before.free_swap="463.8 GiB" now.total="127.9 GiB" now.free="88.6 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:18.372+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="8.8 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:18.379+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="10.9 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:18.387+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:18.447+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.451+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.462+02:00 level=DEBUG source=server.go:324 msg="adding gpu library" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.464+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.469+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.473+02:00 level=DEBUG source=server.go:332 msg="adding gpu dependency paths" paths="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" time=2025-10-16T04:34:18.476+02:00 level=INFO source=server.go:399 msg="starting runner" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --model H:\\LLM\\blobs\\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 --port 63675" time=2025-10-16T04:34:18.479+02:00 level=DEBUG source=server.go:400 msg=subprocess CUDA_PATH="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" CUDA_PATH_V10_2="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2" CUDA_PATH_V12_8="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8" CUDA_PATH_V12_9="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9" CUDA_PATH_V13_0="C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0" OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=1 OLLAMA_MAX_LOADED_MODELS=9 OLLAMA_MODELS=H:\LLM OLLAMA_SCHED_SPREAD=1 PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v10.2\\libnvvp;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\WINDOWS\\system32;C:\\WINDOWS;C:\\WINDOWS\\System32\\Wbem;C:\\WINDOWS\\System32\\WindowsPowerShell\\v1.0\\;C:\\WINDOWS\\System32\\OpenSSH\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama;C:\\Program Files\\Git\\cmd;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" OLLAMA_LIBRARY_PATH="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.526+02:00 level=DEBUG source=gpu.go:402 msg="updating system memory data" before.total="127.9 GiB" before.free="88.6 GiB" before.free_swap="463.8 GiB" now.total="127.9 GiB" now.free="88.6 GiB" now.free_swap="463.8 GiB" time=2025-10-16T04:34:18.575+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 name="NVIDIA GeForce RTX 5070" overhead="0 B" before.total="11.9 GiB" before.free="8.8 GiB" now.total="11.9 GiB" now.free="8.8 GiB" now.used="3.1 GiB" time=2025-10-16T04:34:18.583+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced name="NVIDIA GeForce RTX 3060" overhead="133.5 MiB" before.total="12.0 GiB" before.free="10.9 GiB" now.total="12.0 GiB" now.free="10.9 GiB" now.used="1.0 GiB" time=2025-10-16T04:34:18.590+02:00 level=DEBUG source=gpu.go:452 msg="updating cuda memory data" gpu=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c name="NVIDIA GeForce RTX 3060" overhead="261.6 MiB" before.total="12.0 GiB" before.free="11.0 GiB" now.total="12.0 GiB" now.free="11.0 GiB" now.used="781.4 MiB" releasing nvml library time=2025-10-16T04:34:18.596+02:00 level=INFO source=server.go:504 msg="system memory" total="127.9 GiB" free="88.6 GiB" free_swap="463.8 GiB" time=2025-10-16T04:34:18.596+02:00 level=DEBUG source=memory.go:181 msg=evaluating library=cuda gpu_count=3 available="[11.0 GiB 10.9 GiB 8.8 GiB]" time=2025-10-16T04:34:18.601+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=qwen3.vision.block_count default=0 time=2025-10-16T04:34:18.606+02:00 level=DEBUG source=ggml.go:609 msg="default cache size estimate" "attention MiB"=10240 "attention bytes"=10737418240 "recurrent MiB"=0 "recurrent bytes"=0 time=2025-10-16T04:34:18.609+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.612+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="10.9 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.617+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="8.8 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.621+02:00 level=DEBUG source=memory.go:411 msg="insufficient VRAM to load any model layers" time=2025-10-16T04:34:18.625+02:00 level=DEBUG source=memory.go:181 msg=evaluating library=cuda gpu_count=3 available="[8.8 GiB 10.9 GiB 11.0 GiB]" time=2025-10-16T04:34:18.627+02:00 level=DEBUG source=ggml.go:274 msg="key with type not found" key=qwen3.vision.block_count default=0 time=2025-10-16T04:34:18.629+02:00 level=DEBUG source=ggml.go:609 msg="default cache size estimate" "attention MiB"=10240 "attention bytes"=10737418240 "recurrent MiB"=0 "recurrent bytes"=0 time=2025-10-16T04:34:18.631+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-7c7493fb-894e-ffa3-0d90-00b83a5731e8 library=cuda variant=v13 compute=12.0 driver=13.0 name="NVIDIA GeForce RTX 5070" total="11.9 GiB" available="8.8 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.634+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-e2200bae-da5e-037d-9fc2-d0d8c6081ced library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="10.9 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.641+02:00 level=DEBUG source=memory.go:269 msg="gpu has too little memory to allocate any layers" id=GPU-67bf118a-52b6-0f51-c4ad-c561cdfbb20c library=cuda variant=v13 compute=8.6 driver=13.0 name="NVIDIA GeForce RTX 3060" total="12.0 GiB" available="11.0 GiB" minimum_memory=479199232 layer_size="461.0 MiB" gpu_zer_overhead="0 B" partial_offload="13.3 GiB" full_offload="13.3 GiB" time=2025-10-16T04:34:18.645+02:00 level=DEBUG source=memory.go:411 msg="insufficient VRAM to load any model layers" time=2025-10-16T04:34:18.646+02:00 level=INFO source=server.go:544 msg=offload library=cuda layers.requested=-1 layers.model=65 layers.offload=0 layers.split=[] memory.available="[8.8 GiB 10.9 GiB 11.0 GiB]" memory.gpu_overhead="0 B" memory.required.full="28.4 GiB" memory.required.partial="0 B" memory.required.kv="10.0 GiB" memory.required.allocations="[0 B 0 B 0 B]" memory.weights.total="18.4 GiB" memory.weights.repeating="17.8 GiB" memory.weights.nonrepeating="608.6 MiB" memory.graph.full="13.3 GiB" memory.graph.partial="13.3 GiB" time=2025-10-16T04:34:18.693+02:00 level=INFO source=runner.go:864 msg="starting go runner" time=2025-10-16T04:34:18.694+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-16T04:34:18.750+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-16T04:34:18.766+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-16T04:34:18.769+02:00 level=INFO source=runner.go:900 msg="Server listening on 127.0.0.1:63675" time=2025-10-16T04:34:18.786+02:00 level=INFO source=runner.go:799 msg=load request="{Operation:commit LoraPath:[] Parallel:1 BatchSize:512 FlashAttention:false KvSize:40960 KvCacheType: NumThreads:28 GPULayers:[] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:false}" time=2025-10-16T04:34:18.786+02:00 level=INFO source=server.go:1251 msg="waiting for llama runner to start responding" time=2025-10-16T04:34:18.794+02:00 level=INFO source=server.go:1285 msg="waiting for server to become available" status="llm server loading model" llama_model_loader: loaded meta data with 27 key-value pairs and 707 tensors from H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen3 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = Qwen3 32B llama_model_loader: - kv 3: general.basename str = Qwen3 llama_model_loader: - kv 4: general.size_label str = 32B llama_model_loader: - kv 5: qwen3.block_count u32 = 64 llama_model_loader: - kv 6: qwen3.context_length u32 = 40960 llama_model_loader: - kv 7: qwen3.embedding_length u32 = 5120 llama_model_loader: - kv 8: qwen3.feed_forward_length u32 = 25600 llama_model_loader: - kv 9: qwen3.attention.head_count u32 = 64 llama_model_loader: - kv 10: qwen3.attention.head_count_kv u32 = 8 llama_model_loader: - kv 11: qwen3.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 12: qwen3.attention.layer_norm_rms_epsilon f32 = 0.000001 llama_model_loader: - kv 13: qwen3.attention.key_length u32 = 128 llama_model_loader: - kv 14: qwen3.attention.value_length u32 = 128 llama_model_loader: - kv 15: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 16: tokenizer.ggml.pre str = qwen2 llama_model_loader: - kv 17: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... llama_model_loader: - kv 18: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... llama_model_loader: - kv 19: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 20: tokenizer.ggml.eos_token_id u32 = 151645 llama_model_loader: - kv 21: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 22: tokenizer.ggml.bos_token_id u32 = 151643 llama_model_loader: - kv 23: tokenizer.ggml.add_bos_token bool = false llama_model_loader: - kv 24: tokenizer.chat_template str = {%- if tools %}\n {{- '<|im_start|>... llama_model_loader: - kv 25: general.quantization_version u32 = 2 llama_model_loader: - kv 26: general.file_type u32 = 15 llama_model_loader: - type f32: 257 tensors llama_model_loader: - type f16: 64 tensors llama_model_loader: - type q4_K: 353 tensors llama_model_loader: - type q6_K: 33 tensors print_info: file format = GGUF V3 (latest) print_info: file type = Q4_K - Medium print_info: file size = 18.81 GiB (4.93 BPW) init_tokenizer: initializing tokenizer for type 2 load: control token: 151659 '<|fim_prefix|>' is not marked as EOG load: control token: 151656 '<|video_pad|>' is not marked as EOG load: control token: 151655 '<|image_pad|>' is not marked as EOG load: control token: 151653 '<|vision_end|>' is not marked as EOG load: control token: 151652 '<|vision_start|>' is not marked as EOG load: control token: 151651 '<|quad_end|>' is not marked as EOG load: control token: 151649 '<|box_end|>' is not marked as EOG load: control token: 151648 '<|box_start|>' is not marked as EOG load: control token: 151646 '<|object_ref_start|>' is not marked as EOG load: control token: 151644 '<|im_start|>' is not marked as EOG load: control token: 151661 '<|fim_suffix|>' is not marked as EOG load: control token: 151647 '<|object_ref_end|>' is not marked as EOG load: control token: 151660 '<|fim_middle|>' is not marked as EOG load: control token: 151654 '<|vision_pad|>' is not marked as EOG load: control token: 151650 '<|quad_start|>' is not marked as EOG load: printing all EOG tokens: load: - 151643 ('<|endoftext|>') load: - 151645 ('<|im_end|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') load: special tokens cache size = 26 load: token to piece cache size = 0.9311 MB print_info: arch = qwen3 print_info: vocab_only = 0 print_info: n_ctx_train = 40960 print_info: n_embd = 5120 print_info: n_layer = 64 print_info: n_head = 64 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: is_swa_any = 0 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 8 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-06 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 25600 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = -1 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 40960 print_info: rope_finetuned = unknown print_info: model type = 32B print_info: model params = 32.76 B print_info: general.name = Qwen3 32B print_info: vocab type = BPE print_info: n_vocab = 151936 print_info: n_merges = 151387 print_info: BOS token = 151643 '<|endoftext|>' print_info: EOS token = 151645 '<|im_end|>' print_info: EOT token = 151645 '<|im_end|>' print_info: PAD token = 151643 '<|endoftext|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|endoftext|>' print_info: EOG token = 151645 '<|im_end|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = false) load_tensors: layer 0 assigned to device CPU, is_swa = 0 load_tensors: layer 1 assigned to device CPU, is_swa = 0 load_tensors: layer 2 assigned to device CPU, is_swa = 0 load_tensors: layer 3 assigned to device CPU, is_swa = 0 load_tensors: layer 4 assigned to device CPU, is_swa = 0 load_tensors: layer 5 assigned to device CPU, is_swa = 0 load_tensors: layer 6 assigned to device CPU, is_swa = 0 load_tensors: layer 7 assigned to device CPU, is_swa = 0 load_tensors: layer 8 assigned to device CPU, is_swa = 0 load_tensors: layer 9 assigned to device CPU, is_swa = 0 load_tensors: layer 10 assigned to device CPU, is_swa = 0 load_tensors: layer 11 assigned to device CPU, is_swa = 0 load_tensors: layer 12 assigned to device CPU, is_swa = 0 load_tensors: layer 13 assigned to device CPU, is_swa = 0 load_tensors: layer 14 assigned to device CPU, is_swa = 0 load_tensors: layer 15 assigned to device CPU, is_swa = 0 load_tensors: layer 16 assigned to device CPU, is_swa = 0 load_tensors: layer 17 assigned to device CPU, is_swa = 0 load_tensors: layer 18 assigned to device CPU, is_swa = 0 load_tensors: layer 19 assigned to device CPU, is_swa = 0 load_tensors: layer 20 assigned to device CPU, is_swa = 0 load_tensors: layer 21 assigned to device CPU, is_swa = 0 load_tensors: layer 22 assigned to device CPU, is_swa = 0 load_tensors: layer 23 assigned to device CPU, is_swa = 0 load_tensors: layer 24 assigned to device CPU, is_swa = 0 load_tensors: layer 25 assigned to device CPU, is_swa = 0 load_tensors: layer 26 assigned to device CPU, is_swa = 0 load_tensors: layer 27 assigned to device CPU, is_swa = 0 load_tensors: layer 28 assigned to device CPU, is_swa = 0 load_tensors: layer 29 assigned to device CPU, is_swa = 0 load_tensors: layer 30 assigned to device CPU, is_swa = 0 load_tensors: layer 31 assigned to device CPU, is_swa = 0 load_tensors: layer 32 assigned to device CPU, is_swa = 0 load_tensors: layer 33 assigned to device CPU, is_swa = 0 load_tensors: layer 34 assigned to device CPU, is_swa = 0 load_tensors: layer 35 assigned to device CPU, is_swa = 0 load_tensors: layer 36 assigned to device CPU, is_swa = 0 load_tensors: layer 37 assigned to device CPU, is_swa = 0 load_tensors: layer 38 assigned to device CPU, is_swa = 0 load_tensors: layer 39 assigned to device CPU, is_swa = 0 load_tensors: layer 40 assigned to device CPU, is_swa = 0 load_tensors: layer 41 assigned to device CPU, is_swa = 0 load_tensors: layer 42 assigned to device CPU, is_swa = 0 load_tensors: layer 43 assigned to device CPU, is_swa = 0 load_tensors: layer 44 assigned to device CPU, is_swa = 0 load_tensors: layer 45 assigned to device CPU, is_swa = 0 load_tensors: layer 46 assigned to device CPU, is_swa = 0 load_tensors: layer 47 assigned to device CPU, is_swa = 0 load_tensors: layer 48 assigned to device CPU, is_swa = 0 load_tensors: layer 49 assigned to device CPU, is_swa = 0 load_tensors: layer 50 assigned to device CPU, is_swa = 0 load_tensors: layer 51 assigned to device CPU, is_swa = 0 load_tensors: layer 52 assigned to device CPU, is_swa = 0 load_tensors: layer 53 assigned to device CPU, is_swa = 0 load_tensors: layer 54 assigned to device CPU, is_swa = 0 load_tensors: layer 55 assigned to device CPU, is_swa = 0 load_tensors: layer 56 assigned to device CPU, is_swa = 0 load_tensors: layer 57 assigned to device CPU, is_swa = 0 load_tensors: layer 58 assigned to device CPU, is_swa = 0 load_tensors: layer 59 assigned to device CPU, is_swa = 0 load_tensors: layer 60 assigned to device CPU, is_swa = 0 load_tensors: layer 61 assigned to device CPU, is_swa = 0 load_tensors: layer 62 assigned to device CPU, is_swa = 0 load_tensors: layer 63 assigned to device CPU, is_swa = 0 load_tensors: layer 64 assigned to device CPU, is_swa = 0 load_tensors: CPU model buffer size = 19259.71 MiB load_all_data: no device found for buffer type CPU for async uploads time=2025-10-16T04:34:22.813+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.02" time=2025-10-16T04:34:27.840+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.05" time=2025-10-16T04:34:28.092+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.06" time=2025-10-16T04:34:28.847+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.06" time=2025-10-16T04:34:29.601+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.106+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.358+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:30.612+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.07" time=2025-10-16T04:34:31.367+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:32.373+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:32.878+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.08" time=2025-10-16T04:34:33.130+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:33.382+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:34.137+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.09" time=2025-10-16T04:34:34.892+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.396+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.648+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:35.901+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.10" time=2025-10-16T04:34:36.405+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.11" time=2025-10-16T04:34:37.161+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.11" time=2025-10-16T04:34:37.665+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:37.917+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:38.171+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:38.674+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.12" time=2025-10-16T04:34:39.178+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:39.682+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:39.933+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.13" time=2025-10-16T04:34:40.438+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.14" time=2025-10-16T04:34:40.942+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.14" time=2025-10-16T04:34:41.194+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.446+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.699+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:41.951+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.15" time=2025-10-16T04:34:42.707+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.16" time=2025-10-16T04:34:42.959+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.16" time=2025-10-16T04:34:43.211+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:43.715+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:44.220+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.17" time=2025-10-16T04:34:44.724+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:44.976+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:45.480+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.18" time=2025-10-16T04:34:45.734+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:45.986+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:46.239+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.19" time=2025-10-16T04:34:46.491+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:46.743+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:47.247+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.20" time=2025-10-16T04:34:47.500+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:47.752+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:48.256+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.21" time=2025-10-16T04:34:48.761+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.013+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.266+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.519+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.22" time=2025-10-16T04:34:49.772+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.23" time=2025-10-16T04:34:50.276+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.23" time=2025-10-16T04:34:50.529+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:50.782+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:51.287+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.24" time=2025-10-16T04:34:51.792+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.045+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.298+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.25" time=2025-10-16T04:34:52.804+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.26" time=2025-10-16T04:34:53.309+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.26" time=2025-10-16T04:34:53.813+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.066+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.319+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.27" time=2025-10-16T04:34:54.824+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.329+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.583+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.28" time=2025-10-16T04:34:55.835+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.341+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.592+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.29" time=2025-10-16T04:34:56.846+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:57.101+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:57.353+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.30" time=2025-10-16T04:34:58.111+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:58.364+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:58.617+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.31" time=2025-10-16T04:34:59.122+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:34:59.375+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:34:59.880+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.32" time=2025-10-16T04:35:00.134+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:00.639+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:00.893+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.33" time=2025-10-16T04:35:01.399+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:01.656+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:01.908+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.34" time=2025-10-16T04:35:02.415+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.35" time=2025-10-16T04:35:02.921+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.35" time=2025-10-16T04:35:03.175+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:03.680+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:03.935+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.36" time=2025-10-16T04:35:04.190+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.443+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.697+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:04.950+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.37" time=2025-10-16T04:35:05.456+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:05.709+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:05.962+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.38" time=2025-10-16T04:35:06.467+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.39" time=2025-10-16T04:35:06.972+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.39" time=2025-10-16T04:35:07.479+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:07.733+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:07.986+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.40" time=2025-10-16T04:35:08.492+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:08.745+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:08.998+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:09.251+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.41" time=2025-10-16T04:35:09.505+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.009+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.264+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.42" time=2025-10-16T04:35:10.516+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:10.770+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:11.023+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.43" time=2025-10-16T04:35:11.530+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.036+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.289+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.44" time=2025-10-16T04:35:12.796+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.45" time=2025-10-16T04:35:13.049+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.45" time=2025-10-16T04:35:13.555+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:13.809+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:14.063+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" time=2025-10-16T04:35:14.569+02:00 level=DEBUG source=server.go:1295 msg="model load progress 0.46" llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 40960 llama_context: n_ctx_per_seq = 40960 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = 0 llama_context: kv_unified = false llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 set_abort_callback: call llama_context: CPU output buffer size = 0.60 MiB create_memory: n_ctx = 40960 (padded) llama_kv_cache_unified: layer 0: dev = CPU llama_kv_cache_unified: layer 1: dev = CPU llama_kv_cache_unified: layer 2: dev = CPU llama_kv_cache_unified: layer 3: dev = CPU llama_kv_cache_unified: layer 4: dev = CPU llama_kv_cache_unified: layer 5: dev = CPU llama_kv_cache_unified: layer 6: dev = CPU llama_kv_cache_unified: layer 7: dev = CPU llama_kv_cache_unified: layer 8: dev = CPU llama_kv_cache_unified: layer 9: dev = CPU llama_kv_cache_unified: layer 10: dev = CPU llama_kv_cache_unified: layer 11: dev = CPU llama_kv_cache_unified: layer 12: dev = CPU llama_kv_cache_unified: layer 13: dev = CPU llama_kv_cache_unified: layer 14: dev = CPU llama_kv_cache_unified: layer 15: dev = CPU llama_kv_cache_unified: layer 16: dev = CPU llama_kv_cache_unified: layer 17: dev = CPU llama_kv_cache_unified: layer 18: dev = CPU llama_kv_cache_unified: layer 19: dev = CPU llama_kv_cache_unified: layer 20: dev = CPU llama_kv_cache_unified: layer 21: dev = CPU llama_kv_cache_unified: layer 22: dev = CPU llama_kv_cache_unified: layer 23: dev = CPU llama_kv_cache_unified: layer 24: dev = CPU llama_kv_cache_unified: layer 25: dev = CPU llama_kv_cache_unified: layer 26: dev = CPU llama_kv_cache_unified: layer 27: dev = CPU llama_kv_cache_unified: layer 28: dev = CPU llama_kv_cache_unified: layer 29: dev = CPU llama_kv_cache_unified: layer 30: dev = CPU llama_kv_cache_unified: layer 31: dev = CPU llama_kv_cache_unified: layer 32: dev = CPU llama_kv_cache_unified: layer 33: dev = CPU llama_kv_cache_unified: layer 34: dev = CPU llama_kv_cache_unified: layer 35: dev = CPU llama_kv_cache_unified: layer 36: dev = CPU llama_kv_cache_unified: layer 37: dev = CPU llama_kv_cache_unified: layer 38: dev = CPU llama_kv_cache_unified: layer 39: dev = CPU llama_kv_cache_unified: layer 40: dev = CPU llama_kv_cache_unified: layer 41: dev = CPU llama_kv_cache_unified: layer 42: dev = CPU llama_kv_cache_unified: layer 43: dev = CPU llama_kv_cache_unified: layer 44: dev = CPU llama_kv_cache_unified: layer 45: dev = CPU llama_kv_cache_unified: layer 46: dev = CPU llama_kv_cache_unified: layer 47: dev = CPU llama_kv_cache_unified: layer 48: dev = CPU llama_kv_cache_unified: layer 49: dev = CPU llama_kv_cache_unified: layer 50: dev = CPU llama_kv_cache_unified: layer 51: dev = CPU llama_kv_cache_unified: layer 52: dev = CPU llama_kv_cache_unified: layer 53: dev = CPU llama_kv_cache_unified: layer 54: dev = CPU llama_kv_cache_unified: layer 55: dev = CPU llama_kv_cache_unified: layer 56: dev = CPU llama_kv_cache_unified: layer 57: dev = CPU llama_kv_cache_unified: layer 58: dev = CPU llama_kv_cache_unified: layer 59: dev = CPU llama_kv_cache_unified: layer 60: dev = CPU llama_kv_cache_unified: layer 61: dev = CPU llama_kv_cache_unified: layer 62: dev = CPU llama_kv_cache_unified: layer 63: dev = CPU llama_kv_cache_unified: CPU KV buffer size = 10240.00 MiB llama_kv_cache_unified: size = 10240.00 MiB ( 40960 cells, 64 layers, 1/1 seqs), K (f16): 5120.00 MiB, V (f16): 5120.00 MiB llama_context: enumerating backends llama_context: backend_ptrs.size() = 1 llama_context: max_nodes = 5656 llama_context: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 0 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 llama_context: CPU compute buffer size = 5256.01 MiB llama_context: graph nodes = 2502 llama_context: graph splits = 1 time=2025-10-16T04:36:51.998+02:00 level=INFO source=server.go:1289 msg="llama runner started in 153.52 seconds" time=2025-10-16T04:36:51.999+02:00 level=INFO source=sched.go:473 msg="loaded runners" count=1 time=2025-10-16T04:36:52.002+02:00 level=INFO source=server.go:1251 msg="waiting for llama runner to start responding" time=2025-10-16T04:36:52.002+02:00 level=INFO source=server.go:1289 msg="llama runner started in 153.53 seconds" time=2025-10-16T04:36:52.003+02:00 level=DEBUG source=sched.go:485 msg="finished setting up" runner.name=registry.ollama.ai/library/qwen3:32b runner.inference=cuda runner.devices=3 runner.size="28.4 GiB" runner.vram="0 B" runner.parallel=1 runner.pid=50652 runner.model=H:\LLM\blobs\sha256-3291abe70f16ee9682de7bfae08db5373ea9d6497e614aaad63340ad421d6312 runner.num_ctx=131072 time=2025-10-16T04:36:52.013+02:00 level=DEBUG source=server.go:1388 msg="completion request" images=0 prompt=4250 format="" time=2025-10-16T04:36:52.022+02:00 level=DEBUG source=cache.go:104 msg="loading cache slot" id=0 cache=0 prompt=891 used=0 remaining=891@Digital-Monk commented on GitHub (Oct 16, 2025):
I'm just jumping on the bandwagon. Haven't been able to get GPU inference since "upgrading" to 12.5 from 12.3. I have uninstalled, purged, and reinstalled everything, and still no luck. I've tried with and without CUDA_VISIBLE_DEVICES. This happened right after I powered up this machine after removing one of my GPUs to put in a different server, so I have been assuming that something went sideways during that change. But I've seen a couple of other notes about 12.5 causing grief. Let me get my system back into an up-to-date software/driver state and then I'll grab logs and send them.
@Digital-Monk commented on GitHub (Oct 16, 2025):
serve.log
smi.log
@dhiltgen commented on GitHub (Oct 16, 2025):
I haven't been able to reproduce on a Windows system with NVIDIA GPUs running the same driver version 581.57. The cuda_v13 library should be able to enumerate the GPUs, but for some reason it's failing silently. It seems like the libraries may not be getting loaded at all. Your log shows
What I see on my test system is:
So it's not loading the CPU or GPU libraries properly.
Can you share what files you see in
C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollamaandC:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13Maybe there's a dependency problem, or an AV system getting in the way?
On the dependency front, perhaps try simplifying your PATH down and see if that makes any difference? Something like
You could also try to install the latest MSVC C++ runtime, although I don't think that should have any impact.
https://aka.ms/vs/17/release/vc_redist.x64.exe
@dhiltgen commented on GitHub (Oct 16, 2025):
@Digital-Monk what type of GPU do you have?
The error in your logs says:
Oops, I see now you included the SMI output too
NVIDIA GeForce GTX 960This should be an easy fix.
@Digital-Monk commented on GitHub (Oct 16, 2025):
Yeah, ancient hardware. Still faster than the CPU, though :)
@dhiltgen commented on GitHub (Oct 16, 2025):
@Digital-Monk #12666 should get this working again. Sorry about that.
@Digital-Monk commented on GitHub (Oct 17, 2025):
0.12.6 fixed it for me. Thank you!
@catclaw commented on GitHub (Oct 17, 2025):
Ollama still refuses to load even the smallest models to the GPU, and falls back to CPU, no matter what I do...
I've tried pretty much everything.
I'm going to try removing all Nvidia drivers and reinstalling them from scratch, hoping it will fix the problem.
I tried:
$env:PATH="C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama"
$env:OLLAMA_DEBUG="2"
ollama serve 2>&1 | % ToString | tee-object serve.log
Didn't help either - but I'll upload the log file.
serve2.log
@dhiltgen commented on GitHub (Oct 17, 2025):
@catclaw the server2.log you attached appears to be from version 0.11.11 not 0.12.6
@catclaw commented on GitHub (Oct 17, 2025):
Yes, because the latest version refuses to detect my GPUs at all...
This is the output 0.12.6:
@catclaw commented on GitHub (Oct 17, 2025):
I've used Nvidia CleanUpTool to remove everything and reinstalled CUDA 12.8, then 12.9, and 13.0.
I tried running Ollama (both version 0.11.11 and 0.12.6) between the installations - but;
0.11.11 detects my GPUs - but falls back to using CPU, no matter how small the model is.
0.12.6 doesn't detect any of my GPUs.
Here is the output from 0.12.6:
@dhiltgen commented on GitHub (Oct 17, 2025):
@catclaw the directory listings you shared don't seem to match the file sizes I see for 0.12.6 - perhaps that was from an older version?
I'm still not sure what's going wrong, but for some reason, the
ggml-*.dlllibraries aren't loading. I'll try to add some more debug logging in a future version which might help troubleshoot. Until then, things to potentially try:@aminalshmu commented on GitHub (Oct 18, 2025):
Have been experiencing this same issue on Gentoo since around 0.12.5, this is from the just compiled git version:
However version 0.12.3 from guru overlay (same use flags, blas+cuda) it detects the GPU without issue:
Edited to include debug level 2 log. Thanks for your help with this! Let me know if I can provide any more info.
@dhiltgen commented on GitHub (Oct 18, 2025):
@aminalshmu do the official binaries correctly discover your GPU? It looks like your build didn't include the compute capability for your GPU. How did you build the native components? What CUDA version do you have installed?
@aminalshmu commented on GitHub (Oct 18, 2025):
Yes, official binary package detects GPU. I have been using the guru gentoo overlay, the -9999 package fetches current repo at build time:
https://github.com/gentoo/guru/tree/master/sci-ml/ollama
It was working with the live builds until about a week ago, around 0.12.5 I believe, and the guru ebuild has not changed in 2 weeks, seemed to work fine with source versions around 0.12.4. I have CUDA 13.0 and NVidia driver 580.95.05 FWIW.
I'm thinking this is a completely separate issue from catclaw as the Windows binary does not detect his GPUs but the 0.12.6 amd64 package works fine for me on Linux, just not building from source. Will use the binary for now but would like to figure out why the source version isn't working.
@miversen33 commented on GitHub (Oct 19, 2025):
Disregard the below. In my case it was a stupid cgroup oversight with
nvidia-uvm.I am noticing identical behavior with Ollama 0.12.5 and 0.12.6 on linux.
I am unsure if my issue is the same, but the result is the same. Ollama seems to refuse to talk to my 2 P40 GPUs. Been pulling my hair out for the past few days trying to debug it but from everything I can tell, the issue is simply that ollama doesn't like the GPUs.
In my case I installed with the installation script found here and when I run it, I notice that ollama doesn't have runners for anything pre cuda_12. For my GPUs, they run cuda_v11 and it seems that while Ollama can see the GPUS, it doesn't know how to talk to them as there is no runner for them
Github is fighting me and won't let me upload the log as a file so here is my log, maybe it will help shed some light on stuff?
time=2025-10-19T14:59:21.474Z level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES: GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:4096 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:/root/.ollama/models OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:false ROCR_VISIBLE_DEVICES: http_proxy: https_proxy: no_proxy:]" time=2025-10-19T14:59:21.474Z level=INFO source=images.go:522 msg="total blobs: 0" time=2025-10-19T14:59:21.474Z level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-19T14:59:21.474Z level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-19T14:59:21.475Z level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-19T14:59:21.475Z level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-19T14:59:21.475Z level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" extra_envs=[] time=2025-10-19T14:59:21.476Z level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash CGO_LDFLAGS=-L/opt/ollama-source/build/lib/ollama -L/usr/local/cuda-11.8/lib64 PWD=/usr/local/lib/ollama LOGNAME=root XDG_SESSION_TYPE=tty CXX=/usr/bin/g++-11 HOME=/root LANG=C LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: SSH_CONNECTION=10.40.1.52 34792 10.40.1.20 22 CUDAHOSTCXX=/usr/bin/g++-11 LESSCLOSE=/usr/bin/lesspipe %s %s XDG_SESSION_CLASS=user TERM=tmux-256color LESSOPEN=| /usr/bin/lesspipe %s USER=root SHLVL=1 XDG_SESSION_ID=3329 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v12:/opt/ollama-source/build/lib/ollama:/usr/local/cuda-11.8/lib64:/usr/local/cuda-11.8/lib64: XDG_RUNTIME_DIR=/run/user/0 SSH_CLIENT=10.40.1.52 34792 22 CUDA_HOME=/usr/local/cuda-11.8 PATH=/usr/local/cuda-11.8/bin:/usr/local/cuda-11.8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin CC=/usr/bin/gcc-11 CGO_ENABLED=1 DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/0/bus SSH_TTY=/dev/pts/3 OLDPWD=/root _=/usr/local/bin/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v12]" cmd="/usr/local/bin/ollama runner --ollama-engine --port 34211" time=2025-10-19T14:59:21.495Z level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-19T14:59:21.497Z level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:34211" time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.498Z level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-19T14:59:21.498Z level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-19T14:59:21.498Z level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-19T14:59:21.498Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu-haswell.so time=2025-10-19T14:59:21.520Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v12 ggml_cuda_init: failed to initialize CUDA: unknown error load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v12/libggml-cuda.so time=2025-10-19T14:59:21.626Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-19T14:59:21.626Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-19T14:59:21.627Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-19T14:59:21.627Z level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=129.726648ms time=2025-10-19T14:59:21.627Z level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=810ns time=2025-10-19T14:59:21.628Z level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" devices=[] time=2025-10-19T14:59:21.628Z level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=153.310479ms OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v12]" extra_envs=[] time=2025-10-19T14:59:21.628Z level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" extra_envs=[] time=2025-10-19T14:59:21.629Z level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[SHELL=/bin/bash CGO_LDFLAGS=-L/opt/ollama-source/build/lib/ollama -L/usr/local/cuda-11.8/lib64 PWD=/usr/local/lib/ollama LOGNAME=root XDG_SESSION_TYPE=tty CXX=/usr/bin/g++-11 HOME=/root LANG=C LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=00:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.avif=01;35:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:*~=00;90:*#=00;90:*.bak=00;90:*.crdownload=00;90:*.dpkg-dist=00;90:*.dpkg-new=00;90:*.dpkg-old=00;90:*.dpkg-tmp=00;90:*.old=00;90:*.orig=00;90:*.part=00;90:*.rej=00;90:*.rpmnew=00;90:*.rpmorig=00;90:*.rpmsave=00;90:*.swp=00;90:*.tmp=00;90:*.ucf-dist=00;90:*.ucf-new=00;90:*.ucf-old=00;90: SSH_CONNECTION=10.40.1.52 34792 10.40.1.20 22 CUDAHOSTCXX=/usr/bin/g++-11 LESSCLOSE=/usr/bin/lesspipe %s %s XDG_SESSION_CLASS=user TERM=tmux-256color LESSOPEN=| /usr/bin/lesspipe %s USER=root SHLVL=1 XDG_SESSION_ID=3329 OLLAMA_DEBUG=2 LD_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v13:/opt/ollama-source/build/lib/ollama:/usr/local/cuda-11.8/lib64:/usr/local/cuda-11.8/lib64: XDG_RUNTIME_DIR=/run/user/0 SSH_CLIENT=10.40.1.52 34792 22 CUDA_HOME=/usr/local/cuda-11.8 PATH=/usr/local/cuda-11.8/bin:/usr/local/cuda-11.8/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin CC=/usr/bin/gcc-11 CGO_ENABLED=1 DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/0/bus SSH_TTY=/dev/pts/3 OLDPWD=/root _=/usr/local/bin/ollama OLLAMA_LIBRARY_PATH=/usr/local/lib/ollama:/usr/local/lib/ollama/cuda_v13]" cmd="/usr/local/bin/ollama runner --ollama-engine --port 41121" time=2025-10-19T14:59:21.678Z level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-19T14:59:21.678Z level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:41121" time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.680Z level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-19T14:59:21.680Z level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-19T14:59:21.680Z level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-19T14:59:21.680Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama load_backend: loaded CPU backend from /usr/local/lib/ollama/libggml-cpu-haswell.so time=2025-10-19T14:59:21.701Z level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path=/usr/local/lib/ollama/cuda_v13 ggml_cuda_init: failed to initialize CUDA: unknown error load_backend: loaded CUDA backend from /usr/local/lib/ollama/cuda_v13/libggml-cuda.so time=2025-10-19T14:59:21.818Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA=1 CPU.0.BMI2=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 compiler=cgo(gcc) time=2025-10-19T14:59:21.818Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-19T14:59:21.819Z level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-19T14:59:21.819Z level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=139.677341ms time=2025-10-19T14:59:21.819Z level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=850ns time=2025-10-19T14:59:21.820Z level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" devices=[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=191.574864ms OLLAMA_LIBRARY_PATH="[/usr/local/lib/ollama /usr/local/lib/ollama/cuda_v13]" extra_envs=[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-19T14:59:21.820Z level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-19T14:59:21.820Z level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=345.517504ms time=2025-10-19T14:59:21.821Z level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="8.0 GiB" available="7.6 GiB" time=2025-10-19T14:59:21.821Z level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"Nvidia details
Card details:
Os details: Ubuntu 24.04 (in an LXC)
@dhiltgen commented on GitHub (Oct 20, 2025):
@catclaw the logging update is merged, so when 0.12.7 ships (later this week) please give it a try with OLLAMA_DEBUG=2 and hopefully we'll get a little more detail on why none of the libraries are loading properly on your system.
@comicchang commented on GitHub (Oct 21, 2025):
I am experiencing the same issue.
I am on Arch Linux, using the following packages:
cuda 13.0.2-1ollama 0.12.6-1ollama-cuda 0.12.6-1Here is my
nvidia-smioutput:And here is my
OLLAMA_DEBUG=2 ollama serveoutput:debug.txt
@fernvenue commented on GitHub (Oct 21, 2025):
Same here:
@dhiltgen commented on GitHub (Oct 21, 2025):
@comicchang it looks like the Arch packages aren't being built correctly. Until that's resolved by those maintainers, please try to use the Ollama official binaries and it should detect your GPU. If not, please share an updated log.
@fernvenue commented on GitHub (Oct 21, 2025):
Hi @dhiltgen, thx for your explain, just tried the latest official build, and seems it can now detect GPU correctly, thx again :)
@comicchang commented on GitHub (Oct 22, 2025):
Thanks, I checked the ollama PKGBUILD file and found that support for compute capability 6.1 (sm_61) was indeed removed. I will try to build the package myself.
@catclaw commented on GitHub (Oct 25, 2025):
Will do! Thank you so much!
I've made a lot of changes to my system, but the problem still persists.
This is the current output from "ollama serve" in a console:
time=2025-10-25T20:18:21.577+02:00 level=INFO source=routes.go:1511 msg="server config" env="map[CUDA_VISIBLE_DEVICES:0,1,2 GGML_VK_VISIBLE_DEVICES: GPU_DEVICE_ORDINAL: HIP_VISIBLE_DEVICES: HSA_OVERRIDE_GFX_VERSION: HTTPS_PROXY: HTTP_PROXY: NO_PROXY: OLLAMA_CONTEXT_LENGTH:131072 OLLAMA_DEBUG:DEBUG-4 OLLAMA_FLASH_ATTENTION:false OLLAMA_GPU_OVERHEAD:0 OLLAMA_HOST:http://127.0.0.1:11434 OLLAMA_INTEL_GPU:false OLLAMA_KEEP_ALIVE:5m0s OLLAMA_KV_CACHE_TYPE: OLLAMA_LLM_LIBRARY: OLLAMA_LOAD_TIMEOUT:5m0s OLLAMA_MAX_LOADED_MODELS:0 OLLAMA_MAX_QUEUE:512 OLLAMA_MODELS:H:\\LLM OLLAMA_MULTIUSER_CACHE:false OLLAMA_NEW_ENGINE:false OLLAMA_NOHISTORY:false OLLAMA_NOPRUNE:false OLLAMA_NUM_PARALLEL:1 OLLAMA_ORIGINS:[http://localhost https://localhost http://localhost:* https://localhost:* http://127.0.0.1 https://127.0.0.1 http://127.0.0.1:* https://127.0.0.1:* http://0.0.0.0 https://0.0.0.0 http://0.0.0.0:* https://0.0.0.0:* app://* file://* tauri://* vscode-webview://* vscode-file://*] OLLAMA_REMOTES:[ollama.com] OLLAMA_SCHED_SPREAD:true ROCR_VISIBLE_DEVICES:]" time=2025-10-25T20:18:21.619+02:00 level=INFO source=images.go:522 msg="total blobs: 146" time=2025-10-25T20:18:21.636+02:00 level=INFO source=images.go:529 msg="total unused blobs removed: 0" time=2025-10-25T20:18:21.651+02:00 level=INFO source=routes.go:1564 msg="Listening on 127.0.0.1:11434 (version 0.12.6)" time=2025-10-25T20:18:21.652+02:00 level=DEBUG source=sched.go:123 msg="starting llm scheduler" time=2025-10-25T20:18:21.658+02:00 level=INFO source=runner.go:80 msg="discovering available GPUs..." time=2025-10-25T20:18:21.658+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-25T20:18:21.709+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62569" time=2025-10-25T20:18:21.870+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:21.877+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62569" time=2025-10-25T20:18:21.883+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:21.885+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:21.887+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:21.888+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:21.888+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:21.890+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:21.893+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:21.896+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:21.898+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:21.927+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm" time=2025-10-25T20:18:21.962+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:21.964+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:21.967+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:21.969+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:21.971+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:21.973+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:21.974+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:21.977+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:21.979+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:21.979+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:21.981+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:21.982+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:21.984+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:21.988+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:21.990+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:21.992+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:21.992+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:21.995+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:21.996+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:21.998+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.001+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.002+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.003+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.005+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=122.7994ms time=2025-10-25T20:18:22.005+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.009+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" devices=[] time=2025-10-25T20:18:22.012+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=353.0328ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm]" extra_envs=[] time=2025-10-25T20:18:22.015+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-25T20:18:22.023+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62584" time=2025-10-25T20:18:22.194+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:22.199+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62584" time=2025-10-25T20:18:22.211+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.213+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:22.214+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:22.215+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.216+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:22.219+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:22.220+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:22.221+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:22.223+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:22.252+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12" time=2025-10-25T20:18:22.301+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:22.302+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.304+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:22.306+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:22.307+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:22.308+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:22.310+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:22.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:22.311+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:22.315+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:22.317+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:22.318+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:22.319+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:22.320+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:22.321+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:22.323+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:22.329+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:22.329+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:22.331+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.332+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.333+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.333+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.335+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=124.9775ms time=2025-10-25T20:18:22.339+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.342+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" devices=[] time=2025-10-25T20:18:22.344+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=325.1639ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v12]" extra_envs=[] time=2025-10-25T20:18:22.345+02:00 level=DEBUG source=runner.go:448 msg="spawning runner with" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-25T20:18:22.349+02:00 level=TRACE source=runner.go:529 msg="starting runner for device discovery" env="[=::=::\\ =C:=C:\\Windows\\System32 ALLUSERSPROFILE=C:\\ProgramData APPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Roaming ChocolateyInstall=C:\\ProgramData\\chocolatey ChocolateyLastPathUpdate=133956273910003280 ChocolateyToolsLocation=C:\\tools CL=C:\\Program Files\\Microsoft Visual Studio\\2022\\Community\\VC\\Tools\\MSVC\\14.42.34433\\bin\\Hostx64 CommonProgramFiles=C:\\Program Files\\Common Files CommonProgramFiles(x86)=C:\\Program Files (x86)\\Common Files CommonProgramW6432=C:\\Program Files\\Common Files COMPUTERNAME=SINGULARITY ComSpec=C:\\WINDOWS\\system32\\cmd.exe CUDA_PATH=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_PATH_V12_8=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8 CUDA_PATH_V12_9=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9 CUDA_PATH_V13_0=C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v13.0 CUDA_VISIBLE_DEVICES=0,1,2 DriverData=C:\\Windows\\System32\\Drivers\\DriverData EFC_22116_1592913036=1 GOPATH=C:\\Users\\Iman Mirbioki\\go GOROOT=C:\\Program Files\\Go HOMEDRIVE=C: HOMEPATH=\\Users\\Iman Mirbioki INTEL_DEV_REDIST=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\ JAVA_HOME=C:\\Program Files\\Zulu\\zulu-23\\ LOCALAPPDATA=C:\\Users\\Iman Mirbioki\\AppData\\Local LOGONSERVER=\\\\SINGULARITY MIC_LD_LIBRARY_PATH=C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\compiler\\lib\\mic NUMBER_OF_PROCESSORS=56 NVVFX_MODEL_DIR=C:\\Program Files\\NVIDIA Corporation\\NVIDIA Video Effects\\models OculusBase=C:\\Program Files\\Oculus\\ OLLAMA_CONTEXT_LENGTH=131072 OLLAMA_DEBUG=2 OLLAMA_ENABLE_CUDA=1 OLLAMA_MODELS=H:\\LLM OLLAMA_SCHED_SPREAD=1 OneDrive=C:\\Users\\Iman Mirbioki\\OneDrive OneDriveConsumer=C:\\Users\\Iman Mirbioki\\OneDrive OpenSC=C:\\Program Files\\OpenSC Project\\OpenSC\\tools OS=Windows_NT PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\rocm;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\tools\\ruby34\\bin;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;H:\\Fooocus\\python_embeded\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Program Files\\dotnet\\;C:\\Program Files\\PuTTY\\;C:\\Program Files\\Process Lasso\\;C:\\Users\\Iman Mirbioki\\miniconda3\\Scripts;C:\\Users\\Iman Mirbioki\\miniconda3;C:\\FPC\\3.2.2\\bin\\i386-Win32;C:\\Program Files\\Microsoft SQL Server\\Client SDK\\ODBC\\170\\Tools\\Binn\\;C:\\Program Files\\Cloudflare\\Cloudflare WARP\\;C:\\Program Files\\gsudo\\Current;C:\\Program Files (x86)\\Windows Kits\\10\\Windows Performance Toolkit\\;C:\\Program Files\\CrowdSec\\;C:\\Program Files\\GitHub CLI\\;C:\\Program Files\\Go\\bin;C:\\Program Files\\gs\\gs10.05.1\\bin;C:\\Program Files\\nodejs\\;C:\\Program Files\\CMake\\bin;C:\\Program Files\\Calibre2\\;C:\\Program Files\\PowerShell\\7\\;C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common;C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2025.3.0\\;C:\\Program Files\\NVIDIA Corporation\\NVIDIA App\\NvDLISR;C:\\Program Files\\Git\\cmd;C:\\Program Files\\Python\\Scripts\\;C:\\Program Files\\Python\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\Scripts\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Python313\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Python\\Launcher\\;C:\\Cargo\\bin;C:\\Users\\Iman Mirbioki\\scoop\\shims;C:\\VulkanSDK\\1.4.313.0\\Bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Python312\\Scripts\\;C:\\Python312\\;C:\\Program Files\\Zulu\\zulu-23\\bin\\;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.9\\libnvvp;C:\\Program Files\\Oculus\\Support\\oculus-runtime;C:\\Program Files\\NVIDIA\\CUDNN\\v9.7\\bin;C:\\Program Files\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\java8path;C:\\Program Files (x86)\\Common Files\\Oracle\\Java\\javapath;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\bin;C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.8\\libnvvp;C:\\Program Files (x86)\\Common Files\\Intel\\Shared Libraries\\redist\\intel64\\compiler;C:\\Windows\\system32;C:\\Windows;C:\\Windows\\System32\\Wbem;C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\;C:\\Windows\\System32\\OpenSSH\\;C:\\Program Files\\WinMerge;C:\\Program Files\\Microsoft SQL Server\\150\\Tools\\Binn\\;C:\\Program Files\\Microsoft VS Code\\bin;C:\\ProgramData\\chocolatey\\bin;E:\\Documents\\PowerShell\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WinGet\\Links;C:\\msys64\\ucrt64\\bin;C:\\Users\\Iman Mirbioki\\.dotnet\\tools;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\Python\\Python312\\Scripts;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Microsoft\\WindowsApps;C:\\Program Files (x86)\\GnuWin32\\bin\\;c:\\users\\iman mirbioki\\appdata\\roaming\\python\\python312\\site-packages;C:\\Program Files\\Git\\usr\\bin\\;C:\\Program Files (x86)\\Nmap;C:\\Users\\Iman Mirbioki\\go\\bin;C:\\Program Files\\GPAC;C:\\Users\\Iman Mirbioki\\AppData\\Roaming\\npm;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Runtimes\\6.1.2\\usr\\bin\\;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Toolchains\\6.1.2+Asserts\\usr\\bin\\;;C:\\Users\\Iman Mirbioki\\.lmstudio\\bin;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama PATHEXT=.COM;.EXE;.BAT;.CMD;.VBS;.VBE;.JS;.JSE;.WSF;.WSH;.MSC;.PY;.PYW;.RB;.RBW PM_PACKAGES_ROOT=C:\\packman-repo POWERSHELL_DISTRIBUTION_CHANNEL=MSI:Windows 10 Enterprise PROCESSOR_ARCHITECTURE=AMD64 PROCESSOR_IDENTIFIER=Intel64 Family 6 Model 79 Stepping 1, GenuineIntel PROCESSOR_LEVEL=6 PROCESSOR_REVISION=4f01 PROGF81DEF27053=1 ProgramData=C:\\ProgramData ProgramFiles=C:\\Program Files ProgramFiles(x86)=C:\\Program Files (x86) ProgramW6432=C:\\Program Files PROMPT=$P$G PSModulePath=C:\\Program Files\\WindowsPowerShell\\Modules;C:\\WINDOWS\\system32\\WindowsPowerShell\\v1.0\\Modules PUBLIC=C:\\Users\\Public SDKROOT=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Swift\\Platforms\\6.1.2\\Windows.platform\\Developer\\SDKs\\Windows.sdk\\ SESSIONNAME=Console SystemDrive=C: SystemRoot=C:\\WINDOWS TEMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp TMP=C:\\Users\\IMANMI~1\\AppData\\Local\\Temp USERDOMAIN=SINGULARITY USERDOMAIN_ROAMINGPROFILE=SINGULARITY USERNAME=Iman Mirbioki USERPROFILE=C:\\Users\\Iman Mirbioki VK_SDK_PATH=C:\\VulkanSDK\\1.4.313.0 VS140COMNTOOLS=C:\\Program Files (x86)\\Microsoft Visual Studio 14.0\\Common7\\Tools\\ VULKAN_SDK=C:\\VulkanSDK\\1.4.313.0 windir=C:\\WINDOWS __PSLockDownPolicy=0 OLLAMA_LIBRARY_PATH=C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama;C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" cmd="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\ollama.exe runner --ollama-engine --port 62602" time=2025-10-25T20:18:22.526+02:00 level=INFO source=runner.go:1332 msg="starting ollama engine" time=2025-10-25T20:18:22.532+02:00 level=INFO source=runner.go:1367 msg="Server listening on 127.0.0.1:62602" time=2025-10-25T20:18:22.544+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.545+02:00 level=DEBUG source=gguf.go:578 msg=general.architecture type=string time=2025-10-25T20:18:22.546+02:00 level=DEBUG source=gguf.go:578 msg=tokenizer.ggml.model type=string time=2025-10-25T20:18:22.547+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.alignment default=32 time=2025-10-25T20:18:22.549+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.file_type default=0 time=2025-10-25T20:18:22.552+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.name default="" time=2025-10-25T20:18:22.553+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=general.description default="" time=2025-10-25T20:18:22.554+02:00 level=INFO source=ggml.go:134 msg="" architecture=llama file_type=unknown name="" description="" num_tensors=0 num_key_values=3 time=2025-10-25T20:18:22.555+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama" time=2025-10-25T20:18:22.584+02:00 level=DEBUG source=ggml.go:94 msg="ggml backend load all from path" path="C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13" time=2025-10-25T20:18:22.607+02:00 level=INFO source=ggml.go:104 msg=system CPU.0.LLAMAFILE=1 compiler=cgo(clang) time=2025-10-25T20:18:22.607+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.610+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.pooling_type default=0 time=2025-10-25T20:18:22.612+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.expert_count default=0 time=2025-10-25T20:18:22.612+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.tokens default="&{size:0 values:[]}" time=2025-10-25T20:18:22.614+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.scores default="&{size:0 values:[]}" time=2025-10-25T20:18:22.615+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.token_type default="&{size:0 values:[]}" time=2025-10-25T20:18:22.616+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.merges default="&{size:0 values:[]}" time=2025-10-25T20:18:22.617+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_bos_token default=true time=2025-10-25T20:18:22.621+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.bos_token_id default=0 time=2025-10-25T20:18:22.623+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.add_eos_token default=false time=2025-10-25T20:18:22.623+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_id default=0 time=2025-10-25T20:18:22.625+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.eos_token_ids default="&{size:0 values:[]}" time=2025-10-25T20:18:22.625+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=tokenizer.ggml.pre default="" time=2025-10-25T20:18:22.628+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.block_count default=0 time=2025-10-25T20:18:22.628+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.embedding_length default=0 time=2025-10-25T20:18:22.629+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count default=0 time=2025-10-25T20:18:22.634+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.head_count_kv default=0 time=2025-10-25T20:18:22.635+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.key_length default=0 time=2025-10-25T20:18:22.636+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.dimension_count default=0 time=2025-10-25T20:18:22.637+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.attention.layer_norm_rms_epsilon default=0 time=2025-10-25T20:18:22.638+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.freq_base default=100000 time=2025-10-25T20:18:22.639+02:00 level=DEBUG source=ggml.go:276 msg="key with type not found" key=llama.rope.scaling.factor default=1 time=2025-10-25T20:18:22.640+02:00 level=DEBUG source=runner.go:1307 msg="dummy model load took" duration=97.5824ms time=2025-10-25T20:18:22.642+02:00 level=DEBUG source=runner.go:1312 msg="gathering device infos took" duration=0s time=2025-10-25T20:18:22.649+02:00 level=TRACE source=runner.go:548 msg="runner enumerated devices" OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" devices=[] time=2025-10-25T20:18:22.651+02:00 level=DEBUG source=runner.go:451 msg="bootstrap discovery took" duration=304.9441ms OLLAMA_LIBRARY_PATH="[C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama C:\\Users\\Iman Mirbioki\\AppData\\Local\\Programs\\Ollama\\lib\\ollama\\cuda_v13]" extra_envs=[] time=2025-10-25T20:18:22.652+02:00 level=DEBUG source=runner.go:118 msg="filtering out unsupported or overlapping GPU library combinations" count=0 time=2025-10-25T20:18:22.653+02:00 level=TRACE source=runner.go:171 msg="supported GPU library combinations" supported=map[] time=2025-10-25T20:18:22.654+02:00 level=DEBUG source=runner.go:45 msg="GPU bootstrap discovery took" duration=1.0015121s time=2025-10-25T20:18:22.655+02:00 level=INFO source=types.go:129 msg="inference compute" id=cpu library=cpu compute="" name=cpu description=cpu libdirs=ollama driver="" pci_id="" type="" total="127.9 GiB" available="85.5 GiB" time=2025-10-25T20:18:22.659+02:00 level=INFO source=routes.go:1605 msg="entering low vram mode" "total vram"="0 B" threshold="20.0 GiB"This is the list of all files of 0.12.6 installation:
` Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama
2025-10-25 20:14
.2025-10-22 02:44 ..
2025-10-16 23:00 7 502 app.ico
2025-10-25 20:14 lib
2025-10-16 23:09 13 897 208 ollama app.exe
2025-10-16 23:09 32 979 448 ollama.exe
2025-10-25 20:14 446 008 unins000.dat
2025-10-25 20:12 3 706 360 unins000.exe
2025-10-25 20:14 25 291 unins000.msg
6 File(s) 51 061 817 bytes
Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib
2025-10-25 20:14
.2025-10-25 20:14 ..
2025-10-25 20:13 ollama
2025-10-25 20:14 2 138 Ollama.lnk
1 File(s) 2 138 bytes
Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama
2025-10-25 20:13
.2025-10-25 20:14 ..
2025-10-25 20:13 cuda_v12
2025-10-25 20:13 cuda_v13
2025-10-16 23:09 902 136 ggml-base.dll
2025-10-16 23:09 1 067 512 ggml-cpu-alderlake.dll
2025-10-16 23:09 1 070 584 ggml-cpu-haswell.dll
2025-10-16 23:09 1 296 888 ggml-cpu-icelake.dll
2025-10-16 23:09 955 896 ggml-cpu-sandybridge.dll
2025-10-16 23:09 1 295 352 ggml-cpu-skylakex.dll
2025-10-16 23:09 794 616 ggml-cpu-sse42.dll
2025-10-16 23:09 783 352 ggml-cpu-x64.dll
2025-10-25 20:13 rocm
8 File(s) 8 166 336 bytes
Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v12
2025-10-25 20:13
.2025-10-25 20:13 ..
2025-10-16 23:09 113 720 824 cublas64_12.dll
2025-10-16 23:09 692 449 784 cublasLt64_12.dll
2025-10-16 23:09 582 136 cudart64_12.dll
2025-10-16 23:09 1 569 384 952 ggml-cuda.dll
4 File(s) 2 376 137 696 bytes
Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\cuda_v13
2025-10-25 20:13
.2025-10-25 20:13 ..
2025-10-16 23:09 50 286 072 cublas64_13.dll
2025-10-16 23:09 477 894 648 cublasLt64_13.dll
2025-10-16 23:09 293 704 696 ggml-cuda.dll
3 File(s) 821 885 416 bytes
Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm
2025-10-25 20:13
.2025-10-25 20:13 ..
2025-10-16 23:09 23 660 536 amdhip64_6.dll
2025-10-16 23:09 111 860 216 amd_comgr_2.dll
2025-10-16 23:09 426 622 968 ggml-hip.dll
2025-10-16 23:09 908 792 hipblas.dll
2025-10-25 20:13 rocblas
2025-10-16 23:09 416 692 728 rocblas.dll
5 File(s) 979 745 240 bytes
Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\rocblas
2025-10-25 20:13
.2025-10-25 20:13 ..
2025-10-25 20:14 library
0 File(s) 0 bytes
Directory of C:\Users\Iman Mirbioki\AppData\Local\Programs\Ollama\lib\ollama\rocm\rocblas\library
2025-10-25 20:14
.2025-10-25 20:13 ..
2025-10-16 23:03 248 288 Kernels.so-000-gfx1030.hsaco
2025-10-16 23:03 262 120 Kernels.so-000-gfx1100.hsaco
2025-10-16 23:03 262 376 Kernels.so-000-gfx1101.hsaco
2025-10-16 23:03 262 120 Kernels.so-000-gfx1102.hsaco
2025-10-16 23:03 260 840 Kernels.so-000-gfx1151.hsaco
2025-10-16 23:03 23 026 TensileLibrary_lazy_gfx1030.dat
2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1100.dat
2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1101.dat
2025-10-16 23:03 24 186 TensileLibrary_lazy_gfx1102.dat
2025-10-16 23:03 17 653 TensileLibrary_lazy_gfx1151.dat
2025-10-16 23:03 43 504 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 161 304 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 170 784 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 37 230 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 134 320 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 142 264 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 141 240 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 43 848 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 161 048 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 170 016 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 169 760 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 170 272 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 169 760 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 37 398 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 133 808 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 141 752 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 141 496 TensileLibrary_Type_4xi8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 56 121 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 474 768 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 500 632 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 500 376 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 16 798 632 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 549 603 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 17 275 880 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 583 354 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 11 487 568 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 244 064 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 53 167 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 437 856 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 472 168 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 469 864 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 16 199 336 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 645 565 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 6 410 600 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 309 377 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 8 035 592 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 215 028 TensileLibrary_Type_BB_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 39 542 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 189 456 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 199 448 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 199 448 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 198 936 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 198 424 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 15 258 976 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 590 496 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 12 111 384 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 477 855 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 516 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 317 608 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 53 294 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 434 272 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 468 328 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 467 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 468 328 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 467 048 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 11 708 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 576 526 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 6 024 760 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 382 857 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 4 275 560 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 145 331 TensileLibrary_Type_BB_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 54 356 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 449 936 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 465 048 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 465 048 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 466 072 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 464 536 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 32 004 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 240 096 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 252 392 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 251 624 TensileLibrary_Type_BS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 17 086 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 81 496 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 85 344 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 84 832 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 28 317 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 195 440 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 204 152 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 203 384 TensileLibrary_Type_BS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 298 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 286 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 960 TensileLibrary_Type_CC_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 326 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 49 024 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 49 024 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 48 768 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 48 768 TensileLibrary_Type_CC_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 43 888 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 200 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 45 944 TensileLibrary_Type_CC_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 314 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 912 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 712 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 968 TensileLibrary_Type_CC_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 904 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 704 TensileLibrary_Type_CC_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 302 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 43 880 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 45 936 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 46 192 TensileLibrary_Type_CC_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 094 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 20 008 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 8 802 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 33 864 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 35 664 TensileLibrary_Type_DD_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 110 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 20 008 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 20 784 TensileLibrary_Type_DD_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 110 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 19 752 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 20 528 TensileLibrary_Type_DD_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 27 664 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 186 632 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 200 720 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 200 464 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 5 667 184 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 320 632 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 12 455 016 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 518 011 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 16 953 096 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 626 281 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 9 981 064 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 257 641 TensileLibrary_Type_HH_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 20 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 90 288 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 97 976 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 97 720 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 10 899 056 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 606 844 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 171 584 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 568 803 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 5 949 184 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 336 152 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 9 584 152 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 322 795 TensileLibrary_Type_HH_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 20 980 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 104 888 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 112 832 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 113 088 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 112 832 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 113 088 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 5 846 216 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 346 808 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 478 496 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 507 834 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 11 279 360 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 504 971 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 7 762 144 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 240 050 TensileLibrary_Type_HH_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 20 976 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 89 776 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 96 696 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 96 952 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 96 952 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 96 696 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 8 805 984 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 433 493 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 7 842 920 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 526 415 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 4 909 552 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 375 873 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 3 627 104 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 164 717 TensileLibrary_Type_HH_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 74 464 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 380 904 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 395 504 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 394 992 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 9 250 488 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 624 107 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 13 129 136 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 540 665 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 14 238 888 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 523 364 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 4 892 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 139 468 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 52 394 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 196 904 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 206 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 206 640 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 207 920 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 206 384 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 13 248 936 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 159 612 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 984 424 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 562 937 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 5 833 984 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 329 670 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 12 036 552 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 391 906 TensileLibrary_Type_HH_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 45 640 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 108 496 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 112 856 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 112 600 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 1 538 176 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 143 490 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 11 003 984 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 537 678 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 9 958 440 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 473 572 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 8 283 312 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 251 527 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 45 664 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 164 056 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 171 488 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 171 488 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 172 768 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 171 232 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 11 981 432 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 060 470 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 8 804 720 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 563 577 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 4 893 264 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 376 677 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 5 387 384 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 226 414 TensileLibrary_Type_HH_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 46 618 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 340 912 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 349 624 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 349 624 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 348 856 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 348 856 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 28 091 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 201 336 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 211 072 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 208 768 TensileLibrary_Type_HS_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 20 519 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 102 576 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 106 936 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 106 168 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 31 766 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 236 000 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 244 968 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 244 712 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 246 504 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 244 712 TensileLibrary_Type_HS_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 70 739 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 491 840 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 539 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 539 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 539 976 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 538 952 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 3 253 936 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 318 729 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 22 255 688 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 679 573 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 15 519 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 571 645 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 17 097 848 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 327 608 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 56 959 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 334 288 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 364 760 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 364 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 365 528 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 364 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 2 191 064 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 197 238 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 17 182 176 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 698 938 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 7 241 096 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 413 453 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 15 757 872 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 397 205 TensileLibrary_Type_I8I_HPA_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 67 638 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 463 320 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 509 664 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 509 664 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 510 432 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 509 408 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 2 486 384 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 234 821 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 16 680 632 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 633 900 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 5 576 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 337 236 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 14 921 928 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 375 463 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 65 854 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 411 816 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 451 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 451 248 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 2 302 504 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 234 701 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 10 926 528 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 710 893 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 3 804 720 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 387 494 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 7 114 904 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 318 061 TensileLibrary_Type_I8I_HPA_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 53 296 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 58 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 15 770 432 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 263 969 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 15 770 432 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 1 263 969 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 976 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 016 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 224 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Ailk_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 110 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 34 856 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 37 936 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 16 283 560 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 447 637 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 16 283 560 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 1 447 637 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 984 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 017 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 232 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Ailk_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 34 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 37 432 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 8 104 512 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 563 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 8 104 512 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 563 352 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 984 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 016 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 232 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Alik_Bjlk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 114 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 44 080 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 47 416 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 47 416 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 47 672 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 47 160 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 17 987 376 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.co
2025-10-16 23:03 1 423 043 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1030.dat
2025-10-16 23:03 17 987 376 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.co
2025-10-16 23:03 1 423 043 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1100.dat
2025-10-16 23:03 27 992 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.co
2025-10-16 23:03 2 017 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1101.dat
2025-10-16 23:03 13 240 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.co
2025-10-16 23:03 2 007 TensileLibrary_Type_SS_Contraction_l_Alik_Bljk_Cijk_Dijk_gfx1102.dat
2025-10-16 23:03 7 330 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Ailk_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 326 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Ailk_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 358 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 760 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 304 TensileLibrary_Type_ZZ_Contraction_l_AlikC_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 47 984 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 040 TensileLibrary_Type_ZZ_Contraction_l_AlikC_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 346 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 752 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 296 TensileLibrary_Type_ZZ_Contraction_l_Alik_BjlkC_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 48 744 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 288 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bjlk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 7 334 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback.dat
2025-10-16 23:03 47 976 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1030.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1100.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1101.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1102.hsaco
2025-10-16 23:03 50 032 TensileLibrary_Type_ZZ_Contraction_l_Alik_Bljk_Cijk_Dijk_fallback_gfx1151.hsaco
2025-10-16 23:03 115 291 TensileManifest.txt
487 File(s) 788 770 641 bytes
Meanwhile, I'm using LM Studio on the same port as Ollama (Not at the same time) to make things easier, and linked my existing models using Gollama -L, and it's working without any issues, besides I feel that I don't have the same capabilities, and LM Studio feels slower, and I get random "Out of memory" errors on 70B models...
Can't wait till 0.12.7 release
Output from LM Studio server:
[2025-10-25 20:35:43][DEBUG][LM Studio] GPU Configuration: Strategy: evenly Priority: [] Disabled GPUs: [] Limit weight offload to dedicated GPU Memory: OFF Offload KV Cache to GPU: ON [2025-10-25 20:35:43][DEBUG][LM Studio] Live GPU memory info (source 'LMS Core'): GPU 0: NVIDIA GeForce RTX 5070 (Used: 4.12 GB, Total: 12.82 GB, Free: 8.70 GB) GPU 1: NVIDIA GeForce RTX 3060 (Used: 970.50 MB, Total: 12.88 GB, Free: 11.91 GB) GPU 2: NVIDIA GeForce RTX 3060 (Used: 998.65 MB, Total: 12.88 GB, Free: 11.89 GB) [2025-10-25 20:35:43][DEBUG][LM Studio] Model load size estimate with raw num offload layers 'max' and context length '131072': Model: 18.40 GB Context: 46.55 GB Total: 64.96 GB [2025-10-25 20:35:43][DEBUG][LM Studio] Strict GPU VRAM cap is OFF: GPU offload layers will not be checked for adjustment [LM Studio] Resolved GPU config options: Num Offload Layers: max Num CPU Expert Layers: 0 Main GPU: 0 Tensor Split: [0] Disabled GPUs: [] [2025-10-25 20:35:43][DEBUG] ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no ggml_cuda_init: found 3 CUDA devices: Device 0: NVIDIA GeForce RTX 5070, compute capability 12.0, VMM: yes Device 1: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes Device 2: NVIDIA GeForce RTX 3060, compute capability 8.6, VMM: yes [2025-10-25 20:35:43][DEBUG] CUDA : ARCHS = 750,800,890,900,1000,1200 | USE_GRAPHS = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | BMI2 = 1 | LLAMAFILE = 1 | OPENMP = 1 | REPACK = 1 | [2025-10-25 20:35:43][DEBUG] llama_model_load_from_file_impl: using device CUDA0 (NVIDIA GeForce RTX 5070) (0000:03:00.0) - 10929 MiB free llama_model_load_from_file_impl: using device CUDA1 (NVIDIA GeForce RTX 3060) (0000:04:00.0) - 11245 MiB free llama_model_load_from_file_impl: using device CUDA2 (NVIDIA GeForce RTX 3060) (0000:83:00.0) - 11245 MiB free [2025-10-25 20:35:44][DEBUG] llama_model_loader: loaded meta data with 44 key-value pairs and 771 tensors from C:\Users\Iman Mirbioki\.lmstudio\models\aratan\DeepSeek-R1-32B-Uncensored\DeepSeek-R1-32B-Uncensored.gguf (version GGUF V3 (latest)) llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. llama_model_loader: - kv 0: general.architecture str = qwen2 llama_model_loader: - kv 1: general.type str = model llama_model_loader: - kv 2: general.name str = DeepSeek R1 Distill Qwen 32B Uncensored llama_model_loader: - kv 3: general.finetune str = Uncensored llama_model_loader: - kv 4: general.basename str = DeepSeek-R1-Distill-Qwen llama_model_loader: - kv 5: general.size_label str = 32B llama_model_loader: - kv 6: general.license str = mit llama_model_loader: - kv 7: general.base_model.count u32 = 1 llama_model_loader: - kv 8: general.base_model.0.name str = DeepSeek R1 Distill Qwen 32B llama_model_loader: - kv 9: general.base_model.0.organization str = Deepseek Ai llama_model_loader: - kv 10: general.base_model.0.repo_url str = https://huggingface.co/deepseek-ai/De... llama_model_loader: - kv 11: general.dataset.count u32 = 1 llama_model_loader: - kv 12: general.dataset.0.name str = Uncensor llama_model_loader: - kv 13: general.dataset.0.organization str = Guilherme34 llama_model_loader: - kv 14: general.dataset.0.repo_url str = https://huggingface.co/Guilherme34/un... llama_model_loader: - kv 15: general.tags arr[str,1] = ["generated_from_trainer"] llama_model_loader: - kv 16: qwen2.block_count u32 = 64 llama_model_loader: - kv 17: qwen2.context_length u32 = 131072 llama_model_loader: - kv 18: qwen2.embedding_length u32 = 5120 llama_model_loader: - kv 19: qwen2.feed_forward_length u32 = 27648 llama_model_loader: - kv 20: qwen2.attention.head_count u32 = 40 llama_model_loader: - kv 21: qwen2.attention.head_count_kv u32 = 8 llama_model_loader: - kv 22: qwen2.rope.freq_base f32 = 1000000.000000 llama_model_loader: - kv 23: qwen2.attention.layer_norm_rms_epsilon f32 = 0.000010 llama_model_loader: - kv 24: tokenizer.ggml.model str = gpt2 llama_model_loader: - kv 25: tokenizer.ggml.pre str = deepseek-r1-qwen [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 26: tokenizer.ggml.tokens arr[str,151665] = ["!", "\"", "#", "$", "%", "&", "'", ... [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 27: tokenizer.ggml.token_type arr[i32,151665] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... [2025-10-25 20:35:44][DEBUG] llama_model_loader: - kv 28: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... llama_model_loader: - kv 29: tokenizer.ggml.bos_token_id u32 = 151646 llama_model_loader: - kv 30: tokenizer.ggml.eos_token_id u32 = 151643 llama_model_loader: - kv 31: tokenizer.ggml.padding_token_id u32 = 151643 llama_model_loader: - kv 32: tokenizer.ggml.add_bos_token bool = true llama_model_loader: - kv 33: tokenizer.ggml.add_eos_token bool = false llama_model_loader: - kv 34: tokenizer.chat_template str = {% if not add_generation_prompt is de... llama_model_loader: - kv 35: general.quantization_version u32 = 2 llama_model_loader: - kv 36: general.file_type u32 = 30 llama_model_loader: - kv 37: general.url str = https://huggingface.co/mradermacher/D... llama_model_loader: - kv 38: mradermacher.quantize_version str = 2 llama_model_loader: - kv 39: mradermacher.quantized_by str = mradermacher llama_model_loader: - kv 40: mradermacher.quantized_at str = 2025-01-26T04:57:59+01:00 llama_model_loader: - kv 41: mradermacher.quantized_on str = marco llama_model_loader: - kv 42: general.source.url str = https://huggingface.co/nicoboss/DeepS... llama_model_loader: - kv 43: mradermacher.convert_type str = hf llama_model_loader: - type f32: 321 tensors llama_model_loader: - type q5_K: 72 tensors llama_model_loader: - type q6_K: 1 tensors llama_model_loader: - type iq4_xs: 377 tensors print_info: file format = GGUF V3 (latest) print_info: file type = IQ4_XS - 4.25 bpw print_info: file size = 16.63 GiB (4.36 BPW) [2025-10-25 20:35:44][DEBUG] load: special_eos_id is not in special_eog_ids - the tokenizer config may be incorrect load: printing all EOG tokens: load: - 151643 ('<|end▁of▁sentence|>') load: - 151662 ('<|fim_pad|>') load: - 151663 ('<|repo_name|>') load: - 151664 ('<|file_sep|>') [2025-10-25 20:35:44][DEBUG] load: special tokens cache size = 22 [2025-10-25 20:35:44][DEBUG] load: token to piece cache size = 0.9310 MB print_info: arch = qwen2 print_info: vocab_only = 0 print_info: n_ctx_train = 131072 print_info: n_embd = 5120 print_info: n_layer = 64 print_info: n_head = 40 print_info: n_head_kv = 8 print_info: n_rot = 128 print_info: n_swa = 0 print_info: is_swa_any = 0 print_info: n_embd_head_k = 128 print_info: n_embd_head_v = 128 print_info: n_gqa = 5 print_info: n_embd_k_gqa = 1024 print_info: n_embd_v_gqa = 1024 print_info: f_norm_eps = 0.0e+00 print_info: f_norm_rms_eps = 1.0e-05 print_info: f_clamp_kqv = 0.0e+00 print_info: f_max_alibi_bias = 0.0e+00 print_info: f_logit_scale = 0.0e+00 [2025-10-25 20:35:44][DEBUG] print_info: f_attn_scale = 0.0e+00 print_info: n_ff = 27648 print_info: n_expert = 0 print_info: n_expert_used = 0 print_info: causal attn = 1 print_info: pooling type = -1 print_info: rope type = 2 print_info: rope scaling = linear print_info: freq_base_train = 1000000.0 print_info: freq_scale_train = 1 print_info: n_ctx_orig_yarn = 131072 print_info: rope_finetuned = unknown print_info: model type = 32B print_info: model params = 32.76 B print_info: general.name = DeepSeek R1 Distill Qwen 32B Uncensored print_info: vocab type = BPE print_info: n_vocab = 151665 print_info: n_merges = 151387 print_info: BOS token = 151646 '<|begin▁of▁sentence|>' print_info: EOS token = 151643 '<|end▁of▁sentence|>' print_info: EOT token = 151643 '<|end▁of▁sentence|>' print_info: PAD token = 151643 '<|end▁of▁sentence|>' print_info: LF token = 198 'Ċ' print_info: FIM PRE token = 151659 '<|fim_prefix|>' print_info: FIM SUF token = 151661 '<|fim_suffix|>' print_info: FIM MID token = 151660 '<|fim_middle|>' print_info: FIM PAD token = 151662 '<|fim_pad|>' print_info: FIM REP token = 151663 '<|repo_name|>' print_info: FIM SEP token = 151664 '<|file_sep|>' print_info: EOG token = 151643 '<|end▁of▁sentence|>' print_info: EOG token = 151662 '<|fim_pad|>' print_info: EOG token = 151663 '<|repo_name|>' print_info: EOG token = 151664 '<|file_sep|>' print_info: max token length = 256 load_tensors: loading model tensors, this can take a while... (mmap = true) [2025-10-25 20:37:03][DEBUG] load_tensors: offloading 64 repeating layers to GPU load_tensors: offloading output layer to GPU load_tensors: offloaded 65/65 layers to GPU load_tensors: CUDA2 model buffer size = 5565.08 MiB load_tensors: CUDA1 model buffer size = 5453.34 MiB load_tensors: CUDA0 model buffer size = 5622.09 MiB load_tensors: CPU_Mapped model buffer size = 393.42 MiB [2025-10-25 20:37:19][DEBUG] llama_context: constructing llama_context llama_context: n_seq_max = 1 llama_context: n_ctx = 131072 llama_context: n_ctx_per_seq = 131072 llama_context: n_batch = 512 llama_context: n_ubatch = 512 llama_context: causal_attn = 1 llama_context: flash_attn = disabled llama_context: kv_unified = false llama_context: freq_base = 1000000.0 llama_context: freq_scale = 1 [2025-10-25 20:37:19][DEBUG] llama_context: CUDA_Host output buffer size = 0.58 MiBI've been going through all my logs, and it seems like the problem started when Win11 update 25H2 was released - but it's just a guess.
@pisarevel commented on GitHub (Oct 31, 2025):
I had the same problem.
While searching for a solution, I found a table of CUDA requirements for graphics card driver versions. I realized that my driver was outdated for the current CUDA version.
After updating the driver, everything begin to work as expected.
You can find the table here: https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html
@catclaw commented on GitHub (Nov 3, 2025):
I've updated to the latest CUDA for Windows 11, my drivers match, too, MSVC C++ runtime is the latest (I've tried "repair", too) but after upgrading to the latest Ollama version (0.12.9), it still refuses to detect my GPUs.
Version 0.11.11 detects the GPUs, but falls back to the CPUs when trying to load a model.
The log is included. (For ver. 0.12.9)
I've tried several other tools like LM Studio, and they detect the GPUs with no problems?
Please help - I really need to get Ollama up and working... 🙏🙏🙏
serve.log
@dhiltgen commented on GitHub (Nov 4, 2025):
@catclaw with the updated logging code we have a little more detail, but unfortunately still not an exact root cause on what's causing this problem on your system. None of the dynamic libraries will load properly
For a missing dependency, I would expect the error to say "The specified module could not be found" so the fact you're seeing "procedure" seems to imply the way we're linking our libraries is causing an incompatibility with something on your system.
Do you have any other software on your system that's doing preload/injection? Any unusual AV software?
Maybe we'll be able to get a little more insight if you can scan the libraries on your system? You might try https://github.com/lucasg/Dependencies/ which has a nice GUI that can show dependencies, and in this case, maybe it will help us understand what is causing all the DLLs listed above to not load?
@catclaw commented on GitHub (Nov 4, 2025):
Will do! Thank you so much!
I don't have anything unusual on my system. I've been using AVG AV for years (paid version).
I'll try Dependencies (I'm familiar with it), and see what I can find.
I'm also going to try uninstalling AVG Antivirus and see if that helps. (It's impossible to stop the AV services fully.)
However, I don't think the problem is with the AV.
Ollama 0.11.8 and 0.11.11 detect my GPUs, but they fall back to the CPU when I try to load a model. (No matter the size.)
I'm also going to try rebooting with all but MS services and see if it helps.
Meanwhile, I've added the serve.log for both 0.11.8 (which detects my GPUs) and 0.12.9 (which doesn't).
It's strange, as LM Studio works in a similar way, but it detects my GPUs and loads the models to VRAM. (I used Gollama -L to link my 35+ models to LM Studio.)
This all happened overnight after I updated Windows 11 (Enterprise) from 24H2 to 25H2, which I became aware of quickly, as I use OllamaVision and MagicPrompt in SwarmUI to rewrite and optimize my prompts. I noticed how my system became almost unresponsive and CPU usage went sky-high.
I'll be back with more information - and thank you so much again for taking your time to look at this.
I'm really grateful.
serve_0.12_9.log
serve_0_11_8.log
@catclaw commented on GitHub (Nov 5, 2025):
This is what I found out:
After reinstalling the correct Visual C++ Redistributable package (once again - I had to uninstall the previous version), and running sfc /scannow, dism.exe /online /cleanup-image /scanhealth, dism.exe /online /cleanup-image /restorehealth, and dism.exe /online /cleanup-image /startcomponentcleanup - the DLL-files were still missing.
If you look at the end of apisets, some APIs call non-existing DLLs. ([ ]).
I've included the dependencies chain in a log file.
log.txt
@catclaw commented on GitHub (Nov 5, 2025):
I can't explain this - but I started Windows 11 Sandbox and installed Ollama, quit the "ollama app.exe" (in the taskbar), and restarted it in cmd with ollama serve...
...And it detected all my 3 GPUs, no errors, no hassle...
The problem is that as soon as I close the Sandbox, everything will disappear, and the strange thing is that the Sandbox is a "copy" of my existing Windows 11 installation, running in a VM. (With GPU passthrough.) 😳😳😳😳
I feel.... lost!
@dhiltgen commented on GitHub (Nov 5, 2025):
@catclaw one point of clarification - we dynamically load the
ggml-*.dllfiles at runtime, so they won't show up as dependencies of main ollama.exe binary. You'll want to scan any of those files mentioned above withThe specified procedure could not be founderrors.@catclaw commented on GitHub (Nov 7, 2025):
Sorry, guys and gals...
I've installed the latest Ollama version, but it still refuses to detect my GPUs.
I've uninstalled CUDA 12.9 and 12.8 and only kept 13.0.88
Visual C++ Red. x64 2015-2022 is installed (actually ALL of them, down to 2005)
I have tried every solution I can think of, and Dependencies is not giving me any clues besides these files, which dependencies.exe -modules shows as missing in CLI and are marked as orange in the GUI (missing):
(Which also the command line version shows as not found.)
When I install and run Ollama in the Sandbox, it works just fine, but I noticed something important - I thought the Sandbox VM was a copy of my existing Windows 11 installation - which it is not. Winver reports 24H2, while I have 25H2 on my computer. (An no way to go back, without having to reinstall several hundreds programs - some very large and time-consuming like the Adobe family.
This is the debug output from Ollama in Sandbox: (Debug = 2)
It detects my GPUs and loads models to VRAM.
This is the serve.log:
serve.log
However, when I install and run Ollama on my computer (outside Sandbox):
This is the output: (Debug = 2)
And this is the serve.log:
serve.log
I've tried everything I could think of - I've added Ollama LIB folders to PATH, I've copied the DLL files reported missing with Dependencies.exe/DependenciesGUI.exe from other x64 software to the LIB folder...
I don't know what to do anymore? 😖😭😭😭
@Anyeos commented on GitHub (Nov 12, 2025):
Update:
I missed to build the "build" subdirectory part. I issued cmake -B build and cmake --build build and now it is working.
Maybe @catclaw you are forgeting the libraries folder or you are putting them on another place. It must be in "./lib/ollama" path (as indicated by the documentation). That will provide you GPU and CPU acceleration.
My original post:
Same happens to me and I build it from source. The old version detect my GPU automatically.
And I am under Ubuntu 24.04
It can be a bug in source code or some dependency that directly is skipped.
I use go run . serve
I tried go build again but nothing changed.
Call my attention that it says "total vram"="0 B" <- Why? What is happening there?
@VSOL3008 commented on GitHub (Nov 12, 2025):
I had the same problem,
for some reason Ollama 0.12.4 have some bug with the environment variable OLLAMA_LLM_LIBRARY=CUDA, basically its causing an error preventing the detection of the multiple GPUs. But this variable was working in older versions of Ollama, now in new version it probably have different behavior.
So what helped for me, was just removing this OLLAMA_LLM_LIBRARY variable.
I hope it helps.
@dhiltgen commented on GitHub (Nov 12, 2025):
@catclaw so far, I haven't seen anyone else report the error
The specified procedure could not be foundcausing all of the libraries not to load. Given you have development tools installed on your system, perhaps you could try to build from source and see if the problem disappears? My best guess at this point is your system DLLs are somehow different than the standard ones we link against in our CI.@catclaw commented on GitHub (Nov 13, 2025):
I've created a clean install of Windows 11 (Enterprise) ver. 24H2, on a new computer, and moved my GPUs over to the new motherboard.
Ollama worked fine - till I updated to 25H2, and the same problem occurred once again.
I've been diving into some MS technical data, and Nvidia developers forum - I see "CUDA context issue introduced in the 25H2 driver layer" repeating everywhere.
I'm not entirely sure what it means, but it seems like MS has made some low-layer changes to how Windows handles GPU/CUDA drivers.
I'm currently running Ollama (latest version) on WSL2 (Ubuntu), with the following in my .bashrc
export OLLAMA_MODELS=/mnt/h/LLM
export CUDA_VISIBLE_DEVICES=0,1,2 #3 is a PNY Nvidia Tesla A100 80GB, which I use in ComfyUI, while using Ollama for writing my prompts.
export OLLAMA_SCHED_SPREAD=1
export OLLAMA_NUM_PARALLEL=3
export OLLAMA_MAX_QUEUE=3
export OLLAMA_NO_MMAP=1
export OLLAMA_LOAD_TIMEOUT=10m
export OLLAMA_DEBUG=2
However, I can't use localhost on Windows anymore to reach Ollama, and have to type my LAN IP.
If anyone is going to create a similar setup, make sure to include OLLAMA_NO_MMAP=1 as ext4 fs is faster than NTFS, and without it, you might run into trouble loading larger models.
That said...
I hope you find the solution to this problem, as WSL uses a lot of my system resources and RAM, slowing down video generation for example.
@dhiltgen commented on GitHub (Nov 13, 2025):
At least that narrows it down a bit.
I don't think it's related to CUDA, since the logs you shared showed even the CPU libraries failed to load. It appears there's something about the way we're building the windows binaries which makes them incompatible with Windows Enterprise 25H2. We have Windows Pro 25H2 test systems that all seem to work OK. In our CI, we build on Windows Server github runners, and the compiler used for the CPU libraries is
C:/Program Files/Microsoft Visual Studio/2022/Enterprise/VC/Tools/Llvm/x64/bin/clang.exewhich I would have thought would produce compatible binaries.I still think it would be interesting to see if locally built binaries work on your system if you have a chance to try that.
@dhiltgen commented on GitHub (Nov 14, 2025):
I just spun up a Windows 11 Enterprise 25H2 VM on Azure, and it is able to load the libraries without error. @catclaw could there be some other piece of software intervening, like AV software on your system? If you have 3rd party AV software, could you try disabling that temporarily and see if the libraries load correctly? On my test VM, I see MsMpEng.exe chewing up a ton of CPU cycles as we're doing our initial bootstrapping, but it appears to be letting the libraries load. All the DLLs look correctly signed by our signing key.
If you are in a corporate environment and can't disable the AV software, can you check with your IT to see if there are logs showing it is blocking the libraries from loading?
@catclaw commented on GitHub (Nov 15, 2025):
Hi!

Thank you for taking the time to test it on a VM. I really appreciate it.
Perhaps the updates have something to do with the libs not being loaded?
My firmware was also updated from 2006 to 2009.
No, I'm not in a corporate environment - it's my own personal computer. However, I had some trouble using PNY Nvidia Tesla A100 80GB on the Pro-edition (I could either use my 3 RTX GPUs or A100 - not all 4 at the same time), so I contacted both PNY and Microsoft support. They both suggested that I use a server edition of Windows, or the Enterprise edition, which I did, and now I can use all 4 GPUs at the same time. (I never got an explanation for why it would matter?)
Yes, I installed a bare Windows 11 Enterprise edition, without any AV software, or anything else.
The Ollama setup EXE file installed vc_redist.x64.exe, but nothing else.
Despite this, Ollama fails to detect my GPUs.
However, I installed Windows 11 Enterprise 24H2 (with all updates, except 25H2) - and Ollama worked just fine; it detected my GPUs and loaded models to VRAM.
(Yes, MsMpEng.exe spins up pretty good during the initial loading of a model, but it's a part of Windows Security - so I guess it's screening for something.)
As I mentioned earlier, some people are reporting "CUDA context issue introduced in the 25H2 driver layer" related to software that uses CUDA 13.0.x and Win11 25H2 - but I have no idea what it's about?
I'm going to try WinDbg and step through the libs loading and see if I can find anything.
@dhiltgen commented on GitHub (Nov 15, 2025):
I was hoping I could repro and then I was going to try
gflags -i ollama.exe +slsand try running it under a debugger, which apparently will show more detail on LoadLibrary failures. One wrinkle to be aware of is we use subprocesses, so you'll have to get the debugger to follow child processes. (The initialollama servespawns another copy ofollamawhich is what calls LoadLibrary). You can look at the logs to see the commands it's running and try running the runner subprocess manually.@dhiltgen commented on GitHub (Nov 17, 2025):
@catclaw based on this comment a possible explanation is another version of
ggml-base.dllsomeplace on your system that's conflicting. Can you look around to see if you have another copy of this anywhere in the PATH? Try removing it and see if that resolves the problem.@catclaw commented on GitHub (Nov 22, 2025):
Well... The new 0.13.0 version finally detects my GPUs, but still loads models into RAM and uses the CPU. (~70-80% - please se attached image at the end of this post.)
(No, there are no other ggml-base.dll in PATH; ComfyUI, SwarmUI, Whisper, etc. are all in their own folders outside the path.)
Also, when I run ollama serve, it seems like my Nvidia drivers crash and restart, as all my 4 monitors go black, and the same thing happens as when I restart the drivers manually by pressing CTRL+SHIFT+WIN+B.
@catclaw commented on GitHub (Nov 23, 2025):
I've finally fixed it!!
I copied
ggml-base.dllfrom Ollama installation folder toC:\Windowsand now all the models are loading to VRAM of my GPUs!@dhiltgen commented on GitHub (Dec 1, 2025):
@catclaw that "fix" will likely break on the next update. You should remove ggml-base.dll from any system-wide locations so that the version that is found is within the Ollama install directory, as that file changes on every release, and is not binary compatible between releases.
@mwprado commented on GitHub (Dec 15, 2025):
I Use fedora 43, cuda 12 and tesla P4 the last after 0.12.5, it stops GPU use.