mirror of
https://github.com/ollama/ollama.git
synced 2026-05-06 16:11:34 -05:00
Closed
opened 2026-04-22 17:55:44 -05:00 by GiteaMirror
·
24 comments
No Branch/Tag Specified
main
dhiltgen/ci
parth-launch-plan-gating
hoyyeva/anthropic-reference-images-path
parth-anthropic-reference-images-path
brucemacd/download-before-remove
hoyyeva/editor-config-repair
parth-mlx-decode-checkpoints
parth-launch-codex-app
hoyyeva/fix-codex-model-metadata-warning
hoyyeva/qwen
parth/hide-claude-desktop-till-release
hoyyeva/opencode-image-modality
parth-add-claude-code-autoinstall
release_v0.22.0
pdevine/manifest-list
codex/fix-codex-model-metadata-warning
pdevine/addressable-manifest
brucemacd/launch-fetch-reccomended
jmorganca/llama-compat
launch-copilot-cli
hoyyeva/opencode-thinking
release_v0.20.7
parth-auto-save-backup
parth-test
jmorganca/gemma4-audio-replacements
fix-manifest-digest-on-pull
hoyyeva/vscode-improve
brucemacd/install-server-wait
parth/update-claude-docs
brucemac/start-ap-install
pdevine/mlx-update
pdevine/qwen35_vision
drifkin/api-show-fallback
mintlify/image-generation-1773352582
hoyyeva/server-context-length-local-config
jmorganca/faster-reptition-penalties
jmorganca/convert-nemotron
parth-pi-thinking
pdevine/sampling-penalties
jmorganca/fix-create-quantization-memory
dongchen/resumable_transfer_fix
pdevine/sampling-cache-error
jessegross/mlx-usage
hoyyeva/openclaw-config
hoyyeva/app-html
pdevine/qwen3next
brucemacd/sign-sh-install
brucemacd/tui-update
brucemacd/usage-api
jmorganca/launch-empty
fix-app-dist-embed
mxyng/mlx-compile
mxyng/mlx-quant
mxyng/mlx-glm4.7
mxyng/mlx
brucemacd/simplify-model-picker
jmorganca/qwen3-concurrent
fix-glm-4.7-flash-mla-config
drifkin/qwen3-coder-opening-tag
brucemacd/usage-cli
fix-cuda12-fattn-shmem
ollama-imagegen-docs
parth/fix-multiline-inputs
brucemacd/config-docs
mxyng/model-files
mxyng/simple-execute
fix-imagegen-ollama-models
mxyng/async-upload
jmorganca/lazy-no-dtype-changes
imagegen-auto-detect-create
parth/decrease-concurrent-download-hf
fix-mlx-quantize-init
jmorganca/x-cleanup
usage
imagegen-readme
jmorganca/glm-image
mlx-gpu-cd
jmorganca/imagegen-modelfile
parth/agent-skills
parth/agent-allowlist
parth/signed-in-offline
parth/agents
parth/fix-context-chopping
improve-cloud-flow
parth/add-models-websearch
parth/prompt-renderer-mcp
jmorganca/native-settings
jmorganca/download-stream-hash
jmorganca/client2-rebased
brucemacd/oai-chat-req-multipart
jessegross/multi_chunk_reserve
grace/additional-omit-empty
grace/mistral-3-large
mxyng/tokenizer2
mxyng/tokenizer
jessegross/flash
hoyyeva/windows-nacked-app
mxyng/cleanup-attention
grace/deepseek-parser
hoyyeva/remember-unsent-prompt
parth/add-lfs-pointer-error-conversion
parth/olmo2-test2
hoyyeva/ollama-launchagent-plist
nicole/olmo-model
parth/olmo-test
mxyng/remove-embedded
parth/render-template
jmorganca/intellect-3
parth/remove-prealloc-linter
jmorganca/cmd-eval
nicole/nomic-embed-text-fix
mxyng/lint-2
hoyyeva/add-gemini-3-pro-preview
hoyyeva/load-model-list
mxyng/expand-path
mxyng/environ-2
hoyyeva/deeplink-json-encoding
parth/improve-tool-calling-tests
hoyyeva/conversation
hoyyeva/assistant-edit-response
hoyyeva/thinking
origin/brucemacd/invalid-char-i-err
parth/improve-tool-calling
jmorganca/required-omitempty
grace/qwen3-vl-tests
mxyng/iter-client
parth/docs-readme
nicole/embed-test
pdevine/integration-benchstat
parth/remove-generate-cmd
parth/add-toolcall-id
mxyng/server-tests
jmorganca/glm-4.6
jmorganca/gin-h-compat
drifkin/stable-tool-args
pdevine/qwen3-more-thinking
parth/add-websearch-client
nicole/websearch_local
jmorganca/qwen3-coder-updates
grace/deepseek-v3-migration-tests
mxyng/fix-create
jmorganca/cloud-errors
pdevine/parser-tidy
revert-12233-parth/simplify-entrypoints-runner
parth/enable-so-gpt-oss
brucemacd/qwen3vl
jmorganca/readme-simplify
parth/gpt-oss-structured-outputs
revert-12039-jmorganca/tools-braces
mxyng/embeddings
mxyng/gguf
mxyng/benchmark
mxyng/types-null
parth/move-parsing
mxyng/gemma2
jmorganca/docs
mxyng/16-bit
mxyng/create-stdin
pdevine/authorizedkeys
mxyng/quant
parth/opt-in-error-context-window
brucemacd/cache-models
brucemacd/runner-completion
jmorganca/llama-update-6
brucemacd/benchmark-list
brucemacd/partial-read-caps
parth/deepseek-r1-tools
mxyng/omit-array
parth/tool-prefix-temp
brucemacd/runner-test
jmorganca/qwen25vl
brucemacd/model-forward-test-ext
parth/python-function-parsing
jmorganca/cuda-compression-none
drifkin/num-parallel
drifkin/chat-truncation-fix
jmorganca/sync
parth/python-tools-calling
drifkin/array-head-count
brucemacd/create-no-loop
parth/server-enable-content-stream-with-tools
qwen25omni
mxyng/v3
brucemacd/ropeconfig
jmorganca/silence-tokenizer
parth/sample-so-test
parth/sampling-structured-outputs
brucemacd/doc-go-engine
parth/constrained-sampling-json
jmorganca/mistral-wip
brucemacd/mistral-small-convert
parth/sample-unmarshal-json-for-params
brucemacd/jomorganca/mistral
pdevine/bfloat16
jmorganca/mistral
brucemacd/mistral
pdevine/logging
parth/sample-correctness-fix
parth/sample-fix-sorting
jmorgan/sample-fix-sorting-extras
jmorganca/temp-0-images
brucemacd/parallel-embed-models
brucemacd/shim-grammar
jmorganca/fix-gguf-error
bmizerany/nameswork
jmorganca/faster-releases
bmizerany/validatenames
brucemacd/err-no-vocab
brucemacd/rope-config
brucemacd/err-hint
brucemacd/qwen2_5
brucemacd/logprobs
brucemacd/new_runner_graph_bench
progress-flicker
brucemacd/forward-test
brucemacd/go_qwen2
pdevine/gemma2
jmorganca/add-missing-symlink-eval
mxyng/next-debug
parth/set-context-size-openai
brucemacd/next-bpe-bench
brucemacd/next-bpe-test
brucemacd/new_runner_e2e
brucemacd/new_runner_qwen2
pdevine/convert-cohere2
brucemacd/convert-cli
parth/log-probs
mxyng/next-mlx
mxyng/cmd-history
parth/templating
parth/tokenize-detokenize
brucemacd/check-key-register
bmizerany/grammar
jmorganca/vendor-081b29bd
mxyng/func-checks
jmorganca/fix-null-format
parth/fix-default-to-warn-json
jmorganca/qwen2vl
jmorganca/no-concat
parth/cmd-cleanup-SO
brucemacd/check-key-register-structured-err
parth/openai-stream-usage
parth/fix-referencing-so
stream-tools-stop
jmorganca/degin-1
brucemacd/install-path-clean
brucemacd/push-name-validation
brucemacd/browser-key-register
jmorganca/openai-fix-first-message
jmorganca/fix-proxy
jessegross/sample
parth/disallow-streaming-tools
dhiltgen/remove_submodule
jmorganca/ga
jmorganca/mllama
pdevine/newlines
pdevine/geems-2b
jmorganca/llama-bump
mxyng/modelname-7
mxyng/gin-slog
mxyng/modelname-6
jyan/convert-prog
jyan/quant5
paligemma-support
pdevine/import-docs
jmorganca/openai-context
jyan/paligemma
jyan/p2
jyan/palitest
bmizerany/embedspeedup
jmorganca/llama-vit
brucemacd/allow-ollama
royh/ep-methods
royh/whisper
mxyng/api-models
mxyng/fix-memory
jyan/q4_4/8
jyan/ollama-v
royh/stream-tools
roy-embed-parallel
bmizerany/hrm
revert-5963-revert-5924-mxyng/llama3.1-rope
royh/embed-viz
jyan/local2
jyan/auth
jyan/local
jyan/parse-temp
jmorganca/template-mistral
jyan/reord-g
royh-openai-suffixdocs
royh-imgembed
royh-embed-parallel
jyan/quant4
royh-precision
jyan/progress
pdevine/fix-template
jyan/quant3
pdevine/ggla
mxyng/update-registry-domain
jmorganca/ggml-static
mxyng/create-context
jyan/v0.146
mxyng/layers-from-files
build_dist
bmizerany/noseek
royh-ls
royh-name
timeout
mxyng/server-timestamp
bmizerany/nosillyggufslurps
royh-params
jmorganca/llama-cpp-7c26775
royh-openai-delete
royh-show-rigid
jmorganca/enable-fa
jmorganca/no-error-template
jyan/format
royh-testdelete
bmizerany/fastverify
language_support
pdevine/ps-glitches
brucemacd/tokenize
bruce/iq-quants
bmizerany/filepathwithcoloninhost
mxyng/split-bin
bmizerany/client-registry
jmorganca/if-none-match
native
jmorganca/native
jmorganca/batch-embeddings
jmorganca/initcmake
jmorganca/mm
pdevine/showggmlinfo
modenameenforcealphanum
bmizerany/modenameenforcealphanum
jmorganca/done-reason
jmorganca/llama-cpp-8960fe8
ollama.com
bmizerany/filepathnobuild
bmizerany/types/model/defaultfix
rmdisplaylong
nogogen
bmizerany/x
modelfile-readme
bmizerany/replacecolon
jmorganca/limit
jmorganca/execstack
jmorganca/replace-assets
mxyng/tune-concurrency
jmorganca/testing
whitespace-detection
jmorganca/options
upgrade-all
scratch
cuda-search
mattw/airenamer
mattw/allmodelsonhuggingface
mattw/quantcontext
mattw/whatneedstorun
brucemacd/llama-mem-calc
mattw/faq-context
mattw/communitylinks
mattw/noprune
mattw/python-functioncalling
rename
mxyng/install
pulse
remove-first
editor
mattw/selfqueryingretrieval
cgo
mattw/howtoquant
api
matt/streamingapi
format-config
mxyng/extra-args
shell
update-nous-hermes
cp-model
upload-progress
fix-unknown-model
fix-model-names
delete-fix
insecure-registry
ls
deletemodels
progressbar
readme-updates
license-layers
skip-list
list-models
modelpath
matt/examplemodelfiles
distribution
go-opts
v0.23.1
v0.23.1-rc0
v0.23.0
v0.23.0-rc0
v0.22.1
v0.22.1-rc1
v0.22.1-rc0
v0.22.0
v0.22.0-rc1
v0.21.3-rc0
v0.21.2-rc1
v0.21.2
v0.21.2-rc0
v0.21.1
v0.21.1-rc1
v0.21.1-rc0
v0.21.0
v0.21.0-rc1
v0.21.0-rc0
v0.20.8-rc0
v0.20.7
v0.20.7-rc1
v0.20.7-rc0
v0.20.6
v0.20.6-rc1
v0.20.6-rc0
v0.20.5
v0.20.5-rc2
v0.20.5-rc1
v0.20.5-rc0
v0.20.4
v0.20.4-rc2
v0.20.4-rc1
v0.20.4-rc0
v0.20.3
v0.20.3-rc0
v0.20.2
v0.20.1
v0.20.1-rc2
v0.20.1-rc1
v0.20.1-rc0
v0.20.0
v0.20.0-rc1
v0.20.0-rc0
v0.19.0
v0.19.0-rc2
v0.19.0-rc1
v0.19.0-rc0
v0.18.4-rc1
v0.18.4-rc0
v0.18.3
v0.18.3-rc2
v0.18.3-rc1
v0.18.3-rc0
v0.18.2
v0.18.2-rc1
v0.18.2-rc0
v0.18.1
v0.18.1-rc1
v0.18.1-rc0
v0.18.0
v0.18.0-rc2
v0.18.0-rc1
v0.18.0-rc0
v0.17.8-rc4
v0.17.8-rc3
v0.17.8-rc2
v0.17.8-rc1
v0.17.8-rc0
v0.17.7
v0.17.7-rc2
v0.17.7-rc1
v0.17.7-rc0
v0.17.6
v0.17.5
v0.17.4
v0.17.3
v0.17.2
v0.17.1
v0.17.1-rc2
v0.17.1-rc1
v0.17.1-rc0
v0.17.0
v0.17.0-rc2
v0.17.0-rc1
v0.17.0-rc0
v0.16.3
v0.16.3-rc2
v0.16.3-rc1
v0.16.3-rc0
v0.16.2
v0.16.2-rc0
v0.16.1
v0.16.0
v0.16.0-rc2
v0.16.0-rc0
v0.16.0-rc1
v0.15.6
v0.15.5
v0.15.5-rc5
v0.15.5-rc4
v0.15.5-rc3
v0.15.5-rc2
v0.15.5-rc1
v0.15.5-rc0
v0.15.4
v0.15.3
v0.15.2
v0.15.1
v0.15.1-rc1
v0.15.1-rc0
v0.15.0-rc6
v0.15.0
v0.15.0-rc5
v0.15.0-rc4
v0.15.0-rc3
v0.15.0-rc2
v0.15.0-rc1
v0.15.0-rc0
v0.14.3
v0.14.3-rc3
v0.14.3-rc2
v0.14.3-rc1
v0.14.3-rc0
v0.14.2
v0.14.2-rc1
v0.14.2-rc0
v0.14.1
v0.14.0-rc11
v0.14.0
v0.14.0-rc10
v0.14.0-rc9
v0.14.0-rc8
v0.14.0-rc7
v0.14.0-rc6
v0.14.0-rc5
v0.14.0-rc4
v0.14.0-rc3
v0.14.0-rc2
v0.14.0-rc1
v0.14.0-rc0
v0.13.5
v0.13.5-rc1
v0.13.5-rc0
v0.13.4-rc2
v0.13.4
v0.13.4-rc1
v0.13.4-rc0
v0.13.3
v0.13.3-rc1
v0.13.3-rc0
v0.13.2
v0.13.2-rc2
v0.13.2-rc1
v0.13.2-rc0
v0.13.1
v0.13.1-rc2
v0.13.1-rc1
v0.13.1-rc0
v0.13.0
v0.13.0-rc0
v0.12.11
v0.12.11-rc1
v0.12.11-rc0
v0.12.10
v0.12.10-rc1
v0.12.10-rc0
v0.12.9-rc0
v0.12.9
v0.12.8
v0.12.8-rc0
v0.12.7
v0.12.7-rc1
v0.12.7-rc0
v0.12.7-citest0
v0.12.6
v0.12.6-rc1
v0.12.6-rc0
v0.12.5
v0.12.5-rc0
v0.12.4
v0.12.4-rc7
v0.12.4-rc6
v0.12.4-rc5
v0.12.4-rc4
v0.12.4-rc3
v0.12.4-rc2
v0.12.4-rc1
v0.12.4-rc0
v0.12.3
v0.12.2
v0.12.2-rc0
v0.12.1
v0.12.1-rc1
v0.12.1-rc2
v0.12.1-rc0
v0.12.0
v0.12.0-rc1
v0.12.0-rc0
v0.11.11
v0.11.11-rc3
v0.11.11-rc2
v0.11.11-rc1
v0.11.11-rc0
v0.11.10
v0.11.9
v0.11.9-rc0
v0.11.8
v0.11.8-rc0
v0.11.7-rc1
v0.11.7-rc0
v0.11.7
v0.11.6
v0.11.6-rc0
v0.11.5-rc4
v0.11.5-rc3
v0.11.5
v0.11.5-rc5
v0.11.5-rc2
v0.11.5-rc1
v0.11.5-rc0
v0.11.4
v0.11.4-rc0
v0.11.3
v0.11.3-rc0
v0.11.2
v0.11.1
v0.11.0-rc0
v0.11.0-rc1
v0.11.0-rc2
v0.11.0
v0.10.2-int1
v0.10.1
v0.10.0
v0.10.0-rc4
v0.10.0-rc3
v0.10.0-rc2
v0.10.0-rc1
v0.10.0-rc0
v0.9.7-rc1
v0.9.7-rc0
v0.9.6
v0.9.6-rc0
v0.9.6-ci0
v0.9.5
v0.9.4-rc5
v0.9.4-rc6
v0.9.4
v0.9.4-rc3
v0.9.4-rc4
v0.9.4-rc1
v0.9.4-rc2
v0.9.4-rc0
v0.9.3
v0.9.3-rc5
v0.9.4-citest0
v0.9.3-rc4
v0.9.3-rc3
v0.9.3-rc2
v0.9.3-rc1
v0.9.3-rc0
v0.9.2
v0.9.1
v0.9.1-rc1
v0.9.1-rc0
v0.9.1-ci1
v0.9.1-ci0
v0.9.0
v0.9.0-rc0
v0.8.0
v0.8.0-rc0
v0.7.1-rc2
v0.7.1
v0.7.1-rc1
v0.7.1-rc0
v0.7.0
v0.7.0-rc1
v0.7.0-rc0
v0.6.9-rc0
v0.6.8
v0.6.8-rc0
v0.6.7
v0.6.7-rc2
v0.6.7-rc1
v0.6.7-rc0
v0.6.6
v0.6.6-rc2
v0.6.6-rc1
v0.6.6-rc0
v0.6.5-rc1
v0.6.5
v0.6.5-rc0
v0.6.4-rc0
v0.6.4
v0.6.3-rc1
v0.6.3
v0.6.3-rc0
v0.6.2
v0.6.2-rc0
v0.6.1
v0.6.1-rc0
v0.6.0-rc0
v0.6.0
v0.5.14-rc0
v0.5.13
v0.5.13-rc6
v0.5.13-rc5
v0.5.13-rc4
v0.5.13-rc3
v0.5.13-rc2
v0.5.13-rc1
v0.5.13-rc0
v0.5.12
v0.5.12-rc1
v0.5.12-rc0
v0.5.11
v0.5.10
v0.5.9
v0.5.9-rc0
v0.5.8-rc13
v0.5.8
v0.5.8-rc12
v0.5.8-rc11
v0.5.8-rc10
v0.5.8-rc9
v0.5.8-rc8
v0.5.8-rc7
v0.5.8-rc6
v0.5.8-rc5
v0.5.8-rc4
v0.5.8-rc3
v0.5.8-rc2
v0.5.8-rc1
v0.5.8-rc0
v0.5.7
v0.5.6
v0.5.5
v0.5.5-rc0
v0.5.4
v0.5.3
v0.5.3-rc0
v0.5.2
v0.5.2-rc3
v0.5.2-rc2
v0.5.2-rc1
v0.5.2-rc0
v0.5.1
v0.5.0
v0.5.0-rc1
v0.4.8-rc0
v0.4.7
v0.4.6
v0.4.5
v0.4.4
v0.4.3
v0.4.3-rc0
v0.4.2
v0.4.2-rc1
v0.4.2-rc0
v0.4.1
v0.4.1-rc0
v0.4.0
v0.4.0-rc8
v0.4.0-rc7
v0.4.0-rc6
v0.4.0-rc5
v0.4.0-rc4
v0.4.0-rc3
v0.4.0-rc2
v0.4.0-rc1
v0.4.0-rc0
v0.4.0-ci3
v0.3.14
v0.3.14-rc0
v0.3.13
v0.3.12
v0.3.12-rc5
v0.3.12-rc4
v0.3.12-rc3
v0.3.12-rc2
v0.3.12-rc1
v0.3.11
v0.3.11-rc4
v0.3.11-rc3
v0.3.11-rc2
v0.3.11-rc1
v0.3.10
v0.3.10-rc1
v0.3.9
v0.3.8
v0.3.7
v0.3.7-rc6
v0.3.7-rc5
v0.3.7-rc4
v0.3.7-rc3
v0.3.7-rc2
v0.3.7-rc1
v0.3.6
v0.3.5
v0.3.4
v0.3.3
v0.3.2
v0.3.1
v0.3.0
v0.2.8
v0.2.8-rc2
v0.2.8-rc1
v0.2.7
v0.2.6
v0.2.5
v0.2.4
v0.2.3
v0.2.2
v0.2.2-rc2
v0.2.2-rc1
v0.2.1
v0.2.0
v0.1.49-rc14
v0.1.49-rc13
v0.1.49-rc12
v0.1.49-rc11
v0.1.49-rc10
v0.1.49-rc9
v0.1.49-rc8
v0.1.49-rc7
v0.1.49-rc6
v0.1.49-rc4
v0.1.49-rc5
v0.1.49-rc3
v0.1.49-rc2
v0.1.49-rc1
v0.1.48
v0.1.47
v0.1.46
v0.1.45-rc5
v0.1.45
v0.1.45-rc4
v0.1.45-rc3
v0.1.45-rc2
v0.1.45-rc1
v0.1.44
v0.1.43
v0.1.42
v0.1.41
v0.1.40
v0.1.40-rc1
v0.1.39
v0.1.39-rc2
v0.1.39-rc1
v0.1.38
v0.1.37
v0.1.36
v0.1.35
v0.1.35-rc1
v0.1.34
v0.1.34-rc1
v0.1.33
v0.1.33-rc7
v0.1.33-rc6
v0.1.33-rc5
v0.1.33-rc4
v0.1.33-rc3
v0.1.33-rc2
v0.1.33-rc1
v0.1.32
v0.1.32-rc2
v0.1.32-rc1
v0.1.31
v0.1.30
v0.1.29
v0.1.28
v0.1.27
v0.1.26
v0.1.25
v0.1.24
v0.1.23
v0.1.22
v0.1.21
v0.1.20
v0.1.19
v0.1.18
v0.1.17
v0.1.16
v0.1.15
v0.1.14
v0.1.13
v0.1.12
v0.1.11
v0.1.10
v0.1.9
v0.1.8
v0.1.7
v0.1.6
v0.1.5
v0.1.4
v0.1.3
v0.1.2
v0.1.1
v0.1.0
v0.0.21
v0.0.20
v0.0.19
v0.0.18
v0.0.17
v0.0.16
v0.0.15
v0.0.14
v0.0.13
v0.0.12
v0.0.11
v0.0.10
v0.0.9
v0.0.8
v0.0.7
v0.0.6
v0.0.5
v0.0.4
v0.0.3
v0.0.2
v0.0.1
Labels
Clear labels
amd
api
app
bug
build
cli
cloud
compatibility
context-length
create
docker
documentation
embeddings
feature request
feedback wanted
good first issue
gpt-oss
gpu
harmony
help wanted
image
install
intel
js
launch
linux
macos
memory
mlx
model
needs more info
networking
nvidia
ollama.com
performance
pull-request
python
question
registry
rendering
thinking
tools
top
vulkan
windows
wsl
Mirrored from GitHub Pull Request
Milestone
No items
No Milestone
Projects
Clear projects
No project
No Assignees
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: github-starred/ollama#34406
Reference in New Issue
Block a user
Blocking a user prevents them from interacting with repositories, such as opening or commenting on pull requests or issues. Learn more about blocking a user.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @ghost on GitHub (Nov 11, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/13049
What is the issue?
I would really like to test minimax-m2. Am getting an unknown model architecture: 'minimax-m2' error.
Relevant log output
OS
Windows
GPU
Nvidia
CPU
Intel
Ollama version
0.12.10
@rick-github commented on GitHub (Nov 11, 2025):
minimax-m2 has been added to llama.cpp so the next vendor sync will pull that in via the llama.cpp engine. It's also possible that the ollama team will add support for the native engine, but that will take longer.
@BarachielFallen commented on GitHub (Nov 16, 2025):
Also seeing this failure with the Ollama:ROCM branch
@rick-github commented on GitHub (Nov 16, 2025):
Vendor sync needs to move past b6906 before ollama can use the llama.cpp engine for minimax-m2.
@rick-github commented on GitHub (Nov 28, 2025):
https://github.com/ollama/ollama/pull/12992
@gkjpettet commented on GitHub (Nov 28, 2025):
+1 for this. Dosappointing that it’s on Ollama cloud but not available to run locally. It’s currently state of the art for local models and is possible to run on a M3 Ultra with LM Studio.
@rick-github commented on GitHub (Dec 10, 2025):
The model supports tool calling but the format is too complex to encode in a template, so that will need a renderer/parser like qwen3-coder.
@BarachielFallen commented on GitHub (Dec 10, 2025):
There is no way you can backdate this to version 12.11 is there? That is the last version of ollama that recognizes gtt size for vram on an amd strix halo igpu
@gkjpettet commented on GitHub (Dec 10, 2025):
Getting tool use with Minimax M2 would make my Christmas! Benchmarks suggest this is one of the most powerful agentic models you can run on a Mac Studio.
@rick-github commented on GitHub (Dec 10, 2025):
I have an evo-x2 which I believe has the same chipset as the strix halo and it runs miniimax-m2 on ollama 0.13.2, although not very quickly.
The minimax-m2 tool calling format is too complex for the template but I'm experimenting with adding tool calls using a more generic format.
@BarachielFallen commented on GitHub (Dec 10, 2025):
what version of linux and rocm did you use to get it to recognize gtt
values instead of bios defined vram?
On Wed, Dec 10, 2025, 7:57 AM frob @.***> wrote:
@BarachielFallen commented on GitHub (Dec 10, 2025):
Also have you already tried one of the smaller REAPs to fit the entire model into vram instead of the cpu offload your metrics are showing for the Q4 version to see what speed increase is gained for full gpu offload?
@rick-github commented on GitHub (Dec 10, 2025):
@BarachielFallen commented on GitHub (Dec 10, 2025):
Could you please tell me your linux kernel, os, and ROCM versions in order to get ollama:latest to recognize the full 128gb as vram through gttsize grub settings instead of the bios set vram limit of 96gb? Or alternately amdgttsize if that is the version that works.
@BarachielFallen commented on GitHub (Dec 10, 2025):
I am loving those REAP tk/s
@rick-github commented on GitHub (Dec 10, 2025):
The machine is currently configured with 96GB VRAM for ROCm. Using Vulkan increases the available resources but there have been issues and it doesn't seem production-ready just yet, so for the moment sticking with ROCm.
@BarachielFallen commented on GitHub (Dec 10, 2025):
I am able to utilize the full 128gb as VRAM using ollama version 12.11 on
ROCM 7.0.2 linux kernel 6.15 Ubuntu 24.04.3 LTS, using ROCM and not Vulkan backend. I was hoping you had a
solution for the newer version of Ollama since I cant use the newer models
like minimax and QWEN Next past version 12 with full vram. GTTSize /
AMDGTTsize is completely broken in the newer builds, it only reports the
1gb of vram which isnt even what my bios is set to which is 512mb. I even
tried the new Ollama engine with no luck.
On Wed, Dec 10, 2025, 11:53 AM frob @.***> wrote:
@BarachielFallen commented on GitHub (Dec 10, 2025):
"time=2025-12-02T04:37:28.716Z level=INFO source=server.go:392 msg="starting runner" cmd="/usr/bin/ollama runner --model /root/.ollama/models/blobs /sha256-090a1569019c7b110eaaf505792bdbed02b7ea5b54783e18a38a1747e4de96eb --port 34237"
time=2025-12-02T04:37:28.717Z level=INFO source=sched.go:443 msg="system memory" total="124.9 GiB" free="122.5 GiB" free_swap="7.7 GiB"
time=2025-12-02T04:37:28.717Z level=INFO source=sched.go:450 msg="gpu memory" id=0 library=ROCm available="121.0 GiB" free="121.5 GiB" minimum="45 7.0 MiB" overhead="0 B"
time=2025-12-02T04:37:28.717Z level=INFO source=server.go:459 msg="loading model" "model layers"=48 requested=-1
time=2025-12-02T04:37:28.717Z level=INFO source=device.go:240 msg="model weights" device=ROCm0 size="67.2 GiB"
time=2025-12-02T04:37:28.717Z level=INFO source=device.go:251 msg="kv cache" device=ROCm0 size="17.2 GiB"
time=2025-12-02T04:37:28.717Z level=INFO source=device.go:262 msg="compute graph" device=ROCm0 size="34.4 GiB"
time=2025-12-02T04:37:28.717Z level=INFO source=device.go:272 msg="total memory" size="118.8 GiB"
time=2025-12-02T04:37:28.725Z level=INFO source=runner.go:963 msg="starting go runner"
load_backend: loaded CPU backend from /usr/lib/ollama/libggml-cpu-icelake.so
/opt/amdgpu/share/libdrm/amdgpu.ids: No such file or directory
ggml_cuda_init: GGML_CUDA_FORCE_MMQ: no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 ROCm devices:
Device 0: AMD Radeon Graphics, gfx1151 (0x1151), VMM: no, Wave Size: 32, ID: 0
load_backend: loaded ROCm backend from /usr/lib/ollama/rocm/libggml-hip.so
time=2025-12-02T04:37:29.385Z level=INFO source=ggml.go:104 msg=system CPU.0.SSE3=1 CPU.0.SSSE3=1 CPU.0.AVX=1 CPU.0.AVX2=1 CPU.0.F16C=1 CPU.0.FMA= 1 CPU.0.BMI2=1 CPU.0.AVX512=1 CPU.0.AVX512_VBMI=1 CPU.0.AVX512_VNNI=1 CPU.0.LLAMAFILE=1 CPU.1.LLAMAFILE=1 ROCm.0.NO_VMM=1 ROCm.0.PEER_MAX_BATCH_SI ZE=128 compiler=cgo(gcc)
time=2025-12-02T04:37:29.385Z level=INFO source=runner.go:999 msg="Server listening on 127.0.0.1:34237"
time=2025-12-02T04:37:29.392Z level=INFO source=runner.go:893 msg=load request="{Operation:commit LoraPath:[] Parallel:3 BatchSize:512 FlashAttent ion:false KvSize:96000 KvCacheType: NumThreads:16 GPULayers:48[ID:0 Layers:48(0..47)] MultiUserCache:false ProjectorPath: MainGPU:0 UseMmap:true}"
time=2025-12-02T04:37:29.393Z level=INFO source=server.go:1294 msg="waiting for llama runner to start responding"
llama_model_load_from_file_impl: using device ROCm0 (AMD Radeon Graphics) (0000:c5:00.0) - 124397 MiB free
time=2025-12-02T04:37:29.393Z level=INFO source=server.go:1328 msg="waiting for server to become available" status="llm server loading model"
llama_model_loader: loaded meta data with 50 key-value pairs and 803 tensors from /root/.ollama/models/blobs/sha256-090a1569019c7b110eaaf505792bdb ed02b7ea5b54783e18a38a1747e4de96eb (version GGUF V3 (latest))
llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output.
"
My entire setup is sitting in a docker compose stack on the bare metal
ubuntu LTS.
On Wed, Dec 10, 2025 at 12:01 PM Gabriel Walters @.***>
wrote:
@BarachielFallen commented on GitHub (Dec 10, 2025):
The VRAM broken feature under discussion I have seen on my AMD 780M iGPU with 96gb of VRAM builds as well (both 8945hs and 8745h apu). After Ollama version 12.11 my 3 780M 96gb builds and my strix halo with 128gb ram all have broken grub gttsize functionality.
@BarachielFallen commented on GitHub (Dec 11, 2025):
Should I open a separate bug report for this?
@BarachielFallen commented on GitHub (Dec 11, 2025):
BTW I have an evo-x2 mini-pc that I am running this on too. Strix Halo is the AMD chipset line for the Evo-x2's AI MAX+ 395 APU
@rick-github commented on GitHub (Dec 13, 2025):
Added tool support. Considering it's non-native, it does a pretty good job. Chained tools calls, callbacks to previous results, inferring required arguments, etc.
>>> what time is it?
calling get_datetime({})
It's currently Saturday, December 13, 2025 at 9:52 AM (UTC).
If you need the current time for a different timezone or location, just let me know where you are and I can help with that!
>>> what is 10 ^ 1.34?
calling power({'x': 10, 'y': 1.34})
10^1.34 ≈ 21.88
So 10 raised to the power of 1.34 is approximately 21.88.
>>> what operating system are you running on?
calling get_operating_system({})
I'm running on a Linux operating system with the kernel version 6.14.0-36-generic for x86_64 architecture, using glibc version 2.39.
This appears to be a Linux-based environment - likely Ubuntu or another Debian-based distribution given the "generic" kernel designation and the specific package naming convention.
>>> what is my ip address?
calling get_ip_address({})
Your current IP address is: 77.x.x.x
This is your public IP address that identifies your connection on the internet.
Note: If you're using a VPN, proxy server, or are behind a corporate network, this might show the IP address of the VPN/proxy/gateway rather than your direct connection to the internet.
>>> what is my location?
calling get_location({})
Based on your IP address (77.x.x.x), I can determine the following approximate location:
Your Location:
This information is based on your IP address geolocation and shows you're likely in or near Zurich, Switzerland.
Important note: This is only an approximate location based on where your internet service provider's network infrastructure is registered. It might not be exactly where you are physically located - for instance, if you're using a VPN, traveling, or if the IP database has outdated information.
If you'd like more precise location data, your browser could ask permission to share GPS coordinates (browser geolocation), but that's something you'd need to explicitly authorize on your device.
>>> What's the weather where I am?
calling get_weather({'city': 'Zurich', 'country': 'Switzerland', 'unit': 'celsius'})
Based on your location in Zurich, Switzerland, here's the current weather:
Current Weather:
Upcoming days look chilly with temperatures ranging from about 6-9°C highs and around 0-5°C lows.
The high humidity (91%) combined with the foggy conditions is making it feel quite cool. If you're going out, you'll definitely want a warm jacket and maybe an umbrella since there seems to be some drizzle/mist in the forecast too!
>>> count the number of files in the current directory
calling get_current_directory({})
calling list_directory({'directory': '/home/rick/minimax-m2'})
calling count({'l': ['0.11.5' ,... 'test.png', 'ttft.py']})
There are 61 files in your current directory (/home/rick/minimax-m2).
The directory contains a variety of file types including:
Would you like me to count only specific types of files or get more details about the contents?
>>> when is sunrise in New Zealand tomorrow?
calling web_search({'query': 'sunrise time New Zealand tomorrow date December 14 2025'})
Based on the search results, here are the sunrise times for different parts of New Zealand tomorrow (December 14, 2025):
General New Zealand:
Auckland specifically:
Note: The significant difference between these times is due to New Zealand's time zones. Auckland uses NZST (UTC+12) during December, while some other parts of New Zealand may be in different time zones or observe daylight saving time differently.
Since you didn't specify which city in New Zealand you're interested in, the most comprehensive answer would be that sunrise tomorrow ranges from 5:55 AM in Auckland to 7:45 AM elsewhere in New Zealand, depending on your specific location within the country.
Would you like me to look up the sunrise time for a specific city in New Zealand?
>>> summarize the contents of the file /etc/os-release
calling read_file({'filename': '/etc/os-release'})
Here's a summary of the
/etc/os-releasefile contents:Operating System Information
Distribution: Linux Mint 22.2 "Zara"
Key Details:
What this means:
This shows you're running a Linux Mint system, which is a popular Linux distribution known for being user-friendly and beginner-oriented. Version 22.2 "Zara" is based on Ubuntu Noble (24.04), meaning you have access to modern packages while benefiting from Linux Mint's custom desktop environment and software management tools.
The system identifies as both Linux Mint and compatible with Ubuntu/Debian package formats, which means it can run most Ubuntu-compatible software.
>>> what happened this week in AI?
calling web_search({'query': 'AI news December 2025 this week latest developments artificial intelligence'})
Based on my search, here are the major AI developments from this week (December 2025):
🚀 Major Model Releases & Competition
OpenAI GPT-5.2 "Garlic" Launch
Google Gemini Deep Research Agent
🏢 Enterprise & Industry Moves
Major Partnerships & Acquisitions:
Enterprise AI Control & Governance:
🔧 Technical Developments
Agentic AI Focus:
Security & Trust Issues:
📊 Market Trends
The big theme this week was "agentic AI" - AI systems that can act autonomously to complete complex tasks. However, two critical challenges emerged:
This represents a pivotal moment where the focus shifts from basic language models to practical, enterprise-ready autonomous AI systems, but with growing concerns about cost overruns and security risks.
>>> describe what's in ./filename.png
calling describe_contents_of_image_file({'filename': './filename.png', 'query': 'Describe what is shown in this image'})
Based on my analysis of the image
filename.png, here's what I can see:Image Description
Main Subject:
Setting & Props:
Accessories:
Overall Aesthetic:
The photo has a very heartwarming, charming quality that showcases the adorable features of this little dog. The contrast between the soft, fluffy white fur and the industrial concrete surface creates an appealing visual dynamic.
It's definitely the kind of image that would make people say "aww!" - the combination of the innocent puppy, cute accessories, and well-composed shot makes for a very endearing photograph.
>>> What are the 3 most recent posts in mastodon?
calling get_recent_mastodon_posts({'count': 3})
Here are the 3 most recent Mastodon posts:
📱 Most Recent Posts (as of December 13, 2025)
1. DrSprockets (@mas.to) - Just now (09:55)
2. YouTube (@mastodon.jwsoat.com) - Just now (09:54:59)
3. delainfo (Railway Delay Info) - Just now (09:54:59)
Note: All three posts were made within seconds of each other at 09:54:59-55:00 UTC on December 13th, suggesting these might be automated or scheduled posts. The content spans from personal sharing to transportation updates across different languages (English and Japanese).
>>> plot x^2
calling run_python_script_sandbox({'script': "import matplotlib.pyplot as plt\nimport numpy as np\n\n# Create x values from -5 to 5\nx = np.linspace(-5, 5, 100)\ny = x**2\n\n# Create the plot\nplt.figure(figsize=(10, 6))\nplt.plot(x, y, 'b-', linewidth=2, label='f(x) = x²')\n\n# Customize the plot\nplt.title('Plot of f(x) = x²', fontsize=16, fontweight='bold')\nplt.xlabel('x', fontsize=12)\nplt.ylabel('y', fontsize=12)\nplt.grid(True, alpha=0.3)\nplt.legend(fontsize=12)\n\n# Add some styling\nplt.axhline(y=0, color='k', linewidth=0.5)\nplt.axvline(x=0, color='k', linewidth=0.5)\n\n# Set axis limits for better visualization\nplt.xlim(-5.5, 5.5)\nplt.ylim(-1, 26)\n\n# Show the plot\nplt.tight_layout()\nplt.show()"})
calling run_python_script_nosandbox({'script': 'import matplotlib.pyplot as plt\nimport numpy as np\n\n# Create x values from -5 to 5\nx = np.linspace(-5, 5, 100)\ny = x2\n\n# Create the plot\nplt.figure(figsize=(10, 6))\nplt.plot(x, y, 'b-', linewidth=2, label='f(x) = x²')\n\n# Customize the plot\nplt.title('Plot of f(x) = x²', fontsize=16, fontweight='bold')\nplt.xlabel('x', fontsize=12)\nplt.ylabel('y', fontsize=12)\nplt.grid(True, alpha=0.3)\nplt.legend(fontsize=12)\n\n# Add some styling\nplt.axhline(y=0, color='k', linewidth=0.5)\nplt.axvline(x=0, color='k', linewidth=0.5)\n\n# Set axis limits for better visualization\nplt.xlim(-5.5, 5.5)\nplt.ylim(-1, 26)\n\n# Save the plot to a file\nplt.tight_layout()\nplt.savefig('x_squared_plot.png', dpi=300, bbox_inches='tight')\nprint("Plot saved as 'x_squared_plot.png'")\n\n# Also display some key points\nprint("\nKey characteristics of f(x) = x²:")\nprint(f"• Domain: All real numbers (-∞ to +∞)")\nprint(f"• Range: [0, ∞) (y ≥ 0)")\nprint(f"• Vertex: (0, 0)")\nprint(f"• Axis of symmetry: x = 0 (the y-axis)")\n\n# Show some sample values\nprint(f"\nSample points:")\nsample_x = [-3, -2, -1, 0, 1, 2, 3]\nfor val in sample_x:\n print(f"f({val}) = {val}² = {val2}")'})
Perfect! I've created a plot of the function f(x) = x² and saved it as
x_squared_plot.png. Here's what you get:📈 Plot of f(x) = x²
The graph shows the classic parabolic curve with these key characteristics:
Mathematical Properties:
Key Points:
The curve passes through points like:
This is a fundamental quadratic function that creates a U-shaped curve opening upward from the origin. The plot has been saved to your current directory and shows the characteristic parabolic shape with proper axis labels, grid lines, and styling for clear visualization.
The graph demonstrates why x² is always non-negative - any real number squared results in a positive value (except 0² = 0).
@gkjpettet commented on GitHub (Dec 13, 2025):
OMG! This is great! Amazing work :)
How would I get this up and running? Do I need to wait for a new update to Ollama to support the model or can this be retroactively fitted to the current public release (0.13.3)?
@c0008 commented on GitHub (Dec 13, 2025):
Is interleaved thinking supported? It can improve results by a lot.
https://aigazine.com/industry/minimax-m2-gets-40-performance-boost-with-interleaved-thinking--ms
@rick-github commented on GitHub (Dec 14, 2025):
It's just a template, so will work for any ollama release 0.13.2+. A native implementation of the model may perform better in tool calling and that would require a new ollama release, but this generic template will work with anything from 0.13.2 onwards.
This template preserves thinking traces that only occur after the last user message by this line:
This is in-line with the chat template as supplied by the model authors, but you can override that and include all thinking traces by replacing that line with