mirror of
https://github.com/ollama/ollama.git
synced 2026-05-06 08:02:14 -05:00
Closed
opened 2026-04-28 04:43:02 -05:00 by GiteaMirror
·
81 comments
No Branch/Tag Specified
main
parth-mlx-decode-checkpoints
dhiltgen/ci
hoyyeva/editor-config-repair
parth-launch-codex-app
hoyyeva/fix-codex-model-metadata-warning
hoyyeva/qwen
hoyyeva/launch-backup-ux
parth/hide-claude-desktop-till-release
hoyyeva/opencode-image-modality
parth-add-claude-code-autoinstall
release_v0.22.0
pdevine/manifest-list
codex/fix-codex-model-metadata-warning
pdevine/addressable-manifest
brucemacd/launch-fetch-reccomended
jmorganca/llama-compat
launch-copilot-cli
hoyyeva/opencode-thinking
release_v0.20.7
parth-auto-save-backup
parth-test
jmorganca/gemma4-audio-replacements
fix-manifest-digest-on-pull
hoyyeva/vscode-improve
brucemacd/install-server-wait
brucemacd/download-before-remove
parth/update-claude-docs
parth-anthropic-reference-images-path
brucemac/start-ap-install
pdevine/mlx-update
pdevine/qwen35_vision
drifkin/api-show-fallback
mintlify/image-generation-1773352582
hoyyeva/server-context-length-local-config
jmorganca/faster-reptition-penalties
jmorganca/convert-nemotron
parth-pi-thinking
pdevine/sampling-penalties
jmorganca/fix-create-quantization-memory
dongchen/resumable_transfer_fix
pdevine/sampling-cache-error
jessegross/mlx-usage
hoyyeva/openclaw-config
hoyyeva/app-html
pdevine/qwen3next
brucemacd/sign-sh-install
brucemacd/tui-update
brucemacd/usage-api
jmorganca/launch-empty
fix-app-dist-embed
mxyng/mlx-compile
mxyng/mlx-quant
mxyng/mlx-glm4.7
mxyng/mlx
brucemacd/simplify-model-picker
jmorganca/qwen3-concurrent
fix-glm-4.7-flash-mla-config
drifkin/qwen3-coder-opening-tag
brucemacd/usage-cli
fix-cuda12-fattn-shmem
ollama-imagegen-docs
parth/fix-multiline-inputs
brucemacd/config-docs
mxyng/model-files
mxyng/simple-execute
fix-imagegen-ollama-models
mxyng/async-upload
jmorganca/lazy-no-dtype-changes
imagegen-auto-detect-create
parth/decrease-concurrent-download-hf
fix-mlx-quantize-init
jmorganca/x-cleanup
usage
imagegen-readme
jmorganca/glm-image
mlx-gpu-cd
jmorganca/imagegen-modelfile
parth/agent-skills
parth/agent-allowlist
parth/signed-in-offline
parth/agents
parth/fix-context-chopping
improve-cloud-flow
parth/add-models-websearch
parth/prompt-renderer-mcp
jmorganca/native-settings
jmorganca/download-stream-hash
jmorganca/client2-rebased
brucemacd/oai-chat-req-multipart
jessegross/multi_chunk_reserve
grace/additional-omit-empty
grace/mistral-3-large
mxyng/tokenizer2
mxyng/tokenizer
jessegross/flash
hoyyeva/windows-nacked-app
mxyng/cleanup-attention
grace/deepseek-parser
hoyyeva/remember-unsent-prompt
parth/add-lfs-pointer-error-conversion
parth/olmo2-test2
hoyyeva/ollama-launchagent-plist
nicole/olmo-model
parth/olmo-test
mxyng/remove-embedded
parth/render-template
jmorganca/intellect-3
parth/remove-prealloc-linter
jmorganca/cmd-eval
nicole/nomic-embed-text-fix
mxyng/lint-2
hoyyeva/add-gemini-3-pro-preview
hoyyeva/load-model-list
mxyng/expand-path
mxyng/environ-2
hoyyeva/deeplink-json-encoding
parth/improve-tool-calling-tests
hoyyeva/conversation
hoyyeva/assistant-edit-response
hoyyeva/thinking
origin/brucemacd/invalid-char-i-err
parth/improve-tool-calling
jmorganca/required-omitempty
grace/qwen3-vl-tests
mxyng/iter-client
parth/docs-readme
nicole/embed-test
pdevine/integration-benchstat
parth/remove-generate-cmd
parth/add-toolcall-id
mxyng/server-tests
jmorganca/glm-4.6
jmorganca/gin-h-compat
drifkin/stable-tool-args
pdevine/qwen3-more-thinking
parth/add-websearch-client
nicole/websearch_local
jmorganca/qwen3-coder-updates
grace/deepseek-v3-migration-tests
mxyng/fix-create
jmorganca/cloud-errors
pdevine/parser-tidy
revert-12233-parth/simplify-entrypoints-runner
parth/enable-so-gpt-oss
brucemacd/qwen3vl
jmorganca/readme-simplify
parth/gpt-oss-structured-outputs
revert-12039-jmorganca/tools-braces
mxyng/embeddings
mxyng/gguf
mxyng/benchmark
mxyng/types-null
parth/move-parsing
mxyng/gemma2
jmorganca/docs
mxyng/16-bit
mxyng/create-stdin
pdevine/authorizedkeys
mxyng/quant
parth/opt-in-error-context-window
brucemacd/cache-models
brucemacd/runner-completion
jmorganca/llama-update-6
brucemacd/benchmark-list
brucemacd/partial-read-caps
parth/deepseek-r1-tools
mxyng/omit-array
parth/tool-prefix-temp
brucemacd/runner-test
jmorganca/qwen25vl
brucemacd/model-forward-test-ext
parth/python-function-parsing
jmorganca/cuda-compression-none
drifkin/num-parallel
drifkin/chat-truncation-fix
jmorganca/sync
parth/python-tools-calling
drifkin/array-head-count
brucemacd/create-no-loop
parth/server-enable-content-stream-with-tools
qwen25omni
mxyng/v3
brucemacd/ropeconfig
jmorganca/silence-tokenizer
parth/sample-so-test
parth/sampling-structured-outputs
brucemacd/doc-go-engine
parth/constrained-sampling-json
jmorganca/mistral-wip
brucemacd/mistral-small-convert
parth/sample-unmarshal-json-for-params
brucemacd/jomorganca/mistral
pdevine/bfloat16
jmorganca/mistral
brucemacd/mistral
pdevine/logging
parth/sample-correctness-fix
parth/sample-fix-sorting
jmorgan/sample-fix-sorting-extras
jmorganca/temp-0-images
brucemacd/parallel-embed-models
brucemacd/shim-grammar
jmorganca/fix-gguf-error
bmizerany/nameswork
jmorganca/faster-releases
bmizerany/validatenames
brucemacd/err-no-vocab
brucemacd/rope-config
brucemacd/err-hint
brucemacd/qwen2_5
brucemacd/logprobs
brucemacd/new_runner_graph_bench
progress-flicker
brucemacd/forward-test
brucemacd/go_qwen2
pdevine/gemma2
jmorganca/add-missing-symlink-eval
mxyng/next-debug
parth/set-context-size-openai
brucemacd/next-bpe-bench
brucemacd/next-bpe-test
brucemacd/new_runner_e2e
brucemacd/new_runner_qwen2
pdevine/convert-cohere2
brucemacd/convert-cli
parth/log-probs
mxyng/next-mlx
mxyng/cmd-history
parth/templating
parth/tokenize-detokenize
brucemacd/check-key-register
bmizerany/grammar
jmorganca/vendor-081b29bd
mxyng/func-checks
jmorganca/fix-null-format
parth/fix-default-to-warn-json
jmorganca/qwen2vl
jmorganca/no-concat
parth/cmd-cleanup-SO
brucemacd/check-key-register-structured-err
parth/openai-stream-usage
parth/fix-referencing-so
stream-tools-stop
jmorganca/degin-1
brucemacd/install-path-clean
brucemacd/push-name-validation
brucemacd/browser-key-register
jmorganca/openai-fix-first-message
jmorganca/fix-proxy
jessegross/sample
parth/disallow-streaming-tools
dhiltgen/remove_submodule
jmorganca/ga
jmorganca/mllama
pdevine/newlines
pdevine/geems-2b
jmorganca/llama-bump
mxyng/modelname-7
mxyng/gin-slog
mxyng/modelname-6
jyan/convert-prog
jyan/quant5
paligemma-support
pdevine/import-docs
jmorganca/openai-context
jyan/paligemma
jyan/p2
jyan/palitest
bmizerany/embedspeedup
jmorganca/llama-vit
brucemacd/allow-ollama
royh/ep-methods
royh/whisper
mxyng/api-models
mxyng/fix-memory
jyan/q4_4/8
jyan/ollama-v
royh/stream-tools
roy-embed-parallel
bmizerany/hrm
revert-5963-revert-5924-mxyng/llama3.1-rope
royh/embed-viz
jyan/local2
jyan/auth
jyan/local
jyan/parse-temp
jmorganca/template-mistral
jyan/reord-g
royh-openai-suffixdocs
royh-imgembed
royh-embed-parallel
jyan/quant4
royh-precision
jyan/progress
pdevine/fix-template
jyan/quant3
pdevine/ggla
mxyng/update-registry-domain
jmorganca/ggml-static
mxyng/create-context
jyan/v0.146
mxyng/layers-from-files
build_dist
bmizerany/noseek
royh-ls
royh-name
timeout
mxyng/server-timestamp
bmizerany/nosillyggufslurps
royh-params
jmorganca/llama-cpp-7c26775
royh-openai-delete
royh-show-rigid
jmorganca/enable-fa
jmorganca/no-error-template
jyan/format
royh-testdelete
bmizerany/fastverify
language_support
pdevine/ps-glitches
brucemacd/tokenize
bruce/iq-quants
bmizerany/filepathwithcoloninhost
mxyng/split-bin
bmizerany/client-registry
jmorganca/if-none-match
native
jmorganca/native
jmorganca/batch-embeddings
jmorganca/initcmake
jmorganca/mm
pdevine/showggmlinfo
modenameenforcealphanum
bmizerany/modenameenforcealphanum
jmorganca/done-reason
jmorganca/llama-cpp-8960fe8
ollama.com
bmizerany/filepathnobuild
bmizerany/types/model/defaultfix
rmdisplaylong
nogogen
bmizerany/x
modelfile-readme
bmizerany/replacecolon
jmorganca/limit
jmorganca/execstack
jmorganca/replace-assets
mxyng/tune-concurrency
jmorganca/testing
whitespace-detection
jmorganca/options
upgrade-all
scratch
cuda-search
mattw/airenamer
mattw/allmodelsonhuggingface
mattw/quantcontext
mattw/whatneedstorun
brucemacd/llama-mem-calc
mattw/faq-context
mattw/communitylinks
mattw/noprune
mattw/python-functioncalling
rename
mxyng/install
pulse
remove-first
editor
mattw/selfqueryingretrieval
cgo
mattw/howtoquant
api
matt/streamingapi
format-config
mxyng/extra-args
shell
update-nous-hermes
cp-model
upload-progress
fix-unknown-model
fix-model-names
delete-fix
insecure-registry
ls
deletemodels
progressbar
readme-updates
license-layers
skip-list
list-models
modelpath
matt/examplemodelfiles
distribution
go-opts
v0.23.1
v0.23.1-rc0
v0.23.0
v0.23.0-rc0
v0.22.1
v0.22.1-rc1
v0.22.1-rc0
v0.22.0
v0.22.0-rc1
v0.21.3-rc0
v0.21.2-rc1
v0.21.2
v0.21.2-rc0
v0.21.1
v0.21.1-rc1
v0.21.1-rc0
v0.21.0
v0.21.0-rc1
v0.21.0-rc0
v0.20.8-rc0
v0.20.7
v0.20.7-rc1
v0.20.7-rc0
v0.20.6
v0.20.6-rc1
v0.20.6-rc0
v0.20.5
v0.20.5-rc2
v0.20.5-rc1
v0.20.5-rc0
v0.20.4
v0.20.4-rc2
v0.20.4-rc1
v0.20.4-rc0
v0.20.3
v0.20.3-rc0
v0.20.2
v0.20.1
v0.20.1-rc2
v0.20.1-rc1
v0.20.1-rc0
v0.20.0
v0.20.0-rc1
v0.20.0-rc0
v0.19.0
v0.19.0-rc2
v0.19.0-rc1
v0.19.0-rc0
v0.18.4-rc1
v0.18.4-rc0
v0.18.3
v0.18.3-rc2
v0.18.3-rc1
v0.18.3-rc0
v0.18.2
v0.18.2-rc1
v0.18.2-rc0
v0.18.1
v0.18.1-rc1
v0.18.1-rc0
v0.18.0
v0.18.0-rc2
v0.18.0-rc1
v0.18.0-rc0
v0.17.8-rc4
v0.17.8-rc3
v0.17.8-rc2
v0.17.8-rc1
v0.17.8-rc0
v0.17.7
v0.17.7-rc2
v0.17.7-rc1
v0.17.7-rc0
v0.17.6
v0.17.5
v0.17.4
v0.17.3
v0.17.2
v0.17.1
v0.17.1-rc2
v0.17.1-rc1
v0.17.1-rc0
v0.17.0
v0.17.0-rc2
v0.17.0-rc1
v0.17.0-rc0
v0.16.3
v0.16.3-rc2
v0.16.3-rc1
v0.16.3-rc0
v0.16.2
v0.16.2-rc0
v0.16.1
v0.16.0
v0.16.0-rc2
v0.16.0-rc0
v0.16.0-rc1
v0.15.6
v0.15.5
v0.15.5-rc5
v0.15.5-rc4
v0.15.5-rc3
v0.15.5-rc2
v0.15.5-rc1
v0.15.5-rc0
v0.15.4
v0.15.3
v0.15.2
v0.15.1
v0.15.1-rc1
v0.15.1-rc0
v0.15.0-rc6
v0.15.0
v0.15.0-rc5
v0.15.0-rc4
v0.15.0-rc3
v0.15.0-rc2
v0.15.0-rc1
v0.15.0-rc0
v0.14.3
v0.14.3-rc3
v0.14.3-rc2
v0.14.3-rc1
v0.14.3-rc0
v0.14.2
v0.14.2-rc1
v0.14.2-rc0
v0.14.1
v0.14.0-rc11
v0.14.0
v0.14.0-rc10
v0.14.0-rc9
v0.14.0-rc8
v0.14.0-rc7
v0.14.0-rc6
v0.14.0-rc5
v0.14.0-rc4
v0.14.0-rc3
v0.14.0-rc2
v0.14.0-rc1
v0.14.0-rc0
v0.13.5
v0.13.5-rc1
v0.13.5-rc0
v0.13.4-rc2
v0.13.4
v0.13.4-rc1
v0.13.4-rc0
v0.13.3
v0.13.3-rc1
v0.13.3-rc0
v0.13.2
v0.13.2-rc2
v0.13.2-rc1
v0.13.2-rc0
v0.13.1
v0.13.1-rc2
v0.13.1-rc1
v0.13.1-rc0
v0.13.0
v0.13.0-rc0
v0.12.11
v0.12.11-rc1
v0.12.11-rc0
v0.12.10
v0.12.10-rc1
v0.12.10-rc0
v0.12.9-rc0
v0.12.9
v0.12.8
v0.12.8-rc0
v0.12.7
v0.12.7-rc1
v0.12.7-rc0
v0.12.7-citest0
v0.12.6
v0.12.6-rc1
v0.12.6-rc0
v0.12.5
v0.12.5-rc0
v0.12.4
v0.12.4-rc7
v0.12.4-rc6
v0.12.4-rc5
v0.12.4-rc4
v0.12.4-rc3
v0.12.4-rc2
v0.12.4-rc1
v0.12.4-rc0
v0.12.3
v0.12.2
v0.12.2-rc0
v0.12.1
v0.12.1-rc1
v0.12.1-rc2
v0.12.1-rc0
v0.12.0
v0.12.0-rc1
v0.12.0-rc0
v0.11.11
v0.11.11-rc3
v0.11.11-rc2
v0.11.11-rc1
v0.11.11-rc0
v0.11.10
v0.11.9
v0.11.9-rc0
v0.11.8
v0.11.8-rc0
v0.11.7-rc1
v0.11.7-rc0
v0.11.7
v0.11.6
v0.11.6-rc0
v0.11.5-rc4
v0.11.5-rc3
v0.11.5
v0.11.5-rc5
v0.11.5-rc2
v0.11.5-rc1
v0.11.5-rc0
v0.11.4
v0.11.4-rc0
v0.11.3
v0.11.3-rc0
v0.11.2
v0.11.1
v0.11.0-rc0
v0.11.0-rc1
v0.11.0-rc2
v0.11.0
v0.10.2-int1
v0.10.1
v0.10.0
v0.10.0-rc4
v0.10.0-rc3
v0.10.0-rc2
v0.10.0-rc1
v0.10.0-rc0
v0.9.7-rc1
v0.9.7-rc0
v0.9.6
v0.9.6-rc0
v0.9.6-ci0
v0.9.5
v0.9.4-rc5
v0.9.4-rc6
v0.9.4
v0.9.4-rc3
v0.9.4-rc4
v0.9.4-rc1
v0.9.4-rc2
v0.9.4-rc0
v0.9.3
v0.9.3-rc5
v0.9.4-citest0
v0.9.3-rc4
v0.9.3-rc3
v0.9.3-rc2
v0.9.3-rc1
v0.9.3-rc0
v0.9.2
v0.9.1
v0.9.1-rc1
v0.9.1-rc0
v0.9.1-ci1
v0.9.1-ci0
v0.9.0
v0.9.0-rc0
v0.8.0
v0.8.0-rc0
v0.7.1-rc2
v0.7.1
v0.7.1-rc1
v0.7.1-rc0
v0.7.0
v0.7.0-rc1
v0.7.0-rc0
v0.6.9-rc0
v0.6.8
v0.6.8-rc0
v0.6.7
v0.6.7-rc2
v0.6.7-rc1
v0.6.7-rc0
v0.6.6
v0.6.6-rc2
v0.6.6-rc1
v0.6.6-rc0
v0.6.5-rc1
v0.6.5
v0.6.5-rc0
v0.6.4-rc0
v0.6.4
v0.6.3-rc1
v0.6.3
v0.6.3-rc0
v0.6.2
v0.6.2-rc0
v0.6.1
v0.6.1-rc0
v0.6.0-rc0
v0.6.0
v0.5.14-rc0
v0.5.13
v0.5.13-rc6
v0.5.13-rc5
v0.5.13-rc4
v0.5.13-rc3
v0.5.13-rc2
v0.5.13-rc1
v0.5.13-rc0
v0.5.12
v0.5.12-rc1
v0.5.12-rc0
v0.5.11
v0.5.10
v0.5.9
v0.5.9-rc0
v0.5.8-rc13
v0.5.8
v0.5.8-rc12
v0.5.8-rc11
v0.5.8-rc10
v0.5.8-rc9
v0.5.8-rc8
v0.5.8-rc7
v0.5.8-rc6
v0.5.8-rc5
v0.5.8-rc4
v0.5.8-rc3
v0.5.8-rc2
v0.5.8-rc1
v0.5.8-rc0
v0.5.7
v0.5.6
v0.5.5
v0.5.5-rc0
v0.5.4
v0.5.3
v0.5.3-rc0
v0.5.2
v0.5.2-rc3
v0.5.2-rc2
v0.5.2-rc1
v0.5.2-rc0
v0.5.1
v0.5.0
v0.5.0-rc1
v0.4.8-rc0
v0.4.7
v0.4.6
v0.4.5
v0.4.4
v0.4.3
v0.4.3-rc0
v0.4.2
v0.4.2-rc1
v0.4.2-rc0
v0.4.1
v0.4.1-rc0
v0.4.0
v0.4.0-rc8
v0.4.0-rc7
v0.4.0-rc6
v0.4.0-rc5
v0.4.0-rc4
v0.4.0-rc3
v0.4.0-rc2
v0.4.0-rc1
v0.4.0-rc0
v0.4.0-ci3
v0.3.14
v0.3.14-rc0
v0.3.13
v0.3.12
v0.3.12-rc5
v0.3.12-rc4
v0.3.12-rc3
v0.3.12-rc2
v0.3.12-rc1
v0.3.11
v0.3.11-rc4
v0.3.11-rc3
v0.3.11-rc2
v0.3.11-rc1
v0.3.10
v0.3.10-rc1
v0.3.9
v0.3.8
v0.3.7
v0.3.7-rc6
v0.3.7-rc5
v0.3.7-rc4
v0.3.7-rc3
v0.3.7-rc2
v0.3.7-rc1
v0.3.6
v0.3.5
v0.3.4
v0.3.3
v0.3.2
v0.3.1
v0.3.0
v0.2.8
v0.2.8-rc2
v0.2.8-rc1
v0.2.7
v0.2.6
v0.2.5
v0.2.4
v0.2.3
v0.2.2
v0.2.2-rc2
v0.2.2-rc1
v0.2.1
v0.2.0
v0.1.49-rc14
v0.1.49-rc13
v0.1.49-rc12
v0.1.49-rc11
v0.1.49-rc10
v0.1.49-rc9
v0.1.49-rc8
v0.1.49-rc7
v0.1.49-rc6
v0.1.49-rc4
v0.1.49-rc5
v0.1.49-rc3
v0.1.49-rc2
v0.1.49-rc1
v0.1.48
v0.1.47
v0.1.46
v0.1.45-rc5
v0.1.45
v0.1.45-rc4
v0.1.45-rc3
v0.1.45-rc2
v0.1.45-rc1
v0.1.44
v0.1.43
v0.1.42
v0.1.41
v0.1.40
v0.1.40-rc1
v0.1.39
v0.1.39-rc2
v0.1.39-rc1
v0.1.38
v0.1.37
v0.1.36
v0.1.35
v0.1.35-rc1
v0.1.34
v0.1.34-rc1
v0.1.33
v0.1.33-rc7
v0.1.33-rc6
v0.1.33-rc5
v0.1.33-rc4
v0.1.33-rc3
v0.1.33-rc2
v0.1.33-rc1
v0.1.32
v0.1.32-rc2
v0.1.32-rc1
v0.1.31
v0.1.30
v0.1.29
v0.1.28
v0.1.27
v0.1.26
v0.1.25
v0.1.24
v0.1.23
v0.1.22
v0.1.21
v0.1.20
v0.1.19
v0.1.18
v0.1.17
v0.1.16
v0.1.15
v0.1.14
v0.1.13
v0.1.12
v0.1.11
v0.1.10
v0.1.9
v0.1.8
v0.1.7
v0.1.6
v0.1.5
v0.1.4
v0.1.3
v0.1.2
v0.1.1
v0.1.0
v0.0.21
v0.0.20
v0.0.19
v0.0.18
v0.0.17
v0.0.16
v0.0.15
v0.0.14
v0.0.13
v0.0.12
v0.0.11
v0.0.10
v0.0.9
v0.0.8
v0.0.7
v0.0.6
v0.0.5
v0.0.4
v0.0.3
v0.0.2
v0.0.1
Labels
Clear labels
amd
api
app
bug
build
cli
cloud
compatibility
context-length
create
docker
documentation
embeddings
feature request
feedback wanted
good first issue
gpt-oss
gpu
harmony
help wanted
image
install
intel
js
launch
linux
macos
memory
mlx
model
needs more info
networking
nvidia
ollama.com
performance
pull-request
python
question
registry
rendering
thinking
tools
top
vulkan
windows
wsl
Mirrored from GitHub Pull Request
No Label
Milestone
No items
No Milestone
Projects
Clear projects
No project
No Assignees
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: github-starred/ollama#47654
Reference in New Issue
Block a user
Blocking a user prevents them from interacting with repositories, such as opening or commenting on pull requests or issues. Learn more about blocking a user.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @remy415 on GitHub (Jan 13, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/1979
Originally assigned to: @dhiltgen on GitHub.
I've reviewed the great tutorial made by @bnodnarb here:
https://github.com/jmorganca/ollama/blob/main/docs/tutorials/nvidia-jetson.md
The Orin Nano is running Ubuntu 20.04 with Jetpack 5.1.2 (r35.4.1 L4T). The container is also running L4T version 35.4.1. Jetpack 5.1.2 comes with CUDA 11.4 installed with compatibility support for CUDA 11.8.
I also followed along with the other 3 Jetson-related issues and have not found a fix.
I have also:
Run ollama serve
In each of the situations, I used the 'mistral-jetson' generated model. For each of them, I get a similar output:
Key outputs are:
2024/01/13 20:14:03 routes.go:953: no GPU detectedllm_load_tensors: mem required = 3917.98 MiBAgain, would just like to note that the stable-diffusion-webui application works with GPU, as well as the referenced docker container from dustynv. Any suggestions of things to check?
Update: I forgot to mention that I verified CPU and GPU activity using jtop in another terminal. Edited for formatting. Edited to add OS & Jetson versions. Edited to add CUDA version.
@Q-point commented on GitHub (Jan 20, 2024):
@remy415 any solution ? I am observing the same on AGX Orin. Seems like a bug on ollama.
I verified with Jtop also. Ollama only runs on ARM cpus
~$ LD_LIBRARY_PATH=/usr/local/cuda/lib64 ollama serve
2024/01/19 21:25:40 images.go:808: total blobs: 8
2024/01/19 21:25:40 images.go:815: total unused blobs removed: 0
2024/01/19 21:25:40 routes.go:930: Listening on 127.0.0.1:11434 (version 0.1.20)
2024/01/19 21:25:41 shim_ext_server.go:142: Dynamic LLM variants [cuda]
2024/01/19 21:25:41 gpu.go:88: Detecting GPU type
2024/01/19 21:25:41 gpu.go:203: Searching for GPU management library libnvidia-ml.so
2024/01/19 21:25:41 gpu.go:248: Discovered GPU libraries: []
2024/01/19 21:25:41 gpu.go:203: Searching for GPU management library librocm_smi64.so
2024/01/19 21:25:41 gpu.go:248: Discovered GPU libraries: []
2024/01/19 21:25:41 routes.go:953: no GPU detected
@remy415 commented on GitHub (Jan 20, 2024):
@Q-point nothing yet. I haven’t had time to troubleshoot. If I had to guess, I would say there may have been an update in the way Jetpack presents its drivers; I’m not an expert in Linux drivers, it’s just the only thing that makes sense given that @bnodnarb was able to get it working with little customization, and I doubt Ollama made any tweaks if it was already working so the only logical culprit is a change in the drivers.
I do know that NVidia made a change in the way it exposes CUDA to containers. Previously, containers would basically mount the installed drivers in the container. Now, the containers released by dustynv have the drivers baked in to the containers and the expectation from NVidia is the decoupling of host system drivers and container-used drivers.
@dhiltgen commented on GitHub (Jan 26, 2024):
Is there any way to get
libnvidia-ml.soinstalled on the system? What does thenvidial-smioutput look like on these systems?@remy415 commented on GitHub (Jan 26, 2024):
I’ve tried to get it installed, but as dustynv pointed out in another post
somewhere (on my phone, will find it later) the Tegra line of SBCs running
Jetpack are integrated GPUs and aren’t compatible with nvml/
libnvidia-ml.so/nvidial-smi. This is changing in Jetpack 6.0 but that isn’t
releasing until at least March.
I spent some time poking around in the ollama source code to see what
exactly it needed from libnvidia-ml.so, but I was having difficulty finding
comparable syscalls on the Jetson because the system data tools I did find
are just Python scripts that call Python CUDA libraries; I didn’t dive too
far down that rabbit hole.
Another thing is that the llama_cpp that works for the Jetson is a custom
build done by dustynv that leverages the nvcc compiler. I tried injecting
his llama_cpp container and prebuilt binary into the ollama dockerfile
build but it didn’t work; I think there is something gpu_info passes to the
make process but I haven’t worked the kinks out of that yet, and I still
need to find what information the gpu_info.go routine is requiring from the
CUDA api to ensure it’s properly converted to the Jetson format.
Any insights you could provide on that front would be greatly appreciated.
On Fri, Jan 26, 2024 at 4:51 PM Daniel Hiltgen @.***>
wrote:
@dhiltgen commented on GitHub (Jan 26, 2024):
That's unfortunate they didn't implement support for the management library. We've added a dependency on that to discover the available GPUs, and their memory information so we can determine how much we can load into the GPU. We do have a mechanism now to force a specific llm library with
OLLAMA_LLM_LIBRARY(see https://github.com/ollama/ollama/blob/main/docs/troubleshooting.md#llm-libraries for usage) however this doesn't currently play nicely with the memory prediction logic for GPUs - it's largely to force CPU mode at this point.A potential path for us to consider here is to refine the memory prediction logic so you can tell us how much memory to use via env var and bypass the management library checks, and then force the cuda llm library, and that might be sufficient to get us working again on these systems.
@remy415 commented on GitHub (Jan 27, 2024):
There's support for the library, kind of. I've compiled my own .so file essentially wrapping the nvml function parameters returned in the API calls. I'm not terribly great with C/C++ but I got it to compile with NVCC. TL;DR: I used cuda_runtime.h API calls to gather the same information returned by the NVML calls. It seems to have mostly worked but I'm running into errors. I'll be taking another try on it tomorrow some time, but meanwhile I've uploaded file changes if you're interested in taking a look. https://www.github.com/remy415/ollama_tegra_fix
The error I received:
Seems like the GetGpuInfo() function (memInfo call) failed. I'll need to take a look and make sure I implemented the API correctly and make sure the information is passed in the same format the Go routine expects. I haven't had a lot of time to troubleshoot today, this is just an initial draft of what I was thinking could be done to avoid leaving too much in the user's hands, assuming API usage is preferred over user-defined env variables for memory, etc.
@dhiltgen commented on GitHub (Jan 27, 2024):
@remy415 here's the set of APIs we call in the library. The critical ones are
nvmlInit_v2,nvmlDeviceGetCount_v2,nvmlDeviceGetHandleByIndex, andnvmlDeviceGetMemoryInfo. The rest can be stubbed to return "not supported" style errors and we should gracefully handle them.@remy415 commented on GitHub (Jan 27, 2024):
@dhiltgen Thank you, I'll take a look at that. I'll update the repo and let you know here if I get anything working.
@remy415 commented on GitHub (Jan 28, 2024):
@dhiltgen I was able to use the CUDA Runtime Driver API compiled with NVCC to grab an initialize, device count, "gethandlebyindex" was just assumed to be '0' -- there's no Driver API call for it but it's also not needed -- simply query device properties for a device index (can be done in a for loop using devicecount), and the memory info (memory max, memory used, used a diff to figure out the free memory), and the CUDA capability major & minor values. This will work on any CUDA device and doesn't require hooking into the NVML shared object. It only needs to be compiled with NVCC but that's included in the standard CUDA toolkit. The source code for the sample binary I made should work on any linux machine with CUDA drivers and a CUDA device.
Any reason you wouldn't want to switch to using the CUDA Runtime API instead of querying nvml? I'm not super experienced with the -isms/quirks of pinging the Runtime API vs pinging the NVML API, but if all you're doing is gathering device info before loading llama_cpp then leveraging the Runtime API instead of NVML would work great for a solution that is compatible with Jetson and desktop CUDA. I think there's device properties to cover most of what the rest of your typedefs were looking for too.
If you have a linux box with a CUDA device (I'll test it out in Windows at some point too, but I think it should work there as their device API calls seem to be system agnostic), please let me know if this would work as a suitable alternative to libnvidia-ml.so.
Files:
https://raw.githubusercontent.com/remy415/ollama_tegra_fix/master/tegra-ml-test.cu
https://raw.githubusercontent.com/remy415/ollama_tegra_fix/master/gpu_info_tegra.h
Compile command is
nvcc --ptxas-options=-v --compiler-options '-fPIC' -o tegra-ml-test.out tegra-ml-test.cuCommand output:
@dhiltgen commented on GitHub (Jan 28, 2024):
This sort of approach could be viable. A key aspect of our GPU discovery logic is relying on dlopen/dlsym (and LoadLibrary/GetProcAddress on windows) so that we can have a soft dependency on the underlying GPU libraries. This lets us fail gracefully at runtime and try multiple options before ultimately falling back to CPU mode if necessary. I believe this would translate into loading libcudart.so and wiring up these
cuda*routines.@remy415 commented on GitHub (Jan 28, 2024):
I'll try and whip something up as gpu_info_tegra.c & gpu_info_tegra.h with the same structure as gpu_info_cuda.c/etc. I can see the benefits of keeping your code as loosely coupled to CUDA as possible.
Correct me if I'm wrong, but does ollama compile llama_cpp into it when you build it? I didn't see anything in the manual installation instructions. If so, Jetson doesn't work with the default compilations of llama_cpp and requires a build using a syntax I've copied from dustynv's llama_cpp container build. Where is the proper place to inject llama_cpp build flags?
@dhiltgen commented on GitHub (Jan 28, 2024):
We use
go generate ./...as a mechanism to perform the native compilation of llama.cpp and leverage the cmake build rigging from the upstream repo (with some minor modifications for our use). Check out https://github.com/ollama/ollama/tree/main/llm/generate and in particular, gen_linux.sh and gen_common.sh - Build flags are somewhat tunable, but Jetson may require additional refinement of those scripts.@remy415 commented on GitHub (Jan 28, 2024):
I saw something in the comments about all your CUDA builds requiring/using AVX. Tegras ship with ARM64 CPUs that don't have AVX extensions, so if the logic automatically disables GPU support if AVX isn't present (as per the comments) then the GPU library loading will be skipped every time.
@dhiltgen commented on GitHub (Jan 28, 2024):
Good catch! Yes, that recently introduced logic needs to be x86 only. I'll get a PR up for that.
@bnodnarb commented on GitHub (Jan 29, 2024):
Hi @remy415 and @dhiltgen - thanks for all the effort you both are putting into this. I can confirm that the original tutorial no longer works with Jetpack 5.1.2. I will try on the Jetpack 6 Developer Preview.
Thanks all!
@remy415 commented on GitHub (Jan 29, 2024):
Okay so I wrote a .c file for Tegra devices, edited the gpu.go to accommodate, and made a few tweaks here and there. Long story short: I recompiled llama_cpp and ollama, set a couple env variables, and got it to run.
https://github.com/remy415/ollama_tegra_fix.git
@dhiltgen I don't know how you want to approach incorporating this into Ollama for Jetson users, whether you want to incorporate it into the main branch or offer it as a patch or something.
@bnodnarb Try the patch at my github repo and see if it works for you, it worked on my 8gb Orin Nano.
@remy415 commented on GitHub (Jan 30, 2024):
@Q-point @bnodnarb Submitted a PR, should fix the Jetson issues.
@dhiltgen Not sure if you're tracking this or not :)
@Q-point commented on GitHub (Feb 3, 2024):
@remy415 Are you running this within a docker. Even with the instructions above I still get compile errors when issuing :
go generate ./...
@remy415 commented on GitHub (Feb 3, 2024):
No, I didn’t. Did you run go clean first? And did you pull from the repo linked in my pull request?
@remy415 commented on GitHub (Feb 3, 2024):
https://github.com/remy415/ollama.git
@remy415 commented on GitHub (Feb 3, 2024):
I need more info about your setup and what repo you’re using
@Q-point commented on GitHub (Feb 3, 2024):
@remy415
I'm using your repo; Simply followed your instructions on https://github.com/remy415/ollama_tegra_fix
Regarding the comment in bold: Also ensure: IMPORTANT -DCMAKE_CUDA_COMPILER=/usr/local/cuda/bin/nvcc is in the llm/generate/gen_linux.sh file under CUBLAS;
That is already included under:
@remy415 commented on GitHub (Feb 3, 2024):
Yes, I created the guide before I made the package. I need to update the guide. If you’re running Jetpack 5, it should work if you clone the repo and install as is. I am currently working back my edits to make it align with the default install. I think if you ensure your ld_library_path is set it should work
@remy415 commented on GitHub (Feb 3, 2024):
@Q-point use this repo in an empty folder, it’s the whole package https://github.com/remy415/ollama.git
@Q-point commented on GitHub (Feb 3, 2024):
OK, just got it working with that repo. https://github.com/remy415/ollama.git
@remy415 commented on GitHub (Feb 3, 2024):
Okay, let me know if the GPU acceleration works 😃
@Q-point commented on GitHub (Feb 3, 2024):
@remy415 commented on GitHub (Feb 3, 2024):
Awesome! Now I kinda wish I got an AGX instead of 4 Orin Nanos 😂🤣
@dhiltgen commented on GitHub (Feb 5, 2024):
@remy415 could you open a draft PR with your WIP so we can take a look and provide feedback?scratch that, you already did - just missed it.
@remy415 commented on GitHub (Feb 5, 2024):
@dhiltgen
https://github.com/ollama/ollama/pull/2279
@jhkuperus commented on GitHub (Feb 15, 2024):
I was reading along through this thread yesterday when I received my Jetson AGX Devkit. I couldn't get Ollama to use the CUDA-cores and then subsequently bricked my Jetson when upgrading stuff and trying to compile the version from #2279 .
Today I reflashed and upgraded everything on the Jetson, managed to get the version from @remy415's PR compiling and working when I start it as root. However, when I try to start it as a system service, it fails with a permission denied when trying to load the CUDA libraries. Here's the relevant output:
What am I missing here? The files in
/usr/local/cuda/lib64look like they have permissions that are set just fine:@remy415 commented on GitHub (Feb 15, 2024):
@jhkuperus which Jetson AGX do you have? What version of L4T/Jetpack? I see you’re running CUDA 12, did you install that manually? Jetson prior to Jetpack 6 comes with CUDA 11.4, and only the Orin series even supports Jetpack 6. Also JP6 is in beta, if you’re using it I recommend reflashing JP 5Disregard, see below.
Note that if you have continuous issues with JP6, I would recommend reflashing JP5 since JP6 is still under "developer preview" and isn't a full release yet.
@remy415 commented on GitHub (Feb 15, 2024):
@jhkuperus
oh I missed this part. Also I poked more into the errors being displayed and I'm frequently seeing one of two things pop up:
sudo init 6or whichever mechanism you prefersudo lsmod | grep -i nvidiaon your system. Note the-rinsudo modprobe -rhas the same effect asrmmodcommand.Other systems referenced
nvidia_uvm, I didn't have this module on my system but if it's present on yours you can use the same commands:Let me know if any of this works for you.
@jhkuperus commented on GitHub (Feb 15, 2024):
These are the details of my Jetson device:
Device used to test:
Jetson AGX Orin Developer Kit 64GB
Jetpack 6.0DP, L4T 36.2.0
CUDA 12.2.140
CUDA Capability Supported 8.7
Go version 1.21.6
Cmake 3.22.1
nvcc 12.2.140
I received it yesterday and I couldn't get compilation to work on the pre-flashed JP5. So I tried to upgrade the system, but that broke more things than it made better. Today I reflashed the device with JP6 DP. Compilation went smoothly and it runs perfectly when I run it as root, or as my normal user. If I try to start it as the
ollama-user, I get permission denied.This is the output of my
lsmodfornvidia-modules, I don't see a problem here:I'm thinking maybe it's a permission problem on a device, or socket or something like that?
@remy415 commented on GitHub (Feb 15, 2024):
I didn't realize you were running it as a separate user. I'm assuming you added the user manually? May I see the contents of your service file please?
@jhkuperus commented on GitHub (Feb 15, 2024):
Hah! Just figured it out. The problem is actually quite simple: the
ollamauser had to be a member of thevideogroup to fix it.This user is added by the installation script provided by the Ollama repository itself. Maybe this is something we can still add to your PR?
@remy415 commented on GitHub (Feb 15, 2024):
The official script adds the
ollamauser to therendergroup. If/when the PR is merged with the main branch, the rest should automatically work as the PR only affects which shared libraries are loaded when the binary is executed.From the source code
@davidtheITguy commented on GitHub (Feb 18, 2024):
@Q-point Hi, could you please let me know what you did to resolve these build errors, I'm where you were: https://github.com/ollama/ollama/issues/1979#issuecomment-1925028134 Many thanks.
@Q-point commented on GitHub (Feb 18, 2024):
@davidtheITguy I used the suggested repo https://github.com/remy415/ollama.git . Compiled from source. You'll have to install go compiler.
@davidtheITguy commented on GitHub (Feb 18, 2024):
Yes sir, did all that. Also overlaid the
[package_cudart_build]directory per the https://github.com/remy415/ollama_tegra_fix, loaded the suggested env vars too. Compiled from source and got the same errors you posted https://github.com/ollama/ollama/issues/1979#issuecomment-1925028134.Appreciate the response and I'll keep plugging away good to know I'm on the right track.
@remy415 commented on GitHub (Feb 18, 2024):
@davidtheITguy what error messages are you getting? Is it the ones about the half precision posted above? When did you last pull the repo? It should work with any JP5 Jetson
@davidtheITguy commented on GitHub (Feb 18, 2024):
Apologize in advance for posting the entire error output, but want you to see. Actually got pretty far into the build before this happened:
I tried to follow your directions to the letter. Intergated the ollama_tegra_fix repo, cloned your latest fork, etc.
Tool chain versions:
@remy415 commented on GitHub (Feb 18, 2024):
@davidtheITguy Okay that error you're seeing happens to me when I try to compile and the cuda architectures are incorrectly set. Are you manually setting the architectures? Ensure you clear the variable
CUDA_ARCHITECTURESwith the commandexport CUDA_ARCHITECTURES=""and let the script configure it automatically. I've incorporated the necessary fixes into the code base and eliminated the need for most of the ollama_tegra_fix page. The only env variables that are absolutely critical are:export LD_LIBRARY_PATH="/usr/local/cuda/lib64:/usr/local/cuda/compat:/usr/local/cuda/include"and the one regarding cpu generate is just helpful to Jetson users as it skips the AVX builds
export OLLAMA_SKIP_CPU_GENERATE="yes"I would highly recommend deleting your ollama folder completely, re-clone
https://github.com/remy415/ollama.git, ensure these 2 env variables only are set, and rungo generate ./... && go build .For reference, during the build process you should see a reference to CUDA_ARCHITECTURES. The only ones that should be loaded depend on your Jetson device and Jetpack version but
it should detect it automaticallyif you are running any of the 3 Jetpack versions listed here:Nano/TX1 = 5.3, TX2 = 6.2, Xavier = 7.2, Orin = 8.7
L4T_VERSION.major >= 36: # JetPack 6
CUDA_ARCHITECTURES = [87]L4T_VERSION.major >= 34: # JetPack 5
CUDA_ARCHITECTURES = [72, 87]L4T_VERSION.major == 32: # JetPack 4
CUDA_ARCHITECTURES = [53, 62, 72]Edited for formatting
@remy415 commented on GitHub (Feb 18, 2024):
I just noticed this. Sorry, I haven't cleaned up the documentation yet as I'm still working on the code itself to get it cleaned up. This step is no longer needed as that package is no longer valid.
@davidtheITguy commented on GitHub (Feb 19, 2024):
Understood and thank you ill circle back when the build is successful
@davidtheITguy commented on GitHub (Feb 19, 2024):
Looks like the build is very close, but not quite there yet.
go generate ./...seems to have completed successfully.The
go build .errors out with what appears to be an external reference issue:build github.com/jmorganca/ollama: cannot load crypto/ecdh: malformed module path "crypto/ecdh": missing dot in first path element.My environment:
LD_LIBRARY_PATH=/usr/local/cuda/lib64:/usr/local/cuda/compat:/usr/local/cuda/includeOLLAMA_SKIP_CPU_GENERATE=yesBuild steps:
conda activate ollama # activate private build environmentgit clone --depth=1 --recursive https://github.com/remy415/ollamago generate ./...go build .Debug steps: I ran a
go get -uand thengo mod tidyto try to resolve any reference issues here is the output:I couldn't find any local reference to "crypto/ecdh" so wondering if it is an external dependency...
@remy415 commented on GitHub (Feb 19, 2024):
I found the below env on a forum post about some new changes in Go 1.13. Odd that you’re getting that error, I haven’t seen it before. I think it may have something to do with your conda environment. Try checking go env before and after activating your conda environment.
Try this:
export GO111MODULE=off@davidtheITguy commented on GitHub (Feb 20, 2024):
Ok got it to compile. Made a rookie move, my version of golang was something like 1.13, go.mod says 1.21 (but doesn't enforce for some reason). Upgrading go to 1.21.7 worked for me.
So for the next person just to sum up what's needed as of this comment to compile the fork for Nvidia GPUs:
git clone --depth=1 --recursive https://github.com/remy415/ollamacd ollamago generate ./...go build .BUT still no CUDA/GPU unfortunately, looks a perms issue on the CUDA libs:
I'll keep plugging away hopefully, this will all get settled and merged at some point.
Many thanks @remy415 for your assistance on the build.
@remy415 commented on GitHub (Feb 20, 2024):
@davidtheITguy the user running the binary needs to be part of the
renderuser group, and thevideouser group.The command follows this syntax:
sudo usermod -a -G examplegroup exampleusername.sudo usermod -a -G render <user name>sudo usermod -a -G video <user name>If you’re running it as a service using the ollama user, then add the ollama user. If you’re running directly from CLI ensure your user is part of the group.
@davidtheITguy commented on GitHub (Feb 20, 2024):
@remy415 Bingo! That did it, thanks for the help!
@davidtheITguy commented on GitHub (Feb 20, 2024):
@remy415 running lama2-7b cli is blazing fast on my Jetson Orin AGX 64GB. Here's to hoping you merge this soon and again many thanks for your assistance!
@remy415 commented on GitHub (Feb 20, 2024):
@davidtheITguy glad I could help! I’m working on fixing some quirks in the differences between the two CUDA libraries, hope to be done soon.
@telemetrieTP23 commented on GitHub (Feb 21, 2024):
Thanks for your help ( https://github.com/ollama/ollama/issues/2491#issuecomment-1954615027 ).
There with "preliminary build" you meant that i should build your fork ?
I have cmake version 3.28.2
go version go1.27.1 linux/arm64
gcc (Ubuntu 9.4.0-1ubuntu1~20.04.2) 9.4.0
tried your tips from there:
export LD_LIBRARY_PATH="/usr/local/cuda/lib64:/usr/local/cuda/compat:/usr/local/cuda/include"
export OLLAMA_SKIP_CPU_GENERATE="1"
export CUDA_ARCHITECTURES="72;87"
But still when i start "go generate ./..."
Then it generates '-DCMAKE_CUDA_ARCHITECTURES=50;52;61;70;75;80' in the CMAKE _DEFS
and the errors like here https://github.com/ollama/ollama/issues/1979#issuecomment-1951463913
where produced....
@remy415 commented on GitHub (Feb 21, 2024):
@telemetrieTP23 i responded to your other post 😀
I had a typo in my original response and it’s updated now.
@davidtheITguy commented on GitHub (Feb 22, 2024):
@remy415 Hey I had a quick follow up question to all this. Your build is working great on my Orin AGX!
However, after a lot of testing, it appears that (believe it or not) we are offloading to only one GPU (there are two separate devices on the Orin). I've confirmed this with the standard Python HF libs which do use both GPUs during inference.
Here is some output from the ollama service:
Given beggars can't be choosers, have you run across this issue? Is there a switch perhaps for CUDA=all or similar?
Thanks!
@remy415 commented on GitHub (Feb 22, 2024):
@davidtheITguy According to the NVidia Jetson AGX Orin Technical documentation at https://www.nvidia.com/content/dam/en-zz/Solutions/gtcf21/jetson-orin/nvidia-jetson-agx-orin-technical-brief.pdf, it would seem that it's two graphics compute clusters with a unified front end.
I don't have the AGX Orin so I can't personally confirm, but one way I would do it is to run a device query that shows the getDeviceCount results (The ollama service logs do show the results of that API call further up towards the top,
export OLLAMA_DEBUG="1"if you don't see it). Another way is I would run the python script you referenced while running JTOP, and compare the reported GPU usage when running ollama.Last, the lines beginning with
ggml_init_cublasare actually logs reported from the llama.cpp back end. The ollama code we have is purely for querying device information and verifying the device is present & ready to go. Most (if not all) of the performance-related code is in the llama.cpp back end, but based on what your screenshots are showing I believe your AGX Orin is firing on all cylinders.@davidtheITguy commented on GitHub (Feb 22, 2024):
@remy415 Got it, ty. Weird that only one GPU works (I definitely get both with the python/HF scripts). I'll report back when able.
@davidtheITguy commented on GitHub (Feb 22, 2024):
@remy415 quick clarify. You are correct, it does appear that both GPUs are exposed as one interface. The difference appears to be the "GPU Shared RAM" capability not being utilized by the Lama.cpp back end in this case. I'll keep digging.
@remy415 commented on GitHub (Feb 22, 2024):
@davidtheITguy I don't know which model you have loaded but your JTOP is reporting 9.1G GPU Shared Ram being used, which is definitely more than a model like Mistral 7b uses (typically it uses ~4G RAM). I think your hardware is being fully leveraged
@ToeiRei commented on GitHub (Mar 10, 2024):
I cannot reproduce your builds using go version go1.22.1 linux/arm64. Does anyone have a binary for an Orin NX 16 GB?
@remy415 commented on GitHub (Mar 10, 2024):
It should build on Orin NX. Please clone the repo at my repo. Then
cd ollamago generate ./… && go build .When the build is done, either put it in a location on your path or use
./ollama serveand./ollama run <model>. If you have any issues, please include logs so we can troubleshoot.@ToeiRei commented on GitHub (Mar 10, 2024):
Thanks @remy415 , already trying with your repo. Error on go build is
@remy415 commented on GitHub (Mar 10, 2024):
@ToeiRei Yes, those are compiler warnings, but they are not critical errors. Those particular warnings are present even in the Ollama main builds. The binary should still have compiled and should work for you.
@ToeiRei commented on GitHub (Mar 10, 2024):
my bad. I had expected something like a compile done as it just showed the warnings and a prompt.
@remy415 commented on GitHub (Mar 10, 2024):
No worries, I fell down the same rabbit hole myself
@ToeiRei commented on GitHub (Mar 10, 2024):
Thanks. After failing to restart ollama due to a lack of caffeine, we're finally cooking with
gas... errcuda@remy415 commented on GitHub (Mar 10, 2024):
That's great! How is the performance on the Orin NX?
@ToeiRei commented on GitHub (Mar 10, 2024):
It's definitely slower than my RTX4070, but it does not feel too bad. Like a person live typing on a 13b model.
but somehow there's got to be a problem with the model as it says
Error: open /vicuna:13b: no such file or directoryon ollama create - but running the model directly works fine.@remy415 commented on GitHub (Mar 10, 2024):
While in the ollama directory, and with
ollama serverunning in the background:./ollama show vicuna:13b --modelfile > Modelfile(use whatever name you want forModelfile)Edit the model file:
vim Modelfile(or whichever text editor you want to use) -> addPARAMETER num_gpu 999underneath the "FROM" lineCreate the model
./ollama create <NEW MODEL NAME OF YOUR CHOICE> -f ./ModelfileIt would seem the model discovery logic is currently broken as the Modelfile I had made previously also failed with the same error as yours, and I had to use the above commands to generate a new "template" with the correct path auto-populated.
@UserName-wang commented on GitHub (Mar 23, 2024):
I got some error messages:
=> ERROR [ollama-cuda-l4t-base 9/9] RUN go build . 20.3s
Dockerfile2:27
25 | WORKDIR /go/src/github.com/jmorganca/ollama
26 | RUN go generate ./...
27 | >>> RUN go build .
28 |
29 | # Runtime stages
ERROR: failed to solve: process "/bin/sh -c go build ." did not complete successfully: exit code: 1
The docker file I use:
Dockerfile2.txt
@remy415 , Could you please have a look at my error message and give me a hint? thank you!
@hangxingliu commented on GitHub (Mar 23, 2024):
Hi @UserName-wang . I guess that your issue is caused by @remy415 still developing on his main branch (forgot to push the changes for importing
assetspackage)I have run @remy415 's great forked version of ollama on my Jetson AGX Orin with L4T 36.2.0 (Jetpack 6.0 DP) few days. It works great with
llama2:13b-chat-fp16andgemma:7b-instruct-fp16and it is very fast for daily usage. The ollama I built is based on this commit:https://github.com/remy415/ollama/tree/23204a77d27cfdfbf80e0bdaf96116fcf1255039
@remy415 commented on GitHub (Mar 23, 2024):
@UserName-wang @hangxingliu the fork is currently broken due to an incomplete merge. I’m working with @dhiltgen to get it fixed.
@remy415 commented on GitHub (Mar 23, 2024):
@UserName-wang @hangxingliu
I have put in a temp fix here. I am not able to run a test build right now.
Try this branch here
@UserName-wang commented on GitHub (Mar 24, 2024):
@remy415 , Thank you a lot! now it works! and I'm sure ollama now running on GPU. I tried gemma and llama2 and running fast!
And thank you! @hangxingliu ! I hope some day I can be like help full like you to help someone else!
I used **go generate ./... && go build . t**o build this package on host( agx orin) it works. but faild to run in docker. the build process is successful (and detected GPU). but failed to run command: ollama run gemma:2b. the error message attached.
ollama_Docker_build_log.txt
Can you please have a look and give me some suggestions if you have time? thank you!
@remy415 commented on GitHub (Mar 24, 2024):
@UserName-wang Just to confirm:
go generate ./... && go build .from your regular command line, and tried to runollama run gemma:2bfrom a Docker container?I'm going to look at the gpu.go file and see why it didn't print the path to the library it loads as well.
@remy415 commented on GitHub (Mar 24, 2024):
@UserName-wang Just FYI I haven't worked out running this on Docker containers as the Jetson is a bit of an oddity with that. I suggest you look at dustynv's Jetson containers for running GPU accelerated stuff in docker containers as the runtime needs a special configuration to work on Jetsons.
@UserName-wang commented on GitHub (Mar 24, 2024):
answer to question 1, yes! and the docker already detected CUDA.
answer to question 2, yes!
@remy415 commented on GitHub (Mar 24, 2024):
@UserName-wang ok so it runs outside of Docker, but does not run inside of Docker? The error message in your log suggested it is missing a driver, that’s why I asked if you configured it for Docker by following NVidias instructions for CUDA. I don’t have Jetpack 6 installed yet, I was waiting for the official release
@remy415 commented on GitHub (Mar 25, 2024):
Fixed with merge of #2279
@remy415 commented on GitHub (Apr 15, 2024):
@UserName-wang dusty-nv merged a PR with a container for Ollama. You can find it on his github page
@UserName-wang commented on GitHub (Apr 20, 2024):
Yes! I tested and it works!Thank you for your information!在 2024年4月15日,09:20,Jeremy @.***> 写道:
@UserName-wang dusty-nv merged a PR with a container for Ollama. You can find it on his github page
—Reply to this email directly, view it on GitHub, or unsubscribe.You are receiving this because you were mentioned.Message ID: @.***>