mirror of
https://github.com/ollama/ollama.git
synced 2026-05-07 00:22:43 -05:00
Open
opened 2026-04-28 03:57:41 -05:00 by GiteaMirror
·
124 comments
No Branch/Tag Specified
main
hoyyeva/anthropic-local-image-path
dhiltgen/ci
dhiltgen/llama-runner
parth-remove-claude-desktop-launch
hoyyeva/anthropic-reference-images-path
parth-anthropic-reference-images-path
brucemacd/download-before-remove
hoyyeva/editor-config-repair
parth-mlx-decode-checkpoints
parth-launch-codex-app
hoyyeva/fix-codex-model-metadata-warning
hoyyeva/qwen
parth/hide-claude-desktop-till-release
hoyyeva/opencode-image-modality
parth-add-claude-code-autoinstall
release_v0.22.0
pdevine/manifest-list
codex/fix-codex-model-metadata-warning
pdevine/addressable-manifest
brucemacd/launch-fetch-reccomended
jmorganca/llama-compat
launch-copilot-cli
hoyyeva/opencode-thinking
release_v0.20.7
parth-auto-save-backup
parth-test
jmorganca/gemma4-audio-replacements
fix-manifest-digest-on-pull
hoyyeva/vscode-improve
brucemacd/install-server-wait
parth/update-claude-docs
brucemac/start-ap-install
pdevine/mlx-update
pdevine/qwen35_vision
drifkin/api-show-fallback
mintlify/image-generation-1773352582
hoyyeva/server-context-length-local-config
jmorganca/faster-reptition-penalties
jmorganca/convert-nemotron
parth-pi-thinking
pdevine/sampling-penalties
jmorganca/fix-create-quantization-memory
dongchen/resumable_transfer_fix
pdevine/sampling-cache-error
jessegross/mlx-usage
hoyyeva/openclaw-config
hoyyeva/app-html
pdevine/qwen3next
brucemacd/sign-sh-install
brucemacd/tui-update
brucemacd/usage-api
jmorganca/launch-empty
fix-app-dist-embed
mxyng/mlx-compile
mxyng/mlx-quant
mxyng/mlx-glm4.7
mxyng/mlx
brucemacd/simplify-model-picker
jmorganca/qwen3-concurrent
fix-glm-4.7-flash-mla-config
drifkin/qwen3-coder-opening-tag
brucemacd/usage-cli
fix-cuda12-fattn-shmem
ollama-imagegen-docs
parth/fix-multiline-inputs
brucemacd/config-docs
mxyng/model-files
mxyng/simple-execute
fix-imagegen-ollama-models
mxyng/async-upload
jmorganca/lazy-no-dtype-changes
imagegen-auto-detect-create
parth/decrease-concurrent-download-hf
fix-mlx-quantize-init
jmorganca/x-cleanup
usage
imagegen-readme
jmorganca/glm-image
mlx-gpu-cd
jmorganca/imagegen-modelfile
parth/agent-skills
parth/agent-allowlist
parth/signed-in-offline
parth/agents
parth/fix-context-chopping
improve-cloud-flow
parth/add-models-websearch
parth/prompt-renderer-mcp
jmorganca/native-settings
jmorganca/download-stream-hash
jmorganca/client2-rebased
brucemacd/oai-chat-req-multipart
jessegross/multi_chunk_reserve
grace/additional-omit-empty
grace/mistral-3-large
mxyng/tokenizer2
mxyng/tokenizer
jessegross/flash
hoyyeva/windows-nacked-app
mxyng/cleanup-attention
grace/deepseek-parser
hoyyeva/remember-unsent-prompt
parth/add-lfs-pointer-error-conversion
parth/olmo2-test2
hoyyeva/ollama-launchagent-plist
nicole/olmo-model
parth/olmo-test
mxyng/remove-embedded
parth/render-template
jmorganca/intellect-3
parth/remove-prealloc-linter
jmorganca/cmd-eval
nicole/nomic-embed-text-fix
mxyng/lint-2
hoyyeva/add-gemini-3-pro-preview
hoyyeva/load-model-list
mxyng/expand-path
mxyng/environ-2
hoyyeva/deeplink-json-encoding
parth/improve-tool-calling-tests
hoyyeva/conversation
hoyyeva/assistant-edit-response
hoyyeva/thinking
origin/brucemacd/invalid-char-i-err
parth/improve-tool-calling
jmorganca/required-omitempty
grace/qwen3-vl-tests
mxyng/iter-client
parth/docs-readme
nicole/embed-test
pdevine/integration-benchstat
parth/remove-generate-cmd
parth/add-toolcall-id
mxyng/server-tests
jmorganca/glm-4.6
jmorganca/gin-h-compat
drifkin/stable-tool-args
pdevine/qwen3-more-thinking
parth/add-websearch-client
nicole/websearch_local
jmorganca/qwen3-coder-updates
grace/deepseek-v3-migration-tests
mxyng/fix-create
jmorganca/cloud-errors
pdevine/parser-tidy
revert-12233-parth/simplify-entrypoints-runner
parth/enable-so-gpt-oss
brucemacd/qwen3vl
jmorganca/readme-simplify
parth/gpt-oss-structured-outputs
revert-12039-jmorganca/tools-braces
mxyng/embeddings
mxyng/gguf
mxyng/benchmark
mxyng/types-null
parth/move-parsing
mxyng/gemma2
jmorganca/docs
mxyng/16-bit
mxyng/create-stdin
pdevine/authorizedkeys
mxyng/quant
parth/opt-in-error-context-window
brucemacd/cache-models
brucemacd/runner-completion
jmorganca/llama-update-6
brucemacd/benchmark-list
brucemacd/partial-read-caps
parth/deepseek-r1-tools
mxyng/omit-array
parth/tool-prefix-temp
brucemacd/runner-test
jmorganca/qwen25vl
brucemacd/model-forward-test-ext
parth/python-function-parsing
jmorganca/cuda-compression-none
drifkin/num-parallel
drifkin/chat-truncation-fix
jmorganca/sync
parth/python-tools-calling
drifkin/array-head-count
brucemacd/create-no-loop
parth/server-enable-content-stream-with-tools
qwen25omni
mxyng/v3
brucemacd/ropeconfig
jmorganca/silence-tokenizer
parth/sample-so-test
parth/sampling-structured-outputs
brucemacd/doc-go-engine
parth/constrained-sampling-json
jmorganca/mistral-wip
brucemacd/mistral-small-convert
parth/sample-unmarshal-json-for-params
brucemacd/jomorganca/mistral
pdevine/bfloat16
jmorganca/mistral
brucemacd/mistral
pdevine/logging
parth/sample-correctness-fix
parth/sample-fix-sorting
jmorgan/sample-fix-sorting-extras
jmorganca/temp-0-images
brucemacd/parallel-embed-models
brucemacd/shim-grammar
jmorganca/fix-gguf-error
bmizerany/nameswork
jmorganca/faster-releases
bmizerany/validatenames
brucemacd/err-no-vocab
brucemacd/rope-config
brucemacd/err-hint
brucemacd/qwen2_5
brucemacd/logprobs
brucemacd/new_runner_graph_bench
progress-flicker
brucemacd/forward-test
brucemacd/go_qwen2
pdevine/gemma2
jmorganca/add-missing-symlink-eval
mxyng/next-debug
parth/set-context-size-openai
brucemacd/next-bpe-bench
brucemacd/next-bpe-test
brucemacd/new_runner_e2e
brucemacd/new_runner_qwen2
pdevine/convert-cohere2
brucemacd/convert-cli
parth/log-probs
mxyng/next-mlx
mxyng/cmd-history
parth/templating
parth/tokenize-detokenize
brucemacd/check-key-register
bmizerany/grammar
jmorganca/vendor-081b29bd
mxyng/func-checks
jmorganca/fix-null-format
parth/fix-default-to-warn-json
jmorganca/qwen2vl
jmorganca/no-concat
parth/cmd-cleanup-SO
brucemacd/check-key-register-structured-err
parth/openai-stream-usage
parth/fix-referencing-so
stream-tools-stop
jmorganca/degin-1
brucemacd/install-path-clean
brucemacd/push-name-validation
brucemacd/browser-key-register
jmorganca/openai-fix-first-message
jmorganca/fix-proxy
jessegross/sample
parth/disallow-streaming-tools
dhiltgen/remove_submodule
jmorganca/ga
jmorganca/mllama
pdevine/newlines
pdevine/geems-2b
jmorganca/llama-bump
mxyng/modelname-7
mxyng/gin-slog
mxyng/modelname-6
jyan/convert-prog
jyan/quant5
paligemma-support
pdevine/import-docs
jmorganca/openai-context
jyan/paligemma
jyan/p2
jyan/palitest
bmizerany/embedspeedup
jmorganca/llama-vit
brucemacd/allow-ollama
royh/ep-methods
royh/whisper
mxyng/api-models
mxyng/fix-memory
jyan/q4_4/8
jyan/ollama-v
royh/stream-tools
roy-embed-parallel
bmizerany/hrm
revert-5963-revert-5924-mxyng/llama3.1-rope
royh/embed-viz
jyan/local2
jyan/auth
jyan/local
jyan/parse-temp
jmorganca/template-mistral
jyan/reord-g
royh-openai-suffixdocs
royh-imgembed
royh-embed-parallel
jyan/quant4
royh-precision
jyan/progress
pdevine/fix-template
jyan/quant3
pdevine/ggla
mxyng/update-registry-domain
jmorganca/ggml-static
mxyng/create-context
jyan/v0.146
mxyng/layers-from-files
build_dist
bmizerany/noseek
royh-ls
royh-name
timeout
mxyng/server-timestamp
bmizerany/nosillyggufslurps
royh-params
jmorganca/llama-cpp-7c26775
royh-openai-delete
royh-show-rigid
jmorganca/enable-fa
jmorganca/no-error-template
jyan/format
royh-testdelete
bmizerany/fastverify
language_support
pdevine/ps-glitches
brucemacd/tokenize
bruce/iq-quants
bmizerany/filepathwithcoloninhost
mxyng/split-bin
bmizerany/client-registry
jmorganca/if-none-match
native
jmorganca/native
jmorganca/batch-embeddings
jmorganca/initcmake
jmorganca/mm
pdevine/showggmlinfo
modenameenforcealphanum
bmizerany/modenameenforcealphanum
jmorganca/done-reason
jmorganca/llama-cpp-8960fe8
ollama.com
bmizerany/filepathnobuild
bmizerany/types/model/defaultfix
rmdisplaylong
nogogen
bmizerany/x
modelfile-readme
bmizerany/replacecolon
jmorganca/limit
jmorganca/execstack
jmorganca/replace-assets
mxyng/tune-concurrency
jmorganca/testing
whitespace-detection
jmorganca/options
upgrade-all
scratch
cuda-search
mattw/airenamer
mattw/allmodelsonhuggingface
mattw/quantcontext
mattw/whatneedstorun
brucemacd/llama-mem-calc
mattw/faq-context
mattw/communitylinks
mattw/noprune
mattw/python-functioncalling
rename
mxyng/install
pulse
remove-first
editor
mattw/selfqueryingretrieval
cgo
mattw/howtoquant
api
matt/streamingapi
format-config
mxyng/extra-args
shell
update-nous-hermes
cp-model
upload-progress
fix-unknown-model
fix-model-names
delete-fix
insecure-registry
ls
deletemodels
progressbar
readme-updates
license-layers
skip-list
list-models
modelpath
matt/examplemodelfiles
distribution
go-opts
v0.30.0-rc3
v0.30.0-rc2
v0.30.0-rc1
v0.30.0-rc0
v0.23.1
v0.23.1-rc0
v0.23.0
v0.23.0-rc0
v0.22.1
v0.22.1-rc1
v0.22.1-rc0
v0.22.0
v0.22.0-rc1
v0.21.3-rc0
v0.21.2-rc1
v0.21.2
v0.21.2-rc0
v0.21.1
v0.21.1-rc1
v0.21.1-rc0
v0.21.0
v0.21.0-rc1
v0.21.0-rc0
v0.20.8-rc0
v0.20.7
v0.20.7-rc1
v0.20.7-rc0
v0.20.6
v0.20.6-rc1
v0.20.6-rc0
v0.20.5
v0.20.5-rc2
v0.20.5-rc1
v0.20.5-rc0
v0.20.4
v0.20.4-rc2
v0.20.4-rc1
v0.20.4-rc0
v0.20.3
v0.20.3-rc0
v0.20.2
v0.20.1
v0.20.1-rc2
v0.20.1-rc1
v0.20.1-rc0
v0.20.0
v0.20.0-rc1
v0.20.0-rc0
v0.19.0
v0.19.0-rc2
v0.19.0-rc1
v0.19.0-rc0
v0.18.4-rc1
v0.18.4-rc0
v0.18.3
v0.18.3-rc2
v0.18.3-rc1
v0.18.3-rc0
v0.18.2
v0.18.2-rc1
v0.18.2-rc0
v0.18.1
v0.18.1-rc1
v0.18.1-rc0
v0.18.0
v0.18.0-rc2
v0.18.0-rc1
v0.18.0-rc0
v0.17.8-rc4
v0.17.8-rc3
v0.17.8-rc2
v0.17.8-rc1
v0.17.8-rc0
v0.17.7
v0.17.7-rc2
v0.17.7-rc1
v0.17.7-rc0
v0.17.6
v0.17.5
v0.17.4
v0.17.3
v0.17.2
v0.17.1
v0.17.1-rc2
v0.17.1-rc1
v0.17.1-rc0
v0.17.0
v0.17.0-rc2
v0.17.0-rc1
v0.17.0-rc0
v0.16.3
v0.16.3-rc2
v0.16.3-rc1
v0.16.3-rc0
v0.16.2
v0.16.2-rc0
v0.16.1
v0.16.0
v0.16.0-rc2
v0.16.0-rc0
v0.16.0-rc1
v0.15.6
v0.15.5
v0.15.5-rc5
v0.15.5-rc4
v0.15.5-rc3
v0.15.5-rc2
v0.15.5-rc1
v0.15.5-rc0
v0.15.4
v0.15.3
v0.15.2
v0.15.1
v0.15.1-rc1
v0.15.1-rc0
v0.15.0-rc6
v0.15.0
v0.15.0-rc5
v0.15.0-rc4
v0.15.0-rc3
v0.15.0-rc2
v0.15.0-rc1
v0.15.0-rc0
v0.14.3
v0.14.3-rc3
v0.14.3-rc2
v0.14.3-rc1
v0.14.3-rc0
v0.14.2
v0.14.2-rc1
v0.14.2-rc0
v0.14.1
v0.14.0-rc11
v0.14.0
v0.14.0-rc10
v0.14.0-rc9
v0.14.0-rc8
v0.14.0-rc7
v0.14.0-rc6
v0.14.0-rc5
v0.14.0-rc4
v0.14.0-rc3
v0.14.0-rc2
v0.14.0-rc1
v0.14.0-rc0
v0.13.5
v0.13.5-rc1
v0.13.5-rc0
v0.13.4-rc2
v0.13.4
v0.13.4-rc1
v0.13.4-rc0
v0.13.3
v0.13.3-rc1
v0.13.3-rc0
v0.13.2
v0.13.2-rc2
v0.13.2-rc1
v0.13.2-rc0
v0.13.1
v0.13.1-rc2
v0.13.1-rc1
v0.13.1-rc0
v0.13.0
v0.13.0-rc0
v0.12.11
v0.12.11-rc1
v0.12.11-rc0
v0.12.10
v0.12.10-rc1
v0.12.10-rc0
v0.12.9-rc0
v0.12.9
v0.12.8
v0.12.8-rc0
v0.12.7
v0.12.7-rc1
v0.12.7-rc0
v0.12.7-citest0
v0.12.6
v0.12.6-rc1
v0.12.6-rc0
v0.12.5
v0.12.5-rc0
v0.12.4
v0.12.4-rc7
v0.12.4-rc6
v0.12.4-rc5
v0.12.4-rc4
v0.12.4-rc3
v0.12.4-rc2
v0.12.4-rc1
v0.12.4-rc0
v0.12.3
v0.12.2
v0.12.2-rc0
v0.12.1
v0.12.1-rc1
v0.12.1-rc2
v0.12.1-rc0
v0.12.0
v0.12.0-rc1
v0.12.0-rc0
v0.11.11
v0.11.11-rc3
v0.11.11-rc2
v0.11.11-rc1
v0.11.11-rc0
v0.11.10
v0.11.9
v0.11.9-rc0
v0.11.8
v0.11.8-rc0
v0.11.7-rc1
v0.11.7-rc0
v0.11.7
v0.11.6
v0.11.6-rc0
v0.11.5-rc4
v0.11.5-rc3
v0.11.5
v0.11.5-rc5
v0.11.5-rc2
v0.11.5-rc1
v0.11.5-rc0
v0.11.4
v0.11.4-rc0
v0.11.3
v0.11.3-rc0
v0.11.2
v0.11.1
v0.11.0-rc0
v0.11.0-rc1
v0.11.0-rc2
v0.11.0
v0.10.2-int1
v0.10.1
v0.10.0
v0.10.0-rc4
v0.10.0-rc3
v0.10.0-rc2
v0.10.0-rc1
v0.10.0-rc0
v0.9.7-rc1
v0.9.7-rc0
v0.9.6
v0.9.6-rc0
v0.9.6-ci0
v0.9.5
v0.9.4-rc5
v0.9.4-rc6
v0.9.4
v0.9.4-rc3
v0.9.4-rc4
v0.9.4-rc1
v0.9.4-rc2
v0.9.4-rc0
v0.9.3
v0.9.3-rc5
v0.9.4-citest0
v0.9.3-rc4
v0.9.3-rc3
v0.9.3-rc2
v0.9.3-rc1
v0.9.3-rc0
v0.9.2
v0.9.1
v0.9.1-rc1
v0.9.1-rc0
v0.9.1-ci1
v0.9.1-ci0
v0.9.0
v0.9.0-rc0
v0.8.0
v0.8.0-rc0
v0.7.1-rc2
v0.7.1
v0.7.1-rc1
v0.7.1-rc0
v0.7.0
v0.7.0-rc1
v0.7.0-rc0
v0.6.9-rc0
v0.6.8
v0.6.8-rc0
v0.6.7
v0.6.7-rc2
v0.6.7-rc1
v0.6.7-rc0
v0.6.6
v0.6.6-rc2
v0.6.6-rc1
v0.6.6-rc0
v0.6.5-rc1
v0.6.5
v0.6.5-rc0
v0.6.4-rc0
v0.6.4
v0.6.3-rc1
v0.6.3
v0.6.3-rc0
v0.6.2
v0.6.2-rc0
v0.6.1
v0.6.1-rc0
v0.6.0-rc0
v0.6.0
v0.5.14-rc0
v0.5.13
v0.5.13-rc6
v0.5.13-rc5
v0.5.13-rc4
v0.5.13-rc3
v0.5.13-rc2
v0.5.13-rc1
v0.5.13-rc0
v0.5.12
v0.5.12-rc1
v0.5.12-rc0
v0.5.11
v0.5.10
v0.5.9
v0.5.9-rc0
v0.5.8-rc13
v0.5.8
v0.5.8-rc12
v0.5.8-rc11
v0.5.8-rc10
v0.5.8-rc9
v0.5.8-rc8
v0.5.8-rc7
v0.5.8-rc6
v0.5.8-rc5
v0.5.8-rc4
v0.5.8-rc3
v0.5.8-rc2
v0.5.8-rc1
v0.5.8-rc0
v0.5.7
v0.5.6
v0.5.5
v0.5.5-rc0
v0.5.4
v0.5.3
v0.5.3-rc0
v0.5.2
v0.5.2-rc3
v0.5.2-rc2
v0.5.2-rc1
v0.5.2-rc0
v0.5.1
v0.5.0
v0.5.0-rc1
v0.4.8-rc0
v0.4.7
v0.4.6
v0.4.5
v0.4.4
v0.4.3
v0.4.3-rc0
v0.4.2
v0.4.2-rc1
v0.4.2-rc0
v0.4.1
v0.4.1-rc0
v0.4.0
v0.4.0-rc8
v0.4.0-rc7
v0.4.0-rc6
v0.4.0-rc5
v0.4.0-rc4
v0.4.0-rc3
v0.4.0-rc2
v0.4.0-rc1
v0.4.0-rc0
v0.4.0-ci3
v0.3.14
v0.3.14-rc0
v0.3.13
v0.3.12
v0.3.12-rc5
v0.3.12-rc4
v0.3.12-rc3
v0.3.12-rc2
v0.3.12-rc1
v0.3.11
v0.3.11-rc4
v0.3.11-rc3
v0.3.11-rc2
v0.3.11-rc1
v0.3.10
v0.3.10-rc1
v0.3.9
v0.3.8
v0.3.7
v0.3.7-rc6
v0.3.7-rc5
v0.3.7-rc4
v0.3.7-rc3
v0.3.7-rc2
v0.3.7-rc1
v0.3.6
v0.3.5
v0.3.4
v0.3.3
v0.3.2
v0.3.1
v0.3.0
v0.2.8
v0.2.8-rc2
v0.2.8-rc1
v0.2.7
v0.2.6
v0.2.5
v0.2.4
v0.2.3
v0.2.2
v0.2.2-rc2
v0.2.2-rc1
v0.2.1
v0.2.0
v0.1.49-rc14
v0.1.49-rc13
v0.1.49-rc12
v0.1.49-rc11
v0.1.49-rc10
v0.1.49-rc9
v0.1.49-rc8
v0.1.49-rc7
v0.1.49-rc6
v0.1.49-rc4
v0.1.49-rc5
v0.1.49-rc3
v0.1.49-rc2
v0.1.49-rc1
v0.1.48
v0.1.47
v0.1.46
v0.1.45-rc5
v0.1.45
v0.1.45-rc4
v0.1.45-rc3
v0.1.45-rc2
v0.1.45-rc1
v0.1.44
v0.1.43
v0.1.42
v0.1.41
v0.1.40
v0.1.40-rc1
v0.1.39
v0.1.39-rc2
v0.1.39-rc1
v0.1.38
v0.1.37
v0.1.36
v0.1.35
v0.1.35-rc1
v0.1.34
v0.1.34-rc1
v0.1.33
v0.1.33-rc7
v0.1.33-rc6
v0.1.33-rc5
v0.1.33-rc4
v0.1.33-rc3
v0.1.33-rc2
v0.1.33-rc1
v0.1.32
v0.1.32-rc2
v0.1.32-rc1
v0.1.31
v0.1.30
v0.1.29
v0.1.28
v0.1.27
v0.1.26
v0.1.25
v0.1.24
v0.1.23
v0.1.22
v0.1.21
v0.1.20
v0.1.19
v0.1.18
v0.1.17
v0.1.16
v0.1.15
v0.1.14
v0.1.13
v0.1.12
v0.1.11
v0.1.10
v0.1.9
v0.1.8
v0.1.7
v0.1.6
v0.1.5
v0.1.4
v0.1.3
v0.1.2
v0.1.1
v0.1.0
v0.0.21
v0.0.20
v0.0.19
v0.0.18
v0.0.17
v0.0.16
v0.0.15
v0.0.14
v0.0.13
v0.0.12
v0.0.11
v0.0.10
v0.0.9
v0.0.8
v0.0.7
v0.0.6
v0.0.5
v0.0.4
v0.0.3
v0.0.2
v0.0.1
Labels
Clear labels
amd
api
app
bug
build
cli
cloud
compatibility
context-length
create
docker
documentation
embeddings
feature request
feedback wanted
good first issue
gpt-oss
gpu
harmony
help wanted
image
install
intel
js
launch
linux
macos
memory
mlx
model
needs more info
networking
nvidia
ollama.com
performance
pull-request
python
question
registry
rendering
thinking
tools
top
vulkan
windows
wsl
Mirrored from GitHub Pull Request
Milestone
No items
No Milestone
Projects
Clear projects
No project
No Assignees
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: github-starred/ollama#47502
Reference in New Issue
Block a user
Blocking a user prevents them from interacting with repositories, such as opening or commenting on pull requests or issues. Learn more about blocking a user.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @Pugio on GitHub (Dec 29, 2023).
Original GitHub issue: https://github.com/ollama/ollama/issues/1736
For every model I've downloaded, the speed saturates my bandwidth (~13MB/sec) until it hits 98/99%. Then the download slows to a few tens of KB/s and takes hour(s) to finish.
I've tried multiple models and this behavior happens each time. Happy to debug, but I'm not sure what to try.
I'm in Australia, in case that matters.
@Sully233 commented on GitHub (Dec 29, 2023):
I'm also experiencing this exact issue.
The workaround I've found is to stop the download when it slows down and then restart it - this resumes the download at full speed.
@pdevine commented on GitHub (Jan 2, 2024):
@Pugio can you run
ollama --version? Also, can you take a look at the logs? I'm curious if there is something like:@Pugio commented on GitHub (Jan 2, 2024):
Version
0.1.17.Tinyllama pulled fine, so tried it on Mistral and got the slowdown. I'm 3.8/4.1 GB and the download speed went from 13MB/s (saturating my connection) to a consistent 600-700KB/s.
The only log entries for this
pullare:EDIT: PatchingInitiative's trick of cancelling and restarting the
pullhas worked for me fairly well in the past few days, though for some models I needed to do that a couple of times.@Pugio commented on GitHub (Jan 3, 2024):
Additional logs after download finished:
@raidmax21 commented on GitHub (Jan 6, 2024):
Same issue here.
It seems to happen on all the models I download.
It starts out very quick using almost all of my 1gbit connection, but once it passes about 98% it drops right off to below 1mbps.
On a side note I'm using this in a docker on unraid 6.12.4.
ollama version is 0.1.18root@43f90abcec0d:/# ollama run llama2:13b-text pulling manifest pulling a42778cb0676... 99% ▕█████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████████ ▏ 7.3 GB/7.4 GB 42 KB/s 18m17s@FustarReal commented on GitHub (Jan 9, 2024):
I tried to download a model, it went to 48% with 26 mb/s then it slowed to 1 mb/s then I restarted my wifi and it went back to 20% only downloading at 4mb/s, tried to reset my wifi and stop the download as someone suggested and it didnt work, it still downloads only at 4mb/s
@pdevine commented on GitHub (Jan 9, 2024):
cc @mxyng
@mxyng commented on GitHub (Jan 11, 2024):
Some context:
Ollama downloads large files in parts with multiple concurrent workers. This maximizes transfer speed allow users to get their files faster.
The problem seems to be certain parts stall completely and zero data is received from the backend. The connection itself is still healthy so it doesn't trigger a retry. When this happens, the part stands out as other parts finish making it very visible in the last few percentages.
The linked PR aims to mitigate this by detecting these stalls and preemptively reset the connection before it can become a visible problem. The real solution is to address the storage backend so it never gets into this stalled state.
Aside: it seems the EOFs are at least correlated with this stalling behaviour. While the EOFs themselves will not cause any problems (the request will retry and continue where it left off), it's an indicator something's not quite right with the storage backend.
@ewebgh33 commented on GitHub (Jan 31, 2024):
**I have this same issue today and it's very VERY frustrating.
Running latest - just did an update.
I've pulled two 70b models and now a 34b model, and each time it rips until it gets to around 96%
Then it takes an hour to do the last few %.**
As seen in the screenshot it also sits there for 10mins or so (2min at the time I took the screenshot) when it's at 100% already. If it's at 100% why does it keep downloading for 10mins?
Why is this??
I wish we could download models directly into a folder, and not have to wait for Ollama to pull them itself.
Aside -
On that note, I also wish this because I have a bunch of models already, but it seems Ollama wants them in it's own special hashed (no file names) format, instead of just downloading a model from TheBloke directly to the folder, like can be done with textgen-webui (oobabooga).
So if you want to use a model with both apps or share models (heaven forbid, because we're all testing apps each day), we end up with each app wanting to download it's own version of the model and the terabytes just vanish from drives.
@pdevine commented on GitHub (Jan 31, 2024):
@EmmaWebGH sorry you're running into this. It's as @mxyng mentioned though; one of the streams of data is hung, but it's reporting healthy. You should be able to hit
Ctrl + cand then re-pull the image again and it should resume where it left off. Can you double check that you're running0.1.22?@ewebgh33 commented on GitHub (Jan 31, 2024):
@pdevine OK I'll try that next time.
I was afraid it wouldn't resume so yesterday I didn't try that solution! :)
@raidmax21 commented on GitHub (Feb 1, 2024):
There should be a fix in the next release ~ https://github.com/ollama/ollama/pull/1916
@marklysze commented on GitHub (Feb 3, 2024):
I just updated to 0.1.23 and tried to pull a model and it started off at full speed but near the end it slowed down to a crawl (as it had previously). So for me the fix didn't work (I'm in Australia).
Ctrl+C and running the command again finished the download at good speeds.
@raidmax21 commented on GitHub (Feb 3, 2024):
Gave the new version a test and sadly, 0.1.23 did not solve this issue.
@cyphbt commented on GitHub (Feb 27, 2024):
Stop the downloading process, and retry that will continue with high speed network.
@rhuanbarros commented on GitHub (Mar 11, 2024):
I'm using Ollama version 0.1.28 and encountering the same issue. I tried to download the model using the REST API, but the download stalled. So I tried again, but this time I connected to the terminal and attempted to download. This time it worked fine. So perhaps the problem lies solely with using the REST API
@ZaneA commented on GitHub (Mar 12, 2024):
I am getting this with 0.1.28, and I am using the terminal and not the REST API (yet) :)
Hitting ctrl-c and running again helps it along as suggested.
@Kuiriel commented on GitHub (Mar 12, 2024):
Same issue with the latest Ollama as well. Also Ctrl-C and running it again helped. Thank you.
@Qualzz commented on GitHub (Mar 14, 2024):
Confirming the same issue with high 8gbps speed. Downloading mixtral model in a few seconds at around 800mb/s then at 98/99% is goes down to around 100Kb/s and takes forever. Ctrl C and running again works fine.
@marschr commented on GitHub (Mar 25, 2024):
0.1.29 here and still getting this issue..
Ctrl+Cand resuming the download seems to work, but something is flaky..@frosty1492 commented on GitHub (Mar 29, 2024):
I'm having this same issue. I have a slow link, only 10 mbs but I have plenty of time and the connection is extremely stable. Old DSL server line. The download stops after it downloads about 38 MB. By using the terminal commands ctrl C and then restarting the with the command; ollama pull llama2:70b I can get about 80 mb before it craps out. I have concluded that this is an issue with either Ollama or the module server. It seems to be a handshake error. Yes, I am aware that this is a larger file but my system is very capable of managing it. I just live in an area where we do not have great internet service.
@chuklee commented on GitHub (Apr 2, 2024):
I have the same issue, i have this log on my docker:
ollama | time=2024-04-02T14:58:04.942Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 34 stalled; retrying"ollama | time=2024-04-02T14:59:07.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 11 stalled; retrying"
ollama | time=2024-04-02T14:59:07.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 29 stalled; retrying"
ollama | time=2024-04-02T14:59:08.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 3 stalled; retrying"
ollama | time=2024-04-02T14:59:08.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 25 stalled; retrying"
ollama | time=2024-04-02T14:59:08.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 37 stalled; retrying"
ollama | time=2024-04-02T14:59:08.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 20 stalled; retrying"
ollama | time=2024-04-02T14:59:08.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 14 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 40 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 41 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 31 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 33 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 35 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 18 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 39 stalled; retrying"
ollama | time=2024-04-02T14:59:09.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 38 stalled; retrying"
ollama | time=2024-04-02T14:59:09.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 6 stalled; retrying"
ollama | time=2024-04-02T14:59:09.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 5 stalled; retrying"
ollama | time=2024-04-02T14:59:09.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 4 stalled; retrying"
ollama | time=2024-04-02T14:59:09.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 16 stalled; retrying"
ollama | time=2024-04-02T14:59:09.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 12 stalled; retrying"
ollama | time=2024-04-02T14:59:10.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 10 stalled; retrying"
ollama | time=2024-04-02T14:59:49.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 39 stalled; retrying"
ollama | time=2024-04-02T14:59:49.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 40 stalled; retrying"
ollama | time=2024-04-02T14:59:50.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 10 stalled; retrying"
ollama | time=2024-04-02T14:59:50.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 16 stalled; retrying"
ollama | time=2024-04-02T14:59:51.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 28 stalled; retrying"
ollama | time=2024-04-02T14:59:51.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 9 stalled; retrying"
ollama | time=2024-04-02T14:59:51.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 7 stalled; retrying"
ollama | time=2024-04-02T14:59:51.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 25 stalled; retrying"
ollama | time=2024-04-02T14:59:51.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 41 stalled; retrying"
ollama | time=2024-04-02T15:00:05.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 17 stalled; retrying"
ollama | time=2024-04-02T15:00:05.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 38 stalled; retrying"
ollama | time=2024-04-02T15:00:07.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 23 stalled; retrying"
ollama | time=2024-04-02T15:00:08.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 13 stalled; retrying"
ollama | time=2024-04-02T15:00:08.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 8 stalled; retrying"
ollama | time=2024-04-02T15:00:08.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 36 stalled; retrying"
ollama | time=2024-04-02T15:00:08.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 24 stalled; retrying"
ollama | time=2024-04-02T15:00:09.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 32 stalled; retrying"
ollama | time=2024-04-02T15:00:09.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 6 stalled; retrying"
ollama | time=2024-04-02T15:00:10.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 12 stalled; retrying"
ollama | time=2024-04-02T15:00:10.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 5 stalled; retrying"
ollama | time=2024-04-02T15:00:14.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 21 stalled; retrying"
ollama | time=2024-04-02T15:00:37.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 30 stalled; retrying"
ollama | time=2024-04-02T15:00:38.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 5 stalled; retrying"
ollama | time=2024-04-02T15:00:52.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 18 stalled; retrying"
ollama | time=2024-04-02T15:00:54.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 32 stalled; retrying"
ollama | time=2024-04-02T15:00:57.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 20 stalled; retrying"
ollama | time=2024-04-02T15:00:57.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 27 stalled; retrying"
ollama | time=2024-04-02T15:00:58.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 4 stalled; retrying"
ollama | time=2024-04-02T15:01:02.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 25 stalled; retrying"
ollama | time=2024-04-02T15:01:04.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 31 stalled; retrying"
ollama | time=2024-04-02T15:01:06.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 0 stalled; retrying"
ollama | time=2024-04-02T15:01:14.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 6 stalled; retrying"
ollama | time=2024-04-02T15:01:16.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 39 stalled; retrying"
ollama | time=2024-04-02T15:01:17.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 22 stalled; retrying"
ollama | time=2024-04-02T15:01:17.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 19 stalled; retrying"
ollama | time=2024-04-02T15:01:22.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 37 stalled; retrying"
ollama | time=2024-04-02T15:01:23.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 36 stalled; retrying"
ollama | time=2024-04-02T15:01:24.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 33 stalled; retrying"
ollama | time=2024-04-02T15:01:24.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 14 stalled; retrying"
ollama | time=2024-04-02T15:01:24.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 34 stalled; retrying"
ollama | time=2024-04-02T15:01:24.947Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 40 stalled; retrying"
ollama | time=2024-04-02T15:01:25.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 29 stalled; retrying"
ollama | time=2024-04-02T15:01:25.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 25 stalled; retrying"
ollama | time=2024-04-02T15:01:25.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 35 stalled; retrying"
ollama | time=2024-04-02T15:01:25.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 17 stalled; retrying"
ollama | time=2024-04-02T15:01:28.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 27 stalled; retrying"
ollama | time=2024-04-02T15:01:34.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 3 stalled; retrying"
ollama | time=2024-04-02T15:01:34.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 0 stalled; retrying"
ollama | time=2024-04-02T15:01:34.946Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 18 stalled; retrying"
ollama | time=2024-04-02T15:01:35.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 13 stalled; retrying"
ollama | time=2024-04-02T15:04:36.697Z level=INFO source=download.go:178 msg="e8a35b5937a5 part 11 attempt 0 failed: unexpected EOF, retrying in 1s"
ollama | time=2024-04-02T15:04:37.802Z level=INFO source=download.go:178 msg="e8a35b5937a5 part 24 attempt 0 failed: unexpected EOF, retrying in 1s"
ollama | time=2024-04-02T15:04:41.698Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 11 stalled; retrying"
ollama | time=2024-04-02T15:04:52.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 36 stalled; retrying"
ollama | time=2024-04-02T15:04:52.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 34 stalled; retrying"
ollama | time=2024-04-02T15:04:52.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 37 stalled; retrying"
ollama | time=2024-04-02T15:04:53.942Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 28 stalled; retrying"
ollama | time=2024-04-02T15:04:53.942Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 23 stalled; retrying"
ollama | time=2024-04-02T15:04:53.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 17 stalled; retrying"
ollama | time=2024-04-02T15:04:53.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 31 stalled; retrying"
ollama | time=2024-04-02T15:04:53.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 38 stalled; retrying"
ollama | time=2024-04-02T15:04:53.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 29 stalled; retrying"
ollama | time=2024-04-02T15:04:53.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 35 stalled; retrying"
ollama | time=2024-04-02T15:04:53.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 6 stalled; retrying"
ollama | time=2024-04-02T15:04:53.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 39 stalled; retrying"
ollama | time=2024-04-02T15:05:16.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 29 stalled; retrying"
ollama | time=2024-04-02T15:05:25.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 7 stalled; retrying"
ollama | time=2024-04-02T15:05:47.942Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 9 stalled; retrying"
ollama | time=2024-04-02T15:05:47.942Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 30 stalled; retrying"
ollama | time=2024-04-02T15:05:59.854Z level=INFO source=download.go:178 msg="e8a35b5937a5 part 36 attempt 0 failed: unexpected EOF, retrying in 1s"
ollama | time=2024-04-02T15:07:06.904Z level=INFO source=download.go:178 msg="e8a35b5937a5 part 6 attempt 0 failed: unexpected EOF, retrying in 1s"
ollama | time=2024-04-02T15:07:38.614Z level=INFO source=download.go:178 msg="e8a35b5937a5 part 21 attempt 0 failed: unexpected EOF, retrying in 1s"
ollama | time=2024-04-02T15:08:52.542Z level=INFO source=download.go:178 msg="e8a35b5937a5 part 15 attempt 0 failed: unexpected EOF, retrying in 1s"
ollama | time=2024-04-02T15:09:14.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 8 stalled; retrying"
ollama | time=2024-04-02T15:09:16.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 17 stalled; retrying"
ollama | time=2024-04-02T15:09:21.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 23 stalled; retrying"
ollama | time=2024-04-02T15:09:22.699Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 11 stalled; retrying"
ollama | time=2024-04-02T15:09:26.944Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 35 stalled; retrying"
ollama | time=2024-04-02T15:09:48.942Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 22 stalled; retrying"
ollama | time=2024-04-02T15:09:48.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 20 stalled; retrying"
ollama | time=2024-04-02T15:09:48.945Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 35 stalled; retrying"
ollama | time=2024-04-02T15:09:49.615Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 21 stalled; retrying"
ollama | time=2024-04-02T15:10:05.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 30 stalled; retrying"
ollama | time=2024-04-02T15:10:05.943Z level=INFO source=download.go:250 msg="e8a35b5937a5 part 13 stalled; retrying"
@frosty1492 commented on GitHub (Apr 2, 2024):
I don't know why we can't just download a file from the server using
regular download protocols. It seems that for some reason Ollama requires
the Ollama app to pull the files. Doesn't make sense. I download work
press and Mediawiki files then install the add ons etc.
On Tue, Apr 2, 2024 at 10:10 AM chuklee @.***> wrote:
--
John Foster
@dinakar17 commented on GitHub (Apr 3, 2024):
The simple solution is to run your command prompt or Powershell as administrator. It will work like a charm.
@frosty1492 commented on GitHub (Apr 3, 2024):
I use Linux and have tried as root. No success.
On Wed, Apr 3, 2024 at 2:20 AM Dinakar Chennupati @.***>
wrote:
--
John Foster
@wlhee commented on GitHub (Apr 9, 2024):
+1, it's pretty confusing about why Ollama pull slows down at 99%.
@pdevine commented on GitHub (Apr 10, 2024):
@wlhee see my comment above. I think it's a problem we're having w/ Cloudflare. The only way to really get around it is to hit Ctrl+C and retry; you won't lose any progress.
@wlhee commented on GitHub (Apr 10, 2024):
Thanks for the tip. ctrl+c and retry does resume the progress, but I still see the extreme slowness issue.
More importantly, I'd like to have the root cause to this problem as it would affect using Ollama to serve inference workloads in prod.
@sammcj commented on GitHub (Apr 10, 2024):
Seems to be a constant issue for me as well, also in Australia
@frosty1492 commented on GitHub (Apr 11, 2024):
So I tried several times to download any of the mods. No success. I am
fully aware that I have a slow internet link, but it is all I have and
works fine for other stuff. The OP was, why can't we download the mods as a
regular file from Git Hub, etc. then have Ollama pull them from our drives
and install the mods. I already do this on many hubs and sites for many
applications.
On Tue, Apr 9, 2024 at 10:54 PM Sam @.***> wrote:
--
John Foster
@asitjain commented on GitHub (Apr 12, 2024):
On Windows 11 here stuck at 99%, I stopped the transfer using CTRL+C, exited the power shell (x86) and then started Power Shell (No x86 so that must be 64 bit) as administrator. Restarted again and it worked like a charm and finished in 5 seconds.
@LagSlug commented on GitHub (Apr 13, 2024):
Been having the same problem, but I think the problem is with the models/blobs not updating unless either:
So the status objects aren't being sent until the very end, and it makes it appear that the system has halted. If you look over at the stats window in docker you can see, however, that the download speed remains constant.
@matheusnienow commented on GitHub (Apr 20, 2024):
Same issue. Restarting the download fixed it.
@nielpattin commented on GitHub (Apr 20, 2024):
same issue. Version 0.1.32 when trying to
ollama run llama3for example. It just randomly slowed down at 88% ICtrl+Cquit it and re-run again it started at full speed and then the same thing happened after at 96% it slowed down again. OMG@gauravagnihotristla commented on GitHub (Apr 24, 2024):
Having same issue
ollama version is 0.1.32re-running the command did not fix the issue
@nielpattin commented on GitHub (Apr 24, 2024):
sOmE TiMe tHeY did, you have to keep trying... (Your luck is 50/50)
@Gurpreethgnis commented on GitHub (Apr 25, 2024):
This not fixed. I have a 1 GB Fibre network.
Apple M1 Pro
ollama -vollama version is 0.1.32First stop
ollama run llama2pulling manifestpulling 8934d96d3f08... 12% ▕█ ▏ 454 MB/3.8 GB 3.3 MB/s 16m58sI get this error
Error: max retries exceeded: Get "https:// .... : net/http: TLS handshake timeoutSecond Stop
ollama run llama2pulling manifestpulling 8934d96d3f08... 94% ▕███████████████ ▏ 3.6 GB/3.8 GB 3.6 MB/s 1m1sError: max retries exceeded: 400: <?xml version="1.0" encoding="UTF-8"?><Error><Code>InvalidArgument</Code><Message>Invalid Argument: range must be positive.</Message></Error>Third Stop
ollama run llama2pulling manifestpulling 8934d96d3f08... 100% ▕████████████████▏ 3.8 GBError: max retries exceeded: 400: <?xml version="1.0" encoding="UTF-8"?><Error><Code>InvalidArgument</Code><Message>Invalid Argument: range must be positive.</Message></Error>Fourth time
ollama run llama2pulling manifestpulling 8934d96d3f08... 100% ▕███████████████████████████████████████████████████████▏ 3.8 GBIt has been stuck for the past half an hour and nothing is happening!
@Gurpreethgnis commented on GitHub (Apr 25, 2024):
** Update**
pulling 8934d96d3f08... 100% ▕████████████▏ 3.8 GBError: max retries exceeded: unexpected EOFSo, not working!
@Gurpreethgnis commented on GitHub (Apr 25, 2024):
Tried with another model. All models when they reach 100% start giving this error infinitely
Error: max retries exceeded: Get ... : net/http: TLS handshake timeout@gauravagnihotristla commented on GitHub (Apr 25, 2024):
Are you on a VPN?
@Gurpreethgnis commented on GitHub (Apr 25, 2024):
no
@taozhiyuai commented on GitHub (May 7, 2024):
just ctrl+c,and restart pulling
@agomezh commented on GitHub (May 8, 2024):
I can also say, that disconnecting & reconnecting to the wifi network did the trick.
Thanks!
@sychedelix commented on GitHub (May 8, 2024):
+1 and stopping the command and re-running it works.
@bmizerany commented on GitHub (May 9, 2024):
Hello, Everyone!
At Ollama we're working on a solution to this issue, and have been seeing some positive results!
Now we need your help testing in your enviroments as well!
How to help:
Run a test pull through our staging server
From the list below, pick one (or many) of the models that you have not pulled already, and perform a pull.
Remove and retry 2 or 3 more times
Report back!
Please respond here answering these questions to the best of your ability:
ollama pullcommand you ran including model?30-50 MB/s)ollama pull <model>for the same model(s)?Thank you all so much in advance. We look forward to hearing back from you.
@mchiang0610 commented on GitHub (May 9, 2024):
Thank you @bmizerany!
ollama pull issue1736.ollama.dev/library/llama3:8bMacOS Sonoma 14.4.1
Speed starts at around 58MB/s; after 1% it reduces to 25-30MB/s; after about 10% the speed increases to 58MB/s and fluctuates between 45-60MB/s. Same results all 3 times
ollama v0.1.34
Canada, Toronto (East)
Top speed is around 600Mbps (75MB/s) down on my wifi
Slightly faster than normal pull
@adriens commented on GitHub (May 9, 2024):
ℹ️ Settings
ollama pull issue1736.ollama.dev/library/gemma:2bLinux Mint 21.3 Virginia0.1.34Nouméa, New-Caledonia9.7 MB/sdown on my wifi`3.7 MB/s9.4 MB/s9.4 MB/s📷 Details
1️⃣ Try
2️⃣ Try
3️⃣ Try
@MaxJa4 commented on GitHub (May 9, 2024):
Command:
ollama pull issue1736.ollama.dev/library/dolphin-phi(1.6 GB)OS:
Win 11 ProSpeed range:
28-30 MB/sVersion:
v0.1.34Region:
Europe, Germany / domestic internetMax Line Speed:
~30-31 MB/sSpeed compared to normal pull:
SameCommand:
ollama pull issue1736.ollama.dev/library/dolphin-phi(1.6 GB)OS:
Ubuntu 22 LTS ServerSpeed range:
57-77 MB/sVersion:
v0.1.34Region:
Europe, Germany / cloud server infrastructureMax Line Speed:
~1000 MB/s at time of testing (varies, usually 600-1200)Speed compared to normal pull:
Slower, normal pull had 100-250 MB/sCommand:
ollama pull issue1736.ollama.dev/library/llama3:8b(4.7 GB)OS:
Ubuntu 22 LTS ServerSpeed range:
68-80 MB/sVersion:
v0.1.34Region:
Europe, Germany / cloud server infrastructureMax Line Speed:
~1000 MB/s at time of testing (varies, usually 600-1200)Speed compared to normal pull:
Slower, normal pull had 200-420 MB/sIt used to be super slow after ~95%, but since a few days or so it got a lot better!
Hope this helps. Thanks for working on Ollama!
@bmizerany commented on GitHub (May 9, 2024):
@MaxJa4 Do you mind trying again? It's likely the cache wasn't warm for your region when you tried it. Hopefully it's faster now?
@MaxJa4 commented on GitHub (May 9, 2024):
Sure, tried it again three times for the llama3 model (staging). This time it was 320-340 MB/s and pretty stable throughout the whole download, including the very last percentages.
@bmizerany commented on GitHub (May 9, 2024):
@MaxJa4 thank you so much for the extra help!
@alwqx commented on GitHub (May 10, 2024):
start with 150KB/s -> 1MB/s -> 2MB/s -> 4.6MB/s in 3 second and remain this speed until pull finished.
@taozhiyuai commented on GitHub (May 10, 2024):
taozhiyu@603e5f4a42f1 ~ % ollama pull issue1736.ollama.dev/library/llama3:8b
pulling manifest
Error: pull model manifest: Get "https://issue1736.ollama.dev/v2/library/llama3/manifests/8b": net/http: TLS handshake timeout
ollama pull issue1736.ollama.dev/library/llama3:8b ; FAIL,net/http: TLS handshake timeout
ollama pull issue1736.ollama.dev/library/gemma:2b ;CAN DOWNLOAD AT 25MB/S MAX
ollama pull issue1736.ollama.dev/library/mistral ; CAN DOWNLOAD AT 25MB/S MAX, SOMETIMES DROP TO 0MB/S, MUST CTRL+C AND RESTART WHEN 98%
ollama pull issue1736.ollama.dev/library/dolphin-mistral ; CAN DOWNLOAD AT 31MB/S MAX;SOMETIMES DROP TO 0MB/S,
ollama pull issue1736.ollama.dev/library/wizardlm2
ollama pull issue1736.ollama.dev/library/llava-phi3 ; CAN DOWNLOAD AT 29MB/S MAX
ollama pull issue1736.ollama.dev/library/llava-llama3
ollama pull issue1736.ollama.dev/library/dolphin-phi
ollama pull issue1736.ollama.dev/library/nomic-embed-text ; CAN DOWNLOAD AT 19MB/S MAX
ollama pull issue1736.ollama.dev/library/phi3
ollama pull issue1736.ollama.dev/library/orca-mini ; CAN DOWNLOAD AT 29MB/S MAX
taozhiyu@603e5f4a42f1 ~ % ping issue1736.ollama.dev
PING issue1736.ollama.dev (66.241.124.100): 56 data bytes
Request timeout for icmp_seq 0
64 bytes from 66.241.124.100: icmp_seq=1 ttl=50 time=180.223 ms
64 bytes from 66.241.124.100: icmp_seq=2 ttl=50 time=185.270 ms
64 bytes from 66.241.124.100: icmp_seq=3 ttl=50 time=191.586 ms
64 bytes from 66.241.124.100: icmp_seq=4 ttl=50 time=193.099 ms
64 bytes from 66.241.124.100: icmp_seq=5 ttl=50 time=169.806 ms
Request timeout for icmp_seq 6
64 bytes from 66.241.124.100: icmp_seq=7 ttl=50 time=136.826 ms
64 bytes from 66.241.124.100: icmp_seq=8 ttl=50 time=167.838 ms
64 bytes from 66.241.124.100: icmp_seq=9 ttl=50 time=171.574 ms
Request timeout for icmp_seq 10
64 bytes from 66.241.124.100: icmp_seq=11 ttl=50 time=136.509 ms
64 bytes from 66.241.124.100: icmp_seq=12 ttl=50 time=165.411 ms
Request timeout for icmp_seq 13
64 bytes from 66.241.124.100: icmp_seq=14 ttl=50 time=166.876 ms
64 bytes from 66.241.124.100: icmp_seq=15 ttl=50 time=190.036 ms
Request timeout for icmp_seq 16
64 bytes from 66.241.124.100: icmp_seq=17 ttl=50 time=166.609 ms
Request timeout for icmp_seq 18
Request timeout for icmp_seq 19
64 bytes from 66.241.124.100: icmp_seq=20 ttl=50 time=175.115 ms
64 bytes from 66.241.124.100: icmp_seq=21 ttl=50 time=178.726 ms
64 bytes from 66.241.124.100: icmp_seq=22 ttl=50 time=175.443 ms
^C
--- issue1736.ollama.dev ping statistics ---
23 packets transmitted, 16 packets received, 30.4% packet loss
round-trip min/avg/max/stddev = 136.509/171.934/193.099/15.973 ms
taozhiyu@603e5f4a42f1 ~ % networkQuality
==== SUMMARY ====
Uplink capacity: 432.862 Kbps
Downlink capacity: 795.363 Mbps
Responsiveness: Low (714.286 milliseconds | 84 RPM)
Idle Latency: 20.667 milliseconds | 3000 RPM
taozhiyu@603e5f4a42f1 ~ % traceroute issue1736.ollama.dev
traceroute to issue1736.ollama.dev (66.241.124.100), 64 hops max, 40 byte packets
1 xiaoqiang (192.168.31.1) 3.657 ms 2.969 ms 2.667 ms
2 192.168.1.1 (192.168.1.1) 4.323 ms 5.977 ms 4.195 ms
3 100.64.0.1 (100.64.0.1) 8.548 ms 7.425 ms 6.664 ms
4 58.50.123.141 (58.50.123.141) 7.843 ms 6.885 ms 6.698 ms
5 * * 58.50.122.1 (58.50.122.1) 10.428 ms
6 202.97.68.101 (202.97.68.101) 27.711 ms
202.97.19.61 (202.97.19.61) 23.056 ms
202.97.29.153 (202.97.29.153) 24.625 ms
7 * * 202.97.71.34 (202.97.71.34) 23.779 ms
8 * 202.97.83.22 (202.97.83.22) 25.817 ms
202.97.39.201 (202.97.39.201) 30.011 ms
9 203.86.97.6 (203.86.97.6) 154.323 ms
xe-1-5-2-3.a03.tokyjp05.jp.bb.gin.ntt.net (129.250.9.125) 139.994 ms
203.86.97.6 (203.86.97.6) 187.262 ms
10 ae-3.r33.tokyjp05.jp.bb.gin.ntt.net (129.250.5.95) 123.159 ms 119.312 ms
ae-5.r32.tokyjp05.jp.bb.gin.ntt.net (129.250.5.93) 145.196 ms
11 * ae-0.a00.tokyjp09.jp.bb.gin.ntt.net (129.250.7.14) 128.363 ms 156.647 ms
12 * * *
13 * * *
14 * 103.84.154.10 (103.84.154.10) 134.021 ms *
15 * * *
16 * * *
17 * * *
18 * * *
19 * * *
20 * * *
21 * * *
22 * * *
23 * * *
24 * * *
25 * * *
26 * * *
27 * * *
28 * * *
29 * * *
30 * * *
31 * * *
32 * * *
33 * * *
34 * * *
35 * * *
36 * * *
37 * * *
38 * * *
39 * * *
40 * * *
41 * * *
42 * * *
43 * * *
44 * * *
45 * * *
46 * *^C
@bmizerany
@tao12345666333 commented on GitHub (May 10, 2024):
@bmizerany commented on GitHub (May 10, 2024):
@tao12345666333 By any chance did you try pulling one of those models through the staging server after waiting maybe 10 minutes or more? You may have not hit the cache on the first pull, which we have been seeing is slower than pulling from ollama.com instead, but the pulls through the staging server after the first are much faster than the first.
@tao12345666333 commented on GitHub (May 11, 2024):
@bmizerany After waiting for a while, I tried again and it was indeed much faster than the first time.
However, based on my current test results, they are consistently slower than the existing repo (ollama.com).
@sammcj commented on GitHub (May 19, 2024):
I've been hitting this problem for a few months now, it used to be I could just cmd+c and resume the download to fix it - but something interesting changed over the last week.
Now instead of stalling at 95-99%, now pulling models runs until about 95%~ and then the progress starts jumping around, the speed bounces around and the percentage complete even goes backwards at times!
Then after a minute or so it crashes with the error:
Error: max retries exceeded: EOFIt looks like the Ollama is pulling models from
2606:4700::6812:85aor for IPv4172.67.182.229which appears to be a Cloudflare IP and I'm not seeing any packet loss or high latency with mtr/ping.I've tested on:
Here's a screen recording of the progress bar at the end of a pull: https://github.com/ollama/ollama/assets/862951/fdf830d0-c122-444f-95c4-300c5058193c
Edit: I just saw mention of issue1736.ollama.dev/library/, and tried it - it appeared to work, then got to 99% and slowed to a crawl, cmd+c and started it again and it did actually finish downloading the image - but the SHA is corrupt:
See also
*Edit 2024-05-25: This PR seems to fix the broken pulls for me: https://github.com/ollama/ollama/pull/4619
@hangxingliu commented on GitHub (May 25, 2024):
I believe that this issue has been fixed by the patch https://github.com/ollama/ollama/issues/4520, which was merged just now.
I had the same issue (The download progress is always stuck at the last 1 gb) today. Fortunately, @noxer just fixed this issue, the download process became available when I used ollama compiled from the latest code.
@davidtranjs commented on GitHub (May 30, 2024):
This issue still happens after I updated to latest version. Start fast untill the last G, I have to cancel and resume it about 10 times to finish the download
@metamec commented on GitHub (May 30, 2024):
Same problem with the latest version. In addition, by downloading so many chunks simultaneously, it's overwhelming my connection. Everything else grinds to a halt. I had to ctrl+c just to load this page. (And again just to submit this comment. 😅)
@MaxJa4 commented on GitHub (May 30, 2024):
Same here, it also consumes a rather high amount of CPU for "just" a download.
Tried it on a GPU cloud server... it reports 20 Gbit/s uplink, but is throttled to ~1 GBit/s due to 100% CPU usage just from model downloading.
@dhaneshdutta commented on GitHub (Jun 22, 2024):
same problem with verision 0.1.45, download speed slows down when approaching the end of downloading a model
@IAkumaI commented on GitHub (Jun 27, 2024):
Same problem on 0.1.47. Retry it does not change anything.
Also there are logs:
@frosty1492 commented on GitHub (Jun 27, 2024):
Perhaps the downloads need to be on a different server with a proper
connection.
On Thu, Jun 27, 2024 at 7:11 AM Valery Ozarnichuk @.***>
wrote:
--
John Foster
@ahaggard2013 commented on GitHub (Jul 7, 2024):
I was only experiencing this on windows. worked on linux/mac fine. I bypassed this for windows by running ollama in docker and mounting the dockers model directory to the windows model directory for downloads. Worked as a temporary work around for anyone seeing this.
edit: this worked once. seemed to be able to download via docker, but had the same issue after one successful model download.
@metamec commented on GitHub (Jul 11, 2024):
It's not hard to understand why it's saturating home networks when according to
netstat -aon, it's establishing 63 sockets to download a single file.@frosty1492 commented on GitHub (Jul 11, 2024):
I do not understand why anyone would set up a server for software downloads
and not have the DL files to be in a compressed format to be
de-compressed and installed after a single file download. Pardon me but
that causes concern about the efficiency of the software. On a second note,
the actual Ollama files were downloaded and installed in less than a minute
on my system. So is there another explanation? I'm just VERY curious....
On Thu, Jul 11, 2024 at 12:41 PM metamec @.***> wrote:
--
John Foster
@goodie1974 commented on GitHub (Jul 20, 2024):
it stinks! i'm on DSL and before I got to bed I start downloading. But it's constantly giving me this shit:
wsarecv: A connection attempt failed because the connected party did not properly respond after a period of time, or established connection failed because connected host has failed to respond.
@frosty1492 commented on GitHub (Jul 23, 2024):
So I'm on a very slow DSL hard line, but just because I have it on my
T-Mobil service. I decided to tether my system to my cell phone wifi. It
took 2 attempts but I got all of llama3 and llava installed along with
Docker and WebUI. Now experimenting but it all works. Got to say Llama3 is
impressive but VERY SLOW to respond. Llava is much faster but seems less
astute. Any tips ore comments?
On Sat, Jul 20, 2024 at 12:03 PM goodie1974 @.***>
wrote:
--
John Foster
@vamsi-panchada commented on GitHub (Jul 25, 2024):
I've facing the issue while pulling the model, the model download speed is suddenly dropping from 300 Mbps to 10 Mbps. when the model is at 40-50% downloaded, along with that it's gradually going down to 1 Mbps duing the last 10-5%, could someone help me in resolving this issue.
thank you
@vineetgupta0 commented on GitHub (Jul 25, 2024):
I was facing the same issue after 3 retries and turning off windows firewall it worked. I don't face any similar problem with my macbook m2 so can say network is not at the fault and my windows machine is powerful enough, we can see the speeds were in kb/s before and later it was higher, even after getting the high speed it suddenly drips back to kb's

@rampageservices commented on GitHub (Aug 2, 2024):
As mentioned in the quoted message above from earlier in this thread. The PR #4625 that is yet to be pulled resolves the remaining issue for some users.
@sammcj commented on GitHub (Sep 11, 2024):
Looks like the stalled at 90%~ bug is back!
@2102819372 commented on GitHub (Sep 20, 2024):
i don't know how to describe the problem, i follow the log recommendations retried so many times,but is useless


@sammcj commented on GitHub (Oct 29, 2024):
FYI this is still a an issue as of 2024-10-29, currently causing the following issues (other than generally being annoying 😅):
@FreshLucas-git commented on GitHub (Nov 8, 2024):
Same issue here. This issue still exists on v0.4.0.
@sammcj commented on GitHub (Nov 10, 2024):
This is impacting the new create model from Huggingface feature as well:
@charlieamer commented on GitHub (Dec 25, 2024):
Hello all. One of the 2 things solved this issue for me (probably not both though):
I am not sure which of the 2 things above helped me, but i don't have the issuse anymore
@rampageservices commented on GitHub (Dec 25, 2024):
That is helpful feedback Amer. It could mean that some sort of checking
needs to be done on the I/O side or check to ensure the download function
be done in a separate thread if not done already. (It's been quite a few
months since I've seen the codebase.)
On Wed, Dec 25, 2024, 2:48 PM Amer @.***> wrote:
@ngocdd commented on GitHub (Dec 29, 2024):
the same issue fro my side the speed too slow

@oyaro-tech commented on GitHub (Jan 7, 2025):
I had the same issue on MacOS only when using an external SSD drive to store models. After formatting it to APFS, everything works fine
@eduanmoldeep commented on GitHub (Jan 12, 2025):
I found a much simple solution that worked for me
from stalling 200 kbps, jumped back to 5Mbps
Hope it helps you, who had to scroll this deep here
@Nitish-bot commented on GitHub (Jan 20, 2025):
@eduanmoldeep it is the first comment
@harisiqbal12 commented on GitHub (Jan 25, 2025):
Still facing this issue, my mistral model was downloaded of 50% then my internet fluctuated and it's back to 10% i'm on latest version. 0.5.7
@eduanmoldeep commented on GitHub (Jan 25, 2025):
@Nitish-bot
might sound odd, but stopping download didn't help,
I had to close the whole terminal and in a fresh instance only was able to get the throttling solved.
Thus wrote a seperate comment
@krecco commented on GitHub (Jan 30, 2025):
Is there any other simple way to download a model? Downloading mixtral, downloaded 100+ GB with restarts (a lot is lost after restart) but still not finished....
@HildaJ commented on GitHub (Jan 30, 2025):
I closed the terminal, and entered
ollama pull [your model name]instead ofollama run [your model name]. It works for me.@pdevine commented on GitHub (Jan 30, 2025):
@HildaJ The mechanism for pulling is the same for both
ollama pullandollama run, so I think you just got lucky.@krecco commented on GitHub (Jan 30, 2025):
agreed, i always use pull and its not working ... normally i try to load instruct model, not the default one
@rampageservices commented on GitHub (Jan 31, 2025):
I am going to take a look at this over the next week and suggest a PR. This
issue continues to persist after such a long time. To ease initial use for
users who are new to the environment of which I am positive there will be
thousands more, this will continue to be a wall that may delay future AI
innovators from creating the next big thing.
On Thu, Jan 30, 2025, 11:18 PM BoRut @.***> wrote:
@flyfox666 commented on GitHub (Feb 1, 2025):
I've run into the same problem, every time I reopen and start it the speed is really fast, but after a while it slows down and the drop is visible to the naked eye.The download speed goes down in steps.
@rampageservices commented on GitHub (Feb 1, 2025):
The slow down may actually be a symptom of your ISP performing QoS on large
downloads to ensure bandwidth for other users on your network. This is
outside of Ollama's control and not a result of anything in particular
wrong with the software itself.
On Sat, Feb 1, 2025, 3:00 PM bbw1986 @.***> wrote:
@arunofhyd commented on GitHub (Feb 1, 2025):
This is the ultimate fix!
@krecco commented on GitHub (Feb 2, 2025):
I have tested it on ubuntu server, cli only and I havent had any problems since.
@imyhxy commented on GitHub (Feb 11, 2025):
It's frustrated that I have to stop and start the pull command over and over again once the download speed is lower than 1MB/s. 😢
@arunofhyd commented on GitHub (Feb 11, 2025):
You can just stop and restart, it works!
@imyhxy commented on GitHub (Feb 11, 2025):
@IngeniousThomas Yes, but it will slow down again very soon.
@shrijayan commented on GitHub (Feb 17, 2025):
Now this work around is also not working.
@frosty1492 commented on GitHub (Feb 18, 2025):
Can anyone explain why we can not simply download this as an installable
file? This process is not good at all.
On Mon, Feb 17, 2025 at 6:09 AM Shrijayan @.***> wrote:
--
John Foster
@dmilneconcordinfiniti commented on GitHub (Mar 6, 2025):
Ctrl + C then closing the terminal before reapplying "ollama pull [model]" worked for me.
I had also added the "OLLAMA_NOPRUNE=true" environment variable to enable (encourage?) resumable downloads.
I'm also downloading from Australia and I don't think it's QoS - as it happens consistently on the last gigabyte of an Ollama download regardless of the size of the model. 9GB model throttles to dial-up speed at 8GB downloaded; while a 19GB model throttles to dial-up speed at 18GB downloaded.
@seeadoog commented on GitHub (Mar 10, 2025):
just like playing monkey
@imyhxy commented on GitHub (Mar 11, 2025):
Try this:
Restart the download process every 600s until the download is finished.
@sandeep-tukaram commented on GitHub (Mar 15, 2025):
Don't understand why 15MB/s , order of 10MBps, limit on the download speed when I do ollama pull on a model.. Is it because of the registry host - Docker hub?
Tops this my ISP speed drops from about 100MBps to 15MBps (speedtest ookla) when downloading from Ollama. Fail to wrap my head around this crawling speed.
@sandeep-tukaram commented on GitHub (Mar 15, 2025):
Every ISP caping downloads at 15 MB/s and crawling at 99%. Sounds unrealistic.
@rampageservices commented on GitHub (Mar 15, 2025):
I have achieved 60MB/s on every recent model pull so... not sure what to
tell you. It was a suggestion that I was providing. No need to directly
attack me. If you have this issue, please look into it on your end as well.
On Sat, Mar 15, 2025, 6:41 PM Sandeep T @.***> wrote:
@pdevine commented on GitHub (Mar 16, 2025):
Hey guys, using ollama 0.6.1 can you set the
OLLAMA_EXPERIMENTenv toclient2and try pulling? Instructions for setting the environment variables are here.You'll get something that looks like this:

When it's finished you won't see the
verifying sha256 digestmessage as the checksums are validated as the data is streamed.If you want to set it back to the default client you can set
OLLAMA_EXPERIMENTtoclient1or remove the environment variable.@ViZAlice commented on GitHub (Jun 6, 2025):
Yeah, it works! And when I repeat the last pull command, it will keep processing.
@SPOOKEXE commented on GitHub (Jun 11, 2025):
setting the OLLAMA_EXPERIMENT to client2 fixed my issue it seems!
@curlydingo commented on GitHub (Jul 19, 2025):
client2 has fixed the issue for me too, thanks @pdevine !
@ferrouswheel commented on GitHub (Aug 3, 2025):
client2fixed mine too. Downloaded most of qwen3-coder, but it would get stuck at 80kb/s and 30 minutes remaining despite almost being complete. Restarting ollama and/or the download withollama pulldidn't change anything.But adding
-e OLLAMA_EXPERIMENT=client2to docker run makes the download complete instantly (suggesting almost all parts were already downloaded)This seems like a high prority bug since the ease of downloading models is the primary reason I use ollama.
@div002123 commented on GitHub (Aug 6, 2025):
not working...
0%-95% 30min
95%~ 30min~1hour
It turns out I was being throttled; switching networks solved the problem. Just ignore me.😂
@Rizmi commented on GitHub (Aug 28, 2025):
Oh yeah? U genuinely think tht isp gonna block extractly at 98% of a 5gb model? This issue old af and still not fixed, instead blame the isp 😆
@DaRacci commented on GitHub (Sep 10, 2025):
Still relevant, downloading any model seems to result in a crawl past ~96% unless its a super small model.
Its not an ISP issue, DNS issue or network related, this is something directly related to ollama.
Ollama version: 0.11.7
@ferrouswheel commented on GitHub (Sep 14, 2025):
After updating ollama in docker
OLLAMA_EXPERIMENT=client2results in a stream of "context cancelled" log errors and model downloading doesn't occur.model name: deepseek-r1:8b
ollama version: 0.11.10
Not sure if any of the tokens/signatures in the query string are non-public so have stripped them.
Removing the
OLLAMA_EXPERIMENTenvironment variable allows downloading to occur again but exhibits the old behaviour of slowing to a halt at ~95%. Here is a network graph of the hosts traffic to show how extreme the slow down is.(the first two traffic peaks are docker pulls, the last is the ollama pull stuck at 95%)
Edit: after 40 minutes it said:
time=2025-09-14T21:03:30.184Z level=INFO source=download.go:295 msg="e6a7edc1a4d7 part 12 attempt 0 failed: unexpected EOF, retrying in 1s"and appears to have now completed the download.I also tried pulling the smallest model I could quickly find and
embeddinggemma:300mdownloaded with stalling.@barbuz commented on GitHub (Sep 15, 2025):
Same here (version 0.11.10), download speed pulls the brakes hard after 90% on a 3GB model (gemma3).
OLLAMA_EXPERIMENT=client2as suggested above did not fix it, the only valid workaround seems to be stopping the download and restarting it.@adriens commented on GitHub (Sep 15, 2025):
Same here
@omalagouzNW commented on GitHub (Sep 17, 2025):
I tried all the methods above and was stuck at 95% at 155 KB/s downloading a 5GB model, the only thing that actually fixed the issue was switching from Wi-Fi to my phone's 5G personal hotspot, went back to at least 50 MB/s and finished downloading in 30 seconds.
@Tarun047 commented on GitHub (Oct 18, 2025):
Wow its full 2 years and this problem is not resolved yet!
So much for running models locally!
@FakeUser2121 commented on GitHub (Oct 19, 2025):
This problem is still not fixed and its been almost 2 years. Lm studio downloads 100s of Gb of models just fine. I dont know why isnt this fixed yet.
PS C:\Users\user1> ollama pull llama3.2-vision
pulling manifest
pulling 9999d473417a: 99% ▕█████████████████████████████████████████████████████████ ▏ 7.8 GB/7.8 GB
PS C:\Users\user1> ollama pull llama3.2-vision
pulling manifest
pulling 9999d473417a: 99% ▕█████████████████████████████████████████████████████████ ▏ 7.8 GB/7.8 GB 17 KB/s 57m49s
PS C:\Users\user1> ollama pull llama3.2-vision
pulling manifest
pulling 9999d473417a: 99% ▕█████████████████████████████████████████████████████████ ▏ 7.8 GB/7.8 GB
PS C:\Users\user1> ollama pull llama3.2-vision
pulling manifest
pulling 9999d473417a: 99% ▕█████████████████████████████████████████████████████████ ▏ 7.8 GB/7.8 GB
PS C:\Users\user1> ollama pull llama3.2-vision
pulling manifest
pulling 9999d473417a: 100% ▕█████████████████████████████████████████████████████████ ▏ 7.8 GB/7.8 GB 116 KB/s 2m17s
@rampageservices commented on GitHub (Oct 20, 2025):
I just want to point out that this is not the way to ask for help with Open Source software. You are not entitled to free fixes. To others reading this, don't be this person.
@mendi80 commented on GitHub (Oct 23, 2025):
working solution. after the slow down start, Ctrl+C, switch the network to your phone hotspot (or other network with different public ip address), continue the downloading (ollama pull mymodel) for some seconds, Ctrl+C again. back to you home router, continue downloading. now it will continue at max speed.
very strange bug. but i guess it is easy to fix.
@Phizicks commented on GitHub (Oct 24, 2025):
instead of going through cloudflare etc, could think the possibility in future to allow torrenting of the models.
@sudhir-asuracore commented on GitHub (Oct 26, 2025):
I just switch on my vpn, connect to a zone in USA and the download completes the last few percent at full bandwidth. Still not sure what the cause is, but switching to a vpn spot helps.
@AbhishekRH commented on GitHub (Oct 27, 2025):
This ACTUALLY WORKS!! 🔥
Thanks man!