[GH-ISSUE #6531] Prebuilt ollama-linux-amd64.tgz without cuda libs, please? #29872

Open
opened 2026-04-22 09:08:59 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @sevaseva on GitHub (Aug 27, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/6531

Originally assigned to: @dhiltgen on GitHub.

I occasionally update ollama on a linux box by downloading URLs like https://github.com/ollama/ollama/releases/download/v0.3.7-rc6/ollama-linux-amd64.tgz and extracting/overwriting files into a local directory (not into /usr as a root mind you, just into a local directory as a non-privileged user; that is how I prefer to use it).

I have necessary cuda libs installed in the system.

I don't care to use the libs distributed with ollama to begin with (and if bin/ollama defaults to searching libs in ../lib first I don't love that but that's fine).

But I certainly don't care to download the same 1GB of libs every time I update.

(I wonder how many users are like me).

I can haz a version of linux-amd64 without cuda libs included in https://github.com/ollama/ollama/releases prebuilt assets?

...or should I instead just git pull and build the binary from source whenever I want to update (which would be fine with me) or what would you guys recommend?

Originally created by @sevaseva on GitHub (Aug 27, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/6531 Originally assigned to: @dhiltgen on GitHub. I occasionally update ollama on a linux box by downloading URLs like `https://github.com/ollama/ollama/releases/download/v0.3.7-rc6/ollama-linux-amd64.tgz` and extracting/overwriting files into a local directory (not into `/usr` as a root mind you, just into a local directory as a non-privileged user; that is how I prefer to use it). I have necessary cuda libs installed in the system. I don't care to use the libs distributed with ollama to begin with (and if `bin/ollama` defaults to searching libs in `../lib` first I don't love that but that's fine). But I certainly don't care to download the same 1GB of libs every time I update. (I wonder how many users are like me). **I can haz a version of `linux-amd64` without cuda libs included in https://github.com/ollama/ollama/releases prebuilt assets?** ...or should I instead just `git pull` and build the binary from source whenever I want to update (which would be fine with me) or what would you guys recommend?
GiteaMirror added the linuxfeature request labels 2026-04-22 09:08:59 -05:00
Author
Owner

@robert-mcdermott commented on GitHub (Aug 28, 2024):

I have a similar issue. I run Ollama as a non-privileged user on an HPC cluster node. I need to run Ollama in my home folder and don't have access to add the library files to /usr/lib.

<!-- gh-comment-id:2316193973 --> @robert-mcdermott commented on GitHub (Aug 28, 2024): I have a similar issue. I run Ollama as a non-privileged user on an HPC cluster node. I need to run Ollama in my home folder and don't have access to add the library files to /usr/lib.
Author
Owner

@kth8 commented on GitHub (Aug 31, 2024):

I have been using this one liner to extract the latest ollama binary into my local bin folder:

curl -L https://github.com/ollama/ollama/releases/download/$(curl -s https://api.github.com/repos/ollama/ollama/releases/latest | jq -r .tag_name)/ollama-linux-amd64.tgz | tar -xz --strip 2 ./bin/ollama -C ~/.local/bin 
<!-- gh-comment-id:2323053653 --> @kth8 commented on GitHub (Aug 31, 2024): I have been using this one liner to extract the latest ollama binary into my local bin folder: ``` curl -L https://github.com/ollama/ollama/releases/download/$(curl -s https://api.github.com/repos/ollama/ollama/releases/latest | jq -r .tag_name)/ollama-linux-amd64.tgz | tar -xz --strip 2 ./bin/ollama -C ~/.local/bin ```
Author
Owner

@dhiltgen commented on GitHub (Sep 5, 2024):

We're continuing to refine how we package our official builds. PR #6547 will bring some additional improvements to the container image to optimize CUDA and ROCm sizes and startup times for container users. Due to the size of ROCm, we've had to split the dependencies out in a separate tar file as combined it exceeds github's artifact size limit. It's possible we could consider splitting the cuda dependencies out in a similar fashion and then add smarts into the install script to make sure it grabs the correct bundles for the GPUs we detect.

<!-- gh-comment-id:2332267797 --> @dhiltgen commented on GitHub (Sep 5, 2024): We're continuing to refine how we package our official builds. PR #6547 will bring some additional improvements to the container image to optimize CUDA and ROCm sizes and startup times for container users. Due to the size of ROCm, we've had to split the dependencies out in a separate tar file as combined it exceeds github's artifact size limit. It's possible we could consider splitting the cuda dependencies out in a similar fashion and then add smarts into the install script to make sure it grabs the correct bundles for the GPUs we detect.
Author
Owner

@lowlyocean commented on GitHub (Dec 2, 2024):

+1, there are such frequent releases that it's a bit of a waste of bandwidth to be re-downloading CUDA so often (especially if the machine already has CUDA installed system-wide). Have APT / YUM package management (with proper dependency management of shared libs) been decided against?

<!-- gh-comment-id:2511260594 --> @lowlyocean commented on GitHub (Dec 2, 2024): +1, there are such frequent releases that it's a bit of a waste of bandwidth to be re-downloading CUDA so often (especially if the machine already has CUDA installed system-wide). Have APT / YUM package management (with proper dependency management of shared libs) been decided against?
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#29872