[PR #7217] [MERGED] Add arm64 cuda jetpack variants #43602

Closed
opened 2026-04-24 23:12:42 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/7217
Author: @dhiltgen
Created: 10/15/2024
Status: Merged
Merged: 11/12/2024
Merged by: @dhiltgen

Base: mainHead: go_server_jetson


📝 Commits (1)

  • ce1567a Jetpack support for Go server

📊 Changes

8 files changed (+78 additions, -20 deletions)

View changed files

📝 Dockerfile (+64 -8)
📝 discover/amd_linux.go (+1 -1)
📝 discover/amd_windows.go (+1 -1)
📝 discover/gpu.go (+4 -4)
📝 discover/types.go (+1 -1)
📝 llama/llama.go (+4 -2)
📝 llama/make/cuda.make (+1 -1)
📝 llm/server.go (+2 -2)

📄 Description

This adds 2 new variants for the arm64 build to support nvidia jetson systems based on jetpack 5 and 6. Jetpack 4 is too old to be built with our toolchain (the older cuda requires an old gcc which can't build llama.cpp) and will remain unsupported.

The sbsa discrete GPU cuda libraries we bundle in the existing arm64 build are incompatible with jetson iGPU systems. Unfortunately swapping them at runtime isn't viable given the way nvcc compilation/linking works, so we need to actually build and link against those specific cuda libraries, and bundle them.

File sizes are too large to try to combine into a single unified tgz bundle, so this splits things out into a main bundle which contains all the runners, and then two auxiliary bundles, one for jetpack 5 and 6 which contain all the libraries specific to those versions.

Fixes https://github.com/ollama/ollama/issues/2408
Fixes https://github.com/ollama/ollama/issues/4693
Fixes https://github.com/ollama/ollama/issues/5100
Fixes https://github.com/ollama/ollama/issues/4861
Fixes #6999
Fixes #7293


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/7217 **Author:** [@dhiltgen](https://github.com/dhiltgen) **Created:** 10/15/2024 **Status:** ✅ Merged **Merged:** 11/12/2024 **Merged by:** [@dhiltgen](https://github.com/dhiltgen) **Base:** `main` ← **Head:** `go_server_jetson` --- ### 📝 Commits (1) - [`ce1567a`](https://github.com/ollama/ollama/commit/ce1567ae5d4197be34a964ae50b1233616af4476) Jetpack support for Go server ### 📊 Changes **8 files changed** (+78 additions, -20 deletions) <details> <summary>View changed files</summary> 📝 `Dockerfile` (+64 -8) 📝 `discover/amd_linux.go` (+1 -1) 📝 `discover/amd_windows.go` (+1 -1) 📝 `discover/gpu.go` (+4 -4) 📝 `discover/types.go` (+1 -1) 📝 `llama/llama.go` (+4 -2) 📝 `llama/make/cuda.make` (+1 -1) 📝 `llm/server.go` (+2 -2) </details> ### 📄 Description This adds 2 new variants for the arm64 build to support nvidia jetson systems based on jetpack 5 and 6. Jetpack 4 is too old to be built with our toolchain (the older cuda requires an old gcc which can't build llama.cpp) and will remain unsupported. The sbsa discrete GPU cuda libraries we bundle in the existing arm64 build are incompatible with jetson iGPU systems. Unfortunately swapping them at runtime isn't viable given the way nvcc compilation/linking works, so we need to actually build and link against those specific cuda libraries, and bundle them. File sizes are too large to try to combine into a single unified tgz bundle, so this splits things out into a main bundle which contains all the runners, and then two auxiliary bundles, one for jetpack 5 and 6 which contain all the libraries specific to those versions. Fixes https://github.com/ollama/ollama/issues/2408 Fixes https://github.com/ollama/ollama/issues/4693 Fixes https://github.com/ollama/ollama/issues/5100 Fixes https://github.com/ollama/ollama/issues/4861 Fixes #6999 Fixes #7293 --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-24 23:12:42 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#43602