[PR #3792] [CLOSED] docs: added Ollama Operator into README.md as one of community projects #16553

Closed
opened 2026-04-16 05:34:38 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/3792
Author: @nekomeowww
Created: 4/21/2024
Status: Closed

Base: mainHead: dev/docs-ollama-operator-add


📝 Commits (2)

  • 54d4f01 docs: added Ollama Operator into README.md as one of community projects
  • c9a159b update entry

📊 Changes

1 file changed (+5 additions, -1 deletions)

View changed files

📝 README.md (+5 -1)

📄 Description

First of all, huge thank you to all the wonderful work and awesome contributors to you folks for both ollama, llama.cpp and researchers worked on it.

It made it easy for us to deploy and host our own large language model.

Summary

I'm Neko Ayaka (https://github.com/nekomeowww) from China.
I'm currently working as a full-stack senior developer at @DaoCloud, diving deep into cloud native tech, AI, and UI/UX design.

Ollama is decent to deploy as single instance on machines like Macbook, Mac Studio. I'm really pumped about how we designed the ollama cli. It keeps things simple for users, much like what Docker does (it reminds me Docker CLI every time I use it). And the implementation of Modelfile in the project really streamlines the process, again reminds me Dockerfile.

Additionally, I researched the users needs, and interested in over the official Discord server of Ollama, found that many users are researching to find a way to deploy it concurrently.

Therefore, inspired by the awesome user experience with ollama, I wanted to bring that same vibe to my own Kubernetes setup at home. That's where the idea for this project started, fueled by conversations with friends who are also into cloud native projects.

This is the open-source project called Ollama Operator (GitHub: https://github.com/nekomeowww/ollama-operator, documentation site: https://ollama-operator.ayaka.io/pages/en/ ) I want to introduce and add to README of Ollama here.

It's built around a Kubernetes operator concept by leveraging both ollama pull and ollama serve to make it possible to deploy multiple instances of ollama serve to inference multiple models across nodes of clusters by introducing Model CRD, like this:

apiVersion: ollama.ayaka.io/v1
kind: Model
metadata:
  name: phi
spec:
  # Scale the model to 2 replicas
  replicas: 2
  # Use the model image `phi`
  image: phi

I've mapped out the specs and CRDs for deploying ollama instances on Kubernetes. There are still a few kinks to iron out, but it's looking good as a proof of concept. I've put together all the docs, the architectural design, and even got it up on a neat documentation site powered by VitePress!

Besides Ollama Operator, and the introduced Model CRD to simplify the process of multi-instance deployment, I've also made a CLI tool called kollama (source at: https://github.com/nekomeowww/ollama-operator/tree/main/cmd/kollama , documentation site: https://ollama-operator.ayaka.io/pages/en/references/cli/commands/deploy.html ) to even simplify the interactions with introduced Model CRD with just a single command, like this:

kollama deploy phi --expose

I have Ollama Operator running on our team's dedicated server, my own development kind cluster on my Macbook, and the general K8s cluster on my Homelab, little testing k3s cluster on my two Respberry Pi. They are running smoothly for fine for the past 10 days, and I consider it is a general available project for users to try out and feedback to me to improve the project continously.

There are still much work to do and many things worth me to research and test, I really want to share the concepts, design, simplified concepts with Kubernets.

Proposal

Add and include Ollama Operator into Ollama's README.md file as one of the community driven project.


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/3792 **Author:** [@nekomeowww](https://github.com/nekomeowww) **Created:** 4/21/2024 **Status:** ❌ Closed **Base:** `main` ← **Head:** `dev/docs-ollama-operator-add` --- ### 📝 Commits (2) - [`54d4f01`](https://github.com/ollama/ollama/commit/54d4f01311b382d688e3810024ecb053085520ed) docs: added Ollama Operator into README.md as one of community projects - [`c9a159b`](https://github.com/ollama/ollama/commit/c9a159b357c7d91adb99744b5ce7a5e521625d20) update entry ### 📊 Changes **1 file changed** (+5 additions, -1 deletions) <details> <summary>View changed files</summary> 📝 `README.md` (+5 -1) </details> ### 📄 Description First of all, huge thank you to all the wonderful work and awesome contributors to you folks for both ollama, llama.cpp and researchers worked on it. It made it easy for us to deploy and host our own large language model. ## Summary I'm Neko Ayaka (https://github.com/nekomeowww) from China. I'm currently working as a full-stack senior developer at @DaoCloud, diving deep into cloud native tech, AI, and UI/UX design. Ollama is decent to deploy as single instance on machines like Macbook, Mac Studio. I'm really pumped about how we designed the ollama cli. It keeps things simple for users, much like what Docker does (it reminds me Docker CLI every time I use it). And the implementation of `Modelfile` in the project really streamlines the process, again reminds me `Dockerfile`. Additionally, I researched the users needs, and interested in over the official Discord server of Ollama, found that many users are researching to find a way to deploy it concurrently. Therefore, inspired by the awesome user experience with ollama, I wanted to bring that same vibe to my own Kubernetes setup at home. That's where the idea for this project started, fueled by conversations with friends who are also into cloud native projects. This is the open-source project called Ollama Operator (GitHub: https://github.com/nekomeowww/ollama-operator, documentation site: https://ollama-operator.ayaka.io/pages/en/ ) I want to introduce and add to README of Ollama here. It's built around a Kubernetes operator concept by leveraging both `ollama pull` and `ollama serve` to make it possible to deploy multiple instances of `ollama serve` to inference multiple models across nodes of clusters by introducing [`Model` CRD](https://ollama-operator.ayaka.io/pages/en/references/crd/model.html), like this: ```yaml apiVersion: ollama.ayaka.io/v1 kind: Model metadata: name: phi spec: # Scale the model to 2 replicas replicas: 2 # Use the model image `phi` image: phi ``` I've mapped out the specs and CRDs for deploying ollama instances on Kubernetes. There are still a few kinks to iron out, but it's looking good as a proof of concept. I've put together all the docs, the architectural design, and even got it up on a neat documentation site powered by VitePress! Besides Ollama Operator, and the introduced `Model` CRD to simplify the process of multi-instance deployment, I've also made a CLI tool called `kollama` (source at: https://github.com/nekomeowww/ollama-operator/tree/main/cmd/kollama , documentation site: https://ollama-operator.ayaka.io/pages/en/references/cli/commands/deploy.html ) to even simplify the interactions with introduced `Model` CRD with just a single command, like this: ```shell kollama deploy phi --expose ``` I have Ollama Operator running on our team's dedicated server, my own development `kind` cluster on my Macbook, and the general K8s cluster on my Homelab, little testing k3s cluster on my two Respberry Pi. They are running smoothly for fine for the past 10 days, and I consider it is a general available project for users to try out and feedback to me to improve the project continously. > There are still much work to do and many things worth me to research and test, I really want to share the concepts, design, simplified concepts with Kubernets. ## Proposal Add and include [Ollama Operator](https://github.com/nekomeowww/ollama-operator) into Ollama's `README.md` file as one of the community driven project. --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-16 05:34:38 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#16553