[GH-ISSUE #3910] Add OpenELM #64460

Open
opened 2026-05-03 17:44:58 -05:00 by GiteaMirror · 15 comments
Owner

Originally created by @3Samourai on GitHub (Apr 25, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/3910

Apple released several open source LLMs that are designed to run on-device.
Huggingface Link

Originally created by @3Samourai on GitHub (Apr 25, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/3910 Apple released several open source LLMs that are designed to run on-device. [Huggingface Link](https://huggingface.co/apple/OpenELM)
GiteaMirror added the model label 2026-05-03 17:44:58 -05:00
Author
Owner

@thinkverse commented on GitHub (Apr 25, 2024):

Not supported in llama.cpp yet, there's an issue for it, https://github.com/ggerganov/llama.cpp/issues/6868, labeled as a good first issue, if someone with C++ and Python experience wants to tackle it. 👍

<!-- gh-comment-id:2077624837 --> @thinkverse commented on GitHub (Apr 25, 2024): Not supported in llama.cpp yet, there's an issue for it, https://github.com/ggerganov/llama.cpp/issues/6868, labeled as a good first issue, if someone with C++ and Python experience wants to tackle it. 👍
Author
Owner

@mertbozkir commented on GitHub (Apr 26, 2024):

Interesting, I haven't seen this issue, and was trying to upload this model. 🫣

<!-- gh-comment-id:2080056210 --> @mertbozkir commented on GitHub (Apr 26, 2024): Interesting, I haven't seen this issue, and was trying to upload this model. 🫣
Author
Owner

@olumolu commented on GitHub (Apr 27, 2024):

What is the requirement for llama.cpp.
As far as i see ollama does not depend on llama.cpp

<!-- gh-comment-id:2080422387 --> @olumolu commented on GitHub (Apr 27, 2024): What is the requirement for llama.cpp. As far as i see ollama does not depend on llama.cpp
Author
Owner

@olumolu commented on GitHub (Apr 27, 2024):

What is the requirement for llama.cpp.
As far as i see ollama does not depend on llama.cpp

<!-- gh-comment-id:2080431685 --> @olumolu commented on GitHub (Apr 27, 2024): What is the requirement for llama.cpp. As far as i see ollama does not depend on llama.cpp
Author
Owner

@thinkverse commented on GitHub (Apr 27, 2024):

As far as i see ollama does not depend on llama.cpp

Ollama makes heavy use of llama.ccp, it's the backend Ollama uses. When you start Ollama, it starts a llama.cpp server. When you chat with an LLM using Ollama is forwards it to the llama.cpp server.

2bed62926e/llm/server.go (L73)

2bed62926e/server/routes.go (L1315-L1320)

You can see the llama.cpp submodude under https://github.com/ollama/ollama/tree/main/llm.

<!-- gh-comment-id:2080823018 --> @thinkverse commented on GitHub (Apr 27, 2024): > As far as i see ollama does not depend on llama.cpp Ollama makes heavy use of llama.ccp, it's the backend Ollama uses. When you start Ollama, it starts a llama.cpp server. When you chat with an LLM using Ollama is forwards it to the llama.cpp server. https://github.com/ollama/ollama/blob/2bed62926e0add8e5033b61c801cea3d61f8b0d4/llm/server.go#L73 https://github.com/ollama/ollama/blob/2bed62926e0add8e5033b61c801cea3d61f8b0d4/server/routes.go#L1315-L1320 You can see the llama.cpp submodude under https://github.com/ollama/ollama/tree/main/llm.
Author
Owner

@olumolu commented on GitHub (Apr 28, 2024):

https://github.com/ggerganov/llama.cpp/issues/6960

<!-- gh-comment-id:2081399464 --> @olumolu commented on GitHub (Apr 28, 2024): https://github.com/ggerganov/llama.cpp/issues/6960
Author
Owner

@thinkverse commented on GitHub (Apr 29, 2024):

Update on the OpenELM support, a draft PR has been opened https://github.com/ggerganov/llama.cpp/pull/6986, and @joshcarp is looking for anyone to help out. I'm certain it would be appreciated if anyone who has experience with C++, Python, or something related can help. 👍

<!-- gh-comment-id:2083454647 --> @thinkverse commented on GitHub (Apr 29, 2024): Update on the OpenELM support, a draft PR has been opened https://github.com/ggerganov/llama.cpp/pull/6986, and @joshcarp is looking for anyone to help out. I'm certain it would be appreciated if anyone who has experience with C++, Python, or something related can help. 👍
Author
Owner

@UmutAlihan commented on GitHub (Jun 17, 2024):

looking very forward to be able to infer using ollama with these model ggufs: https://huggingface.co/apple/OpenELM

<!-- gh-comment-id:2173023222 --> @UmutAlihan commented on GitHub (Jun 17, 2024): looking very forward to be able to infer using ollama with these model ggufs: https://huggingface.co/apple/OpenELM
Author
Owner

@olumolu commented on GitHub (Jul 4, 2024):

https://github.com/ggerganov/llama.cpp/pull/7359
Now this has been merged so this model can now easily featured in ollama.

<!-- gh-comment-id:2209380884 --> @olumolu commented on GitHub (Jul 4, 2024): https://github.com/ggerganov/llama.cpp/pull/7359 Now this has been merged so this model can now easily featured in ollama.
Author
Owner

@thinkverse commented on GitHub (Jul 4, 2024):

this model can now easily featured in ollama

Ollama needs to update its version of llama.cpp first, maybe https://github.com/ollama/ollama/pull/5475 could be updated to include the OpenELM PR, @jmorganca?

<!-- gh-comment-id:2209385755 --> @thinkverse commented on GitHub (Jul 4, 2024): > this model can now easily featured in ollama Ollama needs to update its version of llama.cpp first, maybe https://github.com/ollama/ollama/pull/5475 could be updated to include the OpenELM PR, @jmorganca?
Author
Owner

@CamJN commented on GitHub (Jul 10, 2024):

llama.cpp d7fd29f adds OpenELM support

the latest release of ollama is commit e4ff73297d, the llama.cpp submodule is pinned to a8db2a9 for that commit

a8db2a9 is after and contains the changes of d7fd29f as can be seen here: a8db2a9ce6/

So adding OpenELM support should be possible now.

<!-- gh-comment-id:2220721537 --> @CamJN commented on GitHub (Jul 10, 2024): llama.cpp [d7fd29f](https://github.com/ggerganov/llama.cpp/commit/d7fd29fff16456ce9c3a23fd2d09a66256b05aff) adds OpenELM support the latest release of ollama is commit e4ff73297db2f53f1ea4b603df5670c5bde6a944, the llama.cpp submodule is pinned to [a8db2a9](https://github.com/ggerganov/llama.cpp/tree/a8db2a9ce64cd4417f6a312ab61858f17f0f8584) for that commit [a8db2a9](https://github.com/ggerganov/llama.cpp/tree/a8db2a9ce64cd4417f6a312ab61858f17f0f8584) is after and contains the changes of [d7fd29f](https://github.com/ggerganov/llama.cpp/commit/d7fd29fff16456ce9c3a23fd2d09a66256b05aff) as can be seen here: https://github.com/ggerganov/llama.cpp/commits/a8db2a9ce64cd4417f6a312ab61858f17f0f8584/ So adding OpenELM support should be possible now.
Author
Owner

@olumolu commented on GitHub (Jul 15, 2024):

What is the status of this when i can assume to have support for this.?

<!-- gh-comment-id:2229108777 --> @olumolu commented on GitHub (Jul 15, 2024): What is the status of this when i can assume to have support for this.?
Author
Owner

@tomasmcm commented on GitHub (Jul 25, 2024):

I've uploaded the 3b-intruct-q5_K_M version here https://ollama.com/tomasmcm/openelm
But it does not seem these models are that great. Unless you ask something simple like Can you translate this to french? "Hi there, how are you doing?" it hallucinates a lot.

<!-- gh-comment-id:2250944272 --> @tomasmcm commented on GitHub (Jul 25, 2024): I've uploaded the **3b-intruct-q5_K_M** version here https://ollama.com/tomasmcm/openelm But it does not seem these models are that great. Unless you ask something simple like `Can you translate this to french? "Hi there, how are you doing?"` it hallucinates a lot.
Author
Owner

@JeremyLaurenson commented on GitHub (Jul 25, 2024):

I think they are really expecting people to fine-tune this for their particular application and so basic sentence understanding is about it

<!-- gh-comment-id:2251026632 --> @JeremyLaurenson commented on GitHub (Jul 25, 2024): I think they are really expecting people to fine-tune this for their particular application and so basic sentence understanding is about it
Author
Owner

@DiazOlveraCo commented on GitHub (Aug 11, 2024):

It's meant to be like a model for edge devices like closer to where the data is created on device, on iPhones.

https://www.reddit.com/r/OpenELM/

<!-- gh-comment-id:2282344167 --> @DiazOlveraCo commented on GitHub (Aug 11, 2024): It's meant to be like a model for edge devices like closer to where the data is created on device, on iPhones. https://www.reddit.com/r/OpenELM/
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#64460