[GH-ISSUE #1930] Support for CogVLM wanted. CogVLM is an alternative for LLaVA #26872

Open
opened 2026-04-22 03:34:29 -05:00 by GiteaMirror · 7 comments
Owner

Originally created by @henryclw on GitHub (Jan 11, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/1930

Currently ollama is supporting LLaVA, which is super great.
I wonder is there a chance to load other similar models like CogVLM?

https://github.com/THUDM/CogVLM

Originally created by @henryclw on GitHub (Jan 11, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/1930 Currently ollama is supporting LLaVA, which is super great. I wonder is there a chance to load other similar models like CogVLM? https://github.com/THUDM/CogVLM
GiteaMirror added the model label 2026-04-22 03:34:29 -05:00
Author
Owner

@easp commented on GitHub (Jan 11, 2024):

At this point the path to Ollama support is via Llama.cpp. It looks like CogVLM hasn't really gained traction there. The one dev who expressed an interest in it also said they all ready have a lot on their plate. Plus it sounds like it could take a lot of work.

https://github.com/ggerganov/llama.cpp/issues/4387

<!-- gh-comment-id:1887949275 --> @easp commented on GitHub (Jan 11, 2024): At this point the path to Ollama support is via Llama.cpp. It looks like CogVLM hasn't really gained traction there. The one dev who expressed an interest in it also said they all ready have a lot on their plate. Plus it sounds like it could take a lot of work. https://github.com/ggerganov/llama.cpp/issues/4387
Author
Owner

@geroldmeisinger commented on GitHub (Jun 4, 2024):

try https://github.com/jhc13/taggui if you look for a simple tool (i can load CogVLM v1 with 12GB VRAM using 4-bit)

<!-- gh-comment-id:2147124422 --> @geroldmeisinger commented on GitHub (Jun 4, 2024): try https://github.com/jhc13/taggui if you look for a simple tool (i can load CogVLM v1 with 12GB VRAM using 4-bit)
Author
Owner

@geroldmeisinger commented on GitHub (Jun 4, 2024):

CogVLM2: https://github.com/THUDM/CogVLM2
int4 version: https://huggingface.co/THUDM/cogvlm2-llama3-chat-19B-int4

supported in https://github.com/jhc13/taggui

<!-- gh-comment-id:2148217697 --> @geroldmeisinger commented on GitHub (Jun 4, 2024): CogVLM2: https://github.com/THUDM/CogVLM2 int4 version: https://huggingface.co/THUDM/cogvlm2-llama3-chat-19B-int4 supported in https://github.com/jhc13/taggui
Author
Owner

@geroldmeisinger commented on GitHub (Jun 23, 2024):

At this point the path to Ollama support is via Llama.cpp.

although I quite don't stand why it HAS to be this way if it doesn't get support in llama.cpp . CogVLM2 works in python and the mentioned taggui tool already. is there something fundamental in Ollama which makes it difficult to just load it via python bindings and forego llama.cpp for the time being?

<!-- gh-comment-id:2185306705 --> @geroldmeisinger commented on GitHub (Jun 23, 2024): > At this point the path to Ollama support is via Llama.cpp. although I quite don't stand why it HAS to be this way if it doesn't get support in llama.cpp . CogVLM2 works in python and the mentioned taggui tool already. is there something fundamental in Ollama which makes it difficult to just load it via python bindings and forego llama.cpp for the time being?
Author
Owner

@easp commented on GitHub (Jun 24, 2024):

I haven't been following the project as closely as I used to, but so far as I know, Ollama doesn't load anything via python bindings and forgo llama.cpp.

<!-- gh-comment-id:2185466133 --> @easp commented on GitHub (Jun 24, 2024): I haven't been following the project as closely as I used to, but so far as I know, Ollama doesn't load anything via python bindings and forgo llama.cpp.
Author
Owner

@geroldmeisinger commented on GitHub (Jun 24, 2024):

but is there anything which keeps us from doing that?

<!-- gh-comment-id:2185735896 --> @geroldmeisinger commented on GitHub (Jun 24, 2024): but is there anything which keeps us from doing that?
Author
Owner

@easp commented on GitHub (Jun 24, 2024):

Who is "us?"

The Ollama maintainers ultimately decide what the project does. I don't think they've supported any backends other than llama.cpp at this point. In addition, last I'd looked, they were pretty conservative about the sorts of community contributions they've accepted. There is, of course, the option of forking the project.

<!-- gh-comment-id:2187142631 --> @easp commented on GitHub (Jun 24, 2024): Who is "us?" The Ollama maintainers ultimately decide what the project does. I don't think they've supported any backends other than llama.cpp at this point. In addition, last I'd looked, they were pretty conservative about the sorts of community contributions they've accepted. There is, of course, the option of forking the project.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#26872