[GH-ISSUE #8300] How to use the locally downloaded model in Ollama? #31074

Closed
opened 2026-04-22 11:13:32 -05:00 by GiteaMirror · 4 comments
Owner

Originally created by @OTLCLoud on GitHub (Jan 4, 2025).
Original GitHub issue: https://github.com/ollama/ollama/issues/8300

I have downloaded a model on my server (Qwen2-VL-72B-Instruct), how can I run this model using ollama?

Originally created by @OTLCLoud on GitHub (Jan 4, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/8300 I have downloaded a model on my server (Qwen2-VL-72B-Instruct), how can I run this model using ollama?
GiteaMirror added the model label 2026-04-22 11:13:32 -05:00
Author
Owner

@rick-github commented on GitHub (Jan 4, 2025):

Qwen vision models are not currently supported. https://github.com/ollama/ollama/issues/6564

<!-- gh-comment-id:2570138197 --> @rick-github commented on GitHub (Jan 4, 2025): Qwen vision models are not currently supported. https://github.com/ollama/ollama/issues/6564
Author
Owner

@vYLQs6 commented on GitHub (Jan 5, 2025):

Qwen vision models are not currently supported. #6564

So this commit didn't enable qwen2vl support?

8c9fb8eb73

Refactor mllama image processing code, and add pixtral and qwen2vl

<!-- gh-comment-id:2571663149 --> @vYLQs6 commented on GitHub (Jan 5, 2025): > Qwen vision models are not currently supported. #6564 So this commit didn't enable qwen2vl support? https://github.com/ollama/ollama/commit/8c9fb8eb73afc220e8bf99772572096b6498b748 `Refactor mllama image processing code, and add pixtral and qwen2vl`
Author
Owner

@rick-github commented on GitHub (Jan 5, 2025):

Ah, I hadn't seen that had been merged. The model conversion is next and then we should be good.

<!-- gh-comment-id:2571672350 --> @rick-github commented on GitHub (Jan 5, 2025): Ah, I hadn't seen that had been merged. The model conversion is next and then we should be good.
Author
Owner

@pdevine commented on GitHub (Jan 8, 2025):

That code was only for the image processing (i.e. the part that puts an image into a format that can be read by the vision encoder), not for complete qwen2vl support.

Unfortunately it's still going to be a little bit longer until we can support it. We tried to get this in w/ the old backend engine (llama.cpp/clip) but ended up running into a lot of issues and it was just too buggy for us to release. This should be a lot easier with the new engine (#8301 and #8194).

I'll go ahead and close this as a dupe of #6564

<!-- gh-comment-id:2578278359 --> @pdevine commented on GitHub (Jan 8, 2025): That code was only for the image processing (i.e. the part that puts an image into a format that can be read by the vision encoder), not for complete qwen2vl support. Unfortunately it's still going to be a little bit longer until we can support it. We tried to get this in w/ the old backend engine (llama.cpp/clip) but ended up running into a lot of issues and it was just too buggy for us to release. This should be a lot easier with the new engine (#8301 and #8194). I'll go ahead and close this as a dupe of #6564
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#31074