[GH-ISSUE #12349] Qwen3-Next #8203

Closed
opened 2026-04-12 20:39:55 -05:00 by GiteaMirror · 9 comments
Owner
Originally created by @chigkim on GitHub (Sep 19, 2025). Original GitHub issue: https://github.com/ollama/ollama/issues/12349 https://huggingface.co/collections/Qwen/qwen3-next-68c25fd6838e585db8eeea9d?spm=a2ty_o06.30285417.0.0.4766c921yrdQpe https://qwen.ai/blog?id=4074cca80393150c248e508aa62983f9cb7d27cd&from=research.latest-advancements-list Thanks!
GiteaMirror added the model label 2026-04-12 20:39:55 -05:00
Author
Owner

@rick-github commented on GitHub (Sep 19, 2025):

https://github.com/ggml-org/llama.cpp/pull/16095

<!-- gh-comment-id:3314177913 --> @rick-github commented on GitHub (Sep 19, 2025): https://github.com/ggml-org/llama.cpp/pull/16095
Author
Owner

@timbmg commented on GitHub (Sep 23, 2025):

ggml-org/llama.cpp#16095

Sorry, maybe a dumb question, but do you mean it can only be added once it has been adopted by llama.cpp?

<!-- gh-comment-id:3324143416 --> @timbmg commented on GitHub (Sep 23, 2025): > [ggml-org/llama.cpp#16095](https://github.com/ggml-org/llama.cpp/pull/16095) Sorry, maybe a dumb question, but do you mean it can only be added once it has been adopted by llama.cpp?
Author
Owner

@rick-github commented on GitHub (Sep 23, 2025):

There's ongoing work to include it in llama.cpp, so it will eventually be available through the old engine.

It's possible that the ollama developers are also working on including it in the new engine (like they did for gpt-oss), but if that work is underway it's not visible through the public repo.

<!-- gh-comment-id:3324162285 --> @rick-github commented on GitHub (Sep 23, 2025): There's ongoing work to include it in llama.cpp, so it will eventually be available through the old engine. It's possible that the ollama developers are also working on including it in the new engine (like they did for gpt-oss), but if that work is underway it's not visible through the public repo.
Author
Owner

@paolss commented on GitHub (Oct 1, 2025):

this is hard model i can say :) fast good but ... as f hard to implement in vllm llama.cpp etc... ehh.. but worth it

<!-- gh-comment-id:3354310551 --> @paolss commented on GitHub (Oct 1, 2025): this is hard model i can say :) fast good but ... as f hard to implement in vllm llama.cpp etc... ehh.. but worth it
Author
Owner

@chigkim commented on GitHub (Oct 5, 2025):

It works nicely in MLX.

<!-- gh-comment-id:3369517008 --> @chigkim commented on GitHub (Oct 5, 2025): It works nicely in MLX.
Author
Owner

@paolss commented on GitHub (Oct 6, 2025):

or in vllm... basically im using it with vllm there is no support for nvidia in lmstudio and ollama - vllm and sglang are working and this model is really good and fast as ...

<!-- gh-comment-id:3369550228 --> @paolss commented on GitHub (Oct 6, 2025): or in vllm... basically im using it with vllm there is no support for nvidia in lmstudio and ollama - vllm and sglang are working and this model is really good and fast as ...
Author
Owner

@ghecko commented on GitHub (Nov 28, 2025):

Llama.cpp now support it. Any timeframe on when this will be supported by ollama ?

<!-- gh-comment-id:3589600861 --> @ghecko commented on GitHub (Nov 28, 2025): Llama.cpp now support it. Any timeframe on when this will be supported by ollama ?
Author
Owner

@mehditahmasebi commented on GitHub (Dec 3, 2025):

Qwen3 next 80b 3A is one of the best

<!-- gh-comment-id:3607865249 --> @mehditahmasebi commented on GitHub (Dec 3, 2025): Qwen3 next 80b 3A is one of the best
Author
Owner

@rick-github commented on GitHub (Dec 3, 2025):

The next vendor sync after b7186 will make support available in ollama.

<!-- gh-comment-id:3607876331 --> @rick-github commented on GitHub (Dec 3, 2025): The next vendor sync after [b7186](https://github.com/ggml-org/llama.cpp/releases/tag/b7186) will make support available in ollama.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#8203