[GH-ISSUE #14385] [Feature Request] Add Minimax local models to Ollama library #9346

Closed
opened 2026-04-12 22:12:43 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @25r72gnj22-hub on GitHub (Feb 24, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/14385

Feature Request

It would be great to have Minimax models available as local models in Ollama, similar to how Qwen, Llama, and
DeepSeek are available.

Context

  • Currently, Ollama only offers Minimax as a cloud model (minimax-m2.5:cloud)
  • Many users (especially in China) would prefer to run Minimax locally for:
    • Better privacy (data doesn't leave the machine)
    • Lower latency
    • Offline capability
    • Cost savings on API calls

Request

  1. Add Minimax M2.5 and future versions to the Ollama library
  2. Provide both standard and quantized versions for different hardware specs

Why Minimax?

  • Strong reasoning capabilities (comparable to GPT-4)
  • Good Chinese language support
  • Growing user base, especially for bilingual (Chinese/English) tasks
  • Competitive pricing for API, but local deployment would be even better

Alternatives Considered

Using the cloud API is an option, but:

  • Privacy concerns for sensitive conversations
  • Internet dependency
  • Ongoing API costs

Thank you for considering this request! 🚀

Originally created by @25r72gnj22-hub on GitHub (Feb 24, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/14385 ## Feature Request It would be great to have Minimax models available as local models in Ollama, similar to how Qwen, Llama, and DeepSeek are available. ## Context - Currently, Ollama only offers Minimax as a cloud model (`minimax-m2.5:cloud`) - Many users (especially in China) would prefer to run Minimax locally for: - Better privacy (data doesn't leave the machine) - Lower latency - Offline capability - Cost savings on API calls ## Request 1. Add Minimax M2.5 and future versions to the Ollama library 2. Provide both standard and quantized versions for different hardware specs ## Why Minimax? - Strong reasoning capabilities (comparable to GPT-4) - Good Chinese language support - Growing user base, especially for bilingual (Chinese/English) tasks - Competitive pricing for API, but local deployment would be even better ## Alternatives Considered Using the cloud API is an option, but: - Privacy concerns for sensitive conversations - Internet dependency - Ongoing API costs --- Thank you for considering this request! 🚀 ```
GiteaMirror added the feature request label 2026-04-12 22:12:43 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#9346