[GH-ISSUE #15519] Add MERaLiON-2 multimodal audio model support #56432

Open
opened 2026-04-29 10:49:09 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @SiruiHe on GitHub (Apr 12, 2026).
Original GitHub issue: https://github.com/ollama/ollama/issues/15519

MERaLiON-2 is a speech-text model from A*STAR Singapore. It pairs a localized Whisper-large-v3 encoder with a Gemma-2 decoder, and covers English, Mandarin, Malay, and Tamil reasonably well, with a focus on Singapore's multilingual context. There are 3B and 10B variants.

The inference support landed in llama.cpp on April 11 (build b8762, ggml-org/llama.cpp#21756). GGUF files are on HuggingFace:

The architecture adds one new projector type (PROJECTOR_TYPE_MERALION) to the existing mtmd framework, following the same pattern as Ultravox and Voxtral. It needs llama.cpp at b8762 or later to run.

I noticed Ollama is currently pinned to ec98e200 from December 2025, so this would require a llama.cpp bump to pick up. Posting here mainly to flag interest and check if there is a preferred process for getting this into the library. Happy to help with a Modelfile or anything else on our end.

Originally created by @SiruiHe on GitHub (Apr 12, 2026). Original GitHub issue: https://github.com/ollama/ollama/issues/15519 MERaLiON-2 is a speech-text model from A*STAR Singapore. It pairs a localized Whisper-large-v3 encoder with a Gemma-2 decoder, and covers English, Mandarin, Malay, and Tamil reasonably well, with a focus on Singapore's multilingual context. There are 3B and 10B variants. The inference support landed in llama.cpp on April 11 (build b8762, ggml-org/llama.cpp#21756). GGUF files are on HuggingFace: - 3B: https://huggingface.co/MERaLiON/MERaLiON-2-3B-GGUF - 10B: https://huggingface.co/MERaLiON/MERaLiON-2-10B-GGUF The architecture adds one new projector type (PROJECTOR_TYPE_MERALION) to the existing mtmd framework, following the same pattern as Ultravox and Voxtral. It needs llama.cpp at b8762 or later to run. I noticed Ollama is currently pinned to ec98e200 from December 2025, so this would require a llama.cpp bump to pick up. Posting here mainly to flag interest and check if there is a preferred process for getting this into the library. Happy to help with a Modelfile or anything else on our end.
Author
Owner

@PureBlissAK commented on GitHub (Apr 18, 2026):

🤖 Automated Triage & Analysis Report

Issue: #15519
Analyzed: 2026-04-18T18:21:23.149787

Analysis

  • Type: unknown
  • Severity: medium
  • Components: unknown

Implementation Plan

  • Effort: medium
  • Steps:

This issue has been triaged and marked for implementation.

<!-- gh-comment-id:4274308218 --> @PureBlissAK commented on GitHub (Apr 18, 2026): <!-- ollama-issue-orchestrator:v1 issue:15519 --> ## 🤖 Automated Triage & Analysis Report **Issue**: #15519 **Analyzed**: 2026-04-18T18:21:23.149787 ### Analysis - **Type**: unknown - **Severity**: medium - **Components**: unknown ### Implementation Plan - **Effort**: medium - **Steps**: *This issue has been triaged and marked for implementation.*
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#56432