[PR #10716] [MERGED] fix mllama conversion #39211

Closed
opened 2026-04-22 23:51:30 -05:00 by GiteaMirror · 0 comments
Owner

📋 Pull Request Information

Original PR: https://github.com/ollama/ollama/pull/10716
Author: @mxyng
Created: 5/15/2025
Status: Merged
Merged: 5/15/2025
Merged by: @mxyng

Base: mainHead: mxyng/mllama


📝 Commits (1)

📊 Changes

1 file changed (+4 additions, -3 deletions)

View changed files

📝 convert/convert_llama.go (+4 -3)

📄 Description

cross attention Q and K projections needs to have their heads swapped, similar to non-cross attention Q and K tensors


🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.

## 📋 Pull Request Information **Original PR:** https://github.com/ollama/ollama/pull/10716 **Author:** [@mxyng](https://github.com/mxyng) **Created:** 5/15/2025 **Status:** ✅ Merged **Merged:** 5/15/2025 **Merged by:** [@mxyng](https://github.com/mxyng) **Base:** `main` ← **Head:** `mxyng/mllama` --- ### 📝 Commits (1) - [`262c075`](https://github.com/ollama/ollama/commit/262c0757d074fb2ba4b2428f774f8471d7d46d4e) fix mllama conversion ### 📊 Changes **1 file changed** (+4 additions, -3 deletions) <details> <summary>View changed files</summary> 📝 `convert/convert_llama.go` (+4 -3) </details> ### 📄 Description cross attention Q and K projections needs to have their heads swapped, similar to non-cross attention Q and K tensors --- <sub>🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.</sub>
GiteaMirror added the pull-request label 2026-04-22 23:51:30 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#39211