[GH-ISSUE #3285] gemma accuracy down from 0.128 to 0.129 #64062

Open
opened 2026-05-03 16:01:04 -05:00 by GiteaMirror · 3 comments
Owner

Originally created by @RamiKassouf on GitHub (Mar 21, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/3285

What is the issue?

Prompts that were producing the correct results are now producing different (false) outputs

What did you expect to see?

Corretly formatted yaml with correct inputs based on custgom promt

-> Got a yaml with indentation issues and missing fields with wrong structure even though there are json schemas provided

Note that this has also happened on my collegue's pc as well as on a Remote Machine

Steps to reproduce

Upgrade ollama using linux download command provided in website

ollama run gemma

After I rm the model then pull it becomes better for a bit then it get's worse after a while

Are there any recent changes that introduced the issue?

Version Upgrade to 0.129

OS

Linux

Architecture

No response

Platform

WSL2

Ollama version

0.1.29

GPU

Nvidia

GPU info

Quadro P1000

CPU

Intel

Other software

Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz 2.59 GHz

Originally created by @RamiKassouf on GitHub (Mar 21, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/3285 ### What is the issue? Prompts that were producing the correct results are now producing different (false) outputs ### What did you expect to see? Corretly formatted yaml with correct inputs based on custgom promt -> Got a yaml with indentation issues and missing fields with wrong structure even though there are json schemas provided Note that this has also happened on my collegue's pc as well as on a Remote Machine ### Steps to reproduce Upgrade ollama using linux download command provided in website ollama run gemma After I rm the model then pull it becomes better for a bit then it get's worse after a while ### Are there any recent changes that introduced the issue? Version Upgrade to 0.129 ### OS Linux ### Architecture _No response_ ### Platform WSL2 ### Ollama version 0.1.29 ### GPU Nvidia ### GPU info Quadro P1000 ### CPU Intel ### Other software Intel(R) Core(TM) i7-8850H CPU @ 2.60GHz 2.59 GHz
GiteaMirror added the questionbug labels 2026-05-03 16:01:19 -05:00
Author
Owner

@BruceMacD commented on GitHub (Mar 25, 2024):

Hi @RamiKassouf, sorry you're seeing this. I've heard in the past about Gemma getting focused on initial prompts and I'm wondering if that could be the case here. Removing the model and pulling it again shouldn't change the behavior, are you able to get better responses after stopping the model and re-running it too?

<!-- gh-comment-id:2018549161 --> @BruceMacD commented on GitHub (Mar 25, 2024): Hi @RamiKassouf, sorry you're seeing this. I've heard in the past about Gemma getting focused on initial prompts and I'm wondering if that could be the case here. Removing the model and pulling it again shouldn't change the behavior, are you able to get better responses after stopping the model and re-running it too?
Author
Owner

@RamiKassouf commented on GitHub (Mar 26, 2024):

Hi @RamiKassouf, sorry you're seeing this. I've heard in the past about Gemma getting focused on initial prompts and I'm wondering if that could be the case here. Removing the model and pulling it again shouldn't change the behavior, are you able to get better responses after stopping the model and re-running it too?

Yess I do for a little while then it becomes worse
It feels like there's a memory leak or something like that?

<!-- gh-comment-id:2019549914 --> @RamiKassouf commented on GitHub (Mar 26, 2024): > Hi @RamiKassouf, sorry you're seeing this. I've heard in the past about Gemma getting focused on initial prompts and I'm wondering if that could be the case here. Removing the model and pulling it again shouldn't change the behavior, are you able to get better responses after stopping the model and re-running it too? Yess I do for a little while then it becomes worse It feels like there's a memory leak or something like that?
Author
Owner

@oldgithubman commented on GitHub (Mar 30, 2024):

After updating, ollama is definitely using more memory for me. Even before loading any models

<!-- gh-comment-id:2027908107 --> @oldgithubman commented on GitHub (Mar 30, 2024): After updating, ollama is definitely using more memory for me. Even before loading any models
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#64062