mirror of
https://github.com/ollama/ollama.git
synced 2025-12-05 18:46:22 -06:00
Gemma 2 9B and 27B is not behaving right #3347
Closed
opened 2025-11-12 11:33:29 -06:00 by GiteaMirror
·
20 comments
No Branch/Tag Specified
main
parth/olmo3-thinking-renderer
drifkin/responses
jmorganca/write-int32
parth/render-template
jmorganca/intellect-3
parth/remove-prealloc-linter
nicole/olmo-model
jmorganca/cmd-eval
grace/deepseek-parser
nicole/nomic-embed-text-fix
jessegross/multi_chunk_reserve
grace/deepseek-renderer-parser
hoyyeva/proxy-ollama-serve
hoyyeva/model-capabilities-after-download
mxyng/ropeopts
mxyng/lint-2
nicole/nomic-embed-text-v2
hoyyeva/add-gemini-3-pro-preview
hoyyeva/load-model-list
hoyyeva/app-ollama-apis
mxyng/expand-path
mxyng/environ-2
hoyyeva/thinking-rendering
nicole/truncation
hoyyeva/deeplink-json-encoding
parth/improve-tool-calling-tests
hoyyeva/conversation
hoyyeva/assistant-edit-response
hoyyeva/thinking
origin/brucemacd/invalid-char-i-err
parth/improve-tool-calling
jmorganca/required-omitempty
grace/qwen3-vl-tests
mxyng/iter-client
parth/docs-readme
nicole/embed-test
pdevine/integration-benchstat
parth/remove-generate-cmd
parth/add-toolcall-id
mxyng/server-tests
jmorganca/glm-4.6
jmorganca/gin-h-compat
drifkin/stable-tool-args
pdevine/qwen3-more-thinking
parth/add-websearch-client
nicole/websearch_local
jmorganca/qwen3-coder-updates
grace/deepseek-v3-migration-tests
mxyng/fix-create
jmorganca/cloud-errors
pdevine/parser-tidy
revert-12233-parth/simplify-entrypoints-runner
parth/enable-so-gpt-oss
brucemacd/qwen3vl
jmorganca/readme-simplify
parth/gpt-oss-structured-outputs
revert-12039-jmorganca/tools-braces
mxyng/embeddings
mxyng/gguf
mxyng/benchmark
mxyng/types-null
parth/move-parsing
mxyng/gemma2
jmorganca/docs
mxyng/16-bit
mxyng/create-stdin
pdevine/authorizedkeys
mxyng/quant
parth/opt-in-error-context-window
brucemacd/cache-models
brucemacd/runner-completion
jmorganca/llama-update-6
brucemacd/benchmark-list
brucemacd/partial-read-caps
parth/deepseek-r1-tools
mxyng/omit-array
parth/tool-prefix-temp
brucemacd/runner-test
jmorganca/qwen25vl
brucemacd/model-forward-test-ext
parth/python-function-parsing
jmorganca/cuda-compression-none
drifkin/num-parallel
drifkin/chat-truncation-fix
jmorganca/sync
parth/python-tools-calling
drifkin/array-head-count
brucemacd/create-no-loop
parth/server-enable-content-stream-with-tools
qwen25omni
mxyng/v3
brucemacd/ropeconfig
jmorganca/silence-tokenizer
parth/sample-so-test
parth/sampling-structured-outputs
brucemacd/doc-go-engine
parth/constrained-sampling-json
jmorganca/mistral-wip
brucemacd/mistral-small-convert
parth/sample-unmarshal-json-for-params
brucemacd/jomorganca/mistral
pdevine/bfloat16
jmorganca/mistral
brucemacd/mistral
pdevine/logging
parth/sample-correctness-fix
parth/sample-fix-sorting
jmorgan/sample-fix-sorting-extras
jmorganca/temp-0-images
brucemacd/parallel-embed-models
brucemacd/shim-grammar
jmorganca/fix-gguf-error
bmizerany/nameswork
jmorganca/faster-releases
bmizerany/validatenames
brucemacd/err-no-vocab
brucemacd/rope-config
brucemacd/err-hint
brucemacd/qwen2_5
brucemacd/logprobs
brucemacd/new_runner_graph_bench
progress-flicker
brucemacd/forward-test
brucemacd/go_qwen2
pdevine/gemma2
jmorganca/add-missing-symlink-eval
mxyng/next-debug
parth/set-context-size-openai
brucemacd/next-bpe-bench
brucemacd/next-bpe-test
brucemacd/new_runner_e2e
brucemacd/new_runner_qwen2
pdevine/convert-cohere2
brucemacd/convert-cli
parth/log-probs
mxyng/next-mlx
mxyng/cmd-history
parth/templating
parth/tokenize-detokenize
brucemacd/check-key-register
bmizerany/grammar
jmorganca/vendor-081b29bd
mxyng/func-checks
jmorganca/fix-null-format
parth/fix-default-to-warn-json
jmorganca/qwen2vl
jmorganca/no-concat
parth/cmd-cleanup-SO
brucemacd/check-key-register-structured-err
parth/openai-stream-usage
parth/fix-referencing-so
stream-tools-stop
jmorganca/degin-1
brucemacd/install-path-clean
brucemacd/push-name-validation
brucemacd/browser-key-register
jmorganca/openai-fix-first-message
jmorganca/fix-proxy
jessegross/sample
parth/disallow-streaming-tools
dhiltgen/remove_submodule
jmorganca/ga
jmorganca/mllama
pdevine/newlines
pdevine/geems-2b
jmorganca/llama-bump
mxyng/modelname-7
mxyng/gin-slog
mxyng/modelname-6
jyan/convert-prog
jyan/quant5
paligemma-support
pdevine/import-docs
jmorganca/openai-context
jyan/paligemma
jyan/p2
jyan/palitest
bmizerany/embedspeedup
jmorganca/llama-vit
brucemacd/allow-ollama
royh/ep-methods
royh/whisper
mxyng/api-models
mxyng/fix-memory
jyan/q4_4/8
jyan/ollama-v
royh/stream-tools
roy-embed-parallel
bmizerany/hrm
revert-5963-revert-5924-mxyng/llama3.1-rope
royh/embed-viz
jyan/local2
jyan/auth
jyan/local
jyan/parse-temp
jmorganca/template-mistral
jyan/reord-g
royh-openai-suffixdocs
royh-imgembed
royh-embed-parallel
jyan/quant4
royh-precision
jyan/progress
pdevine/fix-template
jyan/quant3
pdevine/ggla
mxyng/update-registry-domain
jmorganca/ggml-static
mxyng/create-context
jyan/v0.146
mxyng/layers-from-files
build_dist
bmizerany/noseek
royh-ls
royh-name
timeout
mxyng/server-timestamp
bmizerany/nosillyggufslurps
royh-params
jmorganca/llama-cpp-7c26775
royh-openai-delete
royh-show-rigid
jmorganca/enable-fa
jmorganca/no-error-template
jyan/format
royh-testdelete
bmizerany/fastverify
language_support
pdevine/ps-glitches
brucemacd/tokenize
bruce/iq-quants
bmizerany/filepathwithcoloninhost
mxyng/split-bin
bmizerany/client-registry
jmorganca/if-none-match
native
jmorganca/native
jmorganca/batch-embeddings
jmorganca/initcmake
jmorganca/mm
pdevine/showggmlinfo
modenameenforcealphanum
bmizerany/modenameenforcealphanum
jmorganca/done-reason
jmorganca/llama-cpp-8960fe8
ollama.com
bmizerany/filepathnobuild
bmizerany/types/model/defaultfix
rmdisplaylong
nogogen
bmizerany/x
modelfile-readme
bmizerany/replacecolon
jmorganca/limit
jmorganca/execstack
jmorganca/replace-assets
mxyng/tune-concurrency
jmorganca/testing
whitespace-detection
jmorganca/options
upgrade-all
scratch
cuda-search
mattw/airenamer
mattw/allmodelsonhuggingface
mattw/quantcontext
mattw/whatneedstorun
brucemacd/llama-mem-calc
mattw/faq-context
mattw/communitylinks
mattw/noprune
mattw/python-functioncalling
rename
mxyng/install
pulse
remove-first
editor
mattw/selfqueryingretrieval
cgo
mattw/howtoquant
api
matt/streamingapi
format-config
mxyng/extra-args
shell
update-nous-hermes
cp-model
upload-progress
fix-unknown-model
fix-model-names
delete-fix
insecure-registry
ls
deletemodels
progressbar
readme-updates
license-layers
skip-list
list-models
modelpath
matt/examplemodelfiles
distribution
go-opts
v0.13.2-rc2
v0.13.2-rc1
v0.13.2-rc0
v0.13.1
v0.13.1-rc2
v0.13.1-rc1
v0.13.1-rc0
v0.13.0
v0.13.0-rc0
v0.12.11
v0.12.11-rc1
v0.12.11-rc0
v0.12.10
v0.12.10-rc1
v0.12.10-rc0
v0.12.9-rc0
v0.12.9
v0.12.8
v0.12.8-rc0
v0.12.7
v0.12.7-rc1
v0.12.7-rc0
v0.12.6
v0.12.6-rc1
v0.12.6-rc0
v0.12.5
v0.12.5-rc0
v0.12.4
v0.12.4-rc7
v0.12.4-rc6
v0.12.4-rc5
v0.12.4-rc4
v0.12.4-rc3
v0.12.4-rc2
v0.12.4-rc1
v0.12.4-rc0
v0.12.3
v0.12.2
v0.12.2-rc0
v0.12.1
v0.12.1-rc1
v0.12.1-rc2
v0.12.1-rc0
v0.12.0
v0.12.0-rc1
v0.12.0-rc0
v0.11.11
v0.11.11-rc3
v0.11.11-rc2
v0.11.11-rc1
v0.11.11-rc0
v0.11.10
v0.11.9
v0.11.9-rc0
v0.11.8
v0.11.8-rc0
v0.11.7-rc1
v0.11.7-rc0
v0.11.7
v0.11.6
v0.11.6-rc0
v0.11.5-rc4
v0.11.5-rc3
v0.11.5
v0.11.5-rc5
v0.11.5-rc2
v0.11.5-rc1
v0.11.5-rc0
v0.11.4
v0.11.4-rc0
v0.11.3
v0.11.3-rc0
v0.11.2
v0.11.1
v0.11.0
v0.10.1
v0.10.0
v0.10.0-rc4
v0.10.0-rc3
v0.10.0-rc2
v0.10.0-rc1
v0.10.0-rc0
v0.9.7-rc1
v0.9.7-rc0
v0.9.6
v0.9.6-rc0
v0.9.5
v0.9.4-rc5
v0.9.4-rc6
v0.9.4
v0.9.4-rc3
v0.9.4-rc4
v0.9.4-rc1
v0.9.4-rc2
v0.9.4-rc0
v0.9.3
v0.9.3-rc5
v0.9.4-citest0
v0.9.3-rc4
v0.9.3-rc3
v0.9.3-rc2
v0.9.3-rc1
v0.9.3-rc0
v0.9.2
v0.9.1
v0.9.1-rc1
v0.9.1-rc0
v0.9.0
v0.9.0-rc0
v0.8.0
v0.8.0-rc0
v0.7.1-rc2
v0.7.1
v0.7.1-rc1
v0.7.1-rc0
v0.7.0
v0.7.0-rc1
v0.7.0-rc0
v0.6.8
v0.6.8-rc0
v0.6.7
v0.6.7-rc2
v0.6.7-rc1
v0.6.7-rc0
v0.6.6
v0.6.6-rc2
v0.6.6-rc1
v0.6.6-rc0
v0.6.5-rc1
v0.6.5
v0.6.5-rc0
v0.6.4-rc0
v0.6.4
v0.6.3-rc1
v0.6.3
v0.6.3-rc0
v0.6.2
v0.6.2-rc0
v0.6.1
v0.6.1-rc0
v0.6.0-rc0
v0.6.0
v0.5.13
v0.5.13-rc6
v0.5.13-rc5
v0.5.13-rc4
v0.5.13-rc3
v0.5.13-rc2
v0.5.13-rc1
v0.5.13-rc0
v0.5.12
v0.5.12-rc1
v0.5.12-rc0
v0.5.11
v0.5.10
v0.5.9
v0.5.9-rc0
v0.5.8-rc13
v0.5.8
v0.5.8-rc12
v0.5.8-rc11
v0.5.8-rc10
v0.5.8-rc9
v0.5.8-rc8
v0.5.8-rc7
v0.5.8-rc6
v0.5.8-rc5
v0.5.8-rc4
v0.5.8-rc3
v0.5.8-rc2
v0.5.8-rc1
v0.5.8-rc0
v0.5.7
v0.5.6
v0.5.5
v0.5.5-rc0
v0.5.4
v0.5.3
v0.5.3-rc0
v0.5.2
v0.5.2-rc3
v0.5.2-rc2
v0.5.2-rc1
v0.5.2-rc0
v0.5.1
v0.5.0
v0.5.0-rc1
v0.4.8-rc0
v0.4.7
v0.4.6
v0.4.5
v0.4.4
v0.4.3
v0.4.3-rc0
v0.4.2
v0.4.2-rc1
v0.4.2-rc0
v0.4.1
v0.4.1-rc0
v0.4.0
v0.4.0-rc8
v0.4.0-rc7
v0.4.0-rc6
v0.4.0-rc5
v0.4.0-rc4
v0.4.0-rc3
v0.4.0-rc2
v0.4.0-rc1
v0.4.0-rc0
v0.4.0-ci3
v0.3.14
v0.3.14-rc0
v0.3.13
v0.3.12
v0.3.12-rc5
v0.3.12-rc4
v0.3.12-rc3
v0.3.12-rc2
v0.3.12-rc1
v0.3.11
v0.3.11-rc4
v0.3.11-rc3
v0.3.11-rc2
v0.3.11-rc1
v0.3.10
v0.3.10-rc1
v0.3.9
v0.3.8
v0.3.7
v0.3.7-rc6
v0.3.7-rc5
v0.3.7-rc4
v0.3.7-rc3
v0.3.7-rc2
v0.3.7-rc1
v0.3.6
v0.3.5
v0.3.4
v0.3.3
v0.3.2
v0.3.1
v0.3.0
v0.2.8
v0.2.8-rc2
v0.2.8-rc1
v0.2.7
v0.2.6
v0.2.5
v0.2.4
v0.2.3
v0.2.2
v0.2.2-rc2
v0.2.2-rc1
v0.2.1
v0.2.0
v0.1.49-rc14
v0.1.49-rc13
v0.1.49-rc12
v0.1.49-rc11
v0.1.49-rc10
v0.1.49-rc9
v0.1.49-rc8
v0.1.49-rc7
v0.1.49-rc6
v0.1.49-rc4
v0.1.49-rc5
v0.1.49-rc3
v0.1.49-rc2
v0.1.49-rc1
v0.1.48
v0.1.47
v0.1.46
v0.1.45-rc5
v0.1.45
v0.1.45-rc4
v0.1.45-rc3
v0.1.45-rc2
v0.1.45-rc1
v0.1.44
v0.1.43
v0.1.42
v0.1.41
v0.1.40
v0.1.40-rc1
v0.1.39
v0.1.39-rc2
v0.1.39-rc1
v0.1.38
v0.1.37
v0.1.36
v0.1.35
v0.1.35-rc1
v0.1.34
v0.1.34-rc1
v0.1.33
v0.1.33-rc7
v0.1.33-rc6
v0.1.33-rc5
v0.1.33-rc4
v0.1.33-rc3
v0.1.33-rc2
v0.1.33-rc1
v0.1.32
v0.1.32-rc2
v0.1.32-rc1
v0.1.31
v0.1.30
v0.1.29
v0.1.28
v0.1.27
v0.1.26
v0.1.25
v0.1.24
v0.1.23
v0.1.22
v0.1.21
v0.1.20
v0.1.19
v0.1.18
v0.1.17
v0.1.16
v0.1.15
v0.1.14
v0.1.13
v0.1.12
v0.1.11
v0.1.10
v0.1.9
v0.1.8
v0.1.7
v0.1.6
v0.1.5
v0.1.4
v0.1.3
v0.1.2
v0.1.1
v0.1.0
v0.0.21
v0.0.20
v0.0.19
v0.0.18
v0.0.17
v0.0.16
v0.0.15
v0.0.14
v0.0.13
v0.0.12
v0.0.11
v0.0.10
v0.0.9
v0.0.8
v0.0.7
v0.0.6
v0.0.5
v0.0.4
v0.0.3
v0.0.2
v0.0.1
Labels
Clear labels
amd
api
app
bug
build
cli
client2
cloud
compatibility
context-length
create
docker
documentation
embeddings
engine
feature request
feedback wanted
good first issue
gpt-oss
gpu
harmony
help wanted
install
integration
intel
js
linux
macos
memory
model
needs more info
networking
nvidia
ollama.com
performance
pull-request
Mirrored from GitHub Pull Request
python
question
registry
rendering
thinking
tools
top
windows
wsl
No Label
bug
Milestone
No items
No Milestone
Projects
Clear projects
No project
No Assignees
Notifications
Due Date
No due date set.
Dependencies
No dependencies set.
Reference: github-starred/ollama-ollama#3347
Reference in New Issue
Block a user
Blocking a user prevents them from interacting with repositories, such as opening or commenting on pull requests or issues. Learn more about blocking a user.
No description provided.
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
Originally created by @jayakumark on GitHub (Jun 27, 2024).
Originally assigned to: @jmorganca on GitHub.
What is the issue?
Try this in ollama Gemma 2 9B or 27B, it just never stops.
Give a succinct summary of the entire email conversation in not more than 40 words,
Emails To Andrew Fastow:
An 11 million dollar financial deal:
| william.giuliani@enron.com | andrew.fastow@enron.com | 2001-06-07 07:48:00 | Gentlemen: Attached is the DASH for the approval of the DPR Accelerated Put transaction. This partial divestiture allows us to put $11 million of our equity interest back to DPR Holding Company, LLC and its subsidiary, Dakota, LLC. Both entities are controlled by Chris Cline. In addition to redeeming part of our equity interest, the deal provides us 900,000 tons of coal priced below market, an option which could lead to a very profitable synfuel project, and the potential for more marketing fees from other Cline entities. The DASH has been approved and signed by RAC and JEDI II, and is now awaiting Mark Haedicke s review and approval. I wanted to give you the opportunity to review the DASH and become familiar with the provisions of the deal. If you have any questions on the transaction, feel free to contact me at (412) 490-9048. Others familiar with the deal are Mike Beyer, George McClellan, and Wayne Gresham. Thank you. Bill Giuliani
—-
Enron to form one corporate equity investment unit, Enron Principal Investments:
| steven.kean@enron.com | andrew.fastow@enron.com | 2001-06-12 02:09:00 | As we discussed… The other memo will follow shortly from Maureen McVicker (my assistant).———————- Forwarded by Steven J Kean/NA/Enron on 06/12/2001 09:08 AM —————————From: Sherri Sera/ENRON@enronXgate on 06/11/2001 04:51 PMTo: Steven J Kean/NA/Enron@Enroncc: Subject: FW: Draft of Organizational AnnouncementSteve, Kevin Garland sent this to me hoping to get Jeff s approval to send it out from the office of the chairman. Would it make sense to incorporate it into the memo you re working on? Please advise. Thanks, SRS —–Original Message—–From: Garland, Kevin Sent: Monday, June 11, 2001 3:28 PMTo: Sera, SherriSubject: FW: Draft of Organizational AnnouncementAnnouncing the Formation of One Corporate Equity Investing UnitTo better develop and manage equity investment opportunities related to our core businesses, Enron has formed one corporate equity investment unit. This new unit, Enron Principal Investments, will combine the existing investment units of ENA, EBS and Enron Investment Partners. Additionally, the Enron Special Asset Group will also become part of Enron Principal Investments. The strategy of Enron Principal Investments will be to work with all the business units of Enron to identify, execute, and manage equity investments, which leverage Enron s unique and proprietary knowledge. These investments may be in the form of venture capital, LBO s, traditional private equity and distressed debt positions. Kevin Garland will serve as Managing Director, overseeing all activities of Enron Principal Investments. Gene Humphrey, Michael Miller, Dick Lydecker, and their groups, will join Kevin and his group to form Enron Principal Investments. This new business unit will report to an investment committee, consisting of Greg Whalley, Ken Rice and Dave Delainey. Please join me in congratulating and supporting Kevin, Gene, Michael, Dick and the other members of this group in this effort.Jeff Skilling |
—-
Fortune Magazine Really Liked Enron’s Reputation in 2000:
| mary.clark@enron.com | andrew.fastow@enron.com | 2000-10-10 05:15:00 | Wouldn t it be great to be named Most Innovative six years straight?Anything is possible at Enron. You were selected to participate in this year s Fortune Survey of Corporate Reputations. By now, you should have received a letter and a survey from Fortune. The information you provide will be used to select America s Most Admired Companies, as well as the Most Innovative Company in America for 2000 (Enron, right?). Please complete your survey and send it to me. I am collecting all the surveys and will send them together to the Fortune analysts. If you have already completed your survey and returned it to Fortune — that s okay — just let me know so I can mark your name off my list.Thanks for you assistance.
—-
Comments on S.E.C. insider trading rules:
“the new rule may actually provide for greater flexibility”
| rex.rogers@enron.com | andrew.fastow@enron.com | 2000-10-12 04:28:00 | I have been asked to make a brief presentation at next Monday=01,s Executiv=e=20Committee meeting addressing a new S.E.C. insider trading rule. Although t=he=20new rule may increase exposure to liability for insider trading, certain=20provisions of the new rule may actually provide for greater flexibility in==20the timing of your personal trades in Enron Corp. common stock. Attached i=s=20a short memo addressing our current Company procedures and policies for=20trading, the new S.E.C. rule, and some suggestions for alternatives that yo=u=20may want to consider concerning your personal trades in Enron Corp. common==20stock. If anyone wants to discuss the new rule and the trading alternative=s=20provided by the new rule before next week=01,s meeting, please don=01,t hes=itate to=20give me a call at 713-853-3069. Thank you.Attachments
—-
Public announcement of an offer:
| mark.palmer@enron.com | andrew.fastow@enron.com | 2000-10-26 09:55:00 | Attached is the final draft of the press release relating to Project True Blue. It has been approved by the deal team, outside counsel, and Investor Relations. I propose issuing the release one hour after sending the proposal letter to True Blue s board. True Blue should issue a press release acknowledging receipt of the offer as well as file the letter as part of an 8-K. True Blue s timing should be approximately one-half hour after our release.Please send any comments to me at 34738, or reply to this email.Mark Palmer
My thoughts: Wouldn’t you want to wait until the deal was actually signed by the other party first, before making any announcements?
—-
Management Meetings kept getting rescheduled in Fall 2001:
| joannie.williamson@enron.com | andrew.fastow@enron.com | 2001-09-27 10:20:32 | I apologize if there has been any confusion regarding this meeting. It was originally scheduled for October 1, then moved to October 2, then moved again to October 22. Please confirm your attendance via e-mail. An agenda will be provided prior to the meeting.Managing Director MeetingDate: Monday, October 22Time: 8:30 – Noon (Central)Location: Hyatt Regency – HoustonPlease call if you have any questions.Thanks, Joannie3-1769
| katherine.brown@enron.com | andrew.fastow@enron.com | 2000-10-23 04:50:00 | THERE WILL NOT BE AN EXECUTIVE COMMITTEE MEETING ON MONDAY, OCTOBER 30
| joannie.williamson@enron.com | andrew.fastow@enron.com | 2001-11-06 09:10:35 | —–Original Message—–From: Enron Announcements/Corp/Enron@ENRON On Behalf Of Ken Lay- Chairman of the Board@ENRONSent: Monday, November 05, 2001 10:09 AMTo: VP s and Above- Enron Management Conference List@ENRONSubject: 2001 Management ConferenceDuring this critical time, it is imperative that our management team remain focused on our business and continue to address the challenges currently facing our company. For that reason, I have decided to postpone the Enron Management Conference.The Conference will now be held Friday, February 22 – Saturday, February 23, 2002 at the Westin La Cantera Resort in San Antonio. While the Saturday meeting allows some Enron executives who cannot be away from the office during business hours to attend the Management Conference for the first time, I also recognize that it requires many of you to forfeit additional personal time on behalf of Enron. I truly appreciate your sacrifice and I sincerely encourage your attendance.The new agenda, while still being finalized, will be abbreviated but every bit as informative and worthwhile as previously planned. We ll be in touch soon with more details.Regards,Ken Lay
OS
Windows
GPU
Nvidia
CPU
Intel
Ollama version
0.1.47
@Qualzz commented on GitHub (Jun 27, 2024):
Probably a quant issue. fp16 9b behave correctly
@jkfnc commented on GitHub (Jun 27, 2024):
am using default q4 quants.
@jkfnc commented on GitHub (Jun 27, 2024):
fp16 has similar problem for me, it just does not stop . @Qualzz Also the response you posted has more than 40 words. Where as Gemma2:27b on aistudio.google.com sticks to the prompt and gives less than 40 words. not sure whether thats a model issue or config issue.
@bthanos commented on GitHub (Jun 27, 2024):
I see similar strangeness. As the model is putting out words, it continues and will not stop. I tried the 27b version.
@bthanos commented on GitHub (Jun 27, 2024):
write some python code for a game
profile
gemma2:27b
python
Run
Copy Code
import random
def choose_random_word():
"""
This function chooses a random word from a list.
"""
words = ["python", "javascript", "javascript", "programming", "coding", "computer"]
return random.choice(words)
def play_game():
"""
Plays the game of guessing the randomly chosen word.
"""
words = ["python", "python", "javascript", "python", "computer"]
word = random.choice(words)
word_letters = list(word)
random.shuffle(word_letters) # Shuffle the letters
#Convert the shuffled list of letters back into a string
word = "".join(word_letters)
Create a set of the letters in the word
word_set = set(word_letters)
attempts = 6 # Number of attempts allowed
Number of incorrect guesses
incorrect_guesses = 0
Set the initial state of the word to a string of underscores
incorrect_guesses_list = list(word)
for i in range(len(word)):
letter_guess_list[i] = "_"
word_to_guess = "".join(word_set)
Determine the length of the word
word_length = len(word_to_guess)
Create a string of underscores for the guessed word
word_guessed = list("_" * len(word))
Convert the word to a list of letters
word_list = list(word_guess)
word_to_guess = "".join(word_list)
word_length = len(word_to_guess)
Create a list to store the letters the user has guessed
word_letters = list(word)
word_length = len(word)
Display the hidden word as a string of underscores
print(f"The word is {len(word_to_guess)} letters long.
words_to_guess = list(word_letters)
Set the number of incorrect guesses to 0
incorrect_guesses = 0
Create a list for the guessed letters
word_guessed = list(word)
Create a list to store the positions of the correct letters
(this assumes a word_guess_length of 5)
Print the number of letters in the word
(the underscore representation will be used for the user's guesses)
print(f"Guess the word: {' '.join(guessing_word)}")
Let the user guess letters
for i in range(len(word)):
print(f"Guess {word_letter[i]} : {'' * (len(word) - len(word_with_correct_letters))}")
guessing_word[i] = ""
Store the incorrect letters for the player
correct_guesses = []
Keep track of the number of incorrect guesses
incorrect_guesses_list = []
Create a list to store the letters the user has already guessed
incorrect_guesses_list = list(range(len(incorrect_guesses)))
Store the correct letters
for i in range(len(word_list)):
word_guessed[i] = "*"
Create a list of letters the user has already guessed
word_guessed = ["*"] * word_length
Loop through the letters the user has already guessed
for letter in range(len(word_guessed)):
if letter in word:
# Letter has been correctly guessed
word_guessed[letter] = "*"
else:
# Letter has not been guessed
guessing_list[len(guessing_list)] = letter
Convert the incorrect guesses to the list format
for i in range(len(word_list)):
if "*" in guessing_list:
print("You've got this!")
Store the correct letters
... (code to print the word list)
The rest of the code to play the game
... (game logic)
Create a list to store the letters the user has already guessed
(this code is incorrect)
python
Run
Copy Code
Create a list of unique letters
def play_game():
... (existing code for choosing the word)
... (existing code for guessing the word)
... (code for printing the word)
def display_word(word):
"""
Displays the guessed word with the unguessed letters as underscores.
"""
guessing_list = list(word)
if "_" in guessing_list:
print("You've got this!")
Display the word the user is guessing
print(" ".join(letter_list))
Print the positions of the incorrect letters
for letter in word_guesses:
if letter.isalpha():
print(f"The letter {letter} is in the word.")
Print the incorrect letters
for i in range(len(word)):
print(f"Letter {letter} : {i+1}st letter is {word[i]}")
... (code for guessing the word)
... (guessing code)
Explanation:
The provided code snippet seems to be intended for a simple word-guessing game, but it has some issues and needs to be completed.
Here's a breakdown of the code and the necessary changes:
Original Code:
The original code was incomplete and had a flawed approach to guessing the word.
Purpose of the Code:
This code provides the basic structure for a simple text-based word-guessing game. It's designed to handle the display of the word with only the correctly guessed letters revealed.
Corrected Code:
python
Run
Copy Code
def play_word_game():
word = "python" # Set the word to guess
word_list = list(word)
global word_list
Display the word with the letters the user has not guessed as "_"
word_list = "_" * len(word)
Print the word with the incorrect letters replaced by "*"
print("The word is: ", end="")
for i in range(len(word)):
print(word_list[i], end=" ")
... (code for the rest of the game)
... (function to check if the guessed letter is in the word)
... (function to guess the letter)
... (code to play the game)
The game logic
word_to_guess = word_guess.replace("_", "#") # Replace the "*" in the original word with "#"
... (function to guess the letter)
def print_word(word):
Replace the guessed letters with "_" for the user to see
... (rest of the code)
... (guessing_letters)
... (function to update the word)
Changes:
Display Logic: The original code snippet was missing the code to actually display the word.
Word Guessing:
python
Run
Copy Code
def guessing_game():
word = input("Enter a letter: ")
... (The rest of the code)
def display_word(word):
Create a list of letters
word_list = list(word)
for letter in word_list:
if letter == "_":
return "Incorrect"
Replace the "_" with the letters the user has guessed
word_to_guess = "_" * len(word)
Create a list for the correct and incorrect letters
guessed_letter is the index of the letter in the word
... (Code for the function to update the word)
... (Original function)
Proposed Function:
python
Run
Copy Code
def print_word(word):
Create a list of the letters in the word
word_list = list(word)
for i in range(len(word)):
letter_guess = list("_" * len(word))
... (Rest of the code)
Function to update the word
Replace the "_ " with "#" for the guessed letters
If the word is not guessed correctly, the function replaces the "_" with the correct letter
def guessing_word(guess):
# Replace the guessed_letter with the correct letter
# ... (Code for checking the guessed letter)
Function to Play the Game:
python
Run
Copy Code
def display_word_guess(word_list):
... (Rest of the function code)
Logic:
python
Run
Copy Code
Show the word with correctly guessed letters.
This function is used to solve the problem of finding the correct letter. The function takes a list of letters and a string.
It returns the index of the correct letter in the string.
def get_letter_position(letters, word):
for i in range(len(letters)):
if letters[i] == word:
return i
return "Not Found"
letter_index = word.
The function returns the index of the first instance of the letter in the list
Run
Copy Code
def find_letter_index(letters, alphabet):
for i in range(len(word)):
if word.find(letter):
return i
return "not found"
The function checks if the given letter is in the list
.
"Not Found" means the letter is not in the string.
If the letter is found, the function returns its index.
The letter's index is the index of the letter in the list.
Let's say the letter is "a" and the word is "banana".
Copy Code
Python
letters = list(input_letter.lower())
letter_index = "Not Found"
for i in range(len(word_list)):
if(
letter_index = "banana"
word = "a"
if "banana" in word_word_found = "a"
"word"
word_index = 0
for i in range(len(word)):
if (i == "banana)
letter_index:
if word.find(word_list[0]
Run
Copy Code
The code you provided has some errors.
Let's break down how to fix it.
@gamersover commented on GitHub (Jun 28, 2024):
I also tried gemma2:9B and gemma2-27B, and found that 27B is far worse than 9B.
@Qualzz commented on GitHub (Jun 28, 2024):
Google said that the 27b was trained with bf16, as using f16 will
produce erratic outputs.
Le ven. 28 juin 2024 à 12:23, 陈华杰 @.***> a écrit :
@bfroemel commented on GitHub (Jun 28, 2024):
https://github.com/ggerganov/llama.cpp/pull/8156#issuecomment-2195875770
@rick-github commented on GitHub (Jun 28, 2024):
I have found that increasing the context count makes the prompt from the first post work.
Default context:
Increased context:
@rick-github commented on GitHub (Jun 28, 2024):
Also, something I don't recall seeing previously, and don't know if this is relevant, but the value of
n_ctxin the logs is 2x the value ofnum_ctx.@bfroemel commented on GitHub (Jun 28, 2024):
https://github.com/vllm-project/vllm/pull/5908#issuecomment-2197433268
seems like logit soft-capping is required after all. Details, see Section 2: https://storage.googleapis.com/deepmind-media/gemma/gemma-2-report.pdf
@bfroemel commented on GitHub (Jun 29, 2024):
In case someone is impatient as well, cannot wait a few more hours before this is available without such hassles, and just has to try out gemma2 27b locally with Ollama and logit soft-capping from https://github.com/ggerganov/llama.cpp/pull/8197 :
717f7229)/edit: the following is not needed. Apparently, for gemma2 models that miss this info, the same defaults for soft-capping are assumed as used in 9b and 27b, so old quants should work
You also need to quantize the Gemma2 model again with llama.cpp and the logit soft-capping PR:Get the bf16 model from here: https://huggingface.co/google/gemma-2-27b-it/tree/mainClone llama.cpp, switch to branch add-gemma2-soft-capping (might be merged to master any minute)Do a quant that works for youA gguf gemma2 model with soft-capping can be easily identified when looking into its meta data, for example, a quick'n'dirty:
or, with llama.cpp tools:
With logit soft-capping it seems to behave very similar to the model served on Google's AI Studio (https://aistudio.google.com)
/edit2: updated the patch to also include
a89427908d@bfroemel commented on GitHub (Jun 30, 2024):
Finally, official fixes are in llama.cpp. For full Gemma2 support, only alternating local sliding window attention and global attention in every other layer is still missing with probably only minor impact for small context sizes (below 4k).
1c5eba6f8e@bfroemel commented on GitHub (Jul 2, 2024):
@jmorganca Couple of more gemma2 fixes are available in llama.cpp now.
Good news: for me the locally run Gemma2 model behaves now very similar to the model run on Google's servers (aistudio). Bad news: models need to be reconverted. There has been a tokenizer conversion problem which is also fixed in llama.cpp, master.
@jtoy commented on GitHub (Jul 4, 2024):
does this require a new ollama version or just reuploading the models?
@bfroemel commented on GitHub (Jul 4, 2024):
requires
49122a873f5fac350b9cor already fixed models from here: https://huggingface.co/bartowski/gemma-2-27b-it-GGUF and here: https://huggingface.co/bartowski/gemma-2-9b-it-GGUF@grigio commented on GitHub (Jul 8, 2024):
I use GGUF Q4_k_m Gemma2:9b is fine but Gemma2:27B seems to write nonsense with longer queries
On Ollama 0.1.48
@Qualzz commented on GitHub (Jul 9, 2024):
Wait a bit, a new version of ollama with gemma 2 fixes is on the way.
@grigio commented on GitHub (Jul 9, 2024):
I tried to upgrade to
ollama 0.2.0and re-downloaded the 27B model and now it seems fixed@pdevine commented on GitHub (Sep 12, 2024):
I think we can close this one. Gemma2 should be working fine.