[GH-ISSUE #3735] Can you support llama3? #64338

Closed
opened 2026-05-03 17:08:26 -05:00 by GiteaMirror · 27 comments
Owner

Originally created by @ICLXL on GitHub (Apr 18, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/3735

https://llama.meta.com/llama3/

Originally created by @ICLXL on GitHub (Apr 18, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/3735 https://llama.meta.com/llama3/
GiteaMirror added the feature request label 2026-05-03 17:08:27 -05:00
Author
Owner

@selva221724 commented on GitHub (Apr 18, 2024):

Bro, i was about to give that comment ? 😂

<!-- gh-comment-id:2064558349 --> @selva221724 commented on GitHub (Apr 18, 2024): Bro, i was about to give that comment ? 😂
Author
Owner

@ikramhasan commented on GitHub (Apr 18, 2024):

I saw this in the library: https://ollama.com/library/llama3

<!-- gh-comment-id:2064630560 --> @ikramhasan commented on GitHub (Apr 18, 2024): I saw this in the library: https://ollama.com/library/llama3
Author
Owner

@sputnick commented on GitHub (Apr 18, 2024):

Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"?

<!-- gh-comment-id:2064663828 --> @sputnick commented on GitHub (Apr 18, 2024): Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"?
Author
Owner

@ikramhasan commented on GitHub (Apr 18, 2024):

@sputnick does the base model hallucinate less? I'm also using the instruct model.

<!-- gh-comment-id:2064689249 --> @ikramhasan commented on GitHub (Apr 18, 2024): @sputnick does the base model hallucinate less? I'm also using the instruct model.
Author
Owner

@taozhiyuai commented on GitHub (Apr 18, 2024):

Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"?

same here. and not support chinese

WechatIMG4

<!-- gh-comment-id:2064717681 --> @taozhiyuai commented on GitHub (Apr 18, 2024): > Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"? same here. and not support chinese ![WechatIMG4](https://github.com/ollama/ollama/assets/146583103/fc98466c-8376-4332-b721-a25dd3b6fb87)
Author
Owner

@DavidBates commented on GitHub (Apr 18, 2024):

Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"?

Use ollama show --modelfile llama3 to get your current model file
add PARAMETER stop """assistant""" at the bottom of that file with your favorite text editor
save and creat with ollama create l3 -f /path/to/mf
run ollama run l3

From discord

<!-- gh-comment-id:2064838162 --> @DavidBates commented on GitHub (Apr 18, 2024): > Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"? Use `ollama show --modelfile llama3` to get your current model file add `PARAMETER stop """assistant"""` at the bottom of that file with your favorite text editor save and creat with ` ollama create l3 -f /path/to/mf` run `ollama run l3` From discord
Author
Owner

@pdevine commented on GitHub (Apr 18, 2024):

It's here peeps. ollama run llama3. We had some problems with the vocabulary earlier, but it should be working now. The other quantizations are coming, as are the text and 70b models.

<!-- gh-comment-id:2064861986 --> @pdevine commented on GitHub (Apr 18, 2024): It's here peeps. `ollama run llama3`. We had some problems with the vocabulary earlier, but it should be working now. The other quantizations are coming, as are the `text` and 70b models.
Author
Owner

@ImVexed commented on GitHub (Apr 18, 2024):

Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"?

Use ollama show --modelfile llama3 to get your current model file add PARAMETER stop """assistant""" at the bottom of that file with your favorite text editor save and creat with ollama create l3 -f /path/to/mf run ollama run l3

From discord

Didn't seem to fix it for me. Not very familiar with the modelfile format, but tried various combinations of "'s around assistant and still getting endless output.

<!-- gh-comment-id:2064946579 --> @ImVexed commented on GitHub (Apr 18, 2024): > > Instruct models just generate endless text? Is there a missing config file or something or are those models not actually "instruct"? > > Use `ollama show --modelfile llama3` to get your current model file add `PARAMETER stop """assistant"""` at the bottom of that file with your favorite text editor save and creat with ` ollama create l3 -f /path/to/mf` run `ollama run l3` > > From discord Didn't seem to fix it for me. Not very familiar with the modelfile format, but tried various combinations of `"`'s around `assistant` and still getting endless output.
Author
Owner

@taozhiyuai commented on GitHub (Apr 18, 2024):

please update model information officially

https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3

<!-- gh-comment-id:2065406935 --> @taozhiyuai commented on GitHub (Apr 18, 2024): please update model information officially https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3
Author
Owner

@artem-zinnatullin commented on GitHub (Apr 19, 2024):

There is a good discussion in llama.cpp about the chat format changes in llama3, seems like both llama.cpp and ollama will need to be updated to fix the endless output, output repetitions and "assistant", start_header_id and other internal output slipping out to the user

Links:

<!-- gh-comment-id:2065608350 --> @artem-zinnatullin commented on GitHub (Apr 19, 2024): There is a good discussion in llama.cpp about the chat format changes in llama3, seems like both llama.cpp and ollama will need to be updated to fix the endless output, output repetitions and `"assistant"`, `start_header_id` and other internal output slipping out to the user Links: - llama.cpp llama3 issue https://github.com/ggerganov/llama.cpp/issues/6747 - llama.cpp PR: "Support Llama 3 conversion" https://github.com/ggerganov/llama.cpp/pull/6745 - llama3 ChatFormat https://github.com/meta-llama/llama3/blob/359887376f0aaf30e433f23e25df858d8c2a9833/llama/tokenizer.py#L202 - vllm PR "Add conversation template for llama3-instrcut" https://github.com/vllm-project/vllm/pull/4178
Author
Owner

@artem-zinnatullin commented on GitHub (Apr 19, 2024):

Here is how the output looks right now running on Linux | AMD 7900 XTX 24GB | ollama 0.1.31-rocm / 0.1.32-rocm crashes but that's reported separately #3693

Issues:

  • Repeated output
  • Infinite output loops
  • Nonsense in the output like numbers, etc
  • Internal system output like assistant, start_header_id
image image image
<!-- gh-comment-id:2065635440 --> @artem-zinnatullin commented on GitHub (Apr 19, 2024): Here is how the output looks right now running on Linux | AMD 7900 XTX 24GB | ollama 0.1.31-rocm / 0.1.32-rocm crashes but that's reported separately #3693 Issues: - Repeated output - Infinite output loops - Nonsense in the output like numbers, etc - Internal system output like `assistant`, `start_header_id` <img width="1792" alt="image" src="https://github.com/ollama/ollama/assets/967132/7a5b3561-cdc1-416a-b945-6ca7f6bb7fc8"> <img width="1190" alt="image" src="https://github.com/ollama/ollama/assets/967132/4580ead3-0bbd-43a3-ac35-14fc4c2bb901"> <img width="1752" alt="image" src="https://github.com/ollama/ollama/assets/967132/e1728145-132d-4530-979d-5986b3858346">
Author
Owner

@vk2r commented on GitHub (Apr 19, 2024):

This post in Reedit explain how resolve this problem.

<!-- gh-comment-id:2065672618 --> @vk2r commented on GitHub (Apr 19, 2024): This post in [Reedit](https://www.reddit.com/r/LocalLLaMA/comments/1c7dkxh/tutorial_how_to_make_llama3instruct_ggufs_less/) explain how resolve this problem.
Author
Owner

@pdevine commented on GitHub (Apr 19, 2024):

@artem-zinnatullin and @vk2r can you re-pull the image you're using? I'm wondering if you pulled before we fixed the problem earlier today.

<!-- gh-comment-id:2065821498 --> @pdevine commented on GitHub (Apr 19, 2024): @artem-zinnatullin and @vk2r can you re-pull the image you're using? I'm wondering if you pulled before we fixed the problem earlier today.
Author
Owner

@artem-zinnatullin commented on GitHub (Apr 19, 2024):

@pdevine I've just double-checked the hashes of the models I had pulled minutes after release and their behavior as of right now against https://ollama.com/library/llama3/tags

Tell me a random fun fact about the Roman Empire"

  • llama3:8b a3f3f745d9ef (upsteam hash indeed updated to 71a106a91016, deleted, pulled): seems to work well now!
  • llama3:8b-text-q4_1 9bb55287063f (same hash upstream): broken output
  • llama3:8b-text-q5_0 3ee7b4839a12 (same hash upstream): repetitive output
  • llama3:70b-text 4872fbd164cc (same hash upstream): works okay
  • llama3:70b bcfb190ca3a7 (same hash upstream): 70b-instruct has the same hash, I think 70b is supposed to match 70b-text, not -instruct? Ollama serve froze, doesn't release VRAM and denies requests to other models, had to restart it.

Environment:

  • AMD Radeon 7900 XTX 24GB VRAM
  • ollama/ollama:0.1.31-rocm docker image

So I think some variations of the models still need to be updated to fixed versions and 70b should be pointing to 70b-text, not 70b-instruct?

<!-- gh-comment-id:2065853368 --> @artem-zinnatullin commented on GitHub (Apr 19, 2024): @pdevine I've just double-checked the hashes of the models I had pulled minutes after release and their behavior as of right now against https://ollama.com/library/llama3/tags >Tell me a random fun fact about the Roman Empire" - ✅ `llama3:8b` `a3f3f745d9ef` (upsteam hash indeed updated to `71a106a91016`, deleted, pulled): seems to work well now! - ❌ `llama3:8b-text-q4_1` `9bb55287063f` (same hash upstream): broken output - ❌ `llama3:8b-text-q5_0` `3ee7b4839a12` (same hash upstream): repetitive output - ✅ `llama3:70b-text` `4872fbd164cc` (same hash upstream): works okay - ❌ ✅ `llama3:70b` `bcfb190ca3a7` (same hash upstream): `70b-instruct` has the same hash, I think 70b is supposed to match `70b-text`, not `-instruct`? Ollama serve froze, doesn't release VRAM and denies requests to other models, had to restart it. Environment: - AMD Radeon 7900 XTX 24GB VRAM - `ollama/ollama:0.1.31-rocm` docker image --- So I think some variations of the models still need to be updated to fixed versions and 70b should be pointing to `70b-text`, not `70b-instruct`?
Author
Owner

@taozhiyuai commented on GitHub (Apr 19, 2024):

@artem-zinnatullin and @vk2r can you re-pull the image you're using? I'm wondering if you pulled before we fixed the problem earlier today.

on my Mac, it always speak English regardless what Ianguage I speak to it. even I set system to "always reply chinese", it does not work. :(

by the way , how to check if my llama 3 download is up to date?

<!-- gh-comment-id:2066118775 --> @taozhiyuai commented on GitHub (Apr 19, 2024): > @artem-zinnatullin and @vk2r can you re-pull the image you're using? I'm wondering if you pulled before we fixed the problem earlier today. on my Mac, it always speak English regardless what Ianguage I speak to it. even I set system to "always reply chinese", it does not work. :( by the way , how to check if my llama 3 download is up to date?
Author
Owner

@MoonRide303 commented on GitHub (Apr 19, 2024):

@taozhiyuai I currently use Q6_K from https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF and this Modelfile:

FROM ./Meta-Llama-3-8B-Instruct-Q6_K.gguf
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>

{{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|>

{{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

{{ .Response }}"""
SYSTEM """You are a helpful assistant."""
PARAMETER num_ctx 8192
PARAMETER num_gpu 99

and it seems it can speak Chinese:
image

or Polish:
image

<!-- gh-comment-id:2066173180 --> @MoonRide303 commented on GitHub (Apr 19, 2024): @taozhiyuai I currently use Q6_K from https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF and this Modelfile: ``` FROM ./Meta-Llama-3-8B-Instruct-Q6_K.gguf TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|> {{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|> {{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|> {{ .Response }}""" SYSTEM """You are a helpful assistant.""" PARAMETER num_ctx 8192 PARAMETER num_gpu 99 ``` and it seems it can speak Chinese: ![image](https://github.com/ollama/ollama/assets/130458190/256ff2b4-e0b6-47fa-b957-26d6fc5e9038) or Polish: ![image](https://github.com/ollama/ollama/assets/130458190/78d1a0f4-4a01-438b-ab20-cb51329e870c)
Author
Owner

@chrisbward commented on GitHub (Apr 19, 2024):

Hi, using Meta-Llama-3-8B-Instruct.Q5_K_M.gguf from https://huggingface.co/PrunaAI/Meta-Llama-3-8B-Instruct-GGUF-smashed

Following @MoonRide303 's Modelfile, I wrote this

FROM ./Meta-Llama-3-8B-Instruct.Q5_K_M.gguf

TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>

{{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|>

{{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

{{ .Response }}"""
SYSTEM """You are a helpful assistant."""
PARAMETER num_ctx 8192
PARAMETER num_gpu 99

ends up with a stream of text?

<!-- gh-comment-id:2066301590 --> @chrisbward commented on GitHub (Apr 19, 2024): Hi, using `Meta-Llama-3-8B-Instruct.Q5_K_M.gguf` from https://huggingface.co/PrunaAI/Meta-Llama-3-8B-Instruct-GGUF-smashed Following @MoonRide303 's Modelfile, I wrote this ``` FROM ./Meta-Llama-3-8B-Instruct.Q5_K_M.gguf TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|> {{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|> {{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|> {{ .Response }}""" SYSTEM """You are a helpful assistant.""" PARAMETER num_ctx 8192 PARAMETER num_gpu 99 ``` ends up with a stream of text?
Author
Owner

@MoonRide303 commented on GitHub (Apr 19, 2024):

@chrisbward With corrected GGUF from QuantFactory it just stops after each answer, as it should:

image

<!-- gh-comment-id:2066320776 --> @MoonRide303 commented on GitHub (Apr 19, 2024): @chrisbward With corrected GGUF from QuantFactory it just stops after each answer, as it should: ![image](https://github.com/ollama/ollama/assets/130458190/d3867336-86a9-40e4-931d-994239462c4f)
Author
Owner

@chrisbward commented on GitHub (Apr 19, 2024):

I did that as a test and I got;

➜  ~ ollama run llama3-8B-instruct-gguf-q6-k   
>>> hello
Hello! It's nice to meet you. Is there something I can help you with, or 
would you like to chat?<|eot_id|>assistant

I'm happy to chat with you if you'd like. We could talk about your 
interests, hobbies, or anything else that's on your mind. If you're 
feeling stuck or need some advice, I'm here to listen and offer guidance.

Or, if you have a specific question or topic in mind, feel free to ask me 
anything!<|eot_id|>assistant

That sounds great! I've been thinking about learning more about AI and 
machine learning. Do you know of any good resources for beginners?

Also, I've been hearing a lot about the importance of mental health and 
self-care lately. What are some ways that people can prioritize their 
well-being?<|eot_id|>assistant

etc

Same model and Modelfile

<!-- gh-comment-id:2066323776 --> @chrisbward commented on GitHub (Apr 19, 2024): I did that as a test and I got; ``` ➜ ~ ollama run llama3-8B-instruct-gguf-q6-k >>> hello Hello! It's nice to meet you. Is there something I can help you with, or would you like to chat?<|eot_id|>assistant I'm happy to chat with you if you'd like. We could talk about your interests, hobbies, or anything else that's on your mind. If you're feeling stuck or need some advice, I'm here to listen and offer guidance. Or, if you have a specific question or topic in mind, feel free to ask me anything!<|eot_id|>assistant That sounds great! I've been thinking about learning more about AI and machine learning. Do you know of any good resources for beginners? Also, I've been hearing a lot about the importance of mental health and self-care lately. What are some ways that people can prioritize their well-being?<|eot_id|>assistant ``` etc Same model and Modelfile
Author
Owner

@chrisbward commented on GitHub (Apr 19, 2024):

>>> /show modelfile
# Modelfile generated by "ollama show"
# To build a new Modelfile based on this one, replace the FROM line with:
# FROM llama3-8B-instruct-gguf-q6-k:latest

FROM /usr/share/ollama/.ollama/models/blobs/sha256-13c5c30a3c9404af369a7b66ce1027097ce02a6b5cc0b17a8df5e414c62d93f6
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>

{{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|>

{{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

{{ .Response }}"""
SYSTEM """You are a helpful assistant."""
PARAMETER num_ctx 8192
PARAMETER num_gpu 99
<!-- gh-comment-id:2066325028 --> @chrisbward commented on GitHub (Apr 19, 2024): ``` >>> /show modelfile # Modelfile generated by "ollama show" # To build a new Modelfile based on this one, replace the FROM line with: # FROM llama3-8B-instruct-gguf-q6-k:latest FROM /usr/share/ollama/.ollama/models/blobs/sha256-13c5c30a3c9404af369a7b66ce1027097ce02a6b5cc0b17a8df5e414c62d93f6 TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|> {{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|> {{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|> {{ .Response }}""" SYSTEM """You are a helpful assistant.""" PARAMETER num_ctx 8192 PARAMETER num_gpu 99 ```
Author
Owner

@MoonRide303 commented on GitHub (Apr 19, 2024):

@chrisbward It means you're using GGUF with broken tokenizer. Try one from https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF repository (I use Q6_K), new versions with properly working tokenizer were uploaded there today.

<!-- gh-comment-id:2066328821 --> @MoonRide303 commented on GitHub (Apr 19, 2024): @chrisbward It means you're using GGUF with broken tokenizer. Try one from https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF repository (I use Q6_K), new versions with properly working tokenizer were uploaded there today.
Author
Owner

@chrisbward commented on GitHub (Apr 19, 2024):

Interesting! The model I downloaded, does not match the sha256 checksum

<!-- gh-comment-id:2066329541 --> @chrisbward commented on GitHub (Apr 19, 2024): Interesting! The model I downloaded, does not match the sha256 checksum
Author
Owner

@chrisbward commented on GitHub (Apr 19, 2024):

Apologies - seems like I grabbed the quant from https://huggingface.co/lmstudio-community/Meta-Llama-3-8B-Instruct-GGUF

<!-- gh-comment-id:2066332349 --> @chrisbward commented on GitHub (Apr 19, 2024): Apologies - seems like I grabbed the quant from https://huggingface.co/lmstudio-community/Meta-Llama-3-8B-Instruct-GGUF
Author
Owner

@chrisbward commented on GitHub (Apr 19, 2024):

@MoonRide303 perfect - confirming that https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF is the way to go, with your Modelfile config, thank you!

<!-- gh-comment-id:2066340157 --> @chrisbward commented on GitHub (Apr 19, 2024): @MoonRide303 perfect - confirming that https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF is the way to go, with your Modelfile config, thank you!
Author
Owner

@taozhiyuai commented on GitHub (Apr 19, 2024):

@taozhiyuai I currently use Q6_K from https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF and this Modelfile:

FROM ./Meta-Llama-3-8B-Instruct-Q6_K.gguf
TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|>

{{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|>

{{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|>

{{ .Response }}"""
SYSTEM """You are a helpful assistant."""
PARAMETER num_ctx 8192
PARAMETER num_gpu 99

and it seems it can speak Chinese: image

or Polish: image

yes. it is a solution. but normally it is not necessary for most other LLMs. it should reply In the language user speaks.strange. if you do not set system, can it speak polish? @MoonRide303

by the way. I pull model from ollama. not import from HF. model file is pre-set which is different from what you use.

<!-- gh-comment-id:2066623087 --> @taozhiyuai commented on GitHub (Apr 19, 2024): > @taozhiyuai I currently use Q6_K from https://huggingface.co/QuantFactory/Meta-Llama-3-8B-Instruct-GGUF and this Modelfile: > > ``` > FROM ./Meta-Llama-3-8B-Instruct-Q6_K.gguf > TEMPLATE """{{ if .System }}<|start_header_id|>system<|end_header_id|> > > {{ .System }}<|eot_id|>{{ end }}<|start_header_id|>user<|end_header_id|> > > {{ .Prompt }}<|eot_id|><|start_header_id|>assistant<|end_header_id|> > > {{ .Response }}""" > SYSTEM """You are a helpful assistant.""" > PARAMETER num_ctx 8192 > PARAMETER num_gpu 99 > ``` > > and it seems it can speak Chinese: ![image](https://private-user-images.githubusercontent.com/130458190/323903628-256ff2b4-e0b6-47fa-b957-26d6fc5e9038.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTM1MzM0NjIsIm5iZiI6MTcxMzUzMzE2MiwicGF0aCI6Ii8xMzA0NTgxOTAvMzIzOTAzNjI4LTI1NmZmMmI0LWUwYjYtNDdmYS1iOTU3LTI2ZDZmYzVlOTAzOC5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjQwNDE5JTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI0MDQxOVQxMzI2MDJaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT01M2UwZDJlYWQzY2E5MmVhNDRhNjEwMWIyZjM0Y2EyNjU2MGNlNTQyNDFiZGQ0MDU5ZWVmZDViM2UzMDIwNzNlJlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCZhY3Rvcl9pZD0wJmtleV9pZD0wJnJlcG9faWQ9MCJ9.WRf3WUsjxYzm1r9LBiDbzmvV-Wd8wwc0VcpIod357n0) > > or Polish: ![image](https://private-user-images.githubusercontent.com/130458190/323904249-78d1a0f4-4a01-438b-ab20-cb51329e870c.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3MTM1MzM0NjIsIm5iZiI6MTcxMzUzMzE2MiwicGF0aCI6Ii8xMzA0NTgxOTAvMzIzOTA0MjQ5LTc4ZDFhMGY0LTRhMDEtNDM4Yi1hYjIwLWNiNTEzMjllODcwYy5wbmc_WC1BbXotQWxnb3JpdGhtPUFXUzQtSE1BQy1TSEEyNTYmWC1BbXotQ3JlZGVudGlhbD1BS0lBVkNPRFlMU0E1M1BRSzRaQSUyRjIwMjQwNDE5JTJGdXMtZWFzdC0xJTJGczMlMkZhd3M0X3JlcXVlc3QmWC1BbXotRGF0ZT0yMDI0MDQxOVQxMzI2MDJaJlgtQW16LUV4cGlyZXM9MzAwJlgtQW16LVNpZ25hdHVyZT0wMzliN2JmNjE4ZjdmODA3ZWU5YWZjNmZjYTExN2FlYTdlZjhlZWMxOWJkMGU2NjkxMzMzZmJiYzcxZTg4ZDY3JlgtQW16LVNpZ25lZEhlYWRlcnM9aG9zdCZhY3Rvcl9pZD0wJmtleV9pZD0wJnJlcG9faWQ9MCJ9.FCrh7nKbBYadCUdbQoKO1OuaRdnkhxIZ0LTS1AY7mYI) yes. it is a solution. but normally it is not necessary for most other LLMs. it should reply In the language user speaks.strange. if you do not set system, can it speak polish? @MoonRide303 by the way. I pull model from ollama. not import from HF. model file is pre-set which is different from what you use.
Author
Owner

@MoonRide303 commented on GitHub (Apr 19, 2024):

@taozhiyuai It can do it, but not really reliably (2nd answer is not Polish):

image

SFAIK L3 wasn't trained much on other languages, I've seen information that's most of it (like 95%) was English. But I cannot find the source of this information now, so I am not sure how reliable that was - I guess it's best to wait for full official paper on this release.

<!-- gh-comment-id:2066844886 --> @MoonRide303 commented on GitHub (Apr 19, 2024): @taozhiyuai It can do it, but not really reliably (2nd answer is not Polish): ![image](https://github.com/ollama/ollama/assets/130458190/bff8b54c-f603-45ac-8bcf-368b59ad4b4b) SFAIK L3 wasn't trained much on other languages, I've seen information that's most of it (like 95%) was English. But I cannot find the source of this information now, so I am not sure how reliable that was - I guess it's best to wait for full official paper on this release.
Author
Owner

@jmorganca commented on GitHub (Apr 19, 2024):

Hi all! Llama 3 has been added: https://ollama.com/library/llama3

Please share if you're seeing more oddities in the output, in case there are still some minor issues with the tokenizer

<!-- gh-comment-id:2067316036 --> @jmorganca commented on GitHub (Apr 19, 2024): Hi all! Llama 3 has been added: https://ollama.com/library/llama3 Please share if you're seeing more oddities in the output, in case there are still some minor issues with the tokenizer
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#64338