[GH-ISSUE #140] Start/stop tokens seem to bug out sometimes in long winded sessions #62087

Closed
opened 2026-05-03 07:28:44 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @nathanleclaire on GitHub (Jul 20, 2023).
Original GitHub issue: https://github.com/ollama/ollama/issues/140

stuff like:

>>> ... user prompt ...
...some response here...
<<SYS>>

You are an expert at summarizing text documents step by step and preserving
information. Between each of our interactions, summarize my message in a bullet
point summary, including all previously summarized information.

<</SYS>>

>>> ...

I've also seen it have conversations back and forth with itself lol but that might have been my fault due to mucking with instruction formats

Originally created by @nathanleclaire on GitHub (Jul 20, 2023). Original GitHub issue: https://github.com/ollama/ollama/issues/140 stuff like: ``` >>> ... user prompt ... ...some response here... <<SYS>> You are an expert at summarizing text documents step by step and preserving information. Between each of our interactions, summarize my message in a bullet point summary, including all previously summarized information. <</SYS>> >>> ... ``` I've also seen it have conversations back and forth with itself lol but that might have been my fault due to mucking with instruction formats
GiteaMirror added the bug label 2026-05-03 07:28:44 -05:00
Author
Owner

@adamjmurray commented on GitHub (Jul 26, 2023):

I installed Ollama and pulled llama2 for the first time tonight, and a few interactions into a mundane conversation, it printed this after an otherwise normal response:

<<SYS>>

You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.

If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information.

<</SYS>>

I wonder if it is a problem with Ollama or the underlying LLM?

<!-- gh-comment-id:1651019569 --> @adamjmurray commented on GitHub (Jul 26, 2023): I installed Ollama and pulled llama2 for the first time tonight, and a few interactions into a mundane conversation, it printed this after an otherwise normal response: ``` <<SYS>> You are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature. If a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information. <</SYS>> ``` I wonder if it is a problem with Ollama or the underlying LLM?
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#62087