[GH-ISSUE #2475] Request to add leo-hessianai to ollama #27208

Open
opened 2026-04-22 04:17:58 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @arsenij-ust on GitHub (Feb 13, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/2475

Hi guys,
I tried to use the leo-hessianai-7B model on Ollama. I use the GGUF file (Q4_K_M.gguf from here https://huggingface.co/TheBloke/leo-hessianai-7B-GGUF/tree/main) and followed the instructions from Ollama (https://github.com/ollama/ollama/blob/main/docs/import.md). I already managed to generate answers with the model, but they are extremely wrong and hallucinating (you can say crazy). Unfortunately, I don't know what I'm doing wrong. I assume that the parameters or the template (in the Modelfile you have to create for Ollama) are incorrect.
Hope you can help me out 🙂

I tried the following Modelfiles:

FROM ./leo-hessianai-7b.Q4_K_M.gguf
TEMPLATE """{{- if .System }}
<|im_start|>system {{ .System }}<|im_end|>
{{- end }}
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assistant
"""

SYSTEM """"""

PARAMETER stop <|im_start|>
PARAMETER stop <|im_end|>
FROM ./leo-hessianai-7b.Q4_K_M.gguf
TEMPLATE "[INST] {{ .Prompt }} [/INST]"

(The same problem occurred when I used the safetensors from this repo and used the ollama tools to convert and quantize the model.)

--> So it would be great if leo-hessianai-7B, leo-hessianai-13B, and leo-hessianai-70B could be added to ollama - find the models at https://huggingface.co/LeoLM

Originally created by @arsenij-ust on GitHub (Feb 13, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/2475 Hi guys, I tried to use the leo-hessianai-7B model on Ollama. I use the GGUF file (Q4_K_M.gguf from here https://huggingface.co/TheBloke/leo-hessianai-7B-GGUF/tree/main) and followed the instructions from Ollama (https://github.com/ollama/ollama/blob/main/docs/import.md). I already managed to generate answers with the model, but they are extremely wrong and hallucinating (you can say crazy). Unfortunately, I don't know what I'm doing wrong. I assume that the parameters or the template (in the Modelfile you have to create for Ollama) are incorrect. Hope you can help me out 🙂 I tried the following Modelfiles: ``` FROM ./leo-hessianai-7b.Q4_K_M.gguf TEMPLATE """{{- if .System }} <|im_start|>system {{ .System }}<|im_end|> {{- end }} <|im_start|>user {{ .Prompt }}<|im_end|> <|im_start|>assistant """ SYSTEM """""" PARAMETER stop <|im_start|> PARAMETER stop <|im_end|> ``` ``` FROM ./leo-hessianai-7b.Q4_K_M.gguf TEMPLATE "[INST] {{ .Prompt }} [/INST]" ``` (The same problem occurred when I used the safetensors from this repo and used the ollama tools to convert and quantize the model.) --> So it would be great if `leo-hessianai-7B`, `leo-hessianai-13B`, and `leo-hessianai-70B` could be added to ollama - find the models at https://huggingface.co/LeoLM
GiteaMirror added the help wanted label 2026-04-22 04:17:58 -05:00
Author
Owner

@n-bluefield commented on GitHub (Mar 8, 2024):

I ran into the same problem using https://huggingface.co/TheBloke/leo-hessianai-13B-chat-bilingual-GGUF in Ollama.
On HuggingFace the template is given as a ChatML prompt template, as you showed above. I tried different ways of formatting the modelfile, always with hallucinating results. Seems to me it's revealing training data.

Did you find a solution? Does anyone else know how to make it work? Help would be much appreciated!

<!-- gh-comment-id:1986005759 --> @n-bluefield commented on GitHub (Mar 8, 2024): I ran into the same problem using https://huggingface.co/TheBloke/leo-hessianai-13B-chat-bilingual-GGUF in Ollama. On HuggingFace the template is given as a ChatML prompt template, as you showed above. I tried different ways of formatting the modelfile, always with hallucinating results. Seems to me it's revealing training data. Did you find a solution? Does anyone else know how to make it work? Help would be much appreciated!
Author
Owner

@arsenij-ust commented on GitHub (Mar 11, 2024):

@n-bluefield Hi, no, unfortunately I didn't find a solution. My Discord request, as well as the same question on https://huggingface.co/LeoLM/leo-hessianai-7b-chat/discussions/7 was not answered yet.

<!-- gh-comment-id:1988346423 --> @arsenij-ust commented on GitHub (Mar 11, 2024): @n-bluefield Hi, no, unfortunately I didn't find a solution. My Discord request, as well as the same question on https://huggingface.co/LeoLM/leo-hessianai-7b-chat/discussions/7 was not answered yet.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#27208