[GH-ISSUE #4666] ollama doesn't create a model from modelfile and gives an error #64970

Closed
opened 2026-05-03 19:26:32 -05:00 by GiteaMirror · 7 comments
Owner

Originally created by @tMrMorgan on GitHub (May 27, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/4666

What is the issue?

Sorry in advance for any mistakes in text
when I trying to create a model in terminal, no matter what it based on, and even if the "modelfile" is a stock template of downloaded llm, after command "ollama create test" i got same output everytime
" Error: command must be one of "from", "license", "template", "system", "adapter", "parameter", or "message" "
i tried to change names of file, change command to e.g. "ollama create test -t modelfile" it's not working at all.
also i tried to edit commands in modelfile, so it doesn't get any reaction too. I've read the recent topic with similar error here, as it been suggested - converted modelfile in utf8, download file from topic - didn't helped. And also, i re-installed Ollama.

i tried to create modelfile thru Open WebUI from my models, and it works fine. And if i write in dialoge with llm in terminal: "/set system " you are something " and then " /save test " it saves and create new model without any trouble.

it is clear that the problem is not with modelfile itself, so what is it?

OS is WIN10pro. 16gb RAM, fx8320, gtx1650 with CUDA drivers installed.

OS

Windows

GPU

Nvidia

CPU

AMD

Ollama version

0.1.38

Originally created by @tMrMorgan on GitHub (May 27, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/4666 ### What is the issue? Sorry in advance for any mistakes in text when I trying to create a model in terminal, no matter what it based on, and even if the "modelfile" is a stock template of downloaded llm, after command "ollama create test" i got same output everytime " Error: command must be one of "from", "license", "template", "system", "adapter", "parameter", or "message" " i tried to change names of file, change command to e.g. "ollama create test -t modelfile" it's not working at all. also i tried to edit commands in modelfile, so it doesn't get any reaction too. I've read the recent topic with similar error here, as it been suggested - converted modelfile in utf8, download file from topic - didn't helped. And also, i re-installed Ollama. i tried to create modelfile thru Open WebUI from my models, and it works fine. And if i write in dialoge with llm in terminal: "/set system " you are something " and then " /save test " it saves and create new model without any trouble. it is clear that the problem is not with modelfile itself, so what is it? OS is WIN10pro. 16gb RAM, fx8320, gtx1650 with CUDA drivers installed. ### OS Windows ### GPU Nvidia ### CPU AMD ### Ollama version 0.1.38
GiteaMirror added the bugwindows labels 2026-05-03 19:26:33 -05:00
Author
Owner

@sealad886 commented on GitHub (May 27, 2024):

Please be specific in reporting this issue. What specific commands did you try, with which file(s) / version(s) (include the modelfile contents and filename/extension)? We can't help you unless you give us the details, please.

<!-- gh-comment-id:2134048985 --> @sealad886 commented on GitHub (May 27, 2024): Please be specific in reporting this issue. What specific commands did you try, with which file(s) / version(s) (include the modelfile contents and filename/extension)? We can't help you unless you give us the details, please.
Author
Owner

@tMrMorgan commented on GitHub (May 28, 2024):

Please be specific in reporting this issue. What specific commands did you try, with which file(s) / version(s) (include the modelfile contents and filename/extension)? We can't help you unless you give us the details, please.

sure, here is the modelfile commands. I got the modelfile by "ollama show dolphin-llama3 --modelfile"
then i created a copy by "ollama show dolphin-llama3 --modelfile > modelfile"
to create a model i used "ollama create modelfile" and "ollama create modelfile -f modelfile" and got the error.

I don't know exact file extension cause it marks just as a "file" in folder

(i just removed "#" from first 3 lines for this message cause it displayed weird here. i'm a noob, sorry!)

Modelfile generated by "ollama show"
To build a new Modelfile based on this, replace FROM with:
FROM dolphin-llama3:latest

FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580
TEMPLATE "{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>
"
SYSTEM "You are Dolphin, a helpful AI assistant.
"
PARAMETER stop <|im_start|>
PARAMETER stop <|im_end|>
LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT

then goes a huge license agreement, i didn't see a reason to copy it here, if you need i can just load modelfile here.
I tried to do all this on several models, like: llama3, phi3, dolphin-mistral. result is the same.
Also, i tred to paste "FROM dolphin-llama3:latest" instead path to model.

<!-- gh-comment-id:2135903662 --> @tMrMorgan commented on GitHub (May 28, 2024): > Please be specific in reporting this issue. What specific commands did you try, with which file(s) / version(s) (include the modelfile contents and filename/extension)? We can't help you unless you give us the details, please. sure, here is the modelfile commands. I got the modelfile by "ollama show dolphin-llama3 --modelfile" then i created a copy by "ollama show dolphin-llama3 --modelfile > modelfile" to create a model i used "ollama create modelfile" and "ollama create modelfile -f modelfile" and got the error. I don't know exact file extension cause it marks just as a "file" in folder (i just removed "#" from first 3 lines for this message cause it displayed weird here. i'm a noob, sorry!) Modelfile generated by "ollama show" To build a new Modelfile based on this, replace FROM with: FROM dolphin-llama3:latest FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580 TEMPLATE "{{ if .System }}<|im_start|>system {{ .System }}<|im_end|> {{ end }}{{ if .Prompt }}<|im_start|>user {{ .Prompt }}<|im_end|> {{ end }}<|im_start|>assistant {{ .Response }}<|im_end|> " SYSTEM "You are Dolphin, a helpful AI assistant. " PARAMETER stop <|im_start|> PARAMETER stop <|im_end|> LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT then goes a huge license agreement, i didn't see a reason to copy it here, if you need i can just load modelfile here. I tried to do all this on several models, like: llama3, phi3, dolphin-mistral. result is the same. Also, i tred to paste "FROM dolphin-llama3:latest" instead path to model.
Author
Owner

@sealad886 commented on GitHub (May 28, 2024):

Thanks for getting this back. Two things, off the bat:

  1. High value read in this document on Markdown so that both you and we can read your response easier.
  2. Yes, it's helpful to see verbatim what exactly your files look like, are named, and what commands you type in and what the output/error codes you get are. Please copy and paste it in.

Now, all of that said--let's have a look.
I'm not an expert in Windows machines myself, but my first suggestion to you would be to make sure you have triple-quotes (i.e. """) at the beginning and end of your multi-line strings.

So something like this:

# Modelfile generated by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
FROM dolphin-llama3:latest
# (^^I'd recommend using this format just because there are other blob files in the cache that get included when you reference it this way)

# FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580
TEMPLATE """{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>
"""
SYSTEM """You are Dolphin, a helpful AI assistant.
"""
PARAMETER stop <|im_start|>
PARAMETER stop <|im_end|>
LICENSE """META LLAMA 3 COMMUNITY LICENSE AGREEMENT

then goes a huge license agreement, i didn't see a reason to copy it here, if you need i can just load modelfile here.
I tried to do all this on several models, like: llama3, phi3, dolphin-mistral. result is the same.
Also, i tred to paste "FROM dolphin-llama3:latest" instead path to model.
# truncated legal stuff
"""

Technically, all that you need to include in the Modelfile is the path to the files that Ollama can understand. Everything else will just be defaulted from the model itself.

# Contents of Modelfile
FROM dolphin-llama3

is sufficient to create a new model in Ollama using the command:

ollama create -f /path/to/modelfile dolphin-llama3
ollama create -f /path/to/modelfile username/dolphin-llama3
ollama create -f /path/to/modelfile username/whacky-inflatable-balloon-man:q4_k

(and other permutations of that same command)

<!-- gh-comment-id:2136000703 --> @sealad886 commented on GitHub (May 28, 2024): Thanks for getting this back. Two things, off the bat: 1. High value read in [this document on Markdown](https://docs.github.com/en/get-started/writing-on-github/getting-started-with-writing-and-formatting-on-github/basic-writing-and-formatting-syntax) so that both you and we can read your response easier. 2. Yes, it's helpful to see *verbatim* what exactly your files look like, are named, and what commands you type in and what the output/error codes you get are. Please copy and paste it in. * [Collapsed sections](https://docs.github.com/en/get-started/writing-on-github/working-with-advanced-formatting/organizing-information-with-collapsed-sections) might be useful here Now, all of that said--let's have a look. I'm not an expert in Windows machines myself, but my first suggestion to you would be to make sure you have triple-quotes (i.e. \""") at the beginning and end of your multi-line strings. So something like this: ```script # Modelfile generated by "ollama show" # To build a new Modelfile based on this, replace FROM with: FROM dolphin-llama3:latest # (^^I'd recommend using this format just because there are other blob files in the cache that get included when you reference it this way) # FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580 TEMPLATE """{{ if .System }}<|im_start|>system {{ .System }}<|im_end|> {{ end }}{{ if .Prompt }}<|im_start|>user {{ .Prompt }}<|im_end|> {{ end }}<|im_start|>assistant {{ .Response }}<|im_end|> """ SYSTEM """You are Dolphin, a helpful AI assistant. """ PARAMETER stop <|im_start|> PARAMETER stop <|im_end|> LICENSE """META LLAMA 3 COMMUNITY LICENSE AGREEMENT then goes a huge license agreement, i didn't see a reason to copy it here, if you need i can just load modelfile here. I tried to do all this on several models, like: llama3, phi3, dolphin-mistral. result is the same. Also, i tred to paste "FROM dolphin-llama3:latest" instead path to model. # truncated legal stuff """ ``` Technically, all that you need to include in the Modelfile is the path to the files that Ollama can understand. Everything else will just be defaulted from the model itself. ```scripts # Contents of Modelfile FROM dolphin-llama3 ``` is sufficient to create a new model in Ollama using the command: ```shell ollama create -f /path/to/modelfile dolphin-llama3 ``` ```shell ollama create -f /path/to/modelfile username/dolphin-llama3 ``` ```shell ollama create -f /path/to/modelfile username/whacky-inflatable-balloon-man:q4_k ``` (and other permutations of that same command)
Author
Owner

@tMrMorgan commented on GitHub (May 28, 2024):

Thanks for getting this back. Two things, off the bat:

1. High value read in [this document on Markdown](https://docs.github.com/en/get-started/writing-on-github/getting-started-with-writing-and-formatting-on-github/basic-writing-and-formatting-syntax) so that both you and we can read your response easier.

2. Yes, it's helpful to see _verbatim_ what exactly your files look like, are named, and what commands you type in and what the output/error codes you get are. Please copy and paste it in.
   
   * [Collapsed sections](https://docs.github.com/en/get-started/writing-on-github/working-with-advanced-formatting/organizing-information-with-collapsed-sections) might be useful here

Now, all of that said--let's have a look. I'm not an expert in Windows machines myself, but my first suggestion to you would be to make sure you have triple-quotes (i.e. """) at the beginning and end of your multi-line strings.

So something like this:

# Modelfile generated by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
FROM dolphin-llama3:latest
# (^^I'd recommend using this format just because there are other blob files in the cache that get included when you reference it this way)

# FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580
TEMPLATE """{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>
"""
SYSTEM """You are Dolphin, a helpful AI assistant.
"""
PARAMETER stop <|im_start|>
PARAMETER stop <|im_end|>
LICENSE """META LLAMA 3 COMMUNITY LICENSE AGREEMENT

then goes a huge license agreement, i didn't see a reason to copy it here, if you need i can just load modelfile here.
I tried to do all this on several models, like: llama3, phi3, dolphin-mistral. result is the same.
Also, i tred to paste "FROM dolphin-llama3:latest" instead path to model.
# truncated legal stuff
"""

Technically, all that you need to include in the Modelfile is the path to the files that Ollama can understand. Everything else will just be defaulted from the model itself.

# Contents of Modelfile
FROM dolphin-llama3

is sufficient to create a new model in Ollama using the command:

ollama create -f /path/to/modelfile dolphin-llama3
ollama create -f /path/to/modelfile username/dolphin-llama3
ollama create -f /path/to/modelfile username/whacky-inflatable-balloon-man:q4_k

(and other permutations of that same command)

Well, i tried to fix midelfile, but it still gives me the error. But! i noticed, that when i create modelfile template, the first line of code is failed to get console mode for stdout: The handle is invalid.
without any marks before it.. and even if i remove that, error will appear anyway.

so unedited modelfile looks like this

failed to get console mode for stdout: The handle is invalid.
# Modelfile generated by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
# FROM dolphin-llama3:latest

FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580
TEMPLATE "{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>
"
SYSTEM "You are Dolphin, a helpful AI assistant.
"
PARAMETER stop <|im_start|>
PARAMETER stop <|im_end|>
LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT

By the way, thank you for links to tutorials!

<!-- gh-comment-id:2136076140 --> @tMrMorgan commented on GitHub (May 28, 2024): > Thanks for getting this back. Two things, off the bat: > > 1. High value read in [this document on Markdown](https://docs.github.com/en/get-started/writing-on-github/getting-started-with-writing-and-formatting-on-github/basic-writing-and-formatting-syntax) so that both you and we can read your response easier. > > 2. Yes, it's helpful to see _verbatim_ what exactly your files look like, are named, and what commands you type in and what the output/error codes you get are. Please copy and paste it in. > > * [Collapsed sections](https://docs.github.com/en/get-started/writing-on-github/working-with-advanced-formatting/organizing-information-with-collapsed-sections) might be useful here > > > Now, all of that said--let's have a look. I'm not an expert in Windows machines myself, but my first suggestion to you would be to make sure you have triple-quotes (i.e. """) at the beginning and end of your multi-line strings. > > So something like this: > > ``` > # Modelfile generated by "ollama show" > # To build a new Modelfile based on this, replace FROM with: > FROM dolphin-llama3:latest > # (^^I'd recommend using this format just because there are other blob files in the cache that get included when you reference it this way) > > # FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580 > TEMPLATE """{{ if .System }}<|im_start|>system > {{ .System }}<|im_end|> > {{ end }}{{ if .Prompt }}<|im_start|>user > {{ .Prompt }}<|im_end|> > {{ end }}<|im_start|>assistant > {{ .Response }}<|im_end|> > """ > SYSTEM """You are Dolphin, a helpful AI assistant. > """ > PARAMETER stop <|im_start|> > PARAMETER stop <|im_end|> > LICENSE """META LLAMA 3 COMMUNITY LICENSE AGREEMENT > > then goes a huge license agreement, i didn't see a reason to copy it here, if you need i can just load modelfile here. > I tried to do all this on several models, like: llama3, phi3, dolphin-mistral. result is the same. > Also, i tred to paste "FROM dolphin-llama3:latest" instead path to model. > # truncated legal stuff > """ > ``` > > Technically, all that you need to include in the Modelfile is the path to the files that Ollama can understand. Everything else will just be defaulted from the model itself. > > ``` > # Contents of Modelfile > FROM dolphin-llama3 > ``` > > is sufficient to create a new model in Ollama using the command: > > ```shell > ollama create -f /path/to/modelfile dolphin-llama3 > ``` > > ```shell > ollama create -f /path/to/modelfile username/dolphin-llama3 > ``` > > ```shell > ollama create -f /path/to/modelfile username/whacky-inflatable-balloon-man:q4_k > ``` > > (and other permutations of that same command) Well, i tried to fix midelfile, but it still gives me the error. But! i noticed, that when i create modelfile template, the first line of code is `failed to get console mode for stdout: The handle is invalid.` without any marks before it.. and even if i remove that, error will appear anyway. so unedited modelfile looks like this ``` failed to get console mode for stdout: The handle is invalid. # Modelfile generated by "ollama show" # To build a new Modelfile based on this, replace FROM with: # FROM dolphin-llama3:latest FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580 TEMPLATE "{{ if .System }}<|im_start|>system {{ .System }}<|im_end|> {{ end }}{{ if .Prompt }}<|im_start|>user {{ .Prompt }}<|im_end|> {{ end }}<|im_start|>assistant {{ .Response }}<|im_end|> " SYSTEM "You are Dolphin, a helpful AI assistant. " PARAMETER stop <|im_start|> PARAMETER stop <|im_end|> LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT ``` By the way, thank you for links to tutorials!
Author
Owner

@ddutjnrevenge-universe commented on GitHub (Oct 26, 2024):

Try modifying your Modelfile by removing that failed to get console mode for stdout: The handle is invalid. line and removing the # in front of the FROM dolphin-llama3:latest, then put the # before the FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580

Your complete modified Modelfile code should look like this:

# Modelfile generated by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
FROM dolphin-llama3:latest

# FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580
TEMPLATE "{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>
"
SYSTEM "You are Dolphin, a helpful AI assistant.
"
PARAMETER stop <|im_start|>
PARAMETER stop <|im_end|>
LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT

In my case I have problem with the qwen2m and trying to modify the Modelfile like what I explained above works.
Hope this helps! ^^

<!-- gh-comment-id:2439407492 --> @ddutjnrevenge-universe commented on GitHub (Oct 26, 2024): Try modifying your Modelfile by removing that `failed to get console mode for stdout: The handle is invalid.` line and removing the `#` in front of the `FROM dolphin-llama3:latest`, then put the `#` before the `FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580` Your complete modified Modelfile code should look like this: ``` # Modelfile generated by "ollama show" # To build a new Modelfile based on this, replace FROM with: FROM dolphin-llama3:latest # FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580 TEMPLATE "{{ if .System }}<|im_start|>system {{ .System }}<|im_end|> {{ end }}{{ if .Prompt }}<|im_start|>user {{ .Prompt }}<|im_end|> {{ end }}<|im_start|>assistant {{ .Response }}<|im_end|> " SYSTEM "You are Dolphin, a helpful AI assistant. " PARAMETER stop <|im_start|> PARAMETER stop <|im_end|> LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT ``` In my case I have problem with the `qwen2m` and trying to modify the Modelfile like what I explained above works. Hope this helps! ^^
Author
Owner

@josiahsak commented on GitHub (Feb 19, 2025):

Instead of using a .modelfile format use a .txt format. it worked for me.

<!-- gh-comment-id:2667443451 --> @josiahsak commented on GitHub (Feb 19, 2025): Instead of using a .modelfile format use a .txt format. it worked for me.
Author
Owner

@tMrMorgan commented on GitHub (Feb 20, 2025):

Try modifying your Modelfile by removing that failed to get console mode for stdout: The handle is invalid. line and removing the # in front of the FROM dolphin-llama3:latest, then put the # before the FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580

Your complete modified Modelfile code should look like this:

# Modelfile generated by "ollama show"
# To build a new Modelfile based on this, replace FROM with:
FROM dolphin-llama3:latest

# FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580
TEMPLATE "{{ if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>
"
SYSTEM "You are Dolphin, a helpful AI assistant.
"
PARAMETER stop <|im_start|>
PARAMETER stop <|im_end|>
LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT

In my case I have problem with the qwen2m and trying to modify the Modelfile like what I explained above works. Hope this helps! ^^

after a long time i tried to create model. don't know if it's due to new updates or your way of solving really helps, but thank you sm, now it's working.

<!-- gh-comment-id:2670790462 --> @tMrMorgan commented on GitHub (Feb 20, 2025): > Try modifying your Modelfile by removing that `failed to get console mode for stdout: The handle is invalid.` line and removing the `#` in front of the `FROM dolphin-llama3:latest`, then put the `#` before the `FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580` > > Your complete modified Modelfile code should look like this: > > ``` > # Modelfile generated by "ollama show" > # To build a new Modelfile based on this, replace FROM with: > FROM dolphin-llama3:latest > > # FROM F:\Installed\Ollama\models\blobs\sha256-ea025c107c1c3e5a87380b25e205db3b19a4d2ed56f1954f531573a963621580 > TEMPLATE "{{ if .System }}<|im_start|>system > {{ .System }}<|im_end|> > {{ end }}{{ if .Prompt }}<|im_start|>user > {{ .Prompt }}<|im_end|> > {{ end }}<|im_start|>assistant > {{ .Response }}<|im_end|> > " > SYSTEM "You are Dolphin, a helpful AI assistant. > " > PARAMETER stop <|im_start|> > PARAMETER stop <|im_end|> > LICENSE "META LLAMA 3 COMMUNITY LICENSE AGREEMENT > ``` > > In my case I have problem with the `qwen2m` and trying to modify the Modelfile like what I explained above works. Hope this helps! ^^ after a long time i tried to create model. don't know if it's due to new updates or your way of solving really helps, but thank you sm, now it's working.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#64970