[GH-ISSUE #1806] Add cli switch to show generation time and tokens/sec output time #47543

Closed
opened 2026-04-28 04:07:45 -05:00 by GiteaMirror · 3 comments
Owner

Originally created by @thawkins on GitHub (Jan 5, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/1806

Would it be possible to add a metrics switch to show net generation time and output time with tokens/seconds. This would make comparing the performance of LLMs easier.

Originally created by @thawkins on GitHub (Jan 5, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/1806 Would it be possible to add a metrics switch to show net generation time and output time with tokens/seconds. This would make comparing the performance of LLMs easier.
Author
Owner

@jukofyork commented on GitHub (Jan 5, 2024):

You can use the - -verbose command line option to do this:

> ollama run --help
Run a model

Usage:
  ollama run MODEL [PROMPT] [flags]

Flags:
      --format string   Response format (e.g. json)
  -h, --help            help for run
      --insecure        Use an insecure registry
      --nowordwrap      Don't wrap words to the next line automatically
      --verbose         Show timings for response

I originally didn't realise and was piping in a text file to start it off with the command "/set verbose" each time I ran the CLI!

<!-- gh-comment-id:1878911456 --> @jukofyork commented on GitHub (Jan 5, 2024): You can use the - -verbose command line option to do this: ``` > ollama run --help Run a model Usage: ollama run MODEL [PROMPT] [flags] Flags: --format string Response format (e.g. json) -h, --help help for run --insecure Use an insecure registry --nowordwrap Don't wrap words to the next line automatically --verbose Show timings for response ``` I originally didn't realise and was piping in a text file to start it off with the command "/set verbose" each time I ran the CLI!
Author
Owner

@nejib1 commented on GitHub (Jan 6, 2024):

ollama run mistral --verbose
>>> hello
 Hello! How can I help you today? Is there a specific question or topic you'd like to discuss? I'm here
to provide information and answer any queries you may have. Let me know if there's something on your
mind, and I'll do my best to assist you. If you don't have a particular question, feel free to ask me
about anything that interests you, or just share some conversation starters if you'd like! I'm here to
make this interaction enjoyable and informative for you. So, what would you like to talk about?

total duration:       5.088275983s
load duration:        1.365523ms
prompt eval count:    11 token(s)
prompt eval duration: 204.563ms
prompt eval rate:     53.77 tokens/s
eval count:           120 token(s)
eval duration:        4.876787s
eval rate:            24.61 tokens/s


<!-- gh-comment-id:1879518492 --> @nejib1 commented on GitHub (Jan 6, 2024): ``` ollama run mistral --verbose >>> hello Hello! How can I help you today? Is there a specific question or topic you'd like to discuss? I'm here to provide information and answer any queries you may have. Let me know if there's something on your mind, and I'll do my best to assist you. If you don't have a particular question, feel free to ask me about anything that interests you, or just share some conversation starters if you'd like! I'm here to make this interaction enjoyable and informative for you. So, what would you like to talk about? total duration: 5.088275983s load duration: 1.365523ms prompt eval count: 11 token(s) prompt eval duration: 204.563ms prompt eval rate: 53.77 tokens/s eval count: 120 token(s) eval duration: 4.876787s eval rate: 24.61 tokens/s
Author
Owner

@mxyng commented on GitHub (Jan 8, 2024):

As the other commenters have already mentioned, --verbose is probably what you're looking for.

<!-- gh-comment-id:1881650439 --> @mxyng commented on GitHub (Jan 8, 2024): As the other commenters have already mentioned, `--verbose` is probably what you're looking for.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#47543