[GH-ISSUE #188] feat: verbose output (show timings per response) #27498

Closed
opened 2026-04-25 02:11:57 -05:00 by GiteaMirror · 6 comments
Owner

Originally created by @johncadengo on GitHub (Dec 7, 2023).
Original GitHub issue: https://github.com/open-webui/open-webui/issues/188

When running Ollama from the command prompt, you can type the --verbose argument to get timings that output like this:

$ ollama run --verbose llama2
>>> Hi
Hello! It's nice to meet you. Is there something I can help you with or would you like to chat?

total duration:       1.279118717s
load duration:        891.933µs
prompt eval count:    21 token(s)
prompt eval duration: 231.416ms
prompt eval rate:     90.75 tokens/s
eval count:           25 token(s)
eval duration:        1.042407s
eval rate:            23.98 tokens/s

Is there a way to get this to appear in the web UI?

Originally created by @johncadengo on GitHub (Dec 7, 2023). Original GitHub issue: https://github.com/open-webui/open-webui/issues/188 When running Ollama from the command prompt, you can type the `--verbose` argument to get timings that output like this: ``` $ ollama run --verbose llama2 >>> Hi Hello! It's nice to meet you. Is there something I can help you with or would you like to chat? total duration: 1.279118717s load duration: 891.933µs prompt eval count: 21 token(s) prompt eval duration: 231.416ms prompt eval rate: 90.75 tokens/s eval count: 25 token(s) eval duration: 1.042407s eval rate: 23.98 tokens/s ``` Is there a way to get this to appear in the web UI?
Author
Owner

@tjbck commented on GitHub (Dec 7, 2023):

Hi, I'll think of a way to incorporate this into the webui, stay tuned!

<!-- gh-comment-id:1846036508 --> @tjbck commented on GitHub (Dec 7, 2023): Hi, I'll think of a way to incorporate this into the webui, stay tuned!
Author
Owner

@tjbck commented on GitHub (Dec 19, 2023):

image

The requested feature has been added with #215. Please try out the latest version, and let me know if you encounter any issues. Thanks!

<!-- gh-comment-id:1862027446 --> @tjbck commented on GitHub (Dec 19, 2023): ![image](https://github.com/ollama-webui/ollama-webui/assets/25473318/49b833ed-40da-43a2-ba85-0e6666860e3e) The requested feature has been added with #215. Please try out the latest version, and let me know if you encounter any issues. Thanks!
Author
Owner

@johncadengo commented on GitHub (Dec 21, 2023):

Ty! 👏 👏 👏

<!-- gh-comment-id:1865320329 --> @johncadengo commented on GitHub (Dec 21, 2023): Ty! 👏 👏 👏
Author
Owner

@johncadengo commented on GitHub (Apr 22, 2024):

@tjbck I have just spun up the latest version of this image and so many things have changed. Do you happen to know if this feature is still present? I can't seem to find it in the latest version.

<!-- gh-comment-id:2068685659 --> @johncadengo commented on GitHub (Apr 22, 2024): @tjbck I have just spun up the latest version of this image and so many things have changed. Do you happen to know if this feature is still present? I can't seem to find it in the latest version.
Author
Owner

@tjbck commented on GitHub (Apr 22, 2024):

image

All still there :)

<!-- gh-comment-id:2069845379 --> @tjbck commented on GitHub (Apr 22, 2024): <img width="886" alt="image" src="https://github.com/open-webui/open-webui/assets/25473318/c95022ff-141d-4435-9637-c09bbd366657"> All still there :)
Author
Owner

@johncadengo commented on GitHub (Apr 24, 2024):

@tjbck okay, thanks for showing that.

I think what happened between versions for me is that I was decoupling the model API from the docker container and trying to have a separate API for the models. At first, I was using the LiteLLM models feature. I'm assuming it doesn't work for that?

Anyhow, I figured out how to use Ollama models instead and now I see the generation info again.

<!-- gh-comment-id:2075829713 --> @johncadengo commented on GitHub (Apr 24, 2024): @tjbck okay, thanks for showing that. I think what happened between versions for me is that I was decoupling the model API from the docker container and trying to have a separate API for the models. At first, I was using the LiteLLM models feature. I'm assuming it doesn't work for that? Anyhow, I figured out how to use Ollama models instead and now I see the generation info again.
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#27498