[GH-ISSUE #3181] Suppressing output of all the metadata. #63997

Closed
opened 2026-05-03 15:44:40 -05:00 by GiteaMirror · 0 comments
Owner

Originally created by @phalexo on GitHub (Mar 16, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/3181

What is the issue?

I want to use Ollama to serve a local LLM with OpenAI API to allow Pythogora/gpt-pilot to interact with it.
The back end constantly prints out the crap below:

{"function":"launch_slot_with_data","id_slot":0,"id_task":2842,"level":"INFO","line":1002,"msg":"slot is processing task","tid":"139720604485376","timestamp":1710547752}
{"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":8,"tid":"139720604485376","timestamp":1710547752}
{"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":520,"tid":"139720604485376","timestamp":1710547760}
{"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":1032,"tid":"139720604485376","timestamp":1710547767}
{"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":1544,"tid":"139720604485376","timestamp":1710547775}
DONE{"function":"print_timings","id_slot":0,"id_task":2842,"level":"INFO","line":310,"msg":"prompt eval time     =   26782.10 ms /  1736 tokens (   15.43 ms per token,    64.82 tokens per second)","n_prompt_tokens_processed":1736,"n_tokens_second":64.81942506738177,"t_prompt_processing":26782.095,"t_token":15.427474078341014,"tid":"139720604485376","timestamp":1710547780}
{"function":"print_timings","id_slot":0,"id_task":2842,"level":"INFO","line":326,"msg":"generation eval time =     612.50 ms /     3 runs   (  204.17 ms per token,     4.90 tokens per second)","n_decoded":3,"n_tokens_second":4.897951187018471,"t_token":204.167,"t_token_generation":612.501,"tid":"139720604485376","timestamp":1710547780}
{"function":"print_timings","id_slot":0,"id_task":2842,"level":"INFO","line":337,"msg":"          total time =   27394.60 ms","t_prompt_processing":26782.095,"t_token_generation":612.501,"t_total":27394.596,"tid":"139720604485376","timestamp":1710547780}
{"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1611,"msg":"slot released","n_cache_tokens":1746,"n_ctx":2048,"n_past":1746,"n_system_tokens":0,"tid":"139720604485376","timestamp":1710547780,"truncated":false}
{"function":"update_slots","level":"INFO","line":1637,"msg":"all slots are idle","tid":"139720604485376","timestamp":1710547780}
{"function":"update_slots","level":"INFO","line":1637,"msg":"all slots are idle","tid":"139720604485376","timestamp":1710547780}
[GIN] 2024/03/15 - 20:09:40 | 200 | 27.405338872s |       127.0.0.1 | POST     "/v1/chat/completions"

What did you expect to see?

I want some means to see NO metadata at all. I just to see the prompt text and the answer text and none of the spurious stuff.

Steps to reproduce

It is obvious "ollama run somemodel"

Are there any recent changes that introduced the issue?

No response

OS

No response

Architecture

No response

Platform

No response

Ollama version

No response

GPU

No response

GPU info

No response

CPU

No response

Other software

No response

Originally created by @phalexo on GitHub (Mar 16, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/3181 ### What is the issue? I want to use Ollama to serve a local LLM with OpenAI API to allow Pythogora/gpt-pilot to interact with it. The back end constantly prints out the crap below: ```bash {"function":"launch_slot_with_data","id_slot":0,"id_task":2842,"level":"INFO","line":1002,"msg":"slot is processing task","tid":"139720604485376","timestamp":1710547752} {"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":8,"tid":"139720604485376","timestamp":1710547752} {"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":520,"tid":"139720604485376","timestamp":1710547760} {"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":1032,"tid":"139720604485376","timestamp":1710547767} {"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1916,"msg":"kv cache rm [p0, end)","p0":1544,"tid":"139720604485376","timestamp":1710547775} DONE{"function":"print_timings","id_slot":0,"id_task":2842,"level":"INFO","line":310,"msg":"prompt eval time = 26782.10 ms / 1736 tokens ( 15.43 ms per token, 64.82 tokens per second)","n_prompt_tokens_processed":1736,"n_tokens_second":64.81942506738177,"t_prompt_processing":26782.095,"t_token":15.427474078341014,"tid":"139720604485376","timestamp":1710547780} {"function":"print_timings","id_slot":0,"id_task":2842,"level":"INFO","line":326,"msg":"generation eval time = 612.50 ms / 3 runs ( 204.17 ms per token, 4.90 tokens per second)","n_decoded":3,"n_tokens_second":4.897951187018471,"t_token":204.167,"t_token_generation":612.501,"tid":"139720604485376","timestamp":1710547780} {"function":"print_timings","id_slot":0,"id_task":2842,"level":"INFO","line":337,"msg":" total time = 27394.60 ms","t_prompt_processing":26782.095,"t_token_generation":612.501,"t_total":27394.596,"tid":"139720604485376","timestamp":1710547780} {"function":"update_slots","id_slot":0,"id_task":2842,"level":"INFO","line":1611,"msg":"slot released","n_cache_tokens":1746,"n_ctx":2048,"n_past":1746,"n_system_tokens":0,"tid":"139720604485376","timestamp":1710547780,"truncated":false} {"function":"update_slots","level":"INFO","line":1637,"msg":"all slots are idle","tid":"139720604485376","timestamp":1710547780} {"function":"update_slots","level":"INFO","line":1637,"msg":"all slots are idle","tid":"139720604485376","timestamp":1710547780} [GIN] 2024/03/15 - 20:09:40 | 200 | 27.405338872s | 127.0.0.1 | POST "/v1/chat/completions" ``` ### What did you expect to see? I want some means to see NO metadata at all. I just to see the prompt text and the answer text and none of the spurious stuff. ### Steps to reproduce It is obvious "ollama run somemodel" ### Are there any recent changes that introduced the issue? _No response_ ### OS _No response_ ### Architecture _No response_ ### Platform _No response_ ### Ollama version _No response_ ### GPU _No response_ ### GPU info _No response_ ### CPU _No response_ ### Other software _No response_
GiteaMirror added the bug label 2026-05-03 15:44:40 -05:00
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#63997