[GH-ISSUE #5694] API that breaks model output #65586

Closed
opened 2026-05-03 21:47:50 -05:00 by GiteaMirror · 2 comments
Owner

Originally created by @wltime on GitHub (Jul 15, 2024).
Original GitHub issue: https://github.com/ollama/ollama/issues/5694

Sometimes when there is a lot of output, and I don't need the model to continue to output. Is there an API to interrupt the output of the model?

Originally created by @wltime on GitHub (Jul 15, 2024). Original GitHub issue: https://github.com/ollama/ollama/issues/5694 Sometimes when there is a lot of output, and I don't need the model to continue to output. Is there an API to interrupt the output of the model?
GiteaMirror added the feature request label 2026-05-03 21:47:50 -05:00
Author
Owner

@jmorganca commented on GitHub (Jul 15, 2024):

Hi @wltime you can interrupt the request it by cancelling the http connection or by setting num_predict in the options to set an upper limit

<!-- gh-comment-id:2227652905 --> @jmorganca commented on GitHub (Jul 15, 2024): Hi @wltime you can interrupt the request it by cancelling the http connection or by setting `num_predict` in the options to set an upper limit
Author
Owner

@wltime commented on GitHub (Jul 15, 2024):

Hi @wltime you can interrupt the request it by cancelling the http connection or by setting num_predict in the options to set an upper limit
I am using ollama to deploy a dialogue application and the output of ollama is blocking, I am not able to interrupt the current output to continue a new dialogue

<!-- gh-comment-id:2227662205 --> @wltime commented on GitHub (Jul 15, 2024): > Hi @wltime you can interrupt the request it by cancelling the http connection or by setting `num_predict` in the options to set an upper limit I am using ollama to deploy a dialogue application and the output of ollama is blocking, I am not able to interrupt the current output to continue a new dialogue
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#65586