[GH-ISSUE #318] add a generate option for max response length #62177

Closed
opened 2026-05-03 07:45:39 -05:00 by GiteaMirror · 1 comment
Owner

Originally created by @BruceMacD on GitHub (Aug 10, 2023).
Original GitHub issue: https://github.com/ollama/ollama/issues/318

Add an option to the generate request body that specifies the max tokens a response from the model can be.

split from #281

Originally created by @BruceMacD on GitHub (Aug 10, 2023). Original GitHub issue: https://github.com/ollama/ollama/issues/318 Add an option to the generate request body that specifies the max tokens a response from the model can be. split from #281
GiteaMirror added the feature request label 2026-05-03 07:45:39 -05:00
Author
Owner

@jmorganca commented on GitHub (Sep 7, 2023):

This should be possible with the num_predict option/parameter 🎉

<!-- gh-comment-id:1710181439 --> @jmorganca commented on GitHub (Sep 7, 2023): This should be possible with the `num_predict` option/parameter 🎉
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/ollama#62177