I'm using open-webui for ollama's web pages, but would like to get a faster response time. #1234

Closed
opened 2025-11-11 14:40:44 -06:00 by GiteaMirror · 0 comments
Owner

Originally created by @BA7LWN on GitHub (Jun 12, 2024).

To preload the mistral model using the generate endpoint, use:

curl http://localhost:11434/api/generate -d '{"model": "mistral"}'

To use the chat completions endpoint, use:

curl http://localhost:11434/api/chat -d '{"model": "mistral"}'

I want to know which one I should choose to get faster response time

Originally created by @BA7LWN on GitHub (Jun 12, 2024). ### To preload the mistral model using the generate endpoint, use: ```curl http://localhost:11434/api/generate -d '{"model": "mistral"}'``` ### To use the chat completions endpoint, use: ```curl http://localhost:11434/api/chat -d '{"model": "mistral"}'``` ### I want to know which one I should choose to get faster response time
Sign in to join this conversation.
1 Participants
Notifications
Due Date
No due date set.
Dependencies

No dependencies set.

Reference: github-starred/open-webui#1234