mirror of
https://github.com/open-webui/open-webui.git
synced 2026-03-13 02:24:42 -05:00
[PR #7237] [CLOSED] feat: add cache_prompt option for llama-server completions #8828
Reference in New Issue
Block a user
Delete Branch "%!s()"
Deleting a branch is permanent. Although the deleted branch may continue to exist for a short time before it actually gets removed, it CANNOT be undone in most cases. Continue?
📋 Pull Request Information
Original PR: https://github.com/open-webui/open-webui/pull/7237
Author: @drunnells
Created: 11/22/2024
Status: ❌ Closed
Base:
dev← Head:main📝 Commits (1)
75d13a0feat: add cache_prompt option for llama-server completions📊 Changes
3 files changed (+46 additions, -0 deletions)
View changed files
📝
backend/open_webui/apps/openai/main.py(+4 -0)📝
backend/open_webui/utils/payload.py(+1 -0)📝
src/lib/components/chat/Settings/Advanced/AdvancedParams.svelte(+41 -0)📄 Description
Pull Request Checklist
Note to first-time contributors: Please open a discussion post in Discussions and describe your changes before submitting a pull request.
Before submitting, make sure you've checked the following:
devbranch.Changelog Entry
Description
cache_promptoption for llama-server completions, enabling users to reuse the KV cache when supported, improving completion speed in long conversations when using llama.cpp server.Added
cache_promptparameter in the payload for llama-server completions.cache_promptin Advanced Settings.Changed
Deprecated
Removed
Fixed
Security
Breaking Changes
Additional Information
cache_promptis enabled and disabled.Screenshots or Videos
🔄 This issue represents a GitHub Pull Request. It cannot be merged through Gitea due to API limitations.