Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Support prompt caching API #220

Open
yuzisun opened this issue Jan 29, 2025 · 1 comment
Open

Support prompt caching API #220

yuzisun opened this issue Jan 29, 2025 · 1 comment
Labels
enhancement New feature or request

Comments

@yuzisun
Copy link
Contributor

yuzisun commented Jan 29, 2025

Description:
Extend OpenAI compatible API to support prompt caching API, this feature is supported both by KServe via LMCache and AWS Bedrock

AWS Bedrock: https://docs.aws.amazon.com/bedrock/latest/userguide/prompt-caching.html
LMCache: https://github.com/LMCache/LMCache

[optional Relevant Links:]

Any extra documentation required to understand the issue.

@yuzisun yuzisun added the enhancement New feature or request label Jan 29, 2025
@mathetake
Copy link
Member

mathetake commented Jan 29, 2025

could you share the big picture of what kind of change/implementation is necessary to this repo? (I guess it's transformer impl?)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants