-
Notifications
You must be signed in to change notification settings - Fork 152
Closed as not planned
Labels
duplicateThis issue or pull request already existsThis issue or pull request already existsenhancementNew feature or requestNew feature or request
Description
One of the things it is going to be interesting to have out of the box about AI is the semantic cache of requests.
Actually, it could be used in any method but according to a recent study, 31% of queries to LLM can be cached (or, in other words, 31% of the queries are contextually repeatable), which can significantly improve response time in GenAI apps.
I created a simple example that implements this with Redis: https://github.com/lordofthejars-ai/quarkus-langchain-examples/tree/main/semantic-cache
Do you think it might be interesting to integrate this into Quarkus Cache system for example as Redis-semantic-cache or something like this?
Metadata
Metadata
Assignees
Labels
duplicateThis issue or pull request already existsThis issue or pull request already existsenhancementNew feature or requestNew feature or request