Prompt Caching? #9444
Prompt Caching?
#9444
Replies: 2 comments 8 replies
-
@brandonh-msft - thanks for brining this up. This looks like it is happening on the model end so you should be getting this already. Cosmos DB and Redis also supports Semantic Caching if you want to check that out also. You can also do Semantic Caching with filters also here - |
Beta Was this translation helpful? Give feedback.
8 replies
-
Issue filed on OpenAI .NET SDK: openai/openai-dotnet#281 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Prompt caching for AOAI and OAI has been released, but has some requirements around how to make it most effective, specifically, the order in which things are serialized.
Is SK already serializing objects in this manner so token caching will get maximum hits? If not, is there work on the backlog to do it and when could we expect it (for each lang)?
Beta Was this translation helpful? Give feedback.
All reactions