Describe the feature
AWS Bedrock now supports LLM prompt caching (as a beta feature) and it added a new explicitPromptCaching parameter to invoke_model and invoke_model_with_response_stream. But this param is not yet reflected in the boto request shape so trying to use the param fails validation. It would be great if we could add it!
Use Case
We'd love to use the new bedrock prompt caching feature to reduce time to first token.
Proposed Solution
Add explicitPromptCaching to InvokeModelRequest and InvokeModelWithResponseStream in botocore/data/bedrock-runtime/<date>/service-2.json.
Other Information
No response
Acknowledgements
SDK version used
1.35.*
Environment details (OS name and version, etc.)
Any environment
Describe the feature
AWS Bedrock now supports LLM prompt caching (as a beta feature) and it added a new
explicitPromptCachingparameter toinvoke_modelandinvoke_model_with_response_stream. But this param is not yet reflected in the boto request shape so trying to use the param fails validation. It would be great if we could add it!Use Case
We'd love to use the new bedrock prompt caching feature to reduce time to first token.
Proposed Solution
Add
explicitPromptCachingtoInvokeModelRequestandInvokeModelWithResponseStreaminbotocore/data/bedrock-runtime/<date>/service-2.json.Other Information
No response
Acknowledgements
SDK version used
1.35.*
Environment details (OS name and version, etc.)
Any environment