Skip to content

Support for InvokeModelWithResponseStream param in invoke_model and InvokeModelWithResponseStream in Bedrock #3323

@realyze

Description

@realyze

Describe the feature

AWS Bedrock now supports LLM prompt caching (as a beta feature) and it added a new explicitPromptCaching parameter to invoke_model and invoke_model_with_response_stream. But this param is not yet reflected in the boto request shape so trying to use the param fails validation. It would be great if we could add it!

Use Case

We'd love to use the new bedrock prompt caching feature to reduce time to first token.

Proposed Solution

Add explicitPromptCaching to InvokeModelRequest and InvokeModelWithResponseStream in botocore/data/bedrock-runtime/<date>/service-2.json.

Other Information

No response

Acknowledgements

  • I may be able to implement this feature request
  • This feature might incur a breaking change

SDK version used

1.35.*

Environment details (OS name and version, etc.)

Any environment

Metadata

Metadata

Assignees

Labels

feature-requestThis issue requests a feature.service-apiThis issue is caused by the service API, not the SDK implementation.

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions