@Generated(value="com.amazonaws:aws-java-sdk-code-generator") public interface AmazonBedrockRuntime
Note: Do not directly implement this interface, new methods are added to it regularly. Extend from
AbstractAmazonBedrockRuntime
instead.
Describes the API operations for running inference using Amazon Bedrock models.
Modifier and Type | Field and Description |
---|---|
static String |
ENDPOINT_PREFIX
The region metadata service name for computing region endpoints.
|
Modifier and Type | Method and Description |
---|---|
ResponseMetadata |
getCachedResponseMetadata(AmazonWebServiceRequest request)
Returns additional metadata for a previously executed successful request, typically used for debugging issues
where a service isn't acting as expected.
|
InvokeModelResult |
invokeModel(InvokeModelRequest invokeModelRequest)
Invokes the specified Amazon Bedrock model to run inference using the prompt and inference parameters provided in
the request body.
|
void |
shutdown()
Shuts down this client object, releasing any resources that might be held open.
|
static final String ENDPOINT_PREFIX
InvokeModelResult invokeModel(InvokeModelRequest invokeModelRequest)
Invokes the specified Amazon Bedrock model to run inference using the prompt and inference parameters provided in the request body. You use model inference to generate text, images, and embeddings.
For example code, see Invoke model code examples in the Amazon Bedrock User Guide.
This operation requires permission for the bedrock:InvokeModel
action.
invokeModelRequest
- AccessDeniedException
- The request is denied because of missing access permissions.ResourceNotFoundException
- The specified resource ARN was not found. Check the ARN and try your request again.ThrottlingException
- The number of requests exceeds the limit. Resubmit your request later.ModelTimeoutException
- The request took too long to process. Processing time exceeded the model timeout length.InternalServerException
- An internal server error occurred. Retry your request.ValidationException
- Input validation failed. Check your request parameters and retry the request.ModelNotReadyException
- The model specified in the request is not ready to serve inference requests.ServiceQuotaExceededException
- The number of requests exceeds the service quota. Resubmit your request later.ModelErrorException
- The request failed due to an error while processing the model.void shutdown()
ResponseMetadata getCachedResponseMetadata(AmazonWebServiceRequest request)
Response metadata is only cached for a limited period of time, so if you need to access this extra diagnostic information for an executed request, you should use this method to retrieve it as soon as possible after executing a request.
request
- The originally executed request.