Request Inferences from a Deployed Service (Amazon CLI) - Amazon SageMaker
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China (PDF).

Request Inferences from a Deployed Service (Amazon CLI)

Inference requests can be made with the sagemaker-runtime invoke-endpoint once you have an Amazon SageMaker endpoint InService. You can make inference requests with the Amazon Command Line Interface (Amazon CLI). The following example shows how to send an image for inference:

aws sagemaker-runtime invoke-endpoint --endpoint-name 'insert name of your endpoint here' --body fileb://image.jpg --content-type=application/x-image output_file.txt

An output_file.txt with information about your inference requests is made if the inference was successful.

For TensorFlow submit an input with application/json as the content type.

aws sagemaker-runtime invoke-endpoint --endpoint-name 'insert name of your endpoint here' --body fileb://input.json --content-type=application/json output_file.txt