Request Inferences from a Deployed Service - Amazon SageMaker
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China (PDF).

Request Inferences from a Deployed Service

If you have followed instructions in Deploy a Model, you should have a SageMaker endpoint set up and running. Regardless of how you deployed your Neo-compiled model, there are three ways you can submit inference requests: