Cloud Instances - Amazon SageMaker
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China (PDF).

Cloud Instances

Amazon SageMaker Neo provides compilation support for popular machine learning frameworks such as TensorFlow, PyTorch, MXNet, and more. You can deploy your compiled model to cloud instances and Amazon Inferentia instances. For a full list of supported frameworks and instances types, see Supported Instances Types and Frameworks.

You can compile your model in one of three ways: through the Amazon CLI, the SageMaker Console, or the SageMaker SDK for Python. See, Use Neo to Compile a Model for more information. Once compiled, your model artifacts are stored in the Amazon S3 bucket URI you specified during the compilation job. You can deploy your compiled model to cloud instances and Amazon Inferentia instances using the SageMaker SDK for Python, Amazon SDK for Python (Boto3), Amazon CLI, or the Amazon console.

If you deploy your model using Amazon CLI, the console, or Boto3, you must select a Docker image Amazon ECR URI for your primary container. See Neo Inference Container Images for a list of Amazon ECR URIs.