Instance selection for the Neptune ML stages - Amazon Neptune
Services or capabilities described in Amazon Web Services documentation might vary by Region. To see the differences applicable to the China Regions, see Getting Started with Amazon Web Services in China.

Instance selection for the Neptune ML stages

The different stages of Neptune ML processing use different SageMaker instances. Here, we discuss how to choose the right instance type for each stage. You can find information about SageMaker instance types and pricing at Amazon SageMaker Pricing.

Selecting an instance for data processing

The SageMaker data-processing step requires a processing instance that has enough memory and disk storage for the input, intermediate, and output data. The specific amount of memory and disk storage needed depends on the characteristics of the Neptune ML graph and its exported features.

By default, Neptune ML chooses the smallest ml.r5 instance whose memory is ten times larger than the size of the exported graph data on disk.

Selecting an instance for model training and model transform

Selecting the right instance type for model training or model transform depends on the task type, the graph size, and your turn-around requirements. GPU instances provide the best performance. We generally recommend p3 and g4dn serial instances. You can also use p2 or p4d instances.

By default, Neptune ML chooses the smallest GPU instance with more memory than model training and model transform requires. You can find what that selection is in the train_instance_recommendation.json file, in the Amazon S3 data processing output location. Here is an example of the contents of a train_instance_recommendation.json file:

{ "instance": "(the recommended instance type for model training and transform)", "cpu_instance": "(the recommended instance type for base processing instance)", "disk_size": "(the estimated disk space required)", "mem_size": "(the estimated memory required)" }

Selecting an instance for an inference endpoint

Selecting the right instance type for an inference endpoint depends on the task type, the graph size and your budget. By default, Neptune ML chooses the smallest ml.m5d instance with more memory the inference endpoint requires.

Note

If more than 384 GB of memory is needed, Neptune ML uses an ml.r5d.24xlarge instance.

You can see what instance type Neptune ML recommends in the infer_instance_recommendation.json file located in the Amazon S3 location you are using for model training. Here is an example of that file's contents:

{ "instance" : "(the recommended instance type for an inference endpoint)", "disk_size" : "(the estimated disk space required)", "mem_size" : "(the estimated memory required)" }