The Amazon Inferentia Chip With DLAMI
Amazon Inferentia is a custom machine learning chip designed by Amazon that you can use for high-performance inference predictions. In order to use the chip, set up an Amazon Elastic Compute Cloud instance and use the Amazon Neuron software development kit (SDK) to invoke the Inferentia chip. To provide customers with the best Inferentia experience, Neuron has been built into the Amazon Deep Learning AMIs (DLAMI).
The following topics show you how to get started using Inferentia with the DLAMI.