Note: We simplified this lab into a new repository. https://github.com/awshlabs/Jul2020-Inf1Lab
In this workshop, you gain hands-on experience with Amazon EC2 Inf1 instances, powered by custom AWS Inferentia chips. Amazon EC2 Inf1 instances offer low-latency, high-throughput, and cost-effective machine learning inference in the cloud. This workshop walks you through taking a trained deep learning model to deployment on Amazon EC2 Inf1 instances by using AWS Neuron, an SDK for optimizing inference using AWS Inferentia processors.
Please follow the labs in sequence.
Lab 1. Launch a C5 Instance, install the Neuron development environment, Custom compile a pre-trained model to target the Inferentia Neuron Processor.
Lab 2. Launch an Inf1 Instance, install Neuron run-time and development environment, test and model serve the compiled ResNet package.
Lab 3. Compile on C5 and launch a load test run on Inf1 Instance.
Lab 4. Debug and profile your model on Inf1 Instance.
Reinvent workshop slides at at: slides Directory