This repository contains instructions/examples/tutorials using the neuron-sdk
for running inference and training Hugging Face libraries like transformers, datasets with AWS Accelerators including AWS Inferentia & AWS Trainium.
- Speed up BERT inference with Hugging Face Transformers and AWS Inferentia
- Accelerate Hugging Face Transformers with AWS Inferentia 2 | Benchmarking Transformers for AWS Inferentia 2
Before we can start make sure you have met the following requirements
- AWS Account with quota for
Inf1
,Inf2
,Trn1
instances. - AWS CLI installed
- AWS IAM user configured in CLI with permission to create and manage ec2 instances
inf2:
source ../aws_neuron_venv_pytorch_p37/bin/activate