AWS Trainium & Inferentia documentation
Overview
Optimum Neuron
🤗 Optimum NeuronEC2 SetupQuickstartSupported ArchitecturesOptimum Containers Notebooks
How-To Guides
Neuron model cacheDistributed TrainingExport a model to InferentiaInference pipelines with AWS NeuronInference on Neuron platforms using vLLMDeploying a LLM Model with Inference EndpointsBenchmarking LLM performance with vLLM on AWS Inferentia2
Training Tutorials
Fine-tune BERT for Text Classification
How-to Fine-Tune LLMs
Inference Tutorials
EC2
SageMaker
Inference Endpoints
Inference Benchmarks
Contribute
Set up a development environmentAdd a custom model implementation for trainingAdd inference support for a new model architecture
Training API
Models and Pipelines Inference API