This document is relevant for: Inf1
Neuron Hardware#
AWS Neuron hardware consists of custom-designed machine learning accelerators optimized for deep learning workloads. This section covers the architecture and capabilities of AWS Inferentia and Trainium chips, their NeuronCore processing units, and the EC2 instances that host them.
First-generation inference accelerator chip
Second-generation inference accelerator chip
First-generation training accelerator chip
Second-generation training accelerator chip
Processing unit architecture for Inferentia
Processing unit architecture for Inferentia2 and Trainium
Processing unit architecture for Trainium2
Device management and configuration
EC2 instance types with Neuron accelerators
Inf2 instance architecture and specifications
This document is relevant for: Inf1