Previous Releases Notes (Neuron 2.x)
Contents
This document is relevant for: Inf1
, Inf2
, Trn1
, Trn1n
Previous Releases Notes (Neuron 2.x)#
Table of contents
Neuron 2.8.0 (02/24/2023)#
Table of contents
What’s New#
This release adds support for EC2 Inf2 instances, introduces initial inference support with TensorFlow 2.x Neuron (tensorflow-neuronx
) on Trn1 and Inf2, and introduces minor enhancements and bug fixes.
This release introduces the following:
What’s New |
Details |
---|---|
Support for EC2 Inf2 instances |
|
TensorFlow 2.x Neuron ( |
|
New Neuron GitHub samples |
|
Minor enhancements and bug fixes. |
|
Release included packages |
|
For more detailed release notes of the new features and resolved issues, see Neuron Components Release Notes.
Neuron 2.7.0 (02/08/2023)#
Table of contents
What’s New#
This release introduces new capabilities and libraries, as well as features and tools that improves usability. This release introduces the following:
What’s New |
Details |
---|---|
PyTorch 1.13 |
Support of PyTorch 1.13 version for PyTorch Neuron ( |
PyTorch DistributedDataParallel (DDP) API |
Support of PyTorch DistributedDataParallel (DDP) API in PyTorch Neuron ( |
Inference support in |
For more details please visit pytorch-neuronx-main` page. You can also try Neuron Inference samples https://github.com/aws-neuron/aws-neuron-samples/tree/master/torch-neuronx in the |
Neuron Custom C++ Operators[Experimental] |
Initial support for Neuron Custom C++ Operators [Experimental] , with Neuron Custom C++ Operators (“CustomOps”) you can now write CustomOps that run on NeuronCore-v2 chips. For more resources please check Neuron Custom C++ Operators [Experimental] section. |
|
|
Neuron sysfs filesystem |
Neuron sysfs filesystem exposes Neuron Devices under |
TFLOPS support in Neuron System Tools |
Neuron System Tools now also report model actual TFLOPs rate in both |
New sample scripts for training |
This release adds multiple new sample scripts for training models with |
New sample scripts for inference |
This release adds multiple new sample scripts for deploying models with |
Neuron GitHub samples repository for Amazon EKS |
A new AWS Neuron GitHub samples repository for Amazon EKS, Please check aws-neuron-samples repository |
For more detailed release notes of the new features and resolved issues, see Neuron Components Release Notes.
Neuron 2.6.0 (12/12/2022)#
This release introduces the support of PyTorch 1.12 version, and introduces PyTorch Neuron (torch-neuronx
) profiling through Neuron Plugin for TensorBoard. Pytorch Neuron (torch-neuronx
) users can now profile their models through the following TensorBoard views:
Operator Framework View
Operator HLO View
Operator Trace View
This release introduces the support of LAMB optimizer for FP32 mode, and adds support for capturing snapshots of inputs, outputs and graph HLO for debugging.
In addition, this release introduces the support of new operators and resolves issues that improve stability for Trn1 customers.
For more detailed release notes of the new features and resolved issues, see Neuron Components Release Notes.
Neuron 2.5.0 (11/23/2022)#
Neuron 2.5.0 is a major release which introduces new features and resolves issues that improve stability for Inf1 customers.
Component |
New in this release |
---|---|
PyTorch Neuron |
|
TensorFlow Neuron |
|
This Neuron release is the last release that will include torch-neuron
versions 1.7 and 1.8, and that will include tensorflow-neuron
versions 2.5 and 2.6.
In addition, this release introduces changes to the Neuron packaging and installation instructions for Inf1 customers, see Introducing Neuron packaging and installation changes for Inf1 customers for more information.
For more detailed release notes of the new features and resolved issues, see Neuron Components Release Notes.
Neuron 2.4.0 (10/27/2022)#
This release introduces new features and resolves issues that improve stability. The release introduces “memory utilization breakdown” feature in both Neuron Monitor and Neuron Top system tools. The release introduces support for “NeuronCore Based Sheduling” capability to the Neuron Kubernetes Scheduler and introduces new operators support in Neuron Compiler and PyTorch Neuron. This release introduces also additional eight (8) samples of models’ fine tuning using PyTorch Neuron. The new samples can be found in the AWS Neuron Samples GitHub repository.
Neuron 2.3.0 (10/10/2022)#
This Neuron 2.3.0 release extends Neuron 1.x and adds support for the new AWS Trainium powered Amazon EC2 Trn1 instances. With this release, you can now run deep learning training workloads on Trn1 instances to save training costs by up to 50% over equivalent GPU-based EC2 instances, while getting the highest training performance in AWS cloud for popular NLP models.
What’s New |
|
Tested workloads and known issues |
|
Neural-networks training support
Supported instances: Trn1
Supported Frameworks: PyTorch Neuron (torch-neuronx)
Supported Data-types
FP32, BF16
Supported Rounding Modes
Stochastic Rounding (SR)
Round Nearest ties to Even (RNE)
Supported Automatic Casting Methods
Neuron automatic casting of FP32 tensors / weights / operations to BF16 - Default mode
PyTorch automatic casting
Full BF16 automatic casting (via XLA_USE_BF16=1 environment variable)
PyTorch Neuron (torch-neuronx
)
PyTorch 1.11
Supported instances: Trn1
Supported Python versions: Python 3.7, Python 3.8
Eager Debug Mode
Persistent Cache for compilation
Collective compute operations: AllReduce
Optimizers: AdamW, SGD
Tested loss functions: Negative log-likelihood (NLL), Cross-entropy
Training Libraries/Frameworks
torch.distributed
Megatron-LM Reference for Neuron
Training Examples
For More information:
Neuron Runtime, Drivers and Networking Components
Neuron Runtime 2.9
Supported instances: Trn1, Inf1
Elastic Fabric Adapter (EFA) @ 800Gbps
Collective communication operators
AllReduce
AllGather
ReduceScatter
Release Notes:
Neuron Tools
Neuron system tools - Adding Trn1 support to the following tools:
neuron-monitor
neuron-top
neuron-ls
Release Notes:
Developer Flows
Containers
Deep Learning Containers (DLC) supporting PyTorch Neuron (
torch-neuronx
)
Multi-Instance distributed workloads orchestration:
AWS ParallelCluster (Through custom AMI build)
Amazon Elastic Compute Cloud (ECS)
Supported Amazon Linux Images (AMIs)
Ubuntu 20 Neuron DLAMI-base (Python 3.8)
Amazon Linux2 Neuron DLAMI-base (Python 3.7)
Ubuntu 18 Neuron DLAMI-base (Python 3.7)
Ubuntu 18 AMI (Python 3.7)
Ubuntu 20 AMI (Python 3.8)
Amazon Linux2 AMI (Python 3.7)
The following workloads were tested in this release:
Distributed data-parallel pre-training of Hugging Face BERT model on single Trn1.32xl instance (32 NeuronCores).
Distributed data-parallel pre-training of Hugging Face BERT model on multiple Trn1.32xl instances.
HuggingFace BERT MRPC task finetuning on single NeuronCore or multiple NeuronCores (data-parallel).
Megatron-LM GPT3 (6.7B parameters) pre-training on single Trn1.32xl instance.
Megatron-LM GPT3 (6.7B parameters) pre-training on multi Trn1.32xl instances.
Multi-Layer Perceptron (ML) model training on single NeuronCore or multiple NeuronCores (data-parallel).
For maximum training performance, please set environment variables
XLA_USE_BF16=1
to enable full BF16 and Stochastic Rounding.
This document is relevant for: Inf1
, Inf2
, Trn1
, Trn1n