.. _model_samples_training_trn1: Training Samples/Tutorials (Trn1/Trn1n) ======================================= .. contents:: Table of contents :local: :depth: 1 .. _encoder_model_samples_training_trn1: Encoders -------- .. list-table:: :widths: 20 15 45 :header-rows: 1 :align: left :class: table-smaller-font-size * - Model - Frameworks/Libraries - Samples and Tutorials * - bert-base-cased - torch-neuronx - * `Fine-tune a "bert-base-cased" PyTorch model for Text Classification `_ * `How to fine-tune a "bert base cased" PyTorch model with AWS Trainium (Trn1 instances) for Sentiment Analysis `_ * - bert-base-uncased - torch-neuronx - * `Fine-tune a "bert-base-uncased" PyTorch model `_ * `Fine tuning BERT base model from HuggingFace on Amazon SageMaker `_ * - bert-large-cased - torch-neuronx - * `Fine-tune a "bert-large-cased" PyTorch model `_ * - bert-large-uncased - torch-neuronx - * :ref:`hf-bert-pretraining-tutorial` * `Launch Bert Large Phase 1 pretraining job on Parallel Cluster `_ * `Launch a Multi-Node PyTorch Neuron Training Job on Trainium Using TorchX and EKS `_ * :ref:`torch-hf-bert-finetune` * `Fine-tune a "bert-large-uncased" PyTorch model `_ * - roberta-base - tensorflow-neuronx - * `Fine-tune a "roberta-base" PyTorch model `_ * - roberta-large - torch-neuronx - * `Fine-tune a "roberta-large" PyTorch model `_ * - xlm-roberta-base - torch-neuronx - * `Fine-tune a "xlm-roberta-base" PyTorch model `_ * - alberta-base-v2 - torch-neuronx - * `Fine-tune a "alberta-base-v2" PyTorch model `_ * - distilbert-base-uncased - torch-neuronx - * `Fine-tune a "distilbert-base-uncased" PyTorch model `_ * - camembert-base - torch-neuronx - * `Fine-tune a "camembert-base PyTorch model `_ * - cl-tohoku/bert-base-japanese-whole-word-masking - torch-neuronx - * `Fine-tuning & Deployment Hugging Face BERT Japanese model `_ .. _decoder_model_samples_training_trn1: Decoders -------- .. list-table:: :widths: 20 15 45 :header-rows: 1 :align: left :class: table-smaller-font-size * - Model - Frameworks/Libraries - Samples and Tutorials * - gpt-2 - torch-neuronx - * `How to run training jobs for "gpt2" PyTorch model with AWS Trainium `_ * :ref:`zero1-gpt2-pretraining-tutorial` * - gpt-3 - neuronx-nemo-megatron - * `Launch a GPT-3 23B pretraining job using neuronx-nemo-megatron `_ * `Launch a GPT-3 46B pretraining job using neuronx-nemo-megatron `_ * `Launch a GPT-3 175B pretraining job using neuronx-nemo-megatron `_ * - GPT-NEOX-20B - neuronx-distributed - * :ref:`gpt_neox_20b_tp_zero1_tutorial` * `Training GPT-NEOX 20B model using neuronx-distributed `_ * `Pre-train GPT Neox 20b on Wikicorpus dataset using Neuronx Distributed library `_ * - GPT-NEOX-6.9B - neuronx-distributed - * :ref:`gpt_neox_tp_zero1_tutorial` * `Training GPT-NEOX 6.9B model using neuronx-distributed `_ * `Pre-train GPT Neox 6.9b on Wikicorpus dataset using Neuronx Distributed library `_ * - meta-llama/Llama-2-7b - neuronx-distributed - * :ref:`llama2_7b_tp_zero1_tutorial` * `Training Llama2 7B Model with AWS Batch and Trainium `_ * :ref:`llama2_7b_tp_zero1_ptl_finetune_tutorial` * `Pre-train Llama2-7B on Wikicorpus dataset using Neuronx Distributed library `_ * - meta-llama/Llama-2-13b - neuronx-distributed - * :ref:`llama2_tp_pp_tutorial` * - meta-llama/Llama-2-70b - neuronx-distributed - * :ref:`llama2_tp_pp_tutorial` * - codegen25-7b-mono - neuronx-distributed - * :ref:`codegen25_7b_tp_zero1_tutorial` * - meta-llama/Llama-2 - neuronx-nemo-megatron - * `Launch a Llama-2-7B pretraining job using neuronx-nemo-megatron `_ * `Launch a Llama-2-13B pretraining job using neuronx-nemo-megatron `_ * `Launch a Llama-2-70B pretraining job using neuronx-nemo-megatron `_ * - Mistral-7B - neuronx-nemo-megatron - * `Training Mistral-7B `_ .. _encoder_decoder_model_samples_training_trn1: Encoder-Decoders ---------------- .. list-table:: :widths: 20 15 45 :header-rows: 1 :align: left :class: table-smaller-font-size * - Model - Frameworks/Libraries - Samples and Tutorials * - t5-small - * torch-neuronx * optimum-neuron - * :ref:`torch-hf-t5-finetune` * - facebook/bart-large - * torch-neuronx - * `How to fine-tune a "Bart-Large" PyTorch model with AWS Trainium (trn1 instances) `_ .. _vision_transformer_model_samples_training_trn1: Vision Transformers ------------------- .. list-table:: :widths: 20 15 45 :header-rows: 1 :align: left :class: table-smaller-font-size * - Model - Frameworks/Libraries - Samples and Tutorials * - google/vit-base-patch16-224-in21k - torch-neuronx - * `Fine-tune a pretrained HuggingFace vision transformer PyTorch model `_ * - openai/clip-vit-base-patch32 - torch-neuronx - * `Fine-tune a pretrained HuggingFace CLIP-base PyTorch model with AWS Trainium `_ * - openai/clip-vit-large-patch14 - torch-neuronx - * `Fine-tune a pretrained HuggingFace CLIP-large PyTorch model with AWS Trainium `_ .. _sd_model_samples_training_trn1: Stable Diffusion ---------------- .. list-table:: :widths: 20 15 45 :header-rows: 1 :align: left :class: table-smaller-font-size * - Model - Frameworks/Libraries - Samples and Tutorials * - stabilityai/stable-diffusion-2-1-base - torch-neuronx - * [Beta] `Train stabilityai/stable-diffusion-2-1-base with AWS Trainium (trn1 instances) `_ * - runwayml/stable-diffusion-v1-5 - torch-neuronx - * [Beta] `Train runwayml/stable-diffusion-v1-5 with AWS Trainium (trn1 instances) `_ .. _multi_modal_model_samples_training_trn1: Multi Modal ----------- .. list-table:: :widths: 20 15 45 :header-rows: 1 :align: left :class: table-smaller-font-size * - Model - Frameworks/Libraries - Samples and Tutorials * - language-perceiver - torch-neuronx - * `How to fine-tune a "language perceiver" PyTorch model with AWS Trainium (trn1 instances) `_ * - vision-perceiver-conv - torch-neuronx - * `How to fine-tune a pretrained HuggingFace Vision Perceiver Conv `_ .. _cnn_model_samples_training_trn1: Convolutional Neural Networks(CNN) ---------------------------------- .. list-table:: :widths: 20 15 45 :header-rows: 1 :align: left :class: table-smaller-font-size * - Model - Frameworks/Libraries - Samples and Tutorials * - resnet50 - torch-neuronx - * `How to fine-tune a pretrained ResNet50 Pytorch model with AWS Trainium (trn1 instances) using NeuronSDK `_ * - milesial/Pytorch-UNet - torch-neuronx - * `This notebook shows how to fine-tune a pretrained UNET PyTorch model with AWS Trainium (trn1 instances) using NeuronSDK. `_