Frugality meets Accuracy: Cost-efficient training of GPT NeoX and Pythia models with AWS Trainium
AWS Machine Learning Blog
DECEMBER 12, 2023
Walkthrough Download the pre-tokenized Wikipedia dataset as shown: export DATA_DIR=~/examples_datasets/gpt2 mkdir -p ${DATA_DIR} && cd ${DATA_DIR} wget [link] wget [link] aws s3 cp s3://neuron-s3/training_datasets/gpt/wikipedia/my-gpt2_text_document.bin. Each trn1.32xl has 16 accelerators with two workers per accelerator.
Let's personalize your content