From PyTorch DDP to Accelerate to Trainer, mastery of distributed training with ease
This tutorial assumes you have a basic understanding of PyTorch and how to train a simple model. It will showcase training on multiple GPUs through a process called Distributed Data Parallelism (DDP) through three different levels of increasing abstraction:
- Native PyTorch DDP through the
pytorch.distributedmodule - Utilizing 🤗 Accelerate’s light wrapper around
pytorch.distributedthat also helps ensure the code can be run