PyTorch Distributed: Experiences on Accelerating Data Parallel Training Paper • 2006.15704 • Published Jun 28, 2020 • 4
PyTorch FSDP: Experiences on Scaling Fully Sharded Data Parallel Paper • 2304.11277 • Published Apr 21, 2023 • 5