Data Parallelism: How to Train Deep Learning Models on Multiple GPUs
Monday, November 6, 2023 -
9:00 AM
Monday, November 6, 2023
9:00 AM
Welcome and Introduction
Welcome and Introduction
9:00 AM - 9:15 AM
- Meet the instructor. - Create an account at courses.nvidia.com/join
9:15 AM
Stochastic Gradient Descent and the Effects of Batch Size
Stochastic Gradient Descent and the Effects of Batch Size
9:15 AM - 11:15 AM
Learn the significance of stochastic gradient descent when training on multiple GPUs - Understand the issues with sequential single-thread data processing and the theory behind speeding up applications with parallel processing. - Understand loss function, gradient descent, and stochastic gradient descent (SGD). - Understand the effect of batch size on accuracy and training time with an eye towards its use on multi-GPU systems.
11:15 AM
Break
Break
11:15 AM - 12:15 PM
12:15 PM
Training on Multiple GPUs with PyTorch Distributed Data Parallel (DDP)
Training on Multiple GPUs with PyTorch Distributed Data Parallel (DDP)
12:15 PM - 2:15 PM
Learn to convert single GPU training to multiple GPUs using PyTorch Distributed Data Parallel - Understand how DDP coordinates training among multiple GPUs. - Refactor single-GPU training programs to run on multiple GPUs with DDP.
2:15 PM
Break
Break
2:15 PM - 2:30 PM
2:30 PM
Maintaining Model Accuracy when Scaling to Multiple GPUs
Maintaining Model Accuracy when Scaling to Multiple GPUs
2:30 PM - 4:00 PM
Understand and apply key algorithmic considerations to retain accuracy when training on multiple GPUs - Understand what might cause accuracy to decrease when parallelizing training on multiple GPUs. - Learn and understand techniques for maintaining accuracy when scaling training to multiple GPUs.
4:00 PM
Workshop Assessment
Workshop Assessment
4:00 PM - 4:30 PM
Use what you have learned during the workshop: complete the workshop assessment to earn a certificate of competency
4:30 PM
Final Review
Final Review
4:30 PM - 4:45 PM
- Review key learnings and wrap up questions. - Take the workshop survey.