Share via


Distributed Data Parallel (DDP) training

Important

This feature is in Beta.

This page has notebook examples for using Distributed Data Parallel (DDP) training on Serverless GPU compute. These examples demonstrate how to scale training across multiple GPUs and nodes for improved performance.

Training a simple multilayer perceptron (MLP) neural network on synthetic dataset using DDP.

The following notebook demonstrates distributed training of a simple multilayer perceptron (MLP) neural network using PyTorch's Distributed Data Parallel (DDP) module on Azure Databricks with serverless GPU resources.

Notebook

Get notebook