Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
Important
This feature is in Beta.
This page has notebook examples for using Distributed Data Parallel (DDP) training on Serverless GPU compute. These examples demonstrate how to scale training across multiple GPUs and nodes for improved performance.
Training a simple multilayer perceptron (MLP) neural network on synthetic dataset using DDP.
The following notebook demonstrates distributed training of a simple multilayer perceptron (MLP) neural network using PyTorch's Distributed Data Parallel (DDP) module on Azure Databricks with serverless GPU resources.