Skip to main content

Multi-GPU distributed training

Beta

This feature is in Beta. Workspace admins can control access to this feature from the Previews page. See Manage Databricks previews.

This page has notebook examples for multi-GPU distributed training using AI Runtime. These examples demonstrate how to scale training across multiple GPUs and nodes for improved performance.

note

Multi-GPU distributed training is supported on H100 GPUs.

Choose your parallelism technique

When scaling your model training across multiple GPUs, choosing the right parallelism technique depends on your model size, available GPU memory, and performance requirements.

Technique

When to use

DDP (Distributed Data Parallel)

Full model fits in single GPU memory; need to scale data throughput

FSDP (Fully Sharded Data Parallel)

Very large models that don't fit in single GPU memory

DeepSpeed ZeRO

Large models with advanced memory optimization needs

For detailed information about each technique, see DDP, FSDP, and DeepSpeed.

Example notebooks by technique and framework

The following table organizes example notebooks by the framework/library you're using and the parallelism technique applied. Multiple notebooks may appear in a single cell.

Get started

Use the following tutorials to get started with the serverless GPU Python library for distributed training:

Tutorial

Description

AI Runtime with H100 GPUs

Learn how to use Databricks AI Runtime with H100 accelerators to run distributed GPU workloads using the serverless_gpu Python library.