Skip to main content

Distributed training using DeepSpeed

Beta

This feature is in Beta.

This page has notebook examples for distributed training using DeepSpeed on Serverless GPU compute.

Supervised fine-tuning using DeepSpeed and TRL

This notebook demonstrates how to use the Serverless GPU Python API to run supervised fine-tuning (SFT) using the Transformer Reinforcement Learning (TRL) library with DeepSpeed ZeRO Stage 3 optimization.

Notebook

Open notebook in new tab