Optimum documentation

DeepSpeed for HPUs

You are viewing main version, which requires installation from source. If you'd like regular pip install, checkout the latest stable version (v1.7.1).
Join the Hugging Face community

to get started

# DeepSpeed for HPUs

DeepSpeed enables you to fit and train larger models on HPUs thanks to various optimizations described in the ZeRO paper. In particular, you can use the two following ZeRO configurations that have been validated to be fully functioning with Gaudi:

• ZeRO-1: partitions the optimizer states across processes.
• ZeRO-2: partitions the optimizer states + gradients across processes.

These configurations are fully compatible with Habana Mixed Precision and can thus be used to train your model in bf16 precision.

## Setup

To use DeepSpeed on Gaudi, you need to install Optimum Habana and Habana’s DeepSpeed fork with:

pip install optimum[habana]
pip install git+https://github.com/HabanaAI/DeepSpeed.git@1.8.0

## Using DeepSpeed with Optimum Habana

The GaudiTrainer allows using DeepSpeed as easily as the Transformers Trainer. This can be done in 3 steps:

1. A DeepSpeed configuration has to be defined.
2. The deepspeed training argument enables to specify the path to the DeepSpeed configuration.
3. The deepspeed launcher must be used to run your script.

These steps are detailed below. A comprehensive guide about how to use DeepSpeed with the Transformers Trainer is also available here.

### DeepSpeed configuration

The DeepSpeed configuration to use is passed through a JSON file and enables you to choose the optimizations to apply. Here is an example for applying ZeRO-2 optimizations and bf16 precision:

{
"steps_per_print": 64,
"train_batch_size": "auto",
"train_micro_batch_size_per_gpu": "auto",
"bf16": {
"enabled": true
},
"zero_optimization": {
"stage": 2,
"overlap_comm": false,
"reduce_scatter": false,
}
}

The special value "auto" enables to automatically get the correct or most efficient value. You can also specify the values yourself but, if you do so, you should be careful not to have conflicting values with your training arguments. It is strongly advised to read this section in the Transformers documentation to completely understand how this works.

Other examples of configurations for HPUs are proposed here by Habana.

The Transformers documentation explains how to write a configuration from scratch very well. A more complete description of all configuration possibilities is available here.

### The deepspeed training argument

To use DeepSpeed, you must specify deespeed=path_to_my_deepspeed_configuration in your GaudiTrainingArguments instance:

training_args = GaudiTrainingArguments(
# my usual training arguments...
use_habana=True,
use_lazy_mode=True,
gaudi_config_name=path_to_my_gaudi_config,
deepspeed=path_to_my_deepspeed_config,
)

This argument both indicates that DeepSpeed should be used and points to your DeepSpeed configuration.

Finally, there are two possible ways to launch your script:

1. Using the gaudi_spawn.py script:
python gaudi_spawn.py \
--world_size number_of_hpu_you_have --use_deepspeed \
path_to_script.py --args1 --args2 ... --argsN \
--deepspeed path_to_deepspeed_config

where --argX is an argument of the script to run with DeepSpeed.

1. Using the DistributedRunner directly in code:
from optimum.habana.distributed import DistributedRunner
from optimum.utils import logging

world_size=8 # Number of HPUs to use (1 or 8)

# define distributed runner
distributed_runner = DistributedRunner(
command_list=["scripts/train.py --args1 --args2 ... --argsN --deepspeed path_to_deepspeed_config"],
world_size=world_size,
use_deepspeed=True,
)

# start job
ret_code = distributed_runner.run()

You should set "use_fused_adam": false in your Gaudi configuration because it is not compatible with DeepSpeed yet.