DistributedRunner
class optimum.habana.distributed.DistributedRunner
< source >( command_list: List = [] world_size: int = 1 hostfile: Union = None use_mpi: bool = False use_deepspeed: bool = False master_port: int = 29500 use_env: bool = False map_by: bool = 'socket' multi_hls = None )
Set up training/inference hardware configurations and run distributed commands.
Multi-node configuration setup for DeepSpeed.
Single-card setup.
Single-node multi-card configuration setup.
Single-node multi-card configuration setup for DeepSpeed.
Single-node multi-card configuration setup for mpirun.
Returns the master address to use for multi-node runs with DeepSpeed. Directly inspired from https://github.com/microsoft/DeepSpeed/blob/316c4a43e0802a979951ee17f735daf77ea9780f/deepspeed/autotuning/utils.py#L145.
Runs the desired command with configuration specified by the user.