Optimum documentation

DistributedRunner

You are viewing v1.17.1 version. A newer version v1.19.0 is available.
Hugging Face's logo
Join the Hugging Face community

and get access to the augmented documentation experience

to get started

DistributedRunner

class optimum.habana.distributed.DistributedRunner

< >

( command_list: List = [] world_size: int = 1 hostfile: Union = None use_mpi: bool = False use_deepspeed: bool = False master_port: int = 29500 use_env: bool = False map_by: bool = 'socket' multi_hls = None )

Set up training/inference hardware configurations and run distributed commands.

create_multi_node_setup

< >

( )

Multi-node configuration setup for DeepSpeed.

create_single_card_setup

< >

( use_deepspeed = False )

Single-card setup.

create_single_node_setup

< >

( )

Single-node multi-card configuration setup.

create_single_node_setup_deepspeed

< >

( )

Single-node multi-card configuration setup for DeepSpeed.

create_single_node_setup_mpirun

< >

( )

Single-node multi-card configuration setup for mpirun.

process_hostfile

< >

( ) str

Returns

str

address of the master node.

Returns the master address to use for multi-node runs with DeepSpeed. Directly inspired from https://github.com/microsoft/DeepSpeed/blob/316c4a43e0802a979951ee17f735daf77ea9780f/deepspeed/autotuning/utils.py#L145.

run

< >

( )

Runs the desired command with configuration specified by the user.