| | |
| | |
| |
|
| | |
| |
|
| | import torch |
| |
|
| | from typing import Any, Tuple |
| | from torch import Tensor |
| | from torch.nn import Module |
| |
|
| | import deepspeed.comm as dist |
| |
|
| |
|
| | def single_all_to_all(input, scatter_idx, gather_idx, group): |
| | seq_world_size = dist.get_world_size(group) |
| | inp_shape = list(input.shape) |
| | inp_shape[scatter_idx] = inp_shape[scatter_idx] // seq_world_size |
| | if scatter_idx < 2: |
| | input_t = input.reshape( |
| | [seq_world_size, inp_shape[scatter_idx]] + \ |
| | inp_shape[scatter_idx + 1:] |
| | ).contiguous() |
| | else: |
| | |
| | input_t = input.reshape( |
| | [-1, seq_world_size, inp_shape[scatter_idx]] + \ |
| | inp_shape[scatter_idx + 1:] |
| | ).transpose(0, 1).contiguous() |
| |
|
| | output = torch.empty_like(input_t) |
| | dist.all_to_all_single(output, input_t, group=group) |
| |
|
| | |
| | if scatter_idx < 2: |
| | output = output.transpose(0, 1).contiguous() |
| |
|
| | return output.reshape( |
| | inp_shape[: gather_idx] + \ |
| | [inp_shape[gather_idx] * seq_world_size,] + \ |
| | inp_shape[gather_idx + 1:]).contiguous() |
| |
|
| |
|
| | class _SeqAllToAll(torch.autograd.Function): |
| |
|
| | @staticmethod |
| | def forward(ctx: Any, group: dist.ProcessGroup, input: Tensor, scatter_idx: int, gather_idx: int) -> Tensor: |
| |
|
| | ctx.group = group |
| | ctx.scatter_idx = scatter_idx |
| | ctx.gather_idx = gather_idx |
| |
|
| | return single_all_to_all(input, scatter_idx, gather_idx, group) |
| |
|
| | @staticmethod |
| | def backward(ctx: Any, *grad_output: Tensor) -> Tuple[None, Tensor, None, None]: |
| | return (None, _SeqAllToAll.apply(ctx.group, *grad_output, ctx.gather_idx, ctx.scatter_idx), None, None) |
| |
|
| |
|
| | class DistributedAttention(torch.nn.Module): |
| | """Initialization. |
| | |
| | Arguments: |
| | local_attention (Module): local attention with q,k,v |
| | sequence_process_group (ProcessGroup): sequence parallel process group |
| | scatter_idx (int): scatter_idx for all2all comm |
| | gather_idx (int): gather_idx for all2all comm |
| | """ |
| |
|
| | def __init__( |
| | self, |
| | local_attention: Module, |
| | sequence_process_group: dist.ProcessGroup, |
| | scatter_idx: int = 2, |
| | gather_idx: int = 0, |
| | ) -> None: |
| |
|
| | super(DistributedAttention, self).__init__() |
| | self.local_attn = local_attention |
| | self.spg = sequence_process_group |
| | self.scatter_idx = scatter_idx |
| | self.gather_idx = gather_idx |
| |
|
| | def forward(self, query: Tensor, key: Tensor, value: Tensor, *args: Any) -> Tensor: |
| | """ forward |
| | |
| | Arguments: |
| | query (Tensor): query input to the layer |
| | key (Tensor): key input to the layer |
| | value (Tensor): value input to the layer |
| | args: other args |
| | |
| | Returns: |
| | * output (Tensor): context output |
| | """ |
| | |
| | |
| | |
| | query_layer = _SeqAllToAll.apply(self.spg, query, self.scatter_idx, self.gather_idx) |
| | key_layer = _SeqAllToAll.apply(self.spg, key, self.scatter_idx, self.gather_idx) |
| | value_layer = _SeqAllToAll.apply(self.spg, value, self.scatter_idx, self.gather_idx) |
| |
|
| | |
| | context_layer = self.local_attn(query_layer, key_layer, value_layer, *args) |
| |
|
| | output = _SeqAllToAll.apply(self.spg, context_layer, self.gather_idx, self.scatter_idx) |
| |
|
| | |
| | return output |
| |
|