itay-levy commited on
Commit
46481bc
1 Parent(s): 9ac48ff

add padding_mask

Browse files
Files changed (1) hide show
  1. modeling_decilm.py +1 -0
modeling_decilm.py CHANGED
@@ -62,6 +62,7 @@ class DeciLMAttention(LlamaAttention):
62
  past_key_value: Optional[Tuple[torch.Tensor]] = None,
63
  output_attentions: bool = False,
64
  use_cache: bool = False,
 
65
  ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
66
  bsz, q_len, _ = hidden_states.size()
67
  if past_key_value is None:
 
62
  past_key_value: Optional[Tuple[torch.Tensor]] = None,
63
  output_attentions: bool = False,
64
  use_cache: bool = False,
65
+ padding_mask: Optional[torch.LongTensor] = None,
66
  ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
67
  bsz, q_len, _ = hidden_states.size()
68
  if past_key_value is None: