gugarosa commited on
Commit
d38e6f9
1 Parent(s): 8091327

Update modeling_mixformer_sequential.py

Browse files

Removes print regarding attention_mask to prevent excessive information from being logged.

Files changed (1) hide show
  1. modeling_mixformer_sequential.py +0 -3
modeling_mixformer_sequential.py CHANGED
@@ -756,9 +756,6 @@ class MixFormerSequentialForCausalLM(MixFormerSequentialPreTrainedModel):
756
  labels: Optional[torch.LongTensor] = None,
757
  **kwargs,
758
  ) -> CausalLMOutputWithPast:
759
- if attention_mask is not None and self.training:
760
- print("`attention_mask` is not supported during training. Using it might lead to unexpected results.")
761
-
762
  if past_key_values is None and attention_mask is None:
763
  lm_logits = self.layers(input_ids)
764
  else:
 
756
  labels: Optional[torch.LongTensor] = None,
757
  **kwargs,
758
  ) -> CausalLMOutputWithPast:
 
 
 
759
  if past_key_values is None and attention_mask is None:
760
  lm_logits = self.layers(input_ids)
761
  else: