robinzixuan
commited on
Update modeling_bert.py
Browse files- modeling_bert.py +1 -1
modeling_bert.py
CHANGED
@@ -537,7 +537,7 @@ class BertOutEffHop(nn.Module):
|
|
537 |
|
538 |
# Normalize the attention scores to probabilities.
|
539 |
attention_probs = softmax_1(attention_scores, dim=-1)
|
540 |
-
|
541 |
# This is actually dropping out entire tokens to attend to, which might
|
542 |
# seem a bit unusual, but is taken from the original Transformer paper.
|
543 |
attention_probs = self.dropout(attention_probs)
|
|
|
537 |
|
538 |
# Normalize the attention scores to probabilities.
|
539 |
attention_probs = softmax_1(attention_scores, dim=-1)
|
540 |
+
print(softmax_1)
|
541 |
# This is actually dropping out entire tokens to attend to, which might
|
542 |
# seem a bit unusual, but is taken from the original Transformer paper.
|
543 |
attention_probs = self.dropout(attention_probs)
|