Fill-Mask
Transformers
PyTorch
Safetensors
bert
custom_code
robinzixuan commited on
Commit
9b45691
·
verified ·
1 Parent(s): 08c4c3a

Update modeling_bert.py

Browse files
Files changed (1) hide show
  1. modeling_bert.py +1 -1
modeling_bert.py CHANGED
@@ -537,7 +537,7 @@ class BertOutEffHop(nn.Module):
537
 
538
  # Normalize the attention scores to probabilities.
539
  attention_probs = softmax_1(attention_scores, dim=-1)
540
-
541
  # This is actually dropping out entire tokens to attend to, which might
542
  # seem a bit unusual, but is taken from the original Transformer paper.
543
  attention_probs = self.dropout(attention_probs)
 
537
 
538
  # Normalize the attention scores to probabilities.
539
  attention_probs = softmax_1(attention_scores, dim=-1)
540
+ print(softmax_1)
541
  # This is actually dropping out entire tokens to attend to, which might
542
  # seem a bit unusual, but is taken from the original Transformer paper.
543
  attention_probs = self.dropout(attention_probs)