lhallee commited on
Commit
601fa2b
·
verified ·
1 Parent(s): 1ffa8b1

Upload modeling_fastesm.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. modeling_fastesm.py +6 -3
modeling_fastesm.py CHANGED
@@ -55,7 +55,12 @@ class FastEsmConfig(PretrainedConfig):
55
  emb_layer_norm_before: bool = None,
56
  **kwargs,
57
  ):
58
- super().__init__(pad_token_id=pad_token_id, mask_token_id=mask_token_id, **kwargs)
 
 
 
 
 
59
 
60
  self.vocab_size = vocab_size
61
  self.hidden_size = hidden_size
@@ -945,8 +950,6 @@ class FastEsmModel(FastEsmPreTrainedModel, EmbeddingMixin):
945
 
946
 
947
  class FastEsmForMaskedLM(FastEsmPreTrainedModel, EmbeddingMixin):
948
- _tied_weights_keys = ["lm_head.decoder.weight"]
949
-
950
  def __init__(self, config, **kwargs):
951
  FastEsmPreTrainedModel.__init__(self, config, **kwargs)
952
  self.esm = FAST_ESM_ENCODER(config, add_pooling_layer=False)
 
55
  emb_layer_norm_before: bool = None,
56
  **kwargs,
57
  ):
58
+ super().__init__(
59
+ pad_token_id=pad_token_id,
60
+ mask_token_id=mask_token_id,
61
+ tie_word_embeddings=False,
62
+ **kwargs,
63
+ )
64
 
65
  self.vocab_size = vocab_size
66
  self.hidden_size = hidden_size
 
950
 
951
 
952
  class FastEsmForMaskedLM(FastEsmPreTrainedModel, EmbeddingMixin):
 
 
953
  def __init__(self, config, **kwargs):
954
  FastEsmPreTrainedModel.__init__(self, config, **kwargs)
955
  self.esm = FAST_ESM_ENCODER(config, add_pooling_layer=False)