Upload modeling_fastesm.py with huggingface_hub
Browse files- modeling_fastesm.py +6 -3
modeling_fastesm.py
CHANGED
@@ -55,7 +55,12 @@ class FastEsmConfig(PretrainedConfig):
|
|
55 |
emb_layer_norm_before: bool = None,
|
56 |
**kwargs,
|
57 |
):
|
58 |
-
super().__init__(
|
|
|
|
|
|
|
|
|
|
|
59 |
|
60 |
self.vocab_size = vocab_size
|
61 |
self.hidden_size = hidden_size
|
@@ -945,8 +950,6 @@ class FastEsmModel(FastEsmPreTrainedModel, EmbeddingMixin):
|
|
945 |
|
946 |
|
947 |
class FastEsmForMaskedLM(FastEsmPreTrainedModel, EmbeddingMixin):
|
948 |
-
_tied_weights_keys = ["lm_head.decoder.weight"]
|
949 |
-
|
950 |
def __init__(self, config, **kwargs):
|
951 |
FastEsmPreTrainedModel.__init__(self, config, **kwargs)
|
952 |
self.esm = FAST_ESM_ENCODER(config, add_pooling_layer=False)
|
|
|
55 |
emb_layer_norm_before: bool = None,
|
56 |
**kwargs,
|
57 |
):
|
58 |
+
super().__init__(
|
59 |
+
pad_token_id=pad_token_id,
|
60 |
+
mask_token_id=mask_token_id,
|
61 |
+
tie_word_embeddings=False,
|
62 |
+
**kwargs,
|
63 |
+
)
|
64 |
|
65 |
self.vocab_size = vocab_size
|
66 |
self.hidden_size = hidden_size
|
|
|
950 |
|
951 |
|
952 |
class FastEsmForMaskedLM(FastEsmPreTrainedModel, EmbeddingMixin):
|
|
|
|
|
953 |
def __init__(self, config, **kwargs):
|
954 |
FastEsmPreTrainedModel.__init__(self, config, **kwargs)
|
955 |
self.esm = FAST_ESM_ENCODER(config, add_pooling_layer=False)
|