teddy-f-47
commited on
Commit
•
eeb401c
1
Parent(s):
d0f5976
Update configuration_phi.py
Browse files- configuration_phi.py +9 -9
configuration_phi.py
CHANGED
@@ -116,13 +116,13 @@ class PhiConfig(PretrainedConfig):
|
|
116 |
|
117 |
def __init__(
|
118 |
self,
|
119 |
-
vocab_size=
|
120 |
-
hidden_size=
|
121 |
-
intermediate_size=
|
122 |
-
num_hidden_layers=
|
123 |
num_attention_heads=32,
|
124 |
-
num_key_value_heads=
|
125 |
-
resid_pdrop=0.
|
126 |
embd_pdrop=0.0,
|
127 |
attention_dropout=0.0,
|
128 |
hidden_act="gelu_new",
|
@@ -133,10 +133,10 @@ class PhiConfig(PretrainedConfig):
|
|
133 |
tie_word_embeddings=False,
|
134 |
rope_theta=10000.0,
|
135 |
rope_scaling=None,
|
136 |
-
partial_rotary_factor=0.
|
137 |
qk_layernorm=False,
|
138 |
-
bos_token_id=
|
139 |
-
eos_token_id=
|
140 |
**kwargs,
|
141 |
):
|
142 |
self.vocab_size = vocab_size
|
|
|
116 |
|
117 |
def __init__(
|
118 |
self,
|
119 |
+
vocab_size=50295,
|
120 |
+
hidden_size=2560,
|
121 |
+
intermediate_size=10240,
|
122 |
+
num_hidden_layers=32,
|
123 |
num_attention_heads=32,
|
124 |
+
num_key_value_heads=32,
|
125 |
+
resid_pdrop=0.1,
|
126 |
embd_pdrop=0.0,
|
127 |
attention_dropout=0.0,
|
128 |
hidden_act="gelu_new",
|
|
|
133 |
tie_word_embeddings=False,
|
134 |
rope_theta=10000.0,
|
135 |
rope_scaling=None,
|
136 |
+
partial_rotary_factor=0.4,
|
137 |
qk_layernorm=False,
|
138 |
+
bos_token_id=0,
|
139 |
+
eos_token_id=0,
|
140 |
**kwargs,
|
141 |
):
|
142 |
self.vocab_size = vocab_size
|