teddy-f-47 commited on
Commit
eeb401c
1 Parent(s): d0f5976

Update configuration_phi.py

Browse files
Files changed (1) hide show
  1. configuration_phi.py +9 -9
configuration_phi.py CHANGED
@@ -116,13 +116,13 @@ class PhiConfig(PretrainedConfig):
116
 
117
  def __init__(
118
  self,
119
- vocab_size=51200,
120
- hidden_size=2048,
121
- intermediate_size=8192,
122
- num_hidden_layers=24,
123
  num_attention_heads=32,
124
- num_key_value_heads=None,
125
- resid_pdrop=0.0,
126
  embd_pdrop=0.0,
127
  attention_dropout=0.0,
128
  hidden_act="gelu_new",
@@ -133,10 +133,10 @@ class PhiConfig(PretrainedConfig):
133
  tie_word_embeddings=False,
134
  rope_theta=10000.0,
135
  rope_scaling=None,
136
- partial_rotary_factor=0.5,
137
  qk_layernorm=False,
138
- bos_token_id=1,
139
- eos_token_id=2,
140
  **kwargs,
141
  ):
142
  self.vocab_size = vocab_size
 
116
 
117
  def __init__(
118
  self,
119
+ vocab_size=50295,
120
+ hidden_size=2560,
121
+ intermediate_size=10240,
122
+ num_hidden_layers=32,
123
  num_attention_heads=32,
124
+ num_key_value_heads=32,
125
+ resid_pdrop=0.1,
126
  embd_pdrop=0.0,
127
  attention_dropout=0.0,
128
  hidden_act="gelu_new",
 
133
  tie_word_embeddings=False,
134
  rope_theta=10000.0,
135
  rope_scaling=None,
136
+ partial_rotary_factor=0.4,
137
  qk_layernorm=False,
138
+ bos_token_id=0,
139
+ eos_token_id=0,
140
  **kwargs,
141
  ):
142
  self.vocab_size = vocab_size