Text Generation
Transformers
Safetensors
lola_v1
custom_code
neo-nlp-dev commited on
Commit
0507803
1 Parent(s): bbc824c

update bos and eos config

Browse files
config.json CHANGED
@@ -8,9 +8,9 @@
8
  "AutoConfig": "configuration_lola_gpt2.LOLAConfig",
9
  "AutoModelForCausalLM": "modeling_lola_gpt2.LOLALMHeadModel"
10
  },
11
- "bos_token_id": 100095,
12
  "embd_pdrop": 0.1,
13
- "eos_token_id": 100095,
14
  "initializer_range": 0.02,
15
  "layer_norm_epsilon": 1e-05,
16
  "model_type": "lola_v1",
 
8
  "AutoConfig": "configuration_lola_gpt2.LOLAConfig",
9
  "AutoModelForCausalLM": "modeling_lola_gpt2.LOLALMHeadModel"
10
  },
11
+ "bos_token_id": 0,
12
  "embd_pdrop": 0.1,
13
+ "eos_token_id": 5,
14
  "initializer_range": 0.02,
15
  "layer_norm_epsilon": 1e-05,
16
  "model_type": "lola_v1",
configuration_lola_gpt2.py CHANGED
@@ -22,7 +22,7 @@ class LOLAConfig(PretrainedConfig):
22
 
23
  def __init__(
24
  self,
25
- vocab_size=100096,
26
  n_positions=2048,
27
  n_embd=2048,
28
  n_layer=24,
@@ -41,8 +41,8 @@ class LOLAConfig(PretrainedConfig):
41
  summary_first_dropout=0.1,
42
  scale_attn_weights=True,
43
  use_cache=True,
44
- bos_token_id=100095,
45
- eos_token_id=100095,
46
  scale_attn_by_inverse_layer_idx=False,
47
  reorder_and_upcast_attn=False,
48
  num_experts=16,
 
22
 
23
  def __init__(
24
  self,
25
+ vocab_size=100096, # Padded vocab size, actual size is 100000
26
  n_positions=2048,
27
  n_embd=2048,
28
  n_layer=24,
 
41
  summary_first_dropout=0.1,
42
  scale_attn_weights=True,
43
  use_cache=True,
44
+ bos_token_id=0,
45
+ eos_token_id=5,
46
  scale_attn_by_inverse_layer_idx=False,
47
  reorder_and_upcast_attn=False,
48
  num_experts=16,
generation_config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_from_model_config": true,
3
- "bos_token_id": 100095,
4
- "eos_token_id": 100095,
5
  "transformers_version": "4.39.1"
6
  }
 
1
  {
2
  "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "eos_token_id": 5,
5
  "transformers_version": "4.39.1"
6
  }