ammarnasr commited on
Commit
e39a192
1 Parent(s): a779b76

Upload tokenizer

Browse files
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "[PAD]": 49152
3
+ }
special_tokens_map.json CHANGED
@@ -32,7 +32,13 @@
32
  "rstrip": false,
33
  "single_word": false
34
  },
35
- "pad_token": "<|endoftext|>",
 
 
 
 
 
 
36
  "unk_token": {
37
  "content": "<|endoftext|>",
38
  "lstrip": false,
 
32
  "rstrip": false,
33
  "single_word": false
34
  },
35
+ "pad_token": {
36
+ "content": "[PAD]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false
41
+ },
42
  "unk_token": {
43
  "content": "<|endoftext|>",
44
  "lstrip": false,
tokenizer.json CHANGED
@@ -2,11 +2,18 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 512,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
- "padding": null,
 
 
 
 
 
 
 
10
  "added_tokens": [
11
  {
12
  "id": 0,
@@ -160,6 +167,15 @@
160
  "rstrip": false,
161
  "normalized": false,
162
  "special": true
 
 
 
 
 
 
 
 
 
163
  }
164
  ],
165
  "normalizer": null,
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 256,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
9
+ "padding": {
10
+ "strategy": "BatchLongest",
11
+ "direction": "Left",
12
+ "pad_to_multiple_of": null,
13
+ "pad_id": 49152,
14
+ "pad_type_id": 0,
15
+ "pad_token": "[PAD]"
16
+ },
17
  "added_tokens": [
18
  {
19
  "id": 0,
 
167
  "rstrip": false,
168
  "normalized": false,
169
  "special": true
170
+ },
171
+ {
172
+ "id": 49152,
173
+ "content": "[PAD]",
174
+ "single_word": false,
175
+ "lstrip": false,
176
+ "rstrip": false,
177
+ "normalized": false,
178
+ "special": true
179
  }
180
  ],
181
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -136,6 +136,14 @@
136
  "rstrip": false,
137
  "single_word": false,
138
  "special": true
 
 
 
 
 
 
 
 
139
  }
140
  },
141
  "additional_special_tokens": [
@@ -160,8 +168,8 @@
160
  "bos_token": "<|endoftext|>",
161
  "clean_up_tokenization_spaces": false,
162
  "eos_token": "<|endoftext|>",
163
- "model_max_length": 1000000000000000019884624838656,
164
- "pad_token": "<|endoftext|>",
165
  "tokenizer_class": "GPT2Tokenizer",
166
  "unk_token": "<|endoftext|>",
167
  "vocab_size": 49152
 
136
  "rstrip": false,
137
  "single_word": false,
138
  "special": true
139
+ },
140
+ "49152": {
141
+ "content": "[PAD]",
142
+ "lstrip": false,
143
+ "normalized": false,
144
+ "rstrip": false,
145
+ "single_word": false,
146
+ "special": true
147
  }
148
  },
149
  "additional_special_tokens": [
 
168
  "bos_token": "<|endoftext|>",
169
  "clean_up_tokenization_spaces": false,
170
  "eos_token": "<|endoftext|>",
171
+ "model_max_length": 256,
172
+ "pad_token": "[PAD]",
173
  "tokenizer_class": "GPT2Tokenizer",
174
  "unk_token": "<|endoftext|>",
175
  "vocab_size": 49152