kushaltatariya commited on
Commit
4030f49
1 Parent(s): 6bf26b4

Upload tokenizer

Browse files
special_tokens_map.json CHANGED
@@ -1,7 +1,9 @@
1
  {
2
- "bos_token": "<s>",
3
- "eos_token": "</s>",
4
- "mask_token": "<mask>",
5
- "pad_token": "<pad>",
6
- "unk_token": "<unk>"
 
 
7
  }
 
1
  {
2
+ "bos_token": "[BOS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[EOS]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": "[UNK]"
9
  }
tokenization.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff30dc12b46871e024d8592e6d360e0bcd41733a2ffd3dd30cd2aa393d6b751b
3
- size 248649
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69db0f27962663caa548cf3f946a029813d752e2f4c2be4946b0547db10bd7b0
3
+ size 248674
tokenizer.json CHANGED
@@ -5,7 +5,7 @@
5
  "added_tokens": [
6
  {
7
  "id": 0,
8
- "content": "<pad>",
9
  "single_word": false,
10
  "lstrip": false,
11
  "rstrip": false,
@@ -14,7 +14,7 @@
14
  },
15
  {
16
  "id": 1,
17
- "content": "<unk>",
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
@@ -23,7 +23,7 @@
23
  },
24
  {
25
  "id": 2,
26
- "content": "<s>",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
@@ -32,7 +32,7 @@
32
  },
33
  {
34
  "id": 3,
35
- "content": "</s>",
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
@@ -41,7 +41,25 @@
41
  },
42
  {
43
  "id": 4,
44
- "content": "<mask>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
45
  "single_word": false,
46
  "lstrip": false,
47
  "rstrip": false,
@@ -88,23 +106,31 @@
88
  "unk_id": 1,
89
  "vocab": [
90
  [
91
- "<pad>",
 
 
 
 
 
 
 
 
92
  0.0
93
  ],
94
  [
95
- "<unk>",
96
  0.0
97
  ],
98
  [
99
- "<s>",
100
  0.0
101
  ],
102
  [
103
- "</s>",
104
  0.0
105
  ],
106
  [
107
- "<mask>",
108
  0.0
109
  ],
110
  [
@@ -3163,14 +3189,6 @@
3163
  "▁agbaza",
3164
  -9.1220121383667
3165
  ],
3166
- [
3167
- "aguidi",
3168
- -9.122014045715332
3169
- ],
3170
- [
3171
- "▁iglej",
3172
- -9.122014999389648
3173
- ],
3174
  [
3175
  "’",
3176
  -9.184510231018066
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
8
+ "content": "[PAD]",
9
  "single_word": false,
10
  "lstrip": false,
11
  "rstrip": false,
 
14
  },
15
  {
16
  "id": 1,
17
+ "content": "[UNK]",
18
  "single_word": false,
19
  "lstrip": false,
20
  "rstrip": false,
 
23
  },
24
  {
25
  "id": 2,
26
+ "content": "[BOS]",
27
  "single_word": false,
28
  "lstrip": false,
29
  "rstrip": false,
 
32
  },
33
  {
34
  "id": 3,
35
+ "content": "[EOS]",
36
  "single_word": false,
37
  "lstrip": false,
38
  "rstrip": false,
 
41
  },
42
  {
43
  "id": 4,
44
+ "content": "[CLS]",
45
+ "single_word": false,
46
+ "lstrip": false,
47
+ "rstrip": false,
48
+ "normalized": false,
49
+ "special": true
50
+ },
51
+ {
52
+ "id": 5,
53
+ "content": "[SEP]",
54
+ "single_word": false,
55
+ "lstrip": false,
56
+ "rstrip": false,
57
+ "normalized": false,
58
+ "special": true
59
+ },
60
+ {
61
+ "id": 6,
62
+ "content": "[MASK]",
63
  "single_word": false,
64
  "lstrip": false,
65
  "rstrip": false,
 
106
  "unk_id": 1,
107
  "vocab": [
108
  [
109
+ "[PAD]",
110
+ 0.0
111
+ ],
112
+ [
113
+ "[UNK]",
114
+ 0.0
115
+ ],
116
+ [
117
+ "[BOS]",
118
  0.0
119
  ],
120
  [
121
+ "[EOS]",
122
  0.0
123
  ],
124
  [
125
+ "[CLS]",
126
  0.0
127
  ],
128
  [
129
+ "[SEP]",
130
  0.0
131
  ],
132
  [
133
+ "[MASK]",
134
  0.0
135
  ],
136
  [
 
3189
  "▁agbaza",
3190
  -9.1220121383667
3191
  ],
 
 
 
 
 
 
 
 
3192
  [
3193
  "’",
3194
  -9.184510231018066
tokenizer_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
7
- "content": "<pad>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
@@ -12,7 +12,7 @@
12
  "special": true
13
  },
14
  "1": {
15
- "content": "<unk>",
16
  "lstrip": false,
17
  "normalized": false,
18
  "rstrip": false,
@@ -20,7 +20,7 @@
20
  "special": true
21
  },
22
  "2": {
23
- "content": "<s>",
24
  "lstrip": false,
25
  "normalized": false,
26
  "rstrip": false,
@@ -28,7 +28,7 @@
28
  "special": true
29
  },
30
  "3": {
31
- "content": "</s>",
32
  "lstrip": false,
33
  "normalized": false,
34
  "rstrip": false,
@@ -36,7 +36,23 @@
36
  "special": true
37
  },
38
  "4": {
39
- "content": "<mask>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  "lstrip": false,
41
  "normalized": false,
42
  "rstrip": false,
@@ -44,20 +60,24 @@
44
  "special": false
45
  }
46
  },
47
- "bos_token": "<s>",
48
  "bos_token_id": 2,
49
  "clean_up_tokenization_spaces": false,
50
- "eos_token": "</s>",
 
 
51
  "eos_token_id": 3,
52
  "legacy": true,
53
- "mask_token": "<mask>",
54
- "mask_token_id": 4,
55
  "model_max_length": 1000000000000000019884624838656,
56
- "pad_token": "<pad>",
57
  "pad_token_id": 0,
 
 
58
  "sp_model_kwargs": {},
59
  "spaces_between_special_tokens": false,
60
  "tokenizer_class": "HfSentencePieceTokenizer",
61
- "unk_token": "<unk>",
62
  "unk_token_id": 1
63
  }
 
4
  "add_prefix_space": true,
5
  "added_tokens_decoder": {
6
  "0": {
7
+ "content": "[PAD]",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
 
12
  "special": true
13
  },
14
  "1": {
15
+ "content": "[UNK]",
16
  "lstrip": false,
17
  "normalized": false,
18
  "rstrip": false,
 
20
  "special": true
21
  },
22
  "2": {
23
+ "content": "[BOS]",
24
  "lstrip": false,
25
  "normalized": false,
26
  "rstrip": false,
 
28
  "special": true
29
  },
30
  "3": {
31
+ "content": "[EOS]",
32
  "lstrip": false,
33
  "normalized": false,
34
  "rstrip": false,
 
36
  "special": true
37
  },
38
  "4": {
39
+ "content": "[CLS]",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ },
46
+ "5": {
47
+ "content": "[SEP]",
48
+ "lstrip": false,
49
+ "normalized": false,
50
+ "rstrip": false,
51
+ "single_word": false,
52
+ "special": true
53
+ },
54
+ "6": {
55
+ "content": "[MASK]",
56
  "lstrip": false,
57
  "normalized": false,
58
  "rstrip": false,
 
60
  "special": false
61
  }
62
  },
63
+ "bos_token": "[BOS]",
64
  "bos_token_id": 2,
65
  "clean_up_tokenization_spaces": false,
66
+ "cls_token": "[CLS]",
67
+ "cls_token_id": 4,
68
+ "eos_token": "[EOS]",
69
  "eos_token_id": 3,
70
  "legacy": true,
71
+ "mask_token": "[MASK]",
72
+ "mask_token_id": 6,
73
  "model_max_length": 1000000000000000019884624838656,
74
+ "pad_token": "[PAD]",
75
  "pad_token_id": 0,
76
+ "sep_token": "[SEP]",
77
+ "sep_token_id": 5,
78
  "sp_model_kwargs": {},
79
  "spaces_between_special_tokens": false,
80
  "tokenizer_class": "HfSentencePieceTokenizer",
81
+ "unk_token": "[UNK]",
82
  "unk_token_id": 1
83
  }