ehristoforu commited on
Commit
bf80da3
1 Parent(s): 466d2b4

Upload 4 files

Browse files
tokenizer_2/merges.txt CHANGED
@@ -48892,4 +48892,4 @@ si gue</w>
48892
  fun ko
48893
  rend ous</w>
48894
  ðŁĴ ij</w>
48895
- jeky ll</w>
 
48892
  fun ko
48893
  rend ous</w>
48894
  ðŁĴ ij</w>
48895
+ jeky ll</w>
tokenizer_2/special_tokens_map.json CHANGED
@@ -1 +1,24 @@
1
- {"bos_token": {"content": "<|startoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "eos_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "unk_token": {"content": "<|endoftext|>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, "pad_token": "<|endoftext|>"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "!",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer_2/tokenizer_config.json CHANGED
@@ -1,34 +1,38 @@
1
  {
2
- "unk_token": {
3
- "content": "<|endoftext|>",
4
- "single_word": false,
5
- "lstrip": false,
6
- "rstrip": false,
7
- "normalized": true,
8
- "__type": "AddedToken"
 
 
9
  },
10
- "bos_token": {
11
- "content": "<|startoftext|>",
12
- "single_word": false,
13
- "lstrip": false,
14
- "rstrip": false,
15
- "normalized": true,
16
- "__type": "AddedToken"
17
  },
18
- "eos_token": {
19
- "content": "<|endoftext|>",
20
- "single_word": false,
21
- "lstrip": false,
22
- "rstrip": false,
23
- "normalized": true,
24
- "__type": "AddedToken"
25
- },
26
- "pad_token": "<|endoftext|>",
27
- "add_prefix_space": false,
28
- "errors": "replace",
29
- "do_lower_case": true,
30
- "name_or_path": "openai/clip-vit-base-patch32",
31
- "model_max_length": 77,
32
- "special_tokens_map_file": "./special_tokens_map.json",
33
- "tokenizer_class": "CLIPTokenizer"
34
- }
 
 
 
1
  {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "!",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
  },
12
+ "49406": {
13
+ "content": "<|startoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
  },
20
+ "49407": {
21
+ "content": "<|endoftext|>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ }
28
+ },
29
+ "bos_token": "<|startoftext|>",
30
+ "clean_up_tokenization_spaces": true,
31
+ "do_lower_case": true,
32
+ "eos_token": "<|endoftext|>",
33
+ "errors": "replace",
34
+ "model_max_length": 77,
35
+ "pad_token": "!",
36
+ "tokenizer_class": "CLIPTokenizer",
37
+ "unk_token": "<|endoftext|>"
38
+ }
tokenizer_2/vocab.json CHANGED
The diff for this file is too large to render. See raw diff