File size: 1,583 Bytes
1fe0ffe
 
 
b741397
1fe0ffe
b741397
1fe0ffe
 
 
 
b741397
 
1fe0ffe
b741397
1fe0ffe
 
 
 
b741397
 
1fe0ffe
b741397
1fe0ffe
 
 
 
b741397
 
1fe0ffe
b741397
1fe0ffe
 
 
 
b741397
 
 
 
1fe0ffe
 
 
 
 
b3b76f5
 
1fe0ffe
b741397
b3b76f5
 
 
 
 
 
e4c5d5e
b741397
 
1fe0ffe
e4c5d5e
b3b76f5
b741397
b3b76f5
 
 
 
 
 
 
 
b741397
b3b76f5
 
 
 
 
 
b741397
 
b3b76f5
b741397
1fe0ffe
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
{
  "added_tokens_decoder": {
    "0": {
      "content": "[PAD]",
      "lstrip": false,
      "normalized": false,
      "rstrip": false,
      "single_word": false,
      "special": true
    },
    "100": {
      "content": "[UNK]",
      "lstrip": false,
      "normalized": false,
      "rstrip": false,
      "single_word": false,
      "special": true
    },
    "101": {
      "content": "[CLS]",
      "lstrip": false,
      "normalized": false,
      "rstrip": false,
      "single_word": false,
      "special": true
    },
    "102": {
      "content": "[SEP]",
      "lstrip": false,
      "normalized": false,
      "rstrip": false,
      "single_word": false,
      "special": true
    },
    "103": {
      "content": "[MASK]",
      "lstrip": false,
      "normalized": false,
      "rstrip": false,
      "single_word": false,
      "special": true
    }
  },
  "additional_special_tokens": [],
  "apply_ocr": false,
  "clean_up_tokenization_spaces": true,
  "cls_token": "[CLS]",
  "cls_token_box": [
    0,
    0,
    0,
    0
  ],
  "do_basic_tokenize": true,
  "do_lower_case": true,
  "mask_token": "[MASK]",
  "model_max_length": 512,
  "never_split": null,
  "only_label_first_subword": true,
  "pad_token": "[PAD]",
  "pad_token_box": [
    0,
    0,
    0,
    0
  ],
  "pad_token_label": -100,
  "processor_class": "LayoutLMv2Processor",
  "sep_token": "[SEP]",
  "sep_token_box": [
    1000,
    1000,
    1000,
    1000
  ],
  "strip_accents": null,
  "tokenize_chinese_chars": true,
  "tokenizer_class": "LayoutLMv2Tokenizer",
  "unk_token": "[UNK]"
}