RJ3vans commited on
Commit
d731647
·
verified ·
1 Parent(s): 169595b

Upload TokenClassificationPipeline

Browse files
Files changed (5) hide show
  1. config.json +8 -2
  2. model.safetensors +2 -2
  3. tokenizer.json +0 -0
  4. tokenizer_config.json +1 -1
  5. vocab.txt +0 -0
config.json CHANGED
@@ -1,10 +1,11 @@
1
  {
2
- "_name_or_path": "/tmp/test-richard_bert_base_uncased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
 
8
  "finetuning_task": "ner",
9
  "gradient_checkpointing": false,
10
  "hidden_act": "gelu",
@@ -190,10 +191,15 @@
190
  "num_attention_heads": 16,
191
  "num_hidden_layers": 24,
192
  "pad_token_id": 0,
 
 
 
 
 
193
  "position_embedding_type": "absolute",
194
  "torch_dtype": "float32",
195
  "transformers_version": "4.47.1",
196
  "type_vocab_size": 2,
197
  "use_cache": true,
198
- "vocab_size": 30522
199
  }
 
1
  {
2
+ "_name_or_path": "/tmp/test-richard_bert_large_cased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "classifier_dropout": null,
8
+ "directionality": "bidi",
9
  "finetuning_task": "ner",
10
  "gradient_checkpointing": false,
11
  "hidden_act": "gelu",
 
191
  "num_attention_heads": 16,
192
  "num_hidden_layers": 24,
193
  "pad_token_id": 0,
194
+ "pooler_fc_size": 768,
195
+ "pooler_num_attention_heads": 12,
196
+ "pooler_num_fc_layers": 3,
197
+ "pooler_size_per_head": 128,
198
+ "pooler_type": "first_token_transform",
199
  "position_embedding_type": "absolute",
200
  "torch_dtype": "float32",
201
  "transformers_version": "4.47.1",
202
  "type_vocab_size": 2,
203
  "use_cache": true,
204
+ "vocab_size": 28996
205
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37b44208c4516ec2300eac39efde8bfe3b3a10fba8c26feafab7897d97fb050e
3
- size 1336760368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c1504e3a692070df46ea74d32f064ace0f0e78ab5f6fd0c007ad53960516d58
3
+ size 1330509872
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -43,7 +43,7 @@
43
  },
44
  "clean_up_tokenization_spaces": false,
45
  "cls_token": "[CLS]",
46
- "do_lower_case": true,
47
  "extra_special_tokens": {},
48
  "mask_token": "[MASK]",
49
  "max_length": 512,
 
43
  },
44
  "clean_up_tokenization_spaces": false,
45
  "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
  "extra_special_tokens": {},
48
  "mask_token": "[MASK]",
49
  "max_length": 512,
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff