model update
Browse files- config.json +1 -1
- eval/metric.json +1 -1
- eval/prediction.2020.dev.json +0 -0
- tokenizer_config.json +1 -1
- trainer_config.json +1 -0
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"adapters": {
|
4 |
"adapters": {},
|
5 |
"config_map": {},
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "roberta-large",
|
3 |
"adapters": {
|
4 |
"adapters": {},
|
5 |
"config_map": {},
|
eval/metric.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"2020.dev": {"micro/f1": 0.
|
|
|
1 |
+
{"2020.dev": {"micro/f1": 0.6376109765940274, "micro/f1_ci": {}, "micro/recall": 0.6191222570532915, "micro/precision": 0.6572379367720466, "macro/f1": 0.5786196442940901, "macro/f1_ci": {}, "macro/recall": 0.5654249405318954, "macro/precision": 0.5969422597833425, "per_entity_metric": {"corporation": {"f1": 0.4817927170868347, "f1_ci": {}, "precision": 0.5584415584415584, "recall": 0.4236453201970443}, "creative_work": {"f1": 0.48470588235294115, "f1_ci": {}, "precision": 0.47465437788018433, "recall": 0.4951923076923077}, "event": {"f1": 0.35802469135802467, "f1_ci": {}, "precision": 0.3782608695652174, "recall": 0.33984375}, "group": {"f1": 0.5526932084309134, "f1_ci": {}, "precision": 0.59, "recall": 0.5198237885462555}, "location": {"f1": 0.6326530612244898, "f1_ci": {}, "precision": 0.5876777251184834, "recall": 0.6850828729281768}, "person": {"f1": 0.8824034334763948, "f1_ci": {}, "precision": 0.9065255731922398, "recall": 0.8595317725752508}, "product": {"f1": 0.6580645161290323, "f1_ci": {}, "precision": 0.6830357142857143, "recall": 0.6348547717842323}}}}
|
eval/prediction.2020.dev.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "
|
|
|
1 |
+
{"unk_token": "<unk>", "bos_token": "<s>", "eos_token": "</s>", "add_prefix_space": false, "errors": "replace", "sep_token": "</s>", "cls_token": "<s>", "pad_token": "<pad>", "mask_token": "<mask>", "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "roberta-large", "tokenizer_class": "RobertaTokenizer"}
|
trainer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"data_split": "2020.train", "model": "roberta-large", "crf": true, "max_length": 128, "epoch": 30, "batch_size": 32, "lr": 1e-05, "random_seed": 0, "gradient_accumulation_steps": 1, "weight_decay": 1e-07, "lr_warmup_step_ratio": 0.3, "max_grad_norm": 1}
|