Susana Sotelo Docio commited on
Commit
75e3f58
1 Parent(s): dfcb43b

Update model

Browse files
README.md CHANGED
@@ -9,9 +9,10 @@ tags:
9
  - roberta-base-bne
10
  widget:
11
  - text: "Fue antes de llegar a Sigüeiro, en el Camino de Santiago."
12
- - text: "El proyecto lo financia el Ministerio de Industria y Competitividad."
 
13
  model-index:
14
- - name: roberta-bne-ner-cds
15
  results: []
16
  ---
17
 
@@ -44,13 +45,13 @@ ToDo
44
 
45
  entity|precision|recall|f1
46
  -|-|-|-
47
- PER|0.965|0.924|0.944
48
- ORG|0.900|0.701|0.788
49
- LOC|0.982|0.985|0.983
50
- MISC|0.798|0.874|0.834
51
- micro avg|0.964|0.968|0.966|4265
52
- macro avg|0.911|0.871|0.887
53
- weighted avg|0.965|0.968|0.966
54
 
55
  ## Training procedure
56
 
@@ -67,7 +68,7 @@ The following hyperparameters were used during training:
67
 
68
  ### Framework versions
69
 
70
- - Transformers 4.25.1
71
- - Pytorch 1.13.0+cu117
72
- - Datasets 2.7.1
73
- - Tokenizers 0.13.2
 
9
  - roberta-base-bne
10
  widget:
11
  - text: "Fue antes de llegar a Sigüeiro, en el Camino de Santiago."
12
+ - text: "Si te metes en el Franco desde la Alameda, vas hacia la Catedral."
13
+ - text: "Y allí precisamente es Santiago el patrón del pueblo."
14
  model-index:
15
+ - name: es_trf_ner_cds_bne-base
16
  results: []
17
  ---
18
 
 
45
 
46
  entity|precision|recall|f1
47
  -|-|-|-
48
+ LOC|0.986|0.982|0.984
49
+ MISC|0.800|0.911|0.852
50
+ ORG|0.896|0.779|0.833
51
+ PER|0.953|0.937|0.945
52
+ micro avg|0.967|0.971|0.969
53
+ macro avg|0.909|0.902|0.903
54
+ weighted avg|0.968|0.971|0.969
55
 
56
  ## Training procedure
57
 
 
68
 
69
  ### Framework versions
70
 
71
+ - Transformers 4.28.1
72
+ - Pytorch 2.0.1+cu117
73
+ - Datasets 2.12.0
74
+ - Tokenizers 0.13.3
all_results.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 0.9983546594777942,
4
+ "eval_f1": 0.9674450707465777,
5
+ "eval_loss": 0.007044909987598658,
6
+ "eval_precision": 0.9653351698806244,
7
+ "eval_recall": 0.9695642148950888,
8
+ "eval_runtime": 16.6647,
9
+ "eval_samples": 15178,
10
+ "eval_samples_per_second": 910.788,
11
+ "eval_steps_per_second": 113.893,
12
+ "train_loss": 0.009494402144574594,
13
+ "train_runtime": 283.3545,
14
+ "train_samples": 45533,
15
+ "train_samples_per_second": 482.078,
16
+ "train_steps_per_second": 15.066
17
+ }
config.json CHANGED
@@ -44,7 +44,7 @@
44
  "pad_token_id": 1,
45
  "position_embedding_type": "absolute",
46
  "torch_dtype": "float32",
47
- "transformers_version": "4.25.1",
48
  "type_vocab_size": 1,
49
  "use_cache": true,
50
  "vocab_size": 50262
 
44
  "pad_token_id": 1,
45
  "position_embedding_type": "absolute",
46
  "torch_dtype": "float32",
47
+ "transformers_version": "4.28.1",
48
  "type_vocab_size": 1,
49
  "use_cache": true,
50
  "vocab_size": 50262
merges.txt CHANGED
@@ -1,4 +1,4 @@
1
- #version: 0.2 - Trained by `huggingface/tokenizers`
2
  Ġ d
3
  Ġ e
4
  Ġd e
 
1
+ #version: 0.2
2
  Ġ d
3
  Ġ e
4
  Ġd e
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54853750d9a7acb21f7ab654bfdbed2a01f46f209e9d1aaf9502cf1952e8c4ed
3
  size 496311341
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e8e6d0cf96f427fbddf174a0d9bc41f7fb4d1b73748577b86c83194d9e87fbc
3
  size 496311341
tokenizer.json CHANGED
@@ -87,6 +87,7 @@
87
  "continuing_subword_prefix": "",
88
  "end_of_word_suffix": "",
89
  "fuse_unk": false,
 
90
  "vocab": {
91
  "<s>": 0,
92
  "<pad>": 1,
 
87
  "continuing_subword_prefix": "",
88
  "end_of_word_suffix": "",
89
  "fuse_unk": false,
90
+ "byte_fallback": false,
91
  "vocab": {
92
  "<s>": 0,
93
  "<pad>": 1,
tokenizer_config.json CHANGED
@@ -8,6 +8,7 @@
8
  "rstrip": false,
9
  "single_word": false
10
  },
 
11
  "cls_token": {
12
  "__type": "AddedToken",
13
  "content": "<s>",
@@ -35,7 +36,6 @@
35
  },
36
  "max_len": 512,
37
  "model_max_length": 512,
38
- "name_or_path": "PlanTL-GOB-ES/roberta-base-bne",
39
  "pad_token": {
40
  "__type": "AddedToken",
41
  "content": "<pad>",
@@ -52,7 +52,6 @@
52
  "rstrip": false,
53
  "single_word": false
54
  },
55
- "special_tokens_map_file": null,
56
  "tokenizer_class": "RobertaTokenizer",
57
  "trim_offsets": true,
58
  "unk_token": {
 
8
  "rstrip": false,
9
  "single_word": false
10
  },
11
+ "clean_up_tokenization_spaces": true,
12
  "cls_token": {
13
  "__type": "AddedToken",
14
  "content": "<s>",
 
36
  },
37
  "max_len": 512,
38
  "model_max_length": 512,
 
39
  "pad_token": {
40
  "__type": "AddedToken",
41
  "content": "<pad>",
 
52
  "rstrip": false,
53
  "single_word": false
54
  },
 
55
  "tokenizer_class": "RobertaTokenizer",
56
  "trim_offsets": true,
57
  "unk_token": {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cb1283c01fe20d44d8f566f373f9c0086927abb23b9c7e7c7f8d3a5c64e28dff
3
- size 3515
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f96df4c64e04e69808c64c6b2f9a77af877cb3296b929a31ceaaa1e32e0cb04
3
+ size 3579