haryoaw commited on
Commit
30b5b5f
1 Parent(s): 248b698

Initial Commit

Browse files
Files changed (4) hide show
  1. README.md +88 -110
  2. eval_result_ner.json +1 -1
  3. model.safetensors +1 -1
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,14 +1,14 @@
1
  ---
2
- base_model: microsoft/mdeberta-v3-base
3
  library_name: transformers
4
  license: mit
 
 
 
5
  metrics:
6
  - precision
7
  - recall
8
  - f1
9
  - accuracy
10
- tags:
11
- - generated_from_trainer
12
  model-index:
13
  - name: scenario-non-kd-scr-ner-full-mdeberta_data-univner_full55
14
  results: []
@@ -21,10 +21,10 @@ should probably proofread and complete it, then remove this comment. -->
21
 
22
  This model is a fine-tuned version of [microsoft/mdeberta-v3-base](https://huggingface.co/microsoft/mdeberta-v3-base) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.3716
25
- - Precision: 0.6403
26
- - Recall: 0.6003
27
- - F1: 0.6197
28
  - Accuracy: 0.9630
29
 
30
  ## Model description
@@ -56,109 +56,87 @@ The following hyperparameters were used during training:
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
58
  |:-------------:|:-------:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:|
59
- | 0.3049 | 0.2910 | 500 | 0.2421 | 0.3445 | 0.2023 | 0.2549 | 0.9344 |
60
- | 0.1944 | 0.5821 | 1000 | 0.1887 | 0.4125 | 0.3663 | 0.3880 | 0.9459 |
61
- | 0.1496 | 0.8731 | 1500 | 0.1595 | 0.4675 | 0.4816 | 0.4745 | 0.9521 |
62
- | 0.108 | 1.1641 | 2000 | 0.1674 | 0.5369 | 0.4727 | 0.5027 | 0.9555 |
63
- | 0.0861 | 1.4552 | 2500 | 0.1482 | 0.4975 | 0.5773 | 0.5344 | 0.9553 |
64
- | 0.0799 | 1.7462 | 3000 | 0.1540 | 0.5180 | 0.5859 | 0.5499 | 0.9565 |
65
- | 0.0712 | 2.0373 | 3500 | 0.1607 | 0.5804 | 0.5491 | 0.5643 | 0.9602 |
66
- | 0.0468 | 2.3283 | 4000 | 0.1684 | 0.5888 | 0.5618 | 0.5750 | 0.9610 |
67
- | 0.0451 | 2.6193 | 4500 | 0.1660 | 0.6045 | 0.5788 | 0.5914 | 0.9619 |
68
- | 0.0448 | 2.9104 | 5000 | 0.1607 | 0.5980 | 0.5835 | 0.5906 | 0.9622 |
69
- | 0.0304 | 3.2014 | 5500 | 0.1758 | 0.6017 | 0.5817 | 0.5915 | 0.9623 |
70
- | 0.026 | 3.4924 | 6000 | 0.1844 | 0.6058 | 0.5722 | 0.5885 | 0.9614 |
71
- | 0.0267 | 3.7835 | 6500 | 0.1856 | 0.5987 | 0.5908 | 0.5947 | 0.9620 |
72
- | 0.0239 | 4.0745 | 7000 | 0.1959 | 0.5982 | 0.6009 | 0.5995 | 0.9626 |
73
- | 0.0144 | 4.3655 | 7500 | 0.2080 | 0.6233 | 0.5812 | 0.6015 | 0.9628 |
74
- | 0.0161 | 4.6566 | 8000 | 0.1997 | 0.6053 | 0.5926 | 0.5989 | 0.9622 |
75
- | 0.0168 | 4.9476 | 8500 | 0.2029 | 0.5840 | 0.6276 | 0.6050 | 0.9625 |
76
- | 0.0108 | 5.2386 | 9000 | 0.2222 | 0.5902 | 0.6191 | 0.6043 | 0.9623 |
77
- | 0.0099 | 5.5297 | 9500 | 0.2334 | 0.6012 | 0.5993 | 0.6002 | 0.9624 |
78
- | 0.0099 | 5.8207 | 10000 | 0.2226 | 0.6103 | 0.6117 | 0.6110 | 0.9625 |
79
- | 0.0093 | 6.1118 | 10500 | 0.2447 | 0.6089 | 0.5924 | 0.6006 | 0.9622 |
80
- | 0.0066 | 6.4028 | 11000 | 0.2507 | 0.5975 | 0.6090 | 0.6032 | 0.9621 |
81
- | 0.0074 | 6.6938 | 11500 | 0.2553 | 0.5984 | 0.6001 | 0.5992 | 0.9625 |
82
- | 0.0074 | 6.9849 | 12000 | 0.2498 | 0.6079 | 0.5876 | 0.5976 | 0.9620 |
83
- | 0.0045 | 7.2759 | 12500 | 0.2704 | 0.5944 | 0.5985 | 0.5964 | 0.9622 |
84
- | 0.0054 | 7.5669 | 13000 | 0.2619 | 0.6076 | 0.5972 | 0.6023 | 0.9621 |
85
- | 0.0057 | 7.8580 | 13500 | 0.2755 | 0.6113 | 0.5812 | 0.5959 | 0.9618 |
86
- | 0.0052 | 8.1490 | 14000 | 0.2589 | 0.6026 | 0.6133 | 0.6079 | 0.9623 |
87
- | 0.0037 | 8.4400 | 14500 | 0.2706 | 0.6169 | 0.5941 | 0.6053 | 0.9621 |
88
- | 0.0042 | 8.7311 | 15000 | 0.2766 | 0.5952 | 0.6154 | 0.6051 | 0.9612 |
89
- | 0.0039 | 9.0221 | 15500 | 0.2786 | 0.6111 | 0.5930 | 0.6019 | 0.9624 |
90
- | 0.0026 | 9.3132 | 16000 | 0.2878 | 0.6176 | 0.5839 | 0.6003 | 0.9616 |
91
- | 0.0029 | 9.6042 | 16500 | 0.2880 | 0.6274 | 0.5897 | 0.6080 | 0.9626 |
92
- | 0.0034 | 9.8952 | 17000 | 0.2865 | 0.6119 | 0.6076 | 0.6097 | 0.9625 |
93
- | 0.0028 | 10.1863 | 17500 | 0.2835 | 0.6128 | 0.6109 | 0.6118 | 0.9626 |
94
- | 0.0025 | 10.4773 | 18000 | 0.2833 | 0.6062 | 0.6104 | 0.6083 | 0.9626 |
95
- | 0.0022 | 10.7683 | 18500 | 0.2938 | 0.6124 | 0.6178 | 0.6151 | 0.9624 |
96
- | 0.0029 | 11.0594 | 19000 | 0.3017 | 0.6224 | 0.5933 | 0.6075 | 0.9622 |
97
- | 0.0023 | 11.3504 | 19500 | 0.3013 | 0.6277 | 0.5918 | 0.6092 | 0.9626 |
98
- | 0.0022 | 11.6414 | 20000 | 0.3003 | 0.6241 | 0.5910 | 0.6071 | 0.9625 |
99
- | 0.0021 | 11.9325 | 20500 | 0.3098 | 0.6284 | 0.5888 | 0.6080 | 0.9625 |
100
- | 0.0016 | 12.2235 | 21000 | 0.3075 | 0.6182 | 0.5973 | 0.6076 | 0.9625 |
101
- | 0.0015 | 12.5146 | 21500 | 0.3202 | 0.6382 | 0.5917 | 0.6141 | 0.9629 |
102
- | 0.0017 | 12.8056 | 22000 | 0.3060 | 0.6194 | 0.6165 | 0.6179 | 0.9630 |
103
- | 0.0017 | 13.0966 | 22500 | 0.3139 | 0.6260 | 0.6048 | 0.6152 | 0.9629 |
104
- | 0.0011 | 13.3877 | 23000 | 0.3247 | 0.6306 | 0.5848 | 0.6068 | 0.9628 |
105
- | 0.0014 | 13.6787 | 23500 | 0.3184 | 0.6369 | 0.5973 | 0.6165 | 0.9629 |
106
- | 0.0016 | 13.9697 | 24000 | 0.3177 | 0.6334 | 0.5928 | 0.6125 | 0.9631 |
107
- | 0.0012 | 14.2608 | 24500 | 0.3162 | 0.6316 | 0.6060 | 0.6185 | 0.9632 |
108
- | 0.001 | 14.5518 | 25000 | 0.3166 | 0.6305 | 0.5947 | 0.6121 | 0.9625 |
109
- | 0.0013 | 14.8428 | 25500 | 0.3338 | 0.6144 | 0.5797 | 0.5965 | 0.9622 |
110
- | 0.001 | 15.1339 | 26000 | 0.3287 | 0.6161 | 0.6024 | 0.6092 | 0.9627 |
111
- | 0.0008 | 15.4249 | 26500 | 0.3321 | 0.6194 | 0.6132 | 0.6163 | 0.9626 |
112
- | 0.0011 | 15.7159 | 27000 | 0.3292 | 0.6195 | 0.6018 | 0.6105 | 0.9624 |
113
- | 0.0013 | 16.0070 | 27500 | 0.3255 | 0.6214 | 0.6100 | 0.6157 | 0.9626 |
114
- | 0.0007 | 16.2980 | 28000 | 0.3410 | 0.6298 | 0.5894 | 0.6089 | 0.9625 |
115
- | 0.0008 | 16.5891 | 28500 | 0.3310 | 0.6202 | 0.6136 | 0.6169 | 0.9628 |
116
- | 0.0008 | 16.8801 | 29000 | 0.3366 | 0.6266 | 0.6040 | 0.6150 | 0.9627 |
117
- | 0.0007 | 17.1711 | 29500 | 0.3331 | 0.6270 | 0.6070 | 0.6168 | 0.9627 |
118
- | 0.0006 | 17.4622 | 30000 | 0.3410 | 0.6421 | 0.6028 | 0.6218 | 0.9628 |
119
- | 0.0007 | 17.7532 | 30500 | 0.3459 | 0.6444 | 0.5882 | 0.6150 | 0.9626 |
120
- | 0.0008 | 18.0442 | 31000 | 0.3577 | 0.6452 | 0.5814 | 0.6117 | 0.9627 |
121
- | 0.0007 | 18.3353 | 31500 | 0.3473 | 0.6250 | 0.5947 | 0.6095 | 0.9621 |
122
- | 0.0005 | 18.6263 | 32000 | 0.3449 | 0.6184 | 0.6107 | 0.6145 | 0.9623 |
123
- | 0.0005 | 18.9173 | 32500 | 0.3459 | 0.6322 | 0.6008 | 0.6161 | 0.9625 |
124
- | 0.0005 | 19.2084 | 33000 | 0.3526 | 0.6247 | 0.6048 | 0.6146 | 0.9628 |
125
- | 0.0003 | 19.4994 | 33500 | 0.3578 | 0.6383 | 0.5923 | 0.6144 | 0.9626 |
126
- | 0.0007 | 19.7905 | 34000 | 0.3438 | 0.6239 | 0.6166 | 0.6203 | 0.9626 |
127
- | 0.0005 | 20.0815 | 34500 | 0.3454 | 0.6294 | 0.6115 | 0.6203 | 0.9629 |
128
- | 0.0004 | 20.3725 | 35000 | 0.3505 | 0.6380 | 0.5996 | 0.6182 | 0.9630 |
129
- | 0.0003 | 20.6636 | 35500 | 0.3525 | 0.6327 | 0.5920 | 0.6117 | 0.9626 |
130
- | 0.0003 | 20.9546 | 36000 | 0.3543 | 0.6191 | 0.6265 | 0.6228 | 0.9629 |
131
- | 0.0002 | 21.2456 | 36500 | 0.3570 | 0.6285 | 0.6057 | 0.6169 | 0.9628 |
132
- | 0.0003 | 21.5367 | 37000 | 0.3683 | 0.6162 | 0.6136 | 0.6149 | 0.9630 |
133
- | 0.0004 | 21.8277 | 37500 | 0.3562 | 0.6156 | 0.6126 | 0.6141 | 0.9627 |
134
- | 0.0003 | 22.1187 | 38000 | 0.3599 | 0.6278 | 0.6073 | 0.6174 | 0.9626 |
135
- | 0.0001 | 22.4098 | 38500 | 0.3587 | 0.6371 | 0.5973 | 0.6166 | 0.9626 |
136
- | 0.0004 | 22.7008 | 39000 | 0.3620 | 0.6148 | 0.6071 | 0.6109 | 0.9623 |
137
- | 0.0003 | 22.9919 | 39500 | 0.3568 | 0.6347 | 0.5990 | 0.6163 | 0.9628 |
138
- | 0.0002 | 23.2829 | 40000 | 0.3555 | 0.6363 | 0.6019 | 0.6186 | 0.9629 |
139
- | 0.0002 | 23.5739 | 40500 | 0.3560 | 0.6231 | 0.6198 | 0.6214 | 0.9631 |
140
- | 0.0002 | 23.8650 | 41000 | 0.3622 | 0.6166 | 0.6083 | 0.6124 | 0.9626 |
141
- | 0.0003 | 24.1560 | 41500 | 0.3534 | 0.6284 | 0.6084 | 0.6182 | 0.9631 |
142
- | 0.0002 | 24.4470 | 42000 | 0.3669 | 0.6367 | 0.5891 | 0.6120 | 0.9625 |
143
- | 0.0001 | 24.7381 | 42500 | 0.3583 | 0.6211 | 0.6115 | 0.6163 | 0.9628 |
144
- | 0.0001 | 25.0291 | 43000 | 0.3654 | 0.6297 | 0.5992 | 0.6141 | 0.9627 |
145
- | 0.0001 | 25.3201 | 43500 | 0.3659 | 0.6356 | 0.6057 | 0.6203 | 0.9630 |
146
- | 0.0001 | 25.6112 | 44000 | 0.3638 | 0.6310 | 0.6087 | 0.6197 | 0.9630 |
147
- | 0.0002 | 25.9022 | 44500 | 0.3663 | 0.6265 | 0.6103 | 0.6183 | 0.9629 |
148
- | 0.0001 | 26.1932 | 45000 | 0.3651 | 0.6357 | 0.6044 | 0.6196 | 0.9631 |
149
- | 0.0001 | 26.4843 | 45500 | 0.3644 | 0.6370 | 0.6128 | 0.6247 | 0.9633 |
150
- | 0.0002 | 26.7753 | 46000 | 0.3637 | 0.6325 | 0.6066 | 0.6192 | 0.9629 |
151
- | 0.0001 | 27.0664 | 46500 | 0.3645 | 0.6435 | 0.5967 | 0.6193 | 0.9630 |
152
- | 0.0 | 27.3574 | 47000 | 0.3657 | 0.6328 | 0.6018 | 0.6169 | 0.9629 |
153
- | 0.0001 | 27.6484 | 47500 | 0.3742 | 0.6392 | 0.5901 | 0.6137 | 0.9629 |
154
- | 0.0001 | 27.9395 | 48000 | 0.3682 | 0.6400 | 0.5999 | 0.6193 | 0.9629 |
155
- | 0.0 | 28.2305 | 48500 | 0.3670 | 0.6317 | 0.6112 | 0.6213 | 0.9631 |
156
- | 0.0 | 28.5215 | 49000 | 0.3707 | 0.6394 | 0.6035 | 0.6209 | 0.9632 |
157
- | 0.0001 | 28.8126 | 49500 | 0.3669 | 0.6369 | 0.6080 | 0.6221 | 0.9632 |
158
- | 0.0001 | 29.1036 | 50000 | 0.3692 | 0.6434 | 0.6009 | 0.6215 | 0.9631 |
159
- | 0.0 | 29.3946 | 50500 | 0.3683 | 0.6374 | 0.6055 | 0.6210 | 0.9631 |
160
- | 0.0001 | 29.6857 | 51000 | 0.3716 | 0.6379 | 0.5993 | 0.6180 | 0.9630 |
161
- | 0.0 | 29.9767 | 51500 | 0.3716 | 0.6403 | 0.6003 | 0.6197 | 0.9630 |
162
 
163
 
164
  ### Framework versions
 
1
  ---
 
2
  library_name: transformers
3
  license: mit
4
+ base_model: microsoft/mdeberta-v3-base
5
+ tags:
6
+ - generated_from_trainer
7
  metrics:
8
  - precision
9
  - recall
10
  - f1
11
  - accuracy
 
 
12
  model-index:
13
  - name: scenario-non-kd-scr-ner-full-mdeberta_data-univner_full55
14
  results: []
 
21
 
22
  This model is a fine-tuned version of [microsoft/mdeberta-v3-base](https://huggingface.co/microsoft/mdeberta-v3-base) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.3730
25
+ - Precision: 0.6215
26
+ - Recall: 0.6022
27
+ - F1: 0.6117
28
  - Accuracy: 0.9630
29
 
30
  ## Model description
 
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
58
  |:-------------:|:-------:|:-----:|:---------------:|:---------:|:------:|:------:|:--------:|
59
+ | 0.3045 | 0.2910 | 500 | 0.2283 | 0.3540 | 0.2161 | 0.2684 | 0.9357 |
60
+ | 0.1895 | 0.5821 | 1000 | 0.1836 | 0.4325 | 0.3881 | 0.4091 | 0.9470 |
61
+ | 0.1453 | 0.8731 | 1500 | 0.1564 | 0.4890 | 0.4855 | 0.4873 | 0.9530 |
62
+ | 0.1066 | 1.1641 | 2000 | 0.1663 | 0.5335 | 0.4569 | 0.4923 | 0.9552 |
63
+ | 0.0852 | 1.4552 | 2500 | 0.1476 | 0.5104 | 0.5826 | 0.5441 | 0.9564 |
64
+ | 0.0793 | 1.7462 | 3000 | 0.1463 | 0.5519 | 0.5879 | 0.5693 | 0.9590 |
65
+ | 0.0707 | 2.0373 | 3500 | 0.1563 | 0.6032 | 0.5569 | 0.5791 | 0.9609 |
66
+ | 0.0466 | 2.3283 | 4000 | 0.1710 | 0.6074 | 0.5667 | 0.5864 | 0.9614 |
67
+ | 0.0458 | 2.6193 | 4500 | 0.1601 | 0.5843 | 0.6106 | 0.5971 | 0.9613 |
68
+ | 0.0456 | 2.9104 | 5000 | 0.1588 | 0.5943 | 0.5926 | 0.5934 | 0.9624 |
69
+ | 0.0308 | 3.2014 | 5500 | 0.1774 | 0.5950 | 0.5965 | 0.5957 | 0.9624 |
70
+ | 0.0257 | 3.4924 | 6000 | 0.1898 | 0.5864 | 0.5852 | 0.5858 | 0.9614 |
71
+ | 0.027 | 3.7835 | 6500 | 0.1869 | 0.5966 | 0.5963 | 0.5964 | 0.9619 |
72
+ | 0.0248 | 4.0745 | 7000 | 0.2036 | 0.6006 | 0.5920 | 0.5962 | 0.9623 |
73
+ | 0.015 | 4.3655 | 7500 | 0.2156 | 0.6220 | 0.5713 | 0.5956 | 0.9624 |
74
+ | 0.0172 | 4.6566 | 8000 | 0.1987 | 0.6221 | 0.5996 | 0.6106 | 0.9632 |
75
+ | 0.0171 | 4.9476 | 8500 | 0.2058 | 0.5867 | 0.6334 | 0.6091 | 0.9622 |
76
+ | 0.0107 | 5.2386 | 9000 | 0.2216 | 0.5917 | 0.6125 | 0.6019 | 0.9621 |
77
+ | 0.0104 | 5.5297 | 9500 | 0.2355 | 0.6218 | 0.5924 | 0.6068 | 0.9631 |
78
+ | 0.0107 | 5.8207 | 10000 | 0.2279 | 0.6144 | 0.6096 | 0.6120 | 0.9625 |
79
+ | 0.0099 | 6.1118 | 10500 | 0.2473 | 0.6093 | 0.6009 | 0.6051 | 0.9626 |
80
+ | 0.0069 | 6.4028 | 11000 | 0.2488 | 0.5906 | 0.6133 | 0.6017 | 0.9623 |
81
+ | 0.0081 | 6.6938 | 11500 | 0.2435 | 0.5928 | 0.6154 | 0.6039 | 0.9624 |
82
+ | 0.0083 | 6.9849 | 12000 | 0.2470 | 0.6183 | 0.5973 | 0.6076 | 0.9627 |
83
+ | 0.0046 | 7.2759 | 12500 | 0.2558 | 0.6143 | 0.5913 | 0.6026 | 0.9627 |
84
+ | 0.0052 | 7.5669 | 13000 | 0.2637 | 0.5962 | 0.6205 | 0.6081 | 0.9617 |
85
+ | 0.0055 | 7.8580 | 13500 | 0.2688 | 0.6019 | 0.5941 | 0.5980 | 0.9621 |
86
+ | 0.0052 | 8.1490 | 14000 | 0.2702 | 0.5943 | 0.6126 | 0.6033 | 0.9618 |
87
+ | 0.0036 | 8.4400 | 14500 | 0.2707 | 0.5897 | 0.6289 | 0.6087 | 0.9618 |
88
+ | 0.0044 | 8.7311 | 15000 | 0.2772 | 0.6089 | 0.6044 | 0.6066 | 0.9627 |
89
+ | 0.0045 | 9.0221 | 15500 | 0.2781 | 0.6104 | 0.5856 | 0.5977 | 0.9621 |
90
+ | 0.0028 | 9.3132 | 16000 | 0.2813 | 0.6061 | 0.6099 | 0.6080 | 0.9627 |
91
+ | 0.0032 | 9.6042 | 16500 | 0.2962 | 0.6134 | 0.5999 | 0.6066 | 0.9627 |
92
+ | 0.0041 | 9.8952 | 17000 | 0.2819 | 0.6097 | 0.5980 | 0.6038 | 0.9625 |
93
+ | 0.0025 | 10.1863 | 17500 | 0.2859 | 0.6138 | 0.6037 | 0.6087 | 0.9627 |
94
+ | 0.0022 | 10.4773 | 18000 | 0.2976 | 0.6018 | 0.6122 | 0.6069 | 0.9624 |
95
+ | 0.0027 | 10.7683 | 18500 | 0.3066 | 0.6387 | 0.5819 | 0.6090 | 0.9626 |
96
+ | 0.003 | 11.0594 | 19000 | 0.2925 | 0.6402 | 0.5921 | 0.6152 | 0.9632 |
97
+ | 0.002 | 11.3504 | 19500 | 0.3069 | 0.5776 | 0.6094 | 0.5931 | 0.9613 |
98
+ | 0.0023 | 11.6414 | 20000 | 0.2979 | 0.6201 | 0.6063 | 0.6131 | 0.9628 |
99
+ | 0.0023 | 11.9325 | 20500 | 0.3015 | 0.5935 | 0.6181 | 0.6056 | 0.9621 |
100
+ | 0.0015 | 12.2235 | 21000 | 0.3179 | 0.6137 | 0.6070 | 0.6103 | 0.9629 |
101
+ | 0.0016 | 12.5146 | 21500 | 0.3073 | 0.6145 | 0.6211 | 0.6178 | 0.9631 |
102
+ | 0.0017 | 12.8056 | 22000 | 0.3159 | 0.6267 | 0.5914 | 0.6085 | 0.9628 |
103
+ | 0.0016 | 13.0966 | 22500 | 0.3224 | 0.6003 | 0.6154 | 0.6077 | 0.9623 |
104
+ | 0.0015 | 13.3877 | 23000 | 0.3160 | 0.6111 | 0.5884 | 0.5995 | 0.9624 |
105
+ | 0.0016 | 13.6787 | 23500 | 0.3201 | 0.6208 | 0.6057 | 0.6132 | 0.9630 |
106
+ | 0.0016 | 13.9697 | 24000 | 0.3187 | 0.6251 | 0.5954 | 0.6099 | 0.9626 |
107
+ | 0.0011 | 14.2608 | 24500 | 0.3188 | 0.6253 | 0.6094 | 0.6173 | 0.9630 |
108
+ | 0.0013 | 14.5518 | 25000 | 0.3178 | 0.6170 | 0.6165 | 0.6168 | 0.9629 |
109
+ | 0.0011 | 14.8428 | 25500 | 0.3311 | 0.6304 | 0.5830 | 0.6058 | 0.9626 |
110
+ | 0.0011 | 15.1339 | 26000 | 0.3345 | 0.6200 | 0.6077 | 0.6138 | 0.9631 |
111
+ | 0.0009 | 15.4249 | 26500 | 0.3385 | 0.6107 | 0.5970 | 0.6038 | 0.9623 |
112
+ | 0.0011 | 15.7159 | 27000 | 0.3289 | 0.6219 | 0.6192 | 0.6206 | 0.9632 |
113
+ | 0.001 | 16.0070 | 27500 | 0.3345 | 0.6101 | 0.5986 | 0.6043 | 0.9627 |
114
+ | 0.0005 | 16.2980 | 28000 | 0.3388 | 0.6202 | 0.6053 | 0.6126 | 0.9627 |
115
+ | 0.0007 | 16.5891 | 28500 | 0.3375 | 0.6204 | 0.6152 | 0.6178 | 0.9630 |
116
+ | 0.0009 | 16.8801 | 29000 | 0.3439 | 0.6103 | 0.6175 | 0.6139 | 0.9627 |
117
+ | 0.0008 | 17.1711 | 29500 | 0.3406 | 0.6238 | 0.6149 | 0.6193 | 0.9630 |
118
+ | 0.0006 | 17.4622 | 30000 | 0.3436 | 0.6147 | 0.6086 | 0.6116 | 0.9631 |
119
+ | 0.0007 | 17.7532 | 30500 | 0.3336 | 0.6366 | 0.6080 | 0.6219 | 0.9633 |
120
+ | 0.0005 | 18.0442 | 31000 | 0.3510 | 0.6210 | 0.6038 | 0.6123 | 0.9630 |
121
+ | 0.0005 | 18.3353 | 31500 | 0.3560 | 0.6148 | 0.6038 | 0.6093 | 0.9626 |
122
+ | 0.0008 | 18.6263 | 32000 | 0.3578 | 0.6195 | 0.6097 | 0.6146 | 0.9627 |
123
+ | 0.0004 | 18.9173 | 32500 | 0.3573 | 0.6300 | 0.6035 | 0.6165 | 0.9631 |
124
+ | 0.0005 | 19.2084 | 33000 | 0.3565 | 0.6336 | 0.6041 | 0.6185 | 0.9630 |
125
+ | 0.0004 | 19.4994 | 33500 | 0.3627 | 0.6317 | 0.6047 | 0.6179 | 0.9633 |
126
+ | 0.0005 | 19.7905 | 34000 | 0.3632 | 0.6161 | 0.6216 | 0.6188 | 0.9631 |
127
+ | 0.0004 | 20.0815 | 34500 | 0.3581 | 0.6086 | 0.6115 | 0.6100 | 0.9629 |
128
+ | 0.0004 | 20.3725 | 35000 | 0.3638 | 0.6148 | 0.6029 | 0.6088 | 0.9628 |
129
+ | 0.0005 | 20.6636 | 35500 | 0.3579 | 0.6216 | 0.6022 | 0.6118 | 0.9627 |
130
+ | 0.0003 | 20.9546 | 36000 | 0.3601 | 0.6014 | 0.6214 | 0.6112 | 0.9627 |
131
+ | 0.0002 | 21.2456 | 36500 | 0.3671 | 0.6361 | 0.5934 | 0.6140 | 0.9631 |
132
+ | 0.0003 | 21.5367 | 37000 | 0.3706 | 0.6269 | 0.6029 | 0.6147 | 0.9632 |
133
+ | 0.0003 | 21.8277 | 37500 | 0.3645 | 0.6128 | 0.6080 | 0.6104 | 0.9628 |
134
+ | 0.0004 | 22.1187 | 38000 | 0.3660 | 0.6157 | 0.6099 | 0.6128 | 0.9628 |
135
+ | 0.0003 | 22.4098 | 38500 | 0.3625 | 0.6214 | 0.6054 | 0.6133 | 0.9633 |
136
+ | 0.0003 | 22.7008 | 39000 | 0.3602 | 0.6125 | 0.6109 | 0.6117 | 0.9631 |
137
+ | 0.0003 | 22.9919 | 39500 | 0.3692 | 0.6196 | 0.5931 | 0.6061 | 0.9630 |
138
+ | 0.0001 | 23.2829 | 40000 | 0.3713 | 0.6226 | 0.6035 | 0.6129 | 0.9631 |
139
+ | 0.0002 | 23.5739 | 40500 | 0.3730 | 0.6215 | 0.6022 | 0.6117 | 0.9630 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
140
 
141
 
142
  ### Framework versions
eval_result_ner.json CHANGED
@@ -1 +1 @@
1
- {"ceb_gja": {"precision": 0.32967032967032966, "recall": 0.6122448979591837, "f1": 0.4285714285714286, "accuracy": 0.9374517374517375}, "en_pud": {"precision": 0.4904862579281184, "recall": 0.4316279069767442, "f1": 0.4591786244433449, "accuracy": 0.949187759727994}, "de_pud": {"precision": 0.13655640373197625, "recall": 0.30991337824831566, "f1": 0.18957904032970269, "accuracy": 0.8522807182035536}, "pt_pud": {"precision": 0.5718232044198895, "recall": 0.565059144676979, "f1": 0.5684210526315789, "accuracy": 0.9607809629597983}, "ru_pud": {"precision": 0.01642664266426643, "recall": 0.07046332046332046, "f1": 0.026642335766423362, "accuracy": 0.6238698010849909}, "sv_pud": {"precision": 0.5173501577287066, "recall": 0.31875607385811466, "f1": 0.3944678292242934, "accuracy": 0.9455336548542671}, "tl_trg": {"precision": 0.25, "recall": 0.6086956521739131, "f1": 0.35443037974683544, "accuracy": 0.9332425068119891}, "tl_ugnayan": {"precision": 0.033707865168539325, "recall": 0.09090909090909091, "f1": 0.04918032786885246, "accuracy": 0.8979033728350045}, "zh_gsd": {"precision": 0.5717852684144819, "recall": 0.5971316818774446, "f1": 0.5841836734693878, "accuracy": 0.9451381951381952}, "zh_gsdsimp": {"precision": 0.5605095541401274, "recall": 0.5766710353866317, "f1": 0.5684754521963824, "accuracy": 0.9437229437229437}, "hr_set": {"precision": 0.7615330021291696, "recall": 0.7647897362794013, "f1": 0.763157894736842, "accuracy": 0.972835943940643}, "da_ddt": {"precision": 0.6641604010025063, "recall": 0.5928411633109619, "f1": 0.6264775413711584, "accuracy": 0.9735608101366856}, "en_ewt": {"precision": 0.6173469387755102, "recall": 0.5560661764705882, "f1": 0.5851063829787233, "accuracy": 0.9610311989480814}, "pt_bosque": {"precision": 0.6515276630883567, "recall": 0.6493827160493827, "f1": 0.6504534212695795, "accuracy": 0.9676133893638603}, "sr_set": {"precision": 0.7988235294117647, "recall": 0.8016528925619835, "f1": 0.8002357100766058, "accuracy": 0.9712809736450398}, "sk_snk": {"precision": 0.4069767441860465, "recall": 0.2677595628415301, "f1": 0.3230059327620303, "accuracy": 0.9170854271356784}, "sv_talbanken": {"precision": 0.6742857142857143, "recall": 0.6020408163265306, "f1": 0.6361185983827493, "accuracy": 0.993522108259312}}
 
1
+ {"ceb_gja": {"precision": 0.26785714285714285, "recall": 0.6122448979591837, "f1": 0.3726708074534162, "accuracy": 0.9173745173745174}, "en_pud": {"precision": 0.5027322404371585, "recall": 0.42790697674418604, "f1": 0.4623115577889447, "accuracy": 0.9496127691726483}, "de_pud": {"precision": 0.13111962537249894, "recall": 0.2964388835418672, "f1": 0.18181818181818182, "accuracy": 0.8523744784585814}, "pt_pud": {"precision": 0.5862068965517241, "recall": 0.556869881710646, "f1": 0.5711619225384974, "accuracy": 0.9611654633229376}, "ru_pud": {"precision": 0.01607084289931125, "recall": 0.0472972972972973, "f1": 0.023990208078335372, "accuracy": 0.7374838543012141}, "sv_pud": {"precision": 0.5384615384615384, "recall": 0.36054421768707484, "f1": 0.4318975552968568, "accuracy": 0.9472111553784861}, "tl_trg": {"precision": 0.2727272727272727, "recall": 0.6521739130434783, "f1": 0.38461538461538464, "accuracy": 0.9373297002724795}, "tl_ugnayan": {"precision": 0.07, "recall": 0.21212121212121213, "f1": 0.10526315789473686, "accuracy": 0.8842297174111212}, "zh_gsd": {"precision": 0.5868421052631579, "recall": 0.5814863102998696, "f1": 0.5841519318926, "accuracy": 0.9460539460539461}, "zh_gsdsimp": {"precision": 0.59254327563249, "recall": 0.583224115334207, "f1": 0.5878467635402906, "accuracy": 0.944971694971695}, "hr_set": {"precision": 0.7613314447592068, "recall": 0.7662152530292231, "f1": 0.7637655417406749, "accuracy": 0.9730008244023083}, "da_ddt": {"precision": 0.6598984771573604, "recall": 0.5816554809843401, "f1": 0.6183115338882283, "accuracy": 0.9730619574977551}, "en_ewt": {"precision": 0.6018614270941055, "recall": 0.5349264705882353, "f1": 0.5664233576642335, "accuracy": 0.959716300753078}, "pt_bosque": {"precision": 0.6563285834031852, "recall": 0.6444444444444445, "f1": 0.6503322259136213, "accuracy": 0.9680118823358933}, "sr_set": {"precision": 0.8122786304604487, "recall": 0.8122786304604487, "f1": 0.8122786304604487, "accuracy": 0.9725943437527362}, "sk_snk": {"precision": 0.4213740458015267, "recall": 0.3016393442622951, "f1": 0.3515923566878981, "accuracy": 0.9186557788944724}, "sv_talbanken": {"precision": 0.7159090909090909, "recall": 0.6428571428571429, "f1": 0.6774193548387097, "accuracy": 0.9941110075084654}}
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d88ebd8aea8181029fc4bb765025cdb13035cc0cb4958bf29e283df92cbbdb11
3
  size 942800188
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94c798d33249bd160639357d57ccdb1a59f3a5b1a1298ea3c4b88c466cb4bcc5
3
  size 942800188
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:74741181ca413de56bab26118eae716bf783061c7ea5b807ba8d2c0d720fea36
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0660f7adcec985c2b6cc9209ba76316087b5b519afd5f5266cf9963ef1d585e2
3
  size 5304