SimoneJLaudani commited on
Commit
9958ce1
·
verified ·
1 Parent(s): a841c62

End of training

Browse files
README.md CHANGED
@@ -20,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.7582
24
- - Precision: 0.8478
25
- - Recall: 0.8448
26
- - F1: 0.8446
27
- - Accuracy: 0.8448
28
 
29
  ## Model description
30
 
@@ -43,7 +43,7 @@ More information needed
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
46
- - learning_rate: 5e-05
47
  - train_batch_size: 8
48
  - eval_batch_size: 8
49
  - seed: 42
@@ -55,53 +55,53 @@ The following hyperparameters were used during training:
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
- | 1.6945 | 0.08 | 30 | 1.5762 | 0.5100 | 0.5115 | 0.5001 | 0.5115 |
59
- | 1.4363 | 0.17 | 60 | 1.3360 | 0.5888 | 0.5697 | 0.5483 | 0.5697 |
60
- | 1.1347 | 0.25 | 90 | 1.0684 | 0.7150 | 0.6966 | 0.6943 | 0.6966 |
61
- | 0.8997 | 0.34 | 120 | 0.8797 | 0.7419 | 0.7337 | 0.7333 | 0.7337 |
62
- | 0.7458 | 0.42 | 150 | 0.8047 | 0.7659 | 0.7354 | 0.7328 | 0.7354 |
63
- | 0.8419 | 0.51 | 180 | 0.7057 | 0.7912 | 0.7795 | 0.7806 | 0.7795 |
64
- | 0.5818 | 0.59 | 210 | 0.7138 | 0.7896 | 0.7831 | 0.7774 | 0.7831 |
65
- | 0.6419 | 0.67 | 240 | 0.6906 | 0.8012 | 0.7848 | 0.7806 | 0.7848 |
66
- | 0.7079 | 0.76 | 270 | 0.6524 | 0.8114 | 0.7972 | 0.7953 | 0.7972 |
67
- | 0.5156 | 0.84 | 300 | 0.6335 | 0.8372 | 0.8148 | 0.8134 | 0.8148 |
68
- | 0.5888 | 0.93 | 330 | 0.6472 | 0.8282 | 0.8166 | 0.8162 | 0.8166 |
69
- | 0.4568 | 1.01 | 360 | 0.5767 | 0.8562 | 0.8430 | 0.8443 | 0.8430 |
70
- | 0.3246 | 1.1 | 390 | 0.6753 | 0.8152 | 0.8042 | 0.8042 | 0.8042 |
71
- | 0.3564 | 1.18 | 420 | 0.5988 | 0.8273 | 0.8166 | 0.8172 | 0.8166 |
72
- | 0.217 | 1.26 | 450 | 0.6661 | 0.8206 | 0.8078 | 0.8084 | 0.8078 |
73
- | 0.3266 | 1.35 | 480 | 0.6623 | 0.8255 | 0.8095 | 0.8099 | 0.8095 |
74
- | 0.3483 | 1.43 | 510 | 0.6147 | 0.8289 | 0.8219 | 0.8220 | 0.8219 |
75
- | 0.3132 | 1.52 | 540 | 0.6314 | 0.8531 | 0.8466 | 0.8469 | 0.8466 |
76
- | 0.3595 | 1.6 | 570 | 0.6002 | 0.8400 | 0.8360 | 0.8367 | 0.8360 |
77
- | 0.4029 | 1.69 | 600 | 0.6517 | 0.8375 | 0.8307 | 0.8311 | 0.8307 |
78
- | 0.2137 | 1.77 | 630 | 0.6710 | 0.8504 | 0.8448 | 0.8460 | 0.8448 |
79
- | 0.3862 | 1.85 | 660 | 0.6675 | 0.8385 | 0.8307 | 0.8320 | 0.8307 |
80
- | 0.2409 | 1.94 | 690 | 0.6429 | 0.8518 | 0.8466 | 0.8476 | 0.8466 |
81
- | 0.2348 | 2.02 | 720 | 0.6541 | 0.8445 | 0.8360 | 0.8376 | 0.8360 |
82
- | 0.059 | 2.11 | 750 | 0.6353 | 0.8591 | 0.8519 | 0.8520 | 0.8519 |
83
- | 0.1164 | 2.19 | 780 | 0.7147 | 0.8364 | 0.8342 | 0.8326 | 0.8342 |
84
- | 0.1661 | 2.28 | 810 | 0.6468 | 0.8584 | 0.8519 | 0.8528 | 0.8519 |
85
- | 0.1886 | 2.36 | 840 | 0.6984 | 0.8597 | 0.8501 | 0.8510 | 0.8501 |
86
- | 0.1226 | 2.44 | 870 | 0.7298 | 0.8405 | 0.8342 | 0.8344 | 0.8342 |
87
- | 0.0909 | 2.53 | 900 | 0.7092 | 0.8485 | 0.8466 | 0.8469 | 0.8466 |
88
- | 0.1675 | 2.61 | 930 | 0.6954 | 0.8528 | 0.8466 | 0.8476 | 0.8466 |
89
- | 0.1521 | 2.7 | 960 | 0.7429 | 0.8453 | 0.8413 | 0.8415 | 0.8413 |
90
- | 0.1005 | 2.78 | 990 | 0.7826 | 0.8384 | 0.8307 | 0.8311 | 0.8307 |
91
- | 0.2409 | 2.87 | 1020 | 0.8100 | 0.8325 | 0.8219 | 0.8224 | 0.8219 |
92
- | 0.0748 | 2.95 | 1050 | 0.7315 | 0.8418 | 0.8360 | 0.8358 | 0.8360 |
93
- | 0.0819 | 3.03 | 1080 | 0.7462 | 0.8433 | 0.8395 | 0.8399 | 0.8395 |
94
- | 0.0813 | 3.12 | 1110 | 0.7432 | 0.8460 | 0.8430 | 0.8428 | 0.8430 |
95
- | 0.0396 | 3.2 | 1140 | 0.7130 | 0.8456 | 0.8430 | 0.8430 | 0.8430 |
96
- | 0.0846 | 3.29 | 1170 | 0.7621 | 0.8403 | 0.8377 | 0.8376 | 0.8377 |
97
- | 0.0455 | 3.37 | 1200 | 0.7667 | 0.8373 | 0.8342 | 0.8341 | 0.8342 |
98
- | 0.022 | 3.46 | 1230 | 0.7427 | 0.8379 | 0.8342 | 0.8342 | 0.8342 |
99
- | 0.0586 | 3.54 | 1260 | 0.7552 | 0.8432 | 0.8377 | 0.8377 | 0.8377 |
100
- | 0.0423 | 3.62 | 1290 | 0.7524 | 0.8443 | 0.8395 | 0.8393 | 0.8395 |
101
- | 0.0366 | 3.71 | 1320 | 0.7506 | 0.8462 | 0.8430 | 0.8429 | 0.8430 |
102
- | 0.0435 | 3.79 | 1350 | 0.7439 | 0.8454 | 0.8430 | 0.8428 | 0.8430 |
103
- | 0.046 | 3.88 | 1380 | 0.7534 | 0.8452 | 0.8430 | 0.8429 | 0.8430 |
104
- | 0.0279 | 3.96 | 1410 | 0.7576 | 0.8445 | 0.8413 | 0.8412 | 0.8413 |
105
 
106
 
107
  ### Framework versions
 
20
 
21
  This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 1.0618
24
+ - Precision: 0.8364
25
+ - Recall: 0.8325
26
+ - F1: 0.8329
27
+ - Accuracy: 0.8325
28
 
29
  ## Model description
30
 
 
43
  ### Training hyperparameters
44
 
45
  The following hyperparameters were used during training:
46
+ - learning_rate: 2e-05
47
  - train_batch_size: 8
48
  - eval_batch_size: 8
49
  - seed: 42
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
+ | 0.0336 | 0.08 | 30 | 0.8889 | 0.8378 | 0.8272 | 0.8274 | 0.8272 |
59
+ | 0.0653 | 0.17 | 60 | 0.8057 | 0.8477 | 0.8448 | 0.8446 | 0.8448 |
60
+ | 0.0424 | 0.25 | 90 | 0.8184 | 0.8437 | 0.8430 | 0.8430 | 0.8430 |
61
+ | 0.0725 | 0.34 | 120 | 0.8139 | 0.8490 | 0.8430 | 0.8432 | 0.8430 |
62
+ | 0.0887 | 0.42 | 150 | 0.8627 | 0.8411 | 0.8342 | 0.8341 | 0.8342 |
63
+ | 0.0441 | 0.51 | 180 | 0.8865 | 0.8505 | 0.8413 | 0.8420 | 0.8413 |
64
+ | 0.0085 | 0.59 | 210 | 0.8662 | 0.8458 | 0.8430 | 0.8428 | 0.8430 |
65
+ | 0.0785 | 0.67 | 240 | 0.9317 | 0.8431 | 0.8395 | 0.8398 | 0.8395 |
66
+ | 0.0075 | 0.76 | 270 | 0.9654 | 0.8538 | 0.8448 | 0.8446 | 0.8448 |
67
+ | 0.0259 | 0.84 | 300 | 0.9987 | 0.8405 | 0.8342 | 0.8342 | 0.8342 |
68
+ | 0.0672 | 0.93 | 330 | 1.0019 | 0.8394 | 0.8342 | 0.8346 | 0.8342 |
69
+ | 0.017 | 1.01 | 360 | 1.0248 | 0.8384 | 0.8325 | 0.8332 | 0.8325 |
70
+ | 0.0049 | 1.1 | 390 | 1.0067 | 0.8474 | 0.8377 | 0.8393 | 0.8377 |
71
+ | 0.0296 | 1.18 | 420 | 1.0234 | 0.8438 | 0.8377 | 0.8385 | 0.8377 |
72
+ | 0.0349 | 1.26 | 450 | 1.0185 | 0.8277 | 0.8219 | 0.8223 | 0.8219 |
73
+ | 0.0025 | 1.35 | 480 | 1.0467 | 0.8239 | 0.8166 | 0.8171 | 0.8166 |
74
+ | 0.0041 | 1.43 | 510 | 1.0966 | 0.8315 | 0.8236 | 0.8243 | 0.8236 |
75
+ | 0.0017 | 1.52 | 540 | 1.0549 | 0.8323 | 0.8272 | 0.8279 | 0.8272 |
76
+ | 0.0361 | 1.6 | 570 | 1.0055 | 0.8519 | 0.8483 | 0.8490 | 0.8483 |
77
+ | 0.0644 | 1.69 | 600 | 1.1315 | 0.8371 | 0.8325 | 0.8322 | 0.8325 |
78
+ | 0.0016 | 1.77 | 630 | 1.1434 | 0.8244 | 0.8183 | 0.8188 | 0.8183 |
79
+ | 0.0405 | 1.85 | 660 | 1.0628 | 0.8326 | 0.8272 | 0.8276 | 0.8272 |
80
+ | 0.0005 | 1.94 | 690 | 1.0394 | 0.8391 | 0.8342 | 0.8348 | 0.8342 |
81
+ | 0.0324 | 2.02 | 720 | 1.1081 | 0.8316 | 0.8254 | 0.8264 | 0.8254 |
82
+ | 0.0012 | 2.11 | 750 | 1.0663 | 0.8354 | 0.8325 | 0.8323 | 0.8325 |
83
+ | 0.005 | 2.19 | 780 | 1.0777 | 0.8335 | 0.8307 | 0.8304 | 0.8307 |
84
+ | 0.0264 | 2.28 | 810 | 1.0483 | 0.8361 | 0.8325 | 0.8326 | 0.8325 |
85
+ | 0.0431 | 2.36 | 840 | 1.0193 | 0.8473 | 0.8430 | 0.8435 | 0.8430 |
86
+ | 0.0004 | 2.44 | 870 | 1.0411 | 0.8457 | 0.8413 | 0.8418 | 0.8413 |
87
+ | 0.0045 | 2.53 | 900 | 1.0604 | 0.8319 | 0.8289 | 0.8290 | 0.8289 |
88
+ | 0.0305 | 2.61 | 930 | 1.0808 | 0.8323 | 0.8289 | 0.8292 | 0.8289 |
89
+ | 0.0362 | 2.7 | 960 | 1.0466 | 0.8430 | 0.8395 | 0.8398 | 0.8395 |
90
+ | 0.0004 | 2.78 | 990 | 1.0518 | 0.8429 | 0.8395 | 0.8397 | 0.8395 |
91
+ | 0.0147 | 2.87 | 1020 | 1.0781 | 0.8397 | 0.8360 | 0.8361 | 0.8360 |
92
+ | 0.0034 | 2.95 | 1050 | 1.0696 | 0.8377 | 0.8342 | 0.8344 | 0.8342 |
93
+ | 0.0004 | 3.03 | 1080 | 1.0649 | 0.8395 | 0.8360 | 0.8362 | 0.8360 |
94
+ | 0.0063 | 3.12 | 1110 | 1.0614 | 0.8347 | 0.8325 | 0.8325 | 0.8325 |
95
+ | 0.0014 | 3.2 | 1140 | 1.0433 | 0.8367 | 0.8342 | 0.8345 | 0.8342 |
96
+ | 0.0181 | 3.29 | 1170 | 1.0559 | 0.8339 | 0.8307 | 0.8311 | 0.8307 |
97
+ | 0.0069 | 3.37 | 1200 | 1.0693 | 0.8320 | 0.8289 | 0.8293 | 0.8289 |
98
+ | 0.0004 | 3.46 | 1230 | 1.0666 | 0.8320 | 0.8289 | 0.8293 | 0.8289 |
99
+ | 0.0117 | 3.54 | 1260 | 1.0596 | 0.8316 | 0.8289 | 0.8292 | 0.8289 |
100
+ | 0.0211 | 3.62 | 1290 | 1.0679 | 0.8326 | 0.8289 | 0.8293 | 0.8289 |
101
+ | 0.0124 | 3.71 | 1320 | 1.0688 | 0.8359 | 0.8325 | 0.8328 | 0.8325 |
102
+ | 0.0177 | 3.79 | 1350 | 1.0676 | 0.8346 | 0.8307 | 0.8312 | 0.8307 |
103
+ | 0.0009 | 3.88 | 1380 | 1.0635 | 0.8364 | 0.8325 | 0.8329 | 0.8325 |
104
+ | 0.0012 | 3.96 | 1410 | 1.0616 | 0.8364 | 0.8325 | 0.8329 | 0.8325 |
105
 
106
 
107
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:58b2037ef2d1cffed5274c1cae15b5597c48f6ca89da010494239e083029d7fe
3
  size 267847948
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73325bc0fc693e9d328337ab8ac1990a8f75e75936320bc640a9798abbe66c51
3
  size 267847948
runs/Apr07_09-39-49_b6630175ab9e/events.out.tfevents.1712492291.b6630175ab9e.818.2 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0e4a7d26ba8f1d79e8d64b9398dcd8b82070f1aba157d430c48f177fc93235b2
3
- size 560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3cc91da22c5223a23d4947da0e02313134c48175421f8d21c81ea3dd0c394594
3
+ size 5324
runs/Apr07_12-35-57_b6630175ab9e/events.out.tfevents.1712493366.b6630175ab9e.818.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af382e88209d8cd4398e41ac6cf47d5bfea95558e13c36578d173b67796958e1
3
+ size 37255
runs/Apr07_12-35-57_b6630175ab9e/events.out.tfevents.1712503671.b6630175ab9e.818.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2452abe91d1c1bf677948a7760129ac0d4d57224c90d908b7c49f4e62058a724
3
+ size 560
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b669e89e2a8fdcaee1c1f6efd59357c30103856bedaea28eeac54b3a5c8a836f
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8c98b6fe46c73caddc21e882b59d79d4f0b3060ddbf635b1dbdeeb1b324933e
3
  size 4920