File size: 2,888 Bytes
7db66ec
1ade577
9457c05
cf6c14b
3c1f5ba
b80ec74
45a213a
a6c6390
f842217
f935464
0c1ca99
39c51b0
90fc023
3c09443
359c2b1
66f44ee
e1e1659
ca7d999
191f729
d0ff4b6
627f2bf
1917344
13010fb
2a6e92b
4c3b633
6f8807e
912542f
863c4c2
2b48916
a64a4c8
d90fe4a
121e0e0
8223f8c
d87ef6b
3fc9afa
aac7be9
c34ff6e
e953db0
db83c3b
c99aac5
fb5f173
5c3d37a
7a8dfe2
8cd8763
75df4b2
f8ffcea
ceacc6f
0c84e4a
ed6891e
667e513
f62d511
5f36335
a2f5139
9100b82
34666f6
9d8cee2
23837db
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
Started at: 14:37:27
norbert, 5e-06, 128
({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {})
Epoch: 0
Training loss: 0.49241517592560163 - MAE: 0.5587459042828556
Validation loss : 0.35868610159770864 - MAE: 0.4753577830860895
Epoch: 1
Training loss: 0.3175132931633429 - MAE: 0.445445539977737
Validation loss : 0.25557071695456635 - MAE: 0.3910300926165894
Epoch: 2
Training loss: 0.23261894990097393 - MAE: 0.3786371806608561
Validation loss : 0.20166053240363663 - MAE: 0.3493823568582558
Epoch: 3
Training loss: 0.19200442243706095 - MAE: 0.3470611576431532
Validation loss : 0.1713124976770298 - MAE: 0.3275480687779099
Epoch: 4
Training loss: 0.17018077576702292 - MAE: 0.3278047750176803
Validation loss : 0.1574633888698913 - MAE: 0.31174332485895984
Epoch: 5
Training loss: 0.1535857619209723 - MAE: 0.31193839952824526
Validation loss : 0.15219828848903244 - MAE: 0.3075684372060492
Epoch: 6
Training loss: 0.14406093453819102 - MAE: 0.30176564464643024
Validation loss : 0.13898070478761518 - MAE: 0.29244764946070223
Epoch: 7
Training loss: 0.13362374928864565 - MAE: 0.29026767344805254
Validation loss : 0.13048673622511528 - MAE: 0.2813505453732496
Epoch: 8
Training loss: 0.12214575077999722 - MAE: 0.27611727530666424
Validation loss : 0.12706001525795138 - MAE: 0.27665191233280867
Epoch: 9
Training loss: 0.11591105806556615 - MAE: 0.26916014538434074
Validation loss : 0.11628925296905879 - MAE: 0.2643442778374055
Epoch: 10
Training loss: 0.10905246978456323 - MAE: 0.2604420025940941
Validation loss : 0.11431488414873948 - MAE: 0.25982212380330555
Epoch: 11
Training loss: 0.10331852971152826 - MAE: 0.2536274391469398
Validation loss : 0.10815783006113928 - MAE: 0.2503694937076749
Epoch: 12
Training loss: 0.09894988679073073 - MAE: 0.24673591588906577
Validation loss : 0.10309344207918322 - MAE: 0.24489055743019184
Epoch: 13
Training loss: 0.09555300352248279 - MAE: 0.2422491618605809
Validation loss : 0.09517849277000169 - MAE: 0.23459944038365538
Epoch: 14
Training loss: 0.09006806313991547 - MAE: 0.2344158579264854
Validation loss : 0.09714008787193813 - MAE: 0.23673838294794178
Epoch: 15
Training loss: 0.08482879820195112 - MAE: 0.2277674742544599
Validation loss : 0.09199248536213024 - MAE: 0.22825399727560355
Epoch: 16
Training loss: 0.08355843451890078 - MAE: 0.22517612165359246
Validation loss : 0.08988457132835646 - MAE: 0.2259545000346762
Epoch: 17
Training loss: 0.07894870334050873 - MAE: 0.21817967396293864
Validation loss : 0.09049511318271225 - MAE: 0.22485460323935755