File size: 2,696 Bytes
6a0262b
91dd1ea
c10d212
5e0f2a7
a66f788
c8fa7de
74fd1b5
0b3ae78
6f51af3
06b5032
de16fed
28b8ba2
7c5a45e
bece3c4
8c64a1e
0808e54
0c4a126
7227538
3821e2f
fe2cf47
e65ae93
a5cc751
99d643d
e4ba947
c9ec556
a96e2e0
2ce3c51
02607a8
270aaa7
b393a05
b6f827d
1627b8c
25cd78a
fc07372
03beba3
3340cfa
b73265c
626fdfd
e05a46e
e73f52f
5c15c38
3494494
1ae31f6
3baadc9
177fdc1
d11ad39
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
Started at: 12:15:47
nb-bert-base, 0.001, 128
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.31982501298189164 - MAE: 0.4287338867738713
Validation loss : 0.16746729612350464 - MAE: 0.31536104399438897
Epoch: 1
Training loss: 0.16691179797053338 - MAE: 0.3121238207010245
Validation loss : 0.15773158739594853 - MAE: 0.3046954138541735
Epoch: 2
Training loss: 0.15689186841249467 - MAE: 0.3008464796601783
Validation loss : 0.15233191146570094 - MAE: 0.29779621490764496
Epoch: 3
Training loss: 0.15314741998910905 - MAE: 0.296528327871902
Validation loss : 0.1497987894450917 - MAE: 0.29459793118874833
Epoch: 4
Training loss: 0.14834036141633988 - MAE: 0.2910206240007747
Validation loss : 0.1474542622180546 - MAE: 0.2915430406480437
Epoch: 5
Training loss: 0.14879375964403152 - MAE: 0.2913913004246226
Validation loss : 0.14557460432543473 - MAE: 0.2891730110373066
Epoch: 6
Training loss: 0.14662521675229073 - MAE: 0.2886886095421301
Validation loss : 0.14476492475060856 - MAE: 0.28795673596095495
Epoch: 7
Training loss: 0.14347930431365966 - MAE: 0.2869532495194668
Validation loss : 0.1444830390460351 - MAE: 0.2884910243549571
Epoch: 8
Training loss: 0.14114041782915593 - MAE: 0.2831716332297515
Validation loss : 0.14459675462806926 - MAE: 0.28917789765778673
Epoch: 9
Training loss: 0.1433439274132252 - MAE: 0.2865429012477346
Validation loss : 0.14251315286930868 - MAE: 0.2853422469518363
Epoch: 10
Training loss: 0.1406331904232502 - MAE: 0.28132702192119613
Validation loss : 0.14170135689132354 - MAE: 0.2838292713935197
Epoch: 11
Training loss: 0.14126647457480432 - MAE: 0.28410774431585517
Validation loss : 0.1421863292069996 - MAE: 0.2853570444611879
Epoch: 12
Training loss: 0.13998315826058388 - MAE: 0.282938143440139
Validation loss : 0.1432714159874355 - MAE: 0.2874122980672003
Epoch: 13
Training loss: 0.1395747384428978 - MAE: 0.2831718714019129
Validation loss : 0.14076389372348785 - MAE: 0.283541297177118
Epoch: 14