File size: 2,848 Bytes
ffdff51
ca93d39
8a9dc10
54e7f6f
8344db1
38f51a1
480fedd
f82f0a4
c45369e
5bcb629
96d8a52
f9c61ff
05f907e
a04b204
d682b41
656c0a1
30b5648
242651e
e148f83
7233b82
67b6d25
c14add0
2ee0943
ab40d12
df9b96f
4bc1456
ed075b0
e5cdd3f
4b32231
f3329cf
f98b900
a6865f2
1ad68f3
a3244ec
41f96c1
ad2af4f
67df123
00e4480
87e8292
1ab51e5
c65a2e6
a64d778
9abe17c
e6cbaa5
850f5b8
97b8fdb
e8d10b7
56253a9
4e2e70f
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
Started at: 14:47:29
nb-bert-base, 5e-06, 128
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.46638986387036063 - MAE: 0.5398754368007034
Validation loss : 0.2809767316322069 - MAE: 0.4124626148131228
Epoch: 1
Training loss: 0.21324598220261662 - MAE: 0.36342001646930033
Validation loss : 0.17091604342331757 - MAE: 0.32559828894035
Epoch: 2
Training loss: 0.15148694582960823 - MAE: 0.3101711238743909
Validation loss : 0.14158385267128815 - MAE: 0.29382737678226023
Epoch: 3
Training loss: 0.124982850253582 - MAE: 0.2813409221247067
Validation loss : 0.11344313339607136 - MAE: 0.25976834796862297
Epoch: 4
Training loss: 0.10359308848326856 - MAE: 0.25381723494987607
Validation loss : 0.10265538237384848 - MAE: 0.24230259616167318
Epoch: 5
Training loss: 0.08948654772883112 - MAE: 0.23498061801274692
Validation loss : 0.09245944788327089 - MAE: 0.2265152617992811
Epoch: 6
Training loss: 0.07607155100188472 - MAE: 0.2157861911084654
Validation loss : 0.08652295635358707 - MAE: 0.2184979559256568
Epoch: 7
Training loss: 0.06690294654531913 - MAE: 0.2011048192003143
Validation loss : 0.08006738347781671 - MAE: 0.20853752498405181
Epoch: 8
Training loss: 0.05949505604803562 - MAE: 0.1890938053594342
Validation loss : 0.06948294661737778 - MAE: 0.1920086602951222
Epoch: 9
Training loss: 0.05425354472615502 - MAE: 0.17931217790377968
Validation loss : 0.06705979368573911 - MAE: 0.1872916947876361
Epoch: 10
Training loss: 0.051136760142716495 - MAE: 0.17425557358320554
Validation loss : 0.07039306645055075 - MAE: 0.1957787938835992
Epoch: 11
Training loss: 0.049823644300076096 - MAE: 0.17265991846906573
Validation loss : 0.059706183905537065 - MAE: 0.17449212548911355
Epoch: 12
Training loss: 0.04213783075525002 - MAE: 0.15748476174497095
Validation loss : 0.06050218903535121 - MAE: 0.17571373887778519
Epoch: 13
Training loss: 0.03851394113153219 - MAE: 0.1512245667258023
Validation loss : 0.05904384138616356 - MAE: 0.1736026045715516
Epoch: 14
Training loss: 0.03739640076052059 - MAE: 0.14833851531891173
Validation loss : 0.05757016124757561 - MAE: 0.17082927729316882
Epoch: 15