File size: 2,027 Bytes
ffdff51
ca93d39
8a9dc10
54e7f6f
8344db1
38f51a1
480fedd
f82f0a4
c45369e
5bcb629
96d8a52
f9c61ff
05f907e
a04b204
d682b41
656c0a1
30b5648
242651e
e148f83
7233b82
67b6d25
c14add0
2ee0943
ab40d12
df9b96f
4bc1456
ed075b0
e5cdd3f
4b32231
f3329cf
f98b900
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
Started at: 14:47:29
nb-bert-base, 5e-06, 128
({'_name_or_path': '/disk4/folder1/working/checkpoints/huggingface/native_pytorch/step4_8/', 'attention_probs_dropout_prob': 0.1, 'directionality': 'bidi', 'gradient_checkpointing': False, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'layer_norm_eps': 1e-12, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'pad_token_id': 0, 'pooler_fc_size': 768, 'pooler_num_attention_heads': 12, 'pooler_num_fc_layers': 3, 'pooler_size_per_head': 128, 'pooler_type': 'first_token_transform', 'position_embedding_type': 'absolute', 'type_vocab_size': 2, 'vocab_size': 119547, '_commit_hash': '82b194c0b3ea1fcad65f1eceee04adb26f9f71ac'}, {})
Epoch: 0
Training loss: 0.46638986387036063 - MAE: 0.5398754368007034
Validation loss : 0.2809767316322069 - MAE: 0.4124626148131228
Epoch: 1
Training loss: 0.21324598220261662 - MAE: 0.36342001646930033
Validation loss : 0.17091604342331757 - MAE: 0.32559828894035
Epoch: 2
Training loss: 0.15148694582960823 - MAE: 0.3101711238743909
Validation loss : 0.14158385267128815 - MAE: 0.29382737678226023
Epoch: 3
Training loss: 0.124982850253582 - MAE: 0.2813409221247067
Validation loss : 0.11344313339607136 - MAE: 0.25976834796862297
Epoch: 4
Training loss: 0.10359308848326856 - MAE: 0.25381723494987607
Validation loss : 0.10265538237384848 - MAE: 0.24230259616167318
Epoch: 5
Training loss: 0.08948654772883112 - MAE: 0.23498061801274692
Validation loss : 0.09245944788327089 - MAE: 0.2265152617992811
Epoch: 6
Training loss: 0.07607155100188472 - MAE: 0.2157861911084654
Validation loss : 0.08652295635358707 - MAE: 0.2184979559256568
Epoch: 7
Training loss: 0.06690294654531913 - MAE: 0.2011048192003143
Validation loss : 0.08006738347781671 - MAE: 0.20853752498405181
Epoch: 8
Training loss: 0.05949505604803562 - MAE: 0.1890938053594342
Validation loss : 0.06948294661737778 - MAE: 0.1920086602951222
Epoch: 9