ececet commited on
Commit
69cb8a3
·
1 Parent(s): 0aae357

Delete oversampled_frozen_norbert2_256_0.0001_64.txt

Browse files
oversampled_frozen_norbert2_256_0.0001_64.txt DELETED
@@ -1,61 +0,0 @@
1
- Started at: 15:48:18
2
- norbert2, 0.0001, 256
3
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
4
- Epoch: 0
5
- Training loss: 0.6048455547202717 - MAE: 0.6266300187442916
6
- Validation loss : 0.4582846462726593 - MAE: 0.5477823136637207
7
- Epoch: 1
8
- Training loss: 0.4581279342824763 - MAE: 0.5446929326843661
9
- Validation loss : 0.43447837076689066 - MAE: 0.5302158168090491
10
- Epoch: 2
11
- Training loss: 0.439133803952824 - MAE: 0.5313758711028191
12
- Validation loss : 0.4172007551318721 - MAE: 0.5173525707183267
13
- Epoch: 3
14
- Training loss: 0.4256406155499545 - MAE: 0.5213601684170421
15
- Validation loss : 0.4036315475639544 - MAE: 0.5072787234745049
16
- Epoch: 4
17
- Training loss: 0.4128780511292544 - MAE: 0.5122914042413262
18
- Validation loss : 0.39381301246191325 - MAE: 0.4998854366792771
19
- Epoch: 5
20
- Training loss: 0.403443688696081 - MAE: 0.5064829263635311
21
- Validation loss : 0.3850055195783314 - MAE: 0.49343864394198494
22
- Epoch: 6
23
- Training loss: 0.3943689080801877 - MAE: 0.4988654481063448
24
- Validation loss : 0.3792250125031722 - MAE: 0.48865697242123873
25
- Epoch: 7
26
- Training loss: 0.39008136066523463 - MAE: 0.4961428843061405
27
- Validation loss : 0.37279942788575826 - MAE: 0.48401661704906546
28
- Epoch: 8
29
- Training loss: 0.381467575376684 - MAE: 0.4899541087772683
30
- Validation loss : 0.3682594236574675 - MAE: 0.48040131545362147
31
- Epoch: 9
32
- Training loss: 0.3751734440976923 - MAE: 0.48501710497031664
33
- Validation loss : 0.3617483942132247 - MAE: 0.4762219395614589
34
- Epoch: 10
35
- Started at: 17:35:48
36
- norbert2, 0.0001, 256
37
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
38
- Started at: 17:37:00
39
- norbert2, 0.0001, 256
40
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
41
- Started at: 17:37:31
42
- norbert2, 0.0001, 256
43
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
44
- Training loss: 0.3698521521958438 - MAE: 0.4828537843610755
45
- Validation loss : 0.35875062879763153 - MAE: 0.47371693001384463
46
- Epoch: 11
47
- Started at: 17:38:51
48
- norbert2, 0.0001, 256
49
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
50
- Started at: 17:40:23
51
- norbert2, 0.0001, 256
52
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
53
- Started at: 17:41:33
54
- norbert2, 0.0001, 256
55
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
56
- Started at: 00:24:31
57
- norbert2, 0.0001, 256
58
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})
59
- Started at: 00:50:10
60
- norbert2, 0.0001, 256
61
- ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {})