File size: 3,085 Bytes
fd1f392 3818fca 74d56c4 38dbb06 170c32e e54303a 6d84434 fb53ab9 f2eea60 a955de2 3c17317 e344626 a2fa97c 180c5c1 f8ae289 4ed60da 581031b ff1c738 770c353 dcce380 e4d03fe 30fe58f 0b11992 13127f0 2606fa1 3fc53eb 125e149 cb46dd4 19bb37b a190d71 c54bf5c a33fa5a d1ef7d5 1aab339 e575809 1fa4dbb cf93c8f 0424846 87fe95e e144505 5c44a5e 79d3460 f8f3f82 16c5678 407a522 b64fc47 c38e4cc bcc6b65 88ae8ac f10daa8 76a0720 9437d18 e19f589 a7eab35 433b3b3 a6fe8ca 8b3d5ea b223896 574fb16 2eaf64b d21a09e 234ca7b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 |
Started at: 11:12:13 norbert2, 5e-06, 256 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.41463836485689337 - MAE: 0.5087427320675125 Validation loss : 0.2686559503015719 - MAE: 0.4043045732741782 Epoch: 1 Training loss: 0.24431555514985864 - MAE: 0.38740748680688086 Validation loss : 0.18667236833195938 - MAE: 0.33838491992877257 Epoch: 2 Training loss: 0.18506982191042468 - MAE: 0.3382457084132687 Validation loss : 0.15371774607583097 - MAE: 0.3074420312338156 Epoch: 3 Training loss: 0.1546944863416932 - MAE: 0.31008215112442716 Validation loss : 0.133816171241434 - MAE: 0.2858432389366333 Epoch: 4 Training loss: 0.13350679468024862 - MAE: 0.28923511177879146 Validation loss : 0.10932125621720364 - MAE: 0.256454776400781 Epoch: 5 Training loss: 0.1168869366699999 - MAE: 0.26986456802804265 Validation loss : 0.09649829840973805 - MAE: 0.23724551405129826 Epoch: 6 Training loss: 0.10431941856037487 - MAE: 0.2534326974131011 Validation loss : 0.08749974440587194 - MAE: 0.22284109841832134 Epoch: 7 Training loss: 0.08919526636600494 - MAE: 0.2336376825413137 Validation loss : 0.07944989831824052 - MAE: 0.2086082506448854 Epoch: 8 Training loss: 0.08102413564920426 - MAE: 0.22234788799560584 Validation loss : 0.07408433172263597 - MAE: 0.19957130841128806 Epoch: 9 Training loss: 0.07282835455103354 - MAE: 0.20905587992288474 Validation loss : 0.07321358158400185 - MAE: 0.1958791246236593 Epoch: 10 Training loss: 0.06805636808276176 - MAE: 0.2019703178512447 Validation loss : 0.06818335130810738 - MAE: 0.18784236918567274 Epoch: 11 Training loss: 0.06315255009315231 - MAE: 0.19438868563685022 Validation loss : 0.061382615840748736 - MAE: 0.17627332670686088 Epoch: 12 Training loss: 0.05669153705239296 - MAE: 0.18379919555202343 Validation loss : 0.06136786408330265 - MAE: 0.17450401481941827 Epoch: 13 Training loss: 0.05257220755923878 - MAE: 0.17699028630098032 Validation loss : 0.060792625734680576 - MAE: 0.17298781770388472 Epoch: 14 Training loss: 0.04955695366317576 - MAE: 0.17188387365707222 Validation loss : 0.059647708543037116 - MAE: 0.17080333908403275 Epoch: 15 Training loss: 0.047338411550630226 - MAE: 0.16792417539661747 Validation loss : 0.05730088721764715 - MAE: 0.16665574399906258 Epoch: 16 Training loss: 0.044937838884917175 - MAE: 0.1632461903660146 Validation loss : 0.05832104149617647 - MAE: 0.16912266812807178 Epoch: 17 Training loss: 0.04221029823476618 - MAE: 0.15887206312399307 Validation loss : 0.056816791037195606 - MAE: 0.16741534462301444 Epoch: 18 Training loss: 0.04010521661151539 - MAE: 0.15510032220620396 Validation loss : 0.05663576937819782 - MAE: 0.1699596096888483 Stopped after 18 epochs. Prediction MAE: 0.1728 |