File size: 3,065 Bytes
c9a16c9 a45f1ab 60c04ea 6b95f88 bcd2bae 603d241 3fe10e1 9512e59 8f0c868 45f8f55 4c9c30f 0987dd4 1de3ec9 6d37b16 19ce39c ec06058 b35f200 7127cfe 5b89c81 ce35c81 802d2e3 d4e1e09 9eea91a af3bad4 ab43f09 1c1c715 7dfd04f ea7074c 6200300 28e110f c9b72e4 2878c2d a78d5b2 91273d1 e2567d8 1cd7e44 6d686e2 5908cc1 f261c12 2b961a5 cae9ab9 b77aca7 6243913 aacdcd5 83318b1 6cce714 6ed416b 2c98c5c aa3bd50 d915ac9 4747d28 5cf3c1a cec557a 219898a 18f784a 091b700 29bba41 310780d 1fe92ac 245c177 a5094b2 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 |
Started at: 12:43:11 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '44815f7e109b53547cccdf3c6847f4c28b989816'}, {}) Epoch: 0 Training loss: 0.2008972004489991 - MSE: 0.33518312736635436 Validation loss : 0.1646703563630581 - MSE: 0.3085779357984263 Epoch: 1 Training loss: 0.1659574282256145 - MSE: 0.3064189085675543 Validation loss : 0.157762396974223 - MSE: 0.28898105177130284 Epoch: 2 Training loss: 0.16920435048856783 - MSE: 0.3099259860789372 Validation loss : 0.2597339808940887 - MSE: 0.41778062962694096 Epoch: 3 Training loss: 0.18403085151198997 - MSE: 0.3220045950051232 Validation loss : 0.2158472946711949 - MSE: 0.37326406195393896 Epoch: 4 Training loss: 0.18172540587996974 - MSE: 0.32080418783672165 Validation loss : 0.180368021769183 - MSE: 0.33231508495186324 Epoch: 5 Training loss: 0.17783939802241558 - MSE: 0.31700912658652863 Validation loss : 0.17726878098079135 - MSE: 0.33020463892025875 Epoch: 6 Training loss: 0.17681393569823609 - MSE: 0.3154852375991437 Validation loss : 0.17508771706904683 - MSE: 0.32495340255388455 Epoch: 7 Training loss: 0.18098733559684846 - MSE: 0.3195274022724451 Validation loss : 0.17311148579631533 - MSE: 0.3220135453306804 Epoch: 8 Training loss: 0.18061703660534423 - MSE: 0.3191405354137568 Validation loss : 0.17032980706010545 - MSE: 0.3176248594869061 Epoch: 9 Training loss: 0.18142556370172686 - MSE: 0.319926894033844 Validation loss : 0.1691651236798082 - MSE: 0.31562867300485126 Epoch: 10 Training loss: 0.18212312047631996 - MSE: 0.32070363034661087 Validation loss : 0.16798930146864482 - MSE: 0.31360456355614585 Epoch: 11 Training loss: 0.18196018481428183 - MSE: 0.3207149931643255 Validation loss : 0.16837686087403977 - MSE: 0.31434407412639953 Epoch: 12 Training loss: 0.1814197617248424 - MSE: 0.3202695122298144 Validation loss : 0.16799633439098086 - MSE: 0.3135975342526633 Epoch: 13 Training loss: 0.18225394370197093 - MSE: 0.3212486534658962 Validation loss : 0.16925174317189626 - MSE: 0.3159311910575655 Epoch: 14 Training loss: 0.18245843891958588 - MSE: 0.3215480212902345 Validation loss : 0.16824361362627574 - MSE: 0.31412619200834474 Epoch: 15 Training loss: 0.1803396447624975 - MSE: 0.3195030057718649 Validation loss : 0.1688083845589842 - MSE: 0.3150487916560711 Epoch: 16 Training loss: 0.17986665409311509 - MSE: 0.3195031756770499 Validation loss : 0.22056829929351807 - MSE: 0.37911737345857544 Epoch: 17 Training loss: 0.18849548813208794 - MSE: 0.3279959400740679 Validation loss : 0.1744179646883692 - MSE: 0.32440157115093565 Epoch: 18 Training loss: 0.18259398508997796 - MSE: 0.32226052051809423 Validation loss : 0.17137913001435143 - MSE: 0.32016465139874656 Epoch: 19 Training loss: 0.17060092156662524 - MSE: 0.3103650882393749 |