File size: 2,749 Bytes
1affc1c 18c0138 221e918 5272fa4 47622c4 d42f18b d6b9d9e 6a061eb 4c7031e fec7105 76fe34f 715d94a 78e089a e2810c5 7120b98 3d73bfd 8475310 1ce0d0f 0432dbd 02e32ff ac40506 d6157e1 9719423 221176c 6aa232b 1f1b310 dc07b22 f51ef5e 2594b9a cd2b674 ee5ae6c 3d6bb3e b5fb8d2 f425a3f 3b2e2e4 3842791 6b3ab6b 129a980 b768ab9 41920a2 bdb3a40 ce3c2a4 f3087db 372a28c b4bbd81 b7362ab a7c7a80 30933c7 1c92ebd 50adafe b2716b6 59610ad fbfae16 c639bcf 3d88cd4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 |
Started at: 12:52:23 norbert, 5e-06, 256 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.6053906446153468 - MAE: 0.6184159507940801 Validation loss : 0.4188641560705085 - MAE: 0.5190200870427507 Epoch: 1 Training loss: 0.37205150289969013 - MAE: 0.4891842565138049 Validation loss : 0.3361840279478776 - MAE: 0.4554968620483818 Epoch: 2 Training loss: 0.30589349486611106 - MAE: 0.43993173479624126 Validation loss : 0.27830397376888677 - MAE: 0.41290173204153435 Epoch: 3 Training loss: 0.2584832324223085 - MAE: 0.4034832139365259 Validation loss : 0.2353723849120893 - MAE: 0.378263007542243 Epoch: 4 Training loss: 0.2208004041151567 - MAE: 0.3714158687728299 Validation loss : 0.20094563145386546 - MAE: 0.35106286126524094 Epoch: 5 Training loss: 0.1931754020127383 - MAE: 0.3488251679890293 Validation loss : 0.17955732188726725 - MAE: 0.33182885819892266 Epoch: 6 Training loss: 0.17624386982484297 - MAE: 0.3333650190493288 Validation loss : 0.16801361818062632 - MAE: 0.3229406276757107 Epoch: 7 Training loss: 0.16203992610627954 - MAE: 0.3199497565391834 Validation loss : 0.1510195426250759 - MAE: 0.306880161949399 Epoch: 8 Training loss: 0.15154187828302385 - MAE: 0.3089620812443599 Validation loss : 0.14486609556173025 - MAE: 0.29924708380840065 Epoch: 9 Training loss: 0.14079513807188382 - MAE: 0.29892621874422587 Validation loss : 0.13840533440050326 - MAE: 0.290965015323921 Epoch: 10 Training loss: 0.13332889432256873 - MAE: 0.29082796894473617 Validation loss : 0.12829295977165825 - MAE: 0.27816580614753955 Epoch: 11 Training loss: 0.12650601890954105 - MAE: 0.28307433471783455 Validation loss : 0.12042002497535002 - MAE: 0.26958139319431695 Epoch: 12 Training loss: 0.1180860929868438 - MAE: 0.27275313703304305 Validation loss : 0.1134116300626805 - MAE: 0.2581940768065212 Epoch: 13 Training loss: 0.1120438659732992 - MAE: 0.2657403258096752 Validation loss : 0.11130064725875854 - MAE: 0.2551686039616152 Epoch: 14 Training loss: 0.10740146054462953 - MAE: 0.26015119794445635 Validation loss : 0.10768120618242967 - MAE: 0.2487607443147535 Epoch: 15 Training loss: 0.10318632572889327 - MAE: 0.2533595643372031 Validation loss : 0.10223148920034107 - MAE: 0.2425870576871514 Epoch: 16 Training loss: 0.09851837293668227 - MAE: 0.2475603386691453 Validation loss : 0.10070653259754181 - MAE: 0.2398464623664194 Epoch: 17 |