File size: 3,019 Bytes
c20f414 b2fc17b 01c12e3 2ecef43 396ab70 56862d9 d72bc25 6d53e85 ee873ff 1030e04 db424fe 089be7b b3f91d8 4d1b736 c3d23da 6d355ea 7b31ad3 51fab6f daf9fa5 9132459 add69de af5b932 fe040a9 6d84d6a 04fb77f 45d9452 8cc1242 fe42d1f 8599816 a62e89a d200d6e 1035181 8845b3f 8e8bfbb 5900d89 5830f93 e9ebafd c2ca7e3 279704b 5163a3d 19938fd cd604b9 e0ed1d1 17ec8ee daa5c03 ca5c403 029678b 97bee02 4a4b6c9 1aa6200 dc81459 cbb6fd8 4ad3b58 0ee94a7 11f6541 0cda7c5 d1cd881 fb70416 b07f46f 202a02a 124b5c4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 |
Started at: 12:07:49 norbert2, 1e-06, 256 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 50104, '_commit_hash': 'f22bb47f536f62edfcd86ca9320ade990eafbe22'}, {}) Epoch: 0 Training loss: 0.30783953905105593 - MAE: 0.43719580843584505 Validation loss : 0.21560138960679373 - MAE: 0.3596361274665927 Epoch: 1 Training loss: 0.2266489964723587 - MAE: 0.3668055945655404 Validation loss : 0.2019458363453547 - MAE: 0.34808209428471554 Epoch: 2 Training loss: 0.19879530966281891 - MAE: 0.3426665569097344 Validation loss : 0.18689965373939937 - MAE: 0.3323873764788491 Epoch: 3 Training loss: 0.19044960021972657 - MAE: 0.3341355262436057 Validation loss : 0.17861351039674547 - MAE: 0.32349605653410407 Epoch: 4 Training loss: 0.18339800894260405 - MAE: 0.3276666275734149 Validation loss : 0.17232751680745018 - MAE: 0.316849883292955 Epoch: 5 Training loss: 0.17674655199050904 - MAE: 0.32018679623163826 Validation loss : 0.16676822139157188 - MAE: 0.31070956479175915 Epoch: 6 Training loss: 0.17175395369529725 - MAE: 0.3132561194660417 Validation loss : 0.16173937420050302 - MAE: 0.30571354481973123 Epoch: 7 Training loss: 0.16629139453172684 - MAE: 0.3090430212290697 Validation loss : 0.1575402236647076 - MAE: 0.30089353544828384 Epoch: 8 Training loss: 0.16224935233592988 - MAE: 0.3062369513645003 Validation loss : 0.15363233288129172 - MAE: 0.29686793533262296 Epoch: 9 Training loss: 0.15677126228809357 - MAE: 0.3001369819741856 Validation loss : 0.1501677847570843 - MAE: 0.2930093392614317 Epoch: 10 Training loss: 0.1522154378890991 - MAE: 0.2944758665139634 Validation loss : 0.1469805646273825 - MAE: 0.2896878885467087 Epoch: 11 Training loss: 0.14997684746980666 - MAE: 0.2930107961544775 Validation loss : 0.14408415887090895 - MAE: 0.28684108941263253 Epoch: 12 Training loss: 0.1469377765059471 - MAE: 0.29026566478408594 Validation loss : 0.14135516103770998 - MAE: 0.2838060469660866 Epoch: 13 Training loss: 0.1430877184867859 - MAE: 0.2856650031673201 Validation loss : 0.1389518520898289 - MAE: 0.2805393504731374 Epoch: 14 Training loss: 0.14132562696933745 - MAE: 0.2830484027164898 Validation loss : 0.1361759594745106 - MAE: 0.27795312559437724 Epoch: 15 Training loss: 0.1375071695446968 - MAE: 0.2793070474640609 Validation loss : 0.1337506083978547 - MAE: 0.2752909288563907 Epoch: 16 Training loss: 0.13490652769804 - MAE: 0.27698289906517004 Validation loss : 0.13174073066976336 - MAE: 0.27296385689056446 Epoch: 17 Training loss: 0.13232652187347413 - MAE: 0.2744538024977914 Validation loss : 0.12997501095136008 - MAE: 0.2712409643126546 Epoch: 18 Training loss: 0.12868250519037247 - MAE: 0.27025114055170957 Validation loss : 0.12838428136375216 - MAE: 0.269251110723689 Epoch: 19 |