File size: 3,155 Bytes
9995053 17f2dfa 592498a 3f29404 7886495 b97b048 710cd61 fc09eaa a7981b5 eddb02f d50d711 b5b740a 90dff37 e1590a1 e255196 ffdb4db cb8a78b 7f303dd f80baa5 11b58ec 3c7bd9c 1a572d8 a2ada58 4acbf81 7f2d59d 33b9cda e6d3c03 d9422d0 562a3df 9f70bc6 6a5e155 ce722a4 9dc8bd8 42de57f 0c10325 1895edd 140dc3b ca71f3d 6649741 1f571b6 3d13143 f54692f 2c14963 c34683d 0401392 a5963de 0ffae1d f48f56c a827c86 e9d1008 a7fd2c2 f174b54 a0008ca 434922e 62f8acf 662212d 1747c46 8dcefcb cc7324a 31c6144 e578164 cd5550e 9d387f0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 |
Started at: 16:05:12 norbert, 0.001, 320 ({'architectures': ['BertForMaskedLM'], 'attention_probs_dropout_prob': 0.1, 'hidden_act': 'gelu', 'hidden_dropout_prob': 0.1, 'hidden_size': 768, 'initializer_range': 0.02, 'intermediate_size': 3072, 'max_position_embeddings': 512, 'model_type': 'bert', 'num_attention_heads': 12, 'num_hidden_layers': 12, 'type_vocab_size': 2, 'vocab_size': 32922, '_commit_hash': '075d4e3705390691013e859faffc5696d071e33b'}, {}) Epoch: 0 Training loss: 0.5762677684426307 - MAE: 0.6118854715373047 Validation loss : 0.21141972073486873 - MAE: 0.3609583366423162 Epoch: 1 Training loss: 0.21102749779820443 - MAE: 0.34836694451800443 Validation loss : 0.21293789786951883 - MAE: 0.36360750057199404 Epoch: 2 Training loss: 0.18259856328368188 - MAE: 0.3243689929907013 Validation loss : 0.18117137679031917 - MAE: 0.329247425803855 Epoch: 3 Training loss: 0.17436085492372513 - MAE: 0.31863324386216885 Validation loss : 0.17593795699732645 - MAE: 0.32252189332349507 Epoch: 4 Training loss: 0.17129840925335885 - MAE: 0.31424783128879047 Validation loss : 0.17302861383983068 - MAE: 0.31924465234263655 Epoch: 5 Training loss: 0.16920918673276902 - MAE: 0.31110102172938675 Validation loss : 0.17061551340988704 - MAE: 0.3167244912842947 Epoch: 6 Training loss: 0.1653466336429119 - MAE: 0.30718663648201594 Validation loss : 0.1688840091228485 - MAE: 0.3148823277205399 Epoch: 7 Training loss: 0.1616860084235668 - MAE: 0.30553906160014854 Validation loss : 0.16725684063775198 - MAE: 0.3129958080183389 Epoch: 8 Training loss: 0.1612413488328457 - MAE: 0.30383874548816886 Validation loss : 0.16608954327447073 - MAE: 0.31147184647570914 Epoch: 9 Training loss: 0.16023582145571708 - MAE: 0.30331546575363605 Validation loss : 0.1649702489376068 - MAE: 0.310041077480561 Epoch: 10 Training loss: 0.15981780216097832 - MAE: 0.3027028828885578 Validation loss : 0.16403222509792872 - MAE: 0.3092826478279413 Epoch: 11 Training loss: 0.15804845243692398 - MAE: 0.29963723801837655 Validation loss : 0.16327827743121556 - MAE: 0.3085516138884669 Epoch: 12 Training loss: 0.1555868938565254 - MAE: 0.2992816365654825 Validation loss : 0.16232714908463614 - MAE: 0.30687830980995734 Epoch: 13 Training loss: 0.15700081139802932 - MAE: 0.2990457346311504 Validation loss : 0.16192833653518132 - MAE: 0.3067905196244964 Epoch: 14 Training loss: 0.1548595778644085 - MAE: 0.29765682874344357 Validation loss : 0.1616982775075095 - MAE: 0.3064313819681474 Epoch: 15 Training loss: 0.15383974090218544 - MAE: 0.2965260924465664 Validation loss : 0.1607140749692917 - MAE: 0.30490487384563325 Epoch: 16 Training loss: 0.15208731070160866 - MAE: 0.29383326150297523 Validation loss : 0.16031161376408168 - MAE: 0.30458542586117876 Epoch: 17 Training loss: 0.15210237205028534 - MAE: 0.2941252809307741 Validation loss : 0.16001720300742558 - MAE: 0.30438225845624445 Epoch: 18 Training loss: 0.15230538696050644 - MAE: 0.29471632999962655 Validation loss : 0.15990264500890458 - MAE: 0.30443557919219594 Epoch: 19 Training loss: 0.1517851881682873 - MAE: 0.2945964528897411 Validation loss : 0.15933816773550852 - MAE: 0.30348587056502435 |