|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.08187912594033059, |
|
"eval_steps": 500, |
|
"global_step": 400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.00020469781485082646, |
|
"grad_norm": 1.694807571622714, |
|
"learning_rate": 1.3605442176870747e-08, |
|
"loss": 0.6346, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0004093956297016529, |
|
"grad_norm": 1.7333604659657242, |
|
"learning_rate": 2.7210884353741493e-08, |
|
"loss": 0.5956, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0006140934445524794, |
|
"grad_norm": 1.675170156463719, |
|
"learning_rate": 4.081632653061224e-08, |
|
"loss": 0.6494, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.0008187912594033058, |
|
"grad_norm": 1.4535412459526658, |
|
"learning_rate": 5.442176870748299e-08, |
|
"loss": 0.6097, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.0010234890742541324, |
|
"grad_norm": 1.6204240919715567, |
|
"learning_rate": 6.802721088435375e-08, |
|
"loss": 0.6259, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0012281868891049587, |
|
"grad_norm": 1.6382909584778356, |
|
"learning_rate": 8.163265306122448e-08, |
|
"loss": 0.6049, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.0014328847039557853, |
|
"grad_norm": 1.6591863389965569, |
|
"learning_rate": 9.523809523809523e-08, |
|
"loss": 0.6093, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.0016375825188066117, |
|
"grad_norm": 1.529188807208944, |
|
"learning_rate": 1.0884353741496597e-07, |
|
"loss": 0.625, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.0018422803336574382, |
|
"grad_norm": 1.7414059653199376, |
|
"learning_rate": 1.2244897959183673e-07, |
|
"loss": 0.6148, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.002046978148508265, |
|
"grad_norm": 1.6622320550472127, |
|
"learning_rate": 1.360544217687075e-07, |
|
"loss": 0.5797, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.002251675963359091, |
|
"grad_norm": 1.6508189144245708, |
|
"learning_rate": 1.4965986394557823e-07, |
|
"loss": 0.6484, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.0024563737782099175, |
|
"grad_norm": 1.7202133207821506, |
|
"learning_rate": 1.6326530612244896e-07, |
|
"loss": 0.6216, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.0026610715930607443, |
|
"grad_norm": 1.5235060143030161, |
|
"learning_rate": 1.7687074829931972e-07, |
|
"loss": 0.6452, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.0028657694079115706, |
|
"grad_norm": 1.5363560852946705, |
|
"learning_rate": 1.9047619047619045e-07, |
|
"loss": 0.6405, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.003070467222762397, |
|
"grad_norm": 1.6730999257251689, |
|
"learning_rate": 2.0408163265306121e-07, |
|
"loss": 0.6497, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0032751650376132233, |
|
"grad_norm": 1.5758327138243107, |
|
"learning_rate": 2.1768707482993195e-07, |
|
"loss": 0.6336, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.00347986285246405, |
|
"grad_norm": 1.5492535238923828, |
|
"learning_rate": 2.312925170068027e-07, |
|
"loss": 0.6037, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0036845606673148765, |
|
"grad_norm": 1.6696926699572276, |
|
"learning_rate": 2.4489795918367347e-07, |
|
"loss": 0.6139, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.003889258482165703, |
|
"grad_norm": 1.6544769292475, |
|
"learning_rate": 2.5850340136054423e-07, |
|
"loss": 0.6315, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.00409395629701653, |
|
"grad_norm": 1.6860896587110352, |
|
"learning_rate": 2.72108843537415e-07, |
|
"loss": 0.6324, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0042986541118673555, |
|
"grad_norm": 1.5451565683271684, |
|
"learning_rate": 2.857142857142857e-07, |
|
"loss": 0.6227, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.004503351926718182, |
|
"grad_norm": 1.5658617650258626, |
|
"learning_rate": 2.9931972789115645e-07, |
|
"loss": 0.5873, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.004708049741569009, |
|
"grad_norm": 1.6014268573351107, |
|
"learning_rate": 3.129251700680272e-07, |
|
"loss": 0.6066, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.004912747556419835, |
|
"grad_norm": 1.560124972985785, |
|
"learning_rate": 3.265306122448979e-07, |
|
"loss": 0.6062, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.005117445371270662, |
|
"grad_norm": 1.6662191020723245, |
|
"learning_rate": 3.401360544217687e-07, |
|
"loss": 0.5968, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.0053221431861214885, |
|
"grad_norm": 1.7137595321931511, |
|
"learning_rate": 3.5374149659863944e-07, |
|
"loss": 0.6325, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.0055268410009723145, |
|
"grad_norm": 1.570188696390546, |
|
"learning_rate": 3.673469387755102e-07, |
|
"loss": 0.6375, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.005731538815823141, |
|
"grad_norm": 1.5585726347237283, |
|
"learning_rate": 3.809523809523809e-07, |
|
"loss": 0.6216, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.005936236630673967, |
|
"grad_norm": 1.4585947364133294, |
|
"learning_rate": 3.9455782312925167e-07, |
|
"loss": 0.5975, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.006140934445524794, |
|
"grad_norm": 1.5313204391085877, |
|
"learning_rate": 4.0816326530612243e-07, |
|
"loss": 0.6461, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.006345632260375621, |
|
"grad_norm": 1.4690318982818216, |
|
"learning_rate": 4.217687074829932e-07, |
|
"loss": 0.6054, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.006550330075226447, |
|
"grad_norm": 1.5256726957060316, |
|
"learning_rate": 4.353741496598639e-07, |
|
"loss": 0.6507, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.006755027890077273, |
|
"grad_norm": 1.541131533646238, |
|
"learning_rate": 4.4897959183673465e-07, |
|
"loss": 0.6185, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.0069597257049281, |
|
"grad_norm": 1.5233070330699345, |
|
"learning_rate": 4.625850340136054e-07, |
|
"loss": 0.6541, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.007164423519778926, |
|
"grad_norm": 1.4300240195672376, |
|
"learning_rate": 4.761904761904761e-07, |
|
"loss": 0.6156, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.007369121334629753, |
|
"grad_norm": 1.3386118655838508, |
|
"learning_rate": 4.897959183673469e-07, |
|
"loss": 0.5943, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.00757381914948058, |
|
"grad_norm": 1.3270928639031936, |
|
"learning_rate": 5.034013605442177e-07, |
|
"loss": 0.6139, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.007778516964331406, |
|
"grad_norm": 1.3116129547815811, |
|
"learning_rate": 5.170068027210885e-07, |
|
"loss": 0.6119, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.007983214779182232, |
|
"grad_norm": 1.2451803048665653, |
|
"learning_rate": 5.306122448979592e-07, |
|
"loss": 0.5463, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.00818791259403306, |
|
"grad_norm": 1.2351915311334578, |
|
"learning_rate": 5.4421768707483e-07, |
|
"loss": 0.5762, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.008392610408883885, |
|
"grad_norm": 1.3425104949855924, |
|
"learning_rate": 5.578231292517006e-07, |
|
"loss": 0.5866, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.008597308223734711, |
|
"grad_norm": 1.3464358053560985, |
|
"learning_rate": 5.714285714285714e-07, |
|
"loss": 0.6134, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.008802006038585539, |
|
"grad_norm": 1.3225968492677225, |
|
"learning_rate": 5.850340136054421e-07, |
|
"loss": 0.6034, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.009006703853436365, |
|
"grad_norm": 1.2483346937333237, |
|
"learning_rate": 5.986394557823129e-07, |
|
"loss": 0.5495, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.00921140166828719, |
|
"grad_norm": 1.1648688787665145, |
|
"learning_rate": 6.122448979591837e-07, |
|
"loss": 0.616, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.009416099483138018, |
|
"grad_norm": 1.2616996144445687, |
|
"learning_rate": 6.258503401360544e-07, |
|
"loss": 0.57, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.009620797297988844, |
|
"grad_norm": 1.3108653064941627, |
|
"learning_rate": 6.394557823129252e-07, |
|
"loss": 0.5814, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.00982549511283967, |
|
"grad_norm": 1.1754918916726766, |
|
"learning_rate": 6.530612244897958e-07, |
|
"loss": 0.5754, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.010030192927690498, |
|
"grad_norm": 1.272022559229399, |
|
"learning_rate": 6.666666666666666e-07, |
|
"loss": 0.5944, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.010234890742541324, |
|
"grad_norm": 1.13107848406085, |
|
"learning_rate": 6.802721088435374e-07, |
|
"loss": 0.5945, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01043958855739215, |
|
"grad_norm": 1.1273813534766033, |
|
"learning_rate": 6.938775510204081e-07, |
|
"loss": 0.5538, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.010644286372242977, |
|
"grad_norm": 1.1293664677810216, |
|
"learning_rate": 7.074829931972789e-07, |
|
"loss": 0.5854, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.010848984187093803, |
|
"grad_norm": 0.9728651370750258, |
|
"learning_rate": 7.210884353741496e-07, |
|
"loss": 0.5108, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.011053682001944629, |
|
"grad_norm": 1.0432420839745669, |
|
"learning_rate": 7.346938775510204e-07, |
|
"loss": 0.5346, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.011258379816795457, |
|
"grad_norm": 1.0023551080535893, |
|
"learning_rate": 7.482993197278912e-07, |
|
"loss": 0.5799, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.011463077631646282, |
|
"grad_norm": 0.9638908320867696, |
|
"learning_rate": 7.619047619047618e-07, |
|
"loss": 0.555, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.011667775446497108, |
|
"grad_norm": 1.0398584356633989, |
|
"learning_rate": 7.755102040816326e-07, |
|
"loss": 0.5147, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.011872473261347934, |
|
"grad_norm": 0.9629896909635629, |
|
"learning_rate": 7.891156462585033e-07, |
|
"loss": 0.5413, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.012077171076198762, |
|
"grad_norm": 0.9770292637339174, |
|
"learning_rate": 8.027210884353741e-07, |
|
"loss": 0.5205, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.012281868891049588, |
|
"grad_norm": 0.971945782703798, |
|
"learning_rate": 8.163265306122449e-07, |
|
"loss": 0.5422, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.012486566705900414, |
|
"grad_norm": 0.950398975311517, |
|
"learning_rate": 8.299319727891156e-07, |
|
"loss": 0.5071, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.012691264520751241, |
|
"grad_norm": 0.9049285150490526, |
|
"learning_rate": 8.435374149659864e-07, |
|
"loss": 0.4964, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.012895962335602067, |
|
"grad_norm": 0.8793095995125478, |
|
"learning_rate": 8.57142857142857e-07, |
|
"loss": 0.5331, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.013100660150452893, |
|
"grad_norm": 0.8515461613654705, |
|
"learning_rate": 8.707482993197278e-07, |
|
"loss": 0.5283, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.013305357965303721, |
|
"grad_norm": 0.867859420385022, |
|
"learning_rate": 8.843537414965985e-07, |
|
"loss": 0.5164, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.013510055780154547, |
|
"grad_norm": 0.8786706131313361, |
|
"learning_rate": 8.979591836734693e-07, |
|
"loss": 0.5645, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.013714753595005373, |
|
"grad_norm": 0.8579092596142676, |
|
"learning_rate": 9.115646258503401e-07, |
|
"loss": 0.5399, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.0139194514098562, |
|
"grad_norm": 0.8773908463960428, |
|
"learning_rate": 9.251700680272108e-07, |
|
"loss": 0.5229, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.014124149224707026, |
|
"grad_norm": 0.8528366708567172, |
|
"learning_rate": 9.387755102040816e-07, |
|
"loss": 0.5349, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.014328847039557852, |
|
"grad_norm": 0.9184139371914097, |
|
"learning_rate": 9.523809523809522e-07, |
|
"loss": 0.5331, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.01453354485440868, |
|
"grad_norm": 0.8507461371837629, |
|
"learning_rate": 9.65986394557823e-07, |
|
"loss": 0.4801, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.014738242669259506, |
|
"grad_norm": 0.8374936253263676, |
|
"learning_rate": 9.795918367346939e-07, |
|
"loss": 0.4931, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.014942940484110332, |
|
"grad_norm": 0.8174848059151317, |
|
"learning_rate": 9.931972789115645e-07, |
|
"loss": 0.5248, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.01514763829896116, |
|
"grad_norm": 0.8174077531772923, |
|
"learning_rate": 1.0068027210884354e-06, |
|
"loss": 0.5036, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.015352336113811985, |
|
"grad_norm": 0.7262562022534738, |
|
"learning_rate": 1.020408163265306e-06, |
|
"loss": 0.5232, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.015557033928662811, |
|
"grad_norm": 0.7855250505927771, |
|
"learning_rate": 1.034013605442177e-06, |
|
"loss": 0.5098, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.015761731743513637, |
|
"grad_norm": 0.8278680336215173, |
|
"learning_rate": 1.0476190476190476e-06, |
|
"loss": 0.4829, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.015966429558364463, |
|
"grad_norm": 0.797196328457245, |
|
"learning_rate": 1.0612244897959184e-06, |
|
"loss": 0.5037, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.016171127373215292, |
|
"grad_norm": 0.7507210642711485, |
|
"learning_rate": 1.074829931972789e-06, |
|
"loss": 0.4944, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.01637582518806612, |
|
"grad_norm": 0.826047544790976, |
|
"learning_rate": 1.08843537414966e-06, |
|
"loss": 0.5179, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.016580523002916944, |
|
"grad_norm": 0.7746315656318813, |
|
"learning_rate": 1.1020408163265304e-06, |
|
"loss": 0.5223, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.01678522081776777, |
|
"grad_norm": 0.778762710130468, |
|
"learning_rate": 1.1156462585034013e-06, |
|
"loss": 0.4845, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.016989918632618596, |
|
"grad_norm": 0.749908717861716, |
|
"learning_rate": 1.129251700680272e-06, |
|
"loss": 0.5175, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.017194616447469422, |
|
"grad_norm": 0.7582554704845739, |
|
"learning_rate": 1.1428571428571428e-06, |
|
"loss": 0.4978, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.01739931426232025, |
|
"grad_norm": 0.7595367961287336, |
|
"learning_rate": 1.1564625850340134e-06, |
|
"loss": 0.4966, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.017604012077171077, |
|
"grad_norm": 0.7488555001974914, |
|
"learning_rate": 1.1700680272108843e-06, |
|
"loss": 0.5025, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.017808709892021903, |
|
"grad_norm": 0.8307772703305798, |
|
"learning_rate": 1.183673469387755e-06, |
|
"loss": 0.5144, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.01801340770687273, |
|
"grad_norm": 0.7317615547098743, |
|
"learning_rate": 1.1972789115646258e-06, |
|
"loss": 0.4817, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.018218105521723555, |
|
"grad_norm": 0.8210594860542216, |
|
"learning_rate": 1.2108843537414965e-06, |
|
"loss": 0.5058, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.01842280333657438, |
|
"grad_norm": 0.7250535412206353, |
|
"learning_rate": 1.2244897959183673e-06, |
|
"loss": 0.4796, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.018627501151425207, |
|
"grad_norm": 0.7476633557284366, |
|
"learning_rate": 1.238095238095238e-06, |
|
"loss": 0.4732, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.018832198966276036, |
|
"grad_norm": 0.7245302420505394, |
|
"learning_rate": 1.2517006802721089e-06, |
|
"loss": 0.5085, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.019036896781126862, |
|
"grad_norm": 0.7287781044325405, |
|
"learning_rate": 1.2653061224489795e-06, |
|
"loss": 0.4837, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.019241594595977688, |
|
"grad_norm": 0.7461257075758424, |
|
"learning_rate": 1.2789115646258504e-06, |
|
"loss": 0.4847, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.019446292410828514, |
|
"grad_norm": 0.7500567577642135, |
|
"learning_rate": 1.292517006802721e-06, |
|
"loss": 0.5023, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.01965099022567934, |
|
"grad_norm": 0.7516926737451503, |
|
"learning_rate": 1.3061224489795917e-06, |
|
"loss": 0.4944, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.019855688040530166, |
|
"grad_norm": 0.8160475040600308, |
|
"learning_rate": 1.3197278911564623e-06, |
|
"loss": 0.4707, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.020060385855380995, |
|
"grad_norm": 0.7313987935291313, |
|
"learning_rate": 1.3333333333333332e-06, |
|
"loss": 0.4631, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.02026508367023182, |
|
"grad_norm": 0.7272827048713341, |
|
"learning_rate": 1.3469387755102039e-06, |
|
"loss": 0.4912, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.020469781485082647, |
|
"grad_norm": 0.7148392974765637, |
|
"learning_rate": 1.3605442176870747e-06, |
|
"loss": 0.4686, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.020674479299933473, |
|
"grad_norm": 0.8073254642999934, |
|
"learning_rate": 1.3741496598639456e-06, |
|
"loss": 0.4889, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.0208791771147843, |
|
"grad_norm": 0.7585784341693678, |
|
"learning_rate": 1.3877551020408162e-06, |
|
"loss": 0.4661, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.021083874929635125, |
|
"grad_norm": 0.750059071249337, |
|
"learning_rate": 1.4013605442176871e-06, |
|
"loss": 0.4856, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.021288572744485954, |
|
"grad_norm": 0.7391246566572075, |
|
"learning_rate": 1.4149659863945578e-06, |
|
"loss": 0.4835, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.02149327055933678, |
|
"grad_norm": 0.7961401475792825, |
|
"learning_rate": 1.4285714285714286e-06, |
|
"loss": 0.5078, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.021697968374187606, |
|
"grad_norm": 0.736443177126423, |
|
"learning_rate": 1.4421768707482993e-06, |
|
"loss": 0.4754, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.021902666189038432, |
|
"grad_norm": 0.7433000385873849, |
|
"learning_rate": 1.4557823129251701e-06, |
|
"loss": 0.5051, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.022107364003889258, |
|
"grad_norm": 0.742852894387874, |
|
"learning_rate": 1.4693877551020408e-06, |
|
"loss": 0.4815, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.022312061818740084, |
|
"grad_norm": 0.7321778169129644, |
|
"learning_rate": 1.4829931972789117e-06, |
|
"loss": 0.4883, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.022516759633590913, |
|
"grad_norm": 0.7374200652655346, |
|
"learning_rate": 1.4965986394557823e-06, |
|
"loss": 0.5235, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.02272145744844174, |
|
"grad_norm": 0.772981855244519, |
|
"learning_rate": 1.510204081632653e-06, |
|
"loss": 0.5008, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.022926155263292565, |
|
"grad_norm": 0.7342674795579016, |
|
"learning_rate": 1.5238095238095236e-06, |
|
"loss": 0.4671, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.02313085307814339, |
|
"grad_norm": 0.7795902315585469, |
|
"learning_rate": 1.5374149659863945e-06, |
|
"loss": 0.507, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.023335550892994217, |
|
"grad_norm": 0.7765099211131105, |
|
"learning_rate": 1.5510204081632651e-06, |
|
"loss": 0.5251, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.023540248707845043, |
|
"grad_norm": 0.7386929957340117, |
|
"learning_rate": 1.564625850340136e-06, |
|
"loss": 0.4578, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.02374494652269587, |
|
"grad_norm": 0.7248512159636582, |
|
"learning_rate": 1.5782312925170067e-06, |
|
"loss": 0.48, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.023949644337546698, |
|
"grad_norm": 0.7545806311647761, |
|
"learning_rate": 1.5918367346938775e-06, |
|
"loss": 0.4884, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.024154342152397524, |
|
"grad_norm": 0.7683287783699582, |
|
"learning_rate": 1.6054421768707482e-06, |
|
"loss": 0.4834, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.02435903996724835, |
|
"grad_norm": 0.775426549385026, |
|
"learning_rate": 1.619047619047619e-06, |
|
"loss": 0.4822, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.024563737782099176, |
|
"grad_norm": 0.7149469826873975, |
|
"learning_rate": 1.6326530612244897e-06, |
|
"loss": 0.4721, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.02476843559695, |
|
"grad_norm": 0.7985393152422335, |
|
"learning_rate": 1.6462585034013606e-06, |
|
"loss": 0.5166, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.024973133411800828, |
|
"grad_norm": 0.7885248764092557, |
|
"learning_rate": 1.6598639455782312e-06, |
|
"loss": 0.5091, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.025177831226651657, |
|
"grad_norm": 0.7531097232781883, |
|
"learning_rate": 1.673469387755102e-06, |
|
"loss": 0.5133, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.025382529041502483, |
|
"grad_norm": 0.7097213698861701, |
|
"learning_rate": 1.6870748299319727e-06, |
|
"loss": 0.5001, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.02558722685635331, |
|
"grad_norm": 0.6936318152279768, |
|
"learning_rate": 1.7006802721088434e-06, |
|
"loss": 0.4611, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.025791924671204135, |
|
"grad_norm": 0.7442480820206602, |
|
"learning_rate": 1.714285714285714e-06, |
|
"loss": 0.5107, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.02599662248605496, |
|
"grad_norm": 0.7310368101162509, |
|
"learning_rate": 1.727891156462585e-06, |
|
"loss": 0.4568, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.026201320300905787, |
|
"grad_norm": 0.7723563494615043, |
|
"learning_rate": 1.7414965986394556e-06, |
|
"loss": 0.4976, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.026406018115756616, |
|
"grad_norm": 0.7688284872373655, |
|
"learning_rate": 1.7551020408163264e-06, |
|
"loss": 0.4876, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.026610715930607442, |
|
"grad_norm": 0.7663908612309938, |
|
"learning_rate": 1.768707482993197e-06, |
|
"loss": 0.5089, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.026815413745458268, |
|
"grad_norm": 0.6966352320510637, |
|
"learning_rate": 1.782312925170068e-06, |
|
"loss": 0.4537, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.027020111560309094, |
|
"grad_norm": 0.6933747179682217, |
|
"learning_rate": 1.7959183673469386e-06, |
|
"loss": 0.4431, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.02722480937515992, |
|
"grad_norm": 0.7620187715357651, |
|
"learning_rate": 1.8095238095238095e-06, |
|
"loss": 0.5002, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.027429507190010746, |
|
"grad_norm": 0.694229773433825, |
|
"learning_rate": 1.8231292517006801e-06, |
|
"loss": 0.4602, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.02763420500486157, |
|
"grad_norm": 0.7152627743695282, |
|
"learning_rate": 1.836734693877551e-06, |
|
"loss": 0.46, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.0278389028197124, |
|
"grad_norm": 0.7175467946942147, |
|
"learning_rate": 1.8503401360544217e-06, |
|
"loss": 0.4687, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.028043600634563227, |
|
"grad_norm": 0.7852808070086453, |
|
"learning_rate": 1.8639455782312925e-06, |
|
"loss": 0.5045, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.028248298449414053, |
|
"grad_norm": 0.7212069697520485, |
|
"learning_rate": 1.8775510204081632e-06, |
|
"loss": 0.458, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.02845299626426488, |
|
"grad_norm": 0.6901869666091209, |
|
"learning_rate": 1.891156462585034e-06, |
|
"loss": 0.4873, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.028657694079115704, |
|
"grad_norm": 0.7038286804084832, |
|
"learning_rate": 1.9047619047619045e-06, |
|
"loss": 0.4575, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.02886239189396653, |
|
"grad_norm": 0.7736514303776025, |
|
"learning_rate": 1.918367346938775e-06, |
|
"loss": 0.4989, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.02906708970881736, |
|
"grad_norm": 0.7546506061753928, |
|
"learning_rate": 1.931972789115646e-06, |
|
"loss": 0.6157, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.029271787523668186, |
|
"grad_norm": 0.7502287441885653, |
|
"learning_rate": 1.945578231292517e-06, |
|
"loss": 0.4744, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.02947648533851901, |
|
"grad_norm": 0.7544747394474504, |
|
"learning_rate": 1.9591836734693877e-06, |
|
"loss": 0.4652, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.029681183153369838, |
|
"grad_norm": 0.7400203790224271, |
|
"learning_rate": 1.972789115646258e-06, |
|
"loss": 0.4749, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.029885880968220663, |
|
"grad_norm": 0.7756604473816919, |
|
"learning_rate": 1.986394557823129e-06, |
|
"loss": 0.4879, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.03009057878307149, |
|
"grad_norm": 0.7364241267157726, |
|
"learning_rate": 2e-06, |
|
"loss": 0.4641, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.03029527659792232, |
|
"grad_norm": 0.7509999341558731, |
|
"learning_rate": 1.9999997801737146e-06, |
|
"loss": 0.4716, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.030499974412773145, |
|
"grad_norm": 0.7817167258395246, |
|
"learning_rate": 1.9999991206949555e-06, |
|
"loss": 0.478, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.03070467222762397, |
|
"grad_norm": 0.6975681554994494, |
|
"learning_rate": 1.9999980215640124e-06, |
|
"loss": 0.4698, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.030909370042474796, |
|
"grad_norm": 0.6671508819481775, |
|
"learning_rate": 1.9999964827813685e-06, |
|
"loss": 0.4502, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.031114067857325622, |
|
"grad_norm": 0.7588040820967348, |
|
"learning_rate": 1.9999945043477006e-06, |
|
"loss": 0.4932, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.03131876567217645, |
|
"grad_norm": 0.7918237654214221, |
|
"learning_rate": 1.9999920862638785e-06, |
|
"loss": 0.4676, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.031523463487027274, |
|
"grad_norm": 0.7313481708497578, |
|
"learning_rate": 1.999989228530965e-06, |
|
"loss": 0.458, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.031728161301878104, |
|
"grad_norm": 0.7363675329922608, |
|
"learning_rate": 1.9999859311502164e-06, |
|
"loss": 0.4794, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.031932859116728926, |
|
"grad_norm": 0.7511639384926047, |
|
"learning_rate": 1.999982194123083e-06, |
|
"loss": 0.4811, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.032137556931579755, |
|
"grad_norm": 0.6977601768137399, |
|
"learning_rate": 1.9999780174512074e-06, |
|
"loss": 0.5046, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.032342254746430585, |
|
"grad_norm": 0.6871160044462953, |
|
"learning_rate": 1.999973401136426e-06, |
|
"loss": 0.4473, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.03254695256128141, |
|
"grad_norm": 0.735276225575691, |
|
"learning_rate": 1.999968345180768e-06, |
|
"loss": 0.4769, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.03275165037613224, |
|
"grad_norm": 0.6707772270842888, |
|
"learning_rate": 1.999962849586457e-06, |
|
"loss": 0.4395, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.03295634819098306, |
|
"grad_norm": 0.7714337406838349, |
|
"learning_rate": 1.9999569143559085e-06, |
|
"loss": 0.4658, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.03316104600583389, |
|
"grad_norm": 0.7896606525524605, |
|
"learning_rate": 1.999950539491732e-06, |
|
"loss": 0.4645, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.03336574382068471, |
|
"grad_norm": 0.7359544675011239, |
|
"learning_rate": 1.999943724996731e-06, |
|
"loss": 0.4671, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.03357044163553554, |
|
"grad_norm": 0.6942155729771998, |
|
"learning_rate": 1.9999364708739005e-06, |
|
"loss": 0.4567, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.03377513945038637, |
|
"grad_norm": 0.7156881239994389, |
|
"learning_rate": 1.9999287771264305e-06, |
|
"loss": 0.4871, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.03397983726523719, |
|
"grad_norm": 0.7027330885247588, |
|
"learning_rate": 1.999920643757703e-06, |
|
"loss": 0.4371, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.03418453508008802, |
|
"grad_norm": 0.7022489082551948, |
|
"learning_rate": 1.9999120707712943e-06, |
|
"loss": 0.45, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.034389232894938844, |
|
"grad_norm": 0.7600832151805308, |
|
"learning_rate": 1.9999030581709736e-06, |
|
"loss": 0.4812, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.03459393070978967, |
|
"grad_norm": 0.7757365362216246, |
|
"learning_rate": 1.9998936059607028e-06, |
|
"loss": 0.4951, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.0347986285246405, |
|
"grad_norm": 0.7180624522326351, |
|
"learning_rate": 1.9998837141446378e-06, |
|
"loss": 0.4733, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.035003326339491325, |
|
"grad_norm": 0.8012677566963108, |
|
"learning_rate": 1.9998733827271277e-06, |
|
"loss": 0.4854, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.035208024154342155, |
|
"grad_norm": 0.715314370525801, |
|
"learning_rate": 1.999862611712715e-06, |
|
"loss": 0.4777, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.03541272196919298, |
|
"grad_norm": 0.7191382757417352, |
|
"learning_rate": 1.9998514011061344e-06, |
|
"loss": 0.4637, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.035617419784043806, |
|
"grad_norm": 0.7116753984408628, |
|
"learning_rate": 1.9998397509123154e-06, |
|
"loss": 0.4536, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.03582211759889463, |
|
"grad_norm": 0.7598078550909712, |
|
"learning_rate": 1.9998276611363797e-06, |
|
"loss": 0.4908, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.03602681541374546, |
|
"grad_norm": 0.6563039003765047, |
|
"learning_rate": 1.999815131783643e-06, |
|
"loss": 0.449, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.03623151322859629, |
|
"grad_norm": 0.719561949641505, |
|
"learning_rate": 1.999802162859613e-06, |
|
"loss": 0.4741, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.03643621104344711, |
|
"grad_norm": 0.7109902438469043, |
|
"learning_rate": 1.999788754369993e-06, |
|
"loss": 0.4701, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.03664090885829794, |
|
"grad_norm": 0.7065392449298251, |
|
"learning_rate": 1.9997749063206762e-06, |
|
"loss": 0.4714, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.03684560667314876, |
|
"grad_norm": 0.7696360740535267, |
|
"learning_rate": 1.9997606187177524e-06, |
|
"loss": 0.4875, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.03705030448799959, |
|
"grad_norm": 0.7305783700088637, |
|
"learning_rate": 1.999745891567502e-06, |
|
"loss": 0.4606, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.037255002302850414, |
|
"grad_norm": 0.7270975727384246, |
|
"learning_rate": 1.9997307248764014e-06, |
|
"loss": 0.4198, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.03745970011770124, |
|
"grad_norm": 0.7569285986642791, |
|
"learning_rate": 1.9997151186511173e-06, |
|
"loss": 0.4354, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.03766439793255207, |
|
"grad_norm": 0.7219339206651326, |
|
"learning_rate": 1.9996990728985115e-06, |
|
"loss": 0.4378, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.037869095747402895, |
|
"grad_norm": 0.7690405011750759, |
|
"learning_rate": 1.9996825876256386e-06, |
|
"loss": 0.4791, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.038073793562253724, |
|
"grad_norm": 0.7552362068529521, |
|
"learning_rate": 1.9996656628397466e-06, |
|
"loss": 0.4672, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.03827849137710455, |
|
"grad_norm": 0.7341580278198813, |
|
"learning_rate": 1.999648298548276e-06, |
|
"loss": 0.4677, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.038483189191955376, |
|
"grad_norm": 0.7067121453226938, |
|
"learning_rate": 1.9996304947588612e-06, |
|
"loss": 0.4727, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.038687887006806206, |
|
"grad_norm": 0.7237165727925357, |
|
"learning_rate": 1.99961225147933e-06, |
|
"loss": 0.4446, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.03889258482165703, |
|
"grad_norm": 0.7069486805133093, |
|
"learning_rate": 1.999593568717703e-06, |
|
"loss": 0.4599, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.03909728263650786, |
|
"grad_norm": 0.897481774030034, |
|
"learning_rate": 1.9995744464821936e-06, |
|
"loss": 0.5129, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.03930198045135868, |
|
"grad_norm": 0.7488636278687589, |
|
"learning_rate": 1.9995548847812097e-06, |
|
"loss": 0.5184, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.03950667826620951, |
|
"grad_norm": 0.7194583225483666, |
|
"learning_rate": 1.9995348836233515e-06, |
|
"loss": 0.4915, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.03971137608106033, |
|
"grad_norm": 0.7384378767131218, |
|
"learning_rate": 1.999514443017412e-06, |
|
"loss": 0.4487, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.03991607389591116, |
|
"grad_norm": 0.7577332351147034, |
|
"learning_rate": 1.9994935629723784e-06, |
|
"loss": 0.4842, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.04012077171076199, |
|
"grad_norm": 0.7207002083905842, |
|
"learning_rate": 1.999472243497431e-06, |
|
"loss": 0.4698, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.04032546952561281, |
|
"grad_norm": 0.8010492120535461, |
|
"learning_rate": 1.9994504846019423e-06, |
|
"loss": 0.4561, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.04053016734046364, |
|
"grad_norm": 0.7453701461541147, |
|
"learning_rate": 1.9994282862954787e-06, |
|
"loss": 0.4806, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.040734865155314465, |
|
"grad_norm": 0.7255193966716207, |
|
"learning_rate": 1.9994056485878002e-06, |
|
"loss": 0.4511, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.040939562970165294, |
|
"grad_norm": 0.7957588909816856, |
|
"learning_rate": 1.9993825714888594e-06, |
|
"loss": 0.4775, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.04114426078501612, |
|
"grad_norm": 0.7304580504624026, |
|
"learning_rate": 1.999359055008802e-06, |
|
"loss": 0.4476, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.041348958599866946, |
|
"grad_norm": 0.8052138479705295, |
|
"learning_rate": 1.999335099157967e-06, |
|
"loss": 0.4621, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.041553656414717775, |
|
"grad_norm": 0.7344879094324241, |
|
"learning_rate": 1.999310703946887e-06, |
|
"loss": 0.448, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.0417583542295686, |
|
"grad_norm": 0.804269507197302, |
|
"learning_rate": 1.999285869386287e-06, |
|
"loss": 0.471, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.04196305204441943, |
|
"grad_norm": 0.7284627322104599, |
|
"learning_rate": 1.9992605954870867e-06, |
|
"loss": 0.4418, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.04216774985927025, |
|
"grad_norm": 0.7243013667651625, |
|
"learning_rate": 1.999234882260396e-06, |
|
"loss": 0.4669, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.04237244767412108, |
|
"grad_norm": 0.677583318692503, |
|
"learning_rate": 1.9992087297175213e-06, |
|
"loss": 0.4447, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.04257714548897191, |
|
"grad_norm": 0.7334595699121094, |
|
"learning_rate": 1.9991821378699598e-06, |
|
"loss": 0.4719, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.04278184330382273, |
|
"grad_norm": 0.7351912069847943, |
|
"learning_rate": 1.999155106729403e-06, |
|
"loss": 0.4758, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.04298654111867356, |
|
"grad_norm": 0.7262994043092325, |
|
"learning_rate": 1.9991276363077355e-06, |
|
"loss": 0.4636, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.04319123893352438, |
|
"grad_norm": 0.7170624975773432, |
|
"learning_rate": 1.999099726617034e-06, |
|
"loss": 0.4432, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.04339593674837521, |
|
"grad_norm": 0.7756861925710989, |
|
"learning_rate": 1.9990713776695697e-06, |
|
"loss": 0.4277, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.043600634563226034, |
|
"grad_norm": 0.7766290291464314, |
|
"learning_rate": 1.999042589477806e-06, |
|
"loss": 0.4521, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.043805332378076864, |
|
"grad_norm": 0.7675003445260637, |
|
"learning_rate": 1.9990133620543992e-06, |
|
"loss": 0.4728, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.04401003019292769, |
|
"grad_norm": 0.7502537566865808, |
|
"learning_rate": 1.9989836954122006e-06, |
|
"loss": 0.4919, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.044214728007778516, |
|
"grad_norm": 0.7256221777073304, |
|
"learning_rate": 1.998953589564252e-06, |
|
"loss": 0.4427, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.044419425822629345, |
|
"grad_norm": 0.7209802907423725, |
|
"learning_rate": 1.9989230445237905e-06, |
|
"loss": 0.4482, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.04462412363748017, |
|
"grad_norm": 0.761848969478383, |
|
"learning_rate": 1.9988920603042437e-06, |
|
"loss": 0.4623, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.044828821452331, |
|
"grad_norm": 0.7511377700619639, |
|
"learning_rate": 1.9988606369192357e-06, |
|
"loss": 0.4695, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.045033519267181826, |
|
"grad_norm": 0.684910692983434, |
|
"learning_rate": 1.998828774382581e-06, |
|
"loss": 0.4546, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.04523821708203265, |
|
"grad_norm": 0.8229255435418116, |
|
"learning_rate": 1.998796472708288e-06, |
|
"loss": 0.4736, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.04544291489688348, |
|
"grad_norm": 0.7208349457907924, |
|
"learning_rate": 1.998763731910558e-06, |
|
"loss": 0.4464, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.0456476127117343, |
|
"grad_norm": 0.8196660504458043, |
|
"learning_rate": 1.998730552003786e-06, |
|
"loss": 0.5129, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.04585231052658513, |
|
"grad_norm": 0.7410783352083353, |
|
"learning_rate": 1.99869693300256e-06, |
|
"loss": 0.4716, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.04605700834143595, |
|
"grad_norm": 0.6829979843011463, |
|
"learning_rate": 1.9986628749216598e-06, |
|
"loss": 0.4543, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.04626170615628678, |
|
"grad_norm": 0.6883777439066587, |
|
"learning_rate": 1.9986283777760598e-06, |
|
"loss": 0.4441, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.04646640397113761, |
|
"grad_norm": 0.7411624814334564, |
|
"learning_rate": 1.9985934415809266e-06, |
|
"loss": 0.4902, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.046671101785988434, |
|
"grad_norm": 0.7207442928316582, |
|
"learning_rate": 1.99855806635162e-06, |
|
"loss": 0.4502, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.04687579960083926, |
|
"grad_norm": 0.7535800049970842, |
|
"learning_rate": 1.9985222521036923e-06, |
|
"loss": 0.4588, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.047080497415690085, |
|
"grad_norm": 0.755131046087142, |
|
"learning_rate": 1.9984859988528896e-06, |
|
"loss": 0.478, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.047285195230540915, |
|
"grad_norm": 0.7363267944788255, |
|
"learning_rate": 1.9984493066151515e-06, |
|
"loss": 0.4722, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.04748989304539174, |
|
"grad_norm": 0.7175156333070826, |
|
"learning_rate": 1.9984121754066084e-06, |
|
"loss": 0.4284, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.04769459086024257, |
|
"grad_norm": 0.7021868758602576, |
|
"learning_rate": 1.9983746052435867e-06, |
|
"loss": 0.4549, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.047899288675093396, |
|
"grad_norm": 0.7661651319293146, |
|
"learning_rate": 1.998336596142603e-06, |
|
"loss": 0.4626, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.04810398648994422, |
|
"grad_norm": 0.7200117942966474, |
|
"learning_rate": 1.9982981481203685e-06, |
|
"loss": 0.4602, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.04830868430479505, |
|
"grad_norm": 0.7402636115111145, |
|
"learning_rate": 1.9982592611937875e-06, |
|
"loss": 0.462, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.04851338211964587, |
|
"grad_norm": 0.7289521015317652, |
|
"learning_rate": 1.998219935379956e-06, |
|
"loss": 0.4247, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.0487180799344967, |
|
"grad_norm": 0.8254564955090967, |
|
"learning_rate": 1.9981801706961637e-06, |
|
"loss": 0.445, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.04892277774934753, |
|
"grad_norm": 0.711987818470011, |
|
"learning_rate": 1.9981399671598938e-06, |
|
"loss": 0.4373, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.04912747556419835, |
|
"grad_norm": 0.7578069994316992, |
|
"learning_rate": 1.9980993247888215e-06, |
|
"loss": 0.4397, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.04933217337904918, |
|
"grad_norm": 0.7111156315088227, |
|
"learning_rate": 1.9980582436008155e-06, |
|
"loss": 0.444, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.0495368711939, |
|
"grad_norm": 0.6825451116028749, |
|
"learning_rate": 1.998016723613937e-06, |
|
"loss": 0.4216, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.04974156900875083, |
|
"grad_norm": 0.717357311135891, |
|
"learning_rate": 1.9979747648464406e-06, |
|
"loss": 0.4393, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.049946266823601655, |
|
"grad_norm": 0.7431878666336732, |
|
"learning_rate": 1.9979323673167735e-06, |
|
"loss": 0.4588, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.050150964638452485, |
|
"grad_norm": 0.7393037207206594, |
|
"learning_rate": 1.997889531043576e-06, |
|
"loss": 0.4547, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.050355662453303314, |
|
"grad_norm": 0.751624588819876, |
|
"learning_rate": 1.997846256045681e-06, |
|
"loss": 0.4423, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.050560360268154136, |
|
"grad_norm": 0.7291703435937729, |
|
"learning_rate": 1.9978025423421143e-06, |
|
"loss": 0.4577, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.050765058083004966, |
|
"grad_norm": 0.7730058435275767, |
|
"learning_rate": 1.9977583899520954e-06, |
|
"loss": 0.4725, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.05096975589785579, |
|
"grad_norm": 0.7388265631923211, |
|
"learning_rate": 1.9977137988950354e-06, |
|
"loss": 0.4915, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.05117445371270662, |
|
"grad_norm": 0.7253056655402751, |
|
"learning_rate": 1.9976687691905393e-06, |
|
"loss": 0.4463, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.05137915152755744, |
|
"grad_norm": 0.7279513578282504, |
|
"learning_rate": 1.997623300858404e-06, |
|
"loss": 0.4692, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.05158384934240827, |
|
"grad_norm": 0.7758619445867678, |
|
"learning_rate": 1.99757739391862e-06, |
|
"loss": 0.4359, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.0517885471572591, |
|
"grad_norm": 0.6934936677414176, |
|
"learning_rate": 1.9975310483913706e-06, |
|
"loss": 0.4342, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.05199324497210992, |
|
"grad_norm": 0.785492944076531, |
|
"learning_rate": 1.9974842642970316e-06, |
|
"loss": 0.4762, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.05219794278696075, |
|
"grad_norm": 0.693346587621168, |
|
"learning_rate": 1.9974370416561716e-06, |
|
"loss": 0.4077, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.05240264060181157, |
|
"grad_norm": 0.7036807587452536, |
|
"learning_rate": 1.9973893804895526e-06, |
|
"loss": 0.4559, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.0526073384166624, |
|
"grad_norm": 0.6487321250079171, |
|
"learning_rate": 1.997341280818128e-06, |
|
"loss": 0.4445, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.05281203623151323, |
|
"grad_norm": 0.7581793864091325, |
|
"learning_rate": 1.9972927426630464e-06, |
|
"loss": 0.4189, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.053016734046364054, |
|
"grad_norm": 0.7654003299344445, |
|
"learning_rate": 1.9972437660456465e-06, |
|
"loss": 0.4772, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.053221431861214884, |
|
"grad_norm": 0.7482689765950153, |
|
"learning_rate": 1.9971943509874614e-06, |
|
"loss": 0.4577, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.053426129676065706, |
|
"grad_norm": 0.7328225919609754, |
|
"learning_rate": 1.997144497510217e-06, |
|
"loss": 0.4301, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.053630827490916536, |
|
"grad_norm": 0.734577771662883, |
|
"learning_rate": 1.9970942056358307e-06, |
|
"loss": 0.4721, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.05383552530576736, |
|
"grad_norm": 0.7974471697046129, |
|
"learning_rate": 1.997043475386414e-06, |
|
"loss": 0.4759, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.05404022312061819, |
|
"grad_norm": 0.7204102805022299, |
|
"learning_rate": 1.99699230678427e-06, |
|
"loss": 0.4159, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.05424492093546902, |
|
"grad_norm": 0.8012057502786673, |
|
"learning_rate": 1.996940699851896e-06, |
|
"loss": 0.4784, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.05444961875031984, |
|
"grad_norm": 0.7456181199531785, |
|
"learning_rate": 1.9968886546119805e-06, |
|
"loss": 0.4716, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.05465431656517067, |
|
"grad_norm": 0.7582815001255205, |
|
"learning_rate": 1.996836171087405e-06, |
|
"loss": 0.4561, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.05485901438002149, |
|
"grad_norm": 0.7334168776176787, |
|
"learning_rate": 1.996783249301245e-06, |
|
"loss": 0.4344, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.05506371219487232, |
|
"grad_norm": 0.6872926048341936, |
|
"learning_rate": 1.9967298892767674e-06, |
|
"loss": 0.4293, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.05526841000972314, |
|
"grad_norm": 0.7532210670290133, |
|
"learning_rate": 1.9966760910374313e-06, |
|
"loss": 0.4644, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.05547310782457397, |
|
"grad_norm": 0.7306067456669837, |
|
"learning_rate": 1.99662185460689e-06, |
|
"loss": 0.4618, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.0556778056394248, |
|
"grad_norm": 0.7456313505165526, |
|
"learning_rate": 1.9965671800089887e-06, |
|
"loss": 0.4389, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.055882503454275624, |
|
"grad_norm": 0.6963357390921938, |
|
"learning_rate": 1.9965120672677646e-06, |
|
"loss": 0.4596, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.056087201269126453, |
|
"grad_norm": 0.6963762346496246, |
|
"learning_rate": 1.9964565164074488e-06, |
|
"loss": 0.4452, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.056291899083977276, |
|
"grad_norm": 0.7538437049515703, |
|
"learning_rate": 1.996400527452464e-06, |
|
"loss": 0.442, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.056496596898828105, |
|
"grad_norm": 0.7129437285411915, |
|
"learning_rate": 1.9963441004274265e-06, |
|
"loss": 0.4575, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.056701294713678935, |
|
"grad_norm": 0.7500374688918953, |
|
"learning_rate": 1.9962872353571436e-06, |
|
"loss": 0.4739, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.05690599252852976, |
|
"grad_norm": 0.7695931372733311, |
|
"learning_rate": 1.996229932266617e-06, |
|
"loss": 0.4523, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.05711069034338059, |
|
"grad_norm": 0.8047384042314083, |
|
"learning_rate": 1.99617219118104e-06, |
|
"loss": 0.4541, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.05731538815823141, |
|
"grad_norm": 0.7337412326319969, |
|
"learning_rate": 1.9961140121257978e-06, |
|
"loss": 0.5429, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.05752008597308224, |
|
"grad_norm": 0.7647644479794993, |
|
"learning_rate": 1.99605539512647e-06, |
|
"loss": 0.4479, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.05772478378793306, |
|
"grad_norm": 0.7676226046817726, |
|
"learning_rate": 1.9959963402088274e-06, |
|
"loss": 0.4641, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.05792948160278389, |
|
"grad_norm": 0.7235143819897621, |
|
"learning_rate": 1.9959368473988333e-06, |
|
"loss": 0.4545, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.05813417941763472, |
|
"grad_norm": 0.7247778707795571, |
|
"learning_rate": 1.9958769167226444e-06, |
|
"loss": 0.4527, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.05833887723248554, |
|
"grad_norm": 0.7539665158584474, |
|
"learning_rate": 1.995816548206609e-06, |
|
"loss": 0.4458, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.05854357504733637, |
|
"grad_norm": 0.7532087725317325, |
|
"learning_rate": 1.995755741877269e-06, |
|
"loss": 0.4461, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.058748272862187194, |
|
"grad_norm": 0.7797916526020292, |
|
"learning_rate": 1.995694497761357e-06, |
|
"loss": 0.4665, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.05895297067703802, |
|
"grad_norm": 0.7517569918058703, |
|
"learning_rate": 1.9956328158857992e-06, |
|
"loss": 0.4728, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.059157668491888846, |
|
"grad_norm": 0.7570745421432602, |
|
"learning_rate": 1.995570696277715e-06, |
|
"loss": 0.4563, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.059362366306739675, |
|
"grad_norm": 0.740586734451177, |
|
"learning_rate": 1.995508138964415e-06, |
|
"loss": 0.4709, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.059567064121590504, |
|
"grad_norm": 0.7681915361666245, |
|
"learning_rate": 1.995445143973403e-06, |
|
"loss": 0.4503, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.05977176193644133, |
|
"grad_norm": 0.7455448750612553, |
|
"learning_rate": 1.9953817113323743e-06, |
|
"loss": 0.4529, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.059976459751292156, |
|
"grad_norm": 0.7745938996976398, |
|
"learning_rate": 1.9953178410692174e-06, |
|
"loss": 0.4256, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.06018115756614298, |
|
"grad_norm": 0.7431821602019313, |
|
"learning_rate": 1.9952535332120137e-06, |
|
"loss": 0.4453, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.06038585538099381, |
|
"grad_norm": 0.6903595364669262, |
|
"learning_rate": 1.9951887877890354e-06, |
|
"loss": 0.4339, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.06059055319584464, |
|
"grad_norm": 0.773597292773469, |
|
"learning_rate": 1.9951236048287483e-06, |
|
"loss": 0.4817, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.06079525101069546, |
|
"grad_norm": 0.7263693618591641, |
|
"learning_rate": 1.9950579843598105e-06, |
|
"loss": 0.4572, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.06099994882554629, |
|
"grad_norm": 0.7297961413486055, |
|
"learning_rate": 1.994991926411072e-06, |
|
"loss": 0.4359, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.06120464664039711, |
|
"grad_norm": 0.7855052727291876, |
|
"learning_rate": 1.9949254310115753e-06, |
|
"loss": 0.4568, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.06140934445524794, |
|
"grad_norm": 0.7300504368627593, |
|
"learning_rate": 1.994858498190556e-06, |
|
"loss": 0.4501, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.061614042270098764, |
|
"grad_norm": 0.7096161782700376, |
|
"learning_rate": 1.99479112797744e-06, |
|
"loss": 0.4663, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.06181874008494959, |
|
"grad_norm": 0.7018494412530502, |
|
"learning_rate": 1.9947233204018477e-06, |
|
"loss": 0.4401, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.06202343789980042, |
|
"grad_norm": 0.7907086687325865, |
|
"learning_rate": 1.9946550754935906e-06, |
|
"loss": 0.4394, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.062228135714651245, |
|
"grad_norm": 0.7637305795975494, |
|
"learning_rate": 1.9945863932826727e-06, |
|
"loss": 0.4368, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.062432833529502074, |
|
"grad_norm": 0.7745057196668337, |
|
"learning_rate": 1.9945172737992904e-06, |
|
"loss": 0.4926, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.0626375313443529, |
|
"grad_norm": 0.7367608831304401, |
|
"learning_rate": 1.994447717073832e-06, |
|
"loss": 0.4688, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.06284222915920372, |
|
"grad_norm": 0.7695856962055981, |
|
"learning_rate": 1.9943777231368794e-06, |
|
"loss": 0.4484, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.06304692697405455, |
|
"grad_norm": 0.7187776209986876, |
|
"learning_rate": 1.994307292019204e-06, |
|
"loss": 0.4444, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.06325162478890538, |
|
"grad_norm": 0.7347949665208309, |
|
"learning_rate": 1.994236423751772e-06, |
|
"loss": 0.4485, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.06345632260375621, |
|
"grad_norm": 0.7105615411584904, |
|
"learning_rate": 1.99416511836574e-06, |
|
"loss": 0.4115, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.06366102041860704, |
|
"grad_norm": 0.7092446318381356, |
|
"learning_rate": 1.994093375892459e-06, |
|
"loss": 0.4398, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.06386571823345785, |
|
"grad_norm": 0.770807237418002, |
|
"learning_rate": 1.9940211963634696e-06, |
|
"loss": 0.4413, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.06407041604830868, |
|
"grad_norm": 0.7396584102001305, |
|
"learning_rate": 1.9939485798105057e-06, |
|
"loss": 0.4851, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.06427511386315951, |
|
"grad_norm": 0.7444518429400907, |
|
"learning_rate": 1.9938755262654945e-06, |
|
"loss": 0.4337, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.06447981167801034, |
|
"grad_norm": 0.722720227886292, |
|
"learning_rate": 1.9938020357605527e-06, |
|
"loss": 0.4965, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.06468450949286117, |
|
"grad_norm": 0.7666649202795676, |
|
"learning_rate": 1.993728108327992e-06, |
|
"loss": 0.4532, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.06488920730771199, |
|
"grad_norm": 0.7549308832804502, |
|
"learning_rate": 1.9936537440003134e-06, |
|
"loss": 0.4582, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.06509390512256281, |
|
"grad_norm": 0.6872546619483418, |
|
"learning_rate": 1.993578942810212e-06, |
|
"loss": 0.4695, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.06529860293741364, |
|
"grad_norm": 0.7550879312423509, |
|
"learning_rate": 1.9935037047905748e-06, |
|
"loss": 0.4833, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.06550330075226447, |
|
"grad_norm": 0.7064631296777315, |
|
"learning_rate": 1.99342802997448e-06, |
|
"loss": 0.4724, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.0657079985671153, |
|
"grad_norm": 0.7041398494235039, |
|
"learning_rate": 1.9933519183951977e-06, |
|
"loss": 0.4441, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.06591269638196612, |
|
"grad_norm": 0.7624056534749137, |
|
"learning_rate": 1.9932753700861914e-06, |
|
"loss": 0.465, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.06611739419681695, |
|
"grad_norm": 0.7553645719854849, |
|
"learning_rate": 1.9931983850811155e-06, |
|
"loss": 0.4241, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.06632209201166778, |
|
"grad_norm": 0.7617302308064162, |
|
"learning_rate": 1.9931209634138158e-06, |
|
"loss": 0.4475, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.0665267898265186, |
|
"grad_norm": 0.7484410349937703, |
|
"learning_rate": 1.9930431051183324e-06, |
|
"loss": 0.432, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.06673148764136942, |
|
"grad_norm": 0.7283423040518047, |
|
"learning_rate": 1.9929648102288953e-06, |
|
"loss": 0.4388, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.06693618545622025, |
|
"grad_norm": 0.7852619772614368, |
|
"learning_rate": 1.9928860787799265e-06, |
|
"loss": 0.468, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.06714088327107108, |
|
"grad_norm": 0.7769054079891817, |
|
"learning_rate": 1.992806910806041e-06, |
|
"loss": 0.4579, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.06734558108592191, |
|
"grad_norm": 0.7619008198687942, |
|
"learning_rate": 1.992727306342045e-06, |
|
"loss": 0.4789, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.06755027890077274, |
|
"grad_norm": 0.7607061858221339, |
|
"learning_rate": 1.9926472654229376e-06, |
|
"loss": 0.4241, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.06775497671562355, |
|
"grad_norm": 0.7236097758888326, |
|
"learning_rate": 1.992566788083908e-06, |
|
"loss": 0.4181, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.06795967453047438, |
|
"grad_norm": 0.8114241422692142, |
|
"learning_rate": 1.992485874360338e-06, |
|
"loss": 0.4383, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.06816437234532521, |
|
"grad_norm": 0.7429107582072085, |
|
"learning_rate": 1.992404524287803e-06, |
|
"loss": 0.4418, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.06836907016017604, |
|
"grad_norm": 0.7664592844718724, |
|
"learning_rate": 1.9923227379020674e-06, |
|
"loss": 0.4424, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.06857376797502687, |
|
"grad_norm": 0.7525421136101852, |
|
"learning_rate": 1.9922405152390893e-06, |
|
"loss": 0.4601, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.06877846578987769, |
|
"grad_norm": 0.7924977816871237, |
|
"learning_rate": 1.9921578563350182e-06, |
|
"loss": 0.4481, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.06898316360472852, |
|
"grad_norm": 0.7205455607720451, |
|
"learning_rate": 1.9920747612261953e-06, |
|
"loss": 0.472, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.06918786141957935, |
|
"grad_norm": 0.8099842191878124, |
|
"learning_rate": 1.9919912299491534e-06, |
|
"loss": 0.4522, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.06939255923443018, |
|
"grad_norm": 0.7601204829884295, |
|
"learning_rate": 1.991907262540617e-06, |
|
"loss": 0.4361, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.069597257049281, |
|
"grad_norm": 0.7498165219729875, |
|
"learning_rate": 1.9918228590375034e-06, |
|
"loss": 0.4363, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.06980195486413182, |
|
"grad_norm": 0.7130352034738562, |
|
"learning_rate": 1.9917380194769197e-06, |
|
"loss": 0.4355, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.07000665267898265, |
|
"grad_norm": 0.7653650161746682, |
|
"learning_rate": 1.991652743896167e-06, |
|
"loss": 0.4062, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.07021135049383348, |
|
"grad_norm": 0.8196206368298663, |
|
"learning_rate": 1.991567032332736e-06, |
|
"loss": 0.4614, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.07041604830868431, |
|
"grad_norm": 0.7812062890731587, |
|
"learning_rate": 1.991480884824311e-06, |
|
"loss": 0.4975, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.07062074612353512, |
|
"grad_norm": 0.7206276352235353, |
|
"learning_rate": 1.9913943014087655e-06, |
|
"loss": 0.4157, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.07082544393838595, |
|
"grad_norm": 0.78698897087266, |
|
"learning_rate": 1.9913072821241672e-06, |
|
"loss": 0.4581, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.07103014175323678, |
|
"grad_norm": 0.7764210083677198, |
|
"learning_rate": 1.991219827008775e-06, |
|
"loss": 0.4408, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.07123483956808761, |
|
"grad_norm": 0.7595433195005857, |
|
"learning_rate": 1.9911319361010367e-06, |
|
"loss": 0.4492, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.07143953738293844, |
|
"grad_norm": 0.722334537399672, |
|
"learning_rate": 1.991043609439596e-06, |
|
"loss": 0.4098, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.07164423519778926, |
|
"grad_norm": 0.7424362807959037, |
|
"learning_rate": 1.9909548470632842e-06, |
|
"loss": 0.4169, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.07184893301264009, |
|
"grad_norm": 0.7665542829825039, |
|
"learning_rate": 1.9908656490111267e-06, |
|
"loss": 0.4635, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.07205363082749092, |
|
"grad_norm": 0.7670493050777013, |
|
"learning_rate": 1.9907760153223396e-06, |
|
"loss": 0.4245, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.07225832864234175, |
|
"grad_norm": 0.7082306081348982, |
|
"learning_rate": 1.9906859460363304e-06, |
|
"loss": 0.3885, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.07246302645719258, |
|
"grad_norm": 0.7179046673862461, |
|
"learning_rate": 1.990595441192699e-06, |
|
"loss": 0.4047, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.07266772427204339, |
|
"grad_norm": 0.7785673633417279, |
|
"learning_rate": 1.990504500831235e-06, |
|
"loss": 0.484, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.07287242208689422, |
|
"grad_norm": 0.7657292926077239, |
|
"learning_rate": 1.9904131249919215e-06, |
|
"loss": 0.4444, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.07307711990174505, |
|
"grad_norm": 0.7569753573050043, |
|
"learning_rate": 1.9903213137149313e-06, |
|
"loss": 0.4701, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.07328181771659588, |
|
"grad_norm": 0.8072244283837906, |
|
"learning_rate": 1.99022906704063e-06, |
|
"loss": 0.4409, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.07348651553144671, |
|
"grad_norm": 0.7629374834771911, |
|
"learning_rate": 1.990136385009574e-06, |
|
"loss": 0.4927, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.07369121334629752, |
|
"grad_norm": 0.7385249798311992, |
|
"learning_rate": 1.990043267662511e-06, |
|
"loss": 0.4338, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.07389591116114835, |
|
"grad_norm": 0.775227359922002, |
|
"learning_rate": 1.989949715040381e-06, |
|
"loss": 0.4789, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.07410060897599918, |
|
"grad_norm": 0.7517526442766227, |
|
"learning_rate": 1.9898557271843133e-06, |
|
"loss": 0.4504, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.07430530679085001, |
|
"grad_norm": 0.7375927174293256, |
|
"learning_rate": 1.9897613041356314e-06, |
|
"loss": 0.4108, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.07451000460570083, |
|
"grad_norm": 0.72174871971371, |
|
"learning_rate": 1.9896664459358472e-06, |
|
"loss": 0.4475, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.07471470242055166, |
|
"grad_norm": 0.7329776456768429, |
|
"learning_rate": 1.9895711526266667e-06, |
|
"loss": 0.424, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.07491940023540249, |
|
"grad_norm": 0.7185696661034995, |
|
"learning_rate": 1.9894754242499852e-06, |
|
"loss": 0.4543, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.07512409805025332, |
|
"grad_norm": 0.7049184644044292, |
|
"learning_rate": 1.98937926084789e-06, |
|
"loss": 0.42, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.07532879586510415, |
|
"grad_norm": 0.7293724081932391, |
|
"learning_rate": 1.989282662462659e-06, |
|
"loss": 0.43, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.07553349367995496, |
|
"grad_norm": 0.7059632003848231, |
|
"learning_rate": 1.9891856291367626e-06, |
|
"loss": 0.4275, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.07573819149480579, |
|
"grad_norm": 0.7683789138077434, |
|
"learning_rate": 1.9890881609128618e-06, |
|
"loss": 0.4408, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.07594288930965662, |
|
"grad_norm": 0.7872951646579199, |
|
"learning_rate": 1.9889902578338087e-06, |
|
"loss": 0.4292, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.07614758712450745, |
|
"grad_norm": 0.7924075912782186, |
|
"learning_rate": 1.988891919942646e-06, |
|
"loss": 0.4408, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.07635228493935828, |
|
"grad_norm": 0.7607074410846072, |
|
"learning_rate": 1.9887931472826093e-06, |
|
"loss": 0.4485, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.0765569827542091, |
|
"grad_norm": 0.7592064137840062, |
|
"learning_rate": 1.9886939398971238e-06, |
|
"loss": 0.4404, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.07676168056905992, |
|
"grad_norm": 0.8183614522835663, |
|
"learning_rate": 1.9885942978298054e-06, |
|
"loss": 0.4677, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.07696637838391075, |
|
"grad_norm": 0.7400985748589023, |
|
"learning_rate": 1.9884942211244637e-06, |
|
"loss": 0.4867, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.07717107619876158, |
|
"grad_norm": 0.7657568826393069, |
|
"learning_rate": 1.988393709825096e-06, |
|
"loss": 0.4592, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.07737577401361241, |
|
"grad_norm": 0.7242932934680155, |
|
"learning_rate": 1.988292763975893e-06, |
|
"loss": 0.4037, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.07758047182846323, |
|
"grad_norm": 0.695616397865644, |
|
"learning_rate": 1.9881913836212365e-06, |
|
"loss": 0.4534, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.07778516964331406, |
|
"grad_norm": 0.7286084983103666, |
|
"learning_rate": 1.9880895688056977e-06, |
|
"loss": 0.4267, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.07798986745816489, |
|
"grad_norm": 0.7434704183939207, |
|
"learning_rate": 1.98798731957404e-06, |
|
"loss": 0.4308, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.07819456527301571, |
|
"grad_norm": 0.7803766550265465, |
|
"learning_rate": 1.9878846359712176e-06, |
|
"loss": 0.4455, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.07839926308786653, |
|
"grad_norm": 0.777748381274515, |
|
"learning_rate": 1.9877815180423757e-06, |
|
"loss": 0.4639, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.07860396090271736, |
|
"grad_norm": 0.724760659067026, |
|
"learning_rate": 1.9876779658328503e-06, |
|
"loss": 0.4666, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.07880865871756819, |
|
"grad_norm": 0.6620651646162441, |
|
"learning_rate": 1.9875739793881685e-06, |
|
"loss": 0.4195, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.07901335653241902, |
|
"grad_norm": 0.7524415209772658, |
|
"learning_rate": 1.9874695587540477e-06, |
|
"loss": 0.4431, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.07921805434726985, |
|
"grad_norm": 0.6797630439484897, |
|
"learning_rate": 1.9873647039763975e-06, |
|
"loss": 0.4453, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.07942275216212066, |
|
"grad_norm": 0.7685812672797145, |
|
"learning_rate": 1.987259415101317e-06, |
|
"loss": 0.4623, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.07962744997697149, |
|
"grad_norm": 0.7581895406953137, |
|
"learning_rate": 1.9871536921750965e-06, |
|
"loss": 0.4423, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.07983214779182232, |
|
"grad_norm": 0.7100776210964536, |
|
"learning_rate": 1.987047535244218e-06, |
|
"loss": 0.4171, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.08003684560667315, |
|
"grad_norm": 0.7299757652818287, |
|
"learning_rate": 1.9869409443553535e-06, |
|
"loss": 0.454, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.08024154342152398, |
|
"grad_norm": 0.7439290629641574, |
|
"learning_rate": 1.9868339195553657e-06, |
|
"loss": 0.4426, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.0804462412363748, |
|
"grad_norm": 0.75716235792549, |
|
"learning_rate": 1.9867264608913084e-06, |
|
"loss": 0.4479, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.08065093905122563, |
|
"grad_norm": 0.6968938261452492, |
|
"learning_rate": 1.9866185684104266e-06, |
|
"loss": 0.4335, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.08085563686607646, |
|
"grad_norm": 0.7170546940871543, |
|
"learning_rate": 1.9865102421601545e-06, |
|
"loss": 0.4286, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.08106033468092728, |
|
"grad_norm": 0.7478223773320385, |
|
"learning_rate": 1.986401482188119e-06, |
|
"loss": 0.4202, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.08126503249577811, |
|
"grad_norm": 0.7076328746009948, |
|
"learning_rate": 1.986292288542136e-06, |
|
"loss": 0.4558, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.08146973031062893, |
|
"grad_norm": 0.7591423650217918, |
|
"learning_rate": 1.986182661270213e-06, |
|
"loss": 0.4674, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.08167442812547976, |
|
"grad_norm": 0.7416281412961139, |
|
"learning_rate": 1.9860726004205485e-06, |
|
"loss": 0.4397, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.08187912594033059, |
|
"grad_norm": 0.7416913042041706, |
|
"learning_rate": 1.98596210604153e-06, |
|
"loss": 0.4405, |
|
"step": 400 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 4885, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 200, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 53820235284480.0, |
|
"train_batch_size": 4, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|