{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 4145, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.030156815440289506, "grad_norm": Infinity, "learning_rate": 4.2e-05, "loss": 4.6298, "step": 25 }, { "epoch": 0.06031363088057901, "grad_norm": 40.83469772338867, "learning_rate": 9.200000000000001e-05, "loss": 0.6127, "step": 50 }, { "epoch": 0.09047044632086852, "grad_norm": 79.89573669433594, "learning_rate": 9.948717948717949e-05, "loss": 0.7524, "step": 75 }, { "epoch": 0.12062726176115803, "grad_norm": 446.1002502441406, "learning_rate": 9.89010989010989e-05, "loss": 1.0178, "step": 100 }, { "epoch": 0.15078407720144751, "grad_norm": 128.81219482421875, "learning_rate": 9.829059829059829e-05, "loss": 1.834, "step": 125 }, { "epoch": 0.18094089264173704, "grad_norm": 37.102203369140625, "learning_rate": 9.768009768009768e-05, "loss": 1.7381, "step": 150 }, { "epoch": 0.21109770808202655, "grad_norm": 31.156801223754883, "learning_rate": 9.706959706959707e-05, "loss": 1.3946, "step": 175 }, { "epoch": 0.24125452352231605, "grad_norm": 37.00471115112305, "learning_rate": 9.645909645909647e-05, "loss": 1.2038, "step": 200 }, { "epoch": 0.27141133896260555, "grad_norm": 26.962158203125, "learning_rate": 9.584859584859585e-05, "loss": 1.1359, "step": 225 }, { "epoch": 0.30156815440289503, "grad_norm": 32.251277923583984, "learning_rate": 9.523809523809524e-05, "loss": 1.112, "step": 250 }, { "epoch": 0.33172496984318456, "grad_norm": 26.48785972595215, "learning_rate": 9.462759462759463e-05, "loss": 1.071, "step": 275 }, { "epoch": 0.3618817852834741, "grad_norm": 18.70881462097168, "learning_rate": 9.401709401709401e-05, "loss": 1.0259, "step": 300 }, { "epoch": 0.39203860072376356, "grad_norm": 21.082536697387695, "learning_rate": 9.340659340659341e-05, "loss": 1.0232, "step": 325 }, { "epoch": 0.4221954161640531, "grad_norm": 18.735260009765625, "learning_rate": 9.27960927960928e-05, "loss": 1.0632, "step": 350 }, { "epoch": 0.45235223160434257, "grad_norm": 22.869077682495117, "learning_rate": 9.218559218559219e-05, "loss": 1.0362, "step": 375 }, { "epoch": 0.4825090470446321, "grad_norm": 19.3148250579834, "learning_rate": 9.157509157509158e-05, "loss": 0.9837, "step": 400 }, { "epoch": 0.5126658624849216, "grad_norm": 15.154926300048828, "learning_rate": 9.096459096459096e-05, "loss": 0.957, "step": 425 }, { "epoch": 0.5428226779252111, "grad_norm": 15.833111763000488, "learning_rate": 9.035409035409036e-05, "loss": 0.9336, "step": 450 }, { "epoch": 0.5729794933655006, "grad_norm": 21.135068893432617, "learning_rate": 8.974358974358975e-05, "loss": 0.9431, "step": 475 }, { "epoch": 0.6031363088057901, "grad_norm": 17.726829528808594, "learning_rate": 8.913308913308915e-05, "loss": 0.9775, "step": 500 }, { "epoch": 0.6332931242460796, "grad_norm": 25.220857620239258, "learning_rate": 8.852258852258852e-05, "loss": 1.0455, "step": 525 }, { "epoch": 0.6634499396863691, "grad_norm": 22.668001174926758, "learning_rate": 8.791208791208791e-05, "loss": 0.9617, "step": 550 }, { "epoch": 0.6936067551266586, "grad_norm": 19.432435989379883, "learning_rate": 8.730158730158731e-05, "loss": 0.9314, "step": 575 }, { "epoch": 0.7237635705669482, "grad_norm": 18.803312301635742, "learning_rate": 8.66910866910867e-05, "loss": 0.9559, "step": 600 }, { "epoch": 0.7539203860072377, "grad_norm": 17.0799560546875, "learning_rate": 8.608058608058608e-05, "loss": 0.9768, "step": 625 }, { "epoch": 0.7840772014475271, "grad_norm": 18.541135787963867, "learning_rate": 8.547008547008547e-05, "loss": 0.9384, "step": 650 }, { "epoch": 0.8142340168878166, "grad_norm": 76.61760711669922, "learning_rate": 8.485958485958486e-05, "loss": 0.9088, "step": 675 }, { "epoch": 0.8443908323281062, "grad_norm": 15.41170883178711, "learning_rate": 8.424908424908426e-05, "loss": 0.8698, "step": 700 }, { "epoch": 0.8745476477683957, "grad_norm": 19.12152099609375, "learning_rate": 8.363858363858364e-05, "loss": 0.932, "step": 725 }, { "epoch": 0.9047044632086851, "grad_norm": 13.084074974060059, "learning_rate": 8.302808302808303e-05, "loss": 0.8682, "step": 750 }, { "epoch": 0.9348612786489746, "grad_norm": 17.12626838684082, "learning_rate": 8.241758241758242e-05, "loss": 0.879, "step": 775 }, { "epoch": 0.9650180940892642, "grad_norm": 27.889331817626953, "learning_rate": 8.18070818070818e-05, "loss": 0.8815, "step": 800 }, { "epoch": 0.9951749095295537, "grad_norm": 15.480960845947266, "learning_rate": 8.11965811965812e-05, "loss": 0.8855, "step": 825 }, { "epoch": 1.0253317249698433, "grad_norm": 17.717817306518555, "learning_rate": 8.058608058608059e-05, "loss": 0.8537, "step": 850 }, { "epoch": 1.0554885404101326, "grad_norm": 18.344432830810547, "learning_rate": 7.997557997557998e-05, "loss": 0.9154, "step": 875 }, { "epoch": 1.0856453558504222, "grad_norm": 17.965234756469727, "learning_rate": 7.936507936507937e-05, "loss": 0.91, "step": 900 }, { "epoch": 1.1158021712907118, "grad_norm": 22.685340881347656, "learning_rate": 7.875457875457875e-05, "loss": 0.8627, "step": 925 }, { "epoch": 1.1459589867310012, "grad_norm": 36.94546127319336, "learning_rate": 7.814407814407815e-05, "loss": 0.84, "step": 950 }, { "epoch": 1.1761158021712907, "grad_norm": 98.67118835449219, "learning_rate": 7.753357753357754e-05, "loss": 0.8758, "step": 975 }, { "epoch": 1.2062726176115803, "grad_norm": 22.5080509185791, "learning_rate": 7.694749694749695e-05, "loss": 0.9677, "step": 1000 }, { "epoch": 1.2364294330518697, "grad_norm": 20.48828125, "learning_rate": 7.633699633699634e-05, "loss": 1.2519, "step": 1025 }, { "epoch": 1.2665862484921593, "grad_norm": 16.624662399291992, "learning_rate": 7.572649572649573e-05, "loss": 0.8674, "step": 1050 }, { "epoch": 1.2967430639324489, "grad_norm": 17.774490356445312, "learning_rate": 7.511599511599511e-05, "loss": 0.8505, "step": 1075 }, { "epoch": 1.3268998793727382, "grad_norm": 12.95785903930664, "learning_rate": 7.450549450549451e-05, "loss": 0.8215, "step": 1100 }, { "epoch": 1.3570566948130278, "grad_norm": 14.339399337768555, "learning_rate": 7.38949938949939e-05, "loss": 0.8527, "step": 1125 }, { "epoch": 1.3872135102533172, "grad_norm": 13.43790340423584, "learning_rate": 7.328449328449329e-05, "loss": 0.7856, "step": 1150 }, { "epoch": 1.4173703256936068, "grad_norm": 15.764432907104492, "learning_rate": 7.267399267399268e-05, "loss": 0.8953, "step": 1175 }, { "epoch": 1.4475271411338961, "grad_norm": 17.939727783203125, "learning_rate": 7.206349206349206e-05, "loss": 0.7962, "step": 1200 }, { "epoch": 1.4776839565741857, "grad_norm": 22.913164138793945, "learning_rate": 7.145299145299146e-05, "loss": 0.8549, "step": 1225 }, { "epoch": 1.5078407720144753, "grad_norm": 36.68199157714844, "learning_rate": 7.084249084249085e-05, "loss": 0.7912, "step": 1250 }, { "epoch": 1.5379975874547647, "grad_norm": 14.023738861083984, "learning_rate": 7.025641025641025e-05, "loss": 0.9224, "step": 1275 }, { "epoch": 1.5681544028950543, "grad_norm": 14.844511985778809, "learning_rate": 6.964590964590965e-05, "loss": 0.8707, "step": 1300 }, { "epoch": 1.5983112183353438, "grad_norm": 67.9472427368164, "learning_rate": 6.903540903540904e-05, "loss": 0.8692, "step": 1325 }, { "epoch": 1.6284680337756332, "grad_norm": 13.173006057739258, "learning_rate": 6.842490842490842e-05, "loss": 0.8179, "step": 1350 }, { "epoch": 1.6586248492159228, "grad_norm": 18.236101150512695, "learning_rate": 6.781440781440782e-05, "loss": 0.8281, "step": 1375 }, { "epoch": 1.6887816646562124, "grad_norm": 20.294797897338867, "learning_rate": 6.720390720390721e-05, "loss": 0.8273, "step": 1400 }, { "epoch": 1.7189384800965017, "grad_norm": 18.06429672241211, "learning_rate": 6.65934065934066e-05, "loss": 0.7984, "step": 1425 }, { "epoch": 1.7490952955367913, "grad_norm": 159.88597106933594, "learning_rate": 6.598290598290599e-05, "loss": 0.8045, "step": 1450 }, { "epoch": 1.779252110977081, "grad_norm": 32.71599197387695, "learning_rate": 6.537240537240537e-05, "loss": 0.8615, "step": 1475 }, { "epoch": 1.8094089264173703, "grad_norm": 20.00482177734375, "learning_rate": 6.476190476190477e-05, "loss": 0.8346, "step": 1500 }, { "epoch": 1.8395657418576599, "grad_norm": 29.308307647705078, "learning_rate": 6.415140415140416e-05, "loss": 0.7786, "step": 1525 }, { "epoch": 1.8697225572979495, "grad_norm": 815.1858520507812, "learning_rate": 6.354090354090355e-05, "loss": 0.8377, "step": 1550 }, { "epoch": 1.8998793727382388, "grad_norm": 17.91884994506836, "learning_rate": 6.293040293040293e-05, "loss": 0.7578, "step": 1575 }, { "epoch": 1.9300361881785284, "grad_norm": 53.48311996459961, "learning_rate": 6.231990231990232e-05, "loss": 0.8059, "step": 1600 }, { "epoch": 1.960193003618818, "grad_norm": 16.006927490234375, "learning_rate": 6.170940170940172e-05, "loss": 0.8219, "step": 1625 }, { "epoch": 1.9903498190591074, "grad_norm": 26.18201446533203, "learning_rate": 6.10989010989011e-05, "loss": 0.8062, "step": 1650 }, { "epoch": 2.0205066344993967, "grad_norm": 41.472023010253906, "learning_rate": 6.048840048840049e-05, "loss": 0.782, "step": 1675 }, { "epoch": 2.0506634499396865, "grad_norm": 17.68017578125, "learning_rate": 5.987789987789988e-05, "loss": 0.787, "step": 1700 }, { "epoch": 2.080820265379976, "grad_norm": 16.18324089050293, "learning_rate": 5.9267399267399274e-05, "loss": 0.7642, "step": 1725 }, { "epoch": 2.1109770808202653, "grad_norm": 15.70964527130127, "learning_rate": 5.865689865689866e-05, "loss": 0.793, "step": 1750 }, { "epoch": 2.141133896260555, "grad_norm": 49.760772705078125, "learning_rate": 5.8046398046398054e-05, "loss": 0.8482, "step": 1775 }, { "epoch": 2.1712907117008444, "grad_norm": 15.465973854064941, "learning_rate": 5.7435897435897434e-05, "loss": 0.824, "step": 1800 }, { "epoch": 2.201447527141134, "grad_norm": 27.056657791137695, "learning_rate": 5.682539682539683e-05, "loss": 0.7957, "step": 1825 }, { "epoch": 2.2316043425814236, "grad_norm": 14.316446304321289, "learning_rate": 5.6214896214896215e-05, "loss": 0.807, "step": 1850 }, { "epoch": 2.261761158021713, "grad_norm": 13.680840492248535, "learning_rate": 5.560439560439561e-05, "loss": 0.7543, "step": 1875 }, { "epoch": 2.2919179734620023, "grad_norm": 16.61717987060547, "learning_rate": 5.4993894993895e-05, "loss": 0.7723, "step": 1900 }, { "epoch": 2.3220747889022917, "grad_norm": 21.093223571777344, "learning_rate": 5.438339438339438e-05, "loss": 0.7988, "step": 1925 }, { "epoch": 2.3522316043425815, "grad_norm": 17.148927688598633, "learning_rate": 5.3772893772893775e-05, "loss": 0.7661, "step": 1950 }, { "epoch": 2.382388419782871, "grad_norm": 15.15334415435791, "learning_rate": 5.316239316239316e-05, "loss": 0.8032, "step": 1975 }, { "epoch": 2.4125452352231607, "grad_norm": 15.025490760803223, "learning_rate": 5.2551892551892556e-05, "loss": 0.7666, "step": 2000 }, { "epoch": 2.44270205066345, "grad_norm": 39.71788787841797, "learning_rate": 5.194139194139195e-05, "loss": 0.7697, "step": 2025 }, { "epoch": 2.4728588661037394, "grad_norm": 35.849727630615234, "learning_rate": 5.133089133089133e-05, "loss": 0.7531, "step": 2050 }, { "epoch": 2.5030156815440288, "grad_norm": 34.135833740234375, "learning_rate": 5.072039072039072e-05, "loss": 0.7767, "step": 2075 }, { "epoch": 2.5331724969843186, "grad_norm": 13.79836368560791, "learning_rate": 5.010989010989011e-05, "loss": 0.7961, "step": 2100 }, { "epoch": 2.563329312424608, "grad_norm": 27.408437728881836, "learning_rate": 4.94993894993895e-05, "loss": 0.7926, "step": 2125 }, { "epoch": 2.5934861278648977, "grad_norm": 11.799072265625, "learning_rate": 4.888888888888889e-05, "loss": 0.7315, "step": 2150 }, { "epoch": 2.623642943305187, "grad_norm": 22.04068374633789, "learning_rate": 4.8278388278388283e-05, "loss": 0.749, "step": 2175 }, { "epoch": 2.6537997587454765, "grad_norm": 18.57539939880371, "learning_rate": 4.766788766788767e-05, "loss": 0.7661, "step": 2200 }, { "epoch": 2.683956574185766, "grad_norm": 14.180893898010254, "learning_rate": 4.705738705738706e-05, "loss": 0.7402, "step": 2225 }, { "epoch": 2.7141133896260556, "grad_norm": 15.663522720336914, "learning_rate": 4.644688644688645e-05, "loss": 0.7778, "step": 2250 }, { "epoch": 2.744270205066345, "grad_norm": 14.202746391296387, "learning_rate": 4.583638583638584e-05, "loss": 0.7922, "step": 2275 }, { "epoch": 2.7744270205066344, "grad_norm": 19.26568031311035, "learning_rate": 4.522588522588523e-05, "loss": 0.7454, "step": 2300 }, { "epoch": 2.804583835946924, "grad_norm": 30.88574981689453, "learning_rate": 4.461538461538462e-05, "loss": 0.7383, "step": 2325 }, { "epoch": 2.8347406513872135, "grad_norm": 17.263395309448242, "learning_rate": 4.4004884004884005e-05, "loss": 0.7409, "step": 2350 }, { "epoch": 2.864897466827503, "grad_norm": 15.320577621459961, "learning_rate": 4.33943833943834e-05, "loss": 0.7946, "step": 2375 }, { "epoch": 2.8950542822677923, "grad_norm": 11.715496063232422, "learning_rate": 4.2783882783882785e-05, "loss": 0.7125, "step": 2400 }, { "epoch": 2.925211097708082, "grad_norm": 14.207977294921875, "learning_rate": 4.217338217338218e-05, "loss": 0.7403, "step": 2425 }, { "epoch": 2.9553679131483714, "grad_norm": 16.815935134887695, "learning_rate": 4.1562881562881565e-05, "loss": 0.754, "step": 2450 }, { "epoch": 2.9855247285886612, "grad_norm": 21.52472496032715, "learning_rate": 4.095238095238095e-05, "loss": 0.7502, "step": 2475 }, { "epoch": 3.0156815440289506, "grad_norm": 12.597681999206543, "learning_rate": 4.0341880341880346e-05, "loss": 0.7553, "step": 2500 }, { "epoch": 3.04583835946924, "grad_norm": 15.156243324279785, "learning_rate": 3.973137973137973e-05, "loss": 0.6723, "step": 2525 }, { "epoch": 3.0759951749095293, "grad_norm": 19.262258529663086, "learning_rate": 3.912087912087912e-05, "loss": 0.7203, "step": 2550 }, { "epoch": 3.106151990349819, "grad_norm": 12.042356491088867, "learning_rate": 3.851037851037851e-05, "loss": 0.7715, "step": 2575 }, { "epoch": 3.1363088057901085, "grad_norm": 15.979390144348145, "learning_rate": 3.78998778998779e-05, "loss": 0.7234, "step": 2600 }, { "epoch": 3.166465621230398, "grad_norm": 27.65557861328125, "learning_rate": 3.728937728937729e-05, "loss": 0.6915, "step": 2625 }, { "epoch": 3.1966224366706877, "grad_norm": 28.780155181884766, "learning_rate": 3.667887667887668e-05, "loss": 0.6991, "step": 2650 }, { "epoch": 3.226779252110977, "grad_norm": 23.69841957092285, "learning_rate": 3.606837606837607e-05, "loss": 0.7365, "step": 2675 }, { "epoch": 3.2569360675512664, "grad_norm": 26.912643432617188, "learning_rate": 3.545787545787546e-05, "loss": 0.7455, "step": 2700 }, { "epoch": 3.287092882991556, "grad_norm": 25.788583755493164, "learning_rate": 3.484737484737485e-05, "loss": 0.7622, "step": 2725 }, { "epoch": 3.3172496984318456, "grad_norm": 36.25239181518555, "learning_rate": 3.423687423687424e-05, "loss": 0.7134, "step": 2750 }, { "epoch": 3.347406513872135, "grad_norm": 25.903047561645508, "learning_rate": 3.362637362637363e-05, "loss": 0.7219, "step": 2775 }, { "epoch": 3.3775633293124248, "grad_norm": 53.384639739990234, "learning_rate": 3.3015873015873014e-05, "loss": 0.7554, "step": 2800 }, { "epoch": 3.407720144752714, "grad_norm": 14.4858980178833, "learning_rate": 3.240537240537241e-05, "loss": 0.7368, "step": 2825 }, { "epoch": 3.4378769601930035, "grad_norm": 16.597522735595703, "learning_rate": 3.1794871794871795e-05, "loss": 0.7357, "step": 2850 }, { "epoch": 3.4680337756332933, "grad_norm": 12.060571670532227, "learning_rate": 3.118437118437119e-05, "loss": 0.6797, "step": 2875 }, { "epoch": 3.4981905910735827, "grad_norm": 14.644749641418457, "learning_rate": 3.0573870573870575e-05, "loss": 0.7417, "step": 2900 }, { "epoch": 3.528347406513872, "grad_norm": 12.184840202331543, "learning_rate": 2.9963369963369965e-05, "loss": 0.7043, "step": 2925 }, { "epoch": 3.558504221954162, "grad_norm": 14.599753379821777, "learning_rate": 2.9352869352869355e-05, "loss": 0.7172, "step": 2950 }, { "epoch": 3.588661037394451, "grad_norm": 29.58240509033203, "learning_rate": 2.8742368742368742e-05, "loss": 0.7016, "step": 2975 }, { "epoch": 3.6188178528347406, "grad_norm": 12.626923561096191, "learning_rate": 2.8131868131868132e-05, "loss": 0.719, "step": 3000 }, { "epoch": 3.64897466827503, "grad_norm": 17.7034969329834, "learning_rate": 2.7521367521367526e-05, "loss": 0.6679, "step": 3025 }, { "epoch": 3.6791314837153197, "grad_norm": 16.806182861328125, "learning_rate": 2.6910866910866913e-05, "loss": 0.6795, "step": 3050 }, { "epoch": 3.709288299155609, "grad_norm": 11.212449073791504, "learning_rate": 2.6300366300366303e-05, "loss": 0.6742, "step": 3075 }, { "epoch": 3.739445114595899, "grad_norm": 16.0538272857666, "learning_rate": 2.568986568986569e-05, "loss": 0.6756, "step": 3100 }, { "epoch": 3.7696019300361883, "grad_norm": 18.737091064453125, "learning_rate": 2.507936507936508e-05, "loss": 0.7179, "step": 3125 }, { "epoch": 3.7997587454764776, "grad_norm": 13.685617446899414, "learning_rate": 2.446886446886447e-05, "loss": 0.667, "step": 3150 }, { "epoch": 3.829915560916767, "grad_norm": 16.607234954833984, "learning_rate": 2.385836385836386e-05, "loss": 0.6868, "step": 3175 }, { "epoch": 3.860072376357057, "grad_norm": 16.519487380981445, "learning_rate": 2.324786324786325e-05, "loss": 0.6811, "step": 3200 }, { "epoch": 3.890229191797346, "grad_norm": 20.921655654907227, "learning_rate": 2.2637362637362637e-05, "loss": 0.6893, "step": 3225 }, { "epoch": 3.920386007237636, "grad_norm": 13.694364547729492, "learning_rate": 2.2026862026862027e-05, "loss": 0.6833, "step": 3250 }, { "epoch": 3.9505428226779253, "grad_norm": 15.668880462646484, "learning_rate": 2.1416361416361417e-05, "loss": 0.6897, "step": 3275 }, { "epoch": 3.9806996381182147, "grad_norm": 9.373758316040039, "learning_rate": 2.0805860805860808e-05, "loss": 0.6629, "step": 3300 }, { "epoch": 4.010856453558504, "grad_norm": 17.02540397644043, "learning_rate": 2.0195360195360198e-05, "loss": 0.6492, "step": 3325 }, { "epoch": 4.041013268998793, "grad_norm": 24.518537521362305, "learning_rate": 1.9584859584859585e-05, "loss": 0.6732, "step": 3350 }, { "epoch": 4.071170084439084, "grad_norm": 15.94694995880127, "learning_rate": 1.8974358974358975e-05, "loss": 0.6687, "step": 3375 }, { "epoch": 4.101326899879373, "grad_norm": 12.533055305480957, "learning_rate": 1.8363858363858365e-05, "loss": 0.6745, "step": 3400 }, { "epoch": 4.131483715319662, "grad_norm": 17.95030403137207, "learning_rate": 1.7753357753357755e-05, "loss": 0.6614, "step": 3425 }, { "epoch": 4.161640530759952, "grad_norm": 16.25010108947754, "learning_rate": 1.7142857142857145e-05, "loss": 0.6743, "step": 3450 }, { "epoch": 4.191797346200241, "grad_norm": 16.024572372436523, "learning_rate": 1.6532356532356532e-05, "loss": 0.6436, "step": 3475 }, { "epoch": 4.2219541616405305, "grad_norm": 22.803369522094727, "learning_rate": 1.5921855921855922e-05, "loss": 0.6421, "step": 3500 }, { "epoch": 4.25211097708082, "grad_norm": 14.937376022338867, "learning_rate": 1.5311355311355312e-05, "loss": 0.6565, "step": 3525 }, { "epoch": 4.28226779252111, "grad_norm": 14.354950904846191, "learning_rate": 1.4700854700854703e-05, "loss": 0.6647, "step": 3550 }, { "epoch": 4.3124246079613995, "grad_norm": 17.91471290588379, "learning_rate": 1.4090354090354091e-05, "loss": 0.6313, "step": 3575 }, { "epoch": 4.342581423401689, "grad_norm": 33.2518424987793, "learning_rate": 1.347985347985348e-05, "loss": 0.6264, "step": 3600 }, { "epoch": 4.372738238841978, "grad_norm": 13.307084083557129, "learning_rate": 1.2869352869352868e-05, "loss": 0.6563, "step": 3625 }, { "epoch": 4.402895054282268, "grad_norm": 15.540495872497559, "learning_rate": 1.2258852258852258e-05, "loss": 0.6654, "step": 3650 }, { "epoch": 4.433051869722557, "grad_norm": 14.595681190490723, "learning_rate": 1.1648351648351648e-05, "loss": 0.6633, "step": 3675 }, { "epoch": 4.463208685162847, "grad_norm": 19.460350036621094, "learning_rate": 1.1037851037851039e-05, "loss": 0.6612, "step": 3700 }, { "epoch": 4.4933655006031366, "grad_norm": 13.955097198486328, "learning_rate": 1.0427350427350429e-05, "loss": 0.6768, "step": 3725 }, { "epoch": 4.523522316043426, "grad_norm": 13.924335479736328, "learning_rate": 9.816849816849817e-06, "loss": 0.6451, "step": 3750 }, { "epoch": 4.553679131483715, "grad_norm": 12.562804222106934, "learning_rate": 9.206349206349207e-06, "loss": 0.6408, "step": 3775 }, { "epoch": 4.583835946924005, "grad_norm": 11.879654884338379, "learning_rate": 8.595848595848596e-06, "loss": 0.6746, "step": 3800 }, { "epoch": 4.613992762364294, "grad_norm": 12.780753135681152, "learning_rate": 7.985347985347984e-06, "loss": 0.6322, "step": 3825 }, { "epoch": 4.644149577804583, "grad_norm": 13.602087020874023, "learning_rate": 7.374847374847375e-06, "loss": 0.6343, "step": 3850 }, { "epoch": 4.674306393244874, "grad_norm": 16.94817352294922, "learning_rate": 6.764346764346764e-06, "loss": 0.6351, "step": 3875 }, { "epoch": 4.704463208685163, "grad_norm": 10.593673706054688, "learning_rate": 6.153846153846155e-06, "loss": 0.5892, "step": 3900 }, { "epoch": 4.734620024125452, "grad_norm": 21.8657283782959, "learning_rate": 5.543345543345543e-06, "loss": 0.661, "step": 3925 }, { "epoch": 4.764776839565742, "grad_norm": 14.38430404663086, "learning_rate": 4.932844932844933e-06, "loss": 0.6261, "step": 3950 }, { "epoch": 4.794933655006031, "grad_norm": 9.656220436096191, "learning_rate": 4.322344322344323e-06, "loss": 0.6405, "step": 3975 }, { "epoch": 4.825090470446321, "grad_norm": 13.676799774169922, "learning_rate": 3.711843711843712e-06, "loss": 0.6169, "step": 4000 }, { "epoch": 4.855247285886611, "grad_norm": 11.05395221710205, "learning_rate": 3.1013431013431015e-06, "loss": 0.6447, "step": 4025 }, { "epoch": 4.8854041013269, "grad_norm": 13.141251564025879, "learning_rate": 2.4908424908424913e-06, "loss": 0.6244, "step": 4050 }, { "epoch": 4.915560916767189, "grad_norm": 13.8480863571167, "learning_rate": 1.8803418803418804e-06, "loss": 0.638, "step": 4075 }, { "epoch": 4.945717732207479, "grad_norm": 12.424724578857422, "learning_rate": 1.26984126984127e-06, "loss": 0.6151, "step": 4100 }, { "epoch": 4.975874547647768, "grad_norm": 14.598273277282715, "learning_rate": 6.593406593406594e-07, "loss": 0.6537, "step": 4125 } ], "logging_steps": 25, "max_steps": 4145, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 25, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.058118140105759e+17, "train_batch_size": 1, "trial_name": null, "trial_params": null }