|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 6.806451612903226, |
|
"eval_steps": 16, |
|
"global_step": 434, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.016129032258064516, |
|
"grad_norm": 5.020356178283691, |
|
"learning_rate": 2e-05, |
|
"loss": 2.1533, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.016129032258064516, |
|
"eval_loss": 2.183060884475708, |
|
"eval_runtime": 261.5376, |
|
"eval_samples_per_second": 4.053, |
|
"eval_steps_per_second": 0.677, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.03225806451612903, |
|
"grad_norm": 5.117613315582275, |
|
"learning_rate": 4e-05, |
|
"loss": 2.1211, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.04838709677419355, |
|
"grad_norm": 4.581052303314209, |
|
"learning_rate": 6e-05, |
|
"loss": 2.1036, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.06451612903225806, |
|
"grad_norm": 3.6673312187194824, |
|
"learning_rate": 8e-05, |
|
"loss": 2.0632, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.08064516129032258, |
|
"grad_norm": 3.1003849506378174, |
|
"learning_rate": 0.0001, |
|
"loss": 2.021, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0967741935483871, |
|
"grad_norm": 2.2640504837036133, |
|
"learning_rate": 0.00012, |
|
"loss": 1.9713, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.11290322580645161, |
|
"grad_norm": 2.594027519226074, |
|
"learning_rate": 0.00014, |
|
"loss": 1.8607, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.12903225806451613, |
|
"grad_norm": 2.064988136291504, |
|
"learning_rate": 0.00016, |
|
"loss": 1.8101, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.14516129032258066, |
|
"grad_norm": 1.8494470119476318, |
|
"learning_rate": 0.00018, |
|
"loss": 1.7351, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.16129032258064516, |
|
"grad_norm": 1.8000997304916382, |
|
"learning_rate": 0.0002, |
|
"loss": 1.7415, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.1774193548387097, |
|
"grad_norm": 1.7077090740203857, |
|
"learning_rate": 0.0001999986737997063, |
|
"loss": 1.7155, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.1935483870967742, |
|
"grad_norm": 1.8839540481567383, |
|
"learning_rate": 0.00019999469523400122, |
|
"loss": 1.7474, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.20967741935483872, |
|
"grad_norm": 1.7330329418182373, |
|
"learning_rate": 0.00019998806440841234, |
|
"loss": 1.6775, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.22580645161290322, |
|
"grad_norm": 1.4584600925445557, |
|
"learning_rate": 0.00019997878149881574, |
|
"loss": 1.7176, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.24193548387096775, |
|
"grad_norm": 1.4614368677139282, |
|
"learning_rate": 0.0001999668467514313, |
|
"loss": 1.6047, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.25806451612903225, |
|
"grad_norm": 1.3776947259902954, |
|
"learning_rate": 0.0001999522604828164, |
|
"loss": 1.6416, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.25806451612903225, |
|
"eval_loss": 1.7332706451416016, |
|
"eval_runtime": 431.4384, |
|
"eval_samples_per_second": 2.457, |
|
"eval_steps_per_second": 0.41, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.27419354838709675, |
|
"grad_norm": 1.4929783344268799, |
|
"learning_rate": 0.00019993502307985724, |
|
"loss": 1.73, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.2903225806451613, |
|
"grad_norm": 1.3602776527404785, |
|
"learning_rate": 0.00019991513499975882, |
|
"loss": 1.6958, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.3064516129032258, |
|
"grad_norm": 1.4044336080551147, |
|
"learning_rate": 0.00019989259677003274, |
|
"loss": 1.6475, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.3225806451612903, |
|
"grad_norm": 1.3602012395858765, |
|
"learning_rate": 0.00019986740898848306, |
|
"loss": 1.7063, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.3387096774193548, |
|
"grad_norm": 1.3823881149291992, |
|
"learning_rate": 0.0001998395723231907, |
|
"loss": 1.7079, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.3548387096774194, |
|
"grad_norm": 1.3182601928710938, |
|
"learning_rate": 0.00019980908751249555, |
|
"loss": 1.672, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.3709677419354839, |
|
"grad_norm": 1.2420353889465332, |
|
"learning_rate": 0.00019977595536497687, |
|
"loss": 1.6393, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.3870967741935484, |
|
"grad_norm": 1.3861604928970337, |
|
"learning_rate": 0.00019974017675943192, |
|
"loss": 1.6904, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.4032258064516129, |
|
"grad_norm": 1.537832498550415, |
|
"learning_rate": 0.00019970175264485266, |
|
"loss": 1.6731, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.41935483870967744, |
|
"grad_norm": 1.2361652851104736, |
|
"learning_rate": 0.0001996606840404006, |
|
"loss": 1.6262, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.43548387096774194, |
|
"grad_norm": 1.253831148147583, |
|
"learning_rate": 0.00019961697203537952, |
|
"loss": 1.6068, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.45161290322580644, |
|
"grad_norm": 1.306778073310852, |
|
"learning_rate": 0.00019957061778920701, |
|
"loss": 1.6397, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.46774193548387094, |
|
"grad_norm": 1.1258474588394165, |
|
"learning_rate": 0.0001995216225313833, |
|
"loss": 1.5656, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.4838709677419355, |
|
"grad_norm": 1.198811411857605, |
|
"learning_rate": 0.0001994699875614589, |
|
"loss": 1.6768, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 1.4581488370895386, |
|
"learning_rate": 0.00019941571424900013, |
|
"loss": 1.5823, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.5161290322580645, |
|
"grad_norm": 1.2790226936340332, |
|
"learning_rate": 0.00019935880403355253, |
|
"loss": 1.6154, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.5161290322580645, |
|
"eval_loss": 1.64582359790802, |
|
"eval_runtime": 430.558, |
|
"eval_samples_per_second": 2.462, |
|
"eval_steps_per_second": 0.411, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.532258064516129, |
|
"grad_norm": 1.1969035863876343, |
|
"learning_rate": 0.0001992992584246031, |
|
"loss": 1.6766, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.5483870967741935, |
|
"grad_norm": 1.4593340158462524, |
|
"learning_rate": 0.00019923707900153982, |
|
"loss": 1.6112, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.5645161290322581, |
|
"grad_norm": 1.306646704673767, |
|
"learning_rate": 0.00019917226741361015, |
|
"loss": 1.5671, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.5806451612903226, |
|
"grad_norm": 1.3403998613357544, |
|
"learning_rate": 0.00019910482537987702, |
|
"loss": 1.6703, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.5967741935483871, |
|
"grad_norm": 1.1974356174468994, |
|
"learning_rate": 0.0001990347546891733, |
|
"loss": 1.586, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.6129032258064516, |
|
"grad_norm": 1.3454561233520508, |
|
"learning_rate": 0.0001989620572000544, |
|
"loss": 1.6023, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.6290322580645161, |
|
"grad_norm": 1.2124018669128418, |
|
"learning_rate": 0.000198886734840749, |
|
"loss": 1.6405, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.6451612903225806, |
|
"grad_norm": 1.0802409648895264, |
|
"learning_rate": 0.00019880878960910772, |
|
"loss": 1.5191, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.6612903225806451, |
|
"grad_norm": 1.4059436321258545, |
|
"learning_rate": 0.0001987282235725504, |
|
"loss": 1.5969, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.6774193548387096, |
|
"grad_norm": 1.2623989582061768, |
|
"learning_rate": 0.00019864503886801106, |
|
"loss": 1.5921, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.6935483870967742, |
|
"grad_norm": 1.2654995918273926, |
|
"learning_rate": 0.0001985592377018813, |
|
"loss": 1.5707, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.7096774193548387, |
|
"grad_norm": 1.2158616781234741, |
|
"learning_rate": 0.00019847082234995171, |
|
"loss": 1.5141, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.7258064516129032, |
|
"grad_norm": 1.2419042587280273, |
|
"learning_rate": 0.00019837979515735166, |
|
"loss": 1.5541, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.7419354838709677, |
|
"grad_norm": 1.1441471576690674, |
|
"learning_rate": 0.00019828615853848688, |
|
"loss": 1.4788, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.7580645161290323, |
|
"grad_norm": 1.1924492120742798, |
|
"learning_rate": 0.00019818991497697565, |
|
"loss": 1.5335, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.7741935483870968, |
|
"grad_norm": 1.133042812347412, |
|
"learning_rate": 0.00019809106702558277, |
|
"loss": 1.5155, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.7741935483870968, |
|
"eval_loss": 1.5807298421859741, |
|
"eval_runtime": 433.4129, |
|
"eval_samples_per_second": 2.446, |
|
"eval_steps_per_second": 0.408, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.7903225806451613, |
|
"grad_norm": 1.2338794469833374, |
|
"learning_rate": 0.0001979896173061518, |
|
"loss": 1.611, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.8064516129032258, |
|
"grad_norm": 1.306178331375122, |
|
"learning_rate": 0.0001978855685095358, |
|
"loss": 1.5046, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8225806451612904, |
|
"grad_norm": 1.174604058265686, |
|
"learning_rate": 0.00019777892339552559, |
|
"loss": 1.5431, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.8387096774193549, |
|
"grad_norm": 1.3554447889328003, |
|
"learning_rate": 0.00019766968479277683, |
|
"loss": 1.5713, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.8548387096774194, |
|
"grad_norm": 1.111693263053894, |
|
"learning_rate": 0.00019755785559873488, |
|
"loss": 1.5173, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.8709677419354839, |
|
"grad_norm": 1.1916875839233398, |
|
"learning_rate": 0.00019744343877955788, |
|
"loss": 1.5706, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.8870967741935484, |
|
"grad_norm": 1.1925556659698486, |
|
"learning_rate": 0.00019732643737003827, |
|
"loss": 1.5466, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.9032258064516129, |
|
"grad_norm": 1.174657940864563, |
|
"learning_rate": 0.00019720685447352209, |
|
"loss": 1.5281, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.9193548387096774, |
|
"grad_norm": 1.1206492185592651, |
|
"learning_rate": 0.00019708469326182678, |
|
"loss": 1.5871, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.9354838709677419, |
|
"grad_norm": 1.089790940284729, |
|
"learning_rate": 0.0001969599569751571, |
|
"loss": 1.5988, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.9516129032258065, |
|
"grad_norm": 1.146849274635315, |
|
"learning_rate": 0.000196832648922019, |
|
"loss": 1.5066, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.967741935483871, |
|
"grad_norm": 1.0739450454711914, |
|
"learning_rate": 0.00019670277247913205, |
|
"loss": 1.5486, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.9838709677419355, |
|
"grad_norm": 1.0958797931671143, |
|
"learning_rate": 0.00019657033109133975, |
|
"loss": 1.5212, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 1.134975552558899, |
|
"learning_rate": 0.0001964353282715183, |
|
"loss": 1.5581, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 1.0161290322580645, |
|
"grad_norm": 1.1783857345581055, |
|
"learning_rate": 0.0001962977676004832, |
|
"loss": 1.4905, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 1.032258064516129, |
|
"grad_norm": 1.0987132787704468, |
|
"learning_rate": 0.00019615765272689461, |
|
"loss": 1.5359, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.032258064516129, |
|
"eval_loss": 1.5370668172836304, |
|
"eval_runtime": 379.6837, |
|
"eval_samples_per_second": 2.792, |
|
"eval_steps_per_second": 0.466, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 1.0161290322580645, |
|
"grad_norm": 1.0743085145950317, |
|
"learning_rate": 0.00019601498736716017, |
|
"loss": 1.1431, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.032258064516129, |
|
"grad_norm": 0.9797853231430054, |
|
"learning_rate": 0.00019586977530533677, |
|
"loss": 1.0969, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 1.0483870967741935, |
|
"grad_norm": 1.1307501792907715, |
|
"learning_rate": 0.00019572202039303, |
|
"loss": 1.0826, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 1.064516129032258, |
|
"grad_norm": 1.0152018070220947, |
|
"learning_rate": 0.00019557172654929196, |
|
"loss": 1.1224, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 1.0806451612903225, |
|
"grad_norm": 1.0957362651824951, |
|
"learning_rate": 0.0001954188977605175, |
|
"loss": 1.0978, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 1.096774193548387, |
|
"grad_norm": 1.0177264213562012, |
|
"learning_rate": 0.00019526353808033825, |
|
"loss": 1.0882, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.1129032258064515, |
|
"grad_norm": 1.0174063444137573, |
|
"learning_rate": 0.00019510565162951537, |
|
"loss": 1.0331, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 1.129032258064516, |
|
"grad_norm": 1.0872232913970947, |
|
"learning_rate": 0.00019494524259582992, |
|
"loss": 1.0552, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 1.1451612903225807, |
|
"grad_norm": 1.0081063508987427, |
|
"learning_rate": 0.00019478231523397215, |
|
"loss": 1.023, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 1.1612903225806452, |
|
"grad_norm": 1.083362340927124, |
|
"learning_rate": 0.00019461687386542826, |
|
"loss": 1.0639, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 1.1774193548387097, |
|
"grad_norm": 1.1271260976791382, |
|
"learning_rate": 0.00019444892287836613, |
|
"loss": 1.053, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.1935483870967742, |
|
"grad_norm": 1.1567646265029907, |
|
"learning_rate": 0.00019427846672751873, |
|
"loss": 1.0214, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 1.2096774193548387, |
|
"grad_norm": 1.1053016185760498, |
|
"learning_rate": 0.00019410550993406603, |
|
"loss": 1.0552, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 1.2258064516129032, |
|
"grad_norm": 1.0573008060455322, |
|
"learning_rate": 0.00019393005708551498, |
|
"loss": 1.0223, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 1.2419354838709677, |
|
"grad_norm": 1.081141471862793, |
|
"learning_rate": 0.00019375211283557798, |
|
"loss": 1.0422, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 1.2580645161290323, |
|
"grad_norm": 1.154768943786621, |
|
"learning_rate": 0.00019357168190404936, |
|
"loss": 1.0746, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.2580645161290323, |
|
"eval_loss": 1.5887596607208252, |
|
"eval_runtime": 379.2194, |
|
"eval_samples_per_second": 2.795, |
|
"eval_steps_per_second": 0.467, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.2741935483870968, |
|
"grad_norm": 1.0481725931167603, |
|
"learning_rate": 0.00019338876907668026, |
|
"loss": 1.0684, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 1.2903225806451613, |
|
"grad_norm": 1.1526888608932495, |
|
"learning_rate": 0.00019320337920505153, |
|
"loss": 1.104, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 1.3064516129032258, |
|
"grad_norm": 1.108069658279419, |
|
"learning_rate": 0.00019301551720644523, |
|
"loss": 1.0907, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 1.3225806451612903, |
|
"grad_norm": 1.1018728017807007, |
|
"learning_rate": 0.00019282518806371414, |
|
"loss": 1.0699, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 1.3387096774193548, |
|
"grad_norm": 1.157453179359436, |
|
"learning_rate": 0.00019263239682514952, |
|
"loss": 1.0589, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.3548387096774195, |
|
"grad_norm": 1.116068959236145, |
|
"learning_rate": 0.0001924371486043473, |
|
"loss": 1.0847, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 1.370967741935484, |
|
"grad_norm": 1.0575870275497437, |
|
"learning_rate": 0.00019223944858007253, |
|
"loss": 1.0688, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 1.3870967741935485, |
|
"grad_norm": 1.1120954751968384, |
|
"learning_rate": 0.0001920393019961217, |
|
"loss": 1.0829, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 1.403225806451613, |
|
"grad_norm": 1.0838677883148193, |
|
"learning_rate": 0.00019183671416118405, |
|
"loss": 1.0751, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 1.4193548387096775, |
|
"grad_norm": 1.1601375341415405, |
|
"learning_rate": 0.0001916316904487005, |
|
"loss": 1.0348, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.435483870967742, |
|
"grad_norm": 1.080993413925171, |
|
"learning_rate": 0.00019142423629672117, |
|
"loss": 1.0512, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 1.4516129032258065, |
|
"grad_norm": 1.4107691049575806, |
|
"learning_rate": 0.00019121435720776122, |
|
"loss": 1.0834, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 1.467741935483871, |
|
"grad_norm": 1.10901939868927, |
|
"learning_rate": 0.00019100205874865485, |
|
"loss": 1.127, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 1.4838709677419355, |
|
"grad_norm": 1.040759563446045, |
|
"learning_rate": 0.0001907873465504076, |
|
"loss": 1.0838, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 1.117697834968567, |
|
"learning_rate": 0.00019057022630804716, |
|
"loss": 1.094, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.5161290322580645, |
|
"grad_norm": 1.1124318838119507, |
|
"learning_rate": 0.00019035070378047204, |
|
"loss": 1.0806, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 1.5161290322580645, |
|
"eval_loss": 1.569594144821167, |
|
"eval_runtime": 379.9276, |
|
"eval_samples_per_second": 2.79, |
|
"eval_steps_per_second": 0.466, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 1.532258064516129, |
|
"grad_norm": 1.0274182558059692, |
|
"learning_rate": 0.00019012878479029906, |
|
"loss": 1.0322, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 1.5483870967741935, |
|
"grad_norm": 1.214774250984192, |
|
"learning_rate": 0.00018990447522370884, |
|
"loss": 1.1029, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 1.564516129032258, |
|
"grad_norm": 1.146145224571228, |
|
"learning_rate": 0.00018967778103028967, |
|
"loss": 1.0939, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 1.5806451612903225, |
|
"grad_norm": 1.115272045135498, |
|
"learning_rate": 0.00018944870822287956, |
|
"loss": 1.0795, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.596774193548387, |
|
"grad_norm": 1.0580235719680786, |
|
"learning_rate": 0.0001892172628774071, |
|
"loss": 1.0887, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 1.6129032258064515, |
|
"grad_norm": 1.1681277751922607, |
|
"learning_rate": 0.00018898345113272998, |
|
"loss": 1.0781, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 1.629032258064516, |
|
"grad_norm": 1.1023615598678589, |
|
"learning_rate": 0.00018874727919047227, |
|
"loss": 1.0679, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 1.6451612903225805, |
|
"grad_norm": 1.139622688293457, |
|
"learning_rate": 0.00018850875331485995, |
|
"loss": 1.0795, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 1.661290322580645, |
|
"grad_norm": 1.0712463855743408, |
|
"learning_rate": 0.00018826787983255473, |
|
"loss": 1.11, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.6774193548387095, |
|
"grad_norm": 1.0515589714050293, |
|
"learning_rate": 0.00018802466513248632, |
|
"loss": 1.0896, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 1.6935483870967742, |
|
"grad_norm": 1.0597035884857178, |
|
"learning_rate": 0.00018777911566568282, |
|
"loss": 1.0761, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 1.7096774193548387, |
|
"grad_norm": 1.11446213722229, |
|
"learning_rate": 0.00018753123794509974, |
|
"loss": 1.0813, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 1.7258064516129032, |
|
"grad_norm": 1.0365442037582397, |
|
"learning_rate": 0.0001872810385454472, |
|
"loss": 1.1037, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 1.7419354838709677, |
|
"grad_norm": 1.0622906684875488, |
|
"learning_rate": 0.00018702852410301554, |
|
"loss": 1.085, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.7580645161290323, |
|
"grad_norm": 1.0445847511291504, |
|
"learning_rate": 0.0001867737013154993, |
|
"loss": 1.1109, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 1.7741935483870968, |
|
"grad_norm": 1.066573977470398, |
|
"learning_rate": 0.0001865165769418196, |
|
"loss": 1.0348, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 1.7741935483870968, |
|
"eval_loss": 1.5535695552825928, |
|
"eval_runtime": 385.6979, |
|
"eval_samples_per_second": 2.748, |
|
"eval_steps_per_second": 0.459, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 1.7903225806451613, |
|
"grad_norm": 1.1161065101623535, |
|
"learning_rate": 0.00018625715780194485, |
|
"loss": 1.0733, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 1.8064516129032258, |
|
"grad_norm": 1.0873162746429443, |
|
"learning_rate": 0.00018599545077670985, |
|
"loss": 1.0875, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 1.8225806451612905, |
|
"grad_norm": 2.7381904125213623, |
|
"learning_rate": 0.00018573146280763324, |
|
"loss": 1.1284, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.838709677419355, |
|
"grad_norm": 1.0686196088790894, |
|
"learning_rate": 0.0001854652008967335, |
|
"loss": 1.064, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 1.8548387096774195, |
|
"grad_norm": 1.0897347927093506, |
|
"learning_rate": 0.0001851966721063431, |
|
"loss": 1.0418, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 1.870967741935484, |
|
"grad_norm": 1.092289686203003, |
|
"learning_rate": 0.00018492588355892124, |
|
"loss": 1.1247, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 1.8870967741935485, |
|
"grad_norm": 1.1446326971054077, |
|
"learning_rate": 0.00018465284243686494, |
|
"loss": 1.0789, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 1.903225806451613, |
|
"grad_norm": 1.052269458770752, |
|
"learning_rate": 0.00018437755598231856, |
|
"loss": 1.0792, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.9193548387096775, |
|
"grad_norm": 1.1960537433624268, |
|
"learning_rate": 0.00018410003149698162, |
|
"loss": 1.0551, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 1.935483870967742, |
|
"grad_norm": 1.2509639263153076, |
|
"learning_rate": 0.00018382027634191524, |
|
"loss": 1.1336, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 1.9516129032258065, |
|
"grad_norm": 1.1205637454986572, |
|
"learning_rate": 0.00018353829793734669, |
|
"loss": 1.082, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 1.967741935483871, |
|
"grad_norm": 1.087641954421997, |
|
"learning_rate": 0.00018325410376247294, |
|
"loss": 1.0999, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 1.9838709677419355, |
|
"grad_norm": 1.141137957572937, |
|
"learning_rate": 0.0001829677013552619, |
|
"loss": 1.1358, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 1.1019104719161987, |
|
"learning_rate": 0.0001826790983122527, |
|
"loss": 1.0671, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 2.0161290322580645, |
|
"grad_norm": 1.0500136613845825, |
|
"learning_rate": 0.00018238830228835417, |
|
"loss": 1.1061, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 2.032258064516129, |
|
"grad_norm": 1.0191640853881836, |
|
"learning_rate": 0.00018209532099664174, |
|
"loss": 1.0769, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 2.032258064516129, |
|
"eval_loss": 1.5341366529464722, |
|
"eval_runtime": 377.1489, |
|
"eval_samples_per_second": 2.811, |
|
"eval_steps_per_second": 0.469, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 2.0161290322580645, |
|
"grad_norm": 0.8983094692230225, |
|
"learning_rate": 0.00018180016220815292, |
|
"loss": 0.6901, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 2.032258064516129, |
|
"grad_norm": 0.9102672934532166, |
|
"learning_rate": 0.00018150283375168114, |
|
"loss": 0.6643, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.0483870967741935, |
|
"grad_norm": 0.9043310284614563, |
|
"learning_rate": 0.00018120334351356813, |
|
"loss": 0.6813, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 2.064516129032258, |
|
"grad_norm": 0.920064389705658, |
|
"learning_rate": 0.00018090169943749476, |
|
"loss": 0.6422, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 2.0806451612903225, |
|
"grad_norm": 1.0108799934387207, |
|
"learning_rate": 0.0001805979095242702, |
|
"loss": 0.6727, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 2.096774193548387, |
|
"grad_norm": 0.9978906512260437, |
|
"learning_rate": 0.00018029198183161998, |
|
"loss": 0.6705, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 2.1129032258064515, |
|
"grad_norm": 0.9658149480819702, |
|
"learning_rate": 0.00017998392447397197, |
|
"loss": 0.6519, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.129032258064516, |
|
"grad_norm": 0.9641355872154236, |
|
"learning_rate": 0.00017967374562224132, |
|
"loss": 0.6408, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 2.1451612903225805, |
|
"grad_norm": 1.057494878768921, |
|
"learning_rate": 0.0001793614535036137, |
|
"loss": 0.6903, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 2.161290322580645, |
|
"grad_norm": 1.0118696689605713, |
|
"learning_rate": 0.00017904705640132718, |
|
"loss": 0.6606, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 2.1774193548387095, |
|
"grad_norm": 0.9846657514572144, |
|
"learning_rate": 0.0001787305626544523, |
|
"loss": 0.6712, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 2.193548387096774, |
|
"grad_norm": 1.151206374168396, |
|
"learning_rate": 0.00017841198065767107, |
|
"loss": 0.6781, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.2096774193548385, |
|
"grad_norm": 0.9539623856544495, |
|
"learning_rate": 0.0001780913188610542, |
|
"loss": 0.6253, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 2.225806451612903, |
|
"grad_norm": 1.044904351234436, |
|
"learning_rate": 0.00017776858576983712, |
|
"loss": 0.6617, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 2.241935483870968, |
|
"grad_norm": 0.9991489052772522, |
|
"learning_rate": 0.0001774437899441942, |
|
"loss": 0.6703, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 2.258064516129032, |
|
"grad_norm": 1.0453065633773804, |
|
"learning_rate": 0.0001771169399990119, |
|
"loss": 0.6608, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.258064516129032, |
|
"eval_loss": 1.6200847625732422, |
|
"eval_runtime": 371.4426, |
|
"eval_samples_per_second": 2.854, |
|
"eval_steps_per_second": 0.477, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 2.274193548387097, |
|
"grad_norm": 0.9984713196754456, |
|
"learning_rate": 0.00017678804460366, |
|
"loss": 0.6976, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.2903225806451615, |
|
"grad_norm": 0.9991441369056702, |
|
"learning_rate": 0.00017645711248176195, |
|
"loss": 0.6872, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 2.306451612903226, |
|
"grad_norm": 2.5695972442626953, |
|
"learning_rate": 0.00017612415241096327, |
|
"loss": 0.6672, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 2.3225806451612905, |
|
"grad_norm": 0.9834401607513428, |
|
"learning_rate": 0.00017578917322269886, |
|
"loss": 0.6764, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 2.338709677419355, |
|
"grad_norm": 1.0587729215621948, |
|
"learning_rate": 0.00017545218380195864, |
|
"loss": 0.6741, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 2.3548387096774195, |
|
"grad_norm": 1.0295051336288452, |
|
"learning_rate": 0.00017511319308705198, |
|
"loss": 0.6963, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.370967741935484, |
|
"grad_norm": 1.0167349576950073, |
|
"learning_rate": 0.00017477221006937066, |
|
"loss": 0.7085, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 2.3870967741935485, |
|
"grad_norm": 1.0259524583816528, |
|
"learning_rate": 0.0001744292437931502, |
|
"loss": 0.7143, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 2.403225806451613, |
|
"grad_norm": 1.0057259798049927, |
|
"learning_rate": 0.00017408430335523012, |
|
"loss": 0.708, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 2.4193548387096775, |
|
"grad_norm": 0.975724995136261, |
|
"learning_rate": 0.00017373739790481262, |
|
"loss": 0.696, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 2.435483870967742, |
|
"grad_norm": 1.0695990324020386, |
|
"learning_rate": 0.00017338853664321992, |
|
"loss": 0.6734, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.4516129032258065, |
|
"grad_norm": 1.3752026557922363, |
|
"learning_rate": 0.00017303772882365016, |
|
"loss": 0.6976, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 2.467741935483871, |
|
"grad_norm": 2.8574488162994385, |
|
"learning_rate": 0.00017268498375093202, |
|
"loss": 0.7288, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 2.4838709677419355, |
|
"grad_norm": 1.0393046140670776, |
|
"learning_rate": 0.00017233031078127788, |
|
"loss": 0.6664, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 1.456668496131897, |
|
"learning_rate": 0.00017197371932203568, |
|
"loss": 0.6944, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 2.5161290322580645, |
|
"grad_norm": 1.0640977621078491, |
|
"learning_rate": 0.00017161521883143934, |
|
"loss": 0.6918, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.5161290322580645, |
|
"eval_loss": 1.6184865236282349, |
|
"eval_runtime": 239.7727, |
|
"eval_samples_per_second": 4.421, |
|
"eval_steps_per_second": 0.738, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.532258064516129, |
|
"grad_norm": 1.1856733560562134, |
|
"learning_rate": 0.000171254818818358, |
|
"loss": 0.7186, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 2.5483870967741935, |
|
"grad_norm": 1.0068649053573608, |
|
"learning_rate": 0.00017089252884204377, |
|
"loss": 0.6656, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 2.564516129032258, |
|
"grad_norm": 1.055477499961853, |
|
"learning_rate": 0.00017052835851187804, |
|
"loss": 0.7235, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 2.5806451612903225, |
|
"grad_norm": 1.0402213335037231, |
|
"learning_rate": 0.0001701623174871168, |
|
"loss": 0.7212, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 2.596774193548387, |
|
"grad_norm": 1.0087556838989258, |
|
"learning_rate": 0.00016979441547663435, |
|
"loss": 0.7124, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.6129032258064515, |
|
"grad_norm": 1.9748854637145996, |
|
"learning_rate": 0.0001694246622386658, |
|
"loss": 0.732, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 2.629032258064516, |
|
"grad_norm": 1.0986181497573853, |
|
"learning_rate": 0.0001690530675805482, |
|
"loss": 0.7063, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 2.6451612903225805, |
|
"grad_norm": 0.9828779101371765, |
|
"learning_rate": 0.00016867964135846043, |
|
"loss": 0.7026, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 2.661290322580645, |
|
"grad_norm": 1.0073943138122559, |
|
"learning_rate": 0.0001683043934771618, |
|
"loss": 0.7084, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 2.6774193548387095, |
|
"grad_norm": 1.0138689279556274, |
|
"learning_rate": 0.00016792733388972932, |
|
"loss": 0.7144, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.693548387096774, |
|
"grad_norm": 1.0346741676330566, |
|
"learning_rate": 0.00016754847259729369, |
|
"loss": 0.686, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 2.709677419354839, |
|
"grad_norm": 1.0213960409164429, |
|
"learning_rate": 0.0001671678196487741, |
|
"loss": 0.6537, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 2.725806451612903, |
|
"grad_norm": 0.9593885540962219, |
|
"learning_rate": 0.00016678538514061154, |
|
"loss": 0.7029, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 2.741935483870968, |
|
"grad_norm": 1.0341320037841797, |
|
"learning_rate": 0.00016640117921650117, |
|
"loss": 0.6969, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 2.758064516129032, |
|
"grad_norm": 1.1095786094665527, |
|
"learning_rate": 0.00016601521206712318, |
|
"loss": 0.7048, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.774193548387097, |
|
"grad_norm": 0.9925594329833984, |
|
"learning_rate": 0.00016562749392987254, |
|
"loss": 0.7203, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 2.774193548387097, |
|
"eval_loss": 1.6154118776321411, |
|
"eval_runtime": 232.8684, |
|
"eval_samples_per_second": 4.552, |
|
"eval_steps_per_second": 0.76, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 2.790322580645161, |
|
"grad_norm": 0.9950226545333862, |
|
"learning_rate": 0.00016523803508858732, |
|
"loss": 0.6853, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 2.806451612903226, |
|
"grad_norm": 1.0214858055114746, |
|
"learning_rate": 0.0001648468458732762, |
|
"loss": 0.7098, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 2.8225806451612905, |
|
"grad_norm": 1.0195788145065308, |
|
"learning_rate": 0.00016445393665984417, |
|
"loss": 0.6775, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 2.838709677419355, |
|
"grad_norm": 0.9992055892944336, |
|
"learning_rate": 0.00016405931786981755, |
|
"loss": 0.6981, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.8548387096774195, |
|
"grad_norm": 1.0090147256851196, |
|
"learning_rate": 0.0001636629999700674, |
|
"loss": 0.7083, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 2.870967741935484, |
|
"grad_norm": 1.0279518365859985, |
|
"learning_rate": 0.00016326499347253207, |
|
"loss": 0.7097, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 2.8870967741935485, |
|
"grad_norm": 0.9872741103172302, |
|
"learning_rate": 0.00016286530893393818, |
|
"loss": 0.7185, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 2.903225806451613, |
|
"grad_norm": 1.0596610307693481, |
|
"learning_rate": 0.00016246395695552085, |
|
"loss": 0.6818, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 2.9193548387096775, |
|
"grad_norm": 5.755076885223389, |
|
"learning_rate": 0.00016206094818274229, |
|
"loss": 0.7244, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 2.935483870967742, |
|
"grad_norm": 1.0628125667572021, |
|
"learning_rate": 0.00016165629330500952, |
|
"loss": 0.7094, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 2.9516129032258065, |
|
"grad_norm": 1.0117448568344116, |
|
"learning_rate": 0.00016125000305539094, |
|
"loss": 0.7052, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 2.967741935483871, |
|
"grad_norm": 1.0038176774978638, |
|
"learning_rate": 0.0001608420882103315, |
|
"loss": 0.6832, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 2.9838709677419355, |
|
"grad_norm": 1.0732207298278809, |
|
"learning_rate": 0.00016043255958936693, |
|
"loss": 0.7367, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"grad_norm": 1.056867003440857, |
|
"learning_rate": 0.00016002142805483685, |
|
"loss": 0.674, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 3.0161290322580645, |
|
"grad_norm": 1.0539047718048096, |
|
"learning_rate": 0.00015960870451159638, |
|
"loss": 0.724, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 3.032258064516129, |
|
"grad_norm": 1.0756276845932007, |
|
"learning_rate": 0.0001591943999067273, |
|
"loss": 0.7172, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 3.032258064516129, |
|
"eval_loss": 1.6202313899993896, |
|
"eval_runtime": 227.1127, |
|
"eval_samples_per_second": 4.667, |
|
"eval_steps_per_second": 0.779, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 3.0161290322580645, |
|
"grad_norm": 0.8275887966156006, |
|
"learning_rate": 0.00015877852522924732, |
|
"loss": 0.4273, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 3.032258064516129, |
|
"grad_norm": 0.8425289988517761, |
|
"learning_rate": 0.00015836109150981886, |
|
"loss": 0.3975, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 3.0483870967741935, |
|
"grad_norm": 0.8054454922676086, |
|
"learning_rate": 0.00015794210982045636, |
|
"loss": 0.4053, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 3.064516129032258, |
|
"grad_norm": 0.9864973425865173, |
|
"learning_rate": 0.00015752159127423263, |
|
"loss": 0.3912, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 3.0806451612903225, |
|
"grad_norm": 4.44207763671875, |
|
"learning_rate": 0.00015709954702498397, |
|
"loss": 0.4295, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 3.096774193548387, |
|
"grad_norm": 1.4279245138168335, |
|
"learning_rate": 0.0001566759882670146, |
|
"loss": 0.458, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 3.1129032258064515, |
|
"grad_norm": 0.9974954724311829, |
|
"learning_rate": 0.0001562509262347995, |
|
"loss": 0.4105, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 3.129032258064516, |
|
"grad_norm": 1.0291105508804321, |
|
"learning_rate": 0.00015582437220268647, |
|
"loss": 0.3779, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.1451612903225805, |
|
"grad_norm": 0.9659109711647034, |
|
"learning_rate": 0.00015539633748459724, |
|
"loss": 0.4046, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 3.161290322580645, |
|
"grad_norm": 1.0539616346359253, |
|
"learning_rate": 0.0001549668334337271, |
|
"loss": 0.4101, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 3.1774193548387095, |
|
"grad_norm": 1.0015851259231567, |
|
"learning_rate": 0.00015453587144224407, |
|
"loss": 0.4184, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 3.193548387096774, |
|
"grad_norm": 0.9041028618812561, |
|
"learning_rate": 0.0001541034629409865, |
|
"loss": 0.3943, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 3.2096774193548385, |
|
"grad_norm": 0.922798216342926, |
|
"learning_rate": 0.00015366961939916008, |
|
"loss": 0.3937, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 3.225806451612903, |
|
"grad_norm": 0.9305877685546875, |
|
"learning_rate": 0.00015323435232403337, |
|
"loss": 0.3796, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 3.241935483870968, |
|
"grad_norm": 0.9297038316726685, |
|
"learning_rate": 0.00015279767326063296, |
|
"loss": 0.404, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 3.258064516129032, |
|
"grad_norm": 0.8899175524711609, |
|
"learning_rate": 0.00015235959379143678, |
|
"loss": 0.3914, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 3.258064516129032, |
|
"eval_loss": 1.7161870002746582, |
|
"eval_runtime": 361.1949, |
|
"eval_samples_per_second": 2.935, |
|
"eval_steps_per_second": 0.49, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 3.274193548387097, |
|
"grad_norm": 0.9059929847717285, |
|
"learning_rate": 0.0001519201255360673, |
|
"loss": 0.3761, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 3.2903225806451615, |
|
"grad_norm": 1.0878820419311523, |
|
"learning_rate": 0.0001514792801509831, |
|
"loss": 0.4018, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.306451612903226, |
|
"grad_norm": 0.9320698976516724, |
|
"learning_rate": 0.00015103706932916974, |
|
"loss": 0.398, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 3.3225806451612905, |
|
"grad_norm": 0.8879979252815247, |
|
"learning_rate": 0.00015059350479982965, |
|
"loss": 0.3848, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 3.338709677419355, |
|
"grad_norm": 0.8591129779815674, |
|
"learning_rate": 0.00015014859832807105, |
|
"loss": 0.3935, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 3.3548387096774195, |
|
"grad_norm": 0.8328726887702942, |
|
"learning_rate": 0.0001497023617145958, |
|
"loss": 0.4025, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 3.370967741935484, |
|
"grad_norm": 0.9114816188812256, |
|
"learning_rate": 0.00014925480679538647, |
|
"loss": 0.3978, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 3.3870967741935485, |
|
"grad_norm": 0.8772809505462646, |
|
"learning_rate": 0.0001488059454413923, |
|
"loss": 0.3936, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 3.403225806451613, |
|
"grad_norm": 0.96305251121521, |
|
"learning_rate": 0.00014835578955821457, |
|
"loss": 0.4098, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 3.4193548387096775, |
|
"grad_norm": 0.8828062415122986, |
|
"learning_rate": 0.00014790435108579048, |
|
"loss": 0.3994, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 3.435483870967742, |
|
"grad_norm": 1.0111607313156128, |
|
"learning_rate": 0.00014745164199807682, |
|
"loss": 0.4002, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 3.4516129032258065, |
|
"grad_norm": 0.9364872574806213, |
|
"learning_rate": 0.000146997674302732, |
|
"loss": 0.4115, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.467741935483871, |
|
"grad_norm": 0.9206238389015198, |
|
"learning_rate": 0.00014654246004079793, |
|
"loss": 0.3979, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 3.4838709677419355, |
|
"grad_norm": 0.910442590713501, |
|
"learning_rate": 0.00014608601128638027, |
|
"loss": 0.4041, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 3.5, |
|
"grad_norm": 0.9023163914680481, |
|
"learning_rate": 0.00014562834014632853, |
|
"loss": 0.3936, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 3.5161290322580645, |
|
"grad_norm": 1.0501468181610107, |
|
"learning_rate": 0.00014516945875991472, |
|
"loss": 0.4111, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 3.5161290322580645, |
|
"eval_loss": 1.711409091949463, |
|
"eval_runtime": 360.3679, |
|
"eval_samples_per_second": 2.941, |
|
"eval_steps_per_second": 0.491, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 3.532258064516129, |
|
"grad_norm": 1.386006474494934, |
|
"learning_rate": 0.0001447093792985114, |
|
"loss": 0.4127, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.5483870967741935, |
|
"grad_norm": 0.8904193639755249, |
|
"learning_rate": 0.00014424811396526892, |
|
"loss": 0.4089, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 3.564516129032258, |
|
"grad_norm": 0.8737647533416748, |
|
"learning_rate": 0.00014378567499479168, |
|
"loss": 0.3884, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 3.5806451612903225, |
|
"grad_norm": 0.914535641670227, |
|
"learning_rate": 0.00014332207465281364, |
|
"loss": 0.4217, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 3.596774193548387, |
|
"grad_norm": 0.9375549554824829, |
|
"learning_rate": 0.00014285732523587304, |
|
"loss": 0.4092, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 3.6129032258064515, |
|
"grad_norm": 0.8925461173057556, |
|
"learning_rate": 0.0001423914390709861, |
|
"loss": 0.3956, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.629032258064516, |
|
"grad_norm": 0.8751139640808105, |
|
"learning_rate": 0.00014192442851532018, |
|
"loss": 0.413, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 3.6451612903225805, |
|
"grad_norm": 0.8749066591262817, |
|
"learning_rate": 0.00014145630595586607, |
|
"loss": 0.4015, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 3.661290322580645, |
|
"grad_norm": 0.9000979661941528, |
|
"learning_rate": 0.0001409870838091092, |
|
"loss": 0.4117, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 3.6774193548387095, |
|
"grad_norm": 0.909804105758667, |
|
"learning_rate": 0.00014051677452070065, |
|
"loss": 0.3782, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 3.693548387096774, |
|
"grad_norm": 1.7925716638565063, |
|
"learning_rate": 0.00014004539056512667, |
|
"loss": 0.4451, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 3.709677419354839, |
|
"grad_norm": 0.9784117937088013, |
|
"learning_rate": 0.00013957294444537808, |
|
"loss": 0.4252, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 3.725806451612903, |
|
"grad_norm": 0.9990687370300293, |
|
"learning_rate": 0.00013909944869261855, |
|
"loss": 0.4384, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 3.741935483870968, |
|
"grad_norm": 1.0078152418136597, |
|
"learning_rate": 0.0001386249158658522, |
|
"loss": 0.4193, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 3.758064516129032, |
|
"grad_norm": 0.9793704152107239, |
|
"learning_rate": 0.00013814935855159055, |
|
"loss": 0.4253, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 3.774193548387097, |
|
"grad_norm": 0.9232134819030762, |
|
"learning_rate": 0.00013767278936351854, |
|
"loss": 0.4091, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.774193548387097, |
|
"eval_loss": 1.717702865600586, |
|
"eval_runtime": 359.9193, |
|
"eval_samples_per_second": 2.945, |
|
"eval_steps_per_second": 0.492, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.790322580645161, |
|
"grad_norm": 1.0362968444824219, |
|
"learning_rate": 0.00013719522094216013, |
|
"loss": 0.418, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 3.806451612903226, |
|
"grad_norm": 0.9517565369606018, |
|
"learning_rate": 0.00013671666595454295, |
|
"loss": 0.4343, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 3.8225806451612905, |
|
"grad_norm": 0.9661825895309448, |
|
"learning_rate": 0.00013623713709386227, |
|
"loss": 0.3973, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 3.838709677419355, |
|
"grad_norm": 0.9564940333366394, |
|
"learning_rate": 0.00013575664707914448, |
|
"loss": 0.3935, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 3.8548387096774195, |
|
"grad_norm": 0.9845961928367615, |
|
"learning_rate": 0.0001352752086549095, |
|
"loss": 0.4374, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 3.870967741935484, |
|
"grad_norm": 0.9213180541992188, |
|
"learning_rate": 0.0001347928345908329, |
|
"loss": 0.4253, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 3.8870967741935485, |
|
"grad_norm": 0.8919222950935364, |
|
"learning_rate": 0.00013430953768140723, |
|
"loss": 0.4148, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 3.903225806451613, |
|
"grad_norm": 0.899038553237915, |
|
"learning_rate": 0.00013382533074560255, |
|
"loss": 0.419, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 3.9193548387096775, |
|
"grad_norm": 0.9399131536483765, |
|
"learning_rate": 0.00013334022662652649, |
|
"loss": 0.4081, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 3.935483870967742, |
|
"grad_norm": 0.8963966965675354, |
|
"learning_rate": 0.0001328542381910835, |
|
"loss": 0.4036, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 3.9516129032258065, |
|
"grad_norm": 0.8994773626327515, |
|
"learning_rate": 0.0001323673783296337, |
|
"loss": 0.4138, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 3.967741935483871, |
|
"grad_norm": 0.9611584544181824, |
|
"learning_rate": 0.00013187965995565098, |
|
"loss": 0.4416, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 3.9838709677419355, |
|
"grad_norm": 0.9271295666694641, |
|
"learning_rate": 0.00013139109600538028, |
|
"loss": 0.4009, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"grad_norm": 0.9294149875640869, |
|
"learning_rate": 0.00013090169943749476, |
|
"loss": 0.4187, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 4.016129032258064, |
|
"grad_norm": 0.9005634188652039, |
|
"learning_rate": 0.0001304114832327518, |
|
"loss": 0.4068, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 4.032258064516129, |
|
"grad_norm": 0.9278081655502319, |
|
"learning_rate": 0.00012992046039364893, |
|
"loss": 0.4103, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 4.032258064516129, |
|
"eval_loss": 1.719123363494873, |
|
"eval_runtime": 366.7154, |
|
"eval_samples_per_second": 2.891, |
|
"eval_steps_per_second": 0.483, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 4.016129032258065, |
|
"grad_norm": 0.6807416677474976, |
|
"learning_rate": 0.00012942864394407879, |
|
"loss": 0.2192, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 4.032258064516129, |
|
"grad_norm": 0.7075343132019043, |
|
"learning_rate": 0.0001289360469289838, |
|
"loss": 0.2174, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 4.048387096774194, |
|
"grad_norm": 0.7534266114234924, |
|
"learning_rate": 0.0001284426824140101, |
|
"loss": 0.2102, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 4.064516129032258, |
|
"grad_norm": 0.80478435754776, |
|
"learning_rate": 0.00012794856348516095, |
|
"loss": 0.1996, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 4.080645161290323, |
|
"grad_norm": 0.8782503604888916, |
|
"learning_rate": 0.00012745370324844974, |
|
"loss": 0.2231, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 4.096774193548387, |
|
"grad_norm": 0.9023078083992004, |
|
"learning_rate": 0.00012695811482955227, |
|
"loss": 0.2127, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 4.112903225806452, |
|
"grad_norm": 1.4141674041748047, |
|
"learning_rate": 0.0001264618113734587, |
|
"loss": 0.2308, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 4.129032258064516, |
|
"grad_norm": 0.7676845192909241, |
|
"learning_rate": 0.00012596480604412484, |
|
"loss": 0.2044, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 4.145161290322581, |
|
"grad_norm": 0.7796346545219421, |
|
"learning_rate": 0.00012546711202412287, |
|
"loss": 0.2103, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 4.161290322580645, |
|
"grad_norm": 0.7459124326705933, |
|
"learning_rate": 0.000124968742514292, |
|
"loss": 0.1979, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 4.17741935483871, |
|
"grad_norm": 0.7819743156433105, |
|
"learning_rate": 0.000124469710733388, |
|
"loss": 0.1989, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 4.193548387096774, |
|
"grad_norm": 0.8068183660507202, |
|
"learning_rate": 0.00012397002991773275, |
|
"loss": 0.211, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 4.209677419354839, |
|
"grad_norm": 0.76580810546875, |
|
"learning_rate": 0.00012346971332086317, |
|
"loss": 0.2161, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 4.225806451612903, |
|
"grad_norm": 0.7457690238952637, |
|
"learning_rate": 0.0001229687742131796, |
|
"loss": 0.2082, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 4.241935483870968, |
|
"grad_norm": 0.7540544867515564, |
|
"learning_rate": 0.0001224672258815938, |
|
"loss": 0.1897, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 4.258064516129032, |
|
"grad_norm": 0.769862174987793, |
|
"learning_rate": 0.00012196508162917677, |
|
"loss": 0.1996, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 4.258064516129032, |
|
"eval_loss": 1.8386796712875366, |
|
"eval_runtime": 360.7673, |
|
"eval_samples_per_second": 2.938, |
|
"eval_steps_per_second": 0.491, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 4.274193548387097, |
|
"grad_norm": 0.7516776919364929, |
|
"learning_rate": 0.00012146235477480559, |
|
"loss": 0.2048, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 4.290322580645161, |
|
"grad_norm": 0.8582746386528015, |
|
"learning_rate": 0.00012095905865281025, |
|
"loss": 0.2017, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 4.306451612903226, |
|
"grad_norm": 0.7552902102470398, |
|
"learning_rate": 0.0001204552066126201, |
|
"loss": 0.205, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.32258064516129, |
|
"grad_norm": 0.7861020565032959, |
|
"learning_rate": 0.00011995081201840956, |
|
"loss": 0.21, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 4.338709677419355, |
|
"grad_norm": 0.8263004422187805, |
|
"learning_rate": 0.00011944588824874388, |
|
"loss": 0.2006, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 4.354838709677419, |
|
"grad_norm": 3.678614616394043, |
|
"learning_rate": 0.00011894044869622403, |
|
"loss": 0.19, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 4.370967741935484, |
|
"grad_norm": 0.7940691709518433, |
|
"learning_rate": 0.0001184345067671317, |
|
"loss": 0.2141, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 4.387096774193548, |
|
"grad_norm": 0.7446163296699524, |
|
"learning_rate": 0.00011792807588107357, |
|
"loss": 0.192, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.403225806451613, |
|
"grad_norm": 0.7419606447219849, |
|
"learning_rate": 0.00011742116947062545, |
|
"loss": 0.2085, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 4.419354838709677, |
|
"grad_norm": 0.7716213464736938, |
|
"learning_rate": 0.00011691380098097597, |
|
"loss": 0.2126, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 4.435483870967742, |
|
"grad_norm": 0.9422786831855774, |
|
"learning_rate": 0.00011640598386956992, |
|
"loss": 0.2048, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 4.451612903225806, |
|
"grad_norm": 0.8216850161552429, |
|
"learning_rate": 0.0001158977316057513, |
|
"loss": 0.1977, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 4.467741935483871, |
|
"grad_norm": 0.7419984936714172, |
|
"learning_rate": 0.0001153890576704062, |
|
"loss": 0.2047, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 4.483870967741936, |
|
"grad_norm": 0.7556256651878357, |
|
"learning_rate": 0.00011487997555560503, |
|
"loss": 0.2094, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 4.5, |
|
"grad_norm": 0.8195300102233887, |
|
"learning_rate": 0.0001143704987642448, |
|
"loss": 0.199, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 4.516129032258064, |
|
"grad_norm": 0.7117184996604919, |
|
"learning_rate": 0.00011386064080969094, |
|
"loss": 0.1932, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 4.516129032258064, |
|
"eval_loss": 1.8438739776611328, |
|
"eval_runtime": 361.0378, |
|
"eval_samples_per_second": 2.936, |
|
"eval_steps_per_second": 0.49, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 4.532258064516129, |
|
"grad_norm": 0.7270099520683289, |
|
"learning_rate": 0.00011335041521541885, |
|
"loss": 0.1916, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 4.548387096774194, |
|
"grad_norm": 0.7429494261741638, |
|
"learning_rate": 0.00011283983551465511, |
|
"loss": 0.1984, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 4.564516129032258, |
|
"grad_norm": 0.7204171419143677, |
|
"learning_rate": 0.00011232891525001876, |
|
"loss": 0.1814, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 4.580645161290323, |
|
"grad_norm": 0.7861455678939819, |
|
"learning_rate": 0.0001118176679731619, |
|
"loss": 0.2097, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 4.596774193548387, |
|
"grad_norm": 0.8191571831703186, |
|
"learning_rate": 0.0001113061072444103, |
|
"loss": 0.221, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 4.612903225806452, |
|
"grad_norm": 1.3524975776672363, |
|
"learning_rate": 0.00011079424663240372, |
|
"loss": 0.2066, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 4.629032258064516, |
|
"grad_norm": 0.7753690481185913, |
|
"learning_rate": 0.00011028209971373605, |
|
"loss": 0.2112, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 4.645161290322581, |
|
"grad_norm": 0.7331258654594421, |
|
"learning_rate": 0.00010976968007259519, |
|
"loss": 0.1993, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 4.661290322580645, |
|
"grad_norm": 0.7499488592147827, |
|
"learning_rate": 0.00010925700130040266, |
|
"loss": 0.1937, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 4.67741935483871, |
|
"grad_norm": 0.813107967376709, |
|
"learning_rate": 0.00010874407699545328, |
|
"loss": 0.2157, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 4.693548387096774, |
|
"grad_norm": 0.8471450805664062, |
|
"learning_rate": 0.0001082309207625543, |
|
"loss": 0.215, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 4.709677419354839, |
|
"grad_norm": 0.7541672587394714, |
|
"learning_rate": 0.00010771754621266466, |
|
"loss": 0.1968, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.725806451612903, |
|
"grad_norm": 0.8032663464546204, |
|
"learning_rate": 0.00010720396696253393, |
|
"loss": 0.2161, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 4.741935483870968, |
|
"grad_norm": 0.7688977718353271, |
|
"learning_rate": 0.00010669019663434117, |
|
"loss": 0.202, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 4.758064516129032, |
|
"grad_norm": 0.8105716109275818, |
|
"learning_rate": 0.00010617624885533356, |
|
"loss": 0.228, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 4.774193548387097, |
|
"grad_norm": 0.7785431146621704, |
|
"learning_rate": 0.00010566213725746506, |
|
"loss": 0.2185, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 4.774193548387097, |
|
"eval_loss": 1.8509935140609741, |
|
"eval_runtime": 361.2404, |
|
"eval_samples_per_second": 2.934, |
|
"eval_steps_per_second": 0.49, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 4.790322580645161, |
|
"grad_norm": 0.865378737449646, |
|
"learning_rate": 0.00010514787547703466, |
|
"loss": 0.2095, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 4.806451612903226, |
|
"grad_norm": 1.0173976421356201, |
|
"learning_rate": 0.00010463347715432488, |
|
"loss": 0.2202, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 4.82258064516129, |
|
"grad_norm": 0.7536367774009705, |
|
"learning_rate": 0.00010411895593323981, |
|
"loss": 0.2072, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 4.838709677419355, |
|
"grad_norm": 0.7996053695678711, |
|
"learning_rate": 0.00010360432546094341, |
|
"loss": 0.209, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 4.854838709677419, |
|
"grad_norm": 0.8005889654159546, |
|
"learning_rate": 0.00010308959938749729, |
|
"loss": 0.217, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 4.870967741935484, |
|
"grad_norm": 0.7493289709091187, |
|
"learning_rate": 0.00010257479136549889, |
|
"loss": 0.2037, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 4.887096774193548, |
|
"grad_norm": 0.8069470524787903, |
|
"learning_rate": 0.00010205991504971912, |
|
"loss": 0.2189, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 4.903225806451613, |
|
"grad_norm": 0.7922778129577637, |
|
"learning_rate": 0.00010154498409674051, |
|
"loss": 0.2241, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 4.919354838709677, |
|
"grad_norm": 0.7668143510818481, |
|
"learning_rate": 0.00010103001216459469, |
|
"loss": 0.2202, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 4.935483870967742, |
|
"grad_norm": 0.7700707316398621, |
|
"learning_rate": 0.00010051501291240008, |
|
"loss": 0.2149, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 4.951612903225806, |
|
"grad_norm": 0.7833395004272461, |
|
"learning_rate": 0.0001, |
|
"loss": 0.2176, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 4.967741935483871, |
|
"grad_norm": 0.7745612263679504, |
|
"learning_rate": 9.948498708759993e-05, |
|
"loss": 0.2121, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 4.983870967741936, |
|
"grad_norm": 0.7606418132781982, |
|
"learning_rate": 9.896998783540536e-05, |
|
"loss": 0.1974, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"grad_norm": 0.8445224165916443, |
|
"learning_rate": 9.845501590325948e-05, |
|
"loss": 0.2194, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 5.016129032258064, |
|
"grad_norm": 0.7644294500350952, |
|
"learning_rate": 9.794008495028087e-05, |
|
"loss": 0.2037, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 5.032258064516129, |
|
"grad_norm": 0.7571164965629578, |
|
"learning_rate": 9.742520863450115e-05, |
|
"loss": 0.2221, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 5.032258064516129, |
|
"eval_loss": 1.851494550704956, |
|
"eval_runtime": 366.7021, |
|
"eval_samples_per_second": 2.891, |
|
"eval_steps_per_second": 0.483, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 5.016129032258065, |
|
"grad_norm": 0.591090977191925, |
|
"learning_rate": 9.691040061250273e-05, |
|
"loss": 0.1024, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 5.032258064516129, |
|
"grad_norm": 0.5239948630332947, |
|
"learning_rate": 9.639567453905661e-05, |
|
"loss": 0.0993, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 5.048387096774194, |
|
"grad_norm": 0.4719286561012268, |
|
"learning_rate": 9.58810440667602e-05, |
|
"loss": 0.0905, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 5.064516129032258, |
|
"grad_norm": 0.5142775177955627, |
|
"learning_rate": 9.536652284567513e-05, |
|
"loss": 0.0916, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 5.080645161290323, |
|
"grad_norm": 0.6494867205619812, |
|
"learning_rate": 9.485212452296535e-05, |
|
"loss": 0.1001, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.096774193548387, |
|
"grad_norm": 0.6289418935775757, |
|
"learning_rate": 9.433786274253495e-05, |
|
"loss": 0.0945, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 5.112903225806452, |
|
"grad_norm": 0.6664146780967712, |
|
"learning_rate": 9.382375114466644e-05, |
|
"loss": 0.0919, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 5.129032258064516, |
|
"grad_norm": 0.6793842911720276, |
|
"learning_rate": 9.330980336565887e-05, |
|
"loss": 0.1023, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 5.145161290322581, |
|
"grad_norm": 0.6552854776382446, |
|
"learning_rate": 9.279603303746608e-05, |
|
"loss": 0.0956, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 5.161290322580645, |
|
"grad_norm": 0.5973914861679077, |
|
"learning_rate": 9.228245378733537e-05, |
|
"loss": 0.0978, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 5.17741935483871, |
|
"grad_norm": 0.5700520277023315, |
|
"learning_rate": 9.176907923744571e-05, |
|
"loss": 0.0909, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 5.193548387096774, |
|
"grad_norm": 0.5783142447471619, |
|
"learning_rate": 9.125592300454676e-05, |
|
"loss": 0.094, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 5.209677419354839, |
|
"grad_norm": 0.5391398072242737, |
|
"learning_rate": 9.074299869959738e-05, |
|
"loss": 0.0885, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 5.225806451612903, |
|
"grad_norm": 0.5473701357841492, |
|
"learning_rate": 9.023031992740488e-05, |
|
"loss": 0.0917, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 5.241935483870968, |
|
"grad_norm": 0.5759130716323853, |
|
"learning_rate": 8.971790028626395e-05, |
|
"loss": 0.1002, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 5.258064516129032, |
|
"grad_norm": 0.5708579421043396, |
|
"learning_rate": 8.920575336759629e-05, |
|
"loss": 0.0968, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 5.258064516129032, |
|
"eval_loss": 2.0317347049713135, |
|
"eval_runtime": 365.1587, |
|
"eval_samples_per_second": 2.903, |
|
"eval_steps_per_second": 0.485, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 5.274193548387097, |
|
"grad_norm": 0.6354628205299377, |
|
"learning_rate": 8.869389275558972e-05, |
|
"loss": 0.1, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 5.290322580645161, |
|
"grad_norm": 0.6346054077148438, |
|
"learning_rate": 8.818233202683814e-05, |
|
"loss": 0.097, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 5.306451612903226, |
|
"grad_norm": 0.5868246555328369, |
|
"learning_rate": 8.767108474998128e-05, |
|
"loss": 0.098, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 5.32258064516129, |
|
"grad_norm": 0.5682688355445862, |
|
"learning_rate": 8.71601644853449e-05, |
|
"loss": 0.0887, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 5.338709677419355, |
|
"grad_norm": 0.643573522567749, |
|
"learning_rate": 8.664958478458117e-05, |
|
"loss": 0.0969, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 5.354838709677419, |
|
"grad_norm": 1.490026593208313, |
|
"learning_rate": 8.613935919030907e-05, |
|
"loss": 0.0983, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 5.370967741935484, |
|
"grad_norm": 0.629753828048706, |
|
"learning_rate": 8.562950123575523e-05, |
|
"loss": 0.0991, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 5.387096774193548, |
|
"grad_norm": 0.6013917326927185, |
|
"learning_rate": 8.512002444439502e-05, |
|
"loss": 0.1026, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 5.403225806451613, |
|
"grad_norm": 0.5838581323623657, |
|
"learning_rate": 8.461094232959381e-05, |
|
"loss": 0.089, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 5.419354838709677, |
|
"grad_norm": 0.5835732221603394, |
|
"learning_rate": 8.410226839424871e-05, |
|
"loss": 0.0939, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 5.435483870967742, |
|
"grad_norm": 0.616823136806488, |
|
"learning_rate": 8.359401613043012e-05, |
|
"loss": 0.0936, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 5.451612903225806, |
|
"grad_norm": 0.6824636459350586, |
|
"learning_rate": 8.308619901902406e-05, |
|
"loss": 0.1023, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 5.467741935483871, |
|
"grad_norm": 0.5994443297386169, |
|
"learning_rate": 8.257883052937456e-05, |
|
"loss": 0.0975, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 5.483870967741936, |
|
"grad_norm": 0.7672178149223328, |
|
"learning_rate": 8.207192411892646e-05, |
|
"loss": 0.0983, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.5, |
|
"grad_norm": 1.1863083839416504, |
|
"learning_rate": 8.156549323286831e-05, |
|
"loss": 0.0951, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 5.516129032258064, |
|
"grad_norm": 0.5679525136947632, |
|
"learning_rate": 8.1059551303776e-05, |
|
"loss": 0.0937, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 5.516129032258064, |
|
"eval_loss": 2.013756036758423, |
|
"eval_runtime": 365.2484, |
|
"eval_samples_per_second": 2.902, |
|
"eval_steps_per_second": 0.485, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 5.532258064516129, |
|
"grad_norm": 0.5959299802780151, |
|
"learning_rate": 8.055411175125616e-05, |
|
"loss": 0.086, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 5.548387096774194, |
|
"grad_norm": 0.5877125859260559, |
|
"learning_rate": 8.004918798159045e-05, |
|
"loss": 0.0911, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 5.564516129032258, |
|
"grad_norm": 0.5466898083686829, |
|
"learning_rate": 7.954479338737995e-05, |
|
"loss": 0.0845, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 5.580645161290323, |
|
"grad_norm": 0.5609007477760315, |
|
"learning_rate": 7.904094134718976e-05, |
|
"loss": 0.0883, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 5.596774193548387, |
|
"grad_norm": 0.6323772668838501, |
|
"learning_rate": 7.853764522519445e-05, |
|
"loss": 0.0944, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 5.612903225806452, |
|
"grad_norm": 0.6980358958244324, |
|
"learning_rate": 7.803491837082324e-05, |
|
"loss": 0.1007, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 5.629032258064516, |
|
"grad_norm": 0.6050729751586914, |
|
"learning_rate": 7.753277411840622e-05, |
|
"loss": 0.0927, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 5.645161290322581, |
|
"grad_norm": 0.5835074782371521, |
|
"learning_rate": 7.703122578682046e-05, |
|
"loss": 0.0876, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.661290322580645, |
|
"grad_norm": 0.5104186534881592, |
|
"learning_rate": 7.653028667913686e-05, |
|
"loss": 0.084, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 5.67741935483871, |
|
"grad_norm": 0.5861672163009644, |
|
"learning_rate": 7.602997008226726e-05, |
|
"loss": 0.0932, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 5.693548387096774, |
|
"grad_norm": 0.5980664491653442, |
|
"learning_rate": 7.553028926661201e-05, |
|
"loss": 0.0833, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 5.709677419354839, |
|
"grad_norm": 0.6397245526313782, |
|
"learning_rate": 7.5031257485708e-05, |
|
"loss": 0.088, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 5.725806451612903, |
|
"grad_norm": 0.6025028228759766, |
|
"learning_rate": 7.453288797587714e-05, |
|
"loss": 0.0949, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 5.741935483870968, |
|
"grad_norm": 0.6607884168624878, |
|
"learning_rate": 7.403519395587521e-05, |
|
"loss": 0.1029, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 5.758064516129032, |
|
"grad_norm": 0.6854023933410645, |
|
"learning_rate": 7.353818862654129e-05, |
|
"loss": 0.0984, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 5.774193548387097, |
|
"grad_norm": 0.5980477929115295, |
|
"learning_rate": 7.304188517044774e-05, |
|
"loss": 0.0973, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 5.774193548387097, |
|
"eval_loss": 2.027407646179199, |
|
"eval_runtime": 229.0955, |
|
"eval_samples_per_second": 4.627, |
|
"eval_steps_per_second": 0.773, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 5.790322580645161, |
|
"grad_norm": 0.5738844871520996, |
|
"learning_rate": 7.254629675155027e-05, |
|
"loss": 0.0872, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 5.806451612903226, |
|
"grad_norm": 0.6093122363090515, |
|
"learning_rate": 7.205143651483906e-05, |
|
"loss": 0.0939, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 5.82258064516129, |
|
"grad_norm": 0.5988517999649048, |
|
"learning_rate": 7.155731758598992e-05, |
|
"loss": 0.0922, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 5.838709677419355, |
|
"grad_norm": 0.6104899644851685, |
|
"learning_rate": 7.106395307101621e-05, |
|
"loss": 0.0829, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 5.854838709677419, |
|
"grad_norm": 0.8592974543571472, |
|
"learning_rate": 7.057135605592121e-05, |
|
"loss": 0.1204, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 5.870967741935484, |
|
"grad_norm": 0.5683754682540894, |
|
"learning_rate": 7.007953960635109e-05, |
|
"loss": 0.0865, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 5.887096774193548, |
|
"grad_norm": 0.5832166075706482, |
|
"learning_rate": 6.958851676724823e-05, |
|
"loss": 0.0911, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 5.903225806451613, |
|
"grad_norm": 0.534551739692688, |
|
"learning_rate": 6.909830056250527e-05, |
|
"loss": 0.0861, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 5.919354838709677, |
|
"grad_norm": 0.546940803527832, |
|
"learning_rate": 6.860890399461974e-05, |
|
"loss": 0.0893, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 5.935483870967742, |
|
"grad_norm": 0.5596715211868286, |
|
"learning_rate": 6.812034004434903e-05, |
|
"loss": 0.0913, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 5.951612903225806, |
|
"grad_norm": 0.6050382852554321, |
|
"learning_rate": 6.76326216703663e-05, |
|
"loss": 0.0853, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 5.967741935483871, |
|
"grad_norm": 0.563949465751648, |
|
"learning_rate": 6.714576180891654e-05, |
|
"loss": 0.0814, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 5.983870967741936, |
|
"grad_norm": 0.6186781525611877, |
|
"learning_rate": 6.665977337347354e-05, |
|
"loss": 0.1003, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"grad_norm": 0.583235502243042, |
|
"learning_rate": 6.617466925439746e-05, |
|
"loss": 0.0911, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 6.016129032258064, |
|
"grad_norm": 0.6332904696464539, |
|
"learning_rate": 6.569046231859281e-05, |
|
"loss": 0.0923, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 6.032258064516129, |
|
"grad_norm": 0.6747143864631653, |
|
"learning_rate": 6.520716540916709e-05, |
|
"loss": 0.083, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 6.032258064516129, |
|
"eval_loss": 2.0257039070129395, |
|
"eval_runtime": 219.95, |
|
"eval_samples_per_second": 4.819, |
|
"eval_steps_per_second": 0.805, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 6.016129032258065, |
|
"grad_norm": 0.33682334423065186, |
|
"learning_rate": 6.472479134509052e-05, |
|
"loss": 0.0476, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 6.032258064516129, |
|
"grad_norm": 0.33974146842956543, |
|
"learning_rate": 6.424335292085553e-05, |
|
"loss": 0.0462, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 6.048387096774194, |
|
"grad_norm": 0.335627019405365, |
|
"learning_rate": 6.376286290613776e-05, |
|
"loss": 0.0437, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 6.064516129032258, |
|
"grad_norm": 0.3977552652359009, |
|
"learning_rate": 6.32833340454571e-05, |
|
"loss": 0.0446, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 6.080645161290323, |
|
"grad_norm": 0.3793430030345917, |
|
"learning_rate": 6.280477905783988e-05, |
|
"loss": 0.0437, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 6.096774193548387, |
|
"grad_norm": 0.38144728541374207, |
|
"learning_rate": 6.232721063648148e-05, |
|
"loss": 0.0388, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 6.112903225806452, |
|
"grad_norm": 0.3977310359477997, |
|
"learning_rate": 6.185064144840948e-05, |
|
"loss": 0.0412, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 6.129032258064516, |
|
"grad_norm": 0.4477882385253906, |
|
"learning_rate": 6.137508413414784e-05, |
|
"loss": 0.0422, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 6.145161290322581, |
|
"grad_norm": 0.4987168312072754, |
|
"learning_rate": 6.0900551307381484e-05, |
|
"loss": 0.0391, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 6.161290322580645, |
|
"grad_norm": 0.4499349892139435, |
|
"learning_rate": 6.0427055554621913e-05, |
|
"loss": 0.0382, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 6.17741935483871, |
|
"grad_norm": 0.39157646894454956, |
|
"learning_rate": 5.9954609434873344e-05, |
|
"loss": 0.0368, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 6.193548387096774, |
|
"grad_norm": 0.38924601674079895, |
|
"learning_rate": 5.948322547929939e-05, |
|
"loss": 0.0406, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 6.209677419354839, |
|
"grad_norm": 0.4060990810394287, |
|
"learning_rate": 5.901291619089081e-05, |
|
"loss": 0.0453, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 6.225806451612903, |
|
"grad_norm": 0.5091999173164368, |
|
"learning_rate": 5.854369404413398e-05, |
|
"loss": 0.0415, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 6.241935483870968, |
|
"grad_norm": 0.38994643092155457, |
|
"learning_rate": 5.807557148467984e-05, |
|
"loss": 0.0429, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 6.258064516129032, |
|
"grad_norm": 0.4015688896179199, |
|
"learning_rate": 5.7608560929013946e-05, |
|
"loss": 0.0385, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.258064516129032, |
|
"eval_loss": 2.1730642318725586, |
|
"eval_runtime": 220.4701, |
|
"eval_samples_per_second": 4.808, |
|
"eval_steps_per_second": 0.803, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.274193548387097, |
|
"grad_norm": 0.41079944372177124, |
|
"learning_rate": 5.7142674764127e-05, |
|
"loss": 0.0426, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 6.290322580645161, |
|
"grad_norm": 0.382231205701828, |
|
"learning_rate": 5.667792534718639e-05, |
|
"loss": 0.0387, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 6.306451612903226, |
|
"grad_norm": 0.3789753019809723, |
|
"learning_rate": 5.6214325005208355e-05, |
|
"loss": 0.0411, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 6.32258064516129, |
|
"grad_norm": 0.36216092109680176, |
|
"learning_rate": 5.5751886034731115e-05, |
|
"loss": 0.0411, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 6.338709677419355, |
|
"grad_norm": 0.44374072551727295, |
|
"learning_rate": 5.5290620701488594e-05, |
|
"loss": 0.0408, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 6.354838709677419, |
|
"grad_norm": 0.46136826276779175, |
|
"learning_rate": 5.483054124008528e-05, |
|
"loss": 0.045, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 6.370967741935484, |
|
"grad_norm": 0.3711608052253723, |
|
"learning_rate": 5.437165985367145e-05, |
|
"loss": 0.0369, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 6.387096774193548, |
|
"grad_norm": 0.41239285469055176, |
|
"learning_rate": 5.391398871361972e-05, |
|
"loss": 0.0379, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 6.403225806451613, |
|
"grad_norm": 0.5803573727607727, |
|
"learning_rate": 5.3457539959202086e-05, |
|
"loss": 0.0515, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 6.419354838709677, |
|
"grad_norm": 0.4359411597251892, |
|
"learning_rate": 5.300232569726804e-05, |
|
"loss": 0.0427, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 6.435483870967742, |
|
"grad_norm": 0.4077586233615875, |
|
"learning_rate": 5.2548358001923204e-05, |
|
"loss": 0.0399, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 6.451612903225806, |
|
"grad_norm": 0.3488485813140869, |
|
"learning_rate": 5.2095648914209525e-05, |
|
"loss": 0.0396, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 6.467741935483871, |
|
"grad_norm": 0.39937523007392883, |
|
"learning_rate": 5.1644210441785467e-05, |
|
"loss": 0.039, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 6.483870967741936, |
|
"grad_norm": 0.43873974680900574, |
|
"learning_rate": 5.119405455860772e-05, |
|
"loss": 0.0366, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 6.5, |
|
"grad_norm": 0.356111079454422, |
|
"learning_rate": 5.074519320461357e-05, |
|
"loss": 0.0372, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 6.516129032258064, |
|
"grad_norm": 0.42101165652275085, |
|
"learning_rate": 5.029763828540419e-05, |
|
"loss": 0.0411, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 6.516129032258064, |
|
"eval_loss": 2.2114336490631104, |
|
"eval_runtime": 358.1731, |
|
"eval_samples_per_second": 2.959, |
|
"eval_steps_per_second": 0.494, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 6.532258064516129, |
|
"grad_norm": 0.3797813653945923, |
|
"learning_rate": 4.9851401671928934e-05, |
|
"loss": 0.0431, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 6.548387096774194, |
|
"grad_norm": 0.48072004318237305, |
|
"learning_rate": 4.940649520017035e-05, |
|
"loss": 0.0359, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 6.564516129032258, |
|
"grad_norm": 0.38997748494148254, |
|
"learning_rate": 4.896293067083032e-05, |
|
"loss": 0.0386, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 6.580645161290323, |
|
"grad_norm": 0.35900595784187317, |
|
"learning_rate": 4.852071984901696e-05, |
|
"loss": 0.0412, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 6.596774193548387, |
|
"grad_norm": 0.3810931444168091, |
|
"learning_rate": 4.807987446393275e-05, |
|
"loss": 0.0356, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 6.612903225806452, |
|
"grad_norm": 0.6008711457252502, |
|
"learning_rate": 4.7640406208563224e-05, |
|
"loss": 0.0433, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 6.629032258064516, |
|
"grad_norm": 0.5079227685928345, |
|
"learning_rate": 4.720232673936706e-05, |
|
"loss": 0.0369, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 6.645161290322581, |
|
"grad_norm": 0.3621387183666229, |
|
"learning_rate": 4.676564767596663e-05, |
|
"loss": 0.0373, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 6.661290322580645, |
|
"grad_norm": 0.418442964553833, |
|
"learning_rate": 4.633038060083996e-05, |
|
"loss": 0.0445, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 6.67741935483871, |
|
"grad_norm": 0.37667208909988403, |
|
"learning_rate": 4.5896537059013536e-05, |
|
"loss": 0.0368, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 6.693548387096774, |
|
"grad_norm": 0.3964817523956299, |
|
"learning_rate": 4.546412855775595e-05, |
|
"loss": 0.0395, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 6.709677419354839, |
|
"grad_norm": 0.3593597710132599, |
|
"learning_rate": 4.503316656627294e-05, |
|
"loss": 0.038, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 6.725806451612903, |
|
"grad_norm": 0.6818132400512695, |
|
"learning_rate": 4.460366251540281e-05, |
|
"loss": 0.0504, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 6.741935483870968, |
|
"grad_norm": 0.41429397463798523, |
|
"learning_rate": 4.417562779731355e-05, |
|
"loss": 0.0405, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 6.758064516129032, |
|
"grad_norm": 0.5057131052017212, |
|
"learning_rate": 4.374907376520054e-05, |
|
"loss": 0.0395, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 6.774193548387097, |
|
"grad_norm": 0.5655283331871033, |
|
"learning_rate": 4.3324011732985433e-05, |
|
"loss": 0.0446, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 6.774193548387097, |
|
"eval_loss": 2.2080209255218506, |
|
"eval_runtime": 357.0367, |
|
"eval_samples_per_second": 2.969, |
|
"eval_steps_per_second": 0.496, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 6.790322580645161, |
|
"grad_norm": 0.5865839719772339, |
|
"learning_rate": 4.2900452975016037e-05, |
|
"loss": 0.0479, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 6.806451612903226, |
|
"grad_norm": 0.43115484714508057, |
|
"learning_rate": 4.247840872576739e-05, |
|
"loss": 0.0429, |
|
"step": 434 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 620, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 62, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.1927216431929754e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|