WikiLinking-bi-gliner-base / trainer_state.json
BioMike's picture
Upload folder using huggingface_hub
9b83565 verified
raw
history blame
202 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.04848151844115758,
"eval_steps": 500,
"global_step": 12000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 4.040126536763131e-05,
"grad_norm": 13786306.0,
"learning_rate": 2e-08,
"loss": 1195071.7,
"step": 10
},
{
"epoch": 8.080253073526263e-05,
"grad_norm": 3502914.5,
"learning_rate": 4e-08,
"loss": 1488604.5,
"step": 20
},
{
"epoch": 0.00012120379610289395,
"grad_norm": 5249533.5,
"learning_rate": 6e-08,
"loss": 1363003.1,
"step": 30
},
{
"epoch": 0.00016160506147052525,
"grad_norm": 17823226.0,
"learning_rate": 8e-08,
"loss": 1539778.9,
"step": 40
},
{
"epoch": 0.00020200632683815657,
"grad_norm": 14010826.0,
"learning_rate": 1.0000000000000001e-07,
"loss": 1378274.9,
"step": 50
},
{
"epoch": 0.0002424075922057879,
"grad_norm": 23003992.0,
"learning_rate": 1.2e-07,
"loss": 1250654.1,
"step": 60
},
{
"epoch": 0.0002828088575734192,
"grad_norm": 7809646.0,
"learning_rate": 1.4e-07,
"loss": 1819393.6,
"step": 70
},
{
"epoch": 0.0003232101229410505,
"grad_norm": 43233020.0,
"learning_rate": 1.6e-07,
"loss": 1770900.4,
"step": 80
},
{
"epoch": 0.0003636113883086818,
"grad_norm": 6094604.5,
"learning_rate": 1.8e-07,
"loss": 1529363.0,
"step": 90
},
{
"epoch": 0.00040401265367631315,
"grad_norm": 8696879.0,
"learning_rate": 2.0000000000000002e-07,
"loss": 1028300.1,
"step": 100
},
{
"epoch": 0.00044441391904394446,
"grad_norm": 7774881.5,
"learning_rate": 2.2e-07,
"loss": 1786590.6,
"step": 110
},
{
"epoch": 0.0004848151844115758,
"grad_norm": 2879459.0,
"learning_rate": 2.4e-07,
"loss": 1377895.4,
"step": 120
},
{
"epoch": 0.000525216449779207,
"grad_norm": 19783298.0,
"learning_rate": 2.6e-07,
"loss": 1599422.0,
"step": 130
},
{
"epoch": 0.0005656177151468384,
"grad_norm": 2185791.0,
"learning_rate": 2.8e-07,
"loss": 1243788.4,
"step": 140
},
{
"epoch": 0.0006060189805144697,
"grad_norm": 9816560.0,
"learning_rate": 3.0000000000000004e-07,
"loss": 1245093.3,
"step": 150
},
{
"epoch": 0.000646420245882101,
"grad_norm": 9404809.0,
"learning_rate": 3.2e-07,
"loss": 1288713.5,
"step": 160
},
{
"epoch": 0.0006868215112497323,
"grad_norm": 8446407.0,
"learning_rate": 3.4e-07,
"loss": 1147532.8,
"step": 170
},
{
"epoch": 0.0007272227766173637,
"grad_norm": 3988919.0,
"learning_rate": 3.6e-07,
"loss": 704286.9,
"step": 180
},
{
"epoch": 0.000767624041984995,
"grad_norm": 2574691.5,
"learning_rate": 3.8e-07,
"loss": 545635.6,
"step": 190
},
{
"epoch": 0.0008080253073526263,
"grad_norm": 21092594.0,
"learning_rate": 4.0000000000000003e-07,
"loss": 728793.05,
"step": 200
},
{
"epoch": 0.0008484265727202576,
"grad_norm": 2704460.5,
"learning_rate": 4.2e-07,
"loss": 467207.7,
"step": 210
},
{
"epoch": 0.0008888278380878889,
"grad_norm": 4876784.0,
"learning_rate": 4.4e-07,
"loss": 369614.55,
"step": 220
},
{
"epoch": 0.0009292291034555202,
"grad_norm": 10803276.0,
"learning_rate": 4.6e-07,
"loss": 428869.6,
"step": 230
},
{
"epoch": 0.0009696303688231516,
"grad_norm": 5475451.0,
"learning_rate": 4.8e-07,
"loss": 404623.45,
"step": 240
},
{
"epoch": 0.0010100316341907828,
"grad_norm": 3401077.25,
"learning_rate": 5.000000000000001e-07,
"loss": 606814.6,
"step": 250
},
{
"epoch": 0.001050432899558414,
"grad_norm": 1785725.0,
"learning_rate": 5.2e-07,
"loss": 230520.125,
"step": 260
},
{
"epoch": 0.0010908341649260454,
"grad_norm": 26237066.0,
"learning_rate": 5.4e-07,
"loss": 251476.375,
"step": 270
},
{
"epoch": 0.0011312354302936767,
"grad_norm": 3433710.25,
"learning_rate": 5.6e-07,
"loss": 378523.225,
"step": 280
},
{
"epoch": 0.001171636695661308,
"grad_norm": 3058920.5,
"learning_rate": 5.8e-07,
"loss": 118036.575,
"step": 290
},
{
"epoch": 0.0012120379610289394,
"grad_norm": 2775939.25,
"learning_rate": 6.000000000000001e-07,
"loss": 128199.575,
"step": 300
},
{
"epoch": 0.0012524392263965707,
"grad_norm": 2798658.75,
"learning_rate": 6.2e-07,
"loss": 145799.6125,
"step": 310
},
{
"epoch": 0.001292840491764202,
"grad_norm": 3050425.75,
"learning_rate": 6.4e-07,
"loss": 170385.65,
"step": 320
},
{
"epoch": 0.0013332417571318333,
"grad_norm": 483486.6875,
"learning_rate": 6.6e-07,
"loss": 79056.7,
"step": 330
},
{
"epoch": 0.0013736430224994647,
"grad_norm": 903465.9375,
"learning_rate": 6.8e-07,
"loss": 69717.8562,
"step": 340
},
{
"epoch": 0.001414044287867096,
"grad_norm": 166525.765625,
"learning_rate": 7.000000000000001e-07,
"loss": 29574.1,
"step": 350
},
{
"epoch": 0.0014544455532347273,
"grad_norm": 83730.7109375,
"learning_rate": 7.2e-07,
"loss": 9268.3289,
"step": 360
},
{
"epoch": 0.0014948468186023586,
"grad_norm": 13107.71484375,
"learning_rate": 7.400000000000001e-07,
"loss": 16814.9484,
"step": 370
},
{
"epoch": 0.00153524808396999,
"grad_norm": 199187.046875,
"learning_rate": 7.6e-07,
"loss": 4728.9383,
"step": 380
},
{
"epoch": 0.0015756493493376213,
"grad_norm": 209663.171875,
"learning_rate": 7.8e-07,
"loss": 3811.3047,
"step": 390
},
{
"epoch": 0.0016160506147052526,
"grad_norm": 261596.96875,
"learning_rate": 8.000000000000001e-07,
"loss": 1873.8627,
"step": 400
},
{
"epoch": 0.001656451880072884,
"grad_norm": 9995.125,
"learning_rate": 8.200000000000001e-07,
"loss": 1428.6535,
"step": 410
},
{
"epoch": 0.0016968531454405152,
"grad_norm": 2573.125732421875,
"learning_rate": 8.4e-07,
"loss": 533.159,
"step": 420
},
{
"epoch": 0.0017372544108081465,
"grad_norm": 547.4088134765625,
"learning_rate": 8.6e-07,
"loss": 702.8197,
"step": 430
},
{
"epoch": 0.0017776556761757779,
"grad_norm": 734.5255126953125,
"learning_rate": 8.8e-07,
"loss": 485.3759,
"step": 440
},
{
"epoch": 0.0018180569415434092,
"grad_norm": 5114.30615234375,
"learning_rate": 9e-07,
"loss": 545.9343,
"step": 450
},
{
"epoch": 0.0018584582069110405,
"grad_norm": 1266.8800048828125,
"learning_rate": 9.2e-07,
"loss": 482.805,
"step": 460
},
{
"epoch": 0.0018988594722786718,
"grad_norm": 5457.6689453125,
"learning_rate": 9.400000000000001e-07,
"loss": 550.8278,
"step": 470
},
{
"epoch": 0.0019392607376463031,
"grad_norm": 1092.0926513671875,
"learning_rate": 9.6e-07,
"loss": 572.7061,
"step": 480
},
{
"epoch": 0.0019796620030139342,
"grad_norm": 1708.60546875,
"learning_rate": 9.8e-07,
"loss": 447.9137,
"step": 490
},
{
"epoch": 0.0020200632683815656,
"grad_norm": 833.2102661132812,
"learning_rate": 1.0000000000000002e-06,
"loss": 452.3768,
"step": 500
},
{
"epoch": 0.002060464533749197,
"grad_norm": 854.2012329101562,
"learning_rate": 1.0200000000000002e-06,
"loss": 453.9597,
"step": 510
},
{
"epoch": 0.002100865799116828,
"grad_norm": 883.29052734375,
"learning_rate": 1.04e-06,
"loss": 399.3994,
"step": 520
},
{
"epoch": 0.0021412670644844595,
"grad_norm": 876.2081909179688,
"learning_rate": 1.06e-06,
"loss": 508.956,
"step": 530
},
{
"epoch": 0.002181668329852091,
"grad_norm": 1216.003662109375,
"learning_rate": 1.08e-06,
"loss": 279.6047,
"step": 540
},
{
"epoch": 0.002222069595219722,
"grad_norm": 20951.376953125,
"learning_rate": 1.1e-06,
"loss": 523.4955,
"step": 550
},
{
"epoch": 0.0022624708605873535,
"grad_norm": 6249.7998046875,
"learning_rate": 1.12e-06,
"loss": 373.9713,
"step": 560
},
{
"epoch": 0.002302872125954985,
"grad_norm": 8211.34375,
"learning_rate": 1.14e-06,
"loss": 322.8237,
"step": 570
},
{
"epoch": 0.002343273391322616,
"grad_norm": 1029.943359375,
"learning_rate": 1.16e-06,
"loss": 443.8321,
"step": 580
},
{
"epoch": 0.0023836746566902474,
"grad_norm": 5415.279296875,
"learning_rate": 1.18e-06,
"loss": 580.9646,
"step": 590
},
{
"epoch": 0.0024240759220578788,
"grad_norm": 683.0673217773438,
"learning_rate": 1.2000000000000002e-06,
"loss": 322.4693,
"step": 600
},
{
"epoch": 0.00246447718742551,
"grad_norm": 3990.54931640625,
"learning_rate": 1.2200000000000002e-06,
"loss": 543.5786,
"step": 610
},
{
"epoch": 0.0025048784527931414,
"grad_norm": 771.86767578125,
"learning_rate": 1.24e-06,
"loss": 464.2772,
"step": 620
},
{
"epoch": 0.0025452797181607727,
"grad_norm": 830.4238891601562,
"learning_rate": 1.26e-06,
"loss": 502.5308,
"step": 630
},
{
"epoch": 0.002585680983528404,
"grad_norm": 4528.138671875,
"learning_rate": 1.28e-06,
"loss": 368.3498,
"step": 640
},
{
"epoch": 0.0026260822488960354,
"grad_norm": 701.6304321289062,
"learning_rate": 1.3e-06,
"loss": 433.9283,
"step": 650
},
{
"epoch": 0.0026664835142636667,
"grad_norm": 593.0372924804688,
"learning_rate": 1.32e-06,
"loss": 327.8368,
"step": 660
},
{
"epoch": 0.002706884779631298,
"grad_norm": 718.9718627929688,
"learning_rate": 1.34e-06,
"loss": 363.8515,
"step": 670
},
{
"epoch": 0.0027472860449989293,
"grad_norm": 1276.8353271484375,
"learning_rate": 1.36e-06,
"loss": 457.8593,
"step": 680
},
{
"epoch": 0.0027876873103665606,
"grad_norm": 9622.8271484375,
"learning_rate": 1.3800000000000001e-06,
"loss": 448.6425,
"step": 690
},
{
"epoch": 0.002828088575734192,
"grad_norm": 795.946044921875,
"learning_rate": 1.4000000000000001e-06,
"loss": 370.567,
"step": 700
},
{
"epoch": 0.0028684898411018233,
"grad_norm": 2605.2119140625,
"learning_rate": 1.4200000000000002e-06,
"loss": 399.8033,
"step": 710
},
{
"epoch": 0.0029088911064694546,
"grad_norm": 3644.47607421875,
"learning_rate": 1.44e-06,
"loss": 519.6482,
"step": 720
},
{
"epoch": 0.002949292371837086,
"grad_norm": 859.5913696289062,
"learning_rate": 1.46e-06,
"loss": 366.024,
"step": 730
},
{
"epoch": 0.0029896936372047172,
"grad_norm": 9709.5673828125,
"learning_rate": 1.4800000000000002e-06,
"loss": 387.851,
"step": 740
},
{
"epoch": 0.0030300949025723486,
"grad_norm": 1355.1983642578125,
"learning_rate": 1.5e-06,
"loss": 287.8213,
"step": 750
},
{
"epoch": 0.00307049616793998,
"grad_norm": 1910.957275390625,
"learning_rate": 1.52e-06,
"loss": 263.4618,
"step": 760
},
{
"epoch": 0.003110897433307611,
"grad_norm": 1562.11669921875,
"learning_rate": 1.54e-06,
"loss": 514.2329,
"step": 770
},
{
"epoch": 0.0031512986986752425,
"grad_norm": 765.830078125,
"learning_rate": 1.56e-06,
"loss": 332.8841,
"step": 780
},
{
"epoch": 0.003191699964042874,
"grad_norm": 13385.1689453125,
"learning_rate": 1.5800000000000003e-06,
"loss": 456.157,
"step": 790
},
{
"epoch": 0.003232101229410505,
"grad_norm": 969.9119262695312,
"learning_rate": 1.6000000000000001e-06,
"loss": 300.1112,
"step": 800
},
{
"epoch": 0.0032725024947781365,
"grad_norm": 1840.109619140625,
"learning_rate": 1.62e-06,
"loss": 384.0655,
"step": 810
},
{
"epoch": 0.003312903760145768,
"grad_norm": 1270.9483642578125,
"learning_rate": 1.6400000000000002e-06,
"loss": 355.3219,
"step": 820
},
{
"epoch": 0.003353305025513399,
"grad_norm": 668.4849243164062,
"learning_rate": 1.6600000000000002e-06,
"loss": 247.1219,
"step": 830
},
{
"epoch": 0.0033937062908810304,
"grad_norm": 1676.130615234375,
"learning_rate": 1.68e-06,
"loss": 312.9415,
"step": 840
},
{
"epoch": 0.0034341075562486618,
"grad_norm": 480.2029724121094,
"learning_rate": 1.7000000000000002e-06,
"loss": 388.1627,
"step": 850
},
{
"epoch": 0.003474508821616293,
"grad_norm": 579.8500366210938,
"learning_rate": 1.72e-06,
"loss": 423.7644,
"step": 860
},
{
"epoch": 0.0035149100869839244,
"grad_norm": 780.9789428710938,
"learning_rate": 1.7399999999999999e-06,
"loss": 335.4401,
"step": 870
},
{
"epoch": 0.0035553113523515557,
"grad_norm": 500.0970153808594,
"learning_rate": 1.76e-06,
"loss": 375.4284,
"step": 880
},
{
"epoch": 0.003595712617719187,
"grad_norm": 663.9608154296875,
"learning_rate": 1.7800000000000001e-06,
"loss": 432.7744,
"step": 890
},
{
"epoch": 0.0036361138830868184,
"grad_norm": 738.2296142578125,
"learning_rate": 1.8e-06,
"loss": 330.4268,
"step": 900
},
{
"epoch": 0.0036765151484544497,
"grad_norm": 888.3779907226562,
"learning_rate": 1.8200000000000002e-06,
"loss": 328.1051,
"step": 910
},
{
"epoch": 0.003716916413822081,
"grad_norm": 681.369873046875,
"learning_rate": 1.84e-06,
"loss": 455.4454,
"step": 920
},
{
"epoch": 0.0037573176791897123,
"grad_norm": 719.9152221679688,
"learning_rate": 1.86e-06,
"loss": 291.1318,
"step": 930
},
{
"epoch": 0.0037977189445573436,
"grad_norm": 765.2400512695312,
"learning_rate": 1.8800000000000002e-06,
"loss": 453.5919,
"step": 940
},
{
"epoch": 0.003838120209924975,
"grad_norm": 586.1494140625,
"learning_rate": 1.9e-06,
"loss": 338.6631,
"step": 950
},
{
"epoch": 0.0038785214752926063,
"grad_norm": 559.566650390625,
"learning_rate": 1.92e-06,
"loss": 309.9115,
"step": 960
},
{
"epoch": 0.003918922740660237,
"grad_norm": 2274.710205078125,
"learning_rate": 1.94e-06,
"loss": 366.6058,
"step": 970
},
{
"epoch": 0.0039593240060278685,
"grad_norm": 2413.857666015625,
"learning_rate": 1.96e-06,
"loss": 323.5407,
"step": 980
},
{
"epoch": 0.0039997252713955,
"grad_norm": 466.5931396484375,
"learning_rate": 1.98e-06,
"loss": 332.9158,
"step": 990
},
{
"epoch": 0.004040126536763131,
"grad_norm": 1368.92626953125,
"learning_rate": 2.0000000000000003e-06,
"loss": 309.8214,
"step": 1000
},
{
"epoch": 0.0040805278021307624,
"grad_norm": 578.7489624023438,
"learning_rate": 2.02e-06,
"loss": 293.1464,
"step": 1010
},
{
"epoch": 0.004120929067498394,
"grad_norm": 932.1323852539062,
"learning_rate": 2.0400000000000004e-06,
"loss": 382.6469,
"step": 1020
},
{
"epoch": 0.004161330332866025,
"grad_norm": 861.5737915039062,
"learning_rate": 2.06e-06,
"loss": 394.3037,
"step": 1030
},
{
"epoch": 0.004201731598233656,
"grad_norm": 540.3167114257812,
"learning_rate": 2.08e-06,
"loss": 356.1217,
"step": 1040
},
{
"epoch": 0.004242132863601288,
"grad_norm": 645.2182006835938,
"learning_rate": 2.1000000000000002e-06,
"loss": 375.9719,
"step": 1050
},
{
"epoch": 0.004282534128968919,
"grad_norm": 3072.72607421875,
"learning_rate": 2.12e-06,
"loss": 385.9836,
"step": 1060
},
{
"epoch": 0.00432293539433655,
"grad_norm": 1441.8131103515625,
"learning_rate": 2.14e-06,
"loss": 447.29,
"step": 1070
},
{
"epoch": 0.004363336659704182,
"grad_norm": 531.906494140625,
"learning_rate": 2.16e-06,
"loss": 307.2454,
"step": 1080
},
{
"epoch": 0.004403737925071813,
"grad_norm": 4712.17236328125,
"learning_rate": 2.1800000000000003e-06,
"loss": 599.3445,
"step": 1090
},
{
"epoch": 0.004444139190439444,
"grad_norm": 543.140380859375,
"learning_rate": 2.2e-06,
"loss": 236.8443,
"step": 1100
},
{
"epoch": 0.004484540455807076,
"grad_norm": 724.6696166992188,
"learning_rate": 2.2200000000000003e-06,
"loss": 402.6835,
"step": 1110
},
{
"epoch": 0.004524941721174707,
"grad_norm": 890.041015625,
"learning_rate": 2.24e-06,
"loss": 335.9333,
"step": 1120
},
{
"epoch": 0.004565342986542338,
"grad_norm": 517.4255981445312,
"learning_rate": 2.26e-06,
"loss": 377.6742,
"step": 1130
},
{
"epoch": 0.00460574425190997,
"grad_norm": 1773.51025390625,
"learning_rate": 2.28e-06,
"loss": 387.5232,
"step": 1140
},
{
"epoch": 0.004646145517277601,
"grad_norm": 1350.251220703125,
"learning_rate": 2.3e-06,
"loss": 430.9959,
"step": 1150
},
{
"epoch": 0.004686546782645232,
"grad_norm": 1267.9136962890625,
"learning_rate": 2.32e-06,
"loss": 417.8642,
"step": 1160
},
{
"epoch": 0.0047269480480128636,
"grad_norm": 1760.7581787109375,
"learning_rate": 2.34e-06,
"loss": 335.7049,
"step": 1170
},
{
"epoch": 0.004767349313380495,
"grad_norm": 471.7587890625,
"learning_rate": 2.36e-06,
"loss": 321.8847,
"step": 1180
},
{
"epoch": 0.004807750578748126,
"grad_norm": 1277.0931396484375,
"learning_rate": 2.38e-06,
"loss": 339.8787,
"step": 1190
},
{
"epoch": 0.0048481518441157575,
"grad_norm": 766.57568359375,
"learning_rate": 2.4000000000000003e-06,
"loss": 343.6708,
"step": 1200
},
{
"epoch": 0.004888553109483389,
"grad_norm": 733.90234375,
"learning_rate": 2.42e-06,
"loss": 284.1223,
"step": 1210
},
{
"epoch": 0.00492895437485102,
"grad_norm": 705.3512573242188,
"learning_rate": 2.4400000000000004e-06,
"loss": 290.2316,
"step": 1220
},
{
"epoch": 0.0049693556402186515,
"grad_norm": 1195.5079345703125,
"learning_rate": 2.46e-06,
"loss": 351.2376,
"step": 1230
},
{
"epoch": 0.005009756905586283,
"grad_norm": 1140.5634765625,
"learning_rate": 2.48e-06,
"loss": 364.3263,
"step": 1240
},
{
"epoch": 0.005050158170953914,
"grad_norm": 1419.2325439453125,
"learning_rate": 2.5e-06,
"loss": 420.2925,
"step": 1250
},
{
"epoch": 0.0050905594363215454,
"grad_norm": 945.1781616210938,
"learning_rate": 2.52e-06,
"loss": 377.5975,
"step": 1260
},
{
"epoch": 0.005130960701689177,
"grad_norm": 1156.38818359375,
"learning_rate": 2.54e-06,
"loss": 438.8611,
"step": 1270
},
{
"epoch": 0.005171361967056808,
"grad_norm": 803.0604858398438,
"learning_rate": 2.56e-06,
"loss": 275.1979,
"step": 1280
},
{
"epoch": 0.005211763232424439,
"grad_norm": 4027.070556640625,
"learning_rate": 2.5800000000000003e-06,
"loss": 417.5213,
"step": 1290
},
{
"epoch": 0.005252164497792071,
"grad_norm": 762.65771484375,
"learning_rate": 2.6e-06,
"loss": 338.176,
"step": 1300
},
{
"epoch": 0.005292565763159702,
"grad_norm": 514.9345092773438,
"learning_rate": 2.6200000000000003e-06,
"loss": 360.1529,
"step": 1310
},
{
"epoch": 0.005332967028527333,
"grad_norm": 0.0,
"learning_rate": 2.64e-06,
"loss": 437.6661,
"step": 1320
},
{
"epoch": 0.005373368293894965,
"grad_norm": 964.8443603515625,
"learning_rate": 2.66e-06,
"loss": 348.5103,
"step": 1330
},
{
"epoch": 0.005413769559262596,
"grad_norm": 1033.2138671875,
"learning_rate": 2.68e-06,
"loss": 351.3489,
"step": 1340
},
{
"epoch": 0.005454170824630227,
"grad_norm": 621.0077514648438,
"learning_rate": 2.7e-06,
"loss": 297.7263,
"step": 1350
},
{
"epoch": 0.005494572089997859,
"grad_norm": 670.5285034179688,
"learning_rate": 2.72e-06,
"loss": 266.751,
"step": 1360
},
{
"epoch": 0.00553497335536549,
"grad_norm": 1063.011962890625,
"learning_rate": 2.74e-06,
"loss": 388.5708,
"step": 1370
},
{
"epoch": 0.005575374620733121,
"grad_norm": 1352.453369140625,
"learning_rate": 2.7600000000000003e-06,
"loss": 413.9111,
"step": 1380
},
{
"epoch": 0.005615775886100753,
"grad_norm": 1039.40966796875,
"learning_rate": 2.78e-06,
"loss": 425.6994,
"step": 1390
},
{
"epoch": 0.005656177151468384,
"grad_norm": 852.0213012695312,
"learning_rate": 2.8000000000000003e-06,
"loss": 379.7008,
"step": 1400
},
{
"epoch": 0.005696578416836015,
"grad_norm": 560.5407104492188,
"learning_rate": 2.82e-06,
"loss": 373.5774,
"step": 1410
},
{
"epoch": 0.0057369796822036466,
"grad_norm": 2221.54248046875,
"learning_rate": 2.8400000000000003e-06,
"loss": 377.4932,
"step": 1420
},
{
"epoch": 0.005777380947571278,
"grad_norm": 986.341064453125,
"learning_rate": 2.86e-06,
"loss": 381.8052,
"step": 1430
},
{
"epoch": 0.005817782212938909,
"grad_norm": 2288.12060546875,
"learning_rate": 2.88e-06,
"loss": 396.2358,
"step": 1440
},
{
"epoch": 0.0058581834783065405,
"grad_norm": 506.0441589355469,
"learning_rate": 2.9e-06,
"loss": 306.8475,
"step": 1450
},
{
"epoch": 0.005898584743674172,
"grad_norm": 1451.433837890625,
"learning_rate": 2.92e-06,
"loss": 421.7085,
"step": 1460
},
{
"epoch": 0.005938986009041803,
"grad_norm": 410.0963439941406,
"learning_rate": 2.9400000000000002e-06,
"loss": 245.0907,
"step": 1470
},
{
"epoch": 0.0059793872744094345,
"grad_norm": 644.2957153320312,
"learning_rate": 2.9600000000000005e-06,
"loss": 256.7748,
"step": 1480
},
{
"epoch": 0.006019788539777066,
"grad_norm": 587.813232421875,
"learning_rate": 2.9800000000000003e-06,
"loss": 307.0451,
"step": 1490
},
{
"epoch": 0.006060189805144697,
"grad_norm": 661.851806640625,
"learning_rate": 3e-06,
"loss": 240.7283,
"step": 1500
},
{
"epoch": 0.0061005910705123284,
"grad_norm": 9922.1162109375,
"learning_rate": 3.0200000000000003e-06,
"loss": 314.2891,
"step": 1510
},
{
"epoch": 0.00614099233587996,
"grad_norm": 681.8471069335938,
"learning_rate": 3.04e-06,
"loss": 304.0082,
"step": 1520
},
{
"epoch": 0.006181393601247591,
"grad_norm": 1679.9188232421875,
"learning_rate": 3.06e-06,
"loss": 316.8349,
"step": 1530
},
{
"epoch": 0.006221794866615222,
"grad_norm": 0.0,
"learning_rate": 3.08e-06,
"loss": 273.9679,
"step": 1540
},
{
"epoch": 0.006262196131982854,
"grad_norm": 934.2025756835938,
"learning_rate": 3.1e-06,
"loss": 281.7329,
"step": 1550
},
{
"epoch": 0.006302597397350485,
"grad_norm": 1820.7791748046875,
"learning_rate": 3.12e-06,
"loss": 292.2456,
"step": 1560
},
{
"epoch": 0.006342998662718116,
"grad_norm": 1122.6767578125,
"learning_rate": 3.14e-06,
"loss": 327.0943,
"step": 1570
},
{
"epoch": 0.006383399928085748,
"grad_norm": 3407.740234375,
"learning_rate": 3.1600000000000007e-06,
"loss": 282.118,
"step": 1580
},
{
"epoch": 0.006423801193453379,
"grad_norm": 556.7740478515625,
"learning_rate": 3.1800000000000005e-06,
"loss": 332.0084,
"step": 1590
},
{
"epoch": 0.00646420245882101,
"grad_norm": 1229.8035888671875,
"learning_rate": 3.2000000000000003e-06,
"loss": 344.9013,
"step": 1600
},
{
"epoch": 0.006504603724188642,
"grad_norm": 949.2359008789062,
"learning_rate": 3.22e-06,
"loss": 371.0438,
"step": 1610
},
{
"epoch": 0.006545004989556273,
"grad_norm": 1869.7620849609375,
"learning_rate": 3.24e-06,
"loss": 388.4585,
"step": 1620
},
{
"epoch": 0.006585406254923904,
"grad_norm": 527.3277587890625,
"learning_rate": 3.2599999999999997e-06,
"loss": 372.7793,
"step": 1630
},
{
"epoch": 0.006625807520291536,
"grad_norm": 1427.3751220703125,
"learning_rate": 3.2800000000000004e-06,
"loss": 386.9533,
"step": 1640
},
{
"epoch": 0.006666208785659167,
"grad_norm": 1839.386474609375,
"learning_rate": 3.3e-06,
"loss": 414.8817,
"step": 1650
},
{
"epoch": 0.006706610051026798,
"grad_norm": 767.32373046875,
"learning_rate": 3.3200000000000004e-06,
"loss": 262.8998,
"step": 1660
},
{
"epoch": 0.0067470113163944296,
"grad_norm": 487.2662353515625,
"learning_rate": 3.34e-06,
"loss": 349.8497,
"step": 1670
},
{
"epoch": 0.006787412581762061,
"grad_norm": 1609.6903076171875,
"learning_rate": 3.36e-06,
"loss": 328.737,
"step": 1680
},
{
"epoch": 0.006827813847129692,
"grad_norm": 824.4249267578125,
"learning_rate": 3.38e-06,
"loss": 285.9903,
"step": 1690
},
{
"epoch": 0.0068682151124973235,
"grad_norm": 3939.00244140625,
"learning_rate": 3.4000000000000005e-06,
"loss": 360.8893,
"step": 1700
},
{
"epoch": 0.006908616377864955,
"grad_norm": 611.8134155273438,
"learning_rate": 3.4200000000000003e-06,
"loss": 316.7729,
"step": 1710
},
{
"epoch": 0.006949017643232586,
"grad_norm": 3673.492919921875,
"learning_rate": 3.44e-06,
"loss": 446.8052,
"step": 1720
},
{
"epoch": 0.0069894189086002175,
"grad_norm": 580.3406982421875,
"learning_rate": 3.46e-06,
"loss": 306.8866,
"step": 1730
},
{
"epoch": 0.007029820173967849,
"grad_norm": 797.6139526367188,
"learning_rate": 3.4799999999999997e-06,
"loss": 360.2713,
"step": 1740
},
{
"epoch": 0.00707022143933548,
"grad_norm": 1184.318115234375,
"learning_rate": 3.5000000000000004e-06,
"loss": 360.695,
"step": 1750
},
{
"epoch": 0.0071106227047031114,
"grad_norm": 659.1159057617188,
"learning_rate": 3.52e-06,
"loss": 393.7832,
"step": 1760
},
{
"epoch": 0.007151023970070743,
"grad_norm": 611.4116821289062,
"learning_rate": 3.5400000000000004e-06,
"loss": 302.0126,
"step": 1770
},
{
"epoch": 0.007191425235438374,
"grad_norm": 2148.29541015625,
"learning_rate": 3.5600000000000002e-06,
"loss": 411.9108,
"step": 1780
},
{
"epoch": 0.007231826500806005,
"grad_norm": 635.7216796875,
"learning_rate": 3.58e-06,
"loss": 236.3176,
"step": 1790
},
{
"epoch": 0.007272227766173637,
"grad_norm": 367.6564025878906,
"learning_rate": 3.6e-06,
"loss": 293.0558,
"step": 1800
},
{
"epoch": 0.007312629031541268,
"grad_norm": 712.876220703125,
"learning_rate": 3.6200000000000005e-06,
"loss": 349.5894,
"step": 1810
},
{
"epoch": 0.007353030296908899,
"grad_norm": 505.4992980957031,
"learning_rate": 3.6400000000000003e-06,
"loss": 300.7681,
"step": 1820
},
{
"epoch": 0.007393431562276531,
"grad_norm": 852.1588745117188,
"learning_rate": 3.66e-06,
"loss": 408.5499,
"step": 1830
},
{
"epoch": 0.007433832827644162,
"grad_norm": 1410.185791015625,
"learning_rate": 3.68e-06,
"loss": 253.6884,
"step": 1840
},
{
"epoch": 0.007474234093011793,
"grad_norm": 889.8245849609375,
"learning_rate": 3.7e-06,
"loss": 408.2236,
"step": 1850
},
{
"epoch": 0.007514635358379425,
"grad_norm": 906.4953002929688,
"learning_rate": 3.72e-06,
"loss": 397.8984,
"step": 1860
},
{
"epoch": 0.007555036623747056,
"grad_norm": 0.0,
"learning_rate": 3.7400000000000006e-06,
"loss": 238.5037,
"step": 1870
},
{
"epoch": 0.007595437889114687,
"grad_norm": 13512.4267578125,
"learning_rate": 3.7600000000000004e-06,
"loss": 348.5812,
"step": 1880
},
{
"epoch": 0.007635839154482319,
"grad_norm": 4588.23193359375,
"learning_rate": 3.7800000000000002e-06,
"loss": 364.4923,
"step": 1890
},
{
"epoch": 0.00767624041984995,
"grad_norm": 803.103759765625,
"learning_rate": 3.8e-06,
"loss": 372.9157,
"step": 1900
},
{
"epoch": 0.007716641685217581,
"grad_norm": 1621.907958984375,
"learning_rate": 3.82e-06,
"loss": 360.8791,
"step": 1910
},
{
"epoch": 0.0077570429505852126,
"grad_norm": 592.7820434570312,
"learning_rate": 3.84e-06,
"loss": 375.443,
"step": 1920
},
{
"epoch": 0.007797444215952844,
"grad_norm": 2514.99462890625,
"learning_rate": 3.86e-06,
"loss": 332.9206,
"step": 1930
},
{
"epoch": 0.007837845481320474,
"grad_norm": 1829.7589111328125,
"learning_rate": 3.88e-06,
"loss": 442.6601,
"step": 1940
},
{
"epoch": 0.007878246746688106,
"grad_norm": 897.420166015625,
"learning_rate": 3.9e-06,
"loss": 288.4365,
"step": 1950
},
{
"epoch": 0.007918648012055737,
"grad_norm": 1314.1402587890625,
"learning_rate": 3.92e-06,
"loss": 334.6603,
"step": 1960
},
{
"epoch": 0.007959049277423368,
"grad_norm": 2495.09130859375,
"learning_rate": 3.9399999999999995e-06,
"loss": 434.7513,
"step": 1970
},
{
"epoch": 0.007999450542791,
"grad_norm": 402.0351867675781,
"learning_rate": 3.96e-06,
"loss": 307.8936,
"step": 1980
},
{
"epoch": 0.008039851808158631,
"grad_norm": 751.7537841796875,
"learning_rate": 3.98e-06,
"loss": 308.6638,
"step": 1990
},
{
"epoch": 0.008080253073526262,
"grad_norm": 604.2171630859375,
"learning_rate": 4.000000000000001e-06,
"loss": 392.2409,
"step": 2000
},
{
"epoch": 0.008120654338893894,
"grad_norm": 1126.428955078125,
"learning_rate": 4.0200000000000005e-06,
"loss": 318.5683,
"step": 2010
},
{
"epoch": 0.008161055604261525,
"grad_norm": 4945.88037109375,
"learning_rate": 4.04e-06,
"loss": 311.324,
"step": 2020
},
{
"epoch": 0.008201456869629156,
"grad_norm": 745.3814697265625,
"learning_rate": 4.06e-06,
"loss": 381.683,
"step": 2030
},
{
"epoch": 0.008241858134996788,
"grad_norm": 556.3217163085938,
"learning_rate": 4.080000000000001e-06,
"loss": 335.636,
"step": 2040
},
{
"epoch": 0.008282259400364419,
"grad_norm": 1249.5528564453125,
"learning_rate": 4.1000000000000006e-06,
"loss": 216.6521,
"step": 2050
},
{
"epoch": 0.00832266066573205,
"grad_norm": 4082.633544921875,
"learning_rate": 4.12e-06,
"loss": 422.9517,
"step": 2060
},
{
"epoch": 0.008363061931099681,
"grad_norm": 1134.3583984375,
"learning_rate": 4.14e-06,
"loss": 368.1808,
"step": 2070
},
{
"epoch": 0.008403463196467313,
"grad_norm": 473.27655029296875,
"learning_rate": 4.16e-06,
"loss": 323.5618,
"step": 2080
},
{
"epoch": 0.008443864461834944,
"grad_norm": 1147.6612548828125,
"learning_rate": 4.18e-06,
"loss": 358.5086,
"step": 2090
},
{
"epoch": 0.008484265727202575,
"grad_norm": 716.1121826171875,
"learning_rate": 4.2000000000000004e-06,
"loss": 289.7142,
"step": 2100
},
{
"epoch": 0.008524666992570207,
"grad_norm": 595.6019287109375,
"learning_rate": 4.22e-06,
"loss": 249.1531,
"step": 2110
},
{
"epoch": 0.008565068257937838,
"grad_norm": 1831.8653564453125,
"learning_rate": 4.24e-06,
"loss": 336.2978,
"step": 2120
},
{
"epoch": 0.00860546952330547,
"grad_norm": 549.8616943359375,
"learning_rate": 4.26e-06,
"loss": 310.2739,
"step": 2130
},
{
"epoch": 0.0086458707886731,
"grad_norm": 888.9124755859375,
"learning_rate": 4.28e-06,
"loss": 372.0684,
"step": 2140
},
{
"epoch": 0.008686272054040732,
"grad_norm": 654.9121704101562,
"learning_rate": 4.2999999999999995e-06,
"loss": 218.5617,
"step": 2150
},
{
"epoch": 0.008726673319408363,
"grad_norm": 1140.73486328125,
"learning_rate": 4.32e-06,
"loss": 396.7231,
"step": 2160
},
{
"epoch": 0.008767074584775995,
"grad_norm": 818.670166015625,
"learning_rate": 4.34e-06,
"loss": 335.7865,
"step": 2170
},
{
"epoch": 0.008807475850143626,
"grad_norm": 792.7094116210938,
"learning_rate": 4.360000000000001e-06,
"loss": 321.234,
"step": 2180
},
{
"epoch": 0.008847877115511257,
"grad_norm": 2842.64697265625,
"learning_rate": 4.38e-06,
"loss": 247.4936,
"step": 2190
},
{
"epoch": 0.008888278380878889,
"grad_norm": 756.8375244140625,
"learning_rate": 4.4e-06,
"loss": 358.118,
"step": 2200
},
{
"epoch": 0.00892867964624652,
"grad_norm": 447.1920166015625,
"learning_rate": 4.420000000000001e-06,
"loss": 243.4523,
"step": 2210
},
{
"epoch": 0.008969080911614151,
"grad_norm": 598.7188110351562,
"learning_rate": 4.440000000000001e-06,
"loss": 317.7062,
"step": 2220
},
{
"epoch": 0.009009482176981783,
"grad_norm": 782.6261596679688,
"learning_rate": 4.4600000000000005e-06,
"loss": 423.8147,
"step": 2230
},
{
"epoch": 0.009049883442349414,
"grad_norm": 650.9381103515625,
"learning_rate": 4.48e-06,
"loss": 372.274,
"step": 2240
},
{
"epoch": 0.009090284707717045,
"grad_norm": 911.272216796875,
"learning_rate": 4.5e-06,
"loss": 309.0599,
"step": 2250
},
{
"epoch": 0.009130685973084677,
"grad_norm": 1821.9324951171875,
"learning_rate": 4.52e-06,
"loss": 291.9244,
"step": 2260
},
{
"epoch": 0.009171087238452308,
"grad_norm": 1428.2496337890625,
"learning_rate": 4.540000000000001e-06,
"loss": 299.4853,
"step": 2270
},
{
"epoch": 0.00921148850381994,
"grad_norm": 474.319580078125,
"learning_rate": 4.56e-06,
"loss": 284.1177,
"step": 2280
},
{
"epoch": 0.00925188976918757,
"grad_norm": 521.900146484375,
"learning_rate": 4.58e-06,
"loss": 305.8711,
"step": 2290
},
{
"epoch": 0.009292291034555202,
"grad_norm": 742.1409301757812,
"learning_rate": 4.6e-06,
"loss": 233.3259,
"step": 2300
},
{
"epoch": 0.009332692299922833,
"grad_norm": 702.3360595703125,
"learning_rate": 4.62e-06,
"loss": 351.8805,
"step": 2310
},
{
"epoch": 0.009373093565290464,
"grad_norm": 969.7785034179688,
"learning_rate": 4.64e-06,
"loss": 273.1551,
"step": 2320
},
{
"epoch": 0.009413494830658096,
"grad_norm": 7805.5947265625,
"learning_rate": 4.66e-06,
"loss": 303.4039,
"step": 2330
},
{
"epoch": 0.009453896096025727,
"grad_norm": 683.2047119140625,
"learning_rate": 4.68e-06,
"loss": 299.0721,
"step": 2340
},
{
"epoch": 0.009494297361393358,
"grad_norm": 433.588134765625,
"learning_rate": 4.7e-06,
"loss": 275.1868,
"step": 2350
},
{
"epoch": 0.00953469862676099,
"grad_norm": 970.4158325195312,
"learning_rate": 4.72e-06,
"loss": 508.6281,
"step": 2360
},
{
"epoch": 0.009575099892128621,
"grad_norm": 1364.8466796875,
"learning_rate": 4.74e-06,
"loss": 333.4644,
"step": 2370
},
{
"epoch": 0.009615501157496252,
"grad_norm": 1310.9354248046875,
"learning_rate": 4.76e-06,
"loss": 271.3296,
"step": 2380
},
{
"epoch": 0.009655902422863884,
"grad_norm": 551.7647094726562,
"learning_rate": 4.780000000000001e-06,
"loss": 339.265,
"step": 2390
},
{
"epoch": 0.009696303688231515,
"grad_norm": 542.0902709960938,
"learning_rate": 4.800000000000001e-06,
"loss": 375.706,
"step": 2400
},
{
"epoch": 0.009736704953599146,
"grad_norm": 517.1104736328125,
"learning_rate": 4.8200000000000004e-06,
"loss": 183.125,
"step": 2410
},
{
"epoch": 0.009777106218966778,
"grad_norm": 1165.0673828125,
"learning_rate": 4.84e-06,
"loss": 303.7644,
"step": 2420
},
{
"epoch": 0.009817507484334409,
"grad_norm": 4149.34130859375,
"learning_rate": 4.86e-06,
"loss": 361.1945,
"step": 2430
},
{
"epoch": 0.00985790874970204,
"grad_norm": 957.4842529296875,
"learning_rate": 4.880000000000001e-06,
"loss": 238.9853,
"step": 2440
},
{
"epoch": 0.009898310015069672,
"grad_norm": 493.18341064453125,
"learning_rate": 4.9000000000000005e-06,
"loss": 253.369,
"step": 2450
},
{
"epoch": 0.009938711280437303,
"grad_norm": 913.3916015625,
"learning_rate": 4.92e-06,
"loss": 197.3718,
"step": 2460
},
{
"epoch": 0.009979112545804934,
"grad_norm": 641.3722534179688,
"learning_rate": 4.94e-06,
"loss": 359.7244,
"step": 2470
},
{
"epoch": 0.010019513811172566,
"grad_norm": 805.710693359375,
"learning_rate": 4.96e-06,
"loss": 314.5141,
"step": 2480
},
{
"epoch": 0.010059915076540197,
"grad_norm": 979.5321655273438,
"learning_rate": 4.98e-06,
"loss": 402.7761,
"step": 2490
},
{
"epoch": 0.010100316341907828,
"grad_norm": 434.0482177734375,
"learning_rate": 5e-06,
"loss": 190.8965,
"step": 2500
},
{
"epoch": 0.01014071760727546,
"grad_norm": 907.0119018554688,
"learning_rate": 5.02e-06,
"loss": 269.7839,
"step": 2510
},
{
"epoch": 0.010181118872643091,
"grad_norm": 363.0292053222656,
"learning_rate": 5.04e-06,
"loss": 258.706,
"step": 2520
},
{
"epoch": 0.010221520138010722,
"grad_norm": 794.2672729492188,
"learning_rate": 5.06e-06,
"loss": 220.4875,
"step": 2530
},
{
"epoch": 0.010261921403378354,
"grad_norm": 956.221923828125,
"learning_rate": 5.08e-06,
"loss": 283.6568,
"step": 2540
},
{
"epoch": 0.010302322668745985,
"grad_norm": 779.488037109375,
"learning_rate": 5.1e-06,
"loss": 356.618,
"step": 2550
},
{
"epoch": 0.010342723934113616,
"grad_norm": 505.3084716796875,
"learning_rate": 5.12e-06,
"loss": 195.6782,
"step": 2560
},
{
"epoch": 0.010383125199481247,
"grad_norm": 1260.047119140625,
"learning_rate": 5.140000000000001e-06,
"loss": 321.1118,
"step": 2570
},
{
"epoch": 0.010423526464848879,
"grad_norm": 776.1239624023438,
"learning_rate": 5.1600000000000006e-06,
"loss": 303.4421,
"step": 2580
},
{
"epoch": 0.01046392773021651,
"grad_norm": 899.60009765625,
"learning_rate": 5.18e-06,
"loss": 267.6064,
"step": 2590
},
{
"epoch": 0.010504328995584141,
"grad_norm": 1198.875732421875,
"learning_rate": 5.2e-06,
"loss": 311.9983,
"step": 2600
},
{
"epoch": 0.010544730260951773,
"grad_norm": 2578.52734375,
"learning_rate": 5.220000000000001e-06,
"loss": 285.3186,
"step": 2610
},
{
"epoch": 0.010585131526319404,
"grad_norm": 613.3502807617188,
"learning_rate": 5.240000000000001e-06,
"loss": 407.8097,
"step": 2620
},
{
"epoch": 0.010625532791687035,
"grad_norm": 724.2944946289062,
"learning_rate": 5.2600000000000005e-06,
"loss": 375.3231,
"step": 2630
},
{
"epoch": 0.010665934057054667,
"grad_norm": 625.8546752929688,
"learning_rate": 5.28e-06,
"loss": 337.0123,
"step": 2640
},
{
"epoch": 0.010706335322422298,
"grad_norm": 2968.8515625,
"learning_rate": 5.3e-06,
"loss": 279.1451,
"step": 2650
},
{
"epoch": 0.01074673658778993,
"grad_norm": 507.51885986328125,
"learning_rate": 5.32e-06,
"loss": 214.3604,
"step": 2660
},
{
"epoch": 0.01078713785315756,
"grad_norm": 738.8612060546875,
"learning_rate": 5.3400000000000005e-06,
"loss": 332.2985,
"step": 2670
},
{
"epoch": 0.010827539118525192,
"grad_norm": 684.1400146484375,
"learning_rate": 5.36e-06,
"loss": 330.1413,
"step": 2680
},
{
"epoch": 0.010867940383892823,
"grad_norm": 6400.771484375,
"learning_rate": 5.38e-06,
"loss": 344.3884,
"step": 2690
},
{
"epoch": 0.010908341649260455,
"grad_norm": 986.3028564453125,
"learning_rate": 5.4e-06,
"loss": 354.2751,
"step": 2700
},
{
"epoch": 0.010948742914628086,
"grad_norm": 933.0596313476562,
"learning_rate": 5.42e-06,
"loss": 367.7645,
"step": 2710
},
{
"epoch": 0.010989144179995717,
"grad_norm": 956.349365234375,
"learning_rate": 5.44e-06,
"loss": 293.7472,
"step": 2720
},
{
"epoch": 0.011029545445363349,
"grad_norm": 473.7074890136719,
"learning_rate": 5.46e-06,
"loss": 281.8573,
"step": 2730
},
{
"epoch": 0.01106994671073098,
"grad_norm": 1302.126953125,
"learning_rate": 5.48e-06,
"loss": 275.9536,
"step": 2740
},
{
"epoch": 0.011110347976098611,
"grad_norm": 557.58203125,
"learning_rate": 5.500000000000001e-06,
"loss": 205.2503,
"step": 2750
},
{
"epoch": 0.011150749241466243,
"grad_norm": 3368.84228515625,
"learning_rate": 5.5200000000000005e-06,
"loss": 328.5797,
"step": 2760
},
{
"epoch": 0.011191150506833874,
"grad_norm": 1747.6932373046875,
"learning_rate": 5.54e-06,
"loss": 246.7141,
"step": 2770
},
{
"epoch": 0.011231551772201505,
"grad_norm": 871.47802734375,
"learning_rate": 5.56e-06,
"loss": 261.7437,
"step": 2780
},
{
"epoch": 0.011271953037569137,
"grad_norm": 1074.76416015625,
"learning_rate": 5.580000000000001e-06,
"loss": 292.0746,
"step": 2790
},
{
"epoch": 0.011312354302936768,
"grad_norm": 575.856689453125,
"learning_rate": 5.600000000000001e-06,
"loss": 247.1602,
"step": 2800
},
{
"epoch": 0.0113527555683044,
"grad_norm": 847.0436401367188,
"learning_rate": 5.62e-06,
"loss": 244.562,
"step": 2810
},
{
"epoch": 0.01139315683367203,
"grad_norm": 479.6651306152344,
"learning_rate": 5.64e-06,
"loss": 196.8626,
"step": 2820
},
{
"epoch": 0.011433558099039662,
"grad_norm": 1245.398681640625,
"learning_rate": 5.66e-06,
"loss": 338.9747,
"step": 2830
},
{
"epoch": 0.011473959364407293,
"grad_norm": 502.24951171875,
"learning_rate": 5.680000000000001e-06,
"loss": 243.1993,
"step": 2840
},
{
"epoch": 0.011514360629774924,
"grad_norm": 790.6267700195312,
"learning_rate": 5.7000000000000005e-06,
"loss": 289.827,
"step": 2850
},
{
"epoch": 0.011554761895142556,
"grad_norm": 996.7022094726562,
"learning_rate": 5.72e-06,
"loss": 285.2487,
"step": 2860
},
{
"epoch": 0.011595163160510187,
"grad_norm": 1381.54541015625,
"learning_rate": 5.74e-06,
"loss": 299.6274,
"step": 2870
},
{
"epoch": 0.011635564425877818,
"grad_norm": 674.7173461914062,
"learning_rate": 5.76e-06,
"loss": 212.4647,
"step": 2880
},
{
"epoch": 0.01167596569124545,
"grad_norm": 1115.6590576171875,
"learning_rate": 5.78e-06,
"loss": 350.1185,
"step": 2890
},
{
"epoch": 0.011716366956613081,
"grad_norm": 967.9352416992188,
"learning_rate": 5.8e-06,
"loss": 303.8514,
"step": 2900
},
{
"epoch": 0.011756768221980712,
"grad_norm": 826.4132690429688,
"learning_rate": 5.82e-06,
"loss": 390.5727,
"step": 2910
},
{
"epoch": 0.011797169487348344,
"grad_norm": 554.1842651367188,
"learning_rate": 5.84e-06,
"loss": 228.6073,
"step": 2920
},
{
"epoch": 0.011837570752715975,
"grad_norm": 555.0242919921875,
"learning_rate": 5.86e-06,
"loss": 237.2674,
"step": 2930
},
{
"epoch": 0.011877972018083606,
"grad_norm": 554.800048828125,
"learning_rate": 5.8800000000000005e-06,
"loss": 307.6302,
"step": 2940
},
{
"epoch": 0.011918373283451238,
"grad_norm": 7435.06005859375,
"learning_rate": 5.9e-06,
"loss": 363.9746,
"step": 2950
},
{
"epoch": 0.011958774548818869,
"grad_norm": 3039.5869140625,
"learning_rate": 5.920000000000001e-06,
"loss": 294.7946,
"step": 2960
},
{
"epoch": 0.0119991758141865,
"grad_norm": 9456.8857421875,
"learning_rate": 5.940000000000001e-06,
"loss": 348.9202,
"step": 2970
},
{
"epoch": 0.012039577079554132,
"grad_norm": 833.540771484375,
"learning_rate": 5.9600000000000005e-06,
"loss": 302.1644,
"step": 2980
},
{
"epoch": 0.012079978344921763,
"grad_norm": 1317.080810546875,
"learning_rate": 5.98e-06,
"loss": 346.3184,
"step": 2990
},
{
"epoch": 0.012120379610289394,
"grad_norm": 546.1310424804688,
"learning_rate": 6e-06,
"loss": 185.3799,
"step": 3000
},
{
"epoch": 0.012160780875657026,
"grad_norm": 1277.5128173828125,
"learning_rate": 6.02e-06,
"loss": 271.2112,
"step": 3010
},
{
"epoch": 0.012201182141024657,
"grad_norm": 651.244384765625,
"learning_rate": 6.040000000000001e-06,
"loss": 206.7404,
"step": 3020
},
{
"epoch": 0.012241583406392288,
"grad_norm": 886.7020874023438,
"learning_rate": 6.0600000000000004e-06,
"loss": 309.8574,
"step": 3030
},
{
"epoch": 0.01228198467175992,
"grad_norm": 765.9307861328125,
"learning_rate": 6.08e-06,
"loss": 412.906,
"step": 3040
},
{
"epoch": 0.01232238593712755,
"grad_norm": 1054.11669921875,
"learning_rate": 6.1e-06,
"loss": 356.0523,
"step": 3050
},
{
"epoch": 0.012362787202495182,
"grad_norm": 786.2109375,
"learning_rate": 6.12e-06,
"loss": 260.9853,
"step": 3060
},
{
"epoch": 0.012403188467862813,
"grad_norm": 1215.10400390625,
"learning_rate": 6.1400000000000005e-06,
"loss": 259.1632,
"step": 3070
},
{
"epoch": 0.012443589733230445,
"grad_norm": 943.6842651367188,
"learning_rate": 6.16e-06,
"loss": 216.0937,
"step": 3080
},
{
"epoch": 0.012483990998598076,
"grad_norm": 1460.5831298828125,
"learning_rate": 6.18e-06,
"loss": 337.1417,
"step": 3090
},
{
"epoch": 0.012524392263965707,
"grad_norm": 3781.843017578125,
"learning_rate": 6.2e-06,
"loss": 374.3547,
"step": 3100
},
{
"epoch": 0.012564793529333339,
"grad_norm": 676.937744140625,
"learning_rate": 6.22e-06,
"loss": 350.8003,
"step": 3110
},
{
"epoch": 0.01260519479470097,
"grad_norm": 1300.2470703125,
"learning_rate": 6.24e-06,
"loss": 354.9542,
"step": 3120
},
{
"epoch": 0.012645596060068601,
"grad_norm": 715.6893310546875,
"learning_rate": 6.26e-06,
"loss": 225.5772,
"step": 3130
},
{
"epoch": 0.012685997325436233,
"grad_norm": 438.3706359863281,
"learning_rate": 6.28e-06,
"loss": 262.8397,
"step": 3140
},
{
"epoch": 0.012726398590803864,
"grad_norm": 5010.02392578125,
"learning_rate": 6.300000000000001e-06,
"loss": 282.8991,
"step": 3150
},
{
"epoch": 0.012766799856171495,
"grad_norm": 1642.5548095703125,
"learning_rate": 6.320000000000001e-06,
"loss": 273.512,
"step": 3160
},
{
"epoch": 0.012807201121539127,
"grad_norm": 484.15069580078125,
"learning_rate": 6.34e-06,
"loss": 190.9169,
"step": 3170
},
{
"epoch": 0.012847602386906758,
"grad_norm": 1157.349365234375,
"learning_rate": 6.360000000000001e-06,
"loss": 378.3686,
"step": 3180
},
{
"epoch": 0.01288800365227439,
"grad_norm": 547.7994384765625,
"learning_rate": 6.38e-06,
"loss": 241.6315,
"step": 3190
},
{
"epoch": 0.01292840491764202,
"grad_norm": 0.0,
"learning_rate": 6.4000000000000006e-06,
"loss": 233.8554,
"step": 3200
},
{
"epoch": 0.012968806183009652,
"grad_norm": 953.7396850585938,
"learning_rate": 6.4199999999999995e-06,
"loss": 316.3731,
"step": 3210
},
{
"epoch": 0.013009207448377283,
"grad_norm": 539.7330932617188,
"learning_rate": 6.44e-06,
"loss": 234.0129,
"step": 3220
},
{
"epoch": 0.013049608713744915,
"grad_norm": 736.7625732421875,
"learning_rate": 6.460000000000001e-06,
"loss": 335.0041,
"step": 3230
},
{
"epoch": 0.013090009979112546,
"grad_norm": 1194.715576171875,
"learning_rate": 6.48e-06,
"loss": 276.9117,
"step": 3240
},
{
"epoch": 0.013130411244480177,
"grad_norm": 8021.32763671875,
"learning_rate": 6.5000000000000004e-06,
"loss": 374.7519,
"step": 3250
},
{
"epoch": 0.013170812509847809,
"grad_norm": 931.0237426757812,
"learning_rate": 6.519999999999999e-06,
"loss": 278.108,
"step": 3260
},
{
"epoch": 0.01321121377521544,
"grad_norm": 953.587890625,
"learning_rate": 6.54e-06,
"loss": 368.7168,
"step": 3270
},
{
"epoch": 0.013251615040583071,
"grad_norm": 1098.186767578125,
"learning_rate": 6.560000000000001e-06,
"loss": 276.9152,
"step": 3280
},
{
"epoch": 0.013292016305950703,
"grad_norm": 1590.1026611328125,
"learning_rate": 6.58e-06,
"loss": 422.8925,
"step": 3290
},
{
"epoch": 0.013332417571318334,
"grad_norm": 857.4612426757812,
"learning_rate": 6.6e-06,
"loss": 343.3155,
"step": 3300
},
{
"epoch": 0.013372818836685965,
"grad_norm": 941.4984130859375,
"learning_rate": 6.62e-06,
"loss": 187.542,
"step": 3310
},
{
"epoch": 0.013413220102053596,
"grad_norm": 938.3494262695312,
"learning_rate": 6.640000000000001e-06,
"loss": 237.5327,
"step": 3320
},
{
"epoch": 0.013453621367421228,
"grad_norm": 936.9948120117188,
"learning_rate": 6.660000000000001e-06,
"loss": 275.3572,
"step": 3330
},
{
"epoch": 0.013494022632788859,
"grad_norm": 1025.8851318359375,
"learning_rate": 6.68e-06,
"loss": 346.2491,
"step": 3340
},
{
"epoch": 0.01353442389815649,
"grad_norm": 2196.15478515625,
"learning_rate": 6.700000000000001e-06,
"loss": 242.7346,
"step": 3350
},
{
"epoch": 0.013574825163524122,
"grad_norm": 843.5136108398438,
"learning_rate": 6.72e-06,
"loss": 188.1487,
"step": 3360
},
{
"epoch": 0.013615226428891753,
"grad_norm": 714.3571166992188,
"learning_rate": 6.740000000000001e-06,
"loss": 358.2911,
"step": 3370
},
{
"epoch": 0.013655627694259384,
"grad_norm": 689.7639770507812,
"learning_rate": 6.76e-06,
"loss": 150.4375,
"step": 3380
},
{
"epoch": 0.013696028959627016,
"grad_norm": 1244.08837890625,
"learning_rate": 6.78e-06,
"loss": 300.2315,
"step": 3390
},
{
"epoch": 0.013736430224994647,
"grad_norm": 1171.05419921875,
"learning_rate": 6.800000000000001e-06,
"loss": 241.5314,
"step": 3400
},
{
"epoch": 0.013776831490362278,
"grad_norm": 1291.3603515625,
"learning_rate": 6.82e-06,
"loss": 449.0854,
"step": 3410
},
{
"epoch": 0.01381723275572991,
"grad_norm": 1006.1640014648438,
"learning_rate": 6.840000000000001e-06,
"loss": 346.3529,
"step": 3420
},
{
"epoch": 0.013857634021097541,
"grad_norm": 6614.48583984375,
"learning_rate": 6.8599999999999995e-06,
"loss": 289.8388,
"step": 3430
},
{
"epoch": 0.013898035286465172,
"grad_norm": 574.2523803710938,
"learning_rate": 6.88e-06,
"loss": 187.4786,
"step": 3440
},
{
"epoch": 0.013938436551832804,
"grad_norm": 577.289794921875,
"learning_rate": 6.900000000000001e-06,
"loss": 280.2674,
"step": 3450
},
{
"epoch": 0.013978837817200435,
"grad_norm": 1416.127197265625,
"learning_rate": 6.92e-06,
"loss": 330.4,
"step": 3460
},
{
"epoch": 0.014019239082568066,
"grad_norm": 973.31689453125,
"learning_rate": 6.9400000000000005e-06,
"loss": 251.0885,
"step": 3470
},
{
"epoch": 0.014059640347935698,
"grad_norm": 2889.614013671875,
"learning_rate": 6.9599999999999994e-06,
"loss": 288.5922,
"step": 3480
},
{
"epoch": 0.014100041613303329,
"grad_norm": 734.4885864257812,
"learning_rate": 6.98e-06,
"loss": 247.6378,
"step": 3490
},
{
"epoch": 0.01414044287867096,
"grad_norm": 1920.7076416015625,
"learning_rate": 7.000000000000001e-06,
"loss": 175.3216,
"step": 3500
},
{
"epoch": 0.014180844144038592,
"grad_norm": 676.3031005859375,
"learning_rate": 7.0200000000000006e-06,
"loss": 209.18,
"step": 3510
},
{
"epoch": 0.014221245409406223,
"grad_norm": 1840.922119140625,
"learning_rate": 7.04e-06,
"loss": 241.9013,
"step": 3520
},
{
"epoch": 0.014261646674773854,
"grad_norm": 1004.5390625,
"learning_rate": 7.06e-06,
"loss": 219.0837,
"step": 3530
},
{
"epoch": 0.014302047940141486,
"grad_norm": 2489.928955078125,
"learning_rate": 7.080000000000001e-06,
"loss": 207.0227,
"step": 3540
},
{
"epoch": 0.014342449205509117,
"grad_norm": 713.509033203125,
"learning_rate": 7.1e-06,
"loss": 208.6873,
"step": 3550
},
{
"epoch": 0.014382850470876748,
"grad_norm": 453.5292053222656,
"learning_rate": 7.1200000000000004e-06,
"loss": 281.9877,
"step": 3560
},
{
"epoch": 0.01442325173624438,
"grad_norm": 900.6409912109375,
"learning_rate": 7.140000000000001e-06,
"loss": 235.5468,
"step": 3570
},
{
"epoch": 0.01446365300161201,
"grad_norm": 438.5726623535156,
"learning_rate": 7.16e-06,
"loss": 162.5427,
"step": 3580
},
{
"epoch": 0.014504054266979642,
"grad_norm": 861.850341796875,
"learning_rate": 7.180000000000001e-06,
"loss": 236.0929,
"step": 3590
},
{
"epoch": 0.014544455532347273,
"grad_norm": 892.953369140625,
"learning_rate": 7.2e-06,
"loss": 249.5097,
"step": 3600
},
{
"epoch": 0.014584856797714905,
"grad_norm": 1040.248046875,
"learning_rate": 7.22e-06,
"loss": 268.9143,
"step": 3610
},
{
"epoch": 0.014625258063082536,
"grad_norm": 774.4244384765625,
"learning_rate": 7.240000000000001e-06,
"loss": 235.6966,
"step": 3620
},
{
"epoch": 0.014665659328450167,
"grad_norm": 540.7510986328125,
"learning_rate": 7.26e-06,
"loss": 222.3721,
"step": 3630
},
{
"epoch": 0.014706060593817799,
"grad_norm": 3457.84912109375,
"learning_rate": 7.280000000000001e-06,
"loss": 262.0487,
"step": 3640
},
{
"epoch": 0.01474646185918543,
"grad_norm": 802.951416015625,
"learning_rate": 7.2999999999999996e-06,
"loss": 277.8951,
"step": 3650
},
{
"epoch": 0.014786863124553061,
"grad_norm": 1426.4849853515625,
"learning_rate": 7.32e-06,
"loss": 273.6454,
"step": 3660
},
{
"epoch": 0.014827264389920693,
"grad_norm": 763.5048828125,
"learning_rate": 7.340000000000001e-06,
"loss": 324.4447,
"step": 3670
},
{
"epoch": 0.014867665655288324,
"grad_norm": 2634.498046875,
"learning_rate": 7.36e-06,
"loss": 285.5299,
"step": 3680
},
{
"epoch": 0.014908066920655955,
"grad_norm": 1109.640380859375,
"learning_rate": 7.3800000000000005e-06,
"loss": 421.3942,
"step": 3690
},
{
"epoch": 0.014948468186023587,
"grad_norm": 1876.8143310546875,
"learning_rate": 7.4e-06,
"loss": 345.6086,
"step": 3700
},
{
"epoch": 0.014988869451391218,
"grad_norm": 820.827880859375,
"learning_rate": 7.420000000000001e-06,
"loss": 241.7691,
"step": 3710
},
{
"epoch": 0.01502927071675885,
"grad_norm": 1500.841552734375,
"learning_rate": 7.44e-06,
"loss": 247.7824,
"step": 3720
},
{
"epoch": 0.01506967198212648,
"grad_norm": 1121.291259765625,
"learning_rate": 7.4600000000000006e-06,
"loss": 302.3958,
"step": 3730
},
{
"epoch": 0.015110073247494112,
"grad_norm": 1057.3134765625,
"learning_rate": 7.480000000000001e-06,
"loss": 271.7261,
"step": 3740
},
{
"epoch": 0.015150474512861743,
"grad_norm": 1772.322509765625,
"learning_rate": 7.5e-06,
"loss": 404.0533,
"step": 3750
},
{
"epoch": 0.015190875778229375,
"grad_norm": 547.7960815429688,
"learning_rate": 7.520000000000001e-06,
"loss": 255.6221,
"step": 3760
},
{
"epoch": 0.015231277043597006,
"grad_norm": 980.1959228515625,
"learning_rate": 7.54e-06,
"loss": 310.8835,
"step": 3770
},
{
"epoch": 0.015271678308964637,
"grad_norm": 1686.9091796875,
"learning_rate": 7.5600000000000005e-06,
"loss": 246.4056,
"step": 3780
},
{
"epoch": 0.015312079574332269,
"grad_norm": 867.049072265625,
"learning_rate": 7.580000000000001e-06,
"loss": 246.8477,
"step": 3790
},
{
"epoch": 0.0153524808396999,
"grad_norm": 1169.5635986328125,
"learning_rate": 7.6e-06,
"loss": 184.4648,
"step": 3800
},
{
"epoch": 0.015392882105067531,
"grad_norm": 689.9214477539062,
"learning_rate": 7.620000000000001e-06,
"loss": 323.659,
"step": 3810
},
{
"epoch": 0.015433283370435162,
"grad_norm": 519.1124267578125,
"learning_rate": 7.64e-06,
"loss": 225.6972,
"step": 3820
},
{
"epoch": 0.015473684635802794,
"grad_norm": 1257.6265869140625,
"learning_rate": 7.660000000000001e-06,
"loss": 337.094,
"step": 3830
},
{
"epoch": 0.015514085901170425,
"grad_norm": 5566.001953125,
"learning_rate": 7.68e-06,
"loss": 191.7031,
"step": 3840
},
{
"epoch": 0.015554487166538056,
"grad_norm": 734.2090454101562,
"learning_rate": 7.7e-06,
"loss": 221.7874,
"step": 3850
},
{
"epoch": 0.015594888431905688,
"grad_norm": 1421.68359375,
"learning_rate": 7.72e-06,
"loss": 266.4913,
"step": 3860
},
{
"epoch": 0.01563528969727332,
"grad_norm": 2573.87353515625,
"learning_rate": 7.74e-06,
"loss": 301.2037,
"step": 3870
},
{
"epoch": 0.01567569096264095,
"grad_norm": 492.0390319824219,
"learning_rate": 7.76e-06,
"loss": 234.7707,
"step": 3880
},
{
"epoch": 0.01571609222800858,
"grad_norm": 600.4080200195312,
"learning_rate": 7.78e-06,
"loss": 267.7107,
"step": 3890
},
{
"epoch": 0.01575649349337621,
"grad_norm": 1099.673828125,
"learning_rate": 7.8e-06,
"loss": 292.7474,
"step": 3900
},
{
"epoch": 0.015796894758743844,
"grad_norm": 541.96875,
"learning_rate": 7.820000000000001e-06,
"loss": 229.897,
"step": 3910
},
{
"epoch": 0.015837296024111474,
"grad_norm": 3221.322509765625,
"learning_rate": 7.84e-06,
"loss": 200.1891,
"step": 3920
},
{
"epoch": 0.015877697289479107,
"grad_norm": 2533.82177734375,
"learning_rate": 7.860000000000001e-06,
"loss": 240.3558,
"step": 3930
},
{
"epoch": 0.015918098554846737,
"grad_norm": 909.9031372070312,
"learning_rate": 7.879999999999999e-06,
"loss": 231.9597,
"step": 3940
},
{
"epoch": 0.01595849982021437,
"grad_norm": 669.5313110351562,
"learning_rate": 7.9e-06,
"loss": 316.9614,
"step": 3950
},
{
"epoch": 0.015998901085582,
"grad_norm": 694.2930297851562,
"learning_rate": 7.92e-06,
"loss": 272.3228,
"step": 3960
},
{
"epoch": 0.016039302350949632,
"grad_norm": 528.3792114257812,
"learning_rate": 7.94e-06,
"loss": 329.1032,
"step": 3970
},
{
"epoch": 0.016079703616317262,
"grad_norm": 579.0252075195312,
"learning_rate": 7.96e-06,
"loss": 258.9907,
"step": 3980
},
{
"epoch": 0.016120104881684895,
"grad_norm": 645.7807006835938,
"learning_rate": 7.98e-06,
"loss": 242.0821,
"step": 3990
},
{
"epoch": 0.016160506147052525,
"grad_norm": 543.9231567382812,
"learning_rate": 8.000000000000001e-06,
"loss": 284.7925,
"step": 4000
},
{
"epoch": 0.016200907412420158,
"grad_norm": 610.683349609375,
"learning_rate": 8.02e-06,
"loss": 220.6932,
"step": 4010
},
{
"epoch": 0.016241308677787787,
"grad_norm": 0.0,
"learning_rate": 8.040000000000001e-06,
"loss": 230.9501,
"step": 4020
},
{
"epoch": 0.01628170994315542,
"grad_norm": 1050.20947265625,
"learning_rate": 8.06e-06,
"loss": 292.7622,
"step": 4030
},
{
"epoch": 0.01632211120852305,
"grad_norm": 1064.6005859375,
"learning_rate": 8.08e-06,
"loss": 346.057,
"step": 4040
},
{
"epoch": 0.016362512473890683,
"grad_norm": 1817.19482421875,
"learning_rate": 8.1e-06,
"loss": 229.4932,
"step": 4050
},
{
"epoch": 0.016402913739258312,
"grad_norm": 1169.279052734375,
"learning_rate": 8.12e-06,
"loss": 346.2139,
"step": 4060
},
{
"epoch": 0.016443315004625945,
"grad_norm": 1004.5889282226562,
"learning_rate": 8.14e-06,
"loss": 257.1552,
"step": 4070
},
{
"epoch": 0.016483716269993575,
"grad_norm": 660.7535400390625,
"learning_rate": 8.160000000000001e-06,
"loss": 247.7336,
"step": 4080
},
{
"epoch": 0.016524117535361208,
"grad_norm": 638.9656372070312,
"learning_rate": 8.18e-06,
"loss": 360.9811,
"step": 4090
},
{
"epoch": 0.016564518800728838,
"grad_norm": 865.1559448242188,
"learning_rate": 8.200000000000001e-06,
"loss": 267.7802,
"step": 4100
},
{
"epoch": 0.01660492006609647,
"grad_norm": 528.1417236328125,
"learning_rate": 8.22e-06,
"loss": 211.1934,
"step": 4110
},
{
"epoch": 0.0166453213314641,
"grad_norm": 602.795166015625,
"learning_rate": 8.24e-06,
"loss": 342.2845,
"step": 4120
},
{
"epoch": 0.016685722596831733,
"grad_norm": 537.5540161132812,
"learning_rate": 8.26e-06,
"loss": 222.6927,
"step": 4130
},
{
"epoch": 0.016726123862199363,
"grad_norm": 1009.3809814453125,
"learning_rate": 8.28e-06,
"loss": 278.8901,
"step": 4140
},
{
"epoch": 0.016766525127566996,
"grad_norm": 1916.926513671875,
"learning_rate": 8.3e-06,
"loss": 335.051,
"step": 4150
},
{
"epoch": 0.016806926392934626,
"grad_norm": 759.0120849609375,
"learning_rate": 8.32e-06,
"loss": 292.8282,
"step": 4160
},
{
"epoch": 0.01684732765830226,
"grad_norm": 1126.7659912109375,
"learning_rate": 8.34e-06,
"loss": 356.8466,
"step": 4170
},
{
"epoch": 0.016887728923669888,
"grad_norm": 4519.04833984375,
"learning_rate": 8.36e-06,
"loss": 254.2794,
"step": 4180
},
{
"epoch": 0.01692813018903752,
"grad_norm": 625.6685791015625,
"learning_rate": 8.380000000000001e-06,
"loss": 256.8828,
"step": 4190
},
{
"epoch": 0.01696853145440515,
"grad_norm": 901.2313842773438,
"learning_rate": 8.400000000000001e-06,
"loss": 173.8549,
"step": 4200
},
{
"epoch": 0.017008932719772784,
"grad_norm": 612.3013916015625,
"learning_rate": 8.42e-06,
"loss": 278.5784,
"step": 4210
},
{
"epoch": 0.017049333985140414,
"grad_norm": 500.49169921875,
"learning_rate": 8.44e-06,
"loss": 197.3738,
"step": 4220
},
{
"epoch": 0.017089735250508047,
"grad_norm": 1060.2108154296875,
"learning_rate": 8.46e-06,
"loss": 298.4693,
"step": 4230
},
{
"epoch": 0.017130136515875676,
"grad_norm": 612.9854736328125,
"learning_rate": 8.48e-06,
"loss": 329.1869,
"step": 4240
},
{
"epoch": 0.01717053778124331,
"grad_norm": 1364.545654296875,
"learning_rate": 8.500000000000002e-06,
"loss": 334.3979,
"step": 4250
},
{
"epoch": 0.01721093904661094,
"grad_norm": 696.9126586914062,
"learning_rate": 8.52e-06,
"loss": 271.8317,
"step": 4260
},
{
"epoch": 0.017251340311978572,
"grad_norm": 688.8264770507812,
"learning_rate": 8.540000000000001e-06,
"loss": 203.2293,
"step": 4270
},
{
"epoch": 0.0172917415773462,
"grad_norm": 1147.045654296875,
"learning_rate": 8.56e-06,
"loss": 349.6926,
"step": 4280
},
{
"epoch": 0.017332142842713835,
"grad_norm": 477.3202209472656,
"learning_rate": 8.580000000000001e-06,
"loss": 146.8848,
"step": 4290
},
{
"epoch": 0.017372544108081464,
"grad_norm": 1310.89599609375,
"learning_rate": 8.599999999999999e-06,
"loss": 247.6962,
"step": 4300
},
{
"epoch": 0.017412945373449097,
"grad_norm": 1124.722900390625,
"learning_rate": 8.62e-06,
"loss": 347.9728,
"step": 4310
},
{
"epoch": 0.017453346638816727,
"grad_norm": 1394.403564453125,
"learning_rate": 8.64e-06,
"loss": 261.2298,
"step": 4320
},
{
"epoch": 0.01749374790418436,
"grad_norm": 889.3135375976562,
"learning_rate": 8.66e-06,
"loss": 176.3773,
"step": 4330
},
{
"epoch": 0.01753414916955199,
"grad_norm": 1988.6873779296875,
"learning_rate": 8.68e-06,
"loss": 260.4097,
"step": 4340
},
{
"epoch": 0.017574550434919622,
"grad_norm": 1403.53955078125,
"learning_rate": 8.7e-06,
"loss": 222.8027,
"step": 4350
},
{
"epoch": 0.017614951700287252,
"grad_norm": 1020.5213623046875,
"learning_rate": 8.720000000000001e-06,
"loss": 202.7748,
"step": 4360
},
{
"epoch": 0.017655352965654885,
"grad_norm": 4106.30126953125,
"learning_rate": 8.740000000000001e-06,
"loss": 257.8873,
"step": 4370
},
{
"epoch": 0.017695754231022515,
"grad_norm": 458.9610290527344,
"learning_rate": 8.76e-06,
"loss": 274.0975,
"step": 4380
},
{
"epoch": 0.017736155496390148,
"grad_norm": 1450.5250244140625,
"learning_rate": 8.78e-06,
"loss": 212.7885,
"step": 4390
},
{
"epoch": 0.017776556761757777,
"grad_norm": 658.0136108398438,
"learning_rate": 8.8e-06,
"loss": 313.3228,
"step": 4400
},
{
"epoch": 0.01781695802712541,
"grad_norm": 607.8909301757812,
"learning_rate": 8.82e-06,
"loss": 220.5824,
"step": 4410
},
{
"epoch": 0.01785735929249304,
"grad_norm": 964.3424682617188,
"learning_rate": 8.840000000000002e-06,
"loss": 214.6224,
"step": 4420
},
{
"epoch": 0.017897760557860673,
"grad_norm": 1267.5516357421875,
"learning_rate": 8.86e-06,
"loss": 231.4749,
"step": 4430
},
{
"epoch": 0.017938161823228303,
"grad_norm": 852.99462890625,
"learning_rate": 8.880000000000001e-06,
"loss": 209.5046,
"step": 4440
},
{
"epoch": 0.017978563088595936,
"grad_norm": 895.4765014648438,
"learning_rate": 8.9e-06,
"loss": 227.3372,
"step": 4450
},
{
"epoch": 0.018018964353963565,
"grad_norm": 658.2288818359375,
"learning_rate": 8.920000000000001e-06,
"loss": 235.962,
"step": 4460
},
{
"epoch": 0.0180593656193312,
"grad_norm": 1219.0494384765625,
"learning_rate": 8.939999999999999e-06,
"loss": 222.9912,
"step": 4470
},
{
"epoch": 0.018099766884698828,
"grad_norm": 976.2411499023438,
"learning_rate": 8.96e-06,
"loss": 251.0392,
"step": 4480
},
{
"epoch": 0.01814016815006646,
"grad_norm": 1230.3253173828125,
"learning_rate": 8.98e-06,
"loss": 208.2115,
"step": 4490
},
{
"epoch": 0.01818056941543409,
"grad_norm": 738.1622314453125,
"learning_rate": 9e-06,
"loss": 248.237,
"step": 4500
},
{
"epoch": 0.018220970680801724,
"grad_norm": 825.9674072265625,
"learning_rate": 9.02e-06,
"loss": 231.5556,
"step": 4510
},
{
"epoch": 0.018261371946169353,
"grad_norm": 782.737060546875,
"learning_rate": 9.04e-06,
"loss": 213.6435,
"step": 4520
},
{
"epoch": 0.018301773211536986,
"grad_norm": 712.6553344726562,
"learning_rate": 9.06e-06,
"loss": 234.0471,
"step": 4530
},
{
"epoch": 0.018342174476904616,
"grad_norm": 1101.6629638671875,
"learning_rate": 9.080000000000001e-06,
"loss": 198.3389,
"step": 4540
},
{
"epoch": 0.01838257574227225,
"grad_norm": 1747.929443359375,
"learning_rate": 9.100000000000001e-06,
"loss": 332.4391,
"step": 4550
},
{
"epoch": 0.01842297700763988,
"grad_norm": 2980.9208984375,
"learning_rate": 9.12e-06,
"loss": 337.9396,
"step": 4560
},
{
"epoch": 0.01846337827300751,
"grad_norm": 1282.68115234375,
"learning_rate": 9.14e-06,
"loss": 237.8889,
"step": 4570
},
{
"epoch": 0.01850377953837514,
"grad_norm": 948.6932983398438,
"learning_rate": 9.16e-06,
"loss": 288.0977,
"step": 4580
},
{
"epoch": 0.018544180803742774,
"grad_norm": 501.8581237792969,
"learning_rate": 9.180000000000002e-06,
"loss": 329.9324,
"step": 4590
},
{
"epoch": 0.018584582069110404,
"grad_norm": 1850.9791259765625,
"learning_rate": 9.2e-06,
"loss": 319.7892,
"step": 4600
},
{
"epoch": 0.018624983334478037,
"grad_norm": 730.577392578125,
"learning_rate": 9.220000000000002e-06,
"loss": 217.722,
"step": 4610
},
{
"epoch": 0.018665384599845666,
"grad_norm": 408.53619384765625,
"learning_rate": 9.24e-06,
"loss": 190.9173,
"step": 4620
},
{
"epoch": 0.0187057858652133,
"grad_norm": 884.1920776367188,
"learning_rate": 9.260000000000001e-06,
"loss": 193.8064,
"step": 4630
},
{
"epoch": 0.01874618713058093,
"grad_norm": 1684.18408203125,
"learning_rate": 9.28e-06,
"loss": 197.2281,
"step": 4640
},
{
"epoch": 0.018786588395948562,
"grad_norm": 1591.9307861328125,
"learning_rate": 9.3e-06,
"loss": 282.6273,
"step": 4650
},
{
"epoch": 0.01882698966131619,
"grad_norm": 1377.8363037109375,
"learning_rate": 9.32e-06,
"loss": 306.7148,
"step": 4660
},
{
"epoch": 0.018867390926683825,
"grad_norm": 600.4591674804688,
"learning_rate": 9.34e-06,
"loss": 290.5179,
"step": 4670
},
{
"epoch": 0.018907792192051454,
"grad_norm": 1007.29296875,
"learning_rate": 9.36e-06,
"loss": 211.2058,
"step": 4680
},
{
"epoch": 0.018948193457419087,
"grad_norm": 1834.026611328125,
"learning_rate": 9.38e-06,
"loss": 214.4897,
"step": 4690
},
{
"epoch": 0.018988594722786717,
"grad_norm": 895.9391479492188,
"learning_rate": 9.4e-06,
"loss": 241.2341,
"step": 4700
},
{
"epoch": 0.01902899598815435,
"grad_norm": 940.978271484375,
"learning_rate": 9.420000000000001e-06,
"loss": 255.7137,
"step": 4710
},
{
"epoch": 0.01906939725352198,
"grad_norm": 0.0,
"learning_rate": 9.44e-06,
"loss": 202.176,
"step": 4720
},
{
"epoch": 0.019109798518889613,
"grad_norm": 1429.115234375,
"learning_rate": 9.460000000000001e-06,
"loss": 199.2825,
"step": 4730
},
{
"epoch": 0.019150199784257242,
"grad_norm": 1368.30322265625,
"learning_rate": 9.48e-06,
"loss": 265.0907,
"step": 4740
},
{
"epoch": 0.019190601049624875,
"grad_norm": 773.6226196289062,
"learning_rate": 9.5e-06,
"loss": 272.2148,
"step": 4750
},
{
"epoch": 0.019231002314992505,
"grad_norm": 799.7369384765625,
"learning_rate": 9.52e-06,
"loss": 259.8739,
"step": 4760
},
{
"epoch": 0.019271403580360138,
"grad_norm": 718.1935424804688,
"learning_rate": 9.54e-06,
"loss": 255.6761,
"step": 4770
},
{
"epoch": 0.019311804845727767,
"grad_norm": 1052.681396484375,
"learning_rate": 9.560000000000002e-06,
"loss": 240.1528,
"step": 4780
},
{
"epoch": 0.0193522061110954,
"grad_norm": 351.5966491699219,
"learning_rate": 9.58e-06,
"loss": 170.535,
"step": 4790
},
{
"epoch": 0.01939260737646303,
"grad_norm": 759.4265747070312,
"learning_rate": 9.600000000000001e-06,
"loss": 185.2903,
"step": 4800
},
{
"epoch": 0.019433008641830663,
"grad_norm": 763.7293701171875,
"learning_rate": 9.62e-06,
"loss": 233.2488,
"step": 4810
},
{
"epoch": 0.019473409907198293,
"grad_norm": 1342.382568359375,
"learning_rate": 9.640000000000001e-06,
"loss": 275.3923,
"step": 4820
},
{
"epoch": 0.019513811172565926,
"grad_norm": 3298.531005859375,
"learning_rate": 9.66e-06,
"loss": 226.5765,
"step": 4830
},
{
"epoch": 0.019554212437933555,
"grad_norm": 1122.7933349609375,
"learning_rate": 9.68e-06,
"loss": 178.3835,
"step": 4840
},
{
"epoch": 0.01959461370330119,
"grad_norm": 1547.0048828125,
"learning_rate": 9.7e-06,
"loss": 305.318,
"step": 4850
},
{
"epoch": 0.019635014968668818,
"grad_norm": 364.29541015625,
"learning_rate": 9.72e-06,
"loss": 170.9627,
"step": 4860
},
{
"epoch": 0.01967541623403645,
"grad_norm": 1955.25634765625,
"learning_rate": 9.74e-06,
"loss": 302.9251,
"step": 4870
},
{
"epoch": 0.01971581749940408,
"grad_norm": 2048.748291015625,
"learning_rate": 9.760000000000001e-06,
"loss": 202.4392,
"step": 4880
},
{
"epoch": 0.019756218764771714,
"grad_norm": 660.9871215820312,
"learning_rate": 9.78e-06,
"loss": 187.9075,
"step": 4890
},
{
"epoch": 0.019796620030139343,
"grad_norm": 689.9225463867188,
"learning_rate": 9.800000000000001e-06,
"loss": 311.5491,
"step": 4900
},
{
"epoch": 0.019837021295506976,
"grad_norm": 953.7089233398438,
"learning_rate": 9.820000000000001e-06,
"loss": 299.3845,
"step": 4910
},
{
"epoch": 0.019877422560874606,
"grad_norm": 614.9231567382812,
"learning_rate": 9.84e-06,
"loss": 168.2812,
"step": 4920
},
{
"epoch": 0.01991782382624224,
"grad_norm": 1575.7044677734375,
"learning_rate": 9.86e-06,
"loss": 273.8306,
"step": 4930
},
{
"epoch": 0.01995822509160987,
"grad_norm": 826.9859619140625,
"learning_rate": 9.88e-06,
"loss": 277.1096,
"step": 4940
},
{
"epoch": 0.0199986263569775,
"grad_norm": 934.8515625,
"learning_rate": 9.900000000000002e-06,
"loss": 230.0742,
"step": 4950
},
{
"epoch": 0.02003902762234513,
"grad_norm": 2514.587646484375,
"learning_rate": 9.92e-06,
"loss": 143.4857,
"step": 4960
},
{
"epoch": 0.020079428887712764,
"grad_norm": 0.0,
"learning_rate": 9.940000000000001e-06,
"loss": 292.2021,
"step": 4970
},
{
"epoch": 0.020119830153080394,
"grad_norm": 1051.95068359375,
"learning_rate": 9.96e-06,
"loss": 243.0733,
"step": 4980
},
{
"epoch": 0.020160231418448027,
"grad_norm": 522.6783447265625,
"learning_rate": 9.980000000000001e-06,
"loss": 255.3195,
"step": 4990
},
{
"epoch": 0.020200632683815656,
"grad_norm": 1430.8314208984375,
"learning_rate": 1e-05,
"loss": 181.9884,
"step": 5000
},
{
"epoch": 0.02024103394918329,
"grad_norm": 2610.226806640625,
"learning_rate": 1.002e-05,
"loss": 240.1806,
"step": 5010
},
{
"epoch": 0.02028143521455092,
"grad_norm": 706.251220703125,
"learning_rate": 1.004e-05,
"loss": 154.691,
"step": 5020
},
{
"epoch": 0.020321836479918552,
"grad_norm": 1110.47021484375,
"learning_rate": 1.006e-05,
"loss": 270.7392,
"step": 5030
},
{
"epoch": 0.020362237745286182,
"grad_norm": 1037.9814453125,
"learning_rate": 1.008e-05,
"loss": 260.6087,
"step": 5040
},
{
"epoch": 0.020402639010653815,
"grad_norm": 988.474609375,
"learning_rate": 1.0100000000000002e-05,
"loss": 195.7535,
"step": 5050
},
{
"epoch": 0.020443040276021444,
"grad_norm": 0.0,
"learning_rate": 1.012e-05,
"loss": 292.6674,
"step": 5060
},
{
"epoch": 0.020483441541389077,
"grad_norm": 1622.9769287109375,
"learning_rate": 1.0140000000000001e-05,
"loss": 226.0767,
"step": 5070
},
{
"epoch": 0.020523842806756707,
"grad_norm": 2696.41845703125,
"learning_rate": 1.016e-05,
"loss": 242.0753,
"step": 5080
},
{
"epoch": 0.02056424407212434,
"grad_norm": 2494.242431640625,
"learning_rate": 1.018e-05,
"loss": 310.918,
"step": 5090
},
{
"epoch": 0.02060464533749197,
"grad_norm": 549.554443359375,
"learning_rate": 1.02e-05,
"loss": 238.1347,
"step": 5100
},
{
"epoch": 0.020645046602859603,
"grad_norm": 753.8233642578125,
"learning_rate": 1.022e-05,
"loss": 271.2417,
"step": 5110
},
{
"epoch": 0.020685447868227232,
"grad_norm": 905.9545288085938,
"learning_rate": 1.024e-05,
"loss": 204.41,
"step": 5120
},
{
"epoch": 0.020725849133594865,
"grad_norm": 1900.184326171875,
"learning_rate": 1.026e-05,
"loss": 203.9288,
"step": 5130
},
{
"epoch": 0.020766250398962495,
"grad_norm": 533.8387451171875,
"learning_rate": 1.0280000000000002e-05,
"loss": 186.7773,
"step": 5140
},
{
"epoch": 0.020806651664330128,
"grad_norm": 671.30322265625,
"learning_rate": 1.03e-05,
"loss": 164.2021,
"step": 5150
},
{
"epoch": 0.020847052929697758,
"grad_norm": 764.4326171875,
"learning_rate": 1.0320000000000001e-05,
"loss": 234.4929,
"step": 5160
},
{
"epoch": 0.02088745419506539,
"grad_norm": 921.5126953125,
"learning_rate": 1.0340000000000001e-05,
"loss": 257.7942,
"step": 5170
},
{
"epoch": 0.02092785546043302,
"grad_norm": 1028.72412109375,
"learning_rate": 1.036e-05,
"loss": 219.407,
"step": 5180
},
{
"epoch": 0.020968256725800653,
"grad_norm": 857.1285400390625,
"learning_rate": 1.038e-05,
"loss": 267.7374,
"step": 5190
},
{
"epoch": 0.021008657991168283,
"grad_norm": 733.449951171875,
"learning_rate": 1.04e-05,
"loss": 259.0882,
"step": 5200
},
{
"epoch": 0.021049059256535916,
"grad_norm": 853.8319091796875,
"learning_rate": 1.042e-05,
"loss": 236.0859,
"step": 5210
},
{
"epoch": 0.021089460521903546,
"grad_norm": 780.4057006835938,
"learning_rate": 1.0440000000000002e-05,
"loss": 235.1703,
"step": 5220
},
{
"epoch": 0.02112986178727118,
"grad_norm": 1369.0264892578125,
"learning_rate": 1.046e-05,
"loss": 262.7009,
"step": 5230
},
{
"epoch": 0.021170263052638808,
"grad_norm": 802.2042236328125,
"learning_rate": 1.0480000000000001e-05,
"loss": 317.5121,
"step": 5240
},
{
"epoch": 0.02121066431800644,
"grad_norm": 3069.8564453125,
"learning_rate": 1.05e-05,
"loss": 266.7954,
"step": 5250
},
{
"epoch": 0.02125106558337407,
"grad_norm": 1437.6041259765625,
"learning_rate": 1.0520000000000001e-05,
"loss": 290.8906,
"step": 5260
},
{
"epoch": 0.021291466848741704,
"grad_norm": 1116.5682373046875,
"learning_rate": 1.0539999999999999e-05,
"loss": 217.5333,
"step": 5270
},
{
"epoch": 0.021331868114109333,
"grad_norm": 674.6525268554688,
"learning_rate": 1.056e-05,
"loss": 262.4274,
"step": 5280
},
{
"epoch": 0.021372269379476967,
"grad_norm": 980.0327758789062,
"learning_rate": 1.058e-05,
"loss": 253.3151,
"step": 5290
},
{
"epoch": 0.021412670644844596,
"grad_norm": 1546.599609375,
"learning_rate": 1.06e-05,
"loss": 284.0318,
"step": 5300
},
{
"epoch": 0.02145307191021223,
"grad_norm": 4072.65478515625,
"learning_rate": 1.062e-05,
"loss": 299.3221,
"step": 5310
},
{
"epoch": 0.02149347317557986,
"grad_norm": 835.0826416015625,
"learning_rate": 1.064e-05,
"loss": 217.7071,
"step": 5320
},
{
"epoch": 0.021533874440947492,
"grad_norm": 825.3660278320312,
"learning_rate": 1.0660000000000001e-05,
"loss": 250.3456,
"step": 5330
},
{
"epoch": 0.02157427570631512,
"grad_norm": 1196.41455078125,
"learning_rate": 1.0680000000000001e-05,
"loss": 217.1434,
"step": 5340
},
{
"epoch": 0.021614676971682754,
"grad_norm": 556.876220703125,
"learning_rate": 1.0700000000000001e-05,
"loss": 239.3144,
"step": 5350
},
{
"epoch": 0.021655078237050384,
"grad_norm": 602.3602905273438,
"learning_rate": 1.072e-05,
"loss": 308.0275,
"step": 5360
},
{
"epoch": 0.021695479502418017,
"grad_norm": 747.0403442382812,
"learning_rate": 1.074e-05,
"loss": 249.5218,
"step": 5370
},
{
"epoch": 0.021735880767785647,
"grad_norm": 2065.047607421875,
"learning_rate": 1.076e-05,
"loss": 289.0896,
"step": 5380
},
{
"epoch": 0.02177628203315328,
"grad_norm": 1931.2476806640625,
"learning_rate": 1.0780000000000002e-05,
"loss": 270.5146,
"step": 5390
},
{
"epoch": 0.02181668329852091,
"grad_norm": 5352.14013671875,
"learning_rate": 1.08e-05,
"loss": 330.6276,
"step": 5400
},
{
"epoch": 0.021857084563888542,
"grad_norm": 750.8521728515625,
"learning_rate": 1.0820000000000001e-05,
"loss": 260.1782,
"step": 5410
},
{
"epoch": 0.021897485829256172,
"grad_norm": 1566.5867919921875,
"learning_rate": 1.084e-05,
"loss": 277.7564,
"step": 5420
},
{
"epoch": 0.021937887094623805,
"grad_norm": 4719.2412109375,
"learning_rate": 1.0860000000000001e-05,
"loss": 270.119,
"step": 5430
},
{
"epoch": 0.021978288359991435,
"grad_norm": 886.023681640625,
"learning_rate": 1.088e-05,
"loss": 246.386,
"step": 5440
},
{
"epoch": 0.022018689625359068,
"grad_norm": 15289.64453125,
"learning_rate": 1.09e-05,
"loss": 360.9848,
"step": 5450
},
{
"epoch": 0.022059090890726697,
"grad_norm": 854.2662353515625,
"learning_rate": 1.092e-05,
"loss": 234.6689,
"step": 5460
},
{
"epoch": 0.02209949215609433,
"grad_norm": 1195.251953125,
"learning_rate": 1.094e-05,
"loss": 233.2883,
"step": 5470
},
{
"epoch": 0.02213989342146196,
"grad_norm": 2974.2724609375,
"learning_rate": 1.096e-05,
"loss": 256.327,
"step": 5480
},
{
"epoch": 0.022180294686829593,
"grad_norm": 1651.259765625,
"learning_rate": 1.098e-05,
"loss": 258.9452,
"step": 5490
},
{
"epoch": 0.022220695952197222,
"grad_norm": 1866.4769287109375,
"learning_rate": 1.1000000000000001e-05,
"loss": 277.225,
"step": 5500
},
{
"epoch": 0.022261097217564856,
"grad_norm": 2012.682861328125,
"learning_rate": 1.1020000000000001e-05,
"loss": 213.3877,
"step": 5510
},
{
"epoch": 0.022301498482932485,
"grad_norm": 787.662353515625,
"learning_rate": 1.1040000000000001e-05,
"loss": 223.892,
"step": 5520
},
{
"epoch": 0.022341899748300118,
"grad_norm": 9214.21484375,
"learning_rate": 1.106e-05,
"loss": 217.7184,
"step": 5530
},
{
"epoch": 0.022382301013667748,
"grad_norm": 1571.4190673828125,
"learning_rate": 1.108e-05,
"loss": 268.1608,
"step": 5540
},
{
"epoch": 0.02242270227903538,
"grad_norm": 1564.1744384765625,
"learning_rate": 1.11e-05,
"loss": 259.3926,
"step": 5550
},
{
"epoch": 0.02246310354440301,
"grad_norm": 676.7450561523438,
"learning_rate": 1.112e-05,
"loss": 205.3203,
"step": 5560
},
{
"epoch": 0.022503504809770643,
"grad_norm": 1126.58837890625,
"learning_rate": 1.114e-05,
"loss": 127.3151,
"step": 5570
},
{
"epoch": 0.022543906075138273,
"grad_norm": 3336.22607421875,
"learning_rate": 1.1160000000000002e-05,
"loss": 226.2071,
"step": 5580
},
{
"epoch": 0.022584307340505906,
"grad_norm": 416.8356018066406,
"learning_rate": 1.118e-05,
"loss": 257.1062,
"step": 5590
},
{
"epoch": 0.022624708605873536,
"grad_norm": 653.0436401367188,
"learning_rate": 1.1200000000000001e-05,
"loss": 198.1556,
"step": 5600
},
{
"epoch": 0.02266510987124117,
"grad_norm": 5374.44091796875,
"learning_rate": 1.122e-05,
"loss": 230.6045,
"step": 5610
},
{
"epoch": 0.0227055111366088,
"grad_norm": 1517.1585693359375,
"learning_rate": 1.124e-05,
"loss": 217.2467,
"step": 5620
},
{
"epoch": 0.02274591240197643,
"grad_norm": 1149.62646484375,
"learning_rate": 1.126e-05,
"loss": 294.9917,
"step": 5630
},
{
"epoch": 0.02278631366734406,
"grad_norm": 1075.8157958984375,
"learning_rate": 1.128e-05,
"loss": 273.2365,
"step": 5640
},
{
"epoch": 0.022826714932711694,
"grad_norm": 1706.23388671875,
"learning_rate": 1.13e-05,
"loss": 239.202,
"step": 5650
},
{
"epoch": 0.022867116198079324,
"grad_norm": 939.1124877929688,
"learning_rate": 1.132e-05,
"loss": 262.5748,
"step": 5660
},
{
"epoch": 0.022907517463446957,
"grad_norm": 1178.756591796875,
"learning_rate": 1.134e-05,
"loss": 205.6744,
"step": 5670
},
{
"epoch": 0.022947918728814586,
"grad_norm": 402.2773132324219,
"learning_rate": 1.1360000000000001e-05,
"loss": 192.2235,
"step": 5680
},
{
"epoch": 0.02298831999418222,
"grad_norm": 0.0,
"learning_rate": 1.1380000000000001e-05,
"loss": 258.3543,
"step": 5690
},
{
"epoch": 0.02302872125954985,
"grad_norm": 1016.2149658203125,
"learning_rate": 1.1400000000000001e-05,
"loss": 217.5345,
"step": 5700
},
{
"epoch": 0.023069122524917482,
"grad_norm": 1061.48828125,
"learning_rate": 1.142e-05,
"loss": 227.8594,
"step": 5710
},
{
"epoch": 0.02310952379028511,
"grad_norm": 481.3313293457031,
"learning_rate": 1.144e-05,
"loss": 194.8761,
"step": 5720
},
{
"epoch": 0.023149925055652745,
"grad_norm": 782.8972778320312,
"learning_rate": 1.146e-05,
"loss": 173.4156,
"step": 5730
},
{
"epoch": 0.023190326321020374,
"grad_norm": 672.4293212890625,
"learning_rate": 1.148e-05,
"loss": 206.9097,
"step": 5740
},
{
"epoch": 0.023230727586388007,
"grad_norm": 2217.28369140625,
"learning_rate": 1.1500000000000002e-05,
"loss": 238.6014,
"step": 5750
},
{
"epoch": 0.023271128851755637,
"grad_norm": 3414.3115234375,
"learning_rate": 1.152e-05,
"loss": 324.1321,
"step": 5760
},
{
"epoch": 0.02331153011712327,
"grad_norm": 864.6550903320312,
"learning_rate": 1.1540000000000001e-05,
"loss": 213.1406,
"step": 5770
},
{
"epoch": 0.0233519313824909,
"grad_norm": 752.6487426757812,
"learning_rate": 1.156e-05,
"loss": 136.1037,
"step": 5780
},
{
"epoch": 0.023392332647858533,
"grad_norm": 550.0790405273438,
"learning_rate": 1.1580000000000001e-05,
"loss": 295.7111,
"step": 5790
},
{
"epoch": 0.023432733913226162,
"grad_norm": 918.3283081054688,
"learning_rate": 1.16e-05,
"loss": 234.3793,
"step": 5800
},
{
"epoch": 0.023473135178593795,
"grad_norm": 1468.6207275390625,
"learning_rate": 1.162e-05,
"loss": 191.9201,
"step": 5810
},
{
"epoch": 0.023513536443961425,
"grad_norm": 1899.0367431640625,
"learning_rate": 1.164e-05,
"loss": 234.0571,
"step": 5820
},
{
"epoch": 0.023553937709329058,
"grad_norm": 684.876953125,
"learning_rate": 1.166e-05,
"loss": 180.286,
"step": 5830
},
{
"epoch": 0.023594338974696687,
"grad_norm": 1083.017578125,
"learning_rate": 1.168e-05,
"loss": 300.8844,
"step": 5840
},
{
"epoch": 0.02363474024006432,
"grad_norm": 1754.2984619140625,
"learning_rate": 1.1700000000000001e-05,
"loss": 391.4959,
"step": 5850
},
{
"epoch": 0.02367514150543195,
"grad_norm": 945.9032592773438,
"learning_rate": 1.172e-05,
"loss": 272.7161,
"step": 5860
},
{
"epoch": 0.023715542770799583,
"grad_norm": 1564.7225341796875,
"learning_rate": 1.1740000000000001e-05,
"loss": 233.5665,
"step": 5870
},
{
"epoch": 0.023755944036167213,
"grad_norm": 523.8766479492188,
"learning_rate": 1.1760000000000001e-05,
"loss": 204.2223,
"step": 5880
},
{
"epoch": 0.023796345301534846,
"grad_norm": 1801.8782958984375,
"learning_rate": 1.178e-05,
"loss": 215.9124,
"step": 5890
},
{
"epoch": 0.023836746566902475,
"grad_norm": 2132.8076171875,
"learning_rate": 1.18e-05,
"loss": 328.9819,
"step": 5900
},
{
"epoch": 0.02387714783227011,
"grad_norm": 1254.7607421875,
"learning_rate": 1.182e-05,
"loss": 249.0257,
"step": 5910
},
{
"epoch": 0.023917549097637738,
"grad_norm": 873.1114501953125,
"learning_rate": 1.1840000000000002e-05,
"loss": 156.6504,
"step": 5920
},
{
"epoch": 0.02395795036300537,
"grad_norm": 2675.64892578125,
"learning_rate": 1.186e-05,
"loss": 213.5535,
"step": 5930
},
{
"epoch": 0.023998351628373,
"grad_norm": 744.88671875,
"learning_rate": 1.1880000000000001e-05,
"loss": 189.1653,
"step": 5940
},
{
"epoch": 0.024038752893740634,
"grad_norm": 979.5391845703125,
"learning_rate": 1.19e-05,
"loss": 223.2056,
"step": 5950
},
{
"epoch": 0.024079154159108263,
"grad_norm": 872.666015625,
"learning_rate": 1.1920000000000001e-05,
"loss": 194.5632,
"step": 5960
},
{
"epoch": 0.024119555424475896,
"grad_norm": 1239.266845703125,
"learning_rate": 1.1940000000000001e-05,
"loss": 221.0386,
"step": 5970
},
{
"epoch": 0.024159956689843526,
"grad_norm": 1718.065673828125,
"learning_rate": 1.196e-05,
"loss": 237.8773,
"step": 5980
},
{
"epoch": 0.02420035795521116,
"grad_norm": 1780.784423828125,
"learning_rate": 1.198e-05,
"loss": 119.0116,
"step": 5990
},
{
"epoch": 0.02424075922057879,
"grad_norm": 921.1470947265625,
"learning_rate": 1.2e-05,
"loss": 225.7212,
"step": 6000
},
{
"epoch": 0.02428116048594642,
"grad_norm": 2161.641845703125,
"learning_rate": 1.202e-05,
"loss": 181.6732,
"step": 6010
},
{
"epoch": 0.02432156175131405,
"grad_norm": 1104.2542724609375,
"learning_rate": 1.204e-05,
"loss": 187.8781,
"step": 6020
},
{
"epoch": 0.024361963016681684,
"grad_norm": 1214.7061767578125,
"learning_rate": 1.206e-05,
"loss": 221.4846,
"step": 6030
},
{
"epoch": 0.024402364282049314,
"grad_norm": 1222.162841796875,
"learning_rate": 1.2080000000000001e-05,
"loss": 172.2742,
"step": 6040
},
{
"epoch": 0.024442765547416947,
"grad_norm": 828.8873291015625,
"learning_rate": 1.2100000000000001e-05,
"loss": 227.2261,
"step": 6050
},
{
"epoch": 0.024483166812784576,
"grad_norm": 621.658447265625,
"learning_rate": 1.2120000000000001e-05,
"loss": 236.3511,
"step": 6060
},
{
"epoch": 0.02452356807815221,
"grad_norm": 1036.8626708984375,
"learning_rate": 1.214e-05,
"loss": 233.8868,
"step": 6070
},
{
"epoch": 0.02456396934351984,
"grad_norm": 1811.7554931640625,
"learning_rate": 1.216e-05,
"loss": 238.5441,
"step": 6080
},
{
"epoch": 0.02460437060888747,
"grad_norm": 1012.0717163085938,
"learning_rate": 1.2180000000000002e-05,
"loss": 234.7135,
"step": 6090
},
{
"epoch": 0.0246447718742551,
"grad_norm": 1480.242919921875,
"learning_rate": 1.22e-05,
"loss": 275.5092,
"step": 6100
},
{
"epoch": 0.02468517313962273,
"grad_norm": 1884.02001953125,
"learning_rate": 1.2220000000000002e-05,
"loss": 242.7474,
"step": 6110
},
{
"epoch": 0.024725574404990364,
"grad_norm": 1144.38916015625,
"learning_rate": 1.224e-05,
"loss": 368.5085,
"step": 6120
},
{
"epoch": 0.024765975670357994,
"grad_norm": 1017.46875,
"learning_rate": 1.2260000000000001e-05,
"loss": 224.2549,
"step": 6130
},
{
"epoch": 0.024806376935725627,
"grad_norm": 1646.821044921875,
"learning_rate": 1.2280000000000001e-05,
"loss": 221.9995,
"step": 6140
},
{
"epoch": 0.024846778201093257,
"grad_norm": 719.86376953125,
"learning_rate": 1.23e-05,
"loss": 221.8839,
"step": 6150
},
{
"epoch": 0.02488717946646089,
"grad_norm": 865.7079467773438,
"learning_rate": 1.232e-05,
"loss": 216.4899,
"step": 6160
},
{
"epoch": 0.02492758073182852,
"grad_norm": 837.6893310546875,
"learning_rate": 1.234e-05,
"loss": 212.6082,
"step": 6170
},
{
"epoch": 0.024967981997196152,
"grad_norm": 619.9283447265625,
"learning_rate": 1.236e-05,
"loss": 231.417,
"step": 6180
},
{
"epoch": 0.025008383262563782,
"grad_norm": 938.736328125,
"learning_rate": 1.238e-05,
"loss": 222.0059,
"step": 6190
},
{
"epoch": 0.025048784527931415,
"grad_norm": 768.3204956054688,
"learning_rate": 1.24e-05,
"loss": 269.6617,
"step": 6200
},
{
"epoch": 0.025089185793299044,
"grad_norm": 25211.0859375,
"learning_rate": 1.2420000000000001e-05,
"loss": 283.9776,
"step": 6210
},
{
"epoch": 0.025129587058666678,
"grad_norm": 1000.73046875,
"learning_rate": 1.244e-05,
"loss": 237.4733,
"step": 6220
},
{
"epoch": 0.025169988324034307,
"grad_norm": 1383.859375,
"learning_rate": 1.2460000000000001e-05,
"loss": 207.5625,
"step": 6230
},
{
"epoch": 0.02521038958940194,
"grad_norm": 1157.95654296875,
"learning_rate": 1.248e-05,
"loss": 188.4448,
"step": 6240
},
{
"epoch": 0.02525079085476957,
"grad_norm": 437.2555847167969,
"learning_rate": 1.25e-05,
"loss": 253.4308,
"step": 6250
},
{
"epoch": 0.025291192120137203,
"grad_norm": 735.6442260742188,
"learning_rate": 1.252e-05,
"loss": 239.4786,
"step": 6260
},
{
"epoch": 0.025331593385504832,
"grad_norm": 3905.33935546875,
"learning_rate": 1.2540000000000002e-05,
"loss": 340.7768,
"step": 6270
},
{
"epoch": 0.025371994650872465,
"grad_norm": 1644.240234375,
"learning_rate": 1.256e-05,
"loss": 221.2143,
"step": 6280
},
{
"epoch": 0.025412395916240095,
"grad_norm": 745.5984497070312,
"learning_rate": 1.258e-05,
"loss": 154.9017,
"step": 6290
},
{
"epoch": 0.025452797181607728,
"grad_norm": 2822.624267578125,
"learning_rate": 1.2600000000000001e-05,
"loss": 193.9914,
"step": 6300
},
{
"epoch": 0.025493198446975358,
"grad_norm": 415.230712890625,
"learning_rate": 1.2620000000000001e-05,
"loss": 150.345,
"step": 6310
},
{
"epoch": 0.02553359971234299,
"grad_norm": 2347.072021484375,
"learning_rate": 1.2640000000000003e-05,
"loss": 239.6525,
"step": 6320
},
{
"epoch": 0.02557400097771062,
"grad_norm": 1561.8658447265625,
"learning_rate": 1.2659999999999999e-05,
"loss": 288.6605,
"step": 6330
},
{
"epoch": 0.025614402243078253,
"grad_norm": 651.1387329101562,
"learning_rate": 1.268e-05,
"loss": 266.1772,
"step": 6340
},
{
"epoch": 0.025654803508445883,
"grad_norm": 544.3804931640625,
"learning_rate": 1.27e-05,
"loss": 273.7383,
"step": 6350
},
{
"epoch": 0.025695204773813516,
"grad_norm": 1525.947021484375,
"learning_rate": 1.2720000000000002e-05,
"loss": 251.1667,
"step": 6360
},
{
"epoch": 0.025735606039181146,
"grad_norm": 1018.3317260742188,
"learning_rate": 1.2740000000000002e-05,
"loss": 210.4547,
"step": 6370
},
{
"epoch": 0.02577600730454878,
"grad_norm": 911.9342651367188,
"learning_rate": 1.276e-05,
"loss": 278.7565,
"step": 6380
},
{
"epoch": 0.025816408569916408,
"grad_norm": 3710.921630859375,
"learning_rate": 1.278e-05,
"loss": 229.9252,
"step": 6390
},
{
"epoch": 0.02585680983528404,
"grad_norm": 709.6052856445312,
"learning_rate": 1.2800000000000001e-05,
"loss": 197.896,
"step": 6400
},
{
"epoch": 0.02589721110065167,
"grad_norm": 688.3367919921875,
"learning_rate": 1.2820000000000001e-05,
"loss": 205.0635,
"step": 6410
},
{
"epoch": 0.025937612366019304,
"grad_norm": 1232.222412109375,
"learning_rate": 1.2839999999999999e-05,
"loss": 196.4835,
"step": 6420
},
{
"epoch": 0.025978013631386934,
"grad_norm": 665.0555419921875,
"learning_rate": 1.286e-05,
"loss": 211.4123,
"step": 6430
},
{
"epoch": 0.026018414896754567,
"grad_norm": 2699.49755859375,
"learning_rate": 1.288e-05,
"loss": 266.4242,
"step": 6440
},
{
"epoch": 0.026058816162122196,
"grad_norm": 1464.1005859375,
"learning_rate": 1.29e-05,
"loss": 271.645,
"step": 6450
},
{
"epoch": 0.02609921742748983,
"grad_norm": 605.3104858398438,
"learning_rate": 1.2920000000000002e-05,
"loss": 158.6754,
"step": 6460
},
{
"epoch": 0.02613961869285746,
"grad_norm": 643.6430053710938,
"learning_rate": 1.294e-05,
"loss": 246.8146,
"step": 6470
},
{
"epoch": 0.026180019958225092,
"grad_norm": 780.9293823242188,
"learning_rate": 1.296e-05,
"loss": 270.3225,
"step": 6480
},
{
"epoch": 0.02622042122359272,
"grad_norm": 2428.1328125,
"learning_rate": 1.2980000000000001e-05,
"loss": 218.6574,
"step": 6490
},
{
"epoch": 0.026260822488960354,
"grad_norm": 737.8549194335938,
"learning_rate": 1.3000000000000001e-05,
"loss": 241.2201,
"step": 6500
},
{
"epoch": 0.026301223754327984,
"grad_norm": 690.761474609375,
"learning_rate": 1.3020000000000002e-05,
"loss": 215.8646,
"step": 6510
},
{
"epoch": 0.026341625019695617,
"grad_norm": 693.7470092773438,
"learning_rate": 1.3039999999999999e-05,
"loss": 226.2056,
"step": 6520
},
{
"epoch": 0.026382026285063247,
"grad_norm": 1125.502197265625,
"learning_rate": 1.306e-05,
"loss": 250.0323,
"step": 6530
},
{
"epoch": 0.02642242755043088,
"grad_norm": 3172.219482421875,
"learning_rate": 1.308e-05,
"loss": 247.7785,
"step": 6540
},
{
"epoch": 0.02646282881579851,
"grad_norm": 1940.4234619140625,
"learning_rate": 1.3100000000000002e-05,
"loss": 178.4448,
"step": 6550
},
{
"epoch": 0.026503230081166142,
"grad_norm": 632.4578857421875,
"learning_rate": 1.3120000000000001e-05,
"loss": 206.7619,
"step": 6560
},
{
"epoch": 0.026543631346533772,
"grad_norm": 1491.89306640625,
"learning_rate": 1.314e-05,
"loss": 164.1459,
"step": 6570
},
{
"epoch": 0.026584032611901405,
"grad_norm": 831.048828125,
"learning_rate": 1.316e-05,
"loss": 202.0544,
"step": 6580
},
{
"epoch": 0.026624433877269035,
"grad_norm": 812.6109619140625,
"learning_rate": 1.3180000000000001e-05,
"loss": 127.2249,
"step": 6590
},
{
"epoch": 0.026664835142636668,
"grad_norm": 678.7022094726562,
"learning_rate": 1.32e-05,
"loss": 262.2538,
"step": 6600
},
{
"epoch": 0.026705236408004297,
"grad_norm": 1193.4014892578125,
"learning_rate": 1.3220000000000002e-05,
"loss": 258.1862,
"step": 6610
},
{
"epoch": 0.02674563767337193,
"grad_norm": 881.2777099609375,
"learning_rate": 1.324e-05,
"loss": 181.1106,
"step": 6620
},
{
"epoch": 0.02678603893873956,
"grad_norm": 751.0634765625,
"learning_rate": 1.326e-05,
"loss": 328.8563,
"step": 6630
},
{
"epoch": 0.026826440204107193,
"grad_norm": 810.8790893554688,
"learning_rate": 1.3280000000000002e-05,
"loss": 175.6053,
"step": 6640
},
{
"epoch": 0.026866841469474823,
"grad_norm": 544.2944946289062,
"learning_rate": 1.3300000000000001e-05,
"loss": 172.9304,
"step": 6650
},
{
"epoch": 0.026907242734842456,
"grad_norm": 1003.13818359375,
"learning_rate": 1.3320000000000001e-05,
"loss": 180.5678,
"step": 6660
},
{
"epoch": 0.026947644000210085,
"grad_norm": 1979.7833251953125,
"learning_rate": 1.334e-05,
"loss": 212.5934,
"step": 6670
},
{
"epoch": 0.026988045265577718,
"grad_norm": 660.824462890625,
"learning_rate": 1.336e-05,
"loss": 170.068,
"step": 6680
},
{
"epoch": 0.027028446530945348,
"grad_norm": 753.3848876953125,
"learning_rate": 1.338e-05,
"loss": 226.8918,
"step": 6690
},
{
"epoch": 0.02706884779631298,
"grad_norm": 1572.1107177734375,
"learning_rate": 1.3400000000000002e-05,
"loss": 285.1505,
"step": 6700
},
{
"epoch": 0.02710924906168061,
"grad_norm": 727.326416015625,
"learning_rate": 1.3420000000000002e-05,
"loss": 259.9044,
"step": 6710
},
{
"epoch": 0.027149650327048244,
"grad_norm": 1423.973388671875,
"learning_rate": 1.344e-05,
"loss": 208.3222,
"step": 6720
},
{
"epoch": 0.027190051592415873,
"grad_norm": 3651.573974609375,
"learning_rate": 1.346e-05,
"loss": 276.9352,
"step": 6730
},
{
"epoch": 0.027230452857783506,
"grad_norm": 986.3583984375,
"learning_rate": 1.3480000000000001e-05,
"loss": 233.0887,
"step": 6740
},
{
"epoch": 0.027270854123151136,
"grad_norm": 1490.3424072265625,
"learning_rate": 1.3500000000000001e-05,
"loss": 308.2051,
"step": 6750
},
{
"epoch": 0.02731125538851877,
"grad_norm": 1850.4970703125,
"learning_rate": 1.352e-05,
"loss": 220.4273,
"step": 6760
},
{
"epoch": 0.0273516566538864,
"grad_norm": 558.9097900390625,
"learning_rate": 1.3539999999999999e-05,
"loss": 192.5242,
"step": 6770
},
{
"epoch": 0.02739205791925403,
"grad_norm": 1742.455322265625,
"learning_rate": 1.356e-05,
"loss": 228.9638,
"step": 6780
},
{
"epoch": 0.02743245918462166,
"grad_norm": 6109.70947265625,
"learning_rate": 1.358e-05,
"loss": 218.1945,
"step": 6790
},
{
"epoch": 0.027472860449989294,
"grad_norm": 2294.8359375,
"learning_rate": 1.3600000000000002e-05,
"loss": 259.6226,
"step": 6800
},
{
"epoch": 0.027513261715356924,
"grad_norm": 1345.371826171875,
"learning_rate": 1.362e-05,
"loss": 208.2849,
"step": 6810
},
{
"epoch": 0.027553662980724557,
"grad_norm": 631.337158203125,
"learning_rate": 1.364e-05,
"loss": 174.1924,
"step": 6820
},
{
"epoch": 0.027594064246092186,
"grad_norm": 2324.395751953125,
"learning_rate": 1.3660000000000001e-05,
"loss": 217.651,
"step": 6830
},
{
"epoch": 0.02763446551145982,
"grad_norm": 1169.88916015625,
"learning_rate": 1.3680000000000001e-05,
"loss": 152.5202,
"step": 6840
},
{
"epoch": 0.02767486677682745,
"grad_norm": 580.0879516601562,
"learning_rate": 1.3700000000000001e-05,
"loss": 260.7849,
"step": 6850
},
{
"epoch": 0.027715268042195082,
"grad_norm": 1336.279052734375,
"learning_rate": 1.3719999999999999e-05,
"loss": 184.8877,
"step": 6860
},
{
"epoch": 0.02775566930756271,
"grad_norm": 1506.610107421875,
"learning_rate": 1.374e-05,
"loss": 236.7283,
"step": 6870
},
{
"epoch": 0.027796070572930345,
"grad_norm": 1100.15234375,
"learning_rate": 1.376e-05,
"loss": 263.8396,
"step": 6880
},
{
"epoch": 0.027836471838297974,
"grad_norm": 794.3699340820312,
"learning_rate": 1.3780000000000002e-05,
"loss": 290.786,
"step": 6890
},
{
"epoch": 0.027876873103665607,
"grad_norm": 2009.521240234375,
"learning_rate": 1.3800000000000002e-05,
"loss": 173.6581,
"step": 6900
},
{
"epoch": 0.027917274369033237,
"grad_norm": 1658.059814453125,
"learning_rate": 1.382e-05,
"loss": 210.0458,
"step": 6910
},
{
"epoch": 0.02795767563440087,
"grad_norm": 632.90869140625,
"learning_rate": 1.384e-05,
"loss": 179.4226,
"step": 6920
},
{
"epoch": 0.0279980768997685,
"grad_norm": 1309.6959228515625,
"learning_rate": 1.3860000000000001e-05,
"loss": 168.1927,
"step": 6930
},
{
"epoch": 0.028038478165136133,
"grad_norm": 1135.08935546875,
"learning_rate": 1.3880000000000001e-05,
"loss": 189.0599,
"step": 6940
},
{
"epoch": 0.028078879430503762,
"grad_norm": 642.1088256835938,
"learning_rate": 1.3900000000000002e-05,
"loss": 191.8304,
"step": 6950
},
{
"epoch": 0.028119280695871395,
"grad_norm": 1067.7398681640625,
"learning_rate": 1.3919999999999999e-05,
"loss": 153.804,
"step": 6960
},
{
"epoch": 0.028159681961239025,
"grad_norm": 0.0,
"learning_rate": 1.394e-05,
"loss": 131.2499,
"step": 6970
},
{
"epoch": 0.028200083226606658,
"grad_norm": 4239.3876953125,
"learning_rate": 1.396e-05,
"loss": 219.0053,
"step": 6980
},
{
"epoch": 0.028240484491974287,
"grad_norm": 1219.2093505859375,
"learning_rate": 1.3980000000000002e-05,
"loss": 216.1493,
"step": 6990
},
{
"epoch": 0.02828088575734192,
"grad_norm": 829.5226440429688,
"learning_rate": 1.4000000000000001e-05,
"loss": 197.1629,
"step": 7000
},
{
"epoch": 0.02832128702270955,
"grad_norm": 783.1107177734375,
"learning_rate": 1.402e-05,
"loss": 197.8331,
"step": 7010
},
{
"epoch": 0.028361688288077183,
"grad_norm": 1040.990966796875,
"learning_rate": 1.4040000000000001e-05,
"loss": 234.4111,
"step": 7020
},
{
"epoch": 0.028402089553444813,
"grad_norm": 952.650390625,
"learning_rate": 1.4060000000000001e-05,
"loss": 138.1043,
"step": 7030
},
{
"epoch": 0.028442490818812446,
"grad_norm": 2120.21337890625,
"learning_rate": 1.408e-05,
"loss": 190.9683,
"step": 7040
},
{
"epoch": 0.028482892084180075,
"grad_norm": 1032.8970947265625,
"learning_rate": 1.4099999999999999e-05,
"loss": 211.108,
"step": 7050
},
{
"epoch": 0.02852329334954771,
"grad_norm": 845.619384765625,
"learning_rate": 1.412e-05,
"loss": 249.2893,
"step": 7060
},
{
"epoch": 0.028563694614915338,
"grad_norm": 2072.714599609375,
"learning_rate": 1.414e-05,
"loss": 148.9613,
"step": 7070
},
{
"epoch": 0.02860409588028297,
"grad_norm": 1205.255615234375,
"learning_rate": 1.4160000000000002e-05,
"loss": 208.2003,
"step": 7080
},
{
"epoch": 0.0286444971456506,
"grad_norm": 1514.6177978515625,
"learning_rate": 1.4180000000000001e-05,
"loss": 211.4245,
"step": 7090
},
{
"epoch": 0.028684898411018234,
"grad_norm": 599.4893188476562,
"learning_rate": 1.42e-05,
"loss": 270.9207,
"step": 7100
},
{
"epoch": 0.028725299676385863,
"grad_norm": 622.9778442382812,
"learning_rate": 1.422e-05,
"loss": 248.9024,
"step": 7110
},
{
"epoch": 0.028765700941753496,
"grad_norm": 827.0371704101562,
"learning_rate": 1.4240000000000001e-05,
"loss": 228.1889,
"step": 7120
},
{
"epoch": 0.028806102207121126,
"grad_norm": 1157.1502685546875,
"learning_rate": 1.426e-05,
"loss": 181.2518,
"step": 7130
},
{
"epoch": 0.02884650347248876,
"grad_norm": 1035.8681640625,
"learning_rate": 1.4280000000000002e-05,
"loss": 223.2644,
"step": 7140
},
{
"epoch": 0.02888690473785639,
"grad_norm": 1198.443359375,
"learning_rate": 1.43e-05,
"loss": 193.8474,
"step": 7150
},
{
"epoch": 0.02892730600322402,
"grad_norm": 1005.276611328125,
"learning_rate": 1.432e-05,
"loss": 250.6531,
"step": 7160
},
{
"epoch": 0.02896770726859165,
"grad_norm": 2052.37158203125,
"learning_rate": 1.434e-05,
"loss": 184.0699,
"step": 7170
},
{
"epoch": 0.029008108533959284,
"grad_norm": 22271.3046875,
"learning_rate": 1.4360000000000001e-05,
"loss": 298.0969,
"step": 7180
},
{
"epoch": 0.029048509799326914,
"grad_norm": 1663.1221923828125,
"learning_rate": 1.4380000000000001e-05,
"loss": 256.0283,
"step": 7190
},
{
"epoch": 0.029088911064694547,
"grad_norm": 5215.501953125,
"learning_rate": 1.44e-05,
"loss": 208.372,
"step": 7200
},
{
"epoch": 0.029129312330062176,
"grad_norm": 1427.092529296875,
"learning_rate": 1.4420000000000001e-05,
"loss": 196.7174,
"step": 7210
},
{
"epoch": 0.02916971359542981,
"grad_norm": 755.31494140625,
"learning_rate": 1.444e-05,
"loss": 207.5049,
"step": 7220
},
{
"epoch": 0.02921011486079744,
"grad_norm": 1840.75927734375,
"learning_rate": 1.4460000000000002e-05,
"loss": 246.7446,
"step": 7230
},
{
"epoch": 0.029250516126165072,
"grad_norm": 1352.86767578125,
"learning_rate": 1.4480000000000002e-05,
"loss": 193.7286,
"step": 7240
},
{
"epoch": 0.029290917391532702,
"grad_norm": 1088.9520263671875,
"learning_rate": 1.45e-05,
"loss": 213.7931,
"step": 7250
},
{
"epoch": 0.029331318656900335,
"grad_norm": 2126.0322265625,
"learning_rate": 1.452e-05,
"loss": 227.5255,
"step": 7260
},
{
"epoch": 0.029371719922267964,
"grad_norm": 0.0,
"learning_rate": 1.4540000000000001e-05,
"loss": 149.3132,
"step": 7270
},
{
"epoch": 0.029412121187635597,
"grad_norm": 582.3840942382812,
"learning_rate": 1.4560000000000001e-05,
"loss": 219.7635,
"step": 7280
},
{
"epoch": 0.029452522453003227,
"grad_norm": 935.360107421875,
"learning_rate": 1.4580000000000003e-05,
"loss": 219.2881,
"step": 7290
},
{
"epoch": 0.02949292371837086,
"grad_norm": 2865.860595703125,
"learning_rate": 1.4599999999999999e-05,
"loss": 170.7084,
"step": 7300
},
{
"epoch": 0.02953332498373849,
"grad_norm": 1290.148681640625,
"learning_rate": 1.462e-05,
"loss": 273.1594,
"step": 7310
},
{
"epoch": 0.029573726249106123,
"grad_norm": 1073.5101318359375,
"learning_rate": 1.464e-05,
"loss": 276.2752,
"step": 7320
},
{
"epoch": 0.029614127514473752,
"grad_norm": 2523.898681640625,
"learning_rate": 1.4660000000000002e-05,
"loss": 229.5607,
"step": 7330
},
{
"epoch": 0.029654528779841385,
"grad_norm": 2364.947021484375,
"learning_rate": 1.4680000000000002e-05,
"loss": 194.6274,
"step": 7340
},
{
"epoch": 0.029694930045209015,
"grad_norm": 1139.0018310546875,
"learning_rate": 1.47e-05,
"loss": 143.1356,
"step": 7350
},
{
"epoch": 0.029735331310576648,
"grad_norm": 1392.792236328125,
"learning_rate": 1.472e-05,
"loss": 220.7791,
"step": 7360
},
{
"epoch": 0.029775732575944278,
"grad_norm": 1605.3038330078125,
"learning_rate": 1.4740000000000001e-05,
"loss": 159.6861,
"step": 7370
},
{
"epoch": 0.02981613384131191,
"grad_norm": 1095.6702880859375,
"learning_rate": 1.4760000000000001e-05,
"loss": 190.0395,
"step": 7380
},
{
"epoch": 0.02985653510667954,
"grad_norm": 3128.165283203125,
"learning_rate": 1.4779999999999999e-05,
"loss": 217.6996,
"step": 7390
},
{
"epoch": 0.029896936372047173,
"grad_norm": 868.65966796875,
"learning_rate": 1.48e-05,
"loss": 303.5258,
"step": 7400
},
{
"epoch": 0.029937337637414803,
"grad_norm": 1097.5875244140625,
"learning_rate": 1.482e-05,
"loss": 208.767,
"step": 7410
},
{
"epoch": 0.029977738902782436,
"grad_norm": 1184.8807373046875,
"learning_rate": 1.4840000000000002e-05,
"loss": 212.3098,
"step": 7420
},
{
"epoch": 0.030018140168150065,
"grad_norm": 1545.9532470703125,
"learning_rate": 1.4860000000000002e-05,
"loss": 142.4259,
"step": 7430
},
{
"epoch": 0.0300585414335177,
"grad_norm": 1245.4107666015625,
"learning_rate": 1.488e-05,
"loss": 164.8042,
"step": 7440
},
{
"epoch": 0.030098942698885328,
"grad_norm": 577.9031982421875,
"learning_rate": 1.49e-05,
"loss": 216.6332,
"step": 7450
},
{
"epoch": 0.03013934396425296,
"grad_norm": 1043.52392578125,
"learning_rate": 1.4920000000000001e-05,
"loss": 247.0811,
"step": 7460
},
{
"epoch": 0.03017974522962059,
"grad_norm": 1351.240966796875,
"learning_rate": 1.4940000000000001e-05,
"loss": 198.2244,
"step": 7470
},
{
"epoch": 0.030220146494988224,
"grad_norm": 1484.0284423828125,
"learning_rate": 1.4960000000000002e-05,
"loss": 197.885,
"step": 7480
},
{
"epoch": 0.030260547760355853,
"grad_norm": 1378.754638671875,
"learning_rate": 1.4979999999999999e-05,
"loss": 201.7562,
"step": 7490
},
{
"epoch": 0.030300949025723486,
"grad_norm": 7344.61767578125,
"learning_rate": 1.5e-05,
"loss": 231.1991,
"step": 7500
},
{
"epoch": 0.030341350291091116,
"grad_norm": 3804.617919921875,
"learning_rate": 1.502e-05,
"loss": 310.0097,
"step": 7510
},
{
"epoch": 0.03038175155645875,
"grad_norm": 1033.6324462890625,
"learning_rate": 1.5040000000000002e-05,
"loss": 198.1659,
"step": 7520
},
{
"epoch": 0.03042215282182638,
"grad_norm": 653.5598754882812,
"learning_rate": 1.5060000000000001e-05,
"loss": 178.0301,
"step": 7530
},
{
"epoch": 0.030462554087194012,
"grad_norm": 1495.2060546875,
"learning_rate": 1.508e-05,
"loss": 250.4168,
"step": 7540
},
{
"epoch": 0.03050295535256164,
"grad_norm": 3768.185791015625,
"learning_rate": 1.51e-05,
"loss": 161.5591,
"step": 7550
},
{
"epoch": 0.030543356617929274,
"grad_norm": 966.1246337890625,
"learning_rate": 1.5120000000000001e-05,
"loss": 173.9351,
"step": 7560
},
{
"epoch": 0.030583757883296904,
"grad_norm": 1045.052490234375,
"learning_rate": 1.514e-05,
"loss": 140.5521,
"step": 7570
},
{
"epoch": 0.030624159148664537,
"grad_norm": 1348.0595703125,
"learning_rate": 1.5160000000000002e-05,
"loss": 318.5935,
"step": 7580
},
{
"epoch": 0.030664560414032167,
"grad_norm": 798.9227294921875,
"learning_rate": 1.518e-05,
"loss": 269.7896,
"step": 7590
},
{
"epoch": 0.0307049616793998,
"grad_norm": 1470.73583984375,
"learning_rate": 1.52e-05,
"loss": 282.1007,
"step": 7600
},
{
"epoch": 0.03074536294476743,
"grad_norm": 770.0698852539062,
"learning_rate": 1.5220000000000002e-05,
"loss": 262.2653,
"step": 7610
},
{
"epoch": 0.030785764210135062,
"grad_norm": 571.9855346679688,
"learning_rate": 1.5240000000000001e-05,
"loss": 192.3983,
"step": 7620
},
{
"epoch": 0.030826165475502692,
"grad_norm": 1054.919921875,
"learning_rate": 1.5260000000000003e-05,
"loss": 219.8108,
"step": 7630
},
{
"epoch": 0.030866566740870325,
"grad_norm": 556.494873046875,
"learning_rate": 1.528e-05,
"loss": 165.5094,
"step": 7640
},
{
"epoch": 0.030906968006237955,
"grad_norm": 1616.6512451171875,
"learning_rate": 1.53e-05,
"loss": 183.0181,
"step": 7650
},
{
"epoch": 0.030947369271605588,
"grad_norm": 678.8358764648438,
"learning_rate": 1.5320000000000002e-05,
"loss": 138.8231,
"step": 7660
},
{
"epoch": 0.030987770536973217,
"grad_norm": 817.9168090820312,
"learning_rate": 1.5340000000000002e-05,
"loss": 166.3831,
"step": 7670
},
{
"epoch": 0.03102817180234085,
"grad_norm": 1391.115966796875,
"learning_rate": 1.536e-05,
"loss": 265.5596,
"step": 7680
},
{
"epoch": 0.03106857306770848,
"grad_norm": 2713.052490234375,
"learning_rate": 1.538e-05,
"loss": 221.1131,
"step": 7690
},
{
"epoch": 0.031108974333076113,
"grad_norm": 1156.0615234375,
"learning_rate": 1.54e-05,
"loss": 188.7715,
"step": 7700
},
{
"epoch": 0.031149375598443742,
"grad_norm": 1744.91845703125,
"learning_rate": 1.542e-05,
"loss": 203.3742,
"step": 7710
},
{
"epoch": 0.031189776863811376,
"grad_norm": 1208.4666748046875,
"learning_rate": 1.544e-05,
"loss": 146.0077,
"step": 7720
},
{
"epoch": 0.031230178129179005,
"grad_norm": 643.0628662109375,
"learning_rate": 1.546e-05,
"loss": 195.1551,
"step": 7730
},
{
"epoch": 0.03127057939454664,
"grad_norm": 1256.745849609375,
"learning_rate": 1.548e-05,
"loss": 128.5509,
"step": 7740
},
{
"epoch": 0.03131098065991427,
"grad_norm": 1408.4251708984375,
"learning_rate": 1.55e-05,
"loss": 202.1482,
"step": 7750
},
{
"epoch": 0.0313513819252819,
"grad_norm": 2253.969970703125,
"learning_rate": 1.552e-05,
"loss": 258.4842,
"step": 7760
},
{
"epoch": 0.03139178319064953,
"grad_norm": 1037.827392578125,
"learning_rate": 1.554e-05,
"loss": 167.7397,
"step": 7770
},
{
"epoch": 0.03143218445601716,
"grad_norm": 1066.5501708984375,
"learning_rate": 1.556e-05,
"loss": 218.4733,
"step": 7780
},
{
"epoch": 0.031472585721384796,
"grad_norm": 1018.4628295898438,
"learning_rate": 1.558e-05,
"loss": 196.4911,
"step": 7790
},
{
"epoch": 0.03151298698675242,
"grad_norm": 1526.3201904296875,
"learning_rate": 1.56e-05,
"loss": 214.0186,
"step": 7800
},
{
"epoch": 0.031553388252120056,
"grad_norm": 829.8909301757812,
"learning_rate": 1.5620000000000003e-05,
"loss": 150.7465,
"step": 7810
},
{
"epoch": 0.03159378951748769,
"grad_norm": 1091.463623046875,
"learning_rate": 1.5640000000000003e-05,
"loss": 240.4906,
"step": 7820
},
{
"epoch": 0.03163419078285532,
"grad_norm": 470.00250244140625,
"learning_rate": 1.566e-05,
"loss": 159.8959,
"step": 7830
},
{
"epoch": 0.03167459204822295,
"grad_norm": 1311.1846923828125,
"learning_rate": 1.568e-05,
"loss": 205.6226,
"step": 7840
},
{
"epoch": 0.03171499331359058,
"grad_norm": 1698.6754150390625,
"learning_rate": 1.5700000000000002e-05,
"loss": 217.6704,
"step": 7850
},
{
"epoch": 0.031755394578958214,
"grad_norm": 1376.8931884765625,
"learning_rate": 1.5720000000000002e-05,
"loss": 172.5748,
"step": 7860
},
{
"epoch": 0.03179579584432585,
"grad_norm": 1433.7232666015625,
"learning_rate": 1.5740000000000002e-05,
"loss": 185.2836,
"step": 7870
},
{
"epoch": 0.03183619710969347,
"grad_norm": 4025.76220703125,
"learning_rate": 1.5759999999999998e-05,
"loss": 315.4204,
"step": 7880
},
{
"epoch": 0.031876598375061106,
"grad_norm": 853.8228759765625,
"learning_rate": 1.578e-05,
"loss": 153.3957,
"step": 7890
},
{
"epoch": 0.03191699964042874,
"grad_norm": 944.7000732421875,
"learning_rate": 1.58e-05,
"loss": 128.8845,
"step": 7900
},
{
"epoch": 0.03195740090579637,
"grad_norm": 643.9895629882812,
"learning_rate": 1.582e-05,
"loss": 223.8146,
"step": 7910
},
{
"epoch": 0.031997802171164,
"grad_norm": 793.1214599609375,
"learning_rate": 1.584e-05,
"loss": 231.9548,
"step": 7920
},
{
"epoch": 0.03203820343653163,
"grad_norm": 1593.3084716796875,
"learning_rate": 1.586e-05,
"loss": 222.1474,
"step": 7930
},
{
"epoch": 0.032078604701899265,
"grad_norm": 0.0,
"learning_rate": 1.588e-05,
"loss": 156.401,
"step": 7940
},
{
"epoch": 0.0321190059672669,
"grad_norm": 941.607177734375,
"learning_rate": 1.59e-05,
"loss": 239.5231,
"step": 7950
},
{
"epoch": 0.032159407232634524,
"grad_norm": 1561.1961669921875,
"learning_rate": 1.592e-05,
"loss": 269.0299,
"step": 7960
},
{
"epoch": 0.03219980849800216,
"grad_norm": 4084.514892578125,
"learning_rate": 1.594e-05,
"loss": 269.6239,
"step": 7970
},
{
"epoch": 0.03224020976336979,
"grad_norm": 1185.47314453125,
"learning_rate": 1.596e-05,
"loss": 217.7421,
"step": 7980
},
{
"epoch": 0.03228061102873742,
"grad_norm": 1212.916748046875,
"learning_rate": 1.598e-05,
"loss": 187.7411,
"step": 7990
},
{
"epoch": 0.03232101229410505,
"grad_norm": 658.2976684570312,
"learning_rate": 1.6000000000000003e-05,
"loss": 230.9428,
"step": 8000
},
{
"epoch": 0.03236141355947268,
"grad_norm": 1249.109130859375,
"learning_rate": 1.6020000000000002e-05,
"loss": 222.3688,
"step": 8010
},
{
"epoch": 0.032401814824840315,
"grad_norm": 3268.414306640625,
"learning_rate": 1.604e-05,
"loss": 330.1671,
"step": 8020
},
{
"epoch": 0.03244221609020795,
"grad_norm": 2235.10595703125,
"learning_rate": 1.606e-05,
"loss": 239.3141,
"step": 8030
},
{
"epoch": 0.032482617355575574,
"grad_norm": 461.5077819824219,
"learning_rate": 1.6080000000000002e-05,
"loss": 210.1967,
"step": 8040
},
{
"epoch": 0.03252301862094321,
"grad_norm": 754.9512939453125,
"learning_rate": 1.6100000000000002e-05,
"loss": 196.5837,
"step": 8050
},
{
"epoch": 0.03256341988631084,
"grad_norm": 1378.9425048828125,
"learning_rate": 1.612e-05,
"loss": 261.1862,
"step": 8060
},
{
"epoch": 0.03260382115167847,
"grad_norm": 12839.7548828125,
"learning_rate": 1.6139999999999998e-05,
"loss": 188.6295,
"step": 8070
},
{
"epoch": 0.0326442224170461,
"grad_norm": 1415.6925048828125,
"learning_rate": 1.616e-05,
"loss": 212.514,
"step": 8080
},
{
"epoch": 0.03268462368241373,
"grad_norm": 633.0122680664062,
"learning_rate": 1.618e-05,
"loss": 167.5946,
"step": 8090
},
{
"epoch": 0.032725024947781366,
"grad_norm": 1421.470947265625,
"learning_rate": 1.62e-05,
"loss": 141.4309,
"step": 8100
},
{
"epoch": 0.032765426213149,
"grad_norm": 7646.71875,
"learning_rate": 1.622e-05,
"loss": 228.34,
"step": 8110
},
{
"epoch": 0.032805827478516625,
"grad_norm": 1138.2156982421875,
"learning_rate": 1.624e-05,
"loss": 209.1626,
"step": 8120
},
{
"epoch": 0.03284622874388426,
"grad_norm": 649.6055908203125,
"learning_rate": 1.626e-05,
"loss": 204.4883,
"step": 8130
},
{
"epoch": 0.03288663000925189,
"grad_norm": 646.7942504882812,
"learning_rate": 1.628e-05,
"loss": 155.1816,
"step": 8140
},
{
"epoch": 0.032927031274619524,
"grad_norm": 632.5813598632812,
"learning_rate": 1.63e-05,
"loss": 144.9363,
"step": 8150
},
{
"epoch": 0.03296743253998715,
"grad_norm": 859.72900390625,
"learning_rate": 1.6320000000000003e-05,
"loss": 207.8656,
"step": 8160
},
{
"epoch": 0.03300783380535478,
"grad_norm": 661.2268676757812,
"learning_rate": 1.634e-05,
"loss": 156.5535,
"step": 8170
},
{
"epoch": 0.033048235070722416,
"grad_norm": 1062.429931640625,
"learning_rate": 1.636e-05,
"loss": 206.9411,
"step": 8180
},
{
"epoch": 0.03308863633609005,
"grad_norm": 1048.27294921875,
"learning_rate": 1.6380000000000002e-05,
"loss": 186.6454,
"step": 8190
},
{
"epoch": 0.033129037601457675,
"grad_norm": 1131.9554443359375,
"learning_rate": 1.6400000000000002e-05,
"loss": 255.3936,
"step": 8200
},
{
"epoch": 0.03316943886682531,
"grad_norm": 1163.3974609375,
"learning_rate": 1.6420000000000002e-05,
"loss": 188.8816,
"step": 8210
},
{
"epoch": 0.03320984013219294,
"grad_norm": 555.4834594726562,
"learning_rate": 1.644e-05,
"loss": 134.8741,
"step": 8220
},
{
"epoch": 0.033250241397560575,
"grad_norm": 837.699951171875,
"learning_rate": 1.646e-05,
"loss": 183.4674,
"step": 8230
},
{
"epoch": 0.0332906426629282,
"grad_norm": 731.1207275390625,
"learning_rate": 1.648e-05,
"loss": 117.9002,
"step": 8240
},
{
"epoch": 0.033331043928295834,
"grad_norm": 747.77880859375,
"learning_rate": 1.65e-05,
"loss": 222.1833,
"step": 8250
},
{
"epoch": 0.03337144519366347,
"grad_norm": 1067.3509521484375,
"learning_rate": 1.652e-05,
"loss": 186.3794,
"step": 8260
},
{
"epoch": 0.0334118464590311,
"grad_norm": 1084.0224609375,
"learning_rate": 1.654e-05,
"loss": 236.8639,
"step": 8270
},
{
"epoch": 0.033452247724398726,
"grad_norm": 1279.375732421875,
"learning_rate": 1.656e-05,
"loss": 259.652,
"step": 8280
},
{
"epoch": 0.03349264898976636,
"grad_norm": 947.9122314453125,
"learning_rate": 1.658e-05,
"loss": 233.7726,
"step": 8290
},
{
"epoch": 0.03353305025513399,
"grad_norm": 439.6305236816406,
"learning_rate": 1.66e-05,
"loss": 177.2815,
"step": 8300
},
{
"epoch": 0.033573451520501625,
"grad_norm": 815.047119140625,
"learning_rate": 1.662e-05,
"loss": 176.6726,
"step": 8310
},
{
"epoch": 0.03361385278586925,
"grad_norm": 6470.5361328125,
"learning_rate": 1.664e-05,
"loss": 222.3292,
"step": 8320
},
{
"epoch": 0.033654254051236884,
"grad_norm": 790.7086791992188,
"learning_rate": 1.666e-05,
"loss": 194.9121,
"step": 8330
},
{
"epoch": 0.03369465531660452,
"grad_norm": 794.2681884765625,
"learning_rate": 1.668e-05,
"loss": 175.7233,
"step": 8340
},
{
"epoch": 0.03373505658197215,
"grad_norm": 1229.4151611328125,
"learning_rate": 1.6700000000000003e-05,
"loss": 227.8069,
"step": 8350
},
{
"epoch": 0.033775457847339777,
"grad_norm": 792.4901123046875,
"learning_rate": 1.672e-05,
"loss": 198.4519,
"step": 8360
},
{
"epoch": 0.03381585911270741,
"grad_norm": 1083.63623046875,
"learning_rate": 1.674e-05,
"loss": 164.7306,
"step": 8370
},
{
"epoch": 0.03385626037807504,
"grad_norm": 1077.84765625,
"learning_rate": 1.6760000000000002e-05,
"loss": 154.6505,
"step": 8380
},
{
"epoch": 0.033896661643442676,
"grad_norm": 2413.804931640625,
"learning_rate": 1.6780000000000002e-05,
"loss": 247.3373,
"step": 8390
},
{
"epoch": 0.0339370629088103,
"grad_norm": 2946.898681640625,
"learning_rate": 1.6800000000000002e-05,
"loss": 161.9533,
"step": 8400
},
{
"epoch": 0.033977464174177935,
"grad_norm": 1092.52392578125,
"learning_rate": 1.6819999999999998e-05,
"loss": 216.0875,
"step": 8410
},
{
"epoch": 0.03401786543954557,
"grad_norm": 930.1837158203125,
"learning_rate": 1.684e-05,
"loss": 169.9454,
"step": 8420
},
{
"epoch": 0.0340582667049132,
"grad_norm": 514.6612548828125,
"learning_rate": 1.686e-05,
"loss": 202.2469,
"step": 8430
},
{
"epoch": 0.03409866797028083,
"grad_norm": 747.2769775390625,
"learning_rate": 1.688e-05,
"loss": 176.8761,
"step": 8440
},
{
"epoch": 0.03413906923564846,
"grad_norm": 1148.2232666015625,
"learning_rate": 1.69e-05,
"loss": 124.8942,
"step": 8450
},
{
"epoch": 0.03417947050101609,
"grad_norm": 1272.0814208984375,
"learning_rate": 1.692e-05,
"loss": 201.3207,
"step": 8460
},
{
"epoch": 0.034219871766383726,
"grad_norm": 658.9476318359375,
"learning_rate": 1.694e-05,
"loss": 193.7989,
"step": 8470
},
{
"epoch": 0.03426027303175135,
"grad_norm": 1220.1839599609375,
"learning_rate": 1.696e-05,
"loss": 204.7586,
"step": 8480
},
{
"epoch": 0.034300674297118985,
"grad_norm": 650.1453857421875,
"learning_rate": 1.698e-05,
"loss": 243.6956,
"step": 8490
},
{
"epoch": 0.03434107556248662,
"grad_norm": 654.3552856445312,
"learning_rate": 1.7000000000000003e-05,
"loss": 200.6542,
"step": 8500
},
{
"epoch": 0.03438147682785425,
"grad_norm": 831.0885009765625,
"learning_rate": 1.702e-05,
"loss": 245.5201,
"step": 8510
},
{
"epoch": 0.03442187809322188,
"grad_norm": 945.2589111328125,
"learning_rate": 1.704e-05,
"loss": 192.208,
"step": 8520
},
{
"epoch": 0.03446227935858951,
"grad_norm": 929.2937622070312,
"learning_rate": 1.706e-05,
"loss": 195.5855,
"step": 8530
},
{
"epoch": 0.034502680623957144,
"grad_norm": 1585.8245849609375,
"learning_rate": 1.7080000000000002e-05,
"loss": 161.8802,
"step": 8540
},
{
"epoch": 0.03454308188932478,
"grad_norm": 1783.3887939453125,
"learning_rate": 1.7100000000000002e-05,
"loss": 259.6359,
"step": 8550
},
{
"epoch": 0.0345834831546924,
"grad_norm": 1169.6343994140625,
"learning_rate": 1.712e-05,
"loss": 126.7247,
"step": 8560
},
{
"epoch": 0.034623884420060036,
"grad_norm": 967.0234375,
"learning_rate": 1.7140000000000002e-05,
"loss": 228.0373,
"step": 8570
},
{
"epoch": 0.03466428568542767,
"grad_norm": 1204.86865234375,
"learning_rate": 1.7160000000000002e-05,
"loss": 183.3604,
"step": 8580
},
{
"epoch": 0.0347046869507953,
"grad_norm": 3491.818603515625,
"learning_rate": 1.718e-05,
"loss": 151.8515,
"step": 8590
},
{
"epoch": 0.03474508821616293,
"grad_norm": 698.922607421875,
"learning_rate": 1.7199999999999998e-05,
"loss": 364.246,
"step": 8600
},
{
"epoch": 0.03478548948153056,
"grad_norm": 754.37646484375,
"learning_rate": 1.722e-05,
"loss": 214.1979,
"step": 8610
},
{
"epoch": 0.034825890746898194,
"grad_norm": 1443.7174072265625,
"learning_rate": 1.724e-05,
"loss": 146.2507,
"step": 8620
},
{
"epoch": 0.03486629201226583,
"grad_norm": 1486.6044921875,
"learning_rate": 1.726e-05,
"loss": 209.7959,
"step": 8630
},
{
"epoch": 0.034906693277633453,
"grad_norm": 1869.3212890625,
"learning_rate": 1.728e-05,
"loss": 216.6505,
"step": 8640
},
{
"epoch": 0.03494709454300109,
"grad_norm": 1274.484130859375,
"learning_rate": 1.73e-05,
"loss": 175.598,
"step": 8650
},
{
"epoch": 0.03498749580836872,
"grad_norm": 906.1784057617188,
"learning_rate": 1.732e-05,
"loss": 194.2832,
"step": 8660
},
{
"epoch": 0.03502789707373635,
"grad_norm": 897.7321166992188,
"learning_rate": 1.734e-05,
"loss": 167.8126,
"step": 8670
},
{
"epoch": 0.03506829833910398,
"grad_norm": 1329.6224365234375,
"learning_rate": 1.736e-05,
"loss": 100.7456,
"step": 8680
},
{
"epoch": 0.03510869960447161,
"grad_norm": 2856.0009765625,
"learning_rate": 1.7380000000000003e-05,
"loss": 136.0745,
"step": 8690
},
{
"epoch": 0.035149100869839245,
"grad_norm": 851.291748046875,
"learning_rate": 1.74e-05,
"loss": 276.2555,
"step": 8700
},
{
"epoch": 0.03518950213520688,
"grad_norm": 1677.5799560546875,
"learning_rate": 1.742e-05,
"loss": 209.6706,
"step": 8710
},
{
"epoch": 0.035229903400574504,
"grad_norm": 640.1288452148438,
"learning_rate": 1.7440000000000002e-05,
"loss": 223.2772,
"step": 8720
},
{
"epoch": 0.03527030466594214,
"grad_norm": 4403.36962890625,
"learning_rate": 1.7460000000000002e-05,
"loss": 219.3949,
"step": 8730
},
{
"epoch": 0.03531070593130977,
"grad_norm": 2118.526123046875,
"learning_rate": 1.7480000000000002e-05,
"loss": 264.3927,
"step": 8740
},
{
"epoch": 0.0353511071966774,
"grad_norm": 1355.910400390625,
"learning_rate": 1.75e-05,
"loss": 172.0311,
"step": 8750
},
{
"epoch": 0.03539150846204503,
"grad_norm": 1693.513671875,
"learning_rate": 1.752e-05,
"loss": 185.7292,
"step": 8760
},
{
"epoch": 0.03543190972741266,
"grad_norm": 1187.20263671875,
"learning_rate": 1.754e-05,
"loss": 157.557,
"step": 8770
},
{
"epoch": 0.035472310992780295,
"grad_norm": 893.8792114257812,
"learning_rate": 1.756e-05,
"loss": 191.817,
"step": 8780
},
{
"epoch": 0.03551271225814793,
"grad_norm": 1077.1021728515625,
"learning_rate": 1.758e-05,
"loss": 189.9844,
"step": 8790
},
{
"epoch": 0.035553113523515555,
"grad_norm": 929.8483276367188,
"learning_rate": 1.76e-05,
"loss": 197.5466,
"step": 8800
},
{
"epoch": 0.03559351478888319,
"grad_norm": 2271.410400390625,
"learning_rate": 1.762e-05,
"loss": 228.7887,
"step": 8810
},
{
"epoch": 0.03563391605425082,
"grad_norm": 2570.052001953125,
"learning_rate": 1.764e-05,
"loss": 155.6589,
"step": 8820
},
{
"epoch": 0.035674317319618454,
"grad_norm": 657.0897216796875,
"learning_rate": 1.766e-05,
"loss": 225.2949,
"step": 8830
},
{
"epoch": 0.03571471858498608,
"grad_norm": 1940.9866943359375,
"learning_rate": 1.7680000000000004e-05,
"loss": 162.1829,
"step": 8840
},
{
"epoch": 0.03575511985035371,
"grad_norm": 1256.1148681640625,
"learning_rate": 1.77e-05,
"loss": 174.6637,
"step": 8850
},
{
"epoch": 0.035795521115721346,
"grad_norm": 394.8336181640625,
"learning_rate": 1.772e-05,
"loss": 173.8255,
"step": 8860
},
{
"epoch": 0.03583592238108898,
"grad_norm": 2153.534423828125,
"learning_rate": 1.774e-05,
"loss": 171.3043,
"step": 8870
},
{
"epoch": 0.035876323646456605,
"grad_norm": 760.2359619140625,
"learning_rate": 1.7760000000000003e-05,
"loss": 200.2851,
"step": 8880
},
{
"epoch": 0.03591672491182424,
"grad_norm": 1478.953369140625,
"learning_rate": 1.7780000000000003e-05,
"loss": 225.8184,
"step": 8890
},
{
"epoch": 0.03595712617719187,
"grad_norm": 1087.418212890625,
"learning_rate": 1.78e-05,
"loss": 185.8715,
"step": 8900
},
{
"epoch": 0.035997527442559504,
"grad_norm": 2434.63623046875,
"learning_rate": 1.7820000000000002e-05,
"loss": 178.0469,
"step": 8910
},
{
"epoch": 0.03603792870792713,
"grad_norm": 1043.3377685546875,
"learning_rate": 1.7840000000000002e-05,
"loss": 179.4526,
"step": 8920
},
{
"epoch": 0.036078329973294763,
"grad_norm": 1111.5899658203125,
"learning_rate": 1.7860000000000002e-05,
"loss": 210.2127,
"step": 8930
},
{
"epoch": 0.0361187312386624,
"grad_norm": 688.8521728515625,
"learning_rate": 1.7879999999999998e-05,
"loss": 126.2401,
"step": 8940
},
{
"epoch": 0.03615913250403003,
"grad_norm": 958.8832397460938,
"learning_rate": 1.79e-05,
"loss": 221.2132,
"step": 8950
},
{
"epoch": 0.036199533769397656,
"grad_norm": 1401.521728515625,
"learning_rate": 1.792e-05,
"loss": 236.2084,
"step": 8960
},
{
"epoch": 0.03623993503476529,
"grad_norm": 10308.7021484375,
"learning_rate": 1.794e-05,
"loss": 186.4272,
"step": 8970
},
{
"epoch": 0.03628033630013292,
"grad_norm": 1097.2364501953125,
"learning_rate": 1.796e-05,
"loss": 178.1187,
"step": 8980
},
{
"epoch": 0.036320737565500555,
"grad_norm": 1978.429931640625,
"learning_rate": 1.798e-05,
"loss": 196.0737,
"step": 8990
},
{
"epoch": 0.03636113883086818,
"grad_norm": 3581.975341796875,
"learning_rate": 1.8e-05,
"loss": 209.1971,
"step": 9000
},
{
"epoch": 0.036401540096235814,
"grad_norm": 1180.078125,
"learning_rate": 1.802e-05,
"loss": 220.5676,
"step": 9010
},
{
"epoch": 0.03644194136160345,
"grad_norm": 1235.55859375,
"learning_rate": 1.804e-05,
"loss": 124.504,
"step": 9020
},
{
"epoch": 0.03648234262697108,
"grad_norm": 1020.11865234375,
"learning_rate": 1.8060000000000003e-05,
"loss": 99.5129,
"step": 9030
},
{
"epoch": 0.036522743892338706,
"grad_norm": 962.7813110351562,
"learning_rate": 1.808e-05,
"loss": 178.8402,
"step": 9040
},
{
"epoch": 0.03656314515770634,
"grad_norm": 1318.480224609375,
"learning_rate": 1.81e-05,
"loss": 180.0119,
"step": 9050
},
{
"epoch": 0.03660354642307397,
"grad_norm": 2261.136962890625,
"learning_rate": 1.812e-05,
"loss": 329.5452,
"step": 9060
},
{
"epoch": 0.036643947688441605,
"grad_norm": 3996.27197265625,
"learning_rate": 1.8140000000000003e-05,
"loss": 224.336,
"step": 9070
},
{
"epoch": 0.03668434895380923,
"grad_norm": 724.8762817382812,
"learning_rate": 1.8160000000000002e-05,
"loss": 148.4422,
"step": 9080
},
{
"epoch": 0.036724750219176865,
"grad_norm": 838.4283447265625,
"learning_rate": 1.818e-05,
"loss": 160.2637,
"step": 9090
},
{
"epoch": 0.0367651514845445,
"grad_norm": 1999.3927001953125,
"learning_rate": 1.8200000000000002e-05,
"loss": 252.0707,
"step": 9100
},
{
"epoch": 0.03680555274991213,
"grad_norm": 1008.3331298828125,
"learning_rate": 1.8220000000000002e-05,
"loss": 112.7752,
"step": 9110
},
{
"epoch": 0.03684595401527976,
"grad_norm": 1373.5531005859375,
"learning_rate": 1.824e-05,
"loss": 140.497,
"step": 9120
},
{
"epoch": 0.03688635528064739,
"grad_norm": 698.5148315429688,
"learning_rate": 1.826e-05,
"loss": 132.642,
"step": 9130
},
{
"epoch": 0.03692675654601502,
"grad_norm": 1003.8239135742188,
"learning_rate": 1.828e-05,
"loss": 172.7703,
"step": 9140
},
{
"epoch": 0.03696715781138265,
"grad_norm": 695.0194702148438,
"learning_rate": 1.83e-05,
"loss": 188.6919,
"step": 9150
},
{
"epoch": 0.03700755907675028,
"grad_norm": 844.4266967773438,
"learning_rate": 1.832e-05,
"loss": 253.677,
"step": 9160
},
{
"epoch": 0.037047960342117915,
"grad_norm": 1122.217041015625,
"learning_rate": 1.834e-05,
"loss": 197.4021,
"step": 9170
},
{
"epoch": 0.03708836160748555,
"grad_norm": 1105.47314453125,
"learning_rate": 1.8360000000000004e-05,
"loss": 146.558,
"step": 9180
},
{
"epoch": 0.037128762872853174,
"grad_norm": 1299.706298828125,
"learning_rate": 1.838e-05,
"loss": 196.9577,
"step": 9190
},
{
"epoch": 0.03716916413822081,
"grad_norm": 4088.533447265625,
"learning_rate": 1.84e-05,
"loss": 178.9004,
"step": 9200
},
{
"epoch": 0.03720956540358844,
"grad_norm": 2833.98388671875,
"learning_rate": 1.842e-05,
"loss": 220.8664,
"step": 9210
},
{
"epoch": 0.037249966668956074,
"grad_norm": 649.87548828125,
"learning_rate": 1.8440000000000003e-05,
"loss": 163.9712,
"step": 9220
},
{
"epoch": 0.0372903679343237,
"grad_norm": 1362.6846923828125,
"learning_rate": 1.846e-05,
"loss": 186.3454,
"step": 9230
},
{
"epoch": 0.03733076919969133,
"grad_norm": 4132.52099609375,
"learning_rate": 1.848e-05,
"loss": 205.0476,
"step": 9240
},
{
"epoch": 0.037371170465058966,
"grad_norm": 1643.118896484375,
"learning_rate": 1.85e-05,
"loss": 209.0955,
"step": 9250
},
{
"epoch": 0.0374115717304266,
"grad_norm": 5062.59716796875,
"learning_rate": 1.8520000000000002e-05,
"loss": 145.1507,
"step": 9260
},
{
"epoch": 0.037451972995794225,
"grad_norm": 390.54925537109375,
"learning_rate": 1.8540000000000002e-05,
"loss": 221.6778,
"step": 9270
},
{
"epoch": 0.03749237426116186,
"grad_norm": 593.5599975585938,
"learning_rate": 1.856e-05,
"loss": 188.7002,
"step": 9280
},
{
"epoch": 0.03753277552652949,
"grad_norm": 1883.616455078125,
"learning_rate": 1.858e-05,
"loss": 189.2378,
"step": 9290
},
{
"epoch": 0.037573176791897124,
"grad_norm": 1161.8848876953125,
"learning_rate": 1.86e-05,
"loss": 197.2501,
"step": 9300
},
{
"epoch": 0.03761357805726475,
"grad_norm": 692.7471313476562,
"learning_rate": 1.862e-05,
"loss": 171.7747,
"step": 9310
},
{
"epoch": 0.03765397932263238,
"grad_norm": 644.8750610351562,
"learning_rate": 1.864e-05,
"loss": 243.3304,
"step": 9320
},
{
"epoch": 0.037694380588000016,
"grad_norm": 1095.04443359375,
"learning_rate": 1.866e-05,
"loss": 262.2983,
"step": 9330
},
{
"epoch": 0.03773478185336765,
"grad_norm": 1455.4674072265625,
"learning_rate": 1.868e-05,
"loss": 179.6782,
"step": 9340
},
{
"epoch": 0.037775183118735275,
"grad_norm": 788.5756225585938,
"learning_rate": 1.87e-05,
"loss": 108.4792,
"step": 9350
},
{
"epoch": 0.03781558438410291,
"grad_norm": 1122.2181396484375,
"learning_rate": 1.872e-05,
"loss": 261.3737,
"step": 9360
},
{
"epoch": 0.03785598564947054,
"grad_norm": 1429.47607421875,
"learning_rate": 1.8740000000000004e-05,
"loss": 143.9239,
"step": 9370
},
{
"epoch": 0.037896386914838175,
"grad_norm": 1072.0924072265625,
"learning_rate": 1.876e-05,
"loss": 217.753,
"step": 9380
},
{
"epoch": 0.0379367881802058,
"grad_norm": 2434.5927734375,
"learning_rate": 1.878e-05,
"loss": 144.4152,
"step": 9390
},
{
"epoch": 0.037977189445573434,
"grad_norm": 803.2135620117188,
"learning_rate": 1.88e-05,
"loss": 167.2345,
"step": 9400
},
{
"epoch": 0.03801759071094107,
"grad_norm": 1207.8824462890625,
"learning_rate": 1.8820000000000003e-05,
"loss": 211.6839,
"step": 9410
},
{
"epoch": 0.0380579919763087,
"grad_norm": 1132.8394775390625,
"learning_rate": 1.8840000000000003e-05,
"loss": 147.7339,
"step": 9420
},
{
"epoch": 0.038098393241676326,
"grad_norm": 802.6298828125,
"learning_rate": 1.886e-05,
"loss": 139.8525,
"step": 9430
},
{
"epoch": 0.03813879450704396,
"grad_norm": 3992.421875,
"learning_rate": 1.888e-05,
"loss": 199.5186,
"step": 9440
},
{
"epoch": 0.03817919577241159,
"grad_norm": 1528.7784423828125,
"learning_rate": 1.8900000000000002e-05,
"loss": 186.0713,
"step": 9450
},
{
"epoch": 0.038219597037779225,
"grad_norm": 853.1768798828125,
"learning_rate": 1.8920000000000002e-05,
"loss": 222.1405,
"step": 9460
},
{
"epoch": 0.03825999830314685,
"grad_norm": 991.3099975585938,
"learning_rate": 1.894e-05,
"loss": 219.1491,
"step": 9470
},
{
"epoch": 0.038300399568514484,
"grad_norm": 2823.091064453125,
"learning_rate": 1.896e-05,
"loss": 216.1606,
"step": 9480
},
{
"epoch": 0.03834080083388212,
"grad_norm": 792.9014892578125,
"learning_rate": 1.898e-05,
"loss": 185.3038,
"step": 9490
},
{
"epoch": 0.03838120209924975,
"grad_norm": 760.1093139648438,
"learning_rate": 1.9e-05,
"loss": 213.5252,
"step": 9500
},
{
"epoch": 0.03842160336461738,
"grad_norm": 1229.9395751953125,
"learning_rate": 1.902e-05,
"loss": 99.1262,
"step": 9510
},
{
"epoch": 0.03846200462998501,
"grad_norm": 1547.040771484375,
"learning_rate": 1.904e-05,
"loss": 193.0997,
"step": 9520
},
{
"epoch": 0.03850240589535264,
"grad_norm": 5546.650390625,
"learning_rate": 1.906e-05,
"loss": 191.2586,
"step": 9530
},
{
"epoch": 0.038542807160720276,
"grad_norm": 1049.4068603515625,
"learning_rate": 1.908e-05,
"loss": 126.0632,
"step": 9540
},
{
"epoch": 0.0385832084260879,
"grad_norm": 670.5220947265625,
"learning_rate": 1.91e-05,
"loss": 152.6221,
"step": 9550
},
{
"epoch": 0.038623609691455535,
"grad_norm": 1974.0662841796875,
"learning_rate": 1.9120000000000003e-05,
"loss": 175.8063,
"step": 9560
},
{
"epoch": 0.03866401095682317,
"grad_norm": 4551.0517578125,
"learning_rate": 1.914e-05,
"loss": 188.1946,
"step": 9570
},
{
"epoch": 0.0387044122221908,
"grad_norm": 806.855712890625,
"learning_rate": 1.916e-05,
"loss": 166.843,
"step": 9580
},
{
"epoch": 0.03874481348755843,
"grad_norm": 531.7847290039062,
"learning_rate": 1.918e-05,
"loss": 141.4506,
"step": 9590
},
{
"epoch": 0.03878521475292606,
"grad_norm": 931.3214111328125,
"learning_rate": 1.9200000000000003e-05,
"loss": 184.4449,
"step": 9600
},
{
"epoch": 0.03882561601829369,
"grad_norm": 728.2243041992188,
"learning_rate": 1.9220000000000002e-05,
"loss": 144.1032,
"step": 9610
},
{
"epoch": 0.038866017283661326,
"grad_norm": 1086.5455322265625,
"learning_rate": 1.924e-05,
"loss": 166.8239,
"step": 9620
},
{
"epoch": 0.03890641854902895,
"grad_norm": 0.0,
"learning_rate": 1.9260000000000002e-05,
"loss": 195.737,
"step": 9630
},
{
"epoch": 0.038946819814396585,
"grad_norm": 1560.2904052734375,
"learning_rate": 1.9280000000000002e-05,
"loss": 165.3053,
"step": 9640
},
{
"epoch": 0.03898722107976422,
"grad_norm": 1141.6810302734375,
"learning_rate": 1.93e-05,
"loss": 170.9885,
"step": 9650
},
{
"epoch": 0.03902762234513185,
"grad_norm": 0.0,
"learning_rate": 1.932e-05,
"loss": 154.8165,
"step": 9660
},
{
"epoch": 0.03906802361049948,
"grad_norm": 1502.9287109375,
"learning_rate": 1.934e-05,
"loss": 161.04,
"step": 9670
},
{
"epoch": 0.03910842487586711,
"grad_norm": 1250.09619140625,
"learning_rate": 1.936e-05,
"loss": 140.4067,
"step": 9680
},
{
"epoch": 0.039148826141234744,
"grad_norm": 1319.486328125,
"learning_rate": 1.938e-05,
"loss": 183.2919,
"step": 9690
},
{
"epoch": 0.03918922740660238,
"grad_norm": 1285.4862060546875,
"learning_rate": 1.94e-05,
"loss": 200.0876,
"step": 9700
},
{
"epoch": 0.03922962867197,
"grad_norm": 935.0762939453125,
"learning_rate": 1.942e-05,
"loss": 135.3171,
"step": 9710
},
{
"epoch": 0.039270029937337636,
"grad_norm": 1619.843017578125,
"learning_rate": 1.944e-05,
"loss": 193.6001,
"step": 9720
},
{
"epoch": 0.03931043120270527,
"grad_norm": 1271.864990234375,
"learning_rate": 1.946e-05,
"loss": 115.0868,
"step": 9730
},
{
"epoch": 0.0393508324680729,
"grad_norm": 724.2216186523438,
"learning_rate": 1.948e-05,
"loss": 184.4641,
"step": 9740
},
{
"epoch": 0.03939123373344053,
"grad_norm": 1635.4261474609375,
"learning_rate": 1.9500000000000003e-05,
"loss": 208.8505,
"step": 9750
},
{
"epoch": 0.03943163499880816,
"grad_norm": 2965.97265625,
"learning_rate": 1.9520000000000003e-05,
"loss": 172.8497,
"step": 9760
},
{
"epoch": 0.039472036264175794,
"grad_norm": 2287.69287109375,
"learning_rate": 1.954e-05,
"loss": 182.5987,
"step": 9770
},
{
"epoch": 0.03951243752954343,
"grad_norm": 812.2120971679688,
"learning_rate": 1.956e-05,
"loss": 125.8975,
"step": 9780
},
{
"epoch": 0.039552838794911054,
"grad_norm": 2255.82275390625,
"learning_rate": 1.9580000000000002e-05,
"loss": 205.0292,
"step": 9790
},
{
"epoch": 0.03959324006027869,
"grad_norm": 794.7918701171875,
"learning_rate": 1.9600000000000002e-05,
"loss": 140.7012,
"step": 9800
},
{
"epoch": 0.03963364132564632,
"grad_norm": 1928.4591064453125,
"learning_rate": 1.9620000000000002e-05,
"loss": 230.6109,
"step": 9810
},
{
"epoch": 0.03967404259101395,
"grad_norm": 497.63433837890625,
"learning_rate": 1.9640000000000002e-05,
"loss": 175.5,
"step": 9820
},
{
"epoch": 0.03971444385638158,
"grad_norm": 605.4525146484375,
"learning_rate": 1.966e-05,
"loss": 146.0165,
"step": 9830
},
{
"epoch": 0.03975484512174921,
"grad_norm": 12817.353515625,
"learning_rate": 1.968e-05,
"loss": 223.3047,
"step": 9840
},
{
"epoch": 0.039795246387116845,
"grad_norm": 1107.6016845703125,
"learning_rate": 1.97e-05,
"loss": 164.6401,
"step": 9850
},
{
"epoch": 0.03983564765248448,
"grad_norm": 1794.5758056640625,
"learning_rate": 1.972e-05,
"loss": 211.4972,
"step": 9860
},
{
"epoch": 0.039876048917852104,
"grad_norm": 1120.5679931640625,
"learning_rate": 1.974e-05,
"loss": 198.396,
"step": 9870
},
{
"epoch": 0.03991645018321974,
"grad_norm": 627.087646484375,
"learning_rate": 1.976e-05,
"loss": 142.4417,
"step": 9880
},
{
"epoch": 0.03995685144858737,
"grad_norm": 415.39337158203125,
"learning_rate": 1.978e-05,
"loss": 109.121,
"step": 9890
},
{
"epoch": 0.039997252713955,
"grad_norm": 1138.5850830078125,
"learning_rate": 1.9800000000000004e-05,
"loss": 190.1185,
"step": 9900
},
{
"epoch": 0.04003765397932263,
"grad_norm": 506.6181945800781,
"learning_rate": 1.982e-05,
"loss": 152.6702,
"step": 9910
},
{
"epoch": 0.04007805524469026,
"grad_norm": 1317.2239990234375,
"learning_rate": 1.984e-05,
"loss": 210.7801,
"step": 9920
},
{
"epoch": 0.040118456510057895,
"grad_norm": 1474.5733642578125,
"learning_rate": 1.986e-05,
"loss": 224.4348,
"step": 9930
},
{
"epoch": 0.04015885777542553,
"grad_norm": 1309.1876220703125,
"learning_rate": 1.9880000000000003e-05,
"loss": 87.4433,
"step": 9940
},
{
"epoch": 0.040199259040793155,
"grad_norm": 861.6743774414062,
"learning_rate": 1.9900000000000003e-05,
"loss": 136.5576,
"step": 9950
},
{
"epoch": 0.04023966030616079,
"grad_norm": 847.05517578125,
"learning_rate": 1.992e-05,
"loss": 182.9425,
"step": 9960
},
{
"epoch": 0.04028006157152842,
"grad_norm": 1133.978515625,
"learning_rate": 1.994e-05,
"loss": 205.3185,
"step": 9970
},
{
"epoch": 0.040320462836896054,
"grad_norm": 670.1548461914062,
"learning_rate": 1.9960000000000002e-05,
"loss": 190.8365,
"step": 9980
},
{
"epoch": 0.04036086410226368,
"grad_norm": 887.6602783203125,
"learning_rate": 1.9980000000000002e-05,
"loss": 180.067,
"step": 9990
},
{
"epoch": 0.04040126536763131,
"grad_norm": 687.7095336914062,
"learning_rate": 2e-05,
"loss": 127.6409,
"step": 10000
},
{
"epoch": 0.040441666632998946,
"grad_norm": 385.455810546875,
"learning_rate": 2.002e-05,
"loss": 161.4435,
"step": 10010
},
{
"epoch": 0.04048206789836658,
"grad_norm": 1409.2718505859375,
"learning_rate": 2.004e-05,
"loss": 182.3646,
"step": 10020
},
{
"epoch": 0.040522469163734205,
"grad_norm": 803.884765625,
"learning_rate": 2.006e-05,
"loss": 210.5829,
"step": 10030
},
{
"epoch": 0.04056287042910184,
"grad_norm": 1560.3782958984375,
"learning_rate": 2.008e-05,
"loss": 164.0962,
"step": 10040
},
{
"epoch": 0.04060327169446947,
"grad_norm": 2029.135009765625,
"learning_rate": 2.01e-05,
"loss": 140.0395,
"step": 10050
},
{
"epoch": 0.040643672959837104,
"grad_norm": 0.0,
"learning_rate": 2.012e-05,
"loss": 144.4245,
"step": 10060
},
{
"epoch": 0.04068407422520473,
"grad_norm": 1140.849853515625,
"learning_rate": 2.014e-05,
"loss": 179.6542,
"step": 10070
},
{
"epoch": 0.040724475490572364,
"grad_norm": 1746.8056640625,
"learning_rate": 2.016e-05,
"loss": 222.9251,
"step": 10080
},
{
"epoch": 0.04076487675594,
"grad_norm": 722.617919921875,
"learning_rate": 2.0180000000000003e-05,
"loss": 175.7524,
"step": 10090
},
{
"epoch": 0.04080527802130763,
"grad_norm": 837.6021118164062,
"learning_rate": 2.0200000000000003e-05,
"loss": 180.1091,
"step": 10100
},
{
"epoch": 0.040845679286675256,
"grad_norm": 1460.4134521484375,
"learning_rate": 2.022e-05,
"loss": 188.8666,
"step": 10110
},
{
"epoch": 0.04088608055204289,
"grad_norm": 770.96484375,
"learning_rate": 2.024e-05,
"loss": 164.6192,
"step": 10120
},
{
"epoch": 0.04092648181741052,
"grad_norm": 887.3283081054688,
"learning_rate": 2.0260000000000003e-05,
"loss": 181.0503,
"step": 10130
},
{
"epoch": 0.040966883082778155,
"grad_norm": 1267.9124755859375,
"learning_rate": 2.0280000000000002e-05,
"loss": 191.1377,
"step": 10140
},
{
"epoch": 0.04100728434814578,
"grad_norm": 785.875244140625,
"learning_rate": 2.0300000000000002e-05,
"loss": 175.139,
"step": 10150
},
{
"epoch": 0.041047685613513414,
"grad_norm": 3146.860107421875,
"learning_rate": 2.032e-05,
"loss": 189.9724,
"step": 10160
},
{
"epoch": 0.04108808687888105,
"grad_norm": 1048.8988037109375,
"learning_rate": 2.0340000000000002e-05,
"loss": 219.648,
"step": 10170
},
{
"epoch": 0.04112848814424868,
"grad_norm": 534.6448364257812,
"learning_rate": 2.036e-05,
"loss": 165.6375,
"step": 10180
},
{
"epoch": 0.041168889409616306,
"grad_norm": 2196.583984375,
"learning_rate": 2.038e-05,
"loss": 219.21,
"step": 10190
},
{
"epoch": 0.04120929067498394,
"grad_norm": 868.2059326171875,
"learning_rate": 2.04e-05,
"loss": 186.5126,
"step": 10200
},
{
"epoch": 0.04124969194035157,
"grad_norm": 1177.8323974609375,
"learning_rate": 2.042e-05,
"loss": 173.2809,
"step": 10210
},
{
"epoch": 0.041290093205719205,
"grad_norm": 1213.236572265625,
"learning_rate": 2.044e-05,
"loss": 195.4992,
"step": 10220
},
{
"epoch": 0.04133049447108683,
"grad_norm": 712.7450561523438,
"learning_rate": 2.046e-05,
"loss": 123.3338,
"step": 10230
},
{
"epoch": 0.041370895736454465,
"grad_norm": 1610.07373046875,
"learning_rate": 2.048e-05,
"loss": 182.489,
"step": 10240
},
{
"epoch": 0.0414112970018221,
"grad_norm": 1756.5374755859375,
"learning_rate": 2.05e-05,
"loss": 271.8728,
"step": 10250
},
{
"epoch": 0.04145169826718973,
"grad_norm": 1240.2843017578125,
"learning_rate": 2.052e-05,
"loss": 166.9299,
"step": 10260
},
{
"epoch": 0.04149209953255736,
"grad_norm": 803.8770141601562,
"learning_rate": 2.054e-05,
"loss": 155.5844,
"step": 10270
},
{
"epoch": 0.04153250079792499,
"grad_norm": 954.6376953125,
"learning_rate": 2.0560000000000003e-05,
"loss": 199.4898,
"step": 10280
},
{
"epoch": 0.04157290206329262,
"grad_norm": 667.2396240234375,
"learning_rate": 2.0580000000000003e-05,
"loss": 112.1381,
"step": 10290
},
{
"epoch": 0.041613303328660256,
"grad_norm": 1343.771240234375,
"learning_rate": 2.06e-05,
"loss": 200.9802,
"step": 10300
},
{
"epoch": 0.04165370459402788,
"grad_norm": 1224.6522216796875,
"learning_rate": 2.062e-05,
"loss": 128.1033,
"step": 10310
},
{
"epoch": 0.041694105859395515,
"grad_norm": 845.0104370117188,
"learning_rate": 2.0640000000000002e-05,
"loss": 165.8742,
"step": 10320
},
{
"epoch": 0.04173450712476315,
"grad_norm": 532.292236328125,
"learning_rate": 2.0660000000000002e-05,
"loss": 131.5288,
"step": 10330
},
{
"epoch": 0.04177490839013078,
"grad_norm": 876.7532958984375,
"learning_rate": 2.0680000000000002e-05,
"loss": 136.8921,
"step": 10340
},
{
"epoch": 0.04181530965549841,
"grad_norm": 2460.5791015625,
"learning_rate": 2.07e-05,
"loss": 152.2601,
"step": 10350
},
{
"epoch": 0.04185571092086604,
"grad_norm": 1794.477783203125,
"learning_rate": 2.072e-05,
"loss": 201.8371,
"step": 10360
},
{
"epoch": 0.041896112186233674,
"grad_norm": 1977.7452392578125,
"learning_rate": 2.074e-05,
"loss": 247.7894,
"step": 10370
},
{
"epoch": 0.04193651345160131,
"grad_norm": 667.1467895507812,
"learning_rate": 2.076e-05,
"loss": 222.6323,
"step": 10380
},
{
"epoch": 0.04197691471696893,
"grad_norm": 351.2049865722656,
"learning_rate": 2.078e-05,
"loss": 138.2745,
"step": 10390
},
{
"epoch": 0.042017315982336566,
"grad_norm": 944.7986450195312,
"learning_rate": 2.08e-05,
"loss": 196.8081,
"step": 10400
},
{
"epoch": 0.0420577172477042,
"grad_norm": 1119.5870361328125,
"learning_rate": 2.082e-05,
"loss": 163.2965,
"step": 10410
},
{
"epoch": 0.04209811851307183,
"grad_norm": 684.2236938476562,
"learning_rate": 2.084e-05,
"loss": 154.9308,
"step": 10420
},
{
"epoch": 0.04213851977843946,
"grad_norm": 1214.9677734375,
"learning_rate": 2.086e-05,
"loss": 181.0213,
"step": 10430
},
{
"epoch": 0.04217892104380709,
"grad_norm": 979.2085571289062,
"learning_rate": 2.0880000000000003e-05,
"loss": 214.0994,
"step": 10440
},
{
"epoch": 0.042219322309174724,
"grad_norm": 1195.45947265625,
"learning_rate": 2.09e-05,
"loss": 184.8288,
"step": 10450
},
{
"epoch": 0.04225972357454236,
"grad_norm": 3266.998291015625,
"learning_rate": 2.092e-05,
"loss": 154.9474,
"step": 10460
},
{
"epoch": 0.04230012483990998,
"grad_norm": 1256.9425048828125,
"learning_rate": 2.0940000000000003e-05,
"loss": 119.0928,
"step": 10470
},
{
"epoch": 0.042340526105277616,
"grad_norm": 816.3329467773438,
"learning_rate": 2.0960000000000003e-05,
"loss": 138.4739,
"step": 10480
},
{
"epoch": 0.04238092737064525,
"grad_norm": 1145.7486572265625,
"learning_rate": 2.098e-05,
"loss": 179.8457,
"step": 10490
},
{
"epoch": 0.04242132863601288,
"grad_norm": 4142.47998046875,
"learning_rate": 2.1e-05,
"loss": 190.1194,
"step": 10500
},
{
"epoch": 0.04246172990138051,
"grad_norm": 748.7318115234375,
"learning_rate": 2.1020000000000002e-05,
"loss": 156.9302,
"step": 10510
},
{
"epoch": 0.04250213116674814,
"grad_norm": 1273.5777587890625,
"learning_rate": 2.1040000000000002e-05,
"loss": 175.272,
"step": 10520
},
{
"epoch": 0.042542532432115775,
"grad_norm": 551.0802612304688,
"learning_rate": 2.106e-05,
"loss": 167.7439,
"step": 10530
},
{
"epoch": 0.04258293369748341,
"grad_norm": 604.2559814453125,
"learning_rate": 2.1079999999999998e-05,
"loss": 153.7686,
"step": 10540
},
{
"epoch": 0.042623334962851034,
"grad_norm": 460.89154052734375,
"learning_rate": 2.11e-05,
"loss": 117.5873,
"step": 10550
},
{
"epoch": 0.04266373622821867,
"grad_norm": 1551.710693359375,
"learning_rate": 2.112e-05,
"loss": 165.0876,
"step": 10560
},
{
"epoch": 0.0427041374935863,
"grad_norm": 934.1738891601562,
"learning_rate": 2.114e-05,
"loss": 143.8318,
"step": 10570
},
{
"epoch": 0.04274453875895393,
"grad_norm": 1267.439453125,
"learning_rate": 2.116e-05,
"loss": 187.996,
"step": 10580
},
{
"epoch": 0.04278494002432156,
"grad_norm": 1843.2279052734375,
"learning_rate": 2.118e-05,
"loss": 181.2761,
"step": 10590
},
{
"epoch": 0.04282534128968919,
"grad_norm": 2143.365966796875,
"learning_rate": 2.12e-05,
"loss": 158.841,
"step": 10600
},
{
"epoch": 0.042865742555056825,
"grad_norm": 925.6973266601562,
"learning_rate": 2.122e-05,
"loss": 210.331,
"step": 10610
},
{
"epoch": 0.04290614382042446,
"grad_norm": 2171.398681640625,
"learning_rate": 2.124e-05,
"loss": 225.7547,
"step": 10620
},
{
"epoch": 0.042946545085792084,
"grad_norm": 1160.7347412109375,
"learning_rate": 2.1260000000000003e-05,
"loss": 153.8757,
"step": 10630
},
{
"epoch": 0.04298694635115972,
"grad_norm": 1135.9134521484375,
"learning_rate": 2.128e-05,
"loss": 188.1846,
"step": 10640
},
{
"epoch": 0.04302734761652735,
"grad_norm": 2029.5738525390625,
"learning_rate": 2.13e-05,
"loss": 201.2368,
"step": 10650
},
{
"epoch": 0.043067748881894984,
"grad_norm": 814.3689575195312,
"learning_rate": 2.1320000000000003e-05,
"loss": 206.1595,
"step": 10660
},
{
"epoch": 0.04310815014726261,
"grad_norm": 422.2183837890625,
"learning_rate": 2.1340000000000002e-05,
"loss": 157.6857,
"step": 10670
},
{
"epoch": 0.04314855141263024,
"grad_norm": 880.4464111328125,
"learning_rate": 2.1360000000000002e-05,
"loss": 163.0372,
"step": 10680
},
{
"epoch": 0.043188952677997876,
"grad_norm": 484.959228515625,
"learning_rate": 2.138e-05,
"loss": 144.4623,
"step": 10690
},
{
"epoch": 0.04322935394336551,
"grad_norm": 1225.7841796875,
"learning_rate": 2.1400000000000002e-05,
"loss": 226.4971,
"step": 10700
},
{
"epoch": 0.043269755208733135,
"grad_norm": 569.0413208007812,
"learning_rate": 2.142e-05,
"loss": 186.6787,
"step": 10710
},
{
"epoch": 0.04331015647410077,
"grad_norm": 1226.3690185546875,
"learning_rate": 2.144e-05,
"loss": 219.7577,
"step": 10720
},
{
"epoch": 0.0433505577394684,
"grad_norm": 1915.441650390625,
"learning_rate": 2.146e-05,
"loss": 194.0689,
"step": 10730
},
{
"epoch": 0.043390959004836034,
"grad_norm": 1193.3553466796875,
"learning_rate": 2.148e-05,
"loss": 149.6157,
"step": 10740
},
{
"epoch": 0.04343136027020366,
"grad_norm": 562.625,
"learning_rate": 2.15e-05,
"loss": 204.4145,
"step": 10750
},
{
"epoch": 0.04347176153557129,
"grad_norm": 12496.0029296875,
"learning_rate": 2.152e-05,
"loss": 248.7909,
"step": 10760
},
{
"epoch": 0.043512162800938926,
"grad_norm": 1115.0479736328125,
"learning_rate": 2.154e-05,
"loss": 183.6795,
"step": 10770
},
{
"epoch": 0.04355256406630656,
"grad_norm": 1180.6798095703125,
"learning_rate": 2.1560000000000004e-05,
"loss": 161.4541,
"step": 10780
},
{
"epoch": 0.043592965331674186,
"grad_norm": 2096.56298828125,
"learning_rate": 2.158e-05,
"loss": 177.0552,
"step": 10790
},
{
"epoch": 0.04363336659704182,
"grad_norm": 1031.798828125,
"learning_rate": 2.16e-05,
"loss": 160.7948,
"step": 10800
},
{
"epoch": 0.04367376786240945,
"grad_norm": 1001.6660766601562,
"learning_rate": 2.162e-05,
"loss": 279.4042,
"step": 10810
},
{
"epoch": 0.043714169127777085,
"grad_norm": 5103.36669921875,
"learning_rate": 2.1640000000000003e-05,
"loss": 198.244,
"step": 10820
},
{
"epoch": 0.04375457039314471,
"grad_norm": 813.843017578125,
"learning_rate": 2.166e-05,
"loss": 204.9555,
"step": 10830
},
{
"epoch": 0.043794971658512344,
"grad_norm": 1556.0152587890625,
"learning_rate": 2.168e-05,
"loss": 178.9818,
"step": 10840
},
{
"epoch": 0.04383537292387998,
"grad_norm": 1136.54443359375,
"learning_rate": 2.1700000000000002e-05,
"loss": 214.708,
"step": 10850
},
{
"epoch": 0.04387577418924761,
"grad_norm": 913.5957641601562,
"learning_rate": 2.1720000000000002e-05,
"loss": 162.8733,
"step": 10860
},
{
"epoch": 0.043916175454615236,
"grad_norm": 1809.7088623046875,
"learning_rate": 2.1740000000000002e-05,
"loss": 202.5214,
"step": 10870
},
{
"epoch": 0.04395657671998287,
"grad_norm": 1365.0262451171875,
"learning_rate": 2.176e-05,
"loss": 197.6029,
"step": 10880
},
{
"epoch": 0.0439969779853505,
"grad_norm": 1124.8267822265625,
"learning_rate": 2.178e-05,
"loss": 195.3639,
"step": 10890
},
{
"epoch": 0.044037379250718135,
"grad_norm": 933.0280151367188,
"learning_rate": 2.18e-05,
"loss": 166.1828,
"step": 10900
},
{
"epoch": 0.04407778051608576,
"grad_norm": 893.4417724609375,
"learning_rate": 2.182e-05,
"loss": 125.791,
"step": 10910
},
{
"epoch": 0.044118181781453394,
"grad_norm": 1642.8094482421875,
"learning_rate": 2.184e-05,
"loss": 144.994,
"step": 10920
},
{
"epoch": 0.04415858304682103,
"grad_norm": 1519.8751220703125,
"learning_rate": 2.186e-05,
"loss": 158.417,
"step": 10930
},
{
"epoch": 0.04419898431218866,
"grad_norm": 1085.8243408203125,
"learning_rate": 2.188e-05,
"loss": 189.8353,
"step": 10940
},
{
"epoch": 0.04423938557755629,
"grad_norm": 907.6950073242188,
"learning_rate": 2.19e-05,
"loss": 131.8621,
"step": 10950
},
{
"epoch": 0.04427978684292392,
"grad_norm": 988.0367431640625,
"learning_rate": 2.192e-05,
"loss": 110.0259,
"step": 10960
},
{
"epoch": 0.04432018810829155,
"grad_norm": 358.17889404296875,
"learning_rate": 2.1940000000000003e-05,
"loss": 144.1581,
"step": 10970
},
{
"epoch": 0.044360589373659186,
"grad_norm": 3126.972900390625,
"learning_rate": 2.196e-05,
"loss": 242.1895,
"step": 10980
},
{
"epoch": 0.04440099063902681,
"grad_norm": 660.0758056640625,
"learning_rate": 2.198e-05,
"loss": 307.6868,
"step": 10990
},
{
"epoch": 0.044441391904394445,
"grad_norm": 1388.6563720703125,
"learning_rate": 2.2000000000000003e-05,
"loss": 191.1774,
"step": 11000
},
{
"epoch": 0.04448179316976208,
"grad_norm": 0.0,
"learning_rate": 2.2020000000000003e-05,
"loss": 182.281,
"step": 11010
},
{
"epoch": 0.04452219443512971,
"grad_norm": 1133.305419921875,
"learning_rate": 2.2040000000000002e-05,
"loss": 141.6396,
"step": 11020
},
{
"epoch": 0.04456259570049734,
"grad_norm": 760.71728515625,
"learning_rate": 2.206e-05,
"loss": 184.8224,
"step": 11030
},
{
"epoch": 0.04460299696586497,
"grad_norm": 700.0611572265625,
"learning_rate": 2.2080000000000002e-05,
"loss": 174.1201,
"step": 11040
},
{
"epoch": 0.0446433982312326,
"grad_norm": 464.8121643066406,
"learning_rate": 2.2100000000000002e-05,
"loss": 147.1738,
"step": 11050
},
{
"epoch": 0.044683799496600236,
"grad_norm": 659.0347900390625,
"learning_rate": 2.212e-05,
"loss": 121.4146,
"step": 11060
},
{
"epoch": 0.04472420076196786,
"grad_norm": 495.3475646972656,
"learning_rate": 2.214e-05,
"loss": 122.6828,
"step": 11070
},
{
"epoch": 0.044764602027335496,
"grad_norm": 899.0453491210938,
"learning_rate": 2.216e-05,
"loss": 171.483,
"step": 11080
},
{
"epoch": 0.04480500329270313,
"grad_norm": 1652.7977294921875,
"learning_rate": 2.218e-05,
"loss": 136.9625,
"step": 11090
},
{
"epoch": 0.04484540455807076,
"grad_norm": 1164.9923095703125,
"learning_rate": 2.22e-05,
"loss": 218.6428,
"step": 11100
},
{
"epoch": 0.04488580582343839,
"grad_norm": 1860.8387451171875,
"learning_rate": 2.222e-05,
"loss": 128.8451,
"step": 11110
},
{
"epoch": 0.04492620708880602,
"grad_norm": 1793.767578125,
"learning_rate": 2.224e-05,
"loss": 127.518,
"step": 11120
},
{
"epoch": 0.044966608354173654,
"grad_norm": 910.3096923828125,
"learning_rate": 2.226e-05,
"loss": 102.4282,
"step": 11130
},
{
"epoch": 0.04500700961954129,
"grad_norm": 3313.176025390625,
"learning_rate": 2.228e-05,
"loss": 205.4728,
"step": 11140
},
{
"epoch": 0.04504741088490891,
"grad_norm": 959.5852661132812,
"learning_rate": 2.23e-05,
"loss": 181.3004,
"step": 11150
},
{
"epoch": 0.045087812150276546,
"grad_norm": 3117.2705078125,
"learning_rate": 2.2320000000000003e-05,
"loss": 159.9985,
"step": 11160
},
{
"epoch": 0.04512821341564418,
"grad_norm": 4501.052734375,
"learning_rate": 2.234e-05,
"loss": 150.3723,
"step": 11170
},
{
"epoch": 0.04516861468101181,
"grad_norm": 981.978515625,
"learning_rate": 2.236e-05,
"loss": 142.0348,
"step": 11180
},
{
"epoch": 0.04520901594637944,
"grad_norm": 1109.785400390625,
"learning_rate": 2.2380000000000003e-05,
"loss": 163.3822,
"step": 11190
},
{
"epoch": 0.04524941721174707,
"grad_norm": 1274.3033447265625,
"learning_rate": 2.2400000000000002e-05,
"loss": 268.0728,
"step": 11200
},
{
"epoch": 0.045289818477114704,
"grad_norm": 750.956298828125,
"learning_rate": 2.2420000000000002e-05,
"loss": 164.3101,
"step": 11210
},
{
"epoch": 0.04533021974248234,
"grad_norm": 908.2940063476562,
"learning_rate": 2.244e-05,
"loss": 205.5986,
"step": 11220
},
{
"epoch": 0.045370621007849964,
"grad_norm": 1060.370849609375,
"learning_rate": 2.2460000000000002e-05,
"loss": 152.6007,
"step": 11230
},
{
"epoch": 0.0454110222732176,
"grad_norm": 1556.2314453125,
"learning_rate": 2.248e-05,
"loss": 204.8753,
"step": 11240
},
{
"epoch": 0.04545142353858523,
"grad_norm": 683.9859619140625,
"learning_rate": 2.25e-05,
"loss": 142.627,
"step": 11250
},
{
"epoch": 0.04549182480395286,
"grad_norm": 1313.3026123046875,
"learning_rate": 2.252e-05,
"loss": 183.1745,
"step": 11260
},
{
"epoch": 0.04553222606932049,
"grad_norm": 1381.1295166015625,
"learning_rate": 2.254e-05,
"loss": 175.9728,
"step": 11270
},
{
"epoch": 0.04557262733468812,
"grad_norm": 509.716796875,
"learning_rate": 2.256e-05,
"loss": 210.1509,
"step": 11280
},
{
"epoch": 0.045613028600055755,
"grad_norm": 2492.5234375,
"learning_rate": 2.258e-05,
"loss": 231.53,
"step": 11290
},
{
"epoch": 0.04565342986542339,
"grad_norm": 1633.21630859375,
"learning_rate": 2.26e-05,
"loss": 152.2005,
"step": 11300
},
{
"epoch": 0.045693831130791014,
"grad_norm": 768.939208984375,
"learning_rate": 2.2620000000000004e-05,
"loss": 185.2856,
"step": 11310
},
{
"epoch": 0.04573423239615865,
"grad_norm": 799.730712890625,
"learning_rate": 2.264e-05,
"loss": 194.0909,
"step": 11320
},
{
"epoch": 0.04577463366152628,
"grad_norm": 3679.43310546875,
"learning_rate": 2.266e-05,
"loss": 354.1977,
"step": 11330
},
{
"epoch": 0.04581503492689391,
"grad_norm": 1576.691650390625,
"learning_rate": 2.268e-05,
"loss": 181.5113,
"step": 11340
},
{
"epoch": 0.04585543619226154,
"grad_norm": 1108.6512451171875,
"learning_rate": 2.2700000000000003e-05,
"loss": 184.2569,
"step": 11350
},
{
"epoch": 0.04589583745762917,
"grad_norm": 1905.447998046875,
"learning_rate": 2.2720000000000003e-05,
"loss": 217.2384,
"step": 11360
},
{
"epoch": 0.045936238722996806,
"grad_norm": 884.3139038085938,
"learning_rate": 2.274e-05,
"loss": 141.6524,
"step": 11370
},
{
"epoch": 0.04597663998836444,
"grad_norm": 617.2454223632812,
"learning_rate": 2.2760000000000002e-05,
"loss": 148.7823,
"step": 11380
},
{
"epoch": 0.046017041253732065,
"grad_norm": 551.2994384765625,
"learning_rate": 2.2780000000000002e-05,
"loss": 133.9882,
"step": 11390
},
{
"epoch": 0.0460574425190997,
"grad_norm": 700.8192138671875,
"learning_rate": 2.2800000000000002e-05,
"loss": 154.8847,
"step": 11400
},
{
"epoch": 0.04609784378446733,
"grad_norm": 1914.3916015625,
"learning_rate": 2.282e-05,
"loss": 145.8075,
"step": 11410
},
{
"epoch": 0.046138245049834964,
"grad_norm": 2445.005859375,
"learning_rate": 2.284e-05,
"loss": 199.9154,
"step": 11420
},
{
"epoch": 0.04617864631520259,
"grad_norm": 1541.644775390625,
"learning_rate": 2.286e-05,
"loss": 168.6186,
"step": 11430
},
{
"epoch": 0.04621904758057022,
"grad_norm": 816.3759765625,
"learning_rate": 2.288e-05,
"loss": 181.8408,
"step": 11440
},
{
"epoch": 0.046259448845937856,
"grad_norm": 1267.0755615234375,
"learning_rate": 2.29e-05,
"loss": 172.8114,
"step": 11450
},
{
"epoch": 0.04629985011130549,
"grad_norm": 855.4586791992188,
"learning_rate": 2.292e-05,
"loss": 197.7767,
"step": 11460
},
{
"epoch": 0.046340251376673115,
"grad_norm": 1094.00634765625,
"learning_rate": 2.294e-05,
"loss": 171.698,
"step": 11470
},
{
"epoch": 0.04638065264204075,
"grad_norm": 1205.0,
"learning_rate": 2.296e-05,
"loss": 215.9526,
"step": 11480
},
{
"epoch": 0.04642105390740838,
"grad_norm": 1176.0413818359375,
"learning_rate": 2.298e-05,
"loss": 163.7831,
"step": 11490
},
{
"epoch": 0.046461455172776014,
"grad_norm": 786.1920776367188,
"learning_rate": 2.3000000000000003e-05,
"loss": 167.4065,
"step": 11500
},
{
"epoch": 0.04650185643814364,
"grad_norm": 516.4776000976562,
"learning_rate": 2.302e-05,
"loss": 165.6811,
"step": 11510
},
{
"epoch": 0.046542257703511274,
"grad_norm": 396.16497802734375,
"learning_rate": 2.304e-05,
"loss": 168.2221,
"step": 11520
},
{
"epoch": 0.04658265896887891,
"grad_norm": 882.178466796875,
"learning_rate": 2.306e-05,
"loss": 140.923,
"step": 11530
},
{
"epoch": 0.04662306023424654,
"grad_norm": 934.631591796875,
"learning_rate": 2.3080000000000003e-05,
"loss": 186.3469,
"step": 11540
},
{
"epoch": 0.046663461499614166,
"grad_norm": 1052.7349853515625,
"learning_rate": 2.3100000000000002e-05,
"loss": 201.8761,
"step": 11550
},
{
"epoch": 0.0467038627649818,
"grad_norm": 1325.5751953125,
"learning_rate": 2.312e-05,
"loss": 255.0451,
"step": 11560
},
{
"epoch": 0.04674426403034943,
"grad_norm": 827.1654663085938,
"learning_rate": 2.3140000000000002e-05,
"loss": 162.8754,
"step": 11570
},
{
"epoch": 0.046784665295717065,
"grad_norm": 506.81097412109375,
"learning_rate": 2.3160000000000002e-05,
"loss": 99.88,
"step": 11580
},
{
"epoch": 0.04682506656108469,
"grad_norm": 1403.5689697265625,
"learning_rate": 2.318e-05,
"loss": 157.3405,
"step": 11590
},
{
"epoch": 0.046865467826452324,
"grad_norm": 1264.109375,
"learning_rate": 2.32e-05,
"loss": 135.8957,
"step": 11600
},
{
"epoch": 0.04690586909181996,
"grad_norm": 851.1090698242188,
"learning_rate": 2.322e-05,
"loss": 98.8174,
"step": 11610
},
{
"epoch": 0.04694627035718759,
"grad_norm": 1114.6529541015625,
"learning_rate": 2.324e-05,
"loss": 174.8662,
"step": 11620
},
{
"epoch": 0.046986671622555216,
"grad_norm": 997.0178833007812,
"learning_rate": 2.326e-05,
"loss": 202.974,
"step": 11630
},
{
"epoch": 0.04702707288792285,
"grad_norm": 2686.275146484375,
"learning_rate": 2.328e-05,
"loss": 180.5854,
"step": 11640
},
{
"epoch": 0.04706747415329048,
"grad_norm": 1341.779541015625,
"learning_rate": 2.3300000000000004e-05,
"loss": 188.3465,
"step": 11650
},
{
"epoch": 0.047107875418658116,
"grad_norm": 1174.360595703125,
"learning_rate": 2.332e-05,
"loss": 100.4389,
"step": 11660
},
{
"epoch": 0.04714827668402574,
"grad_norm": 485.88134765625,
"learning_rate": 2.334e-05,
"loss": 84.7096,
"step": 11670
},
{
"epoch": 0.047188677949393375,
"grad_norm": 700.1830444335938,
"learning_rate": 2.336e-05,
"loss": 252.0006,
"step": 11680
},
{
"epoch": 0.04722907921476101,
"grad_norm": 1874.8160400390625,
"learning_rate": 2.3380000000000003e-05,
"loss": 271.8778,
"step": 11690
},
{
"epoch": 0.04726948048012864,
"grad_norm": 780.0234985351562,
"learning_rate": 2.3400000000000003e-05,
"loss": 156.8313,
"step": 11700
},
{
"epoch": 0.04730988174549627,
"grad_norm": 1217.7091064453125,
"learning_rate": 2.342e-05,
"loss": 139.3272,
"step": 11710
},
{
"epoch": 0.0473502830108639,
"grad_norm": 1002.0595092773438,
"learning_rate": 2.344e-05,
"loss": 155.965,
"step": 11720
},
{
"epoch": 0.04739068427623153,
"grad_norm": 597.17529296875,
"learning_rate": 2.3460000000000002e-05,
"loss": 172.1847,
"step": 11730
},
{
"epoch": 0.047431085541599166,
"grad_norm": 910.2625732421875,
"learning_rate": 2.3480000000000002e-05,
"loss": 148.3666,
"step": 11740
},
{
"epoch": 0.04747148680696679,
"grad_norm": 1215.2615966796875,
"learning_rate": 2.35e-05,
"loss": 194.8361,
"step": 11750
},
{
"epoch": 0.047511888072334425,
"grad_norm": 762.500244140625,
"learning_rate": 2.3520000000000002e-05,
"loss": 113.7633,
"step": 11760
},
{
"epoch": 0.04755228933770206,
"grad_norm": 3147.351318359375,
"learning_rate": 2.354e-05,
"loss": 194.7475,
"step": 11770
},
{
"epoch": 0.04759269060306969,
"grad_norm": 1164.023681640625,
"learning_rate": 2.356e-05,
"loss": 108.1381,
"step": 11780
},
{
"epoch": 0.04763309186843732,
"grad_norm": 2093.1826171875,
"learning_rate": 2.358e-05,
"loss": 165.2831,
"step": 11790
},
{
"epoch": 0.04767349313380495,
"grad_norm": 628.909423828125,
"learning_rate": 2.36e-05,
"loss": 172.9029,
"step": 11800
},
{
"epoch": 0.047713894399172584,
"grad_norm": 400.00811767578125,
"learning_rate": 2.362e-05,
"loss": 181.3741,
"step": 11810
},
{
"epoch": 0.04775429566454022,
"grad_norm": 1536.272216796875,
"learning_rate": 2.364e-05,
"loss": 154.7509,
"step": 11820
},
{
"epoch": 0.04779469692990784,
"grad_norm": 1407.3763427734375,
"learning_rate": 2.366e-05,
"loss": 174.5441,
"step": 11830
},
{
"epoch": 0.047835098195275476,
"grad_norm": 726.475830078125,
"learning_rate": 2.3680000000000004e-05,
"loss": 171.3695,
"step": 11840
},
{
"epoch": 0.04787549946064311,
"grad_norm": 1060.8369140625,
"learning_rate": 2.37e-05,
"loss": 150.6572,
"step": 11850
},
{
"epoch": 0.04791590072601074,
"grad_norm": 2245.559814453125,
"learning_rate": 2.372e-05,
"loss": 172.8109,
"step": 11860
},
{
"epoch": 0.04795630199137837,
"grad_norm": 2129.251708984375,
"learning_rate": 2.374e-05,
"loss": 191.5867,
"step": 11870
},
{
"epoch": 0.047996703256746,
"grad_norm": 1103.0263671875,
"learning_rate": 2.3760000000000003e-05,
"loss": 213.8663,
"step": 11880
},
{
"epoch": 0.048037104522113634,
"grad_norm": 1063.9124755859375,
"learning_rate": 2.3780000000000003e-05,
"loss": 194.2018,
"step": 11890
},
{
"epoch": 0.04807750578748127,
"grad_norm": 7122.65185546875,
"learning_rate": 2.38e-05,
"loss": 229.5696,
"step": 11900
},
{
"epoch": 0.04811790705284889,
"grad_norm": 1017.9201049804688,
"learning_rate": 2.3820000000000002e-05,
"loss": 146.0323,
"step": 11910
},
{
"epoch": 0.048158308318216526,
"grad_norm": 1932.71435546875,
"learning_rate": 2.3840000000000002e-05,
"loss": 143.6282,
"step": 11920
},
{
"epoch": 0.04819870958358416,
"grad_norm": 753.262939453125,
"learning_rate": 2.3860000000000002e-05,
"loss": 151.7239,
"step": 11930
},
{
"epoch": 0.04823911084895179,
"grad_norm": 1837.5714111328125,
"learning_rate": 2.3880000000000002e-05,
"loss": 192.4802,
"step": 11940
},
{
"epoch": 0.04827951211431942,
"grad_norm": 1542.8756103515625,
"learning_rate": 2.39e-05,
"loss": 239.2184,
"step": 11950
},
{
"epoch": 0.04831991337968705,
"grad_norm": 1126.86572265625,
"learning_rate": 2.392e-05,
"loss": 132.8466,
"step": 11960
},
{
"epoch": 0.048360314645054685,
"grad_norm": 650.4667358398438,
"learning_rate": 2.394e-05,
"loss": 121.4513,
"step": 11970
},
{
"epoch": 0.04840071591042232,
"grad_norm": 708.6920776367188,
"learning_rate": 2.396e-05,
"loss": 172.2358,
"step": 11980
},
{
"epoch": 0.048441117175789944,
"grad_norm": 1389.237548828125,
"learning_rate": 2.398e-05,
"loss": 220.8339,
"step": 11990
},
{
"epoch": 0.04848151844115758,
"grad_norm": 580.6951293945312,
"learning_rate": 2.4e-05,
"loss": 178.8768,
"step": 12000
}
],
"logging_steps": 10,
"max_steps": 250000,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 2000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}