{
  "best_metric": 42.71378740060809,
  "best_model_checkpoint": "./whisper-small-ka/checkpoint-14500",
  "epoch": 380.4347826086956,
  "eval_steps": 500,
  "global_step": 17500,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.008451656524678837,
      "grad_norm": 7.251798629760742,
      "learning_rate": 4.6000000000000004e-07,
      "loss": 1.7697,
      "step": 25
    },
    {
      "epoch": 0.016903313049357674,
      "grad_norm": 6.12908935546875,
      "learning_rate": 9.600000000000001e-07,
      "loss": 1.6874,
      "step": 50
    },
    {
      "epoch": 0.02535496957403651,
      "grad_norm": 6.023613452911377,
      "learning_rate": 1.46e-06,
      "loss": 1.5915,
      "step": 75
    },
    {
      "epoch": 0.03380662609871535,
      "grad_norm": 8.941411972045898,
      "learning_rate": 1.9600000000000003e-06,
      "loss": 1.3679,
      "step": 100
    },
    {
      "epoch": 0.042258282623394185,
      "grad_norm": 6.101173400878906,
      "learning_rate": 2.46e-06,
      "loss": 0.9502,
      "step": 125
    },
    {
      "epoch": 0.05070993914807302,
      "grad_norm": 6.879887580871582,
      "learning_rate": 2.96e-06,
      "loss": 0.6954,
      "step": 150
    },
    {
      "epoch": 0.05916159567275186,
      "grad_norm": 5.129401206970215,
      "learning_rate": 3.46e-06,
      "loss": 0.5745,
      "step": 175
    },
    {
      "epoch": 0.0676132521974307,
      "grad_norm": 6.113542079925537,
      "learning_rate": 3.96e-06,
      "loss": 0.4604,
      "step": 200
    },
    {
      "epoch": 0.07606490872210954,
      "grad_norm": 5.392640590667725,
      "learning_rate": 4.4600000000000005e-06,
      "loss": 0.4114,
      "step": 225
    },
    {
      "epoch": 0.08451656524678837,
      "grad_norm": 5.14113712310791,
      "learning_rate": 4.960000000000001e-06,
      "loss": 0.3859,
      "step": 250
    },
    {
      "epoch": 0.09296822177146721,
      "grad_norm": 3.8609211444854736,
      "learning_rate": 5.460000000000001e-06,
      "loss": 0.3692,
      "step": 275
    },
    {
      "epoch": 0.10141987829614604,
      "grad_norm": 4.686000823974609,
      "learning_rate": 5.9600000000000005e-06,
      "loss": 0.3298,
      "step": 300
    },
    {
      "epoch": 0.10987153482082489,
      "grad_norm": 5.032454490661621,
      "learning_rate": 6.460000000000001e-06,
      "loss": 0.2976,
      "step": 325
    },
    {
      "epoch": 0.11832319134550372,
      "grad_norm": 3.724012851715088,
      "learning_rate": 6.96e-06,
      "loss": 0.309,
      "step": 350
    },
    {
      "epoch": 0.12677484787018256,
      "grad_norm": 3.2685935497283936,
      "learning_rate": 7.4600000000000006e-06,
      "loss": 0.2897,
      "step": 375
    },
    {
      "epoch": 0.1352265043948614,
      "grad_norm": 3.796637535095215,
      "learning_rate": 7.960000000000002e-06,
      "loss": 0.2763,
      "step": 400
    },
    {
      "epoch": 0.14367816091954022,
      "grad_norm": 6.331181049346924,
      "learning_rate": 8.46e-06,
      "loss": 0.2714,
      "step": 425
    },
    {
      "epoch": 0.15212981744421908,
      "grad_norm": 4.792943000793457,
      "learning_rate": 8.96e-06,
      "loss": 0.2718,
      "step": 450
    },
    {
      "epoch": 0.1605814739688979,
      "grad_norm": 3.872354507446289,
      "learning_rate": 9.460000000000001e-06,
      "loss": 0.2692,
      "step": 475
    },
    {
      "epoch": 0.16903313049357674,
      "grad_norm": 3.859339475631714,
      "learning_rate": 9.960000000000001e-06,
      "loss": 0.2612,
      "step": 500
    },
    {
      "epoch": 0.16903313049357674,
      "eval_loss": 0.26686015725135803,
      "eval_runtime": 8169.6122,
      "eval_samples_per_second": 1.79,
      "eval_steps_per_second": 0.224,
      "eval_wer": 61.611321105313856,
      "step": 500
    },
    {
      "epoch": 0.17748478701825557,
      "grad_norm": 3.6627910137176514,
      "learning_rate": 9.975789473684211e-06,
      "loss": 0.2347,
      "step": 525
    },
    {
      "epoch": 0.18593644354293443,
      "grad_norm": 3.168164014816284,
      "learning_rate": 9.949473684210526e-06,
      "loss": 0.2438,
      "step": 550
    },
    {
      "epoch": 0.19438810006761326,
      "grad_norm": 2.521146297454834,
      "learning_rate": 9.923157894736844e-06,
      "loss": 0.2238,
      "step": 575
    },
    {
      "epoch": 0.2028397565922921,
      "grad_norm": 4.583806991577148,
      "learning_rate": 9.89684210526316e-06,
      "loss": 0.2412,
      "step": 600
    },
    {
      "epoch": 0.21129141311697092,
      "grad_norm": 3.045239210128784,
      "learning_rate": 9.870526315789474e-06,
      "loss": 0.2155,
      "step": 625
    },
    {
      "epoch": 0.21974306964164977,
      "grad_norm": 3.3001856803894043,
      "learning_rate": 9.84421052631579e-06,
      "loss": 0.2382,
      "step": 650
    },
    {
      "epoch": 0.2281947261663286,
      "grad_norm": 3.613844871520996,
      "learning_rate": 9.817894736842106e-06,
      "loss": 0.2266,
      "step": 675
    },
    {
      "epoch": 0.23664638269100743,
      "grad_norm": 3.4819045066833496,
      "learning_rate": 9.791578947368422e-06,
      "loss": 0.199,
      "step": 700
    },
    {
      "epoch": 0.24509803921568626,
      "grad_norm": 3.3764617443084717,
      "learning_rate": 9.765263157894737e-06,
      "loss": 0.1981,
      "step": 725
    },
    {
      "epoch": 0.2535496957403651,
      "grad_norm": 2.9908406734466553,
      "learning_rate": 9.738947368421054e-06,
      "loss": 0.2129,
      "step": 750
    },
    {
      "epoch": 0.26200135226504395,
      "grad_norm": 3.408409833908081,
      "learning_rate": 9.712631578947369e-06,
      "loss": 0.2171,
      "step": 775
    },
    {
      "epoch": 0.2704530087897228,
      "grad_norm": 3.7549092769622803,
      "learning_rate": 9.686315789473684e-06,
      "loss": 0.2112,
      "step": 800
    },
    {
      "epoch": 0.2789046653144016,
      "grad_norm": 3.1782822608947754,
      "learning_rate": 9.66e-06,
      "loss": 0.2224,
      "step": 825
    },
    {
      "epoch": 0.28735632183908044,
      "grad_norm": 3.1413159370422363,
      "learning_rate": 9.633684210526316e-06,
      "loss": 0.1974,
      "step": 850
    },
    {
      "epoch": 0.29580797836375927,
      "grad_norm": 3.027810573577881,
      "learning_rate": 9.607368421052632e-06,
      "loss": 0.2078,
      "step": 875
    },
    {
      "epoch": 0.30425963488843816,
      "grad_norm": 2.7142438888549805,
      "learning_rate": 9.581052631578947e-06,
      "loss": 0.209,
      "step": 900
    },
    {
      "epoch": 0.312711291413117,
      "grad_norm": 2.405505895614624,
      "learning_rate": 9.554736842105264e-06,
      "loss": 0.1896,
      "step": 925
    },
    {
      "epoch": 0.3211629479377958,
      "grad_norm": 3.2011518478393555,
      "learning_rate": 9.52842105263158e-06,
      "loss": 0.1906,
      "step": 950
    },
    {
      "epoch": 0.32961460446247465,
      "grad_norm": 3.0706121921539307,
      "learning_rate": 9.502105263157896e-06,
      "loss": 0.1677,
      "step": 975
    },
    {
      "epoch": 0.3380662609871535,
      "grad_norm": 2.0819201469421387,
      "learning_rate": 9.475789473684212e-06,
      "loss": 0.1919,
      "step": 1000
    },
    {
      "epoch": 0.3380662609871535,
      "eval_loss": 0.2120433747768402,
      "eval_runtime": 8173.558,
      "eval_samples_per_second": 1.79,
      "eval_steps_per_second": 0.224,
      "eval_wer": 54.19627095935024,
      "step": 1000
    },
    {
      "epoch": 0.3465179175118323,
      "grad_norm": 4.2216315269470215,
      "learning_rate": 9.449473684210527e-06,
      "loss": 0.181,
      "step": 1025
    },
    {
      "epoch": 0.35496957403651114,
      "grad_norm": 3.4836745262145996,
      "learning_rate": 9.423157894736842e-06,
      "loss": 0.187,
      "step": 1050
    },
    {
      "epoch": 0.36342123056118997,
      "grad_norm": 2.6206018924713135,
      "learning_rate": 9.396842105263159e-06,
      "loss": 0.1634,
      "step": 1075
    },
    {
      "epoch": 0.37187288708586885,
      "grad_norm": 1.9824305772781372,
      "learning_rate": 9.370526315789474e-06,
      "loss": 0.1835,
      "step": 1100
    },
    {
      "epoch": 0.3803245436105477,
      "grad_norm": 3.2038474082946777,
      "learning_rate": 9.34421052631579e-06,
      "loss": 0.1849,
      "step": 1125
    },
    {
      "epoch": 0.3887762001352265,
      "grad_norm": 2.054469108581543,
      "learning_rate": 9.317894736842105e-06,
      "loss": 0.1739,
      "step": 1150
    },
    {
      "epoch": 0.39722785665990534,
      "grad_norm": 3.952493190765381,
      "learning_rate": 9.291578947368422e-06,
      "loss": 0.193,
      "step": 1175
    },
    {
      "epoch": 0.4056795131845842,
      "grad_norm": 2.7407076358795166,
      "learning_rate": 9.265263157894737e-06,
      "loss": 0.1709,
      "step": 1200
    },
    {
      "epoch": 0.414131169709263,
      "grad_norm": 2.922785758972168,
      "learning_rate": 9.238947368421052e-06,
      "loss": 0.1782,
      "step": 1225
    },
    {
      "epoch": 0.42258282623394183,
      "grad_norm": 2.69055438041687,
      "learning_rate": 9.21263157894737e-06,
      "loss": 0.1596,
      "step": 1250
    },
    {
      "epoch": 0.43103448275862066,
      "grad_norm": 2.2836110591888428,
      "learning_rate": 9.186315789473685e-06,
      "loss": 0.1747,
      "step": 1275
    },
    {
      "epoch": 0.43948613928329955,
      "grad_norm": 3.7077369689941406,
      "learning_rate": 9.16e-06,
      "loss": 0.1669,
      "step": 1300
    },
    {
      "epoch": 0.4479377958079784,
      "grad_norm": 2.6970176696777344,
      "learning_rate": 9.133684210526317e-06,
      "loss": 0.1501,
      "step": 1325
    },
    {
      "epoch": 0.4563894523326572,
      "grad_norm": 2.819770097732544,
      "learning_rate": 9.107368421052632e-06,
      "loss": 0.1559,
      "step": 1350
    },
    {
      "epoch": 0.46484110885733604,
      "grad_norm": 2.728736400604248,
      "learning_rate": 9.081052631578949e-06,
      "loss": 0.1865,
      "step": 1375
    },
    {
      "epoch": 0.47329276538201487,
      "grad_norm": 2.9660186767578125,
      "learning_rate": 9.054736842105264e-06,
      "loss": 0.1664,
      "step": 1400
    },
    {
      "epoch": 0.4817444219066937,
      "grad_norm": 2.8889880180358887,
      "learning_rate": 9.02842105263158e-06,
      "loss": 0.1533,
      "step": 1425
    },
    {
      "epoch": 0.49019607843137253,
      "grad_norm": 2.158355474472046,
      "learning_rate": 9.002105263157895e-06,
      "loss": 0.1494,
      "step": 1450
    },
    {
      "epoch": 0.49864773495605136,
      "grad_norm": 2.2264366149902344,
      "learning_rate": 8.97578947368421e-06,
      "loss": 0.1551,
      "step": 1475
    },
    {
      "epoch": 0.5070993914807302,
      "grad_norm": 2.8769357204437256,
      "learning_rate": 8.949473684210527e-06,
      "loss": 0.1706,
      "step": 1500
    },
    {
      "epoch": 0.5070993914807302,
      "eval_loss": 0.19238848984241486,
      "eval_runtime": 8175.5745,
      "eval_samples_per_second": 1.789,
      "eval_steps_per_second": 0.224,
      "eval_wer": 51.696552128819626,
      "step": 1500
    },
    {
      "epoch": 0.5155510480054091,
      "grad_norm": 2.3073925971984863,
      "learning_rate": 8.923157894736842e-06,
      "loss": 0.1645,
      "step": 1525
    },
    {
      "epoch": 0.5240027045300879,
      "grad_norm": 2.7228150367736816,
      "learning_rate": 8.896842105263159e-06,
      "loss": 0.1333,
      "step": 1550
    },
    {
      "epoch": 0.5324543610547667,
      "grad_norm": 1.97380530834198,
      "learning_rate": 8.870526315789474e-06,
      "loss": 0.1557,
      "step": 1575
    },
    {
      "epoch": 0.5409060175794456,
      "grad_norm": 2.8810389041900635,
      "learning_rate": 8.84421052631579e-06,
      "loss": 0.1578,
      "step": 1600
    },
    {
      "epoch": 0.5493576741041244,
      "grad_norm": 2.2714059352874756,
      "learning_rate": 8.817894736842107e-06,
      "loss": 0.138,
      "step": 1625
    },
    {
      "epoch": 0.5578093306288032,
      "grad_norm": 1.7381818294525146,
      "learning_rate": 8.791578947368422e-06,
      "loss": 0.1486,
      "step": 1650
    },
    {
      "epoch": 0.566260987153482,
      "grad_norm": 3.1913645267486572,
      "learning_rate": 8.765263157894739e-06,
      "loss": 0.1431,
      "step": 1675
    },
    {
      "epoch": 0.5747126436781609,
      "grad_norm": 2.8816912174224854,
      "learning_rate": 8.738947368421053e-06,
      "loss": 0.1592,
      "step": 1700
    },
    {
      "epoch": 0.5831643002028397,
      "grad_norm": 2.4529218673706055,
      "learning_rate": 8.712631578947368e-06,
      "loss": 0.1554,
      "step": 1725
    },
    {
      "epoch": 0.5916159567275185,
      "grad_norm": 1.9601649045944214,
      "learning_rate": 8.686315789473685e-06,
      "loss": 0.135,
      "step": 1750
    },
    {
      "epoch": 0.6000676132521975,
      "grad_norm": 2.27895188331604,
      "learning_rate": 8.66e-06,
      "loss": 0.1559,
      "step": 1775
    },
    {
      "epoch": 0.6085192697768763,
      "grad_norm": 2.6185851097106934,
      "learning_rate": 8.633684210526317e-06,
      "loss": 0.1491,
      "step": 1800
    },
    {
      "epoch": 0.6169709263015551,
      "grad_norm": 2.6344733238220215,
      "learning_rate": 8.607368421052632e-06,
      "loss": 0.1277,
      "step": 1825
    },
    {
      "epoch": 0.625422582826234,
      "grad_norm": 2.459585189819336,
      "learning_rate": 8.581052631578948e-06,
      "loss": 0.1367,
      "step": 1850
    },
    {
      "epoch": 0.6338742393509128,
      "grad_norm": 2.8337063789367676,
      "learning_rate": 8.554736842105263e-06,
      "loss": 0.1354,
      "step": 1875
    },
    {
      "epoch": 0.6423258958755916,
      "grad_norm": 1.8884564638137817,
      "learning_rate": 8.528421052631578e-06,
      "loss": 0.1335,
      "step": 1900
    },
    {
      "epoch": 0.6507775524002705,
      "grad_norm": 1.8263589143753052,
      "learning_rate": 8.502105263157897e-06,
      "loss": 0.1465,
      "step": 1925
    },
    {
      "epoch": 0.6592292089249493,
      "grad_norm": 1.995797038078308,
      "learning_rate": 8.475789473684212e-06,
      "loss": 0.141,
      "step": 1950
    },
    {
      "epoch": 0.6676808654496281,
      "grad_norm": 2.030308961868286,
      "learning_rate": 8.449473684210527e-06,
      "loss": 0.1368,
      "step": 1975
    },
    {
      "epoch": 0.676132521974307,
      "grad_norm": 1.9444704055786133,
      "learning_rate": 8.423157894736843e-06,
      "loss": 0.1398,
      "step": 2000
    },
    {
      "epoch": 0.676132521974307,
      "eval_loss": 0.18893150985240936,
      "eval_runtime": 9002.7886,
      "eval_samples_per_second": 1.625,
      "eval_steps_per_second": 0.203,
      "eval_wer": 48.848672146618874,
      "step": 2000
    },
    {
      "epoch": 0.6845841784989858,
      "grad_norm": 2.6425070762634277,
      "learning_rate": 8.396842105263158e-06,
      "loss": 0.1373,
      "step": 2025
    },
    {
      "epoch": 0.6930358350236646,
      "grad_norm": 2.495431900024414,
      "learning_rate": 8.370526315789475e-06,
      "loss": 0.143,
      "step": 2050
    },
    {
      "epoch": 0.7014874915483434,
      "grad_norm": 1.92854642868042,
      "learning_rate": 8.34421052631579e-06,
      "loss": 0.144,
      "step": 2075
    },
    {
      "epoch": 0.7099391480730223,
      "grad_norm": 2.4910595417022705,
      "learning_rate": 8.317894736842107e-06,
      "loss": 0.1222,
      "step": 2100
    },
    {
      "epoch": 0.7183908045977011,
      "grad_norm": 2.873384475708008,
      "learning_rate": 8.291578947368422e-06,
      "loss": 0.1341,
      "step": 2125
    },
    {
      "epoch": 0.7268424611223799,
      "grad_norm": 2.113551616668701,
      "learning_rate": 8.265263157894737e-06,
      "loss": 0.1372,
      "step": 2150
    },
    {
      "epoch": 0.7352941176470589,
      "grad_norm": 2.322167158126831,
      "learning_rate": 8.238947368421053e-06,
      "loss": 0.1451,
      "step": 2175
    },
    {
      "epoch": 0.7437457741717377,
      "grad_norm": 1.3647667169570923,
      "learning_rate": 8.212631578947368e-06,
      "loss": 0.1334,
      "step": 2200
    },
    {
      "epoch": 0.7521974306964165,
      "grad_norm": 1.4476187229156494,
      "learning_rate": 8.186315789473685e-06,
      "loss": 0.1293,
      "step": 2225
    },
    {
      "epoch": 0.7606490872210954,
      "grad_norm": 2.103031635284424,
      "learning_rate": 8.16e-06,
      "loss": 0.1414,
      "step": 2250
    },
    {
      "epoch": 0.7691007437457742,
      "grad_norm": 2.4936468601226807,
      "learning_rate": 8.133684210526316e-06,
      "loss": 0.1333,
      "step": 2275
    },
    {
      "epoch": 0.777552400270453,
      "grad_norm": 2.8524506092071533,
      "learning_rate": 8.107368421052633e-06,
      "loss": 0.1502,
      "step": 2300
    },
    {
      "epoch": 0.7860040567951319,
      "grad_norm": 1.78004789352417,
      "learning_rate": 8.081052631578948e-06,
      "loss": 0.1416,
      "step": 2325
    },
    {
      "epoch": 0.7944557133198107,
      "grad_norm": 2.4518020153045654,
      "learning_rate": 8.054736842105265e-06,
      "loss": 0.1305,
      "step": 2350
    },
    {
      "epoch": 0.8029073698444895,
      "grad_norm": 1.6946748495101929,
      "learning_rate": 8.02842105263158e-06,
      "loss": 0.1277,
      "step": 2375
    },
    {
      "epoch": 0.8113590263691683,
      "grad_norm": 2.044149160385132,
      "learning_rate": 8.002105263157895e-06,
      "loss": 0.1342,
      "step": 2400
    },
    {
      "epoch": 0.8198106828938472,
      "grad_norm": 2.5751612186431885,
      "learning_rate": 7.975789473684211e-06,
      "loss": 0.1382,
      "step": 2425
    },
    {
      "epoch": 0.828262339418526,
      "grad_norm": 2.957038164138794,
      "learning_rate": 7.949473684210526e-06,
      "loss": 0.1238,
      "step": 2450
    },
    {
      "epoch": 0.8367139959432048,
      "grad_norm": 1.6040945053100586,
      "learning_rate": 7.923157894736843e-06,
      "loss": 0.1382,
      "step": 2475
    },
    {
      "epoch": 0.8451656524678837,
      "grad_norm": 2.2693052291870117,
      "learning_rate": 7.896842105263158e-06,
      "loss": 0.1317,
      "step": 2500
    },
    {
      "epoch": 0.8451656524678837,
      "eval_loss": 0.1813763976097107,
      "eval_runtime": 8337.7114,
      "eval_samples_per_second": 1.754,
      "eval_steps_per_second": 0.219,
      "eval_wer": 47.81250152809494,
      "step": 2500
    },
    {
      "epoch": 0.8536173089925625,
      "grad_norm": 1.6725122928619385,
      "learning_rate": 7.870526315789475e-06,
      "loss": 0.123,
      "step": 2525
    },
    {
      "epoch": 0.8620689655172413,
      "grad_norm": 2.4779107570648193,
      "learning_rate": 7.84421052631579e-06,
      "loss": 0.1202,
      "step": 2550
    },
    {
      "epoch": 0.8705206220419203,
      "grad_norm": 2.3893909454345703,
      "learning_rate": 7.817894736842105e-06,
      "loss": 0.1197,
      "step": 2575
    },
    {
      "epoch": 0.8789722785665991,
      "grad_norm": 1.9995834827423096,
      "learning_rate": 7.791578947368423e-06,
      "loss": 0.1396,
      "step": 2600
    },
    {
      "epoch": 0.8874239350912779,
      "grad_norm": 2.6336281299591064,
      "learning_rate": 7.765263157894738e-06,
      "loss": 0.1295,
      "step": 2625
    },
    {
      "epoch": 0.8958755916159568,
      "grad_norm": 2.0509071350097656,
      "learning_rate": 7.738947368421053e-06,
      "loss": 0.1184,
      "step": 2650
    },
    {
      "epoch": 0.9043272481406356,
      "grad_norm": 2.0094475746154785,
      "learning_rate": 7.71263157894737e-06,
      "loss": 0.1275,
      "step": 2675
    },
    {
      "epoch": 0.9127789046653144,
      "grad_norm": 1.5212738513946533,
      "learning_rate": 7.686315789473685e-06,
      "loss": 0.1129,
      "step": 2700
    },
    {
      "epoch": 0.9212305611899932,
      "grad_norm": 1.1331150531768799,
      "learning_rate": 7.660000000000001e-06,
      "loss": 0.1195,
      "step": 2725
    },
    {
      "epoch": 0.9296822177146721,
      "grad_norm": 2.1462574005126953,
      "learning_rate": 7.633684210526316e-06,
      "loss": 0.1191,
      "step": 2750
    },
    {
      "epoch": 0.9381338742393509,
      "grad_norm": 2.8891022205352783,
      "learning_rate": 7.607368421052632e-06,
      "loss": 0.1369,
      "step": 2775
    },
    {
      "epoch": 0.9465855307640297,
      "grad_norm": 1.5433547496795654,
      "learning_rate": 7.581052631578948e-06,
      "loss": 0.1233,
      "step": 2800
    },
    {
      "epoch": 0.9550371872887086,
      "grad_norm": 1.6972781419754028,
      "learning_rate": 7.554736842105264e-06,
      "loss": 0.1358,
      "step": 2825
    },
    {
      "epoch": 0.9634888438133874,
      "grad_norm": 1.8598979711532593,
      "learning_rate": 7.5284210526315794e-06,
      "loss": 0.1203,
      "step": 2850
    },
    {
      "epoch": 0.9719405003380662,
      "grad_norm": 2.7386326789855957,
      "learning_rate": 7.502105263157895e-06,
      "loss": 0.1122,
      "step": 2875
    },
    {
      "epoch": 0.9803921568627451,
      "grad_norm": 1.9333444833755493,
      "learning_rate": 7.475789473684211e-06,
      "loss": 0.1278,
      "step": 2900
    },
    {
      "epoch": 0.9888438133874239,
      "grad_norm": 2.209989070892334,
      "learning_rate": 7.449473684210526e-06,
      "loss": 0.1176,
      "step": 2925
    },
    {
      "epoch": 0.9972954699121027,
      "grad_norm": 1.7954308986663818,
      "learning_rate": 7.4231578947368436e-06,
      "loss": 0.123,
      "step": 2950
    },
    {
      "epoch": 1.0057471264367817,
      "grad_norm": 2.3497698307037354,
      "learning_rate": 7.3968421052631585e-06,
      "loss": 0.1234,
      "step": 2975
    },
    {
      "epoch": 1.0141987829614605,
      "grad_norm": 2.3470804691314697,
      "learning_rate": 7.370526315789474e-06,
      "loss": 0.1015,
      "step": 3000
    },
    {
      "epoch": 1.0141987829614605,
      "eval_loss": 0.18137025833129883,
      "eval_runtime": 8405.7063,
      "eval_samples_per_second": 1.74,
      "eval_steps_per_second": 0.218,
      "eval_wer": 47.43940186696528,
      "step": 3000
    },
    {
      "epoch": 1.0226504394861393,
      "grad_norm": 1.895668864250183,
      "learning_rate": 7.34421052631579e-06,
      "loss": 0.1106,
      "step": 3025
    },
    {
      "epoch": 1.0311020960108181,
      "grad_norm": 1.3763364553451538,
      "learning_rate": 7.317894736842106e-06,
      "loss": 0.0896,
      "step": 3050
    },
    {
      "epoch": 1.039553752535497,
      "grad_norm": 1.778601884841919,
      "learning_rate": 7.291578947368422e-06,
      "loss": 0.1068,
      "step": 3075
    },
    {
      "epoch": 1.0480054090601758,
      "grad_norm": 1.7413523197174072,
      "learning_rate": 7.265263157894738e-06,
      "loss": 0.0941,
      "step": 3100
    },
    {
      "epoch": 1.0564570655848546,
      "grad_norm": 1.7844879627227783,
      "learning_rate": 7.2389473684210534e-06,
      "loss": 0.098,
      "step": 3125
    },
    {
      "epoch": 1.0649087221095335,
      "grad_norm": 1.9690433740615845,
      "learning_rate": 7.212631578947369e-06,
      "loss": 0.1019,
      "step": 3150
    },
    {
      "epoch": 1.0733603786342123,
      "grad_norm": 1.6368364095687866,
      "learning_rate": 7.186315789473684e-06,
      "loss": 0.0997,
      "step": 3175
    },
    {
      "epoch": 1.0818120351588911,
      "grad_norm": 1.9386001825332642,
      "learning_rate": 7.16e-06,
      "loss": 0.0877,
      "step": 3200
    },
    {
      "epoch": 1.09026369168357,
      "grad_norm": 1.983993411064148,
      "learning_rate": 7.133684210526316e-06,
      "loss": 0.107,
      "step": 3225
    },
    {
      "epoch": 1.0987153482082488,
      "grad_norm": 1.3441216945648193,
      "learning_rate": 7.107368421052632e-06,
      "loss": 0.1085,
      "step": 3250
    },
    {
      "epoch": 1.1071670047329276,
      "grad_norm": 2.490309238433838,
      "learning_rate": 7.0810526315789475e-06,
      "loss": 0.1067,
      "step": 3275
    },
    {
      "epoch": 1.1156186612576064,
      "grad_norm": 1.403867483139038,
      "learning_rate": 7.054736842105264e-06,
      "loss": 0.0971,
      "step": 3300
    },
    {
      "epoch": 1.1240703177822853,
      "grad_norm": 2.164853096008301,
      "learning_rate": 7.02842105263158e-06,
      "loss": 0.1001,
      "step": 3325
    },
    {
      "epoch": 1.132521974306964,
      "grad_norm": 1.6618404388427734,
      "learning_rate": 7.002105263157896e-06,
      "loss": 0.1215,
      "step": 3350
    },
    {
      "epoch": 1.140973630831643,
      "grad_norm": 0.8422718048095703,
      "learning_rate": 6.975789473684212e-06,
      "loss": 0.0979,
      "step": 3375
    },
    {
      "epoch": 1.1494252873563218,
      "grad_norm": 2.008251905441284,
      "learning_rate": 6.9494736842105275e-06,
      "loss": 0.0979,
      "step": 3400
    },
    {
      "epoch": 1.1578769438810006,
      "grad_norm": 1.493073582649231,
      "learning_rate": 6.9231578947368424e-06,
      "loss": 0.1067,
      "step": 3425
    },
    {
      "epoch": 1.1663286004056794,
      "grad_norm": 1.2973432540893555,
      "learning_rate": 6.896842105263158e-06,
      "loss": 0.0844,
      "step": 3450
    },
    {
      "epoch": 1.1747802569303583,
      "grad_norm": 1.9988205432891846,
      "learning_rate": 6.870526315789474e-06,
      "loss": 0.1018,
      "step": 3475
    },
    {
      "epoch": 1.183231913455037,
      "grad_norm": 2.0083346366882324,
      "learning_rate": 6.84421052631579e-06,
      "loss": 0.1093,
      "step": 3500
    },
    {
      "epoch": 1.183231913455037,
      "eval_loss": 0.1796853244304657,
      "eval_runtime": 8340.7798,
      "eval_samples_per_second": 1.754,
      "eval_steps_per_second": 0.219,
      "eval_wer": 46.34797533532515,
      "step": 3500
    },
    {
      "epoch": 1.1916835699797161,
      "grad_norm": 1.823431372642517,
      "learning_rate": 6.817894736842106e-06,
      "loss": 0.0996,
      "step": 3525
    },
    {
      "epoch": 1.2001352265043947,
      "grad_norm": 2.2880189418792725,
      "learning_rate": 6.7915789473684215e-06,
      "loss": 0.1032,
      "step": 3550
    },
    {
      "epoch": 1.2085868830290738,
      "grad_norm": 1.070098876953125,
      "learning_rate": 6.765263157894737e-06,
      "loss": 0.1062,
      "step": 3575
    },
    {
      "epoch": 1.2170385395537526,
      "grad_norm": 2.1815688610076904,
      "learning_rate": 6.738947368421052e-06,
      "loss": 0.1148,
      "step": 3600
    },
    {
      "epoch": 1.2254901960784315,
      "grad_norm": 2.817481756210327,
      "learning_rate": 6.71263157894737e-06,
      "loss": 0.0933,
      "step": 3625
    },
    {
      "epoch": 1.2339418526031103,
      "grad_norm": 1.8840892314910889,
      "learning_rate": 6.686315789473685e-06,
      "loss": 0.0965,
      "step": 3650
    },
    {
      "epoch": 1.2423935091277891,
      "grad_norm": 2.0833046436309814,
      "learning_rate": 6.660000000000001e-06,
      "loss": 0.1168,
      "step": 3675
    },
    {
      "epoch": 1.250845165652468,
      "grad_norm": 1.6042765378952026,
      "learning_rate": 6.6336842105263164e-06,
      "loss": 0.1033,
      "step": 3700
    },
    {
      "epoch": 1.2592968221771468,
      "grad_norm": 1.8703746795654297,
      "learning_rate": 6.607368421052632e-06,
      "loss": 0.0969,
      "step": 3725
    },
    {
      "epoch": 1.2677484787018256,
      "grad_norm": 1.789585828781128,
      "learning_rate": 6.581052631578948e-06,
      "loss": 0.1084,
      "step": 3750
    },
    {
      "epoch": 1.2762001352265044,
      "grad_norm": 1.974095106124878,
      "learning_rate": 6.554736842105264e-06,
      "loss": 0.1023,
      "step": 3775
    },
    {
      "epoch": 1.2846517917511833,
      "grad_norm": 2.729954481124878,
      "learning_rate": 6.52842105263158e-06,
      "loss": 0.1109,
      "step": 3800
    },
    {
      "epoch": 1.293103448275862,
      "grad_norm": 1.8322004079818726,
      "learning_rate": 6.5021052631578955e-06,
      "loss": 0.0994,
      "step": 3825
    },
    {
      "epoch": 1.301555104800541,
      "grad_norm": 2.05082368850708,
      "learning_rate": 6.4757894736842105e-06,
      "loss": 0.105,
      "step": 3850
    },
    {
      "epoch": 1.3100067613252198,
      "grad_norm": 1.8818949460983276,
      "learning_rate": 6.449473684210526e-06,
      "loss": 0.0872,
      "step": 3875
    },
    {
      "epoch": 1.3184584178498986,
      "grad_norm": 1.996830940246582,
      "learning_rate": 6.423157894736842e-06,
      "loss": 0.0928,
      "step": 3900
    },
    {
      "epoch": 1.3269100743745774,
      "grad_norm": 3.266448736190796,
      "learning_rate": 6.396842105263158e-06,
      "loss": 0.0972,
      "step": 3925
    },
    {
      "epoch": 1.3353617308992562,
      "grad_norm": 1.4689350128173828,
      "learning_rate": 6.370526315789474e-06,
      "loss": 0.0968,
      "step": 3950
    },
    {
      "epoch": 1.343813387423935,
      "grad_norm": 1.166890263557434,
      "learning_rate": 6.3442105263157904e-06,
      "loss": 0.1024,
      "step": 3975
    },
    {
      "epoch": 1.352265043948614,
      "grad_norm": 2.1290385723114014,
      "learning_rate": 6.317894736842106e-06,
      "loss": 0.1011,
      "step": 4000
    },
    {
      "epoch": 1.352265043948614,
      "eval_loss": 0.18109461665153503,
      "eval_runtime": 8386.0632,
      "eval_samples_per_second": 1.744,
      "eval_steps_per_second": 0.218,
      "eval_wer": 46.233062595658744,
      "step": 4000
    },
    {
      "epoch": 1.3607167004732927,
      "grad_norm": 1.8343077898025513,
      "learning_rate": 6.291578947368422e-06,
      "loss": 0.0864,
      "step": 4025
    },
    {
      "epoch": 1.3691683569979716,
      "grad_norm": 1.3331023454666138,
      "learning_rate": 6.265263157894738e-06,
      "loss": 0.0833,
      "step": 4050
    },
    {
      "epoch": 1.3776200135226504,
      "grad_norm": 2.407951831817627,
      "learning_rate": 6.238947368421054e-06,
      "loss": 0.1027,
      "step": 4075
    },
    {
      "epoch": 1.3860716700473292,
      "grad_norm": 1.5978944301605225,
      "learning_rate": 6.212631578947369e-06,
      "loss": 0.0994,
      "step": 4100
    },
    {
      "epoch": 1.394523326572008,
      "grad_norm": 1.399279236793518,
      "learning_rate": 6.1863157894736845e-06,
      "loss": 0.0916,
      "step": 4125
    },
    {
      "epoch": 1.402974983096687,
      "grad_norm": 1.4253430366516113,
      "learning_rate": 6.16e-06,
      "loss": 0.0999,
      "step": 4150
    },
    {
      "epoch": 1.4114266396213657,
      "grad_norm": 1.8193767070770264,
      "learning_rate": 6.133684210526316e-06,
      "loss": 0.099,
      "step": 4175
    },
    {
      "epoch": 1.4198782961460445,
      "grad_norm": 1.1313689947128296,
      "learning_rate": 6.107368421052632e-06,
      "loss": 0.1019,
      "step": 4200
    },
    {
      "epoch": 1.4283299526707234,
      "grad_norm": 2.423612594604492,
      "learning_rate": 6.081052631578948e-06,
      "loss": 0.0976,
      "step": 4225
    },
    {
      "epoch": 1.4367816091954024,
      "grad_norm": 1.1028584241867065,
      "learning_rate": 6.054736842105264e-06,
      "loss": 0.0904,
      "step": 4250
    },
    {
      "epoch": 1.445233265720081,
      "grad_norm": 1.811155915260315,
      "learning_rate": 6.0284210526315786e-06,
      "loss": 0.0951,
      "step": 4275
    },
    {
      "epoch": 1.45368492224476,
      "grad_norm": 1.2637566328048706,
      "learning_rate": 6.002105263157896e-06,
      "loss": 0.0963,
      "step": 4300
    },
    {
      "epoch": 1.4621365787694387,
      "grad_norm": 1.9464035034179688,
      "learning_rate": 5.975789473684212e-06,
      "loss": 0.089,
      "step": 4325
    },
    {
      "epoch": 1.4705882352941178,
      "grad_norm": 1.8550575971603394,
      "learning_rate": 5.949473684210527e-06,
      "loss": 0.0902,
      "step": 4350
    },
    {
      "epoch": 1.4790398918187964,
      "grad_norm": 2.030001640319824,
      "learning_rate": 5.923157894736843e-06,
      "loss": 0.0945,
      "step": 4375
    },
    {
      "epoch": 1.4874915483434754,
      "grad_norm": 1.992264986038208,
      "learning_rate": 5.8968421052631585e-06,
      "loss": 0.1017,
      "step": 4400
    },
    {
      "epoch": 1.495943204868154,
      "grad_norm": 1.6505546569824219,
      "learning_rate": 5.870526315789474e-06,
      "loss": 0.0944,
      "step": 4425
    },
    {
      "epoch": 1.504394861392833,
      "grad_norm": 1.857921838760376,
      "learning_rate": 5.84421052631579e-06,
      "loss": 0.1005,
      "step": 4450
    },
    {
      "epoch": 1.5128465179175117,
      "grad_norm": 1.2483346462249756,
      "learning_rate": 5.817894736842106e-06,
      "loss": 0.0851,
      "step": 4475
    },
    {
      "epoch": 1.5212981744421907,
      "grad_norm": 2.141049861907959,
      "learning_rate": 5.791578947368422e-06,
      "loss": 0.0952,
      "step": 4500
    },
    {
      "epoch": 1.5212981744421907,
      "eval_loss": 0.18032518029212952,
      "eval_runtime": 8277.5716,
      "eval_samples_per_second": 1.767,
      "eval_steps_per_second": 0.221,
      "eval_wer": 45.674635579918146,
      "step": 4500
    },
    {
      "epoch": 1.5297498309668696,
      "grad_norm": 2.129145622253418,
      "learning_rate": 5.765263157894737e-06,
      "loss": 0.0998,
      "step": 4525
    },
    {
      "epoch": 1.5382014874915484,
      "grad_norm": 1.1393293142318726,
      "learning_rate": 5.7389473684210526e-06,
      "loss": 0.1009,
      "step": 4550
    },
    {
      "epoch": 1.5466531440162272,
      "grad_norm": 2.5733561515808105,
      "learning_rate": 5.712631578947368e-06,
      "loss": 0.0986,
      "step": 4575
    },
    {
      "epoch": 1.555104800540906,
      "grad_norm": 1.9032436609268188,
      "learning_rate": 5.686315789473684e-06,
      "loss": 0.0871,
      "step": 4600
    },
    {
      "epoch": 1.5635564570655849,
      "grad_norm": 1.4245878458023071,
      "learning_rate": 5.66e-06,
      "loss": 0.1071,
      "step": 4625
    },
    {
      "epoch": 1.5720081135902637,
      "grad_norm": 1.322015404701233,
      "learning_rate": 5.633684210526317e-06,
      "loss": 0.0856,
      "step": 4650
    },
    {
      "epoch": 1.5804597701149425,
      "grad_norm": 2.069826126098633,
      "learning_rate": 5.6073684210526325e-06,
      "loss": 0.0802,
      "step": 4675
    },
    {
      "epoch": 1.5889114266396214,
      "grad_norm": 1.566107988357544,
      "learning_rate": 5.581052631578948e-06,
      "loss": 0.0896,
      "step": 4700
    },
    {
      "epoch": 1.5973630831643002,
      "grad_norm": 2.0934951305389404,
      "learning_rate": 5.554736842105264e-06,
      "loss": 0.0948,
      "step": 4725
    },
    {
      "epoch": 1.605814739688979,
      "grad_norm": 1.2778680324554443,
      "learning_rate": 5.52842105263158e-06,
      "loss": 0.0914,
      "step": 4750
    },
    {
      "epoch": 1.6142663962136579,
      "grad_norm": 2.1123738288879395,
      "learning_rate": 5.502105263157895e-06,
      "loss": 0.0981,
      "step": 4775
    },
    {
      "epoch": 1.6227180527383367,
      "grad_norm": 0.9122495651245117,
      "learning_rate": 5.475789473684211e-06,
      "loss": 0.0832,
      "step": 4800
    },
    {
      "epoch": 1.6311697092630155,
      "grad_norm": 1.6515990495681763,
      "learning_rate": 5.4494736842105266e-06,
      "loss": 0.0975,
      "step": 4825
    },
    {
      "epoch": 1.6396213657876944,
      "grad_norm": 1.507886528968811,
      "learning_rate": 5.423157894736842e-06,
      "loss": 0.0934,
      "step": 4850
    },
    {
      "epoch": 1.6480730223123732,
      "grad_norm": 1.307924747467041,
      "learning_rate": 5.396842105263158e-06,
      "loss": 0.0855,
      "step": 4875
    },
    {
      "epoch": 1.656524678837052,
      "grad_norm": 1.2642782926559448,
      "learning_rate": 5.370526315789474e-06,
      "loss": 0.1,
      "step": 4900
    },
    {
      "epoch": 1.664976335361731,
      "grad_norm": 1.7240660190582275,
      "learning_rate": 5.34421052631579e-06,
      "loss": 0.0925,
      "step": 4925
    },
    {
      "epoch": 1.6734279918864097,
      "grad_norm": 1.4809589385986328,
      "learning_rate": 5.317894736842105e-06,
      "loss": 0.1017,
      "step": 4950
    },
    {
      "epoch": 1.6818796484110887,
      "grad_norm": 2.2714500427246094,
      "learning_rate": 5.291578947368422e-06,
      "loss": 0.0908,
      "step": 4975
    },
    {
      "epoch": 1.6903313049357673,
      "grad_norm": 1.7016249895095825,
      "learning_rate": 5.265263157894738e-06,
      "loss": 0.0959,
      "step": 5000
    },
    {
      "epoch": 1.6903313049357673,
      "eval_loss": 0.18442150950431824,
      "eval_runtime": 8017.4704,
      "eval_samples_per_second": 1.824,
      "eval_steps_per_second": 0.228,
      "eval_wer": 45.91424086688215,
      "step": 5000
    },
    {
      "epoch": 1.6987829614604464,
      "grad_norm": 1.7705456018447876,
      "learning_rate": 5.238947368421053e-06,
      "loss": 0.1014,
      "step": 5025
    },
    {
      "epoch": 1.707234617985125,
      "grad_norm": 1.8674064874649048,
      "learning_rate": 5.212631578947369e-06,
      "loss": 0.1012,
      "step": 5050
    },
    {
      "epoch": 1.715686274509804,
      "grad_norm": 1.4433462619781494,
      "learning_rate": 5.186315789473685e-06,
      "loss": 0.0859,
      "step": 5075
    },
    {
      "epoch": 1.7241379310344827,
      "grad_norm": 1.2248002290725708,
      "learning_rate": 5.1600000000000006e-06,
      "loss": 0.0884,
      "step": 5100
    },
    {
      "epoch": 1.7325895875591617,
      "grad_norm": 1.763519525527954,
      "learning_rate": 5.133684210526316e-06,
      "loss": 0.1033,
      "step": 5125
    },
    {
      "epoch": 1.7410412440838403,
      "grad_norm": 1.8742742538452148,
      "learning_rate": 5.107368421052632e-06,
      "loss": 0.0994,
      "step": 5150
    },
    {
      "epoch": 1.7494929006085194,
      "grad_norm": 1.9813201427459717,
      "learning_rate": 5.081052631578948e-06,
      "loss": 0.0912,
      "step": 5175
    },
    {
      "epoch": 1.757944557133198,
      "grad_norm": 2.3702893257141113,
      "learning_rate": 5.054736842105263e-06,
      "loss": 0.0934,
      "step": 5200
    },
    {
      "epoch": 1.766396213657877,
      "grad_norm": 1.5483486652374268,
      "learning_rate": 5.028421052631579e-06,
      "loss": 0.1052,
      "step": 5225
    },
    {
      "epoch": 1.7748478701825556,
      "grad_norm": 1.4521353244781494,
      "learning_rate": 5.002105263157895e-06,
      "loss": 0.1046,
      "step": 5250
    },
    {
      "epoch": 1.7832995267072347,
      "grad_norm": 2.0621144771575928,
      "learning_rate": 4.975789473684211e-06,
      "loss": 0.0996,
      "step": 5275
    },
    {
      "epoch": 1.7917511832319133,
      "grad_norm": 2.86627197265625,
      "learning_rate": 4.949473684210527e-06,
      "loss": 0.0974,
      "step": 5300
    },
    {
      "epoch": 1.8002028397565923,
      "grad_norm": 1.1360923051834106,
      "learning_rate": 4.923157894736842e-06,
      "loss": 0.0874,
      "step": 5325
    },
    {
      "epoch": 1.8086544962812712,
      "grad_norm": 1.4977036714553833,
      "learning_rate": 4.896842105263158e-06,
      "loss": 0.0926,
      "step": 5350
    },
    {
      "epoch": 1.81710615280595,
      "grad_norm": 1.642228126525879,
      "learning_rate": 4.870526315789474e-06,
      "loss": 0.0878,
      "step": 5375
    },
    {
      "epoch": 1.8255578093306288,
      "grad_norm": 1.6537584066390991,
      "learning_rate": 4.84421052631579e-06,
      "loss": 0.088,
      "step": 5400
    },
    {
      "epoch": 1.8340094658553077,
      "grad_norm": 1.3952895402908325,
      "learning_rate": 4.817894736842106e-06,
      "loss": 0.0888,
      "step": 5425
    },
    {
      "epoch": 1.8424611223799865,
      "grad_norm": 1.2289971113204956,
      "learning_rate": 4.791578947368421e-06,
      "loss": 0.0917,
      "step": 5450
    },
    {
      "epoch": 1.8509127789046653,
      "grad_norm": 1.2816951274871826,
      "learning_rate": 4.765263157894737e-06,
      "loss": 0.0851,
      "step": 5475
    },
    {
      "epoch": 1.8593644354293442,
      "grad_norm": 1.385846495628357,
      "learning_rate": 4.738947368421053e-06,
      "loss": 0.0828,
      "step": 5500
    },
    {
      "epoch": 1.8593644354293442,
      "eval_loss": 0.18521299958229065,
      "eval_runtime": 8001.8096,
      "eval_samples_per_second": 1.828,
      "eval_steps_per_second": 0.229,
      "eval_wer": 45.38564226441666,
      "step": 5500
    },
    {
      "epoch": 1.867816091954023,
      "grad_norm": 1.7502315044403076,
      "learning_rate": 4.712631578947369e-06,
      "loss": 0.0931,
      "step": 5525
    },
    {
      "epoch": 1.8762677484787018,
      "grad_norm": 1.384947657585144,
      "learning_rate": 4.6863157894736845e-06,
      "loss": 0.081,
      "step": 5550
    },
    {
      "epoch": 1.8847194050033806,
      "grad_norm": 1.736101508140564,
      "learning_rate": 4.66e-06,
      "loss": 0.1042,
      "step": 5575
    },
    {
      "epoch": 1.8931710615280595,
      "grad_norm": 1.5998270511627197,
      "learning_rate": 4.633684210526316e-06,
      "loss": 0.0784,
      "step": 5600
    },
    {
      "epoch": 1.9016227180527383,
      "grad_norm": 1.7052019834518433,
      "learning_rate": 4.607368421052632e-06,
      "loss": 0.0867,
      "step": 5625
    },
    {
      "epoch": 1.9100743745774171,
      "grad_norm": 1.8424391746520996,
      "learning_rate": 4.581052631578948e-06,
      "loss": 0.1042,
      "step": 5650
    },
    {
      "epoch": 1.918526031102096,
      "grad_norm": 1.484885573387146,
      "learning_rate": 4.5547368421052636e-06,
      "loss": 0.089,
      "step": 5675
    },
    {
      "epoch": 1.9269776876267748,
      "grad_norm": 1.7894026041030884,
      "learning_rate": 4.528421052631579e-06,
      "loss": 0.0847,
      "step": 5700
    },
    {
      "epoch": 1.9354293441514536,
      "grad_norm": 1.5164501667022705,
      "learning_rate": 4.502105263157895e-06,
      "loss": 0.1001,
      "step": 5725
    },
    {
      "epoch": 1.9438810006761327,
      "grad_norm": 0.9590908885002136,
      "learning_rate": 4.475789473684211e-06,
      "loss": 0.0858,
      "step": 5750
    },
    {
      "epoch": 1.9523326572008113,
      "grad_norm": 1.887405514717102,
      "learning_rate": 4.449473684210527e-06,
      "loss": 0.079,
      "step": 5775
    },
    {
      "epoch": 1.9607843137254903,
      "grad_norm": 2.0071308612823486,
      "learning_rate": 4.423157894736843e-06,
      "loss": 0.094,
      "step": 5800
    },
    {
      "epoch": 1.969235970250169,
      "grad_norm": 2.262946605682373,
      "learning_rate": 4.3968421052631585e-06,
      "loss": 0.0992,
      "step": 5825
    },
    {
      "epoch": 1.977687626774848,
      "grad_norm": 1.1825026273727417,
      "learning_rate": 4.370526315789474e-06,
      "loss": 0.0831,
      "step": 5850
    },
    {
      "epoch": 1.9861392832995266,
      "grad_norm": 1.4014040231704712,
      "learning_rate": 4.344210526315789e-06,
      "loss": 0.0983,
      "step": 5875
    },
    {
      "epoch": 1.9945909398242057,
      "grad_norm": 1.7053906917572021,
      "learning_rate": 4.317894736842105e-06,
      "loss": 0.0937,
      "step": 5900
    },
    {
      "epoch": 2.0030425963488843,
      "grad_norm": 1.2354754209518433,
      "learning_rate": 4.291578947368422e-06,
      "loss": 0.0794,
      "step": 5925
    },
    {
      "epoch": 2.0114942528735633,
      "grad_norm": 1.5480238199234009,
      "learning_rate": 4.2652631578947376e-06,
      "loss": 0.0621,
      "step": 5950
    },
    {
      "epoch": 2.019945909398242,
      "grad_norm": 1.6204942464828491,
      "learning_rate": 4.238947368421053e-06,
      "loss": 0.0679,
      "step": 5975
    },
    {
      "epoch": 2.028397565922921,
      "grad_norm": 1.8844307661056519,
      "learning_rate": 4.212631578947368e-06,
      "loss": 0.0814,
      "step": 6000
    },
    {
      "epoch": 2.028397565922921,
      "eval_loss": 0.18354202806949615,
      "eval_runtime": 8220.3894,
      "eval_samples_per_second": 1.779,
      "eval_steps_per_second": 0.222,
      "eval_wer": 45.245302024909165,
      "step": 6000
    },
    {
      "epoch": 2.0368492224475996,
      "grad_norm": 1.34647798538208,
      "learning_rate": 4.186315789473684e-06,
      "loss": 0.0707,
      "step": 6025
    },
    {
      "epoch": 2.0453008789722786,
      "grad_norm": 2.5605897903442383,
      "learning_rate": 4.16e-06,
      "loss": 0.0658,
      "step": 6050
    },
    {
      "epoch": 2.0537525354969572,
      "grad_norm": 0.988107442855835,
      "learning_rate": 4.133684210526316e-06,
      "loss": 0.0787,
      "step": 6075
    },
    {
      "epoch": 2.0622041920216363,
      "grad_norm": 1.3938881158828735,
      "learning_rate": 4.1073684210526325e-06,
      "loss": 0.0772,
      "step": 6100
    },
    {
      "epoch": 2.070655848546315,
      "grad_norm": 0.3137703835964203,
      "learning_rate": 4.0810526315789474e-06,
      "loss": 0.0728,
      "step": 6125
    },
    {
      "epoch": 2.079107505070994,
      "grad_norm": 1.0805671215057373,
      "learning_rate": 4.054736842105263e-06,
      "loss": 0.0723,
      "step": 6150
    },
    {
      "epoch": 2.0875591615956726,
      "grad_norm": 1.8240491151809692,
      "learning_rate": 4.029473684210527e-06,
      "loss": 0.0824,
      "step": 6175
    },
    {
      "epoch": 2.0960108181203516,
      "grad_norm": 0.9932010173797607,
      "learning_rate": 4.0031578947368424e-06,
      "loss": 0.0808,
      "step": 6200
    },
    {
      "epoch": 2.1044624746450302,
      "grad_norm": 1.7233359813690186,
      "learning_rate": 3.976842105263158e-06,
      "loss": 0.0656,
      "step": 6225
    },
    {
      "epoch": 2.1129141311697093,
      "grad_norm": 1.3749310970306396,
      "learning_rate": 3.950526315789474e-06,
      "loss": 0.0782,
      "step": 6250
    },
    {
      "epoch": 2.121365787694388,
      "grad_norm": 1.1162022352218628,
      "learning_rate": 3.92421052631579e-06,
      "loss": 0.0764,
      "step": 6275
    },
    {
      "epoch": 2.129817444219067,
      "grad_norm": 2.195281505584717,
      "learning_rate": 3.897894736842106e-06,
      "loss": 0.0694,
      "step": 6300
    },
    {
      "epoch": 2.138269100743746,
      "grad_norm": 1.5847153663635254,
      "learning_rate": 3.8715789473684215e-06,
      "loss": 0.0778,
      "step": 6325
    },
    {
      "epoch": 2.1467207572684246,
      "grad_norm": 1.5322215557098389,
      "learning_rate": 3.845263157894737e-06,
      "loss": 0.068,
      "step": 6350
    },
    {
      "epoch": 2.1551724137931036,
      "grad_norm": 2.1042368412017822,
      "learning_rate": 3.818947368421053e-06,
      "loss": 0.063,
      "step": 6375
    },
    {
      "epoch": 2.1636240703177823,
      "grad_norm": 2.016451120376587,
      "learning_rate": 3.792631578947369e-06,
      "loss": 0.0757,
      "step": 6400
    },
    {
      "epoch": 2.1720757268424613,
      "grad_norm": 1.9875035285949707,
      "learning_rate": 3.766315789473685e-06,
      "loss": 0.0633,
      "step": 6425
    },
    {
      "epoch": 2.18052738336714,
      "grad_norm": 1.0881574153900146,
      "learning_rate": 3.74e-06,
      "loss": 0.0779,
      "step": 6450
    },
    {
      "epoch": 2.188979039891819,
      "grad_norm": 1.4989287853240967,
      "learning_rate": 3.713684210526316e-06,
      "loss": 0.0672,
      "step": 6475
    },
    {
      "epoch": 2.1974306964164976,
      "grad_norm": 1.430418848991394,
      "learning_rate": 3.687368421052632e-06,
      "loss": 0.0903,
      "step": 6500
    },
    {
      "epoch": 2.1974306964164976,
      "eval_loss": 0.18892446160316467,
      "eval_runtime": 8703.8448,
      "eval_samples_per_second": 1.681,
      "eval_steps_per_second": 0.21,
      "eval_wer": 45.42427250455984,
      "step": 6500
    },
    {
      "epoch": 2.2058823529411766,
      "grad_norm": 1.1100589036941528,
      "learning_rate": 3.6610526315789472e-06,
      "loss": 0.0682,
      "step": 6525
    },
    {
      "epoch": 2.2143340094658552,
      "grad_norm": 1.5231108665466309,
      "learning_rate": 3.6347368421052635e-06,
      "loss": 0.0772,
      "step": 6550
    },
    {
      "epoch": 2.2227856659905343,
      "grad_norm": 1.5531678199768066,
      "learning_rate": 3.6084210526315793e-06,
      "loss": 0.0769,
      "step": 6575
    },
    {
      "epoch": 2.231237322515213,
      "grad_norm": 2.173323154449463,
      "learning_rate": 3.582105263157895e-06,
      "loss": 0.0753,
      "step": 6600
    },
    {
      "epoch": 2.239688979039892,
      "grad_norm": 1.8664982318878174,
      "learning_rate": 3.555789473684211e-06,
      "loss": 0.0732,
      "step": 6625
    },
    {
      "epoch": 2.2481406355645706,
      "grad_norm": 1.264046549797058,
      "learning_rate": 3.5294736842105263e-06,
      "loss": 0.0727,
      "step": 6650
    },
    {
      "epoch": 2.2565922920892496,
      "grad_norm": 0.6508150100708008,
      "learning_rate": 3.503157894736842e-06,
      "loss": 0.0627,
      "step": 6675
    },
    {
      "epoch": 2.265043948613928,
      "grad_norm": 2.036755323410034,
      "learning_rate": 3.476842105263158e-06,
      "loss": 0.0703,
      "step": 6700
    },
    {
      "epoch": 2.2734956051386073,
      "grad_norm": 1.4108197689056396,
      "learning_rate": 3.450526315789474e-06,
      "loss": 0.0815,
      "step": 6725
    },
    {
      "epoch": 2.281947261663286,
      "grad_norm": 1.388468623161316,
      "learning_rate": 3.42421052631579e-06,
      "loss": 0.0678,
      "step": 6750
    },
    {
      "epoch": 2.290398918187965,
      "grad_norm": 1.7254157066345215,
      "learning_rate": 3.3978947368421054e-06,
      "loss": 0.0713,
      "step": 6775
    },
    {
      "epoch": 2.2988505747126435,
      "grad_norm": 1.8082466125488281,
      "learning_rate": 3.3715789473684212e-06,
      "loss": 0.0643,
      "step": 6800
    },
    {
      "epoch": 2.3073022312373226,
      "grad_norm": 1.5543231964111328,
      "learning_rate": 3.345263157894737e-06,
      "loss": 0.0693,
      "step": 6825
    },
    {
      "epoch": 2.315753887762001,
      "grad_norm": 2.165055513381958,
      "learning_rate": 3.318947368421053e-06,
      "loss": 0.0718,
      "step": 6850
    },
    {
      "epoch": 2.3242055442866802,
      "grad_norm": 1.7937675714492798,
      "learning_rate": 3.292631578947369e-06,
      "loss": 0.0694,
      "step": 6875
    },
    {
      "epoch": 2.332657200811359,
      "grad_norm": 1.602084994316101,
      "learning_rate": 3.2663157894736845e-06,
      "loss": 0.0575,
      "step": 6900
    },
    {
      "epoch": 2.341108857336038,
      "grad_norm": 0.9363685846328735,
      "learning_rate": 3.2400000000000003e-06,
      "loss": 0.0663,
      "step": 6925
    },
    {
      "epoch": 2.3495605138607165,
      "grad_norm": 1.2168769836425781,
      "learning_rate": 3.213684210526316e-06,
      "loss": 0.0755,
      "step": 6950
    },
    {
      "epoch": 2.3580121703853956,
      "grad_norm": 1.4504033327102661,
      "learning_rate": 3.187368421052632e-06,
      "loss": 0.079,
      "step": 6975
    },
    {
      "epoch": 2.366463826910074,
      "grad_norm": 1.5413113832473755,
      "learning_rate": 3.1610526315789474e-06,
      "loss": 0.0712,
      "step": 7000
    },
    {
      "epoch": 2.366463826910074,
      "eval_loss": 0.19058801233768463,
      "eval_runtime": 8456.0898,
      "eval_samples_per_second": 1.73,
      "eval_steps_per_second": 0.216,
      "eval_wer": 45.145058996689535,
      "step": 7000
    },
    {
      "epoch": 2.3749154834347532,
      "grad_norm": 1.0616259574890137,
      "learning_rate": 3.134736842105263e-06,
      "loss": 0.066,
      "step": 7025
    },
    {
      "epoch": 2.3833671399594323,
      "grad_norm": 1.821663737297058,
      "learning_rate": 3.1084210526315794e-06,
      "loss": 0.0692,
      "step": 7050
    },
    {
      "epoch": 2.391818796484111,
      "grad_norm": 1.2044411897659302,
      "learning_rate": 3.0821052631578952e-06,
      "loss": 0.0631,
      "step": 7075
    },
    {
      "epoch": 2.4002704530087895,
      "grad_norm": 1.1496453285217285,
      "learning_rate": 3.055789473684211e-06,
      "loss": 0.0598,
      "step": 7100
    },
    {
      "epoch": 2.4087221095334685,
      "grad_norm": 1.834873914718628,
      "learning_rate": 3.0294736842105264e-06,
      "loss": 0.0696,
      "step": 7125
    },
    {
      "epoch": 2.4171737660581476,
      "grad_norm": 1.9280352592468262,
      "learning_rate": 3.0031578947368423e-06,
      "loss": 0.0937,
      "step": 7150
    },
    {
      "epoch": 2.425625422582826,
      "grad_norm": 1.8316441774368286,
      "learning_rate": 2.976842105263158e-06,
      "loss": 0.0843,
      "step": 7175
    },
    {
      "epoch": 2.4340770791075053,
      "grad_norm": 1.6660900115966797,
      "learning_rate": 2.9505263157894735e-06,
      "loss": 0.0747,
      "step": 7200
    },
    {
      "epoch": 2.442528735632184,
      "grad_norm": 1.2995704412460327,
      "learning_rate": 2.92421052631579e-06,
      "loss": 0.0697,
      "step": 7225
    },
    {
      "epoch": 2.450980392156863,
      "grad_norm": 0.837418794631958,
      "learning_rate": 2.8978947368421055e-06,
      "loss": 0.0532,
      "step": 7250
    },
    {
      "epoch": 2.4594320486815415,
      "grad_norm": 1.582984447479248,
      "learning_rate": 2.8715789473684214e-06,
      "loss": 0.0785,
      "step": 7275
    },
    {
      "epoch": 2.4678837052062206,
      "grad_norm": 2.5198307037353516,
      "learning_rate": 2.845263157894737e-06,
      "loss": 0.0695,
      "step": 7300
    },
    {
      "epoch": 2.476335361730899,
      "grad_norm": 1.9426721334457397,
      "learning_rate": 2.8189473684210526e-06,
      "loss": 0.0683,
      "step": 7325
    },
    {
      "epoch": 2.4847870182555782,
      "grad_norm": 1.8814282417297363,
      "learning_rate": 2.7926315789473684e-06,
      "loss": 0.0635,
      "step": 7350
    },
    {
      "epoch": 2.493238674780257,
      "grad_norm": 1.7943637371063232,
      "learning_rate": 2.766315789473684e-06,
      "loss": 0.0703,
      "step": 7375
    },
    {
      "epoch": 2.501690331304936,
      "grad_norm": 1.7339930534362793,
      "learning_rate": 2.7400000000000004e-06,
      "loss": 0.0834,
      "step": 7400
    },
    {
      "epoch": 2.5101419878296145,
      "grad_norm": 0.6923316121101379,
      "learning_rate": 2.7136842105263163e-06,
      "loss": 0.0749,
      "step": 7425
    },
    {
      "epoch": 2.5185936443542936,
      "grad_norm": 1.8081867694854736,
      "learning_rate": 2.6873684210526317e-06,
      "loss": 0.0832,
      "step": 7450
    },
    {
      "epoch": 2.527045300878972,
      "grad_norm": 0.8502326607704163,
      "learning_rate": 2.6610526315789475e-06,
      "loss": 0.072,
      "step": 7475
    },
    {
      "epoch": 2.535496957403651,
      "grad_norm": 2.073284864425659,
      "learning_rate": 2.6347368421052633e-06,
      "loss": 0.0937,
      "step": 7500
    },
    {
      "epoch": 2.535496957403651,
      "eval_loss": 0.18810197710990906,
      "eval_runtime": 8587.5354,
      "eval_samples_per_second": 1.703,
      "eval_steps_per_second": 0.213,
      "eval_wer": 44.96070962284172,
      "step": 7500
    },
    {
      "epoch": 2.54394861392833,
      "grad_norm": 1.5389723777770996,
      "learning_rate": 2.608421052631579e-06,
      "loss": 0.0695,
      "step": 7525
    },
    {
      "epoch": 2.552400270453009,
      "grad_norm": 1.7400459051132202,
      "learning_rate": 2.5821052631578954e-06,
      "loss": 0.0752,
      "step": 7550
    },
    {
      "epoch": 2.5608519269776875,
      "grad_norm": 0.9405755400657654,
      "learning_rate": 2.5557894736842108e-06,
      "loss": 0.0728,
      "step": 7575
    },
    {
      "epoch": 2.5693035835023665,
      "grad_norm": 2.09281849861145,
      "learning_rate": 2.5294736842105266e-06,
      "loss": 0.0785,
      "step": 7600
    },
    {
      "epoch": 2.577755240027045,
      "grad_norm": 1.1472514867782593,
      "learning_rate": 2.5031578947368424e-06,
      "loss": 0.0763,
      "step": 7625
    },
    {
      "epoch": 2.586206896551724,
      "grad_norm": 1.5787365436553955,
      "learning_rate": 2.476842105263158e-06,
      "loss": 0.0782,
      "step": 7650
    },
    {
      "epoch": 2.5946585530764033,
      "grad_norm": 0.8288754820823669,
      "learning_rate": 2.4505263157894736e-06,
      "loss": 0.0617,
      "step": 7675
    },
    {
      "epoch": 2.603110209601082,
      "grad_norm": 1.3931238651275635,
      "learning_rate": 2.42421052631579e-06,
      "loss": 0.0652,
      "step": 7700
    },
    {
      "epoch": 2.6115618661257605,
      "grad_norm": 1.8449468612670898,
      "learning_rate": 2.3978947368421052e-06,
      "loss": 0.0708,
      "step": 7725
    },
    {
      "epoch": 2.6200135226504395,
      "grad_norm": 1.462990641593933,
      "learning_rate": 2.371578947368421e-06,
      "loss": 0.0799,
      "step": 7750
    },
    {
      "epoch": 2.6284651791751186,
      "grad_norm": 1.80518639087677,
      "learning_rate": 2.3452631578947373e-06,
      "loss": 0.0712,
      "step": 7775
    },
    {
      "epoch": 2.636916835699797,
      "grad_norm": 1.4477194547653198,
      "learning_rate": 2.3189473684210527e-06,
      "loss": 0.0748,
      "step": 7800
    },
    {
      "epoch": 2.645368492224476,
      "grad_norm": 1.8457978963851929,
      "learning_rate": 2.2926315789473685e-06,
      "loss": 0.0652,
      "step": 7825
    },
    {
      "epoch": 2.653820148749155,
      "grad_norm": 1.2285178899765015,
      "learning_rate": 2.2663157894736843e-06,
      "loss": 0.051,
      "step": 7850
    },
    {
      "epoch": 2.662271805273834,
      "grad_norm": 1.664722204208374,
      "learning_rate": 2.24e-06,
      "loss": 0.0696,
      "step": 7875
    },
    {
      "epoch": 2.6707234617985125,
      "grad_norm": 1.1485116481781006,
      "learning_rate": 2.213684210526316e-06,
      "loss": 0.078,
      "step": 7900
    },
    {
      "epoch": 2.679175118323191,
      "grad_norm": 1.6024932861328125,
      "learning_rate": 2.187368421052632e-06,
      "loss": 0.0811,
      "step": 7925
    },
    {
      "epoch": 2.68762677484787,
      "grad_norm": 1.3482303619384766,
      "learning_rate": 2.1610526315789476e-06,
      "loss": 0.0744,
      "step": 7950
    },
    {
      "epoch": 2.696078431372549,
      "grad_norm": 1.4488401412963867,
      "learning_rate": 2.1347368421052634e-06,
      "loss": 0.0717,
      "step": 7975
    },
    {
      "epoch": 2.704530087897228,
      "grad_norm": 2.0738377571105957,
      "learning_rate": 2.1084210526315792e-06,
      "loss": 0.0637,
      "step": 8000
    },
    {
      "epoch": 2.704530087897228,
      "eval_loss": 0.19148404896259308,
      "eval_runtime": 8546.7321,
      "eval_samples_per_second": 1.711,
      "eval_steps_per_second": 0.214,
      "eval_wer": 44.85362072928025,
      "step": 8000
    },
    {
      "epoch": 2.7129817444219064,
      "grad_norm": 1.608497142791748,
      "learning_rate": 2.082105263157895e-06,
      "loss": 0.0725,
      "step": 8025
    },
    {
      "epoch": 2.7214334009465855,
      "grad_norm": 2.1245532035827637,
      "learning_rate": 2.055789473684211e-06,
      "loss": 0.0726,
      "step": 8050
    },
    {
      "epoch": 2.7298850574712645,
      "grad_norm": 1.3856103420257568,
      "learning_rate": 2.0294736842105263e-06,
      "loss": 0.0731,
      "step": 8075
    },
    {
      "epoch": 2.738336713995943,
      "grad_norm": 1.2372052669525146,
      "learning_rate": 2.003157894736842e-06,
      "loss": 0.0756,
      "step": 8100
    },
    {
      "epoch": 2.746788370520622,
      "grad_norm": 1.6612149477005005,
      "learning_rate": 1.976842105263158e-06,
      "loss": 0.0636,
      "step": 8125
    },
    {
      "epoch": 2.755240027045301,
      "grad_norm": 1.719821572303772,
      "learning_rate": 1.9505263157894737e-06,
      "loss": 0.0804,
      "step": 8150
    },
    {
      "epoch": 2.76369168356998,
      "grad_norm": 1.575764536857605,
      "learning_rate": 1.9242105263157896e-06,
      "loss": 0.0778,
      "step": 8175
    },
    {
      "epoch": 2.7721433400946585,
      "grad_norm": 1.507033348083496,
      "learning_rate": 1.8978947368421056e-06,
      "loss": 0.0643,
      "step": 8200
    },
    {
      "epoch": 2.7805949966193375,
      "grad_norm": 1.1668611764907837,
      "learning_rate": 1.8715789473684212e-06,
      "loss": 0.0867,
      "step": 8225
    },
    {
      "epoch": 2.789046653144016,
      "grad_norm": 0.9894188046455383,
      "learning_rate": 1.845263157894737e-06,
      "loss": 0.0648,
      "step": 8250
    },
    {
      "epoch": 2.797498309668695,
      "grad_norm": 1.6405720710754395,
      "learning_rate": 1.8189473684210528e-06,
      "loss": 0.0708,
      "step": 8275
    },
    {
      "epoch": 2.805949966193374,
      "grad_norm": 1.034415364265442,
      "learning_rate": 1.7926315789473686e-06,
      "loss": 0.0546,
      "step": 8300
    },
    {
      "epoch": 2.814401622718053,
      "grad_norm": 0.9419425129890442,
      "learning_rate": 1.7663157894736843e-06,
      "loss": 0.07,
      "step": 8325
    },
    {
      "epoch": 2.8228532792427314,
      "grad_norm": 2.342827558517456,
      "learning_rate": 1.74e-06,
      "loss": 0.0697,
      "step": 8350
    },
    {
      "epoch": 2.8313049357674105,
      "grad_norm": 1.535322904586792,
      "learning_rate": 1.713684210526316e-06,
      "loss": 0.0696,
      "step": 8375
    },
    {
      "epoch": 2.839756592292089,
      "grad_norm": 1.6388691663742065,
      "learning_rate": 1.6873684210526317e-06,
      "loss": 0.0736,
      "step": 8400
    },
    {
      "epoch": 2.848208248816768,
      "grad_norm": 1.7451198101043701,
      "learning_rate": 1.6610526315789473e-06,
      "loss": 0.0751,
      "step": 8425
    },
    {
      "epoch": 2.8566599053414468,
      "grad_norm": 1.4245884418487549,
      "learning_rate": 1.6347368421052633e-06,
      "loss": 0.0797,
      "step": 8450
    },
    {
      "epoch": 2.865111561866126,
      "grad_norm": 0.799862265586853,
      "learning_rate": 1.6084210526315792e-06,
      "loss": 0.0664,
      "step": 8475
    },
    {
      "epoch": 2.873563218390805,
      "grad_norm": 1.09870183467865,
      "learning_rate": 1.5821052631578948e-06,
      "loss": 0.062,
      "step": 8500
    },
    {
      "epoch": 2.873563218390805,
      "eval_loss": 0.1915096938610077,
      "eval_runtime": 8477.0001,
      "eval_samples_per_second": 1.725,
      "eval_steps_per_second": 0.216,
      "eval_wer": 44.89322895018655,
      "step": 8500
    },
    {
      "epoch": 2.8820148749154835,
      "grad_norm": 1.4032130241394043,
      "learning_rate": 1.5557894736842106e-06,
      "loss": 0.0743,
      "step": 8525
    },
    {
      "epoch": 2.890466531440162,
      "grad_norm": 0.7570385932922363,
      "learning_rate": 1.5294736842105264e-06,
      "loss": 0.0687,
      "step": 8550
    },
    {
      "epoch": 2.898918187964841,
      "grad_norm": 1.0193418264389038,
      "learning_rate": 1.5031578947368422e-06,
      "loss": 0.0615,
      "step": 8575
    },
    {
      "epoch": 2.90736984448952,
      "grad_norm": 1.9574233293533325,
      "learning_rate": 1.4768421052631578e-06,
      "loss": 0.0835,
      "step": 8600
    },
    {
      "epoch": 2.915821501014199,
      "grad_norm": 1.0290601253509521,
      "learning_rate": 1.4505263157894739e-06,
      "loss": 0.0678,
      "step": 8625
    },
    {
      "epoch": 2.9242731575388774,
      "grad_norm": 1.3625876903533936,
      "learning_rate": 1.4242105263157897e-06,
      "loss": 0.0621,
      "step": 8650
    },
    {
      "epoch": 2.9327248140635565,
      "grad_norm": 1.5794146060943604,
      "learning_rate": 1.3978947368421053e-06,
      "loss": 0.0657,
      "step": 8675
    },
    {
      "epoch": 2.9411764705882355,
      "grad_norm": 1.314706563949585,
      "learning_rate": 1.3715789473684213e-06,
      "loss": 0.069,
      "step": 8700
    },
    {
      "epoch": 2.949628127112914,
      "grad_norm": 1.9198169708251953,
      "learning_rate": 1.345263157894737e-06,
      "loss": 0.0701,
      "step": 8725
    },
    {
      "epoch": 2.9580797836375927,
      "grad_norm": 1.3500406742095947,
      "learning_rate": 1.3189473684210527e-06,
      "loss": 0.076,
      "step": 8750
    },
    {
      "epoch": 2.9665314401622718,
      "grad_norm": 1.7573151588439941,
      "learning_rate": 1.2926315789473683e-06,
      "loss": 0.0697,
      "step": 8775
    },
    {
      "epoch": 2.974983096686951,
      "grad_norm": 0.8199857473373413,
      "learning_rate": 1.2663157894736844e-06,
      "loss": 0.0744,
      "step": 8800
    },
    {
      "epoch": 2.9834347532116294,
      "grad_norm": 2.14540696144104,
      "learning_rate": 1.2400000000000002e-06,
      "loss": 0.0686,
      "step": 8825
    },
    {
      "epoch": 2.991886409736308,
      "grad_norm": 1.5464670658111572,
      "learning_rate": 1.213684210526316e-06,
      "loss": 0.07,
      "step": 8850
    },
    {
      "epoch": 3.000338066260987,
      "grad_norm": 1.2003583908081055,
      "learning_rate": 1.188421052631579e-06,
      "loss": 0.0781,
      "step": 8875
    },
    {
      "epoch": 3.008789722785666,
      "grad_norm": 0.630574107170105,
      "learning_rate": 1.1621052631578948e-06,
      "loss": 0.0565,
      "step": 8900
    },
    {
      "epoch": 3.0172413793103448,
      "grad_norm": 1.4540669918060303,
      "learning_rate": 1.1357894736842106e-06,
      "loss": 0.0547,
      "step": 8925
    },
    {
      "epoch": 3.025693035835024,
      "grad_norm": 1.408868670463562,
      "learning_rate": 1.1094736842105264e-06,
      "loss": 0.0604,
      "step": 8950
    },
    {
      "epoch": 3.0341446923597024,
      "grad_norm": 1.040751576423645,
      "learning_rate": 1.0831578947368422e-06,
      "loss": 0.063,
      "step": 8975
    },
    {
      "epoch": 3.0425963488843815,
      "grad_norm": 1.449309229850769,
      "learning_rate": 1.0568421052631578e-06,
      "loss": 0.0593,
      "step": 9000
    },
    {
      "epoch": 3.0425963488843815,
      "eval_loss": 0.19757185876369476,
      "eval_runtime": 8436.2186,
      "eval_samples_per_second": 1.734,
      "eval_steps_per_second": 0.217,
      "eval_wer": 44.964132555512634,
      "step": 9000
    },
    {
      "epoch": 3.05104800540906,
      "grad_norm": 2.647589921951294,
      "learning_rate": 1.0305263157894739e-06,
      "loss": 0.0602,
      "step": 9025
    },
    {
      "epoch": 3.059499661933739,
      "grad_norm": 1.1657071113586426,
      "learning_rate": 1.0042105263157897e-06,
      "loss": 0.0637,
      "step": 9050
    },
    {
      "epoch": 3.0679513184584177,
      "grad_norm": 1.9827054738998413,
      "learning_rate": 9.778947368421053e-07,
      "loss": 0.0667,
      "step": 9075
    },
    {
      "epoch": 3.076402974983097,
      "grad_norm": 1.7273396253585815,
      "learning_rate": 9.515789473684212e-07,
      "loss": 0.0592,
      "step": 9100
    },
    {
      "epoch": 3.0848546315077754,
      "grad_norm": 1.3955703973770142,
      "learning_rate": 9.252631578947368e-07,
      "loss": 0.0522,
      "step": 9125
    },
    {
      "epoch": 3.0933062880324544,
      "grad_norm": 1.3141138553619385,
      "learning_rate": 8.989473684210527e-07,
      "loss": 0.0627,
      "step": 9150
    },
    {
      "epoch": 3.101757944557133,
      "grad_norm": 0.7096537351608276,
      "learning_rate": 8.726315789473686e-07,
      "loss": 0.0583,
      "step": 9175
    },
    {
      "epoch": 3.110209601081812,
      "grad_norm": 1.3626320362091064,
      "learning_rate": 8.463157894736843e-07,
      "loss": 0.0556,
      "step": 9200
    },
    {
      "epoch": 3.1186612576064907,
      "grad_norm": 1.7827398777008057,
      "learning_rate": 8.200000000000001e-07,
      "loss": 0.0549,
      "step": 9225
    },
    {
      "epoch": 3.1271129141311698,
      "grad_norm": 1.8511258363723755,
      "learning_rate": 7.936842105263158e-07,
      "loss": 0.052,
      "step": 9250
    },
    {
      "epoch": 3.1355645706558484,
      "grad_norm": 1.5785564184188843,
      "learning_rate": 7.673684210526316e-07,
      "loss": 0.0554,
      "step": 9275
    },
    {
      "epoch": 3.1440162271805274,
      "grad_norm": 1.3359767198562622,
      "learning_rate": 7.410526315789475e-07,
      "loss": 0.0599,
      "step": 9300
    },
    {
      "epoch": 3.152467883705206,
      "grad_norm": 1.2103520631790161,
      "learning_rate": 7.147368421052632e-07,
      "loss": 0.0564,
      "step": 9325
    },
    {
      "epoch": 3.160919540229885,
      "grad_norm": 1.3505812883377075,
      "learning_rate": 6.884210526315791e-07,
      "loss": 0.0642,
      "step": 9350
    },
    {
      "epoch": 3.1693711967545637,
      "grad_norm": 1.9448202848434448,
      "learning_rate": 6.621052631578948e-07,
      "loss": 0.0567,
      "step": 9375
    },
    {
      "epoch": 3.1778228532792427,
      "grad_norm": 1.4661892652511597,
      "learning_rate": 6.357894736842106e-07,
      "loss": 0.063,
      "step": 9400
    },
    {
      "epoch": 3.186274509803922,
      "grad_norm": 2.1202943325042725,
      "learning_rate": 6.094736842105263e-07,
      "loss": 0.0589,
      "step": 9425
    },
    {
      "epoch": 3.1947261663286004,
      "grad_norm": 1.8288671970367432,
      "learning_rate": 5.831578947368421e-07,
      "loss": 0.0601,
      "step": 9450
    },
    {
      "epoch": 3.203177822853279,
      "grad_norm": 1.0134177207946777,
      "learning_rate": 5.56842105263158e-07,
      "loss": 0.0546,
      "step": 9475
    },
    {
      "epoch": 3.211629479377958,
      "grad_norm": 2.195054292678833,
      "learning_rate": 5.305263157894737e-07,
      "loss": 0.0598,
      "step": 9500
    },
    {
      "epoch": 3.211629479377958,
      "eval_loss": 0.1982024759054184,
      "eval_runtime": 8467.5483,
      "eval_samples_per_second": 1.727,
      "eval_steps_per_second": 0.216,
      "eval_wer": 44.87513630606886,
      "step": 9500
    },
    {
      "epoch": 3.220081135902637,
      "grad_norm": 1.3598229885101318,
      "learning_rate": 5.042105263157895e-07,
      "loss": 0.0544,
      "step": 9525
    },
    {
      "epoch": 3.2285327924273157,
      "grad_norm": 1.9829254150390625,
      "learning_rate": 4.778947368421053e-07,
      "loss": 0.0583,
      "step": 9550
    },
    {
      "epoch": 3.2369844489519948,
      "grad_norm": 1.2495145797729492,
      "learning_rate": 4.5157894736842107e-07,
      "loss": 0.0659,
      "step": 9575
    },
    {
      "epoch": 3.2454361054766734,
      "grad_norm": 1.0571123361587524,
      "learning_rate": 4.2526315789473684e-07,
      "loss": 0.0548,
      "step": 9600
    },
    {
      "epoch": 3.2538877620013524,
      "grad_norm": 1.3266067504882812,
      "learning_rate": 3.9894736842105266e-07,
      "loss": 0.0606,
      "step": 9625
    },
    {
      "epoch": 3.262339418526031,
      "grad_norm": 2.1030797958374023,
      "learning_rate": 3.726315789473685e-07,
      "loss": 0.0651,
      "step": 9650
    },
    {
      "epoch": 3.27079107505071,
      "grad_norm": 0.6798356175422668,
      "learning_rate": 3.4631578947368424e-07,
      "loss": 0.0567,
      "step": 9675
    },
    {
      "epoch": 3.2792427315753887,
      "grad_norm": 1.9583333730697632,
      "learning_rate": 3.2e-07,
      "loss": 0.0581,
      "step": 9700
    },
    {
      "epoch": 3.2876943881000678,
      "grad_norm": 2.115421772003174,
      "learning_rate": 2.936842105263158e-07,
      "loss": 0.0502,
      "step": 9725
    },
    {
      "epoch": 3.2961460446247464,
      "grad_norm": 1.1875923871994019,
      "learning_rate": 2.6736842105263164e-07,
      "loss": 0.0621,
      "step": 9750
    },
    {
      "epoch": 3.3045977011494254,
      "grad_norm": 1.232844352722168,
      "learning_rate": 2.410526315789474e-07,
      "loss": 0.0591,
      "step": 9775
    },
    {
      "epoch": 3.313049357674104,
      "grad_norm": 1.3175997734069824,
      "learning_rate": 2.1473684210526317e-07,
      "loss": 0.0547,
      "step": 9800
    },
    {
      "epoch": 3.321501014198783,
      "grad_norm": 1.7271957397460938,
      "learning_rate": 1.8842105263157897e-07,
      "loss": 0.0611,
      "step": 9825
    },
    {
      "epoch": 3.3299526707234617,
      "grad_norm": 1.2155121564865112,
      "learning_rate": 1.6210526315789476e-07,
      "loss": 0.0645,
      "step": 9850
    },
    {
      "epoch": 3.3384043272481407,
      "grad_norm": 1.6896015405654907,
      "learning_rate": 1.3578947368421055e-07,
      "loss": 0.0526,
      "step": 9875
    },
    {
      "epoch": 3.3468559837728193,
      "grad_norm": 1.4547408819198608,
      "learning_rate": 1.0947368421052632e-07,
      "loss": 0.0512,
      "step": 9900
    },
    {
      "epoch": 3.3553076402974984,
      "grad_norm": 0.5560055375099182,
      "learning_rate": 8.315789473684211e-08,
      "loss": 0.0665,
      "step": 9925
    },
    {
      "epoch": 3.363759296822177,
      "grad_norm": 1.9773001670837402,
      "learning_rate": 5.68421052631579e-08,
      "loss": 0.0522,
      "step": 9950
    },
    {
      "epoch": 3.372210953346856,
      "grad_norm": 1.2007287740707397,
      "learning_rate": 3.0526315789473686e-08,
      "loss": 0.0623,
      "step": 9975
    },
    {
      "epoch": 3.3806626098715347,
      "grad_norm": 1.430873990058899,
      "learning_rate": 4.210526315789474e-09,
      "loss": 0.0553,
      "step": 10000
    },
    {
      "epoch": 3.3806626098715347,
      "eval_loss": 0.19793640077114105,
      "eval_runtime": 8419.2841,
      "eval_samples_per_second": 1.737,
      "eval_steps_per_second": 0.217,
      "eval_wer": 44.86535649843768,
      "step": 10000
    },
    {
      "epoch": 2.519477255591857,
      "grad_norm": 2.5541698932647705,
      "learning_rate": 3.4337931034482762e-06,
      "loss": 0.1093,
      "step": 10025
    },
    {
      "epoch": 2.5257602412666498,
      "grad_norm": 2.943728446960449,
      "learning_rate": 3.4165517241379315e-06,
      "loss": 0.0942,
      "step": 10050
    },
    {
      "epoch": 2.5320432269414423,
      "grad_norm": 1.3236327171325684,
      "learning_rate": 3.3993103448275864e-06,
      "loss": 0.0831,
      "step": 10075
    },
    {
      "epoch": 2.5383262126162354,
      "grad_norm": 1.9076201915740967,
      "learning_rate": 3.3820689655172417e-06,
      "loss": 0.0808,
      "step": 10100
    },
    {
      "epoch": 2.544609198291028,
      "grad_norm": 1.5796664953231812,
      "learning_rate": 3.364827586206897e-06,
      "loss": 0.0738,
      "step": 10125
    },
    {
      "epoch": 2.5508921839658205,
      "grad_norm": 1.349244475364685,
      "learning_rate": 3.347586206896552e-06,
      "loss": 0.0854,
      "step": 10150
    },
    {
      "epoch": 2.557175169640613,
      "grad_norm": 1.5423616170883179,
      "learning_rate": 3.330344827586207e-06,
      "loss": 0.0694,
      "step": 10175
    },
    {
      "epoch": 2.5634581553154057,
      "grad_norm": 1.882533073425293,
      "learning_rate": 3.3131034482758624e-06,
      "loss": 0.0804,
      "step": 10200
    },
    {
      "epoch": 2.5697411409901987,
      "grad_norm": 1.6644792556762695,
      "learning_rate": 3.2958620689655173e-06,
      "loss": 0.0818,
      "step": 10225
    },
    {
      "epoch": 2.5760241266649913,
      "grad_norm": 1.648979663848877,
      "learning_rate": 3.2786206896551726e-06,
      "loss": 0.0806,
      "step": 10250
    },
    {
      "epoch": 2.582307112339784,
      "grad_norm": 1.8302336931228638,
      "learning_rate": 3.261379310344828e-06,
      "loss": 0.0685,
      "step": 10275
    },
    {
      "epoch": 2.5885900980145764,
      "grad_norm": 1.5764237642288208,
      "learning_rate": 3.2441379310344828e-06,
      "loss": 0.0669,
      "step": 10300
    },
    {
      "epoch": 2.594873083689369,
      "grad_norm": 1.609424352645874,
      "learning_rate": 3.226896551724138e-06,
      "loss": 0.0703,
      "step": 10325
    },
    {
      "epoch": 2.601156069364162,
      "grad_norm": 2.0463669300079346,
      "learning_rate": 3.209655172413793e-06,
      "loss": 0.0812,
      "step": 10350
    },
    {
      "epoch": 2.6074390550389546,
      "grad_norm": 1.0998157262802124,
      "learning_rate": 3.1924137931034486e-06,
      "loss": 0.0811,
      "step": 10375
    },
    {
      "epoch": 2.613722040713747,
      "grad_norm": 2.0969502925872803,
      "learning_rate": 3.175172413793104e-06,
      "loss": 0.0831,
      "step": 10400
    },
    {
      "epoch": 2.6200050263885397,
      "grad_norm": 1.6050735712051392,
      "learning_rate": 3.1579310344827592e-06,
      "loss": 0.0856,
      "step": 10425
    },
    {
      "epoch": 2.6262880120633323,
      "grad_norm": 1.6979131698608398,
      "learning_rate": 3.140689655172414e-06,
      "loss": 0.0774,
      "step": 10450
    },
    {
      "epoch": 2.6325709977381253,
      "grad_norm": 2.0969455242156982,
      "learning_rate": 3.1234482758620694e-06,
      "loss": 0.0871,
      "step": 10475
    },
    {
      "epoch": 2.638853983412918,
      "grad_norm": 1.3703320026397705,
      "learning_rate": 3.1062068965517243e-06,
      "loss": 0.0677,
      "step": 10500
    },
    {
      "epoch": 2.638853983412918,
      "eval_loss": 0.169998437166214,
      "eval_runtime": 11305.71,
      "eval_samples_per_second": 1.728,
      "eval_steps_per_second": 0.216,
      "eval_wer": 44.5221605147797,
      "step": 10500
    },
    {
      "epoch": 2.6451369690877105,
      "grad_norm": 1.9054275751113892,
      "learning_rate": 3.0889655172413796e-06,
      "loss": 0.0703,
      "step": 10525
    },
    {
      "epoch": 2.651419954762503,
      "grad_norm": 1.9211713075637817,
      "learning_rate": 3.071724137931035e-06,
      "loss": 0.078,
      "step": 10550
    },
    {
      "epoch": 2.6577029404372956,
      "grad_norm": 1.7208181619644165,
      "learning_rate": 3.0544827586206897e-06,
      "loss": 0.0732,
      "step": 10575
    },
    {
      "epoch": 2.6639859261120886,
      "grad_norm": 1.2526366710662842,
      "learning_rate": 3.037241379310345e-06,
      "loss": 0.0827,
      "step": 10600
    },
    {
      "epoch": 2.670268911786881,
      "grad_norm": 1.8517524003982544,
      "learning_rate": 3.0200000000000003e-06,
      "loss": 0.0758,
      "step": 10625
    },
    {
      "epoch": 2.6765518974616738,
      "grad_norm": 1.4826529026031494,
      "learning_rate": 3.002758620689655e-06,
      "loss": 0.0831,
      "step": 10650
    },
    {
      "epoch": 2.6828348831364663,
      "grad_norm": 2.6525490283966064,
      "learning_rate": 2.9855172413793105e-06,
      "loss": 0.0783,
      "step": 10675
    },
    {
      "epoch": 2.689117868811259,
      "grad_norm": 1.8993303775787354,
      "learning_rate": 2.9682758620689658e-06,
      "loss": 0.0821,
      "step": 10700
    },
    {
      "epoch": 2.695400854486052,
      "grad_norm": 1.4494125843048096,
      "learning_rate": 2.9510344827586206e-06,
      "loss": 0.0771,
      "step": 10725
    },
    {
      "epoch": 2.7016838401608445,
      "grad_norm": 1.8887574672698975,
      "learning_rate": 2.933793103448276e-06,
      "loss": 0.0705,
      "step": 10750
    },
    {
      "epoch": 2.707966825835637,
      "grad_norm": 2.0872342586517334,
      "learning_rate": 2.9165517241379316e-06,
      "loss": 0.0707,
      "step": 10775
    },
    {
      "epoch": 2.7142498115104297,
      "grad_norm": 1.3491803407669067,
      "learning_rate": 2.8993103448275865e-06,
      "loss": 0.0733,
      "step": 10800
    },
    {
      "epoch": 2.7205327971852222,
      "grad_norm": 1.6230404376983643,
      "learning_rate": 2.882068965517242e-06,
      "loss": 0.0703,
      "step": 10825
    },
    {
      "epoch": 2.7268157828600152,
      "grad_norm": 1.4654004573822021,
      "learning_rate": 2.864827586206897e-06,
      "loss": 0.0705,
      "step": 10850
    },
    {
      "epoch": 2.733098768534808,
      "grad_norm": 1.7205106019973755,
      "learning_rate": 2.847586206896552e-06,
      "loss": 0.0713,
      "step": 10875
    },
    {
      "epoch": 2.7393817542096004,
      "grad_norm": 1.7660115957260132,
      "learning_rate": 2.8303448275862073e-06,
      "loss": 0.0759,
      "step": 10900
    },
    {
      "epoch": 2.745664739884393,
      "grad_norm": 1.3411388397216797,
      "learning_rate": 2.813103448275862e-06,
      "loss": 0.0819,
      "step": 10925
    },
    {
      "epoch": 2.7519477255591855,
      "grad_norm": 1.66763436794281,
      "learning_rate": 2.7958620689655174e-06,
      "loss": 0.0714,
      "step": 10950
    },
    {
      "epoch": 2.7582307112339786,
      "grad_norm": 1.856402039527893,
      "learning_rate": 2.7786206896551727e-06,
      "loss": 0.0737,
      "step": 10975
    },
    {
      "epoch": 2.764513696908771,
      "grad_norm": 1.4149645566940308,
      "learning_rate": 2.7613793103448276e-06,
      "loss": 0.0698,
      "step": 11000
    },
    {
      "epoch": 2.764513696908771,
      "eval_loss": 0.16873179376125336,
      "eval_runtime": 11438.4823,
      "eval_samples_per_second": 1.708,
      "eval_steps_per_second": 0.213,
      "eval_wer": 44.161424743801774,
      "step": 11000
    },
    {
      "epoch": 2.7707966825835637,
      "grad_norm": 1.936444878578186,
      "learning_rate": 2.744137931034483e-06,
      "loss": 0.0757,
      "step": 11025
    },
    {
      "epoch": 2.7770796682583563,
      "grad_norm": 1.3079452514648438,
      "learning_rate": 2.726896551724138e-06,
      "loss": 0.0773,
      "step": 11050
    },
    {
      "epoch": 2.783362653933149,
      "grad_norm": 1.8395476341247559,
      "learning_rate": 2.709655172413793e-06,
      "loss": 0.069,
      "step": 11075
    },
    {
      "epoch": 2.789645639607942,
      "grad_norm": 1.5000851154327393,
      "learning_rate": 2.6924137931034483e-06,
      "loss": 0.0808,
      "step": 11100
    },
    {
      "epoch": 2.7959286252827344,
      "grad_norm": 1.7593291997909546,
      "learning_rate": 2.6751724137931036e-06,
      "loss": 0.0742,
      "step": 11125
    },
    {
      "epoch": 2.802211610957527,
      "grad_norm": 1.671397089958191,
      "learning_rate": 2.6579310344827585e-06,
      "loss": 0.0742,
      "step": 11150
    },
    {
      "epoch": 2.8084945966323196,
      "grad_norm": 2.4030308723449707,
      "learning_rate": 2.6406896551724142e-06,
      "loss": 0.0715,
      "step": 11175
    },
    {
      "epoch": 2.814777582307112,
      "grad_norm": 1.1083357334136963,
      "learning_rate": 2.6234482758620695e-06,
      "loss": 0.0685,
      "step": 11200
    },
    {
      "epoch": 2.821060567981905,
      "grad_norm": 2.314483404159546,
      "learning_rate": 2.6062068965517244e-06,
      "loss": 0.0709,
      "step": 11225
    },
    {
      "epoch": 2.8273435536566978,
      "grad_norm": 2.6994457244873047,
      "learning_rate": 2.5889655172413797e-06,
      "loss": 0.0716,
      "step": 11250
    },
    {
      "epoch": 2.8336265393314903,
      "grad_norm": 1.5674411058425903,
      "learning_rate": 2.571724137931035e-06,
      "loss": 0.0751,
      "step": 11275
    },
    {
      "epoch": 2.839909525006283,
      "grad_norm": 1.6868212223052979,
      "learning_rate": 2.55448275862069e-06,
      "loss": 0.0775,
      "step": 11300
    },
    {
      "epoch": 2.8461925106810755,
      "grad_norm": 1.8248207569122314,
      "learning_rate": 2.537241379310345e-06,
      "loss": 0.0682,
      "step": 11325
    },
    {
      "epoch": 2.8524754963558685,
      "grad_norm": 1.6669248342514038,
      "learning_rate": 2.52e-06,
      "loss": 0.0706,
      "step": 11350
    },
    {
      "epoch": 2.858758482030661,
      "grad_norm": 2.579449415206909,
      "learning_rate": 2.5027586206896553e-06,
      "loss": 0.0711,
      "step": 11375
    },
    {
      "epoch": 2.8650414677054536,
      "grad_norm": 2.2403104305267334,
      "learning_rate": 2.4855172413793106e-06,
      "loss": 0.0731,
      "step": 11400
    },
    {
      "epoch": 2.8713244533802462,
      "grad_norm": 1.7787814140319824,
      "learning_rate": 2.4682758620689655e-06,
      "loss": 0.0683,
      "step": 11425
    },
    {
      "epoch": 2.877607439055039,
      "grad_norm": 1.8239678144454956,
      "learning_rate": 2.4510344827586208e-06,
      "loss": 0.0752,
      "step": 11450
    },
    {
      "epoch": 2.883890424729832,
      "grad_norm": 3.1103549003601074,
      "learning_rate": 2.433793103448276e-06,
      "loss": 0.074,
      "step": 11475
    },
    {
      "epoch": 2.8901734104046244,
      "grad_norm": 1.3953099250793457,
      "learning_rate": 2.4165517241379314e-06,
      "loss": 0.069,
      "step": 11500
    },
    {
      "epoch": 2.8901734104046244,
      "eval_loss": 0.16406717896461487,
      "eval_runtime": 11006.7152,
      "eval_samples_per_second": 1.775,
      "eval_steps_per_second": 0.222,
      "eval_wer": 43.59414156441607,
      "step": 11500
    },
    {
      "epoch": 2.896456396079417,
      "grad_norm": 1.4673534631729126,
      "learning_rate": 2.3993103448275866e-06,
      "loss": 0.072,
      "step": 11525
    },
    {
      "epoch": 2.9027393817542095,
      "grad_norm": 1.790159821510315,
      "learning_rate": 2.3820689655172415e-06,
      "loss": 0.0846,
      "step": 11550
    },
    {
      "epoch": 2.909022367429002,
      "grad_norm": 1.482283115386963,
      "learning_rate": 2.364827586206897e-06,
      "loss": 0.0718,
      "step": 11575
    },
    {
      "epoch": 2.915305353103795,
      "grad_norm": 1.344299077987671,
      "learning_rate": 2.3475862068965517e-06,
      "loss": 0.0647,
      "step": 11600
    },
    {
      "epoch": 2.9215883387785877,
      "grad_norm": 2.296725034713745,
      "learning_rate": 2.330344827586207e-06,
      "loss": 0.0729,
      "step": 11625
    },
    {
      "epoch": 2.9278713244533803,
      "grad_norm": 1.4769775867462158,
      "learning_rate": 2.3131034482758623e-06,
      "loss": 0.0665,
      "step": 11650
    },
    {
      "epoch": 2.934154310128173,
      "grad_norm": 1.83705735206604,
      "learning_rate": 2.2958620689655176e-06,
      "loss": 0.0777,
      "step": 11675
    },
    {
      "epoch": 2.9404372958029654,
      "grad_norm": 1.8100663423538208,
      "learning_rate": 2.278620689655173e-06,
      "loss": 0.0717,
      "step": 11700
    },
    {
      "epoch": 2.9467202814777584,
      "grad_norm": 1.144631028175354,
      "learning_rate": 2.2613793103448277e-06,
      "loss": 0.0622,
      "step": 11725
    },
    {
      "epoch": 2.953003267152551,
      "grad_norm": 1.8608859777450562,
      "learning_rate": 2.244137931034483e-06,
      "loss": 0.0706,
      "step": 11750
    },
    {
      "epoch": 2.9592862528273436,
      "grad_norm": 2.04830002784729,
      "learning_rate": 2.2268965517241383e-06,
      "loss": 0.065,
      "step": 11775
    },
    {
      "epoch": 2.965569238502136,
      "grad_norm": 1.9950125217437744,
      "learning_rate": 2.209655172413793e-06,
      "loss": 0.0734,
      "step": 11800
    },
    {
      "epoch": 2.9718522241769287,
      "grad_norm": 1.2800809144973755,
      "learning_rate": 2.1924137931034485e-06,
      "loss": 0.0626,
      "step": 11825
    },
    {
      "epoch": 2.9781352098517218,
      "grad_norm": 2.180931568145752,
      "learning_rate": 2.1751724137931033e-06,
      "loss": 0.0694,
      "step": 11850
    },
    {
      "epoch": 2.9844181955265143,
      "grad_norm": 1.5621517896652222,
      "learning_rate": 2.157931034482759e-06,
      "loss": 0.0789,
      "step": 11875
    },
    {
      "epoch": 2.990701181201307,
      "grad_norm": 1.7615336179733276,
      "learning_rate": 2.140689655172414e-06,
      "loss": 0.0714,
      "step": 11900
    },
    {
      "epoch": 2.9969841668760995,
      "grad_norm": 1.793643593788147,
      "learning_rate": 2.1234482758620692e-06,
      "loss": 0.0663,
      "step": 11925
    },
    {
      "epoch": 3.003267152550892,
      "grad_norm": 1.6309877634048462,
      "learning_rate": 2.1062068965517245e-06,
      "loss": 0.0638,
      "step": 11950
    },
    {
      "epoch": 3.009550138225685,
      "grad_norm": 1.6418914794921875,
      "learning_rate": 2.0889655172413794e-06,
      "loss": 0.0601,
      "step": 11975
    },
    {
      "epoch": 3.0158331239004776,
      "grad_norm": 1.7936900854110718,
      "learning_rate": 2.0717241379310347e-06,
      "loss": 0.0575,
      "step": 12000
    },
    {
      "epoch": 3.0158331239004776,
      "eval_loss": 0.1658269762992859,
      "eval_runtime": 10966.6675,
      "eval_samples_per_second": 1.781,
      "eval_steps_per_second": 0.223,
      "eval_wer": 43.487979085269416,
      "step": 12000
    },
    {
      "epoch": 3.02211610957527,
      "grad_norm": 1.462873935699463,
      "learning_rate": 2.0544827586206896e-06,
      "loss": 0.0606,
      "step": 12025
    },
    {
      "epoch": 3.028399095250063,
      "grad_norm": 1.7744050025939941,
      "learning_rate": 2.037241379310345e-06,
      "loss": 0.07,
      "step": 12050
    },
    {
      "epoch": 3.0346820809248554,
      "grad_norm": 2.0455141067504883,
      "learning_rate": 2.02e-06,
      "loss": 0.063,
      "step": 12075
    },
    {
      "epoch": 3.040965066599648,
      "grad_norm": 1.5644878149032593,
      "learning_rate": 2.0027586206896554e-06,
      "loss": 0.0688,
      "step": 12100
    },
    {
      "epoch": 3.047248052274441,
      "grad_norm": 1.5917677879333496,
      "learning_rate": 1.9855172413793107e-06,
      "loss": 0.0751,
      "step": 12125
    },
    {
      "epoch": 3.0535310379492335,
      "grad_norm": 1.8695135116577148,
      "learning_rate": 1.9682758620689656e-06,
      "loss": 0.0683,
      "step": 12150
    },
    {
      "epoch": 3.059814023624026,
      "grad_norm": 1.6950143575668335,
      "learning_rate": 1.951034482758621e-06,
      "loss": 0.0707,
      "step": 12175
    },
    {
      "epoch": 3.0660970092988187,
      "grad_norm": 1.7341082096099854,
      "learning_rate": 1.933793103448276e-06,
      "loss": 0.0705,
      "step": 12200
    },
    {
      "epoch": 3.0723799949736113,
      "grad_norm": 2.4413957595825195,
      "learning_rate": 1.916551724137931e-06,
      "loss": 0.0732,
      "step": 12225
    },
    {
      "epoch": 3.0786629806484043,
      "grad_norm": 1.6173579692840576,
      "learning_rate": 1.8993103448275864e-06,
      "loss": 0.0693,
      "step": 12250
    },
    {
      "epoch": 3.084945966323197,
      "grad_norm": 1.606766939163208,
      "learning_rate": 1.8820689655172416e-06,
      "loss": 0.0648,
      "step": 12275
    },
    {
      "epoch": 3.0912289519979894,
      "grad_norm": 1.7197469472885132,
      "learning_rate": 1.8648275862068967e-06,
      "loss": 0.0607,
      "step": 12300
    },
    {
      "epoch": 3.097511937672782,
      "grad_norm": 1.3897112607955933,
      "learning_rate": 1.847586206896552e-06,
      "loss": 0.0722,
      "step": 12325
    },
    {
      "epoch": 3.1037949233475746,
      "grad_norm": 1.53862726688385,
      "learning_rate": 1.830344827586207e-06,
      "loss": 0.0635,
      "step": 12350
    },
    {
      "epoch": 3.1100779090223676,
      "grad_norm": 2.654273509979248,
      "learning_rate": 1.8131034482758622e-06,
      "loss": 0.0604,
      "step": 12375
    },
    {
      "epoch": 3.11636089469716,
      "grad_norm": 1.34363853931427,
      "learning_rate": 1.7958620689655173e-06,
      "loss": 0.0563,
      "step": 12400
    },
    {
      "epoch": 3.1226438803719527,
      "grad_norm": 1.0993194580078125,
      "learning_rate": 1.7786206896551726e-06,
      "loss": 0.0689,
      "step": 12425
    },
    {
      "epoch": 3.1289268660467453,
      "grad_norm": 1.6156988143920898,
      "learning_rate": 1.7613793103448276e-06,
      "loss": 0.0563,
      "step": 12450
    },
    {
      "epoch": 3.135209851721538,
      "grad_norm": 1.7135562896728516,
      "learning_rate": 1.744137931034483e-06,
      "loss": 0.0606,
      "step": 12475
    },
    {
      "epoch": 3.141492837396331,
      "grad_norm": 1.968897819519043,
      "learning_rate": 1.7268965517241382e-06,
      "loss": 0.0653,
      "step": 12500
    },
    {
      "epoch": 3.141492837396331,
      "eval_loss": 0.1663118600845337,
      "eval_runtime": 11490.7091,
      "eval_samples_per_second": 1.7,
      "eval_steps_per_second": 0.213,
      "eval_wer": 43.23918334332368,
      "step": 12500
    },
    {
      "epoch": 3.1477758230711235,
      "grad_norm": 2.3499624729156494,
      "learning_rate": 1.7096551724137933e-06,
      "loss": 0.0694,
      "step": 12525
    },
    {
      "epoch": 3.154058808745916,
      "grad_norm": 2.061582088470459,
      "learning_rate": 1.6924137931034484e-06,
      "loss": 0.0585,
      "step": 12550
    },
    {
      "epoch": 3.1603417944207086,
      "grad_norm": 1.3034265041351318,
      "learning_rate": 1.6751724137931037e-06,
      "loss": 0.0585,
      "step": 12575
    },
    {
      "epoch": 3.166624780095501,
      "grad_norm": 1.5961774587631226,
      "learning_rate": 1.6579310344827588e-06,
      "loss": 0.0623,
      "step": 12600
    },
    {
      "epoch": 3.172907765770294,
      "grad_norm": 1.5717999935150146,
      "learning_rate": 1.6406896551724138e-06,
      "loss": 0.0801,
      "step": 12625
    },
    {
      "epoch": 3.179190751445087,
      "grad_norm": 0.8206491470336914,
      "learning_rate": 1.623448275862069e-06,
      "loss": 0.0645,
      "step": 12650
    },
    {
      "epoch": 3.1854737371198794,
      "grad_norm": 1.6915777921676636,
      "learning_rate": 1.6062068965517244e-06,
      "loss": 0.0639,
      "step": 12675
    },
    {
      "epoch": 3.191756722794672,
      "grad_norm": 1.749098300933838,
      "learning_rate": 1.5889655172413795e-06,
      "loss": 0.0613,
      "step": 12700
    },
    {
      "epoch": 3.1980397084694645,
      "grad_norm": 1.5118510723114014,
      "learning_rate": 1.5717241379310346e-06,
      "loss": 0.0654,
      "step": 12725
    },
    {
      "epoch": 3.2043226941442575,
      "grad_norm": 1.413273811340332,
      "learning_rate": 1.55448275862069e-06,
      "loss": 0.071,
      "step": 12750
    },
    {
      "epoch": 3.21060567981905,
      "grad_norm": 2.4434258937835693,
      "learning_rate": 1.537241379310345e-06,
      "loss": 0.0611,
      "step": 12775
    },
    {
      "epoch": 3.2168886654938427,
      "grad_norm": 1.5421768426895142,
      "learning_rate": 1.52e-06,
      "loss": 0.0626,
      "step": 12800
    },
    {
      "epoch": 3.2231716511686352,
      "grad_norm": 1.3737552165985107,
      "learning_rate": 1.5027586206896551e-06,
      "loss": 0.0644,
      "step": 12825
    },
    {
      "epoch": 3.229454636843428,
      "grad_norm": 1.0774625539779663,
      "learning_rate": 1.4855172413793104e-06,
      "loss": 0.0628,
      "step": 12850
    },
    {
      "epoch": 3.235737622518221,
      "grad_norm": 1.323012113571167,
      "learning_rate": 1.4682758620689657e-06,
      "loss": 0.0681,
      "step": 12875
    },
    {
      "epoch": 3.2420206081930134,
      "grad_norm": 1.7004013061523438,
      "learning_rate": 1.4510344827586208e-06,
      "loss": 0.0575,
      "step": 12900
    },
    {
      "epoch": 3.248303593867806,
      "grad_norm": 1.1748865842819214,
      "learning_rate": 1.433793103448276e-06,
      "loss": 0.0686,
      "step": 12925
    },
    {
      "epoch": 3.2545865795425986,
      "grad_norm": 1.9509600400924683,
      "learning_rate": 1.4165517241379312e-06,
      "loss": 0.064,
      "step": 12950
    },
    {
      "epoch": 3.260869565217391,
      "grad_norm": 1.4009268283843994,
      "learning_rate": 1.3993103448275863e-06,
      "loss": 0.07,
      "step": 12975
    },
    {
      "epoch": 3.267152550892184,
      "grad_norm": 2.449444532394409,
      "learning_rate": 1.3820689655172416e-06,
      "loss": 0.0759,
      "step": 13000
    },
    {
      "epoch": 3.267152550892184,
      "eval_loss": 0.16605377197265625,
      "eval_runtime": 11752.4146,
      "eval_samples_per_second": 1.662,
      "eval_steps_per_second": 0.208,
      "eval_wer": 43.109910664634896,
      "step": 13000
    },
    {
      "epoch": 3.2734355365669767,
      "grad_norm": 1.3904393911361694,
      "learning_rate": 1.3648275862068966e-06,
      "loss": 0.0607,
      "step": 13025
    },
    {
      "epoch": 3.2797185222417693,
      "grad_norm": 1.4722603559494019,
      "learning_rate": 1.3475862068965517e-06,
      "loss": 0.0646,
      "step": 13050
    },
    {
      "epoch": 3.286001507916562,
      "grad_norm": 1.7087178230285645,
      "learning_rate": 1.3303448275862072e-06,
      "loss": 0.056,
      "step": 13075
    },
    {
      "epoch": 3.2922844935913544,
      "grad_norm": 1.5727674961090088,
      "learning_rate": 1.3131034482758623e-06,
      "loss": 0.0635,
      "step": 13100
    },
    {
      "epoch": 3.2985674792661475,
      "grad_norm": 1.9577021598815918,
      "learning_rate": 1.2958620689655174e-06,
      "loss": 0.064,
      "step": 13125
    },
    {
      "epoch": 3.30485046494094,
      "grad_norm": 1.86566960811615,
      "learning_rate": 1.2786206896551725e-06,
      "loss": 0.062,
      "step": 13150
    },
    {
      "epoch": 3.3111334506157326,
      "grad_norm": 1.787929892539978,
      "learning_rate": 1.2613793103448278e-06,
      "loss": 0.0587,
      "step": 13175
    },
    {
      "epoch": 3.317416436290525,
      "grad_norm": 1.5625816583633423,
      "learning_rate": 1.2441379310344829e-06,
      "loss": 0.0638,
      "step": 13200
    },
    {
      "epoch": 3.3236994219653178,
      "grad_norm": 1.4375091791152954,
      "learning_rate": 1.226896551724138e-06,
      "loss": 0.066,
      "step": 13225
    },
    {
      "epoch": 3.3299824076401103,
      "grad_norm": 1.6865044832229614,
      "learning_rate": 1.2096551724137932e-06,
      "loss": 0.0584,
      "step": 13250
    },
    {
      "epoch": 3.3362653933149033,
      "grad_norm": 2.0026865005493164,
      "learning_rate": 1.1924137931034483e-06,
      "loss": 0.0627,
      "step": 13275
    },
    {
      "epoch": 3.342548378989696,
      "grad_norm": 0.9974460601806641,
      "learning_rate": 1.1751724137931036e-06,
      "loss": 0.0638,
      "step": 13300
    },
    {
      "epoch": 3.3488313646644885,
      "grad_norm": 1.6640186309814453,
      "learning_rate": 1.1579310344827587e-06,
      "loss": 0.0599,
      "step": 13325
    },
    {
      "epoch": 3.355114350339281,
      "grad_norm": 1.8832957744598389,
      "learning_rate": 1.140689655172414e-06,
      "loss": 0.0609,
      "step": 13350
    },
    {
      "epoch": 3.3613973360140736,
      "grad_norm": 1.3539857864379883,
      "learning_rate": 1.123448275862069e-06,
      "loss": 0.0647,
      "step": 13375
    },
    {
      "epoch": 3.3676803216888667,
      "grad_norm": 1.54263174533844,
      "learning_rate": 1.1062068965517241e-06,
      "loss": 0.0615,
      "step": 13400
    },
    {
      "epoch": 3.3739633073636592,
      "grad_norm": 1.5486729145050049,
      "learning_rate": 1.0889655172413794e-06,
      "loss": 0.0543,
      "step": 13425
    },
    {
      "epoch": 3.380246293038452,
      "grad_norm": 1.3974565267562866,
      "learning_rate": 1.0717241379310345e-06,
      "loss": 0.0661,
      "step": 13450
    },
    {
      "epoch": 3.3865292787132444,
      "grad_norm": 1.3940776586532593,
      "learning_rate": 1.0544827586206898e-06,
      "loss": 0.062,
      "step": 13475
    },
    {
      "epoch": 3.392812264388037,
      "grad_norm": 2.181025981903076,
      "learning_rate": 1.0372413793103449e-06,
      "loss": 0.0668,
      "step": 13500
    },
    {
      "epoch": 3.392812264388037,
      "eval_loss": 0.16507452726364136,
      "eval_runtime": 11681.06,
      "eval_samples_per_second": 1.672,
      "eval_steps_per_second": 0.209,
      "eval_wer": 43.06982891230401,
      "step": 13500
    },
    {
      "epoch": 3.39909525006283,
      "grad_norm": 1.663004755973816,
      "learning_rate": 1.02e-06,
      "loss": 0.0597,
      "step": 13525
    },
    {
      "epoch": 3.4053782357376225,
      "grad_norm": 1.9507300853729248,
      "learning_rate": 1.0027586206896553e-06,
      "loss": 0.0623,
      "step": 13550
    },
    {
      "epoch": 3.411661221412415,
      "grad_norm": 2.0604774951934814,
      "learning_rate": 9.855172413793104e-07,
      "loss": 0.061,
      "step": 13575
    },
    {
      "epoch": 3.4179442070872077,
      "grad_norm": 1.7244702577590942,
      "learning_rate": 9.682758620689656e-07,
      "loss": 0.0609,
      "step": 13600
    },
    {
      "epoch": 3.4242271927620003,
      "grad_norm": 1.3366154432296753,
      "learning_rate": 9.510344827586207e-07,
      "loss": 0.0623,
      "step": 13625
    },
    {
      "epoch": 3.4305101784367933,
      "grad_norm": 1.6767691373825073,
      "learning_rate": 9.33793103448276e-07,
      "loss": 0.0597,
      "step": 13650
    },
    {
      "epoch": 3.436793164111586,
      "grad_norm": 1.580768346786499,
      "learning_rate": 9.165517241379311e-07,
      "loss": 0.073,
      "step": 13675
    },
    {
      "epoch": 3.4430761497863784,
      "grad_norm": 1.6998015642166138,
      "learning_rate": 8.993103448275863e-07,
      "loss": 0.0591,
      "step": 13700
    },
    {
      "epoch": 3.449359135461171,
      "grad_norm": 1.5298068523406982,
      "learning_rate": 8.820689655172414e-07,
      "loss": 0.0598,
      "step": 13725
    },
    {
      "epoch": 3.4556421211359636,
      "grad_norm": 2.080108165740967,
      "learning_rate": 8.648275862068967e-07,
      "loss": 0.0668,
      "step": 13750
    },
    {
      "epoch": 3.4619251068107566,
      "grad_norm": 1.8349488973617554,
      "learning_rate": 8.475862068965517e-07,
      "loss": 0.0671,
      "step": 13775
    },
    {
      "epoch": 3.468208092485549,
      "grad_norm": 2.383136034011841,
      "learning_rate": 8.303448275862069e-07,
      "loss": 0.0647,
      "step": 13800
    },
    {
      "epoch": 3.4744910781603418,
      "grad_norm": 1.7982120513916016,
      "learning_rate": 8.131034482758621e-07,
      "loss": 0.0708,
      "step": 13825
    },
    {
      "epoch": 3.4807740638351343,
      "grad_norm": 1.5413868427276611,
      "learning_rate": 7.958620689655173e-07,
      "loss": 0.0624,
      "step": 13850
    },
    {
      "epoch": 3.487057049509927,
      "grad_norm": 2.7131173610687256,
      "learning_rate": 7.786206896551725e-07,
      "loss": 0.0709,
      "step": 13875
    },
    {
      "epoch": 3.49334003518472,
      "grad_norm": 1.5475999116897583,
      "learning_rate": 7.613793103448276e-07,
      "loss": 0.0588,
      "step": 13900
    },
    {
      "epoch": 3.4996230208595125,
      "grad_norm": 0.9188130497932434,
      "learning_rate": 7.441379310344828e-07,
      "loss": 0.0541,
      "step": 13925
    },
    {
      "epoch": 3.505906006534305,
      "grad_norm": 1.8402552604675293,
      "learning_rate": 7.268965517241381e-07,
      "loss": 0.0611,
      "step": 13950
    },
    {
      "epoch": 3.5121889922090976,
      "grad_norm": 1.9850478172302246,
      "learning_rate": 7.096551724137931e-07,
      "loss": 0.0586,
      "step": 13975
    },
    {
      "epoch": 3.51847197788389,
      "grad_norm": 1.791929006576538,
      "learning_rate": 6.924137931034483e-07,
      "loss": 0.0582,
      "step": 14000
    },
    {
      "epoch": 3.51847197788389,
      "eval_loss": 0.16593758761882782,
      "eval_runtime": 11717.6125,
      "eval_samples_per_second": 1.667,
      "eval_steps_per_second": 0.208,
      "eval_wer": 42.9048076436985,
      "step": 14000
    },
    {
      "epoch": 3.5247549635586832,
      "grad_norm": 1.5299566984176636,
      "learning_rate": 6.758620689655172e-07,
      "loss": 0.0644,
      "step": 14025
    },
    {
      "epoch": 3.531037949233476,
      "grad_norm": 1.8366338014602661,
      "learning_rate": 6.586206896551725e-07,
      "loss": 0.0621,
      "step": 14050
    },
    {
      "epoch": 3.5373209349082684,
      "grad_norm": 1.5470293760299683,
      "learning_rate": 6.413793103448277e-07,
      "loss": 0.0624,
      "step": 14075
    },
    {
      "epoch": 3.543603920583061,
      "grad_norm": 1.9549955129623413,
      "learning_rate": 6.241379310344828e-07,
      "loss": 0.064,
      "step": 14100
    },
    {
      "epoch": 3.5498869062578535,
      "grad_norm": 1.6928914785385132,
      "learning_rate": 6.068965517241379e-07,
      "loss": 0.0688,
      "step": 14125
    },
    {
      "epoch": 3.5561698919326465,
      "grad_norm": 1.8158848285675049,
      "learning_rate": 5.896551724137931e-07,
      "loss": 0.0666,
      "step": 14150
    },
    {
      "epoch": 3.562452877607439,
      "grad_norm": 1.2600568532943726,
      "learning_rate": 5.724137931034483e-07,
      "loss": 0.057,
      "step": 14175
    },
    {
      "epoch": 3.5687358632822317,
      "grad_norm": 1.5076507329940796,
      "learning_rate": 5.551724137931035e-07,
      "loss": 0.0643,
      "step": 14200
    },
    {
      "epoch": 3.5750188489570243,
      "grad_norm": 1.486598014831543,
      "learning_rate": 5.379310344827587e-07,
      "loss": 0.063,
      "step": 14225
    },
    {
      "epoch": 3.581301834631817,
      "grad_norm": 1.381836175918579,
      "learning_rate": 5.206896551724138e-07,
      "loss": 0.06,
      "step": 14250
    },
    {
      "epoch": 3.58758482030661,
      "grad_norm": 1.3430190086364746,
      "learning_rate": 5.03448275862069e-07,
      "loss": 0.0551,
      "step": 14275
    },
    {
      "epoch": 3.5938678059814024,
      "grad_norm": 1.7509955167770386,
      "learning_rate": 4.862068965517241e-07,
      "loss": 0.0555,
      "step": 14300
    },
    {
      "epoch": 3.600150791656195,
      "grad_norm": 1.3928794860839844,
      "learning_rate": 4.6896551724137934e-07,
      "loss": 0.0577,
      "step": 14325
    },
    {
      "epoch": 3.6064337773309876,
      "grad_norm": 1.7978532314300537,
      "learning_rate": 4.5172413793103447e-07,
      "loss": 0.0668,
      "step": 14350
    },
    {
      "epoch": 3.61271676300578,
      "grad_norm": 1.4505749940872192,
      "learning_rate": 4.344827586206897e-07,
      "loss": 0.0608,
      "step": 14375
    },
    {
      "epoch": 3.618999748680573,
      "grad_norm": 1.6973689794540405,
      "learning_rate": 4.1724137931034485e-07,
      "loss": 0.0501,
      "step": 14400
    },
    {
      "epoch": 3.6252827343553657,
      "grad_norm": 2.222531318664551,
      "learning_rate": 4.0000000000000003e-07,
      "loss": 0.0645,
      "step": 14425
    },
    {
      "epoch": 3.6315657200301583,
      "grad_norm": 2.4952709674835205,
      "learning_rate": 3.8275862068965517e-07,
      "loss": 0.0537,
      "step": 14450
    },
    {
      "epoch": 3.637848705704951,
      "grad_norm": 2.3859033584594727,
      "learning_rate": 3.6551724137931036e-07,
      "loss": 0.0652,
      "step": 14475
    },
    {
      "epoch": 3.6441316913797435,
      "grad_norm": 1.629402756690979,
      "learning_rate": 3.4827586206896555e-07,
      "loss": 0.055,
      "step": 14500
    },
    {
      "epoch": 3.6441316913797435,
      "eval_loss": 0.16478079557418823,
      "eval_runtime": 11635.7899,
      "eval_samples_per_second": 1.679,
      "eval_steps_per_second": 0.21,
      "eval_wer": 42.71378740060809,
      "step": 14500
    },
    {
      "epoch": 3.6504146770545365,
      "grad_norm": 1.4962036609649658,
      "learning_rate": 3.3103448275862073e-07,
      "loss": 0.0589,
      "step": 14525
    },
    {
      "epoch": 3.656697662729329,
      "grad_norm": 1.3538328409194946,
      "learning_rate": 3.1379310344827587e-07,
      "loss": 0.0614,
      "step": 14550
    },
    {
      "epoch": 3.6629806484041216,
      "grad_norm": 1.8565229177474976,
      "learning_rate": 2.9655172413793106e-07,
      "loss": 0.082,
      "step": 14575
    },
    {
      "epoch": 3.669263634078914,
      "grad_norm": 1.5180363655090332,
      "learning_rate": 2.7931034482758624e-07,
      "loss": 0.064,
      "step": 14600
    },
    {
      "epoch": 3.675546619753707,
      "grad_norm": 2.005053997039795,
      "learning_rate": 2.6206896551724143e-07,
      "loss": 0.0734,
      "step": 14625
    },
    {
      "epoch": 3.6818296054285,
      "grad_norm": 1.190696120262146,
      "learning_rate": 2.4482758620689657e-07,
      "loss": 0.0606,
      "step": 14650
    },
    {
      "epoch": 3.6881125911032924,
      "grad_norm": 1.364521861076355,
      "learning_rate": 2.2758620689655175e-07,
      "loss": 0.0652,
      "step": 14675
    },
    {
      "epoch": 3.694395576778085,
      "grad_norm": 1.3325939178466797,
      "learning_rate": 2.1034482758620692e-07,
      "loss": 0.0635,
      "step": 14700
    },
    {
      "epoch": 3.7006785624528775,
      "grad_norm": 2.166093587875366,
      "learning_rate": 1.931034482758621e-07,
      "loss": 0.0638,
      "step": 14725
    },
    {
      "epoch": 3.70696154812767,
      "grad_norm": 1.497362494468689,
      "learning_rate": 1.7586206896551726e-07,
      "loss": 0.0595,
      "step": 14750
    },
    {
      "epoch": 3.713244533802463,
      "grad_norm": 1.7004494667053223,
      "learning_rate": 1.5862068965517243e-07,
      "loss": 0.0591,
      "step": 14775
    },
    {
      "epoch": 3.7195275194772557,
      "grad_norm": 1.4566409587860107,
      "learning_rate": 1.413793103448276e-07,
      "loss": 0.0621,
      "step": 14800
    },
    {
      "epoch": 3.7258105051520483,
      "grad_norm": 0.7791139483451843,
      "learning_rate": 1.2413793103448277e-07,
      "loss": 0.0569,
      "step": 14825
    },
    {
      "epoch": 3.732093490826841,
      "grad_norm": 1.436057209968567,
      "learning_rate": 1.0689655172413794e-07,
      "loss": 0.0636,
      "step": 14850
    },
    {
      "epoch": 3.7383764765016334,
      "grad_norm": 1.0641875267028809,
      "learning_rate": 8.965517241379311e-08,
      "loss": 0.0595,
      "step": 14875
    },
    {
      "epoch": 3.7446594621764264,
      "grad_norm": 1.1096875667572021,
      "learning_rate": 7.241379310344829e-08,
      "loss": 0.0588,
      "step": 14900
    },
    {
      "epoch": 3.750942447851219,
      "grad_norm": 1.2665530443191528,
      "learning_rate": 5.517241379310345e-08,
      "loss": 0.0627,
      "step": 14925
    },
    {
      "epoch": 3.7572254335260116,
      "grad_norm": 1.4025217294692993,
      "learning_rate": 3.793103448275862e-08,
      "loss": 0.056,
      "step": 14950
    },
    {
      "epoch": 3.763508419200804,
      "grad_norm": 1.0198439359664917,
      "learning_rate": 2.0689655172413796e-08,
      "loss": 0.0584,
      "step": 14975
    },
    {
      "epoch": 3.7697914048755967,
      "grad_norm": 1.531874656677246,
      "learning_rate": 3.4482758620689654e-09,
      "loss": 0.0577,
      "step": 15000
    },
    {
      "epoch": 3.7697914048755967,
      "eval_loss": 0.16528290510177612,
      "eval_runtime": 11264.7491,
      "eval_samples_per_second": 1.734,
      "eval_steps_per_second": 0.217,
      "eval_wer": 42.7856456773094,
      "step": 15000
    },
    {
      "epoch": 326.6304347826087,
      "grad_norm": 5.988070964813232,
      "learning_rate": 1.4594117647058826e-06,
      "loss": 0.6353,
      "step": 15025
    },
    {
      "epoch": 327.17391304347825,
      "grad_norm": 4.414798736572266,
      "learning_rate": 1.4447058823529413e-06,
      "loss": 0.4939,
      "step": 15050
    },
    {
      "epoch": 327.7173913043478,
      "grad_norm": 5.318791389465332,
      "learning_rate": 1.43e-06,
      "loss": 0.4263,
      "step": 15075
    },
    {
      "epoch": 328.2608695652174,
      "grad_norm": 3.9256298542022705,
      "learning_rate": 1.4152941176470588e-06,
      "loss": 0.4097,
      "step": 15100
    },
    {
      "epoch": 328.80434782608694,
      "grad_norm": 5.1088948249816895,
      "learning_rate": 1.4005882352941178e-06,
      "loss": 0.358,
      "step": 15125
    },
    {
      "epoch": 329.3478260869565,
      "grad_norm": 4.497746467590332,
      "learning_rate": 1.3858823529411767e-06,
      "loss": 0.3366,
      "step": 15150
    },
    {
      "epoch": 329.89130434782606,
      "grad_norm": 3.794694662094116,
      "learning_rate": 1.3711764705882355e-06,
      "loss": 0.3104,
      "step": 15175
    },
    {
      "epoch": 330.4347826086956,
      "grad_norm": 4.165380954742432,
      "learning_rate": 1.3564705882352942e-06,
      "loss": 0.2858,
      "step": 15200
    },
    {
      "epoch": 330.9782608695652,
      "grad_norm": 3.9095475673675537,
      "learning_rate": 1.341764705882353e-06,
      "loss": 0.2827,
      "step": 15225
    },
    {
      "epoch": 331.5217391304348,
      "grad_norm": 3.926863670349121,
      "learning_rate": 1.3270588235294119e-06,
      "loss": 0.2574,
      "step": 15250
    },
    {
      "epoch": 332.0652173913044,
      "grad_norm": 4.20143985748291,
      "learning_rate": 1.3123529411764706e-06,
      "loss": 0.2428,
      "step": 15275
    },
    {
      "epoch": 332.60869565217394,
      "grad_norm": 3.6490259170532227,
      "learning_rate": 1.2976470588235296e-06,
      "loss": 0.2254,
      "step": 15300
    },
    {
      "epoch": 333.1521739130435,
      "grad_norm": 3.7255146503448486,
      "learning_rate": 1.2829411764705885e-06,
      "loss": 0.2145,
      "step": 15325
    },
    {
      "epoch": 333.69565217391306,
      "grad_norm": 3.6490657329559326,
      "learning_rate": 1.268235294117647e-06,
      "loss": 0.1916,
      "step": 15350
    },
    {
      "epoch": 334.2391304347826,
      "grad_norm": 3.2931156158447266,
      "learning_rate": 1.253529411764706e-06,
      "loss": 0.1845,
      "step": 15375
    },
    {
      "epoch": 334.7826086956522,
      "grad_norm": 2.9920129776000977,
      "learning_rate": 1.2388235294117648e-06,
      "loss": 0.173,
      "step": 15400
    },
    {
      "epoch": 335.32608695652175,
      "grad_norm": 3.3761260509490967,
      "learning_rate": 1.2241176470588237e-06,
      "loss": 0.1579,
      "step": 15425
    },
    {
      "epoch": 335.8695652173913,
      "grad_norm": 3.4968042373657227,
      "learning_rate": 1.2094117647058824e-06,
      "loss": 0.1404,
      "step": 15450
    },
    {
      "epoch": 336.4130434782609,
      "grad_norm": 3.300462007522583,
      "learning_rate": 1.1947058823529414e-06,
      "loss": 0.1338,
      "step": 15475
    },
    {
      "epoch": 336.95652173913044,
      "grad_norm": 3.894838809967041,
      "learning_rate": 1.1800000000000001e-06,
      "loss": 0.1218,
      "step": 15500
    },
    {
      "epoch": 336.95652173913044,
      "eval_loss": 0.48647284507751465,
      "eval_runtime": 92.508,
      "eval_samples_per_second": 1.978,
      "eval_steps_per_second": 0.249,
      "eval_wer": 72.84011916583913,
      "step": 15500
    },
    {
      "epoch": 337.5,
      "grad_norm": 2.9717953205108643,
      "learning_rate": 1.1652941176470589e-06,
      "loss": 0.1067,
      "step": 15525
    },
    {
      "epoch": 338.04347826086956,
      "grad_norm": 2.478729724884033,
      "learning_rate": 1.1505882352941178e-06,
      "loss": 0.1015,
      "step": 15550
    },
    {
      "epoch": 338.5869565217391,
      "grad_norm": 3.0605814456939697,
      "learning_rate": 1.1358823529411766e-06,
      "loss": 0.0852,
      "step": 15575
    },
    {
      "epoch": 339.1304347826087,
      "grad_norm": 2.1662635803222656,
      "learning_rate": 1.1211764705882353e-06,
      "loss": 0.0824,
      "step": 15600
    },
    {
      "epoch": 339.67391304347825,
      "grad_norm": 3.0278255939483643,
      "learning_rate": 1.1064705882352943e-06,
      "loss": 0.0691,
      "step": 15625
    },
    {
      "epoch": 340.2173913043478,
      "grad_norm": 2.406172275543213,
      "learning_rate": 1.091764705882353e-06,
      "loss": 0.0586,
      "step": 15650
    },
    {
      "epoch": 340.7608695652174,
      "grad_norm": 1.9660167694091797,
      "learning_rate": 1.077058823529412e-06,
      "loss": 0.0554,
      "step": 15675
    },
    {
      "epoch": 341.30434782608694,
      "grad_norm": 1.6723525524139404,
      "learning_rate": 1.0623529411764707e-06,
      "loss": 0.0461,
      "step": 15700
    },
    {
      "epoch": 341.8478260869565,
      "grad_norm": 2.474976062774658,
      "learning_rate": 1.0476470588235294e-06,
      "loss": 0.0409,
      "step": 15725
    },
    {
      "epoch": 342.39130434782606,
      "grad_norm": 1.8706433773040771,
      "learning_rate": 1.0329411764705884e-06,
      "loss": 0.0348,
      "step": 15750
    },
    {
      "epoch": 342.9347826086956,
      "grad_norm": 1.806139588356018,
      "learning_rate": 1.0182352941176471e-06,
      "loss": 0.0313,
      "step": 15775
    },
    {
      "epoch": 343.4782608695652,
      "grad_norm": 1.2860676050186157,
      "learning_rate": 1.003529411764706e-06,
      "loss": 0.0243,
      "step": 15800
    },
    {
      "epoch": 344.0217391304348,
      "grad_norm": 1.7633931636810303,
      "learning_rate": 9.888235294117648e-07,
      "loss": 0.0226,
      "step": 15825
    },
    {
      "epoch": 344.5652173913044,
      "grad_norm": 1.3817349672317505,
      "learning_rate": 9.741176470588236e-07,
      "loss": 0.0179,
      "step": 15850
    },
    {
      "epoch": 345.10869565217394,
      "grad_norm": 0.9751259684562683,
      "learning_rate": 9.594117647058825e-07,
      "loss": 0.0147,
      "step": 15875
    },
    {
      "epoch": 345.6521739130435,
      "grad_norm": 1.0445324182510376,
      "learning_rate": 9.447058823529411e-07,
      "loss": 0.0117,
      "step": 15900
    },
    {
      "epoch": 346.19565217391306,
      "grad_norm": 0.6687153577804565,
      "learning_rate": 9.300000000000001e-07,
      "loss": 0.0108,
      "step": 15925
    },
    {
      "epoch": 346.7391304347826,
      "grad_norm": 1.2395334243774414,
      "learning_rate": 9.152941176470589e-07,
      "loss": 0.0089,
      "step": 15950
    },
    {
      "epoch": 347.2826086956522,
      "grad_norm": 1.6808990240097046,
      "learning_rate": 9.005882352941177e-07,
      "loss": 0.0073,
      "step": 15975
    },
    {
      "epoch": 347.82608695652175,
      "grad_norm": 0.4099910259246826,
      "learning_rate": 8.858823529411765e-07,
      "loss": 0.0061,
      "step": 16000
    },
    {
      "epoch": 347.82608695652175,
      "eval_loss": 0.7469403743743896,
      "eval_runtime": 89.3932,
      "eval_samples_per_second": 2.047,
      "eval_steps_per_second": 0.257,
      "eval_wer": 73.63455809334657,
      "step": 16000
    },
    {
      "epoch": 348.3695652173913,
      "grad_norm": 0.28456559777259827,
      "learning_rate": 8.711764705882355e-07,
      "loss": 0.0049,
      "step": 16025
    },
    {
      "epoch": 348.9130434782609,
      "grad_norm": 0.3505021929740906,
      "learning_rate": 8.564705882352941e-07,
      "loss": 0.0048,
      "step": 16050
    },
    {
      "epoch": 349.45652173913044,
      "grad_norm": 0.3143045902252197,
      "learning_rate": 8.417647058823531e-07,
      "loss": 0.0041,
      "step": 16075
    },
    {
      "epoch": 350.0,
      "grad_norm": 0.5929996371269226,
      "learning_rate": 8.270588235294118e-07,
      "loss": 0.0039,
      "step": 16100
    },
    {
      "epoch": 350.54347826086956,
      "grad_norm": 0.3261135220527649,
      "learning_rate": 8.123529411764707e-07,
      "loss": 0.003,
      "step": 16125
    },
    {
      "epoch": 351.0869565217391,
      "grad_norm": 0.2300560474395752,
      "learning_rate": 7.976470588235295e-07,
      "loss": 0.0034,
      "step": 16150
    },
    {
      "epoch": 351.6304347826087,
      "grad_norm": 0.23772715032100677,
      "learning_rate": 7.829411764705882e-07,
      "loss": 0.0028,
      "step": 16175
    },
    {
      "epoch": 352.17391304347825,
      "grad_norm": 0.21742898225784302,
      "learning_rate": 7.682352941176471e-07,
      "loss": 0.0027,
      "step": 16200
    },
    {
      "epoch": 352.7173913043478,
      "grad_norm": 0.2496710568666458,
      "learning_rate": 7.53529411764706e-07,
      "loss": 0.0025,
      "step": 16225
    },
    {
      "epoch": 353.2608695652174,
      "grad_norm": 0.18565493822097778,
      "learning_rate": 7.388235294117648e-07,
      "loss": 0.0024,
      "step": 16250
    },
    {
      "epoch": 353.80434782608694,
      "grad_norm": 0.173939049243927,
      "learning_rate": 7.241176470588236e-07,
      "loss": 0.0023,
      "step": 16275
    },
    {
      "epoch": 354.3478260869565,
      "grad_norm": 0.16787025332450867,
      "learning_rate": 7.094117647058824e-07,
      "loss": 0.0021,
      "step": 16300
    },
    {
      "epoch": 354.89130434782606,
      "grad_norm": 0.16970224678516388,
      "learning_rate": 6.947058823529412e-07,
      "loss": 0.002,
      "step": 16325
    },
    {
      "epoch": 355.4347826086956,
      "grad_norm": 0.152974471449852,
      "learning_rate": 6.800000000000001e-07,
      "loss": 0.0019,
      "step": 16350
    },
    {
      "epoch": 355.9782608695652,
      "grad_norm": 0.15272676944732666,
      "learning_rate": 6.652941176470588e-07,
      "loss": 0.0018,
      "step": 16375
    },
    {
      "epoch": 356.5217391304348,
      "grad_norm": 0.15291406214237213,
      "learning_rate": 6.505882352941177e-07,
      "loss": 0.0017,
      "step": 16400
    },
    {
      "epoch": 357.0652173913044,
      "grad_norm": 0.11471806466579437,
      "learning_rate": 6.358823529411766e-07,
      "loss": 0.0017,
      "step": 16425
    },
    {
      "epoch": 357.60869565217394,
      "grad_norm": 0.14267627894878387,
      "learning_rate": 6.211764705882353e-07,
      "loss": 0.0016,
      "step": 16450
    },
    {
      "epoch": 358.1521739130435,
      "grad_norm": 0.12022901326417923,
      "learning_rate": 6.064705882352942e-07,
      "loss": 0.0017,
      "step": 16475
    },
    {
      "epoch": 358.69565217391306,
      "grad_norm": 0.11677850037813187,
      "learning_rate": 5.91764705882353e-07,
      "loss": 0.0015,
      "step": 16500
    },
    {
      "epoch": 358.69565217391306,
      "eval_loss": 0.8427050113677979,
      "eval_runtime": 88.306,
      "eval_samples_per_second": 2.072,
      "eval_steps_per_second": 0.26,
      "eval_wer": 74.87586891757697,
      "step": 16500
    },
    {
      "epoch": 359.2391304347826,
      "grad_norm": 0.10254109650850296,
      "learning_rate": 5.770588235294118e-07,
      "loss": 0.0014,
      "step": 16525
    },
    {
      "epoch": 359.7826086956522,
      "grad_norm": 0.13138434290885925,
      "learning_rate": 5.623529411764706e-07,
      "loss": 0.0014,
      "step": 16550
    },
    {
      "epoch": 360.32608695652175,
      "grad_norm": 0.11177344620227814,
      "learning_rate": 5.476470588235295e-07,
      "loss": 0.0013,
      "step": 16575
    },
    {
      "epoch": 360.8695652173913,
      "grad_norm": 0.08989901840686798,
      "learning_rate": 5.329411764705883e-07,
      "loss": 0.0013,
      "step": 16600
    },
    {
      "epoch": 361.4130434782609,
      "grad_norm": 0.09234068542718887,
      "learning_rate": 5.182352941176471e-07,
      "loss": 0.0013,
      "step": 16625
    },
    {
      "epoch": 361.95652173913044,
      "grad_norm": 0.09787598997354507,
      "learning_rate": 5.035294117647059e-07,
      "loss": 0.0013,
      "step": 16650
    },
    {
      "epoch": 362.5,
      "grad_norm": 0.08090800791978836,
      "learning_rate": 4.888235294117647e-07,
      "loss": 0.0012,
      "step": 16675
    },
    {
      "epoch": 363.04347826086956,
      "grad_norm": 0.09755431860685349,
      "learning_rate": 4.741176470588236e-07,
      "loss": 0.0012,
      "step": 16700
    },
    {
      "epoch": 363.5869565217391,
      "grad_norm": 0.08762162923812866,
      "learning_rate": 4.5941176470588237e-07,
      "loss": 0.0012,
      "step": 16725
    },
    {
      "epoch": 364.1304347826087,
      "grad_norm": 0.0933767557144165,
      "learning_rate": 4.447058823529412e-07,
      "loss": 0.0011,
      "step": 16750
    },
    {
      "epoch": 364.67391304347825,
      "grad_norm": 0.09859120100736618,
      "learning_rate": 4.3e-07,
      "loss": 0.0011,
      "step": 16775
    },
    {
      "epoch": 365.2173913043478,
      "grad_norm": 0.08420507609844208,
      "learning_rate": 4.1529411764705886e-07,
      "loss": 0.0011,
      "step": 16800
    },
    {
      "epoch": 365.7608695652174,
      "grad_norm": 0.10694745182991028,
      "learning_rate": 4.005882352941177e-07,
      "loss": 0.0011,
      "step": 16825
    },
    {
      "epoch": 366.30434782608694,
      "grad_norm": 0.08253324031829834,
      "learning_rate": 3.858823529411765e-07,
      "loss": 0.0011,
      "step": 16850
    },
    {
      "epoch": 366.8478260869565,
      "grad_norm": 0.07511500269174576,
      "learning_rate": 3.711764705882353e-07,
      "loss": 0.001,
      "step": 16875
    },
    {
      "epoch": 367.39130434782606,
      "grad_norm": 0.08655669540166855,
      "learning_rate": 3.564705882352942e-07,
      "loss": 0.001,
      "step": 16900
    },
    {
      "epoch": 367.9347826086956,
      "grad_norm": 0.0791034922003746,
      "learning_rate": 3.41764705882353e-07,
      "loss": 0.001,
      "step": 16925
    },
    {
      "epoch": 368.4782608695652,
      "grad_norm": 0.07935862243175507,
      "learning_rate": 3.270588235294118e-07,
      "loss": 0.001,
      "step": 16950
    },
    {
      "epoch": 369.0217391304348,
      "grad_norm": 0.07234132289886475,
      "learning_rate": 3.123529411764706e-07,
      "loss": 0.001,
      "step": 16975
    },
    {
      "epoch": 369.5652173913044,
      "grad_norm": 0.07198411971330643,
      "learning_rate": 2.976470588235294e-07,
      "loss": 0.0009,
      "step": 17000
    },
    {
      "epoch": 369.5652173913044,
      "eval_loss": 0.8779386281967163,
      "eval_runtime": 88.1801,
      "eval_samples_per_second": 2.075,
      "eval_steps_per_second": 0.261,
      "eval_wer": 75.42204568023834,
      "step": 17000
    },
    {
      "epoch": 370.10869565217394,
      "grad_norm": 0.07644437998533249,
      "learning_rate": 2.8294117647058826e-07,
      "loss": 0.0009,
      "step": 17025
    },
    {
      "epoch": 370.6521739130435,
      "grad_norm": 0.06603951007127762,
      "learning_rate": 2.6823529411764705e-07,
      "loss": 0.0009,
      "step": 17050
    },
    {
      "epoch": 371.19565217391306,
      "grad_norm": 0.08132907003164291,
      "learning_rate": 2.535294117647059e-07,
      "loss": 0.001,
      "step": 17075
    },
    {
      "epoch": 371.7391304347826,
      "grad_norm": 0.07193853706121445,
      "learning_rate": 2.388235294117647e-07,
      "loss": 0.0009,
      "step": 17100
    },
    {
      "epoch": 372.2826086956522,
      "grad_norm": 0.10432302951812744,
      "learning_rate": 2.2411764705882354e-07,
      "loss": 0.0009,
      "step": 17125
    },
    {
      "epoch": 372.82608695652175,
      "grad_norm": 0.0752280056476593,
      "learning_rate": 2.0941176470588236e-07,
      "loss": 0.0009,
      "step": 17150
    },
    {
      "epoch": 373.3695652173913,
      "grad_norm": 0.07817768305540085,
      "learning_rate": 1.947058823529412e-07,
      "loss": 0.0009,
      "step": 17175
    },
    {
      "epoch": 373.9130434782609,
      "grad_norm": 0.06475182622671127,
      "learning_rate": 1.8e-07,
      "loss": 0.0009,
      "step": 17200
    },
    {
      "epoch": 374.45652173913044,
      "grad_norm": 0.06815166771411896,
      "learning_rate": 1.6529411764705884e-07,
      "loss": 0.0008,
      "step": 17225
    },
    {
      "epoch": 375.0,
      "grad_norm": 0.10759684443473816,
      "learning_rate": 1.5058823529411766e-07,
      "loss": 0.0008,
      "step": 17250
    },
    {
      "epoch": 375.54347826086956,
      "grad_norm": 0.06912076473236084,
      "learning_rate": 1.3588235294117648e-07,
      "loss": 0.0008,
      "step": 17275
    },
    {
      "epoch": 376.0869565217391,
      "grad_norm": 0.06856230646371841,
      "learning_rate": 1.211764705882353e-07,
      "loss": 0.0009,
      "step": 17300
    },
    {
      "epoch": 376.6304347826087,
      "grad_norm": 0.06209594011306763,
      "learning_rate": 1.0647058823529413e-07,
      "loss": 0.0009,
      "step": 17325
    },
    {
      "epoch": 377.17391304347825,
      "grad_norm": 0.0590556301176548,
      "learning_rate": 9.176470588235295e-08,
      "loss": 0.0008,
      "step": 17350
    },
    {
      "epoch": 377.7173913043478,
      "grad_norm": 0.06120636314153671,
      "learning_rate": 7.705882352941177e-08,
      "loss": 0.0008,
      "step": 17375
    },
    {
      "epoch": 378.2608695652174,
      "grad_norm": 0.07146897912025452,
      "learning_rate": 6.235294117647059e-08,
      "loss": 0.0008,
      "step": 17400
    },
    {
      "epoch": 378.80434782608694,
      "grad_norm": 0.06837107986211777,
      "learning_rate": 4.764705882352941e-08,
      "loss": 0.0008,
      "step": 17425
    },
    {
      "epoch": 379.3478260869565,
      "grad_norm": 0.061426736414432526,
      "learning_rate": 3.294117647058824e-08,
      "loss": 0.0008,
      "step": 17450
    },
    {
      "epoch": 379.89130434782606,
      "grad_norm": 0.07778898626565933,
      "learning_rate": 1.8235294117647058e-08,
      "loss": 0.0008,
      "step": 17475
    },
    {
      "epoch": 380.4347826086956,
      "grad_norm": 0.07241355627775192,
      "learning_rate": 3.5294117647058827e-09,
      "loss": 0.0008,
      "step": 17500
    },
    {
      "epoch": 380.4347826086956,
      "eval_loss": 0.8890466094017029,
      "eval_runtime": 88.7625,
      "eval_samples_per_second": 2.062,
      "eval_steps_per_second": 0.259,
      "eval_wer": 75.76961271102284,
      "step": 17500
    }
  ],
  "logging_steps": 25,
  "max_steps": 17500,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 381,
  "save_steps": 500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 8.063480104501248e+19,
  "train_batch_size": 16,
  "trial_name": null,
  "trial_params": null
}