{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.5782491261518907,
  "global_step": 7279,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0,
      "learning_rate": 7.936507936507937e-07,
      "loss": 2.9309,
      "theoretical_loss": 3.4868973533572363,
      "tokens_seen": 1650130944
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.5873015873015873e-06,
      "loss": 3.0573,
      "theoretical_loss": 3.4868733789973354,
      "tokens_seen": 1650262016
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.3809523809523808e-06,
      "loss": 2.9048,
      "theoretical_loss": 3.4868494070746388,
      "tokens_seen": 1650393088
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.1746031746031746e-06,
      "loss": 3.1852,
      "theoretical_loss": 3.4868254375887053,
      "tokens_seen": 1650524160
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.968253968253968e-06,
      "loss": 2.9406,
      "theoretical_loss": 3.4868014705390937,
      "tokens_seen": 1650655232
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7619047619047615e-06,
      "loss": 2.9301,
      "theoretical_loss": 3.4867775059253625,
      "tokens_seen": 1650786304
    },
    {
      "epoch": 0.0,
      "learning_rate": 5.555555555555556e-06,
      "loss": 2.9694,
      "theoretical_loss": 3.4867535437470716,
      "tokens_seen": 1650917376
    },
    {
      "epoch": 0.0,
      "learning_rate": 6.349206349206349e-06,
      "loss": 2.8537,
      "theoretical_loss": 3.486729584003779,
      "tokens_seen": 1651048448
    },
    {
      "epoch": 0.0,
      "learning_rate": 7.142857142857143e-06,
      "loss": 3.1275,
      "theoretical_loss": 3.4867056266950454,
      "tokens_seen": 1651179520
    },
    {
      "epoch": 0.0,
      "learning_rate": 7.936507936507936e-06,
      "loss": 2.8591,
      "theoretical_loss": 3.4866816718204294,
      "tokens_seen": 1651310592
    },
    {
      "epoch": 0.0,
      "learning_rate": 8.73015873015873e-06,
      "loss": 2.9591,
      "theoretical_loss": 3.4866577193794903,
      "tokens_seen": 1651441664
    },
    {
      "epoch": 0.0,
      "learning_rate": 9.523809523809523e-06,
      "loss": 2.9381,
      "theoretical_loss": 3.486633769371788,
      "tokens_seen": 1651572736
    },
    {
      "epoch": 0.0,
      "objective/train/docs_used": 911303,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.985459327697754,
      "objective/train/theoretical_loss": 3.486621795280263,
      "objective/train/tokens_used": 22097376,
      "theoretical_loss": 3.486621795280263,
      "tokens_seen": 1651638272
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.0317460317460318e-05,
      "loss": 3.0636,
      "theoretical_loss": 3.4866098217968826,
      "tokens_seen": 1651703808
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.1111111111111112e-05,
      "loss": 2.9173,
      "theoretical_loss": 3.486585876654333,
      "tokens_seen": 1651834880
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.1904761904761905e-05,
      "loss": 2.9235,
      "theoretical_loss": 3.4865619339437,
      "tokens_seen": 1651965952
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.2698412698412699e-05,
      "loss": 2.9036,
      "theoretical_loss": 3.4865379936645438,
      "tokens_seen": 1652097024
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.3492063492063492e-05,
      "loss": 2.8327,
      "theoretical_loss": 3.486514055816424,
      "tokens_seen": 1652228096
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.4285714285714285e-05,
      "loss": 2.8069,
      "theoretical_loss": 3.4864901203989014,
      "tokens_seen": 1652359168
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.5079365079365079e-05,
      "loss": 2.6994,
      "theoretical_loss": 3.4864661874115357,
      "tokens_seen": 1652490240
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.5873015873015872e-05,
      "loss": 2.8277,
      "theoretical_loss": 3.486442256853888,
      "tokens_seen": 1652621312
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.6666666666666667e-05,
      "loss": 2.8779,
      "theoretical_loss": 3.4864183287255193,
      "tokens_seen": 1652752384
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.746031746031746e-05,
      "loss": 2.929,
      "theoretical_loss": 3.4863944030259897,
      "tokens_seen": 1652883456
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.8253968253968254e-05,
      "loss": 2.7098,
      "theoretical_loss": 3.4863704797548607,
      "tokens_seen": 1653014528
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.9047619047619046e-05,
      "loss": 2.7617,
      "theoretical_loss": 3.4863465589116927,
      "tokens_seen": 1653145600
    },
    {
      "epoch": 0.0,
      "objective/train/docs_used": 912643,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.1410162448883057,
      "objective/train/theoretical_loss": 3.486322640496047,
      "objective/train/tokens_used": 23735776,
      "theoretical_loss": 3.486322640496047,
      "tokens_seen": 1653276672
    },
    {
      "epoch": 0.0,
      "learning_rate": 1.984126984126984e-05,
      "loss": 2.8668,
      "theoretical_loss": 3.486322640496047,
      "tokens_seen": 1653276672
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.0634920634920636e-05,
      "loss": 2.7427,
      "theoretical_loss": 3.486298724507485,
      "tokens_seen": 1653407744
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.1428571428571428e-05,
      "loss": 2.9108,
      "theoretical_loss": 3.4862748109455675,
      "tokens_seen": 1653538816
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.2222222222222223e-05,
      "loss": 2.8479,
      "theoretical_loss": 3.486250899809857,
      "tokens_seen": 1653669888
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.3015873015873015e-05,
      "loss": 2.9489,
      "theoretical_loss": 3.4862269910999135,
      "tokens_seen": 1653800960
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.380952380952381e-05,
      "loss": 2.7904,
      "theoretical_loss": 3.4862030848153003,
      "tokens_seen": 1653932032
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.4603174603174602e-05,
      "loss": 2.6043,
      "theoretical_loss": 3.4861791809555784,
      "tokens_seen": 1654063104
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.5396825396825397e-05,
      "loss": 2.9128,
      "theoretical_loss": 3.48615527952031,
      "tokens_seen": 1654194176
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.6190476190476192e-05,
      "loss": 2.7843,
      "theoretical_loss": 3.486131380509057,
      "tokens_seen": 1654325248
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.6984126984126984e-05,
      "loss": 2.8452,
      "theoretical_loss": 3.4861074839213813,
      "tokens_seen": 1654456320
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.777777777777778e-05,
      "loss": 2.6435,
      "theoretical_loss": 3.4860835897568454,
      "tokens_seen": 1654587392
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.857142857142857e-05,
      "loss": 2.7978,
      "theoretical_loss": 3.4860596980150116,
      "tokens_seen": 1654718464
    },
    {
      "epoch": 0.0,
      "learning_rate": 2.9365079365079366e-05,
      "loss": 2.9673,
      "theoretical_loss": 3.4860358086954424,
      "tokens_seen": 1654849536
    },
    {
      "epoch": 0.0,
      "objective/train/docs_used": 913892,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.85264253616333,
      "objective/train/theoretical_loss": 3.48602386494387,
      "objective/train/tokens_used": 25374176,
      "theoretical_loss": 3.48602386494387,
      "tokens_seen": 1654915072
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.0158730158730158e-05,
      "loss": 2.9861,
      "theoretical_loss": 3.4860119217977004,
      "tokens_seen": 1654980608
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.095238095238095e-05,
      "loss": 2.9108,
      "theoretical_loss": 3.485988037321348,
      "tokens_seen": 1655111680
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.1746031746031745e-05,
      "loss": 2.7628,
      "theoretical_loss": 3.4859641552659486,
      "tokens_seen": 1655242752
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.253968253968254e-05,
      "loss": 2.6715,
      "theoretical_loss": 3.485940275631065,
      "tokens_seen": 1655373824
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.3333333333333335e-05,
      "loss": 2.9045,
      "theoretical_loss": 3.4859163984162596,
      "tokens_seen": 1655504896
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.412698412698413e-05,
      "loss": 2.7567,
      "theoretical_loss": 3.4858925236210965,
      "tokens_seen": 1655635968
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.492063492063492e-05,
      "loss": 2.787,
      "theoretical_loss": 3.4858686512451387,
      "tokens_seen": 1655767040
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.571428571428572e-05,
      "loss": 2.6928,
      "theoretical_loss": 3.4858447812879487,
      "tokens_seen": 1655898112
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.650793650793651e-05,
      "loss": 2.7383,
      "theoretical_loss": 3.485820913749091,
      "tokens_seen": 1656029184
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.730158730158731e-05,
      "loss": 2.6014,
      "theoretical_loss": 3.4857970486281293,
      "tokens_seen": 1656160256
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.809523809523809e-05,
      "loss": 2.8116,
      "theoretical_loss": 3.485773185924627,
      "tokens_seen": 1656291328
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.888888888888889e-05,
      "loss": 2.6689,
      "theoretical_loss": 3.4857493256381473,
      "tokens_seen": 1656422400
    },
    {
      "epoch": 0.0,
      "objective/train/docs_used": 914561,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.774996280670166,
      "objective/train/theoretical_loss": 3.4857254677682548,
      "objective/train/tokens_used": 27012576,
      "theoretical_loss": 3.4857254677682548,
      "tokens_seen": 1656553472
    },
    {
      "epoch": 0.0,
      "learning_rate": 3.968253968253968e-05,
      "loss": 2.6872,
      "theoretical_loss": 3.4857254677682548,
      "tokens_seen": 1656553472
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.047619047619048e-05,
      "loss": 2.5722,
      "theoretical_loss": 3.485701612314514,
      "tokens_seen": 1656684544
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.126984126984127e-05,
      "loss": 2.6782,
      "theoretical_loss": 3.4856777592764883,
      "tokens_seen": 1656815616
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.2063492063492065e-05,
      "loss": 2.6618,
      "theoretical_loss": 3.4856539086537426,
      "tokens_seen": 1656946688
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.2857142857142856e-05,
      "loss": 2.6818,
      "theoretical_loss": 3.485630060445841,
      "tokens_seen": 1657077760
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.3650793650793655e-05,
      "loss": 2.4444,
      "theoretical_loss": 3.485606214652347,
      "tokens_seen": 1657208832
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4444444444444447e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.4855823712728276,
      "tokens_seen": 1657339904
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.523809523809524e-05,
      "loss": 2.5876,
      "theoretical_loss": 3.4855585303068453,
      "tokens_seen": 1657470976
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.603174603174603e-05,
      "loss": 2.5061,
      "theoretical_loss": 3.4855346917539665,
      "tokens_seen": 1657602048
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.682539682539683e-05,
      "loss": 2.6817,
      "theoretical_loss": 3.4855108556137546,
      "tokens_seen": 1657733120
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.761904761904762e-05,
      "loss": 2.6159,
      "theoretical_loss": 3.4854870218857763,
      "tokens_seen": 1657864192
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.841269841269841e-05,
      "loss": 2.5539,
      "theoretical_loss": 3.485463190569596,
      "tokens_seen": 1657995264
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9206349206349204e-05,
      "loss": 2.6999,
      "theoretical_loss": 3.485439361664779,
      "tokens_seen": 1658126336
    },
    {
      "epoch": 0.0,
      "objective/train/docs_used": 915678,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6932287216186523,
      "objective/train/theoretical_loss": 3.4854274481164964,
      "objective/train/tokens_used": 28650976,
      "theoretical_loss": 3.4854274481164964,
      "tokens_seen": 1658191872
    },
    {
      "epoch": 0.01,
      "learning_rate": 5e-05,
      "loss": 2.5566,
      "theoretical_loss": 3.4854155351708913,
      "tokens_seen": 1658257408
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.0793650793650794e-05,
      "loss": 2.7198,
      "theoretical_loss": 3.4853917110874972,
      "tokens_seen": 1658388480
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.158730158730159e-05,
      "loss": 2.4993,
      "theoretical_loss": 3.4853678894141638,
      "tokens_seen": 1658519552
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.2380952380952384e-05,
      "loss": 2.5541,
      "theoretical_loss": 3.4853440701504557,
      "tokens_seen": 1658650624
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.3174603174603176e-05,
      "loss": 2.5189,
      "theoretical_loss": 3.4853202532959395,
      "tokens_seen": 1658781696
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.396825396825397e-05,
      "loss": 2.322,
      "theoretical_loss": 3.4852964388501806,
      "tokens_seen": 1658912768
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.4761904761904766e-05,
      "loss": 2.6172,
      "theoretical_loss": 3.485272626812746,
      "tokens_seen": 1659043840
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.555555555555556e-05,
      "loss": 2.5944,
      "theoretical_loss": 3.4852488171832006,
      "tokens_seen": 1659174912
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.634920634920635e-05,
      "loss": 2.5423,
      "theoretical_loss": 3.4852250099611117,
      "tokens_seen": 1659305984
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.714285714285714e-05,
      "loss": 2.5272,
      "theoretical_loss": 3.485201205146046,
      "tokens_seen": 1659437056
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.793650793650795e-05,
      "loss": 2.6078,
      "theoretical_loss": 3.4851774027375693,
      "tokens_seen": 1659568128
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.873015873015873e-05,
      "loss": 2.5119,
      "theoretical_loss": 3.4851536027352483,
      "tokens_seen": 1659699200
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 916417,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.456707239151001,
      "objective/train/theoretical_loss": 3.48512980513865,
      "objective/train/tokens_used": 30289376,
      "theoretical_loss": 3.48512980513865,
      "tokens_seen": 1659830272
    },
    {
      "epoch": 0.01,
      "learning_rate": 5.9523809523809524e-05,
      "loss": 2.6637,
      "theoretical_loss": 3.48512980513865,
      "tokens_seen": 1659830272
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.0317460317460316e-05,
      "loss": 2.3894,
      "theoretical_loss": 3.4851060099473417,
      "tokens_seen": 1659961344
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.111111111111112e-05,
      "loss": 2.5928,
      "theoretical_loss": 3.4850822171608895,
      "tokens_seen": 1660092416
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.19047619047619e-05,
      "loss": 2.4428,
      "theoretical_loss": 3.4850584267788607,
      "tokens_seen": 1660223488
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.26984126984127e-05,
      "loss": 2.7376,
      "theoretical_loss": 3.4850346388008235,
      "tokens_seen": 1660354560
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.349206349206349e-05,
      "loss": 2.4582,
      "theoretical_loss": 3.4850108532263437,
      "tokens_seen": 1660485632
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.428571428571429e-05,
      "loss": 2.409,
      "theoretical_loss": 3.48498707005499,
      "tokens_seen": 1660616704
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.507936507936509e-05,
      "loss": 2.5177,
      "theoretical_loss": 3.4849632892863296,
      "tokens_seen": 1660747776
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.587301587301587e-05,
      "loss": 2.6281,
      "theoretical_loss": 3.48493951091993,
      "tokens_seen": 1660878848
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.666666666666667e-05,
      "loss": 2.5806,
      "theoretical_loss": 3.4849157349553592,
      "tokens_seen": 1661009920
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.746031746031747e-05,
      "loss": 2.5608,
      "theoretical_loss": 3.4848919613921847,
      "tokens_seen": 1661140992
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.825396825396825e-05,
      "loss": 2.5385,
      "theoretical_loss": 3.484868190229975,
      "tokens_seen": 1661272064
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.904761904761905e-05,
      "loss": 2.5632,
      "theoretical_loss": 3.484844421468298,
      "tokens_seen": 1661403136
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 917689,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.0480103492736816,
      "objective/train/theoretical_loss": 3.484832537987524,
      "objective/train/tokens_used": 31927776,
      "theoretical_loss": 3.484832537987524,
      "tokens_seen": 1661468672
    },
    {
      "epoch": 0.01,
      "learning_rate": 6.984126984126984e-05,
      "loss": 2.3842,
      "theoretical_loss": 3.4848206551067213,
      "tokens_seen": 1661534208
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.063492063492065e-05,
      "loss": 2.4595,
      "theoretical_loss": 3.484796891144814,
      "tokens_seen": 1661665280
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.142857142857143e-05,
      "loss": 2.5882,
      "theoretical_loss": 3.4847731295821447,
      "tokens_seen": 1661796352
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.222222222222222e-05,
      "loss": 2.3931,
      "theoretical_loss": 3.4847493704182817,
      "tokens_seen": 1661927424
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.301587301587302e-05,
      "loss": 2.622,
      "theoretical_loss": 3.4847256136527935,
      "tokens_seen": 1662058496
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.380952380952382e-05,
      "loss": 2.5425,
      "theoretical_loss": 3.4847018592852494,
      "tokens_seen": 1662189568
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.460317460317461e-05,
      "loss": 2.601,
      "theoretical_loss": 3.4846781073152173,
      "tokens_seen": 1662320640
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.53968253968254e-05,
      "loss": 2.5712,
      "theoretical_loss": 3.484654357742267,
      "tokens_seen": 1662451712
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.619047619047618e-05,
      "loss": 2.2686,
      "theoretical_loss": 3.4846306105659677,
      "tokens_seen": 1662582784
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.6984126984127e-05,
      "loss": 2.3971,
      "theoretical_loss": 3.4846068657858877,
      "tokens_seen": 1662713856
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.777777777777778e-05,
      "loss": 2.5189,
      "theoretical_loss": 3.4845831234015976,
      "tokens_seen": 1662844928
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.857142857142858e-05,
      "loss": 2.3924,
      "theoretical_loss": 3.484559383412666,
      "tokens_seen": 1662976000
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 918239,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3232572078704834,
      "objective/train/theoretical_loss": 3.4845356458186627,
      "objective/train/tokens_used": 33566176,
      "theoretical_loss": 3.4845356458186627,
      "tokens_seen": 1663107072
    },
    {
      "epoch": 0.01,
      "learning_rate": 7.936507936507937e-05,
      "loss": 2.5743,
      "theoretical_loss": 3.4845356458186627,
      "tokens_seen": 1663107072
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.015873015873016e-05,
      "loss": 2.4116,
      "theoretical_loss": 3.4845119106191578,
      "tokens_seen": 1663238144
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.095238095238096e-05,
      "loss": 2.5243,
      "theoretical_loss": 3.48448817781372,
      "tokens_seen": 1663369216
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.174603174603175e-05,
      "loss": 2.4798,
      "theoretical_loss": 3.484464447401921,
      "tokens_seen": 1663500288
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.253968253968255e-05,
      "loss": 2.3456,
      "theoretical_loss": 3.4844407193833282,
      "tokens_seen": 1663631360
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.333333333333334e-05,
      "loss": 2.4678,
      "theoretical_loss": 3.484416993757514,
      "tokens_seen": 1663762432
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.412698412698413e-05,
      "loss": 2.4106,
      "theoretical_loss": 3.4843932705240483,
      "tokens_seen": 1663893504
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.492063492063493e-05,
      "loss": 2.3669,
      "theoretical_loss": 3.4843695496825005,
      "tokens_seen": 1664024576
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.571428571428571e-05,
      "loss": 2.5092,
      "theoretical_loss": 3.4843458312324413,
      "tokens_seen": 1664155648
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.650793650793651e-05,
      "loss": 2.473,
      "theoretical_loss": 3.484322115173442,
      "tokens_seen": 1664286720
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.730158730158731e-05,
      "loss": 2.5359,
      "theoretical_loss": 3.4842984015050726,
      "tokens_seen": 1664417792
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.80952380952381e-05,
      "loss": 2.6096,
      "theoretical_loss": 3.484274690226904,
      "tokens_seen": 1664548864
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.888888888888889e-05,
      "loss": 2.4428,
      "theoretical_loss": 3.484250981338507,
      "tokens_seen": 1664679936
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 919306,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.603062629699707,
      "objective/train/theoretical_loss": 3.484239127790339,
      "objective/train/tokens_used": 35204576,
      "theoretical_loss": 3.484239127790339,
      "tokens_seen": 1664745472
    },
    {
      "epoch": 0.01,
      "learning_rate": 8.968253968253969e-05,
      "loss": 2.5448,
      "theoretical_loss": 3.484227274839453,
      "tokens_seen": 1664811008
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.047619047619048e-05,
      "loss": 2.5698,
      "theoretical_loss": 3.484203570729313,
      "tokens_seen": 1664942080
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.126984126984128e-05,
      "loss": 2.4996,
      "theoretical_loss": 3.484179869007658,
      "tokens_seen": 1665073152
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.206349206349206e-05,
      "loss": 2.5152,
      "theoretical_loss": 3.4841561696740597,
      "tokens_seen": 1665204224
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.285714285714286e-05,
      "loss": 2.4996,
      "theoretical_loss": 3.4841324727280893,
      "tokens_seen": 1665335296
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.365079365079366e-05,
      "loss": 2.47,
      "theoretical_loss": 3.484108778169318,
      "tokens_seen": 1665466368
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.444444444444444e-05,
      "loss": 2.3922,
      "theoretical_loss": 3.484085085997318,
      "tokens_seen": 1665597440
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.523809523809524e-05,
      "loss": 2.4767,
      "theoretical_loss": 3.484061396211661,
      "tokens_seen": 1665728512
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.603174603174604e-05,
      "loss": 2.5377,
      "theoretical_loss": 3.4840377088119188,
      "tokens_seen": 1665859584
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.682539682539682e-05,
      "loss": 2.61,
      "theoretical_loss": 3.4840140237976636,
      "tokens_seen": 1665990656
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.761904761904762e-05,
      "loss": 2.4506,
      "theoretical_loss": 3.483990341168467,
      "tokens_seen": 1666121728
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.841269841269841e-05,
      "loss": 2.4311,
      "theoretical_loss": 3.483966660923902,
      "tokens_seen": 1666252800
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 919806,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3898022174835205,
      "objective/train/theoretical_loss": 3.4839429830635407,
      "objective/train/tokens_used": 36842976,
      "theoretical_loss": 3.4839429830635407,
      "tokens_seen": 1666383872
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.920634920634922e-05,
      "loss": 2.543,
      "theoretical_loss": 3.4839429830635407,
      "tokens_seen": 1666383872
    },
    {
      "epoch": 0.01,
      "learning_rate": 0.0001,
      "loss": 2.4395,
      "theoretical_loss": 3.4839193075869543,
      "tokens_seen": 1666514944
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.999197560584176e-05,
      "loss": 2.4916,
      "theoretical_loss": 3.4838956344937175,
      "tokens_seen": 1666646016
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.998395121168352e-05,
      "loss": 2.3888,
      "theoretical_loss": 3.4838719637834012,
      "tokens_seen": 1666777088
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.997592681752528e-05,
      "loss": 2.5063,
      "theoretical_loss": 3.483848295455579,
      "tokens_seen": 1666908160
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.996790242336704e-05,
      "loss": 2.4951,
      "theoretical_loss": 3.483824629509824,
      "tokens_seen": 1667039232
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.99598780292088e-05,
      "loss": 2.4226,
      "theoretical_loss": 3.4838009659457088,
      "tokens_seen": 1667170304
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.995185363505056e-05,
      "loss": 2.4431,
      "theoretical_loss": 3.4837773047628064,
      "tokens_seen": 1667301376
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.994382924089232e-05,
      "loss": 2.4434,
      "theoretical_loss": 3.4837536459606904,
      "tokens_seen": 1667432448
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.993580484673407e-05,
      "loss": 2.5021,
      "theoretical_loss": 3.4837299895389333,
      "tokens_seen": 1667563520
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.992778045257584e-05,
      "loss": 2.5622,
      "theoretical_loss": 3.48370633549711,
      "tokens_seen": 1667694592
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.991975605841759e-05,
      "loss": 2.524,
      "theoretical_loss": 3.4836826838347923,
      "tokens_seen": 1667825664
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.991173166425936e-05,
      "loss": 2.5055,
      "theoretical_loss": 3.483659034551555,
      "tokens_seen": 1667956736
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 921005,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6571173667907715,
      "objective/train/theoretical_loss": 3.4836472108019585,
      "objective/train/tokens_used": 38481376,
      "theoretical_loss": 3.4836472108019585,
      "tokens_seen": 1668022272
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.990370727010111e-05,
      "loss": 2.5956,
      "theoretical_loss": 3.483635387646972,
      "tokens_seen": 1668087808
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.989568287594288e-05,
      "loss": 2.5279,
      "theoretical_loss": 3.4836117431206164,
      "tokens_seen": 1668218880
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.988765848178463e-05,
      "loss": 2.45,
      "theoretical_loss": 3.483588100972063,
      "tokens_seen": 1668349952
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.987963408762638e-05,
      "loss": 2.5056,
      "theoretical_loss": 3.483564461200885,
      "tokens_seen": 1668481024
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.987160969346815e-05,
      "loss": 2.56,
      "theoretical_loss": 3.4835408238066567,
      "tokens_seen": 1668612096
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.98635852993099e-05,
      "loss": 2.4673,
      "theoretical_loss": 3.4835171887889533,
      "tokens_seen": 1668743168
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.985556090515167e-05,
      "loss": 2.3887,
      "theoretical_loss": 3.483493556147349,
      "tokens_seen": 1668874240
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.984753651099342e-05,
      "loss": 2.478,
      "theoretical_loss": 3.4834699258814172,
      "tokens_seen": 1669005312
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.983951211683519e-05,
      "loss": 2.6136,
      "theoretical_loss": 3.483446297990734,
      "tokens_seen": 1669136384
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.983148772267694e-05,
      "loss": 2.5876,
      "theoretical_loss": 3.4834226724748736,
      "tokens_seen": 1669267456
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.982346332851871e-05,
      "loss": 2.5149,
      "theoretical_loss": 3.4833990493334106,
      "tokens_seen": 1669398528
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.981543893436046e-05,
      "loss": 2.5755,
      "theoretical_loss": 3.48337542856592,
      "tokens_seen": 1669529600
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 921680,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.329754590988159,
      "objective/train/theoretical_loss": 3.4833518101719774,
      "objective/train/tokens_used": 40119776,
      "theoretical_loss": 3.4833518101719774,
      "tokens_seen": 1669660672
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.980741454020222e-05,
      "loss": 2.4671,
      "theoretical_loss": 3.4833518101719774,
      "tokens_seen": 1669660672
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.979939014604398e-05,
      "loss": 2.3942,
      "theoretical_loss": 3.4833281941511576,
      "tokens_seen": 1669791744
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.979136575188574e-05,
      "loss": 2.4857,
      "theoretical_loss": 3.4833045805030363,
      "tokens_seen": 1669922816
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.97833413577275e-05,
      "loss": 2.4502,
      "theoretical_loss": 3.483280969227188,
      "tokens_seen": 1670053888
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.977531696356926e-05,
      "loss": 2.5831,
      "theoretical_loss": 3.4832573603231887,
      "tokens_seen": 1670184960
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.976729256941102e-05,
      "loss": 2.561,
      "theoretical_loss": 3.4832337537906146,
      "tokens_seen": 1670316032
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.975926817525277e-05,
      "loss": 2.5869,
      "theoretical_loss": 3.4832101496290404,
      "tokens_seen": 1670447104
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.975124378109453e-05,
      "loss": 2.4206,
      "theoretical_loss": 3.4831865478380433,
      "tokens_seen": 1670578176
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.97432193869363e-05,
      "loss": 2.3991,
      "theoretical_loss": 3.4831629484171978,
      "tokens_seen": 1670709248
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.973519499277805e-05,
      "loss": 2.3821,
      "theoretical_loss": 3.4831393513660807,
      "tokens_seen": 1670840320
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.972717059861981e-05,
      "loss": 2.5258,
      "theoretical_loss": 3.483115756684268,
      "tokens_seen": 1670971392
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.971914620446157e-05,
      "loss": 2.4604,
      "theoretical_loss": 3.4830921643713366,
      "tokens_seen": 1671102464
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.971112181030333e-05,
      "loss": 2.5027,
      "theoretical_loss": 3.483068574426862,
      "tokens_seen": 1671233536
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 922598,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.674171209335327,
      "objective/train/theoretical_loss": 3.4830567803426637,
      "objective/train/tokens_used": 41758176,
      "theoretical_loss": 3.4830567803426637,
      "tokens_seen": 1671299072
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.970309741614509e-05,
      "loss": 2.5194,
      "theoretical_loss": 3.483044986850421,
      "tokens_seen": 1671364608
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.969507302198684e-05,
      "loss": 2.4984,
      "theoretical_loss": 3.48302140164159,
      "tokens_seen": 1671495680
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.968704862782861e-05,
      "loss": 2.6113,
      "theoretical_loss": 3.482997818799947,
      "tokens_seen": 1671626752
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.967902423367036e-05,
      "loss": 2.6013,
      "theoretical_loss": 3.4829742383250673,
      "tokens_seen": 1671757824
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.967099983951213e-05,
      "loss": 2.4007,
      "theoretical_loss": 3.4829506602165283,
      "tokens_seen": 1671888896
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.966297544535388e-05,
      "loss": 2.4473,
      "theoretical_loss": 3.482927084473907,
      "tokens_seen": 1672019968
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.965495105119565e-05,
      "loss": 2.3941,
      "theoretical_loss": 3.482903511096781,
      "tokens_seen": 1672151040
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.96469266570374e-05,
      "loss": 2.5564,
      "theoretical_loss": 3.482879940084727,
      "tokens_seen": 1672282112
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.963890226287915e-05,
      "loss": 2.4979,
      "theoretical_loss": 3.4828563714373226,
      "tokens_seen": 1672413184
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.963087786872092e-05,
      "loss": 2.4506,
      "theoretical_loss": 3.4828328051541453,
      "tokens_seen": 1672544256
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.962285347456267e-05,
      "loss": 2.5131,
      "theoretical_loss": 3.482809241234773,
      "tokens_seen": 1672675328
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.961482908040444e-05,
      "loss": 2.4079,
      "theoretical_loss": 3.482785679678783,
      "tokens_seen": 1672806400
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 923236,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.478933095932007,
      "objective/train/theoretical_loss": 3.482762120485753,
      "objective/train/tokens_used": 43396576,
      "theoretical_loss": 3.482762120485753,
      "tokens_seen": 1672937472
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.960680468624619e-05,
      "loss": 2.5005,
      "theoretical_loss": 3.482762120485753,
      "tokens_seen": 1672937472
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.959878029208796e-05,
      "loss": 2.669,
      "theoretical_loss": 3.482738563655261,
      "tokens_seen": 1673068544
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.959075589792971e-05,
      "loss": 2.5706,
      "theoretical_loss": 3.4827150091868853,
      "tokens_seen": 1673199616
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.958273150377147e-05,
      "loss": 2.5295,
      "theoretical_loss": 3.482691457080204,
      "tokens_seen": 1673330688
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.957470710961323e-05,
      "loss": 2.4358,
      "theoretical_loss": 3.482667907334795,
      "tokens_seen": 1673461760
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.956668271545499e-05,
      "loss": 2.4467,
      "theoretical_loss": 3.482644359950237,
      "tokens_seen": 1673592832
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.955865832129675e-05,
      "loss": 2.3988,
      "theoretical_loss": 3.4826208149261078,
      "tokens_seen": 1673723904
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.95506339271385e-05,
      "loss": 2.521,
      "theoretical_loss": 3.482597272261987,
      "tokens_seen": 1673854976
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.954260953298027e-05,
      "loss": 2.409,
      "theoretical_loss": 3.4825737319574523,
      "tokens_seen": 1673986048
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.953458513882203e-05,
      "loss": 2.5548,
      "theoretical_loss": 3.4825501940120835,
      "tokens_seen": 1674117120
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.952656074466378e-05,
      "loss": 2.4541,
      "theoretical_loss": 3.4825266584254586,
      "tokens_seen": 1674248192
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.951853635050554e-05,
      "loss": 2.5549,
      "theoretical_loss": 3.482503125197157,
      "tokens_seen": 1674379264
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.95105119563473e-05,
      "loss": 2.4687,
      "theoretical_loss": 3.4824795943267577,
      "tokens_seen": 1674510336
    },
    {
      "epoch": 0.01,
      "objective/train/docs_used": 924425,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.032245397567749,
      "objective/train/theoretical_loss": 3.48246782977564,
      "objective/train/tokens_used": 45034976,
      "theoretical_loss": 3.48246782977564,
      "tokens_seen": 1674575872
    },
    {
      "epoch": 0.01,
      "learning_rate": 9.950248756218906e-05,
      "loss": 2.4781,
      "theoretical_loss": 3.4824560658138397,
      "tokens_seen": 1674641408
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.949446316803082e-05,
      "loss": 2.6007,
      "theoretical_loss": 3.482432539657983,
      "tokens_seen": 1674772480
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.948643877387258e-05,
      "loss": 2.4198,
      "theoretical_loss": 3.4824090158587664,
      "tokens_seen": 1674903552
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.947841437971434e-05,
      "loss": 2.49,
      "theoretical_loss": 3.4823854944157695,
      "tokens_seen": 1675034624
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.94703899855561e-05,
      "loss": 2.4733,
      "theoretical_loss": 3.482361975328572,
      "tokens_seen": 1675165696
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.946236559139786e-05,
      "loss": 2.6164,
      "theoretical_loss": 3.482338458596754,
      "tokens_seen": 1675296768
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.945434119723961e-05,
      "loss": 2.4729,
      "theoretical_loss": 3.482314944219895,
      "tokens_seen": 1675427840
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.944631680308138e-05,
      "loss": 2.5106,
      "theoretical_loss": 3.482291432197575,
      "tokens_seen": 1675558912
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.943829240892313e-05,
      "loss": 2.4255,
      "theoretical_loss": 3.482267922529374,
      "tokens_seen": 1675689984
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.94302680147649e-05,
      "loss": 2.4172,
      "theoretical_loss": 3.482244415214873,
      "tokens_seen": 1675821056
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.942224362060665e-05,
      "loss": 2.4485,
      "theoretical_loss": 3.482220910253651,
      "tokens_seen": 1675952128
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.941421922644842e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.482197407645289,
      "tokens_seen": 1676083200
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 924948,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1431517601013184,
      "objective/train/theoretical_loss": 3.4821739073893676,
      "objective/train/tokens_used": 46673376,
      "theoretical_loss": 3.4821739073893676,
      "tokens_seen": 1676214272
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.940619483229017e-05,
      "loss": 2.4842,
      "theoretical_loss": 3.4821739073893676,
      "tokens_seen": 1676214272
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.939817043813192e-05,
      "loss": 2.5048,
      "theoretical_loss": 3.482150409485467,
      "tokens_seen": 1676345344
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.939014604397369e-05,
      "loss": 2.4242,
      "theoretical_loss": 3.482126913933169,
      "tokens_seen": 1676476416
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.938212164981544e-05,
      "loss": 2.4716,
      "theoretical_loss": 3.482103420732053,
      "tokens_seen": 1676607488
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.937409725565721e-05,
      "loss": 2.4006,
      "theoretical_loss": 3.482079929881701,
      "tokens_seen": 1676738560
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.936607286149896e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.482056441381694,
      "tokens_seen": 1676869632
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.935804846734073e-05,
      "loss": 2.4948,
      "theoretical_loss": 3.4820329552316123,
      "tokens_seen": 1677000704
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.935002407318248e-05,
      "loss": 2.4135,
      "theoretical_loss": 3.482009471431038,
      "tokens_seen": 1677131776
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.934199967902424e-05,
      "loss": 2.4302,
      "theoretical_loss": 3.4819859899795516,
      "tokens_seen": 1677262848
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.9333975284866e-05,
      "loss": 2.4168,
      "theoretical_loss": 3.481962510876736,
      "tokens_seen": 1677393920
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.932595089070776e-05,
      "loss": 2.648,
      "theoretical_loss": 3.481939034122171,
      "tokens_seen": 1677524992
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.931792649654952e-05,
      "loss": 2.4332,
      "theoretical_loss": 3.4819155597154396,
      "tokens_seen": 1677656064
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.930990210239128e-05,
      "loss": 2.4625,
      "theoretical_loss": 3.4818920876561235,
      "tokens_seen": 1677787136
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 926165,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.494450569152832,
      "objective/train/theoretical_loss": 3.4818803525066153,
      "objective/train/tokens_used": 48311776,
      "theoretical_loss": 3.4818803525066153,
      "tokens_seen": 1677852672
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.930187770823304e-05,
      "loss": 2.4562,
      "theoretical_loss": 3.481868617943804,
      "tokens_seen": 1677918208
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.92938533140748e-05,
      "loss": 2.4934,
      "theoretical_loss": 3.481845150578063,
      "tokens_seen": 1678049280
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.928582891991655e-05,
      "loss": 2.4625,
      "theoretical_loss": 3.481821685558484,
      "tokens_seen": 1678180352
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.927780452575832e-05,
      "loss": 2.3664,
      "theoretical_loss": 3.4817982228846476,
      "tokens_seen": 1678311424
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.926978013160007e-05,
      "loss": 2.4554,
      "theoretical_loss": 3.481774762556137,
      "tokens_seen": 1678442496
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.926175573744183e-05,
      "loss": 2.6398,
      "theoretical_loss": 3.4817513045725343,
      "tokens_seen": 1678573568
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.925373134328359e-05,
      "loss": 2.3632,
      "theoretical_loss": 3.4817278489334225,
      "tokens_seen": 1678704640
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.924570694912535e-05,
      "loss": 2.6002,
      "theoretical_loss": 3.481704395638383,
      "tokens_seen": 1678835712
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.923768255496711e-05,
      "loss": 2.5935,
      "theoretical_loss": 3.4816809446870005,
      "tokens_seen": 1678966784
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.922965816080886e-05,
      "loss": 2.6889,
      "theoretical_loss": 3.4816574960788564,
      "tokens_seen": 1679097856
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.922163376665063e-05,
      "loss": 2.2718,
      "theoretical_loss": 3.4816340498135343,
      "tokens_seen": 1679228928
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.921360937249238e-05,
      "loss": 2.4179,
      "theoretical_loss": 3.4816106058906175,
      "tokens_seen": 1679360000
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 926923,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7133426666259766,
      "objective/train/theoretical_loss": 3.481587164309688,
      "objective/train/tokens_used": 49950176,
      "theoretical_loss": 3.481587164309688,
      "tokens_seen": 1679491072
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.920558497833415e-05,
      "loss": 2.4835,
      "theoretical_loss": 3.481587164309688,
      "tokens_seen": 1679491072
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.91975605841759e-05,
      "loss": 2.5025,
      "theoretical_loss": 3.4815637250703304,
      "tokens_seen": 1679622144
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.918953619001767e-05,
      "loss": 2.4519,
      "theoretical_loss": 3.4815402881721274,
      "tokens_seen": 1679753216
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.918151179585942e-05,
      "loss": 2.5021,
      "theoretical_loss": 3.4815168536146626,
      "tokens_seen": 1679884288
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.917348740170119e-05,
      "loss": 2.4785,
      "theoretical_loss": 3.48149342139752,
      "tokens_seen": 1680015360
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.916546300754294e-05,
      "loss": 2.5138,
      "theoretical_loss": 3.481469991520283,
      "tokens_seen": 1680146432
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.91574386133847e-05,
      "loss": 2.4019,
      "theoretical_loss": 3.4814465639825354,
      "tokens_seen": 1680277504
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.914941421922646e-05,
      "loss": 2.5178,
      "theoretical_loss": 3.481423138783861,
      "tokens_seen": 1680408576
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.914138982506821e-05,
      "loss": 2.4615,
      "theoretical_loss": 3.4813997159238443,
      "tokens_seen": 1680539648
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.913336543090998e-05,
      "loss": 2.533,
      "theoretical_loss": 3.481376295402069,
      "tokens_seen": 1680670720
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.912534103675173e-05,
      "loss": 2.5408,
      "theoretical_loss": 3.4813528772181193,
      "tokens_seen": 1680801792
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.91173166425935e-05,
      "loss": 2.4864,
      "theoretical_loss": 3.4813294613715797,
      "tokens_seen": 1680932864
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.910929224843525e-05,
      "loss": 2.567,
      "theoretical_loss": 3.4813060478620352,
      "tokens_seen": 1681063936
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 927976,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.612856864929199,
      "objective/train/theoretical_loss": 3.4812943419835065,
      "objective/train/tokens_used": 51588576,
      "theoretical_loss": 3.4812943419835065,
      "tokens_seen": 1681129472
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.9101267854277e-05,
      "loss": 2.515,
      "theoretical_loss": 3.48128263668907,
      "tokens_seen": 1681195008
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.909324346011877e-05,
      "loss": 2.5095,
      "theoretical_loss": 3.4812592278522687,
      "tokens_seen": 1681326080
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.908521906596053e-05,
      "loss": 2.6043,
      "theoretical_loss": 3.4812358213512162,
      "tokens_seen": 1681457152
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.907719467180229e-05,
      "loss": 2.6276,
      "theoretical_loss": 3.4812124171854966,
      "tokens_seen": 1681588224
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.906917027764405e-05,
      "loss": 2.3657,
      "theoretical_loss": 3.481189015354696,
      "tokens_seen": 1681719296
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.906114588348581e-05,
      "loss": 2.4345,
      "theoretical_loss": 3.481165615858399,
      "tokens_seen": 1681850368
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.905312148932757e-05,
      "loss": 2.6024,
      "theoretical_loss": 3.481142218696191,
      "tokens_seen": 1681981440
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.904509709516932e-05,
      "loss": 2.5844,
      "theoretical_loss": 3.481118823867657,
      "tokens_seen": 1682112512
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.903707270101109e-05,
      "loss": 2.4985,
      "theoretical_loss": 3.481095431372383,
      "tokens_seen": 1682243584
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.902904830685284e-05,
      "loss": 2.4409,
      "theoretical_loss": 3.481072041209954,
      "tokens_seen": 1682374656
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.90210239126946e-05,
      "loss": 2.4786,
      "theoretical_loss": 3.481048653379955,
      "tokens_seen": 1682505728
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.901299951853636e-05,
      "loss": 2.4033,
      "theoretical_loss": 3.4810252678819733,
      "tokens_seen": 1682636800
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 929273,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6835052967071533,
      "objective/train/theoretical_loss": 3.481001884715594,
      "objective/train/tokens_used": 53226976,
      "theoretical_loss": 3.481001884715594,
      "tokens_seen": 1682767872
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.900497512437812e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.481001884715594,
      "tokens_seen": 1682767872
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.899695073021988e-05,
      "loss": 2.4493,
      "theoretical_loss": 3.4809785038804026,
      "tokens_seen": 1682898944
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.898892633606163e-05,
      "loss": 2.5062,
      "theoretical_loss": 3.480955125375986,
      "tokens_seen": 1683030016
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.89809019419034e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.48093174920193,
      "tokens_seen": 1683161088
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.897287754774515e-05,
      "loss": 2.5248,
      "theoretical_loss": 3.48090837535782,
      "tokens_seen": 1683292160
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.896485315358692e-05,
      "loss": 2.5576,
      "theoretical_loss": 3.4808850038432437,
      "tokens_seen": 1683423232
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.895682875942867e-05,
      "loss": 2.398,
      "theoretical_loss": 3.4808616346577868,
      "tokens_seen": 1683554304
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.894880436527044e-05,
      "loss": 2.5147,
      "theoretical_loss": 3.4808382678010363,
      "tokens_seen": 1683685376
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.894077997111219e-05,
      "loss": 2.5106,
      "theoretical_loss": 3.480814903272579,
      "tokens_seen": 1683816448
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.893275557695394e-05,
      "loss": 2.4724,
      "theoretical_loss": 3.4807915410720005,
      "tokens_seen": 1683947520
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.892473118279571e-05,
      "loss": 2.5067,
      "theoretical_loss": 3.4807681811988895,
      "tokens_seen": 1684078592
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.891670678863746e-05,
      "loss": 2.5444,
      "theoretical_loss": 3.4807448236528318,
      "tokens_seen": 1684209664
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.890868239447923e-05,
      "loss": 2.5985,
      "theoretical_loss": 3.4807214684334147,
      "tokens_seen": 1684340736
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 929850,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.568603992462158,
      "objective/train/theoretical_loss": 3.480709791696068,
      "objective/train/tokens_used": 54865376,
      "theoretical_loss": 3.480709791696068,
      "tokens_seen": 1684406272
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.890065800032098e-05,
      "loss": 2.4515,
      "theoretical_loss": 3.4806981155402257,
      "tokens_seen": 1684471808
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.889263360616275e-05,
      "loss": 2.4635,
      "theoretical_loss": 3.4806747649728518,
      "tokens_seen": 1684602880
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.88846092120045e-05,
      "loss": 2.5249,
      "theoretical_loss": 3.480651416730881,
      "tokens_seen": 1684733952
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.887658481784626e-05,
      "loss": 2.5362,
      "theoretical_loss": 3.4806280708139,
      "tokens_seen": 1684865024
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.886856042368802e-05,
      "loss": 2.4881,
      "theoretical_loss": 3.480604727221497,
      "tokens_seen": 1684996096
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.886053602952978e-05,
      "loss": 2.555,
      "theoretical_loss": 3.4805813859532595,
      "tokens_seen": 1685127168
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.885251163537154e-05,
      "loss": 2.6367,
      "theoretical_loss": 3.480558047008776,
      "tokens_seen": 1685258240
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.88444872412133e-05,
      "loss": 2.4256,
      "theoretical_loss": 3.4805347103876327,
      "tokens_seen": 1685389312
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.883646284705506e-05,
      "loss": 2.2893,
      "theoretical_loss": 3.48051137608942,
      "tokens_seen": 1685520384
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.882843845289682e-05,
      "loss": 2.5847,
      "theoretical_loss": 3.4804880441137245,
      "tokens_seen": 1685651456
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.882041405873857e-05,
      "loss": 2.6211,
      "theoretical_loss": 3.4804647144601346,
      "tokens_seen": 1685782528
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.881238966458034e-05,
      "loss": 2.4622,
      "theoretical_loss": 3.4804413871282396,
      "tokens_seen": 1685913600
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 931063,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.578721523284912,
      "objective/train/theoretical_loss": 3.480418062117627,
      "objective/train/tokens_used": 56503776,
      "theoretical_loss": 3.480418062117627,
      "tokens_seen": 1686044672
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.880436527042209e-05,
      "loss": 2.5689,
      "theoretical_loss": 3.480418062117627,
      "tokens_seen": 1686044672
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.879634087626386e-05,
      "loss": 2.4969,
      "theoretical_loss": 3.4803947394278856,
      "tokens_seen": 1686175744
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.878831648210561e-05,
      "loss": 2.5164,
      "theoretical_loss": 3.4803714190586037,
      "tokens_seen": 1686306816
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.878029208794736e-05,
      "loss": 2.4945,
      "theoretical_loss": 3.4803481010093718,
      "tokens_seen": 1686437888
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.877226769378913e-05,
      "loss": 2.399,
      "theoretical_loss": 3.4803247852797767,
      "tokens_seen": 1686568960
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.876424329963088e-05,
      "loss": 2.4799,
      "theoretical_loss": 3.4803014718694087,
      "tokens_seen": 1686700032
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.875621890547265e-05,
      "loss": 2.35,
      "theoretical_loss": 3.480278160777856,
      "tokens_seen": 1686831104
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.87481945113144e-05,
      "loss": 2.4886,
      "theoretical_loss": 3.480254852004709,
      "tokens_seen": 1686962176
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.874017011715617e-05,
      "loss": 2.5754,
      "theoretical_loss": 3.4802315455495565,
      "tokens_seen": 1687093248
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.873214572299792e-05,
      "loss": 2.5741,
      "theoretical_loss": 3.480208241411987,
      "tokens_seen": 1687224320
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.872412132883967e-05,
      "loss": 2.5755,
      "theoretical_loss": 3.480184939591591,
      "tokens_seen": 1687355392
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.871609693468144e-05,
      "loss": 2.4047,
      "theoretical_loss": 3.4801616400879585,
      "tokens_seen": 1687486464
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.87080725405232e-05,
      "loss": 2.4871,
      "theoretical_loss": 3.480138342900678,
      "tokens_seen": 1687617536
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 931367,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5317652225494385,
      "objective/train/theoretical_loss": 3.480126695175542,
      "objective/train/tokens_used": 58142176,
      "theoretical_loss": 3.480126695175542,
      "tokens_seen": 1687683072
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.870004814636496e-05,
      "loss": 2.6203,
      "theoretical_loss": 3.48011504802934,
      "tokens_seen": 1687748608
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.869202375220671e-05,
      "loss": 2.4963,
      "theoretical_loss": 3.4800917554735347,
      "tokens_seen": 1687879680
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.868399935804847e-05,
      "loss": 2.5753,
      "theoretical_loss": 3.480068465232852,
      "tokens_seen": 1688010752
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.867597496389023e-05,
      "loss": 2.546,
      "theoretical_loss": 3.4800451773068817,
      "tokens_seen": 1688141824
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.866795056973199e-05,
      "loss": 2.3913,
      "theoretical_loss": 3.4800218916952144,
      "tokens_seen": 1688272896
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.865992617557375e-05,
      "loss": 2.5884,
      "theoretical_loss": 3.47999860839744,
      "tokens_seen": 1688403968
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.86519017814155e-05,
      "loss": 2.5369,
      "theoretical_loss": 3.4799753274131495,
      "tokens_seen": 1688535040
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.864387738725727e-05,
      "loss": 2.6298,
      "theoretical_loss": 3.4799520487419335,
      "tokens_seen": 1688666112
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.863585299309903e-05,
      "loss": 2.567,
      "theoretical_loss": 3.4799287723833823,
      "tokens_seen": 1688797184
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.862782859894078e-05,
      "loss": 2.4449,
      "theoretical_loss": 3.4799054983370867,
      "tokens_seen": 1688928256
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.861980420478255e-05,
      "loss": 2.5075,
      "theoretical_loss": 3.4798822266026384,
      "tokens_seen": 1689059328
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.86117798106243e-05,
      "loss": 2.5648,
      "theoretical_loss": 3.4798589571796272,
      "tokens_seen": 1689190400
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 932516,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2863895893096924,
      "objective/train/theoretical_loss": 3.4798356900676444,
      "objective/train/tokens_used": 59780576,
      "theoretical_loss": 3.4798356900676444,
      "tokens_seen": 1689321472
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.860375541646607e-05,
      "loss": 2.4437,
      "theoretical_loss": 3.4798356900676444,
      "tokens_seen": 1689321472
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.859573102230782e-05,
      "loss": 2.4915,
      "theoretical_loss": 3.479812425266282,
      "tokens_seen": 1689452544
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.858770662814957e-05,
      "loss": 2.5486,
      "theoretical_loss": 3.479789162775131,
      "tokens_seen": 1689583616
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.857968223399134e-05,
      "loss": 2.4756,
      "theoretical_loss": 3.479765902593782,
      "tokens_seen": 1689714688
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.857165783983309e-05,
      "loss": 2.4487,
      "theoretical_loss": 3.4797426447218274,
      "tokens_seen": 1689845760
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.856363344567486e-05,
      "loss": 2.5253,
      "theoretical_loss": 3.4797193891588583,
      "tokens_seen": 1689976832
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.855560905151661e-05,
      "loss": 2.4716,
      "theoretical_loss": 3.479696135904467,
      "tokens_seen": 1690107904
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.854758465735838e-05,
      "loss": 2.5138,
      "theoretical_loss": 3.479672884958245,
      "tokens_seen": 1690238976
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.853956026320013e-05,
      "loss": 2.4722,
      "theoretical_loss": 3.4796496363197837,
      "tokens_seen": 1690370048
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.853153586904188e-05,
      "loss": 2.4039,
      "theoretical_loss": 3.4796263899886757,
      "tokens_seen": 1690501120
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.852351147488365e-05,
      "loss": 2.2929,
      "theoretical_loss": 3.4796031459645134,
      "tokens_seen": 1690632192
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.85154870807254e-05,
      "loss": 2.5445,
      "theoretical_loss": 3.4795799042468882,
      "tokens_seen": 1690763264
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.850746268656717e-05,
      "loss": 2.4633,
      "theoretical_loss": 3.4795566648353935,
      "tokens_seen": 1690894336
    },
    {
      "epoch": 0.02,
      "objective/train/docs_used": 933032,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.596432685852051,
      "objective/train/theoretical_loss": 3.4795450459943167,
      "objective/train/tokens_used": 61418976,
      "theoretical_loss": 3.4795450459943167,
      "tokens_seen": 1690959872
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.849943829240892e-05,
      "loss": 2.5766,
      "theoretical_loss": 3.4795334277296206,
      "tokens_seen": 1691025408
    },
    {
      "epoch": 0.02,
      "learning_rate": 9.849141389825068e-05,
      "loss": 2.4845,
      "theoretical_loss": 3.4795101929291627,
      "tokens_seen": 1691156480
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.848338950409244e-05,
      "loss": 2.4419,
      "theoretical_loss": 3.4794869604336123,
      "tokens_seen": 1691287552
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.84753651099342e-05,
      "loss": 2.6624,
      "theoretical_loss": 3.4794637302425624,
      "tokens_seen": 1691418624
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.846734071577596e-05,
      "loss": 2.4082,
      "theoretical_loss": 3.4794405023556054,
      "tokens_seen": 1691549696
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.845931632161772e-05,
      "loss": 2.5021,
      "theoretical_loss": 3.4794172767723346,
      "tokens_seen": 1691680768
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.845129192745948e-05,
      "loss": 2.6212,
      "theoretical_loss": 3.479394053492343,
      "tokens_seen": 1691811840
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.844326753330124e-05,
      "loss": 2.4537,
      "theoretical_loss": 3.4793708325152237,
      "tokens_seen": 1691942912
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.843524313914299e-05,
      "loss": 2.5129,
      "theoretical_loss": 3.47934761384057,
      "tokens_seen": 1692073984
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.842721874498476e-05,
      "loss": 2.5562,
      "theoretical_loss": 3.4793243974679755,
      "tokens_seen": 1692205056
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.841919435082651e-05,
      "loss": 2.409,
      "theoretical_loss": 3.479301183397033,
      "tokens_seen": 1692336128
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.841116995666828e-05,
      "loss": 2.5037,
      "theoretical_loss": 3.4792779716273365,
      "tokens_seen": 1692467200
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 934026,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.466470241546631,
      "objective/train/theoretical_loss": 3.4792547621584804,
      "objective/train/tokens_used": 63057376,
      "theoretical_loss": 3.4792547621584804,
      "tokens_seen": 1692598272
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.840314556251003e-05,
      "loss": 2.4919,
      "theoretical_loss": 3.4792547621584804,
      "tokens_seen": 1692598272
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.839512116835178e-05,
      "loss": 2.5155,
      "theoretical_loss": 3.479231554990057,
      "tokens_seen": 1692729344
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.838709677419355e-05,
      "loss": 2.4349,
      "theoretical_loss": 3.4792083501216613,
      "tokens_seen": 1692860416
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.83790723800353e-05,
      "loss": 2.2746,
      "theoretical_loss": 3.4791851475528874,
      "tokens_seen": 1692991488
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.837104798587707e-05,
      "loss": 2.6368,
      "theoretical_loss": 3.4791619472833286,
      "tokens_seen": 1693122560
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.836302359171882e-05,
      "loss": 2.4889,
      "theoretical_loss": 3.479138749312579,
      "tokens_seen": 1693253632
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.835499919756059e-05,
      "loss": 2.5798,
      "theoretical_loss": 3.4791155536402343,
      "tokens_seen": 1693384704
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.834697480340234e-05,
      "loss": 2.6683,
      "theoretical_loss": 3.479092360265887,
      "tokens_seen": 1693515776
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.83389504092441e-05,
      "loss": 2.4851,
      "theoretical_loss": 3.479069169189133,
      "tokens_seen": 1693646848
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.833092601508586e-05,
      "loss": 2.3859,
      "theoretical_loss": 3.4790459804095666,
      "tokens_seen": 1693777920
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.832290162092762e-05,
      "loss": 2.5253,
      "theoretical_loss": 3.4790227939267826,
      "tokens_seen": 1693908992
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.831487722676938e-05,
      "loss": 2.2977,
      "theoretical_loss": 3.478999609740375,
      "tokens_seen": 1694040064
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.830685283261113e-05,
      "loss": 2.4333,
      "theoretical_loss": 3.4789764278499398,
      "tokens_seen": 1694171136
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 934482,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2957074642181396,
      "objective/train/theoretical_loss": 3.478964837765585,
      "objective/train/tokens_used": 64695776,
      "theoretical_loss": 3.478964837765585,
      "tokens_seen": 1694236672
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.82988284384529e-05,
      "loss": 2.5303,
      "theoretical_loss": 3.4789532482550714,
      "tokens_seen": 1694302208
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.829080404429465e-05,
      "loss": 2.4688,
      "theoretical_loss": 3.478930070955365,
      "tokens_seen": 1694433280
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.828277965013641e-05,
      "loss": 2.443,
      "theoretical_loss": 3.478906895950416,
      "tokens_seen": 1694564352
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.827475525597817e-05,
      "loss": 2.5702,
      "theoretical_loss": 3.4788837232398198,
      "tokens_seen": 1694695424
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.826673086181993e-05,
      "loss": 2.3373,
      "theoretical_loss": 3.4788605528231713,
      "tokens_seen": 1694826496
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.82587064676617e-05,
      "loss": 2.4548,
      "theoretical_loss": 3.4788373847000664,
      "tokens_seen": 1694957568
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.825068207350345e-05,
      "loss": 2.6429,
      "theoretical_loss": 3.478814218870101,
      "tokens_seen": 1695088640
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.824265767934521e-05,
      "loss": 2.3599,
      "theoretical_loss": 3.4787910553328705,
      "tokens_seen": 1695219712
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.823463328518697e-05,
      "loss": 2.3772,
      "theoretical_loss": 3.4787678940879707,
      "tokens_seen": 1695350784
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.822660889102873e-05,
      "loss": 2.5326,
      "theoretical_loss": 3.478744735134998,
      "tokens_seen": 1695481856
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.821858449687049e-05,
      "loss": 2.6442,
      "theoretical_loss": 3.4787215784735475,
      "tokens_seen": 1695612928
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.821056010271224e-05,
      "loss": 2.5572,
      "theoretical_loss": 3.478698424103216,
      "tokens_seen": 1695744000
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 935631,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5070106983184814,
      "objective/train/theoretical_loss": 3.4786752720236005,
      "objective/train/tokens_used": 66334176,
      "theoretical_loss": 3.4786752720236005,
      "tokens_seen": 1695875072
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.820253570855401e-05,
      "loss": 2.6289,
      "theoretical_loss": 3.4786752720236005,
      "tokens_seen": 1695875072
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.819451131439576e-05,
      "loss": 2.659,
      "theoretical_loss": 3.478652122234296,
      "tokens_seen": 1696006144
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.818648692023753e-05,
      "loss": 2.575,
      "theoretical_loss": 3.4786289747348995,
      "tokens_seen": 1696137216
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.817846252607928e-05,
      "loss": 2.637,
      "theoretical_loss": 3.4786058295250077,
      "tokens_seen": 1696268288
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.817043813192105e-05,
      "loss": 2.5304,
      "theoretical_loss": 3.4785826866042173,
      "tokens_seen": 1696399360
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.81624137377628e-05,
      "loss": 2.4708,
      "theoretical_loss": 3.4785595459721246,
      "tokens_seen": 1696530432
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.815438934360455e-05,
      "loss": 2.5821,
      "theoretical_loss": 3.4785364076283267,
      "tokens_seen": 1696661504
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.814636494944632e-05,
      "loss": 2.6878,
      "theoretical_loss": 3.478513271572421,
      "tokens_seen": 1696792576
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.813834055528807e-05,
      "loss": 2.5641,
      "theoretical_loss": 3.4784901378040036,
      "tokens_seen": 1696923648
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.813031616112984e-05,
      "loss": 2.5774,
      "theoretical_loss": 3.4784670063226732,
      "tokens_seen": 1697054720
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.812229176697159e-05,
      "loss": 2.5069,
      "theoretical_loss": 3.4784438771280257,
      "tokens_seen": 1697185792
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.811426737281336e-05,
      "loss": 2.4408,
      "theoretical_loss": 3.4784207502196587,
      "tokens_seen": 1697316864
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.810624297865511e-05,
      "loss": 2.6084,
      "theoretical_loss": 3.47839762559717,
      "tokens_seen": 1697447936
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 936086,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.701629877090454,
      "objective/train/theoretical_loss": 3.478386064143004,
      "objective/train/tokens_used": 67972576,
      "theoretical_loss": 3.478386064143004,
      "tokens_seen": 1697513472
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.809821858449687e-05,
      "loss": 2.5616,
      "theoretical_loss": 3.4783745032601567,
      "tokens_seen": 1697579008
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.809019419033863e-05,
      "loss": 2.5896,
      "theoretical_loss": 3.4783513832082176,
      "tokens_seen": 1697710080
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.808216979618039e-05,
      "loss": 2.4651,
      "theoretical_loss": 3.4783282654409495,
      "tokens_seen": 1697841152
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.807414540202215e-05,
      "loss": 2.5471,
      "theoretical_loss": 3.4783051499579507,
      "tokens_seen": 1697972224
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.80661210078639e-05,
      "loss": 2.5858,
      "theoretical_loss": 3.4782820367588183,
      "tokens_seen": 1698103296
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.805809661370567e-05,
      "loss": 2.4588,
      "theoretical_loss": 3.4782589258431518,
      "tokens_seen": 1698234368
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.805007221954742e-05,
      "loss": 2.4407,
      "theoretical_loss": 3.4782358172105483,
      "tokens_seen": 1698365440
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.804204782538918e-05,
      "loss": 2.6123,
      "theoretical_loss": 3.478212710860607,
      "tokens_seen": 1698496512
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.803402343123094e-05,
      "loss": 2.3523,
      "theoretical_loss": 3.4781896067929257,
      "tokens_seen": 1698627584
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.80259990370727e-05,
      "loss": 2.5561,
      "theoretical_loss": 3.4781665050071027,
      "tokens_seen": 1698758656
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.801797464291446e-05,
      "loss": 2.5973,
      "theoretical_loss": 3.4781434055027365,
      "tokens_seen": 1698889728
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.800995024875622e-05,
      "loss": 2.3148,
      "theoretical_loss": 3.4781203082794265,
      "tokens_seen": 1699020800
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 937360,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2998149394989014,
      "objective/train/theoretical_loss": 3.4780972133367714,
      "objective/train/tokens_used": 69610976,
      "theoretical_loss": 3.4780972133367714,
      "tokens_seen": 1699151872
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.800192585459798e-05,
      "loss": 2.5517,
      "theoretical_loss": 3.4780972133367714,
      "tokens_seen": 1699151872
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.799390146043974e-05,
      "loss": 2.5891,
      "theoretical_loss": 3.47807412067437,
      "tokens_seen": 1699282944
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.798587706628149e-05,
      "loss": 2.5741,
      "theoretical_loss": 3.478051030291821,
      "tokens_seen": 1699414016
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.797785267212326e-05,
      "loss": 2.6164,
      "theoretical_loss": 3.4780279421887235,
      "tokens_seen": 1699545088
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.796982827796501e-05,
      "loss": 2.6742,
      "theoretical_loss": 3.478004856364677,
      "tokens_seen": 1699676160
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.796180388380678e-05,
      "loss": 2.5587,
      "theoretical_loss": 3.4779817728192803,
      "tokens_seen": 1699807232
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.795377948964853e-05,
      "loss": 2.41,
      "theoretical_loss": 3.477958691552134,
      "tokens_seen": 1699938304
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.79457550954903e-05,
      "loss": 2.4044,
      "theoretical_loss": 3.4779356125628365,
      "tokens_seen": 1700069376
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.793773070133205e-05,
      "loss": 2.6259,
      "theoretical_loss": 3.4779125358509875,
      "tokens_seen": 1700200448
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.792970630717382e-05,
      "loss": 2.54,
      "theoretical_loss": 3.4778894614161873,
      "tokens_seen": 1700331520
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.792168191301557e-05,
      "loss": 2.4259,
      "theoretical_loss": 3.4778663892580353,
      "tokens_seen": 1700462592
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.791365751885732e-05,
      "loss": 2.7077,
      "theoretical_loss": 3.477843319376131,
      "tokens_seen": 1700593664
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.790563312469909e-05,
      "loss": 2.6439,
      "theoretical_loss": 3.4778202517700754,
      "tokens_seen": 1700724736
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 938428,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8109610080718994,
      "objective/train/theoretical_loss": 3.4778087188203655,
      "objective/train/tokens_used": 71249376,
      "theoretical_loss": 3.4778087188203655,
      "tokens_seen": 1700790272
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.789760873054084e-05,
      "loss": 2.4903,
      "theoretical_loss": 3.477797186439468,
      "tokens_seen": 1700855808
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.788958433638261e-05,
      "loss": 2.6681,
      "theoretical_loss": 3.477774123383909,
      "tokens_seen": 1700986880
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.788155994222436e-05,
      "loss": 2.4544,
      "theoretical_loss": 3.477751062602999,
      "tokens_seen": 1701117952
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.787353554806613e-05,
      "loss": 2.627,
      "theoretical_loss": 3.4777280040963383,
      "tokens_seen": 1701249024
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.786551115390788e-05,
      "loss": 2.5838,
      "theoretical_loss": 3.477704947863527,
      "tokens_seen": 1701380096
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.785748675974964e-05,
      "loss": 2.5787,
      "theoretical_loss": 3.4776818939041663,
      "tokens_seen": 1701511168
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.78494623655914e-05,
      "loss": 2.6475,
      "theoretical_loss": 3.477658842217857,
      "tokens_seen": 1701642240
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.784143797143316e-05,
      "loss": 2.529,
      "theoretical_loss": 3.4776357928041994,
      "tokens_seen": 1701773312
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.783341357727492e-05,
      "loss": 2.4775,
      "theoretical_loss": 3.4776127456627948,
      "tokens_seen": 1701904384
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.782538918311668e-05,
      "loss": 2.6175,
      "theoretical_loss": 3.477589700793244,
      "tokens_seen": 1702035456
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.781736478895844e-05,
      "loss": 2.5555,
      "theoretical_loss": 3.477566658195148,
      "tokens_seen": 1702166528
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.78093403948002e-05,
      "loss": 2.5131,
      "theoretical_loss": 3.4775436178681085,
      "tokens_seen": 1702297600
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 939017,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7329940795898438,
      "objective/train/theoretical_loss": 3.477520579811727,
      "objective/train/tokens_used": 72887776,
      "theoretical_loss": 3.477520579811727,
      "tokens_seen": 1702428672
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.780131600064195e-05,
      "loss": 2.5521,
      "theoretical_loss": 3.477520579811727,
      "tokens_seen": 1702428672
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.779329160648371e-05,
      "loss": 2.6362,
      "theoretical_loss": 3.4774975440256037,
      "tokens_seen": 1702559744
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.778526721232547e-05,
      "loss": 2.6066,
      "theoretical_loss": 3.4774745105093414,
      "tokens_seen": 1702690816
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.777724281816723e-05,
      "loss": 2.5312,
      "theoretical_loss": 3.477451479262541,
      "tokens_seen": 1702821888
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.776921842400899e-05,
      "loss": 2.5702,
      "theoretical_loss": 3.477428450284805,
      "tokens_seen": 1702952960
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.776119402985075e-05,
      "loss": 2.6024,
      "theoretical_loss": 3.4774054235757346,
      "tokens_seen": 1703084032
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.775316963569251e-05,
      "loss": 2.7145,
      "theoretical_loss": 3.4773823991349317,
      "tokens_seen": 1703215104
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.774514524153426e-05,
      "loss": 2.4479,
      "theoretical_loss": 3.477359376961999,
      "tokens_seen": 1703346176
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.773712084737603e-05,
      "loss": 2.5557,
      "theoretical_loss": 3.4773363570565374,
      "tokens_seen": 1703477248
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.772909645321778e-05,
      "loss": 2.514,
      "theoretical_loss": 3.47731333941815,
      "tokens_seen": 1703608320
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.772107205905955e-05,
      "loss": 2.4775,
      "theoretical_loss": 3.4772903240464395,
      "tokens_seen": 1703739392
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.77130476649013e-05,
      "loss": 2.5532,
      "theoretical_loss": 3.4772673109410075,
      "tokens_seen": 1703870464
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.770502327074307e-05,
      "loss": 2.5049,
      "theoretical_loss": 3.4772443001014564,
      "tokens_seen": 1704001536
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 939656,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5736076831817627,
      "objective/train/theoretical_loss": 3.4772327955312625,
      "objective/train/tokens_used": 74526176,
      "theoretical_loss": 3.4772327955312625,
      "tokens_seen": 1704067072
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.769699887658482e-05,
      "loss": 2.5509,
      "theoretical_loss": 3.47722129152739,
      "tokens_seen": 1704132608
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.768897448242657e-05,
      "loss": 2.5471,
      "theoretical_loss": 3.4771982852184102,
      "tokens_seen": 1704263680
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.768095008826834e-05,
      "loss": 2.4507,
      "theoretical_loss": 3.4771752811741194,
      "tokens_seen": 1704394752
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.767292569411009e-05,
      "loss": 2.5674,
      "theoretical_loss": 3.477152279394122,
      "tokens_seen": 1704525824
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.766490129995186e-05,
      "loss": 2.618,
      "theoretical_loss": 3.4771292798780196,
      "tokens_seen": 1704656896
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.765687690579361e-05,
      "loss": 2.5407,
      "theoretical_loss": 3.4771062826254155,
      "tokens_seen": 1704787968
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.764885251163538e-05,
      "loss": 2.6236,
      "theoretical_loss": 3.477083287635914,
      "tokens_seen": 1704919040
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.764082811747713e-05,
      "loss": 2.7451,
      "theoretical_loss": 3.4770602949091174,
      "tokens_seen": 1705050112
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.763280372331889e-05,
      "loss": 2.5501,
      "theoretical_loss": 3.4770373044446297,
      "tokens_seen": 1705181184
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.762477932916065e-05,
      "loss": 2.6444,
      "theoretical_loss": 3.4770143162420544,
      "tokens_seen": 1705312256
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.76167549350024e-05,
      "loss": 2.6054,
      "theoretical_loss": 3.4769913303009945,
      "tokens_seen": 1705443328
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.760873054084417e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.4769683466210544,
      "tokens_seen": 1705574400
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 940998,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.642338752746582,
      "objective/train/theoretical_loss": 3.4769453652018374,
      "objective/train/tokens_used": 76164576,
      "theoretical_loss": 3.4769453652018374,
      "tokens_seen": 1705705472
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.760070614668593e-05,
      "loss": 2.5956,
      "theoretical_loss": 3.4769453652018374,
      "tokens_seen": 1705705472
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.759268175252769e-05,
      "loss": 2.5025,
      "theoretical_loss": 3.4769223860429483,
      "tokens_seen": 1705836544
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.758465735836945e-05,
      "loss": 2.6917,
      "theoretical_loss": 3.47689940914399,
      "tokens_seen": 1705967616
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.757663296421121e-05,
      "loss": 2.5622,
      "theoretical_loss": 3.4768764345045677,
      "tokens_seen": 1706098688
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.756860857005296e-05,
      "loss": 2.5032,
      "theoretical_loss": 3.4768534621242853,
      "tokens_seen": 1706229760
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.756058417589472e-05,
      "loss": 2.6986,
      "theoretical_loss": 3.4768304920027466,
      "tokens_seen": 1706360832
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.755255978173648e-05,
      "loss": 2.5713,
      "theoretical_loss": 3.4768075241395566,
      "tokens_seen": 1706491904
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.754453538757824e-05,
      "loss": 2.6232,
      "theoretical_loss": 3.4767845585343196,
      "tokens_seen": 1706622976
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.753651099342e-05,
      "loss": 2.5417,
      "theoretical_loss": 3.47676159518664,
      "tokens_seen": 1706754048
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.752848659926176e-05,
      "loss": 2.5311,
      "theoretical_loss": 3.4767386340961233,
      "tokens_seen": 1706885120
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.752046220510352e-05,
      "loss": 2.5849,
      "theoretical_loss": 3.4767156752623736,
      "tokens_seen": 1707016192
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.751243781094528e-05,
      "loss": 2.6568,
      "theoretical_loss": 3.476692718684996,
      "tokens_seen": 1707147264
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.750441341678703e-05,
      "loss": 2.4312,
      "theoretical_loss": 3.4766697643635958,
      "tokens_seen": 1707278336
    },
    {
      "epoch": 0.03,
      "objective/train/docs_used": 942016,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.860006809234619,
      "objective/train/theoretical_loss": 3.4766582880487635,
      "objective/train/tokens_used": 77802976,
      "theoretical_loss": 3.4766582880487635,
      "tokens_seen": 1707343872
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.74963890226288e-05,
      "loss": 2.6738,
      "theoretical_loss": 3.4766468122977776,
      "tokens_seen": 1707409408
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.748836462847055e-05,
      "loss": 2.507,
      "theoretical_loss": 3.4766238624871475,
      "tokens_seen": 1707540480
    },
    {
      "epoch": 0.03,
      "learning_rate": 9.748034023431232e-05,
      "loss": 2.6324,
      "theoretical_loss": 3.4766009149313097,
      "tokens_seen": 1707671552
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.747231584015407e-05,
      "loss": 2.5417,
      "theoretical_loss": 3.4765779696298704,
      "tokens_seen": 1707802624
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.746429144599584e-05,
      "loss": 2.568,
      "theoretical_loss": 3.476555026582435,
      "tokens_seen": 1707933696
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.745626705183759e-05,
      "loss": 2.4916,
      "theoretical_loss": 3.4765320857886097,
      "tokens_seen": 1708064768
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.744824265767934e-05,
      "loss": 2.4715,
      "theoretical_loss": 3.476509147247999,
      "tokens_seen": 1708195840
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.744021826352111e-05,
      "loss": 2.4469,
      "theoretical_loss": 3.4764862109602097,
      "tokens_seen": 1708326912
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.743219386936286e-05,
      "loss": 2.6352,
      "theoretical_loss": 3.476463276924847,
      "tokens_seen": 1708457984
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.742416947520463e-05,
      "loss": 2.6409,
      "theoretical_loss": 3.4764403451415173,
      "tokens_seen": 1708589056
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.741614508104638e-05,
      "loss": 2.6966,
      "theoretical_loss": 3.476417415609827,
      "tokens_seen": 1708720128
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.740812068688815e-05,
      "loss": 2.5196,
      "theoretical_loss": 3.476394488329382,
      "tokens_seen": 1708851200
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 942685,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.43986439704895,
      "objective/train/theoretical_loss": 3.4763715632997885,
      "objective/train/tokens_used": 79441376,
      "theoretical_loss": 3.4763715632997885,
      "tokens_seen": 1708982272
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.74000962927299e-05,
      "loss": 2.5859,
      "theoretical_loss": 3.4763715632997885,
      "tokens_seen": 1708982272
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.739207189857166e-05,
      "loss": 2.669,
      "theoretical_loss": 3.476348640520653,
      "tokens_seen": 1709113344
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.738404750441342e-05,
      "loss": 2.6132,
      "theoretical_loss": 3.476325719991582,
      "tokens_seen": 1709244416
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.737602311025518e-05,
      "loss": 2.5786,
      "theoretical_loss": 3.4763028017121824,
      "tokens_seen": 1709375488
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.736799871609694e-05,
      "loss": 2.4671,
      "theoretical_loss": 3.476279885682061,
      "tokens_seen": 1709506560
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.73599743219387e-05,
      "loss": 2.8103,
      "theoretical_loss": 3.476256971900824,
      "tokens_seen": 1709637632
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.735194992778046e-05,
      "loss": 2.6216,
      "theoretical_loss": 3.476234060368079,
      "tokens_seen": 1709768704
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.734392553362222e-05,
      "loss": 2.5491,
      "theoretical_loss": 3.476211151083432,
      "tokens_seen": 1709899776
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.733590113946397e-05,
      "loss": 2.5661,
      "theoretical_loss": 3.476188244046491,
      "tokens_seen": 1710030848
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.732787674530574e-05,
      "loss": 2.7132,
      "theoretical_loss": 3.4761653392568634,
      "tokens_seen": 1710161920
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.731985235114749e-05,
      "loss": 2.5686,
      "theoretical_loss": 3.476142436714156,
      "tokens_seen": 1710292992
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.731182795698925e-05,
      "loss": 2.7215,
      "theoretical_loss": 3.4761195364179764,
      "tokens_seen": 1710424064
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.730380356283101e-05,
      "loss": 2.4812,
      "theoretical_loss": 3.476096638367932,
      "tokens_seen": 1710555136
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 944008,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.545758008956909,
      "objective/train/theoretical_loss": 3.4760851901850875,
      "objective/train/tokens_used": 81079776,
      "theoretical_loss": 3.4760851901850875,
      "tokens_seen": 1710620672
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.729577916867277e-05,
      "loss": 2.5449,
      "theoretical_loss": 3.4760737425636297,
      "tokens_seen": 1710686208
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.728775477451453e-05,
      "loss": 2.7649,
      "theoretical_loss": 3.476050849004679,
      "tokens_seen": 1710817280
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.72797303803563e-05,
      "loss": 2.4776,
      "theoretical_loss": 3.476027957690686,
      "tokens_seen": 1710948352
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.727170598619805e-05,
      "loss": 2.4808,
      "theoretical_loss": 3.4760050686212596,
      "tokens_seen": 1711079424
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.72636815920398e-05,
      "loss": 2.5526,
      "theoretical_loss": 3.475982181796007,
      "tokens_seen": 1711210496
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.725565719788157e-05,
      "loss": 2.5859,
      "theoretical_loss": 3.475959297214537,
      "tokens_seen": 1711341568
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.724763280372332e-05,
      "loss": 2.6257,
      "theoretical_loss": 3.4759364148764575,
      "tokens_seen": 1711472640
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.723960840956509e-05,
      "loss": 2.7139,
      "theoretical_loss": 3.4759135347813768,
      "tokens_seen": 1711603712
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.723158401540684e-05,
      "loss": 2.6359,
      "theoretical_loss": 3.4758906569289034,
      "tokens_seen": 1711734784
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.722355962124861e-05,
      "loss": 2.6534,
      "theoretical_loss": 3.475867781318646,
      "tokens_seen": 1711865856
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.721553522709036e-05,
      "loss": 2.6935,
      "theoretical_loss": 3.4758449079502123,
      "tokens_seen": 1711996928
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.720751083293211e-05,
      "loss": 2.6434,
      "theoretical_loss": 3.475822036823212,
      "tokens_seen": 1712128000
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 944514,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2325708866119385,
      "objective/train/theoretical_loss": 3.4757991679372533,
      "objective/train/tokens_used": 82718176,
      "theoretical_loss": 3.4757991679372533,
      "tokens_seen": 1712259072
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.719948643877388e-05,
      "loss": 2.3,
      "theoretical_loss": 3.4757991679372533,
      "tokens_seen": 1712259072
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.719146204461563e-05,
      "loss": 2.7386,
      "theoretical_loss": 3.4757763012919454,
      "tokens_seen": 1712390144
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.71834376504574e-05,
      "loss": 2.6512,
      "theoretical_loss": 3.4757534368868974,
      "tokens_seen": 1712521216
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.717541325629915e-05,
      "loss": 2.5087,
      "theoretical_loss": 3.4757305747217178,
      "tokens_seen": 1712652288
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.716738886214092e-05,
      "loss": 2.5342,
      "theoretical_loss": 3.475707714796016,
      "tokens_seen": 1712783360
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.715936446798267e-05,
      "loss": 2.4806,
      "theoretical_loss": 3.4756848571094014,
      "tokens_seen": 1712914432
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.715134007382443e-05,
      "loss": 2.6971,
      "theoretical_loss": 3.475662001661483,
      "tokens_seen": 1713045504
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.714331567966619e-05,
      "loss": 2.6806,
      "theoretical_loss": 3.4756391484518714,
      "tokens_seen": 1713176576
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.713529128550795e-05,
      "loss": 2.554,
      "theoretical_loss": 3.475616297480175,
      "tokens_seen": 1713307648
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.712726689134971e-05,
      "loss": 2.579,
      "theoretical_loss": 3.4755934487460047,
      "tokens_seen": 1713438720
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.711924249719147e-05,
      "loss": 2.6336,
      "theoretical_loss": 3.4755706022489687,
      "tokens_seen": 1713569792
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.711121810303323e-05,
      "loss": 2.4975,
      "theoretical_loss": 3.4755477579886778,
      "tokens_seen": 1713700864
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.710319370887499e-05,
      "loss": 2.5121,
      "theoretical_loss": 3.4755249159647414,
      "tokens_seen": 1713831936
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 945657,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.636183261871338,
      "objective/train/theoretical_loss": 3.4755134957912848,
      "objective/train/tokens_used": 84356576,
      "theoretical_loss": 3.4755134957912848,
      "tokens_seen": 1713897472
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.709516931471674e-05,
      "loss": 2.5273,
      "theoretical_loss": 3.4755020761767703,
      "tokens_seen": 1713963008
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.70871449205585e-05,
      "loss": 2.4317,
      "theoretical_loss": 3.4754792386243745,
      "tokens_seen": 1714094080
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.707912052640026e-05,
      "loss": 2.4492,
      "theoretical_loss": 3.475456403307164,
      "tokens_seen": 1714225152
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.707109613224202e-05,
      "loss": 2.6308,
      "theoretical_loss": 3.475433570224749,
      "tokens_seen": 1714356224
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.706307173808378e-05,
      "loss": 2.5065,
      "theoretical_loss": 3.47541073937674,
      "tokens_seen": 1714487296
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.705504734392554e-05,
      "loss": 2.4417,
      "theoretical_loss": 3.4753879107627483,
      "tokens_seen": 1714618368
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.70470229497673e-05,
      "loss": 2.5615,
      "theoretical_loss": 3.4753650843823833,
      "tokens_seen": 1714749440
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.703899855560905e-05,
      "loss": 2.5363,
      "theoretical_loss": 3.475342260235257,
      "tokens_seen": 1714880512
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.703097416145082e-05,
      "loss": 2.6402,
      "theoretical_loss": 3.4753194383209793,
      "tokens_seen": 1715011584
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.702294976729257e-05,
      "loss": 2.6409,
      "theoretical_loss": 3.4752966186391614,
      "tokens_seen": 1715142656
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.701492537313434e-05,
      "loss": 2.5318,
      "theoretical_loss": 3.4752738011894144,
      "tokens_seen": 1715273728
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.700690097897609e-05,
      "loss": 2.5512,
      "theoretical_loss": 3.47525098597135,
      "tokens_seen": 1715404800
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 946326,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3408730030059814,
      "objective/train/theoretical_loss": 3.4752281729845786,
      "objective/train/tokens_used": 85994976,
      "theoretical_loss": 3.4752281729845786,
      "tokens_seen": 1715535872
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.699887658481786e-05,
      "loss": 2.4938,
      "theoretical_loss": 3.4752281729845786,
      "tokens_seen": 1715535872
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.699085219065961e-05,
      "loss": 2.5903,
      "theoretical_loss": 3.475205362228712,
      "tokens_seen": 1715666944
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.698282779650138e-05,
      "loss": 2.6315,
      "theoretical_loss": 3.4751825537033607,
      "tokens_seen": 1715798016
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.697480340234313e-05,
      "loss": 2.5496,
      "theoretical_loss": 3.4751597474081377,
      "tokens_seen": 1715929088
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.696677900818488e-05,
      "loss": 2.5582,
      "theoretical_loss": 3.4751369433426538,
      "tokens_seen": 1716060160
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.695875461402665e-05,
      "loss": 2.4528,
      "theoretical_loss": 3.4751141415065208,
      "tokens_seen": 1716191232
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.69507302198684e-05,
      "loss": 2.4352,
      "theoretical_loss": 3.4750913418993505,
      "tokens_seen": 1716322304
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.694270582571017e-05,
      "loss": 2.4735,
      "theoretical_loss": 3.475068544520755,
      "tokens_seen": 1716453376
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.693468143155192e-05,
      "loss": 2.6383,
      "theoretical_loss": 3.475045749370346,
      "tokens_seen": 1716584448
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.692665703739369e-05,
      "loss": 2.6037,
      "theoretical_loss": 3.475022956447736,
      "tokens_seen": 1716715520
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.691863264323544e-05,
      "loss": 2.5211,
      "theoretical_loss": 3.4750001657525367,
      "tokens_seen": 1716846592
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.69106082490772e-05,
      "loss": 2.5279,
      "theoretical_loss": 3.474977377284361,
      "tokens_seen": 1716977664
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.690258385491896e-05,
      "loss": 2.471,
      "theoretical_loss": 3.4749545910428212,
      "tokens_seen": 1717108736
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 947520,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7320938110351562,
      "objective/train/theoretical_loss": 3.474943198756918,
      "objective/train/tokens_used": 87633376,
      "theoretical_loss": 3.474943198756918,
      "tokens_seen": 1717174272
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.689455946076072e-05,
      "loss": 2.5489,
      "theoretical_loss": 3.474931807027529,
      "tokens_seen": 1717239808
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.688653506660248e-05,
      "loss": 2.3712,
      "theoretical_loss": 3.474909025238098,
      "tokens_seen": 1717370880
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.687851067244424e-05,
      "loss": 2.5661,
      "theoretical_loss": 3.4748862456741403,
      "tokens_seen": 1717501952
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.6870486278286e-05,
      "loss": 2.6223,
      "theoretical_loss": 3.474863468335269,
      "tokens_seen": 1717633024
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.686246188412776e-05,
      "loss": 2.6434,
      "theoretical_loss": 3.474840693221096,
      "tokens_seen": 1717764096
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.685443748996951e-05,
      "loss": 2.6146,
      "theoretical_loss": 3.474817920331236,
      "tokens_seen": 1717895168
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.684641309581128e-05,
      "loss": 2.5673,
      "theoretical_loss": 3.474795149665301,
      "tokens_seen": 1718026240
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.683838870165303e-05,
      "loss": 2.5127,
      "theoretical_loss": 3.4747723812229045,
      "tokens_seen": 1718157312
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.68303643074948e-05,
      "loss": 2.5503,
      "theoretical_loss": 3.4747496150036596,
      "tokens_seen": 1718288384
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.682233991333655e-05,
      "loss": 2.6373,
      "theoretical_loss": 3.47472685100718,
      "tokens_seen": 1718419456
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.681431551917831e-05,
      "loss": 2.5278,
      "theoretical_loss": 3.4747040892330787,
      "tokens_seen": 1718550528
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.680629112502007e-05,
      "loss": 2.5697,
      "theoretical_loss": 3.474681329680969,
      "tokens_seen": 1718681600
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 948221,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1624414920806885,
      "objective/train/theoretical_loss": 3.4746585723504655,
      "objective/train/tokens_used": 89271776,
      "theoretical_loss": 3.4746585723504655,
      "tokens_seen": 1718812672
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.679826673086182e-05,
      "loss": 2.4717,
      "theoretical_loss": 3.4746585723504655,
      "tokens_seen": 1718812672
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.679024233670359e-05,
      "loss": 2.5336,
      "theoretical_loss": 3.4746358172411815,
      "tokens_seen": 1718943744
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.678221794254534e-05,
      "loss": 2.5264,
      "theoretical_loss": 3.4746130643527304,
      "tokens_seen": 1719074816
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.677419354838711e-05,
      "loss": 2.598,
      "theoretical_loss": 3.474590313684727,
      "tokens_seen": 1719205888
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.676616915422886e-05,
      "loss": 2.492,
      "theoretical_loss": 3.474567565236785,
      "tokens_seen": 1719336960
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.675814476007063e-05,
      "loss": 2.4323,
      "theoretical_loss": 3.4745448190085177,
      "tokens_seen": 1719468032
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.675012036591238e-05,
      "loss": 2.5456,
      "theoretical_loss": 3.4745220749995407,
      "tokens_seen": 1719599104
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.674209597175413e-05,
      "loss": 2.4039,
      "theoretical_loss": 3.474499333209468,
      "tokens_seen": 1719730176
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.67340715775959e-05,
      "loss": 2.4198,
      "theoretical_loss": 3.474476593637913,
      "tokens_seen": 1719861248
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.672604718343765e-05,
      "loss": 2.6933,
      "theoretical_loss": 3.4744538562844913,
      "tokens_seen": 1719992320
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.671802278927942e-05,
      "loss": 2.5493,
      "theoretical_loss": 3.4744311211488172,
      "tokens_seen": 1720123392
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.670999839512117e-05,
      "loss": 2.6969,
      "theoretical_loss": 3.4744083882305055,
      "tokens_seen": 1720254464
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.670197400096294e-05,
      "loss": 2.4579,
      "theoretical_loss": 3.474385657529171,
      "tokens_seen": 1720385536
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 949201,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1461188793182373,
      "objective/train/theoretical_loss": 3.47437429300975,
      "objective/train/tokens_used": 90910176,
      "theoretical_loss": 3.47437429300975,
      "tokens_seen": 1720451072
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.669394960680469e-05,
      "loss": 2.4628,
      "theoretical_loss": 3.4743629290444282,
      "tokens_seen": 1720516608
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.668592521264645e-05,
      "loss": 2.4322,
      "theoretical_loss": 3.474340202775893,
      "tokens_seen": 1720647680
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.667790081848821e-05,
      "loss": 2.5498,
      "theoretical_loss": 3.47431747872318,
      "tokens_seen": 1720778752
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.666987642432997e-05,
      "loss": 2.5226,
      "theoretical_loss": 3.474294756885904,
      "tokens_seen": 1720909824
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.666185203017173e-05,
      "loss": 2.4954,
      "theoretical_loss": 3.474272037263681,
      "tokens_seen": 1721040896
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.665382763601349e-05,
      "loss": 2.6002,
      "theoretical_loss": 3.474249319856126,
      "tokens_seen": 1721171968
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.664580324185525e-05,
      "loss": 2.4917,
      "theoretical_loss": 3.4742266046628543,
      "tokens_seen": 1721303040
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.6637778847697e-05,
      "loss": 2.479,
      "theoretical_loss": 3.474203891683482,
      "tokens_seen": 1721434112
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.662975445353877e-05,
      "loss": 2.479,
      "theoretical_loss": 3.474181180917624,
      "tokens_seen": 1721565184
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.662173005938053e-05,
      "loss": 2.717,
      "theoretical_loss": 3.4741584723648975,
      "tokens_seen": 1721696256
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.661370566522228e-05,
      "loss": 2.6332,
      "theoretical_loss": 3.474135766024917,
      "tokens_seen": 1721827328
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.660568127106405e-05,
      "loss": 2.6836,
      "theoretical_loss": 3.474113061897299,
      "tokens_seen": 1721958400
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 949909,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.497939348220825,
      "objective/train/theoretical_loss": 3.474090359981659,
      "objective/train/tokens_used": 92548576,
      "theoretical_loss": 3.474090359981659,
      "tokens_seen": 1722089472
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.65976568769058e-05,
      "loss": 2.5073,
      "theoretical_loss": 3.474090359981659,
      "tokens_seen": 1722089472
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.658963248274757e-05,
      "loss": 2.4187,
      "theoretical_loss": 3.4740676602776146,
      "tokens_seen": 1722220544
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.658160808858932e-05,
      "loss": 2.5354,
      "theoretical_loss": 3.4740449627847805,
      "tokens_seen": 1722351616
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.657358369443108e-05,
      "loss": 2.428,
      "theoretical_loss": 3.4740222675027734,
      "tokens_seen": 1722482688
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.656555930027284e-05,
      "loss": 2.5802,
      "theoretical_loss": 3.4739995744312107,
      "tokens_seen": 1722613760
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.655753490611459e-05,
      "loss": 2.6047,
      "theoretical_loss": 3.473976883569708,
      "tokens_seen": 1722744832
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.654951051195636e-05,
      "loss": 2.6242,
      "theoretical_loss": 3.4739541949178827,
      "tokens_seen": 1722875904
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.654148611779811e-05,
      "loss": 2.4371,
      "theoretical_loss": 3.47393150847535,
      "tokens_seen": 1723006976
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.653346172363988e-05,
      "loss": 2.3743,
      "theoretical_loss": 3.473908824241729,
      "tokens_seen": 1723138048
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.652543732948163e-05,
      "loss": 2.7518,
      "theoretical_loss": 3.4738861422166343,
      "tokens_seen": 1723269120
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.65174129353234e-05,
      "loss": 2.44,
      "theoretical_loss": 3.4738634623996845,
      "tokens_seen": 1723400192
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.650938854116515e-05,
      "loss": 2.5699,
      "theoretical_loss": 3.4738407847904966,
      "tokens_seen": 1723531264
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.65013641470069e-05,
      "loss": 2.4851,
      "theoretical_loss": 3.473818109388687,
      "tokens_seen": 1723662336
    },
    {
      "epoch": 0.04,
      "objective/train/docs_used": 951235,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2583699226379395,
      "objective/train/theoretical_loss": 3.47380677251543,
      "objective/train/tokens_used": 94186976,
      "theoretical_loss": 3.47380677251543,
      "tokens_seen": 1723727872
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.649333975284867e-05,
      "loss": 2.5272,
      "theoretical_loss": 3.473795436193874,
      "tokens_seen": 1723793408
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.648531535869042e-05,
      "loss": 2.5135,
      "theoretical_loss": 3.4737727652056742,
      "tokens_seen": 1723924480
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.647729096453219e-05,
      "loss": 2.6295,
      "theoretical_loss": 3.473750096423705,
      "tokens_seen": 1724055552
    },
    {
      "epoch": 0.04,
      "learning_rate": 9.646926657037394e-05,
      "loss": 2.6414,
      "theoretical_loss": 3.4737274298475844,
      "tokens_seen": 1724186624
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.646124217621571e-05,
      "loss": 2.5256,
      "theoretical_loss": 3.4737047654769304,
      "tokens_seen": 1724317696
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.645321778205746e-05,
      "loss": 2.664,
      "theoretical_loss": 3.4736821033113605,
      "tokens_seen": 1724448768
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.644519338789922e-05,
      "loss": 2.5327,
      "theoretical_loss": 3.473659443350493,
      "tokens_seen": 1724579840
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.643716899374098e-05,
      "loss": 2.6298,
      "theoretical_loss": 3.4736367855939445,
      "tokens_seen": 1724710912
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.642914459958274e-05,
      "loss": 2.5402,
      "theoretical_loss": 3.4736141300413337,
      "tokens_seen": 1724841984
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.64211202054245e-05,
      "loss": 2.4692,
      "theoretical_loss": 3.47359147669228,
      "tokens_seen": 1724973056
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.641309581126626e-05,
      "loss": 2.6825,
      "theoretical_loss": 3.4735688255464003,
      "tokens_seen": 1725104128
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.640507141710802e-05,
      "loss": 2.4039,
      "theoretical_loss": 3.4735461766033136,
      "tokens_seen": 1725235200
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 951881,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6248931884765625,
      "objective/train/theoretical_loss": 3.4735235298626375,
      "objective/train/tokens_used": 95825376,
      "theoretical_loss": 3.4735235298626375,
      "tokens_seen": 1725366272
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.639704702294978e-05,
      "loss": 2.55,
      "theoretical_loss": 3.4735235298626375,
      "tokens_seen": 1725366272
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.638902262879153e-05,
      "loss": 2.3494,
      "theoretical_loss": 3.473500885323992,
      "tokens_seen": 1725497344
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.63809982346333e-05,
      "loss": 2.6021,
      "theoretical_loss": 3.473478242986994,
      "tokens_seen": 1725628416
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.637297384047505e-05,
      "loss": 2.4924,
      "theoretical_loss": 3.473455602851264,
      "tokens_seen": 1725759488
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.636494944631682e-05,
      "loss": 2.4347,
      "theoretical_loss": 3.4734329649164195,
      "tokens_seen": 1725890560
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.635692505215857e-05,
      "loss": 2.3637,
      "theoretical_loss": 3.47341032918208,
      "tokens_seen": 1726021632
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.634890065800034e-05,
      "loss": 2.5177,
      "theoretical_loss": 3.4733876956478644,
      "tokens_seen": 1726152704
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.634087626384209e-05,
      "loss": 2.5704,
      "theoretical_loss": 3.473365064313392,
      "tokens_seen": 1726283776
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.633285186968385e-05,
      "loss": 2.6099,
      "theoretical_loss": 3.4733424351782816,
      "tokens_seen": 1726414848
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.632482747552561e-05,
      "loss": 2.6702,
      "theoretical_loss": 3.4733198082421533,
      "tokens_seen": 1726545920
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.631680308136736e-05,
      "loss": 2.5298,
      "theoretical_loss": 3.4732971835046254,
      "tokens_seen": 1726676992
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.630877868720913e-05,
      "loss": 2.4978,
      "theoretical_loss": 3.4732745609653177,
      "tokens_seen": 1726808064
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.630075429305088e-05,
      "loss": 2.4702,
      "theoretical_loss": 3.4732519406238507,
      "tokens_seen": 1726939136
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 953135,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.060582160949707,
      "objective/train/theoretical_loss": 3.4732406312771884,
      "objective/train/tokens_used": 97463776,
      "theoretical_loss": 3.4732406312771884,
      "tokens_seen": 1727004672
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.629272989889265e-05,
      "loss": 2.5876,
      "theoretical_loss": 3.473229322479843,
      "tokens_seen": 1727070208
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.62847055047344e-05,
      "loss": 2.6043,
      "theoretical_loss": 3.473206706532915,
      "tokens_seen": 1727201280
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.627668111057617e-05,
      "loss": 2.5915,
      "theoretical_loss": 3.4731840927826867,
      "tokens_seen": 1727332352
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.626865671641792e-05,
      "loss": 2.4664,
      "theoretical_loss": 3.4731614812287774,
      "tokens_seen": 1727463424
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.626063232225967e-05,
      "loss": 2.5555,
      "theoretical_loss": 3.4731388718708076,
      "tokens_seen": 1727594496
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.625260792810144e-05,
      "loss": 2.5768,
      "theoretical_loss": 3.4731162647083975,
      "tokens_seen": 1727725568
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.62445835339432e-05,
      "loss": 2.6285,
      "theoretical_loss": 3.473093659741167,
      "tokens_seen": 1727856640
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.623655913978496e-05,
      "loss": 2.6888,
      "theoretical_loss": 3.473071056968737,
      "tokens_seen": 1727987712
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.622853474562671e-05,
      "loss": 2.6184,
      "theoretical_loss": 3.4730484563907273,
      "tokens_seen": 1728118784
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.622051035146848e-05,
      "loss": 2.6123,
      "theoretical_loss": 3.473025858006759,
      "tokens_seen": 1728249856
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.621248595731023e-05,
      "loss": 2.4683,
      "theoretical_loss": 3.4730032618164524,
      "tokens_seen": 1728380928
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.620446156315199e-05,
      "loss": 2.4986,
      "theoretical_loss": 3.472980667819428,
      "tokens_seen": 1728512000
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 953685,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4338219165802,
      "objective/train/theoretical_loss": 3.472958076015307,
      "objective/train/tokens_used": 99102176,
      "theoretical_loss": 3.472958076015307,
      "tokens_seen": 1728643072
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.619643716899375e-05,
      "loss": 2.6793,
      "theoretical_loss": 3.472958076015307,
      "tokens_seen": 1728643072
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.61884127748355e-05,
      "loss": 2.3993,
      "theoretical_loss": 3.4729354864037107,
      "tokens_seen": 1728774144
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.618038838067727e-05,
      "loss": 2.5727,
      "theoretical_loss": 3.4729128989842595,
      "tokens_seen": 1728905216
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.617236398651903e-05,
      "loss": 2.5781,
      "theoretical_loss": 3.4728903137565745,
      "tokens_seen": 1729036288
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.616433959236079e-05,
      "loss": 2.5062,
      "theoretical_loss": 3.472867730720277,
      "tokens_seen": 1729167360
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.615631519820255e-05,
      "loss": 2.6331,
      "theoretical_loss": 3.4728451498749884,
      "tokens_seen": 1729298432
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.61482908040443e-05,
      "loss": 2.6771,
      "theoretical_loss": 3.4728225712203304,
      "tokens_seen": 1729429504
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.614026640988607e-05,
      "loss": 2.562,
      "theoretical_loss": 3.472799994755924,
      "tokens_seen": 1729560576
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.613224201572782e-05,
      "loss": 2.5274,
      "theoretical_loss": 3.4727774204813904,
      "tokens_seen": 1729691648
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.612421762156959e-05,
      "loss": 2.4391,
      "theoretical_loss": 3.472754848396352,
      "tokens_seen": 1729822720
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.611619322741134e-05,
      "loss": 2.5883,
      "theoretical_loss": 3.4727322785004304,
      "tokens_seen": 1729953792
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.610816883325309e-05,
      "loss": 2.7006,
      "theoretical_loss": 3.4727097107932474,
      "tokens_seen": 1730084864
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.610014443909486e-05,
      "loss": 2.5706,
      "theoretical_loss": 3.472687145274425,
      "tokens_seen": 1730215936
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 954893,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3295490741729736,
      "objective/train/theoretical_loss": 3.472675863335531,
      "objective/train/tokens_used": 100740576,
      "theoretical_loss": 3.472675863335531,
      "tokens_seen": 1730281472
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.609212004493661e-05,
      "loss": 2.5245,
      "theoretical_loss": 3.472664581943585,
      "tokens_seen": 1730347008
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.608409565077838e-05,
      "loss": 2.5552,
      "theoretical_loss": 3.4726420208003494,
      "tokens_seen": 1730478080
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.607607125662013e-05,
      "loss": 2.5185,
      "theoretical_loss": 3.4726194618443413,
      "tokens_seen": 1730609152
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.60680468624619e-05,
      "loss": 2.53,
      "theoretical_loss": 3.472596905075182,
      "tokens_seen": 1730740224
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.606002246830365e-05,
      "loss": 2.6368,
      "theoretical_loss": 3.472574350492495,
      "tokens_seen": 1730871296
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.60519980741454e-05,
      "loss": 2.4409,
      "theoretical_loss": 3.4725517980959015,
      "tokens_seen": 1731002368
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.604397367998717e-05,
      "loss": 2.6554,
      "theoretical_loss": 3.472529247885025,
      "tokens_seen": 1731133440
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.603594928582892e-05,
      "loss": 2.6305,
      "theoretical_loss": 3.4725066998594882,
      "tokens_seen": 1731264512
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.602792489167069e-05,
      "loss": 2.5356,
      "theoretical_loss": 3.4724841540189137,
      "tokens_seen": 1731395584
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.601990049751244e-05,
      "loss": 2.5259,
      "theoretical_loss": 3.4724616103629242,
      "tokens_seen": 1731526656
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.60118761033542e-05,
      "loss": 2.5755,
      "theoretical_loss": 3.472439068891143,
      "tokens_seen": 1731657728
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.600385170919596e-05,
      "loss": 2.6055,
      "theoretical_loss": 3.4724165296031924,
      "tokens_seen": 1731788800
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 956184,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5680079460144043,
      "objective/train/theoretical_loss": 3.472393992498697,
      "objective/train/tokens_used": 102378976,
      "theoretical_loss": 3.472393992498697,
      "tokens_seen": 1731919872
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.599582731503772e-05,
      "loss": 2.6725,
      "theoretical_loss": 3.472393992498697,
      "tokens_seen": 1731919872
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.598780292087948e-05,
      "loss": 2.5937,
      "theoretical_loss": 3.472371457577279,
      "tokens_seen": 1732050944
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.597977852672124e-05,
      "loss": 2.613,
      "theoretical_loss": 3.472348924838562,
      "tokens_seen": 1732182016
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.5971754132563e-05,
      "loss": 2.5003,
      "theoretical_loss": 3.4723263942821694,
      "tokens_seen": 1732313088
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.596372973840476e-05,
      "loss": 2.5422,
      "theoretical_loss": 3.4723038659077248,
      "tokens_seen": 1732444160
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.595570534424651e-05,
      "loss": 2.5334,
      "theoretical_loss": 3.4722813397148515,
      "tokens_seen": 1732575232
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.594768095008828e-05,
      "loss": 2.4992,
      "theoretical_loss": 3.4722588157031744,
      "tokens_seen": 1732706304
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.593965655593003e-05,
      "loss": 2.5993,
      "theoretical_loss": 3.4722362938723155,
      "tokens_seen": 1732837376
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.59316321617718e-05,
      "loss": 2.5522,
      "theoretical_loss": 3.472213774221901,
      "tokens_seen": 1732968448
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.592360776761355e-05,
      "loss": 2.5903,
      "theoretical_loss": 3.472191256751552,
      "tokens_seen": 1733099520
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.59155833734553e-05,
      "loss": 2.55,
      "theoretical_loss": 3.4721687414608953,
      "tokens_seen": 1733230592
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.590755897929707e-05,
      "loss": 2.6655,
      "theoretical_loss": 3.472146228349554,
      "tokens_seen": 1733361664
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.589953458513882e-05,
      "loss": 2.5992,
      "theoretical_loss": 3.472123717417152,
      "tokens_seen": 1733492736
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 956719,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.863294839859009,
      "objective/train/theoretical_loss": 3.472112462767936,
      "objective/train/tokens_used": 104017376,
      "theoretical_loss": 3.472112462767936,
      "tokens_seen": 1733558272
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.589151019098059e-05,
      "loss": 2.592,
      "theoretical_loss": 3.4721012086633145,
      "tokens_seen": 1733623808
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.588348579682234e-05,
      "loss": 2.6002,
      "theoretical_loss": 3.472078702087665,
      "tokens_seen": 1733754880
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.587546140266411e-05,
      "loss": 2.396,
      "theoretical_loss": 3.4720561976898288,
      "tokens_seen": 1733885952
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.586743700850586e-05,
      "loss": 2.5257,
      "theoretical_loss": 3.47203369546943,
      "tokens_seen": 1734017024
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.585941261434761e-05,
      "loss": 2.2733,
      "theoretical_loss": 3.472011195426095,
      "tokens_seen": 1734148096
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.585138822018938e-05,
      "loss": 2.5829,
      "theoretical_loss": 3.4719886975594463,
      "tokens_seen": 1734279168
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.584336382603113e-05,
      "loss": 2.5333,
      "theoretical_loss": 3.47196620186911,
      "tokens_seen": 1734410240
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.58353394318729e-05,
      "loss": 2.6064,
      "theoretical_loss": 3.4719437083547113,
      "tokens_seen": 1734541312
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.582731503771465e-05,
      "loss": 2.5828,
      "theoretical_loss": 3.4719212170158755,
      "tokens_seen": 1734672384
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.581929064355641e-05,
      "loss": 2.5385,
      "theoretical_loss": 3.471898727852227,
      "tokens_seen": 1734803456
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.581126624939817e-05,
      "loss": 2.5934,
      "theoretical_loss": 3.4718762408633914,
      "tokens_seen": 1734934528
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.580324185523993e-05,
      "loss": 2.6143,
      "theoretical_loss": 3.471853756048994,
      "tokens_seen": 1735065600
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 957813,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.456376314163208,
      "objective/train/theoretical_loss": 3.471831273408661,
      "objective/train/tokens_used": 105655776,
      "theoretical_loss": 3.471831273408661,
      "tokens_seen": 1735196672
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.57952174610817e-05,
      "loss": 2.5648,
      "theoretical_loss": 3.471831273408661,
      "tokens_seen": 1735196672
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.578719306692345e-05,
      "loss": 2.6123,
      "theoretical_loss": 3.471808792942017,
      "tokens_seen": 1735327744
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.57791686727652e-05,
      "loss": 2.5439,
      "theoretical_loss": 3.4717863146486887,
      "tokens_seen": 1735458816
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.577114427860697e-05,
      "loss": 2.5435,
      "theoretical_loss": 3.471763838528301,
      "tokens_seen": 1735589888
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.576311988444872e-05,
      "loss": 2.6144,
      "theoretical_loss": 3.47174136458048,
      "tokens_seen": 1735720960
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.575509549029049e-05,
      "loss": 2.6676,
      "theoretical_loss": 3.4717188928048524,
      "tokens_seen": 1735852032
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.574707109613224e-05,
      "loss": 2.6404,
      "theoretical_loss": 3.471696423201043,
      "tokens_seen": 1735983104
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.5739046701974e-05,
      "loss": 2.42,
      "theoretical_loss": 3.4716739557686793,
      "tokens_seen": 1736114176
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.573102230781576e-05,
      "loss": 2.5927,
      "theoretical_loss": 3.4716514905073863,
      "tokens_seen": 1736245248
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.572299791365751e-05,
      "loss": 2.5552,
      "theoretical_loss": 3.4716290274167907,
      "tokens_seen": 1736376320
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.571497351949928e-05,
      "loss": 2.5069,
      "theoretical_loss": 3.4716065664965194,
      "tokens_seen": 1736507392
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.570694912534103e-05,
      "loss": 2.6358,
      "theoretical_loss": 3.4715841077461986,
      "tokens_seen": 1736638464
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.56989247311828e-05,
      "loss": 2.5779,
      "theoretical_loss": 3.4715616511654552,
      "tokens_seen": 1736769536
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 958321,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.462451934814453,
      "objective/train/theoretical_loss": 3.4715504236885577,
      "objective/train/tokens_used": 107294176,
      "theoretical_loss": 3.4715504236885577,
      "tokens_seen": 1736835072
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.569090033702455e-05,
      "loss": 2.5436,
      "theoretical_loss": 3.4715391967539153,
      "tokens_seen": 1736900608
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.568287594286632e-05,
      "loss": 2.7413,
      "theoretical_loss": 3.471516744511206,
      "tokens_seen": 1737031680
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.567485154870807e-05,
      "loss": 2.4977,
      "theoretical_loss": 3.471494294436954,
      "tokens_seen": 1737162752
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.566682715454983e-05,
      "loss": 2.4846,
      "theoretical_loss": 3.471471846530787,
      "tokens_seen": 1737293824
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.565880276039159e-05,
      "loss": 2.5388,
      "theoretical_loss": 3.4714494007923307,
      "tokens_seen": 1737424896
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.565077836623335e-05,
      "loss": 2.6694,
      "theoretical_loss": 3.4714269572212135,
      "tokens_seen": 1737555968
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.564275397207511e-05,
      "loss": 2.5805,
      "theoretical_loss": 3.4714045158170626,
      "tokens_seen": 1737687040
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.563472957791687e-05,
      "loss": 2.4512,
      "theoretical_loss": 3.4713820765795047,
      "tokens_seen": 1737818112
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.562670518375863e-05,
      "loss": 2.4375,
      "theoretical_loss": 3.471359639508168,
      "tokens_seen": 1737949184
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.561868078960038e-05,
      "loss": 2.5466,
      "theoretical_loss": 3.471337204602679,
      "tokens_seen": 1738080256
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.561065639544214e-05,
      "loss": 2.6156,
      "theoretical_loss": 3.471314771862666,
      "tokens_seen": 1738211328
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.56026320012839e-05,
      "loss": 2.7741,
      "theoretical_loss": 3.4712923412877563,
      "tokens_seen": 1738342400
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 959398,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5910706520080566,
      "objective/train/theoretical_loss": 3.4712699128775784,
      "objective/train/tokens_used": 108932576,
      "theoretical_loss": 3.4712699128775784,
      "tokens_seen": 1738473472
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.559460760712566e-05,
      "loss": 2.5787,
      "theoretical_loss": 3.4712699128775784,
      "tokens_seen": 1738473472
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.558658321296742e-05,
      "loss": 2.7085,
      "theoretical_loss": 3.4712474866317597,
      "tokens_seen": 1738604544
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.557855881880918e-05,
      "loss": 2.5791,
      "theoretical_loss": 3.4712250625499284,
      "tokens_seen": 1738735616
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.557053442465094e-05,
      "loss": 2.5154,
      "theoretical_loss": 3.471202640631713,
      "tokens_seen": 1738866688
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.55625100304927e-05,
      "loss": 2.6328,
      "theoretical_loss": 3.4711802208767404,
      "tokens_seen": 1738997760
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.555448563633445e-05,
      "loss": 2.5779,
      "theoretical_loss": 3.4711578032846404,
      "tokens_seen": 1739128832
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.554646124217622e-05,
      "loss": 2.4909,
      "theoretical_loss": 3.47113538785504,
      "tokens_seen": 1739259904
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.553843684801797e-05,
      "loss": 2.6613,
      "theoretical_loss": 3.471112974587568,
      "tokens_seen": 1739390976
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.553041245385974e-05,
      "loss": 2.647,
      "theoretical_loss": 3.471090563481854,
      "tokens_seen": 1739522048
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.552238805970149e-05,
      "loss": 2.5482,
      "theoretical_loss": 3.4710681545375253,
      "tokens_seen": 1739653120
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.551436366554326e-05,
      "loss": 2.6415,
      "theoretical_loss": 3.471045747754211,
      "tokens_seen": 1739784192
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.550633927138501e-05,
      "loss": 2.5234,
      "theoretical_loss": 3.4710233431315407,
      "tokens_seen": 1739915264
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.549831487722676e-05,
      "loss": 2.4932,
      "theoretical_loss": 3.471000940669142,
      "tokens_seen": 1740046336
    },
    {
      "epoch": 0.05,
      "objective/train/docs_used": 960697,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7677383422851562,
      "objective/train/theoretical_loss": 3.470989740247929,
      "objective/train/tokens_used": 110570976,
      "theoretical_loss": 3.470989740247929,
      "tokens_seen": 1740111872
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.549029048306853e-05,
      "loss": 2.5452,
      "theoretical_loss": 3.4709785403666453,
      "tokens_seen": 1740177408
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.548226608891028e-05,
      "loss": 2.5708,
      "theoretical_loss": 3.4709561422236783,
      "tokens_seen": 1740308480
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.547424169475205e-05,
      "loss": 2.7149,
      "theoretical_loss": 3.4709337462398713,
      "tokens_seen": 1740439552
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.54662173005938e-05,
      "loss": 2.5015,
      "theoretical_loss": 3.470911352414853,
      "tokens_seen": 1740570624
    },
    {
      "epoch": 0.05,
      "learning_rate": 9.545819290643557e-05,
      "loss": 2.6489,
      "theoretical_loss": 3.470888960748253,
      "tokens_seen": 1740701696
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.545016851227732e-05,
      "loss": 2.6192,
      "theoretical_loss": 3.4708665712397004,
      "tokens_seen": 1740832768
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.544214411811908e-05,
      "loss": 2.6038,
      "theoretical_loss": 3.4708441838888255,
      "tokens_seen": 1740963840
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.543411972396084e-05,
      "loss": 2.6892,
      "theoretical_loss": 3.4708217986952574,
      "tokens_seen": 1741094912
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.54260953298026e-05,
      "loss": 2.5735,
      "theoretical_loss": 3.4707994156586253,
      "tokens_seen": 1741225984
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.541807093564436e-05,
      "loss": 2.6655,
      "theoretical_loss": 3.47077703477856,
      "tokens_seen": 1741357056
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.541004654148612e-05,
      "loss": 2.5846,
      "theoretical_loss": 3.470754656054691,
      "tokens_seen": 1741488128
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.540202214732788e-05,
      "loss": 2.539,
      "theoretical_loss": 3.470732279486649,
      "tokens_seen": 1741619200
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 961372,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.166609525680542,
      "objective/train/theoretical_loss": 3.470709905074062,
      "objective/train/tokens_used": 112209376,
      "theoretical_loss": 3.470709905074062,
      "tokens_seen": 1741750272
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.539399775316964e-05,
      "loss": 2.6481,
      "theoretical_loss": 3.470709905074062,
      "tokens_seen": 1741750272
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.53859733590114e-05,
      "loss": 2.4939,
      "theoretical_loss": 3.470687532816563,
      "tokens_seen": 1741881344
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.537794896485315e-05,
      "loss": 2.488,
      "theoretical_loss": 3.4706651627137806,
      "tokens_seen": 1742012416
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.536992457069491e-05,
      "loss": 2.5908,
      "theoretical_loss": 3.470642794765345,
      "tokens_seen": 1742143488
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.536190017653667e-05,
      "loss": 2.5974,
      "theoretical_loss": 3.4706204289708875,
      "tokens_seen": 1742274560
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.535387578237843e-05,
      "loss": 2.5184,
      "theoretical_loss": 3.4705980653300386,
      "tokens_seen": 1742405632
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.53458513882202e-05,
      "loss": 2.4658,
      "theoretical_loss": 3.4705757038424285,
      "tokens_seen": 1742536704
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.533782699406195e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.470553344507688,
      "tokens_seen": 1742667776
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.532980259990371e-05,
      "loss": 2.5697,
      "theoretical_loss": 3.470530987325448,
      "tokens_seen": 1742798848
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.532177820574547e-05,
      "loss": 2.7383,
      "theoretical_loss": 3.4705086322953393,
      "tokens_seen": 1742929920
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.531375381158722e-05,
      "loss": 2.5499,
      "theoretical_loss": 3.4704862794169937,
      "tokens_seen": 1743060992
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.530572941742899e-05,
      "loss": 2.5394,
      "theoretical_loss": 3.470463928690041,
      "tokens_seen": 1743192064
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.529770502327074e-05,
      "loss": 2.5673,
      "theoretical_loss": 3.4704415801141133,
      "tokens_seen": 1743323136
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 962529,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.988215923309326,
      "objective/train/theoretical_loss": 3.4704304066326683,
      "objective/train/tokens_used": 113847776,
      "theoretical_loss": 3.4704304066326683,
      "tokens_seen": 1743388672
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.528968062911251e-05,
      "loss": 2.7066,
      "theoretical_loss": 3.4704192336888413,
      "tokens_seen": 1743454208
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.528165623495426e-05,
      "loss": 2.6361,
      "theoretical_loss": 3.4703968894138573,
      "tokens_seen": 1743585280
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.527363184079603e-05,
      "loss": 2.6597,
      "theoretical_loss": 3.470374547288792,
      "tokens_seen": 1743716352
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.526560744663778e-05,
      "loss": 2.7227,
      "theoretical_loss": 3.4703522073132773,
      "tokens_seen": 1743847424
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.525758305247953e-05,
      "loss": 2.4906,
      "theoretical_loss": 3.470329869486944,
      "tokens_seen": 1743978496
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.52495586583213e-05,
      "loss": 2.6045,
      "theoretical_loss": 3.4703075338094256,
      "tokens_seen": 1744109568
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.524153426416305e-05,
      "loss": 2.5417,
      "theoretical_loss": 3.470285200280352,
      "tokens_seen": 1744240640
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.523350987000482e-05,
      "loss": 2.6117,
      "theoretical_loss": 3.470262868899357,
      "tokens_seen": 1744371712
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.522548547584657e-05,
      "loss": 2.6588,
      "theoretical_loss": 3.4702405396660705,
      "tokens_seen": 1744502784
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.521746108168834e-05,
      "loss": 2.6923,
      "theoretical_loss": 3.470218212580126,
      "tokens_seen": 1744633856
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.520943668753009e-05,
      "loss": 2.7159,
      "theoretical_loss": 3.4701958876411556,
      "tokens_seen": 1744764928
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.520141229337185e-05,
      "loss": 2.5938,
      "theoretical_loss": 3.4701735648487912,
      "tokens_seen": 1744896000
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 963086,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.850843667984009,
      "objective/train/theoretical_loss": 3.4701512442026656,
      "objective/train/tokens_used": 115486176,
      "theoretical_loss": 3.4701512442026656,
      "tokens_seen": 1745027072
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.519338789921361e-05,
      "loss": 2.575,
      "theoretical_loss": 3.4701512442026656,
      "tokens_seen": 1745027072
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.518536350505537e-05,
      "loss": 2.6056,
      "theoretical_loss": 3.4701289257024106,
      "tokens_seen": 1745158144
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.517733911089713e-05,
      "loss": 2.5319,
      "theoretical_loss": 3.47010660934766,
      "tokens_seen": 1745289216
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.516931471673889e-05,
      "loss": 2.5956,
      "theoretical_loss": 3.4700842951380446,
      "tokens_seen": 1745420288
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.516129032258065e-05,
      "loss": 2.5044,
      "theoretical_loss": 3.4700619830731982,
      "tokens_seen": 1745551360
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.51532659284224e-05,
      "loss": 2.499,
      "theoretical_loss": 3.4700396731527543,
      "tokens_seen": 1745682432
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.514524153426416e-05,
      "loss": 2.6185,
      "theoretical_loss": 3.4700173653763446,
      "tokens_seen": 1745813504
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.513721714010593e-05,
      "loss": 2.5509,
      "theoretical_loss": 3.469995059743603,
      "tokens_seen": 1745944576
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.512919274594768e-05,
      "loss": 2.5919,
      "theoretical_loss": 3.469972756254162,
      "tokens_seen": 1746075648
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.512116835178944e-05,
      "loss": 2.4592,
      "theoretical_loss": 3.4699504549076545,
      "tokens_seen": 1746206720
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.51131439576312e-05,
      "loss": 2.6279,
      "theoretical_loss": 3.4699281557037147,
      "tokens_seen": 1746337792
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.510511956347296e-05,
      "loss": 2.6709,
      "theoretical_loss": 3.4699058586419755,
      "tokens_seen": 1746468864
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.509709516931472e-05,
      "loss": 2.5483,
      "theoretical_loss": 3.46988356372207,
      "tokens_seen": 1746599936
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 964151,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.168619394302368,
      "objective/train/theoretical_loss": 3.469872417065191,
      "objective/train/tokens_used": 117124576,
      "theoretical_loss": 3.469872417065191,
      "tokens_seen": 1746665472
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.508907077515648e-05,
      "loss": 2.4492,
      "theoretical_loss": 3.4698612709436327,
      "tokens_seen": 1746731008
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.508104638099824e-05,
      "loss": 2.6619,
      "theoretical_loss": 3.469838980306297,
      "tokens_seen": 1746862080
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.507302198683999e-05,
      "loss": 2.6721,
      "theoretical_loss": 3.4698166918096955,
      "tokens_seen": 1746993152
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.506499759268176e-05,
      "loss": 2.502,
      "theoretical_loss": 3.469794405453463,
      "tokens_seen": 1747124224
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.505697319852351e-05,
      "loss": 2.599,
      "theoretical_loss": 3.4697721212372334,
      "tokens_seen": 1747255296
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.504894880436528e-05,
      "loss": 2.6716,
      "theoretical_loss": 3.4697498391606403,
      "tokens_seen": 1747386368
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.504092441020703e-05,
      "loss": 2.5568,
      "theoretical_loss": 3.469727559223318,
      "tokens_seen": 1747517440
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.50329000160488e-05,
      "loss": 2.6531,
      "theoretical_loss": 3.469705281424901,
      "tokens_seen": 1747648512
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.502487562189055e-05,
      "loss": 2.653,
      "theoretical_loss": 3.469683005765023,
      "tokens_seen": 1747779584
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.50168512277323e-05,
      "loss": 2.547,
      "theoretical_loss": 3.469660732243319,
      "tokens_seen": 1747910656
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.500882683357407e-05,
      "loss": 2.5207,
      "theoretical_loss": 3.469638460859423,
      "tokens_seen": 1748041728
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.500080243941582e-05,
      "loss": 2.6592,
      "theoretical_loss": 3.4696161916129693,
      "tokens_seen": 1748172800
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 964837,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.9922051429748535,
      "objective/train/theoretical_loss": 3.469593924503593,
      "objective/train/tokens_used": 118762976,
      "theoretical_loss": 3.469593924503593,
      "tokens_seen": 1748303872
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.499277804525759e-05,
      "loss": 2.5553,
      "theoretical_loss": 3.469593924503593,
      "tokens_seen": 1748303872
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.498475365109934e-05,
      "loss": 2.5655,
      "theoretical_loss": 3.4695716595309287,
      "tokens_seen": 1748434944
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.497672925694111e-05,
      "loss": 2.6482,
      "theoretical_loss": 3.469549396694611,
      "tokens_seen": 1748566016
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.496870486278286e-05,
      "loss": 2.5774,
      "theoretical_loss": 3.469527135994275,
      "tokens_seen": 1748697088
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.496068046862462e-05,
      "loss": 2.4727,
      "theoretical_loss": 3.469504877429556,
      "tokens_seen": 1748828160
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.495265607446638e-05,
      "loss": 2.5101,
      "theoretical_loss": 3.4694826210000884,
      "tokens_seen": 1748959232
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.494463168030814e-05,
      "loss": 2.6768,
      "theoretical_loss": 3.4694603667055075,
      "tokens_seen": 1749090304
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.49366072861499e-05,
      "loss": 2.553,
      "theoretical_loss": 3.4694381145454494,
      "tokens_seen": 1749221376
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.492858289199166e-05,
      "loss": 2.4986,
      "theoretical_loss": 3.469415864519548,
      "tokens_seen": 1749352448
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.492055849783342e-05,
      "loss": 2.66,
      "theoretical_loss": 3.469393616627441,
      "tokens_seen": 1749483520
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.491253410367518e-05,
      "loss": 2.5936,
      "theoretical_loss": 3.4693713708687612,
      "tokens_seen": 1749614592
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.490450970951693e-05,
      "loss": 2.5735,
      "theoretical_loss": 3.4693491272431456,
      "tokens_seen": 1749745664
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.48964853153587e-05,
      "loss": 2.6236,
      "theoretical_loss": 3.46932688575023,
      "tokens_seen": 1749876736
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 965967,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4741082191467285,
      "objective/train/theoretical_loss": 3.4693157658034206,
      "objective/train/tokens_used": 120401376,
      "theoretical_loss": 3.4693157658034206,
      "tokens_seen": 1749942272
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.488846092120045e-05,
      "loss": 2.6312,
      "theoretical_loss": 3.46930464638965,
      "tokens_seen": 1750007808
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.488043652704221e-05,
      "loss": 2.6914,
      "theoretical_loss": 3.4692824091610412,
      "tokens_seen": 1750138880
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.487241213288397e-05,
      "loss": 2.4969,
      "theoretical_loss": 3.4692601740640403,
      "tokens_seen": 1750269952
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.486438773872573e-05,
      "loss": 2.568,
      "theoretical_loss": 3.4692379410982825,
      "tokens_seen": 1750401024
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.485636334456749e-05,
      "loss": 2.5101,
      "theoretical_loss": 3.4692157102634047,
      "tokens_seen": 1750532096
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.484833895040924e-05,
      "loss": 2.6857,
      "theoretical_loss": 3.4691934815590426,
      "tokens_seen": 1750663168
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.484031455625101e-05,
      "loss": 2.4291,
      "theoretical_loss": 3.4691712549848326,
      "tokens_seen": 1750794240
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.483229016209276e-05,
      "loss": 2.6526,
      "theoretical_loss": 3.469149030540412,
      "tokens_seen": 1750925312
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.482426576793453e-05,
      "loss": 2.5627,
      "theoretical_loss": 3.4691268082254156,
      "tokens_seen": 1751056384
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.481624137377628e-05,
      "loss": 2.5353,
      "theoretical_loss": 3.4691045880394817,
      "tokens_seen": 1751187456
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.480821697961805e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.4690823699822455,
      "tokens_seen": 1751318528
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.48001925854598e-05,
      "loss": 2.6184,
      "theoretical_loss": 3.469060154053345,
      "tokens_seen": 1751449600
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 966884,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.789921760559082,
      "objective/train/theoretical_loss": 3.4690379402524165,
      "objective/train/tokens_used": 122039776,
      "theoretical_loss": 3.4690379402524165,
      "tokens_seen": 1751580672
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.479216819130155e-05,
      "loss": 2.5307,
      "theoretical_loss": 3.4690379402524165,
      "tokens_seen": 1751580672
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.478414379714332e-05,
      "loss": 2.5257,
      "theoretical_loss": 3.4690157285790972,
      "tokens_seen": 1751711744
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.477611940298507e-05,
      "loss": 2.561,
      "theoretical_loss": 3.468993519033024,
      "tokens_seen": 1751842816
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.476809500882684e-05,
      "loss": 2.3812,
      "theoretical_loss": 3.468971311613834,
      "tokens_seen": 1751973888
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.476007061466859e-05,
      "loss": 2.5491,
      "theoretical_loss": 3.468949106321164,
      "tokens_seen": 1752104960
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.475204622051036e-05,
      "loss": 2.5825,
      "theoretical_loss": 3.468926903154652,
      "tokens_seen": 1752236032
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.474402182635211e-05,
      "loss": 2.608,
      "theoretical_loss": 3.4689047021139356,
      "tokens_seen": 1752367104
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.473599743219388e-05,
      "loss": 2.6951,
      "theoretical_loss": 3.4688825031986514,
      "tokens_seen": 1752498176
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.472797303803563e-05,
      "loss": 2.5484,
      "theoretical_loss": 3.4688603064084376,
      "tokens_seen": 1752629248
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.471994864387739e-05,
      "loss": 2.7742,
      "theoretical_loss": 3.4688381117429317,
      "tokens_seen": 1752760320
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.471192424971915e-05,
      "loss": 2.5422,
      "theoretical_loss": 3.4688159192017713,
      "tokens_seen": 1752891392
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.47038998555609e-05,
      "loss": 2.637,
      "theoretical_loss": 3.4687937287845947,
      "tokens_seen": 1753022464
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.469587546140267e-05,
      "loss": 2.5073,
      "theoretical_loss": 3.4687715404910398,
      "tokens_seen": 1753153536
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 967417,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.34035325050354,
      "objective/train/theoretical_loss": 3.4687604471405065,
      "objective/train/tokens_used": 123678176,
      "theoretical_loss": 3.4687604471405065,
      "tokens_seen": 1753219072
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.468785106724443e-05,
      "loss": 2.5376,
      "theoretical_loss": 3.4687493543207433,
      "tokens_seen": 1753284608
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.467982667308619e-05,
      "loss": 2.6251,
      "theoretical_loss": 3.468727170273345,
      "tokens_seen": 1753415680
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.467180227892795e-05,
      "loss": 2.8049,
      "theoretical_loss": 3.468704988348483,
      "tokens_seen": 1753546752
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.46637778847697e-05,
      "loss": 2.5002,
      "theoretical_loss": 3.468682808545794,
      "tokens_seen": 1753677824
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.465575349061147e-05,
      "loss": 2.5396,
      "theoretical_loss": 3.4686606308649184,
      "tokens_seen": 1753808896
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.464772909645322e-05,
      "loss": 2.5852,
      "theoretical_loss": 3.468638455305493,
      "tokens_seen": 1753939968
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.463970470229499e-05,
      "loss": 2.5664,
      "theoretical_loss": 3.468616281867157,
      "tokens_seen": 1754071040
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.463168030813674e-05,
      "loss": 2.6855,
      "theoretical_loss": 3.46859411054955,
      "tokens_seen": 1754202112
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.46236559139785e-05,
      "loss": 2.6069,
      "theoretical_loss": 3.468571941352309,
      "tokens_seen": 1754333184
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.461563151982026e-05,
      "loss": 2.5943,
      "theoretical_loss": 3.4685497742750737,
      "tokens_seen": 1754464256
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.460760712566201e-05,
      "loss": 2.5968,
      "theoretical_loss": 3.4685276093174835,
      "tokens_seen": 1754595328
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.459958273150378e-05,
      "loss": 2.4725,
      "theoretical_loss": 3.4685054464791767,
      "tokens_seen": 1754726400
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 968410,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6490659713745117,
      "objective/train/theoretical_loss": 3.4684832857597923,
      "objective/train/tokens_used": 125316576,
      "theoretical_loss": 3.4684832857597923,
      "tokens_seen": 1754857472
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.459155833734553e-05,
      "loss": 2.5628,
      "theoretical_loss": 3.4684832857597923,
      "tokens_seen": 1754857472
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.45835339431873e-05,
      "loss": 2.5849,
      "theoretical_loss": 3.46846112715897,
      "tokens_seen": 1754988544
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.457550954902905e-05,
      "loss": 2.6675,
      "theoretical_loss": 3.468438970676348,
      "tokens_seen": 1755119616
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.456748515487082e-05,
      "loss": 2.5566,
      "theoretical_loss": 3.468416816311567,
      "tokens_seen": 1755250688
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.455946076071257e-05,
      "loss": 2.4097,
      "theoretical_loss": 3.468394664064266,
      "tokens_seen": 1755381760
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.455143636655432e-05,
      "loss": 2.6195,
      "theoretical_loss": 3.468372513934084,
      "tokens_seen": 1755512832
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.454341197239609e-05,
      "loss": 2.5105,
      "theoretical_loss": 3.468350365920662,
      "tokens_seen": 1755643904
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.453538757823784e-05,
      "loss": 2.5745,
      "theoretical_loss": 3.468328220023638,
      "tokens_seen": 1755774976
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.452736318407961e-05,
      "loss": 2.5794,
      "theoretical_loss": 3.4683060762426523,
      "tokens_seen": 1755906048
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.451933878992136e-05,
      "loss": 2.563,
      "theoretical_loss": 3.4682839345773457,
      "tokens_seen": 1756037120
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.451131439576313e-05,
      "loss": 2.5961,
      "theoretical_loss": 3.468261795027357,
      "tokens_seen": 1756168192
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.450329000160488e-05,
      "loss": 2.6463,
      "theoretical_loss": 3.4682396575923264,
      "tokens_seen": 1756299264
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.449526560744664e-05,
      "loss": 2.5798,
      "theoretical_loss": 3.468217522271895,
      "tokens_seen": 1756430336
    },
    {
      "epoch": 0.06,
      "objective/train/docs_used": 968954,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.701629638671875,
      "objective/train/theoretical_loss": 3.468206455404541,
      "objective/train/tokens_used": 126954976,
      "theoretical_loss": 3.468206455404541,
      "tokens_seen": 1756495872
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.44872412132884e-05,
      "loss": 2.4543,
      "theoretical_loss": 3.468195389065702,
      "tokens_seen": 1756561408
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.447921681913016e-05,
      "loss": 2.5349,
      "theoretical_loss": 3.4681732579733886,
      "tokens_seen": 1756692480
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.447119242497192e-05,
      "loss": 2.5532,
      "theoretical_loss": 3.4681511289945943,
      "tokens_seen": 1756823552
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.446316803081368e-05,
      "loss": 2.5709,
      "theoretical_loss": 3.46812900212896,
      "tokens_seen": 1756954624
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.445514363665544e-05,
      "loss": 2.5352,
      "theoretical_loss": 3.468106877376126,
      "tokens_seen": 1757085696
    },
    {
      "epoch": 0.06,
      "learning_rate": 9.44471192424972e-05,
      "loss": 2.6435,
      "theoretical_loss": 3.468084754735734,
      "tokens_seen": 1757216768
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.443909484833896e-05,
      "loss": 2.6461,
      "theoretical_loss": 3.468062634207424,
      "tokens_seen": 1757347840
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.443107045418072e-05,
      "loss": 2.655,
      "theoretical_loss": 3.4680405157908365,
      "tokens_seen": 1757478912
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.442304606002247e-05,
      "loss": 2.6677,
      "theoretical_loss": 3.4680183994856133,
      "tokens_seen": 1757609984
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.441502166586424e-05,
      "loss": 2.6228,
      "theoretical_loss": 3.4679962852913944,
      "tokens_seen": 1757741056
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.440699727170599e-05,
      "loss": 2.6249,
      "theoretical_loss": 3.4679741732078213,
      "tokens_seen": 1757872128
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.439897287754776e-05,
      "loss": 2.5379,
      "theoretical_loss": 3.467952063234536,
      "tokens_seen": 1758003200
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 970319,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6145591735839844,
      "objective/train/theoretical_loss": 3.4679299553711793,
      "objective/train/tokens_used": 128593376,
      "theoretical_loss": 3.4679299553711793,
      "tokens_seen": 1758134272
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.439094848338951e-05,
      "loss": 2.6071,
      "theoretical_loss": 3.4679299553711793,
      "tokens_seen": 1758134272
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.438292408923127e-05,
      "loss": 2.4915,
      "theoretical_loss": 3.467907849617392,
      "tokens_seen": 1758265344
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.437489969507303e-05,
      "loss": 2.4659,
      "theoretical_loss": 3.467885745972816,
      "tokens_seen": 1758396416
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.436687530091478e-05,
      "loss": 2.5481,
      "theoretical_loss": 3.467863644437093,
      "tokens_seen": 1758527488
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.435885090675655e-05,
      "loss": 2.6001,
      "theoretical_loss": 3.4678415450098643,
      "tokens_seen": 1758658560
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.43508265125983e-05,
      "loss": 2.5004,
      "theoretical_loss": 3.467819447690772,
      "tokens_seen": 1758789632
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.434280211844007e-05,
      "loss": 2.6864,
      "theoretical_loss": 3.467797352479458,
      "tokens_seen": 1758920704
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.433477772428182e-05,
      "loss": 2.6084,
      "theoretical_loss": 3.4677752593755633,
      "tokens_seen": 1759051776
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.432675333012359e-05,
      "loss": 2.553,
      "theoretical_loss": 3.4677531683787306,
      "tokens_seen": 1759182848
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.431872893596534e-05,
      "loss": 2.6378,
      "theoretical_loss": 3.4677310794886025,
      "tokens_seen": 1759313920
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.43107045418071e-05,
      "loss": 2.38,
      "theoretical_loss": 3.46770899270482,
      "tokens_seen": 1759444992
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.430268014764886e-05,
      "loss": 2.6261,
      "theoretical_loss": 3.467686908027026,
      "tokens_seen": 1759576064
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.429465575349061e-05,
      "loss": 2.6348,
      "theoretical_loss": 3.4676648254548628,
      "tokens_seen": 1759707136
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 970997,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7837326526641846,
      "objective/train/theoretical_loss": 3.467653784958281,
      "objective/train/tokens_used": 130231776,
      "theoretical_loss": 3.467653784958281,
      "tokens_seen": 1759772672
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.428663135933238e-05,
      "loss": 2.6637,
      "theoretical_loss": 3.467642744987973,
      "tokens_seen": 1759838208
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.427860696517413e-05,
      "loss": 2.6002,
      "theoretical_loss": 3.467620666625999,
      "tokens_seen": 1759969280
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.42705825710159e-05,
      "loss": 2.3965,
      "theoretical_loss": 3.4675985903685826,
      "tokens_seen": 1760100352
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.426255817685765e-05,
      "loss": 2.4942,
      "theoretical_loss": 3.4675765162153676,
      "tokens_seen": 1760231424
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.42545337826994e-05,
      "loss": 2.6354,
      "theoretical_loss": 3.4675544441659962,
      "tokens_seen": 1760362496
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.424650938854117e-05,
      "loss": 2.5839,
      "theoretical_loss": 3.467532374220112,
      "tokens_seen": 1760493568
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.423848499438293e-05,
      "loss": 2.5664,
      "theoretical_loss": 3.4675103063773567,
      "tokens_seen": 1760624640
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.423046060022469e-05,
      "loss": 2.6749,
      "theoretical_loss": 3.4674882406373744,
      "tokens_seen": 1760755712
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.422243620606645e-05,
      "loss": 2.5527,
      "theoretical_loss": 3.4674661769998076,
      "tokens_seen": 1760886784
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.421441181190821e-05,
      "loss": 2.4963,
      "theoretical_loss": 3.4674441154643,
      "tokens_seen": 1761017856
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.420638741774997e-05,
      "loss": 2.5026,
      "theoretical_loss": 3.467422056030494,
      "tokens_seen": 1761148928
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.419836302359172e-05,
      "loss": 2.546,
      "theoretical_loss": 3.467399998698034,
      "tokens_seen": 1761280000
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 972106,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7436439990997314,
      "objective/train/theoretical_loss": 3.4673779434665635,
      "objective/train/tokens_used": 131870176,
      "theoretical_loss": 3.4673779434665635,
      "tokens_seen": 1761411072
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.419033862943349e-05,
      "loss": 2.5274,
      "theoretical_loss": 3.4673779434665635,
      "tokens_seen": 1761411072
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.418231423527524e-05,
      "loss": 2.7645,
      "theoretical_loss": 3.4673558903357247,
      "tokens_seen": 1761542144
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.4174289841117e-05,
      "loss": 2.498,
      "theoretical_loss": 3.467333839305163,
      "tokens_seen": 1761673216
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.416626544695876e-05,
      "loss": 2.6431,
      "theoretical_loss": 3.467311790374521,
      "tokens_seen": 1761804288
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.415824105280053e-05,
      "loss": 2.6502,
      "theoretical_loss": 3.467289743543443,
      "tokens_seen": 1761935360
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.415021665864228e-05,
      "loss": 2.5922,
      "theoretical_loss": 3.4672676988115727,
      "tokens_seen": 1762066432
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.414219226448404e-05,
      "loss": 2.5547,
      "theoretical_loss": 3.4672456561785543,
      "tokens_seen": 1762197504
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.41341678703258e-05,
      "loss": 2.5285,
      "theoretical_loss": 3.4672236156440315,
      "tokens_seen": 1762328576
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.412614347616755e-05,
      "loss": 2.7133,
      "theoretical_loss": 3.467201577207649,
      "tokens_seen": 1762459648
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.411811908200932e-05,
      "loss": 2.516,
      "theoretical_loss": 3.4671795408690507,
      "tokens_seen": 1762590720
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.411009468785107e-05,
      "loss": 2.4072,
      "theoretical_loss": 3.467157506627881,
      "tokens_seen": 1762721792
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.410207029369284e-05,
      "loss": 2.3683,
      "theoretical_loss": 3.4671354744837837,
      "tokens_seen": 1762852864
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.409404589953459e-05,
      "loss": 2.6088,
      "theoretical_loss": 3.4671134444364045,
      "tokens_seen": 1762983936
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 972675,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4672791957855225,
      "objective/train/theoretical_loss": 3.467102430198873,
      "objective/train/tokens_used": 133508576,
      "theoretical_loss": 3.467102430198873,
      "tokens_seen": 1763049472
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.408602150537636e-05,
      "loss": 2.4902,
      "theoretical_loss": 3.4670914164853874,
      "tokens_seen": 1763115008
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.407799711121811e-05,
      "loss": 2.5745,
      "theoretical_loss": 3.4670693906303773,
      "tokens_seen": 1763246080
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.406997271705986e-05,
      "loss": 2.6415,
      "theoretical_loss": 3.4670473668710184,
      "tokens_seen": 1763377152
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.406194832290163e-05,
      "loss": 2.6236,
      "theoretical_loss": 3.467025345206956,
      "tokens_seen": 1763508224
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.405392392874338e-05,
      "loss": 2.5835,
      "theoretical_loss": 3.4670033256378354,
      "tokens_seen": 1763639296
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.404589953458515e-05,
      "loss": 2.5896,
      "theoretical_loss": 3.4669813081633007,
      "tokens_seen": 1763770368
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.40378751404269e-05,
      "loss": 2.4005,
      "theoretical_loss": 3.466959292782998,
      "tokens_seen": 1763901440
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.402985074626867e-05,
      "loss": 2.6698,
      "theoretical_loss": 3.4669372794965723,
      "tokens_seen": 1764032512
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.402182635211042e-05,
      "loss": 2.4925,
      "theoretical_loss": 3.4669152683036684,
      "tokens_seen": 1764163584
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.401380195795218e-05,
      "loss": 2.4732,
      "theoretical_loss": 3.466893259203932,
      "tokens_seen": 1764294656
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.400577756379394e-05,
      "loss": 2.4619,
      "theoretical_loss": 3.4668712521970084,
      "tokens_seen": 1764425728
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.39977531696357e-05,
      "loss": 2.4386,
      "theoretical_loss": 3.466849247282543,
      "tokens_seen": 1764556800
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 973758,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.705296039581299,
      "objective/train/theoretical_loss": 3.466827244460182,
      "objective/train/tokens_used": 135146976,
      "theoretical_loss": 3.466827244460182,
      "tokens_seen": 1764687872
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.398972877547746e-05,
      "loss": 2.6076,
      "theoretical_loss": 3.466827244460182,
      "tokens_seen": 1764687872
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.398170438131922e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.466805243729571,
      "tokens_seen": 1764818944
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.397367998716098e-05,
      "loss": 2.4123,
      "theoretical_loss": 3.4667832450903555,
      "tokens_seen": 1764950016
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.396565559300274e-05,
      "loss": 2.5775,
      "theoretical_loss": 3.4667612485421815,
      "tokens_seen": 1765081088
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.395763119884449e-05,
      "loss": 2.443,
      "theoretical_loss": 3.466739254084695,
      "tokens_seen": 1765212160
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.394960680468626e-05,
      "loss": 2.436,
      "theoretical_loss": 3.4667172617175424,
      "tokens_seen": 1765343232
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.394158241052801e-05,
      "loss": 2.5424,
      "theoretical_loss": 3.4666952714403694,
      "tokens_seen": 1765474304
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.393355801636978e-05,
      "loss": 2.4734,
      "theoretical_loss": 3.466673283252822,
      "tokens_seen": 1765605376
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.392553362221153e-05,
      "loss": 2.5327,
      "theoretical_loss": 3.4666512971545473,
      "tokens_seen": 1765736448
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.39175092280533e-05,
      "loss": 2.5953,
      "theoretical_loss": 3.4666293131451917,
      "tokens_seen": 1765867520
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.390948483389505e-05,
      "loss": 2.5874,
      "theoretical_loss": 3.4666073312244006,
      "tokens_seen": 1765998592
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.39014604397368e-05,
      "loss": 2.4698,
      "theoretical_loss": 3.4665853513918217,
      "tokens_seen": 1766129664
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.389343604557857e-05,
      "loss": 2.4093,
      "theoretical_loss": 3.4665633736471015,
      "tokens_seen": 1766260736
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 974221,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.551424026489258,
      "objective/train/theoretical_loss": 3.466552385557578,
      "objective/train/tokens_used": 136785376,
      "theoretical_loss": 3.466552385557578,
      "tokens_seen": 1766326272
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.388541165142032e-05,
      "loss": 2.5791,
      "theoretical_loss": 3.4665413979898863,
      "tokens_seen": 1766391808
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.387738725726209e-05,
      "loss": 2.4563,
      "theoretical_loss": 3.466519424419823,
      "tokens_seen": 1766522880
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.386936286310384e-05,
      "loss": 2.4674,
      "theoretical_loss": 3.4664974529365593,
      "tokens_seen": 1766653952
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.386133846894561e-05,
      "loss": 2.6283,
      "theoretical_loss": 3.4664754835397416,
      "tokens_seen": 1766785024
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.385331407478736e-05,
      "loss": 2.4859,
      "theoretical_loss": 3.4664535162290164,
      "tokens_seen": 1766916096
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.384528968062911e-05,
      "loss": 2.4277,
      "theoretical_loss": 3.466431551004032,
      "tokens_seen": 1767047168
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.383726528647088e-05,
      "loss": 2.4934,
      "theoretical_loss": 3.466409587864436,
      "tokens_seen": 1767178240
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.382924089231263e-05,
      "loss": 2.476,
      "theoretical_loss": 3.4663876268098734,
      "tokens_seen": 1767309312
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.38212164981544e-05,
      "loss": 2.6492,
      "theoretical_loss": 3.4663656678399946,
      "tokens_seen": 1767440384
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.381319210399615e-05,
      "loss": 2.4337,
      "theoretical_loss": 3.4663437109544453,
      "tokens_seen": 1767571456
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.380516770983792e-05,
      "loss": 2.5528,
      "theoretical_loss": 3.4663217561528734,
      "tokens_seen": 1767702528
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.379714331567967e-05,
      "loss": 2.5765,
      "theoretical_loss": 3.466299803434927,
      "tokens_seen": 1767833600
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 975253,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.134099006652832,
      "objective/train/theoretical_loss": 3.4662778528002534,
      "objective/train/tokens_used": 138423776,
      "theoretical_loss": 3.4662778528002534,
      "tokens_seen": 1767964672
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.378911892152144e-05,
      "loss": 2.5428,
      "theoretical_loss": 3.4662778528002534,
      "tokens_seen": 1767964672
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.37810945273632e-05,
      "loss": 2.4819,
      "theoretical_loss": 3.4662559042485013,
      "tokens_seen": 1768095744
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.377307013320495e-05,
      "loss": 2.6061,
      "theoretical_loss": 3.466233957779318,
      "tokens_seen": 1768226816
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.376504573904671e-05,
      "loss": 2.6773,
      "theoretical_loss": 3.466212013392351,
      "tokens_seen": 1768357888
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.375702134488847e-05,
      "loss": 2.5121,
      "theoretical_loss": 3.46619007108725,
      "tokens_seen": 1768488960
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.374899695073023e-05,
      "loss": 2.5764,
      "theoretical_loss": 3.466168130863662,
      "tokens_seen": 1768620032
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.374097255657199e-05,
      "loss": 2.594,
      "theoretical_loss": 3.466146192721235,
      "tokens_seen": 1768751104
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.373294816241375e-05,
      "loss": 2.6474,
      "theoretical_loss": 3.4661242566596187,
      "tokens_seen": 1768882176
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.37249237682555e-05,
      "loss": 2.6483,
      "theoretical_loss": 3.46610232267846,
      "tokens_seen": 1769013248
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.371689937409726e-05,
      "loss": 2.4411,
      "theoretical_loss": 3.4660803907774094,
      "tokens_seen": 1769144320
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.370887497993903e-05,
      "loss": 2.5386,
      "theoretical_loss": 3.4660584609561145,
      "tokens_seen": 1769275392
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.370085058578078e-05,
      "loss": 2.427,
      "theoretical_loss": 3.4660365332142233,
      "tokens_seen": 1769406464
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.369282619162255e-05,
      "loss": 2.6418,
      "theoretical_loss": 3.4660146075513856,
      "tokens_seen": 1769537536
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 975853,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6023433208465576,
      "objective/train/theoretical_loss": 3.466003645499502,
      "objective/train/tokens_used": 140062176,
      "theoretical_loss": 3.466003645499502,
      "tokens_seen": 1769603072
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.36848017974643e-05,
      "loss": 2.4855,
      "theoretical_loss": 3.46599268396725,
      "tokens_seen": 1769668608
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.367677740330607e-05,
      "loss": 2.5534,
      "theoretical_loss": 3.465970762461466,
      "tokens_seen": 1769799680
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.366875300914782e-05,
      "loss": 2.7021,
      "theoretical_loss": 3.4659488430336816,
      "tokens_seen": 1769930752
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.366072861498957e-05,
      "loss": 2.5897,
      "theoretical_loss": 3.4659269256835468,
      "tokens_seen": 1770061824
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.365270422083134e-05,
      "loss": 2.5387,
      "theoretical_loss": 3.4659050104107103,
      "tokens_seen": 1770192896
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.364467982667309e-05,
      "loss": 2.4854,
      "theoretical_loss": 3.465883097214822,
      "tokens_seen": 1770323968
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.363665543251486e-05,
      "loss": 2.6062,
      "theoretical_loss": 3.465861186095531,
      "tokens_seen": 1770455040
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.362863103835661e-05,
      "loss": 2.6616,
      "theoretical_loss": 3.4658392770524866,
      "tokens_seen": 1770586112
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.362060664419838e-05,
      "loss": 2.54,
      "theoretical_loss": 3.4658173700853387,
      "tokens_seen": 1770717184
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.361258225004013e-05,
      "loss": 2.4839,
      "theoretical_loss": 3.465795465193737,
      "tokens_seen": 1770848256
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.360455785588188e-05,
      "loss": 2.5301,
      "theoretical_loss": 3.465773562377331,
      "tokens_seen": 1770979328
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.359653346172365e-05,
      "loss": 2.4655,
      "theoretical_loss": 3.4657516616357706,
      "tokens_seen": 1771110400
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 976931,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.991319179534912,
      "objective/train/theoretical_loss": 3.4657297629687056,
      "objective/train/tokens_used": 141700576,
      "theoretical_loss": 3.4657297629687056,
      "tokens_seen": 1771241472
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.35885090675654e-05,
      "loss": 2.6133,
      "theoretical_loss": 3.4657297629687056,
      "tokens_seen": 1771241472
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.358048467340717e-05,
      "loss": 2.5932,
      "theoretical_loss": 3.4657078663757868,
      "tokens_seen": 1771372544
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.357246027924892e-05,
      "loss": 2.5936,
      "theoretical_loss": 3.4656859718566633,
      "tokens_seen": 1771503616
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.356443588509069e-05,
      "loss": 2.5908,
      "theoretical_loss": 3.4656640794109856,
      "tokens_seen": 1771634688
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.355641149093244e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.4656421890384044,
      "tokens_seen": 1771765760
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.35483870967742e-05,
      "loss": 2.4126,
      "theoretical_loss": 3.465620300738569,
      "tokens_seen": 1771896832
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.354036270261596e-05,
      "loss": 2.6821,
      "theoretical_loss": 3.465598414511131,
      "tokens_seen": 1772027904
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.353233830845772e-05,
      "loss": 2.5196,
      "theoretical_loss": 3.46557653035574,
      "tokens_seen": 1772158976
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.352431391429948e-05,
      "loss": 2.6704,
      "theoretical_loss": 3.4655546482720476,
      "tokens_seen": 1772290048
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.351628952014124e-05,
      "loss": 2.6822,
      "theoretical_loss": 3.4655327682597035,
      "tokens_seen": 1772421120
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.3508265125983e-05,
      "loss": 2.4813,
      "theoretical_loss": 3.465510890318359,
      "tokens_seen": 1772552192
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.350024073182476e-05,
      "loss": 2.5966,
      "theoretical_loss": 3.465489014447665,
      "tokens_seen": 1772683264
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.349221633766652e-05,
      "loss": 2.6041,
      "theoretical_loss": 3.465467140647272,
      "tokens_seen": 1772814336
    },
    {
      "epoch": 0.07,
      "objective/train/docs_used": 977871,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.325096368789673,
      "objective/train/theoretical_loss": 3.4654562045233295,
      "objective/train/tokens_used": 143338976,
      "theoretical_loss": 3.4654562045233295,
      "tokens_seen": 1772879872
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.348419194350828e-05,
      "loss": 2.5723,
      "theoretical_loss": 3.465445268916832,
      "tokens_seen": 1772945408
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.347616754935003e-05,
      "loss": 2.5643,
      "theoretical_loss": 3.4654233992559944,
      "tokens_seen": 1773076480
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.34681431551918e-05,
      "loss": 2.7092,
      "theoretical_loss": 3.465401531664412,
      "tokens_seen": 1773207552
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.346011876103355e-05,
      "loss": 2.5261,
      "theoretical_loss": 3.465379666141735,
      "tokens_seen": 1773338624
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.345209436687532e-05,
      "loss": 2.6486,
      "theoretical_loss": 3.4653578026876155,
      "tokens_seen": 1773469696
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.344406997271707e-05,
      "loss": 2.4702,
      "theoretical_loss": 3.4653359413017046,
      "tokens_seen": 1773600768
    },
    {
      "epoch": 0.07,
      "learning_rate": 9.343604557855882e-05,
      "loss": 2.6074,
      "theoretical_loss": 3.465314081983654,
      "tokens_seen": 1773731840
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.342802118440059e-05,
      "loss": 2.7033,
      "theoretical_loss": 3.4652922247331155,
      "tokens_seen": 1773862912
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.341999679024234e-05,
      "loss": 2.6171,
      "theoretical_loss": 3.4652703695497404,
      "tokens_seen": 1773993984
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.341197239608411e-05,
      "loss": 2.6683,
      "theoretical_loss": 3.4652485164331805,
      "tokens_seen": 1774125056
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.340394800192586e-05,
      "loss": 2.5473,
      "theoretical_loss": 3.4652266653830877,
      "tokens_seen": 1774256128
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.339592360776763e-05,
      "loss": 2.503,
      "theoretical_loss": 3.465204816399114,
      "tokens_seen": 1774387200
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 978597,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.45396089553833,
      "objective/train/theoretical_loss": 3.4651829694809115,
      "objective/train/tokens_used": 144977376,
      "theoretical_loss": 3.4651829694809115,
      "tokens_seen": 1774518272
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.338789921360938e-05,
      "loss": 2.3952,
      "theoretical_loss": 3.4651829694809115,
      "tokens_seen": 1774518272
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.337987481945113e-05,
      "loss": 2.6399,
      "theoretical_loss": 3.4651611246281324,
      "tokens_seen": 1774649344
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.33718504252929e-05,
      "loss": 2.53,
      "theoretical_loss": 3.465139281840429,
      "tokens_seen": 1774780416
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.336382603113465e-05,
      "loss": 2.4813,
      "theoretical_loss": 3.465117441117453,
      "tokens_seen": 1774911488
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.335580163697642e-05,
      "loss": 2.6267,
      "theoretical_loss": 3.4650956024588577,
      "tokens_seen": 1775042560
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.334777724281817e-05,
      "loss": 2.4972,
      "theoretical_loss": 3.465073765864295,
      "tokens_seen": 1775173632
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.333975284865993e-05,
      "loss": 2.6277,
      "theoretical_loss": 3.4650519313334174,
      "tokens_seen": 1775304704
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.33317284545017e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.465030098865877,
      "tokens_seen": 1775435776
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.332370406034345e-05,
      "loss": 2.4372,
      "theoretical_loss": 3.465008268461328,
      "tokens_seen": 1775566848
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.331567966618521e-05,
      "loss": 2.3739,
      "theoretical_loss": 3.4649864401194215,
      "tokens_seen": 1775697920
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.330765527202697e-05,
      "loss": 2.4286,
      "theoretical_loss": 3.464964613839812,
      "tokens_seen": 1775828992
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.329963087786872e-05,
      "loss": 2.6974,
      "theoretical_loss": 3.464942789622151,
      "tokens_seen": 1775960064
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.329160648371049e-05,
      "loss": 2.4616,
      "theoretical_loss": 3.4649209674660923,
      "tokens_seen": 1776091136
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 979583,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8104419708251953,
      "objective/train/theoretical_loss": 3.464910057161056,
      "objective/train/tokens_used": 146615776,
      "theoretical_loss": 3.464910057161056,
      "tokens_seen": 1776156672
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.328358208955224e-05,
      "loss": 2.6414,
      "theoretical_loss": 3.4648991473712893,
      "tokens_seen": 1776222208
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.3275557695394e-05,
      "loss": 2.524,
      "theoretical_loss": 3.464877329337395,
      "tokens_seen": 1776353280
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.326753330123576e-05,
      "loss": 2.7874,
      "theoretical_loss": 3.4648555133640624,
      "tokens_seen": 1776484352
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.325950890707753e-05,
      "loss": 2.5007,
      "theoretical_loss": 3.4648336994509448,
      "tokens_seen": 1776615424
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.325148451291928e-05,
      "loss": 2.6567,
      "theoretical_loss": 3.464811887597696,
      "tokens_seen": 1776746496
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.324346011876103e-05,
      "loss": 2.5699,
      "theoretical_loss": 3.4647900778039693,
      "tokens_seen": 1776877568
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.32354357246028e-05,
      "loss": 2.5771,
      "theoretical_loss": 3.4647682700694187,
      "tokens_seen": 1777008640
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.322741133044455e-05,
      "loss": 2.6564,
      "theoretical_loss": 3.4647464643936976,
      "tokens_seen": 1777139712
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.321938693628632e-05,
      "loss": 2.5478,
      "theoretical_loss": 3.4647246607764606,
      "tokens_seen": 1777270784
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.321136254212807e-05,
      "loss": 2.5323,
      "theoretical_loss": 3.4647028592173603,
      "tokens_seen": 1777401856
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.320333814796983e-05,
      "loss": 2.6142,
      "theoretical_loss": 3.4646810597160513,
      "tokens_seen": 1777532928
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.319531375381159e-05,
      "loss": 2.6264,
      "theoretical_loss": 3.4646592622721872,
      "tokens_seen": 1777664000
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 979848,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.569380044937134,
      "objective/train/theoretical_loss": 3.4646374668854234,
      "objective/train/tokens_used": 148254176,
      "theoretical_loss": 3.4646374668854234,
      "tokens_seen": 1777795072
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.318728935965335e-05,
      "loss": 2.6984,
      "theoretical_loss": 3.4646374668854234,
      "tokens_seen": 1777795072
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.317926496549511e-05,
      "loss": 2.4607,
      "theoretical_loss": 3.464615673555413,
      "tokens_seen": 1777926144
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.317124057133686e-05,
      "loss": 2.5602,
      "theoretical_loss": 3.4645938822818105,
      "tokens_seen": 1778057216
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.316321617717863e-05,
      "loss": 2.5766,
      "theoretical_loss": 3.4645720930642705,
      "tokens_seen": 1778188288
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.315519178302038e-05,
      "loss": 2.5064,
      "theoretical_loss": 3.4645503059024474,
      "tokens_seen": 1778319360
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.314716738886214e-05,
      "loss": 2.6483,
      "theoretical_loss": 3.4645285207959953,
      "tokens_seen": 1778450432
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.31391429947039e-05,
      "loss": 2.437,
      "theoretical_loss": 3.4645067377445695,
      "tokens_seen": 1778581504
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.313111860054566e-05,
      "loss": 2.7707,
      "theoretical_loss": 3.4644849567478246,
      "tokens_seen": 1778712576
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.312309420638742e-05,
      "loss": 2.6932,
      "theoretical_loss": 3.464463177805415,
      "tokens_seen": 1778843648
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.311506981222918e-05,
      "loss": 2.5334,
      "theoretical_loss": 3.4644414009169964,
      "tokens_seen": 1778974720
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.310704541807093e-05,
      "loss": 2.6573,
      "theoretical_loss": 3.464419626082223,
      "tokens_seen": 1779105792
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.30990210239127e-05,
      "loss": 2.5421,
      "theoretical_loss": 3.46439785330075,
      "tokens_seen": 1779236864
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.309099662975445e-05,
      "loss": 2.6231,
      "theoretical_loss": 3.464376082572233,
      "tokens_seen": 1779367936
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 981185,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7941625118255615,
      "objective/train/theoretical_loss": 3.464365197977725,
      "objective/train/tokens_used": 149892576,
      "theoretical_loss": 3.464365197977725,
      "tokens_seen": 1779433472
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.308297223559622e-05,
      "loss": 2.5164,
      "theoretical_loss": 3.464354313896327,
      "tokens_seen": 1779499008
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.307494784143797e-05,
      "loss": 2.5201,
      "theoretical_loss": 3.4643325472726865,
      "tokens_seen": 1779630080
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.306692344727974e-05,
      "loss": 2.4844,
      "theoretical_loss": 3.4643107827009674,
      "tokens_seen": 1779761152
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.305889905312149e-05,
      "loss": 2.6485,
      "theoretical_loss": 3.464289020180826,
      "tokens_seen": 1779892224
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.305087465896324e-05,
      "loss": 2.5563,
      "theoretical_loss": 3.4642672597119173,
      "tokens_seen": 1780023296
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.304285026480501e-05,
      "loss": 2.4439,
      "theoretical_loss": 3.464245501293896,
      "tokens_seen": 1780154368
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.303482587064676e-05,
      "loss": 2.5935,
      "theoretical_loss": 3.4642237449264193,
      "tokens_seen": 1780285440
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.302680147648853e-05,
      "loss": 2.6076,
      "theoretical_loss": 3.4642019906091424,
      "tokens_seen": 1780416512
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.301877708233028e-05,
      "loss": 2.6269,
      "theoretical_loss": 3.464180238341721,
      "tokens_seen": 1780547584
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.301075268817204e-05,
      "loss": 2.5585,
      "theoretical_loss": 3.464158488123811,
      "tokens_seen": 1780678656
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.30027282940138e-05,
      "loss": 2.6408,
      "theoretical_loss": 3.464136739955069,
      "tokens_seen": 1780809728
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.299470389985556e-05,
      "loss": 2.7265,
      "theoretical_loss": 3.4641149938351505,
      "tokens_seen": 1780940800
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 981883,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.705595016479492,
      "objective/train/theoretical_loss": 3.4640932497637125,
      "objective/train/tokens_used": 151530976,
      "theoretical_loss": 3.4640932497637125,
      "tokens_seen": 1781071872
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.298667950569732e-05,
      "loss": 2.6347,
      "theoretical_loss": 3.4640932497637125,
      "tokens_seen": 1781071872
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.297865511153908e-05,
      "loss": 2.4974,
      "theoretical_loss": 3.4640715077404103,
      "tokens_seen": 1781202944
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.297063071738084e-05,
      "loss": 2.5395,
      "theoretical_loss": 3.464049767764901,
      "tokens_seen": 1781334016
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.29626063232226e-05,
      "loss": 2.5233,
      "theoretical_loss": 3.4640280298368413,
      "tokens_seen": 1781465088
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.295458192906435e-05,
      "loss": 2.5728,
      "theoretical_loss": 3.4640062939558867,
      "tokens_seen": 1781596160
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.294655753490612e-05,
      "loss": 2.7589,
      "theoretical_loss": 3.4639845601216948,
      "tokens_seen": 1781727232
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.293853314074787e-05,
      "loss": 2.6934,
      "theoretical_loss": 3.4639628283339223,
      "tokens_seen": 1781858304
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.293050874658963e-05,
      "loss": 2.5922,
      "theoretical_loss": 3.463941098592225,
      "tokens_seen": 1781989376
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.292248435243139e-05,
      "loss": 2.5203,
      "theoretical_loss": 3.4639193708962615,
      "tokens_seen": 1782120448
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.291445995827315e-05,
      "loss": 2.6153,
      "theoretical_loss": 3.4638976452456873,
      "tokens_seen": 1782251520
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.290643556411491e-05,
      "loss": 2.5865,
      "theoretical_loss": 3.4638759216401596,
      "tokens_seen": 1782382592
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.289841116995666e-05,
      "loss": 2.4681,
      "theoretical_loss": 3.4638542000793358,
      "tokens_seen": 1782513664
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.289038677579843e-05,
      "loss": 2.6832,
      "theoretical_loss": 3.463832480562873,
      "tokens_seen": 1782644736
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 982960,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1652934551239014,
      "objective/train/theoretical_loss": 3.463821621571171,
      "objective/train/tokens_used": 153169376,
      "theoretical_loss": 3.463821621571171,
      "tokens_seen": 1782710272
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.288236238164018e-05,
      "loss": 2.4893,
      "theoretical_loss": 3.4638107630904296,
      "tokens_seen": 1782775808
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.287433798748195e-05,
      "loss": 2.5598,
      "theoretical_loss": 3.4637890476616615,
      "tokens_seen": 1782906880
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.28663135933237e-05,
      "loss": 2.5903,
      "theoretical_loss": 3.4637673342762265,
      "tokens_seen": 1783037952
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.285828919916547e-05,
      "loss": 2.4675,
      "theoretical_loss": 3.463745622933782,
      "tokens_seen": 1783169024
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.285026480500722e-05,
      "loss": 2.426,
      "theoretical_loss": 3.463723913633987,
      "tokens_seen": 1783300096
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.284224041084899e-05,
      "loss": 2.5681,
      "theoretical_loss": 3.4637022063764977,
      "tokens_seen": 1783431168
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.283421601669074e-05,
      "loss": 2.6727,
      "theoretical_loss": 3.4636805011609724,
      "tokens_seen": 1783562240
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.28261916225325e-05,
      "loss": 2.682,
      "theoretical_loss": 3.4636587979870685,
      "tokens_seen": 1783693312
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.281816722837426e-05,
      "loss": 2.4822,
      "theoretical_loss": 3.4636370968544448,
      "tokens_seen": 1783824384
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.281014283421601e-05,
      "loss": 2.5931,
      "theoretical_loss": 3.4636153977627586,
      "tokens_seen": 1783955456
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.280211844005778e-05,
      "loss": 2.5113,
      "theoretical_loss": 3.4635937007116686,
      "tokens_seen": 1784086528
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.279409404589953e-05,
      "loss": 2.4104,
      "theoretical_loss": 3.4635720057008332,
      "tokens_seen": 1784217600
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 984043,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.182506799697876,
      "objective/train/theoretical_loss": 3.4635503127299097,
      "objective/train/tokens_used": 154807776,
      "theoretical_loss": 3.4635503127299097,
      "tokens_seen": 1784348672
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.27860696517413e-05,
      "loss": 2.5613,
      "theoretical_loss": 3.4635503127299097,
      "tokens_seen": 1784348672
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.277804525758305e-05,
      "loss": 2.5171,
      "theoretical_loss": 3.4635286217985572,
      "tokens_seen": 1784479744
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.27700208634248e-05,
      "loss": 2.5051,
      "theoretical_loss": 3.463506932906434,
      "tokens_seen": 1784610816
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.276199646926657e-05,
      "loss": 2.6653,
      "theoretical_loss": 3.4634852460531986,
      "tokens_seen": 1784741888
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.275397207510833e-05,
      "loss": 2.6551,
      "theoretical_loss": 3.46346356123851,
      "tokens_seen": 1784872960
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.274594768095009e-05,
      "loss": 2.5599,
      "theoretical_loss": 3.463441878462026,
      "tokens_seen": 1785004032
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.273792328679185e-05,
      "loss": 2.4816,
      "theoretical_loss": 3.463420197723406,
      "tokens_seen": 1785135104
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.272989889263361e-05,
      "loss": 2.5629,
      "theoretical_loss": 3.4633985190223084,
      "tokens_seen": 1785266176
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.272187449847537e-05,
      "loss": 2.5634,
      "theoretical_loss": 3.463376842358393,
      "tokens_seen": 1785397248
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.271385010431712e-05,
      "loss": 2.6212,
      "theoretical_loss": 3.4633551677313186,
      "tokens_seen": 1785528320
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.270582571015889e-05,
      "loss": 2.5097,
      "theoretical_loss": 3.463333495140744,
      "tokens_seen": 1785659392
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.269780131600064e-05,
      "loss": 2.5634,
      "theoretical_loss": 3.463311824586328,
      "tokens_seen": 1785790464
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.26897769218424e-05,
      "loss": 2.545,
      "theoretical_loss": 3.463290156067731,
      "tokens_seen": 1785921536
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 984688,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8167052268981934,
      "objective/train/theoretical_loss": 3.4632793225717573,
      "objective/train/tokens_used": 156446176,
      "theoretical_loss": 3.4632793225717573,
      "tokens_seen": 1785987072
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.268175252768416e-05,
      "loss": 2.7015,
      "theoretical_loss": 3.463268489584611,
      "tokens_seen": 1786052608
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.267372813352592e-05,
      "loss": 2.6884,
      "theoretical_loss": 3.463246825136628,
      "tokens_seen": 1786183680
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.266570373936768e-05,
      "loss": 2.5157,
      "theoretical_loss": 3.4632251627234423,
      "tokens_seen": 1786314752
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.265767934520943e-05,
      "loss": 2.5907,
      "theoretical_loss": 3.4632035023447125,
      "tokens_seen": 1786445824
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.26496549510512e-05,
      "loss": 2.6504,
      "theoretical_loss": 3.4631818440000988,
      "tokens_seen": 1786576896
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.264163055689295e-05,
      "loss": 2.5668,
      "theoretical_loss": 3.4631601876892604,
      "tokens_seen": 1786707968
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.263360616273472e-05,
      "loss": 2.6424,
      "theoretical_loss": 3.4631385334118576,
      "tokens_seen": 1786839040
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.262558176857647e-05,
      "loss": 2.6048,
      "theoretical_loss": 3.463116881167551,
      "tokens_seen": 1786970112
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.261755737441824e-05,
      "loss": 2.4931,
      "theoretical_loss": 3.463095230955999,
      "tokens_seen": 1787101184
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.260953298025999e-05,
      "loss": 2.5467,
      "theoretical_loss": 3.463073582776863,
      "tokens_seen": 1787232256
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.260150858610174e-05,
      "loss": 2.3647,
      "theoretical_loss": 3.4630519366298023,
      "tokens_seen": 1787363328
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.259348419194351e-05,
      "loss": 2.4978,
      "theoretical_loss": 3.4630302925144782,
      "tokens_seen": 1787494400
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 985933,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.6754649877548218,
      "objective/train/theoretical_loss": 3.4630086504305497,
      "objective/train/tokens_used": 158084576,
      "theoretical_loss": 3.4630086504305497,
      "tokens_seen": 1787625472
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.258545979778526e-05,
      "loss": 2.4287,
      "theoretical_loss": 3.4630086504305497,
      "tokens_seen": 1787625472
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.257743540362703e-05,
      "loss": 2.4368,
      "theoretical_loss": 3.4629870103776783,
      "tokens_seen": 1787756544
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.256941100946878e-05,
      "loss": 2.5856,
      "theoretical_loss": 3.4629653723555247,
      "tokens_seen": 1787887616
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.256138661531055e-05,
      "loss": 2.4903,
      "theoretical_loss": 3.4629437363637483,
      "tokens_seen": 1788018688
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.25533622211523e-05,
      "loss": 2.4972,
      "theoretical_loss": 3.4629221024020107,
      "tokens_seen": 1788149760
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.254533782699407e-05,
      "loss": 2.4449,
      "theoretical_loss": 3.4629004704699726,
      "tokens_seen": 1788280832
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.253731343283582e-05,
      "loss": 2.5606,
      "theoretical_loss": 3.462878840567294,
      "tokens_seen": 1788411904
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.252928903867758e-05,
      "loss": 2.577,
      "theoretical_loss": 3.4628572126936366,
      "tokens_seen": 1788542976
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.252126464451934e-05,
      "loss": 2.5612,
      "theoretical_loss": 3.4628355868486613,
      "tokens_seen": 1788674048
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.25132402503611e-05,
      "loss": 2.4463,
      "theoretical_loss": 3.462813963032029,
      "tokens_seen": 1788805120
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.250521585620286e-05,
      "loss": 2.64,
      "theoretical_loss": 3.462792341243401,
      "tokens_seen": 1788936192
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.249719146204462e-05,
      "loss": 2.543,
      "theoretical_loss": 3.462770721482438,
      "tokens_seen": 1789067264
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.248916706788638e-05,
      "loss": 2.6057,
      "theoretical_loss": 3.462749103748802,
      "tokens_seen": 1789198336
    },
    {
      "epoch": 0.08,
      "objective/train/docs_used": 986505,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2175519466400146,
      "objective/train/theoretical_loss": 3.4627382956421258,
      "objective/train/tokens_used": 159722976,
      "theoretical_loss": 3.4627382956421258,
      "tokens_seen": 1789263872
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.248114267372814e-05,
      "loss": 2.491,
      "theoretical_loss": 3.4627274880421544,
      "tokens_seen": 1789329408
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.247311827956989e-05,
      "loss": 2.6415,
      "theoretical_loss": 3.462705874362156,
      "tokens_seen": 1789460480
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.246509388541166e-05,
      "loss": 2.3435,
      "theoretical_loss": 3.4626842627084695,
      "tokens_seen": 1789591552
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.245706949125341e-05,
      "loss": 2.4189,
      "theoretical_loss": 3.4626626530807547,
      "tokens_seen": 1789722624
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.244904509709518e-05,
      "loss": 2.4021,
      "theoretical_loss": 3.4626410454786756,
      "tokens_seen": 1789853696
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.244102070293693e-05,
      "loss": 2.4551,
      "theoretical_loss": 3.462619439901892,
      "tokens_seen": 1789984768
    },
    {
      "epoch": 0.08,
      "learning_rate": 9.24329963087787e-05,
      "loss": 2.6548,
      "theoretical_loss": 3.462597836350067,
      "tokens_seen": 1790115840
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.242497191462045e-05,
      "loss": 2.5905,
      "theoretical_loss": 3.462576234822863,
      "tokens_seen": 1790246912
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.24169475204622e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.46255463531994,
      "tokens_seen": 1790377984
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.240892312630397e-05,
      "loss": 2.4857,
      "theoretical_loss": 3.462533037840962,
      "tokens_seen": 1790509056
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.240089873214572e-05,
      "loss": 2.6329,
      "theoretical_loss": 3.4625114423855905,
      "tokens_seen": 1790640128
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.239287433798749e-05,
      "loss": 2.4891,
      "theoretical_loss": 3.4624898489534877,
      "tokens_seen": 1790771200
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 987700,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4643940925598145,
      "objective/train/theoretical_loss": 3.462468257544317,
      "objective/train/tokens_used": 161361376,
      "theoretical_loss": 3.462468257544317,
      "tokens_seen": 1790902272
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.238484994382924e-05,
      "loss": 2.3413,
      "theoretical_loss": 3.462468257544317,
      "tokens_seen": 1790902272
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.237682554967101e-05,
      "loss": 2.4009,
      "theoretical_loss": 3.462446668157739,
      "tokens_seen": 1791033344
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.236880115551276e-05,
      "loss": 2.665,
      "theoretical_loss": 3.462425080793418,
      "tokens_seen": 1791164416
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.236077676135451e-05,
      "loss": 2.5981,
      "theoretical_loss": 3.4624034954510154,
      "tokens_seen": 1791295488
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.235275236719628e-05,
      "loss": 2.6992,
      "theoretical_loss": 3.462381912130194,
      "tokens_seen": 1791426560
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.234472797303803e-05,
      "loss": 2.6392,
      "theoretical_loss": 3.462360330830618,
      "tokens_seen": 1791557632
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.23367035788798e-05,
      "loss": 2.5319,
      "theoretical_loss": 3.4623387515519486,
      "tokens_seen": 1791688704
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.232867918472155e-05,
      "loss": 2.487,
      "theoretical_loss": 3.462317174293849,
      "tokens_seen": 1791819776
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.232065479056332e-05,
      "loss": 2.4134,
      "theoretical_loss": 3.462295599055983,
      "tokens_seen": 1791950848
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.231263039640507e-05,
      "loss": 2.387,
      "theoretical_loss": 3.4622740258380134,
      "tokens_seen": 1792081920
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.230460600224683e-05,
      "loss": 2.5494,
      "theoretical_loss": 3.4622524546396027,
      "tokens_seen": 1792212992
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.229658160808859e-05,
      "loss": 2.4921,
      "theoretical_loss": 3.4622308854604156,
      "tokens_seen": 1792344064
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.228855721393035e-05,
      "loss": 2.4341,
      "theoretical_loss": 3.462209318300114,
      "tokens_seen": 1792475136
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 988349,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3251655101776123,
      "objective/train/theoretical_loss": 3.46219853547694,
      "objective/train/tokens_used": 162999776,
      "theoretical_loss": 3.46219853547694,
      "tokens_seen": 1792540672
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.228053281977211e-05,
      "loss": 2.5972,
      "theoretical_loss": 3.4621877531583616,
      "tokens_seen": 1792606208
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.227250842561387e-05,
      "loss": 2.5254,
      "theoretical_loss": 3.4621661900348224,
      "tokens_seen": 1792737280
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.226448403145563e-05,
      "loss": 2.4832,
      "theoretical_loss": 3.46214462892916,
      "tokens_seen": 1792868352
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.225645963729739e-05,
      "loss": 2.4044,
      "theoretical_loss": 3.462123069841038,
      "tokens_seen": 1792999424
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.224843524313914e-05,
      "loss": 2.6646,
      "theoretical_loss": 3.4621015127701193,
      "tokens_seen": 1793130496
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.22404108489809e-05,
      "loss": 2.4458,
      "theoretical_loss": 3.462079957716069,
      "tokens_seen": 1793261568
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.223238645482266e-05,
      "loss": 2.4208,
      "theoretical_loss": 3.46205840467855,
      "tokens_seen": 1793392640
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.222436206066443e-05,
      "loss": 2.5817,
      "theoretical_loss": 3.462036853657227,
      "tokens_seen": 1793523712
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.221633766650618e-05,
      "loss": 2.5792,
      "theoretical_loss": 3.4620153046517634,
      "tokens_seen": 1793654784
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.220831327234795e-05,
      "loss": 2.5938,
      "theoretical_loss": 3.461993757661824,
      "tokens_seen": 1793785856
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.22002888781897e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.4619722126870727,
      "tokens_seen": 1793916928
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.219226448403146e-05,
      "loss": 2.5115,
      "theoretical_loss": 3.461950669727174,
      "tokens_seen": 1794048000
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 989683,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.0829696655273438,
      "objective/train/theoretical_loss": 3.461929128781792,
      "objective/train/tokens_used": 164638176,
      "theoretical_loss": 3.461929128781792,
      "tokens_seen": 1794179072
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.218424008987322e-05,
      "loss": 2.343,
      "theoretical_loss": 3.461929128781792,
      "tokens_seen": 1794179072
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.217621569571497e-05,
      "loss": 2.5634,
      "theoretical_loss": 3.4619075898505915,
      "tokens_seen": 1794310144
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.216819130155674e-05,
      "loss": 2.5899,
      "theoretical_loss": 3.4618860529332363,
      "tokens_seen": 1794441216
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.216016690739849e-05,
      "loss": 2.4196,
      "theoretical_loss": 3.461864518029392,
      "tokens_seen": 1794572288
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.215214251324026e-05,
      "loss": 2.5227,
      "theoretical_loss": 3.4618429851387233,
      "tokens_seen": 1794703360
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.214411811908201e-05,
      "loss": 2.442,
      "theoretical_loss": 3.4618214542608943,
      "tokens_seen": 1794834432
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.213609372492378e-05,
      "loss": 2.5729,
      "theoretical_loss": 3.46179992539557,
      "tokens_seen": 1794965504
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.212806933076553e-05,
      "loss": 2.5655,
      "theoretical_loss": 3.4617783985424158,
      "tokens_seen": 1795096576
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.212004493660728e-05,
      "loss": 2.4973,
      "theoretical_loss": 3.461756873701096,
      "tokens_seen": 1795227648
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.211202054244905e-05,
      "loss": 2.7353,
      "theoretical_loss": 3.461735350871277,
      "tokens_seen": 1795358720
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.21039961482908e-05,
      "loss": 2.5919,
      "theoretical_loss": 3.4617138300526222,
      "tokens_seen": 1795489792
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.209597175413257e-05,
      "loss": 2.4516,
      "theoretical_loss": 3.461692311244798,
      "tokens_seen": 1795620864
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.208794735997432e-05,
      "loss": 2.4701,
      "theoretical_loss": 3.4616707944474703,
      "tokens_seen": 1795751936
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 990185,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.836434841156006,
      "objective/train/theoretical_loss": 3.4616600368026376,
      "objective/train/tokens_used": 166276576,
      "theoretical_loss": 3.4616600368026376,
      "tokens_seen": 1795817472
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.207992296581609e-05,
      "loss": 2.5007,
      "theoretical_loss": 3.4616492796603033,
      "tokens_seen": 1795883008
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.207189857165784e-05,
      "loss": 2.6287,
      "theoretical_loss": 3.4616277668829634,
      "tokens_seen": 1796014080
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.20638741774996e-05,
      "loss": 2.6319,
      "theoretical_loss": 3.461606256115116,
      "tokens_seen": 1796145152
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.205584978334136e-05,
      "loss": 2.3697,
      "theoretical_loss": 3.4615847473564263,
      "tokens_seen": 1796276224
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.204782538918312e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.4615632406065604,
      "tokens_seen": 1796407296
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.203980099502488e-05,
      "loss": 2.4935,
      "theoretical_loss": 3.4615417358651843,
      "tokens_seen": 1796538368
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.203177660086664e-05,
      "loss": 2.6099,
      "theoretical_loss": 3.4615202331319637,
      "tokens_seen": 1796669440
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.20237522067084e-05,
      "loss": 2.6371,
      "theoretical_loss": 3.4614987324065645,
      "tokens_seen": 1796800512
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.201572781255016e-05,
      "loss": 2.464,
      "theoretical_loss": 3.461477233688653,
      "tokens_seen": 1796931584
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.200770341839191e-05,
      "loss": 2.609,
      "theoretical_loss": 3.4614557369778947,
      "tokens_seen": 1797062656
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.199967902423368e-05,
      "loss": 2.5767,
      "theoretical_loss": 3.461434242273957,
      "tokens_seen": 1797193728
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.199165463007543e-05,
      "loss": 2.641,
      "theoretical_loss": 3.461412749576505,
      "tokens_seen": 1797324800
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 991515,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3866214752197266,
      "objective/train/theoretical_loss": 3.4613912588852065,
      "objective/train/tokens_used": 167914976,
      "theoretical_loss": 3.4613912588852065,
      "tokens_seen": 1797455872
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.19836302359172e-05,
      "loss": 2.6222,
      "theoretical_loss": 3.4613912588852065,
      "tokens_seen": 1797455872
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.197560584175895e-05,
      "loss": 2.6715,
      "theoretical_loss": 3.4613697701997266,
      "tokens_seen": 1797586944
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.196758144760072e-05,
      "loss": 2.5473,
      "theoretical_loss": 3.4613482835197322,
      "tokens_seen": 1797718016
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.195955705344247e-05,
      "loss": 2.5753,
      "theoretical_loss": 3.4613267988448913,
      "tokens_seen": 1797849088
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.195153265928422e-05,
      "loss": 2.6014,
      "theoretical_loss": 3.4613053161748684,
      "tokens_seen": 1797980160
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.194350826512599e-05,
      "loss": 2.5159,
      "theoretical_loss": 3.461283835509332,
      "tokens_seen": 1798111232
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.193548387096774e-05,
      "loss": 2.5353,
      "theoretical_loss": 3.4612623568479477,
      "tokens_seen": 1798242304
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.192745947680951e-05,
      "loss": 2.5962,
      "theoretical_loss": 3.4612408801903833,
      "tokens_seen": 1798373376
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.191943508265126e-05,
      "loss": 2.5236,
      "theoretical_loss": 3.4612194055363057,
      "tokens_seen": 1798504448
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.191141068849303e-05,
      "loss": 2.5268,
      "theoretical_loss": 3.461197932885382,
      "tokens_seen": 1798635520
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.190338629433478e-05,
      "loss": 2.5341,
      "theoretical_loss": 3.4611764622372796,
      "tokens_seen": 1798766592
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.189536190017655e-05,
      "loss": 2.538,
      "theoretical_loss": 3.4611549935916646,
      "tokens_seen": 1798897664
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.18873375060183e-05,
      "loss": 2.6186,
      "theoretical_loss": 3.461133526948206,
      "tokens_seen": 1799028736
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 992149,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.483299732208252,
      "objective/train/theoretical_loss": 3.4611227943771814,
      "objective/train/tokens_used": 169553376,
      "theoretical_loss": 3.4611227943771814,
      "tokens_seen": 1799094272
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.187931311186005e-05,
      "loss": 2.565,
      "theoretical_loss": 3.4611120623065705,
      "tokens_seen": 1799159808
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.187128871770182e-05,
      "loss": 2.5304,
      "theoretical_loss": 3.4610905996664254,
      "tokens_seen": 1799290880
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.186326432354357e-05,
      "loss": 2.5325,
      "theoretical_loss": 3.461069139027438,
      "tokens_seen": 1799421952
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.185523992938534e-05,
      "loss": 2.6046,
      "theoretical_loss": 3.461047680389277,
      "tokens_seen": 1799553024
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.18472155352271e-05,
      "loss": 2.5433,
      "theoretical_loss": 3.4610262237516096,
      "tokens_seen": 1799684096
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.183919114106886e-05,
      "loss": 2.5845,
      "theoretical_loss": 3.461004769114103,
      "tokens_seen": 1799815168
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.183116674691061e-05,
      "loss": 2.5856,
      "theoretical_loss": 3.460983316476426,
      "tokens_seen": 1799946240
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.182314235275237e-05,
      "loss": 2.6222,
      "theoretical_loss": 3.4609618658382466,
      "tokens_seen": 1800077312
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.181511795859413e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.4609404171992324,
      "tokens_seen": 1800208384
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.180709356443589e-05,
      "loss": 2.5571,
      "theoretical_loss": 3.4609189705590513,
      "tokens_seen": 1800339456
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.179906917027765e-05,
      "loss": 2.3876,
      "theoretical_loss": 3.460897525917373,
      "tokens_seen": 1800470528
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.17910447761194e-05,
      "loss": 2.5827,
      "theoretical_loss": 3.4608760832738636,
      "tokens_seen": 1800601600
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 993279,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.028233766555786,
      "objective/train/theoretical_loss": 3.4608546426281928,
      "objective/train/tokens_used": 171191776,
      "theoretical_loss": 3.4608546426281928,
      "tokens_seen": 1800732672
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.178302038196117e-05,
      "loss": 2.5501,
      "theoretical_loss": 3.4608546426281928,
      "tokens_seen": 1800732672
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.177499598780293e-05,
      "loss": 2.4478,
      "theoretical_loss": 3.4608332039800294,
      "tokens_seen": 1800863744
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.176697159364468e-05,
      "loss": 2.4669,
      "theoretical_loss": 3.460811767329041,
      "tokens_seen": 1800994816
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.175894719948645e-05,
      "loss": 2.6295,
      "theoretical_loss": 3.4607903326748968,
      "tokens_seen": 1801125888
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.17509228053282e-05,
      "loss": 2.6089,
      "theoretical_loss": 3.460768900017265,
      "tokens_seen": 1801256960
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.174289841116997e-05,
      "loss": 2.6017,
      "theoretical_loss": 3.460747469355815,
      "tokens_seen": 1801388032
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.173487401701172e-05,
      "loss": 2.6287,
      "theoretical_loss": 3.460726040690215,
      "tokens_seen": 1801519104
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.172684962285349e-05,
      "loss": 2.5425,
      "theoretical_loss": 3.460704614020135,
      "tokens_seen": 1801650176
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.171882522869524e-05,
      "loss": 2.4997,
      "theoretical_loss": 3.4606831893452425,
      "tokens_seen": 1801781248
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.171080083453699e-05,
      "loss": 2.5585,
      "theoretical_loss": 3.4606617666652078,
      "tokens_seen": 1801912320
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.170277644037876e-05,
      "loss": 2.5326,
      "theoretical_loss": 3.4606403459796997,
      "tokens_seen": 1802043392
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.169475204622051e-05,
      "loss": 2.6015,
      "theoretical_loss": 3.4606189272883876,
      "tokens_seen": 1802174464
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.168672765206228e-05,
      "loss": 2.5776,
      "theoretical_loss": 3.46059751059094,
      "tokens_seen": 1802305536
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 993741,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3329663276672363,
      "objective/train/theoretical_loss": 3.4605868029898126,
      "objective/train/tokens_used": 172830176,
      "theoretical_loss": 3.4605868029898126,
      "tokens_seen": 1802371072
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.167870325790403e-05,
      "loss": 2.5359,
      "theoretical_loss": 3.4605760958870277,
      "tokens_seen": 1802436608
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.16706788637458e-05,
      "loss": 2.6917,
      "theoretical_loss": 3.4605546831763188,
      "tokens_seen": 1802567680
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.166265446958755e-05,
      "loss": 2.5287,
      "theoretical_loss": 3.460533272458484,
      "tokens_seen": 1802698752
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.16546300754293e-05,
      "loss": 2.6202,
      "theoretical_loss": 3.460511863733192,
      "tokens_seen": 1802829824
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.164660568127107e-05,
      "loss": 2.6136,
      "theoretical_loss": 3.4604904570001134,
      "tokens_seen": 1802960896
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.163858128711282e-05,
      "loss": 2.7156,
      "theoretical_loss": 3.460469052258917,
      "tokens_seen": 1803091968
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.163055689295459e-05,
      "loss": 2.6449,
      "theoretical_loss": 3.460447649509274,
      "tokens_seen": 1803223040
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.162253249879634e-05,
      "loss": 2.6126,
      "theoretical_loss": 3.4604262487508533,
      "tokens_seen": 1803354112
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.161450810463811e-05,
      "loss": 2.6197,
      "theoretical_loss": 3.460404849983325,
      "tokens_seen": 1803485184
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.160648371047986e-05,
      "loss": 2.4627,
      "theoretical_loss": 3.4603834532063598,
      "tokens_seen": 1803616256
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.159845931632163e-05,
      "loss": 2.5741,
      "theoretical_loss": 3.4603620584196273,
      "tokens_seen": 1803747328
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.159043492216338e-05,
      "loss": 2.4163,
      "theoretical_loss": 3.460340665622798,
      "tokens_seen": 1803878400
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 994982,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4623970985412598,
      "objective/train/theoretical_loss": 3.460319274815543,
      "objective/train/tokens_used": 174468576,
      "theoretical_loss": 3.460319274815543,
      "tokens_seen": 1804009472
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.158241052800514e-05,
      "loss": 2.5558,
      "theoretical_loss": 3.460319274815543,
      "tokens_seen": 1804009472
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.15743861338469e-05,
      "loss": 2.6833,
      "theoretical_loss": 3.460297885997531,
      "tokens_seen": 1804140544
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.156636173968866e-05,
      "loss": 2.5265,
      "theoretical_loss": 3.4602764991684345,
      "tokens_seen": 1804271616
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.155833734553042e-05,
      "loss": 2.3257,
      "theoretical_loss": 3.4602551143279228,
      "tokens_seen": 1804402688
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.155031295137218e-05,
      "loss": 2.5691,
      "theoretical_loss": 3.4602337314756664,
      "tokens_seen": 1804533760
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.154228855721394e-05,
      "loss": 2.5498,
      "theoretical_loss": 3.4602123506113376,
      "tokens_seen": 1804664832
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.15342641630557e-05,
      "loss": 2.545,
      "theoretical_loss": 3.460190971734605,
      "tokens_seen": 1804795904
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.152623976889745e-05,
      "loss": 2.6298,
      "theoretical_loss": 3.4601695948451416,
      "tokens_seen": 1804926976
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.151821537473922e-05,
      "loss": 2.6275,
      "theoretical_loss": 3.460148219942617,
      "tokens_seen": 1805058048
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.151019098058097e-05,
      "loss": 2.5375,
      "theoretical_loss": 3.4601268470267024,
      "tokens_seen": 1805189120
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.150216658642274e-05,
      "loss": 2.4901,
      "theoretical_loss": 3.4601054760970698,
      "tokens_seen": 1805320192
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.149414219226449e-05,
      "loss": 2.6574,
      "theoretical_loss": 3.4600841071533894,
      "tokens_seen": 1805451264
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.148611779810626e-05,
      "loss": 2.508,
      "theoretical_loss": 3.460062740195333,
      "tokens_seen": 1805582336
    },
    {
      "epoch": 0.09,
      "objective/train/docs_used": 995914,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.526226758956909,
      "objective/train/theoretical_loss": 3.4600520574608113,
      "objective/train/tokens_used": 176106976,
      "theoretical_loss": 3.4600520574608113,
      "tokens_seen": 1805647872
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.147809340394801e-05,
      "loss": 2.4625,
      "theoretical_loss": 3.4600413752225725,
      "tokens_seen": 1805713408
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.147006900978976e-05,
      "loss": 2.4556,
      "theoretical_loss": 3.460020012234778,
      "tokens_seen": 1805844480
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.146204461563153e-05,
      "loss": 2.56,
      "theoretical_loss": 3.459998651231622,
      "tokens_seen": 1805975552
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.145402022147328e-05,
      "loss": 2.5144,
      "theoretical_loss": 3.4599772922127765,
      "tokens_seen": 1806106624
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.144599582731505e-05,
      "loss": 2.5989,
      "theoretical_loss": 3.459955935177912,
      "tokens_seen": 1806237696
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.14379714331568e-05,
      "loss": 2.6055,
      "theoretical_loss": 3.459934580126701,
      "tokens_seen": 1806368768
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.142994703899857e-05,
      "loss": 2.5955,
      "theoretical_loss": 3.4599132270588155,
      "tokens_seen": 1806499840
    },
    {
      "epoch": 0.09,
      "learning_rate": 9.142192264484032e-05,
      "loss": 2.6991,
      "theoretical_loss": 3.459891875973927,
      "tokens_seen": 1806630912
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.141389825068207e-05,
      "loss": 2.6054,
      "theoretical_loss": 3.4598705268717076,
      "tokens_seen": 1806761984
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.140587385652384e-05,
      "loss": 2.6119,
      "theoretical_loss": 3.4598491797518287,
      "tokens_seen": 1806893056
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.13978494623656e-05,
      "loss": 2.6178,
      "theoretical_loss": 3.459827834613964,
      "tokens_seen": 1807024128
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.138982506820736e-05,
      "loss": 2.7865,
      "theoretical_loss": 3.4598064914577846,
      "tokens_seen": 1807155200
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 996601,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5091497898101807,
      "objective/train/theoretical_loss": 3.459785150282963,
      "objective/train/tokens_used": 177745376,
      "theoretical_loss": 3.459785150282963,
      "tokens_seen": 1807286272
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.138180067404911e-05,
      "loss": 2.6171,
      "theoretical_loss": 3.459785150282963,
      "tokens_seen": 1807286272
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.137377627989088e-05,
      "loss": 2.4956,
      "theoretical_loss": 3.4597638110891715,
      "tokens_seen": 1807417344
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.136575188573263e-05,
      "loss": 2.5369,
      "theoretical_loss": 3.459742473876083,
      "tokens_seen": 1807548416
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.135772749157439e-05,
      "loss": 2.5218,
      "theoretical_loss": 3.4597211386433697,
      "tokens_seen": 1807679488
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.134970309741615e-05,
      "loss": 2.54,
      "theoretical_loss": 3.459699805390705,
      "tokens_seen": 1807810560
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.13416787032579e-05,
      "loss": 2.5169,
      "theoretical_loss": 3.45967847411776,
      "tokens_seen": 1807941632
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.133365430909967e-05,
      "loss": 2.6301,
      "theoretical_loss": 3.4596571448242086,
      "tokens_seen": 1808072704
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.132562991494143e-05,
      "loss": 2.6115,
      "theoretical_loss": 3.4596358175097235,
      "tokens_seen": 1808203776
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.131760552078319e-05,
      "loss": 2.5861,
      "theoretical_loss": 3.4596144921739773,
      "tokens_seen": 1808334848
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.130958112662495e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.459593168816644,
      "tokens_seen": 1808465920
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.13015567324667e-05,
      "loss": 2.571,
      "theoretical_loss": 3.4595718474373953,
      "tokens_seen": 1808596992
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.129353233830847e-05,
      "loss": 2.6187,
      "theoretical_loss": 3.459550528035905,
      "tokens_seen": 1808728064
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.128550794415022e-05,
      "loss": 2.5396,
      "theoretical_loss": 3.459529210611847,
      "tokens_seen": 1808859136
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 997711,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7451651096343994,
      "objective/train/theoretical_loss": 3.4595185526412524,
      "objective/train/tokens_used": 179383776,
      "theoretical_loss": 3.4595185526412524,
      "tokens_seen": 1808924672
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.127748354999199e-05,
      "loss": 2.5304,
      "theoretical_loss": 3.4595078951648937,
      "tokens_seen": 1808990208
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.126945915583374e-05,
      "loss": 2.5649,
      "theoretical_loss": 3.4594865816947182,
      "tokens_seen": 1809121280
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.12614347616755e-05,
      "loss": 2.6362,
      "theoretical_loss": 3.4594652702009956,
      "tokens_seen": 1809252352
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.125341036751726e-05,
      "loss": 2.6765,
      "theoretical_loss": 3.4594439606833975,
      "tokens_seen": 1809383424
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.124538597335903e-05,
      "loss": 2.5887,
      "theoretical_loss": 3.4594226531415986,
      "tokens_seen": 1809514496
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.123736157920078e-05,
      "loss": 2.5145,
      "theoretical_loss": 3.459401347575273,
      "tokens_seen": 1809645568
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.122933718504253e-05,
      "loss": 2.6369,
      "theoretical_loss": 3.4593800439840936,
      "tokens_seen": 1809776640
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.12213127908843e-05,
      "loss": 2.6258,
      "theoretical_loss": 3.4593587423677348,
      "tokens_seen": 1809907712
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.121328839672605e-05,
      "loss": 2.3457,
      "theoretical_loss": 3.45933744272587,
      "tokens_seen": 1810038784
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.120526400256782e-05,
      "loss": 2.4792,
      "theoretical_loss": 3.4593161450581738,
      "tokens_seen": 1810169856
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.119723960840957e-05,
      "loss": 2.5877,
      "theoretical_loss": 3.45929484936432,
      "tokens_seen": 1810300928
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.118921521425134e-05,
      "loss": 2.4965,
      "theoretical_loss": 3.459273555643983,
      "tokens_seen": 1810432000
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 998245,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.329852342605591,
      "objective/train/theoretical_loss": 3.459252263896837,
      "objective/train/tokens_used": 181022176,
      "theoretical_loss": 3.459252263896837,
      "tokens_seen": 1810563072
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.118119082009309e-05,
      "loss": 2.5769,
      "theoretical_loss": 3.459252263896837,
      "tokens_seen": 1810563072
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.117316642593484e-05,
      "loss": 2.6172,
      "theoretical_loss": 3.4592309741225558,
      "tokens_seen": 1810694144
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.116514203177661e-05,
      "loss": 2.6355,
      "theoretical_loss": 3.4592096863208144,
      "tokens_seen": 1810825216
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.115711763761836e-05,
      "loss": 2.6663,
      "theoretical_loss": 3.4591884004912874,
      "tokens_seen": 1810956288
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.114909324346013e-05,
      "loss": 2.4647,
      "theoretical_loss": 3.4591671166336484,
      "tokens_seen": 1811087360
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.114106884930188e-05,
      "loss": 2.5586,
      "theoretical_loss": 3.4591458347475736,
      "tokens_seen": 1811218432
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.113304445514365e-05,
      "loss": 2.6566,
      "theoretical_loss": 3.4591245548327363,
      "tokens_seen": 1811349504
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.11250200609854e-05,
      "loss": 2.4653,
      "theoretical_loss": 3.4591032768888117,
      "tokens_seen": 1811480576
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.111699566682716e-05,
      "loss": 2.5683,
      "theoretical_loss": 3.459082000915475,
      "tokens_seen": 1811611648
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.110897127266892e-05,
      "loss": 2.5607,
      "theoretical_loss": 3.4590607269124005,
      "tokens_seen": 1811742720
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.110094687851068e-05,
      "loss": 2.577,
      "theoretical_loss": 3.4590394548792642,
      "tokens_seen": 1811873792
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.109292248435244e-05,
      "loss": 2.5945,
      "theoretical_loss": 3.459018184815741,
      "tokens_seen": 1812004864
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.10848980901942e-05,
      "loss": 2.5688,
      "theoretical_loss": 3.458996916721505,
      "tokens_seen": 1812135936
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 999249,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.812991142272949,
      "objective/train/theoretical_loss": 3.4589862834127683,
      "objective/train/tokens_used": 182660576,
      "theoretical_loss": 3.4589862834127683,
      "tokens_seen": 1812201472
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.107687369603596e-05,
      "loss": 2.5249,
      "theoretical_loss": 3.458975650596232,
      "tokens_seen": 1812267008
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.106884930187772e-05,
      "loss": 2.6084,
      "theoretical_loss": 3.458954386439598,
      "tokens_seen": 1812398080
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.106082490771947e-05,
      "loss": 2.5619,
      "theoretical_loss": 3.4589331242512777,
      "tokens_seen": 1812529152
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.105280051356124e-05,
      "loss": 2.4964,
      "theoretical_loss": 3.458911864030947,
      "tokens_seen": 1812660224
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.104477611940299e-05,
      "loss": 2.6568,
      "theoretical_loss": 3.458890605778281,
      "tokens_seen": 1812791296
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.103675172524476e-05,
      "loss": 2.5938,
      "theoretical_loss": 3.4588693494929563,
      "tokens_seen": 1812922368
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.102872733108651e-05,
      "loss": 2.7681,
      "theoretical_loss": 3.4588480951746474,
      "tokens_seen": 1813053440
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.102070293692828e-05,
      "loss": 2.5224,
      "theoretical_loss": 3.4588268428230307,
      "tokens_seen": 1813184512
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.101267854277003e-05,
      "loss": 2.6445,
      "theoretical_loss": 3.458805592437782,
      "tokens_seen": 1813315584
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.100465414861178e-05,
      "loss": 2.6035,
      "theoretical_loss": 3.458784344018577,
      "tokens_seen": 1813446656
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.099662975445355e-05,
      "loss": 2.6548,
      "theoretical_loss": 3.4587630975650923,
      "tokens_seen": 1813577728
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.09886053602953e-05,
      "loss": 2.4354,
      "theoretical_loss": 3.4587418530770035,
      "tokens_seen": 1813708800
    },
    {
      "debugging/Self-BLEU-5": 0.3156689391415616,
      "debugging/distinct-1-grams": 0.796493726160272,
      "debugging/distinct-2-grams": 0.9767311252314004,
      "debugging/entropy-1-grams": 5.133892638116764,
      "debugging/entropy-2-grams": 5.762233051450052,
      "debugging/length": 475.0,
      "debugging/num_segments": 5,
      "debugging/score": 0.006195986574498804,
      "debugging/score_std": 0.00502469901082792,
      "epoch": 0.1,
      "objective/train/docs_used": 999799,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.311352252960205,
      "objective/train/theoretical_loss": 3.458720610553987,
      "objective/train/tokens_used": 184298976,
      "theoretical_loss": 3.458720610553987,
      "tokens_seen": 1813839872
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.098058096613707e-05,
      "loss": 2.6938,
      "theoretical_loss": 3.458720610553987,
      "tokens_seen": 1813839872
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.097255657197882e-05,
      "loss": 2.2563,
      "theoretical_loss": 3.4586993699957187,
      "tokens_seen": 1813970944
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.096453217782059e-05,
      "loss": 2.4608,
      "theoretical_loss": 3.458678131401876,
      "tokens_seen": 1814102016
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.095650778366234e-05,
      "loss": 2.5433,
      "theoretical_loss": 3.4586568947721337,
      "tokens_seen": 1814233088
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.094848338950411e-05,
      "loss": 2.5471,
      "theoretical_loss": 3.4586356601061694,
      "tokens_seen": 1814364160
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.094045899534586e-05,
      "loss": 2.4761,
      "theoretical_loss": 3.4586144274036594,
      "tokens_seen": 1814495232
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.093243460118761e-05,
      "loss": 2.4453,
      "theoretical_loss": 3.4585931966642804,
      "tokens_seen": 1814626304
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.092441020702938e-05,
      "loss": 2.5264,
      "theoretical_loss": 3.4585719678877087,
      "tokens_seen": 1814757376
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.091638581287113e-05,
      "loss": 2.4763,
      "theoretical_loss": 3.458550741073622,
      "tokens_seen": 1814888448
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.09083614187129e-05,
      "loss": 2.5676,
      "theoretical_loss": 3.458529516221696,
      "tokens_seen": 1815019520
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.090033702455465e-05,
      "loss": 2.6084,
      "theoretical_loss": 3.458508293331609,
      "tokens_seen": 1815150592
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.089231263039642e-05,
      "loss": 2.3971,
      "theoretical_loss": 3.4584870724030363,
      "tokens_seen": 1815281664
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.088428823623817e-05,
      "loss": 2.4878,
      "theoretical_loss": 3.4584658534356567,
      "tokens_seen": 1815412736
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 1000512,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2753448486328125,
      "objective/train/theoretical_loss": 3.458455244687313,
      "objective/train/tokens_used": 185937376,
      "theoretical_loss": 3.458455244687313,
      "tokens_seen": 1815478272
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.087626384207993e-05,
      "loss": 2.5719,
      "theoretical_loss": 3.458444636429146,
      "tokens_seen": 1815543808
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.08682394479217e-05,
      "loss": 2.5104,
      "theoretical_loss": 3.4584234213831824,
      "tokens_seen": 1815674880
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.086021505376345e-05,
      "loss": 2.5449,
      "theoretical_loss": 3.458402208297443,
      "tokens_seen": 1815805952
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.085219065960521e-05,
      "loss": 2.3611,
      "theoretical_loss": 3.458380997171605,
      "tokens_seen": 1815937024
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.084416626544697e-05,
      "loss": 2.706,
      "theoretical_loss": 3.458359788005346,
      "tokens_seen": 1816068096
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.083614187128873e-05,
      "loss": 2.609,
      "theoretical_loss": 3.458338580798343,
      "tokens_seen": 1816199168
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.082811747713049e-05,
      "loss": 2.4692,
      "theoretical_loss": 3.458317375550275,
      "tokens_seen": 1816330240
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.082009308297224e-05,
      "loss": 2.5132,
      "theoretical_loss": 3.4582961722608188,
      "tokens_seen": 1816461312
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.0812068688814e-05,
      "loss": 2.6601,
      "theoretical_loss": 3.4582749709296516,
      "tokens_seen": 1816592384
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.080404429465576e-05,
      "loss": 2.4721,
      "theoretical_loss": 3.4582537715564525,
      "tokens_seen": 1816723456
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.079601990049753e-05,
      "loss": 2.5989,
      "theoretical_loss": 3.458232574140899,
      "tokens_seen": 1816854528
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.078799550633928e-05,
      "loss": 2.6828,
      "theoretical_loss": 3.4582113786826683,
      "tokens_seen": 1816985600
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 1001919,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5981926918029785,
      "objective/train/theoretical_loss": 3.4581901851814396,
      "objective/train/tokens_used": 187575776,
      "theoretical_loss": 3.4581901851814396,
      "tokens_seen": 1817116672
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.077997111218105e-05,
      "loss": 2.5979,
      "theoretical_loss": 3.4581901851814396,
      "tokens_seen": 1817116672
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.07719467180228e-05,
      "loss": 2.5523,
      "theoretical_loss": 3.4581689936368907,
      "tokens_seen": 1817247744
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.076392232386455e-05,
      "loss": 2.5556,
      "theoretical_loss": 3.458147804048699,
      "tokens_seen": 1817378816
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.075589792970632e-05,
      "loss": 2.611,
      "theoretical_loss": 3.4581266164165445,
      "tokens_seen": 1817509888
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.074787353554807e-05,
      "loss": 2.4397,
      "theoretical_loss": 3.4581054307401047,
      "tokens_seen": 1817640960
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.073984914138984e-05,
      "loss": 2.5673,
      "theoretical_loss": 3.458084247019057,
      "tokens_seen": 1817772032
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.073182474723159e-05,
      "loss": 2.7918,
      "theoretical_loss": 3.458063065253082,
      "tokens_seen": 1817903104
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.072380035307334e-05,
      "loss": 2.456,
      "theoretical_loss": 3.4580418854418573,
      "tokens_seen": 1818034176
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.071577595891511e-05,
      "loss": 2.5651,
      "theoretical_loss": 3.458020707585061,
      "tokens_seen": 1818165248
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.070775156475686e-05,
      "loss": 2.4939,
      "theoretical_loss": 3.457999531682373,
      "tokens_seen": 1818296320
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.069972717059863e-05,
      "loss": 2.6449,
      "theoretical_loss": 3.457978357733471,
      "tokens_seen": 1818427392
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.069170277644038e-05,
      "loss": 2.6295,
      "theoretical_loss": 3.457957185738035,
      "tokens_seen": 1818558464
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.068367838228215e-05,
      "loss": 2.6178,
      "theoretical_loss": 3.4579360156957435,
      "tokens_seen": 1818689536
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 1002350,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.173239231109619,
      "objective/train/theoretical_loss": 3.4579254314069265,
      "objective/train/tokens_used": 189214176,
      "theoretical_loss": 3.4579254314069265,
      "tokens_seen": 1818755072
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.06756539881239e-05,
      "loss": 2.5297,
      "theoretical_loss": 3.4579148476062755,
      "tokens_seen": 1818820608
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.066762959396566e-05,
      "loss": 2.7194,
      "theoretical_loss": 3.4578936814693106,
      "tokens_seen": 1818951680
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.065960519980742e-05,
      "loss": 2.5981,
      "theoretical_loss": 3.4578725172845273,
      "tokens_seen": 1819082752
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.065158080564918e-05,
      "loss": 2.6458,
      "theoretical_loss": 3.4578513550516057,
      "tokens_seen": 1819213824
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.064355641149094e-05,
      "loss": 2.7472,
      "theoretical_loss": 3.457830194770225,
      "tokens_seen": 1819344896
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.06355320173327e-05,
      "loss": 2.4691,
      "theoretical_loss": 3.457809036440064,
      "tokens_seen": 1819475968
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.062750762317445e-05,
      "loss": 2.6148,
      "theoretical_loss": 3.4577878800608026,
      "tokens_seen": 1819607040
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.061948322901622e-05,
      "loss": 2.5812,
      "theoretical_loss": 3.457766725632121,
      "tokens_seen": 1819738112
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.061145883485797e-05,
      "loss": 2.7097,
      "theoretical_loss": 3.457745573153698,
      "tokens_seen": 1819869184
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.060343444069974e-05,
      "loss": 2.554,
      "theoretical_loss": 3.4577244226252137,
      "tokens_seen": 1820000256
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.059541004654149e-05,
      "loss": 2.6685,
      "theoretical_loss": 3.4577032740463483,
      "tokens_seen": 1820131328
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.058738565238326e-05,
      "loss": 2.4642,
      "theoretical_loss": 3.4576821274167813,
      "tokens_seen": 1820262400
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 1003363,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7602131366729736,
      "objective/train/theoretical_loss": 3.4576609827361926,
      "objective/train/tokens_used": 190852576,
      "theoretical_loss": 3.4576609827361926,
      "tokens_seen": 1820393472
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.057936125822501e-05,
      "loss": 2.6968,
      "theoretical_loss": 3.4576609827361926,
      "tokens_seen": 1820393472
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.057133686406676e-05,
      "loss": 2.5316,
      "theoretical_loss": 3.457639840004262,
      "tokens_seen": 1820524544
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.056331246990853e-05,
      "loss": 2.6249,
      "theoretical_loss": 3.4576186992206708,
      "tokens_seen": 1820655616
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.055528807575028e-05,
      "loss": 2.7948,
      "theoretical_loss": 3.457597560385098,
      "tokens_seen": 1820786688
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.054726368159205e-05,
      "loss": 2.6169,
      "theoretical_loss": 3.4575764234972253,
      "tokens_seen": 1820917760
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.05392392874338e-05,
      "loss": 2.6168,
      "theoretical_loss": 3.457555288556731,
      "tokens_seen": 1821048832
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.053121489327556e-05,
      "loss": 2.6352,
      "theoretical_loss": 3.4575341555632972,
      "tokens_seen": 1821179904
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.052319049911732e-05,
      "loss": 2.5961,
      "theoretical_loss": 3.457513024516604,
      "tokens_seen": 1821310976
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.051516610495908e-05,
      "loss": 2.6443,
      "theoretical_loss": 3.4574918954163314,
      "tokens_seen": 1821442048
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.050714171080084e-05,
      "loss": 2.6054,
      "theoretical_loss": 3.4574707682621613,
      "tokens_seen": 1821573120
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.04991173166426e-05,
      "loss": 2.5453,
      "theoretical_loss": 3.4574496430537733,
      "tokens_seen": 1821704192
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.049109292248436e-05,
      "loss": 2.583,
      "theoretical_loss": 3.4574285197908483,
      "tokens_seen": 1821835264
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.048306852832611e-05,
      "loss": 2.6039,
      "theoretical_loss": 3.4574073984730678,
      "tokens_seen": 1821966336
    },
    {
      "epoch": 0.1,
      "objective/train/docs_used": 1003939,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.497610569000244,
      "objective/train/theoretical_loss": 3.4573968385435068,
      "objective/train/tokens_used": 192490976,
      "theoretical_loss": 3.4573968385435068,
      "tokens_seen": 1822031872
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.047504413416787e-05,
      "loss": 2.4516,
      "theoretical_loss": 3.4573862791001124,
      "tokens_seen": 1822097408
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.046701974000963e-05,
      "loss": 2.6655,
      "theoretical_loss": 3.457365161671663,
      "tokens_seen": 1822228480
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.045899534585139e-05,
      "loss": 2.5194,
      "theoretical_loss": 3.457344046187401,
      "tokens_seen": 1822359552
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.045097095169315e-05,
      "loss": 2.7839,
      "theoretical_loss": 3.4573229326470076,
      "tokens_seen": 1822490624
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.044294655753491e-05,
      "loss": 2.6435,
      "theoretical_loss": 3.457301821050164,
      "tokens_seen": 1822621696
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.043492216337666e-05,
      "loss": 2.49,
      "theoretical_loss": 3.4572807113965514,
      "tokens_seen": 1822752768
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.042689776921843e-05,
      "loss": 2.6145,
      "theoretical_loss": 3.457259603685851,
      "tokens_seen": 1822883840
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.041887337506018e-05,
      "loss": 2.575,
      "theoretical_loss": 3.4572384979177455,
      "tokens_seen": 1823014912
    },
    {
      "epoch": 0.1,
      "learning_rate": 9.041084898090195e-05,
      "loss": 2.5775,
      "theoretical_loss": 3.457217394091915,
      "tokens_seen": 1823145984
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.04028245867437e-05,
      "loss": 2.7332,
      "theoretical_loss": 3.457196292208042,
      "tokens_seen": 1823277056
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.039480019258547e-05,
      "loss": 2.5805,
      "theoretical_loss": 3.457175192265808,
      "tokens_seen": 1823408128
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.038677579842722e-05,
      "loss": 2.6114,
      "theoretical_loss": 3.4571540942648946,
      "tokens_seen": 1823539200
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1005294,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7300758361816406,
      "objective/train/theoretical_loss": 3.457132998204984,
      "objective/train/tokens_used": 194129376,
      "theoretical_loss": 3.457132998204984,
      "tokens_seen": 1823670272
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.037875140426897e-05,
      "loss": 2.4603,
      "theoretical_loss": 3.457132998204984,
      "tokens_seen": 1823670272
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.037072701011074e-05,
      "loss": 2.6913,
      "theoretical_loss": 3.4571119040857576,
      "tokens_seen": 1823801344
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.036270261595249e-05,
      "loss": 2.5986,
      "theoretical_loss": 3.457090811906898,
      "tokens_seen": 1823932416
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.035467822179426e-05,
      "loss": 2.6846,
      "theoretical_loss": 3.457069721668087,
      "tokens_seen": 1824063488
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.034665382763601e-05,
      "loss": 2.4226,
      "theoretical_loss": 3.4570486333690065,
      "tokens_seen": 1824194560
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.033862943347777e-05,
      "loss": 2.5405,
      "theoretical_loss": 3.4570275470093397,
      "tokens_seen": 1824325632
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.033060503931953e-05,
      "loss": 2.5503,
      "theoretical_loss": 3.457006462588768,
      "tokens_seen": 1824456704
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.032258064516129e-05,
      "loss": 2.6584,
      "theoretical_loss": 3.4569853801069743,
      "tokens_seen": 1824587776
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.031455625100305e-05,
      "loss": 2.5576,
      "theoretical_loss": 3.4569642995636407,
      "tokens_seen": 1824718848
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.03065318568448e-05,
      "loss": 2.5872,
      "theoretical_loss": 3.4569432209584496,
      "tokens_seen": 1824849920
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.029850746268657e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.4569221442910845,
      "tokens_seen": 1824980992
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.029048306852833e-05,
      "loss": 2.764,
      "theoretical_loss": 3.4569010695612272,
      "tokens_seen": 1825112064
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.028245867437008e-05,
      "loss": 2.614,
      "theoretical_loss": 3.456879996768561,
      "tokens_seen": 1825243136
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1005912,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.9645118713378906,
      "objective/train/theoretical_loss": 3.4568694610985755,
      "objective/train/tokens_used": 195767776,
      "theoretical_loss": 3.4568694610985755,
      "tokens_seen": 1825308672
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.027443428021185e-05,
      "loss": 2.4857,
      "theoretical_loss": 3.456858925912768,
      "tokens_seen": 1825374208
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.02664098860536e-05,
      "loss": 2.4868,
      "theoretical_loss": 3.4568378569935323,
      "tokens_seen": 1825505280
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.025838549189537e-05,
      "loss": 2.5564,
      "theoretical_loss": 3.456816790010536,
      "tokens_seen": 1825636352
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.025036109773712e-05,
      "loss": 2.5284,
      "theoretical_loss": 3.456795724963462,
      "tokens_seen": 1825767424
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.024233670357888e-05,
      "loss": 2.6118,
      "theoretical_loss": 3.456774661851995,
      "tokens_seen": 1825898496
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.023431230942064e-05,
      "loss": 2.5575,
      "theoretical_loss": 3.4567536006758157,
      "tokens_seen": 1826029568
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.022628791526239e-05,
      "loss": 2.5875,
      "theoretical_loss": 3.4567325414346093,
      "tokens_seen": 1826160640
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.021826352110416e-05,
      "loss": 2.5273,
      "theoretical_loss": 3.456711484128059,
      "tokens_seen": 1826291712
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.021023912694591e-05,
      "loss": 2.6429,
      "theoretical_loss": 3.456690428755847,
      "tokens_seen": 1826422784
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.020221473278768e-05,
      "loss": 2.6974,
      "theoretical_loss": 3.4566693753176576,
      "tokens_seen": 1826553856
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.019419033862943e-05,
      "loss": 2.4875,
      "theoretical_loss": 3.4566483238131753,
      "tokens_seen": 1826684928
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.01861659444712e-05,
      "loss": 2.562,
      "theoretical_loss": 3.4566272742420825,
      "tokens_seen": 1826816000
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1007268,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4686145782470703,
      "objective/train/theoretical_loss": 3.4566062266040634,
      "objective/train/tokens_used": 197406176,
      "theoretical_loss": 3.4566062266040634,
      "tokens_seen": 1826947072
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.017814155031295e-05,
      "loss": 2.4638,
      "theoretical_loss": 3.4566062266040634,
      "tokens_seen": 1826947072
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.01701171561547e-05,
      "loss": 2.4429,
      "theoretical_loss": 3.4565851808988013,
      "tokens_seen": 1827078144
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.016209276199647e-05,
      "loss": 2.7966,
      "theoretical_loss": 3.456564137125981,
      "tokens_seen": 1827209216
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.015406836783822e-05,
      "loss": 2.4905,
      "theoretical_loss": 3.4565430952852862,
      "tokens_seen": 1827340288
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.014604397367999e-05,
      "loss": 2.517,
      "theoretical_loss": 3.4565220553764004,
      "tokens_seen": 1827471360
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.013801957952174e-05,
      "loss": 2.5002,
      "theoretical_loss": 3.456501017399008,
      "tokens_seen": 1827602432
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.012999518536351e-05,
      "loss": 2.5377,
      "theoretical_loss": 3.4564799813527936,
      "tokens_seen": 1827733504
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.012197079120526e-05,
      "loss": 2.5164,
      "theoretical_loss": 3.4564589472374405,
      "tokens_seen": 1827864576
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.011394639704702e-05,
      "loss": 2.6276,
      "theoretical_loss": 3.456437915052634,
      "tokens_seen": 1827995648
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.010592200288878e-05,
      "loss": 2.5136,
      "theoretical_loss": 3.456416884798058,
      "tokens_seen": 1828126720
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.009789760873054e-05,
      "loss": 2.4933,
      "theoretical_loss": 3.456395856473397,
      "tokens_seen": 1828257792
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.00898732145723e-05,
      "loss": 2.5778,
      "theoretical_loss": 3.456374830078336,
      "tokens_seen": 1828388864
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.008184882041406e-05,
      "loss": 2.4973,
      "theoretical_loss": 3.456353805612559,
      "tokens_seen": 1828519936
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1007941,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.821141481399536,
      "objective/train/theoretical_loss": 3.4563432941030534,
      "objective/train/tokens_used": 199044576,
      "theoretical_loss": 3.4563432941030534,
      "tokens_seen": 1828585472
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.007382442625582e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.4563327830757506,
      "tokens_seen": 1828651008
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.006580003209758e-05,
      "loss": 2.6534,
      "theoretical_loss": 3.4563117624675965,
      "tokens_seen": 1828782080
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.005777563793933e-05,
      "loss": 2.4451,
      "theoretical_loss": 3.456290743787781,
      "tokens_seen": 1828913152
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.00497512437811e-05,
      "loss": 2.5534,
      "theoretical_loss": 3.4562697270359886,
      "tokens_seen": 1829044224
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.004172684962285e-05,
      "loss": 2.6204,
      "theoretical_loss": 3.456248712211905,
      "tokens_seen": 1829175296
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.003370245546462e-05,
      "loss": 2.5441,
      "theoretical_loss": 3.456227699315215,
      "tokens_seen": 1829306368
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.002567806130637e-05,
      "loss": 2.5198,
      "theoretical_loss": 3.4562066883456035,
      "tokens_seen": 1829437440
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.001765366714814e-05,
      "loss": 2.3917,
      "theoretical_loss": 3.4561856793027568,
      "tokens_seen": 1829568512
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.000962927298989e-05,
      "loss": 2.5708,
      "theoretical_loss": 3.4561646721863584,
      "tokens_seen": 1829699584
    },
    {
      "epoch": 0.11,
      "learning_rate": 9.000160487883166e-05,
      "loss": 2.6934,
      "theoretical_loss": 3.4561436669960948,
      "tokens_seen": 1829830656
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.999358048467341e-05,
      "loss": 2.5966,
      "theoretical_loss": 3.4561226637316516,
      "tokens_seen": 1829961728
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.998555609051516e-05,
      "loss": 2.5411,
      "theoretical_loss": 3.456101662392714,
      "tokens_seen": 1830092800
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1009247,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.360668659210205,
      "objective/train/theoretical_loss": 3.456080662978967,
      "objective/train/tokens_used": 200682976,
      "theoretical_loss": 3.456080662978967,
      "tokens_seen": 1830223872
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.997753169635693e-05,
      "loss": 2.4355,
      "theoretical_loss": 3.456080662978967,
      "tokens_seen": 1830223872
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.996950730219868e-05,
      "loss": 2.5936,
      "theoretical_loss": 3.4560596654900975,
      "tokens_seen": 1830354944
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.996148290804045e-05,
      "loss": 2.5363,
      "theoretical_loss": 3.45603866992579,
      "tokens_seen": 1830486016
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.99534585138822e-05,
      "loss": 2.514,
      "theoretical_loss": 3.4560176762857315,
      "tokens_seen": 1830617088
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.994543411972397e-05,
      "loss": 2.4908,
      "theoretical_loss": 3.455996684569607,
      "tokens_seen": 1830748160
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.993740972556572e-05,
      "loss": 2.5477,
      "theoretical_loss": 3.4559756947771025,
      "tokens_seen": 1830879232
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.992938533140747e-05,
      "loss": 2.4969,
      "theoretical_loss": 3.4559547069079044,
      "tokens_seen": 1831010304
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.992136093724924e-05,
      "loss": 2.5748,
      "theoretical_loss": 3.4559337209616987,
      "tokens_seen": 1831141376
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.9913336543091e-05,
      "loss": 2.5423,
      "theoretical_loss": 3.4559127369381715,
      "tokens_seen": 1831272448
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.990531214893276e-05,
      "loss": 2.7193,
      "theoretical_loss": 3.455891754837009,
      "tokens_seen": 1831403520
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.989728775477451e-05,
      "loss": 2.4593,
      "theoretical_loss": 3.455870774657898,
      "tokens_seen": 1831534592
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.988926336061628e-05,
      "loss": 2.6618,
      "theoretical_loss": 3.4558497964005244,
      "tokens_seen": 1831665664
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.988123896645803e-05,
      "loss": 2.5454,
      "theoretical_loss": 3.455828820064575,
      "tokens_seen": 1831796736
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1009780,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6938259601593018,
      "objective/train/theoretical_loss": 3.455818332617036,
      "objective/train/tokens_used": 202321376,
      "theoretical_loss": 3.455818332617036,
      "tokens_seen": 1831862272
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.987321457229979e-05,
      "loss": 2.6755,
      "theoretical_loss": 3.455807845649735,
      "tokens_seen": 1831927808
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.986519017814155e-05,
      "loss": 2.6631,
      "theoretical_loss": 3.4557868731556933,
      "tokens_seen": 1832058880
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.98571657839833e-05,
      "loss": 2.7067,
      "theoretical_loss": 3.4557659025821352,
      "tokens_seen": 1832189952
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.984914138982507e-05,
      "loss": 2.4591,
      "theoretical_loss": 3.455744933928748,
      "tokens_seen": 1832321024
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.984111699566683e-05,
      "loss": 2.3704,
      "theoretical_loss": 3.455723967195218,
      "tokens_seen": 1832452096
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.983309260150859e-05,
      "loss": 2.5915,
      "theoretical_loss": 3.4557030023812323,
      "tokens_seen": 1832583168
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.982506820735035e-05,
      "loss": 2.576,
      "theoretical_loss": 3.455682039486478,
      "tokens_seen": 1832714240
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.98170438131921e-05,
      "loss": 2.5794,
      "theoretical_loss": 3.4556610785106425,
      "tokens_seen": 1832845312
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.980901941903387e-05,
      "loss": 2.6461,
      "theoretical_loss": 3.4556401194534123,
      "tokens_seen": 1832976384
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.980099502487562e-05,
      "loss": 2.632,
      "theoretical_loss": 3.4556191623144747,
      "tokens_seen": 1833107456
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.979297063071739e-05,
      "loss": 2.6446,
      "theoretical_loss": 3.4555982070935176,
      "tokens_seen": 1833238528
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.978494623655914e-05,
      "loss": 2.5226,
      "theoretical_loss": 3.455577253790228,
      "tokens_seen": 1833369600
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1010758,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.585409164428711,
      "objective/train/theoretical_loss": 3.455556302404293,
      "objective/train/tokens_used": 203959776,
      "theoretical_loss": 3.455556302404293,
      "tokens_seen": 1833500672
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.97769218424009e-05,
      "loss": 2.5823,
      "theoretical_loss": 3.455556302404293,
      "tokens_seen": 1833500672
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.976889744824266e-05,
      "loss": 2.422,
      "theoretical_loss": 3.4555353529354003,
      "tokens_seen": 1833631744
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.976087305408441e-05,
      "loss": 2.6723,
      "theoretical_loss": 3.455514405383237,
      "tokens_seen": 1833762816
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.975284865992618e-05,
      "loss": 2.5495,
      "theoretical_loss": 3.4554934597474922,
      "tokens_seen": 1833893888
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.974482426576793e-05,
      "loss": 2.5599,
      "theoretical_loss": 3.4554725160278528,
      "tokens_seen": 1834024960
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.97367998716097e-05,
      "loss": 2.482,
      "theoretical_loss": 3.4554515742240057,
      "tokens_seen": 1834156032
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.972877547745145e-05,
      "loss": 2.6911,
      "theoretical_loss": 3.45543063433564,
      "tokens_seen": 1834287104
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.972075108329322e-05,
      "loss": 2.6872,
      "theoretical_loss": 3.455409696362443,
      "tokens_seen": 1834418176
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.971272668913497e-05,
      "loss": 2.5098,
      "theoretical_loss": 3.4553887603041034,
      "tokens_seen": 1834549248
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.970470229497674e-05,
      "loss": 2.7034,
      "theoretical_loss": 3.4553678261603085,
      "tokens_seen": 1834680320
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.969667790081849e-05,
      "loss": 2.482,
      "theoretical_loss": 3.455346893930747,
      "tokens_seen": 1834811392
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.968865350666024e-05,
      "loss": 2.5007,
      "theoretical_loss": 3.4553259636151066,
      "tokens_seen": 1834942464
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.968062911250201e-05,
      "loss": 2.5495,
      "theoretical_loss": 3.455305035213076,
      "tokens_seen": 1835073536
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1011050,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.749027729034424,
      "objective/train/theoretical_loss": 3.455294571729567,
      "objective/train/tokens_used": 205598176,
      "theoretical_loss": 3.455294571729567,
      "tokens_seen": 1835139072
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.967260471834376e-05,
      "loss": 2.7461,
      "theoretical_loss": 3.455284108724344,
      "tokens_seen": 1835204608
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.966458032418553e-05,
      "loss": 2.8043,
      "theoretical_loss": 3.455263184148598,
      "tokens_seen": 1835335680
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.965655593002728e-05,
      "loss": 2.5974,
      "theoretical_loss": 3.455242261485527,
      "tokens_seen": 1835466752
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.964853153586905e-05,
      "loss": 2.4865,
      "theoretical_loss": 3.45522134073482,
      "tokens_seen": 1835597824
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.96405071417108e-05,
      "loss": 2.6225,
      "theoretical_loss": 3.455200421896165,
      "tokens_seen": 1835728896
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.963248274755256e-05,
      "loss": 2.6943,
      "theoretical_loss": 3.4551795049692515,
      "tokens_seen": 1835859968
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.962445835339432e-05,
      "loss": 2.5456,
      "theoretical_loss": 3.455158589953768,
      "tokens_seen": 1835991040
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.961643395923608e-05,
      "loss": 2.5191,
      "theoretical_loss": 3.455137676849403,
      "tokens_seen": 1836122112
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.960840956507784e-05,
      "loss": 2.6647,
      "theoretical_loss": 3.4551167656558457,
      "tokens_seen": 1836253184
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.96003851709196e-05,
      "loss": 2.5848,
      "theoretical_loss": 3.4550958563727856,
      "tokens_seen": 1836384256
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.959236077676136e-05,
      "loss": 2.5178,
      "theoretical_loss": 3.4550749489999113,
      "tokens_seen": 1836515328
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.958433638260312e-05,
      "loss": 2.5931,
      "theoretical_loss": 3.455054043536912,
      "tokens_seen": 1836646400
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1012285,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6171936988830566,
      "objective/train/theoretical_loss": 3.4550331399834766,
      "objective/train/tokens_used": 207236576,
      "theoretical_loss": 3.4550331399834766,
      "tokens_seen": 1836777472
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.957631198844487e-05,
      "loss": 2.5262,
      "theoretical_loss": 3.4550331399834766,
      "tokens_seen": 1836777472
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.956828759428664e-05,
      "loss": 2.6416,
      "theoretical_loss": 3.455012238339296,
      "tokens_seen": 1836908544
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.956026320012839e-05,
      "loss": 2.5346,
      "theoretical_loss": 3.4549913386040574,
      "tokens_seen": 1837039616
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.955223880597016e-05,
      "loss": 2.6475,
      "theoretical_loss": 3.4549704407774517,
      "tokens_seen": 1837170688
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.954421441181191e-05,
      "loss": 2.7004,
      "theoretical_loss": 3.4549495448591685,
      "tokens_seen": 1837301760
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.953619001765368e-05,
      "loss": 2.6106,
      "theoretical_loss": 3.4549286508488963,
      "tokens_seen": 1837432832
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.952816562349543e-05,
      "loss": 2.5996,
      "theoretical_loss": 3.4549077587463257,
      "tokens_seen": 1837563904
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.952014122933718e-05,
      "loss": 2.8679,
      "theoretical_loss": 3.4548868685511467,
      "tokens_seen": 1837694976
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.951211683517895e-05,
      "loss": 2.6206,
      "theoretical_loss": 3.4548659802630484,
      "tokens_seen": 1837826048
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.95040924410207e-05,
      "loss": 2.4902,
      "theoretical_loss": 3.4548450938817212,
      "tokens_seen": 1837957120
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.949606804686247e-05,
      "loss": 2.5685,
      "theoretical_loss": 3.454824209406855,
      "tokens_seen": 1838088192
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.948804365270422e-05,
      "loss": 2.6072,
      "theoretical_loss": 3.4548033268381397,
      "tokens_seen": 1838219264
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.948001925854599e-05,
      "loss": 2.5037,
      "theoretical_loss": 3.4547824461752654,
      "tokens_seen": 1838350336
    },
    {
      "epoch": 0.11,
      "objective/train/docs_used": 1012947,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6320648193359375,
      "objective/train/theoretical_loss": 3.4547720065584215,
      "objective/train/tokens_used": 208874976,
      "theoretical_loss": 3.4547720065584215,
      "tokens_seen": 1838415872
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.947199486438774e-05,
      "loss": 2.6496,
      "theoretical_loss": 3.454761567417922,
      "tokens_seen": 1838481408
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.94639704702295e-05,
      "loss": 2.6935,
      "theoretical_loss": 3.4547406905658002,
      "tokens_seen": 1838612480
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.945594607607126e-05,
      "loss": 2.5356,
      "theoretical_loss": 3.454719815618591,
      "tokens_seen": 1838743552
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.944792168191301e-05,
      "loss": 2.5054,
      "theoretical_loss": 3.4546989425759835,
      "tokens_seen": 1838874624
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.943989728775478e-05,
      "loss": 2.663,
      "theoretical_loss": 3.4546780714376695,
      "tokens_seen": 1839005696
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.943187289359653e-05,
      "loss": 2.5547,
      "theoretical_loss": 3.4546572022033377,
      "tokens_seen": 1839136768
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.94238484994383e-05,
      "loss": 2.3293,
      "theoretical_loss": 3.4546363348726805,
      "tokens_seen": 1839267840
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.941582410528005e-05,
      "loss": 2.6209,
      "theoretical_loss": 3.4546154694453883,
      "tokens_seen": 1839398912
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.94077997111218e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.454594605921151,
      "tokens_seen": 1839529984
    },
    {
      "epoch": 0.11,
      "learning_rate": 8.939977531696357e-05,
      "loss": 2.6429,
      "theoretical_loss": 3.454573744299661,
      "tokens_seen": 1839661056
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.939175092280533e-05,
      "loss": 2.6145,
      "theoretical_loss": 3.454552884580607,
      "tokens_seen": 1839792128
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.93837265286471e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.454532026763682,
      "tokens_seen": 1839923200
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1014005,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.342764139175415,
      "objective/train/theoretical_loss": 3.4545111708485763,
      "objective/train/tokens_used": 210513376,
      "theoretical_loss": 3.4545111708485763,
      "tokens_seen": 1840054272
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.937570213448885e-05,
      "loss": 2.5643,
      "theoretical_loss": 3.4545111708485763,
      "tokens_seen": 1840054272
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.936767774033061e-05,
      "loss": 2.7472,
      "theoretical_loss": 3.4544903168349808,
      "tokens_seen": 1840185344
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.935965334617237e-05,
      "loss": 2.6095,
      "theoretical_loss": 3.454469464722587,
      "tokens_seen": 1840316416
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.935162895201413e-05,
      "loss": 2.5621,
      "theoretical_loss": 3.4544486145110858,
      "tokens_seen": 1840447488
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.934360455785589e-05,
      "loss": 2.4646,
      "theoretical_loss": 3.4544277662001694,
      "tokens_seen": 1840578560
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.933558016369764e-05,
      "loss": 2.5613,
      "theoretical_loss": 3.4544069197895286,
      "tokens_seen": 1840709632
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.93275557695394e-05,
      "loss": 2.6725,
      "theoretical_loss": 3.454386075278855,
      "tokens_seen": 1840840704
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.931953137538116e-05,
      "loss": 2.6032,
      "theoretical_loss": 3.4543652326678402,
      "tokens_seen": 1840971776
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.931150698122293e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.454344391956176,
      "tokens_seen": 1841102848
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.930348258706468e-05,
      "loss": 2.541,
      "theoretical_loss": 3.4543235531435537,
      "tokens_seen": 1841233920
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.929545819290645e-05,
      "loss": 2.5714,
      "theoretical_loss": 3.454302716229665,
      "tokens_seen": 1841364992
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.92874337987482e-05,
      "loss": 2.6278,
      "theoretical_loss": 3.454281881214203,
      "tokens_seen": 1841496064
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.927940940458995e-05,
      "loss": 2.5964,
      "theoretical_loss": 3.454261048096858,
      "tokens_seen": 1841627136
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1015263,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.241819381713867,
      "objective/train/theoretical_loss": 3.4542506322498836,
      "objective/train/tokens_used": 212151776,
      "theoretical_loss": 3.4542506322498836,
      "tokens_seen": 1841692672
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.927138501043172e-05,
      "loss": 2.6414,
      "theoretical_loss": 3.454240216877323,
      "tokens_seen": 1841758208
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.926336061627347e-05,
      "loss": 2.5861,
      "theoretical_loss": 3.45421938755529,
      "tokens_seen": 1841889280
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.925533622211524e-05,
      "loss": 2.6131,
      "theoretical_loss": 3.4541985601304503,
      "tokens_seen": 1842020352
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.924731182795699e-05,
      "loss": 2.6827,
      "theoretical_loss": 3.4541777346024976,
      "tokens_seen": 1842151424
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.923928743379876e-05,
      "loss": 2.583,
      "theoretical_loss": 3.4541569109711228,
      "tokens_seen": 1842282496
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.923126303964051e-05,
      "loss": 2.5736,
      "theoretical_loss": 3.4541360892360187,
      "tokens_seen": 1842413568
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.922323864548226e-05,
      "loss": 2.6615,
      "theoretical_loss": 3.4541152693968784,
      "tokens_seen": 1842544640
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.921521425132403e-05,
      "loss": 2.6997,
      "theoretical_loss": 3.4540944514533933,
      "tokens_seen": 1842675712
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.920718985716578e-05,
      "loss": 2.5494,
      "theoretical_loss": 3.4540736354052575,
      "tokens_seen": 1842806784
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.919916546300755e-05,
      "loss": 2.6537,
      "theoretical_loss": 3.4540528212521617,
      "tokens_seen": 1842937856
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.91911410688493e-05,
      "loss": 2.5037,
      "theoretical_loss": 3.4540320089938,
      "tokens_seen": 1843068928
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.918311667469107e-05,
      "loss": 2.5411,
      "theoretical_loss": 3.454011198629865,
      "tokens_seen": 1843200000
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1015948,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7754783630371094,
      "objective/train/theoretical_loss": 3.453990390160049,
      "objective/train/tokens_used": 213790176,
      "theoretical_loss": 3.453990390160049,
      "tokens_seen": 1843331072
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.917509228053282e-05,
      "loss": 2.6023,
      "theoretical_loss": 3.453990390160049,
      "tokens_seen": 1843331072
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.916706788637458e-05,
      "loss": 2.6114,
      "theoretical_loss": 3.453969583584045,
      "tokens_seen": 1843462144
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.915904349221634e-05,
      "loss": 2.6495,
      "theoretical_loss": 3.453948778901547,
      "tokens_seen": 1843593216
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.91510190980581e-05,
      "loss": 2.5029,
      "theoretical_loss": 3.4539279761122472,
      "tokens_seen": 1843724288
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.914299470389986e-05,
      "loss": 2.4838,
      "theoretical_loss": 3.453907175215839,
      "tokens_seen": 1843855360
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.913497030974162e-05,
      "loss": 2.468,
      "theoretical_loss": 3.453886376212015,
      "tokens_seen": 1843986432
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.912694591558338e-05,
      "loss": 2.4687,
      "theoretical_loss": 3.4538655791004693,
      "tokens_seen": 1844117504
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.911892152142514e-05,
      "loss": 2.4653,
      "theoretical_loss": 3.4538447838808954,
      "tokens_seen": 1844248576
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.911089712726689e-05,
      "loss": 2.7463,
      "theoretical_loss": 3.453823990552986,
      "tokens_seen": 1844379648
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.910287273310866e-05,
      "loss": 2.459,
      "theoretical_loss": 3.453803199116435,
      "tokens_seen": 1844510720
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.909484833895041e-05,
      "loss": 2.4994,
      "theoretical_loss": 3.4537824095709366,
      "tokens_seen": 1844641792
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.908682394479218e-05,
      "loss": 2.4943,
      "theoretical_loss": 3.4537616219161835,
      "tokens_seen": 1844772864
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.907879955063393e-05,
      "loss": 2.5477,
      "theoretical_loss": 3.4537408361518693,
      "tokens_seen": 1844903936
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1017203,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5314950942993164,
      "objective/train/theoretical_loss": 3.4537304439785315,
      "objective/train/tokens_used": 215428576,
      "theoretical_loss": 3.4537304439785315,
      "tokens_seen": 1844969472
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.90707751564757e-05,
      "loss": 2.4796,
      "theoretical_loss": 3.4537200522776885,
      "tokens_seen": 1845035008
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.906275076231745e-05,
      "loss": 2.7601,
      "theoretical_loss": 3.4536992702933347,
      "tokens_seen": 1845166080
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.905472636815922e-05,
      "loss": 2.4913,
      "theoretical_loss": 3.4536784901985014,
      "tokens_seen": 1845297152
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.904670197400097e-05,
      "loss": 2.5899,
      "theoretical_loss": 3.4536577119928835,
      "tokens_seen": 1845428224
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.903867757984272e-05,
      "loss": 2.3805,
      "theoretical_loss": 3.4536369356761742,
      "tokens_seen": 1845559296
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.903065318568449e-05,
      "loss": 2.6224,
      "theoretical_loss": 3.4536161612480685,
      "tokens_seen": 1845690368
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.902262879152624e-05,
      "loss": 2.5781,
      "theoretical_loss": 3.4535953887082598,
      "tokens_seen": 1845821440
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.901460439736801e-05,
      "loss": 2.5903,
      "theoretical_loss": 3.453574618056443,
      "tokens_seen": 1845952512
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.900658000320976e-05,
      "loss": 2.5429,
      "theoretical_loss": 3.4535538492923123,
      "tokens_seen": 1846083584
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.899855560905153e-05,
      "loss": 2.5974,
      "theoretical_loss": 3.4535330824155617,
      "tokens_seen": 1846214656
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.899053121489328e-05,
      "loss": 2.464,
      "theoretical_loss": 3.4535123174258864,
      "tokens_seen": 1846345728
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.898250682073503e-05,
      "loss": 2.6167,
      "theoretical_loss": 3.4534915543229805,
      "tokens_seen": 1846476800
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1017618,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7826731204986572,
      "objective/train/theoretical_loss": 3.453470793106539,
      "objective/train/tokens_used": 217066976,
      "theoretical_loss": 3.453470793106539,
      "tokens_seen": 1846607872
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.89744824265768e-05,
      "loss": 2.5319,
      "theoretical_loss": 3.453470793106539,
      "tokens_seen": 1846607872
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.896645803241855e-05,
      "loss": 2.6329,
      "theoretical_loss": 3.453450033776256,
      "tokens_seen": 1846738944
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.895843363826032e-05,
      "loss": 2.4089,
      "theoretical_loss": 3.4534292763318275,
      "tokens_seen": 1846870016
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.895040924410207e-05,
      "loss": 2.5541,
      "theoretical_loss": 3.453408520772947,
      "tokens_seen": 1847001088
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.894238484994384e-05,
      "loss": 2.5275,
      "theoretical_loss": 3.4533877670993096,
      "tokens_seen": 1847132160
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.89343604557856e-05,
      "loss": 2.4709,
      "theoretical_loss": 3.4533670153106115,
      "tokens_seen": 1847263232
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.892633606162735e-05,
      "loss": 2.4936,
      "theoretical_loss": 3.4533462654065463,
      "tokens_seen": 1847394304
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.891831166746911e-05,
      "loss": 2.6244,
      "theoretical_loss": 3.4533255173868103,
      "tokens_seen": 1847525376
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.891028727331087e-05,
      "loss": 2.5891,
      "theoretical_loss": 3.4533047712510982,
      "tokens_seen": 1847656448
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.890226287915263e-05,
      "loss": 2.4386,
      "theoretical_loss": 3.453284026999105,
      "tokens_seen": 1847787520
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.889423848499439e-05,
      "loss": 2.5154,
      "theoretical_loss": 3.453263284630527,
      "tokens_seen": 1847918592
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.888621409083615e-05,
      "loss": 2.5096,
      "theoretical_loss": 3.4532425441450587,
      "tokens_seen": 1848049664
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.88781896966779e-05,
      "loss": 2.5441,
      "theoretical_loss": 3.4532218055423956,
      "tokens_seen": 1848180736
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1018803,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3100316524505615,
      "objective/train/theoretical_loss": 3.4532114369470213,
      "objective/train/tokens_used": 218705376,
      "theoretical_loss": 3.4532114369470213,
      "tokens_seen": 1848246272
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.887016530251966e-05,
      "loss": 2.3515,
      "theoretical_loss": 3.453201068822234,
      "tokens_seen": 1848311808
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.886214090836143e-05,
      "loss": 2.4272,
      "theoretical_loss": 3.4531803339842693,
      "tokens_seen": 1848442880
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.885411651420318e-05,
      "loss": 2.4624,
      "theoretical_loss": 3.4531596010281964,
      "tokens_seen": 1848573952
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.884609212004495e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.453138869953712,
      "tokens_seen": 1848705024
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.88380677258867e-05,
      "loss": 2.622,
      "theoretical_loss": 3.4531181407605116,
      "tokens_seen": 1848836096
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.883004333172847e-05,
      "loss": 2.6775,
      "theoretical_loss": 3.4530974134482917,
      "tokens_seen": 1848967168
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.882201893757022e-05,
      "loss": 2.564,
      "theoretical_loss": 3.4530766880167474,
      "tokens_seen": 1849098240
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.881399454341197e-05,
      "loss": 2.6534,
      "theoretical_loss": 3.4530559644655754,
      "tokens_seen": 1849229312
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.880597014925374e-05,
      "loss": 2.5453,
      "theoretical_loss": 3.4530352427944715,
      "tokens_seen": 1849360384
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.879794575509549e-05,
      "loss": 2.6126,
      "theoretical_loss": 3.453014523003132,
      "tokens_seen": 1849491456
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.878992136093726e-05,
      "loss": 2.5405,
      "theoretical_loss": 3.452993805091253,
      "tokens_seen": 1849622528
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.878189696677901e-05,
      "loss": 2.6833,
      "theoretical_loss": 3.4529730890585313,
      "tokens_seen": 1849753600
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1019322,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4879612922668457,
      "objective/train/theoretical_loss": 3.4529523749046627,
      "objective/train/tokens_used": 220343776,
      "theoretical_loss": 3.4529523749046627,
      "tokens_seen": 1849884672
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.877387257262078e-05,
      "loss": 2.5839,
      "theoretical_loss": 3.4529523749046627,
      "tokens_seen": 1849884672
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.876584817846253e-05,
      "loss": 2.5506,
      "theoretical_loss": 3.4529316626293443,
      "tokens_seen": 1850015744
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.87578237843043e-05,
      "loss": 2.5309,
      "theoretical_loss": 3.4529109522322727,
      "tokens_seen": 1850146816
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.874979939014605e-05,
      "loss": 2.5407,
      "theoretical_loss": 3.452890243713144,
      "tokens_seen": 1850277888
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.87417749959878e-05,
      "loss": 2.4533,
      "theoretical_loss": 3.452869537071655,
      "tokens_seen": 1850408960
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.873375060182957e-05,
      "loss": 2.5144,
      "theoretical_loss": 3.4528488323075024,
      "tokens_seen": 1850540032
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.872572620767132e-05,
      "loss": 2.4951,
      "theoretical_loss": 3.452828129420384,
      "tokens_seen": 1850671104
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.871770181351309e-05,
      "loss": 2.5968,
      "theoretical_loss": 3.4528074284099954,
      "tokens_seen": 1850802176
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.870967741935484e-05,
      "loss": 2.7107,
      "theoretical_loss": 3.452786729276034,
      "tokens_seen": 1850933248
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.870165302519661e-05,
      "loss": 2.4595,
      "theoretical_loss": 3.4527660320181974,
      "tokens_seen": 1851064320
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.869362863103836e-05,
      "loss": 2.5775,
      "theoretical_loss": 3.452745336636182,
      "tokens_seen": 1851195392
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.868560423688012e-05,
      "loss": 2.627,
      "theoretical_loss": 3.4527246431296854,
      "tokens_seen": 1851326464
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.867757984272188e-05,
      "loss": 2.5698,
      "theoretical_loss": 3.452703951498405,
      "tokens_seen": 1851457536
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1019593,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.545363426208496,
      "objective/train/theoretical_loss": 3.452693606385876,
      "objective/train/tokens_used": 221982176,
      "theoretical_loss": 3.452693606385876,
      "tokens_seen": 1851523072
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.866955544856364e-05,
      "loss": 2.6297,
      "theoretical_loss": 3.452683261742038,
      "tokens_seen": 1851588608
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.86615310544054e-05,
      "loss": 2.664,
      "theoretical_loss": 3.452662573860281,
      "tokens_seen": 1851719680
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.865350666024716e-05,
      "loss": 2.5098,
      "theoretical_loss": 3.4526418878528333,
      "tokens_seen": 1851850752
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.864548226608892e-05,
      "loss": 2.5979,
      "theoretical_loss": 3.452621203719391,
      "tokens_seen": 1851981824
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.863745787193068e-05,
      "loss": 2.6701,
      "theoretical_loss": 3.452600521459652,
      "tokens_seen": 1852112896
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.862943347777243e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.4525798410733133,
      "tokens_seen": 1852243968
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.86214090836142e-05,
      "loss": 2.615,
      "theoretical_loss": 3.4525591625600742,
      "tokens_seen": 1852375040
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.861338468945595e-05,
      "loss": 2.4747,
      "theoretical_loss": 3.4525384859196313,
      "tokens_seen": 1852506112
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.860536029529772e-05,
      "loss": 2.5322,
      "theoretical_loss": 3.4525178111516834,
      "tokens_seen": 1852637184
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.859733590113947e-05,
      "loss": 2.6186,
      "theoretical_loss": 3.4524971382559277,
      "tokens_seen": 1852768256
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.858931150698124e-05,
      "loss": 2.5473,
      "theoretical_loss": 3.4524764672320627,
      "tokens_seen": 1852899328
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.858128711282299e-05,
      "loss": 2.5329,
      "theoretical_loss": 3.452455798079786,
      "tokens_seen": 1853030400
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1020710,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7822999954223633,
      "objective/train/theoretical_loss": 3.4524351307987966,
      "objective/train/tokens_used": 223620576,
      "theoretical_loss": 3.4524351307987966,
      "tokens_seen": 1853161472
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.857326271866474e-05,
      "loss": 2.7127,
      "theoretical_loss": 3.4524351307987966,
      "tokens_seen": 1853161472
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.856523832450651e-05,
      "loss": 2.6761,
      "theoretical_loss": 3.452414465388792,
      "tokens_seen": 1853292544
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.855721393034826e-05,
      "loss": 2.5513,
      "theoretical_loss": 3.4523938018494706,
      "tokens_seen": 1853423616
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.854918953619003e-05,
      "loss": 2.5548,
      "theoretical_loss": 3.4523731401805313,
      "tokens_seen": 1853554688
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.854116514203178e-05,
      "loss": 2.7371,
      "theoretical_loss": 3.4523524803816716,
      "tokens_seen": 1853685760
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.853314074787355e-05,
      "loss": 2.5652,
      "theoretical_loss": 3.4523318224525914,
      "tokens_seen": 1853816832
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.85251163537153e-05,
      "loss": 2.6217,
      "theoretical_loss": 3.452311166392988,
      "tokens_seen": 1853947904
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.851709195955705e-05,
      "loss": 2.6013,
      "theoretical_loss": 3.452290512202561,
      "tokens_seen": 1854078976
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.850906756539882e-05,
      "loss": 2.4873,
      "theoretical_loss": 3.4522698598810084,
      "tokens_seen": 1854210048
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.850104317124057e-05,
      "loss": 2.5747,
      "theoretical_loss": 3.4522492094280297,
      "tokens_seen": 1854341120
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.849301877708234e-05,
      "loss": 2.4675,
      "theoretical_loss": 3.4522285608433236,
      "tokens_seen": 1854472192
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.84849943829241e-05,
      "loss": 2.6827,
      "theoretical_loss": 3.4522079141265882,
      "tokens_seen": 1854603264
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.847696998876586e-05,
      "loss": 2.6778,
      "theoretical_loss": 3.4521872692775237,
      "tokens_seen": 1854734336
    },
    {
      "epoch": 0.12,
      "objective/train/docs_used": 1021316,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.845196485519409,
      "objective/train/theoretical_loss": 3.4521769475532738,
      "objective/train/tokens_used": 225258976,
      "theoretical_loss": 3.4521769475532738,
      "tokens_seen": 1854799872
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.846894559460761e-05,
      "loss": 2.6202,
      "theoretical_loss": 3.4521666262958286,
      "tokens_seen": 1854865408
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.846092120044937e-05,
      "loss": 2.4954,
      "theoretical_loss": 3.452145985181202,
      "tokens_seen": 1854996480
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.845289680629113e-05,
      "loss": 2.6478,
      "theoretical_loss": 3.452125345933343,
      "tokens_seen": 1855127552
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.844487241213289e-05,
      "loss": 2.4197,
      "theoretical_loss": 3.4521047085519516,
      "tokens_seen": 1855258624
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.843684801797465e-05,
      "loss": 2.4694,
      "theoretical_loss": 3.4520840730367266,
      "tokens_seen": 1855389696
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.842882362381641e-05,
      "loss": 2.6215,
      "theoretical_loss": 3.4520634393873673,
      "tokens_seen": 1855520768
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.842079922965817e-05,
      "loss": 2.4711,
      "theoretical_loss": 3.452042807603574,
      "tokens_seen": 1855651840
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.841277483549993e-05,
      "loss": 2.5541,
      "theoretical_loss": 3.452022177685045,
      "tokens_seen": 1855782912
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.84047504413417e-05,
      "loss": 2.691,
      "theoretical_loss": 3.452001549631481,
      "tokens_seen": 1855913984
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.839672604718345e-05,
      "loss": 2.7246,
      "theoretical_loss": 3.4519809234425813,
      "tokens_seen": 1856045056
    },
    {
      "epoch": 0.12,
      "learning_rate": 8.83887016530252e-05,
      "loss": 2.3307,
      "theoretical_loss": 3.451960299118046,
      "tokens_seen": 1856176128
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.838067725886697e-05,
      "loss": 2.498,
      "theoretical_loss": 3.4519396766575747,
      "tokens_seen": 1856307200
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1022409,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.1355769634246826,
      "objective/train/theoretical_loss": 3.4519190560608672,
      "objective/train/tokens_used": 226897376,
      "theoretical_loss": 3.4519190560608672,
      "tokens_seen": 1856438272
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.837265286470872e-05,
      "loss": 2.5153,
      "theoretical_loss": 3.4519190560608672,
      "tokens_seen": 1856438272
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.836462847055049e-05,
      "loss": 2.4026,
      "theoretical_loss": 3.4518984373276234,
      "tokens_seen": 1856569344
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.835660407639224e-05,
      "loss": 2.5265,
      "theoretical_loss": 3.4518778204575433,
      "tokens_seen": 1856700416
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.8348579682234e-05,
      "loss": 2.4881,
      "theoretical_loss": 3.4518572054503283,
      "tokens_seen": 1856831488
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.834055528807576e-05,
      "loss": 2.5827,
      "theoretical_loss": 3.4518365923056766,
      "tokens_seen": 1856962560
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.833253089391751e-05,
      "loss": 2.6521,
      "theoretical_loss": 3.45181598102329,
      "tokens_seen": 1857093632
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.832450649975928e-05,
      "loss": 2.5918,
      "theoretical_loss": 3.451795371602868,
      "tokens_seen": 1857224704
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.831648210560103e-05,
      "loss": 2.5057,
      "theoretical_loss": 3.451774764044112,
      "tokens_seen": 1857355776
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.83084577114428e-05,
      "loss": 2.551,
      "theoretical_loss": 3.4517541583467213,
      "tokens_seen": 1857486848
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.830043331728455e-05,
      "loss": 2.5057,
      "theoretical_loss": 3.4517335545103967,
      "tokens_seen": 1857617920
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.829240892312632e-05,
      "loss": 2.4116,
      "theoretical_loss": 3.45171295253484,
      "tokens_seen": 1857748992
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.828438452896807e-05,
      "loss": 2.5689,
      "theoretical_loss": 3.45169235241975,
      "tokens_seen": 1857880064
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.827636013480982e-05,
      "loss": 2.4917,
      "theoretical_loss": 3.4516717541648285,
      "tokens_seen": 1858011136
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1023672,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6751821041107178,
      "objective/train/theoretical_loss": 3.4516614557348375,
      "objective/train/tokens_used": 228535776,
      "theoretical_loss": 3.4516614557348375,
      "tokens_seen": 1858076672
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.826833574065159e-05,
      "loss": 2.6894,
      "theoretical_loss": 3.451651157769777,
      "tokens_seen": 1858142208
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.826031134649334e-05,
      "loss": 2.6551,
      "theoretical_loss": 3.4516305632342945,
      "tokens_seen": 1858273280
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.825228695233511e-05,
      "loss": 2.5355,
      "theoretical_loss": 3.451609970558083,
      "tokens_seen": 1858404352
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.824426255817686e-05,
      "loss": 2.499,
      "theoretical_loss": 3.4515893797408443,
      "tokens_seen": 1858535424
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.823623816401863e-05,
      "loss": 2.6607,
      "theoretical_loss": 3.4515687907822787,
      "tokens_seen": 1858666496
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.822821376986038e-05,
      "loss": 2.6997,
      "theoretical_loss": 3.451548203682088,
      "tokens_seen": 1858797568
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.822018937570214e-05,
      "loss": 2.5268,
      "theoretical_loss": 3.4515276184399717,
      "tokens_seen": 1858928640
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.82121649815439e-05,
      "loss": 2.7007,
      "theoretical_loss": 3.4515070350556334,
      "tokens_seen": 1859059712
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.820414058738566e-05,
      "loss": 2.4948,
      "theoretical_loss": 3.451486453528773,
      "tokens_seen": 1859190784
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.819611619322742e-05,
      "loss": 2.494,
      "theoretical_loss": 3.4514658738590915,
      "tokens_seen": 1859321856
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.818809179906918e-05,
      "loss": 2.7342,
      "theoretical_loss": 3.451445296046292,
      "tokens_seen": 1859452928
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.818006740491094e-05,
      "loss": 2.7209,
      "theoretical_loss": 3.4514247200900754,
      "tokens_seen": 1859584000
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1024297,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4555611610412598,
      "objective/train/theoretical_loss": 3.4514041459901432,
      "objective/train/tokens_used": 230174176,
      "theoretical_loss": 3.4514041459901432,
      "tokens_seen": 1859715072
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.81720430107527e-05,
      "loss": 2.6482,
      "theoretical_loss": 3.4514041459901432,
      "tokens_seen": 1859715072
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.816401861659445e-05,
      "loss": 2.6735,
      "theoretical_loss": 3.451383573746197,
      "tokens_seen": 1859846144
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.815599422243622e-05,
      "loss": 2.5834,
      "theoretical_loss": 3.451363003357939,
      "tokens_seen": 1859977216
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.814796982827797e-05,
      "loss": 2.5174,
      "theoretical_loss": 3.4513424348250705,
      "tokens_seen": 1860108288
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.813994543411974e-05,
      "loss": 2.6091,
      "theoretical_loss": 3.451321868147294,
      "tokens_seen": 1860239360
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.813192103996149e-05,
      "loss": 2.6376,
      "theoretical_loss": 3.4513013033243114,
      "tokens_seen": 1860370432
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.812389664580326e-05,
      "loss": 2.616,
      "theoretical_loss": 3.4512807403558243,
      "tokens_seen": 1860501504
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.811587225164501e-05,
      "loss": 2.5169,
      "theoretical_loss": 3.451260179241536,
      "tokens_seen": 1860632576
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.810784785748678e-05,
      "loss": 2.7286,
      "theoretical_loss": 3.4512396199811466,
      "tokens_seen": 1860763648
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.809982346332853e-05,
      "loss": 2.5962,
      "theoretical_loss": 3.4512190625743604,
      "tokens_seen": 1860894720
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.809179906917028e-05,
      "loss": 2.6989,
      "theoretical_loss": 3.4511985070208793,
      "tokens_seen": 1861025792
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.808377467501205e-05,
      "loss": 2.6519,
      "theoretical_loss": 3.4511779533204043,
      "tokens_seen": 1861156864
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.80757502808538e-05,
      "loss": 2.6076,
      "theoretical_loss": 3.4511574014726403,
      "tokens_seen": 1861287936
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1025545,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7613306045532227,
      "objective/train/theoretical_loss": 3.4511471262434306,
      "objective/train/tokens_used": 231812576,
      "theoretical_loss": 3.4511471262434306,
      "tokens_seen": 1861353472
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.806772588669557e-05,
      "loss": 2.6141,
      "theoretical_loss": 3.4511368514772878,
      "tokens_seen": 1861419008
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.805970149253732e-05,
      "loss": 2.6818,
      "theoretical_loss": 3.4511163033340497,
      "tokens_seen": 1861550080
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.805167709837908e-05,
      "loss": 2.695,
      "theoretical_loss": 3.45109575704263,
      "tokens_seen": 1861681152
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.804365270422084e-05,
      "loss": 2.4897,
      "theoretical_loss": 3.4510752126027295,
      "tokens_seen": 1861812224
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.80356283100626e-05,
      "loss": 2.526,
      "theoretical_loss": 3.451054670014053,
      "tokens_seen": 1861943296
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.802760391590436e-05,
      "loss": 2.564,
      "theoretical_loss": 3.451034129276302,
      "tokens_seen": 1862074368
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.801957952174611e-05,
      "loss": 2.4509,
      "theoretical_loss": 3.45101359038918,
      "tokens_seen": 1862205440
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.801155512758788e-05,
      "loss": 2.6459,
      "theoretical_loss": 3.4509930533523905,
      "tokens_seen": 1862336512
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.800353073342963e-05,
      "loss": 2.5574,
      "theoretical_loss": 3.4509725181656354,
      "tokens_seen": 1862467584
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.799550633927139e-05,
      "loss": 2.6837,
      "theoretical_loss": 3.4509519848286185,
      "tokens_seen": 1862598656
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.798748194511315e-05,
      "loss": 2.6577,
      "theoretical_loss": 3.4509314533410436,
      "tokens_seen": 1862729728
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.797945755095491e-05,
      "loss": 2.6322,
      "theoretical_loss": 3.450910923702613,
      "tokens_seen": 1862860800
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1026014,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.46268630027771,
      "objective/train/theoretical_loss": 3.4508903959130306,
      "objective/train/tokens_used": 233450976,
      "theoretical_loss": 3.4508903959130306,
      "tokens_seen": 1862991872
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.797143315679667e-05,
      "loss": 2.6023,
      "theoretical_loss": 3.4508903959130306,
      "tokens_seen": 1862991872
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.796340876263843e-05,
      "loss": 2.5201,
      "theoretical_loss": 3.450869869972,
      "tokens_seen": 1863122944
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.795538436848018e-05,
      "loss": 2.762,
      "theoretical_loss": 3.4508493458792246,
      "tokens_seen": 1863254016
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.794735997432195e-05,
      "loss": 2.7066,
      "theoretical_loss": 3.4508288236344073,
      "tokens_seen": 1863385088
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.79393355801637e-05,
      "loss": 2.5886,
      "theoretical_loss": 3.4508083032372525,
      "tokens_seen": 1863516160
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.793131118600547e-05,
      "loss": 2.6404,
      "theoretical_loss": 3.4507877846874635,
      "tokens_seen": 1863647232
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.792328679184722e-05,
      "loss": 2.7333,
      "theoretical_loss": 3.4507672679847445,
      "tokens_seen": 1863778304
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.791526239768899e-05,
      "loss": 2.4031,
      "theoretical_loss": 3.450746753128799,
      "tokens_seen": 1863909376
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.790723800353074e-05,
      "loss": 2.5246,
      "theoretical_loss": 3.4507262401193315,
      "tokens_seen": 1864040448
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.789921360937249e-05,
      "loss": 2.3183,
      "theoretical_loss": 3.450705728956045,
      "tokens_seen": 1864171520
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.789118921521426e-05,
      "loss": 2.6952,
      "theoretical_loss": 3.450685219638644,
      "tokens_seen": 1864302592
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.788316482105601e-05,
      "loss": 2.5442,
      "theoretical_loss": 3.4506647121668332,
      "tokens_seen": 1864433664
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.787514042689778e-05,
      "loss": 2.6715,
      "theoretical_loss": 3.450644206540316,
      "tokens_seen": 1864564736
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1027424,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5948421955108643,
      "objective/train/theoretical_loss": 3.4506339544189504,
      "objective/train/tokens_used": 235089376,
      "theoretical_loss": 3.4506339544189504,
      "tokens_seen": 1864630272
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.786711603273953e-05,
      "loss": 2.6985,
      "theoretical_loss": 3.450623702758797,
      "tokens_seen": 1864695808
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.785909163858129e-05,
      "loss": 2.6472,
      "theoretical_loss": 3.450603200821981,
      "tokens_seen": 1864826880
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.785106724442305e-05,
      "loss": 2.7063,
      "theoretical_loss": 3.450582700729571,
      "tokens_seen": 1864957952
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.78430428502648e-05,
      "loss": 2.7536,
      "theoretical_loss": 3.4505622024812728,
      "tokens_seen": 1865089024
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.783501845610657e-05,
      "loss": 2.6416,
      "theoretical_loss": 3.4505417060767902,
      "tokens_seen": 1865220096
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.782699406194833e-05,
      "loss": 2.7129,
      "theoretical_loss": 3.450521211515828,
      "tokens_seen": 1865351168
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.781896966779009e-05,
      "loss": 2.5542,
      "theoretical_loss": 3.4505007187980916,
      "tokens_seen": 1865482240
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.781094527363185e-05,
      "loss": 2.6308,
      "theoretical_loss": 3.4504802279232845,
      "tokens_seen": 1865613312
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.78029208794736e-05,
      "loss": 2.5675,
      "theoretical_loss": 3.4504597388911122,
      "tokens_seen": 1865744384
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.779489648531536e-05,
      "loss": 2.5833,
      "theoretical_loss": 3.4504392517012796,
      "tokens_seen": 1865875456
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.778687209115712e-05,
      "loss": 2.6185,
      "theoretical_loss": 3.450418766353491,
      "tokens_seen": 1866006528
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.777884769699888e-05,
      "loss": 2.7469,
      "theoretical_loss": 3.450398282847452,
      "tokens_seen": 1866137600
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1028516,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7388877868652344,
      "objective/train/theoretical_loss": 3.450377801182868,
      "objective/train/tokens_used": 236727776,
      "theoretical_loss": 3.450377801182868,
      "tokens_seen": 1866268672
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.777082330284064e-05,
      "loss": 2.6726,
      "theoretical_loss": 3.450377801182868,
      "tokens_seen": 1866268672
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.776279890868239e-05,
      "loss": 2.642,
      "theoretical_loss": 3.4503573213594434,
      "tokens_seen": 1866399744
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.775477451452416e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.4503368433768835,
      "tokens_seen": 1866530816
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.774675012036591e-05,
      "loss": 2.4442,
      "theoretical_loss": 3.450316367234894,
      "tokens_seen": 1866661888
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.773872572620768e-05,
      "loss": 2.513,
      "theoretical_loss": 3.45029589293318,
      "tokens_seen": 1866792960
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.773070133204943e-05,
      "loss": 2.5866,
      "theoretical_loss": 3.450275420471447,
      "tokens_seen": 1866924032
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.77226769378912e-05,
      "loss": 2.7162,
      "theoretical_loss": 3.4502549498494,
      "tokens_seen": 1867055104
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.771465254373295e-05,
      "loss": 2.518,
      "theoretical_loss": 3.4502344810667456,
      "tokens_seen": 1867186176
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.77066281495747e-05,
      "loss": 2.6779,
      "theoretical_loss": 3.4502140141231887,
      "tokens_seen": 1867317248
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.769860375541647e-05,
      "loss": 2.6632,
      "theoretical_loss": 3.450193549018435,
      "tokens_seen": 1867448320
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.769057936125822e-05,
      "loss": 2.4661,
      "theoretical_loss": 3.4501730857521906,
      "tokens_seen": 1867579392
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.768255496709999e-05,
      "loss": 2.5559,
      "theoretical_loss": 3.4501526243241605,
      "tokens_seen": 1867710464
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.767453057294174e-05,
      "loss": 2.7043,
      "theoretical_loss": 3.4501321647340513,
      "tokens_seen": 1867841536
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1029117,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6434731483459473,
      "objective/train/theoretical_loss": 3.4501219356281254,
      "objective/train/tokens_used": 238366176,
      "theoretical_loss": 3.4501219356281254,
      "tokens_seen": 1867907072
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.76665061787835e-05,
      "loss": 2.7618,
      "theoretical_loss": 3.450111706981569,
      "tokens_seen": 1867972608
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.765848178462526e-05,
      "loss": 2.5639,
      "theoretical_loss": 3.4500912510664192,
      "tokens_seen": 1868103680
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.765045739046702e-05,
      "loss": 2.6544,
      "theoretical_loss": 3.4500707969883084,
      "tokens_seen": 1868234752
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.764243299630878e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.450050344746942,
      "tokens_seen": 1868365824
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.763440860215054e-05,
      "loss": 2.5622,
      "theoretical_loss": 3.450029894342028,
      "tokens_seen": 1868496896
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.76263842079923e-05,
      "loss": 2.6618,
      "theoretical_loss": 3.4500094457732704,
      "tokens_seen": 1868627968
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.761835981383406e-05,
      "loss": 2.7035,
      "theoretical_loss": 3.449988999040377,
      "tokens_seen": 1868759040
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.761033541967581e-05,
      "loss": 2.6978,
      "theoretical_loss": 3.4499685541430543,
      "tokens_seen": 1868890112
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.760231102551758e-05,
      "loss": 2.5531,
      "theoretical_loss": 3.4499481110810075,
      "tokens_seen": 1869021184
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.759428663135933e-05,
      "loss": 2.5434,
      "theoretical_loss": 3.449927669853945,
      "tokens_seen": 1869152256
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.75862622372011e-05,
      "loss": 2.5474,
      "theoretical_loss": 3.449907230461572,
      "tokens_seen": 1869283328
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.757823784304285e-05,
      "loss": 2.6269,
      "theoretical_loss": 3.4498867929035955,
      "tokens_seen": 1869414400
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1029470,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6646249294281006,
      "objective/train/theoretical_loss": 3.449866357179723,
      "objective/train/tokens_used": 240004576,
      "theoretical_loss": 3.449866357179723,
      "tokens_seen": 1869545472
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.75702134488846e-05,
      "loss": 2.7099,
      "theoretical_loss": 3.449866357179723,
      "tokens_seen": 1869545472
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.756218905472637e-05,
      "loss": 2.635,
      "theoretical_loss": 3.44984592328966,
      "tokens_seen": 1869676544
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.755416466056812e-05,
      "loss": 2.4857,
      "theoretical_loss": 3.4498254912331148,
      "tokens_seen": 1869807616
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.754614026640989e-05,
      "loss": 2.7694,
      "theoretical_loss": 3.4498050610097932,
      "tokens_seen": 1869938688
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.753811587225164e-05,
      "loss": 2.489,
      "theoretical_loss": 3.4497846326194033,
      "tokens_seen": 1870069760
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.753009147809341e-05,
      "loss": 2.479,
      "theoretical_loss": 3.449764206061652,
      "tokens_seen": 1870200832
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.752206708393516e-05,
      "loss": 2.6056,
      "theoretical_loss": 3.4497437813362453,
      "tokens_seen": 1870331904
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.751404268977691e-05,
      "loss": 2.7141,
      "theoretical_loss": 3.449723358442892,
      "tokens_seen": 1870462976
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.750601829561868e-05,
      "loss": 2.6245,
      "theoretical_loss": 3.449702937381298,
      "tokens_seen": 1870594048
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.749799390146043e-05,
      "loss": 2.6135,
      "theoretical_loss": 3.449682518151172,
      "tokens_seen": 1870725120
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.74899695073022e-05,
      "loss": 2.6038,
      "theoretical_loss": 3.449662100752221,
      "tokens_seen": 1870856192
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.748194511314395e-05,
      "loss": 2.5741,
      "theoretical_loss": 3.449641685184152,
      "tokens_seen": 1870987264
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.747392071898572e-05,
      "loss": 2.5113,
      "theoretical_loss": 3.4496212714466727,
      "tokens_seen": 1871118336
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1030778,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5349972248077393,
      "objective/train/theoretical_loss": 3.449611065264313,
      "objective/train/tokens_used": 241642976,
      "theoretical_loss": 3.449611065264313,
      "tokens_seen": 1871183872
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.746589632482747e-05,
      "loss": 2.6365,
      "theoretical_loss": 3.449600859539491,
      "tokens_seen": 1871249408
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.745787193066924e-05,
      "loss": 2.6784,
      "theoretical_loss": 3.4495804494623146,
      "tokens_seen": 1871380480
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.7449847536511e-05,
      "loss": 2.6674,
      "theoretical_loss": 3.449560041214851,
      "tokens_seen": 1871511552
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.744182314235275e-05,
      "loss": 2.6919,
      "theoretical_loss": 3.4495396347968086,
      "tokens_seen": 1871642624
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.743379874819451e-05,
      "loss": 2.5136,
      "theoretical_loss": 3.449519230207895,
      "tokens_seen": 1871773696
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.742577435403627e-05,
      "loss": 2.4687,
      "theoretical_loss": 3.4494988274478184,
      "tokens_seen": 1871904768
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.741774995987803e-05,
      "loss": 2.5254,
      "theoretical_loss": 3.4494784265162863,
      "tokens_seen": 1872035840
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.740972556571979e-05,
      "loss": 2.5739,
      "theoretical_loss": 3.449458027413007,
      "tokens_seen": 1872166912
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.740170117156155e-05,
      "loss": 2.6662,
      "theoretical_loss": 3.4494376301376892,
      "tokens_seen": 1872297984
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.73936767774033e-05,
      "loss": 2.3869,
      "theoretical_loss": 3.4494172346900402,
      "tokens_seen": 1872429056
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.738565238324506e-05,
      "loss": 2.5522,
      "theoretical_loss": 3.4493968410697695,
      "tokens_seen": 1872560128
    },
    {
      "epoch": 0.13,
      "learning_rate": 8.737762798908683e-05,
      "loss": 2.5246,
      "theoretical_loss": 3.449376449276584,
      "tokens_seen": 1872691200
    },
    {
      "epoch": 0.13,
      "objective/train/docs_used": 1031383,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6211163997650146,
      "objective/train/theoretical_loss": 3.449356059310194,
      "objective/train/tokens_used": 243281376,
      "theoretical_loss": 3.449356059310194,
      "tokens_seen": 1872822272
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.736960359492858e-05,
      "loss": 2.53,
      "theoretical_loss": 3.449356059310194,
      "tokens_seen": 1872822272
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.736157920077035e-05,
      "loss": 2.6602,
      "theoretical_loss": 3.4493356711703065,
      "tokens_seen": 1872953344
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.73535548066121e-05,
      "loss": 2.4317,
      "theoretical_loss": 3.44931528485663,
      "tokens_seen": 1873084416
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.734553041245387e-05,
      "loss": 2.6668,
      "theoretical_loss": 3.4492949003688747,
      "tokens_seen": 1873215488
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.733750601829562e-05,
      "loss": 2.6375,
      "theoretical_loss": 3.4492745177067476,
      "tokens_seen": 1873346560
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.732948162413737e-05,
      "loss": 2.6174,
      "theoretical_loss": 3.4492541368699587,
      "tokens_seen": 1873477632
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.732145722997914e-05,
      "loss": 2.4123,
      "theoretical_loss": 3.4492337578582157,
      "tokens_seen": 1873608704
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.731343283582089e-05,
      "loss": 2.5342,
      "theoretical_loss": 3.449213380671229,
      "tokens_seen": 1873739776
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.730540844166266e-05,
      "loss": 2.5533,
      "theoretical_loss": 3.4491930053087065,
      "tokens_seen": 1873870848
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.729738404750441e-05,
      "loss": 2.5133,
      "theoretical_loss": 3.449172631770357,
      "tokens_seen": 1874001920
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.728935965334618e-05,
      "loss": 2.657,
      "theoretical_loss": 3.4491522600558904,
      "tokens_seen": 1874132992
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.728133525918793e-05,
      "loss": 2.5688,
      "theoretical_loss": 3.449131890165016,
      "tokens_seen": 1874264064
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.727331086502968e-05,
      "loss": 2.6587,
      "theoretical_loss": 3.4491115220974424,
      "tokens_seen": 1874395136
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1032518,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.44793438911438,
      "objective/train/theoretical_loss": 3.449101338747303,
      "objective/train/tokens_used": 244919776,
      "theoretical_loss": 3.449101338747303,
      "tokens_seen": 1874460672
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.726528647087145e-05,
      "loss": 2.5988,
      "theoretical_loss": 3.449091155852879,
      "tokens_seen": 1874526208
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.72572620767132e-05,
      "loss": 2.6024,
      "theoretical_loss": 3.4490707914310352,
      "tokens_seen": 1874657280
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.724923768255497e-05,
      "loss": 2.5618,
      "theoretical_loss": 3.4490504288316215,
      "tokens_seen": 1874788352
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.724121328839672e-05,
      "loss": 2.5314,
      "theoretical_loss": 3.4490300680543458,
      "tokens_seen": 1874919424
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.723318889423849e-05,
      "loss": 2.5469,
      "theoretical_loss": 3.4490097090989185,
      "tokens_seen": 1875050496
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.722516450008024e-05,
      "loss": 2.4655,
      "theoretical_loss": 3.448989351965049,
      "tokens_seen": 1875181568
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.7217140105922e-05,
      "loss": 2.6071,
      "theoretical_loss": 3.4489689966524475,
      "tokens_seen": 1875312640
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.720911571176376e-05,
      "loss": 2.5736,
      "theoretical_loss": 3.448948643160823,
      "tokens_seen": 1875443712
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.720109131760552e-05,
      "loss": 2.5253,
      "theoretical_loss": 3.448928291489886,
      "tokens_seen": 1875574784
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.719306692344728e-05,
      "loss": 2.6315,
      "theoretical_loss": 3.448907941639346,
      "tokens_seen": 1875705856
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.718504252928904e-05,
      "loss": 2.6466,
      "theoretical_loss": 3.448887593608914,
      "tokens_seen": 1875836928
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.71770181351308e-05,
      "loss": 2.6256,
      "theoretical_loss": 3.448867247398298,
      "tokens_seen": 1875968000
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1033022,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8692405223846436,
      "objective/train/theoretical_loss": 3.448846903007211,
      "objective/train/tokens_used": 246558176,
      "theoretical_loss": 3.448846903007211,
      "tokens_seen": 1876099072
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.716899374097256e-05,
      "loss": 2.6994,
      "theoretical_loss": 3.448846903007211,
      "tokens_seen": 1876099072
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.716096934681432e-05,
      "loss": 2.6321,
      "theoretical_loss": 3.44882656043536,
      "tokens_seen": 1876230144
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.715294495265608e-05,
      "loss": 2.5969,
      "theoretical_loss": 3.4488062196824574,
      "tokens_seen": 1876361216
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.714492055849783e-05,
      "loss": 2.6608,
      "theoretical_loss": 3.448785880748213,
      "tokens_seen": 1876492288
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.71368961643396e-05,
      "loss": 2.6761,
      "theoretical_loss": 3.4487655436323363,
      "tokens_seen": 1876623360
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.712887177018135e-05,
      "loss": 2.585,
      "theoretical_loss": 3.4487452083345396,
      "tokens_seen": 1876754432
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.712084737602312e-05,
      "loss": 2.4546,
      "theoretical_loss": 3.4487248748545314,
      "tokens_seen": 1876885504
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.711282298186487e-05,
      "loss": 2.7406,
      "theoretical_loss": 3.4487045431920236,
      "tokens_seen": 1877016576
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.710479858770664e-05,
      "loss": 2.6284,
      "theoretical_loss": 3.4486842133467266,
      "tokens_seen": 1877147648
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.709677419354839e-05,
      "loss": 2.6591,
      "theoretical_loss": 3.448663885318351,
      "tokens_seen": 1877278720
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.708874979939014e-05,
      "loss": 2.6345,
      "theoretical_loss": 3.448643559106607,
      "tokens_seen": 1877409792
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.708072540523191e-05,
      "loss": 2.6332,
      "theoretical_loss": 3.4486232347112065,
      "tokens_seen": 1877540864
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.707270101107366e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.4486029121318595,
      "tokens_seen": 1877671936
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1034169,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.514636993408203,
      "objective/train/theoretical_loss": 3.448592751523116,
      "objective/train/tokens_used": 248196576,
      "theoretical_loss": 3.448592751523116,
      "tokens_seen": 1877737472
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.706467661691543e-05,
      "loss": 2.5063,
      "theoretical_loss": 3.4485825913682775,
      "tokens_seen": 1877803008
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.705665222275718e-05,
      "loss": 2.5341,
      "theoretical_loss": 3.448562272420171,
      "tokens_seen": 1877934080
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.704862782859895e-05,
      "loss": 2.6616,
      "theoretical_loss": 3.4485419552872525,
      "tokens_seen": 1878065152
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.70406034344407e-05,
      "loss": 2.5614,
      "theoretical_loss": 3.4485216399692313,
      "tokens_seen": 1878196224
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.703257904028245e-05,
      "loss": 2.5634,
      "theoretical_loss": 3.44850132646582,
      "tokens_seen": 1878327296
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.702455464612422e-05,
      "loss": 2.6708,
      "theoretical_loss": 3.4484810147767293,
      "tokens_seen": 1878458368
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.701653025196597e-05,
      "loss": 2.7177,
      "theoretical_loss": 3.4484607049016702,
      "tokens_seen": 1878589440
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.700850585780774e-05,
      "loss": 2.6269,
      "theoretical_loss": 3.4484403968403554,
      "tokens_seen": 1878720512
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.70004814636495e-05,
      "loss": 2.6973,
      "theoretical_loss": 3.4484200905924953,
      "tokens_seen": 1878851584
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.699245706949126e-05,
      "loss": 2.6511,
      "theoretical_loss": 3.448399786157802,
      "tokens_seen": 1878982656
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.698443267533301e-05,
      "loss": 2.6892,
      "theoretical_loss": 3.448379483535987,
      "tokens_seen": 1879113728
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.697640828117477e-05,
      "loss": 2.6398,
      "theoretical_loss": 3.4483591827267617,
      "tokens_seen": 1879244800
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1034880,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.418511390686035,
      "objective/train/theoretical_loss": 3.4483388837298383,
      "objective/train/tokens_used": 249834976,
      "theoretical_loss": 3.4483388837298383,
      "tokens_seen": 1879375872
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.696838388701653e-05,
      "loss": 2.6351,
      "theoretical_loss": 3.4483388837298383,
      "tokens_seen": 1879375872
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.696035949285829e-05,
      "loss": 2.7967,
      "theoretical_loss": 3.448318586544928,
      "tokens_seen": 1879506944
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.695233509870005e-05,
      "loss": 2.7456,
      "theoretical_loss": 3.4482982911717435,
      "tokens_seen": 1879638016
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.69443107045418e-05,
      "loss": 2.6476,
      "theoretical_loss": 3.4482779976099964,
      "tokens_seen": 1879769088
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.693628631038357e-05,
      "loss": 2.56,
      "theoretical_loss": 3.448257705859399,
      "tokens_seen": 1879900160
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.692826191622533e-05,
      "loss": 2.6287,
      "theoretical_loss": 3.448237415919663,
      "tokens_seen": 1880031232
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.692023752206708e-05,
      "loss": 2.6142,
      "theoretical_loss": 3.4482171277905005,
      "tokens_seen": 1880162304
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.691221312790885e-05,
      "loss": 2.5558,
      "theoretical_loss": 3.448196841471624,
      "tokens_seen": 1880293376
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.69041887337506e-05,
      "loss": 2.4915,
      "theoretical_loss": 3.4481765569627463,
      "tokens_seen": 1880424448
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.689616433959237e-05,
      "loss": 2.5497,
      "theoretical_loss": 3.448156274263578,
      "tokens_seen": 1880555520
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.688813994543412e-05,
      "loss": 2.507,
      "theoretical_loss": 3.4481359933738336,
      "tokens_seen": 1880686592
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.688011555127589e-05,
      "loss": 2.6085,
      "theoretical_loss": 3.448115714293225,
      "tokens_seen": 1880817664
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.687209115711764e-05,
      "loss": 2.612,
      "theoretical_loss": 3.448095437021464,
      "tokens_seen": 1880948736
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1035982,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5726146697998047,
      "objective/train/theoretical_loss": 3.448085299063812,
      "objective/train/tokens_used": 251473376,
      "theoretical_loss": 3.448085299063812,
      "tokens_seen": 1881014272
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.68640667629594e-05,
      "loss": 2.4888,
      "theoretical_loss": 3.448075161558264,
      "tokens_seen": 1881079808
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.685604236880116e-05,
      "loss": 2.6203,
      "theoretical_loss": 3.448054887903337,
      "tokens_seen": 1881210880
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.684801797464291e-05,
      "loss": 2.797,
      "theoretical_loss": 3.448034616056396,
      "tokens_seen": 1881341952
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.683999358048468e-05,
      "loss": 2.6205,
      "theoretical_loss": 3.4480143460171546,
      "tokens_seen": 1881473024
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.683196918632643e-05,
      "loss": 2.4245,
      "theoretical_loss": 3.447994077785325,
      "tokens_seen": 1881604096
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.68239447921682e-05,
      "loss": 2.6329,
      "theoretical_loss": 3.4479738113606198,
      "tokens_seen": 1881735168
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.681592039800995e-05,
      "loss": 2.5879,
      "theoretical_loss": 3.4479535467427525,
      "tokens_seen": 1881866240
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.680789600385172e-05,
      "loss": 2.6507,
      "theoretical_loss": 3.4479332839314365,
      "tokens_seen": 1881997312
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.679987160969347e-05,
      "loss": 2.6719,
      "theoretical_loss": 3.447913022926384,
      "tokens_seen": 1882128384
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.679184721553522e-05,
      "loss": 2.528,
      "theoretical_loss": 3.4478927637273094,
      "tokens_seen": 1882259456
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.678382282137699e-05,
      "loss": 2.6766,
      "theoretical_loss": 3.4478725063339244,
      "tokens_seen": 1882390528
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.677579842721874e-05,
      "loss": 2.5525,
      "theoretical_loss": 3.447852250745944,
      "tokens_seen": 1882521600
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1036495,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.561272144317627,
      "objective/train/theoretical_loss": 3.447831996963081,
      "objective/train/tokens_used": 253111776,
      "theoretical_loss": 3.447831996963081,
      "tokens_seen": 1882652672
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.676777403306051e-05,
      "loss": 2.7183,
      "theoretical_loss": 3.447831996963081,
      "tokens_seen": 1882652672
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.675974963890226e-05,
      "loss": 2.7235,
      "theoretical_loss": 3.4478117449850485,
      "tokens_seen": 1882783744
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.675172524474403e-05,
      "loss": 2.5957,
      "theoretical_loss": 3.44779149481156,
      "tokens_seen": 1882914816
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.674370085058578e-05,
      "loss": 2.5508,
      "theoretical_loss": 3.44777124644233,
      "tokens_seen": 1883045888
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.673567645642754e-05,
      "loss": 2.5098,
      "theoretical_loss": 3.447750999877072,
      "tokens_seen": 1883176960
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.67276520622693e-05,
      "loss": 2.605,
      "theoretical_loss": 3.4477307551154985,
      "tokens_seen": 1883308032
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.671962766811106e-05,
      "loss": 2.7175,
      "theoretical_loss": 3.4477105121573244,
      "tokens_seen": 1883439104
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.671160327395282e-05,
      "loss": 2.5376,
      "theoretical_loss": 3.4476902710022634,
      "tokens_seen": 1883570176
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.670357887979458e-05,
      "loss": 2.5942,
      "theoretical_loss": 3.4476700316500293,
      "tokens_seen": 1883701248
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.669555448563634e-05,
      "loss": 2.6486,
      "theoretical_loss": 3.447649794100336,
      "tokens_seen": 1883832320
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.66875300914781e-05,
      "loss": 2.6782,
      "theoretical_loss": 3.447629558352898,
      "tokens_seen": 1883963392
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.667950569731985e-05,
      "loss": 2.6249,
      "theoretical_loss": 3.447609324407429,
      "tokens_seen": 1884094464
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.667148130316162e-05,
      "loss": 2.5889,
      "theoretical_loss": 3.4475890922636436,
      "tokens_seen": 1884225536
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1037807,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6240508556365967,
      "objective/train/theoretical_loss": 3.447578976867293,
      "objective/train/tokens_used": 254750176,
      "theoretical_loss": 3.447578976867293,
      "tokens_seen": 1884291072
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.666345690900337e-05,
      "loss": 2.5063,
      "theoretical_loss": 3.4475688619212557,
      "tokens_seen": 1884356608
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.665543251484514e-05,
      "loss": 2.8423,
      "theoretical_loss": 3.4475486333799794,
      "tokens_seen": 1884487680
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.664740812068689e-05,
      "loss": 2.6742,
      "theoretical_loss": 3.4475284066395293,
      "tokens_seen": 1884618752
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.663938372652866e-05,
      "loss": 2.7974,
      "theoretical_loss": 3.4475081816996207,
      "tokens_seen": 1884749824
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.663135933237041e-05,
      "loss": 2.5465,
      "theoretical_loss": 3.4474879585599667,
      "tokens_seen": 1884880896
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.662333493821216e-05,
      "loss": 2.6038,
      "theoretical_loss": 3.4474677372202827,
      "tokens_seen": 1885011968
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.661531054405393e-05,
      "loss": 2.6861,
      "theoretical_loss": 3.4474475176802835,
      "tokens_seen": 1885143040
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.660728614989568e-05,
      "loss": 2.4443,
      "theoretical_loss": 3.447427299939683,
      "tokens_seen": 1885274112
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.659926175573745e-05,
      "loss": 2.4831,
      "theoretical_loss": 3.4474070839981974,
      "tokens_seen": 1885405184
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.65912373615792e-05,
      "loss": 2.6502,
      "theoretical_loss": 3.4473868698555403,
      "tokens_seen": 1885536256
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.658321296742097e-05,
      "loss": 2.6768,
      "theoretical_loss": 3.4473666575114263,
      "tokens_seen": 1885667328
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.657518857326272e-05,
      "loss": 2.5946,
      "theoretical_loss": 3.4473464469655717,
      "tokens_seen": 1885798400
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1038424,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4946327209472656,
      "objective/train/theoretical_loss": 3.4473262382176912,
      "objective/train/tokens_used": 256388576,
      "theoretical_loss": 3.4473262382176912,
      "tokens_seen": 1885929472
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.656716417910447e-05,
      "loss": 2.6475,
      "theoretical_loss": 3.4473262382176912,
      "tokens_seen": 1885929472
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.655913978494624e-05,
      "loss": 2.5066,
      "theoretical_loss": 3.447306031267499,
      "tokens_seen": 1886060544
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.6551115390788e-05,
      "loss": 2.7395,
      "theoretical_loss": 3.4472858261147112,
      "tokens_seen": 1886191616
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.654309099662976e-05,
      "loss": 2.6795,
      "theoretical_loss": 3.4472656227590424,
      "tokens_seen": 1886322688
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.653506660247151e-05,
      "loss": 2.8047,
      "theoretical_loss": 3.4472454212002086,
      "tokens_seen": 1886453760
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.652704220831328e-05,
      "loss": 2.4961,
      "theoretical_loss": 3.4472252214379244,
      "tokens_seen": 1886584832
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.651901781415503e-05,
      "loss": 2.5358,
      "theoretical_loss": 3.447205023471906,
      "tokens_seen": 1886715904
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.65109934199968e-05,
      "loss": 2.6046,
      "theoretical_loss": 3.4471848273018684,
      "tokens_seen": 1886846976
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.650296902583855e-05,
      "loss": 2.5741,
      "theoretical_loss": 3.447164632927527,
      "tokens_seen": 1886978048
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.649494463168031e-05,
      "loss": 2.6877,
      "theoretical_loss": 3.4471444403485982,
      "tokens_seen": 1887109120
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.648692023752207e-05,
      "loss": 2.652,
      "theoretical_loss": 3.447124249564797,
      "tokens_seen": 1887240192
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.647889584336383e-05,
      "loss": 2.5121,
      "theoretical_loss": 3.447104060575839,
      "tokens_seen": 1887371264
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.64708714492056e-05,
      "loss": 2.5847,
      "theoretical_loss": 3.44708387338144,
      "tokens_seen": 1887502336
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1039538,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6100270748138428,
      "objective/train/theoretical_loss": 3.447073780457112,
      "objective/train/tokens_used": 258026976,
      "theoretical_loss": 3.447073780457112,
      "tokens_seen": 1887567872
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.646284705504735e-05,
      "loss": 2.7654,
      "theoretical_loss": 3.4470636879813172,
      "tokens_seen": 1887633408
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.645482266088911e-05,
      "loss": 2.5572,
      "theoretical_loss": 3.447043504375185,
      "tokens_seen": 1887764480
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.644679826673087e-05,
      "loss": 2.5704,
      "theoretical_loss": 3.4470233225627602,
      "tokens_seen": 1887895552
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.643877387257262e-05,
      "loss": 2.6712,
      "theoretical_loss": 3.4470031425437586,
      "tokens_seen": 1888026624
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.643074947841439e-05,
      "loss": 2.435,
      "theoretical_loss": 3.4469829643178964,
      "tokens_seen": 1888157696
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.642272508425614e-05,
      "loss": 2.7374,
      "theoretical_loss": 3.44696278788489,
      "tokens_seen": 1888288768
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.64147006900979e-05,
      "loss": 2.6685,
      "theoretical_loss": 3.446942613244455,
      "tokens_seen": 1888419840
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.640667629593966e-05,
      "loss": 2.6827,
      "theoretical_loss": 3.4469224403963086,
      "tokens_seen": 1888550912
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.639865190178143e-05,
      "loss": 2.5129,
      "theoretical_loss": 3.4469022693401667,
      "tokens_seen": 1888681984
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.639062750762318e-05,
      "loss": 2.7268,
      "theoretical_loss": 3.4468821000757464,
      "tokens_seen": 1888813056
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.638260311346493e-05,
      "loss": 2.5836,
      "theoretical_loss": 3.446861932602763,
      "tokens_seen": 1888944128
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.63745787193067e-05,
      "loss": 2.5723,
      "theoretical_loss": 3.446841766920934,
      "tokens_seen": 1889075200
    },
    {
      "epoch": 0.14,
      "objective/train/docs_used": 1040979,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.302525758743286,
      "objective/train/theoretical_loss": 3.446821603029976,
      "objective/train/tokens_used": 259665376,
      "theoretical_loss": 3.446821603029976,
      "tokens_seen": 1889206272
    },
    {
      "epoch": 0.14,
      "learning_rate": 8.636655432514845e-05,
      "loss": 2.5104,
      "theoretical_loss": 3.446821603029976,
      "tokens_seen": 1889206272
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.635852993099022e-05,
      "loss": 2.7808,
      "theoretical_loss": 3.446801440929606,
      "tokens_seen": 1889337344
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.635050553683197e-05,
      "loss": 2.6051,
      "theoretical_loss": 3.4467812806195397,
      "tokens_seen": 1889468416
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.634248114267374e-05,
      "loss": 2.4754,
      "theoretical_loss": 3.4467611220994954,
      "tokens_seen": 1889599488
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.633445674851549e-05,
      "loss": 2.5734,
      "theoretical_loss": 3.4467409653691887,
      "tokens_seen": 1889730560
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.632643235435724e-05,
      "loss": 2.636,
      "theoretical_loss": 3.4467208104283373,
      "tokens_seen": 1889861632
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.631840796019901e-05,
      "loss": 2.5367,
      "theoretical_loss": 3.446700657276658,
      "tokens_seen": 1889992704
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.631038356604076e-05,
      "loss": 2.6323,
      "theoretical_loss": 3.446680505913868,
      "tokens_seen": 1890123776
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.630235917188253e-05,
      "loss": 2.6413,
      "theoretical_loss": 3.4466603563396845,
      "tokens_seen": 1890254848
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.629433477772428e-05,
      "loss": 2.6791,
      "theoretical_loss": 3.4466402085538252,
      "tokens_seen": 1890385920
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.628631038356605e-05,
      "loss": 2.63,
      "theoretical_loss": 3.4466200625560064,
      "tokens_seen": 1890516992
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.62782859894078e-05,
      "loss": 2.6905,
      "theoretical_loss": 3.446599918345946,
      "tokens_seen": 1890648064
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.627026159524956e-05,
      "loss": 2.6284,
      "theoretical_loss": 3.446579775923362,
      "tokens_seen": 1890779136
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1041612,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.944444417953491,
      "objective/train/theoretical_loss": 3.4465697053822844,
      "objective/train/tokens_used": 261303776,
      "theoretical_loss": 3.4465697053822844,
      "tokens_seen": 1890844672
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.626223720109132e-05,
      "loss": 2.6467,
      "theoretical_loss": 3.4465596352879704,
      "tokens_seen": 1890910208
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.625421280693308e-05,
      "loss": 2.4845,
      "theoretical_loss": 3.44653949643949,
      "tokens_seen": 1891041280
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.624618841277484e-05,
      "loss": 2.5807,
      "theoretical_loss": 3.446519359377638,
      "tokens_seen": 1891172352
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.62381640186166e-05,
      "loss": 2.6668,
      "theoretical_loss": 3.4464992241021326,
      "tokens_seen": 1891303424
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.623013962445836e-05,
      "loss": 2.4901,
      "theoretical_loss": 3.4464790906126908,
      "tokens_seen": 1891434496
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.622211523030012e-05,
      "loss": 2.7128,
      "theoretical_loss": 3.446458958909031,
      "tokens_seen": 1891565568
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.621409083614188e-05,
      "loss": 2.6439,
      "theoretical_loss": 3.4464388289908703,
      "tokens_seen": 1891696640
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.620606644198364e-05,
      "loss": 2.6116,
      "theoretical_loss": 3.4464187008579277,
      "tokens_seen": 1891827712
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.619804204782539e-05,
      "loss": 2.5137,
      "theoretical_loss": 3.446398574509921,
      "tokens_seen": 1891958784
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.619001765366716e-05,
      "loss": 2.5921,
      "theoretical_loss": 3.4463784499465673,
      "tokens_seen": 1892089856
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.618199325950891e-05,
      "loss": 2.5911,
      "theoretical_loss": 3.4463583271675855,
      "tokens_seen": 1892220928
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.617396886535068e-05,
      "loss": 2.5032,
      "theoretical_loss": 3.446338206172694,
      "tokens_seen": 1892352000
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1042686,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5551538467407227,
      "objective/train/theoretical_loss": 3.44631808696161,
      "objective/train/tokens_used": 262942176,
      "theoretical_loss": 3.44631808696161,
      "tokens_seen": 1892483072
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.616594447119243e-05,
      "loss": 2.7494,
      "theoretical_loss": 3.44631808696161,
      "tokens_seen": 1892483072
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.61579200770342e-05,
      "loss": 2.5324,
      "theoretical_loss": 3.4462979695340534,
      "tokens_seen": 1892614144
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.614989568287595e-05,
      "loss": 2.5766,
      "theoretical_loss": 3.446277853889742,
      "tokens_seen": 1892745216
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.61418712887177e-05,
      "loss": 2.5912,
      "theoretical_loss": 3.4462577400283934,
      "tokens_seen": 1892876288
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.613384689455947e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.446237627949727,
      "tokens_seen": 1893007360
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.612582250040122e-05,
      "loss": 2.7351,
      "theoretical_loss": 3.4462175176534613,
      "tokens_seen": 1893138432
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.611779810624299e-05,
      "loss": 2.4694,
      "theoretical_loss": 3.446197409139314,
      "tokens_seen": 1893269504
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.610977371208474e-05,
      "loss": 2.7383,
      "theoretical_loss": 3.4461773024070057,
      "tokens_seen": 1893400576
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.610174931792651e-05,
      "loss": 2.2325,
      "theoretical_loss": 3.4461571974562535,
      "tokens_seen": 1893531648
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.609372492376826e-05,
      "loss": 2.5378,
      "theoretical_loss": 3.446137094286777,
      "tokens_seen": 1893662720
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.608570052961001e-05,
      "loss": 2.6805,
      "theoretical_loss": 3.446116992898295,
      "tokens_seen": 1893793792
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.607767613545178e-05,
      "loss": 2.5466,
      "theoretical_loss": 3.4460968932905263,
      "tokens_seen": 1893924864
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.606965174129353e-05,
      "loss": 2.6135,
      "theoretical_loss": 3.44607679546319,
      "tokens_seen": 1894055936
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1043318,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7247536182403564,
      "objective/train/theoretical_loss": 3.4460667472170963,
      "objective/train/tokens_used": 264580576,
      "theoretical_loss": 3.4460667472170963,
      "tokens_seen": 1894121472
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.60616273471353e-05,
      "loss": 2.5337,
      "theoretical_loss": 3.446056699416005,
      "tokens_seen": 1894187008
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.605360295297705e-05,
      "loss": 2.5264,
      "theoretical_loss": 3.4460366051486915,
      "tokens_seen": 1894318080
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.604557855881882e-05,
      "loss": 2.5163,
      "theoretical_loss": 3.446016512660967,
      "tokens_seen": 1894449152
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.603755416466057e-05,
      "loss": 2.7092,
      "theoretical_loss": 3.445996421952552,
      "tokens_seen": 1894580224
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.602952977050233e-05,
      "loss": 2.5607,
      "theoretical_loss": 3.445976333023166,
      "tokens_seen": 1894711296
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.60215053763441e-05,
      "loss": 2.5803,
      "theoretical_loss": 3.4459562458725275,
      "tokens_seen": 1894842368
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.601348098218585e-05,
      "loss": 2.594,
      "theoretical_loss": 3.4459361605003567,
      "tokens_seen": 1894973440
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.600545658802761e-05,
      "loss": 2.8107,
      "theoretical_loss": 3.4459160769063724,
      "tokens_seen": 1895104512
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.599743219386937e-05,
      "loss": 2.5098,
      "theoretical_loss": 3.4458959950902948,
      "tokens_seen": 1895235584
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.598940779971113e-05,
      "loss": 2.6863,
      "theoretical_loss": 3.445875915051844,
      "tokens_seen": 1895366656
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.598138340555289e-05,
      "loss": 2.6153,
      "theoretical_loss": 3.445855836790739,
      "tokens_seen": 1895497728
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.597335901139464e-05,
      "loss": 2.2966,
      "theoretical_loss": 3.4458357603066996,
      "tokens_seen": 1895628800
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1044449,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.796210289001465,
      "objective/train/theoretical_loss": 3.4458156855994457,
      "objective/train/tokens_used": 266218976,
      "theoretical_loss": 3.4458156855994457,
      "tokens_seen": 1895759872
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.59653346172364e-05,
      "loss": 2.6243,
      "theoretical_loss": 3.4458156855994457,
      "tokens_seen": 1895759872
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.595731022307816e-05,
      "loss": 2.5802,
      "theoretical_loss": 3.4457956126686975,
      "tokens_seen": 1895890944
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.594928582891993e-05,
      "loss": 2.5255,
      "theoretical_loss": 3.445775541514175,
      "tokens_seen": 1896022016
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.594126143476168e-05,
      "loss": 2.5969,
      "theoretical_loss": 3.4457554721355983,
      "tokens_seen": 1896153088
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.593323704060345e-05,
      "loss": 2.5436,
      "theoretical_loss": 3.4457354045326873,
      "tokens_seen": 1896284160
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.59252126464452e-05,
      "loss": 2.5276,
      "theoretical_loss": 3.4457153387051616,
      "tokens_seen": 1896415232
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.591718825228697e-05,
      "loss": 2.4972,
      "theoretical_loss": 3.4456952746527425,
      "tokens_seen": 1896546304
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.590916385812872e-05,
      "loss": 2.7186,
      "theoretical_loss": 3.44567521237515,
      "tokens_seen": 1896677376
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.590113946397047e-05,
      "loss": 2.5623,
      "theoretical_loss": 3.4456551518721037,
      "tokens_seen": 1896808448
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.589311506981224e-05,
      "loss": 2.7065,
      "theoretical_loss": 3.4456350931433253,
      "tokens_seen": 1896939520
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.588509067565399e-05,
      "loss": 2.5867,
      "theoretical_loss": 3.445615036188534,
      "tokens_seen": 1897070592
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.587706628149576e-05,
      "loss": 2.5036,
      "theoretical_loss": 3.4455949810074515,
      "tokens_seen": 1897201664
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.586904188733751e-05,
      "loss": 2.444,
      "theoretical_loss": 3.4455749275997976,
      "tokens_seen": 1897332736
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1045116,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.431305408477783,
      "objective/train/theoretical_loss": 3.4455649015609193,
      "objective/train/tokens_used": 267857376,
      "theoretical_loss": 3.4455649015609193,
      "tokens_seen": 1897398272
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.586101749317928e-05,
      "loss": 2.4496,
      "theoretical_loss": 3.4455548759652936,
      "tokens_seen": 1897463808
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.585299309902103e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.4455348261036596,
      "tokens_seen": 1897594880
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.584496870486278e-05,
      "loss": 2.5878,
      "theoretical_loss": 3.4455147780146174,
      "tokens_seen": 1897725952
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.583694431070455e-05,
      "loss": 2.5,
      "theoretical_loss": 3.4454947316978863,
      "tokens_seen": 1897857024
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.58289199165463e-05,
      "loss": 2.6347,
      "theoretical_loss": 3.445474687153189,
      "tokens_seen": 1897988096
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.582089552238807e-05,
      "loss": 2.5567,
      "theoretical_loss": 3.4454546443802454,
      "tokens_seen": 1898119168
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.581287112822982e-05,
      "loss": 2.5346,
      "theoretical_loss": 3.445434603378777,
      "tokens_seen": 1898250240
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.580484673407159e-05,
      "loss": 2.698,
      "theoretical_loss": 3.4454145641485043,
      "tokens_seen": 1898381312
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.579682233991334e-05,
      "loss": 2.5593,
      "theoretical_loss": 3.4453945266891486,
      "tokens_seen": 1898512384
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.57887979457551e-05,
      "loss": 2.5686,
      "theoretical_loss": 3.4453744910004325,
      "tokens_seen": 1898643456
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.578077355159686e-05,
      "loss": 2.5992,
      "theoretical_loss": 3.4453544570820758,
      "tokens_seen": 1898774528
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.577274915743862e-05,
      "loss": 2.5495,
      "theoretical_loss": 3.4453344249338005,
      "tokens_seen": 1898905600
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1046217,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.568467855453491,
      "objective/train/theoretical_loss": 3.4453143945553277,
      "objective/train/tokens_used": 269495776,
      "theoretical_loss": 3.4453143945553277,
      "tokens_seen": 1899036672
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.576472476328038e-05,
      "loss": 2.6635,
      "theoretical_loss": 3.4453143945553277,
      "tokens_seen": 1899036672
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.575670036912214e-05,
      "loss": 2.591,
      "theoretical_loss": 3.4452943659463795,
      "tokens_seen": 1899167744
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.57486759749639e-05,
      "loss": 2.6946,
      "theoretical_loss": 3.445274339106677,
      "tokens_seen": 1899298816
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.574065158080566e-05,
      "loss": 2.5766,
      "theoretical_loss": 3.4452543140359415,
      "tokens_seen": 1899429888
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.573262718664741e-05,
      "loss": 2.6189,
      "theoretical_loss": 3.4452342907338958,
      "tokens_seen": 1899560960
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.572460279248918e-05,
      "loss": 2.556,
      "theoretical_loss": 3.4452142692002603,
      "tokens_seen": 1899692032
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.571657839833093e-05,
      "loss": 2.5013,
      "theoretical_loss": 3.445194249434758,
      "tokens_seen": 1899823104
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.57085540041727e-05,
      "loss": 2.8092,
      "theoretical_loss": 3.4451742314371105,
      "tokens_seen": 1899954176
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.570052961001445e-05,
      "loss": 2.6131,
      "theoretical_loss": 3.4451542152070393,
      "tokens_seen": 1900085248
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.569250521585622e-05,
      "loss": 2.5599,
      "theoretical_loss": 3.445134200744267,
      "tokens_seen": 1900216320
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.568448082169797e-05,
      "loss": 2.4855,
      "theoretical_loss": 3.4451141880485148,
      "tokens_seen": 1900347392
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.567645642753972e-05,
      "loss": 2.5664,
      "theoretical_loss": 3.445094177119505,
      "tokens_seen": 1900478464
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.566843203338149e-05,
      "loss": 2.6017,
      "theoretical_loss": 3.4450741679569608,
      "tokens_seen": 1900609536
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1047224,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.664011001586914,
      "objective/train/theoretical_loss": 3.445064164038026,
      "objective/train/tokens_used": 271134176,
      "theoretical_loss": 3.445064164038026,
      "tokens_seen": 1900675072
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.566040763922324e-05,
      "loss": 2.6229,
      "theoretical_loss": 3.4450541605606038,
      "tokens_seen": 1900740608
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.565238324506501e-05,
      "loss": 2.5607,
      "theoretical_loss": 3.4450341549301564,
      "tokens_seen": 1900871680
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.564435885090676e-05,
      "loss": 2.4586,
      "theoretical_loss": 3.4450141510653403,
      "tokens_seen": 1901002752
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.563633445674853e-05,
      "loss": 2.4271,
      "theoretical_loss": 3.444994148965879,
      "tokens_seen": 1901133824
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.562831006259028e-05,
      "loss": 2.6382,
      "theoretical_loss": 3.444974148631495,
      "tokens_seen": 1901264896
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.562028566843204e-05,
      "loss": 2.6207,
      "theoretical_loss": 3.4449541500619096,
      "tokens_seen": 1901395968
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.56122612742738e-05,
      "loss": 2.6704,
      "theoretical_loss": 3.4449341532568467,
      "tokens_seen": 1901527040
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.560423688011555e-05,
      "loss": 2.7467,
      "theoretical_loss": 3.4449141582160285,
      "tokens_seen": 1901658112
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.559621248595732e-05,
      "loss": 2.6002,
      "theoretical_loss": 3.4448941649391784,
      "tokens_seen": 1901789184
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.558818809179907e-05,
      "loss": 2.6505,
      "theoretical_loss": 3.444874173426018,
      "tokens_seen": 1901920256
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.558016369764084e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.444854183676271,
      "tokens_seen": 1902051328
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.55721393034826e-05,
      "loss": 2.5589,
      "theoretical_loss": 3.44483419568966,
      "tokens_seen": 1902182400
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1047772,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.691408395767212,
      "objective/train/theoretical_loss": 3.444814209465909,
      "objective/train/tokens_used": 272772576,
      "theoretical_loss": 3.444814209465909,
      "tokens_seen": 1902313472
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.556411490932436e-05,
      "loss": 2.6686,
      "theoretical_loss": 3.444814209465909,
      "tokens_seen": 1902313472
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.555609051516611e-05,
      "loss": 2.7427,
      "theoretical_loss": 3.44479422500474,
      "tokens_seen": 1902444544
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.554806612100787e-05,
      "loss": 2.5318,
      "theoretical_loss": 3.4447742423058765,
      "tokens_seen": 1902575616
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.554004172684963e-05,
      "loss": 2.5627,
      "theoretical_loss": 3.4447542613690416,
      "tokens_seen": 1902706688
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.553201733269139e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.4447342821939584,
      "tokens_seen": 1902837760
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.552399293853315e-05,
      "loss": 2.651,
      "theoretical_loss": 3.4447143047803506,
      "tokens_seen": 1902968832
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.551596854437491e-05,
      "loss": 2.5404,
      "theoretical_loss": 3.444694329127941,
      "tokens_seen": 1903099904
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.550794415021667e-05,
      "loss": 2.6022,
      "theoretical_loss": 3.4446743552364545,
      "tokens_seen": 1903230976
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.549991975605843e-05,
      "loss": 2.5531,
      "theoretical_loss": 3.444654383105613,
      "tokens_seen": 1903362048
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.549189536190018e-05,
      "loss": 2.6862,
      "theoretical_loss": 3.44463441273514,
      "tokens_seen": 1903493120
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.548387096774195e-05,
      "loss": 2.7709,
      "theoretical_loss": 3.4446144441247606,
      "tokens_seen": 1903624192
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.54758465735837e-05,
      "loss": 2.5729,
      "theoretical_loss": 3.4445944772741974,
      "tokens_seen": 1903755264
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.546782217942547e-05,
      "loss": 2.6571,
      "theoretical_loss": 3.444574512183175,
      "tokens_seen": 1903886336
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1048883,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7661850452423096,
      "objective/train/theoretical_loss": 3.444564530297405,
      "objective/train/tokens_used": 274410976,
      "theoretical_loss": 3.444564530297405,
      "tokens_seen": 1903951872
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.545979778526722e-05,
      "loss": 2.6316,
      "theoretical_loss": 3.444554548851416,
      "tokens_seen": 1904017408
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.545177339110899e-05,
      "loss": 2.3924,
      "theoretical_loss": 3.4445345872786453,
      "tokens_seen": 1904148480
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.544374899695074e-05,
      "loss": 2.6117,
      "theoretical_loss": 3.444514627464587,
      "tokens_seen": 1904279552
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.543572460279249e-05,
      "loss": 2.5,
      "theoretical_loss": 3.4444946694089635,
      "tokens_seen": 1904410624
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.542770020863426e-05,
      "loss": 2.6789,
      "theoretical_loss": 3.444474713111501,
      "tokens_seen": 1904541696
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.541967581447601e-05,
      "loss": 2.5173,
      "theoretical_loss": 3.444454758571922,
      "tokens_seen": 1904672768
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.541165142031778e-05,
      "loss": 2.6405,
      "theoretical_loss": 3.444434805789952,
      "tokens_seen": 1904803840
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.540362702615953e-05,
      "loss": 2.5492,
      "theoretical_loss": 3.444414854765314,
      "tokens_seen": 1904934912
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.53956026320013e-05,
      "loss": 2.4519,
      "theoretical_loss": 3.4443949054977336,
      "tokens_seen": 1905065984
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.538757823784305e-05,
      "loss": 2.4236,
      "theoretical_loss": 3.444374957986934,
      "tokens_seen": 1905197056
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.53795538436848e-05,
      "loss": 2.5186,
      "theoretical_loss": 3.4443550122326405,
      "tokens_seen": 1905328128
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.537152944952657e-05,
      "loss": 2.5571,
      "theoretical_loss": 3.444335068234577,
      "tokens_seen": 1905459200
    },
    {
      "epoch": 0.15,
      "objective/train/docs_used": 1049481,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8444793224334717,
      "objective/train/theoretical_loss": 3.4443151259924685,
      "objective/train/tokens_used": 276049376,
      "theoretical_loss": 3.4443151259924685,
      "tokens_seen": 1905590272
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.536350505536833e-05,
      "loss": 2.7023,
      "theoretical_loss": 3.4443151259924685,
      "tokens_seen": 1905590272
    },
    {
      "epoch": 0.15,
      "learning_rate": 8.535548066121009e-05,
      "loss": 2.5393,
      "theoretical_loss": 3.4442951855060393,
      "tokens_seen": 1905721344
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.534745626705184e-05,
      "loss": 2.5215,
      "theoretical_loss": 3.444275246775015,
      "tokens_seen": 1905852416
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.533943187289361e-05,
      "loss": 2.6303,
      "theoretical_loss": 3.444255309799119,
      "tokens_seen": 1905983488
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.533140747873536e-05,
      "loss": 2.6693,
      "theoretical_loss": 3.4442353745780765,
      "tokens_seen": 1906114560
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.532338308457712e-05,
      "loss": 2.6018,
      "theoretical_loss": 3.444215441111613,
      "tokens_seen": 1906245632
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.531535869041888e-05,
      "loss": 2.6154,
      "theoretical_loss": 3.444195509399453,
      "tokens_seen": 1906376704
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.530733429626064e-05,
      "loss": 2.6642,
      "theoretical_loss": 3.444175579441322,
      "tokens_seen": 1906507776
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.52993099021024e-05,
      "loss": 2.5281,
      "theoretical_loss": 3.4441556512369442,
      "tokens_seen": 1906638848
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.529128550794416e-05,
      "loss": 2.5748,
      "theoretical_loss": 3.4441357247860456,
      "tokens_seen": 1906769920
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.528326111378591e-05,
      "loss": 2.4962,
      "theoretical_loss": 3.4441158000883507,
      "tokens_seen": 1906900992
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.527523671962768e-05,
      "loss": 2.6538,
      "theoretical_loss": 3.4440958771435852,
      "tokens_seen": 1907032064
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.526721232546943e-05,
      "loss": 2.7252,
      "theoretical_loss": 3.444075955951474,
      "tokens_seen": 1907163136
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1050597,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.0899901390075684,
      "objective/train/theoretical_loss": 3.4440659960125783,
      "objective/train/tokens_used": 277687776,
      "theoretical_loss": 3.4440659960125783,
      "tokens_seen": 1907228672
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.52591879313112e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.4440560365117427,
      "tokens_seen": 1907294208
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.525116353715295e-05,
      "loss": 2.5941,
      "theoretical_loss": 3.4440361188241173,
      "tokens_seen": 1907425280
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.524313914299472e-05,
      "loss": 2.58,
      "theoretical_loss": 3.4440162028883226,
      "tokens_seen": 1907556352
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.523511474883647e-05,
      "loss": 2.3977,
      "theoretical_loss": 3.443996288704084,
      "tokens_seen": 1907687424
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.522709035467822e-05,
      "loss": 2.5847,
      "theoretical_loss": 3.443976376271128,
      "tokens_seen": 1907818496
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.521906596051999e-05,
      "loss": 2.6012,
      "theoretical_loss": 3.4439564655891797,
      "tokens_seen": 1907949568
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.521104156636174e-05,
      "loss": 2.6382,
      "theoretical_loss": 3.4439365566579645,
      "tokens_seen": 1908080640
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.520301717220351e-05,
      "loss": 2.4888,
      "theoretical_loss": 3.4439166494772087,
      "tokens_seen": 1908211712
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.519499277804526e-05,
      "loss": 2.7036,
      "theoretical_loss": 3.4438967440466386,
      "tokens_seen": 1908342784
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.518696838388702e-05,
      "loss": 2.6248,
      "theoretical_loss": 3.443876840365979,
      "tokens_seen": 1908473856
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.517894398972878e-05,
      "loss": 2.6627,
      "theoretical_loss": 3.443856938434957,
      "tokens_seen": 1908604928
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.517091959557054e-05,
      "loss": 2.5851,
      "theoretical_loss": 3.443837038253298,
      "tokens_seen": 1908736000
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1051040,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.524304151535034,
      "objective/train/theoretical_loss": 3.4438171398207285,
      "objective/train/tokens_used": 279326176,
      "theoretical_loss": 3.4438171398207285,
      "tokens_seen": 1908867072
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.51628952014123e-05,
      "loss": 2.5677,
      "theoretical_loss": 3.4438171398207285,
      "tokens_seen": 1908867072
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.515487080725406e-05,
      "loss": 2.5599,
      "theoretical_loss": 3.4437972431369746,
      "tokens_seen": 1908998144
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.514684641309582e-05,
      "loss": 2.7516,
      "theoretical_loss": 3.443777348201762,
      "tokens_seen": 1909129216
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.513882201893758e-05,
      "loss": 2.4558,
      "theoretical_loss": 3.4437574550148176,
      "tokens_seen": 1909260288
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.513079762477933e-05,
      "loss": 2.6508,
      "theoretical_loss": 3.4437375635758674,
      "tokens_seen": 1909391360
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.51227732306211e-05,
      "loss": 2.6752,
      "theoretical_loss": 3.4437176738846382,
      "tokens_seen": 1909522432
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.511474883646285e-05,
      "loss": 2.6269,
      "theoretical_loss": 3.4436977859408566,
      "tokens_seen": 1909653504
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.510672444230461e-05,
      "loss": 2.6323,
      "theoretical_loss": 3.443677899744249,
      "tokens_seen": 1909784576
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.509870004814637e-05,
      "loss": 2.6092,
      "theoretical_loss": 3.4436580152945413,
      "tokens_seen": 1909915648
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.509067565398812e-05,
      "loss": 2.6991,
      "theoretical_loss": 3.4436381325914613,
      "tokens_seen": 1910046720
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.508265125982989e-05,
      "loss": 2.7202,
      "theoretical_loss": 3.443618251634735,
      "tokens_seen": 1910177792
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.507462686567164e-05,
      "loss": 2.397,
      "theoretical_loss": 3.443598372424089,
      "tokens_seen": 1910308864
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.506660247151341e-05,
      "loss": 2.4573,
      "theoretical_loss": 3.443578494959251,
      "tokens_seen": 1910439936
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1051849,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3590188026428223,
      "objective/train/theoretical_loss": 3.4435685568814245,
      "objective/train/tokens_used": 280964576,
      "theoretical_loss": 3.4435685568814245,
      "tokens_seen": 1910505472
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.505857807735516e-05,
      "loss": 2.6063,
      "theoretical_loss": 3.4435586192399477,
      "tokens_seen": 1910571008
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.505055368319693e-05,
      "loss": 2.6925,
      "theoretical_loss": 3.443538745265905,
      "tokens_seen": 1910702080
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.504252928903868e-05,
      "loss": 2.6257,
      "theoretical_loss": 3.443518873036852,
      "tokens_seen": 1910833152
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.503450489488043e-05,
      "loss": 2.5778,
      "theoretical_loss": 3.443499002552514,
      "tokens_seen": 1910964224
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.50264805007222e-05,
      "loss": 2.5841,
      "theoretical_loss": 3.4434791338126187,
      "tokens_seen": 1911095296
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.501845610656395e-05,
      "loss": 2.5606,
      "theoretical_loss": 3.4434592668168933,
      "tokens_seen": 1911226368
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.501043171240572e-05,
      "loss": 2.7309,
      "theoretical_loss": 3.4434394015650653,
      "tokens_seen": 1911357440
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.500240731824747e-05,
      "loss": 2.6035,
      "theoretical_loss": 3.443419538056862,
      "tokens_seen": 1911488512
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.499438292408923e-05,
      "loss": 2.5485,
      "theoretical_loss": 3.4433996762920107,
      "tokens_seen": 1911619584
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.498635852993099e-05,
      "loss": 2.6219,
      "theoretical_loss": 3.443379816270239,
      "tokens_seen": 1911750656
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.497833413577275e-05,
      "loss": 2.6035,
      "theoretical_loss": 3.443359957991275,
      "tokens_seen": 1911881728
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.497030974161451e-05,
      "loss": 2.4346,
      "theoretical_loss": 3.4433401014548446,
      "tokens_seen": 1912012800
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1052490,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.299638032913208,
      "objective/train/theoretical_loss": 3.4433202466606767,
      "objective/train/tokens_used": 282602976,
      "theoretical_loss": 3.4433202466606767,
      "tokens_seen": 1912143872
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.496228534745627e-05,
      "loss": 2.6254,
      "theoretical_loss": 3.4433202466606767,
      "tokens_seen": 1912143872
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.495426095329802e-05,
      "loss": 2.4929,
      "theoretical_loss": 3.4433003936084994,
      "tokens_seen": 1912274944
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.494623655913979e-05,
      "loss": 2.5459,
      "theoretical_loss": 3.44328054229804,
      "tokens_seen": 1912406016
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.493821216498154e-05,
      "loss": 2.4689,
      "theoretical_loss": 3.4432606927290257,
      "tokens_seen": 1912537088
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.49301877708233e-05,
      "loss": 2.6596,
      "theoretical_loss": 3.443240844901185,
      "tokens_seen": 1912668160
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.492216337666506e-05,
      "loss": 2.5501,
      "theoretical_loss": 3.4432209988142457,
      "tokens_seen": 1912799232
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.491413898250683e-05,
      "loss": 2.5588,
      "theoretical_loss": 3.4432011544679364,
      "tokens_seen": 1912930304
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.490611458834858e-05,
      "loss": 2.631,
      "theoretical_loss": 3.4431813118619847,
      "tokens_seen": 1913061376
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.489809019419033e-05,
      "loss": 2.6742,
      "theoretical_loss": 3.4431614709961185,
      "tokens_seen": 1913192448
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.48900658000321e-05,
      "loss": 2.4847,
      "theoretical_loss": 3.4431416318700663,
      "tokens_seen": 1913323520
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.488204140587385e-05,
      "loss": 2.4341,
      "theoretical_loss": 3.443121794483557,
      "tokens_seen": 1913454592
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.487401701171562e-05,
      "loss": 2.7514,
      "theoretical_loss": 3.4431019588363174,
      "tokens_seen": 1913585664
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.486599261755737e-05,
      "loss": 2.5306,
      "theoretical_loss": 3.443082124928077,
      "tokens_seen": 1913716736
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1053877,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7644686698913574,
      "objective/train/theoretical_loss": 3.443072208625997,
      "objective/train/tokens_used": 284241376,
      "theoretical_loss": 3.443072208625997,
      "tokens_seen": 1913782272
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.485796822339914e-05,
      "loss": 2.6458,
      "theoretical_loss": 3.4430622927585643,
      "tokens_seen": 1913847808
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.484994382924089e-05,
      "loss": 2.5744,
      "theoretical_loss": 3.4430424623275075,
      "tokens_seen": 1913978880
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.484191943508264e-05,
      "loss": 2.5332,
      "theoretical_loss": 3.443022633634635,
      "tokens_seen": 1914109952
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.483389504092441e-05,
      "loss": 2.6024,
      "theoretical_loss": 3.443002806679676,
      "tokens_seen": 1914241024
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.482587064676616e-05,
      "loss": 2.6787,
      "theoretical_loss": 3.4429829814623583,
      "tokens_seen": 1914372096
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.481784625260793e-05,
      "loss": 2.4463,
      "theoretical_loss": 3.442963157982412,
      "tokens_seen": 1914503168
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.480982185844968e-05,
      "loss": 2.6267,
      "theoretical_loss": 3.4429433362395643,
      "tokens_seen": 1914634240
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.480179746429145e-05,
      "loss": 2.3903,
      "theoretical_loss": 3.442923516233545,
      "tokens_seen": 1914765312
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.47937730701332e-05,
      "loss": 2.4938,
      "theoretical_loss": 3.4429036979640832,
      "tokens_seen": 1914896384
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.478574867597496e-05,
      "loss": 2.5209,
      "theoretical_loss": 3.442883881430908,
      "tokens_seen": 1915027456
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.477772428181672e-05,
      "loss": 2.6642,
      "theoretical_loss": 3.442864066633747,
      "tokens_seen": 1915158528
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.476969988765848e-05,
      "loss": 2.6275,
      "theoretical_loss": 3.442844253572331,
      "tokens_seen": 1915289600
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1054366,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9697377681732178,
      "objective/train/theoretical_loss": 3.4428244422463887,
      "objective/train/tokens_used": 285879776,
      "theoretical_loss": 3.4428244422463887,
      "tokens_seen": 1915420672
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.476167549350024e-05,
      "loss": 2.6293,
      "theoretical_loss": 3.4428244422463887,
      "tokens_seen": 1915420672
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.4753651099342e-05,
      "loss": 2.4817,
      "theoretical_loss": 3.4428046326556494,
      "tokens_seen": 1915551744
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.474562670518376e-05,
      "loss": 2.4483,
      "theoretical_loss": 3.4427848247998423,
      "tokens_seen": 1915682816
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.473760231102552e-05,
      "loss": 2.6185,
      "theoretical_loss": 3.4427650186786964,
      "tokens_seen": 1915813888
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.472957791686727e-05,
      "loss": 2.5448,
      "theoretical_loss": 3.4427452142919415,
      "tokens_seen": 1915944960
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.472155352270904e-05,
      "loss": 2.5756,
      "theoretical_loss": 3.442725411639307,
      "tokens_seen": 1916076032
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.471352912855079e-05,
      "loss": 2.6482,
      "theoretical_loss": 3.4427056107205223,
      "tokens_seen": 1916207104
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.470550473439256e-05,
      "loss": 2.5003,
      "theoretical_loss": 3.442685811535317,
      "tokens_seen": 1916338176
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.469748034023431e-05,
      "loss": 2.7789,
      "theoretical_loss": 3.4426660140834215,
      "tokens_seen": 1916469248
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.468945594607608e-05,
      "loss": 2.5984,
      "theoretical_loss": 3.442646218364565,
      "tokens_seen": 1916600320
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.468143155191783e-05,
      "loss": 2.5712,
      "theoretical_loss": 3.442626424378477,
      "tokens_seen": 1916731392
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.467340715775958e-05,
      "loss": 2.5202,
      "theoretical_loss": 3.4426066321248876,
      "tokens_seen": 1916862464
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.466538276360135e-05,
      "loss": 2.5165,
      "theoretical_loss": 3.442586841603527,
      "tokens_seen": 1916993536
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1055597,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.398259162902832,
      "objective/train/theoretical_loss": 3.4425769469923475,
      "objective/train/tokens_used": 287518176,
      "theoretical_loss": 3.4425769469923475,
      "tokens_seen": 1917059072
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.46573583694431e-05,
      "loss": 2.6192,
      "theoretical_loss": 3.442567052814124,
      "tokens_seen": 1917124608
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.464933397528487e-05,
      "loss": 2.4034,
      "theoretical_loss": 3.442547265756411,
      "tokens_seen": 1917255680
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.464130958112662e-05,
      "loss": 2.4076,
      "theoretical_loss": 3.4425274804301154,
      "tokens_seen": 1917386752
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.463328518696839e-05,
      "loss": 2.6285,
      "theoretical_loss": 3.442507696834969,
      "tokens_seen": 1917517824
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.462526079281014e-05,
      "loss": 2.4717,
      "theoretical_loss": 3.4424879149707017,
      "tokens_seen": 1917648896
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.461723639865191e-05,
      "loss": 2.649,
      "theoretical_loss": 3.442468134837043,
      "tokens_seen": 1917779968
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.460921200449366e-05,
      "loss": 2.5644,
      "theoretical_loss": 3.4424483564337245,
      "tokens_seen": 1917911040
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.460118761033541e-05,
      "loss": 2.5549,
      "theoretical_loss": 3.4424285797604757,
      "tokens_seen": 1918042112
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.459316321617718e-05,
      "loss": 2.5357,
      "theoretical_loss": 3.4424088048170276,
      "tokens_seen": 1918173184
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.458513882201893e-05,
      "loss": 2.4481,
      "theoretical_loss": 3.4423890316031107,
      "tokens_seen": 1918304256
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.45771144278607e-05,
      "loss": 2.3806,
      "theoretical_loss": 3.442369260118455,
      "tokens_seen": 1918435328
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.456909003370245e-05,
      "loss": 2.5601,
      "theoretical_loss": 3.4423494903627914,
      "tokens_seen": 1918566400
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1056906,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.411397695541382,
      "objective/train/theoretical_loss": 3.4423297223358507,
      "objective/train/tokens_used": 289156576,
      "theoretical_loss": 3.4423297223358507,
      "tokens_seen": 1918697472
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.456106563954422e-05,
      "loss": 2.6341,
      "theoretical_loss": 3.4423297223358507,
      "tokens_seen": 1918697472
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.455304124538597e-05,
      "loss": 2.4965,
      "theoretical_loss": 3.4423099560373633,
      "tokens_seen": 1918828544
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.454501685122773e-05,
      "loss": 2.4888,
      "theoretical_loss": 3.4422901914670607,
      "tokens_seen": 1918959616
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.45369924570695e-05,
      "loss": 2.5232,
      "theoretical_loss": 3.4422704286246733,
      "tokens_seen": 1919090688
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.452896806291125e-05,
      "loss": 2.6028,
      "theoretical_loss": 3.4422506675099322,
      "tokens_seen": 1919221760
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.452094366875301e-05,
      "loss": 2.4306,
      "theoretical_loss": 3.4422309081225686,
      "tokens_seen": 1919352832
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.451291927459477e-05,
      "loss": 2.6496,
      "theoretical_loss": 3.442211150462313,
      "tokens_seen": 1919483904
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.450489488043653e-05,
      "loss": 2.539,
      "theoretical_loss": 3.442191394528897,
      "tokens_seen": 1919614976
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.449687048627829e-05,
      "loss": 2.485,
      "theoretical_loss": 3.4421716403220515,
      "tokens_seen": 1919746048
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.448884609212004e-05,
      "loss": 2.6431,
      "theoretical_loss": 3.442151887841508,
      "tokens_seen": 1919877120
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.44808216979618e-05,
      "loss": 2.7387,
      "theoretical_loss": 3.442132137086997,
      "tokens_seen": 1920008192
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.447279730380356e-05,
      "loss": 2.5542,
      "theoretical_loss": 3.442112388058251,
      "tokens_seen": 1920139264
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.446477290964533e-05,
      "loss": 2.557,
      "theoretical_loss": 3.4420926407550008,
      "tokens_seen": 1920270336
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1057603,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.503713607788086,
      "objective/train/theoretical_loss": 3.442082767750353,
      "objective/train/tokens_used": 290794976,
      "theoretical_loss": 3.442082767750353,
      "tokens_seen": 1920335872
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.445674851548708e-05,
      "loss": 2.6189,
      "theoretical_loss": 3.4420728951769783,
      "tokens_seen": 1920401408
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.444872412132885e-05,
      "loss": 2.5225,
      "theoretical_loss": 3.4420531513239148,
      "tokens_seen": 1920532480
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.44406997271706e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.4420334091955413,
      "tokens_seen": 1920663552
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.443267533301235e-05,
      "loss": 2.6353,
      "theoretical_loss": 3.4420136687915903,
      "tokens_seen": 1920794624
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.442465093885412e-05,
      "loss": 2.4835,
      "theoretical_loss": 3.441993930111793,
      "tokens_seen": 1920925696
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.441662654469587e-05,
      "loss": 2.5794,
      "theoretical_loss": 3.4419741931558816,
      "tokens_seen": 1921056768
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.440860215053764e-05,
      "loss": 2.6229,
      "theoretical_loss": 3.441954457923588,
      "tokens_seen": 1921187840
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.440057775637939e-05,
      "loss": 2.4523,
      "theoretical_loss": 3.4419347244146437,
      "tokens_seen": 1921318912
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.439255336222116e-05,
      "loss": 2.6327,
      "theoretical_loss": 3.4419149926287806,
      "tokens_seen": 1921449984
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.438452896806291e-05,
      "loss": 2.6949,
      "theoretical_loss": 3.4418952625657306,
      "tokens_seen": 1921581056
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.437650457390466e-05,
      "loss": 2.5928,
      "theoretical_loss": 3.4418755342252267,
      "tokens_seen": 1921712128
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.436848017974643e-05,
      "loss": 2.5388,
      "theoretical_loss": 3.4418558076070003,
      "tokens_seen": 1921843200
    },
    {
      "epoch": 0.16,
      "objective/train/docs_used": 1058944,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.295034885406494,
      "objective/train/theoretical_loss": 3.4418360827107835,
      "objective/train/tokens_used": 292433376,
      "theoretical_loss": 3.4418360827107835,
      "tokens_seen": 1921974272
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.436045578558818e-05,
      "loss": 2.5935,
      "theoretical_loss": 3.4418360827107835,
      "tokens_seen": 1921974272
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.435243139142995e-05,
      "loss": 2.6164,
      "theoretical_loss": 3.4418163595363094,
      "tokens_seen": 1922105344
    },
    {
      "epoch": 0.16,
      "learning_rate": 8.43444069972717e-05,
      "loss": 2.6415,
      "theoretical_loss": 3.441796638083309,
      "tokens_seen": 1922236416
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.433638260311347e-05,
      "loss": 2.6415,
      "theoretical_loss": 3.4417769183515157,
      "tokens_seen": 1922367488
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.432835820895522e-05,
      "loss": 2.5213,
      "theoretical_loss": 3.441757200340662,
      "tokens_seen": 1922498560
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.432033381479699e-05,
      "loss": 2.7047,
      "theoretical_loss": 3.4417374840504795,
      "tokens_seen": 1922629632
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.431230942063874e-05,
      "loss": 2.324,
      "theoretical_loss": 3.441717769480701,
      "tokens_seen": 1922760704
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.43042850264805e-05,
      "loss": 2.7517,
      "theoretical_loss": 3.4416980566310604,
      "tokens_seen": 1922891776
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.429626063232226e-05,
      "loss": 2.5888,
      "theoretical_loss": 3.4416783455012885,
      "tokens_seen": 1923022848
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.428823623816402e-05,
      "loss": 2.452,
      "theoretical_loss": 3.4416586360911197,
      "tokens_seen": 1923153920
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.428021184400578e-05,
      "loss": 2.6019,
      "theoretical_loss": 3.4416389284002853,
      "tokens_seen": 1923284992
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.427218744984754e-05,
      "loss": 2.622,
      "theoretical_loss": 3.4416192224285194,
      "tokens_seen": 1923416064
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.42641630556893e-05,
      "loss": 2.642,
      "theoretical_loss": 3.441599518175554,
      "tokens_seen": 1923547136
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1059301,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.181396245956421,
      "objective/train/theoretical_loss": 3.441589666693538,
      "objective/train/tokens_used": 294071776,
      "theoretical_loss": 3.441589666693538,
      "tokens_seen": 1923612672
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.425613866153106e-05,
      "loss": 2.6953,
      "theoretical_loss": 3.4415798156411226,
      "tokens_seen": 1923678208
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.424811426737281e-05,
      "loss": 2.5897,
      "theoretical_loss": 3.4415601148249584,
      "tokens_seen": 1923809280
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.424008987321458e-05,
      "loss": 2.5453,
      "theoretical_loss": 3.441540415726794,
      "tokens_seen": 1923940352
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.423206547905633e-05,
      "loss": 2.4628,
      "theoretical_loss": 3.4415207183463625,
      "tokens_seen": 1924071424
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.42240410848981e-05,
      "loss": 2.5583,
      "theoretical_loss": 3.441501022683398,
      "tokens_seen": 1924202496
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.421601669073985e-05,
      "loss": 2.5335,
      "theoretical_loss": 3.4414813287376327,
      "tokens_seen": 1924333568
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.420799229658162e-05,
      "loss": 2.5424,
      "theoretical_loss": 3.4414616365088007,
      "tokens_seen": 1924464640
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.419996790242337e-05,
      "loss": 2.5134,
      "theoretical_loss": 3.441441945996635,
      "tokens_seen": 1924595712
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.419194350826512e-05,
      "loss": 2.5916,
      "theoretical_loss": 3.441422257200869,
      "tokens_seen": 1924726784
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.418391911410689e-05,
      "loss": 2.4067,
      "theoretical_loss": 3.4414025701212365,
      "tokens_seen": 1924857856
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.417589471994864e-05,
      "loss": 2.5449,
      "theoretical_loss": 3.441382884757471,
      "tokens_seen": 1924988928
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.416787032579041e-05,
      "loss": 2.6262,
      "theoretical_loss": 3.441363201109305,
      "tokens_seen": 1925120000
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1060240,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.727017641067505,
      "objective/train/theoretical_loss": 3.4413435191764745,
      "objective/train/tokens_used": 295710176,
      "theoretical_loss": 3.4413435191764745,
      "tokens_seen": 1925251072
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.415984593163216e-05,
      "loss": 2.5072,
      "theoretical_loss": 3.4413435191764745,
      "tokens_seen": 1925251072
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.415182153747393e-05,
      "loss": 2.5916,
      "theoretical_loss": 3.441323838958712,
      "tokens_seen": 1925382144
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.414379714331568e-05,
      "loss": 2.6457,
      "theoretical_loss": 3.4413041604557506,
      "tokens_seen": 1925513216
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.413577274915743e-05,
      "loss": 2.5979,
      "theoretical_loss": 3.4412844836673253,
      "tokens_seen": 1925644288
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.41277483549992e-05,
      "loss": 2.5639,
      "theoretical_loss": 3.4412648085931696,
      "tokens_seen": 1925775360
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.411972396084095e-05,
      "loss": 2.5317,
      "theoretical_loss": 3.4412451352330176,
      "tokens_seen": 1925906432
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.411169956668272e-05,
      "loss": 2.553,
      "theoretical_loss": 3.441225463586603,
      "tokens_seen": 1926037504
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.410367517252447e-05,
      "loss": 2.5041,
      "theoretical_loss": 3.44120579365366,
      "tokens_seen": 1926168576
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.409565077836624e-05,
      "loss": 2.4359,
      "theoretical_loss": 3.4411861254339238,
      "tokens_seen": 1926299648
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.4087626384208e-05,
      "loss": 2.5611,
      "theoretical_loss": 3.441166458927127,
      "tokens_seen": 1926430720
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.407960199004975e-05,
      "loss": 2.4728,
      "theoretical_loss": 3.441146794133005,
      "tokens_seen": 1926561792
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.407157759589151e-05,
      "loss": 2.4456,
      "theoretical_loss": 3.4411271310512914,
      "tokens_seen": 1926692864
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.406355320173327e-05,
      "loss": 2.5713,
      "theoretical_loss": 3.441107469681721,
      "tokens_seen": 1926823936
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1060815,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9482412338256836,
      "objective/train/theoretical_loss": 3.441097639638907,
      "objective/train/tokens_used": 297348576,
      "theoretical_loss": 3.441097639638907,
      "tokens_seen": 1926889472
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.405552880757503e-05,
      "loss": 2.6823,
      "theoretical_loss": 3.4410878100240287,
      "tokens_seen": 1926955008
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.404750441341679e-05,
      "loss": 2.6304,
      "theoretical_loss": 3.441068152077948,
      "tokens_seen": 1927086080
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.403948001925855e-05,
      "loss": 2.7154,
      "theoretical_loss": 3.441048495843215,
      "tokens_seen": 1927217152
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.403145562510031e-05,
      "loss": 2.6143,
      "theoretical_loss": 3.4410288413195627,
      "tokens_seen": 1927348224
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.402343123094206e-05,
      "loss": 2.5897,
      "theoretical_loss": 3.4410091885067264,
      "tokens_seen": 1927479296
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.401540683678383e-05,
      "loss": 2.7404,
      "theoretical_loss": 3.440989537404441,
      "tokens_seen": 1927610368
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.400738244262558e-05,
      "loss": 2.4568,
      "theoretical_loss": 3.4409698880124413,
      "tokens_seen": 1927741440
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.399935804846735e-05,
      "loss": 2.6438,
      "theoretical_loss": 3.4409502403304626,
      "tokens_seen": 1927872512
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.39913336543091e-05,
      "loss": 2.471,
      "theoretical_loss": 3.440930594358239,
      "tokens_seen": 1928003584
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.398330926015087e-05,
      "loss": 2.4765,
      "theoretical_loss": 3.440910950095506,
      "tokens_seen": 1928134656
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.397528486599262e-05,
      "loss": 2.6399,
      "theoretical_loss": 3.4408913075419987,
      "tokens_seen": 1928265728
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.396726047183439e-05,
      "loss": 2.6799,
      "theoretical_loss": 3.440871666697452,
      "tokens_seen": 1928396800
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1062086,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.882824659347534,
      "objective/train/theoretical_loss": 3.4408520275616006,
      "objective/train/tokens_used": 298986976,
      "theoretical_loss": 3.4408520275616006,
      "tokens_seen": 1928527872
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.395923607767614e-05,
      "loss": 2.5886,
      "theoretical_loss": 3.4408520275616006,
      "tokens_seen": 1928527872
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.395121168351789e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.4408323901341813,
      "tokens_seen": 1928658944
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.394318728935966e-05,
      "loss": 2.5705,
      "theoretical_loss": 3.440812754414927,
      "tokens_seen": 1928790016
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.393516289520141e-05,
      "loss": 2.5066,
      "theoretical_loss": 3.4407931204035753,
      "tokens_seen": 1928921088
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.392713850104318e-05,
      "loss": 2.6055,
      "theoretical_loss": 3.440773488099861,
      "tokens_seen": 1929052160
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.391911410688493e-05,
      "loss": 2.5717,
      "theoretical_loss": 3.440753857503519,
      "tokens_seen": 1929183232
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.39110897127267e-05,
      "loss": 2.5541,
      "theoretical_loss": 3.440734228614285,
      "tokens_seen": 1929314304
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.390306531856845e-05,
      "loss": 2.587,
      "theoretical_loss": 3.4407146014318943,
      "tokens_seen": 1929445376
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.38950409244102e-05,
      "loss": 2.4675,
      "theoretical_loss": 3.4406949759560836,
      "tokens_seen": 1929576448
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.388701653025197e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.440675352186587,
      "tokens_seen": 1929707520
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.387899213609372e-05,
      "loss": 2.6421,
      "theoretical_loss": 3.4406557301231424,
      "tokens_seen": 1929838592
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.387096774193549e-05,
      "loss": 2.5802,
      "theoretical_loss": 3.440636109765484,
      "tokens_seen": 1929969664
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.386294334777724e-05,
      "loss": 2.5423,
      "theoretical_loss": 3.440616491113348,
      "tokens_seen": 1930100736
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1062749,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9142069816589355,
      "objective/train/theoretical_loss": 3.4406066824267683,
      "objective/train/tokens_used": 300625376,
      "theoretical_loss": 3.4406066824267683,
      "tokens_seen": 1930166272
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.385491895361901e-05,
      "loss": 2.5888,
      "theoretical_loss": 3.44059687416647,
      "tokens_seen": 1930231808
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.384689455946076e-05,
      "loss": 2.6012,
      "theoretical_loss": 3.4405772589245873,
      "tokens_seen": 1930362880
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.383887016530252e-05,
      "loss": 2.6279,
      "theoretical_loss": 3.440557645387435,
      "tokens_seen": 1930493952
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.383084577114428e-05,
      "loss": 2.4806,
      "theoretical_loss": 3.4405380335547493,
      "tokens_seen": 1930625024
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.382282137698604e-05,
      "loss": 2.426,
      "theoretical_loss": 3.4405184234262656,
      "tokens_seen": 1930756096
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.38147969828278e-05,
      "loss": 2.6313,
      "theoretical_loss": 3.440498815001722,
      "tokens_seen": 1930887168
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.380677258866956e-05,
      "loss": 2.521,
      "theoretical_loss": 3.440479208280853,
      "tokens_seen": 1931018240
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.379874819451132e-05,
      "loss": 2.6727,
      "theoretical_loss": 3.4404596032633963,
      "tokens_seen": 1931149312
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.379072380035308e-05,
      "loss": 2.4125,
      "theoretical_loss": 3.440439999949087,
      "tokens_seen": 1931280384
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.378269940619483e-05,
      "loss": 2.5713,
      "theoretical_loss": 3.440420398337663,
      "tokens_seen": 1931411456
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.37746750120366e-05,
      "loss": 2.5418,
      "theoretical_loss": 3.4404007984288594,
      "tokens_seen": 1931542528
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.376665061787835e-05,
      "loss": 2.4928,
      "theoretical_loss": 3.4403812002224137,
      "tokens_seen": 1931673600
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1064305,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5660760402679443,
      "objective/train/theoretical_loss": 3.4403616037180624,
      "objective/train/tokens_used": 302263776,
      "theoretical_loss": 3.4403616037180624,
      "tokens_seen": 1931804672
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.375862622372012e-05,
      "loss": 2.5156,
      "theoretical_loss": 3.4403616037180624,
      "tokens_seen": 1931804672
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.375060182956187e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.440342008915542,
      "tokens_seen": 1931935744
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.374257743540364e-05,
      "loss": 2.5074,
      "theoretical_loss": 3.440322415814589,
      "tokens_seen": 1932066816
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.373455304124539e-05,
      "loss": 2.4945,
      "theoretical_loss": 3.4403028244149416,
      "tokens_seen": 1932197888
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.372652864708714e-05,
      "loss": 2.5083,
      "theoretical_loss": 3.440283234716335,
      "tokens_seen": 1932328960
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.371850425292891e-05,
      "loss": 2.6815,
      "theoretical_loss": 3.440263646718507,
      "tokens_seen": 1932460032
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.371047985877066e-05,
      "loss": 2.4038,
      "theoretical_loss": 3.440244060421194,
      "tokens_seen": 1932591104
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.370245546461243e-05,
      "loss": 2.7489,
      "theoretical_loss": 3.4402244758241336,
      "tokens_seen": 1932722176
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.369443107045418e-05,
      "loss": 2.545,
      "theoretical_loss": 3.4402048929270634,
      "tokens_seen": 1932853248
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.368640667629595e-05,
      "loss": 2.616,
      "theoretical_loss": 3.4401853117297194,
      "tokens_seen": 1932984320
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.36783822821377e-05,
      "loss": 2.4904,
      "theoretical_loss": 3.440165732231839,
      "tokens_seen": 1933115392
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.367035788797947e-05,
      "loss": 2.5622,
      "theoretical_loss": 3.440146154433161,
      "tokens_seen": 1933246464
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.366233349382122e-05,
      "loss": 2.7073,
      "theoretical_loss": 3.4401265783334205,
      "tokens_seen": 1933377536
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1064860,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6494178771972656,
      "objective/train/theoretical_loss": 3.4401167909205705,
      "objective/train/tokens_used": 303902176,
      "theoretical_loss": 3.4401167909205705,
      "tokens_seen": 1933443072
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.365430909966297e-05,
      "loss": 2.6403,
      "theoretical_loss": 3.440107003932357,
      "tokens_seen": 1933508608
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.364628470550474e-05,
      "loss": 2.6015,
      "theoretical_loss": 3.4400874312297063,
      "tokens_seen": 1933639680
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.36382603113465e-05,
      "loss": 2.5343,
      "theoretical_loss": 3.4400678602252066,
      "tokens_seen": 1933770752
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.363023591718826e-05,
      "loss": 2.6448,
      "theoretical_loss": 3.440048290918596,
      "tokens_seen": 1933901824
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.362221152303001e-05,
      "loss": 2.5473,
      "theoretical_loss": 3.4400287233096107,
      "tokens_seen": 1934032896
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.361418712887178e-05,
      "loss": 2.6655,
      "theoretical_loss": 3.4400091573979896,
      "tokens_seen": 1934163968
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.360616273471353e-05,
      "loss": 2.5688,
      "theoretical_loss": 3.439989593183471,
      "tokens_seen": 1934295040
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.359813834055529e-05,
      "loss": 2.6169,
      "theoretical_loss": 3.439970030665791,
      "tokens_seen": 1934426112
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.359011394639705e-05,
      "loss": 2.6115,
      "theoretical_loss": 3.4399504698446886,
      "tokens_seen": 1934557184
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.358208955223881e-05,
      "loss": 2.4359,
      "theoretical_loss": 3.4399309107199016,
      "tokens_seen": 1934688256
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.357406515808057e-05,
      "loss": 2.5228,
      "theoretical_loss": 3.439911353291168,
      "tokens_seen": 1934819328
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.356604076392233e-05,
      "loss": 2.558,
      "theoretical_loss": 3.439891797558225,
      "tokens_seen": 1934950400
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1065725,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4864485263824463,
      "objective/train/theoretical_loss": 3.439872243520812,
      "objective/train/tokens_used": 305540576,
      "theoretical_loss": 3.439872243520812,
      "tokens_seen": 1935081472
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.35580163697641e-05,
      "loss": 2.5461,
      "theoretical_loss": 3.439872243520812,
      "tokens_seen": 1935081472
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.354999197560585e-05,
      "loss": 2.4294,
      "theoretical_loss": 3.4398526911786664,
      "tokens_seen": 1935212544
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.35419675814476e-05,
      "loss": 2.48,
      "theoretical_loss": 3.439833140531527,
      "tokens_seen": 1935343616
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.353394318728937e-05,
      "loss": 2.5107,
      "theoretical_loss": 3.4398135915791315,
      "tokens_seen": 1935474688
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.352591879313112e-05,
      "loss": 2.4663,
      "theoretical_loss": 3.439794044321218,
      "tokens_seen": 1935605760
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.351789439897289e-05,
      "loss": 2.6062,
      "theoretical_loss": 3.439774498757526,
      "tokens_seen": 1935736832
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.350987000481464e-05,
      "loss": 2.6143,
      "theoretical_loss": 3.4397549548877926,
      "tokens_seen": 1935867904
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.35018456106564e-05,
      "loss": 2.699,
      "theoretical_loss": 3.439735412711757,
      "tokens_seen": 1935998976
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.349382121649816e-05,
      "loss": 2.5616,
      "theoretical_loss": 3.4397158722291583,
      "tokens_seen": 1936130048
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.348579682233991e-05,
      "loss": 2.5023,
      "theoretical_loss": 3.439696333439734,
      "tokens_seen": 1936261120
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.347777242818168e-05,
      "loss": 2.5506,
      "theoretical_loss": 3.4396767963432238,
      "tokens_seen": 1936392192
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.346974803402343e-05,
      "loss": 2.5643,
      "theoretical_loss": 3.4396572609393656,
      "tokens_seen": 1936523264
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.34617236398652e-05,
      "loss": 2.4963,
      "theoretical_loss": 3.439637727227899,
      "tokens_seen": 1936654336
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1066923,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5231666564941406,
      "objective/train/theoretical_loss": 3.439627961006731,
      "objective/train/tokens_used": 307178976,
      "theoretical_loss": 3.439627961006731,
      "tokens_seen": 1936719872
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.345369924570695e-05,
      "loss": 2.5335,
      "theoretical_loss": 3.4396181952085625,
      "tokens_seen": 1936785408
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.344567485154872e-05,
      "loss": 2.4004,
      "theoretical_loss": 3.4395986648810943,
      "tokens_seen": 1936916480
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.343765045739047e-05,
      "loss": 2.6551,
      "theoretical_loss": 3.439579136245235,
      "tokens_seen": 1937047552
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.342962606323223e-05,
      "loss": 2.5923,
      "theoretical_loss": 3.439559609300722,
      "tokens_seen": 1937178624
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.342160166907399e-05,
      "loss": 2.5727,
      "theoretical_loss": 3.4395400840472954,
      "tokens_seen": 1937309696
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.341357727491575e-05,
      "loss": 2.6667,
      "theoretical_loss": 3.439520560484694,
      "tokens_seen": 1937440768
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.340555288075751e-05,
      "loss": 2.6872,
      "theoretical_loss": 3.439501038612657,
      "tokens_seen": 1937571840
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.339752848659926e-05,
      "loss": 2.6223,
      "theoretical_loss": 3.4394815184309238,
      "tokens_seen": 1937702912
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.338950409244103e-05,
      "loss": 2.5338,
      "theoretical_loss": 3.4394619999392333,
      "tokens_seen": 1937833984
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.338147969828278e-05,
      "loss": 2.6202,
      "theoretical_loss": 3.4394424831373254,
      "tokens_seen": 1937965056
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.337345530412455e-05,
      "loss": 2.556,
      "theoretical_loss": 3.439422968024939,
      "tokens_seen": 1938096128
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.33654309099663e-05,
      "loss": 2.4879,
      "theoretical_loss": 3.4394034546018144,
      "tokens_seen": 1938227200
    },
    {
      "epoch": 0.17,
      "objective/train/docs_used": 1067414,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6843504905700684,
      "objective/train/theoretical_loss": 3.439383942867691,
      "objective/train/tokens_used": 308817376,
      "theoretical_loss": 3.439383942867691,
      "tokens_seen": 1938358272
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.335740651580806e-05,
      "loss": 2.6718,
      "theoretical_loss": 3.439383942867691,
      "tokens_seen": 1938358272
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.334938212164982e-05,
      "loss": 2.4463,
      "theoretical_loss": 3.439364432822307,
      "tokens_seen": 1938489344
    },
    {
      "epoch": 0.17,
      "learning_rate": 8.334135772749158e-05,
      "loss": 2.5893,
      "theoretical_loss": 3.439344924465404,
      "tokens_seen": 1938620416
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.333333333333334e-05,
      "loss": 2.442,
      "theoretical_loss": 3.439325417796721,
      "tokens_seen": 1938751488
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.33253089391751e-05,
      "loss": 2.5347,
      "theoretical_loss": 3.439305912815997,
      "tokens_seen": 1938882560
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.331728454501686e-05,
      "loss": 2.6075,
      "theoretical_loss": 3.439286409522973,
      "tokens_seen": 1939013632
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.330926015085862e-05,
      "loss": 2.453,
      "theoretical_loss": 3.4392669079173883,
      "tokens_seen": 1939144704
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.330123575670037e-05,
      "loss": 2.5933,
      "theoretical_loss": 3.439247407998983,
      "tokens_seen": 1939275776
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.329321136254214e-05,
      "loss": 2.4927,
      "theoretical_loss": 3.439227909767497,
      "tokens_seen": 1939406848
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.328518696838389e-05,
      "loss": 2.5154,
      "theoretical_loss": 3.439208413222671,
      "tokens_seen": 1939537920
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.327716257422566e-05,
      "loss": 2.6075,
      "theoretical_loss": 3.439188918364244,
      "tokens_seen": 1939668992
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.326913818006741e-05,
      "loss": 2.5968,
      "theoretical_loss": 3.439169425191957,
      "tokens_seen": 1939800064
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.326111378590918e-05,
      "loss": 2.4055,
      "theoretical_loss": 3.4391499337055498,
      "tokens_seen": 1939931136
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1068508,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.342050313949585,
      "objective/train/theoretical_loss": 3.4391401885944703,
      "objective/train/tokens_used": 310455776,
      "theoretical_loss": 3.4391401885944703,
      "tokens_seen": 1939996672
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.325308939175093e-05,
      "loss": 2.4388,
      "theoretical_loss": 3.439130443904763,
      "tokens_seen": 1940062208
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.324506499759268e-05,
      "loss": 2.5146,
      "theoretical_loss": 3.439110955789338,
      "tokens_seen": 1940193280
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.323704060343445e-05,
      "loss": 2.6226,
      "theoretical_loss": 3.4390914693590124,
      "tokens_seen": 1940324352
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.32290162092762e-05,
      "loss": 2.5135,
      "theoretical_loss": 3.4390719846135296,
      "tokens_seen": 1940455424
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.322099181511797e-05,
      "loss": 2.5178,
      "theoretical_loss": 3.439052501552628,
      "tokens_seen": 1940586496
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.321296742095972e-05,
      "loss": 2.5175,
      "theoretical_loss": 3.4390330201760495,
      "tokens_seen": 1940717568
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.320494302680149e-05,
      "loss": 2.4956,
      "theoretical_loss": 3.439013540483535,
      "tokens_seen": 1940848640
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.319691863264324e-05,
      "loss": 2.4474,
      "theoretical_loss": 3.4389940624748236,
      "tokens_seen": 1940979712
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.3188894238485e-05,
      "loss": 2.4981,
      "theoretical_loss": 3.4389745861496577,
      "tokens_seen": 1941110784
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.318086984432676e-05,
      "loss": 2.4449,
      "theoretical_loss": 3.438955111507777,
      "tokens_seen": 1941241856
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.317284545016852e-05,
      "loss": 2.5271,
      "theoretical_loss": 3.4389356385489234,
      "tokens_seen": 1941372928
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.316482105601028e-05,
      "loss": 2.5505,
      "theoretical_loss": 3.4389161672728368,
      "tokens_seen": 1941504000
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1069184,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.391522169113159,
      "objective/train/theoretical_loss": 3.438896697679259,
      "objective/train/tokens_used": 312094176,
      "theoretical_loss": 3.438896697679259,
      "tokens_seen": 1941635072
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.315679666185203e-05,
      "loss": 2.5596,
      "theoretical_loss": 3.438896697679259,
      "tokens_seen": 1941635072
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.31487722676938e-05,
      "loss": 2.4975,
      "theoretical_loss": 3.4388772297679306,
      "tokens_seen": 1941766144
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.314074787353555e-05,
      "loss": 2.5002,
      "theoretical_loss": 3.438857763538593,
      "tokens_seen": 1941897216
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.313272347937731e-05,
      "loss": 2.5132,
      "theoretical_loss": 3.4388382989909863,
      "tokens_seen": 1942028288
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.312469908521907e-05,
      "loss": 2.5543,
      "theoretical_loss": 3.4388188361248533,
      "tokens_seen": 1942159360
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.311667469106083e-05,
      "loss": 2.381,
      "theoretical_loss": 3.4387993749399346,
      "tokens_seen": 1942290432
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.31086502969026e-05,
      "loss": 2.3192,
      "theoretical_loss": 3.4387799154359717,
      "tokens_seen": 1942421504
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.310062590274435e-05,
      "loss": 2.5764,
      "theoretical_loss": 3.4387604576127053,
      "tokens_seen": 1942552576
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.309260150858611e-05,
      "loss": 2.3756,
      "theoretical_loss": 3.438741001469878,
      "tokens_seen": 1942683648
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.308457711442787e-05,
      "loss": 2.4454,
      "theoretical_loss": 3.43872154700723,
      "tokens_seen": 1942814720
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.307655272026962e-05,
      "loss": 2.5725,
      "theoretical_loss": 3.438702094224504,
      "tokens_seen": 1942945792
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.306852832611139e-05,
      "loss": 2.4494,
      "theoretical_loss": 3.438682643121441,
      "tokens_seen": 1943076864
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.306050393195314e-05,
      "loss": 2.3641,
      "theoretical_loss": 3.4386631936977827,
      "tokens_seen": 1943207936
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1070317,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5918591022491455,
      "objective/train/theoretical_loss": 3.43865346961565,
      "objective/train/tokens_used": 313732576,
      "theoretical_loss": 3.43865346961565,
      "tokens_seen": 1943273472
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.305247953779491e-05,
      "loss": 2.6276,
      "theoretical_loss": 3.4386437459532715,
      "tokens_seen": 1943339008
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.304445514363666e-05,
      "loss": 2.5815,
      "theoretical_loss": 3.438624299887648,
      "tokens_seen": 1943470080
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.303643074947843e-05,
      "loss": 2.3985,
      "theoretical_loss": 3.438604855500655,
      "tokens_seen": 1943601152
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.302840635532018e-05,
      "loss": 2.5067,
      "theoretical_loss": 3.438585412792034,
      "tokens_seen": 1943732224
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.302038196116195e-05,
      "loss": 2.605,
      "theoretical_loss": 3.438565971761528,
      "tokens_seen": 1943863296
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.30123575670037e-05,
      "loss": 2.4195,
      "theoretical_loss": 3.4385465324088766,
      "tokens_seen": 1943994368
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.300433317284545e-05,
      "loss": 2.5353,
      "theoretical_loss": 3.4385270947338245,
      "tokens_seen": 1944125440
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.299630877868722e-05,
      "loss": 2.5632,
      "theoretical_loss": 3.4385076587361123,
      "tokens_seen": 1944256512
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.298828438452897e-05,
      "loss": 2.4589,
      "theoretical_loss": 3.438488224415483,
      "tokens_seen": 1944387584
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.298025999037074e-05,
      "loss": 2.4504,
      "theoretical_loss": 3.438468791771678,
      "tokens_seen": 1944518656
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.297223559621249e-05,
      "loss": 2.7096,
      "theoretical_loss": 3.4384493608044404,
      "tokens_seen": 1944649728
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.296421120205426e-05,
      "loss": 2.494,
      "theoretical_loss": 3.4384299315135123,
      "tokens_seen": 1944780800
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1070782,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.948735237121582,
      "objective/train/theoretical_loss": 3.438410503898636,
      "objective/train/tokens_used": 315370976,
      "theoretical_loss": 3.438410503898636,
      "tokens_seen": 1944911872
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.295618680789601e-05,
      "loss": 2.5633,
      "theoretical_loss": 3.438410503898636,
      "tokens_seen": 1944911872
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.294816241373777e-05,
      "loss": 2.4825,
      "theoretical_loss": 3.4383910779595537,
      "tokens_seen": 1945042944
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.294013801957953e-05,
      "loss": 2.593,
      "theoretical_loss": 3.438371653696009,
      "tokens_seen": 1945174016
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.293211362542129e-05,
      "loss": 2.4795,
      "theoretical_loss": 3.4383522311077432,
      "tokens_seen": 1945305088
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.292408923126305e-05,
      "loss": 2.6767,
      "theoretical_loss": 3.4383328101945,
      "tokens_seen": 1945436160
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.29160648371048e-05,
      "loss": 2.5267,
      "theoretical_loss": 3.4383133909560213,
      "tokens_seen": 1945567232
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.290804044294657e-05,
      "loss": 2.5432,
      "theoretical_loss": 3.4382939733920503,
      "tokens_seen": 1945698304
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.290001604878832e-05,
      "loss": 2.5006,
      "theoretical_loss": 3.4382745575023295,
      "tokens_seen": 1945829376
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.289199165463008e-05,
      "loss": 2.5887,
      "theoretical_loss": 3.4382551432866024,
      "tokens_seen": 1945960448
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.288396726047184e-05,
      "loss": 2.4243,
      "theoretical_loss": 3.438235730744611,
      "tokens_seen": 1946091520
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.28759428663136e-05,
      "loss": 2.5427,
      "theoretical_loss": 3.4382163198760995,
      "tokens_seen": 1946222592
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.286791847215536e-05,
      "loss": 2.6092,
      "theoretical_loss": 3.43819691068081,
      "tokens_seen": 1946353664
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.285989407799712e-05,
      "loss": 2.4251,
      "theoretical_loss": 3.4381775031584856,
      "tokens_seen": 1946484736
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1072031,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5232903957366943,
      "objective/train/theoretical_loss": 3.4381678000246056,
      "objective/train/tokens_used": 317009376,
      "theoretical_loss": 3.4381678000246056,
      "tokens_seen": 1946550272
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.285186968383888e-05,
      "loss": 2.6291,
      "theoretical_loss": 3.43815809730887,
      "tokens_seen": 1946615808
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.284384528968064e-05,
      "loss": 2.5281,
      "theoretical_loss": 3.438138693131706,
      "tokens_seen": 1946746880
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.283582089552239e-05,
      "loss": 2.4736,
      "theoretical_loss": 3.4381192906267373,
      "tokens_seen": 1946877952
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.282779650136416e-05,
      "loss": 2.538,
      "theoretical_loss": 3.4380998897937065,
      "tokens_seen": 1947009024
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.281977210720591e-05,
      "loss": 2.4369,
      "theoretical_loss": 3.438080490632358,
      "tokens_seen": 1947140096
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.281174771304768e-05,
      "loss": 2.7668,
      "theoretical_loss": 3.4380610931424345,
      "tokens_seen": 1947271168
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.280372331888943e-05,
      "loss": 2.5402,
      "theoretical_loss": 3.438041697323679,
      "tokens_seen": 1947402240
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.27956989247312e-05,
      "loss": 2.499,
      "theoretical_loss": 3.4380223031758366,
      "tokens_seen": 1947533312
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.278767453057295e-05,
      "loss": 2.3226,
      "theoretical_loss": 3.43800291069865,
      "tokens_seen": 1947664384
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.27796501364147e-05,
      "loss": 2.5558,
      "theoretical_loss": 3.4379835198918625,
      "tokens_seen": 1947795456
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.277162574225647e-05,
      "loss": 2.5019,
      "theoretical_loss": 3.4379641307552182,
      "tokens_seen": 1947926528
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.276360134809822e-05,
      "loss": 2.51,
      "theoretical_loss": 3.4379447432884613,
      "tokens_seen": 1948057600
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1072626,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2674527168273926,
      "objective/train/theoretical_loss": 3.437925357491335,
      "objective/train/tokens_used": 318647776,
      "theoretical_loss": 3.437925357491335,
      "tokens_seen": 1948188672
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.275557695393999e-05,
      "loss": 2.5014,
      "theoretical_loss": 3.437925357491335,
      "tokens_seen": 1948188672
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.274755255978174e-05,
      "loss": 2.5272,
      "theoretical_loss": 3.437905973363584,
      "tokens_seen": 1948319744
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.273952816562351e-05,
      "loss": 2.659,
      "theoretical_loss": 3.4378865909049514,
      "tokens_seen": 1948450816
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.273150377146526e-05,
      "loss": 2.5092,
      "theoretical_loss": 3.437867210115181,
      "tokens_seen": 1948581888
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.272347937730703e-05,
      "loss": 2.5668,
      "theoretical_loss": 3.437847830994018,
      "tokens_seen": 1948712960
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.271545498314878e-05,
      "loss": 2.5462,
      "theoretical_loss": 3.437828453541206,
      "tokens_seen": 1948844032
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.270743058899054e-05,
      "loss": 2.7522,
      "theoretical_loss": 3.437809077756489,
      "tokens_seen": 1948975104
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.26994061948323e-05,
      "loss": 2.6027,
      "theoretical_loss": 3.437789703639611,
      "tokens_seen": 1949106176
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.269138180067406e-05,
      "loss": 2.5426,
      "theoretical_loss": 3.4377703311903174,
      "tokens_seen": 1949237248
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.268335740651582e-05,
      "loss": 2.6411,
      "theoretical_loss": 3.437750960408351,
      "tokens_seen": 1949368320
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.267533301235758e-05,
      "loss": 2.4644,
      "theoretical_loss": 3.437731591293457,
      "tokens_seen": 1949499392
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.266730861819934e-05,
      "loss": 2.5176,
      "theoretical_loss": 3.4377122238453803,
      "tokens_seen": 1949630464
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.26592842240411e-05,
      "loss": 2.6194,
      "theoretical_loss": 3.4376928580638646,
      "tokens_seen": 1949761536
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1073976,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5130059719085693,
      "objective/train/theoretical_loss": 3.4376831757979875,
      "objective/train/tokens_used": 320286176,
      "theoretical_loss": 3.4376831757979875,
      "tokens_seen": 1949827072
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.265125982988285e-05,
      "loss": 2.536,
      "theoretical_loss": 3.437673493948655,
      "tokens_seen": 1949892608
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.264323543572461e-05,
      "loss": 2.6313,
      "theoretical_loss": 3.437654131499496,
      "tokens_seen": 1950023680
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.263521104156637e-05,
      "loss": 2.5475,
      "theoretical_loss": 3.437634770716132,
      "tokens_seen": 1950154752
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.262718664740813e-05,
      "loss": 2.6455,
      "theoretical_loss": 3.4376154115983084,
      "tokens_seen": 1950285824
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.261916225324989e-05,
      "loss": 2.3373,
      "theoretical_loss": 3.4375960541457697,
      "tokens_seen": 1950416896
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.261113785909164e-05,
      "loss": 2.4703,
      "theoretical_loss": 3.43757669835826,
      "tokens_seen": 1950547968
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.260311346493341e-05,
      "loss": 2.5127,
      "theoretical_loss": 3.4375573442355254,
      "tokens_seen": 1950679040
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.259508907077516e-05,
      "loss": 2.3909,
      "theoretical_loss": 3.43753799177731,
      "tokens_seen": 1950810112
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.258706467661693e-05,
      "loss": 2.4477,
      "theoretical_loss": 3.437518640983359,
      "tokens_seen": 1950941184
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.257904028245868e-05,
      "loss": 2.3993,
      "theoretical_loss": 3.4374992918534177,
      "tokens_seen": 1951072256
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.257101588830045e-05,
      "loss": 2.5379,
      "theoretical_loss": 3.4374799443872317,
      "tokens_seen": 1951203328
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.25629914941422e-05,
      "loss": 2.5257,
      "theoretical_loss": 3.4374605985845452,
      "tokens_seen": 1951334400
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1074351,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.568186044692993,
      "objective/train/theoretical_loss": 3.437441254445104,
      "objective/train/tokens_used": 321924576,
      "theoretical_loss": 3.437441254445104,
      "tokens_seen": 1951465472
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.255496709998395e-05,
      "loss": 2.6717,
      "theoretical_loss": 3.437441254445104,
      "tokens_seen": 1951465472
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.254694270582572e-05,
      "loss": 2.6209,
      "theoretical_loss": 3.437421911968653,
      "tokens_seen": 1951596544
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.253891831166747e-05,
      "loss": 2.4163,
      "theoretical_loss": 3.4374025711549385,
      "tokens_seen": 1951727616
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.253089391750924e-05,
      "loss": 2.5627,
      "theoretical_loss": 3.437383232003705,
      "tokens_seen": 1951858688
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.252286952335099e-05,
      "loss": 2.4633,
      "theoretical_loss": 3.4373638945146983,
      "tokens_seen": 1951989760
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.251484512919275e-05,
      "loss": 2.5232,
      "theoretical_loss": 3.4373445586876636,
      "tokens_seen": 1952120832
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.250682073503451e-05,
      "loss": 2.5546,
      "theoretical_loss": 3.4373252245223473,
      "tokens_seen": 1952251904
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.249879634087627e-05,
      "loss": 2.5288,
      "theoretical_loss": 3.437305892018494,
      "tokens_seen": 1952382976
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.249077194671803e-05,
      "loss": 2.6947,
      "theoretical_loss": 3.43728656117585,
      "tokens_seen": 1952514048
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.248274755255979e-05,
      "loss": 2.6388,
      "theoretical_loss": 3.437267231994161,
      "tokens_seen": 1952645120
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.247472315840154e-05,
      "loss": 2.5857,
      "theoretical_loss": 3.437247904473173,
      "tokens_seen": 1952776192
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.24666987642433e-05,
      "loss": 2.5235,
      "theoretical_loss": 3.4372285786126318,
      "tokens_seen": 1952907264
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.245867437008506e-05,
      "loss": 2.5955,
      "theoretical_loss": 3.4372092544122825,
      "tokens_seen": 1953038336
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1075675,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2708775997161865,
      "objective/train/theoretical_loss": 3.4371995929346006,
      "objective/train/tokens_used": 323562976,
      "theoretical_loss": 3.4371995929346006,
      "tokens_seen": 1953103872
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.245064997592683e-05,
      "loss": 2.5938,
      "theoretical_loss": 3.437189931871872,
      "tokens_seen": 1953169408
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.244262558176858e-05,
      "loss": 2.6016,
      "theoretical_loss": 3.437170610991146,
      "tokens_seen": 1953300480
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.243460118761035e-05,
      "loss": 2.3854,
      "theoretical_loss": 3.437151291769851,
      "tokens_seen": 1953431552
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.24265767934521e-05,
      "loss": 2.59,
      "theoretical_loss": 3.4371319742077326,
      "tokens_seen": 1953562624
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.241855239929385e-05,
      "loss": 2.4944,
      "theoretical_loss": 3.437112658304537,
      "tokens_seen": 1953693696
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.241052800513562e-05,
      "loss": 2.6604,
      "theoretical_loss": 3.437093344060011,
      "tokens_seen": 1953824768
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.240250361097737e-05,
      "loss": 2.4924,
      "theoretical_loss": 3.437074031473901,
      "tokens_seen": 1953955840
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.239447921681914e-05,
      "loss": 2.5443,
      "theoretical_loss": 3.4370547205459516,
      "tokens_seen": 1954086912
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.238645482266089e-05,
      "loss": 2.5714,
      "theoretical_loss": 3.4370354112759114,
      "tokens_seen": 1954217984
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.237843042850264e-05,
      "loss": 2.5103,
      "theoretical_loss": 3.4370161036635256,
      "tokens_seen": 1954349056
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.237040603434441e-05,
      "loss": 2.5072,
      "theoretical_loss": 3.4369967977085416,
      "tokens_seen": 1954480128
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.236238164018616e-05,
      "loss": 2.4739,
      "theoretical_loss": 3.436977493410706,
      "tokens_seen": 1954611200
    },
    {
      "epoch": 0.18,
      "objective/train/docs_used": 1076263,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2574844360351562,
      "objective/train/theoretical_loss": 3.436958190769764,
      "objective/train/tokens_used": 325201376,
      "theoretical_loss": 3.436958190769764,
      "tokens_seen": 1954742272
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.235435724602793e-05,
      "loss": 2.5272,
      "theoretical_loss": 3.436958190769764,
      "tokens_seen": 1954742272
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.234633285186968e-05,
      "loss": 2.5859,
      "theoretical_loss": 3.4369388897854636,
      "tokens_seen": 1954873344
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.233830845771145e-05,
      "loss": 2.4542,
      "theoretical_loss": 3.436919590457552,
      "tokens_seen": 1955004416
    },
    {
      "epoch": 0.18,
      "learning_rate": 8.23302840635532e-05,
      "loss": 2.6658,
      "theoretical_loss": 3.4369002927857744,
      "tokens_seen": 1955135488
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.232225966939496e-05,
      "loss": 2.5128,
      "theoretical_loss": 3.436880996769879,
      "tokens_seen": 1955266560
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.231423527523672e-05,
      "loss": 2.4474,
      "theoretical_loss": 3.436861702409612,
      "tokens_seen": 1955397632
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.230621088107848e-05,
      "loss": 2.6982,
      "theoretical_loss": 3.4368424097047208,
      "tokens_seen": 1955528704
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.229818648692024e-05,
      "loss": 2.4142,
      "theoretical_loss": 3.4368231186549525,
      "tokens_seen": 1955659776
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.2290162092762e-05,
      "loss": 2.4942,
      "theoretical_loss": 3.436803829260054,
      "tokens_seen": 1955790848
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.228213769860375e-05,
      "loss": 2.5114,
      "theoretical_loss": 3.4367845415197715,
      "tokens_seen": 1955921920
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.227411330444552e-05,
      "loss": 2.4439,
      "theoretical_loss": 3.436765255433854,
      "tokens_seen": 1956052992
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.226608891028727e-05,
      "loss": 2.5427,
      "theoretical_loss": 3.436745971002048,
      "tokens_seen": 1956184064
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.225806451612904e-05,
      "loss": 2.5254,
      "theoretical_loss": 3.4367266882241,
      "tokens_seen": 1956315136
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1077404,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.399874210357666,
      "objective/train/theoretical_loss": 3.436717047455245,
      "objective/train/tokens_used": 326839776,
      "theoretical_loss": 3.436717047455245,
      "tokens_seen": 1956380672
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.225004012197079e-05,
      "loss": 2.5416,
      "theoretical_loss": 3.436707407099759,
      "tokens_seen": 1956446208
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.224201572781256e-05,
      "loss": 2.7036,
      "theoretical_loss": 3.436688127628771,
      "tokens_seen": 1956577280
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.223399133365431e-05,
      "loss": 2.6167,
      "theoretical_loss": 3.436668849810884,
      "tokens_seen": 1956708352
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.222596693949606e-05,
      "loss": 2.5307,
      "theoretical_loss": 3.4366495736458456,
      "tokens_seen": 1956839424
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.221794254533783e-05,
      "loss": 2.6293,
      "theoretical_loss": 3.436630299133403,
      "tokens_seen": 1956970496
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.220991815117958e-05,
      "loss": 2.5141,
      "theoretical_loss": 3.4366110262733045,
      "tokens_seen": 1957101568
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.220189375702135e-05,
      "loss": 2.4892,
      "theoretical_loss": 3.4365917550652973,
      "tokens_seen": 1957232640
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.21938693628631e-05,
      "loss": 2.6137,
      "theoretical_loss": 3.4365724855091293,
      "tokens_seen": 1957363712
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.218584496870485e-05,
      "loss": 2.4596,
      "theoretical_loss": 3.4365532176045486,
      "tokens_seen": 1957494784
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.217782057454662e-05,
      "loss": 2.5889,
      "theoretical_loss": 3.4365339513513025,
      "tokens_seen": 1957625856
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.216979618038837e-05,
      "loss": 2.5195,
      "theoretical_loss": 3.436514686749139,
      "tokens_seen": 1957756928
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.216177178623014e-05,
      "loss": 2.48,
      "theoretical_loss": 3.4364954237978065,
      "tokens_seen": 1957888000
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1077881,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7305192947387695,
      "objective/train/theoretical_loss": 3.436476162497053,
      "objective/train/tokens_used": 328478176,
      "theoretical_loss": 3.436476162497053,
      "tokens_seen": 1958019072
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.21537473920719e-05,
      "loss": 2.5058,
      "theoretical_loss": 3.436476162497053,
      "tokens_seen": 1958019072
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.214572299791366e-05,
      "loss": 2.514,
      "theoretical_loss": 3.4364569028466265,
      "tokens_seen": 1958150144
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.213769860375541e-05,
      "loss": 2.4948,
      "theoretical_loss": 3.436437644846275,
      "tokens_seen": 1958281216
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.212967420959717e-05,
      "loss": 2.4609,
      "theoretical_loss": 3.4364183884957464,
      "tokens_seen": 1958412288
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.212164981543893e-05,
      "loss": 2.6224,
      "theoretical_loss": 3.43639913379479,
      "tokens_seen": 1958543360
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.211362542128069e-05,
      "loss": 2.5121,
      "theoretical_loss": 3.4363798807431527,
      "tokens_seen": 1958674432
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.210560102712245e-05,
      "loss": 2.5246,
      "theoretical_loss": 3.4363606293405837,
      "tokens_seen": 1958805504
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.209757663296421e-05,
      "loss": 2.598,
      "theoretical_loss": 3.436341379586832,
      "tokens_seen": 1958936576
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.208955223880597e-05,
      "loss": 2.6481,
      "theoretical_loss": 3.436322131481645,
      "tokens_seen": 1959067648
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.208152784464773e-05,
      "loss": 2.3418,
      "theoretical_loss": 3.4363028850247717,
      "tokens_seen": 1959198720
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.20735034504895e-05,
      "loss": 2.378,
      "theoretical_loss": 3.4362836402159607,
      "tokens_seen": 1959329792
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.206547905633125e-05,
      "loss": 2.6566,
      "theoretical_loss": 3.43626439705496,
      "tokens_seen": 1959460864
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.2057454662173e-05,
      "loss": 2.5246,
      "theoretical_loss": 3.4362451555415197,
      "tokens_seen": 1959591936
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1079297,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.692366600036621,
      "objective/train/theoretical_loss": 3.436235535402555,
      "objective/train/tokens_used": 330116576,
      "theoretical_loss": 3.436235535402555,
      "tokens_seen": 1959657472
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.204943026801477e-05,
      "loss": 2.4392,
      "theoretical_loss": 3.4362259156753874,
      "tokens_seen": 1959723008
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.204140587385652e-05,
      "loss": 2.5496,
      "theoretical_loss": 3.436206677456312,
      "tokens_seen": 1959854080
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.203338147969829e-05,
      "loss": 2.4338,
      "theoretical_loss": 3.4361874408840425,
      "tokens_seen": 1959985152
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.202535708554004e-05,
      "loss": 2.618,
      "theoretical_loss": 3.436168205958328,
      "tokens_seen": 1960116224
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.20173326913818e-05,
      "loss": 2.4057,
      "theoretical_loss": 3.436148972678917,
      "tokens_seen": 1960247296
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.200930829722356e-05,
      "loss": 2.6132,
      "theoretical_loss": 3.43612974104556,
      "tokens_seen": 1960378368
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.200128390306531e-05,
      "loss": 2.3999,
      "theoretical_loss": 3.436110511058004,
      "tokens_seen": 1960509440
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.199325950890708e-05,
      "loss": 2.5032,
      "theoretical_loss": 3.4360912827159997,
      "tokens_seen": 1960640512
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.198523511474883e-05,
      "loss": 2.4567,
      "theoretical_loss": 3.436072056019295,
      "tokens_seen": 1960771584
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.19772107205906e-05,
      "loss": 2.5605,
      "theoretical_loss": 3.4360528309676406,
      "tokens_seen": 1960902656
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.196918632643235e-05,
      "loss": 2.5381,
      "theoretical_loss": 3.436033607560785,
      "tokens_seen": 1961033728
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.196116193227412e-05,
      "loss": 2.5904,
      "theoretical_loss": 3.4360143857984777,
      "tokens_seen": 1961164800
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1080387,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7845823764801025,
      "objective/train/theoretical_loss": 3.4359951656804677,
      "objective/train/tokens_used": 331754976,
      "theoretical_loss": 3.4359951656804677,
      "tokens_seen": 1961295872
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.195313753811587e-05,
      "loss": 2.6118,
      "theoretical_loss": 3.4359951656804677,
      "tokens_seen": 1961295872
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.194511314395762e-05,
      "loss": 2.5112,
      "theoretical_loss": 3.435975947206505,
      "tokens_seen": 1961426944
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.193708874979939e-05,
      "loss": 2.5616,
      "theoretical_loss": 3.4359567303763385,
      "tokens_seen": 1961558016
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.192906435564114e-05,
      "loss": 2.472,
      "theoretical_loss": 3.4359375151897185,
      "tokens_seen": 1961689088
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.192103996148291e-05,
      "loss": 2.5179,
      "theoretical_loss": 3.4359183016463946,
      "tokens_seen": 1961820160
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.191301556732466e-05,
      "loss": 2.4378,
      "theoretical_loss": 3.435899089746116,
      "tokens_seen": 1961951232
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.190499117316643e-05,
      "loss": 2.3945,
      "theoretical_loss": 3.435879879488633,
      "tokens_seen": 1962082304
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.189696677900818e-05,
      "loss": 2.592,
      "theoretical_loss": 3.4358606708736943,
      "tokens_seen": 1962213376
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.188894238484994e-05,
      "loss": 2.5364,
      "theoretical_loss": 3.435841463901051,
      "tokens_seen": 1962344448
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.18809179906917e-05,
      "loss": 2.5239,
      "theoretical_loss": 3.4358222585704525,
      "tokens_seen": 1962475520
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.187289359653346e-05,
      "loss": 2.5953,
      "theoretical_loss": 3.4358030548816485,
      "tokens_seen": 1962606592
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.186486920237522e-05,
      "loss": 2.4442,
      "theoretical_loss": 3.4357838528343896,
      "tokens_seen": 1962737664
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.185684480821698e-05,
      "loss": 2.5722,
      "theoretical_loss": 3.4357646524284258,
      "tokens_seen": 1962868736
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1080811,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.384798049926758,
      "objective/train/theoretical_loss": 3.435755052840851,
      "objective/train/tokens_used": 333393376,
      "theoretical_loss": 3.435755052840851,
      "tokens_seen": 1962934272
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.184882041405874e-05,
      "loss": 2.4926,
      "theoretical_loss": 3.4357454536635066,
      "tokens_seen": 1962999808
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.18407960199005e-05,
      "loss": 2.5586,
      "theoretical_loss": 3.4357262565393825,
      "tokens_seen": 1963130880
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.183277162574225e-05,
      "loss": 2.5849,
      "theoretical_loss": 3.4357070610558043,
      "tokens_seen": 1963261952
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.182474723158402e-05,
      "loss": 2.5331,
      "theoretical_loss": 3.4356878672125215,
      "tokens_seen": 1963393024
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.181672283742577e-05,
      "loss": 2.433,
      "theoretical_loss": 3.4356686750092846,
      "tokens_seen": 1963524096
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.180869844326754e-05,
      "loss": 2.6144,
      "theoretical_loss": 3.4356494844458445,
      "tokens_seen": 1963655168
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.180067404910929e-05,
      "loss": 2.5068,
      "theoretical_loss": 3.435630295521951,
      "tokens_seen": 1963786240
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.179264965495106e-05,
      "loss": 2.4976,
      "theoretical_loss": 3.4356111082373553,
      "tokens_seen": 1963917312
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.178462526079281e-05,
      "loss": 2.5668,
      "theoretical_loss": 3.435591922591807,
      "tokens_seen": 1964048384
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.177660086663458e-05,
      "loss": 2.5539,
      "theoretical_loss": 3.4355727385850576,
      "tokens_seen": 1964179456
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.176857647247633e-05,
      "loss": 2.4739,
      "theoretical_loss": 3.435553556216858,
      "tokens_seen": 1964310528
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.176055207831808e-05,
      "loss": 2.4657,
      "theoretical_loss": 3.4355343754869576,
      "tokens_seen": 1964441600
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1082085,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1325109004974365,
      "objective/train/theoretical_loss": 3.4355151963951083,
      "objective/train/tokens_used": 335031776,
      "theoretical_loss": 3.4355151963951083,
      "tokens_seen": 1964572672
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.175252768415985e-05,
      "loss": 2.5154,
      "theoretical_loss": 3.4355151963951083,
      "tokens_seen": 1964572672
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.17445032900016e-05,
      "loss": 2.586,
      "theoretical_loss": 3.4354960189410604,
      "tokens_seen": 1964703744
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.173647889584337e-05,
      "loss": 2.6418,
      "theoretical_loss": 3.4354768431245652,
      "tokens_seen": 1964834816
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.172845450168512e-05,
      "loss": 2.566,
      "theoretical_loss": 3.4354576689453733,
      "tokens_seen": 1964965888
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.172043010752689e-05,
      "loss": 2.5053,
      "theoretical_loss": 3.4354384964032363,
      "tokens_seen": 1965096960
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.171240571336864e-05,
      "loss": 2.4657,
      "theoretical_loss": 3.435419325497904,
      "tokens_seen": 1965228032
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.17043813192104e-05,
      "loss": 2.5324,
      "theoretical_loss": 3.4354001562291288,
      "tokens_seen": 1965359104
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.169635692505216e-05,
      "loss": 2.468,
      "theoretical_loss": 3.4353809885966613,
      "tokens_seen": 1965490176
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.168833253089391e-05,
      "loss": 2.5399,
      "theoretical_loss": 3.4353618226002527,
      "tokens_seen": 1965621248
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.168030813673568e-05,
      "loss": 2.5918,
      "theoretical_loss": 3.4353426582396542,
      "tokens_seen": 1965752320
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.167228374257743e-05,
      "loss": 2.5126,
      "theoretical_loss": 3.4353234955146177,
      "tokens_seen": 1965883392
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.16642593484192e-05,
      "loss": 2.4399,
      "theoretical_loss": 3.4353043344248944,
      "tokens_seen": 1966014464
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.165623495426095e-05,
      "loss": 2.465,
      "theoretical_loss": 3.435285174970235,
      "tokens_seen": 1966145536
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1082685,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4286997318267822,
      "objective/train/theoretical_loss": 3.435275595855977,
      "objective/train/tokens_used": 336670176,
      "theoretical_loss": 3.435275595855977,
      "tokens_seen": 1966211072
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.164821056010271e-05,
      "loss": 2.4441,
      "theoretical_loss": 3.4352660171503913,
      "tokens_seen": 1966276608
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.164018616594447e-05,
      "loss": 2.586,
      "theoretical_loss": 3.4352468609651154,
      "tokens_seen": 1966407680
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.163216177178623e-05,
      "loss": 2.5347,
      "theoretical_loss": 3.4352277064141585,
      "tokens_seen": 1966538752
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.1624137377628e-05,
      "loss": 2.4911,
      "theoretical_loss": 3.4352085534972723,
      "tokens_seen": 1966669824
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.161611298346975e-05,
      "loss": 2.6297,
      "theoretical_loss": 3.435189402214208,
      "tokens_seen": 1966800896
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.160808858931151e-05,
      "loss": 2.4255,
      "theoretical_loss": 3.4351702525647183,
      "tokens_seen": 1966931968
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.160006419515327e-05,
      "loss": 2.5653,
      "theoretical_loss": 3.4351511045485545,
      "tokens_seen": 1967063040
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.159203980099502e-05,
      "loss": 2.5436,
      "theoretical_loss": 3.4351319581654685,
      "tokens_seen": 1967194112
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.158401540683679e-05,
      "loss": 2.6022,
      "theoretical_loss": 3.4351128134152122,
      "tokens_seen": 1967325184
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.157599101267854e-05,
      "loss": 2.4349,
      "theoretical_loss": 3.4350936702975376,
      "tokens_seen": 1967456256
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.15679666185203e-05,
      "loss": 2.4711,
      "theoretical_loss": 3.435074528812197,
      "tokens_seen": 1967587328
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.155994222436206e-05,
      "loss": 2.5145,
      "theoretical_loss": 3.435055388958942,
      "tokens_seen": 1967718400
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1083643,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4880809783935547,
      "objective/train/theoretical_loss": 3.435036250737525,
      "objective/train/tokens_used": 338308576,
      "theoretical_loss": 3.435036250737525,
      "tokens_seen": 1967849472
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.155191783020383e-05,
      "loss": 2.5831,
      "theoretical_loss": 3.435036250737525,
      "tokens_seen": 1967849472
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.154389343604558e-05,
      "loss": 2.5233,
      "theoretical_loss": 3.4350171141476986,
      "tokens_seen": 1967980544
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.153586904188733e-05,
      "loss": 2.4237,
      "theoretical_loss": 3.4349979791892142,
      "tokens_seen": 1968111616
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.15278446477291e-05,
      "loss": 2.5631,
      "theoretical_loss": 3.4349788458618247,
      "tokens_seen": 1968242688
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.151982025357085e-05,
      "loss": 2.4931,
      "theoretical_loss": 3.434959714165282,
      "tokens_seen": 1968373760
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.151179585941262e-05,
      "loss": 2.7059,
      "theoretical_loss": 3.4349405840993392,
      "tokens_seen": 1968504832
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.150377146525437e-05,
      "loss": 2.4426,
      "theoretical_loss": 3.434921455663748,
      "tokens_seen": 1968635904
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.149574707109614e-05,
      "loss": 2.5333,
      "theoretical_loss": 3.4349023288582616,
      "tokens_seen": 1968766976
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.148772267693789e-05,
      "loss": 2.5284,
      "theoretical_loss": 3.4348832036826322,
      "tokens_seen": 1968898048
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.147969828277966e-05,
      "loss": 2.5927,
      "theoretical_loss": 3.434864080136612,
      "tokens_seen": 1969029120
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.147167388862141e-05,
      "loss": 2.7439,
      "theoretical_loss": 3.434844958219955,
      "tokens_seen": 1969160192
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.146364949446317e-05,
      "loss": 2.5252,
      "theoretical_loss": 3.434825837932413,
      "tokens_seen": 1969291264
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.145562510030493e-05,
      "loss": 2.674,
      "theoretical_loss": 3.4348067192737384,
      "tokens_seen": 1969422336
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1084001,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6455793380737305,
      "objective/train/theoretical_loss": 3.4347971605551493,
      "objective/train/tokens_used": 339946976,
      "theoretical_loss": 3.4347971605551493,
      "tokens_seen": 1969487872
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.144760070614668e-05,
      "loss": 2.4479,
      "theoretical_loss": 3.4347876022436843,
      "tokens_seen": 1969553408
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.143957631198845e-05,
      "loss": 2.5325,
      "theoretical_loss": 3.434768486842005,
      "tokens_seen": 1969684480
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.14315519178302e-05,
      "loss": 2.6488,
      "theoretical_loss": 3.434749373068451,
      "tokens_seen": 1969815552
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.142352752367197e-05,
      "loss": 2.4881,
      "theoretical_loss": 3.4347302609227772,
      "tokens_seen": 1969946624
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.141550312951372e-05,
      "loss": 2.4335,
      "theoretical_loss": 3.434711150404736,
      "tokens_seen": 1970077696
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.140747873535548e-05,
      "loss": 2.48,
      "theoretical_loss": 3.4346920415140803,
      "tokens_seen": 1970208768
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.139945434119724e-05,
      "loss": 2.6898,
      "theoretical_loss": 3.434672934250564,
      "tokens_seen": 1970339840
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.1391429947039e-05,
      "loss": 2.5244,
      "theoretical_loss": 3.434653828613939,
      "tokens_seen": 1970470912
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.138340555288076e-05,
      "loss": 2.4301,
      "theoretical_loss": 3.434634724603961,
      "tokens_seen": 1970601984
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.137538115872252e-05,
      "loss": 2.5058,
      "theoretical_loss": 3.43461562222038,
      "tokens_seen": 1970733056
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.136735676456428e-05,
      "loss": 2.5456,
      "theoretical_loss": 3.434596521462952,
      "tokens_seen": 1970864128
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.135933237040604e-05,
      "loss": 2.4504,
      "theoretical_loss": 3.4345774223314294,
      "tokens_seen": 1970995200
    },
    {
      "epoch": 0.19,
      "objective/train/docs_used": 1085329,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.782360076904297,
      "objective/train/theoretical_loss": 3.4345583248255656,
      "objective/train/tokens_used": 341585376,
      "theoretical_loss": 3.4345583248255656,
      "tokens_seen": 1971126272
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.135130797624779e-05,
      "loss": 2.4819,
      "theoretical_loss": 3.4345583248255656,
      "tokens_seen": 1971126272
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.134328358208956e-05,
      "loss": 2.6041,
      "theoretical_loss": 3.434539228945115,
      "tokens_seen": 1971257344
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.133525918793131e-05,
      "loss": 2.5261,
      "theoretical_loss": 3.43452013468983,
      "tokens_seen": 1971388416
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.132723479377308e-05,
      "loss": 2.4584,
      "theoretical_loss": 3.4345010420594653,
      "tokens_seen": 1971519488
    },
    {
      "epoch": 0.19,
      "learning_rate": 8.131921039961483e-05,
      "loss": 2.3719,
      "theoretical_loss": 3.434481951053774,
      "tokens_seen": 1971650560
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.13111860054566e-05,
      "loss": 2.5287,
      "theoretical_loss": 3.4344628616725097,
      "tokens_seen": 1971781632
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.130316161129835e-05,
      "loss": 2.7084,
      "theoretical_loss": 3.4344437739154268,
      "tokens_seen": 1971912704
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.12951372171401e-05,
      "loss": 2.4448,
      "theoretical_loss": 3.434424687782279,
      "tokens_seen": 1972043776
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.128711282298187e-05,
      "loss": 2.3985,
      "theoretical_loss": 3.4344056032728196,
      "tokens_seen": 1972174848
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.127908842882362e-05,
      "loss": 2.6247,
      "theoretical_loss": 3.4343865203868034,
      "tokens_seen": 1972305920
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.127106403466539e-05,
      "loss": 2.6056,
      "theoretical_loss": 3.4343674391239842,
      "tokens_seen": 1972436992
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.126303964050714e-05,
      "loss": 2.6542,
      "theoretical_loss": 3.4343483594841158,
      "tokens_seen": 1972568064
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.125501524634891e-05,
      "loss": 2.48,
      "theoretical_loss": 3.4343292814669524,
      "tokens_seen": 1972699136
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1086429,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2790753841400146,
      "objective/train/theoretical_loss": 3.4343197430668084,
      "objective/train/tokens_used": 343223776,
      "theoretical_loss": 3.4343197430668084,
      "tokens_seen": 1972764672
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.124699085219066e-05,
      "loss": 2.6051,
      "theoretical_loss": 3.4343102050722485,
      "tokens_seen": 1972830208
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.123896645803242e-05,
      "loss": 2.5512,
      "theoretical_loss": 3.434291130299758,
      "tokens_seen": 1972961280
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.123094206387418e-05,
      "loss": 2.3245,
      "theoretical_loss": 3.4342720571492356,
      "tokens_seen": 1973092352
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.122291766971594e-05,
      "loss": 2.538,
      "theoretical_loss": 3.434252985620435,
      "tokens_seen": 1973223424
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.12148932755577e-05,
      "loss": 2.6185,
      "theoretical_loss": 3.4342339157131114,
      "tokens_seen": 1973354496
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.120686888139945e-05,
      "loss": 2.5536,
      "theoretical_loss": 3.434214847427018,
      "tokens_seen": 1973485568
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.119884448724122e-05,
      "loss": 2.3628,
      "theoretical_loss": 3.434195780761911,
      "tokens_seen": 1973616640
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.119082009308297e-05,
      "loss": 2.4919,
      "theoretical_loss": 3.434176715717544,
      "tokens_seen": 1973747712
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.118279569892473e-05,
      "loss": 2.5678,
      "theoretical_loss": 3.4341576522936714,
      "tokens_seen": 1973878784
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.11747713047665e-05,
      "loss": 2.6441,
      "theoretical_loss": 3.4341385904900483,
      "tokens_seen": 1974009856
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.116674691060825e-05,
      "loss": 2.5687,
      "theoretical_loss": 3.4341195303064294,
      "tokens_seen": 1974140928
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.115872251645001e-05,
      "loss": 2.5559,
      "theoretical_loss": 3.4341004717425694,
      "tokens_seen": 1974272000
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1087105,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.320133924484253,
      "objective/train/theoretical_loss": 3.4340814147982233,
      "objective/train/tokens_used": 344862176,
      "theoretical_loss": 3.4340814147982233,
      "tokens_seen": 1974403072
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.115069812229177e-05,
      "loss": 2.4467,
      "theoretical_loss": 3.4340814147982233,
      "tokens_seen": 1974403072
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.114267372813353e-05,
      "loss": 2.4401,
      "theoretical_loss": 3.4340623594731454,
      "tokens_seen": 1974534144
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.113464933397529e-05,
      "loss": 2.6396,
      "theoretical_loss": 3.4340433057670907,
      "tokens_seen": 1974665216
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.112662493981705e-05,
      "loss": 2.6161,
      "theoretical_loss": 3.434024253679815,
      "tokens_seen": 1974796288
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.111860054565881e-05,
      "loss": 2.8056,
      "theoretical_loss": 3.434005203211073,
      "tokens_seen": 1974927360
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.111057615150056e-05,
      "loss": 2.5804,
      "theoretical_loss": 3.4339861543606194,
      "tokens_seen": 1975058432
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.110255175734233e-05,
      "loss": 2.4813,
      "theoretical_loss": 3.4339671071282094,
      "tokens_seen": 1975189504
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.109452736318408e-05,
      "loss": 2.5453,
      "theoretical_loss": 3.4339480615135987,
      "tokens_seen": 1975320576
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.108650296902585e-05,
      "loss": 2.618,
      "theoretical_loss": 3.4339290175165424,
      "tokens_seen": 1975451648
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.10784785748676e-05,
      "loss": 2.5069,
      "theoretical_loss": 3.4339099751367956,
      "tokens_seen": 1975582720
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.107045418070937e-05,
      "loss": 2.498,
      "theoretical_loss": 3.433890934374114,
      "tokens_seen": 1975713792
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.106242978655112e-05,
      "loss": 2.5357,
      "theoretical_loss": 3.4338718952282523,
      "tokens_seen": 1975844864
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.105440539239287e-05,
      "loss": 2.4488,
      "theoretical_loss": 3.4338528576989664,
      "tokens_seen": 1975975936
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1088406,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.515301465988159,
      "objective/train/theoretical_loss": 3.433843339540463,
      "objective/train/tokens_used": 346500576,
      "theoretical_loss": 3.433843339540463,
      "tokens_seen": 1976041472
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.104638099823464e-05,
      "loss": 2.4573,
      "theoretical_loss": 3.4338338217860116,
      "tokens_seen": 1976107008
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.103835660407639e-05,
      "loss": 2.5943,
      "theoretical_loss": 3.433814787489144,
      "tokens_seen": 1976238080
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.103033220991816e-05,
      "loss": 2.6775,
      "theoretical_loss": 3.433795754808119,
      "tokens_seen": 1976369152
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.102230781575991e-05,
      "loss": 2.5248,
      "theoretical_loss": 3.4337767237426924,
      "tokens_seen": 1976500224
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.101428342160168e-05,
      "loss": 2.5347,
      "theoretical_loss": 3.43375769429262,
      "tokens_seen": 1976631296
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.100625902744343e-05,
      "loss": 2.4765,
      "theoretical_loss": 3.4337386664576566,
      "tokens_seen": 1976762368
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.099823463328519e-05,
      "loss": 2.4933,
      "theoretical_loss": 3.433719640237559,
      "tokens_seen": 1976893440
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.099021023912695e-05,
      "loss": 2.6944,
      "theoretical_loss": 3.433700615632083,
      "tokens_seen": 1977024512
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.09821858449687e-05,
      "loss": 2.5308,
      "theoretical_loss": 3.4336815926409843,
      "tokens_seen": 1977155584
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.097416145081047e-05,
      "loss": 2.6257,
      "theoretical_loss": 3.433662571264019,
      "tokens_seen": 1977286656
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.096613705665222e-05,
      "loss": 2.4688,
      "theoretical_loss": 3.4336435515009436,
      "tokens_seen": 1977417728
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.095811266249399e-05,
      "loss": 2.4563,
      "theoretical_loss": 3.433624533351513,
      "tokens_seen": 1977548800
    },
    {
      "debugging/Self-BLEU-5": 0.5383387038400951,
      "debugging/distinct-1-grams": 0.7574669480332016,
      "debugging/distinct-2-grams": 0.9486429063642564,
      "debugging/entropy-1-grams": 5.78993480235564,
      "debugging/entropy-2-grams": 6.620072215652616,
      "debugging/length": 591.3,
      "debugging/num_segments": 10,
      "debugging/score": 0.004437156266881241,
      "debugging/score_std": 0.0035169708188302635,
      "epoch": 0.2,
      "objective/train/docs_used": 1088993,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6108832359313965,
      "objective/train/theoretical_loss": 3.4336055168154846,
      "objective/train/tokens_used": 348138976,
      "theoretical_loss": 3.4336055168154846,
      "tokens_seen": 1977679872
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.095008826833574e-05,
      "loss": 2.5314,
      "theoretical_loss": 3.4336055168154846,
      "tokens_seen": 1977679872
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.09420638741775e-05,
      "loss": 2.5313,
      "theoretical_loss": 3.4335865018926146,
      "tokens_seen": 1977810944
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.093403948001926e-05,
      "loss": 2.3723,
      "theoretical_loss": 3.433567488582658,
      "tokens_seen": 1977942016
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.092601508586102e-05,
      "loss": 2.4287,
      "theoretical_loss": 3.433548476885372,
      "tokens_seen": 1978073088
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.091799069170278e-05,
      "loss": 2.49,
      "theoretical_loss": 3.433529466800514,
      "tokens_seen": 1978204160
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.090996629754454e-05,
      "loss": 2.5583,
      "theoretical_loss": 3.433510458327838,
      "tokens_seen": 1978335232
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.09019419033863e-05,
      "loss": 2.4996,
      "theoretical_loss": 3.4334914514671024,
      "tokens_seen": 1978466304
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.089391750922806e-05,
      "loss": 2.5626,
      "theoretical_loss": 3.433472446218063,
      "tokens_seen": 1978597376
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.088589311506981e-05,
      "loss": 2.5652,
      "theoretical_loss": 3.4334534425804764,
      "tokens_seen": 1978728448
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.087786872091158e-05,
      "loss": 2.4335,
      "theoretical_loss": 3.4334344405541,
      "tokens_seen": 1978859520
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.086984432675333e-05,
      "loss": 2.495,
      "theoretical_loss": 3.433415440138689,
      "tokens_seen": 1978990592
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.08618199325951e-05,
      "loss": 2.4582,
      "theoretical_loss": 3.4333964413340015,
      "tokens_seen": 1979121664
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.085379553843685e-05,
      "loss": 2.5634,
      "theoretical_loss": 3.4333774441397935,
      "tokens_seen": 1979252736
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1090226,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6920125484466553,
      "objective/train/theoretical_loss": 3.4333679461465434,
      "objective/train/tokens_used": 349777376,
      "theoretical_loss": 3.4333679461465434,
      "tokens_seen": 1979318272
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.084577114427862e-05,
      "loss": 2.6059,
      "theoretical_loss": 3.433358448555822,
      "tokens_seen": 1979383808
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.083774675012037e-05,
      "loss": 2.5343,
      "theoretical_loss": 3.4333394545818443,
      "tokens_seen": 1979514880
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.082972235596214e-05,
      "loss": 2.588,
      "theoretical_loss": 3.433320462217617,
      "tokens_seen": 1979645952
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.082169796180389e-05,
      "loss": 2.5355,
      "theoretical_loss": 3.4333014714628973,
      "tokens_seen": 1979777024
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.081367356764564e-05,
      "loss": 2.7333,
      "theoretical_loss": 3.433282482317442,
      "tokens_seen": 1979908096
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.080564917348741e-05,
      "loss": 2.4832,
      "theoretical_loss": 3.4332634947810083,
      "tokens_seen": 1980039168
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.079762477932916e-05,
      "loss": 2.5933,
      "theoretical_loss": 3.4332445088533534,
      "tokens_seen": 1980170240
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.078960038517093e-05,
      "loss": 2.5591,
      "theoretical_loss": 3.433225524534235,
      "tokens_seen": 1980301312
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.078157599101268e-05,
      "loss": 2.6487,
      "theoretical_loss": 3.433206541823409,
      "tokens_seen": 1980432384
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.077355159685445e-05,
      "loss": 2.4959,
      "theoretical_loss": 3.433187560720634,
      "tokens_seen": 1980563456
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.07655272026962e-05,
      "loss": 2.6386,
      "theoretical_loss": 3.433168581225667,
      "tokens_seen": 1980694528
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.075750280853796e-05,
      "loss": 2.5598,
      "theoretical_loss": 3.433149603338266,
      "tokens_seen": 1980825600
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1091410,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.233128309249878,
      "objective/train/theoretical_loss": 3.433130627058187,
      "objective/train/tokens_used": 351415776,
      "theoretical_loss": 3.433130627058187,
      "tokens_seen": 1980956672
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.074947841437972e-05,
      "loss": 2.4389,
      "theoretical_loss": 3.433130627058187,
      "tokens_seen": 1980956672
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.074145402022148e-05,
      "loss": 2.6606,
      "theoretical_loss": 3.433111652385189,
      "tokens_seen": 1981087744
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.073342962606324e-05,
      "loss": 2.614,
      "theoretical_loss": 3.433092679319029,
      "tokens_seen": 1981218816
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.0725405231905e-05,
      "loss": 2.5507,
      "theoretical_loss": 3.4330737078594638,
      "tokens_seen": 1981349888
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.071738083774676e-05,
      "loss": 2.7506,
      "theoretical_loss": 3.433054738006253,
      "tokens_seen": 1981480960
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.070935644358851e-05,
      "loss": 2.5217,
      "theoretical_loss": 3.4330357697591527,
      "tokens_seen": 1981612032
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.070133204943027e-05,
      "loss": 2.5975,
      "theoretical_loss": 3.4330168031179213,
      "tokens_seen": 1981743104
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.069330765527203e-05,
      "loss": 2.5798,
      "theoretical_loss": 3.4329978380823167,
      "tokens_seen": 1981874176
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.068528326111379e-05,
      "loss": 2.5297,
      "theoretical_loss": 3.432978874652097,
      "tokens_seen": 1982005248
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.067725886695555e-05,
      "loss": 2.5863,
      "theoretical_loss": 3.4329599128270196,
      "tokens_seen": 1982136320
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.066923447279731e-05,
      "loss": 2.4973,
      "theoretical_loss": 3.4329409526068426,
      "tokens_seen": 1982267392
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.066121007863907e-05,
      "loss": 2.4644,
      "theoretical_loss": 3.4329219939913247,
      "tokens_seen": 1982398464
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.065318568448083e-05,
      "loss": 2.5549,
      "theoretical_loss": 3.432903036980223,
      "tokens_seen": 1982529536
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1091797,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8559112548828125,
      "objective/train/theoretical_loss": 3.4328935590762533,
      "objective/train/tokens_used": 353054176,
      "theoretical_loss": 3.4328935590762533,
      "tokens_seen": 1982595072
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.064516129032258e-05,
      "loss": 2.5303,
      "theoretical_loss": 3.4328840815732966,
      "tokens_seen": 1982660608
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.063713689616435e-05,
      "loss": 2.521,
      "theoretical_loss": 3.4328651277703033,
      "tokens_seen": 1982791680
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.06291125020061e-05,
      "loss": 2.5473,
      "theoretical_loss": 3.432846175571001,
      "tokens_seen": 1982922752
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.062108810784787e-05,
      "loss": 2.6932,
      "theoretical_loss": 3.4328272249751492,
      "tokens_seen": 1983053824
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.061306371368962e-05,
      "loss": 2.6695,
      "theoretical_loss": 3.432808275982505,
      "tokens_seen": 1983184896
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.060503931953139e-05,
      "loss": 2.5086,
      "theoretical_loss": 3.4327893285928273,
      "tokens_seen": 1983315968
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.059701492537314e-05,
      "loss": 2.5303,
      "theoretical_loss": 3.432770382805874,
      "tokens_seen": 1983447040
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.058899053121489e-05,
      "loss": 2.5718,
      "theoretical_loss": 3.432751438621405,
      "tokens_seen": 1983578112
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.058096613705666e-05,
      "loss": 2.6883,
      "theoretical_loss": 3.4327324960391783,
      "tokens_seen": 1983709184
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.057294174289841e-05,
      "loss": 2.6956,
      "theoretical_loss": 3.4327135550589514,
      "tokens_seen": 1983840256
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.056491734874018e-05,
      "loss": 2.6438,
      "theoretical_loss": 3.432694615680485,
      "tokens_seen": 1983971328
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.055689295458193e-05,
      "loss": 2.5278,
      "theoretical_loss": 3.4326756779035357,
      "tokens_seen": 1984102400
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1092754,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2613365650177,
      "objective/train/theoretical_loss": 3.4326567417278637,
      "objective/train/tokens_used": 354692576,
      "theoretical_loss": 3.4326567417278637,
      "tokens_seen": 1984233472
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.05488685604237e-05,
      "loss": 2.54,
      "theoretical_loss": 3.4326567417278637,
      "tokens_seen": 1984233472
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.054084416626545e-05,
      "loss": 2.4222,
      "theoretical_loss": 3.432637807153228,
      "tokens_seen": 1984364544
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.053281977210722e-05,
      "loss": 2.6055,
      "theoretical_loss": 3.4326188741793864,
      "tokens_seen": 1984495616
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.052479537794897e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.4325999428060987,
      "tokens_seen": 1984626688
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.051677098379073e-05,
      "loss": 2.7175,
      "theoretical_loss": 3.432581013033124,
      "tokens_seen": 1984757760
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.050874658963249e-05,
      "loss": 2.4612,
      "theoretical_loss": 3.43256208486022,
      "tokens_seen": 1984888832
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.050072219547425e-05,
      "loss": 2.5532,
      "theoretical_loss": 3.4325431582871473,
      "tokens_seen": 1985019904
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.049269780131601e-05,
      "loss": 2.4733,
      "theoretical_loss": 3.4325242333136643,
      "tokens_seen": 1985150976
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.048467340715777e-05,
      "loss": 2.6379,
      "theoretical_loss": 3.432505309939531,
      "tokens_seen": 1985282048
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.047664901299953e-05,
      "loss": 2.6498,
      "theoretical_loss": 3.432486388164506,
      "tokens_seen": 1985413120
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.046862461884128e-05,
      "loss": 2.5589,
      "theoretical_loss": 3.432467467988348,
      "tokens_seen": 1985544192
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.046060022468304e-05,
      "loss": 2.6433,
      "theoretical_loss": 3.4324485494108177,
      "tokens_seen": 1985675264
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.04525758305248e-05,
      "loss": 2.59,
      "theoretical_loss": 3.4324296324316736,
      "tokens_seen": 1985806336
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1093159,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2847602367401123,
      "objective/train/theoretical_loss": 3.432420174541422,
      "objective/train/tokens_used": 356330976,
      "theoretical_loss": 3.432420174541422,
      "tokens_seen": 1985871872
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.044455143636656e-05,
      "loss": 2.6555,
      "theoretical_loss": 3.4324107170506757,
      "tokens_seen": 1985937408
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.043652704220832e-05,
      "loss": 2.5619,
      "theoretical_loss": 3.432391803267583,
      "tokens_seen": 1986068480
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.042850264805008e-05,
      "loss": 2.7432,
      "theoretical_loss": 3.432372891082156,
      "tokens_seen": 1986199552
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.042047825389184e-05,
      "loss": 2.6317,
      "theoretical_loss": 3.4323539804941534,
      "tokens_seen": 1986330624
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.04124538597336e-05,
      "loss": 2.5587,
      "theoretical_loss": 3.432335071503335,
      "tokens_seen": 1986461696
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.040442946557535e-05,
      "loss": 2.7005,
      "theoretical_loss": 3.4323161641094604,
      "tokens_seen": 1986592768
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.039640507141712e-05,
      "loss": 2.6133,
      "theoretical_loss": 3.4322972583122904,
      "tokens_seen": 1986723840
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.038838067725887e-05,
      "loss": 2.6584,
      "theoretical_loss": 3.4322783541115838,
      "tokens_seen": 1986854912
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.038035628310064e-05,
      "loss": 2.6685,
      "theoretical_loss": 3.4322594515071003,
      "tokens_seen": 1986985984
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.037233188894239e-05,
      "loss": 2.6374,
      "theoretical_loss": 3.4322405504986007,
      "tokens_seen": 1987117056
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.036430749478416e-05,
      "loss": 2.6503,
      "theoretical_loss": 3.432221651085845,
      "tokens_seen": 1987248128
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.035628310062591e-05,
      "loss": 2.637,
      "theoretical_loss": 3.4322027532685926,
      "tokens_seen": 1987379200
    },
    {
      "epoch": 0.2,
      "objective/train/docs_used": 1094053,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.824427366256714,
      "objective/train/theoretical_loss": 3.4321838570466037,
      "objective/train/tokens_used": 357969376,
      "theoretical_loss": 3.4321838570466037,
      "tokens_seen": 1987510272
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.034825870646766e-05,
      "loss": 2.658,
      "theoretical_loss": 3.4321838570466037,
      "tokens_seen": 1987510272
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.034023431230943e-05,
      "loss": 2.568,
      "theoretical_loss": 3.4321649624196384,
      "tokens_seen": 1987641344
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.033220991815118e-05,
      "loss": 2.4641,
      "theoretical_loss": 3.432146069387458,
      "tokens_seen": 1987772416
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.032418552399295e-05,
      "loss": 2.5967,
      "theoretical_loss": 3.432127177949821,
      "tokens_seen": 1987903488
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.03161611298347e-05,
      "loss": 2.6815,
      "theoretical_loss": 3.4321082881064884,
      "tokens_seen": 1988034560
    },
    {
      "epoch": 0.2,
      "learning_rate": 8.030813673567647e-05,
      "loss": 2.6253,
      "theoretical_loss": 3.4320893998572215,
      "tokens_seen": 1988165632
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.030011234151822e-05,
      "loss": 2.5624,
      "theoretical_loss": 3.4320705132017792,
      "tokens_seen": 1988296704
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.029208794735998e-05,
      "loss": 2.6203,
      "theoretical_loss": 3.4320516281399227,
      "tokens_seen": 1988427776
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.028406355320174e-05,
      "loss": 2.6069,
      "theoretical_loss": 3.432032744671413,
      "tokens_seen": 1988558848
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.02760391590435e-05,
      "loss": 2.6759,
      "theoretical_loss": 3.4320138627960097,
      "tokens_seen": 1988689920
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.026801476488526e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.4319949825134737,
      "tokens_seen": 1988820992
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.025999037072702e-05,
      "loss": 2.6477,
      "theoretical_loss": 3.4319761038235663,
      "tokens_seen": 1988952064
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.025196597656878e-05,
      "loss": 2.5959,
      "theoretical_loss": 3.4319572267260474,
      "tokens_seen": 1989083136
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1094680,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8489184379577637,
      "objective/train/theoretical_loss": 3.431947788774359,
      "objective/train/tokens_used": 359607776,
      "theoretical_loss": 3.431947788774359,
      "tokens_seen": 1989148672
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.024394158241054e-05,
      "loss": 2.6793,
      "theoretical_loss": 3.431938351220678,
      "tokens_seen": 1989214208
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.023591718825229e-05,
      "loss": 2.6479,
      "theoretical_loss": 3.4319194773072192,
      "tokens_seen": 1989345280
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.022789279409406e-05,
      "loss": 2.5235,
      "theoretical_loss": 3.4319006049854313,
      "tokens_seen": 1989476352
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.021986839993581e-05,
      "loss": 2.6275,
      "theoretical_loss": 3.431881734255076,
      "tokens_seen": 1989607424
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.021184400577757e-05,
      "loss": 2.6886,
      "theoretical_loss": 3.4318628651159138,
      "tokens_seen": 1989738496
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.020381961161933e-05,
      "loss": 2.5579,
      "theoretical_loss": 3.4318439975677055,
      "tokens_seen": 1989869568
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.01957952174611e-05,
      "loss": 2.5609,
      "theoretical_loss": 3.4318251316102124,
      "tokens_seen": 1990000640
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.018777082330285e-05,
      "loss": 2.5518,
      "theoretical_loss": 3.4318062672431964,
      "tokens_seen": 1990131712
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.017974642914461e-05,
      "loss": 2.7237,
      "theoretical_loss": 3.431787404466417,
      "tokens_seen": 1990262784
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.017172203498637e-05,
      "loss": 2.5871,
      "theoretical_loss": 3.431768543279637,
      "tokens_seen": 1990393856
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.016369764082812e-05,
      "loss": 2.4901,
      "theoretical_loss": 3.431749683682617,
      "tokens_seen": 1990524928
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.015567324666989e-05,
      "loss": 2.6152,
      "theoretical_loss": 3.4317308256751184,
      "tokens_seen": 1990656000
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1095652,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7679972648620605,
      "objective/train/theoretical_loss": 3.4317119692569023,
      "objective/train/tokens_used": 361246176,
      "theoretical_loss": 3.4317119692569023,
      "tokens_seen": 1990787072
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.014764885251164e-05,
      "loss": 2.5604,
      "theoretical_loss": 3.4317119692569023,
      "tokens_seen": 1990787072
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.013962445835341e-05,
      "loss": 2.7062,
      "theoretical_loss": 3.431693114427731,
      "tokens_seen": 1990918144
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.013160006419516e-05,
      "loss": 2.6594,
      "theoretical_loss": 3.431674261187365,
      "tokens_seen": 1991049216
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.012357567003693e-05,
      "loss": 2.5803,
      "theoretical_loss": 3.4316554095355665,
      "tokens_seen": 1991180288
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.011555127587868e-05,
      "loss": 2.6,
      "theoretical_loss": 3.4316365594720963,
      "tokens_seen": 1991311360
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.010752688172043e-05,
      "loss": 2.5181,
      "theoretical_loss": 3.431617710996717,
      "tokens_seen": 1991442432
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.00995024875622e-05,
      "loss": 2.4289,
      "theoretical_loss": 3.4315988641091906,
      "tokens_seen": 1991573504
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.009147809340395e-05,
      "loss": 2.6387,
      "theoretical_loss": 3.431580018809277,
      "tokens_seen": 1991704576
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.008345369924572e-05,
      "loss": 2.6242,
      "theoretical_loss": 3.4315611750967396,
      "tokens_seen": 1991835648
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.007542930508747e-05,
      "loss": 2.5817,
      "theoretical_loss": 3.4315423329713397,
      "tokens_seen": 1991966720
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.006740491092924e-05,
      "loss": 2.5962,
      "theoretical_loss": 3.4315234924328393,
      "tokens_seen": 1992097792
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.005938051677099e-05,
      "loss": 2.5673,
      "theoretical_loss": 3.4315046534810003,
      "tokens_seen": 1992228864
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.005135612261275e-05,
      "loss": 2.4096,
      "theoretical_loss": 3.4314858161155852,
      "tokens_seen": 1992359936
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1096842,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.55733060836792,
      "objective/train/theoretical_loss": 3.431476398027712,
      "objective/train/tokens_used": 362884576,
      "theoretical_loss": 3.431476398027712,
      "tokens_seen": 1992425472
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.004333172845451e-05,
      "loss": 2.5752,
      "theoretical_loss": 3.431466980336355,
      "tokens_seen": 1992491008
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.003530733429627e-05,
      "loss": 2.6069,
      "theoretical_loss": 3.4314481461430724,
      "tokens_seen": 1992622080
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.002728294013803e-05,
      "loss": 2.6405,
      "theoretical_loss": 3.4314293135355003,
      "tokens_seen": 1992753152
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.001925854597979e-05,
      "loss": 2.6135,
      "theoretical_loss": 3.4314104825133995,
      "tokens_seen": 1992884224
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.001123415182155e-05,
      "loss": 2.6832,
      "theoretical_loss": 3.431391653076533,
      "tokens_seen": 1993015296
    },
    {
      "epoch": 0.21,
      "learning_rate": 8.00032097576633e-05,
      "loss": 2.5489,
      "theoretical_loss": 3.4313728252246634,
      "tokens_seen": 1993146368
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.999518536350506e-05,
      "loss": 2.6522,
      "theoretical_loss": 3.431353998957552,
      "tokens_seen": 1993277440
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.998716096934683e-05,
      "loss": 2.7246,
      "theoretical_loss": 3.4313351742749623,
      "tokens_seen": 1993408512
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.997913657518858e-05,
      "loss": 2.7097,
      "theoretical_loss": 3.4313163511766565,
      "tokens_seen": 1993539584
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.997111218103034e-05,
      "loss": 2.7062,
      "theoretical_loss": 3.431297529662397,
      "tokens_seen": 1993670656
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.99630877868721e-05,
      "loss": 2.7169,
      "theoretical_loss": 3.431278709731946,
      "tokens_seen": 1993801728
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.995506339271386e-05,
      "loss": 2.5566,
      "theoretical_loss": 3.431259891385067,
      "tokens_seen": 1993932800
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1097514,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8202970027923584,
      "objective/train/theoretical_loss": 3.4312410746215214,
      "objective/train/tokens_used": 364522976,
      "theoretical_loss": 3.4312410746215214,
      "tokens_seen": 1994063872
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.994703899855562e-05,
      "loss": 2.5633,
      "theoretical_loss": 3.4312410746215214,
      "tokens_seen": 1994063872
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.993901460439737e-05,
      "loss": 2.7012,
      "theoretical_loss": 3.431222259441073,
      "tokens_seen": 1994194944
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.993099021023914e-05,
      "loss": 2.6065,
      "theoretical_loss": 3.431203445843485,
      "tokens_seen": 1994326016
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.992296581608089e-05,
      "loss": 2.6068,
      "theoretical_loss": 3.4311846338285186,
      "tokens_seen": 1994457088
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.991494142192266e-05,
      "loss": 2.527,
      "theoretical_loss": 3.4311658233959372,
      "tokens_seen": 1994588160
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.990691702776441e-05,
      "loss": 2.5969,
      "theoretical_loss": 3.431147014545505,
      "tokens_seen": 1994719232
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.989889263360616e-05,
      "loss": 2.7501,
      "theoretical_loss": 3.4311282072769833,
      "tokens_seen": 1994850304
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.989086823944793e-05,
      "loss": 2.5058,
      "theoretical_loss": 3.4311094015901364,
      "tokens_seen": 1994981376
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.988284384528968e-05,
      "loss": 2.6537,
      "theoretical_loss": 3.431090597484727,
      "tokens_seen": 1995112448
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.987481945113145e-05,
      "loss": 2.6582,
      "theoretical_loss": 3.431071794960517,
      "tokens_seen": 1995243520
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.98667950569732e-05,
      "loss": 2.6121,
      "theoretical_loss": 3.431052994017272,
      "tokens_seen": 1995374592
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.985877066281497e-05,
      "loss": 2.6305,
      "theoretical_loss": 3.431034194654753,
      "tokens_seen": 1995505664
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.985074626865672e-05,
      "loss": 2.5408,
      "theoretical_loss": 3.4310153968727244,
      "tokens_seen": 1995636736
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1098702,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.530543804168701,
      "objective/train/theoretical_loss": 3.43100599857432,
      "objective/train/tokens_used": 366161376,
      "theoretical_loss": 3.43100599857432,
      "tokens_seen": 1995702272
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.984272187449848e-05,
      "loss": 2.6301,
      "theoretical_loss": 3.4309966006709494,
      "tokens_seen": 1995767808
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.983469748034024e-05,
      "loss": 2.6202,
      "theoretical_loss": 3.4309778060491913,
      "tokens_seen": 1995898880
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.9826673086182e-05,
      "loss": 2.5104,
      "theoretical_loss": 3.430959013007213,
      "tokens_seen": 1996029952
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.981864869202376e-05,
      "loss": 2.7801,
      "theoretical_loss": 3.4309402215447795,
      "tokens_seen": 1996161024
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.981062429786552e-05,
      "loss": 2.6228,
      "theoretical_loss": 3.4309214316616528,
      "tokens_seen": 1996292096
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.980259990370727e-05,
      "loss": 2.6714,
      "theoretical_loss": 3.4309026433575966,
      "tokens_seen": 1996423168
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.979457550954904e-05,
      "loss": 2.7458,
      "theoretical_loss": 3.4308838566323754,
      "tokens_seen": 1996554240
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.978655111539079e-05,
      "loss": 2.6498,
      "theoretical_loss": 3.4308650714857523,
      "tokens_seen": 1996685312
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.977852672123256e-05,
      "loss": 2.5197,
      "theoretical_loss": 3.4308462879174915,
      "tokens_seen": 1996816384
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.977050232707431e-05,
      "loss": 2.5544,
      "theoretical_loss": 3.4308275059273563,
      "tokens_seen": 1996947456
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.976247793291608e-05,
      "loss": 2.6762,
      "theoretical_loss": 3.430808725515111,
      "tokens_seen": 1997078528
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.975445353875783e-05,
      "loss": 2.636,
      "theoretical_loss": 3.4307899466805187,
      "tokens_seen": 1997209600
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1099661,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.868697166442871,
      "objective/train/theoretical_loss": 3.4307711694233443,
      "objective/train/tokens_used": 367799776,
      "theoretical_loss": 3.4307711694233443,
      "tokens_seen": 1997340672
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.974642914459958e-05,
      "loss": 2.6351,
      "theoretical_loss": 3.4307711694233443,
      "tokens_seen": 1997340672
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.973840475044135e-05,
      "loss": 2.5566,
      "theoretical_loss": 3.430752393743351,
      "tokens_seen": 1997471744
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.97303803562831e-05,
      "loss": 2.6381,
      "theoretical_loss": 3.4307336196403035,
      "tokens_seen": 1997602816
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.972235596212487e-05,
      "loss": 2.7568,
      "theoretical_loss": 3.4307148471139657,
      "tokens_seen": 1997733888
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.971433156796662e-05,
      "loss": 2.5882,
      "theoretical_loss": 3.4306960761641014,
      "tokens_seen": 1997864960
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.970630717380837e-05,
      "loss": 2.578,
      "theoretical_loss": 3.430677306790475,
      "tokens_seen": 1997996032
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.969828277965014e-05,
      "loss": 2.5211,
      "theoretical_loss": 3.430658538992851,
      "tokens_seen": 1998127104
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.96902583854919e-05,
      "loss": 2.6673,
      "theoretical_loss": 3.4306397727709936,
      "tokens_seen": 1998258176
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.968223399133366e-05,
      "loss": 2.5972,
      "theoretical_loss": 3.4306210081246666,
      "tokens_seen": 1998389248
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.967420959717541e-05,
      "loss": 2.7591,
      "theoretical_loss": 3.4306022450536355,
      "tokens_seen": 1998520320
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.966618520301718e-05,
      "loss": 2.6882,
      "theoretical_loss": 3.4305834835576636,
      "tokens_seen": 1998651392
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.965816080885893e-05,
      "loss": 2.7569,
      "theoretical_loss": 3.4305647236365164,
      "tokens_seen": 1998782464
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.965013641470069e-05,
      "loss": 2.6635,
      "theoretical_loss": 3.4305459652899577,
      "tokens_seen": 1998913536
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1100646,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.641911745071411,
      "objective/train/theoretical_loss": 3.4305365867070754,
      "objective/train/tokens_used": 369438176,
      "theoretical_loss": 3.4305365867070754,
      "tokens_seen": 1998979072
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.964211202054245e-05,
      "loss": 2.6049,
      "theoretical_loss": 3.430527208517752,
      "tokens_seen": 1999044608
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.96340876263842e-05,
      "loss": 2.576,
      "theoretical_loss": 3.430508453319665,
      "tokens_seen": 1999175680
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.962606323222597e-05,
      "loss": 2.6991,
      "theoretical_loss": 3.4304896996954604,
      "tokens_seen": 1999306752
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.961803883806773e-05,
      "loss": 2.5727,
      "theoretical_loss": 3.430470947644903,
      "tokens_seen": 1999437824
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.961001444390948e-05,
      "loss": 2.7093,
      "theoretical_loss": 3.4304521971677584,
      "tokens_seen": 1999568896
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.960199004975125e-05,
      "loss": 2.5668,
      "theoretical_loss": 3.4304334482637904,
      "tokens_seen": 1999699968
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.9593965655593e-05,
      "loss": 2.6501,
      "theoretical_loss": 3.430414700932765,
      "tokens_seen": 1999831040
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.958594126143477e-05,
      "loss": 2.5152,
      "theoretical_loss": 3.430395955174446,
      "tokens_seen": 1999962112
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.957791686727652e-05,
      "loss": 2.6144,
      "theoretical_loss": 3.4303772109885995,
      "tokens_seen": 2000093184
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.956989247311829e-05,
      "loss": 2.5636,
      "theoretical_loss": 3.43035846837499,
      "tokens_seen": 2000224256
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.956186807896004e-05,
      "loss": 2.5419,
      "theoretical_loss": 3.430339727333383,
      "tokens_seen": 2000355328
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.955384368480179e-05,
      "loss": 2.6236,
      "theoretical_loss": 3.4303209878635426,
      "tokens_seen": 2000486400
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1101283,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.551567792892456,
      "objective/train/theoretical_loss": 3.4303022499652354,
      "objective/train/tokens_used": 371076576,
      "theoretical_loss": 3.4303022499652354,
      "tokens_seen": 2000617472
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.954581929064356e-05,
      "loss": 2.6409,
      "theoretical_loss": 3.4303022499652354,
      "tokens_seen": 2000617472
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.953779489648531e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.4302835136382255,
      "tokens_seen": 2000748544
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.952977050232708e-05,
      "loss": 2.5693,
      "theoretical_loss": 3.430264778882279,
      "tokens_seen": 2000879616
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.952174610816883e-05,
      "loss": 2.6041,
      "theoretical_loss": 3.430246045697161,
      "tokens_seen": 2001010688
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.951372171401059e-05,
      "loss": 2.501,
      "theoretical_loss": 3.4302273140826367,
      "tokens_seen": 2001141760
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.950569731985235e-05,
      "loss": 2.7269,
      "theoretical_loss": 3.430208584038472,
      "tokens_seen": 2001272832
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.94976729256941e-05,
      "loss": 2.5952,
      "theoretical_loss": 3.430189855564432,
      "tokens_seen": 2001403904
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.948964853153587e-05,
      "loss": 2.5573,
      "theoretical_loss": 3.430171128660283,
      "tokens_seen": 2001534976
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.948162413737762e-05,
      "loss": 2.6459,
      "theoretical_loss": 3.4301524033257893,
      "tokens_seen": 2001666048
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.947359974321939e-05,
      "loss": 2.7094,
      "theoretical_loss": 3.4301336795607176,
      "tokens_seen": 2001797120
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.946557534906114e-05,
      "loss": 2.5831,
      "theoretical_loss": 3.4301149573648333,
      "tokens_seen": 2001928192
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.94575509549029e-05,
      "loss": 2.5399,
      "theoretical_loss": 3.430096236737902,
      "tokens_seen": 2002059264
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.944952656074466e-05,
      "loss": 2.5772,
      "theoretical_loss": 3.43007751767969,
      "tokens_seen": 2002190336
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1102695,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5117669105529785,
      "objective/train/theoretical_loss": 3.4300681587387807,
      "objective/train/tokens_used": 372714976,
      "theoretical_loss": 3.4300681587387807,
      "tokens_seen": 2002255872
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.944150216658642e-05,
      "loss": 2.5223,
      "theoretical_loss": 3.430058800189963,
      "tokens_seen": 2002321408
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.943347777242818e-05,
      "loss": 2.5086,
      "theoretical_loss": 3.4300400842684864,
      "tokens_seen": 2002452480
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.942545337826994e-05,
      "loss": 2.5602,
      "theoretical_loss": 3.430021369915027,
      "tokens_seen": 2002583552
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.94174289841117e-05,
      "loss": 2.6225,
      "theoretical_loss": 3.43000265712935,
      "tokens_seen": 2002714624
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.940940458995346e-05,
      "loss": 2.5564,
      "theoretical_loss": 3.4299839459112222,
      "tokens_seen": 2002845696
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.940138019579521e-05,
      "loss": 2.5846,
      "theoretical_loss": 3.4299652362604087,
      "tokens_seen": 2002976768
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.939335580163698e-05,
      "loss": 2.578,
      "theoretical_loss": 3.429946528176677,
      "tokens_seen": 2003107840
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.938533140747873e-05,
      "loss": 2.5654,
      "theoretical_loss": 3.429927821659793,
      "tokens_seen": 2003238912
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.93773070133205e-05,
      "loss": 2.5886,
      "theoretical_loss": 3.4299091167095215,
      "tokens_seen": 2003369984
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.936928261916225e-05,
      "loss": 2.5983,
      "theoretical_loss": 3.42989041332563,
      "tokens_seen": 2003501056
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.936125822500402e-05,
      "loss": 2.6702,
      "theoretical_loss": 3.4298717115078854,
      "tokens_seen": 2003632128
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.935323383084577e-05,
      "loss": 2.6316,
      "theoretical_loss": 3.4298530112560535,
      "tokens_seen": 2003763200
    },
    {
      "epoch": 0.21,
      "objective/train/docs_used": 1103208,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2795019149780273,
      "objective/train/theoretical_loss": 3.4298343125699002,
      "objective/train/tokens_used": 374353376,
      "theoretical_loss": 3.4298343125699002,
      "tokens_seen": 2003894272
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.934520943668752e-05,
      "loss": 2.4894,
      "theoretical_loss": 3.4298343125699002,
      "tokens_seen": 2003894272
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.933718504252929e-05,
      "loss": 2.4966,
      "theoretical_loss": 3.4298156154491934,
      "tokens_seen": 2004025344
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.932916064837104e-05,
      "loss": 2.4743,
      "theoretical_loss": 3.429796919893698,
      "tokens_seen": 2004156416
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.932113625421281e-05,
      "loss": 2.5501,
      "theoretical_loss": 3.429778225903182,
      "tokens_seen": 2004287488
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.931311186005456e-05,
      "loss": 2.5934,
      "theoretical_loss": 3.4297595334774114,
      "tokens_seen": 2004418560
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.930508746589633e-05,
      "loss": 2.5649,
      "theoretical_loss": 3.429740842616153,
      "tokens_seen": 2004549632
    },
    {
      "epoch": 0.21,
      "learning_rate": 7.929706307173808e-05,
      "loss": 2.5895,
      "theoretical_loss": 3.4297221533191737,
      "tokens_seen": 2004680704
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.928903867757984e-05,
      "loss": 2.699,
      "theoretical_loss": 3.4297034655862406,
      "tokens_seen": 2004811776
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.92810142834216e-05,
      "loss": 2.6369,
      "theoretical_loss": 3.4296847794171197,
      "tokens_seen": 2004942848
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.927298988926336e-05,
      "loss": 2.5172,
      "theoretical_loss": 3.4296660948115782,
      "tokens_seen": 2005073920
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.926496549510512e-05,
      "loss": 2.4895,
      "theoretical_loss": 3.429647411769384,
      "tokens_seen": 2005204992
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.925694110094687e-05,
      "loss": 2.5339,
      "theoretical_loss": 3.429628730290303,
      "tokens_seen": 2005336064
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.924891670678864e-05,
      "loss": 2.7384,
      "theoretical_loss": 3.429610050374103,
      "tokens_seen": 2005467136
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1104357,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6726927757263184,
      "objective/train/theoretical_loss": 3.4296007110020104,
      "objective/train/tokens_used": 375991776,
      "theoretical_loss": 3.4296007110020104,
      "tokens_seen": 2005532672
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.92408923126304e-05,
      "loss": 2.5657,
      "theoretical_loss": 3.4295913720205506,
      "tokens_seen": 2005598208
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.923286791847216e-05,
      "loss": 2.6706,
      "theoretical_loss": 3.429572695229413,
      "tokens_seen": 2005729280
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.922484352431391e-05,
      "loss": 2.65,
      "theoretical_loss": 3.4295540200004577,
      "tokens_seen": 2005860352
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.921681913015567e-05,
      "loss": 2.5002,
      "theoretical_loss": 3.429535346333452,
      "tokens_seen": 2005991424
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.920879473599743e-05,
      "loss": 2.581,
      "theoretical_loss": 3.4295166742281626,
      "tokens_seen": 2006122496
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.920077034183919e-05,
      "loss": 2.6358,
      "theoretical_loss": 3.4294980036843583,
      "tokens_seen": 2006253568
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.919274594768095e-05,
      "loss": 2.5882,
      "theoretical_loss": 3.4294793347018047,
      "tokens_seen": 2006384640
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.918472155352271e-05,
      "loss": 2.6364,
      "theoretical_loss": 3.4294606672802703,
      "tokens_seen": 2006515712
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.917669715936447e-05,
      "loss": 2.4089,
      "theoretical_loss": 3.4294420014195226,
      "tokens_seen": 2006646784
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.916867276520623e-05,
      "loss": 2.6042,
      "theoretical_loss": 3.429423337119329,
      "tokens_seen": 2006777856
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.916064837104798e-05,
      "loss": 2.5733,
      "theoretical_loss": 3.429404674379457,
      "tokens_seen": 2006908928
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.915262397688975e-05,
      "loss": 2.5246,
      "theoretical_loss": 3.4293860131996743,
      "tokens_seen": 2007040000
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1104987,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.601886749267578,
      "objective/train/theoretical_loss": 3.4293673535797486,
      "objective/train/tokens_used": 377630176,
      "theoretical_loss": 3.4293673535797486,
      "tokens_seen": 2007171072
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.91445995827315e-05,
      "loss": 2.6193,
      "theoretical_loss": 3.4293673535797486,
      "tokens_seen": 2007171072
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.913657518857327e-05,
      "loss": 2.4668,
      "theoretical_loss": 3.4293486955194483,
      "tokens_seen": 2007302144
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.912855079441502e-05,
      "loss": 2.5915,
      "theoretical_loss": 3.42933003901854,
      "tokens_seen": 2007433216
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.912052640025679e-05,
      "loss": 2.5267,
      "theoretical_loss": 3.429311384076793,
      "tokens_seen": 2007564288
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.911250200609854e-05,
      "loss": 2.5647,
      "theoretical_loss": 3.429292730693974,
      "tokens_seen": 2007695360
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.910447761194029e-05,
      "loss": 2.6728,
      "theoretical_loss": 3.4292740788698515,
      "tokens_seen": 2007826432
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.909645321778206e-05,
      "loss": 2.4985,
      "theoretical_loss": 3.4292554286041934,
      "tokens_seen": 2007957504
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.908842882362381e-05,
      "loss": 2.5317,
      "theoretical_loss": 3.4292367798967676,
      "tokens_seen": 2008088576
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.908040442946558e-05,
      "loss": 2.5953,
      "theoretical_loss": 3.4292181327473426,
      "tokens_seen": 2008219648
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.907238003530733e-05,
      "loss": 2.493,
      "theoretical_loss": 3.429199487155686,
      "tokens_seen": 2008350720
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.90643556411491e-05,
      "loss": 2.5788,
      "theoretical_loss": 3.429180843121567,
      "tokens_seen": 2008481792
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.905633124699085e-05,
      "loss": 2.4941,
      "theoretical_loss": 3.429162200644753,
      "tokens_seen": 2008612864
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.90483068528326e-05,
      "loss": 2.5688,
      "theoretical_loss": 3.429143559725013,
      "tokens_seen": 2008743936
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1106102,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6869187355041504,
      "objective/train/theoretical_loss": 3.429134239848973,
      "objective/train/tokens_used": 379268576,
      "theoretical_loss": 3.429134239848973,
      "tokens_seen": 2008809472
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.904028245867437e-05,
      "loss": 2.5055,
      "theoretical_loss": 3.4291249203621144,
      "tokens_seen": 2008875008
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.903225806451613e-05,
      "loss": 2.5025,
      "theoretical_loss": 3.429106282555826,
      "tokens_seen": 2009006080
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.902423367035789e-05,
      "loss": 2.3999,
      "theoretical_loss": 3.429087646305917,
      "tokens_seen": 2009137152
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.901620927619964e-05,
      "loss": 2.5506,
      "theoretical_loss": 3.429069011612155,
      "tokens_seen": 2009268224
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.900818488204141e-05,
      "loss": 2.51,
      "theoretical_loss": 3.4290503784743085,
      "tokens_seen": 2009399296
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.900016048788316e-05,
      "loss": 2.5117,
      "theoretical_loss": 3.429031746892147,
      "tokens_seen": 2009530368
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.899213609372492e-05,
      "loss": 2.6431,
      "theoretical_loss": 3.4290131168654385,
      "tokens_seen": 2009661440
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.898411169956668e-05,
      "loss": 2.5644,
      "theoretical_loss": 3.4289944883939514,
      "tokens_seen": 2009792512
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.897608730540844e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.428975861477455,
      "tokens_seen": 2009923584
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.89680629112502e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.4289572361157186,
      "tokens_seen": 2010054656
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.896003851709196e-05,
      "loss": 2.5642,
      "theoretical_loss": 3.42893861230851,
      "tokens_seen": 2010185728
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.895201412293372e-05,
      "loss": 2.6082,
      "theoretical_loss": 3.428919990055599,
      "tokens_seen": 2010316800
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1107415,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5254108905792236,
      "objective/train/theoretical_loss": 3.428901369356754,
      "objective/train/tokens_used": 380906976,
      "theoretical_loss": 3.428901369356754,
      "tokens_seen": 2010447872
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.894398972877548e-05,
      "loss": 2.5814,
      "theoretical_loss": 3.428901369356754,
      "tokens_seen": 2010447872
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.893596533461724e-05,
      "loss": 2.6239,
      "theoretical_loss": 3.4288827502117436,
      "tokens_seen": 2010578944
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.8927940940459e-05,
      "loss": 2.671,
      "theoretical_loss": 3.4288641326203377,
      "tokens_seen": 2010710016
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.891991654630075e-05,
      "loss": 2.6009,
      "theoretical_loss": 3.428845516582305,
      "tokens_seen": 2010841088
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.891189215214252e-05,
      "loss": 2.6173,
      "theoretical_loss": 3.428826902097415,
      "tokens_seen": 2010972160
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.890386775798427e-05,
      "loss": 2.5199,
      "theoretical_loss": 3.4288082891654366,
      "tokens_seen": 2011103232
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.889584336382604e-05,
      "loss": 2.5374,
      "theoretical_loss": 3.4287896777861384,
      "tokens_seen": 2011234304
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.888781896966779e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.4287710679592913,
      "tokens_seen": 2011365376
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.887979457550956e-05,
      "loss": 2.4322,
      "theoretical_loss": 3.4287524596846635,
      "tokens_seen": 2011496448
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.887177018135131e-05,
      "loss": 2.4975,
      "theoretical_loss": 3.4287338529620244,
      "tokens_seen": 2011627520
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.886374578719306e-05,
      "loss": 2.6787,
      "theoretical_loss": 3.4287152477911436,
      "tokens_seen": 2011758592
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.885572139303483e-05,
      "loss": 2.4772,
      "theoretical_loss": 3.428696644171791,
      "tokens_seen": 2011889664
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.884769699887658e-05,
      "loss": 2.6306,
      "theoretical_loss": 3.428678042103736,
      "tokens_seen": 2012020736
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1108037,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8532683849334717,
      "objective/train/theoretical_loss": 3.428668741651373,
      "objective/train/tokens_used": 382545376,
      "theoretical_loss": 3.428668741651373,
      "tokens_seen": 2012086272
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.883967260471835e-05,
      "loss": 2.5644,
      "theoretical_loss": 3.4286594415867477,
      "tokens_seen": 2012151808
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.88316482105601e-05,
      "loss": 2.5605,
      "theoretical_loss": 3.428640842620596,
      "tokens_seen": 2012282880
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.882362381640187e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.4286222452050508,
      "tokens_seen": 2012413952
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.881559942224362e-05,
      "loss": 2.5207,
      "theoretical_loss": 3.428603649339882,
      "tokens_seen": 2012545024
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.880757502808538e-05,
      "loss": 2.5635,
      "theoretical_loss": 3.4285850550248584,
      "tokens_seen": 2012676096
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.879955063392714e-05,
      "loss": 2.7104,
      "theoretical_loss": 3.4285664622597514,
      "tokens_seen": 2012807168
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.87915262397689e-05,
      "loss": 2.5008,
      "theoretical_loss": 3.4285478710443296,
      "tokens_seen": 2012938240
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.878350184561066e-05,
      "loss": 2.5471,
      "theoretical_loss": 3.4285292813783634,
      "tokens_seen": 2013069312
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.877547745145242e-05,
      "loss": 2.4691,
      "theoretical_loss": 3.4285106932616225,
      "tokens_seen": 2013200384
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.876745305729418e-05,
      "loss": 2.6136,
      "theoretical_loss": 3.428492106693878,
      "tokens_seen": 2013331456
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.875942866313593e-05,
      "loss": 2.4363,
      "theoretical_loss": 3.4284735216748983,
      "tokens_seen": 2013462528
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.875140426897769e-05,
      "loss": 2.4772,
      "theoretical_loss": 3.428454938204455,
      "tokens_seen": 2013593600
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1109293,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.913437604904175,
      "objective/train/theoretical_loss": 3.4284363562823175,
      "objective/train/tokens_used": 384183776,
      "theoretical_loss": 3.4284363562823175,
      "tokens_seen": 2013724672
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.874337987481945e-05,
      "loss": 2.6037,
      "theoretical_loss": 3.4284363562823175,
      "tokens_seen": 2013724672
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.873535548066121e-05,
      "loss": 2.6943,
      "theoretical_loss": 3.4284177759082564,
      "tokens_seen": 2013855744
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.872733108650297e-05,
      "loss": 2.5348,
      "theoretical_loss": 3.428399197082042,
      "tokens_seen": 2013986816
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.871930669234473e-05,
      "loss": 2.5262,
      "theoretical_loss": 3.428380619803444,
      "tokens_seen": 2014117888
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.87112822981865e-05,
      "loss": 2.6206,
      "theoretical_loss": 3.4283620440722333,
      "tokens_seen": 2014248960
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.870325790402825e-05,
      "loss": 2.5586,
      "theoretical_loss": 3.4283434698881807,
      "tokens_seen": 2014380032
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.869523350987e-05,
      "loss": 2.6104,
      "theoretical_loss": 3.4283248972510556,
      "tokens_seen": 2014511104
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.868720911571177e-05,
      "loss": 2.5206,
      "theoretical_loss": 3.42830632616063,
      "tokens_seen": 2014642176
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.867918472155352e-05,
      "loss": 2.5558,
      "theoretical_loss": 3.4282877566166734,
      "tokens_seen": 2014773248
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.867116032739529e-05,
      "loss": 2.5839,
      "theoretical_loss": 3.4282691886189562,
      "tokens_seen": 2014904320
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.866313593323704e-05,
      "loss": 2.5297,
      "theoretical_loss": 3.4282506221672504,
      "tokens_seen": 2015035392
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.865511153907881e-05,
      "loss": 2.3923,
      "theoretical_loss": 3.4282320572613254,
      "tokens_seen": 2015166464
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.864708714492056e-05,
      "loss": 2.5233,
      "theoretical_loss": 3.4282134939009525,
      "tokens_seen": 2015297536
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1109931,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3523659706115723,
      "objective/train/theoretical_loss": 3.4282042128002765,
      "objective/train/tokens_used": 385822176,
      "theoretical_loss": 3.4282042128002765,
      "tokens_seen": 2015363072
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.863906275076233e-05,
      "loss": 2.4561,
      "theoretical_loss": 3.4281949320859026,
      "tokens_seen": 2015428608
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.863103835660408e-05,
      "loss": 2.6495,
      "theoretical_loss": 3.4281763718159466,
      "tokens_seen": 2015559680
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.862301396244583e-05,
      "loss": 2.6334,
      "theoretical_loss": 3.4281578130908548,
      "tokens_seen": 2015690752
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.86149895682876e-05,
      "loss": 2.5526,
      "theoretical_loss": 3.4281392559103994,
      "tokens_seen": 2015821824
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.860696517412935e-05,
      "loss": 2.5854,
      "theoretical_loss": 3.4281207002743503,
      "tokens_seen": 2015952896
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.859894077997112e-05,
      "loss": 2.5802,
      "theoretical_loss": 3.428102146182479,
      "tokens_seen": 2016083968
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.859091638581287e-05,
      "loss": 2.4863,
      "theoretical_loss": 3.4280835936345566,
      "tokens_seen": 2016215040
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.858289199165464e-05,
      "loss": 2.5008,
      "theoretical_loss": 3.4280650426303536,
      "tokens_seen": 2016346112
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.857486759749639e-05,
      "loss": 2.5014,
      "theoretical_loss": 3.428046493169643,
      "tokens_seen": 2016477184
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.856684320333815e-05,
      "loss": 2.6355,
      "theoretical_loss": 3.428027945252194,
      "tokens_seen": 2016608256
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.855881880917991e-05,
      "loss": 2.4802,
      "theoretical_loss": 3.428009398877779,
      "tokens_seen": 2016739328
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.855079441502167e-05,
      "loss": 2.4685,
      "theoretical_loss": 3.427990854046169,
      "tokens_seen": 2016870400
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1110878,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.68257212638855,
      "objective/train/theoretical_loss": 3.427972310757136,
      "objective/train/tokens_used": 387460576,
      "theoretical_loss": 3.427972310757136,
      "tokens_seen": 2017001472
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.854277002086343e-05,
      "loss": 2.5736,
      "theoretical_loss": 3.427972310757136,
      "tokens_seen": 2017001472
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.853474562670519e-05,
      "loss": 2.4837,
      "theoretical_loss": 3.4279537690104513,
      "tokens_seen": 2017132544
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.852672123254695e-05,
      "loss": 2.5344,
      "theoretical_loss": 3.427935228805886,
      "tokens_seen": 2017263616
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.85186968383887e-05,
      "loss": 2.5714,
      "theoretical_loss": 3.4279166901432117,
      "tokens_seen": 2017394688
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.851067244423046e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.4278981530221997,
      "tokens_seen": 2017525760
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.850264805007222e-05,
      "loss": 2.4,
      "theoretical_loss": 3.4278796174426223,
      "tokens_seen": 2017656832
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.849462365591398e-05,
      "loss": 2.5503,
      "theoretical_loss": 3.427861083404251,
      "tokens_seen": 2017787904
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.848659926175574e-05,
      "loss": 2.5524,
      "theoretical_loss": 3.4278425509068575,
      "tokens_seen": 2017918976
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.84785748675975e-05,
      "loss": 2.5015,
      "theoretical_loss": 3.4278240199502137,
      "tokens_seen": 2018050048
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.847055047343926e-05,
      "loss": 2.567,
      "theoretical_loss": 3.4278054905340913,
      "tokens_seen": 2018181120
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.846252607928102e-05,
      "loss": 2.4222,
      "theoretical_loss": 3.427786962658262,
      "tokens_seen": 2018312192
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.845450168512277e-05,
      "loss": 2.6226,
      "theoretical_loss": 3.4277684363224985,
      "tokens_seen": 2018443264
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.844647729096454e-05,
      "loss": 2.3817,
      "theoretical_loss": 3.427749911526572,
      "tokens_seen": 2018574336
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1111506,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.24393630027771,
      "objective/train/theoretical_loss": 3.427740649705976,
      "objective/train/tokens_used": 389098976,
      "theoretical_loss": 3.427740649705976,
      "tokens_seen": 2018639872
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.843845289680629e-05,
      "loss": 2.5344,
      "theoretical_loss": 3.4277313882702547,
      "tokens_seen": 2018705408
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.843042850264806e-05,
      "loss": 2.5531,
      "theoretical_loss": 3.4277128665533194,
      "tokens_seen": 2018836480
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.842240410848981e-05,
      "loss": 2.4688,
      "theoretical_loss": 3.427694346375537,
      "tokens_seen": 2018967552
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.841437971433158e-05,
      "loss": 2.4856,
      "theoretical_loss": 3.4276758277366803,
      "tokens_seen": 2019098624
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.840635532017333e-05,
      "loss": 2.441,
      "theoretical_loss": 3.427657310636522,
      "tokens_seen": 2019229696
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.839833092601508e-05,
      "loss": 2.566,
      "theoretical_loss": 3.4276387950748335,
      "tokens_seen": 2019360768
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.839030653185685e-05,
      "loss": 2.4114,
      "theoretical_loss": 3.427620281051388,
      "tokens_seen": 2019491840
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.83822821376986e-05,
      "loss": 2.4363,
      "theoretical_loss": 3.4276017685659577,
      "tokens_seen": 2019622912
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.837425774354037e-05,
      "loss": 2.4333,
      "theoretical_loss": 3.4275832576183145,
      "tokens_seen": 2019753984
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.836623334938212e-05,
      "loss": 2.5378,
      "theoretical_loss": 3.4275647482082308,
      "tokens_seen": 2019885056
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.835820895522389e-05,
      "loss": 2.5506,
      "theoretical_loss": 3.42754624033548,
      "tokens_seen": 2020016128
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.835018456106564e-05,
      "loss": 2.5029,
      "theoretical_loss": 3.427527733999834,
      "tokens_seen": 2020147200
    },
    {
      "epoch": 0.22,
      "objective/train/docs_used": 1112541,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.220839262008667,
      "objective/train/theoretical_loss": 3.427509229201066,
      "objective/train/tokens_used": 390737376,
      "theoretical_loss": 3.427509229201066,
      "tokens_seen": 2020278272
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.83421601669074e-05,
      "loss": 2.5173,
      "theoretical_loss": 3.427509229201066,
      "tokens_seen": 2020278272
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.833413577274916e-05,
      "loss": 2.5369,
      "theoretical_loss": 3.4274907259389478,
      "tokens_seen": 2020409344
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.832611137859092e-05,
      "loss": 2.5579,
      "theoretical_loss": 3.4274722242132523,
      "tokens_seen": 2020540416
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.831808698443268e-05,
      "loss": 2.5063,
      "theoretical_loss": 3.427453724023753,
      "tokens_seen": 2020671488
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.831006259027444e-05,
      "loss": 2.6443,
      "theoretical_loss": 3.427435225370222,
      "tokens_seen": 2020802560
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.83020381961162e-05,
      "loss": 2.6026,
      "theoretical_loss": 3.427416728252433,
      "tokens_seen": 2020933632
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.829401380195796e-05,
      "loss": 2.4388,
      "theoretical_loss": 3.427398232670158,
      "tokens_seen": 2021064704
    },
    {
      "epoch": 0.22,
      "learning_rate": 7.828598940779972e-05,
      "loss": 2.6021,
      "theoretical_loss": 3.4273797386231704,
      "tokens_seen": 2021195776
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.827796501364148e-05,
      "loss": 2.5579,
      "theoretical_loss": 3.427361246111243,
      "tokens_seen": 2021326848
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.826994061948323e-05,
      "loss": 2.4902,
      "theoretical_loss": 3.4273427551341493,
      "tokens_seen": 2021457920
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.8261916225325e-05,
      "loss": 2.552,
      "theoretical_loss": 3.4273242656916616,
      "tokens_seen": 2021588992
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.825389183116675e-05,
      "loss": 2.5983,
      "theoretical_loss": 3.4273057777835536,
      "tokens_seen": 2021720064
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.824586743700851e-05,
      "loss": 2.4921,
      "theoretical_loss": 3.427287291409599,
      "tokens_seen": 2021851136
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1113127,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.679295539855957,
      "objective/train/theoretical_loss": 3.4272780487978576,
      "objective/train/tokens_used": 392375776,
      "theoretical_loss": 3.4272780487978576,
      "tokens_seen": 2021916672
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.823784304285027e-05,
      "loss": 2.5579,
      "theoretical_loss": 3.42726880656957,
      "tokens_seen": 2021982208
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.822981864869203e-05,
      "loss": 2.3652,
      "theoretical_loss": 3.42725032326324,
      "tokens_seen": 2022113280
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.822179425453379e-05,
      "loss": 2.7038,
      "theoretical_loss": 3.427231841490384,
      "tokens_seen": 2022244352
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.821376986037554e-05,
      "loss": 2.4067,
      "theoretical_loss": 3.4272133612507734,
      "tokens_seen": 2022375424
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.820574546621731e-05,
      "loss": 2.5086,
      "theoretical_loss": 3.427194882544182,
      "tokens_seen": 2022506496
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.819772107205906e-05,
      "loss": 2.5402,
      "theoretical_loss": 3.4271764053703837,
      "tokens_seen": 2022637568
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.818969667790083e-05,
      "loss": 2.4729,
      "theoretical_loss": 3.4271579297291526,
      "tokens_seen": 2022768640
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.818167228374258e-05,
      "loss": 2.4496,
      "theoretical_loss": 3.4271394556202615,
      "tokens_seen": 2022899712
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.817364788958435e-05,
      "loss": 2.3272,
      "theoretical_loss": 3.4271209830434843,
      "tokens_seen": 2023030784
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.81656234954261e-05,
      "loss": 2.6453,
      "theoretical_loss": 3.4271025119985943,
      "tokens_seen": 2023161856
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.815759910126785e-05,
      "loss": 2.5107,
      "theoretical_loss": 3.4270840424853657,
      "tokens_seen": 2023292928
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.814957470710962e-05,
      "loss": 2.5706,
      "theoretical_loss": 3.427065574503572,
      "tokens_seen": 2023424000
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1114473,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.07741641998291,
      "objective/train/theoretical_loss": 3.427047108052988,
      "objective/train/tokens_used": 394014176,
      "theoretical_loss": 3.427047108052988,
      "tokens_seen": 2023555072
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.814155031295137e-05,
      "loss": 2.5202,
      "theoretical_loss": 3.427047108052988,
      "tokens_seen": 2023555072
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.813352591879314e-05,
      "loss": 2.4878,
      "theoretical_loss": 3.4270286431333856,
      "tokens_seen": 2023686144
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.812550152463489e-05,
      "loss": 2.4707,
      "theoretical_loss": 3.4270101797445403,
      "tokens_seen": 2023817216
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.811747713047666e-05,
      "loss": 2.4816,
      "theoretical_loss": 3.4269917178862253,
      "tokens_seen": 2023948288
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.810945273631841e-05,
      "loss": 2.5005,
      "theoretical_loss": 3.426973257558216,
      "tokens_seen": 2024079360
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.810142834216017e-05,
      "loss": 2.58,
      "theoretical_loss": 3.4269547987602844,
      "tokens_seen": 2024210432
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.809340394800193e-05,
      "loss": 2.5284,
      "theoretical_loss": 3.4269363414922056,
      "tokens_seen": 2024341504
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.808537955384369e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.4269178857537543,
      "tokens_seen": 2024472576
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.807735515968545e-05,
      "loss": 2.5637,
      "theoretical_loss": 3.426899431544704,
      "tokens_seen": 2024603648
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.80693307655272e-05,
      "loss": 2.5071,
      "theoretical_loss": 3.4268809788648293,
      "tokens_seen": 2024734720
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.806130637136897e-05,
      "loss": 2.4704,
      "theoretical_loss": 3.4268625277139044,
      "tokens_seen": 2024865792
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.805328197721073e-05,
      "loss": 2.5377,
      "theoretical_loss": 3.4268440780917038,
      "tokens_seen": 2024996864
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.804525758305248e-05,
      "loss": 2.5348,
      "theoretical_loss": 3.4268256299980013,
      "tokens_seen": 2025127936
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1115722,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8050100803375244,
      "objective/train/theoretical_loss": 3.4268164065242663,
      "objective/train/tokens_used": 395652576,
      "theoretical_loss": 3.4268164065242663,
      "tokens_seen": 2025193472
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.803723318889425e-05,
      "loss": 2.5304,
      "theoretical_loss": 3.4268071834325715,
      "tokens_seen": 2025259008
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.8029208794736e-05,
      "loss": 2.4983,
      "theoretical_loss": 3.42678873839519,
      "tokens_seen": 2025390080
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.802118440057776e-05,
      "loss": 2.4642,
      "theoretical_loss": 3.42677029488563,
      "tokens_seen": 2025521152
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.801316000641952e-05,
      "loss": 2.5015,
      "theoretical_loss": 3.426751852903667,
      "tokens_seen": 2025652224
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.800513561226128e-05,
      "loss": 2.3507,
      "theoretical_loss": 3.4267334124490754,
      "tokens_seen": 2025783296
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.799711121810304e-05,
      "loss": 2.6905,
      "theoretical_loss": 3.426714973521629,
      "tokens_seen": 2025914368
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.79890868239448e-05,
      "loss": 2.538,
      "theoretical_loss": 3.426696536121104,
      "tokens_seen": 2026045440
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.798106242978656e-05,
      "loss": 2.598,
      "theoretical_loss": 3.4266781002472744,
      "tokens_seen": 2026176512
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.797303803562831e-05,
      "loss": 2.4248,
      "theoretical_loss": 3.4266596658999147,
      "tokens_seen": 2026307584
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.796501364147008e-05,
      "loss": 2.5141,
      "theoretical_loss": 3.4266412330788008,
      "tokens_seen": 2026438656
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.795698924731183e-05,
      "loss": 2.3992,
      "theoretical_loss": 3.426622801783707,
      "tokens_seen": 2026569728
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.79489648531536e-05,
      "loss": 2.6258,
      "theoretical_loss": 3.426604372014408,
      "tokens_seen": 2026700800
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1116448,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.421637535095215,
      "objective/train/theoretical_loss": 3.426585943770679,
      "objective/train/tokens_used": 397290976,
      "theoretical_loss": 3.426585943770679,
      "tokens_seen": 2026831872
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.794094045899535e-05,
      "loss": 2.5952,
      "theoretical_loss": 3.426585943770679,
      "tokens_seen": 2026831872
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.793291606483712e-05,
      "loss": 2.3854,
      "theoretical_loss": 3.4265675170522956,
      "tokens_seen": 2026962944
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.792489167067887e-05,
      "loss": 2.4204,
      "theoretical_loss": 3.426549091859032,
      "tokens_seen": 2027094016
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.791686727652062e-05,
      "loss": 2.4967,
      "theoretical_loss": 3.4265306681906647,
      "tokens_seen": 2027225088
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.790884288236239e-05,
      "loss": 2.5815,
      "theoretical_loss": 3.4265122460469675,
      "tokens_seen": 2027356160
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.790081848820414e-05,
      "loss": 2.5191,
      "theoretical_loss": 3.4264938254277166,
      "tokens_seen": 2027487232
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.789279409404591e-05,
      "loss": 2.5302,
      "theoretical_loss": 3.4264754063326865,
      "tokens_seen": 2027618304
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.788476969988766e-05,
      "loss": 2.4264,
      "theoretical_loss": 3.426456988761654,
      "tokens_seen": 2027749376
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.787674530572943e-05,
      "loss": 2.5857,
      "theoretical_loss": 3.426438572714393,
      "tokens_seen": 2027880448
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.786872091157118e-05,
      "loss": 2.2381,
      "theoretical_loss": 3.4264201581906795,
      "tokens_seen": 2028011520
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.786069651741294e-05,
      "loss": 2.4836,
      "theoretical_loss": 3.4264017451902893,
      "tokens_seen": 2028142592
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.78526721232547e-05,
      "loss": 2.5449,
      "theoretical_loss": 3.426383333712997,
      "tokens_seen": 2028273664
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.784464772909646e-05,
      "loss": 2.5556,
      "theoretical_loss": 3.4263649237585794,
      "tokens_seen": 2028404736
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1117705,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.295840263366699,
      "objective/train/theoretical_loss": 3.426355719352378,
      "objective/train/tokens_used": 398929376,
      "theoretical_loss": 3.426355719352378,
      "tokens_seen": 2028470272
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.783662333493822e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.4263465153268116,
      "tokens_seen": 2028535808
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.782859894077998e-05,
      "loss": 2.512,
      "theoretical_loss": 3.4263281084174695,
      "tokens_seen": 2028666880
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.782057454662174e-05,
      "loss": 2.5264,
      "theoretical_loss": 3.426309703030328,
      "tokens_seen": 2028797952
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.78125501524635e-05,
      "loss": 2.483,
      "theoretical_loss": 3.4262912991651637,
      "tokens_seen": 2028929024
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.780452575830525e-05,
      "loss": 2.5553,
      "theoretical_loss": 3.426272896821753,
      "tokens_seen": 2029060096
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.779650136414702e-05,
      "loss": 2.4207,
      "theoretical_loss": 3.42625449599987,
      "tokens_seen": 2029191168
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.778847696998877e-05,
      "loss": 2.5201,
      "theoretical_loss": 3.4262360966992924,
      "tokens_seen": 2029322240
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.778045257583053e-05,
      "loss": 2.5239,
      "theoretical_loss": 3.426217698919795,
      "tokens_seen": 2029453312
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.777242818167229e-05,
      "loss": 2.6701,
      "theoretical_loss": 3.426199302661155,
      "tokens_seen": 2029584384
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.776440378751405e-05,
      "loss": 2.6428,
      "theoretical_loss": 3.426180907923147,
      "tokens_seen": 2029715456
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.775637939335581e-05,
      "loss": 2.6025,
      "theoretical_loss": 3.426162514705548,
      "tokens_seen": 2029846528
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.774835499919756e-05,
      "loss": 2.5519,
      "theoretical_loss": 3.426144123008134,
      "tokens_seen": 2029977600
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1118254,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5424864292144775,
      "objective/train/theoretical_loss": 3.426125732830682,
      "objective/train/tokens_used": 400567776,
      "theoretical_loss": 3.426125732830682,
      "tokens_seen": 2030108672
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.774033060503933e-05,
      "loss": 2.4409,
      "theoretical_loss": 3.426125732830682,
      "tokens_seen": 2030108672
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.773230621088108e-05,
      "loss": 2.5376,
      "theoretical_loss": 3.4261073441729666,
      "tokens_seen": 2030239744
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.772428181672285e-05,
      "loss": 2.584,
      "theoretical_loss": 3.4260889570347652,
      "tokens_seen": 2030370816
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.77162574225646e-05,
      "loss": 2.712,
      "theoretical_loss": 3.4260705714158544,
      "tokens_seen": 2030501888
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.770823302840637e-05,
      "loss": 2.4428,
      "theoretical_loss": 3.42605218731601,
      "tokens_seen": 2030632960
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.770020863424812e-05,
      "loss": 2.7793,
      "theoretical_loss": 3.4260338047350087,
      "tokens_seen": 2030764032
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.769218424008987e-05,
      "loss": 2.4869,
      "theoretical_loss": 3.426015423672627,
      "tokens_seen": 2030895104
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.768415984593164e-05,
      "loss": 2.64,
      "theoretical_loss": 3.425997044128641,
      "tokens_seen": 2031026176
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.76761354517734e-05,
      "loss": 2.5178,
      "theoretical_loss": 3.4259786661028278,
      "tokens_seen": 2031157248
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.766811105761516e-05,
      "loss": 2.6034,
      "theoretical_loss": 3.425960289594964,
      "tokens_seen": 2031288320
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.766008666345691e-05,
      "loss": 2.4537,
      "theoretical_loss": 3.4259419146048264,
      "tokens_seen": 2031419392
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.765206226929868e-05,
      "loss": 2.5037,
      "theoretical_loss": 3.425923541132191,
      "tokens_seen": 2031550464
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.764403787514043e-05,
      "loss": 2.4198,
      "theoretical_loss": 3.4259051691768354,
      "tokens_seen": 2031681536
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1119492,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9781367778778076,
      "objective/train/theoretical_loss": 3.425895983768068,
      "objective/train/tokens_used": 402206176,
      "theoretical_loss": 3.425895983768068,
      "tokens_seen": 2031747072
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.76360134809822e-05,
      "loss": 2.6343,
      "theoretical_loss": 3.425886798738537,
      "tokens_seen": 2031812608
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.762798908682395e-05,
      "loss": 2.4851,
      "theoretical_loss": 3.4258684298170703,
      "tokens_seen": 2031943680
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.76199646926657e-05,
      "loss": 2.4681,
      "theoretical_loss": 3.4258500624122146,
      "tokens_seen": 2032074752
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.761194029850747e-05,
      "loss": 2.6024,
      "theoretical_loss": 3.425831696523746,
      "tokens_seen": 2032205824
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.760391590434923e-05,
      "loss": 2.4884,
      "theoretical_loss": 3.4258133321514412,
      "tokens_seen": 2032336896
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.759589151019099e-05,
      "loss": 2.526,
      "theoretical_loss": 3.425794969295078,
      "tokens_seen": 2032467968
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.758786711603275e-05,
      "loss": 2.591,
      "theoretical_loss": 3.4257766079544334,
      "tokens_seen": 2032599040
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.757984272187451e-05,
      "loss": 2.6383,
      "theoretical_loss": 3.4257582481292834,
      "tokens_seen": 2032730112
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.757181832771627e-05,
      "loss": 2.5295,
      "theoretical_loss": 3.4257398898194067,
      "tokens_seen": 2032861184
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.756379393355802e-05,
      "loss": 2.5044,
      "theoretical_loss": 3.4257215330245794,
      "tokens_seen": 2032992256
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.755576953939979e-05,
      "loss": 2.4814,
      "theoretical_loss": 3.4257031777445803,
      "tokens_seen": 2033123328
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.754774514524154e-05,
      "loss": 2.6389,
      "theoretical_loss": 3.425684823979185,
      "tokens_seen": 2033254400
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1120104,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5154623985290527,
      "objective/train/theoretical_loss": 3.4256664717281717,
      "objective/train/tokens_used": 403844576,
      "theoretical_loss": 3.4256664717281717,
      "tokens_seen": 2033385472
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.75397207510833e-05,
      "loss": 2.6288,
      "theoretical_loss": 3.4256664717281717,
      "tokens_seen": 2033385472
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.753169635692506e-05,
      "loss": 2.3721,
      "theoretical_loss": 3.425648120991318,
      "tokens_seen": 2033516544
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.752367196276682e-05,
      "loss": 2.4796,
      "theoretical_loss": 3.4256297717684014,
      "tokens_seen": 2033647616
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.751564756860858e-05,
      "loss": 2.4372,
      "theoretical_loss": 3.4256114240591993,
      "tokens_seen": 2033778688
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.750762317445033e-05,
      "loss": 2.5127,
      "theoretical_loss": 3.425593077863489,
      "tokens_seen": 2033909760
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.74995987802921e-05,
      "loss": 2.485,
      "theoretical_loss": 3.4255747331810484,
      "tokens_seen": 2034040832
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.749157438613385e-05,
      "loss": 2.4735,
      "theoretical_loss": 3.4255563900116552,
      "tokens_seen": 2034171904
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.748354999197562e-05,
      "loss": 2.678,
      "theoretical_loss": 3.4255380483550866,
      "tokens_seen": 2034302976
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.747552559781737e-05,
      "loss": 2.4402,
      "theoretical_loss": 3.4255197082111213,
      "tokens_seen": 2034434048
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.746750120365914e-05,
      "loss": 2.557,
      "theoretical_loss": 3.4255013695795364,
      "tokens_seen": 2034565120
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.745947680950089e-05,
      "loss": 2.5052,
      "theoretical_loss": 3.4254830324601104,
      "tokens_seen": 2034696192
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.745145241534264e-05,
      "loss": 2.4159,
      "theoretical_loss": 3.425464696852621,
      "tokens_seen": 2034827264
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.744342802118441e-05,
      "loss": 2.4283,
      "theoretical_loss": 3.425446362756845,
      "tokens_seen": 2034958336
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1121212,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1749119758605957,
      "objective/train/theoretical_loss": 3.425437196275781,
      "objective/train/tokens_used": 405482976,
      "theoretical_loss": 3.425437196275781,
      "tokens_seen": 2035023872
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.743540362702616e-05,
      "loss": 2.4383,
      "theoretical_loss": 3.4254280301725624,
      "tokens_seen": 2035089408
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.742737923286793e-05,
      "loss": 2.679,
      "theoretical_loss": 3.42540969909955,
      "tokens_seen": 2035220480
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.741935483870968e-05,
      "loss": 2.5305,
      "theoretical_loss": 3.4253913695375857,
      "tokens_seen": 2035351552
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.741133044455145e-05,
      "loss": 2.6681,
      "theoretical_loss": 3.4253730414864485,
      "tokens_seen": 2035482624
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.74033060503932e-05,
      "loss": 2.4899,
      "theoretical_loss": 3.4253547149459163,
      "tokens_seen": 2035613696
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.739528165623496e-05,
      "loss": 2.7024,
      "theoretical_loss": 3.425336389915767,
      "tokens_seen": 2035744768
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.738725726207672e-05,
      "loss": 2.5845,
      "theoretical_loss": 3.4253180663957794,
      "tokens_seen": 2035875840
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.737923286791848e-05,
      "loss": 2.458,
      "theoretical_loss": 3.425299744385731,
      "tokens_seen": 2036006912
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.737120847376024e-05,
      "loss": 2.6014,
      "theoretical_loss": 3.425281423885401,
      "tokens_seen": 2036137984
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.7363184079602e-05,
      "loss": 2.533,
      "theoretical_loss": 3.425263104894568,
      "tokens_seen": 2036269056
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.735515968544376e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.4252447874130096,
      "tokens_seen": 2036400128
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.734713529128552e-05,
      "loss": 2.6825,
      "theoretical_loss": 3.4252264714405047,
      "tokens_seen": 2036531200
    },
    {
      "epoch": 0.23,
      "objective/train/docs_used": 1122341,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8938703536987305,
      "objective/train/theoretical_loss": 3.425208156976832,
      "objective/train/tokens_used": 407121376,
      "theoretical_loss": 3.425208156976832,
      "tokens_seen": 2036662272
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.733911089712728e-05,
      "loss": 2.7145,
      "theoretical_loss": 3.425208156976832,
      "tokens_seen": 2036662272
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.733108650296904e-05,
      "loss": 2.5617,
      "theoretical_loss": 3.4251898440217703,
      "tokens_seen": 2036793344
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.732306210881079e-05,
      "loss": 2.5561,
      "theoretical_loss": 3.4251715325750975,
      "tokens_seen": 2036924416
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.731503771465256e-05,
      "loss": 2.4023,
      "theoretical_loss": 3.4251532226365935,
      "tokens_seen": 2037055488
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.730701332049431e-05,
      "loss": 2.6135,
      "theoretical_loss": 3.425134914206036,
      "tokens_seen": 2037186560
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.729898892633608e-05,
      "loss": 2.6687,
      "theoretical_loss": 3.425116607283204,
      "tokens_seen": 2037317632
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.729096453217783e-05,
      "loss": 2.655,
      "theoretical_loss": 3.4250983018678767,
      "tokens_seen": 2037448704
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.72829401380196e-05,
      "loss": 2.4404,
      "theoretical_loss": 3.4250799979598328,
      "tokens_seen": 2037579776
    },
    {
      "epoch": 0.23,
      "learning_rate": 7.727491574386135e-05,
      "loss": 2.5581,
      "theoretical_loss": 3.425061695558851,
      "tokens_seen": 2037710848
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.72668913497031e-05,
      "loss": 2.5581,
      "theoretical_loss": 3.4250433946647107,
      "tokens_seen": 2037841920
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.725886695554487e-05,
      "loss": 2.5608,
      "theoretical_loss": 3.4250250952771912,
      "tokens_seen": 2037972992
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.725084256138662e-05,
      "loss": 2.624,
      "theoretical_loss": 3.4250067973960707,
      "tokens_seen": 2038104064
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.724281816722839e-05,
      "loss": 2.5705,
      "theoretical_loss": 3.4249885010211285,
      "tokens_seen": 2038235136
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1123098,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.203213691711426,
      "objective/train/theoretical_loss": 3.4249793533984056,
      "objective/train/tokens_used": 408759776,
      "theoretical_loss": 3.4249793533984056,
      "tokens_seen": 2038300672
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.723479377307014e-05,
      "loss": 2.4835,
      "theoretical_loss": 3.4249702061521443,
      "tokens_seen": 2038366208
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.72267693789119e-05,
      "loss": 2.6758,
      "theoretical_loss": 3.4249519127888974,
      "tokens_seen": 2038497280
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.721874498475366e-05,
      "loss": 2.4282,
      "theoretical_loss": 3.4249336209311667,
      "tokens_seen": 2038628352
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.721072059059541e-05,
      "loss": 2.6284,
      "theoretical_loss": 3.4249153305787314,
      "tokens_seen": 2038759424
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.720269619643718e-05,
      "loss": 2.627,
      "theoretical_loss": 3.424897041731371,
      "tokens_seen": 2038890496
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.719467180227893e-05,
      "loss": 2.6302,
      "theoretical_loss": 3.4248787543888652,
      "tokens_seen": 2039021568
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.71866474081207e-05,
      "loss": 2.4353,
      "theoretical_loss": 3.424860468550993,
      "tokens_seen": 2039152640
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.717862301396245e-05,
      "loss": 2.3627,
      "theoretical_loss": 3.4248421842175336,
      "tokens_seen": 2039283712
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.71705986198042e-05,
      "loss": 2.5219,
      "theoretical_loss": 3.424823901388268,
      "tokens_seen": 2039414784
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.716257422564597e-05,
      "loss": 2.5145,
      "theoretical_loss": 3.4248056200629744,
      "tokens_seen": 2039545856
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.715454983148773e-05,
      "loss": 2.599,
      "theoretical_loss": 3.4247873402414326,
      "tokens_seen": 2039676928
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.714652543732949e-05,
      "loss": 2.5982,
      "theoretical_loss": 3.4247690619234223,
      "tokens_seen": 2039808000
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1124234,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5336902141571045,
      "objective/train/theoretical_loss": 3.424750785108724,
      "objective/train/tokens_used": 410398176,
      "theoretical_loss": 3.424750785108724,
      "tokens_seen": 2039939072
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.713850104317125e-05,
      "loss": 2.4938,
      "theoretical_loss": 3.424750785108724,
      "tokens_seen": 2039939072
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.7130476649013e-05,
      "loss": 2.2913,
      "theoretical_loss": 3.424732509797117,
      "tokens_seen": 2040070144
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.712245225485477e-05,
      "loss": 2.5377,
      "theoretical_loss": 3.4247142359883807,
      "tokens_seen": 2040201216
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.711442786069652e-05,
      "loss": 2.4873,
      "theoretical_loss": 3.4246959636822956,
      "tokens_seen": 2040332288
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.710640346653829e-05,
      "loss": 2.3766,
      "theoretical_loss": 3.4246776928786415,
      "tokens_seen": 2040463360
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.709837907238004e-05,
      "loss": 2.6614,
      "theoretical_loss": 3.4246594235771983,
      "tokens_seen": 2040594432
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.70903546782218e-05,
      "loss": 2.5043,
      "theoretical_loss": 3.4246411557777456,
      "tokens_seen": 2040725504
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.708233028406356e-05,
      "loss": 2.4483,
      "theoretical_loss": 3.424622889480064,
      "tokens_seen": 2040856576
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.707430588990531e-05,
      "loss": 2.4879,
      "theoretical_loss": 3.4246046246839335,
      "tokens_seen": 2040987648
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.706628149574708e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.4245863613891343,
      "tokens_seen": 2041118720
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.705825710158883e-05,
      "loss": 2.4924,
      "theoretical_loss": 3.424568099595446,
      "tokens_seen": 2041249792
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.70502327074306e-05,
      "loss": 2.6618,
      "theoretical_loss": 3.4245498393026494,
      "tokens_seen": 2041380864
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.704220831327235e-05,
      "loss": 2.5485,
      "theoretical_loss": 3.424531580510525,
      "tokens_seen": 2041511936
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1124915,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.3148484230041504,
      "objective/train/theoretical_loss": 3.424522451677146,
      "objective/train/tokens_used": 412036576,
      "theoretical_loss": 3.424522451677146,
      "tokens_seen": 2041577472
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.70341839191141e-05,
      "loss": 2.4715,
      "theoretical_loss": 3.4245133232188527,
      "tokens_seen": 2041643008
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.702615952495587e-05,
      "loss": 2.4311,
      "theoretical_loss": 3.424495067427413,
      "tokens_seen": 2041774080
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.701813513079762e-05,
      "loss": 2.5216,
      "theoretical_loss": 3.424476813135986,
      "tokens_seen": 2041905152
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.701011073663939e-05,
      "loss": 2.6073,
      "theoretical_loss": 3.4244585603443527,
      "tokens_seen": 2042036224
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.700208634248114e-05,
      "loss": 2.609,
      "theoretical_loss": 3.4244403090522937,
      "tokens_seen": 2042167296
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.699406194832291e-05,
      "loss": 2.5124,
      "theoretical_loss": 3.424422059259589,
      "tokens_seen": 2042298368
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.698603755416466e-05,
      "loss": 2.4612,
      "theoretical_loss": 3.4244038109660195,
      "tokens_seen": 2042429440
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.697801316000642e-05,
      "loss": 2.4941,
      "theoretical_loss": 3.4243855641713665,
      "tokens_seen": 2042560512
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.696998876584818e-05,
      "loss": 2.7861,
      "theoretical_loss": 3.4243673188754093,
      "tokens_seen": 2042691584
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.696196437168994e-05,
      "loss": 2.5488,
      "theoretical_loss": 3.42434907507793,
      "tokens_seen": 2042822656
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.69539399775317e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.424330832778708,
      "tokens_seen": 2042953728
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.694591558337346e-05,
      "loss": 2.592,
      "theoretical_loss": 3.4243125919775257,
      "tokens_seen": 2043084800
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1125983,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.432866096496582,
      "objective/train/theoretical_loss": 3.424294352674163,
      "objective/train/tokens_used": 413674976,
      "theoretical_loss": 3.424294352674163,
      "tokens_seen": 2043215872
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.693789118921521e-05,
      "loss": 2.4698,
      "theoretical_loss": 3.424294352674163,
      "tokens_seen": 2043215872
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.692986679505698e-05,
      "loss": 2.5398,
      "theoretical_loss": 3.424276114868401,
      "tokens_seen": 2043346944
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.692184240089873e-05,
      "loss": 2.5617,
      "theoretical_loss": 3.424257878560021,
      "tokens_seen": 2043478016
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.69138180067405e-05,
      "loss": 2.4823,
      "theoretical_loss": 3.4242396437488036,
      "tokens_seen": 2043609088
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.690579361258225e-05,
      "loss": 2.4788,
      "theoretical_loss": 3.42422141043453,
      "tokens_seen": 2043740160
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.689776921842402e-05,
      "loss": 2.6387,
      "theoretical_loss": 3.4242031786169815,
      "tokens_seen": 2043871232
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.688974482426577e-05,
      "loss": 2.5571,
      "theoretical_loss": 3.424184948295939,
      "tokens_seen": 2044002304
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.688172043010752e-05,
      "loss": 2.6261,
      "theoretical_loss": 3.4241667194711845,
      "tokens_seen": 2044133376
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.687369603594929e-05,
      "loss": 2.4709,
      "theoretical_loss": 3.424148492142498,
      "tokens_seen": 2044264448
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.686567164179104e-05,
      "loss": 2.561,
      "theoretical_loss": 3.4241302663096613,
      "tokens_seen": 2044395520
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.685764724763281e-05,
      "loss": 2.473,
      "theoretical_loss": 3.4241120419724567,
      "tokens_seen": 2044526592
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.684962285347456e-05,
      "loss": 2.5965,
      "theoretical_loss": 3.424093819130664,
      "tokens_seen": 2044657664
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.684159845931632e-05,
      "loss": 2.4805,
      "theoretical_loss": 3.4240755977840656,
      "tokens_seen": 2044788736
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1126576,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.342308759689331,
      "objective/train/theoretical_loss": 3.4240664876713955,
      "objective/train/tokens_used": 415313376,
      "theoretical_loss": 3.4240664876713955,
      "tokens_seen": 2044854272
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.683357406515808e-05,
      "loss": 2.5274,
      "theoretical_loss": 3.424057377932443,
      "tokens_seen": 2044919808
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.682554967099984e-05,
      "loss": 2.5764,
      "theoretical_loss": 3.4240391595755773,
      "tokens_seen": 2045050880
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.68175252768416e-05,
      "loss": 2.4526,
      "theoretical_loss": 3.42402094271325,
      "tokens_seen": 2045181952
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.680950088268335e-05,
      "loss": 2.5081,
      "theoretical_loss": 3.424002727345244,
      "tokens_seen": 2045313024
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.680147648852512e-05,
      "loss": 2.4939,
      "theoretical_loss": 3.423984513471339,
      "tokens_seen": 2045444096
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.679345209436687e-05,
      "loss": 2.3648,
      "theoretical_loss": 3.423966301091318,
      "tokens_seen": 2045575168
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.678542770020863e-05,
      "loss": 2.5418,
      "theoretical_loss": 3.423948090204963,
      "tokens_seen": 2045706240
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.67774033060504e-05,
      "loss": 2.4757,
      "theoretical_loss": 3.4239298808120546,
      "tokens_seen": 2045837312
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.676937891189215e-05,
      "loss": 2.4403,
      "theoretical_loss": 3.423911672912376,
      "tokens_seen": 2045968384
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.676135451773391e-05,
      "loss": 2.442,
      "theoretical_loss": 3.4238934665057084,
      "tokens_seen": 2046099456
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.675333012357567e-05,
      "loss": 2.3285,
      "theoretical_loss": 3.4238752615918333,
      "tokens_seen": 2046230528
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.674530572941742e-05,
      "loss": 2.6098,
      "theoretical_loss": 3.4238570581705337,
      "tokens_seen": 2046361600
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1127632,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.210137128829956,
      "objective/train/theoretical_loss": 3.423838856241591,
      "objective/train/tokens_used": 416951776,
      "theoretical_loss": 3.423838856241591,
      "tokens_seen": 2046492672
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.673728133525919e-05,
      "loss": 2.4872,
      "theoretical_loss": 3.423838856241591,
      "tokens_seen": 2046492672
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.672925694110094e-05,
      "loss": 2.629,
      "theoretical_loss": 3.4238206558047874,
      "tokens_seen": 2046623744
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.672123254694271e-05,
      "loss": 2.4547,
      "theoretical_loss": 3.423802456859905,
      "tokens_seen": 2046754816
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.671320815278446e-05,
      "loss": 2.4999,
      "theoretical_loss": 3.423784259406726,
      "tokens_seen": 2046885888
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.670518375862623e-05,
      "loss": 2.4866,
      "theoretical_loss": 3.4237660634450324,
      "tokens_seen": 2047016960
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.669715936446798e-05,
      "loss": 2.4262,
      "theoretical_loss": 3.4237478689746066,
      "tokens_seen": 2047148032
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.668913497030975e-05,
      "loss": 2.3548,
      "theoretical_loss": 3.4237296759952316,
      "tokens_seen": 2047279104
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.66811105761515e-05,
      "loss": 2.546,
      "theoretical_loss": 3.423711484506689,
      "tokens_seen": 2047410176
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.667308618199325e-05,
      "loss": 2.5654,
      "theoretical_loss": 3.4236932945087615,
      "tokens_seen": 2047541248
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.666506178783502e-05,
      "loss": 2.4401,
      "theoretical_loss": 3.423675106001231,
      "tokens_seen": 2047672320
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.665703739367677e-05,
      "loss": 2.4974,
      "theoretical_loss": 3.423656918983881,
      "tokens_seen": 2047803392
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.664901299951854e-05,
      "loss": 2.4328,
      "theoretical_loss": 3.4236387334564933,
      "tokens_seen": 2047934464
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.664098860536029e-05,
      "loss": 2.392,
      "theoretical_loss": 3.42362054941885,
      "tokens_seen": 2048065536
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1128264,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3604929447174072,
      "objective/train/theoretical_loss": 3.423611457958615,
      "objective/train/tokens_used": 418590176,
      "theoretical_loss": 3.423611457958615,
      "tokens_seen": 2048131072
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.663296421120206e-05,
      "loss": 2.6137,
      "theoretical_loss": 3.4236023668707354,
      "tokens_seen": 2048196608
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.662493981704381e-05,
      "loss": 2.468,
      "theoretical_loss": 3.4235841858119302,
      "tokens_seen": 2048327680
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.661691542288557e-05,
      "loss": 2.5171,
      "theoretical_loss": 3.4235660062422184,
      "tokens_seen": 2048458752
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.660889102872733e-05,
      "loss": 2.5218,
      "theoretical_loss": 3.423547828161383,
      "tokens_seen": 2048589824
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.660086663456909e-05,
      "loss": 2.4335,
      "theoretical_loss": 3.423529651569206,
      "tokens_seen": 2048720896
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.659284224041085e-05,
      "loss": 2.4221,
      "theoretical_loss": 3.42351147646547,
      "tokens_seen": 2048851968
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.65848178462526e-05,
      "loss": 2.413,
      "theoretical_loss": 3.423493302849959,
      "tokens_seen": 2048983040
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.657679345209437e-05,
      "loss": 2.5214,
      "theoretical_loss": 3.4234751307224554,
      "tokens_seen": 2049114112
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.656876905793612e-05,
      "loss": 2.4643,
      "theoretical_loss": 3.423456960082742,
      "tokens_seen": 2049245184
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.656074466377788e-05,
      "loss": 2.4014,
      "theoretical_loss": 3.423438790930602,
      "tokens_seen": 2049376256
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.655272026961964e-05,
      "loss": 2.6017,
      "theoretical_loss": 3.4234206232658186,
      "tokens_seen": 2049507328
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.65446958754614e-05,
      "loss": 2.4636,
      "theoretical_loss": 3.4234024570881747,
      "tokens_seen": 2049638400
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1129400,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.676476001739502,
      "objective/train/theoretical_loss": 3.4233842923974542,
      "objective/train/tokens_used": 420228576,
      "theoretical_loss": 3.4233842923974542,
      "tokens_seen": 2049769472
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.653667148130316e-05,
      "loss": 2.6725,
      "theoretical_loss": 3.4233842923974542,
      "tokens_seen": 2049769472
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.652864708714492e-05,
      "loss": 2.5167,
      "theoretical_loss": 3.423366129193439,
      "tokens_seen": 2049900544
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.652062269298668e-05,
      "loss": 2.4925,
      "theoretical_loss": 3.4233479674759137,
      "tokens_seen": 2050031616
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.651259829882844e-05,
      "loss": 2.3167,
      "theoretical_loss": 3.4233298072446607,
      "tokens_seen": 2050162688
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.650457390467019e-05,
      "loss": 2.3845,
      "theoretical_loss": 3.423311648499464,
      "tokens_seen": 2050293760
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.649654951051196e-05,
      "loss": 2.5006,
      "theoretical_loss": 3.4232934912401065,
      "tokens_seen": 2050424832
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.648852511635371e-05,
      "loss": 2.4615,
      "theoretical_loss": 3.4232753354663714,
      "tokens_seen": 2050555904
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.648050072219548e-05,
      "loss": 2.4585,
      "theoretical_loss": 3.4232571811780437,
      "tokens_seen": 2050686976
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.647247632803723e-05,
      "loss": 2.5186,
      "theoretical_loss": 3.423239028374905,
      "tokens_seen": 2050818048
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.6464451933879e-05,
      "loss": 2.4754,
      "theoretical_loss": 3.4232208770567403,
      "tokens_seen": 2050949120
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.645642753972075e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.4232027272233325,
      "tokens_seen": 2051080192
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.64484031455625e-05,
      "loss": 2.7089,
      "theoretical_loss": 3.4231845788744657,
      "tokens_seen": 2051211264
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.644037875140427e-05,
      "loss": 2.5746,
      "theoretical_loss": 3.423166432009923,
      "tokens_seen": 2051342336
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1129773,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.70103120803833,
      "objective/train/theoretical_loss": 3.4231573591342057,
      "objective/train/tokens_used": 421866976,
      "theoretical_loss": 3.4231573591342057,
      "tokens_seen": 2051407872
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.643235435724602e-05,
      "loss": 2.5127,
      "theoretical_loss": 3.4231482866294884,
      "tokens_seen": 2051473408
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.642432996308779e-05,
      "loss": 2.4836,
      "theoretical_loss": 3.423130142732947,
      "tokens_seen": 2051604480
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.641630556892954e-05,
      "loss": 2.4251,
      "theoretical_loss": 3.4231120003200806,
      "tokens_seen": 2051735552
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.640828117477131e-05,
      "loss": 2.48,
      "theoretical_loss": 3.4230938593906743,
      "tokens_seen": 2051866624
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.640025678061306e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.423075719944512,
      "tokens_seen": 2051997696
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.639223238645483e-05,
      "loss": 2.416,
      "theoretical_loss": 3.4230575819813778,
      "tokens_seen": 2052128768
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.638420799229658e-05,
      "loss": 2.6944,
      "theoretical_loss": 3.423039445501055,
      "tokens_seen": 2052259840
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.637618359813834e-05,
      "loss": 2.3774,
      "theoretical_loss": 3.423021310503328,
      "tokens_seen": 2052390912
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.63681592039801e-05,
      "loss": 2.3808,
      "theoretical_loss": 3.423003176987981,
      "tokens_seen": 2052521984
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.636013480982186e-05,
      "loss": 2.4532,
      "theoretical_loss": 3.4229850449547987,
      "tokens_seen": 2052653056
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.635211041566362e-05,
      "loss": 2.5069,
      "theoretical_loss": 3.422966914403565,
      "tokens_seen": 2052784128
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.634408602150538e-05,
      "loss": 2.4513,
      "theoretical_loss": 3.4229487853340634,
      "tokens_seen": 2052915200
    },
    {
      "epoch": 0.24,
      "objective/train/docs_used": 1130910,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.115882396697998,
      "objective/train/theoretical_loss": 3.422930657746079,
      "objective/train/tokens_used": 423505376,
      "theoretical_loss": 3.422930657746079,
      "tokens_seen": 2053046272
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.633606162734714e-05,
      "loss": 2.514,
      "theoretical_loss": 3.422930657746079,
      "tokens_seen": 2053046272
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.63280372331889e-05,
      "loss": 2.5533,
      "theoretical_loss": 3.4229125316393967,
      "tokens_seen": 2053177344
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.632001283903065e-05,
      "loss": 2.4303,
      "theoretical_loss": 3.4228944070137994,
      "tokens_seen": 2053308416
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.631198844487241e-05,
      "loss": 2.5642,
      "theoretical_loss": 3.4228762838690727,
      "tokens_seen": 2053439488
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.630396405071417e-05,
      "loss": 2.5084,
      "theoretical_loss": 3.4228581622050007,
      "tokens_seen": 2053570560
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.629593965655593e-05,
      "loss": 2.5071,
      "theoretical_loss": 3.422840042021368,
      "tokens_seen": 2053701632
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.628791526239769e-05,
      "loss": 2.5386,
      "theoretical_loss": 3.422821923317959,
      "tokens_seen": 2053832704
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.627989086823945e-05,
      "loss": 2.5596,
      "theoretical_loss": 3.422803806094559,
      "tokens_seen": 2053963776
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.627186647408121e-05,
      "loss": 2.6299,
      "theoretical_loss": 3.4227856903509517,
      "tokens_seen": 2054094848
    },
    {
      "epoch": 0.24,
      "learning_rate": 7.626384207992296e-05,
      "loss": 2.2966,
      "theoretical_loss": 3.4227675760869225,
      "tokens_seen": 2054225920
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.625581768576473e-05,
      "loss": 2.5026,
      "theoretical_loss": 3.4227494633022557,
      "tokens_seen": 2054356992
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.624779329160648e-05,
      "loss": 2.4105,
      "theoretical_loss": 3.4227313519967364,
      "tokens_seen": 2054488064
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.623976889744825e-05,
      "loss": 2.467,
      "theoretical_loss": 3.4227132421701496,
      "tokens_seen": 2054619136
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1131479,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.358306884765625,
      "objective/train/theoretical_loss": 3.422704187811389,
      "objective/train/tokens_used": 425143776,
      "theoretical_loss": 3.422704187811389,
      "tokens_seen": 2054684672
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.623174450329e-05,
      "loss": 2.499,
      "theoretical_loss": 3.4226951338222804,
      "tokens_seen": 2054750208
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.622372010913177e-05,
      "loss": 2.6132,
      "theoretical_loss": 3.422677026952913,
      "tokens_seen": 2054881280
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.621569571497352e-05,
      "loss": 2.6517,
      "theoretical_loss": 3.4226589215618324,
      "tokens_seen": 2055012352
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.620767132081527e-05,
      "loss": 2.5089,
      "theoretical_loss": 3.4226408176488246,
      "tokens_seen": 2055143424
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.619964692665704e-05,
      "loss": 2.5091,
      "theoretical_loss": 3.4226227152136737,
      "tokens_seen": 2055274496
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.619162253249879e-05,
      "loss": 2.397,
      "theoretical_loss": 3.422604614256165,
      "tokens_seen": 2055405568
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.618359813834056e-05,
      "loss": 2.5668,
      "theoretical_loss": 3.422586514776084,
      "tokens_seen": 2055536640
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.617557374418231e-05,
      "loss": 2.6935,
      "theoretical_loss": 3.422568416773216,
      "tokens_seen": 2055667712
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.616754935002408e-05,
      "loss": 2.3887,
      "theoretical_loss": 3.4225503202473453,
      "tokens_seen": 2055798784
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.615952495586583e-05,
      "loss": 2.48,
      "theoretical_loss": 3.4225322251982586,
      "tokens_seen": 2055929856
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.615150056170759e-05,
      "loss": 2.3379,
      "theoretical_loss": 3.42251413162574,
      "tokens_seen": 2056060928
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.614347616754935e-05,
      "loss": 2.5482,
      "theoretical_loss": 3.4224960395295763,
      "tokens_seen": 2056192000
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1132437,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8114593029022217,
      "objective/train/theoretical_loss": 3.4224779489095516,
      "objective/train/tokens_used": 426782176,
      "theoretical_loss": 3.4224779489095516,
      "tokens_seen": 2056323072
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.61354517733911e-05,
      "loss": 2.4689,
      "theoretical_loss": 3.4224779489095516,
      "tokens_seen": 2056323072
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.612742737923287e-05,
      "loss": 2.5355,
      "theoretical_loss": 3.422459859765452,
      "tokens_seen": 2056454144
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.611940298507463e-05,
      "loss": 2.4067,
      "theoretical_loss": 3.422441772097063,
      "tokens_seen": 2056585216
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.611137859091639e-05,
      "loss": 2.5967,
      "theoretical_loss": 3.42242368590417,
      "tokens_seen": 2056716288
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.610335419675815e-05,
      "loss": 2.5954,
      "theoretical_loss": 3.4224056011865587,
      "tokens_seen": 2056847360
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.609532980259991e-05,
      "loss": 2.5875,
      "theoretical_loss": 3.4223875179440144,
      "tokens_seen": 2056978432
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.608730540844167e-05,
      "loss": 2.7053,
      "theoretical_loss": 3.4223694361763233,
      "tokens_seen": 2057109504
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.607928101428342e-05,
      "loss": 2.5196,
      "theoretical_loss": 3.422351355883271,
      "tokens_seen": 2057240576
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.607125662012518e-05,
      "loss": 2.5351,
      "theoretical_loss": 3.4223332770646437,
      "tokens_seen": 2057371648
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.606323222596694e-05,
      "loss": 2.3518,
      "theoretical_loss": 3.422315199720227,
      "tokens_seen": 2057502720
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.60552078318087e-05,
      "loss": 2.4935,
      "theoretical_loss": 3.422297123849806,
      "tokens_seen": 2057633792
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.604718343765046e-05,
      "loss": 2.5397,
      "theoretical_loss": 3.4222790494531674,
      "tokens_seen": 2057764864
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.603915904349222e-05,
      "loss": 2.4483,
      "theoretical_loss": 3.4222609765300973,
      "tokens_seen": 2057895936
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1133547,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.867115020751953,
      "objective/train/theoretical_loss": 3.4222519406210834,
      "objective/train/tokens_used": 428420576,
      "theoretical_loss": 3.4222519406210834,
      "tokens_seen": 2057961472
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.603113464933398e-05,
      "loss": 2.597,
      "theoretical_loss": 3.422242905080381,
      "tokens_seen": 2058027008
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.602311025517573e-05,
      "loss": 2.6583,
      "theoretical_loss": 3.4222248351038056,
      "tokens_seen": 2058158080
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.60150858610175e-05,
      "loss": 2.541,
      "theoretical_loss": 3.4222067666001563,
      "tokens_seen": 2058289152
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.600706146685925e-05,
      "loss": 2.3219,
      "theoretical_loss": 3.4221886995692197,
      "tokens_seen": 2058420224
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.599903707270102e-05,
      "loss": 2.3508,
      "theoretical_loss": 3.422170634010782,
      "tokens_seen": 2058551296
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.599101267854277e-05,
      "loss": 2.6189,
      "theoretical_loss": 3.422152569924629,
      "tokens_seen": 2058682368
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.598298828438454e-05,
      "loss": 2.4071,
      "theoretical_loss": 3.422134507310548,
      "tokens_seen": 2058813440
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.597496389022629e-05,
      "loss": 2.3938,
      "theoretical_loss": 3.4221164461683236,
      "tokens_seen": 2058944512
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.596693949606804e-05,
      "loss": 2.4802,
      "theoretical_loss": 3.422098386497744,
      "tokens_seen": 2059075584
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.595891510190981e-05,
      "loss": 2.5144,
      "theoretical_loss": 3.4220803282985948,
      "tokens_seen": 2059206656
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.595089070775156e-05,
      "loss": 2.7101,
      "theoretical_loss": 3.4220622715706623,
      "tokens_seen": 2059337728
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.594286631359333e-05,
      "loss": 2.3732,
      "theoretical_loss": 3.4220442163137332,
      "tokens_seen": 2059468800
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1134157,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.846436023712158,
      "objective/train/theoretical_loss": 3.4220261625275943,
      "objective/train/tokens_used": 430058976,
      "theoretical_loss": 3.4220261625275943,
      "tokens_seen": 2059599872
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.593484191943508e-05,
      "loss": 2.6596,
      "theoretical_loss": 3.4220261625275943,
      "tokens_seen": 2059599872
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.592681752527685e-05,
      "loss": 2.5337,
      "theoretical_loss": 3.422008110212032,
      "tokens_seen": 2059730944
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.59187931311186e-05,
      "loss": 2.5602,
      "theoretical_loss": 3.4219900593668324,
      "tokens_seen": 2059862016
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.591076873696036e-05,
      "loss": 2.5941,
      "theoretical_loss": 3.421972009991783,
      "tokens_seen": 2059993088
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.590274434280212e-05,
      "loss": 2.5179,
      "theoretical_loss": 3.4219539620866706,
      "tokens_seen": 2060124160
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.589471994864388e-05,
      "loss": 2.6585,
      "theoretical_loss": 3.4219359156512814,
      "tokens_seen": 2060255232
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.588669555448564e-05,
      "loss": 2.6361,
      "theoretical_loss": 3.4219178706854025,
      "tokens_seen": 2060386304
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.58786711603274e-05,
      "loss": 2.5926,
      "theoretical_loss": 3.4218998271888204,
      "tokens_seen": 2060517376
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.587064676616916e-05,
      "loss": 2.5422,
      "theoretical_loss": 3.4218817851613226,
      "tokens_seen": 2060648448
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.586262237201092e-05,
      "loss": 2.7374,
      "theoretical_loss": 3.4218637446026956,
      "tokens_seen": 2060779520
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.585459797785267e-05,
      "loss": 2.4504,
      "theoretical_loss": 3.4218457055127267,
      "tokens_seen": 2060910592
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.584657358369444e-05,
      "loss": 2.5051,
      "theoretical_loss": 3.421827667891203,
      "tokens_seen": 2061041664
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.583854918953619e-05,
      "loss": 2.5216,
      "theoretical_loss": 3.4218096317379114,
      "tokens_seen": 2061172736
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1135026,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5525903701782227,
      "objective/train/theoretical_loss": 3.4218006142117856,
      "objective/train/tokens_used": 431697376,
      "theoretical_loss": 3.4218006142117856,
      "tokens_seen": 2061238272
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.583052479537795e-05,
      "loss": 2.5612,
      "theoretical_loss": 3.4217915970526387,
      "tokens_seen": 2061303808
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.582250040121971e-05,
      "loss": 2.5427,
      "theoretical_loss": 3.4217735638351727,
      "tokens_seen": 2061434880
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.581447600706147e-05,
      "loss": 2.5848,
      "theoretical_loss": 3.4217555320853004,
      "tokens_seen": 2061565952
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.580645161290323e-05,
      "loss": 2.6088,
      "theoretical_loss": 3.4217375018028084,
      "tokens_seen": 2061697024
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.579842721874498e-05,
      "loss": 2.4747,
      "theoretical_loss": 3.4217194729874856,
      "tokens_seen": 2061828096
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.579040282458675e-05,
      "loss": 2.652,
      "theoretical_loss": 3.4217014456391177,
      "tokens_seen": 2061959168
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.57823784304285e-05,
      "loss": 2.4892,
      "theoretical_loss": 3.421683419757493,
      "tokens_seen": 2062090240
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.577435403627027e-05,
      "loss": 2.5593,
      "theoretical_loss": 3.4216653953423988,
      "tokens_seen": 2062221312
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.576632964211202e-05,
      "loss": 2.5257,
      "theoretical_loss": 3.4216473723936223,
      "tokens_seen": 2062352384
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.575830524795379e-05,
      "loss": 2.7893,
      "theoretical_loss": 3.421629350910951,
      "tokens_seen": 2062483456
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.575028085379554e-05,
      "loss": 2.51,
      "theoretical_loss": 3.421611330894174,
      "tokens_seen": 2062614528
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.574225645963731e-05,
      "loss": 2.6444,
      "theoretical_loss": 3.4215933123430764,
      "tokens_seen": 2062745600
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1136115,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5874297618865967,
      "objective/train/theoretical_loss": 3.421575295257447,
      "objective/train/tokens_used": 433335776,
      "theoretical_loss": 3.421575295257447,
      "tokens_seen": 2062876672
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.573423206547906e-05,
      "loss": 2.6763,
      "theoretical_loss": 3.421575295257447,
      "tokens_seen": 2062876672
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.572620767132081e-05,
      "loss": 2.6042,
      "theoretical_loss": 3.4215572796370743,
      "tokens_seen": 2063007744
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.571818327716258e-05,
      "loss": 2.534,
      "theoretical_loss": 3.421539265481745,
      "tokens_seen": 2063138816
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.571015888300433e-05,
      "loss": 2.4402,
      "theoretical_loss": 3.4215212527912477,
      "tokens_seen": 2063269888
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.57021344888461e-05,
      "loss": 2.5752,
      "theoretical_loss": 3.4215032415653694,
      "tokens_seen": 2063400960
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.569411009468785e-05,
      "loss": 2.5612,
      "theoretical_loss": 3.4214852318038984,
      "tokens_seen": 2063532032
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.568608570052962e-05,
      "loss": 2.5878,
      "theoretical_loss": 3.421467223506623,
      "tokens_seen": 2063663104
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.567806130637137e-05,
      "loss": 2.6003,
      "theoretical_loss": 3.4214492166733304,
      "tokens_seen": 2063794176
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.567003691221313e-05,
      "loss": 2.5608,
      "theoretical_loss": 3.4214312113038092,
      "tokens_seen": 2063925248
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.566201251805489e-05,
      "loss": 2.5979,
      "theoretical_loss": 3.421413207397847,
      "tokens_seen": 2064056320
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.565398812389665e-05,
      "loss": 2.4841,
      "theoretical_loss": 3.4213952049552327,
      "tokens_seen": 2064187392
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.564596372973841e-05,
      "loss": 2.6268,
      "theoretical_loss": 3.4213772039757537,
      "tokens_seen": 2064318464
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.563793933558017e-05,
      "loss": 2.5995,
      "theoretical_loss": 3.4213592044591987,
      "tokens_seen": 2064449536
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1136653,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.423219919204712,
      "objective/train/theoretical_loss": 3.421350205249451,
      "objective/train/tokens_used": 434974176,
      "theoretical_loss": 3.421350205249451,
      "tokens_seen": 2064515072
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.562991494142193e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.421341206405355,
      "tokens_seen": 2064580608
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.562189054726369e-05,
      "loss": 2.5891,
      "theoretical_loss": 3.421323209814012,
      "tokens_seen": 2064711680
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.561386615310544e-05,
      "loss": 2.7013,
      "theoretical_loss": 3.421305214684958,
      "tokens_seen": 2064842752
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.56058417589472e-05,
      "loss": 2.492,
      "theoretical_loss": 3.4212872210179803,
      "tokens_seen": 2064973824
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.559781736478896e-05,
      "loss": 2.4796,
      "theoretical_loss": 3.4212692288128683,
      "tokens_seen": 2065104896
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.558979297063073e-05,
      "loss": 2.5791,
      "theoretical_loss": 3.4212512380694102,
      "tokens_seen": 2065235968
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.558176857647248e-05,
      "loss": 2.5923,
      "theoretical_loss": 3.421233248787394,
      "tokens_seen": 2065367040
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.557374418231424e-05,
      "loss": 2.5457,
      "theoretical_loss": 3.4212152609666093,
      "tokens_seen": 2065498112
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.5565719788156e-05,
      "loss": 2.4568,
      "theoretical_loss": 3.4211972746068438,
      "tokens_seen": 2065629184
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.555769539399775e-05,
      "loss": 2.6654,
      "theoretical_loss": 3.4211792897078865,
      "tokens_seen": 2065760256
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.554967099983952e-05,
      "loss": 2.5968,
      "theoretical_loss": 3.421161306269526,
      "tokens_seen": 2065891328
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.554164660568127e-05,
      "loss": 2.6777,
      "theoretical_loss": 3.4211433242915508,
      "tokens_seen": 2066022400
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1137651,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5705416202545166,
      "objective/train/theoretical_loss": 3.42112534377375,
      "objective/train/tokens_used": 436612576,
      "theoretical_loss": 3.42112534377375,
      "tokens_seen": 2066153472
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.553362221152304e-05,
      "loss": 2.4683,
      "theoretical_loss": 3.42112534377375,
      "tokens_seen": 2066153472
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.552559781736479e-05,
      "loss": 2.6254,
      "theoretical_loss": 3.4211073647159123,
      "tokens_seen": 2066284544
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.551757342320656e-05,
      "loss": 2.5953,
      "theoretical_loss": 3.4210893871178265,
      "tokens_seen": 2066415616
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.550954902904831e-05,
      "loss": 2.5553,
      "theoretical_loss": 3.4210714109792817,
      "tokens_seen": 2066546688
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.550152463489006e-05,
      "loss": 2.7196,
      "theoretical_loss": 3.421053436300067,
      "tokens_seen": 2066677760
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.549350024073183e-05,
      "loss": 2.4659,
      "theoretical_loss": 3.4210354630799706,
      "tokens_seen": 2066808832
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.548547584657358e-05,
      "loss": 2.5927,
      "theoretical_loss": 3.421017491318782,
      "tokens_seen": 2066939904
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.547745145241535e-05,
      "loss": 2.6497,
      "theoretical_loss": 3.4209995210162907,
      "tokens_seen": 2067070976
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.54694270582571e-05,
      "loss": 2.6517,
      "theoretical_loss": 3.420981552172285,
      "tokens_seen": 2067202048
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.546140266409887e-05,
      "loss": 2.595,
      "theoretical_loss": 3.4209635847865547,
      "tokens_seen": 2067333120
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.545337826994062e-05,
      "loss": 2.6581,
      "theoretical_loss": 3.420945618858889,
      "tokens_seen": 2067464192
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.544535387578239e-05,
      "loss": 2.6037,
      "theoretical_loss": 3.420927654389077,
      "tokens_seen": 2067595264
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.543732948162414e-05,
      "loss": 2.6536,
      "theoretical_loss": 3.4209096913769073,
      "tokens_seen": 2067726336
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1138992,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7702243328094482,
      "objective/train/theoretical_loss": 3.420900710417373,
      "objective/train/tokens_used": 438250976,
      "theoretical_loss": 3.420900710417373,
      "tokens_seen": 2067791872
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.54293050874659e-05,
      "loss": 2.6081,
      "theoretical_loss": 3.4208917298221704,
      "tokens_seen": 2067857408
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.542128069330766e-05,
      "loss": 2.7191,
      "theoretical_loss": 3.420873769724655,
      "tokens_seen": 2067988480
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.541325629914942e-05,
      "loss": 2.6877,
      "theoretical_loss": 3.4208558110841505,
      "tokens_seen": 2068119552
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.540523190499118e-05,
      "loss": 2.628,
      "theoretical_loss": 3.420837853900447,
      "tokens_seen": 2068250624
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.539720751083294e-05,
      "loss": 2.6661,
      "theoretical_loss": 3.420819898173333,
      "tokens_seen": 2068381696
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.53891831166747e-05,
      "loss": 2.7085,
      "theoretical_loss": 3.4208019439025987,
      "tokens_seen": 2068512768
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.538115872251646e-05,
      "loss": 2.6818,
      "theoretical_loss": 3.420783991088034,
      "tokens_seen": 2068643840
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.537313432835821e-05,
      "loss": 2.6039,
      "theoretical_loss": 3.420766039729428,
      "tokens_seen": 2068774912
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.536510993419998e-05,
      "loss": 2.6936,
      "theoretical_loss": 3.4207480898265707,
      "tokens_seen": 2068905984
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.535708554004173e-05,
      "loss": 2.5812,
      "theoretical_loss": 3.4207301413792512,
      "tokens_seen": 2069037056
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.53490611458835e-05,
      "loss": 2.5736,
      "theoretical_loss": 3.4207121943872605,
      "tokens_seen": 2069168128
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.534103675172525e-05,
      "loss": 2.5398,
      "theoretical_loss": 3.4206942488503875,
      "tokens_seen": 2069299200
    },
    {
      "epoch": 0.25,
      "objective/train/docs_used": 1139620,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4647276401519775,
      "objective/train/theoretical_loss": 3.420676304768422,
      "objective/train/tokens_used": 439889376,
      "theoretical_loss": 3.420676304768422,
      "tokens_seen": 2069430272
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.533301235756701e-05,
      "loss": 2.4302,
      "theoretical_loss": 3.420676304768422,
      "tokens_seen": 2069430272
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.532498796340877e-05,
      "loss": 2.4488,
      "theoretical_loss": 3.420658362141154,
      "tokens_seen": 2069561344
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.531696356925052e-05,
      "loss": 2.548,
      "theoretical_loss": 3.4206404209683736,
      "tokens_seen": 2069692416
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.530893917509229e-05,
      "loss": 2.6237,
      "theoretical_loss": 3.4206224812498713,
      "tokens_seen": 2069823488
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.530091478093404e-05,
      "loss": 2.6025,
      "theoretical_loss": 3.420604542985436,
      "tokens_seen": 2069954560
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.529289038677581e-05,
      "loss": 2.6769,
      "theoretical_loss": 3.420586606174859,
      "tokens_seen": 2070085632
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.528486599261756e-05,
      "loss": 2.6906,
      "theoretical_loss": 3.4205686708179295,
      "tokens_seen": 2070216704
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.527684159845933e-05,
      "loss": 2.6362,
      "theoretical_loss": 3.4205507369144383,
      "tokens_seen": 2070347776
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.526881720430108e-05,
      "loss": 2.5741,
      "theoretical_loss": 3.420532804464175,
      "tokens_seen": 2070478848
    },
    {
      "epoch": 0.25,
      "learning_rate": 7.526079281014283e-05,
      "loss": 2.5953,
      "theoretical_loss": 3.4205148734669306,
      "tokens_seen": 2070609920
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.52527684159846e-05,
      "loss": 2.5617,
      "theoretical_loss": 3.420496943922495,
      "tokens_seen": 2070740992
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.524474402182635e-05,
      "loss": 2.6418,
      "theoretical_loss": 3.420479015830658,
      "tokens_seen": 2070872064
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.523671962766812e-05,
      "loss": 2.5491,
      "theoretical_loss": 3.4204610891912104,
      "tokens_seen": 2071003136
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1140715,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.805562973022461,
      "objective/train/theoretical_loss": 3.4204521264160674,
      "objective/train/tokens_used": 441527776,
      "theoretical_loss": 3.4204521264160674,
      "tokens_seen": 2071068672
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.522869523350987e-05,
      "loss": 2.6671,
      "theoretical_loss": 3.4204431640039434,
      "tokens_seen": 2071134208
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.522067083935164e-05,
      "loss": 2.6024,
      "theoretical_loss": 3.4204252402686466,
      "tokens_seen": 2071265280
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.521264644519339e-05,
      "loss": 2.6613,
      "theoretical_loss": 3.42040731798511,
      "tokens_seen": 2071396352
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.520462205103515e-05,
      "loss": 2.5674,
      "theoretical_loss": 3.4203893971531256,
      "tokens_seen": 2071527424
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.519659765687691e-05,
      "loss": 2.6061,
      "theoretical_loss": 3.420371477772483,
      "tokens_seen": 2071658496
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.518857326271867e-05,
      "loss": 2.6192,
      "theoretical_loss": 3.4203535598429737,
      "tokens_seen": 2071789568
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.518054886856043e-05,
      "loss": 2.5439,
      "theoretical_loss": 3.4203356433643877,
      "tokens_seen": 2071920640
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.517252447440219e-05,
      "loss": 2.5913,
      "theoretical_loss": 3.4203177283365154,
      "tokens_seen": 2072051712
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.516450008024395e-05,
      "loss": 2.5747,
      "theoretical_loss": 3.420299814759148,
      "tokens_seen": 2072182784
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.51564756860857e-05,
      "loss": 2.4535,
      "theoretical_loss": 3.4202819026320768,
      "tokens_seen": 2072313856
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.514845129192747e-05,
      "loss": 2.5588,
      "theoretical_loss": 3.4202639919550926,
      "tokens_seen": 2072444928
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.514042689776923e-05,
      "loss": 2.5327,
      "theoretical_loss": 3.420246082727985,
      "tokens_seen": 2072576000
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1141285,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8016202449798584,
      "objective/train/theoretical_loss": 3.4202281749505463,
      "objective/train/tokens_used": 443166176,
      "theoretical_loss": 3.4202281749505463,
      "tokens_seen": 2072707072
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.513240250361098e-05,
      "loss": 2.6386,
      "theoretical_loss": 3.4202281749505463,
      "tokens_seen": 2072707072
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.512437810945275e-05,
      "loss": 2.6437,
      "theoretical_loss": 3.420210268622567,
      "tokens_seen": 2072838144
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.51163537152945e-05,
      "loss": 2.5769,
      "theoretical_loss": 3.4201923637438383,
      "tokens_seen": 2072969216
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.510832932113627e-05,
      "loss": 2.5697,
      "theoretical_loss": 3.4201744603141515,
      "tokens_seen": 2073100288
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.510030492697802e-05,
      "loss": 2.6594,
      "theoretical_loss": 3.420156558333297,
      "tokens_seen": 2073231360
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.509228053281979e-05,
      "loss": 2.5883,
      "theoretical_loss": 3.420138657801066,
      "tokens_seen": 2073362432
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.508425613866154e-05,
      "loss": 2.6873,
      "theoretical_loss": 3.4201207587172506,
      "tokens_seen": 2073493504
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.507623174450329e-05,
      "loss": 2.4776,
      "theoretical_loss": 3.4201028610816415,
      "tokens_seen": 2073624576
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.506820735034506e-05,
      "loss": 2.6836,
      "theoretical_loss": 3.4200849648940306,
      "tokens_seen": 2073755648
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.506018295618681e-05,
      "loss": 2.4895,
      "theoretical_loss": 3.420067070154208,
      "tokens_seen": 2073886720
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.505215856202858e-05,
      "loss": 2.6642,
      "theoretical_loss": 3.420049176861966,
      "tokens_seen": 2074017792
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.504413416787033e-05,
      "loss": 2.6899,
      "theoretical_loss": 3.420031285017096,
      "tokens_seen": 2074148864
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.50361097737121e-05,
      "loss": 2.5657,
      "theoretical_loss": 3.4200133946193887,
      "tokens_seen": 2074279936
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1142545,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.676924228668213,
      "objective/train/theoretical_loss": 3.420004449963156,
      "objective/train/tokens_used": 444804576,
      "theoretical_loss": 3.420004449963156,
      "tokens_seen": 2074345472
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.502808537955385e-05,
      "loss": 2.5201,
      "theoretical_loss": 3.419995505668637,
      "tokens_seen": 2074411008
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.50200609853956e-05,
      "loss": 2.8034,
      "theoretical_loss": 3.419977618164631,
      "tokens_seen": 2074542080
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.501203659123737e-05,
      "loss": 2.6983,
      "theoretical_loss": 3.419959732107163,
      "tokens_seen": 2074673152
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.500401219707912e-05,
      "loss": 2.6096,
      "theoretical_loss": 3.419941847496025,
      "tokens_seen": 2074804224
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.499598780292089e-05,
      "loss": 2.5876,
      "theoretical_loss": 3.4199239643310078,
      "tokens_seen": 2074935296
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.498796340876264e-05,
      "loss": 2.69,
      "theoretical_loss": 3.4199060826119037,
      "tokens_seen": 2075066368
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.497993901460441e-05,
      "loss": 2.7312,
      "theoretical_loss": 3.4198882023385044,
      "tokens_seen": 2075197440
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.497191462044616e-05,
      "loss": 2.5642,
      "theoretical_loss": 3.419870323510602,
      "tokens_seen": 2075328512
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.496389022628792e-05,
      "loss": 2.6408,
      "theoretical_loss": 3.4198524461279876,
      "tokens_seen": 2075459584
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.495586583212968e-05,
      "loss": 2.5564,
      "theoretical_loss": 3.419834570190454,
      "tokens_seen": 2075590656
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.494784143797144e-05,
      "loss": 2.6452,
      "theoretical_loss": 3.4198166956977927,
      "tokens_seen": 2075721728
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.49398170438132e-05,
      "loss": 2.7101,
      "theoretical_loss": 3.4197988226497955,
      "tokens_seen": 2075852800
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1143739,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.8349756002426147,
      "objective/train/theoretical_loss": 3.4197809510462545,
      "objective/train/tokens_used": 446442976,
      "theoretical_loss": 3.4197809510462545,
      "tokens_seen": 2075983872
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.493179264965496e-05,
      "loss": 2.4385,
      "theoretical_loss": 3.4197809510462545,
      "tokens_seen": 2075983872
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.492376825549672e-05,
      "loss": 2.5877,
      "theoretical_loss": 3.4197630808869617,
      "tokens_seen": 2076114944
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.491574386133848e-05,
      "loss": 2.6309,
      "theoretical_loss": 3.41974521217171,
      "tokens_seen": 2076246016
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.490771946718023e-05,
      "loss": 2.5745,
      "theoretical_loss": 3.4197273449002905,
      "tokens_seen": 2076377088
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.4899695073022e-05,
      "loss": 2.5886,
      "theoretical_loss": 3.419709479072496,
      "tokens_seen": 2076508160
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.489167067886375e-05,
      "loss": 2.5897,
      "theoretical_loss": 3.4196916146881184,
      "tokens_seen": 2076639232
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.488364628470552e-05,
      "loss": 2.5993,
      "theoretical_loss": 3.419673751746951,
      "tokens_seen": 2076770304
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.487562189054727e-05,
      "loss": 2.5253,
      "theoretical_loss": 3.4196558902487846,
      "tokens_seen": 2076901376
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.486759749638904e-05,
      "loss": 2.5155,
      "theoretical_loss": 3.419638030193413,
      "tokens_seen": 2077032448
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.485957310223079e-05,
      "loss": 2.5878,
      "theoretical_loss": 3.419620171580627,
      "tokens_seen": 2077163520
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.485154870807254e-05,
      "loss": 2.5929,
      "theoretical_loss": 3.419602314410221,
      "tokens_seen": 2077294592
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.484352431391431e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.4195844586819857,
      "tokens_seen": 2077425664
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.483549991975606e-05,
      "loss": 2.5385,
      "theoretical_loss": 3.419566604395715,
      "tokens_seen": 2077556736
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1144416,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.198617935180664,
      "objective/train/theoretical_loss": 3.419557677793251,
      "objective/train/tokens_used": 448081376,
      "theoretical_loss": 3.419557677793251,
      "tokens_seen": 2077622272
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.482747552559783e-05,
      "loss": 2.745,
      "theoretical_loss": 3.4195487515512006,
      "tokens_seen": 2077687808
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.481945113143958e-05,
      "loss": 2.7385,
      "theoretical_loss": 3.4195309001482355,
      "tokens_seen": 2077818880
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.481142673728135e-05,
      "loss": 2.5999,
      "theoretical_loss": 3.4195130501866124,
      "tokens_seen": 2077949952
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.48034023431231e-05,
      "loss": 2.5005,
      "theoretical_loss": 3.419495201666124,
      "tokens_seen": 2078081024
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.479537794896487e-05,
      "loss": 2.6459,
      "theoretical_loss": 3.4194773545865633,
      "tokens_seen": 2078212096
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.478735355480662e-05,
      "loss": 2.5881,
      "theoretical_loss": 3.419459508947723,
      "tokens_seen": 2078343168
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.477932916064837e-05,
      "loss": 2.5664,
      "theoretical_loss": 3.419441664749395,
      "tokens_seen": 2078474240
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.477130476649014e-05,
      "loss": 2.6161,
      "theoretical_loss": 3.4194238219913737,
      "tokens_seen": 2078605312
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.47632803723319e-05,
      "loss": 2.4891,
      "theoretical_loss": 3.4194059806734507,
      "tokens_seen": 2078736384
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.475525597817366e-05,
      "loss": 2.4439,
      "theoretical_loss": 3.41938814079542,
      "tokens_seen": 2078867456
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.474723158401541e-05,
      "loss": 2.6349,
      "theoretical_loss": 3.4193703023570743,
      "tokens_seen": 2078998528
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.473920718985718e-05,
      "loss": 2.5836,
      "theoretical_loss": 3.419352465358206,
      "tokens_seen": 2079129600
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1145721,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.877591133117676,
      "objective/train/theoretical_loss": 3.4193346297986094,
      "objective/train/tokens_used": 449719776,
      "theoretical_loss": 3.4193346297986094,
      "tokens_seen": 2079260672
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.473118279569893e-05,
      "loss": 2.5798,
      "theoretical_loss": 3.4193346297986094,
      "tokens_seen": 2079260672
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.472315840154069e-05,
      "loss": 2.6641,
      "theoretical_loss": 3.4193167956780766,
      "tokens_seen": 2079391744
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.471513400738245e-05,
      "loss": 2.5415,
      "theoretical_loss": 3.4192989629964012,
      "tokens_seen": 2079522816
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.47071096132242e-05,
      "loss": 2.5896,
      "theoretical_loss": 3.419281131753377,
      "tokens_seen": 2079653888
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.469908521906597e-05,
      "loss": 2.5036,
      "theoretical_loss": 3.4192633019487966,
      "tokens_seen": 2079784960
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.469106082490773e-05,
      "loss": 2.5731,
      "theoretical_loss": 3.419245473582453,
      "tokens_seen": 2079916032
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.468303643074949e-05,
      "loss": 2.6267,
      "theoretical_loss": 3.4192276466541403,
      "tokens_seen": 2080047104
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.467501203659125e-05,
      "loss": 2.4308,
      "theoretical_loss": 3.4192098211636512,
      "tokens_seen": 2080178176
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.4666987642433e-05,
      "loss": 2.3999,
      "theoretical_loss": 3.41919199711078,
      "tokens_seen": 2080309248
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.465896324827477e-05,
      "loss": 2.392,
      "theoretical_loss": 3.4191741744953195,
      "tokens_seen": 2080440320
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.465093885411652e-05,
      "loss": 2.5587,
      "theoretical_loss": 3.4191563533170637,
      "tokens_seen": 2080571392
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.464291445995829e-05,
      "loss": 2.6061,
      "theoretical_loss": 3.4191385335758055,
      "tokens_seen": 2080702464
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.463489006580004e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.4191207152713394,
      "tokens_seen": 2080833536
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1146352,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6832168102264404,
      "objective/train/theoretical_loss": 3.4191118066578383,
      "objective/train/tokens_used": 451358176,
      "theoretical_loss": 3.4191118066578383,
      "tokens_seen": 2080899072
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.46268656716418e-05,
      "loss": 2.6432,
      "theoretical_loss": 3.419102898403459,
      "tokens_seen": 2080964608
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.461884127748356e-05,
      "loss": 2.6323,
      "theoretical_loss": 3.4190850829719563,
      "tokens_seen": 2081095680
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.461081688332531e-05,
      "loss": 2.6782,
      "theoretical_loss": 3.4190672689766277,
      "tokens_seen": 2081226752
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.460279248916708e-05,
      "loss": 2.674,
      "theoretical_loss": 3.419049456417265,
      "tokens_seen": 2081357824
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.459476809500883e-05,
      "loss": 2.6025,
      "theoretical_loss": 3.419031645293663,
      "tokens_seen": 2081488896
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.45867437008506e-05,
      "loss": 2.5709,
      "theoretical_loss": 3.419013835605615,
      "tokens_seen": 2081619968
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.457871930669235e-05,
      "loss": 2.6332,
      "theoretical_loss": 3.4189960273529154,
      "tokens_seen": 2081751040
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.457069491253412e-05,
      "loss": 2.4575,
      "theoretical_loss": 3.4189782205353576,
      "tokens_seen": 2081882112
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.456267051837587e-05,
      "loss": 2.4975,
      "theoretical_loss": 3.4189604151527364,
      "tokens_seen": 2082013184
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.455464612421762e-05,
      "loss": 2.4342,
      "theoretical_loss": 3.418942611204845,
      "tokens_seen": 2082144256
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.454662173005939e-05,
      "loss": 2.6555,
      "theoretical_loss": 3.418924808691478,
      "tokens_seen": 2082275328
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.453859733590114e-05,
      "loss": 2.6492,
      "theoretical_loss": 3.4189070076124293,
      "tokens_seen": 2082406400
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1147344,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.602205753326416,
      "objective/train/theoretical_loss": 3.418889207967493,
      "objective/train/tokens_used": 452996576,
      "theoretical_loss": 3.418889207967493,
      "tokens_seen": 2082537472
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.453057294174291e-05,
      "loss": 2.6904,
      "theoretical_loss": 3.418889207967493,
      "tokens_seen": 2082537472
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.452254854758466e-05,
      "loss": 2.7331,
      "theoretical_loss": 3.4188714097564636,
      "tokens_seen": 2082668544
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.451452415342643e-05,
      "loss": 2.4997,
      "theoretical_loss": 3.4188536129791354,
      "tokens_seen": 2082799616
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.450649975926818e-05,
      "loss": 2.6038,
      "theoretical_loss": 3.418835817635303,
      "tokens_seen": 2082930688
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.449847536510994e-05,
      "loss": 2.5311,
      "theoretical_loss": 3.41881802372476,
      "tokens_seen": 2083061760
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.44904509709517e-05,
      "loss": 2.5913,
      "theoretical_loss": 3.4188002312473005,
      "tokens_seen": 2083192832
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.448242657679346e-05,
      "loss": 2.5041,
      "theoretical_loss": 3.41878244020272,
      "tokens_seen": 2083323904
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.447440218263522e-05,
      "loss": 2.5274,
      "theoretical_loss": 3.418764650590812,
      "tokens_seen": 2083454976
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.446637778847698e-05,
      "loss": 2.6918,
      "theoretical_loss": 3.4187468624113717,
      "tokens_seen": 2083586048
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.445835339431873e-05,
      "loss": 2.6158,
      "theoretical_loss": 3.4187290756641935,
      "tokens_seen": 2083717120
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.44503290001605e-05,
      "loss": 2.7229,
      "theoretical_loss": 3.418711290349072,
      "tokens_seen": 2083848192
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.444230460600225e-05,
      "loss": 2.6351,
      "theoretical_loss": 3.418693506465801,
      "tokens_seen": 2083979264
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.443428021184402e-05,
      "loss": 2.7173,
      "theoretical_loss": 3.418675724014177,
      "tokens_seen": 2084110336
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1148002,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8767805099487305,
      "objective/train/theoretical_loss": 3.418666833325168,
      "objective/train/tokens_used": 454634976,
      "theoretical_loss": 3.418666833325168,
      "tokens_seen": 2084175872
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.442625581768577e-05,
      "loss": 2.6401,
      "theoretical_loss": 3.4186579429939927,
      "tokens_seen": 2084241408
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.441823142352754e-05,
      "loss": 2.4244,
      "theoretical_loss": 3.4186401634050445,
      "tokens_seen": 2084372480
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.441020702936929e-05,
      "loss": 2.5651,
      "theoretical_loss": 3.4186223852471262,
      "tokens_seen": 2084503552
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.440218263521104e-05,
      "loss": 2.6412,
      "theoretical_loss": 3.4186046085200332,
      "tokens_seen": 2084634624
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.439415824105281e-05,
      "loss": 2.3889,
      "theoretical_loss": 3.41858683322356,
      "tokens_seen": 2084765696
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.438613384689456e-05,
      "loss": 2.5168,
      "theoretical_loss": 3.4185690593575018,
      "tokens_seen": 2084896768
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.437810945273633e-05,
      "loss": 2.6211,
      "theoretical_loss": 3.418551286921653,
      "tokens_seen": 2085027840
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.437008505857808e-05,
      "loss": 2.6815,
      "theoretical_loss": 3.4185335159158097,
      "tokens_seen": 2085158912
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.436206066441983e-05,
      "loss": 2.4658,
      "theoretical_loss": 3.418515746339766,
      "tokens_seen": 2085289984
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.43540362702616e-05,
      "loss": 2.5338,
      "theoretical_loss": 3.4184979781933174,
      "tokens_seen": 2085421056
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.434601187610335e-05,
      "loss": 2.5857,
      "theoretical_loss": 3.418480211476259,
      "tokens_seen": 2085552128
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.433798748194512e-05,
      "loss": 2.5221,
      "theoretical_loss": 3.4184624461883857,
      "tokens_seen": 2085683200
    },
    {
      "epoch": 0.26,
      "objective/train/docs_used": 1149018,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5401086807250977,
      "objective/train/theoretical_loss": 3.4184446823294934,
      "objective/train/tokens_used": 456273376,
      "theoretical_loss": 3.4184446823294934,
      "tokens_seen": 2085814272
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.432996308778687e-05,
      "loss": 2.5125,
      "theoretical_loss": 3.4184446823294934,
      "tokens_seen": 2085814272
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.432193869362864e-05,
      "loss": 2.5801,
      "theoretical_loss": 3.4184269198993773,
      "tokens_seen": 2085945344
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.43139142994704e-05,
      "loss": 2.4876,
      "theoretical_loss": 3.4184091588978314,
      "tokens_seen": 2086076416
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.430588990531215e-05,
      "loss": 2.8154,
      "theoretical_loss": 3.418391399324653,
      "tokens_seen": 2086207488
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.429786551115391e-05,
      "loss": 2.5371,
      "theoretical_loss": 3.4183736411796364,
      "tokens_seen": 2086338560
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.428984111699567e-05,
      "loss": 2.6052,
      "theoretical_loss": 3.4183558844625765,
      "tokens_seen": 2086469632
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.428181672283743e-05,
      "loss": 2.4728,
      "theoretical_loss": 3.41833812917327,
      "tokens_seen": 2086600704
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.427379232867919e-05,
      "loss": 2.4988,
      "theoretical_loss": 3.418320375311512,
      "tokens_seen": 2086731776
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.426576793452094e-05,
      "loss": 2.7325,
      "theoretical_loss": 3.4183026228770976,
      "tokens_seen": 2086862848
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.425774354036271e-05,
      "loss": 2.6953,
      "theoretical_loss": 3.4182848718698233,
      "tokens_seen": 2086993920
    },
    {
      "epoch": 0.26,
      "learning_rate": 7.424971914620446e-05,
      "loss": 2.5597,
      "theoretical_loss": 3.4182671222894836,
      "tokens_seen": 2087124992
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.424169475204623e-05,
      "loss": 2.5705,
      "theoretical_loss": 3.4182493741358755,
      "tokens_seen": 2087256064
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.423367035788798e-05,
      "loss": 2.5799,
      "theoretical_loss": 3.4182316274087934,
      "tokens_seen": 2087387136
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1149534,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.625910997390747,
      "objective/train/theoretical_loss": 3.4182227545801362,
      "objective/train/tokens_used": 457911776,
      "theoretical_loss": 3.4182227545801362,
      "tokens_seen": 2087452672
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.422564596372975e-05,
      "loss": 2.5105,
      "theoretical_loss": 3.418213882108034,
      "tokens_seen": 2087518208
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.42176215695715e-05,
      "loss": 2.5787,
      "theoretical_loss": 3.4181961382333927,
      "tokens_seen": 2087649280
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.420959717541325e-05,
      "loss": 2.8012,
      "theoretical_loss": 3.4181783957846656,
      "tokens_seen": 2087780352
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.420157278125502e-05,
      "loss": 2.6249,
      "theoretical_loss": 3.4181606547616488,
      "tokens_seen": 2087911424
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.419354838709677e-05,
      "loss": 2.4211,
      "theoretical_loss": 3.418142915164138,
      "tokens_seen": 2088042496
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.418552399293854e-05,
      "loss": 2.5906,
      "theoretical_loss": 3.4181251769919285,
      "tokens_seen": 2088173568
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.417749959878029e-05,
      "loss": 2.5411,
      "theoretical_loss": 3.4181074402448175,
      "tokens_seen": 2088304640
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.416947520462205e-05,
      "loss": 2.4751,
      "theoretical_loss": 3.4180897049226004,
      "tokens_seen": 2088435712
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.416145081046381e-05,
      "loss": 2.6218,
      "theoretical_loss": 3.418071971025073,
      "tokens_seen": 2088566784
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.415342641630557e-05,
      "loss": 2.6089,
      "theoretical_loss": 3.4180542385520325,
      "tokens_seen": 2088697856
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.414540202214733e-05,
      "loss": 2.607,
      "theoretical_loss": 3.4180365075032744,
      "tokens_seen": 2088828928
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.413737762798909e-05,
      "loss": 2.5198,
      "theoretical_loss": 3.4180187778785953,
      "tokens_seen": 2088960000
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1150483,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.777698040008545,
      "objective/train/theoretical_loss": 3.4180010496777906,
      "objective/train/tokens_used": 459550176,
      "theoretical_loss": 3.4180010496777906,
      "tokens_seen": 2089091072
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.412935323383084e-05,
      "loss": 2.5472,
      "theoretical_loss": 3.4180010496777906,
      "tokens_seen": 2089091072
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.41213288396726e-05,
      "loss": 2.4747,
      "theoretical_loss": 3.4179833229006578,
      "tokens_seen": 2089222144
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.411330444551436e-05,
      "loss": 2.588,
      "theoretical_loss": 3.417965597546992,
      "tokens_seen": 2089353216
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.410528005135612e-05,
      "loss": 2.6024,
      "theoretical_loss": 3.4179478736165914,
      "tokens_seen": 2089484288
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.409725565719788e-05,
      "loss": 2.6249,
      "theoretical_loss": 3.4179301511092506,
      "tokens_seen": 2089615360
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.408923126303964e-05,
      "loss": 2.5506,
      "theoretical_loss": 3.417912430024767,
      "tokens_seen": 2089746432
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.40812068688814e-05,
      "loss": 2.6243,
      "theoretical_loss": 3.417894710362937,
      "tokens_seen": 2089877504
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.407318247472315e-05,
      "loss": 2.5144,
      "theoretical_loss": 3.4178769921235572,
      "tokens_seen": 2090008576
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.406515808056492e-05,
      "loss": 2.5705,
      "theoretical_loss": 3.4178592753064243,
      "tokens_seen": 2090139648
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.405713368640667e-05,
      "loss": 2.5887,
      "theoretical_loss": 3.417841559911335,
      "tokens_seen": 2090270720
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.404910929224844e-05,
      "loss": 2.55,
      "theoretical_loss": 3.4178238459380856,
      "tokens_seen": 2090401792
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.404108489809019e-05,
      "loss": 2.6119,
      "theoretical_loss": 3.417806133386473,
      "tokens_seen": 2090532864
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.403306050393196e-05,
      "loss": 2.5615,
      "theoretical_loss": 3.417788422256294,
      "tokens_seen": 2090663936
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1151030,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4571421146392822,
      "objective/train/theoretical_loss": 3.417779567224179,
      "objective/train/tokens_used": 461188576,
      "theoretical_loss": 3.417779567224179,
      "tokens_seen": 2090729472
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.402503610977371e-05,
      "loss": 2.5956,
      "theoretical_loss": 3.417770712547346,
      "tokens_seen": 2090795008
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.401701171561546e-05,
      "loss": 2.6705,
      "theoretical_loss": 3.417753004259425,
      "tokens_seen": 2090926080
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.400898732145723e-05,
      "loss": 2.5801,
      "theoretical_loss": 3.417735297392328,
      "tokens_seen": 2091057152
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.400096292729898e-05,
      "loss": 2.3938,
      "theoretical_loss": 3.417717591945853,
      "tokens_seen": 2091188224
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.399293853314075e-05,
      "loss": 2.5603,
      "theoretical_loss": 3.4176998879197957,
      "tokens_seen": 2091319296
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.39849141389825e-05,
      "loss": 2.653,
      "theoretical_loss": 3.4176821853139536,
      "tokens_seen": 2091450368
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.397688974482427e-05,
      "loss": 2.4816,
      "theoretical_loss": 3.417664484128124,
      "tokens_seen": 2091581440
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.396886535066602e-05,
      "loss": 2.5028,
      "theoretical_loss": 3.4176467843621037,
      "tokens_seen": 2091712512
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.396084095650778e-05,
      "loss": 2.6164,
      "theoretical_loss": 3.4176290860156904,
      "tokens_seen": 2091843584
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.395281656234954e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.4176113890886803,
      "tokens_seen": 2091974656
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.39447921681913e-05,
      "loss": 2.4333,
      "theoretical_loss": 3.417593693580872,
      "tokens_seen": 2092105728
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.393676777403306e-05,
      "loss": 2.5275,
      "theoretical_loss": 3.4175759994920614,
      "tokens_seen": 2092236800
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1152201,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.786438226699829,
      "objective/train/theoretical_loss": 3.4175583068220465,
      "objective/train/tokens_used": 462826976,
      "theoretical_loss": 3.4175583068220465,
      "tokens_seen": 2092367872
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.392874337987482e-05,
      "loss": 2.5778,
      "theoretical_loss": 3.4175583068220465,
      "tokens_seen": 2092367872
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.392071898571658e-05,
      "loss": 2.4786,
      "theoretical_loss": 3.4175406155706245,
      "tokens_seen": 2092498944
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.391269459155834e-05,
      "loss": 2.4362,
      "theoretical_loss": 3.4175229257375936,
      "tokens_seen": 2092630016
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.390467019740009e-05,
      "loss": 2.5344,
      "theoretical_loss": 3.41750523732275,
      "tokens_seen": 2092761088
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.389664580324186e-05,
      "loss": 2.4331,
      "theoretical_loss": 3.417487550325892,
      "tokens_seen": 2092892160
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.388862140908361e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.417469864746817,
      "tokens_seen": 2093023232
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.388059701492537e-05,
      "loss": 2.7274,
      "theoretical_loss": 3.417452180585322,
      "tokens_seen": 2093154304
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.387257262076713e-05,
      "loss": 2.4391,
      "theoretical_loss": 3.4174344978412057,
      "tokens_seen": 2093285376
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.38645482266089e-05,
      "loss": 2.5441,
      "theoretical_loss": 3.417416816514265,
      "tokens_seen": 2093416448
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.385652383245065e-05,
      "loss": 2.6726,
      "theoretical_loss": 3.4173991366042973,
      "tokens_seen": 2093547520
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.384849943829241e-05,
      "loss": 2.7121,
      "theoretical_loss": 3.4173814581111017,
      "tokens_seen": 2093678592
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.384047504413417e-05,
      "loss": 2.4686,
      "theoretical_loss": 3.4173637810344744,
      "tokens_seen": 2093809664
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.383245064997592e-05,
      "loss": 2.4801,
      "theoretical_loss": 3.417346105374214,
      "tokens_seen": 2093940736
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1153513,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.452009439468384,
      "objective/train/theoretical_loss": 3.4173372680751584,
      "objective/train/tokens_used": 464465376,
      "theoretical_loss": 3.4173372680751584,
      "tokens_seen": 2094006272
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.382442625581769e-05,
      "loss": 2.5239,
      "theoretical_loss": 3.4173284311301186,
      "tokens_seen": 2094071808
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.381640186165944e-05,
      "loss": 2.5636,
      "theoretical_loss": 3.4173107583019853,
      "tokens_seen": 2094202880
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.380837746750121e-05,
      "loss": 2.6055,
      "theoretical_loss": 3.4172930868896128,
      "tokens_seen": 2094333952
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.380035307334296e-05,
      "loss": 2.4273,
      "theoretical_loss": 3.4172754168927986,
      "tokens_seen": 2094465024
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.379232867918473e-05,
      "loss": 2.5082,
      "theoretical_loss": 3.417257748311341,
      "tokens_seen": 2094596096
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.378430428502648e-05,
      "loss": 2.5065,
      "theoretical_loss": 3.4172400811450387,
      "tokens_seen": 2094727168
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.377627989086823e-05,
      "loss": 2.4834,
      "theoretical_loss": 3.417222415393688,
      "tokens_seen": 2094858240
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.376825549671e-05,
      "loss": 2.5748,
      "theoretical_loss": 3.4172047510570893,
      "tokens_seen": 2094989312
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.376023110255175e-05,
      "loss": 2.5213,
      "theoretical_loss": 3.417187088135039,
      "tokens_seen": 2095120384
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.375220670839352e-05,
      "loss": 2.5495,
      "theoretical_loss": 3.4171694266273365,
      "tokens_seen": 2095251456
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.374418231423527e-05,
      "loss": 2.5869,
      "theoretical_loss": 3.417151766533779,
      "tokens_seen": 2095382528
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.373615792007704e-05,
      "loss": 2.5185,
      "theoretical_loss": 3.4171341078541664,
      "tokens_seen": 2095513600
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1154141,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.867384433746338,
      "objective/train/theoretical_loss": 3.4171164505882956,
      "objective/train/tokens_used": 466103776,
      "theoretical_loss": 3.4171164505882956,
      "tokens_seen": 2095644672
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.372813352591879e-05,
      "loss": 2.6868,
      "theoretical_loss": 3.4171164505882956,
      "tokens_seen": 2095644672
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.372010913176055e-05,
      "loss": 2.5989,
      "theoretical_loss": 3.4170987947359652,
      "tokens_seen": 2095775744
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.371208473760231e-05,
      "loss": 2.413,
      "theoretical_loss": 3.4170811402969745,
      "tokens_seen": 2095906816
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.370406034344407e-05,
      "loss": 2.5521,
      "theoretical_loss": 3.417063487271121,
      "tokens_seen": 2096037888
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.369603594928583e-05,
      "loss": 2.6069,
      "theoretical_loss": 3.417045835658204,
      "tokens_seen": 2096168960
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.368801155512759e-05,
      "loss": 2.5313,
      "theoretical_loss": 3.417028185458021,
      "tokens_seen": 2096300032
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.367998716096935e-05,
      "loss": 2.5615,
      "theoretical_loss": 3.4170105366703725,
      "tokens_seen": 2096431104
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.36719627668111e-05,
      "loss": 2.5193,
      "theoretical_loss": 3.416992889295055,
      "tokens_seen": 2096562176
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.366393837265286e-05,
      "loss": 2.6729,
      "theoretical_loss": 3.4169752433318688,
      "tokens_seen": 2096693248
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.365591397849463e-05,
      "loss": 2.569,
      "theoretical_loss": 3.416957598780612,
      "tokens_seen": 2096824320
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.364788958433638e-05,
      "loss": 2.5365,
      "theoretical_loss": 3.416939955641083,
      "tokens_seen": 2096955392
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.363986519017815e-05,
      "loss": 2.612,
      "theoretical_loss": 3.4169223139130813,
      "tokens_seen": 2097086464
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.36318407960199e-05,
      "loss": 2.6286,
      "theoretical_loss": 3.416904673596406,
      "tokens_seen": 2097217536
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1155185,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.522921085357666,
      "objective/train/theoretical_loss": 3.416895853967252,
      "objective/train/tokens_used": 467742176,
      "theoretical_loss": 3.416895853967252,
      "tokens_seen": 2097283072
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.362381640186166e-05,
      "loss": 2.5816,
      "theoretical_loss": 3.4168870346908546,
      "tokens_seen": 2097348608
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.361579200770342e-05,
      "loss": 2.6585,
      "theoretical_loss": 3.416869397196227,
      "tokens_seen": 2097479680
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.360776761354517e-05,
      "loss": 2.5419,
      "theoretical_loss": 3.416851761112323,
      "tokens_seen": 2097610752
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.359974321938694e-05,
      "loss": 2.5072,
      "theoretical_loss": 3.41683412643894,
      "tokens_seen": 2097741824
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.359171882522869e-05,
      "loss": 2.6212,
      "theoretical_loss": 3.4168164931758778,
      "tokens_seen": 2097872896
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.358369443107046e-05,
      "loss": 2.5326,
      "theoretical_loss": 3.416798861322936,
      "tokens_seen": 2098003968
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.357567003691221e-05,
      "loss": 2.4969,
      "theoretical_loss": 3.416781230879913,
      "tokens_seen": 2098135040
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.356764564275398e-05,
      "loss": 2.6687,
      "theoretical_loss": 3.416763601846608,
      "tokens_seen": 2098266112
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.355962124859573e-05,
      "loss": 2.6071,
      "theoretical_loss": 3.416745974222821,
      "tokens_seen": 2098397184
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.35515968544375e-05,
      "loss": 2.3758,
      "theoretical_loss": 3.41672834800835,
      "tokens_seen": 2098528256
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.354357246027925e-05,
      "loss": 2.5369,
      "theoretical_loss": 3.416710723202996,
      "tokens_seen": 2098659328
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.3535548066121e-05,
      "loss": 2.5719,
      "theoretical_loss": 3.4166930998065568,
      "tokens_seen": 2098790400
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1155992,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.662437677383423,
      "objective/train/theoretical_loss": 3.416675477818832,
      "objective/train/tokens_used": 469380576,
      "theoretical_loss": 3.416675477818832,
      "tokens_seen": 2098921472
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.352752367196277e-05,
      "loss": 2.5635,
      "theoretical_loss": 3.416675477818832,
      "tokens_seen": 2098921472
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.351949927780452e-05,
      "loss": 2.5932,
      "theoretical_loss": 3.4166578572396222,
      "tokens_seen": 2099052544
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.351147488364629e-05,
      "loss": 2.5664,
      "theoretical_loss": 3.416640238068726,
      "tokens_seen": 2099183616
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.350345048948804e-05,
      "loss": 2.681,
      "theoretical_loss": 3.416622620305943,
      "tokens_seen": 2099314688
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.349542609532981e-05,
      "loss": 2.4963,
      "theoretical_loss": 3.416605003951073,
      "tokens_seen": 2099445760
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.348740170117156e-05,
      "loss": 2.552,
      "theoretical_loss": 3.416587389003915,
      "tokens_seen": 2099576832
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.347937730701332e-05,
      "loss": 2.5934,
      "theoretical_loss": 3.416569775464269,
      "tokens_seen": 2099707904
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.347135291285508e-05,
      "loss": 2.6231,
      "theoretical_loss": 3.416552163331935,
      "tokens_seen": 2099838976
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.346332851869684e-05,
      "loss": 2.7223,
      "theoretical_loss": 3.4165345526067123,
      "tokens_seen": 2099970048
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.34553041245386e-05,
      "loss": 2.4176,
      "theoretical_loss": 3.416516943288401,
      "tokens_seen": 2100101120
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.344727973038036e-05,
      "loss": 2.5539,
      "theoretical_loss": 3.4164993353768005,
      "tokens_seen": 2100232192
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.343925533622212e-05,
      "loss": 2.4916,
      "theoretical_loss": 3.4164817288717106,
      "tokens_seen": 2100363264
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.343123094206388e-05,
      "loss": 2.5582,
      "theoretical_loss": 3.416464123772932,
      "tokens_seen": 2100494336
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1157390,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.862727642059326,
      "objective/train/theoretical_loss": 3.4164553217508464,
      "objective/train/tokens_used": 471018976,
      "theoretical_loss": 3.4164553217508464,
      "tokens_seen": 2100559872
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.342320654790563e-05,
      "loss": 2.6738,
      "theoretical_loss": 3.416446520080264,
      "tokens_seen": 2100625408
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.34151821537474e-05,
      "loss": 2.4995,
      "theoretical_loss": 3.4164289177935063,
      "tokens_seen": 2100756480
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.340715775958915e-05,
      "loss": 2.5321,
      "theoretical_loss": 3.4164113169124595,
      "tokens_seen": 2100887552
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.339913336543092e-05,
      "loss": 2.392,
      "theoretical_loss": 3.4163937174369234,
      "tokens_seen": 2101018624
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.339110897127267e-05,
      "loss": 2.5493,
      "theoretical_loss": 3.4163761193666975,
      "tokens_seen": 2101149696
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.338308457711443e-05,
      "loss": 2.6421,
      "theoretical_loss": 3.416358522701583,
      "tokens_seen": 2101280768
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.337506018295619e-05,
      "loss": 2.4459,
      "theoretical_loss": 3.4163409274413796,
      "tokens_seen": 2101411840
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.336703578879794e-05,
      "loss": 2.5705,
      "theoretical_loss": 3.416323333585888,
      "tokens_seen": 2101542912
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.335901139463971e-05,
      "loss": 2.6294,
      "theoretical_loss": 3.416305741134907,
      "tokens_seen": 2101673984
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.335098700048146e-05,
      "loss": 2.4931,
      "theoretical_loss": 3.4162881500882385,
      "tokens_seen": 2101805056
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.334296260632323e-05,
      "loss": 2.6335,
      "theoretical_loss": 3.416270560445682,
      "tokens_seen": 2101936128
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.333493821216498e-05,
      "loss": 2.42,
      "theoretical_loss": 3.4162529722070385,
      "tokens_seen": 2102067200
    },
    {
      "epoch": 0.27,
      "objective/train/docs_used": 1158091,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.529240369796753,
      "objective/train/theoretical_loss": 3.416235385372107,
      "objective/train/tokens_used": 472657376,
      "theoretical_loss": 3.416235385372107,
      "tokens_seen": 2102198272
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.332691381800675e-05,
      "loss": 2.4835,
      "theoretical_loss": 3.416235385372107,
      "tokens_seen": 2102198272
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.33188894238485e-05,
      "loss": 2.739,
      "theoretical_loss": 3.41621779994069,
      "tokens_seen": 2102329344
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.331086502969025e-05,
      "loss": 2.4934,
      "theoretical_loss": 3.4162002159125864,
      "tokens_seen": 2102460416
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.330284063553202e-05,
      "loss": 2.552,
      "theoretical_loss": 3.4161826332875975,
      "tokens_seen": 2102591488
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.329481624137377e-05,
      "loss": 2.4815,
      "theoretical_loss": 3.4161650520655233,
      "tokens_seen": 2102722560
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.328679184721554e-05,
      "loss": 2.5107,
      "theoretical_loss": 3.4161474722461653,
      "tokens_seen": 2102853632
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.32787674530573e-05,
      "loss": 2.6167,
      "theoretical_loss": 3.416129893829323,
      "tokens_seen": 2102984704
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.327074305889906e-05,
      "loss": 2.5377,
      "theoretical_loss": 3.4161123168147984,
      "tokens_seen": 2103115776
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.326271866474081e-05,
      "loss": 2.4288,
      "theoretical_loss": 3.416094741202391,
      "tokens_seen": 2103246848
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.325469427058258e-05,
      "loss": 2.3633,
      "theoretical_loss": 3.416077166991903,
      "tokens_seen": 2103377920
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.324666987642433e-05,
      "loss": 2.5235,
      "theoretical_loss": 3.416059594183134,
      "tokens_seen": 2103508992
    },
    {
      "epoch": 0.27,
      "learning_rate": 7.323864548226609e-05,
      "loss": 2.5012,
      "theoretical_loss": 3.416042022775885,
      "tokens_seen": 2103640064
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.323062108810785e-05,
      "loss": 2.6528,
      "theoretical_loss": 3.4160244527699577,
      "tokens_seen": 2103771136
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1159419,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5441555976867676,
      "objective/train/theoretical_loss": 3.4160156682924274,
      "objective/train/tokens_used": 474295776,
      "theoretical_loss": 3.4160156682924274,
      "tokens_seen": 2103836672
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.32225966939496e-05,
      "loss": 2.5143,
      "theoretical_loss": 3.416006884165152,
      "tokens_seen": 2103902208
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.321457229979137e-05,
      "loss": 2.5838,
      "theoretical_loss": 3.4159893169612703,
      "tokens_seen": 2104033280
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.320654790563313e-05,
      "loss": 2.4385,
      "theoretical_loss": 3.4159717511581116,
      "tokens_seen": 2104164352
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.319852351147489e-05,
      "loss": 2.4926,
      "theoretical_loss": 3.415954186755479,
      "tokens_seen": 2104295424
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.319049911731665e-05,
      "loss": 2.4845,
      "theoretical_loss": 3.4159366237531725,
      "tokens_seen": 2104426496
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.31824747231584e-05,
      "loss": 2.5013,
      "theoretical_loss": 3.4159190621509934,
      "tokens_seen": 2104557568
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.317445032900017e-05,
      "loss": 2.5026,
      "theoretical_loss": 3.415901501948743,
      "tokens_seen": 2104688640
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.316642593484192e-05,
      "loss": 2.6106,
      "theoretical_loss": 3.4158839431462225,
      "tokens_seen": 2104819712
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.315840154068369e-05,
      "loss": 2.3583,
      "theoretical_loss": 3.415866385743233,
      "tokens_seen": 2104950784
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.315037714652544e-05,
      "loss": 2.5919,
      "theoretical_loss": 3.415848829739576,
      "tokens_seen": 2105081856
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.31423527523672e-05,
      "loss": 2.559,
      "theoretical_loss": 3.4158312751350532,
      "tokens_seen": 2105212928
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.313432835820896e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.4158137219294655,
      "tokens_seen": 2105344000
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1160366,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4362361431121826,
      "objective/train/theoretical_loss": 3.415796170122615,
      "objective/train/tokens_used": 475934176,
      "theoretical_loss": 3.415796170122615,
      "tokens_seen": 2105475072
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.312630396405071e-05,
      "loss": 2.5706,
      "theoretical_loss": 3.415796170122615,
      "tokens_seen": 2105475072
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.311827956989248e-05,
      "loss": 2.4484,
      "theoretical_loss": 3.415778619714302,
      "tokens_seen": 2105606144
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.311025517573423e-05,
      "loss": 2.4628,
      "theoretical_loss": 3.4157610707043284,
      "tokens_seen": 2105737216
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.3102230781576e-05,
      "loss": 2.5434,
      "theoretical_loss": 3.415743523092497,
      "tokens_seen": 2105868288
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.309420638741775e-05,
      "loss": 2.5093,
      "theoretical_loss": 3.4157259768786075,
      "tokens_seen": 2105999360
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.308618199325952e-05,
      "loss": 2.443,
      "theoretical_loss": 3.415708432062463,
      "tokens_seen": 2106130432
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.307815759910127e-05,
      "loss": 2.6816,
      "theoretical_loss": 3.415690888643865,
      "tokens_seen": 2106261504
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.307013320494302e-05,
      "loss": 2.6213,
      "theoretical_loss": 3.415673346622614,
      "tokens_seen": 2106392576
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.306210881078479e-05,
      "loss": 2.6613,
      "theoretical_loss": 3.415655805998513,
      "tokens_seen": 2106523648
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.305408441662654e-05,
      "loss": 2.52,
      "theoretical_loss": 3.415638266771363,
      "tokens_seen": 2106654720
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.304606002246831e-05,
      "loss": 2.6228,
      "theoretical_loss": 3.415620728940967,
      "tokens_seen": 2106785792
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.303803562831006e-05,
      "loss": 2.4382,
      "theoretical_loss": 3.4156031925071257,
      "tokens_seen": 2106916864
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.303001123415183e-05,
      "loss": 2.595,
      "theoretical_loss": 3.4155856574696415,
      "tokens_seen": 2107047936
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1161079,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0612099170684814,
      "objective/train/theoretical_loss": 3.4155768904744717,
      "objective/train/tokens_used": 477572576,
      "theoretical_loss": 3.4155768904744717,
      "tokens_seen": 2107113472
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.302198683999358e-05,
      "loss": 2.6113,
      "theoretical_loss": 3.415568123828317,
      "tokens_seen": 2107179008
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.301396244583534e-05,
      "loss": 2.6254,
      "theoretical_loss": 3.4155505915829525,
      "tokens_seen": 2107310080
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.30059380516771e-05,
      "loss": 2.5267,
      "theoretical_loss": 3.4155330607333516,
      "tokens_seen": 2107441152
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.299791365751886e-05,
      "loss": 2.4114,
      "theoretical_loss": 3.4155155312793157,
      "tokens_seen": 2107572224
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.298988926336062e-05,
      "loss": 2.5213,
      "theoretical_loss": 3.4154980032206472,
      "tokens_seen": 2107703296
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.298186486920238e-05,
      "loss": 2.6768,
      "theoretical_loss": 3.415480476557148,
      "tokens_seen": 2107834368
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.297384047504414e-05,
      "loss": 2.4979,
      "theoretical_loss": 3.4154629512886205,
      "tokens_seen": 2107965440
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.29658160808859e-05,
      "loss": 2.5837,
      "theoretical_loss": 3.415445427414867,
      "tokens_seen": 2108096512
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.295779168672765e-05,
      "loss": 2.5184,
      "theoretical_loss": 3.4154279049356897,
      "tokens_seen": 2108227584
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.294976729256942e-05,
      "loss": 2.487,
      "theoretical_loss": 3.4154103838508907,
      "tokens_seen": 2108358656
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.294174289841117e-05,
      "loss": 2.5156,
      "theoretical_loss": 3.415392864160273,
      "tokens_seen": 2108489728
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.293371850425294e-05,
      "loss": 2.6961,
      "theoretical_loss": 3.4153753458636382,
      "tokens_seen": 2108620800
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1162446,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.430760622024536,
      "objective/train/theoretical_loss": 3.415357828960789,
      "objective/train/tokens_used": 479210976,
      "theoretical_loss": 3.415357828960789,
      "tokens_seen": 2108751872
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.292569411009469e-05,
      "loss": 2.5458,
      "theoretical_loss": 3.415357828960789,
      "tokens_seen": 2108751872
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.291766971593646e-05,
      "loss": 2.6747,
      "theoretical_loss": 3.4153403134515283,
      "tokens_seen": 2108882944
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.290964532177821e-05,
      "loss": 2.5972,
      "theoretical_loss": 3.415322799335658,
      "tokens_seen": 2109014016
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.290162092761998e-05,
      "loss": 2.4214,
      "theoretical_loss": 3.4153052866129814,
      "tokens_seen": 2109145088
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.289359653346173e-05,
      "loss": 2.5883,
      "theoretical_loss": 3.4152877752833,
      "tokens_seen": 2109276160
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.288557213930348e-05,
      "loss": 2.6023,
      "theoretical_loss": 3.4152702653464178,
      "tokens_seen": 2109407232
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.287754774514525e-05,
      "loss": 2.5545,
      "theoretical_loss": 3.4152527568021362,
      "tokens_seen": 2109538304
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.2869523350987e-05,
      "loss": 2.5484,
      "theoretical_loss": 3.4152352496502596,
      "tokens_seen": 2109669376
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.286149895682877e-05,
      "loss": 2.5416,
      "theoretical_loss": 3.4152177438905893,
      "tokens_seen": 2109800448
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.285347456267052e-05,
      "loss": 2.501,
      "theoretical_loss": 3.415200239522928,
      "tokens_seen": 2109931520
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.284545016851229e-05,
      "loss": 2.663,
      "theoretical_loss": 3.4151827365470795,
      "tokens_seen": 2110062592
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.283742577435404e-05,
      "loss": 2.5081,
      "theoretical_loss": 3.4151652349628465,
      "tokens_seen": 2110193664
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.28294013801958e-05,
      "loss": 2.4971,
      "theoretical_loss": 3.415147734770031,
      "tokens_seen": 2110324736
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1163114,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.412344217300415,
      "objective/train/theoretical_loss": 3.4151389851953438,
      "objective/train/tokens_used": 480849376,
      "theoretical_loss": 3.4151389851953438,
      "tokens_seen": 2110390272
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.282137698603756e-05,
      "loss": 2.6189,
      "theoretical_loss": 3.4151302359684372,
      "tokens_seen": 2110455808
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.281335259187931e-05,
      "loss": 2.4918,
      "theoretical_loss": 3.4151127385578675,
      "tokens_seen": 2110586880
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.280532819772108e-05,
      "loss": 2.634,
      "theoretical_loss": 3.415095242538125,
      "tokens_seen": 2110717952
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.279730380356283e-05,
      "loss": 2.5236,
      "theoretical_loss": 3.415077747909013,
      "tokens_seen": 2110849024
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.27892794094046e-05,
      "loss": 2.4392,
      "theoretical_loss": 3.415060254670334,
      "tokens_seen": 2110980096
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.278125501524635e-05,
      "loss": 2.4037,
      "theoretical_loss": 3.415042762821892,
      "tokens_seen": 2111111168
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.27732306210881e-05,
      "loss": 2.5942,
      "theoretical_loss": 3.41502527236349,
      "tokens_seen": 2111242240
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.276520622692987e-05,
      "loss": 2.5675,
      "theoretical_loss": 3.4150077832949304,
      "tokens_seen": 2111373312
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.275718183277163e-05,
      "loss": 2.6567,
      "theoretical_loss": 3.4149902956160174,
      "tokens_seen": 2111504384
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.274915743861339e-05,
      "loss": 2.4766,
      "theoretical_loss": 3.4149728093265543,
      "tokens_seen": 2111635456
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.274113304445515e-05,
      "loss": 2.5684,
      "theoretical_loss": 3.4149553244263444,
      "tokens_seen": 2111766528
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.273310865029691e-05,
      "loss": 2.4782,
      "theoretical_loss": 3.4149378409151905,
      "tokens_seen": 2111897600
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1164441,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.561600923538208,
      "objective/train/theoretical_loss": 3.414920358792897,
      "objective/train/tokens_used": 482487776,
      "theoretical_loss": 3.414920358792897,
      "tokens_seen": 2112028672
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.272508425613867e-05,
      "loss": 2.6107,
      "theoretical_loss": 3.414920358792897,
      "tokens_seen": 2112028672
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.271705986198042e-05,
      "loss": 2.3715,
      "theoretical_loss": 3.4149028780592667,
      "tokens_seen": 2112159744
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.270903546782219e-05,
      "loss": 2.5158,
      "theoretical_loss": 3.414885398714103,
      "tokens_seen": 2112290816
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.270101107366394e-05,
      "loss": 2.5778,
      "theoretical_loss": 3.41486792075721,
      "tokens_seen": 2112421888
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.26929866795057e-05,
      "loss": 2.5923,
      "theoretical_loss": 3.4148504441883913,
      "tokens_seen": 2112552960
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.268496228534746e-05,
      "loss": 2.5593,
      "theoretical_loss": 3.414832969007451,
      "tokens_seen": 2112684032
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.267693789118923e-05,
      "loss": 2.5111,
      "theoretical_loss": 3.414815495214191,
      "tokens_seen": 2112815104
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.266891349703098e-05,
      "loss": 2.6016,
      "theoretical_loss": 3.4147980228084167,
      "tokens_seen": 2112946176
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.266088910287273e-05,
      "loss": 2.488,
      "theoretical_loss": 3.414780551789931,
      "tokens_seen": 2113077248
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.26528647087145e-05,
      "loss": 2.5142,
      "theoretical_loss": 3.4147630821585384,
      "tokens_seen": 2113208320
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.264484031455625e-05,
      "loss": 2.5584,
      "theoretical_loss": 3.414745613914042,
      "tokens_seen": 2113339392
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.263681592039802e-05,
      "loss": 2.5915,
      "theoretical_loss": 3.4147281470562465,
      "tokens_seen": 2113470464
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.262879152623977e-05,
      "loss": 2.5965,
      "theoretical_loss": 3.414710681584956,
      "tokens_seen": 2113601536
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1165048,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.748392105102539,
      "objective/train/theoretical_loss": 3.4147019493691877,
      "objective/train/tokens_used": 484126176,
      "theoretical_loss": 3.4147019493691877,
      "tokens_seen": 2113667072
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.262076713208154e-05,
      "loss": 2.5164,
      "theoretical_loss": 3.414693217499973,
      "tokens_seen": 2113732608
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.261274273792329e-05,
      "loss": 2.5338,
      "theoretical_loss": 3.4146757548011024,
      "tokens_seen": 2113863680
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.260471834376506e-05,
      "loss": 2.5395,
      "theoretical_loss": 3.4146582934881486,
      "tokens_seen": 2113994752
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.259669394960681e-05,
      "loss": 2.6317,
      "theoretical_loss": 3.4146408335609157,
      "tokens_seen": 2114125824
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.258866955544856e-05,
      "loss": 2.4815,
      "theoretical_loss": 3.4146233750192065,
      "tokens_seen": 2114256896
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.258064516129033e-05,
      "loss": 2.6,
      "theoretical_loss": 3.4146059178628265,
      "tokens_seen": 2114387968
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.257262076713208e-05,
      "loss": 2.5097,
      "theoretical_loss": 3.4145884620915803,
      "tokens_seen": 2114519040
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.256459637297385e-05,
      "loss": 2.5873,
      "theoretical_loss": 3.4145710077052707,
      "tokens_seen": 2114650112
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.25565719788156e-05,
      "loss": 2.4999,
      "theoretical_loss": 3.4145535547037023,
      "tokens_seen": 2114781184
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.254854758465737e-05,
      "loss": 2.7295,
      "theoretical_loss": 3.4145361030866805,
      "tokens_seen": 2114912256
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.254052319049912e-05,
      "loss": 2.6493,
      "theoretical_loss": 3.414518652854009,
      "tokens_seen": 2115043328
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.253249879634088e-05,
      "loss": 2.5968,
      "theoretical_loss": 3.414501204005492,
      "tokens_seen": 2115174400
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1166321,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.903796672821045,
      "objective/train/theoretical_loss": 3.4144837565409336,
      "objective/train/tokens_used": 485764576,
      "theoretical_loss": 3.4144837565409336,
      "tokens_seen": 2115305472
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.252447440218264e-05,
      "loss": 2.6134,
      "theoretical_loss": 3.4144837565409336,
      "tokens_seen": 2115305472
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.25164500080244e-05,
      "loss": 2.4382,
      "theoretical_loss": 3.4144663104601394,
      "tokens_seen": 2115436544
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.250842561386616e-05,
      "loss": 2.5556,
      "theoretical_loss": 3.414448865762913,
      "tokens_seen": 2115567616
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.250040121970792e-05,
      "loss": 2.5734,
      "theoretical_loss": 3.4144314224490597,
      "tokens_seen": 2115698688
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.249237682554968e-05,
      "loss": 2.4823,
      "theoretical_loss": 3.4144139805183835,
      "tokens_seen": 2115829760
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.248435243139144e-05,
      "loss": 2.659,
      "theoretical_loss": 3.4143965399706895,
      "tokens_seen": 2115960832
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.247632803723319e-05,
      "loss": 2.3849,
      "theoretical_loss": 3.4143791008057818,
      "tokens_seen": 2116091904
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.246830364307496e-05,
      "loss": 2.5886,
      "theoretical_loss": 3.4143616630234654,
      "tokens_seen": 2116222976
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.246027924891671e-05,
      "loss": 2.5187,
      "theoretical_loss": 3.4143442266235455,
      "tokens_seen": 2116354048
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.245225485475848e-05,
      "loss": 2.4991,
      "theoretical_loss": 3.414326791605826,
      "tokens_seen": 2116485120
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.244423046060023e-05,
      "loss": 2.6432,
      "theoretical_loss": 3.414309357970113,
      "tokens_seen": 2116616192
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.2436206066442e-05,
      "loss": 2.5072,
      "theoretical_loss": 3.41429192571621,
      "tokens_seen": 2116747264
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.242818167228375e-05,
      "loss": 2.4161,
      "theoretical_loss": 3.414274494843923,
      "tokens_seen": 2116878336
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1166896,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.635124921798706,
      "objective/train/theoretical_loss": 3.4142657799258243,
      "objective/train/tokens_used": 487402976,
      "theoretical_loss": 3.4142657799258243,
      "tokens_seen": 2116943872
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.24201572781255e-05,
      "loss": 2.4925,
      "theoretical_loss": 3.4142570653530564,
      "tokens_seen": 2117009408
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.241213288396727e-05,
      "loss": 2.5845,
      "theoretical_loss": 3.414239637243415,
      "tokens_seen": 2117140480
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.240410848980902e-05,
      "loss": 2.4822,
      "theoretical_loss": 3.414222210514805,
      "tokens_seen": 2117271552
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.239608409565079e-05,
      "loss": 2.6126,
      "theoretical_loss": 3.41420478516703,
      "tokens_seen": 2117402624
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.238805970149254e-05,
      "loss": 2.4967,
      "theoretical_loss": 3.414187361199896,
      "tokens_seen": 2117533696
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.238003530733431e-05,
      "loss": 2.5921,
      "theoretical_loss": 3.414169938613208,
      "tokens_seen": 2117664768
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.237201091317606e-05,
      "loss": 2.5172,
      "theoretical_loss": 3.4141525174067704,
      "tokens_seen": 2117795840
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.236398651901781e-05,
      "loss": 2.5096,
      "theoretical_loss": 3.41413509758039,
      "tokens_seen": 2117926912
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.235596212485958e-05,
      "loss": 2.3977,
      "theoretical_loss": 3.414117679133871,
      "tokens_seen": 2118057984
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.234793773070133e-05,
      "loss": 2.5047,
      "theoretical_loss": 3.414100262067019,
      "tokens_seen": 2118189056
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.23399133365431e-05,
      "loss": 2.5251,
      "theoretical_loss": 3.4140828463796398,
      "tokens_seen": 2118320128
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.233188894238485e-05,
      "loss": 2.5052,
      "theoretical_loss": 3.4140654320715376,
      "tokens_seen": 2118451200
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1168203,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.8257986307144165,
      "objective/train/theoretical_loss": 3.414048019142519,
      "objective/train/tokens_used": 489041376,
      "theoretical_loss": 3.414048019142519,
      "tokens_seen": 2118582272
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.232386454822662e-05,
      "loss": 2.3003,
      "theoretical_loss": 3.414048019142519,
      "tokens_seen": 2118582272
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.231584015406837e-05,
      "loss": 2.4695,
      "theoretical_loss": 3.414030607592389,
      "tokens_seen": 2118713344
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.230781575991014e-05,
      "loss": 2.5806,
      "theoretical_loss": 3.414013197420953,
      "tokens_seen": 2118844416
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.22997913657519e-05,
      "loss": 2.6054,
      "theoretical_loss": 3.413995788628017,
      "tokens_seen": 2118975488
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.229176697159365e-05,
      "loss": 2.5005,
      "theoretical_loss": 3.4139783812133864,
      "tokens_seen": 2119106560
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.228374257743541e-05,
      "loss": 2.6139,
      "theoretical_loss": 3.4139609751768663,
      "tokens_seen": 2119237632
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.227571818327717e-05,
      "loss": 2.5732,
      "theoretical_loss": 3.413943570518263,
      "tokens_seen": 2119368704
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.226769378911893e-05,
      "loss": 2.5102,
      "theoretical_loss": 3.413926167237382,
      "tokens_seen": 2119499776
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.225966939496069e-05,
      "loss": 2.4827,
      "theoretical_loss": 3.413908765334029,
      "tokens_seen": 2119630848
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.225164500080245e-05,
      "loss": 2.5705,
      "theoretical_loss": 3.41389136480801,
      "tokens_seen": 2119761920
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.22436206066442e-05,
      "loss": 2.5917,
      "theoretical_loss": 3.413873965659131,
      "tokens_seen": 2119892992
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.223559621248596e-05,
      "loss": 2.3827,
      "theoretical_loss": 3.4138565678871973,
      "tokens_seen": 2120024064
    },
    {
      "epoch": 0.28,
      "learning_rate": 7.222757181832773e-05,
      "loss": 2.4065,
      "theoretical_loss": 3.4138391714920147,
      "tokens_seen": 2120155136
    },
    {
      "epoch": 0.28,
      "objective/train/docs_used": 1168773,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.275192975997925,
      "objective/train/theoretical_loss": 3.413830473810645,
      "objective/train/tokens_used": 490679776,
      "theoretical_loss": 3.413830473810645,
      "tokens_seen": 2120220672
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.221954742416948e-05,
      "loss": 2.4053,
      "theoretical_loss": 3.41382177647339,
      "tokens_seen": 2120286208
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.221152303001125e-05,
      "loss": 2.4353,
      "theoretical_loss": 3.413804382831129,
      "tokens_seen": 2120417280
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.2203498635853e-05,
      "loss": 2.467,
      "theoretical_loss": 3.413786990565037,
      "tokens_seen": 2120548352
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.219547424169477e-05,
      "loss": 2.5341,
      "theoretical_loss": 3.4137695996749207,
      "tokens_seen": 2120679424
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.218744984753652e-05,
      "loss": 2.3319,
      "theoretical_loss": 3.413752210160586,
      "tokens_seen": 2120810496
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.217942545337827e-05,
      "loss": 2.5447,
      "theoretical_loss": 3.413734822021839,
      "tokens_seen": 2120941568
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.217140105922004e-05,
      "loss": 2.7,
      "theoretical_loss": 3.413717435258486,
      "tokens_seen": 2121072640
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.216337666506179e-05,
      "loss": 2.5168,
      "theoretical_loss": 3.413700049870333,
      "tokens_seen": 2121203712
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.215535227090356e-05,
      "loss": 2.4817,
      "theoretical_loss": 3.413682665857187,
      "tokens_seen": 2121334784
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.214732787674531e-05,
      "loss": 2.5098,
      "theoretical_loss": 3.4136652832188528,
      "tokens_seen": 2121465856
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.213930348258708e-05,
      "loss": 2.5541,
      "theoretical_loss": 3.4136479019551382,
      "tokens_seen": 2121596928
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.213127908842883e-05,
      "loss": 2.3622,
      "theoretical_loss": 3.4136305220658496,
      "tokens_seen": 2121728000
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1169526,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4141392707824707,
      "objective/train/theoretical_loss": 3.4136131435507924,
      "objective/train/tokens_used": 492318176,
      "theoretical_loss": 3.4136131435507924,
      "tokens_seen": 2121859072
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.212325469427058e-05,
      "loss": 2.4221,
      "theoretical_loss": 3.4136131435507924,
      "tokens_seen": 2121859072
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.211523030011235e-05,
      "loss": 2.4602,
      "theoretical_loss": 3.413595766409774,
      "tokens_seen": 2121990144
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.21072059059541e-05,
      "loss": 2.3975,
      "theoretical_loss": 3.4135783906425994,
      "tokens_seen": 2122121216
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.209918151179587e-05,
      "loss": 2.443,
      "theoretical_loss": 3.4135610162490773,
      "tokens_seen": 2122252288
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.209115711763762e-05,
      "loss": 2.387,
      "theoretical_loss": 3.4135436432290125,
      "tokens_seen": 2122383360
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.208313272347939e-05,
      "loss": 2.696,
      "theoretical_loss": 3.4135262715822123,
      "tokens_seen": 2122514432
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.207510832932114e-05,
      "loss": 2.6137,
      "theoretical_loss": 3.4135089013084836,
      "tokens_seen": 2122645504
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.20670839351629e-05,
      "loss": 2.4994,
      "theoretical_loss": 3.413491532407633,
      "tokens_seen": 2122776576
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.205905954100466e-05,
      "loss": 2.6495,
      "theoretical_loss": 3.4134741648794664,
      "tokens_seen": 2122907648
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.205103514684642e-05,
      "loss": 2.5628,
      "theoretical_loss": 3.413456798723792,
      "tokens_seen": 2123038720
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.204301075268818e-05,
      "loss": 2.5866,
      "theoretical_loss": 3.4134394339404155,
      "tokens_seen": 2123169792
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.203498635852994e-05,
      "loss": 2.5491,
      "theoretical_loss": 3.413422070529144,
      "tokens_seen": 2123300864
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.20269619643717e-05,
      "loss": 2.5522,
      "theoretical_loss": 3.413404708489785,
      "tokens_seen": 2123431936
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1170253,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3250832557678223,
      "objective/train/theoretical_loss": 3.4133960279845117,
      "objective/train/tokens_used": 493956576,
      "theoretical_loss": 3.4133960279845117,
      "tokens_seen": 2123497472
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.201893757021346e-05,
      "loss": 2.3746,
      "theoretical_loss": 3.4133873478221446,
      "tokens_seen": 2123563008
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.201091317605521e-05,
      "loss": 2.4241,
      "theoretical_loss": 3.41336998852603,
      "tokens_seen": 2123694080
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.200288878189698e-05,
      "loss": 2.6194,
      "theoretical_loss": 3.4133526306012483,
      "tokens_seen": 2123825152
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.199486438773873e-05,
      "loss": 2.455,
      "theoretical_loss": 3.4133352740476064,
      "tokens_seen": 2123956224
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.19868399935805e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.413317918864912,
      "tokens_seen": 2124087296
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.197881559942225e-05,
      "loss": 2.5824,
      "theoretical_loss": 3.4133005650529715,
      "tokens_seen": 2124218368
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.197079120526402e-05,
      "loss": 2.3911,
      "theoretical_loss": 3.413283212611592,
      "tokens_seen": 2124349440
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.196276681110577e-05,
      "loss": 2.5731,
      "theoretical_loss": 3.413265861540582,
      "tokens_seen": 2124480512
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.195474241694754e-05,
      "loss": 2.4229,
      "theoretical_loss": 3.413248511839747,
      "tokens_seen": 2124611584
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.194671802278929e-05,
      "loss": 2.4563,
      "theoretical_loss": 3.413231163508895,
      "tokens_seen": 2124742656
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.193869362863104e-05,
      "loss": 2.3896,
      "theoretical_loss": 3.4132138165478336,
      "tokens_seen": 2124873728
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.193066923447281e-05,
      "loss": 2.5088,
      "theoretical_loss": 3.41319647095637,
      "tokens_seen": 2125004800
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1171472,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.178572177886963,
      "objective/train/theoretical_loss": 3.4131791267343115,
      "objective/train/tokens_used": 495594976,
      "theoretical_loss": 3.4131791267343115,
      "tokens_seen": 2125135872
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.192264484031456e-05,
      "loss": 2.6308,
      "theoretical_loss": 3.4131791267343115,
      "tokens_seen": 2125135872
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.191462044615633e-05,
      "loss": 2.7299,
      "theoretical_loss": 3.413161783881465,
      "tokens_seen": 2125266944
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.190659605199808e-05,
      "loss": 2.5909,
      "theoretical_loss": 3.4131444423976394,
      "tokens_seen": 2125398016
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.189857165783985e-05,
      "loss": 2.5526,
      "theoretical_loss": 3.4131271022826413,
      "tokens_seen": 2125529088
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.18905472636816e-05,
      "loss": 2.5848,
      "theoretical_loss": 3.4131097635362777,
      "tokens_seen": 2125660160
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.188252286952335e-05,
      "loss": 2.4472,
      "theoretical_loss": 3.4130924261583573,
      "tokens_seen": 2125791232
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.187449847536512e-05,
      "loss": 2.3733,
      "theoretical_loss": 3.413075090148687,
      "tokens_seen": 2125922304
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.186647408120687e-05,
      "loss": 2.59,
      "theoretical_loss": 3.413057755507075,
      "tokens_seen": 2126053376
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.185844968704864e-05,
      "loss": 2.6016,
      "theoretical_loss": 3.413040422233328,
      "tokens_seen": 2126184448
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.18504252928904e-05,
      "loss": 2.4636,
      "theoretical_loss": 3.413023090327255,
      "tokens_seen": 2126315520
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.184240089873216e-05,
      "loss": 2.5986,
      "theoretical_loss": 3.413005759788663,
      "tokens_seen": 2126446592
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.183437650457391e-05,
      "loss": 2.439,
      "theoretical_loss": 3.412988430617361,
      "tokens_seen": 2126577664
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.182635211041567e-05,
      "loss": 2.4675,
      "theoretical_loss": 3.4129711028131546,
      "tokens_seen": 2126708736
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1172068,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.215710163116455,
      "objective/train/theoretical_loss": 3.4129624394236533,
      "objective/train/tokens_used": 497233376,
      "theoretical_loss": 3.4129624394236533,
      "tokens_seen": 2126774272
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.181832771625743e-05,
      "loss": 2.4494,
      "theoretical_loss": 3.4129537763758537,
      "tokens_seen": 2126839808
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.181030332209919e-05,
      "loss": 2.4569,
      "theoretical_loss": 3.4129364513052654,
      "tokens_seen": 2126970880
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.180227892794095e-05,
      "loss": 2.6965,
      "theoretical_loss": 3.4129191276011985,
      "tokens_seen": 2127101952
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.17942545337827e-05,
      "loss": 2.5618,
      "theoretical_loss": 3.4129018052634597,
      "tokens_seen": 2127233024
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.178623013962446e-05,
      "loss": 2.5163,
      "theoretical_loss": 3.412884484291858,
      "tokens_seen": 2127364096
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.177820574546623e-05,
      "loss": 2.3182,
      "theoretical_loss": 3.412867164686201,
      "tokens_seen": 2127495168
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.177018135130798e-05,
      "loss": 2.6376,
      "theoretical_loss": 3.4128498464462975,
      "tokens_seen": 2127626240
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.176215695714975e-05,
      "loss": 2.5248,
      "theoretical_loss": 3.412832529571955,
      "tokens_seen": 2127757312
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.17541325629915e-05,
      "loss": 2.6077,
      "theoretical_loss": 3.4128152140629817,
      "tokens_seen": 2127888384
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.174610816883327e-05,
      "loss": 2.599,
      "theoretical_loss": 3.4127978999191866,
      "tokens_seen": 2128019456
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.173808377467502e-05,
      "loss": 2.4737,
      "theoretical_loss": 3.412780587140378,
      "tokens_seen": 2128150528
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.173005938051677e-05,
      "loss": 2.668,
      "theoretical_loss": 3.4127632757263626,
      "tokens_seen": 2128281600
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1173293,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4033796787261963,
      "objective/train/theoretical_loss": 3.4127459656769505,
      "objective/train/tokens_used": 498871776,
      "theoretical_loss": 3.4127459656769505,
      "tokens_seen": 2128412672
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.172203498635854e-05,
      "loss": 2.4953,
      "theoretical_loss": 3.4127459656769505,
      "tokens_seen": 2128412672
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.171401059220029e-05,
      "loss": 2.5535,
      "theoretical_loss": 3.4127286569919493,
      "tokens_seen": 2128543744
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.170598619804206e-05,
      "loss": 2.6874,
      "theoretical_loss": 3.412711349671168,
      "tokens_seen": 2128674816
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.169796180388381e-05,
      "loss": 2.7642,
      "theoretical_loss": 3.4126940437144144,
      "tokens_seen": 2128805888
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.168993740972557e-05,
      "loss": 2.5645,
      "theoretical_loss": 3.4126767391214976,
      "tokens_seen": 2128936960
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.168191301556733e-05,
      "loss": 2.432,
      "theoretical_loss": 3.412659435892226,
      "tokens_seen": 2129068032
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.167388862140908e-05,
      "loss": 2.6999,
      "theoretical_loss": 3.412642134026408,
      "tokens_seen": 2129199104
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.166586422725085e-05,
      "loss": 2.7949,
      "theoretical_loss": 3.4126248335238527,
      "tokens_seen": 2129330176
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.16578398330926e-05,
      "loss": 2.5417,
      "theoretical_loss": 3.412607534384368,
      "tokens_seen": 2129461248
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.164981543893436e-05,
      "loss": 2.44,
      "theoretical_loss": 3.4125902366077634,
      "tokens_seen": 2129592320
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.164179104477612e-05,
      "loss": 2.6147,
      "theoretical_loss": 3.412572940193847,
      "tokens_seen": 2129723392
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.163376665061788e-05,
      "loss": 2.5853,
      "theoretical_loss": 3.412555645142428,
      "tokens_seen": 2129854464
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.162574225645964e-05,
      "loss": 2.5282,
      "theoretical_loss": 3.4125383514533154,
      "tokens_seen": 2129985536
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1173914,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.899099111557007,
      "objective/train/theoretical_loss": 3.412529705119564,
      "objective/train/tokens_used": 500510176,
      "theoretical_loss": 3.412529705119564,
      "tokens_seen": 2130051072
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.16177178623014e-05,
      "loss": 2.7238,
      "theoretical_loss": 3.4125210591263175,
      "tokens_seen": 2130116608
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.160969346814316e-05,
      "loss": 2.476,
      "theoretical_loss": 3.412503768161244,
      "tokens_seen": 2130247680
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.160166907398492e-05,
      "loss": 2.6226,
      "theoretical_loss": 3.412486478557903,
      "tokens_seen": 2130378752
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.159364467982667e-05,
      "loss": 2.5487,
      "theoretical_loss": 3.412469190316104,
      "tokens_seen": 2130509824
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.158562028566844e-05,
      "loss": 2.613,
      "theoretical_loss": 3.412451903435656,
      "tokens_seen": 2130640896
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.157759589151019e-05,
      "loss": 2.5264,
      "theoretical_loss": 3.412434617916368,
      "tokens_seen": 2130771968
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.156957149735196e-05,
      "loss": 2.5246,
      "theoretical_loss": 3.4124173337580492,
      "tokens_seen": 2130903040
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.156154710319371e-05,
      "loss": 2.5873,
      "theoretical_loss": 3.412400050960508,
      "tokens_seen": 2131034112
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.155352270903546e-05,
      "loss": 2.4861,
      "theoretical_loss": 3.4123827695235542,
      "tokens_seen": 2131165184
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.154549831487723e-05,
      "loss": 2.4371,
      "theoretical_loss": 3.412365489446998,
      "tokens_seen": 2131296256
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.153747392071898e-05,
      "loss": 2.6413,
      "theoretical_loss": 3.412348210730647,
      "tokens_seen": 2131427328
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.152944952656075e-05,
      "loss": 2.63,
      "theoretical_loss": 3.412330933374311,
      "tokens_seen": 2131558400
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1174844,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1257615089416504,
      "objective/train/theoretical_loss": 3.4123136573777995,
      "objective/train/tokens_used": 502148576,
      "theoretical_loss": 3.4123136573777995,
      "tokens_seen": 2131689472
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.15214251324025e-05,
      "loss": 2.4612,
      "theoretical_loss": 3.4123136573777995,
      "tokens_seen": 2131689472
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.151340073824427e-05,
      "loss": 2.5941,
      "theoretical_loss": 3.412296382740922,
      "tokens_seen": 2131820544
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.150537634408602e-05,
      "loss": 2.627,
      "theoretical_loss": 3.412279109463488,
      "tokens_seen": 2131951616
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.149735194992778e-05,
      "loss": 2.6469,
      "theoretical_loss": 3.412261837545307,
      "tokens_seen": 2132082688
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.148932755576954e-05,
      "loss": 2.5935,
      "theoretical_loss": 3.412244566986187,
      "tokens_seen": 2132213760
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.14813031616113e-05,
      "loss": 2.6418,
      "theoretical_loss": 3.4122272977859396,
      "tokens_seen": 2132344832
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.147327876745306e-05,
      "loss": 2.3957,
      "theoretical_loss": 3.412210029944374,
      "tokens_seen": 2132475904
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.146525437329482e-05,
      "loss": 2.485,
      "theoretical_loss": 3.4121927634612987,
      "tokens_seen": 2132606976
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.145722997913657e-05,
      "loss": 2.6645,
      "theoretical_loss": 3.4121754983365236,
      "tokens_seen": 2132738048
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.144920558497834e-05,
      "loss": 2.559,
      "theoretical_loss": 3.4121582345698593,
      "tokens_seen": 2132869120
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.144118119082009e-05,
      "loss": 2.4919,
      "theoretical_loss": 3.4121409721611147,
      "tokens_seen": 2133000192
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.143315679666185e-05,
      "loss": 2.6485,
      "theoretical_loss": 3.4121237111101,
      "tokens_seen": 2133131264
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.142513240250361e-05,
      "loss": 2.671,
      "theoretical_loss": 3.4121064514166246,
      "tokens_seen": 2133262336
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1175941,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.085177183151245,
      "objective/train/theoretical_loss": 3.4120978220789047,
      "objective/train/tokens_used": 503786976,
      "theoretical_loss": 3.4120978220789047,
      "tokens_seen": 2133327872
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.141710800834537e-05,
      "loss": 2.5607,
      "theoretical_loss": 3.412089193080498,
      "tokens_seen": 2133393408
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.140908361418713e-05,
      "loss": 2.4615,
      "theoretical_loss": 3.4120719361015315,
      "tokens_seen": 2133524480
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.140105922002888e-05,
      "loss": 2.628,
      "theoretical_loss": 3.4120546804795335,
      "tokens_seen": 2133655552
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.139303482587065e-05,
      "loss": 2.5597,
      "theoretical_loss": 3.412037426214315,
      "tokens_seen": 2133786624
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.13850104317124e-05,
      "loss": 2.605,
      "theoretical_loss": 3.4120201733056854,
      "tokens_seen": 2133917696
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.137698603755417e-05,
      "loss": 2.7064,
      "theoretical_loss": 3.412002921753455,
      "tokens_seen": 2134048768
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.136896164339592e-05,
      "loss": 2.4825,
      "theoretical_loss": 3.4119856715574333,
      "tokens_seen": 2134179840
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.136093724923767e-05,
      "loss": 2.6736,
      "theoretical_loss": 3.411968422717431,
      "tokens_seen": 2134310912
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.135291285507944e-05,
      "loss": 2.6355,
      "theoretical_loss": 3.4119511752332583,
      "tokens_seen": 2134441984
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.13448884609212e-05,
      "loss": 2.6273,
      "theoretical_loss": 3.411933929104725,
      "tokens_seen": 2134573056
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.133686406676296e-05,
      "loss": 2.4734,
      "theoretical_loss": 3.4119166843316413,
      "tokens_seen": 2134704128
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.132883967260471e-05,
      "loss": 2.5344,
      "theoretical_loss": 3.411899440913818,
      "tokens_seen": 2134835200
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1176552,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7227025032043457,
      "objective/train/theoretical_loss": 3.411882198851065,
      "objective/train/tokens_used": 505425376,
      "theoretical_loss": 3.411882198851065,
      "tokens_seen": 2134966272
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.132081527844648e-05,
      "loss": 2.5944,
      "theoretical_loss": 3.411882198851065,
      "tokens_seen": 2134966272
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.131279088428823e-05,
      "loss": 2.5057,
      "theoretical_loss": 3.411864958143192,
      "tokens_seen": 2135097344
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.130476649013e-05,
      "loss": 2.6437,
      "theoretical_loss": 3.4118477187900105,
      "tokens_seen": 2135228416
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.129674209597175e-05,
      "loss": 2.5914,
      "theoretical_loss": 3.41183048079133,
      "tokens_seen": 2135359488
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.12887177018135e-05,
      "loss": 2.6825,
      "theoretical_loss": 3.4118132441469617,
      "tokens_seen": 2135490560
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.128069330765527e-05,
      "loss": 2.5187,
      "theoretical_loss": 3.4117960088567156,
      "tokens_seen": 2135621632
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.127266891349703e-05,
      "loss": 2.599,
      "theoretical_loss": 3.4117787749204025,
      "tokens_seen": 2135752704
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.126464451933879e-05,
      "loss": 2.5976,
      "theoretical_loss": 3.4117615423378327,
      "tokens_seen": 2135883776
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.125662012518055e-05,
      "loss": 2.543,
      "theoretical_loss": 3.4117443111088175,
      "tokens_seen": 2136014848
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.124859573102231e-05,
      "loss": 2.6293,
      "theoretical_loss": 3.411727081233166,
      "tokens_seen": 2136145920
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.124057133686407e-05,
      "loss": 2.4895,
      "theoretical_loss": 3.41170985271069,
      "tokens_seen": 2136276992
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.123254694270582e-05,
      "loss": 2.4808,
      "theoretical_loss": 3.4116926255412006,
      "tokens_seen": 2136408064
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.122452254854759e-05,
      "loss": 2.6091,
      "theoretical_loss": 3.4116753997245075,
      "tokens_seen": 2136539136
    },
    {
      "epoch": 0.29,
      "objective/train/docs_used": 1177438,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7546355724334717,
      "objective/train/theoretical_loss": 3.4116667873234006,
      "objective/train/tokens_used": 507063776,
      "theoretical_loss": 3.4116667873234006,
      "tokens_seen": 2136604672
    },
    {
      "epoch": 0.29,
      "learning_rate": 7.121649815438934e-05,
      "loss": 2.6309,
      "theoretical_loss": 3.4116581752604223,
      "tokens_seen": 2136670208
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.12084737602311e-05,
      "loss": 2.6768,
      "theoretical_loss": 3.411640952148755,
      "tokens_seen": 2136801280
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.120044936607286e-05,
      "loss": 2.6746,
      "theoretical_loss": 3.4116237303893175,
      "tokens_seen": 2136932352
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.119242497191462e-05,
      "loss": 2.5923,
      "theoretical_loss": 3.4116065099819197,
      "tokens_seen": 2137063424
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.118440057775638e-05,
      "loss": 2.6047,
      "theoretical_loss": 3.411589290926374,
      "tokens_seen": 2137194496
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.117637618359813e-05,
      "loss": 2.61,
      "theoretical_loss": 3.411572073222489,
      "tokens_seen": 2137325568
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.11683517894399e-05,
      "loss": 2.686,
      "theoretical_loss": 3.4115548568700778,
      "tokens_seen": 2137456640
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.116032739528165e-05,
      "loss": 2.5895,
      "theoretical_loss": 3.411537641868951,
      "tokens_seen": 2137587712
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.115230300112342e-05,
      "loss": 2.5471,
      "theoretical_loss": 3.411520428218919,
      "tokens_seen": 2137718784
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.114427860696517e-05,
      "loss": 2.399,
      "theoretical_loss": 3.4115032159197938,
      "tokens_seen": 2137849856
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.113625421280694e-05,
      "loss": 2.5753,
      "theoretical_loss": 3.4114860049713855,
      "tokens_seen": 2137980928
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.112822981864869e-05,
      "loss": 2.725,
      "theoretical_loss": 3.4114687953735063,
      "tokens_seen": 2138112000
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1177662,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.830423593521118,
      "objective/train/theoretical_loss": 3.411451587125967,
      "objective/train/tokens_used": 508702176,
      "theoretical_loss": 3.411451587125967,
      "tokens_seen": 2138243072
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.112020542449044e-05,
      "loss": 2.6082,
      "theoretical_loss": 3.411451587125967,
      "tokens_seen": 2138243072
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.111218103033221e-05,
      "loss": 2.5201,
      "theoretical_loss": 3.411434380228579,
      "tokens_seen": 2138374144
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.110415663617396e-05,
      "loss": 2.5066,
      "theoretical_loss": 3.4114171746811537,
      "tokens_seen": 2138505216
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.109613224201573e-05,
      "loss": 2.6115,
      "theoretical_loss": 3.4113999704835023,
      "tokens_seen": 2138636288
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.108810784785748e-05,
      "loss": 2.6089,
      "theoretical_loss": 3.4113827676354362,
      "tokens_seen": 2138767360
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.108008345369925e-05,
      "loss": 2.7135,
      "theoretical_loss": 3.411365566136767,
      "tokens_seen": 2138898432
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.1072059059541e-05,
      "loss": 2.6082,
      "theoretical_loss": 3.4113483659873056,
      "tokens_seen": 2139029504
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.106403466538276e-05,
      "loss": 2.597,
      "theoretical_loss": 3.4113311671868645,
      "tokens_seen": 2139160576
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.105601027122452e-05,
      "loss": 2.7267,
      "theoretical_loss": 3.4113139697352546,
      "tokens_seen": 2139291648
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.104798587706628e-05,
      "loss": 2.533,
      "theoretical_loss": 3.4112967736322872,
      "tokens_seen": 2139422720
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.103996148290804e-05,
      "loss": 2.5938,
      "theoretical_loss": 3.4112795788777746,
      "tokens_seen": 2139553792
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.10319370887498e-05,
      "loss": 2.6091,
      "theoretical_loss": 3.4112623854715283,
      "tokens_seen": 2139684864
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.102391269459156e-05,
      "loss": 2.6935,
      "theoretical_loss": 3.41124519341336,
      "tokens_seen": 2139815936
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1178610,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.592621088027954,
      "objective/train/theoretical_loss": 3.4112365978897463,
      "objective/train/tokens_used": 510340576,
      "theoretical_loss": 3.4112365978897463,
      "tokens_seen": 2139881472
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.101588830043332e-05,
      "loss": 2.612,
      "theoretical_loss": 3.411228002703081,
      "tokens_seen": 2139947008
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.100786390627508e-05,
      "loss": 2.6124,
      "theoretical_loss": 3.4112108133405035,
      "tokens_seen": 2140078080
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.099983951211684e-05,
      "loss": 2.5644,
      "theoretical_loss": 3.4111936253254393,
      "tokens_seen": 2140209152
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.099181511795859e-05,
      "loss": 2.6921,
      "theoretical_loss": 3.4111764386576997,
      "tokens_seen": 2140340224
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.098379072380036e-05,
      "loss": 2.5306,
      "theoretical_loss": 3.4111592533370976,
      "tokens_seen": 2140471296
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.097576632964211e-05,
      "loss": 2.6231,
      "theoretical_loss": 3.4111420693634438,
      "tokens_seen": 2140602368
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.096774193548388e-05,
      "loss": 2.5484,
      "theoretical_loss": 3.411124886736551,
      "tokens_seen": 2140733440
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.095971754132563e-05,
      "loss": 2.5626,
      "theoretical_loss": 3.4111077054562315,
      "tokens_seen": 2140864512
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.09516931471674e-05,
      "loss": 2.5805,
      "theoretical_loss": 3.4110905255222965,
      "tokens_seen": 2140995584
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.094366875300915e-05,
      "loss": 2.5912,
      "theoretical_loss": 3.4110733469345584,
      "tokens_seen": 2141126656
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.09356443588509e-05,
      "loss": 2.5502,
      "theoretical_loss": 3.4110561696928294,
      "tokens_seen": 2141257728
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.092761996469267e-05,
      "loss": 2.6747,
      "theoretical_loss": 3.4110389937969217,
      "tokens_seen": 2141388800
    },
    {
      "debugging/Self-BLEU-5": 0.3562156871264047,
      "debugging/distinct-1-grams": 0.8053620964343855,
      "debugging/distinct-2-grams": 0.9818306010928961,
      "debugging/entropy-1-grams": 5.331752904391725,
      "debugging/entropy-2-grams": 5.892259632920162,
      "debugging/length": 535.8,
      "debugging/num_segments": 5,
      "debugging/score": 0.007634856930105227,
      "debugging/score_std": 0.0023886908693567656,
      "epoch": 0.3,
      "objective/train/docs_used": 1179232,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.219043016433716,
      "objective/train/theoretical_loss": 3.4110218192466473,
      "objective/train/tokens_used": 511978976,
      "theoretical_loss": 3.4110218192466473,
      "tokens_seen": 2141519872
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.091959557053442e-05,
      "loss": 2.6092,
      "theoretical_loss": 3.4110218192466473,
      "tokens_seen": 2141519872
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.091157117637619e-05,
      "loss": 2.7279,
      "theoretical_loss": 3.4110046460418184,
      "tokens_seen": 2141650944
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.090354678221794e-05,
      "loss": 2.6784,
      "theoretical_loss": 3.4109874741822477,
      "tokens_seen": 2141782016
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.089552238805971e-05,
      "loss": 2.6052,
      "theoretical_loss": 3.410970303667747,
      "tokens_seen": 2141913088
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.088749799390146e-05,
      "loss": 2.5963,
      "theoretical_loss": 3.4109531344981283,
      "tokens_seen": 2142044160
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.087947359974321e-05,
      "loss": 2.6833,
      "theoretical_loss": 3.4109359666732053,
      "tokens_seen": 2142175232
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.087144920558498e-05,
      "loss": 2.6276,
      "theoretical_loss": 3.410918800192789,
      "tokens_seen": 2142306304
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.086342481142673e-05,
      "loss": 2.5189,
      "theoretical_loss": 3.4109016350566934,
      "tokens_seen": 2142437376
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.08554004172685e-05,
      "loss": 2.5976,
      "theoretical_loss": 3.410884471264729,
      "tokens_seen": 2142568448
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.084737602311025e-05,
      "loss": 2.5985,
      "theoretical_loss": 3.41086730881671,
      "tokens_seen": 2142699520
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.083935162895202e-05,
      "loss": 2.6729,
      "theoretical_loss": 3.4108501477124484,
      "tokens_seen": 2142830592
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.083132723479377e-05,
      "loss": 2.5663,
      "theoretical_loss": 3.4108329879517565,
      "tokens_seen": 2142961664
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.082330284063553e-05,
      "loss": 2.6607,
      "theoretical_loss": 3.4108158295344473,
      "tokens_seen": 2143092736
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1180509,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.578857421875,
      "objective/train/theoretical_loss": 3.4108072508295026,
      "objective/train/tokens_used": 513617376,
      "theoretical_loss": 3.4108072508295026,
      "tokens_seen": 2143158272
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.081527844647729e-05,
      "loss": 2.563,
      "theoretical_loss": 3.4107986724603334,
      "tokens_seen": 2143223808
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.080725405231905e-05,
      "loss": 2.4593,
      "theoretical_loss": 3.4107815167292275,
      "tokens_seen": 2143354880
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.079922965816081e-05,
      "loss": 2.6285,
      "theoretical_loss": 3.4107643623409425,
      "tokens_seen": 2143485952
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.079120526400257e-05,
      "loss": 2.6234,
      "theoretical_loss": 3.410747209295291,
      "tokens_seen": 2143617024
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.078318086984433e-05,
      "loss": 2.558,
      "theoretical_loss": 3.4107300575920854,
      "tokens_seen": 2143748096
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.077515647568609e-05,
      "loss": 2.4592,
      "theoretical_loss": 3.4107129072311393,
      "tokens_seen": 2143879168
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.076713208152784e-05,
      "loss": 2.6204,
      "theoretical_loss": 3.410695758212266,
      "tokens_seen": 2144010240
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.07591076873696e-05,
      "loss": 2.7057,
      "theoretical_loss": 3.4106786105352773,
      "tokens_seen": 2144141312
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.075108329321136e-05,
      "loss": 2.6211,
      "theoretical_loss": 3.410661464199986,
      "tokens_seen": 2144272384
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.074305889905313e-05,
      "loss": 2.5688,
      "theoretical_loss": 3.410644319206207,
      "tokens_seen": 2144403456
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.073503450489488e-05,
      "loss": 2.6275,
      "theoretical_loss": 3.4106271755537514,
      "tokens_seen": 2144534528
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.072701011073665e-05,
      "loss": 2.5701,
      "theoretical_loss": 3.410610033242433,
      "tokens_seen": 2144665600
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1181762,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5804944038391113,
      "objective/train/theoretical_loss": 3.4105928922720654,
      "objective/train/tokens_used": 515255776,
      "theoretical_loss": 3.4105928922720654,
      "tokens_seen": 2144796672
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.07189857165784e-05,
      "loss": 2.5957,
      "theoretical_loss": 3.4105928922720654,
      "tokens_seen": 2144796672
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.071096132242017e-05,
      "loss": 2.6133,
      "theoretical_loss": 3.410575752642461,
      "tokens_seen": 2144927744
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.070293692826192e-05,
      "loss": 2.5896,
      "theoretical_loss": 3.4105586143534334,
      "tokens_seen": 2145058816
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.069491253410367e-05,
      "loss": 2.4899,
      "theoretical_loss": 3.410541477404796,
      "tokens_seen": 2145189888
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.068688813994544e-05,
      "loss": 2.6221,
      "theoretical_loss": 3.4105243417963615,
      "tokens_seen": 2145320960
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.067886374578719e-05,
      "loss": 2.593,
      "theoretical_loss": 3.4105072075279432,
      "tokens_seen": 2145452032
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.067083935162896e-05,
      "loss": 2.6936,
      "theoretical_loss": 3.410490074599356,
      "tokens_seen": 2145583104
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.066281495747071e-05,
      "loss": 2.5297,
      "theoretical_loss": 3.4104729430104115,
      "tokens_seen": 2145714176
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.065479056331248e-05,
      "loss": 2.6632,
      "theoretical_loss": 3.4104558127609232,
      "tokens_seen": 2145845248
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.064676616915423e-05,
      "loss": 2.6771,
      "theoretical_loss": 3.410438683850706,
      "tokens_seen": 2145976320
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.063874177499598e-05,
      "loss": 2.667,
      "theoretical_loss": 3.410421556279572,
      "tokens_seen": 2146107392
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.063071738083775e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.4104044300473353,
      "tokens_seen": 2146238464
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.06226929866795e-05,
      "loss": 2.626,
      "theoretical_loss": 3.4103873051538094,
      "tokens_seen": 2146369536
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1182263,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.857555866241455,
      "objective/train/theoretical_loss": 3.410378743209005,
      "objective/train/tokens_used": 516894176,
      "theoretical_loss": 3.410378743209005,
      "tokens_seen": 2146435072
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.061466859252127e-05,
      "loss": 2.6154,
      "theoretical_loss": 3.4103701815988083,
      "tokens_seen": 2146500608
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.060664419836302e-05,
      "loss": 2.4926,
      "theoretical_loss": 3.4103530593821447,
      "tokens_seen": 2146631680
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.059861980420479e-05,
      "loss": 2.7829,
      "theoretical_loss": 3.4103359385036334,
      "tokens_seen": 2146762752
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.059059541004654e-05,
      "loss": 2.5032,
      "theoretical_loss": 3.4103188189630878,
      "tokens_seen": 2146893824
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.05825710158883e-05,
      "loss": 2.5789,
      "theoretical_loss": 3.4103017007603205,
      "tokens_seen": 2147024896
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.057454662173006e-05,
      "loss": 2.5056,
      "theoretical_loss": 3.410284583895147,
      "tokens_seen": 2147155968
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.056652222757182e-05,
      "loss": 2.5706,
      "theoretical_loss": 3.4102674683673806,
      "tokens_seen": 2147287040
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.055849783341358e-05,
      "loss": 2.6252,
      "theoretical_loss": 3.4102503541768345,
      "tokens_seen": 2147418112
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.055047343925534e-05,
      "loss": 2.5292,
      "theoretical_loss": 3.4102332413233234,
      "tokens_seen": 2147549184
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.05424490450971e-05,
      "loss": 2.6922,
      "theoretical_loss": 3.4102161298066607,
      "tokens_seen": 2147680256
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.053442465093886e-05,
      "loss": 2.6755,
      "theoretical_loss": 3.410199019626661,
      "tokens_seen": 2147811328
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.052640025678061e-05,
      "loss": 2.5742,
      "theoretical_loss": 3.410181910783138,
      "tokens_seen": 2147942400
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1183255,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4469528198242188,
      "objective/train/theoretical_loss": 3.4101648032759053,
      "objective/train/tokens_used": 518532576,
      "theoretical_loss": 3.4101648032759053,
      "tokens_seen": 2148073472
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.051837586262238e-05,
      "loss": 2.4785,
      "theoretical_loss": 3.4101648032759053,
      "tokens_seen": 2148073472
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.051035146846413e-05,
      "loss": 2.6102,
      "theoretical_loss": 3.4101476971047777,
      "tokens_seen": 2148204544
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.05023270743059e-05,
      "loss": 2.5222,
      "theoretical_loss": 3.410130592269569,
      "tokens_seen": 2148335616
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.049430268014765e-05,
      "loss": 2.665,
      "theoretical_loss": 3.4101134887700937,
      "tokens_seen": 2148466688
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.048627828598942e-05,
      "loss": 2.6828,
      "theoretical_loss": 3.410096386606166,
      "tokens_seen": 2148597760
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.047825389183117e-05,
      "loss": 2.5421,
      "theoretical_loss": 3.4100792857775994,
      "tokens_seen": 2148728832
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.047022949767292e-05,
      "loss": 2.6022,
      "theoretical_loss": 3.4100621862842084,
      "tokens_seen": 2148859904
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.046220510351469e-05,
      "loss": 2.6454,
      "theoretical_loss": 3.410045088125808,
      "tokens_seen": 2148990976
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.045418070935644e-05,
      "loss": 2.5891,
      "theoretical_loss": 3.4100279913022122,
      "tokens_seen": 2149122048
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.044615631519821e-05,
      "loss": 2.6476,
      "theoretical_loss": 3.410010895813236,
      "tokens_seen": 2149253120
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.043813192103996e-05,
      "loss": 2.4586,
      "theoretical_loss": 3.4099938016586924,
      "tokens_seen": 2149384192
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.043010752688173e-05,
      "loss": 2.6862,
      "theoretical_loss": 3.4099767088383968,
      "tokens_seen": 2149515264
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.042208313272348e-05,
      "loss": 2.5898,
      "theoretical_loss": 3.409959617352164,
      "tokens_seen": 2149646336
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1183895,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.68380069732666,
      "objective/train/theoretical_loss": 3.4099510721092625,
      "objective/train/tokens_used": 520170976,
      "theoretical_loss": 3.4099510721092625,
      "tokens_seen": 2149711872
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.041405873856523e-05,
      "loss": 2.6714,
      "theoretical_loss": 3.4099425271998074,
      "tokens_seen": 2149777408
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.0406034344407e-05,
      "loss": 2.5034,
      "theoretical_loss": 3.409925438381143,
      "tokens_seen": 2149908480
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.039800995024875e-05,
      "loss": 2.5538,
      "theoretical_loss": 3.409908350895984,
      "tokens_seen": 2150039552
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.038998555609052e-05,
      "loss": 2.4173,
      "theoretical_loss": 3.4098912647441466,
      "tokens_seen": 2150170624
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.038196116193227e-05,
      "loss": 2.7148,
      "theoretical_loss": 3.4098741799254446,
      "tokens_seen": 2150301696
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.037393676777404e-05,
      "loss": 2.4907,
      "theoretical_loss": 3.4098570964396924,
      "tokens_seen": 2150432768
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.03659123736158e-05,
      "loss": 2.6022,
      "theoretical_loss": 3.409840014286705,
      "tokens_seen": 2150563840
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.035788797945756e-05,
      "loss": 2.487,
      "theoretical_loss": 3.409822933466298,
      "tokens_seen": 2150694912
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.034986358529931e-05,
      "loss": 2.4843,
      "theoretical_loss": 3.4098058539782854,
      "tokens_seen": 2150825984
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.034183919114107e-05,
      "loss": 2.5202,
      "theoretical_loss": 3.4097887758224825,
      "tokens_seen": 2150957056
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.033381479698283e-05,
      "loss": 2.6072,
      "theoretical_loss": 3.4097716989987035,
      "tokens_seen": 2151088128
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.032579040282459e-05,
      "loss": 2.613,
      "theoretical_loss": 3.4097546235067644,
      "tokens_seen": 2151219200
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1184816,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.916866660118103,
      "objective/train/theoretical_loss": 3.409737549346479,
      "objective/train/tokens_used": 521809376,
      "theoretical_loss": 3.409737549346479,
      "tokens_seen": 2151350272
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.031776600866635e-05,
      "loss": 2.57,
      "theoretical_loss": 3.409737549346479,
      "tokens_seen": 2151350272
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.03097416145081e-05,
      "loss": 2.4954,
      "theoretical_loss": 3.409720476517664,
      "tokens_seen": 2151481344
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.030171722034987e-05,
      "loss": 2.6051,
      "theoretical_loss": 3.4097034050201334,
      "tokens_seen": 2151612416
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.029369282619163e-05,
      "loss": 2.4134,
      "theoretical_loss": 3.4096863348537014,
      "tokens_seen": 2151743488
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.028566843203338e-05,
      "loss": 2.695,
      "theoretical_loss": 3.409669266018185,
      "tokens_seen": 2151874560
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.027764403787515e-05,
      "loss": 2.5432,
      "theoretical_loss": 3.409652198513398,
      "tokens_seen": 2152005632
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.02696196437169e-05,
      "loss": 2.6862,
      "theoretical_loss": 3.4096351323391563,
      "tokens_seen": 2152136704
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.026159524955867e-05,
      "loss": 2.6014,
      "theoretical_loss": 3.4096180674952747,
      "tokens_seen": 2152267776
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.025357085540042e-05,
      "loss": 2.6281,
      "theoretical_loss": 3.409601003981569,
      "tokens_seen": 2152398848
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.024554646124219e-05,
      "loss": 2.7042,
      "theoretical_loss": 3.4095839417978544,
      "tokens_seen": 2152529920
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.023752206708394e-05,
      "loss": 2.5141,
      "theoretical_loss": 3.409566880943945,
      "tokens_seen": 2152660992
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.022949767292569e-05,
      "loss": 2.5674,
      "theoretical_loss": 3.4095498214196587,
      "tokens_seen": 2152792064
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.022147327876746e-05,
      "loss": 2.5695,
      "theoretical_loss": 3.409532763224809,
      "tokens_seen": 2152923136
    },
    {
      "epoch": 0.3,
      "objective/train/docs_used": 1185361,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4470932483673096,
      "objective/train/theoretical_loss": 3.4095242346258656,
      "objective/train/tokens_used": 523447776,
      "theoretical_loss": 3.4095242346258656,
      "tokens_seen": 2152988672
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.021344888460921e-05,
      "loss": 2.6913,
      "theoretical_loss": 3.409515706359212,
      "tokens_seen": 2153054208
    },
    {
      "epoch": 0.3,
      "learning_rate": 7.020542449045098e-05,
      "loss": 2.7435,
      "theoretical_loss": 3.409498650822683,
      "tokens_seen": 2153185280
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.019740009629273e-05,
      "loss": 2.4549,
      "theoretical_loss": 3.409481596615038,
      "tokens_seen": 2153316352
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.01893757021345e-05,
      "loss": 2.5406,
      "theoretical_loss": 3.4094645437360915,
      "tokens_seen": 2153447424
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.018135130797625e-05,
      "loss": 2.6382,
      "theoretical_loss": 3.4094474921856603,
      "tokens_seen": 2153578496
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.0173326913818e-05,
      "loss": 2.6475,
      "theoretical_loss": 3.4094304419635595,
      "tokens_seen": 2153709568
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.016530251965977e-05,
      "loss": 2.8102,
      "theoretical_loss": 3.409413393069605,
      "tokens_seen": 2153840640
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.015727812550152e-05,
      "loss": 2.6188,
      "theoretical_loss": 3.4093963455036125,
      "tokens_seen": 2153971712
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.014925373134329e-05,
      "loss": 2.609,
      "theoretical_loss": 3.4093792992653977,
      "tokens_seen": 2154102784
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.014122933718504e-05,
      "loss": 2.7287,
      "theoretical_loss": 3.409362254354776,
      "tokens_seen": 2154233856
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.013320494302681e-05,
      "loss": 2.5641,
      "theoretical_loss": 3.409345210771564,
      "tokens_seen": 2154364928
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.012518054886856e-05,
      "loss": 2.5944,
      "theoretical_loss": 3.409328168515577,
      "tokens_seen": 2154496000
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1186484,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.775474786758423,
      "objective/train/theoretical_loss": 3.4093111275866312,
      "objective/train/tokens_used": 525086176,
      "theoretical_loss": 3.4093111275866312,
      "tokens_seen": 2154627072
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.011715615471032e-05,
      "loss": 2.6251,
      "theoretical_loss": 3.4093111275866312,
      "tokens_seen": 2154627072
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.010913176055208e-05,
      "loss": 2.663,
      "theoretical_loss": 3.4092940879845424,
      "tokens_seen": 2154758144
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.010110736639384e-05,
      "loss": 2.6242,
      "theoretical_loss": 3.409277049709127,
      "tokens_seen": 2154889216
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.00930829722356e-05,
      "loss": 2.548,
      "theoretical_loss": 3.4092600127602006,
      "tokens_seen": 2155020288
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.008505857807736e-05,
      "loss": 2.61,
      "theoretical_loss": 3.4092429771375787,
      "tokens_seen": 2155151360
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.007703418391912e-05,
      "loss": 2.5879,
      "theoretical_loss": 3.4092259428410783,
      "tokens_seen": 2155282432
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.006900978976088e-05,
      "loss": 2.5991,
      "theoretical_loss": 3.409208909870516,
      "tokens_seen": 2155413504
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.006098539560264e-05,
      "loss": 2.5815,
      "theoretical_loss": 3.4091918782257062,
      "tokens_seen": 2155544576
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.00529610014444e-05,
      "loss": 2.6043,
      "theoretical_loss": 3.409174847906467,
      "tokens_seen": 2155675648
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.004493660728615e-05,
      "loss": 2.6625,
      "theoretical_loss": 3.409157818912613,
      "tokens_seen": 2155806720
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.003691221312792e-05,
      "loss": 2.7572,
      "theoretical_loss": 3.4091407912439617,
      "tokens_seen": 2155937792
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.002888781896967e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.409123764900329,
      "tokens_seen": 2156068864
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.002086342481144e-05,
      "loss": 2.6222,
      "theoretical_loss": 3.409106739881531,
      "tokens_seen": 2156199936
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1187201,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.76190185546875,
      "objective/train/theoretical_loss": 3.4090982278688875,
      "objective/train/tokens_used": 526724576,
      "theoretical_loss": 3.4090982278688875,
      "tokens_seen": 2156265472
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.001283903065319e-05,
      "loss": 2.651,
      "theoretical_loss": 3.4090897161873848,
      "tokens_seen": 2156331008
    },
    {
      "epoch": 0.31,
      "learning_rate": 7.000481463649496e-05,
      "loss": 2.5518,
      "theoretical_loss": 3.4090726938177056,
      "tokens_seen": 2156462080
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.999679024233671e-05,
      "loss": 2.4781,
      "theoretical_loss": 3.409055672772311,
      "tokens_seen": 2156593152
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.998876584817846e-05,
      "loss": 2.6187,
      "theoretical_loss": 3.409038653051017,
      "tokens_seen": 2156724224
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.998074145402023e-05,
      "loss": 2.6385,
      "theoretical_loss": 3.40902163465364,
      "tokens_seen": 2156855296
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.997271705986198e-05,
      "loss": 2.6592,
      "theoretical_loss": 3.4090046175799973,
      "tokens_seen": 2156986368
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.996469266570375e-05,
      "loss": 2.6731,
      "theoretical_loss": 3.408987601829905,
      "tokens_seen": 2157117440
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.99566682715455e-05,
      "loss": 2.6475,
      "theoretical_loss": 3.4089705874031795,
      "tokens_seen": 2157248512
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.994864387738727e-05,
      "loss": 2.436,
      "theoretical_loss": 3.4089535742996375,
      "tokens_seen": 2157379584
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.994061948322902e-05,
      "loss": 2.6633,
      "theoretical_loss": 3.408936562519096,
      "tokens_seen": 2157510656
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.993259508907077e-05,
      "loss": 2.6681,
      "theoretical_loss": 3.4089195520613718,
      "tokens_seen": 2157641728
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.992457069491254e-05,
      "loss": 2.573,
      "theoretical_loss": 3.4089025429262816,
      "tokens_seen": 2157772800
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1188127,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6738452911376953,
      "objective/train/theoretical_loss": 3.4088855351136425,
      "objective/train/tokens_used": 528362976,
      "theoretical_loss": 3.4088855351136425,
      "tokens_seen": 2157903872
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.99165463007543e-05,
      "loss": 2.5564,
      "theoretical_loss": 3.4088855351136425,
      "tokens_seen": 2157903872
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.990852190659606e-05,
      "loss": 2.6039,
      "theoretical_loss": 3.4088685286232705,
      "tokens_seen": 2158034944
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.990049751243781e-05,
      "loss": 2.5855,
      "theoretical_loss": 3.4088515234549837,
      "tokens_seen": 2158166016
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.989247311827958e-05,
      "loss": 2.6119,
      "theoretical_loss": 3.408834519608598,
      "tokens_seen": 2158297088
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.988444872412133e-05,
      "loss": 2.5422,
      "theoretical_loss": 3.408817517083931,
      "tokens_seen": 2158428160
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.987642432996309e-05,
      "loss": 2.6445,
      "theoretical_loss": 3.4088005158807992,
      "tokens_seen": 2158559232
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.986839993580485e-05,
      "loss": 2.6504,
      "theoretical_loss": 3.40878351599902,
      "tokens_seen": 2158690304
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.98603755416466e-05,
      "loss": 2.6189,
      "theoretical_loss": 3.408766517438411,
      "tokens_seen": 2158821376
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.985235114748837e-05,
      "loss": 2.5501,
      "theoretical_loss": 3.4087495201987883,
      "tokens_seen": 2158952448
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.984432675333013e-05,
      "loss": 2.4872,
      "theoretical_loss": 3.4087325242799693,
      "tokens_seen": 2159083520
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.98363023591719e-05,
      "loss": 2.5379,
      "theoretical_loss": 3.4087155296817717,
      "tokens_seen": 2159214592
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.982827796501365e-05,
      "loss": 2.4032,
      "theoretical_loss": 3.4086985364040125,
      "tokens_seen": 2159345664
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.98202535708554e-05,
      "loss": 2.6065,
      "theoretical_loss": 3.408681544446509,
      "tokens_seen": 2159476736
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1188847,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7533798217773438,
      "objective/train/theoretical_loss": 3.4086730489627954,
      "objective/train/tokens_used": 530001376,
      "theoretical_loss": 3.4086730489627954,
      "tokens_seen": 2159542272
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.981222917669717e-05,
      "loss": 2.5673,
      "theoretical_loss": 3.4086645538090776,
      "tokens_seen": 2159607808
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.980420478253892e-05,
      "loss": 2.5625,
      "theoretical_loss": 3.4086475644915373,
      "tokens_seen": 2159738880
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.979618038838069e-05,
      "loss": 2.5184,
      "theoretical_loss": 3.408630576493704,
      "tokens_seen": 2159869952
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.978815599422244e-05,
      "loss": 2.6005,
      "theoretical_loss": 3.408613589815396,
      "tokens_seen": 2160001024
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.97801316000642e-05,
      "loss": 2.4839,
      "theoretical_loss": 3.4085966044564304,
      "tokens_seen": 2160132096
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.977210720590596e-05,
      "loss": 2.5711,
      "theoretical_loss": 3.4085796204166243,
      "tokens_seen": 2160263168
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.976408281174773e-05,
      "loss": 2.6037,
      "theoretical_loss": 3.4085626376957965,
      "tokens_seen": 2160394240
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.975605841758948e-05,
      "loss": 2.5763,
      "theoretical_loss": 3.408545656293763,
      "tokens_seen": 2160525312
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.974803402343123e-05,
      "loss": 2.6665,
      "theoretical_loss": 3.4085286762103424,
      "tokens_seen": 2160656384
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.9740009629273e-05,
      "loss": 2.4983,
      "theoretical_loss": 3.408511697445352,
      "tokens_seen": 2160787456
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.973198523511475e-05,
      "loss": 2.5276,
      "theoretical_loss": 3.4084947199986093,
      "tokens_seen": 2160918528
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.972396084095652e-05,
      "loss": 2.6489,
      "theoretical_loss": 3.4084777438699323,
      "tokens_seen": 2161049600
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1190045,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.518320083618164,
      "objective/train/theoretical_loss": 3.408460769059139,
      "objective/train/tokens_used": 531639776,
      "theoretical_loss": 3.408460769059139,
      "tokens_seen": 2161180672
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.971593644679827e-05,
      "loss": 2.538,
      "theoretical_loss": 3.408460769059139,
      "tokens_seen": 2161180672
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.970791205264004e-05,
      "loss": 2.5788,
      "theoretical_loss": 3.408443795566046,
      "tokens_seen": 2161311744
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.969988765848179e-05,
      "loss": 2.6509,
      "theoretical_loss": 3.4084268233904726,
      "tokens_seen": 2161442816
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.969186326432354e-05,
      "loss": 2.578,
      "theoretical_loss": 3.4084098525322357,
      "tokens_seen": 2161573888
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.968383887016531e-05,
      "loss": 2.4955,
      "theoretical_loss": 3.408392882991153,
      "tokens_seen": 2161704960
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.967581447600706e-05,
      "loss": 2.6306,
      "theoretical_loss": 3.408375914767044,
      "tokens_seen": 2161836032
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.966779008184883e-05,
      "loss": 2.6397,
      "theoretical_loss": 3.4083589478597247,
      "tokens_seen": 2161967104
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.965976568769058e-05,
      "loss": 2.5939,
      "theoretical_loss": 3.4083419822690137,
      "tokens_seen": 2162098176
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.965174129353235e-05,
      "loss": 2.6069,
      "theoretical_loss": 3.4083250179947293,
      "tokens_seen": 2162229248
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.96437168993741e-05,
      "loss": 2.5248,
      "theoretical_loss": 3.4083080550366898,
      "tokens_seen": 2162360320
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.963569250521586e-05,
      "loss": 2.6181,
      "theoretical_loss": 3.4082910933947126,
      "tokens_seen": 2162491392
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.962766811105762e-05,
      "loss": 2.5547,
      "theoretical_loss": 3.408274133068617,
      "tokens_seen": 2162622464
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.961964371689938e-05,
      "loss": 2.4597,
      "theoretical_loss": 3.4082571740582197,
      "tokens_seen": 2162753536
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1191405,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.051525115966797,
      "objective/train/theoretical_loss": 3.4082486950463515,
      "objective/train/tokens_used": 533278176,
      "theoretical_loss": 3.4082486950463515,
      "tokens_seen": 2162819072
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.961161932274114e-05,
      "loss": 2.5429,
      "theoretical_loss": 3.4082402163633394,
      "tokens_seen": 2162884608
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.96035949285829e-05,
      "loss": 2.5639,
      "theoretical_loss": 3.408223259983795,
      "tokens_seen": 2163015680
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.959557053442466e-05,
      "loss": 2.5452,
      "theoretical_loss": 3.4082063049194042,
      "tokens_seen": 2163146752
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.958754614026642e-05,
      "loss": 2.5338,
      "theoretical_loss": 3.4081893511699852,
      "tokens_seen": 2163277824
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.957952174610817e-05,
      "loss": 2.605,
      "theoretical_loss": 3.408172398735357,
      "tokens_seen": 2163408896
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.957149735194994e-05,
      "loss": 2.5398,
      "theoretical_loss": 3.4081554476153375,
      "tokens_seen": 2163539968
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.956347295779169e-05,
      "loss": 2.5628,
      "theoretical_loss": 3.4081384978097446,
      "tokens_seen": 2163671040
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.955544856363346e-05,
      "loss": 2.5798,
      "theoretical_loss": 3.408121549318398,
      "tokens_seen": 2163802112
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.954742416947521e-05,
      "loss": 2.5445,
      "theoretical_loss": 3.4081046021411154,
      "tokens_seen": 2163933184
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.953939977531698e-05,
      "loss": 2.5407,
      "theoretical_loss": 3.4080876562777154,
      "tokens_seen": 2164064256
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.953137538115873e-05,
      "loss": 2.5254,
      "theoretical_loss": 3.4080707117280165,
      "tokens_seen": 2164195328
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.952335098700048e-05,
      "loss": 2.5325,
      "theoretical_loss": 3.408053768491838,
      "tokens_seen": 2164326400
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1192132,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1488561630249023,
      "objective/train/theoretical_loss": 3.4080368265689973,
      "objective/train/tokens_used": 534916576,
      "theoretical_loss": 3.4080368265689973,
      "tokens_seen": 2164457472
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.951532659284225e-05,
      "loss": 2.4803,
      "theoretical_loss": 3.4080368265689973,
      "tokens_seen": 2164457472
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.9507302198684e-05,
      "loss": 2.5546,
      "theoretical_loss": 3.4080198859593143,
      "tokens_seen": 2164588544
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.949927780452577e-05,
      "loss": 2.723,
      "theoretical_loss": 3.4080029466626067,
      "tokens_seen": 2164719616
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.949125341036752e-05,
      "loss": 2.4956,
      "theoretical_loss": 3.407986008678694,
      "tokens_seen": 2164850688
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.948322901620929e-05,
      "loss": 2.4595,
      "theoretical_loss": 3.4079690720073943,
      "tokens_seen": 2164981760
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.947520462205104e-05,
      "loss": 2.5846,
      "theoretical_loss": 3.4079521366485275,
      "tokens_seen": 2165112832
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.94671802278928e-05,
      "loss": 2.6315,
      "theoretical_loss": 3.4079352026019114,
      "tokens_seen": 2165243904
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.945915583373456e-05,
      "loss": 2.595,
      "theoretical_loss": 3.4079182698673653,
      "tokens_seen": 2165374976
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.945113143957631e-05,
      "loss": 2.6572,
      "theoretical_loss": 3.407901338444708,
      "tokens_seen": 2165506048
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.944310704541808e-05,
      "loss": 2.4416,
      "theoretical_loss": 3.407884408333759,
      "tokens_seen": 2165637120
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.943508265125983e-05,
      "loss": 2.4823,
      "theoretical_loss": 3.407867479534336,
      "tokens_seen": 2165768192
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.94270582571016e-05,
      "loss": 2.4664,
      "theoretical_loss": 3.40785055204626,
      "tokens_seen": 2165899264
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.941903386294335e-05,
      "loss": 2.5716,
      "theoretical_loss": 3.4078336258693485,
      "tokens_seen": 2166030336
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1193548,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6069717407226562,
      "objective/train/theoretical_loss": 3.407825163272523,
      "objective/train/tokens_used": 536554976,
      "theoretical_loss": 3.407825163272523,
      "tokens_seen": 2166095872
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.941100946878512e-05,
      "loss": 2.4828,
      "theoretical_loss": 3.407816701003421,
      "tokens_seen": 2166161408
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.940298507462687e-05,
      "loss": 2.5359,
      "theoretical_loss": 3.4077997774482967,
      "tokens_seen": 2166292480
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.939496068046863e-05,
      "loss": 2.4822,
      "theoretical_loss": 3.407782855203795,
      "tokens_seen": 2166423552
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.93869362863104e-05,
      "loss": 2.5632,
      "theoretical_loss": 3.4077659342697344,
      "tokens_seen": 2166554624
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.937891189215215e-05,
      "loss": 2.41,
      "theoretical_loss": 3.407749014645935,
      "tokens_seen": 2166685696
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.937088749799391e-05,
      "loss": 2.5752,
      "theoretical_loss": 3.407732096332216,
      "tokens_seen": 2166816768
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.936286310383567e-05,
      "loss": 2.6156,
      "theoretical_loss": 3.407715179328396,
      "tokens_seen": 2166947840
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.935483870967743e-05,
      "loss": 2.512,
      "theoretical_loss": 3.4076982636342956,
      "tokens_seen": 2167078912
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.934681431551919e-05,
      "loss": 2.2435,
      "theoretical_loss": 3.407681349249733,
      "tokens_seen": 2167209984
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.933878992136094e-05,
      "loss": 2.4444,
      "theoretical_loss": 3.4076644361745276,
      "tokens_seen": 2167341056
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.93307655272027e-05,
      "loss": 2.6329,
      "theoretical_loss": 3.4076475244084996,
      "tokens_seen": 2167472128
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.932274113304446e-05,
      "loss": 2.535,
      "theoretical_loss": 3.4076306139514685,
      "tokens_seen": 2167603200
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1194194,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.670583724975586,
      "objective/train/theoretical_loss": 3.4076137048032535,
      "objective/train/tokens_used": 538193376,
      "theoretical_loss": 3.4076137048032535,
      "tokens_seen": 2167734272
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.931471673888623e-05,
      "loss": 2.5218,
      "theoretical_loss": 3.4076137048032535,
      "tokens_seen": 2167734272
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.930669234472798e-05,
      "loss": 2.4784,
      "theoretical_loss": 3.4075967969636745,
      "tokens_seen": 2167865344
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.929866795056975e-05,
      "loss": 2.3791,
      "theoretical_loss": 3.40757989043255,
      "tokens_seen": 2167996416
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.92906435564115e-05,
      "loss": 2.5507,
      "theoretical_loss": 3.4075629852097014,
      "tokens_seen": 2168127488
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.928261916225325e-05,
      "loss": 2.6381,
      "theoretical_loss": 3.4075460812949467,
      "tokens_seen": 2168258560
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.927459476809502e-05,
      "loss": 2.5207,
      "theoretical_loss": 3.4075291786881072,
      "tokens_seen": 2168389632
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.926657037393677e-05,
      "loss": 2.4885,
      "theoretical_loss": 3.407512277389001,
      "tokens_seen": 2168520704
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.925854597977854e-05,
      "loss": 2.5377,
      "theoretical_loss": 3.4074953773974492,
      "tokens_seen": 2168651776
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.925052158562029e-05,
      "loss": 2.3995,
      "theoretical_loss": 3.4074784787132715,
      "tokens_seen": 2168782848
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.924249719146206e-05,
      "loss": 2.4598,
      "theoretical_loss": 3.407461581336287,
      "tokens_seen": 2168913920
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.923447279730381e-05,
      "loss": 2.4744,
      "theoretical_loss": 3.407444685266316,
      "tokens_seen": 2169044992
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.922644840314556e-05,
      "loss": 2.4483,
      "theoretical_loss": 3.4074277905031782,
      "tokens_seen": 2169176064
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.921842400898733e-05,
      "loss": 2.4571,
      "theoretical_loss": 3.4074108970466943,
      "tokens_seen": 2169307136
    },
    {
      "epoch": 0.31,
      "objective/train/docs_used": 1195647,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.518240451812744,
      "objective/train/theoretical_loss": 3.407402450808391,
      "objective/train/tokens_used": 539831776,
      "theoretical_loss": 3.407402450808391,
      "tokens_seen": 2169372672
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.921039961482908e-05,
      "loss": 2.5114,
      "theoretical_loss": 3.4073940048966835,
      "tokens_seen": 2169438208
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.920237522067085e-05,
      "loss": 2.4289,
      "theoretical_loss": 3.4073771140529665,
      "tokens_seen": 2169569280
    },
    {
      "epoch": 0.31,
      "learning_rate": 6.91943508265126e-05,
      "loss": 2.5673,
      "theoretical_loss": 3.4073602245153625,
      "tokens_seen": 2169700352
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.918632643235437e-05,
      "loss": 2.6631,
      "theoretical_loss": 3.4073433362836925,
      "tokens_seen": 2169831424
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.917830203819612e-05,
      "loss": 2.4073,
      "theoretical_loss": 3.407326449357776,
      "tokens_seen": 2169962496
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.917027764403788e-05,
      "loss": 2.6058,
      "theoretical_loss": 3.4073095637374338,
      "tokens_seen": 2170093568
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.916225324987964e-05,
      "loss": 2.6248,
      "theoretical_loss": 3.407292679422486,
      "tokens_seen": 2170224640
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.91542288557214e-05,
      "loss": 2.614,
      "theoretical_loss": 3.407275796412752,
      "tokens_seen": 2170355712
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.914620446156316e-05,
      "loss": 2.5169,
      "theoretical_loss": 3.407258914708053,
      "tokens_seen": 2170486784
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.913818006740492e-05,
      "loss": 2.6178,
      "theoretical_loss": 3.407242034308209,
      "tokens_seen": 2170617856
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.913015567324668e-05,
      "loss": 2.3758,
      "theoretical_loss": 3.40722515521304,
      "tokens_seen": 2170748928
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.912213127908844e-05,
      "loss": 2.5497,
      "theoretical_loss": 3.4072082774223675,
      "tokens_seen": 2170880000
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1196267,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4913604259490967,
      "objective/train/theoretical_loss": 3.4071914009360107,
      "objective/train/tokens_used": 541470176,
      "theoretical_loss": 3.4071914009360107,
      "tokens_seen": 2171011072
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.911410688493019e-05,
      "loss": 2.4815,
      "theoretical_loss": 3.4071914009360107,
      "tokens_seen": 2171011072
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.910608249077196e-05,
      "loss": 2.5382,
      "theoretical_loss": 3.4071745257537907,
      "tokens_seen": 2171142144
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.909805809661371e-05,
      "loss": 2.5229,
      "theoretical_loss": 3.407157651875528,
      "tokens_seen": 2171273216
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.909003370245548e-05,
      "loss": 2.3727,
      "theoretical_loss": 3.407140779301043,
      "tokens_seen": 2171404288
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.908200930829723e-05,
      "loss": 2.4773,
      "theoretical_loss": 3.4071239080301563,
      "tokens_seen": 2171535360
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.907398491413898e-05,
      "loss": 2.5604,
      "theoretical_loss": 3.407107038062688,
      "tokens_seen": 2171666432
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.906596051998075e-05,
      "loss": 2.4941,
      "theoretical_loss": 3.40709016939846,
      "tokens_seen": 2171797504
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.90579361258225e-05,
      "loss": 2.6267,
      "theoretical_loss": 3.4070733020372916,
      "tokens_seen": 2171928576
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.904991173166427e-05,
      "loss": 2.5602,
      "theoretical_loss": 3.4070564359790048,
      "tokens_seen": 2172059648
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.904188733750602e-05,
      "loss": 2.6802,
      "theoretical_loss": 3.407039571223419,
      "tokens_seen": 2172190720
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.903386294334779e-05,
      "loss": 2.5658,
      "theoretical_loss": 3.407022707770356,
      "tokens_seen": 2172321792
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.902583854918954e-05,
      "loss": 2.3892,
      "theoretical_loss": 3.4070058456196364,
      "tokens_seen": 2172452864
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.90178141550313e-05,
      "loss": 2.3801,
      "theoretical_loss": 3.406988984771081,
      "tokens_seen": 2172583936
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1197448,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.376589298248291,
      "objective/train/theoretical_loss": 3.4069805548350587,
      "objective/train/tokens_used": 543108576,
      "theoretical_loss": 3.4069805548350587,
      "tokens_seen": 2172649472
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.900978976087306e-05,
      "loss": 2.4687,
      "theoretical_loss": 3.40697212522451,
      "tokens_seen": 2172715008
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.900176536671482e-05,
      "loss": 2.6256,
      "theoretical_loss": 3.4069552669797454,
      "tokens_seen": 2172846080
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.899374097255658e-05,
      "loss": 2.5264,
      "theoretical_loss": 3.4069384100366076,
      "tokens_seen": 2172977152
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.898571657839833e-05,
      "loss": 2.4931,
      "theoretical_loss": 3.406921554394918,
      "tokens_seen": 2173108224
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.897769218424009e-05,
      "loss": 2.3636,
      "theoretical_loss": 3.406904700054497,
      "tokens_seen": 2173239296
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.896966779008185e-05,
      "loss": 2.4235,
      "theoretical_loss": 3.4068878470151662,
      "tokens_seen": 2173370368
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.896164339592361e-05,
      "loss": 2.4344,
      "theoretical_loss": 3.406870995276747,
      "tokens_seen": 2173501440
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.895361900176537e-05,
      "loss": 2.4578,
      "theoretical_loss": 3.406854144839059,
      "tokens_seen": 2173632512
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.894559460760713e-05,
      "loss": 2.5416,
      "theoretical_loss": 3.406837295701925,
      "tokens_seen": 2173763584
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.89375702134489e-05,
      "loss": 2.5487,
      "theoretical_loss": 3.406820447865166,
      "tokens_seen": 2173894656
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.892954581929065e-05,
      "loss": 2.5288,
      "theoretical_loss": 3.4068036013286025,
      "tokens_seen": 2174025728
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.89215214251324e-05,
      "loss": 2.4718,
      "theoretical_loss": 3.406786756092056,
      "tokens_seen": 2174156800
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1198116,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.546083927154541,
      "objective/train/theoretical_loss": 3.4067699121553483,
      "objective/train/tokens_used": 544746976,
      "theoretical_loss": 3.4067699121553483,
      "tokens_seen": 2174287872
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.891349703097417e-05,
      "loss": 2.4965,
      "theoretical_loss": 3.4067699121553483,
      "tokens_seen": 2174287872
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.890547263681592e-05,
      "loss": 2.5112,
      "theoretical_loss": 3.4067530695183,
      "tokens_seen": 2174418944
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.889744824265769e-05,
      "loss": 2.4467,
      "theoretical_loss": 3.4067362281807334,
      "tokens_seen": 2174550016
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.888942384849944e-05,
      "loss": 2.5762,
      "theoretical_loss": 3.4067193881424696,
      "tokens_seen": 2174681088
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.88813994543412e-05,
      "loss": 2.5666,
      "theoretical_loss": 3.406702549403329,
      "tokens_seen": 2174812160
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.887337506018296e-05,
      "loss": 2.4124,
      "theoretical_loss": 3.406685711963135,
      "tokens_seen": 2174943232
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.886535066602471e-05,
      "loss": 2.4991,
      "theoretical_loss": 3.406668875821708,
      "tokens_seen": 2175074304
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.885732627186648e-05,
      "loss": 2.3663,
      "theoretical_loss": 3.406652040978869,
      "tokens_seen": 2175205376
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.884930187770823e-05,
      "loss": 2.589,
      "theoretical_loss": 3.406635207434441,
      "tokens_seen": 2175336448
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.884127748355e-05,
      "loss": 2.4368,
      "theoretical_loss": 3.4066183751882444,
      "tokens_seen": 2175467520
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.883325308939175e-05,
      "loss": 2.568,
      "theoretical_loss": 3.4066015442401016,
      "tokens_seen": 2175598592
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.88252286952335e-05,
      "loss": 2.515,
      "theoretical_loss": 3.406584714589834,
      "tokens_seen": 2175729664
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.881720430107527e-05,
      "loss": 2.5158,
      "theoretical_loss": 3.4065678862372635,
      "tokens_seen": 2175860736
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1199462,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.0797388553619385,
      "objective/train/theoretical_loss": 3.406559472547559,
      "objective/train/tokens_used": 546385376,
      "theoretical_loss": 3.406559472547559,
      "tokens_seen": 2175926272
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.880917990691703e-05,
      "loss": 2.4784,
      "theoretical_loss": 3.406551059182212,
      "tokens_seen": 2175991808
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.880115551275879e-05,
      "loss": 2.5083,
      "theoretical_loss": 3.406534233424501,
      "tokens_seen": 2176122880
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.879313111860055e-05,
      "loss": 2.5584,
      "theoretical_loss": 3.4065174089639525,
      "tokens_seen": 2176253952
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.87851067244423e-05,
      "loss": 2.573,
      "theoretical_loss": 3.406500585800388,
      "tokens_seen": 2176385024
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.877708233028407e-05,
      "loss": 2.5617,
      "theoretical_loss": 3.40648376393363,
      "tokens_seen": 2176516096
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.876905793612582e-05,
      "loss": 2.5937,
      "theoretical_loss": 3.4064669433635,
      "tokens_seen": 2176647168
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.876103354196759e-05,
      "loss": 2.571,
      "theoretical_loss": 3.4064501240898206,
      "tokens_seen": 2176778240
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.875300914780934e-05,
      "loss": 2.515,
      "theoretical_loss": 3.406433306112413,
      "tokens_seen": 2176909312
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.87449847536511e-05,
      "loss": 2.5359,
      "theoretical_loss": 3.4064164894310998,
      "tokens_seen": 2177040384
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.873696035949286e-05,
      "loss": 2.5196,
      "theoretical_loss": 3.406399674045703,
      "tokens_seen": 2177171456
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.872893596533461e-05,
      "loss": 2.5101,
      "theoretical_loss": 3.406382859956045,
      "tokens_seen": 2177302528
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.872091157117638e-05,
      "loss": 2.4146,
      "theoretical_loss": 3.406366047161947,
      "tokens_seen": 2177433600
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1200438,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.211505174636841,
      "objective/train/theoretical_loss": 3.406349235663232,
      "objective/train/tokens_used": 548023776,
      "theoretical_loss": 3.406349235663232,
      "tokens_seen": 2177564672
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.871288717701813e-05,
      "loss": 2.4019,
      "theoretical_loss": 3.406349235663232,
      "tokens_seen": 2177564672
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.87048627828599e-05,
      "loss": 2.4612,
      "theoretical_loss": 3.4063324254597225,
      "tokens_seen": 2177695744
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.869683838870165e-05,
      "loss": 2.6133,
      "theoretical_loss": 3.40631561655124,
      "tokens_seen": 2177826816
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.86888139945434e-05,
      "loss": 2.4651,
      "theoretical_loss": 3.406298808937607,
      "tokens_seen": 2177957888
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.868078960038517e-05,
      "loss": 2.4464,
      "theoretical_loss": 3.406282002618646,
      "tokens_seen": 2178088960
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.867276520622692e-05,
      "loss": 2.4255,
      "theoretical_loss": 3.4062651975941796,
      "tokens_seen": 2178220032
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.866474081206869e-05,
      "loss": 2.4586,
      "theoretical_loss": 3.40624839386403,
      "tokens_seen": 2178351104
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.865671641791044e-05,
      "loss": 2.4815,
      "theoretical_loss": 3.4062315914280195,
      "tokens_seen": 2178482176
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.864869202375221e-05,
      "loss": 2.6005,
      "theoretical_loss": 3.4062147902859707,
      "tokens_seen": 2178613248
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.864066762959396e-05,
      "loss": 2.5733,
      "theoretical_loss": 3.4061979904377058,
      "tokens_seen": 2178744320
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.863264323543572e-05,
      "loss": 2.4796,
      "theoretical_loss": 3.4061811918830482,
      "tokens_seen": 2178875392
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.862461884127748e-05,
      "loss": 2.5401,
      "theoretical_loss": 3.406164394621819,
      "tokens_seen": 2179006464
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.861659444711924e-05,
      "loss": 2.4538,
      "theoretical_loss": 3.4061475986538428,
      "tokens_seen": 2179137536
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1201179,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.695457696914673,
      "objective/train/theoretical_loss": 3.4061392011547684,
      "objective/train/tokens_used": 549662176,
      "theoretical_loss": 3.4061392011547684,
      "tokens_seen": 2179203072
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.8608570052961e-05,
      "loss": 2.5881,
      "theoretical_loss": 3.4061308039789404,
      "tokens_seen": 2179268608
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.860054565880276e-05,
      "loss": 2.4436,
      "theoretical_loss": 3.406114010596936,
      "tokens_seen": 2179399680
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.859252126464452e-05,
      "loss": 2.5682,
      "theoretical_loss": 3.406097218507651,
      "tokens_seen": 2179530752
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.858449687048628e-05,
      "loss": 2.3501,
      "theoretical_loss": 3.4060804277109087,
      "tokens_seen": 2179661824
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.857647247632803e-05,
      "loss": 2.4841,
      "theoretical_loss": 3.406063638206532,
      "tokens_seen": 2179792896
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.85684480821698e-05,
      "loss": 2.4954,
      "theoretical_loss": 3.4060468499943437,
      "tokens_seen": 2179923968
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.856042368801155e-05,
      "loss": 2.5969,
      "theoretical_loss": 3.4060300630741667,
      "tokens_seen": 2180055040
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.855239929385332e-05,
      "loss": 2.5807,
      "theoretical_loss": 3.4060132774458243,
      "tokens_seen": 2180186112
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.854437489969507e-05,
      "loss": 2.5297,
      "theoretical_loss": 3.4059964931091384,
      "tokens_seen": 2180317184
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.853635050553684e-05,
      "loss": 2.5044,
      "theoretical_loss": 3.4059797100639324,
      "tokens_seen": 2180448256
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.852832611137859e-05,
      "loss": 2.506,
      "theoretical_loss": 3.4059629283100294,
      "tokens_seen": 2180579328
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.852030171722034e-05,
      "loss": 2.4052,
      "theoretical_loss": 3.405946147847253,
      "tokens_seen": 2180710400
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1202357,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7669031620025635,
      "objective/train/theoretical_loss": 3.4059293686754253,
      "objective/train/tokens_used": 551300576,
      "theoretical_loss": 3.4059293686754253,
      "tokens_seen": 2180841472
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.851227732306211e-05,
      "loss": 2.4613,
      "theoretical_loss": 3.4059293686754253,
      "tokens_seen": 2180841472
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.850425292890386e-05,
      "loss": 2.529,
      "theoretical_loss": 3.40591259079437,
      "tokens_seen": 2180972544
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.849622853474563e-05,
      "loss": 2.6914,
      "theoretical_loss": 3.40589581420391,
      "tokens_seen": 2181103616
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.848820414058738e-05,
      "loss": 2.5318,
      "theoretical_loss": 3.4058790389038687,
      "tokens_seen": 2181234688
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.848017974642915e-05,
      "loss": 2.4421,
      "theoretical_loss": 3.4058622648940693,
      "tokens_seen": 2181365760
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.84721553522709e-05,
      "loss": 2.4674,
      "theoretical_loss": 3.4058454921743344,
      "tokens_seen": 2181496832
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.846413095811267e-05,
      "loss": 2.3825,
      "theoretical_loss": 3.405828720744488,
      "tokens_seen": 2181627904
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.845610656395442e-05,
      "loss": 2.5162,
      "theoretical_loss": 3.4058119506043534,
      "tokens_seen": 2181758976
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.844808216979617e-05,
      "loss": 2.3782,
      "theoretical_loss": 3.4057951817537537,
      "tokens_seen": 2181890048
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.844005777563794e-05,
      "loss": 2.4567,
      "theoretical_loss": 3.405778414192512,
      "tokens_seen": 2182021120
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.84320333814797e-05,
      "loss": 2.4216,
      "theoretical_loss": 3.4057616479204524,
      "tokens_seen": 2182152192
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.842400898732146e-05,
      "loss": 2.4975,
      "theoretical_loss": 3.405744882937398,
      "tokens_seen": 2182283264
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.841598459316321e-05,
      "loss": 2.41,
      "theoretical_loss": 3.405728119243172,
      "tokens_seen": 2182414336
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1202887,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.703336000442505,
      "objective/train/theoretical_loss": 3.405719737879315,
      "objective/train/tokens_used": 552938976,
      "theoretical_loss": 3.405719737879315,
      "tokens_seen": 2182479872
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.840796019900498e-05,
      "loss": 2.6198,
      "theoretical_loss": 3.4057113568375987,
      "tokens_seen": 2182545408
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.839993580484673e-05,
      "loss": 2.5698,
      "theoretical_loss": 3.405694595720501,
      "tokens_seen": 2182676480
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.839191141068849e-05,
      "loss": 2.4929,
      "theoretical_loss": 3.4056778358917024,
      "tokens_seen": 2182807552
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.838388701653025e-05,
      "loss": 2.5502,
      "theoretical_loss": 3.405661077351027,
      "tokens_seen": 2182938624
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.8375862622372e-05,
      "loss": 2.6634,
      "theoretical_loss": 3.405644320098298,
      "tokens_seen": 2183069696
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.836783822821377e-05,
      "loss": 2.5867,
      "theoretical_loss": 3.40562756413334,
      "tokens_seen": 2183200768
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.835981383405553e-05,
      "loss": 2.5164,
      "theoretical_loss": 3.405610809455976,
      "tokens_seen": 2183331840
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.835178943989729e-05,
      "loss": 2.4997,
      "theoretical_loss": 3.4055940560660294,
      "tokens_seen": 2183462912
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.834376504573905e-05,
      "loss": 2.4337,
      "theoretical_loss": 3.405577303963325,
      "tokens_seen": 2183593984
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.83357406515808e-05,
      "loss": 2.4849,
      "theoretical_loss": 3.4055605531476854,
      "tokens_seen": 2183725056
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.832771625742257e-05,
      "loss": 2.5738,
      "theoretical_loss": 3.4055438036189356,
      "tokens_seen": 2183856128
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.831969186326432e-05,
      "loss": 2.4967,
      "theoretical_loss": 3.405527055376899,
      "tokens_seen": 2183987200
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1204090,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4657957553863525,
      "objective/train/theoretical_loss": 3.4055103084213996,
      "objective/train/tokens_used": 554577376,
      "theoretical_loss": 3.4055103084213996,
      "tokens_seen": 2184118272
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.831166746910609e-05,
      "loss": 2.5754,
      "theoretical_loss": 3.4055103084213996,
      "tokens_seen": 2184118272
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.830364307494784e-05,
      "loss": 2.4762,
      "theoretical_loss": 3.4054935627522616,
      "tokens_seen": 2184249344
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.82956186807896e-05,
      "loss": 2.5145,
      "theoretical_loss": 3.4054768183693085,
      "tokens_seen": 2184380416
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.828759428663136e-05,
      "loss": 2.5413,
      "theoretical_loss": 3.405460075272365,
      "tokens_seen": 2184511488
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.827956989247311e-05,
      "loss": 2.6336,
      "theoretical_loss": 3.4054433334612546,
      "tokens_seen": 2184642560
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.827154549831488e-05,
      "loss": 2.5995,
      "theoretical_loss": 3.4054265929358016,
      "tokens_seen": 2184773632
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.826352110415663e-05,
      "loss": 2.5385,
      "theoretical_loss": 3.40540985369583,
      "tokens_seen": 2184904704
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.82554967099984e-05,
      "loss": 2.3402,
      "theoretical_loss": 3.4053931157411648,
      "tokens_seen": 2185035776
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.824747231584015e-05,
      "loss": 2.4948,
      "theoretical_loss": 3.405376379071629,
      "tokens_seen": 2185166848
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.823944792168192e-05,
      "loss": 2.6132,
      "theoretical_loss": 3.405359643687048,
      "tokens_seen": 2185297920
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.823142352752367e-05,
      "loss": 2.4663,
      "theoretical_loss": 3.405342909587245,
      "tokens_seen": 2185428992
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.822339913336542e-05,
      "loss": 2.5485,
      "theoretical_loss": 3.405326176772045,
      "tokens_seen": 2185560064
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.821537473920719e-05,
      "loss": 2.5661,
      "theoretical_loss": 3.405309445241272,
      "tokens_seen": 2185691136
    },
    {
      "epoch": 0.32,
      "objective/train/docs_used": 1205389,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9753358364105225,
      "objective/train/theoretical_loss": 3.4053010799574914,
      "objective/train/tokens_used": 556215776,
      "theoretical_loss": 3.4053010799574914,
      "tokens_seen": 2185756672
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.820735034504894e-05,
      "loss": 2.5152,
      "theoretical_loss": 3.405292714994751,
      "tokens_seen": 2185822208
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.819932595089071e-05,
      "loss": 2.7423,
      "theoretical_loss": 3.405275986032306,
      "tokens_seen": 2185953280
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.819130155673246e-05,
      "loss": 2.5975,
      "theoretical_loss": 3.4052592583537615,
      "tokens_seen": 2186084352
    },
    {
      "epoch": 0.32,
      "learning_rate": 6.818327716257423e-05,
      "loss": 2.7106,
      "theoretical_loss": 3.4052425319589417,
      "tokens_seen": 2186215424
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.817525276841598e-05,
      "loss": 2.5553,
      "theoretical_loss": 3.4052258068476715,
      "tokens_seen": 2186346496
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.816722837425775e-05,
      "loss": 2.5986,
      "theoretical_loss": 3.4052090830197757,
      "tokens_seen": 2186477568
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.81592039800995e-05,
      "loss": 2.6214,
      "theoretical_loss": 3.405192360475078,
      "tokens_seen": 2186608640
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.815117958594126e-05,
      "loss": 2.517,
      "theoretical_loss": 3.405175639213404,
      "tokens_seen": 2186739712
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.814315519178302e-05,
      "loss": 2.5678,
      "theoretical_loss": 3.4051589192345775,
      "tokens_seen": 2186870784
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.813513079762478e-05,
      "loss": 2.471,
      "theoretical_loss": 3.405142200538424,
      "tokens_seen": 2187001856
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.812710640346654e-05,
      "loss": 2.6426,
      "theoretical_loss": 3.405125483124768,
      "tokens_seen": 2187132928
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.81190820093083e-05,
      "loss": 2.5891,
      "theoretical_loss": 3.4051087669934343,
      "tokens_seen": 2187264000
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1206012,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.534045696258545,
      "objective/train/theoretical_loss": 3.405092052144247,
      "objective/train/tokens_used": 557854176,
      "theoretical_loss": 3.405092052144247,
      "tokens_seen": 2187395072
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.811105761515006e-05,
      "loss": 2.6112,
      "theoretical_loss": 3.405092052144247,
      "tokens_seen": 2187395072
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.810303322099182e-05,
      "loss": 2.5409,
      "theoretical_loss": 3.405075338577032,
      "tokens_seen": 2187526144
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.809500882683357e-05,
      "loss": 2.596,
      "theoretical_loss": 3.4050586262916136,
      "tokens_seen": 2187657216
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.808698443267534e-05,
      "loss": 2.4925,
      "theoretical_loss": 3.4050419152878164,
      "tokens_seen": 2187788288
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.807896003851709e-05,
      "loss": 2.5335,
      "theoretical_loss": 3.405025205565466,
      "tokens_seen": 2187919360
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.807093564435886e-05,
      "loss": 2.5448,
      "theoretical_loss": 3.405008497124387,
      "tokens_seen": 2188050432
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.806291125020061e-05,
      "loss": 2.5736,
      "theoretical_loss": 3.404991789964405,
      "tokens_seen": 2188181504
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.805488685604238e-05,
      "loss": 2.6279,
      "theoretical_loss": 3.404975084085344,
      "tokens_seen": 2188312576
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.804686246188413e-05,
      "loss": 2.425,
      "theoretical_loss": 3.40495837948703,
      "tokens_seen": 2188443648
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.803883806772588e-05,
      "loss": 2.4579,
      "theoretical_loss": 3.4049416761692872,
      "tokens_seen": 2188574720
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.803081367356765e-05,
      "loss": 2.5621,
      "theoretical_loss": 3.4049249741319416,
      "tokens_seen": 2188705792
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.80227892794094e-05,
      "loss": 2.4914,
      "theoretical_loss": 3.4049082733748186,
      "tokens_seen": 2188836864
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.801476488525117e-05,
      "loss": 2.4783,
      "theoretical_loss": 3.4048915738977423,
      "tokens_seen": 2188967936
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1207268,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3856353759765625,
      "objective/train/theoretical_loss": 3.4048832246391676,
      "objective/train/tokens_used": 559492576,
      "theoretical_loss": 3.4048832246391676,
      "tokens_seen": 2189033472
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.800674049109292e-05,
      "loss": 2.4133,
      "theoretical_loss": 3.4048748757005387,
      "tokens_seen": 2189099008
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.799871609693469e-05,
      "loss": 2.5576,
      "theoretical_loss": 3.4048581787830328,
      "tokens_seen": 2189230080
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.799069170277644e-05,
      "loss": 2.5191,
      "theoretical_loss": 3.4048414831450504,
      "tokens_seen": 2189361152
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.79826673086182e-05,
      "loss": 2.5269,
      "theoretical_loss": 3.4048247887864163,
      "tokens_seen": 2189492224
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.797464291445996e-05,
      "loss": 2.4617,
      "theoretical_loss": 3.404808095706956,
      "tokens_seen": 2189623296
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.796661852030171e-05,
      "loss": 2.5635,
      "theoretical_loss": 3.4047914039064953,
      "tokens_seen": 2189754368
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.795859412614348e-05,
      "loss": 2.5235,
      "theoretical_loss": 3.404774713384859,
      "tokens_seen": 2189885440
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.795056973198523e-05,
      "loss": 2.5943,
      "theoretical_loss": 3.404758024141873,
      "tokens_seen": 2190016512
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.7942545337827e-05,
      "loss": 2.5848,
      "theoretical_loss": 3.4047413361773633,
      "tokens_seen": 2190147584
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.793452094366875e-05,
      "loss": 2.5442,
      "theoretical_loss": 3.4047246494911545,
      "tokens_seen": 2190278656
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.792649654951051e-05,
      "loss": 2.4471,
      "theoretical_loss": 3.404707964083073,
      "tokens_seen": 2190409728
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.791847215535227e-05,
      "loss": 2.3276,
      "theoretical_loss": 3.4046912799529436,
      "tokens_seen": 2190540800
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1207761,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.551201581954956,
      "objective/train/theoretical_loss": 3.404674597100593,
      "objective/train/tokens_used": 561130976,
      "theoretical_loss": 3.404674597100593,
      "tokens_seen": 2190671872
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.791044776119403e-05,
      "loss": 2.5127,
      "theoretical_loss": 3.404674597100593,
      "tokens_seen": 2190671872
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.79024233670358e-05,
      "loss": 2.5596,
      "theoretical_loss": 3.4046579155258456,
      "tokens_seen": 2190802944
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.789439897287755e-05,
      "loss": 2.4916,
      "theoretical_loss": 3.4046412352285285,
      "tokens_seen": 2190934016
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.788637457871931e-05,
      "loss": 2.4777,
      "theoretical_loss": 3.4046245562084665,
      "tokens_seen": 2191065088
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.787835018456107e-05,
      "loss": 2.4466,
      "theoretical_loss": 3.404607878465486,
      "tokens_seen": 2191196160
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.787032579040283e-05,
      "loss": 2.5021,
      "theoretical_loss": 3.4045912019994122,
      "tokens_seen": 2191327232
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.786230139624459e-05,
      "loss": 2.5078,
      "theoretical_loss": 3.4045745268100718,
      "tokens_seen": 2191458304
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.785427700208634e-05,
      "loss": 2.3974,
      "theoretical_loss": 3.40455785289729,
      "tokens_seen": 2191589376
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.78462526079281e-05,
      "loss": 2.5153,
      "theoretical_loss": 3.404541180260893,
      "tokens_seen": 2191720448
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.783822821376986e-05,
      "loss": 2.6615,
      "theoretical_loss": 3.404524508900707,
      "tokens_seen": 2191851520
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.783020381961163e-05,
      "loss": 2.6367,
      "theoretical_loss": 3.404507838816557,
      "tokens_seen": 2191982592
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.782217942545338e-05,
      "loss": 2.4062,
      "theoretical_loss": 3.40449117000827,
      "tokens_seen": 2192113664
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.781415503129515e-05,
      "loss": 2.4533,
      "theoretical_loss": 3.404474502475672,
      "tokens_seen": 2192244736
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1208978,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7618091106414795,
      "objective/train/theoretical_loss": 3.4044661691877023,
      "objective/train/tokens_used": 562769376,
      "theoretical_loss": 3.4044661691877023,
      "tokens_seen": 2192310272
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.78061306371369e-05,
      "loss": 2.6423,
      "theoretical_loss": 3.4044578362185893,
      "tokens_seen": 2192375808
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.779810624297865e-05,
      "loss": 2.3443,
      "theoretical_loss": 3.404441171236847,
      "tokens_seen": 2192506880
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.779008184882042e-05,
      "loss": 2.4826,
      "theoretical_loss": 3.4044245075302726,
      "tokens_seen": 2192637952
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.778205745466217e-05,
      "loss": 2.7126,
      "theoretical_loss": 3.4044078450986914,
      "tokens_seen": 2192769024
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.777403306050394e-05,
      "loss": 2.4795,
      "theoretical_loss": 3.40439118394193,
      "tokens_seen": 2192900096
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.776600866634569e-05,
      "loss": 2.4872,
      "theoretical_loss": 3.4043745240598144,
      "tokens_seen": 2193031168
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.775798427218746e-05,
      "loss": 2.5142,
      "theoretical_loss": 3.4043578654521713,
      "tokens_seen": 2193162240
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.774995987802921e-05,
      "loss": 2.5768,
      "theoretical_loss": 3.404341208118827,
      "tokens_seen": 2193293312
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.774193548387096e-05,
      "loss": 2.3759,
      "theoretical_loss": 3.4043245520596077,
      "tokens_seen": 2193424384
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.773391108971273e-05,
      "loss": 2.4352,
      "theoretical_loss": 3.4043078972743395,
      "tokens_seen": 2193555456
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.772588669555448e-05,
      "loss": 2.562,
      "theoretical_loss": 3.4042912437628496,
      "tokens_seen": 2193686528
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.771786230139625e-05,
      "loss": 2.5281,
      "theoretical_loss": 3.4042745915249637,
      "tokens_seen": 2193817600
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1209592,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4534859657287598,
      "objective/train/theoretical_loss": 3.404257940560509,
      "objective/train/tokens_used": 564407776,
      "theoretical_loss": 3.404257940560509,
      "tokens_seen": 2193948672
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.7709837907238e-05,
      "loss": 2.5249,
      "theoretical_loss": 3.404257940560509,
      "tokens_seen": 2193948672
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.770181351307977e-05,
      "loss": 2.4806,
      "theoretical_loss": 3.4042412908693116,
      "tokens_seen": 2194079744
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.769378911892152e-05,
      "loss": 2.5328,
      "theoretical_loss": 3.404224642451198,
      "tokens_seen": 2194210816
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.768576472476328e-05,
      "loss": 2.4127,
      "theoretical_loss": 3.4042079953059954,
      "tokens_seen": 2194341888
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.767774033060504e-05,
      "loss": 2.3612,
      "theoretical_loss": 3.40419134943353,
      "tokens_seen": 2194472960
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.76697159364468e-05,
      "loss": 2.5744,
      "theoretical_loss": 3.4041747048336286,
      "tokens_seen": 2194604032
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.766169154228856e-05,
      "loss": 2.5248,
      "theoretical_loss": 3.4041580615061178,
      "tokens_seen": 2194735104
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.765366714813032e-05,
      "loss": 2.4543,
      "theoretical_loss": 3.4041414194508244,
      "tokens_seen": 2194866176
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.764564275397208e-05,
      "loss": 2.5248,
      "theoretical_loss": 3.4041247786675752,
      "tokens_seen": 2194997248
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.763761835981384e-05,
      "loss": 2.5018,
      "theoretical_loss": 3.404108139156197,
      "tokens_seen": 2195128320
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.762959396565559e-05,
      "loss": 2.4872,
      "theoretical_loss": 3.4040915009165174,
      "tokens_seen": 2195259392
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.762156957149736e-05,
      "loss": 2.5658,
      "theoretical_loss": 3.404074863948362,
      "tokens_seen": 2195390464
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.761354517733911e-05,
      "loss": 2.5886,
      "theoretical_loss": 3.404058228251558,
      "tokens_seen": 2195521536
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1211067,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.898007869720459,
      "objective/train/theoretical_loss": 3.404049910879859,
      "objective/train/tokens_used": 566046176,
      "theoretical_loss": 3.404049910879859,
      "tokens_seen": 2195587072
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.760552078318088e-05,
      "loss": 2.5062,
      "theoretical_loss": 3.404041593825933,
      "tokens_seen": 2195652608
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.759749638902263e-05,
      "loss": 2.4295,
      "theoretical_loss": 3.4040249606713138,
      "tokens_seen": 2195783680
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.75894719948644e-05,
      "loss": 2.5987,
      "theoretical_loss": 3.404008328787527,
      "tokens_seen": 2195914752
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.758144760070615e-05,
      "loss": 2.5701,
      "theoretical_loss": 3.4039916981744005,
      "tokens_seen": 2196045824
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.75734232065479e-05,
      "loss": 2.5974,
      "theoretical_loss": 3.4039750688317603,
      "tokens_seen": 2196176896
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.756539881238967e-05,
      "loss": 2.5712,
      "theoretical_loss": 3.4039584407594345,
      "tokens_seen": 2196307968
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.755737441823142e-05,
      "loss": 2.4627,
      "theoretical_loss": 3.4039418139572493,
      "tokens_seen": 2196439040
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.754935002407319e-05,
      "loss": 2.61,
      "theoretical_loss": 3.403925188425033,
      "tokens_seen": 2196570112
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.754132562991494e-05,
      "loss": 2.4044,
      "theoretical_loss": 3.4039085641626117,
      "tokens_seen": 2196701184
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.753330123575671e-05,
      "loss": 2.4572,
      "theoretical_loss": 3.4038919411698134,
      "tokens_seen": 2196832256
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.752527684159846e-05,
      "loss": 2.3802,
      "theoretical_loss": 3.403875319446465,
      "tokens_seen": 2196963328
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.751725244744023e-05,
      "loss": 2.538,
      "theoretical_loss": 3.4038586989923942,
      "tokens_seen": 2197094400
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1211765,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.350731134414673,
      "objective/train/theoretical_loss": 3.4038420798074283,
      "objective/train/tokens_used": 567684576,
      "theoretical_loss": 3.4038420798074283,
      "tokens_seen": 2197225472
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.750922805328198e-05,
      "loss": 2.6054,
      "theoretical_loss": 3.4038420798074283,
      "tokens_seen": 2197225472
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.750120365912373e-05,
      "loss": 2.5121,
      "theoretical_loss": 3.403825461891395,
      "tokens_seen": 2197356544
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.74931792649655e-05,
      "loss": 2.3871,
      "theoretical_loss": 3.4038088452441206,
      "tokens_seen": 2197487616
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.748515487080725e-05,
      "loss": 2.5174,
      "theoretical_loss": 3.4037922298654335,
      "tokens_seen": 2197618688
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.747713047664902e-05,
      "loss": 2.5223,
      "theoretical_loss": 3.403775615755161,
      "tokens_seen": 2197749760
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.746910608249077e-05,
      "loss": 2.4984,
      "theoretical_loss": 3.4037590029131306,
      "tokens_seen": 2197880832
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.746108168833254e-05,
      "loss": 2.3523,
      "theoretical_loss": 3.40374239133917,
      "tokens_seen": 2198011904
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.74530572941743e-05,
      "loss": 2.5873,
      "theoretical_loss": 3.4037257810331063,
      "tokens_seen": 2198142976
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.744503290001605e-05,
      "loss": 2.3596,
      "theoretical_loss": 3.403709171994768,
      "tokens_seen": 2198274048
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.743700850585781e-05,
      "loss": 2.5814,
      "theoretical_loss": 3.4036925642239817,
      "tokens_seen": 2198405120
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.742898411169957e-05,
      "loss": 2.3935,
      "theoretical_loss": 3.403675957720576,
      "tokens_seen": 2198536192
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.742095971754133e-05,
      "loss": 2.5633,
      "theoretical_loss": 3.403659352484378,
      "tokens_seen": 2198667264
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.741293532338309e-05,
      "loss": 2.3273,
      "theoretical_loss": 3.403642748515216,
      "tokens_seen": 2198798336
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1213097,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.580597162246704,
      "objective/train/theoretical_loss": 3.40363444700572,
      "objective/train/tokens_used": 569322976,
      "theoretical_loss": 3.40363444700572,
      "tokens_seen": 2198863872
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.740491092922485e-05,
      "loss": 2.4982,
      "theoretical_loss": 3.403626145812918,
      "tokens_seen": 2198929408
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.73968865350666e-05,
      "loss": 2.6289,
      "theoretical_loss": 3.4036095443773107,
      "tokens_seen": 2199060480
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.738886214090836e-05,
      "loss": 2.5871,
      "theoretical_loss": 3.403592944208223,
      "tokens_seen": 2199191552
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.738083774675013e-05,
      "loss": 2.4835,
      "theoretical_loss": 3.403576345305482,
      "tokens_seen": 2199322624
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.737281335259188e-05,
      "loss": 2.4306,
      "theoretical_loss": 3.4035597476689166,
      "tokens_seen": 2199453696
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.736478895843365e-05,
      "loss": 2.6061,
      "theoretical_loss": 3.4035431512983543,
      "tokens_seen": 2199584768
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.73567645642754e-05,
      "loss": 2.5906,
      "theoretical_loss": 3.403526556193623,
      "tokens_seen": 2199715840
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.734874017011717e-05,
      "loss": 2.4728,
      "theoretical_loss": 3.403509962354551,
      "tokens_seen": 2199846912
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.734071577595892e-05,
      "loss": 2.4921,
      "theoretical_loss": 3.403493369780966,
      "tokens_seen": 2199977984
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.733269138180067e-05,
      "loss": 2.6282,
      "theoretical_loss": 3.403476778472696,
      "tokens_seen": 2200109056
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.732466698764244e-05,
      "loss": 2.442,
      "theoretical_loss": 3.4034601884295697,
      "tokens_seen": 2200240128
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.731664259348419e-05,
      "loss": 2.4406,
      "theoretical_loss": 3.403443599651415,
      "tokens_seen": 2200371200
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1214287,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.592477560043335,
      "objective/train/theoretical_loss": 3.40342701213806,
      "objective/train/tokens_used": 570961376,
      "theoretical_loss": 3.40342701213806,
      "tokens_seen": 2200502272
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.730861819932596e-05,
      "loss": 2.493,
      "theoretical_loss": 3.40342701213806,
      "tokens_seen": 2200502272
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.730059380516771e-05,
      "loss": 2.3522,
      "theoretical_loss": 3.403410425889333,
      "tokens_seen": 2200633344
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.729256941100948e-05,
      "loss": 2.3456,
      "theoretical_loss": 3.4033938409050624,
      "tokens_seen": 2200764416
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.728454501685123e-05,
      "loss": 2.5727,
      "theoretical_loss": 3.403377257185076,
      "tokens_seen": 2200895488
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.727652062269298e-05,
      "loss": 2.4925,
      "theoretical_loss": 3.4033606747292033,
      "tokens_seen": 2201026560
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.726849622853475e-05,
      "loss": 2.5255,
      "theoretical_loss": 3.403344093537272,
      "tokens_seen": 2201157632
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.72604718343765e-05,
      "loss": 2.3977,
      "theoretical_loss": 3.4033275136091095,
      "tokens_seen": 2201288704
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.725244744021827e-05,
      "loss": 2.4072,
      "theoretical_loss": 3.4033109349445456,
      "tokens_seen": 2201419776
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.724442304606002e-05,
      "loss": 2.414,
      "theoretical_loss": 3.4032943575434085,
      "tokens_seen": 2201550848
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.723639865190179e-05,
      "loss": 2.488,
      "theoretical_loss": 3.4032777814055266,
      "tokens_seen": 2201681920
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.722837425774354e-05,
      "loss": 2.4856,
      "theoretical_loss": 3.403261206530728,
      "tokens_seen": 2201812992
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.722034986358531e-05,
      "loss": 2.5547,
      "theoretical_loss": 3.4032446329188417,
      "tokens_seen": 2201944064
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.721232546942706e-05,
      "loss": 2.5414,
      "theoretical_loss": 3.4032280605696967,
      "tokens_seen": 2202075136
    },
    {
      "epoch": 0.33,
      "objective/train/docs_used": 1214811,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5659379959106445,
      "objective/train/theoretical_loss": 3.403219774868598,
      "objective/train/tokens_used": 572599776,
      "theoretical_loss": 3.403219774868598,
      "tokens_seen": 2202140672
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.720430107526882e-05,
      "loss": 2.3994,
      "theoretical_loss": 3.403211489483121,
      "tokens_seen": 2202206208
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.719627668111058e-05,
      "loss": 2.5061,
      "theoretical_loss": 3.4031949196589433,
      "tokens_seen": 2202337280
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.718825228695234e-05,
      "loss": 2.4373,
      "theoretical_loss": 3.4031783510969924,
      "tokens_seen": 2202468352
    },
    {
      "epoch": 0.33,
      "learning_rate": 6.71802278927941e-05,
      "loss": 2.4786,
      "theoretical_loss": 3.4031617837970973,
      "tokens_seen": 2202599424
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.717220349863586e-05,
      "loss": 2.429,
      "theoretical_loss": 3.4031452177590866,
      "tokens_seen": 2202730496
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.716417910447762e-05,
      "loss": 2.5734,
      "theoretical_loss": 3.4031286529827893,
      "tokens_seen": 2202861568
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.715615471031938e-05,
      "loss": 2.4266,
      "theoretical_loss": 3.4031120894680336,
      "tokens_seen": 2202992640
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.714813031616113e-05,
      "loss": 2.4192,
      "theoretical_loss": 3.4030955272146493,
      "tokens_seen": 2203123712
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.71401059220029e-05,
      "loss": 2.5598,
      "theoretical_loss": 3.4030789662224645,
      "tokens_seen": 2203254784
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.713208152784465e-05,
      "loss": 2.6017,
      "theoretical_loss": 3.4030624064913084,
      "tokens_seen": 2203385856
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.712405713368642e-05,
      "loss": 2.5483,
      "theoretical_loss": 3.40304584802101,
      "tokens_seen": 2203516928
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.711603273952817e-05,
      "loss": 2.4384,
      "theoretical_loss": 3.4030292908113986,
      "tokens_seen": 2203648000
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1215803,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7451982498168945,
      "objective/train/theoretical_loss": 3.403012734862303,
      "objective/train/tokens_used": 574238176,
      "theoretical_loss": 3.403012734862303,
      "tokens_seen": 2203779072
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.710800834536994e-05,
      "loss": 2.5776,
      "theoretical_loss": 3.403012734862303,
      "tokens_seen": 2203779072
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.709998395121169e-05,
      "loss": 2.4848,
      "theoretical_loss": 3.402996180173552,
      "tokens_seen": 2203910144
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.709195955705344e-05,
      "loss": 2.5394,
      "theoretical_loss": 3.4029796267449752,
      "tokens_seen": 2204041216
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.708393516289521e-05,
      "loss": 2.5138,
      "theoretical_loss": 3.4029630745764012,
      "tokens_seen": 2204172288
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.707591076873696e-05,
      "loss": 2.4304,
      "theoretical_loss": 3.4029465236676595,
      "tokens_seen": 2204303360
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.706788637457873e-05,
      "loss": 2.3949,
      "theoretical_loss": 3.40292997401858,
      "tokens_seen": 2204434432
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.705986198042048e-05,
      "loss": 2.422,
      "theoretical_loss": 3.4029134256289906,
      "tokens_seen": 2204565504
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.705183758626225e-05,
      "loss": 2.4923,
      "theoretical_loss": 3.402896878498721,
      "tokens_seen": 2204696576
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.7043813192104e-05,
      "loss": 2.3677,
      "theoretical_loss": 3.402880332627601,
      "tokens_seen": 2204827648
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.703578879794575e-05,
      "loss": 2.5273,
      "theoretical_loss": 3.4028637880154595,
      "tokens_seen": 2204958720
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.702776440378752e-05,
      "loss": 2.5261,
      "theoretical_loss": 3.4028472446621265,
      "tokens_seen": 2205089792
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.701974000962927e-05,
      "loss": 2.4574,
      "theoretical_loss": 3.4028307025674303,
      "tokens_seen": 2205220864
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.701171561547104e-05,
      "loss": 2.5115,
      "theoretical_loss": 3.402814161731201,
      "tokens_seen": 2205351936
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1216279,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.408686637878418,
      "objective/train/theoretical_loss": 3.4028058917849586,
      "objective/train/tokens_used": 575876576,
      "theoretical_loss": 3.4028058917849586,
      "tokens_seen": 2205417472
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.70036912213128e-05,
      "loss": 2.556,
      "theoretical_loss": 3.4027976221532685,
      "tokens_seen": 2205483008
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.699566682715456e-05,
      "loss": 2.4738,
      "theoretical_loss": 3.402781083833462,
      "tokens_seen": 2205614080
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.698764243299631e-05,
      "loss": 2.4675,
      "theoretical_loss": 3.4027645467716106,
      "tokens_seen": 2205745152
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.697961803883807e-05,
      "loss": 2.4454,
      "theoretical_loss": 3.4027480109675436,
      "tokens_seen": 2205876224
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.697159364467983e-05,
      "loss": 2.6172,
      "theoretical_loss": 3.402731476421092,
      "tokens_seen": 2206007296
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.696356925052159e-05,
      "loss": 2.403,
      "theoretical_loss": 3.402714943132084,
      "tokens_seen": 2206138368
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.695554485636335e-05,
      "loss": 2.431,
      "theoretical_loss": 3.40269841110035,
      "tokens_seen": 2206269440
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.694752046220511e-05,
      "loss": 2.4604,
      "theoretical_loss": 3.40268188032572,
      "tokens_seen": 2206400512
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.693949606804687e-05,
      "loss": 2.4266,
      "theoretical_loss": 3.402665350808023,
      "tokens_seen": 2206531584
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.693147167388863e-05,
      "loss": 2.56,
      "theoretical_loss": 3.4026488225470892,
      "tokens_seen": 2206662656
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.69234472797304e-05,
      "loss": 2.5656,
      "theoretical_loss": 3.402632295542748,
      "tokens_seen": 2206793728
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.691542288557215e-05,
      "loss": 2.4333,
      "theoretical_loss": 3.40261576979483,
      "tokens_seen": 2206924800
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1217338,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7593905925750732,
      "objective/train/theoretical_loss": 3.4025992453031644,
      "objective/train/tokens_used": 577514976,
      "theoretical_loss": 3.4025992453031644,
      "tokens_seen": 2207055872
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.69073984914139e-05,
      "loss": 2.5397,
      "theoretical_loss": 3.4025992453031644,
      "tokens_seen": 2207055872
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.689937409725567e-05,
      "loss": 2.6356,
      "theoretical_loss": 3.402582722067581,
      "tokens_seen": 2207186944
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.689134970309742e-05,
      "loss": 2.5292,
      "theoretical_loss": 3.4025662000879104,
      "tokens_seen": 2207318016
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.688332530893919e-05,
      "loss": 2.4998,
      "theoretical_loss": 3.4025496793639824,
      "tokens_seen": 2207449088
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.687530091478094e-05,
      "loss": 2.4771,
      "theoretical_loss": 3.4025331598956265,
      "tokens_seen": 2207580160
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.68672765206227e-05,
      "loss": 2.6201,
      "theoretical_loss": 3.402516641682673,
      "tokens_seen": 2207711232
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.685925212646446e-05,
      "loss": 2.583,
      "theoretical_loss": 3.4025001247249524,
      "tokens_seen": 2207842304
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.685122773230621e-05,
      "loss": 2.4937,
      "theoretical_loss": 3.402483609022294,
      "tokens_seen": 2207973376
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.684320333814798e-05,
      "loss": 2.3755,
      "theoretical_loss": 3.402467094574529,
      "tokens_seen": 2208104448
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.683517894398973e-05,
      "loss": 2.5587,
      "theoretical_loss": 3.4024505813814865,
      "tokens_seen": 2208235520
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.68271545498315e-05,
      "loss": 2.5725,
      "theoretical_loss": 3.402434069442997,
      "tokens_seen": 2208366592
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.681913015567325e-05,
      "loss": 2.5799,
      "theoretical_loss": 3.4024175587588914,
      "tokens_seen": 2208497664
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.681110576151502e-05,
      "loss": 2.401,
      "theoretical_loss": 3.402401049328999,
      "tokens_seen": 2208628736
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1217846,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5893185138702393,
      "objective/train/theoretical_loss": 3.40239279508433,
      "objective/train/tokens_used": 579153376,
      "theoretical_loss": 3.40239279508433,
      "tokens_seen": 2208694272
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.680308136735677e-05,
      "loss": 2.4925,
      "theoretical_loss": 3.402384541153151,
      "tokens_seen": 2208759808
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.679505697319852e-05,
      "loss": 2.4016,
      "theoretical_loss": 3.402368034231177,
      "tokens_seen": 2208890880
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.678703257904029e-05,
      "loss": 2.3912,
      "theoretical_loss": 3.4023515285629076,
      "tokens_seen": 2209021952
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.677900818488204e-05,
      "loss": 2.4989,
      "theoretical_loss": 3.402335024148173,
      "tokens_seen": 2209153024
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.677098379072381e-05,
      "loss": 2.553,
      "theoretical_loss": 3.4023185209868045,
      "tokens_seen": 2209284096
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.676295939656556e-05,
      "loss": 2.5332,
      "theoretical_loss": 3.4023020190786317,
      "tokens_seen": 2209415168
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.675493500240733e-05,
      "loss": 2.5386,
      "theoretical_loss": 3.4022855184234855,
      "tokens_seen": 2209546240
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.674691060824908e-05,
      "loss": 2.4758,
      "theoretical_loss": 3.4022690190211957,
      "tokens_seen": 2209677312
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.673888621409084e-05,
      "loss": 2.482,
      "theoretical_loss": 3.402252520871594,
      "tokens_seen": 2209808384
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.67308618199326e-05,
      "loss": 2.5679,
      "theoretical_loss": 3.4022360239745106,
      "tokens_seen": 2209939456
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.672283742577436e-05,
      "loss": 2.5476,
      "theoretical_loss": 3.4022195283297756,
      "tokens_seen": 2210070528
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.671481303161612e-05,
      "loss": 2.669,
      "theoretical_loss": 3.4022030339372202,
      "tokens_seen": 2210201600
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1219246,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7586281299591064,
      "objective/train/theoretical_loss": 3.402186540796675,
      "objective/train/tokens_used": 580791776,
      "theoretical_loss": 3.402186540796675,
      "tokens_seen": 2210332672
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.670678863745788e-05,
      "loss": 2.4485,
      "theoretical_loss": 3.402186540796675,
      "tokens_seen": 2210332672
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.669876424329964e-05,
      "loss": 2.3993,
      "theoretical_loss": 3.4021700489079705,
      "tokens_seen": 2210463744
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.66907398491414e-05,
      "loss": 2.4686,
      "theoretical_loss": 3.402153558270938,
      "tokens_seen": 2210594816
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.668271545498315e-05,
      "loss": 2.4943,
      "theoretical_loss": 3.4021370688854073,
      "tokens_seen": 2210725888
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.667469106082492e-05,
      "loss": 2.5063,
      "theoretical_loss": 3.4021205807512107,
      "tokens_seen": 2210856960
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.666666666666667e-05,
      "loss": 2.4981,
      "theoretical_loss": 3.4021040938681777,
      "tokens_seen": 2210988032
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.665864227250844e-05,
      "loss": 2.5183,
      "theoretical_loss": 3.40208760823614,
      "tokens_seen": 2211119104
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.665061787835019e-05,
      "loss": 2.4552,
      "theoretical_loss": 3.402071123854928,
      "tokens_seen": 2211250176
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.664259348419196e-05,
      "loss": 2.5751,
      "theoretical_loss": 3.4020546407243732,
      "tokens_seen": 2211381248
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.663456909003371e-05,
      "loss": 2.5156,
      "theoretical_loss": 3.4020381588443063,
      "tokens_seen": 2211512320
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.662654469587546e-05,
      "loss": 2.5568,
      "theoretical_loss": 3.402021678214558,
      "tokens_seen": 2211643392
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.661852030171723e-05,
      "loss": 2.5904,
      "theoretical_loss": 3.40200519883496,
      "tokens_seen": 2211774464
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.661049590755898e-05,
      "loss": 2.5605,
      "theoretical_loss": 3.4019887207053428,
      "tokens_seen": 2211905536
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1220526,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.844651937484741,
      "objective/train/theoretical_loss": 3.4019804821092245,
      "objective/train/tokens_used": 582430176,
      "theoretical_loss": 3.4019804821092245,
      "tokens_seen": 2211971072
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.660247151340075e-05,
      "loss": 2.461,
      "theoretical_loss": 3.401972243825538,
      "tokens_seen": 2212036608
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.65944471192425e-05,
      "loss": 2.6401,
      "theoretical_loss": 3.4019557681953767,
      "tokens_seen": 2212167680
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.658642272508427e-05,
      "loss": 2.4676,
      "theoretical_loss": 3.4019392938146895,
      "tokens_seen": 2212298752
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.657839833092602e-05,
      "loss": 2.5706,
      "theoretical_loss": 3.4019228206833088,
      "tokens_seen": 2212429824
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.657037393676779e-05,
      "loss": 2.3169,
      "theoretical_loss": 3.401906348801065,
      "tokens_seen": 2212560896
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.656234954260954e-05,
      "loss": 2.6073,
      "theoretical_loss": 3.401889878167789,
      "tokens_seen": 2212691968
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.65543251484513e-05,
      "loss": 2.539,
      "theoretical_loss": 3.401873408783313,
      "tokens_seen": 2212823040
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.654630075429306e-05,
      "loss": 2.5384,
      "theoretical_loss": 3.401856940647468,
      "tokens_seen": 2212954112
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.653827636013481e-05,
      "loss": 2.5948,
      "theoretical_loss": 3.4018404737600854,
      "tokens_seen": 2213085184
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.653025196597658e-05,
      "loss": 2.5626,
      "theoretical_loss": 3.4018240081209963,
      "tokens_seen": 2213216256
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.652222757181833e-05,
      "loss": 2.561,
      "theoretical_loss": 3.401807543730033,
      "tokens_seen": 2213347328
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.65142031776601e-05,
      "loss": 2.4798,
      "theoretical_loss": 3.401791080587026,
      "tokens_seen": 2213478400
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1221127,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5792646408081055,
      "objective/train/theoretical_loss": 3.4017746186918076,
      "objective/train/tokens_used": 584068576,
      "theoretical_loss": 3.4017746186918076,
      "tokens_seen": 2213609472
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.650617878350185e-05,
      "loss": 2.6339,
      "theoretical_loss": 3.4017746186918076,
      "tokens_seen": 2213609472
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.649815438934361e-05,
      "loss": 2.4253,
      "theoretical_loss": 3.401758158044209,
      "tokens_seen": 2213740544
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.649012999518537e-05,
      "loss": 2.5425,
      "theoretical_loss": 3.4017416986440616,
      "tokens_seen": 2213871616
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.648210560102713e-05,
      "loss": 2.5688,
      "theoretical_loss": 3.4017252404911975,
      "tokens_seen": 2214002688
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.64740812068689e-05,
      "loss": 2.5209,
      "theoretical_loss": 3.4017087835854474,
      "tokens_seen": 2214133760
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.646605681271065e-05,
      "loss": 2.5333,
      "theoretical_loss": 3.401692327926644,
      "tokens_seen": 2214264832
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.645803241855241e-05,
      "loss": 2.4757,
      "theoretical_loss": 3.4016758735146193,
      "tokens_seen": 2214395904
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.645000802439417e-05,
      "loss": 2.5999,
      "theoretical_loss": 3.401659420349204,
      "tokens_seen": 2214526976
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.644198363023592e-05,
      "loss": 2.4715,
      "theoretical_loss": 3.40164296843023,
      "tokens_seen": 2214658048
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.643395923607769e-05,
      "loss": 2.4961,
      "theoretical_loss": 3.40162651775753,
      "tokens_seen": 2214789120
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.642593484191944e-05,
      "loss": 2.4219,
      "theoretical_loss": 3.4016100683309345,
      "tokens_seen": 2214920192
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.64179104477612e-05,
      "loss": 2.4605,
      "theoretical_loss": 3.401593620150277,
      "tokens_seen": 2215051264
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.640988605360296e-05,
      "loss": 2.5053,
      "theoretical_loss": 3.401577173215388,
      "tokens_seen": 2215182336
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1222538,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.598327398300171,
      "objective/train/theoretical_loss": 3.4015689502150543,
      "objective/train/tokens_used": 585706976,
      "theoretical_loss": 3.4015689502150543,
      "tokens_seen": 2215247872
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.640186165944471e-05,
      "loss": 2.5099,
      "theoretical_loss": 3.4015607275260997,
      "tokens_seen": 2215313408
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.639383726528648e-05,
      "loss": 2.517,
      "theoretical_loss": 3.401544283082245,
      "tokens_seen": 2215444480
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.638581287112823e-05,
      "loss": 2.6561,
      "theoretical_loss": 3.401527839883655,
      "tokens_seen": 2215575552
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.637778847697e-05,
      "loss": 2.4735,
      "theoretical_loss": 3.401511397930162,
      "tokens_seen": 2215706624
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.636976408281175e-05,
      "loss": 2.5296,
      "theoretical_loss": 3.401494957221598,
      "tokens_seen": 2215837696
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.636173968865352e-05,
      "loss": 2.6785,
      "theoretical_loss": 3.4014785177577953,
      "tokens_seen": 2215968768
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.635371529449527e-05,
      "loss": 2.4534,
      "theoretical_loss": 3.401462079538586,
      "tokens_seen": 2216099840
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.634569090033703e-05,
      "loss": 2.4605,
      "theoretical_loss": 3.401445642563802,
      "tokens_seen": 2216230912
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.633766650617879e-05,
      "loss": 2.4743,
      "theoretical_loss": 3.401429206833276,
      "tokens_seen": 2216361984
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.632964211202055e-05,
      "loss": 2.5421,
      "theoretical_loss": 3.40141277234684,
      "tokens_seen": 2216493056
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.632161771786231e-05,
      "loss": 2.4795,
      "theoretical_loss": 3.401396339104326,
      "tokens_seen": 2216624128
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.631359332370407e-05,
      "loss": 2.3755,
      "theoretical_loss": 3.401379907105566,
      "tokens_seen": 2216755200
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1222919,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.506207227706909,
      "objective/train/theoretical_loss": 3.4013634763503937,
      "objective/train/tokens_used": 587345376,
      "theoretical_loss": 3.4013634763503937,
      "tokens_seen": 2216886272
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.630556892954582e-05,
      "loss": 2.54,
      "theoretical_loss": 3.4013634763503937,
      "tokens_seen": 2216886272
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.629754453538758e-05,
      "loss": 2.6115,
      "theoretical_loss": 3.40134704683864,
      "tokens_seen": 2217017344
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.628952014122934e-05,
      "loss": 2.6062,
      "theoretical_loss": 3.4013306185701384,
      "tokens_seen": 2217148416
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.62814957470711e-05,
      "loss": 2.392,
      "theoretical_loss": 3.4013141915447207,
      "tokens_seen": 2217279488
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.627347135291286e-05,
      "loss": 2.62,
      "theoretical_loss": 3.4012977657622194,
      "tokens_seen": 2217410560
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.626544695875462e-05,
      "loss": 2.5453,
      "theoretical_loss": 3.401281341222467,
      "tokens_seen": 2217541632
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.625742256459638e-05,
      "loss": 2.5269,
      "theoretical_loss": 3.4012649179252965,
      "tokens_seen": 2217672704
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.624939817043813e-05,
      "loss": 2.4656,
      "theoretical_loss": 3.4012484958705396,
      "tokens_seen": 2217803776
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.62413737762799e-05,
      "loss": 2.4634,
      "theoretical_loss": 3.40123207505803,
      "tokens_seen": 2217934848
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.623334938212165e-05,
      "loss": 2.5549,
      "theoretical_loss": 3.401215655487599,
      "tokens_seen": 2218065920
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.622532498796342e-05,
      "loss": 2.4937,
      "theoretical_loss": 3.4011992371590805,
      "tokens_seen": 2218196992
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.621730059380517e-05,
      "loss": 2.5013,
      "theoretical_loss": 3.4011828200723064,
      "tokens_seen": 2218328064
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.620927619964692e-05,
      "loss": 2.6017,
      "theoretical_loss": 3.40116640422711,
      "tokens_seen": 2218459136
    },
    {
      "epoch": 0.34,
      "objective/train/docs_used": 1223993,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.066260814666748,
      "objective/train/theoretical_loss": 3.4011581967700506,
      "objective/train/tokens_used": 588983776,
      "theoretical_loss": 3.4011581967700506,
      "tokens_seen": 2218524672
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.620125180548869e-05,
      "loss": 2.5717,
      "theoretical_loss": 3.401149989623323,
      "tokens_seen": 2218590208
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.619322741133044e-05,
      "loss": 2.7128,
      "theoretical_loss": 3.4011335762607797,
      "tokens_seen": 2218721280
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.618520301717221e-05,
      "loss": 2.6112,
      "theoretical_loss": 3.401117164139312,
      "tokens_seen": 2218852352
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.617717862301396e-05,
      "loss": 2.5601,
      "theoretical_loss": 3.4011007532587527,
      "tokens_seen": 2218983424
    },
    {
      "epoch": 0.34,
      "learning_rate": 6.616915422885573e-05,
      "loss": 2.466,
      "theoretical_loss": 3.401084343618935,
      "tokens_seen": 2219114496
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.616112983469748e-05,
      "loss": 2.5962,
      "theoretical_loss": 3.4010679352196918,
      "tokens_seen": 2219245568
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.615310544053924e-05,
      "loss": 2.6519,
      "theoretical_loss": 3.4010515280608558,
      "tokens_seen": 2219376640
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.6145081046381e-05,
      "loss": 2.6243,
      "theoretical_loss": 3.40103512214226,
      "tokens_seen": 2219507712
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.613705665222276e-05,
      "loss": 2.4531,
      "theoretical_loss": 3.401018717463738,
      "tokens_seen": 2219638784
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.612903225806452e-05,
      "loss": 2.6186,
      "theoretical_loss": 3.401002314025122,
      "tokens_seen": 2219769856
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.612100786390628e-05,
      "loss": 2.3972,
      "theoretical_loss": 3.4009859118262455,
      "tokens_seen": 2219900928
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.611298346974803e-05,
      "loss": 2.6535,
      "theoretical_loss": 3.400969510866942,
      "tokens_seen": 2220032000
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1224628,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7098729610443115,
      "objective/train/theoretical_loss": 3.400953111147044,
      "objective/train/tokens_used": 590622176,
      "theoretical_loss": 3.400953111147044,
      "tokens_seen": 2220163072
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.61049590755898e-05,
      "loss": 2.4975,
      "theoretical_loss": 3.400953111147044,
      "tokens_seen": 2220163072
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.609693468143155e-05,
      "loss": 2.5683,
      "theoretical_loss": 3.4009367126663848,
      "tokens_seen": 2220294144
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.608891028727332e-05,
      "loss": 2.6508,
      "theoretical_loss": 3.4009203154247976,
      "tokens_seen": 2220425216
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.608088589311507e-05,
      "loss": 2.6241,
      "theoretical_loss": 3.4009039194221162,
      "tokens_seen": 2220556288
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.607286149895684e-05,
      "loss": 2.5669,
      "theoretical_loss": 3.4008875246581733,
      "tokens_seen": 2220687360
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.606483710479859e-05,
      "loss": 2.4826,
      "theoretical_loss": 3.400871131132802,
      "tokens_seen": 2220818432
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.605681271064034e-05,
      "loss": 2.3772,
      "theoretical_loss": 3.4008547388458363,
      "tokens_seen": 2220949504
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.604878831648211e-05,
      "loss": 2.6465,
      "theoretical_loss": 3.400838347797109,
      "tokens_seen": 2221080576
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.604076392232386e-05,
      "loss": 2.5433,
      "theoretical_loss": 3.4008219579864543,
      "tokens_seen": 2221211648
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.603273952816563e-05,
      "loss": 2.4825,
      "theoretical_loss": 3.4008055694137047,
      "tokens_seen": 2221342720
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.602471513400738e-05,
      "loss": 2.5396,
      "theoretical_loss": 3.400789182078694,
      "tokens_seen": 2221473792
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.601669073984913e-05,
      "loss": 2.5519,
      "theoretical_loss": 3.400772795981256,
      "tokens_seen": 2221604864
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.60086663456909e-05,
      "loss": 2.5309,
      "theoretical_loss": 3.4007564111212236,
      "tokens_seen": 2221735936
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1225688,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.183415651321411,
      "objective/train/theoretical_loss": 3.400748219155183,
      "objective/train/tokens_used": 592260576,
      "theoretical_loss": 3.400748219155183,
      "tokens_seen": 2221801472
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.600064195153265e-05,
      "loss": 2.4961,
      "theoretical_loss": 3.400740027498431,
      "tokens_seen": 2221867008
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.599261755737442e-05,
      "loss": 2.662,
      "theoretical_loss": 3.400723645112712,
      "tokens_seen": 2221998080
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.598459316321617e-05,
      "loss": 2.5004,
      "theoretical_loss": 3.400707263963899,
      "tokens_seen": 2222129152
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.597656876905794e-05,
      "loss": 2.5408,
      "theoretical_loss": 3.400690884051827,
      "tokens_seen": 2222260224
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.59685443748997e-05,
      "loss": 2.4789,
      "theoretical_loss": 3.4006745053763288,
      "tokens_seen": 2222391296
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.596051998074145e-05,
      "loss": 2.5802,
      "theoretical_loss": 3.4006581279372385,
      "tokens_seen": 2222522368
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.595249558658321e-05,
      "loss": 2.4772,
      "theoretical_loss": 3.4006417517343897,
      "tokens_seen": 2222653440
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.594447119242497e-05,
      "loss": 2.7164,
      "theoretical_loss": 3.4006253767676164,
      "tokens_seen": 2222784512
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.593644679826673e-05,
      "loss": 2.5849,
      "theoretical_loss": 3.4006090030367524,
      "tokens_seen": 2222915584
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.592842240410849e-05,
      "loss": 2.5573,
      "theoretical_loss": 3.4005926305416314,
      "tokens_seen": 2223046656
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.592039800995025e-05,
      "loss": 2.5864,
      "theoretical_loss": 3.4005762592820874,
      "tokens_seen": 2223177728
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.5912373615792e-05,
      "loss": 2.4563,
      "theoretical_loss": 3.4005598892579543,
      "tokens_seen": 2223308800
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1226788,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7860403060913086,
      "objective/train/theoretical_loss": 3.400543520469066,
      "objective/train/tokens_used": 593898976,
      "theoretical_loss": 3.400543520469066,
      "tokens_seen": 2223439872
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.590434922163376e-05,
      "loss": 2.7357,
      "theoretical_loss": 3.400543520469066,
      "tokens_seen": 2223439872
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.589632482747553e-05,
      "loss": 2.5975,
      "theoretical_loss": 3.4005271529152563,
      "tokens_seen": 2223570944
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.588830043331728e-05,
      "loss": 2.5676,
      "theoretical_loss": 3.4005107865963597,
      "tokens_seen": 2223702016
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.588027603915905e-05,
      "loss": 2.3776,
      "theoretical_loss": 3.4004944215122097,
      "tokens_seen": 2223833088
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.58722516450008e-05,
      "loss": 2.4822,
      "theoretical_loss": 3.400478057662641,
      "tokens_seen": 2223964160
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.586422725084257e-05,
      "loss": 2.6994,
      "theoretical_loss": 3.400461695047487,
      "tokens_seen": 2224095232
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.585620285668432e-05,
      "loss": 2.5839,
      "theoretical_loss": 3.4004453336665827,
      "tokens_seen": 2224226304
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.584817846252607e-05,
      "loss": 2.5777,
      "theoretical_loss": 3.4004289735197615,
      "tokens_seen": 2224357376
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.584015406836784e-05,
      "loss": 2.6066,
      "theoretical_loss": 3.4004126146068576,
      "tokens_seen": 2224488448
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.583212967420959e-05,
      "loss": 2.553,
      "theoretical_loss": 3.4003962569277055,
      "tokens_seen": 2224619520
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.582410528005136e-05,
      "loss": 2.5847,
      "theoretical_loss": 3.40037990048214,
      "tokens_seen": 2224750592
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.581608088589311e-05,
      "loss": 2.6418,
      "theoretical_loss": 3.4003635452699945,
      "tokens_seen": 2224881664
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.580805649173488e-05,
      "loss": 2.5826,
      "theoretical_loss": 3.400347191291104,
      "tokens_seen": 2225012736
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1227393,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0688304901123047,
      "objective/train/theoretical_loss": 3.4003390147640773,
      "objective/train/tokens_used": 595537376,
      "theoretical_loss": 3.4003390147640773,
      "tokens_seen": 2225078272
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.580003209757663e-05,
      "loss": 2.7564,
      "theoretical_loss": 3.4003308385453024,
      "tokens_seen": 2225143808
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.579200770341838e-05,
      "loss": 2.5797,
      "theoretical_loss": 3.400314487032424,
      "tokens_seen": 2225274880
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.578398330926015e-05,
      "loss": 2.5858,
      "theoretical_loss": 3.400298136752304,
      "tokens_seen": 2225405952
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.57759589151019e-05,
      "loss": 2.5401,
      "theoretical_loss": 3.400281787704776,
      "tokens_seen": 2225537024
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.576793452094367e-05,
      "loss": 2.6559,
      "theoretical_loss": 3.4002654398896754,
      "tokens_seen": 2225668096
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.575991012678542e-05,
      "loss": 2.4982,
      "theoretical_loss": 3.4002490933068357,
      "tokens_seen": 2225799168
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.575188573262719e-05,
      "loss": 2.4992,
      "theoretical_loss": 3.4002327479560925,
      "tokens_seen": 2225930240
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.574386133846894e-05,
      "loss": 2.5146,
      "theoretical_loss": 3.4002164038372795,
      "tokens_seen": 2226061312
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.57358369443107e-05,
      "loss": 2.6021,
      "theoretical_loss": 3.4002000609502314,
      "tokens_seen": 2226192384
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.572781255015246e-05,
      "loss": 2.5102,
      "theoretical_loss": 3.4001837192947835,
      "tokens_seen": 2226323456
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.571978815599422e-05,
      "loss": 2.5191,
      "theoretical_loss": 3.40016737887077,
      "tokens_seen": 2226454528
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.571176376183598e-05,
      "loss": 2.5819,
      "theoretical_loss": 3.4001510396780255,
      "tokens_seen": 2226585600
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1228788,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.50097393989563,
      "objective/train/theoretical_loss": 3.400134701716385,
      "objective/train/tokens_used": 597175776,
      "theoretical_loss": 3.400134701716385,
      "tokens_seen": 2226716672
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.570373936767774e-05,
      "loss": 2.5784,
      "theoretical_loss": 3.400134701716385,
      "tokens_seen": 2226716672
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.56957149735195e-05,
      "loss": 2.359,
      "theoretical_loss": 3.400118364985684,
      "tokens_seen": 2226847744
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.568769057936126e-05,
      "loss": 2.6825,
      "theoretical_loss": 3.4001020294857556,
      "tokens_seen": 2226978816
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.567966618520301e-05,
      "loss": 2.4631,
      "theoretical_loss": 3.400085695216436,
      "tokens_seen": 2227109888
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.567164179104478e-05,
      "loss": 2.4789,
      "theoretical_loss": 3.40006936217756,
      "tokens_seen": 2227240960
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.566361739688653e-05,
      "loss": 2.3606,
      "theoretical_loss": 3.400053030368962,
      "tokens_seen": 2227372032
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.56555930027283e-05,
      "loss": 2.4429,
      "theoretical_loss": 3.400036699790477,
      "tokens_seen": 2227503104
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.564756860857005e-05,
      "loss": 2.398,
      "theoretical_loss": 3.4000203704419403,
      "tokens_seen": 2227634176
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.563954421441182e-05,
      "loss": 2.6781,
      "theoretical_loss": 3.4000040423231863,
      "tokens_seen": 2227765248
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.563151982025357e-05,
      "loss": 2.4473,
      "theoretical_loss": 3.399987715434051,
      "tokens_seen": 2227896320
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.562349542609534e-05,
      "loss": 2.6737,
      "theoretical_loss": 3.3999713897743686,
      "tokens_seen": 2228027392
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.561547103193709e-05,
      "loss": 2.5061,
      "theoretical_loss": 3.399955065343975,
      "tokens_seen": 2228158464
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.560744663777884e-05,
      "loss": 2.609,
      "theoretical_loss": 3.3999387421427043,
      "tokens_seen": 2228289536
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1229475,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.80594801902771,
      "objective/train/theoretical_loss": 3.3999305810029385,
      "objective/train/tokens_used": 598814176,
      "theoretical_loss": 3.3999305810029385,
      "tokens_seen": 2228355072
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.559942224362061e-05,
      "loss": 2.5804,
      "theoretical_loss": 3.3999224201703924,
      "tokens_seen": 2228420608
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.559139784946236e-05,
      "loss": 2.6122,
      "theoretical_loss": 3.3999060994268744,
      "tokens_seen": 2228551680
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.558337345530413e-05,
      "loss": 2.3885,
      "theoretical_loss": 3.399889779911985,
      "tokens_seen": 2228682752
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.557534906114588e-05,
      "loss": 2.5099,
      "theoretical_loss": 3.3998734616255604,
      "tokens_seen": 2228813824
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.556732466698765e-05,
      "loss": 2.57,
      "theoretical_loss": 3.3998571445674353,
      "tokens_seen": 2228944896
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.55593002728294e-05,
      "loss": 2.554,
      "theoretical_loss": 3.399840828737445,
      "tokens_seen": 2229075968
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.555127587867115e-05,
      "loss": 2.4743,
      "theoretical_loss": 3.3998245141354255,
      "tokens_seen": 2229207040
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.554325148451292e-05,
      "loss": 2.4917,
      "theoretical_loss": 3.3998082007612114,
      "tokens_seen": 2229338112
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.553522709035467e-05,
      "loss": 2.6752,
      "theoretical_loss": 3.3997918886146383,
      "tokens_seen": 2229469184
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.552720269619644e-05,
      "loss": 2.4935,
      "theoretical_loss": 3.399775577695541,
      "tokens_seen": 2229600256
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.55191783020382e-05,
      "loss": 2.5293,
      "theoretical_loss": 3.3997592680037565,
      "tokens_seen": 2229731328
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.551115390787996e-05,
      "loss": 2.4314,
      "theoretical_loss": 3.3997429595391195,
      "tokens_seen": 2229862400
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1230697,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2199413776397705,
      "objective/train/theoretical_loss": 3.3997266523014655,
      "objective/train/tokens_used": 600452576,
      "theoretical_loss": 3.3997266523014655,
      "tokens_seen": 2229993472
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.550312951372171e-05,
      "loss": 2.3866,
      "theoretical_loss": 3.3997266523014655,
      "tokens_seen": 2229993472
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.549510511956347e-05,
      "loss": 2.6148,
      "theoretical_loss": 3.39971034629063,
      "tokens_seen": 2230124544
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.548708072540523e-05,
      "loss": 2.4489,
      "theoretical_loss": 3.399694041506449,
      "tokens_seen": 2230255616
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.547905633124699e-05,
      "loss": 2.6336,
      "theoretical_loss": 3.3996777379487577,
      "tokens_seen": 2230386688
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.547103193708875e-05,
      "loss": 2.5653,
      "theoretical_loss": 3.3996614356173924,
      "tokens_seen": 2230517760
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.54630075429305e-05,
      "loss": 2.5572,
      "theoretical_loss": 3.3996451345121876,
      "tokens_seen": 2230648832
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.545498314877227e-05,
      "loss": 2.4734,
      "theoretical_loss": 3.3996288346329804,
      "tokens_seen": 2230779904
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.544695875461403e-05,
      "loss": 2.5562,
      "theoretical_loss": 3.399612535979606,
      "tokens_seen": 2230910976
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.543893436045578e-05,
      "loss": 2.5883,
      "theoretical_loss": 3.3995962385519,
      "tokens_seen": 2231042048
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.543090996629755e-05,
      "loss": 2.6082,
      "theoretical_loss": 3.399579942349698,
      "tokens_seen": 2231173120
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.54228855721393e-05,
      "loss": 2.5796,
      "theoretical_loss": 3.399563647372837,
      "tokens_seen": 2231304192
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.541486117798107e-05,
      "loss": 2.5459,
      "theoretical_loss": 3.3995473536211516,
      "tokens_seen": 2231435264
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.540683678382282e-05,
      "loss": 2.6118,
      "theoretical_loss": 3.3995310610944784,
      "tokens_seen": 2231566336
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1231344,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.63580322265625,
      "objective/train/theoretical_loss": 3.3995229152904702,
      "objective/train/tokens_used": 602090976,
      "theoretical_loss": 3.3995229152904702,
      "tokens_seen": 2231631872
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.539881238966459e-05,
      "loss": 2.5398,
      "theoretical_loss": 3.3995147697926535,
      "tokens_seen": 2231697408
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.539078799550634e-05,
      "loss": 2.4789,
      "theoretical_loss": 3.399498479715512,
      "tokens_seen": 2231828480
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.538276360134809e-05,
      "loss": 2.7088,
      "theoretical_loss": 3.399482190862891,
      "tokens_seen": 2231959552
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.537473920718986e-05,
      "loss": 2.4904,
      "theoretical_loss": 3.3994659032346264,
      "tokens_seen": 2232090624
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.536671481303161e-05,
      "loss": 2.6163,
      "theoretical_loss": 3.3994496168305535,
      "tokens_seen": 2232221696
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.535869041887338e-05,
      "loss": 2.6065,
      "theoretical_loss": 3.399433331650509,
      "tokens_seen": 2232352768
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.535066602471513e-05,
      "loss": 2.6058,
      "theoretical_loss": 3.399417047694329,
      "tokens_seen": 2232483840
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.53426416305569e-05,
      "loss": 2.5078,
      "theoretical_loss": 3.3994007649618494,
      "tokens_seen": 2232614912
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.533461723639865e-05,
      "loss": 2.5577,
      "theoretical_loss": 3.399384483452907,
      "tokens_seen": 2232745984
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.532659284224042e-05,
      "loss": 2.5347,
      "theoretical_loss": 3.3993682031673376,
      "tokens_seen": 2232877056
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.531856844808217e-05,
      "loss": 2.6003,
      "theoretical_loss": 3.399351924104977,
      "tokens_seen": 2233008128
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.531054405392392e-05,
      "loss": 2.4763,
      "theoretical_loss": 3.399335646265663,
      "tokens_seen": 2233139200
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1232383,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.546870708465576,
      "objective/train/theoretical_loss": 3.3993193696492305,
      "objective/train/tokens_used": 603729376,
      "theoretical_loss": 3.3993193696492305,
      "tokens_seen": 2233270272
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.530251965976569e-05,
      "loss": 2.5945,
      "theoretical_loss": 3.3993193696492305,
      "tokens_seen": 2233270272
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.529449526560744e-05,
      "loss": 2.5399,
      "theoretical_loss": 3.399303094255516,
      "tokens_seen": 2233401344
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.528647087144921e-05,
      "loss": 2.5345,
      "theoretical_loss": 3.3992868200843565,
      "tokens_seen": 2233532416
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.527844647729096e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.399270547135588,
      "tokens_seen": 2233663488
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.527042208313273e-05,
      "loss": 2.4072,
      "theoretical_loss": 3.3992542754090476,
      "tokens_seen": 2233794560
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.526239768897448e-05,
      "loss": 2.6307,
      "theoretical_loss": 3.399238004904571,
      "tokens_seen": 2233925632
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.525437329481624e-05,
      "loss": 2.5343,
      "theoretical_loss": 3.399221735621995,
      "tokens_seen": 2234056704
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.5246348900658e-05,
      "loss": 2.5621,
      "theoretical_loss": 3.3992054675611563,
      "tokens_seen": 2234187776
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.523832450649976e-05,
      "loss": 2.5342,
      "theoretical_loss": 3.3991892007218913,
      "tokens_seen": 2234318848
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.523030011234152e-05,
      "loss": 2.6233,
      "theoretical_loss": 3.399172935104037,
      "tokens_seen": 2234449920
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.522227571818328e-05,
      "loss": 2.464,
      "theoretical_loss": 3.3991566707074297,
      "tokens_seen": 2234580992
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.521425132402504e-05,
      "loss": 2.5425,
      "theoretical_loss": 3.399140407531906,
      "tokens_seen": 2234712064
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.52062269298668e-05,
      "loss": 2.6549,
      "theoretical_loss": 3.3991241455773027,
      "tokens_seen": 2234843136
    },
    {
      "epoch": 0.35,
      "objective/train/docs_used": 1233495,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.886728525161743,
      "objective/train/theoretical_loss": 3.3991160150577953,
      "objective/train/tokens_used": 605367776,
      "theoretical_loss": 3.3991160150577953,
      "tokens_seen": 2234908672
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.519820253570855e-05,
      "loss": 2.5483,
      "theoretical_loss": 3.3991078848434566,
      "tokens_seen": 2234974208
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.519017814155032e-05,
      "loss": 2.5349,
      "theoretical_loss": 3.399091625330205,
      "tokens_seen": 2235105280
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.518215374739207e-05,
      "loss": 2.5666,
      "theoretical_loss": 3.3990753670373834,
      "tokens_seen": 2235236352
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.517412935323384e-05,
      "loss": 2.554,
      "theoretical_loss": 3.3990591099648295,
      "tokens_seen": 2235367424
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.516610495907559e-05,
      "loss": 2.4128,
      "theoretical_loss": 3.3990428541123805,
      "tokens_seen": 2235498496
    },
    {
      "epoch": 0.35,
      "learning_rate": 6.515808056491736e-05,
      "loss": 2.6454,
      "theoretical_loss": 3.399026599479873,
      "tokens_seen": 2235629568
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.515005617075911e-05,
      "loss": 2.5925,
      "theoretical_loss": 3.3990103460671435,
      "tokens_seen": 2235760640
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.514203177660086e-05,
      "loss": 2.4704,
      "theoretical_loss": 3.398994093874029,
      "tokens_seen": 2235891712
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.513400738244263e-05,
      "loss": 2.4929,
      "theoretical_loss": 3.398977842900367,
      "tokens_seen": 2236022784
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.512598298828438e-05,
      "loss": 2.7658,
      "theoretical_loss": 3.3989615931459944,
      "tokens_seen": 2236153856
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.511795859412615e-05,
      "loss": 2.6872,
      "theoretical_loss": 3.3989453446107483,
      "tokens_seen": 2236284928
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.51099341999679e-05,
      "loss": 2.4934,
      "theoretical_loss": 3.3989290972944657,
      "tokens_seen": 2236416000
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1234013,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2317135334014893,
      "objective/train/theoretical_loss": 3.3989128511969833,
      "objective/train/tokens_used": 607006176,
      "theoretical_loss": 3.3989128511969833,
      "tokens_seen": 2236547072
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.510190980580967e-05,
      "loss": 2.5527,
      "theoretical_loss": 3.3989128511969833,
      "tokens_seen": 2236547072
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.509388541165142e-05,
      "loss": 2.527,
      "theoretical_loss": 3.398896606318139,
      "tokens_seen": 2236678144
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.508586101749317e-05,
      "loss": 2.5679,
      "theoretical_loss": 3.3988803626577697,
      "tokens_seen": 2236809216
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.507783662333494e-05,
      "loss": 2.615,
      "theoretical_loss": 3.398864120215712,
      "tokens_seen": 2236940288
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.50698122291767e-05,
      "loss": 2.3885,
      "theoretical_loss": 3.398847878991804,
      "tokens_seen": 2237071360
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.506178783501846e-05,
      "loss": 2.5451,
      "theoretical_loss": 3.398831638985883,
      "tokens_seen": 2237202432
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.505376344086021e-05,
      "loss": 2.547,
      "theoretical_loss": 3.398815400197785,
      "tokens_seen": 2237333504
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.504573904670198e-05,
      "loss": 2.504,
      "theoretical_loss": 3.3987991626273493,
      "tokens_seen": 2237464576
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.503771465254373e-05,
      "loss": 2.6316,
      "theoretical_loss": 3.3987829262744116,
      "tokens_seen": 2237595648
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.50296902583855e-05,
      "loss": 2.5986,
      "theoretical_loss": 3.3987666911388104,
      "tokens_seen": 2237726720
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.502166586422725e-05,
      "loss": 2.3704,
      "theoretical_loss": 3.398750457220382,
      "tokens_seen": 2237857792
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.501364147006901e-05,
      "loss": 2.733,
      "theoretical_loss": 3.3987342245189653,
      "tokens_seen": 2237988864
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.500561707591077e-05,
      "loss": 2.6359,
      "theoretical_loss": 3.398717993034397,
      "tokens_seen": 2238119936
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1235138,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.893874406814575,
      "objective/train/theoretical_loss": 3.39870987774838,
      "objective/train/tokens_used": 608644576,
      "theoretical_loss": 3.39870987774838,
      "tokens_seen": 2238185472
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.499759268175253e-05,
      "loss": 2.5401,
      "theoretical_loss": 3.3987017627665144,
      "tokens_seen": 2238251008
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.49895682875943e-05,
      "loss": 2.3885,
      "theoretical_loss": 3.398685533715155,
      "tokens_seen": 2238382080
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.498154389343605e-05,
      "loss": 2.6437,
      "theoretical_loss": 3.3986693058801576,
      "tokens_seen": 2238513152
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.497351949927781e-05,
      "loss": 2.5716,
      "theoretical_loss": 3.3986530792613583,
      "tokens_seen": 2238644224
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.496549510511957e-05,
      "loss": 2.6233,
      "theoretical_loss": 3.3986368538585956,
      "tokens_seen": 2238775296
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.495747071096132e-05,
      "loss": 2.4871,
      "theoretical_loss": 3.398620629671707,
      "tokens_seen": 2238906368
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.494944631680309e-05,
      "loss": 2.4061,
      "theoretical_loss": 3.3986044067005303,
      "tokens_seen": 2239037440
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.494142192264484e-05,
      "loss": 2.4802,
      "theoretical_loss": 3.398588184944903,
      "tokens_seen": 2239168512
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.49333975284866e-05,
      "loss": 2.4283,
      "theoretical_loss": 3.3985719644046632,
      "tokens_seen": 2239299584
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.492537313432836e-05,
      "loss": 2.5663,
      "theoretical_loss": 3.3985557450796486,
      "tokens_seen": 2239430656
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.491734874017013e-05,
      "loss": 2.4968,
      "theoretical_loss": 3.3985395269696967,
      "tokens_seen": 2239561728
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.490932434601188e-05,
      "loss": 2.6019,
      "theoretical_loss": 3.398523310074646,
      "tokens_seen": 2239692800
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1235870,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.631380558013916,
      "objective/train/theoretical_loss": 3.3985070943943336,
      "objective/train/tokens_used": 610282976,
      "theoretical_loss": 3.3985070943943336,
      "tokens_seen": 2239823872
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.490129995185363e-05,
      "loss": 2.4624,
      "theoretical_loss": 3.3985070943943336,
      "tokens_seen": 2239823872
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.48932755576954e-05,
      "loss": 2.6135,
      "theoretical_loss": 3.398490879928598,
      "tokens_seen": 2239954944
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.488525116353715e-05,
      "loss": 2.3966,
      "theoretical_loss": 3.398474666677277,
      "tokens_seen": 2240086016
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.487722676937892e-05,
      "loss": 2.5696,
      "theoretical_loss": 3.398458454640209,
      "tokens_seen": 2240217088
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.486920237522067e-05,
      "loss": 2.5395,
      "theoretical_loss": 3.3984422438172315,
      "tokens_seen": 2240348160
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.486117798106244e-05,
      "loss": 2.5404,
      "theoretical_loss": 3.3984260342081822,
      "tokens_seen": 2240479232
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.485315358690419e-05,
      "loss": 2.5419,
      "theoretical_loss": 3.3984098258129003,
      "tokens_seen": 2240610304
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.484512919274594e-05,
      "loss": 2.4847,
      "theoretical_loss": 3.398393618631223,
      "tokens_seen": 2240741376
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.483710479858771e-05,
      "loss": 2.6941,
      "theoretical_loss": 3.398377412662989,
      "tokens_seen": 2240872448
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.482908040442946e-05,
      "loss": 2.4717,
      "theoretical_loss": 3.398361207908036,
      "tokens_seen": 2241003520
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.482105601027123e-05,
      "loss": 2.543,
      "theoretical_loss": 3.3983450043662025,
      "tokens_seen": 2241134592
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.481303161611298e-05,
      "loss": 2.4651,
      "theoretical_loss": 3.398328802037327,
      "tokens_seen": 2241265664
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.480500722195475e-05,
      "loss": 2.4543,
      "theoretical_loss": 3.398312600921247,
      "tokens_seen": 2241396736
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1236973,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.588179588317871,
      "objective/train/theoretical_loss": 3.3983045008179555,
      "objective/train/tokens_used": 611921376,
      "theoretical_loss": 3.3983045008179555,
      "tokens_seen": 2241462272
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.47969828277965e-05,
      "loss": 2.3791,
      "theoretical_loss": 3.3982964010178014,
      "tokens_seen": 2241527808
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.478895843363826e-05,
      "loss": 2.6522,
      "theoretical_loss": 3.398280202326829,
      "tokens_seen": 2241658880
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.478093403948002e-05,
      "loss": 2.4594,
      "theoretical_loss": 3.398264004848167,
      "tokens_seen": 2241789952
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.477290964532178e-05,
      "loss": 2.5694,
      "theoretical_loss": 3.3982478085816545,
      "tokens_seen": 2241921024
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.476488525116354e-05,
      "loss": 2.6404,
      "theoretical_loss": 3.3982316135271295,
      "tokens_seen": 2242052096
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.47568608570053e-05,
      "loss": 2.548,
      "theoretical_loss": 3.3982154196844316,
      "tokens_seen": 2242183168
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.474883646284706e-05,
      "loss": 2.5171,
      "theoretical_loss": 3.398199227053398,
      "tokens_seen": 2242314240
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.474081206868882e-05,
      "loss": 2.5647,
      "theoretical_loss": 3.3981830356338674,
      "tokens_seen": 2242445312
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.473278767453057e-05,
      "loss": 2.5136,
      "theoretical_loss": 3.3981668454256786,
      "tokens_seen": 2242576384
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.472476328037234e-05,
      "loss": 2.4726,
      "theoretical_loss": 3.3981506564286708,
      "tokens_seen": 2242707456
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.471673888621409e-05,
      "loss": 2.5908,
      "theoretical_loss": 3.3981344686426818,
      "tokens_seen": 2242838528
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.470871449205586e-05,
      "loss": 2.4583,
      "theoretical_loss": 3.3981182820675504,
      "tokens_seen": 2242969600
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1238252,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0139503479003906,
      "objective/train/theoretical_loss": 3.3981020967031155,
      "objective/train/tokens_used": 613559776,
      "theoretical_loss": 3.3981020967031155,
      "tokens_seen": 2243100672
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.470069009789761e-05,
      "loss": 2.596,
      "theoretical_loss": 3.3981020967031155,
      "tokens_seen": 2243100672
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.469266570373938e-05,
      "loss": 2.4517,
      "theoretical_loss": 3.3980859125492153,
      "tokens_seen": 2243231744
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.468464130958113e-05,
      "loss": 2.5418,
      "theoretical_loss": 3.398069729605689,
      "tokens_seen": 2243362816
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.46766169154229e-05,
      "loss": 2.5578,
      "theoretical_loss": 3.3980535478723755,
      "tokens_seen": 2243493888
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.466859252126465e-05,
      "loss": 2.5241,
      "theoretical_loss": 3.3980373673491133,
      "tokens_seen": 2243624960
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.46605681271064e-05,
      "loss": 2.5103,
      "theoretical_loss": 3.398021188035741,
      "tokens_seen": 2243756032
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.465254373294817e-05,
      "loss": 2.6019,
      "theoretical_loss": 3.398005009932098,
      "tokens_seen": 2243887104
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.464451933878992e-05,
      "loss": 2.4528,
      "theoretical_loss": 3.3979888330380232,
      "tokens_seen": 2244018176
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.463649494463169e-05,
      "loss": 2.4556,
      "theoretical_loss": 3.3979726573533546,
      "tokens_seen": 2244149248
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.462847055047344e-05,
      "loss": 2.6524,
      "theoretical_loss": 3.3979564828779325,
      "tokens_seen": 2244280320
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.462044615631521e-05,
      "loss": 2.6528,
      "theoretical_loss": 3.3979403096115948,
      "tokens_seen": 2244411392
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.461242176215696e-05,
      "loss": 2.5461,
      "theoretical_loss": 3.397924137554181,
      "tokens_seen": 2244542464
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.460439736799871e-05,
      "loss": 2.6122,
      "theoretical_loss": 3.3979079667055303,
      "tokens_seen": 2244673536
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1238737,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.071171522140503,
      "objective/train/theoretical_loss": 3.397899881734441,
      "objective/train/tokens_used": 615198176,
      "theoretical_loss": 3.397899881734441,
      "tokens_seen": 2244739072
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.459637297384048e-05,
      "loss": 2.5194,
      "theoretical_loss": 3.3978917970654816,
      "tokens_seen": 2244804608
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.458834857968223e-05,
      "loss": 2.5799,
      "theoretical_loss": 3.397875628633874,
      "tokens_seen": 2244935680
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.4580324185524e-05,
      "loss": 2.5241,
      "theoretical_loss": 3.397859461410546,
      "tokens_seen": 2245066752
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.457229979136575e-05,
      "loss": 2.4155,
      "theoretical_loss": 3.397843295395338,
      "tokens_seen": 2245197824
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.456427539720752e-05,
      "loss": 2.5593,
      "theoretical_loss": 3.3978271305880883,
      "tokens_seen": 2245328896
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.455625100304927e-05,
      "loss": 2.3937,
      "theoretical_loss": 3.3978109669886365,
      "tokens_seen": 2245459968
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.454822660889103e-05,
      "loss": 2.569,
      "theoretical_loss": 3.3977948045968214,
      "tokens_seen": 2245591040
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.45402022147328e-05,
      "loss": 2.6149,
      "theoretical_loss": 3.397778643412483,
      "tokens_seen": 2245722112
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.453217782057455e-05,
      "loss": 2.669,
      "theoretical_loss": 3.3977624834354603,
      "tokens_seen": 2245853184
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.452415342641631e-05,
      "loss": 2.5393,
      "theoretical_loss": 3.397746324665593,
      "tokens_seen": 2245984256
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.451612903225807e-05,
      "loss": 2.5196,
      "theoretical_loss": 3.3977301671027194,
      "tokens_seen": 2246115328
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.450810463809983e-05,
      "loss": 2.517,
      "theoretical_loss": 3.39771401074668,
      "tokens_seen": 2246246400
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1239806,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.199644088745117,
      "objective/train/theoretical_loss": 3.397697855597314,
      "objective/train/tokens_used": 616836576,
      "theoretical_loss": 3.397697855597314,
      "tokens_seen": 2246377472
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.450008024394159e-05,
      "loss": 2.5675,
      "theoretical_loss": 3.397697855597314,
      "tokens_seen": 2246377472
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.449205584978334e-05,
      "loss": 2.5408,
      "theoretical_loss": 3.3976817016544603,
      "tokens_seen": 2246508544
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.448403145562511e-05,
      "loss": 2.555,
      "theoretical_loss": 3.3976655489179595,
      "tokens_seen": 2246639616
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.447600706146686e-05,
      "loss": 2.4769,
      "theoretical_loss": 3.3976493973876503,
      "tokens_seen": 2246770688
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.446798266730863e-05,
      "loss": 2.5246,
      "theoretical_loss": 3.397633247063372,
      "tokens_seen": 2246901760
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.445995827315038e-05,
      "loss": 2.4229,
      "theoretical_loss": 3.3976170979449654,
      "tokens_seen": 2247032832
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.445193387899215e-05,
      "loss": 2.4574,
      "theoretical_loss": 3.397600950032269,
      "tokens_seen": 2247163904
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.44439094848339e-05,
      "loss": 2.6299,
      "theoretical_loss": 3.3975848033251226,
      "tokens_seen": 2247294976
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.443588509067565e-05,
      "loss": 2.6822,
      "theoretical_loss": 3.3975686578233666,
      "tokens_seen": 2247426048
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.442786069651742e-05,
      "loss": 2.513,
      "theoretical_loss": 3.3975525135268403,
      "tokens_seen": 2247557120
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.441983630235917e-05,
      "loss": 2.6912,
      "theoretical_loss": 3.3975363704353834,
      "tokens_seen": 2247688192
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.441181190820094e-05,
      "loss": 2.4947,
      "theoretical_loss": 3.397520228548836,
      "tokens_seen": 2247819264
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.440378751404269e-05,
      "loss": 2.5929,
      "theoretical_loss": 3.397504087867037,
      "tokens_seen": 2247950336
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1241075,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.425755739212036,
      "objective/train/theoretical_loss": 3.3974960179778684,
      "objective/train/tokens_used": 618474976,
      "theoretical_loss": 3.3974960179778684,
      "tokens_seen": 2248015872
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.439576311988446e-05,
      "loss": 2.5835,
      "theoretical_loss": 3.397487948389827,
      "tokens_seen": 2248081408
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.438773872572621e-05,
      "loss": 2.6561,
      "theoretical_loss": 3.3974718101170462,
      "tokens_seen": 2248212480
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.437971433156798e-05,
      "loss": 2.6205,
      "theoretical_loss": 3.397455673048534,
      "tokens_seen": 2248343552
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.437168993740973e-05,
      "loss": 2.5423,
      "theoretical_loss": 3.3974395371841304,
      "tokens_seen": 2248474624
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.436366554325149e-05,
      "loss": 2.643,
      "theoretical_loss": 3.397423402523675,
      "tokens_seen": 2248605696
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.435564114909325e-05,
      "loss": 2.5475,
      "theoretical_loss": 3.3974072690670085,
      "tokens_seen": 2248736768
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.4347616754935e-05,
      "loss": 2.5904,
      "theoretical_loss": 3.3973911368139706,
      "tokens_seen": 2248867840
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.433959236077677e-05,
      "loss": 2.4771,
      "theoretical_loss": 3.397375005764401,
      "tokens_seen": 2248998912
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.433156796661852e-05,
      "loss": 2.6246,
      "theoretical_loss": 3.397358875918141,
      "tokens_seen": 2249129984
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.432354357246029e-05,
      "loss": 2.6396,
      "theoretical_loss": 3.397342747275029,
      "tokens_seen": 2249261056
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.431551917830204e-05,
      "loss": 2.6124,
      "theoretical_loss": 3.3973266198349066,
      "tokens_seen": 2249392128
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.43074947841438e-05,
      "loss": 2.425,
      "theoretical_loss": 3.397310493597613,
      "tokens_seen": 2249523200
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1241620,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.306302547454834,
      "objective/train/theoretical_loss": 3.397294368562989,
      "objective/train/tokens_used": 620113376,
      "theoretical_loss": 3.397294368562989,
      "tokens_seen": 2249654272
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.429947038998556e-05,
      "loss": 2.5863,
      "theoretical_loss": 3.397294368562989,
      "tokens_seen": 2249654272
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.429144599582732e-05,
      "loss": 2.6603,
      "theoretical_loss": 3.3972782447308747,
      "tokens_seen": 2249785344
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.428342160166908e-05,
      "loss": 2.6064,
      "theoretical_loss": 3.3972621221011106,
      "tokens_seen": 2249916416
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.427539720751084e-05,
      "loss": 2.5384,
      "theoretical_loss": 3.3972460006735363,
      "tokens_seen": 2250047488
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.42673728133526e-05,
      "loss": 2.6714,
      "theoretical_loss": 3.397229880447993,
      "tokens_seen": 2250178560
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.425934841919436e-05,
      "loss": 2.4576,
      "theoretical_loss": 3.3972137614243203,
      "tokens_seen": 2250309632
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.425132402503611e-05,
      "loss": 2.6373,
      "theoretical_loss": 3.397197643602359,
      "tokens_seen": 2250440704
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.424329963087788e-05,
      "loss": 2.4951,
      "theoretical_loss": 3.3971815269819494,
      "tokens_seen": 2250571776
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.423527523671963e-05,
      "loss": 2.4305,
      "theoretical_loss": 3.3971654115629324,
      "tokens_seen": 2250702848
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.42272508425614e-05,
      "loss": 2.6251,
      "theoretical_loss": 3.397149297345148,
      "tokens_seen": 2250833920
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.421922644840315e-05,
      "loss": 2.6587,
      "theoretical_loss": 3.3971331843284367,
      "tokens_seen": 2250964992
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.421120205424492e-05,
      "loss": 2.6134,
      "theoretical_loss": 3.397117072512639,
      "tokens_seen": 2251096064
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.420317766008667e-05,
      "loss": 2.5786,
      "theoretical_loss": 3.397100961897596,
      "tokens_seen": 2251227136
    },
    {
      "epoch": 0.36,
      "objective/train/docs_used": 1242746,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.16178297996521,
      "objective/train/theoretical_loss": 3.3970929070403075,
      "objective/train/tokens_used": 621751776,
      "theoretical_loss": 3.3970929070403075,
      "tokens_seen": 2251292672
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.419515326592842e-05,
      "loss": 2.4616,
      "theoretical_loss": 3.3970848524831476,
      "tokens_seen": 2251358208
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.418712887177019e-05,
      "loss": 2.5961,
      "theoretical_loss": 3.397068744269135,
      "tokens_seen": 2251489280
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.417910447761194e-05,
      "loss": 2.5784,
      "theoretical_loss": 3.397052637255399,
      "tokens_seen": 2251620352
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.417108008345371e-05,
      "loss": 2.4737,
      "theoretical_loss": 3.3970365314417794,
      "tokens_seen": 2251751424
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.416305568929546e-05,
      "loss": 2.4705,
      "theoretical_loss": 3.397020426828118,
      "tokens_seen": 2251882496
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.415503129513723e-05,
      "loss": 2.4693,
      "theoretical_loss": 3.397004323414255,
      "tokens_seen": 2252013568
    },
    {
      "epoch": 0.36,
      "learning_rate": 6.414700690097898e-05,
      "loss": 2.5308,
      "theoretical_loss": 3.396988221200031,
      "tokens_seen": 2252144640
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.413898250682074e-05,
      "loss": 2.5019,
      "theoretical_loss": 3.3969721201852874,
      "tokens_seen": 2252275712
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.41309581126625e-05,
      "loss": 2.6541,
      "theoretical_loss": 3.3969560203698643,
      "tokens_seen": 2252406784
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.412293371850426e-05,
      "loss": 2.644,
      "theoretical_loss": 3.3969399217536034,
      "tokens_seen": 2252537856
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.411490932434602e-05,
      "loss": 2.3905,
      "theoretical_loss": 3.396923824336345,
      "tokens_seen": 2252668928
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.410688493018777e-05,
      "loss": 2.7238,
      "theoretical_loss": 3.39690772811793,
      "tokens_seen": 2252800000
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1243228,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.317286491394043,
      "objective/train/theoretical_loss": 3.3968916330982006,
      "objective/train/tokens_used": 623390176,
      "theoretical_loss": 3.3968916330982006,
      "tokens_seen": 2252931072
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.409886053602954e-05,
      "loss": 2.6101,
      "theoretical_loss": 3.3968916330982006,
      "tokens_seen": 2252931072
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.40908361418713e-05,
      "loss": 2.6364,
      "theoretical_loss": 3.396875539276996,
      "tokens_seen": 2253062144
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.408281174771305e-05,
      "loss": 2.591,
      "theoretical_loss": 3.3968594466541586,
      "tokens_seen": 2253193216
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.407478735355481e-05,
      "loss": 2.5017,
      "theoretical_loss": 3.3968433552295285,
      "tokens_seen": 2253324288
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.406676295939657e-05,
      "loss": 2.5772,
      "theoretical_loss": 3.3968272650029476,
      "tokens_seen": 2253455360
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.405873856523833e-05,
      "loss": 2.5291,
      "theoretical_loss": 3.3968111759742565,
      "tokens_seen": 2253586432
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.405071417108009e-05,
      "loss": 2.4687,
      "theoretical_loss": 3.3967950881432967,
      "tokens_seen": 2253717504
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.404268977692185e-05,
      "loss": 2.6126,
      "theoretical_loss": 3.396779001509909,
      "tokens_seen": 2253848576
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.403466538276361e-05,
      "loss": 2.5522,
      "theoretical_loss": 3.3967629160739348,
      "tokens_seen": 2253979648
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.402664098860537e-05,
      "loss": 2.7067,
      "theoretical_loss": 3.3967468318352156,
      "tokens_seen": 2254110720
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.401861659444713e-05,
      "loss": 2.6029,
      "theoretical_loss": 3.3967307487935923,
      "tokens_seen": 2254241792
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.401059220028888e-05,
      "loss": 2.5112,
      "theoretical_loss": 3.3967146669489066,
      "tokens_seen": 2254372864
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.400256780613065e-05,
      "loss": 2.4022,
      "theoretical_loss": 3.3966985863009995,
      "tokens_seen": 2254503936
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1244407,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6968259811401367,
      "objective/train/theoretical_loss": 3.396690546425788,
      "objective/train/tokens_used": 625028576,
      "theoretical_loss": 3.396690546425788,
      "tokens_seen": 2254569472
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.39945434119724e-05,
      "loss": 2.5944,
      "theoretical_loss": 3.396682506849712,
      "tokens_seen": 2254635008
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.398651901781417e-05,
      "loss": 2.4521,
      "theoretical_loss": 3.3966664285948864,
      "tokens_seen": 2254766080
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.397849462365592e-05,
      "loss": 2.621,
      "theoretical_loss": 3.3966503515363637,
      "tokens_seen": 2254897152
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.397047022949769e-05,
      "loss": 2.6413,
      "theoretical_loss": 3.3966342756739856,
      "tokens_seen": 2255028224
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.396244583533944e-05,
      "loss": 2.6067,
      "theoretical_loss": 3.396618201007593,
      "tokens_seen": 2255159296
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.395442144118119e-05,
      "loss": 2.583,
      "theoretical_loss": 3.3966021275370277,
      "tokens_seen": 2255290368
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.394639704702296e-05,
      "loss": 2.5455,
      "theoretical_loss": 3.3965860552621314,
      "tokens_seen": 2255421440
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.393837265286471e-05,
      "loss": 2.5062,
      "theoretical_loss": 3.3965699841827455,
      "tokens_seen": 2255552512
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.393034825870648e-05,
      "loss": 2.5036,
      "theoretical_loss": 3.3965539142987122,
      "tokens_seen": 2255683584
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.392232386454823e-05,
      "loss": 2.5674,
      "theoretical_loss": 3.3965378456098723,
      "tokens_seen": 2255814656
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.391429947039e-05,
      "loss": 2.6788,
      "theoretical_loss": 3.3965217781160675,
      "tokens_seen": 2255945728
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.390627507623175e-05,
      "loss": 2.6405,
      "theoretical_loss": 3.39650571181714,
      "tokens_seen": 2256076800
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1245462,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.199613094329834,
      "objective/train/theoretical_loss": 3.3964896467129315,
      "objective/train/tokens_used": 626666976,
      "theoretical_loss": 3.3964896467129315,
      "tokens_seen": 2256207872
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.38982506820735e-05,
      "loss": 2.4546,
      "theoretical_loss": 3.3964896467129315,
      "tokens_seen": 2256207872
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.389022628791527e-05,
      "loss": 2.4392,
      "theoretical_loss": 3.3964735828032833,
      "tokens_seen": 2256338944
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.388220189375703e-05,
      "loss": 2.6483,
      "theoretical_loss": 3.396457520088038,
      "tokens_seen": 2256470016
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.387417749959879e-05,
      "loss": 2.6039,
      "theoretical_loss": 3.3964414585670366,
      "tokens_seen": 2256601088
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.386615310544055e-05,
      "loss": 2.6122,
      "theoretical_loss": 3.396425398240121,
      "tokens_seen": 2256732160
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.385812871128231e-05,
      "loss": 2.6314,
      "theoretical_loss": 3.3964093391071337,
      "tokens_seen": 2256863232
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.385010431712406e-05,
      "loss": 2.5326,
      "theoretical_loss": 3.3963932811679163,
      "tokens_seen": 2256994304
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.384207992296582e-05,
      "loss": 2.6952,
      "theoretical_loss": 3.3963772244223107,
      "tokens_seen": 2257125376
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.383405552880758e-05,
      "loss": 2.5161,
      "theoretical_loss": 3.3963611688701585,
      "tokens_seen": 2257256448
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.382603113464934e-05,
      "loss": 2.4891,
      "theoretical_loss": 3.3963451145113024,
      "tokens_seen": 2257387520
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.38180067404911e-05,
      "loss": 2.7088,
      "theoretical_loss": 3.396329061345584,
      "tokens_seen": 2257518592
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.380998234633286e-05,
      "loss": 2.6063,
      "theoretical_loss": 3.3963130093728453,
      "tokens_seen": 2257649664
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.380195795217462e-05,
      "loss": 2.7143,
      "theoretical_loss": 3.396296958592929,
      "tokens_seen": 2257780736
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1246138,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.176940679550171,
      "objective/train/theoretical_loss": 3.3962889336502293,
      "objective/train/tokens_used": 628305376,
      "theoretical_loss": 3.3962889336502293,
      "tokens_seen": 2257846272
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.379393355801638e-05,
      "loss": 2.6306,
      "theoretical_loss": 3.396280909005676,
      "tokens_seen": 2257911808
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.378590916385813e-05,
      "loss": 2.5263,
      "theoretical_loss": 3.39626486061093,
      "tokens_seen": 2258042880
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.37778847696999e-05,
      "loss": 2.6265,
      "theoretical_loss": 3.396248813408532,
      "tokens_seen": 2258173952
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.376986037554165e-05,
      "loss": 2.6052,
      "theoretical_loss": 3.396232767398325,
      "tokens_seen": 2258305024
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.376183598138342e-05,
      "loss": 2.4728,
      "theoretical_loss": 3.3962167225801503,
      "tokens_seen": 2258436096
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.375381158722517e-05,
      "loss": 2.5031,
      "theoretical_loss": 3.3962006789538512,
      "tokens_seen": 2258567168
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.374578719306694e-05,
      "loss": 2.6393,
      "theoretical_loss": 3.396184636519269,
      "tokens_seen": 2258698240
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.373776279890869e-05,
      "loss": 2.598,
      "theoretical_loss": 3.3961685952762473,
      "tokens_seen": 2258829312
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.372973840475044e-05,
      "loss": 2.5913,
      "theoretical_loss": 3.3961525552246274,
      "tokens_seen": 2258960384
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.372171401059221e-05,
      "loss": 2.694,
      "theoretical_loss": 3.396136516364252,
      "tokens_seen": 2259091456
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.371368961643396e-05,
      "loss": 2.6971,
      "theoretical_loss": 3.3961204786949635,
      "tokens_seen": 2259222528
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.370566522227573e-05,
      "loss": 2.4899,
      "theoretical_loss": 3.3961044422166045,
      "tokens_seen": 2259353600
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1246815,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6428818702697754,
      "objective/train/theoretical_loss": 3.3960884069290174,
      "objective/train/tokens_used": 629943776,
      "theoretical_loss": 3.3960884069290174,
      "tokens_seen": 2259484672
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.369764082811748e-05,
      "loss": 2.5699,
      "theoretical_loss": 3.3960884069290174,
      "tokens_seen": 2259484672
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.368961643395925e-05,
      "loss": 2.4269,
      "theoretical_loss": 3.396072372832044,
      "tokens_seen": 2259615744
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.3681592039801e-05,
      "loss": 2.5024,
      "theoretical_loss": 3.3960563399255284,
      "tokens_seen": 2259746816
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.367356764564276e-05,
      "loss": 2.4734,
      "theoretical_loss": 3.396040308209312,
      "tokens_seen": 2259877888
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.366554325148452e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.396024277683238,
      "tokens_seen": 2260008960
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.365751885732628e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.3960082483471483,
      "tokens_seen": 2260140032
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.364949446316804e-05,
      "loss": 2.6712,
      "theoretical_loss": 3.395992220200886,
      "tokens_seen": 2260271104
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.36414700690098e-05,
      "loss": 2.6961,
      "theoretical_loss": 3.395976193244294,
      "tokens_seen": 2260402176
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.363344567485155e-05,
      "loss": 2.6244,
      "theoretical_loss": 3.395960167477215,
      "tokens_seen": 2260533248
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.362542128069332e-05,
      "loss": 2.5395,
      "theoretical_loss": 3.395944142899491,
      "tokens_seen": 2260664320
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.361739688653507e-05,
      "loss": 2.6363,
      "theoretical_loss": 3.3959281195109656,
      "tokens_seen": 2260795392
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.360937249237683e-05,
      "loss": 2.57,
      "theoretical_loss": 3.3959120973114816,
      "tokens_seen": 2260926464
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.360134809821859e-05,
      "loss": 2.4798,
      "theoretical_loss": 3.395896076300881,
      "tokens_seen": 2261057536
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1247928,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4681272506713867,
      "objective/train/theoretical_loss": 3.3958880662413633,
      "objective/train/tokens_used": 631582176,
      "theoretical_loss": 3.3958880662413633,
      "tokens_seen": 2261123072
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.359332370406035e-05,
      "loss": 2.5529,
      "theoretical_loss": 3.3958800564790077,
      "tokens_seen": 2261188608
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.358529930990211e-05,
      "loss": 2.4942,
      "theoretical_loss": 3.3958640378457043,
      "tokens_seen": 2261319680
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.357727491574386e-05,
      "loss": 2.5433,
      "theoretical_loss": 3.3958480204008135,
      "tokens_seen": 2261450752
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.356925052158563e-05,
      "loss": 2.4331,
      "theoretical_loss": 3.395832004144178,
      "tokens_seen": 2261581824
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.356122612742738e-05,
      "loss": 2.4636,
      "theoretical_loss": 3.3958159890756416,
      "tokens_seen": 2261712896
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.355320173326915e-05,
      "loss": 2.5002,
      "theoretical_loss": 3.3957999751950467,
      "tokens_seen": 2261843968
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.35451773391109e-05,
      "loss": 2.5064,
      "theoretical_loss": 3.3957839625022364,
      "tokens_seen": 2261975040
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.353715294495265e-05,
      "loss": 2.4867,
      "theoretical_loss": 3.395767950997054,
      "tokens_seen": 2262106112
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.352912855079442e-05,
      "loss": 2.5062,
      "theoretical_loss": 3.3957519406793426,
      "tokens_seen": 2262237184
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.352110415663617e-05,
      "loss": 2.6116,
      "theoretical_loss": 3.3957359315489453,
      "tokens_seen": 2262368256
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.351307976247794e-05,
      "loss": 2.6306,
      "theoretical_loss": 3.395719923605705,
      "tokens_seen": 2262499328
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.35050553683197e-05,
      "loss": 2.4499,
      "theoretical_loss": 3.395703916849465,
      "tokens_seen": 2262630400
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1248487,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7316555976867676,
      "objective/train/theoretical_loss": 3.395687911280069,
      "objective/train/tokens_used": 633220576,
      "theoretical_loss": 3.395687911280069,
      "tokens_seen": 2262761472
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.349703097416146e-05,
      "loss": 2.6194,
      "theoretical_loss": 3.395687911280069,
      "tokens_seen": 2262761472
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.348900658000321e-05,
      "loss": 2.6273,
      "theoretical_loss": 3.39567190689736,
      "tokens_seen": 2262892544
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.348098218584497e-05,
      "loss": 2.4054,
      "theoretical_loss": 3.3956559037011806,
      "tokens_seen": 2263023616
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.347295779168673e-05,
      "loss": 2.6728,
      "theoretical_loss": 3.395639901691375,
      "tokens_seen": 2263154688
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.346493339752849e-05,
      "loss": 2.587,
      "theoretical_loss": 3.395623900867786,
      "tokens_seen": 2263285760
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.345690900337025e-05,
      "loss": 2.4727,
      "theoretical_loss": 3.395607901230258,
      "tokens_seen": 2263416832
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.3448884609212e-05,
      "loss": 2.5248,
      "theoretical_loss": 3.395591902778633,
      "tokens_seen": 2263547904
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.344086021505376e-05,
      "loss": 2.4696,
      "theoretical_loss": 3.3955759055127555,
      "tokens_seen": 2263678976
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.343283582089553e-05,
      "loss": 2.631,
      "theoretical_loss": 3.3955599094324684,
      "tokens_seen": 2263810048
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.342481142673728e-05,
      "loss": 2.5313,
      "theoretical_loss": 3.395543914537615,
      "tokens_seen": 2263941120
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.341678703257905e-05,
      "loss": 2.4899,
      "theoretical_loss": 3.3955279208280396,
      "tokens_seen": 2264072192
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.34087626384208e-05,
      "loss": 2.5598,
      "theoretical_loss": 3.395511928303585,
      "tokens_seen": 2264203264
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.340073824426257e-05,
      "loss": 2.5354,
      "theoretical_loss": 3.395495936964095,
      "tokens_seen": 2264334336
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1249541,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5170953273773193,
      "objective/train/theoretical_loss": 3.395487941738663,
      "objective/train/tokens_used": 634858976,
      "theoretical_loss": 3.395487941738663,
      "tokens_seen": 2264399872
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.339271385010432e-05,
      "loss": 2.5583,
      "theoretical_loss": 3.395479946809414,
      "tokens_seen": 2264465408
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.338468945594607e-05,
      "loss": 2.5488,
      "theoretical_loss": 3.3954639578393846,
      "tokens_seen": 2264596480
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.337666506178784e-05,
      "loss": 2.554,
      "theoretical_loss": 3.3954479700538505,
      "tokens_seen": 2264727552
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.336864066762959e-05,
      "loss": 2.5779,
      "theoretical_loss": 3.3954319834526565,
      "tokens_seen": 2264858624
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.336061627347136e-05,
      "loss": 2.4888,
      "theoretical_loss": 3.395415998035645,
      "tokens_seen": 2264989696
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.335259187931311e-05,
      "loss": 2.5179,
      "theoretical_loss": 3.3954000138026603,
      "tokens_seen": 2265120768
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.334456748515486e-05,
      "loss": 2.5611,
      "theoretical_loss": 3.395384030753547,
      "tokens_seen": 2265251840
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.333654309099663e-05,
      "loss": 2.5947,
      "theoretical_loss": 3.3953680488881472,
      "tokens_seen": 2265382912
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.332851869683838e-05,
      "loss": 2.6412,
      "theoretical_loss": 3.3953520682063063,
      "tokens_seen": 2265513984
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.332049430268015e-05,
      "loss": 2.4971,
      "theoretical_loss": 3.395336088707867,
      "tokens_seen": 2265645056
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.33124699085219e-05,
      "loss": 2.4861,
      "theoretical_loss": 3.3953201103926745,
      "tokens_seen": 2265776128
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.330444551436366e-05,
      "loss": 2.6189,
      "theoretical_loss": 3.395304133260572,
      "tokens_seen": 2265907200
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1250567,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.461559534072876,
      "objective/train/theoretical_loss": 3.3952881573114033,
      "objective/train/tokens_used": 636497376,
      "theoretical_loss": 3.3952881573114033,
      "tokens_seen": 2266038272
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.329642112020542e-05,
      "loss": 2.4765,
      "theoretical_loss": 3.3952881573114033,
      "tokens_seen": 2266038272
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.328839672604718e-05,
      "loss": 2.4616,
      "theoretical_loss": 3.3952721825450127,
      "tokens_seen": 2266169344
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.328037233188894e-05,
      "loss": 2.6036,
      "theoretical_loss": 3.395256208961244,
      "tokens_seen": 2266300416
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.32723479377307e-05,
      "loss": 2.481,
      "theoretical_loss": 3.3952402365599417,
      "tokens_seen": 2266431488
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.326432354357246e-05,
      "loss": 2.6836,
      "theoretical_loss": 3.3952242653409495,
      "tokens_seen": 2266562560
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.325629914941422e-05,
      "loss": 2.5822,
      "theoretical_loss": 3.3952082953041116,
      "tokens_seen": 2266693632
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.324827475525597e-05,
      "loss": 2.5497,
      "theoretical_loss": 3.395192326449272,
      "tokens_seen": 2266824704
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.324025036109774e-05,
      "loss": 2.5206,
      "theoretical_loss": 3.395176358776275,
      "tokens_seen": 2266955776
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.323222596693949e-05,
      "loss": 2.5478,
      "theoretical_loss": 3.3951603922849656,
      "tokens_seen": 2267086848
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.322420157278126e-05,
      "loss": 2.5811,
      "theoretical_loss": 3.3951444269751865,
      "tokens_seen": 2267217920
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.321617717862301e-05,
      "loss": 2.5322,
      "theoretical_loss": 3.3951284628467833,
      "tokens_seen": 2267348992
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.320815278446478e-05,
      "loss": 2.5236,
      "theoretical_loss": 3.395112499899599,
      "tokens_seen": 2267480064
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.320012839030653e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.3950965381334797,
      "tokens_seen": 2267611136
    },
    {
      "epoch": 0.37,
      "objective/train/docs_used": 1251582,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6694164276123047,
      "objective/train/theoretical_loss": 3.39508855769327,
      "objective/train/tokens_used": 638135776,
      "theoretical_loss": 3.39508855769327,
      "tokens_seen": 2267676672
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.319210399614828e-05,
      "loss": 2.4533,
      "theoretical_loss": 3.395080577548268,
      "tokens_seen": 2267742208
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.318407960199005e-05,
      "loss": 2.7162,
      "theoretical_loss": 3.3950646181438096,
      "tokens_seen": 2267873280
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.31760552078318e-05,
      "loss": 2.6106,
      "theoretical_loss": 3.395048659919948,
      "tokens_seen": 2268004352
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.316803081367357e-05,
      "loss": 2.4827,
      "theoretical_loss": 3.395032702876528,
      "tokens_seen": 2268135424
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.316000641951532e-05,
      "loss": 2.4875,
      "theoretical_loss": 3.3950167470133943,
      "tokens_seen": 2268266496
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.315198202535709e-05,
      "loss": 2.5619,
      "theoretical_loss": 3.3950007923303906,
      "tokens_seen": 2268397568
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.314395763119884e-05,
      "loss": 2.6272,
      "theoretical_loss": 3.3949848388273627,
      "tokens_seen": 2268528640
    },
    {
      "epoch": 0.37,
      "learning_rate": 6.31359332370406e-05,
      "loss": 2.3802,
      "theoretical_loss": 3.394968886504154,
      "tokens_seen": 2268659712
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.312790884288236e-05,
      "loss": 2.5548,
      "theoretical_loss": 3.39495293536061,
      "tokens_seen": 2268790784
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.311988444872411e-05,
      "loss": 2.6083,
      "theoretical_loss": 3.3949369853965745,
      "tokens_seen": 2268921856
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.311186005456588e-05,
      "loss": 2.6153,
      "theoretical_loss": 3.3949210366118923,
      "tokens_seen": 2269052928
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.310383566040763e-05,
      "loss": 2.4796,
      "theoretical_loss": 3.3949050890064085,
      "tokens_seen": 2269184000
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1252189,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5453248023986816,
      "objective/train/theoretical_loss": 3.394889142579968,
      "objective/train/tokens_used": 639774176,
      "theoretical_loss": 3.394889142579968,
      "tokens_seen": 2269315072
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.30958112662494e-05,
      "loss": 2.4309,
      "theoretical_loss": 3.394889142579968,
      "tokens_seen": 2269315072
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.308778687209115e-05,
      "loss": 2.503,
      "theoretical_loss": 3.394873197332415,
      "tokens_seen": 2269446144
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.307976247793292e-05,
      "loss": 2.5122,
      "theoretical_loss": 3.394857253263594,
      "tokens_seen": 2269577216
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.307173808377467e-05,
      "loss": 2.6993,
      "theoretical_loss": 3.3948413103733506,
      "tokens_seen": 2269708288
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.306371368961643e-05,
      "loss": 2.5629,
      "theoretical_loss": 3.394825368661529,
      "tokens_seen": 2269839360
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.30556892954582e-05,
      "loss": 2.4151,
      "theoretical_loss": 3.3948094281279744,
      "tokens_seen": 2269970432
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.304766490129995e-05,
      "loss": 2.6158,
      "theoretical_loss": 3.3947934887725317,
      "tokens_seen": 2270101504
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.303964050714171e-05,
      "loss": 2.5988,
      "theoretical_loss": 3.3947775505950455,
      "tokens_seen": 2270232576
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.303161611298347e-05,
      "loss": 2.4931,
      "theoretical_loss": 3.394761613595361,
      "tokens_seen": 2270363648
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.302359171882523e-05,
      "loss": 2.6879,
      "theoretical_loss": 3.394745677773323,
      "tokens_seen": 2270494720
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.301556732466699e-05,
      "loss": 2.4496,
      "theoretical_loss": 3.394729743128777,
      "tokens_seen": 2270625792
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.300754293050874e-05,
      "loss": 2.7436,
      "theoretical_loss": 3.394713809661568,
      "tokens_seen": 2270756864
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.29995185363505e-05,
      "loss": 2.5725,
      "theoretical_loss": 3.3946978773715397,
      "tokens_seen": 2270887936
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1253382,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.357855796813965,
      "objective/train/theoretical_loss": 3.394689911667921,
      "objective/train/tokens_used": 641412576,
      "theoretical_loss": 3.394689911667921,
      "tokens_seen": 2270953472
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.299149414219226e-05,
      "loss": 2.4205,
      "theoretical_loss": 3.394681946258539,
      "tokens_seen": 2271019008
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.298346974803403e-05,
      "loss": 2.5027,
      "theoretical_loss": 3.39466601632241,
      "tokens_seen": 2271150080
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.297544535387578e-05,
      "loss": 2.5171,
      "theoretical_loss": 3.3946500875629977,
      "tokens_seen": 2271281152
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.296742095971755e-05,
      "loss": 2.6532,
      "theoretical_loss": 3.394634159980148,
      "tokens_seen": 2271412224
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.29593965655593e-05,
      "loss": 2.6409,
      "theoretical_loss": 3.3946182335737056,
      "tokens_seen": 2271543296
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.295137217140105e-05,
      "loss": 2.5364,
      "theoretical_loss": 3.3946023083435164,
      "tokens_seen": 2271674368
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.294334777724282e-05,
      "loss": 2.5256,
      "theoretical_loss": 3.3945863842894246,
      "tokens_seen": 2271805440
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.293532338308457e-05,
      "loss": 2.5709,
      "theoretical_loss": 3.394570461411276,
      "tokens_seen": 2271936512
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.292729898892634e-05,
      "loss": 2.5675,
      "theoretical_loss": 3.3945545397089165,
      "tokens_seen": 2272067584
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.291927459476809e-05,
      "loss": 2.551,
      "theoretical_loss": 3.3945386191821907,
      "tokens_seen": 2272198656
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.291125020060986e-05,
      "loss": 2.593,
      "theoretical_loss": 3.3945226998309446,
      "tokens_seen": 2272329728
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.290322580645161e-05,
      "loss": 2.5302,
      "theoretical_loss": 3.3945067816550223,
      "tokens_seen": 2272460800
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1253927,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2348384857177734,
      "objective/train/theoretical_loss": 3.394490864654271,
      "objective/train/tokens_used": 643050976,
      "theoretical_loss": 3.394490864654271,
      "tokens_seen": 2272591872
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.289520141229336e-05,
      "loss": 2.5167,
      "theoretical_loss": 3.394490864654271,
      "tokens_seen": 2272591872
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.288717701813513e-05,
      "loss": 2.4235,
      "theoretical_loss": 3.394474948828535,
      "tokens_seen": 2272722944
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.287915262397688e-05,
      "loss": 2.3021,
      "theoretical_loss": 3.39445903417766,
      "tokens_seen": 2272854016
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.287112822981865e-05,
      "loss": 2.4075,
      "theoretical_loss": 3.3944431207014922,
      "tokens_seen": 2272985088
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.28631038356604e-05,
      "loss": 2.5877,
      "theoretical_loss": 3.3944272083998763,
      "tokens_seen": 2273116160
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.285507944150217e-05,
      "loss": 2.5949,
      "theoretical_loss": 3.394411297272659,
      "tokens_seen": 2273247232
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.284705504734392e-05,
      "loss": 2.6306,
      "theoretical_loss": 3.394395387319684,
      "tokens_seen": 2273378304
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.283903065318568e-05,
      "loss": 2.4229,
      "theoretical_loss": 3.394379478540799,
      "tokens_seen": 2273509376
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.283100625902744e-05,
      "loss": 2.5252,
      "theoretical_loss": 3.394363570935848,
      "tokens_seen": 2273640448
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.28229818648692e-05,
      "loss": 2.5287,
      "theoretical_loss": 3.3943476645046777,
      "tokens_seen": 2273771520
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.281495747071096e-05,
      "loss": 2.4188,
      "theoretical_loss": 3.394331759247134,
      "tokens_seen": 2273902592
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.280693307655272e-05,
      "loss": 2.4368,
      "theoretical_loss": 3.394315855163062,
      "tokens_seen": 2274033664
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.279890868239448e-05,
      "loss": 2.6502,
      "theoretical_loss": 3.394299952252308,
      "tokens_seen": 2274164736
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1255080,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.468050479888916,
      "objective/train/theoretical_loss": 3.3942920012368765,
      "objective/train/tokens_used": 644689376,
      "theoretical_loss": 3.3942920012368765,
      "tokens_seen": 2274230272
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.279088428823624e-05,
      "loss": 2.672,
      "theoretical_loss": 3.394284050514717,
      "tokens_seen": 2274295808
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.2782859894078e-05,
      "loss": 2.4558,
      "theoretical_loss": 3.394268149950136,
      "tokens_seen": 2274426880
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.277483549991976e-05,
      "loss": 2.3806,
      "theoretical_loss": 3.3942522505584103,
      "tokens_seen": 2274557952
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.276681110576151e-05,
      "loss": 2.5326,
      "theoretical_loss": 3.3942363523393855,
      "tokens_seen": 2274689024
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.275878671160328e-05,
      "loss": 2.4212,
      "theoretical_loss": 3.394220455292908,
      "tokens_seen": 2274820096
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.275076231744503e-05,
      "loss": 2.5278,
      "theoretical_loss": 3.394204559418824,
      "tokens_seen": 2274951168
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.27427379232868e-05,
      "loss": 2.5986,
      "theoretical_loss": 3.394188664716979,
      "tokens_seen": 2275082240
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.273471352912855e-05,
      "loss": 2.511,
      "theoretical_loss": 3.3941727711872187,
      "tokens_seen": 2275213312
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.272668913497032e-05,
      "loss": 2.452,
      "theoretical_loss": 3.3941568788293903,
      "tokens_seen": 2275344384
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.271866474081207e-05,
      "loss": 2.5879,
      "theoretical_loss": 3.394140987643339,
      "tokens_seen": 2275475456
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.271064034665382e-05,
      "loss": 2.6386,
      "theoretical_loss": 3.3941250976289106,
      "tokens_seen": 2275606528
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.270261595249559e-05,
      "loss": 2.5057,
      "theoretical_loss": 3.3941092087859523,
      "tokens_seen": 2275737600
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1255731,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5856258869171143,
      "objective/train/theoretical_loss": 3.3940933211143096,
      "objective/train/tokens_used": 646327776,
      "theoretical_loss": 3.3940933211143096,
      "tokens_seen": 2275868672
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.269459155833734e-05,
      "loss": 2.5352,
      "theoretical_loss": 3.3940933211143096,
      "tokens_seen": 2275868672
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.268656716417911e-05,
      "loss": 2.3659,
      "theoretical_loss": 3.3940774346138287,
      "tokens_seen": 2275999744
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.267854277002086e-05,
      "loss": 2.4207,
      "theoretical_loss": 3.394061549284356,
      "tokens_seen": 2276130816
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.267051837586263e-05,
      "loss": 2.5071,
      "theoretical_loss": 3.394045665125738,
      "tokens_seen": 2276261888
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.266249398170438e-05,
      "loss": 2.5107,
      "theoretical_loss": 3.39402978213782,
      "tokens_seen": 2276392960
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.265446958754613e-05,
      "loss": 2.4449,
      "theoretical_loss": 3.3940139003204495,
      "tokens_seen": 2276524032
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.26464451933879e-05,
      "loss": 2.4925,
      "theoretical_loss": 3.3939980196734725,
      "tokens_seen": 2276655104
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.263842079922965e-05,
      "loss": 2.4598,
      "theoretical_loss": 3.3939821401967354,
      "tokens_seen": 2276786176
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.263039640507142e-05,
      "loss": 2.4818,
      "theoretical_loss": 3.393966261890084,
      "tokens_seen": 2276917248
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.262237201091317e-05,
      "loss": 2.4662,
      "theoretical_loss": 3.3939503847533654,
      "tokens_seen": 2277048320
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.261434761675494e-05,
      "loss": 2.4891,
      "theoretical_loss": 3.3939345087864257,
      "tokens_seen": 2277179392
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.26063232225967e-05,
      "loss": 2.4176,
      "theoretical_loss": 3.3939186339891116,
      "tokens_seen": 2277310464
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.259829882843845e-05,
      "loss": 2.4436,
      "theoretical_loss": 3.3939027603612697,
      "tokens_seen": 2277441536
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1256864,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.405860424041748,
      "objective/train/theoretical_loss": 3.3938948239858524,
      "objective/train/tokens_used": 647966176,
      "theoretical_loss": 3.3938948239858524,
      "tokens_seen": 2277507072
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.259027443428021e-05,
      "loss": 2.5856,
      "theoretical_loss": 3.393886887902746,
      "tokens_seen": 2277572608
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.258225004012197e-05,
      "loss": 2.55,
      "theoretical_loss": 3.393871016613388,
      "tokens_seen": 2277703680
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.257422564596373e-05,
      "loss": 2.6433,
      "theoretical_loss": 3.393855146493041,
      "tokens_seen": 2277834752
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.256620125180549e-05,
      "loss": 2.5091,
      "theoretical_loss": 3.393839277541553,
      "tokens_seen": 2277965824
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.255817685764725e-05,
      "loss": 2.5015,
      "theoretical_loss": 3.3938234097587703,
      "tokens_seen": 2278096896
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.255015246348901e-05,
      "loss": 2.5052,
      "theoretical_loss": 3.393807543144539,
      "tokens_seen": 2278227968
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.254212806933076e-05,
      "loss": 2.4866,
      "theoretical_loss": 3.3937916776987063,
      "tokens_seen": 2278359040
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.253410367517253e-05,
      "loss": 2.4978,
      "theoretical_loss": 3.3937758134211187,
      "tokens_seen": 2278490112
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.252607928101428e-05,
      "loss": 2.413,
      "theoretical_loss": 3.3937599503116234,
      "tokens_seen": 2278621184
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.251805488685605e-05,
      "loss": 2.4349,
      "theoretical_loss": 3.3937440883700667,
      "tokens_seen": 2278752256
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.25100304926978e-05,
      "loss": 2.5007,
      "theoretical_loss": 3.393728227596296,
      "tokens_seen": 2278883328
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.250200609853957e-05,
      "loss": 2.5965,
      "theoretical_loss": 3.3937123679901573,
      "tokens_seen": 2279014400
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1257410,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6075518131256104,
      "objective/train/theoretical_loss": 3.393696509551498,
      "objective/train/tokens_used": 649604576,
      "theoretical_loss": 3.393696509551498,
      "tokens_seen": 2279145472
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.249398170438132e-05,
      "loss": 2.3826,
      "theoretical_loss": 3.393696509551498,
      "tokens_seen": 2279145472
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.248595731022309e-05,
      "loss": 2.5311,
      "theoretical_loss": 3.3936806522801657,
      "tokens_seen": 2279276544
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.247793291606484e-05,
      "loss": 2.5216,
      "theoretical_loss": 3.3936647961760062,
      "tokens_seen": 2279407616
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.246990852190659e-05,
      "loss": 2.5704,
      "theoretical_loss": 3.393648941238867,
      "tokens_seen": 2279538688
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.246188412774836e-05,
      "loss": 2.5134,
      "theoretical_loss": 3.3936330874685954,
      "tokens_seen": 2279669760
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.245385973359011e-05,
      "loss": 2.5912,
      "theoretical_loss": 3.393617234865038,
      "tokens_seen": 2279800832
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.244583533943188e-05,
      "loss": 2.4987,
      "theoretical_loss": 3.393601383428042,
      "tokens_seen": 2279931904
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.243781094527363e-05,
      "loss": 2.5887,
      "theoretical_loss": 3.393585533157455,
      "tokens_seen": 2280062976
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.24297865511154e-05,
      "loss": 2.4106,
      "theoretical_loss": 3.393569684053123,
      "tokens_seen": 2280194048
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.242176215695715e-05,
      "loss": 2.5223,
      "theoretical_loss": 3.393553836114894,
      "tokens_seen": 2280325120
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.24137377627989e-05,
      "loss": 2.5535,
      "theoretical_loss": 3.393537989342615,
      "tokens_seen": 2280456192
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.240571336864067e-05,
      "loss": 2.4859,
      "theoretical_loss": 3.393522143736133,
      "tokens_seen": 2280587264
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.239768897448242e-05,
      "loss": 2.3197,
      "theoretical_loss": 3.3935062992952956,
      "tokens_seen": 2280718336
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1258640,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.744055986404419,
      "objective/train/theoretical_loss": 3.393498377511946,
      "objective/train/tokens_used": 651242976,
      "theoretical_loss": 3.393498377511946,
      "tokens_seen": 2280783872
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.238966458032419e-05,
      "loss": 2.5835,
      "theoretical_loss": 3.39349045601995,
      "tokens_seen": 2280849408
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.238164018616594e-05,
      "loss": 2.4064,
      "theoretical_loss": 3.393474613909943,
      "tokens_seen": 2280980480
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.237361579200771e-05,
      "loss": 2.4895,
      "theoretical_loss": 3.393458772965123,
      "tokens_seen": 2281111552
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.236559139784946e-05,
      "loss": 2.4248,
      "theoretical_loss": 3.3934429331853364,
      "tokens_seen": 2281242624
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.235756700369122e-05,
      "loss": 2.419,
      "theoretical_loss": 3.3934270945704306,
      "tokens_seen": 2281373696
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.234954260953298e-05,
      "loss": 2.3879,
      "theoretical_loss": 3.3934112571202535,
      "tokens_seen": 2281504768
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.234151821537474e-05,
      "loss": 2.6913,
      "theoretical_loss": 3.3933954208346524,
      "tokens_seen": 2281635840
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.23334938212165e-05,
      "loss": 2.4367,
      "theoretical_loss": 3.3933795857134745,
      "tokens_seen": 2281766912
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.232546942705826e-05,
      "loss": 2.5412,
      "theoretical_loss": 3.3933637517565676,
      "tokens_seen": 2281897984
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.231744503290002e-05,
      "loss": 2.533,
      "theoretical_loss": 3.393347918963779,
      "tokens_seen": 2282029056
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.230942063874178e-05,
      "loss": 2.423,
      "theoretical_loss": 3.3933320873349566,
      "tokens_seen": 2282160128
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.230139624458353e-05,
      "loss": 2.5225,
      "theoretical_loss": 3.3933162568699475,
      "tokens_seen": 2282291200
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1259875,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7450172901153564,
      "objective/train/theoretical_loss": 3.3933004275686,
      "objective/train/tokens_used": 652881376,
      "theoretical_loss": 3.3933004275686,
      "tokens_seen": 2282422272
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.22933718504253e-05,
      "loss": 2.6164,
      "theoretical_loss": 3.3933004275686,
      "tokens_seen": 2282422272
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.228534745626705e-05,
      "loss": 2.4909,
      "theoretical_loss": 3.393284599430761,
      "tokens_seen": 2282553344
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.227732306210882e-05,
      "loss": 2.4819,
      "theoretical_loss": 3.3932687724562784,
      "tokens_seen": 2282684416
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.226929866795057e-05,
      "loss": 2.4915,
      "theoretical_loss": 3.3932529466450005,
      "tokens_seen": 2282815488
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.226127427379234e-05,
      "loss": 2.4388,
      "theoretical_loss": 3.3932371219967745,
      "tokens_seen": 2282946560
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.225324987963409e-05,
      "loss": 2.5142,
      "theoretical_loss": 3.393221298511448,
      "tokens_seen": 2283077632
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.224522548547584e-05,
      "loss": 2.6628,
      "theoretical_loss": 3.393205476188869,
      "tokens_seen": 2283208704
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.223720109131761e-05,
      "loss": 2.5043,
      "theoretical_loss": 3.393189655028885,
      "tokens_seen": 2283339776
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.222917669715936e-05,
      "loss": 2.5316,
      "theoretical_loss": 3.3931738350313445,
      "tokens_seen": 2283470848
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.222115230300113e-05,
      "loss": 2.7371,
      "theoretical_loss": 3.393158016196095,
      "tokens_seen": 2283601920
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.221312790884288e-05,
      "loss": 2.5326,
      "theoretical_loss": 3.393142198522984,
      "tokens_seen": 2283732992
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.220510351468465e-05,
      "loss": 2.5107,
      "theoretical_loss": 3.3931263820118605,
      "tokens_seen": 2283864064
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.21970791205264e-05,
      "loss": 2.4426,
      "theoretical_loss": 3.393110566662571,
      "tokens_seen": 2283995136
    },
    {
      "epoch": 0.38,
      "objective/train/docs_used": 1260339,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0385963916778564,
      "objective/train/theoretical_loss": 3.393102659423567,
      "objective/train/tokens_used": 654519776,
      "theoretical_loss": 3.393102659423567,
      "tokens_seen": 2284060672
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.218905472636816e-05,
      "loss": 2.5297,
      "theoretical_loss": 3.393094752474965,
      "tokens_seen": 2284126208
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.218103033220992e-05,
      "loss": 2.5378,
      "theoretical_loss": 3.39307893944889,
      "tokens_seen": 2284257280
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.217300593805168e-05,
      "loss": 2.6055,
      "theoretical_loss": 3.393063127584193,
      "tokens_seen": 2284388352
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.216498154389344e-05,
      "loss": 2.547,
      "theoretical_loss": 3.3930473168807236,
      "tokens_seen": 2284519424
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.21569571497352e-05,
      "loss": 2.4072,
      "theoretical_loss": 3.3930315073383293,
      "tokens_seen": 2284650496
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.214893275557696e-05,
      "loss": 2.4872,
      "theoretical_loss": 3.393015698956858,
      "tokens_seen": 2284781568
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.214090836141871e-05,
      "loss": 2.6139,
      "theoretical_loss": 3.3929998917361575,
      "tokens_seen": 2284912640
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.213288396726048e-05,
      "loss": 2.4484,
      "theoretical_loss": 3.392984085676077,
      "tokens_seen": 2285043712
    },
    {
      "epoch": 0.38,
      "learning_rate": 6.212485957310223e-05,
      "loss": 2.4945,
      "theoretical_loss": 3.3929682807764645,
      "tokens_seen": 2285174784
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.211683517894399e-05,
      "loss": 2.3858,
      "theoretical_loss": 3.3929524770371677,
      "tokens_seen": 2285305856
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.210881078478575e-05,
      "loss": 2.5794,
      "theoretical_loss": 3.392936674458035,
      "tokens_seen": 2285436928
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.210078639062751e-05,
      "loss": 2.5333,
      "theoretical_loss": 3.392920873038915,
      "tokens_seen": 2285568000
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1261840,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0126242637634277,
      "objective/train/theoretical_loss": 3.392905072779656,
      "objective/train/tokens_used": 656158176,
      "theoretical_loss": 3.392905072779656,
      "tokens_seen": 2285699072
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.209276199646927e-05,
      "loss": 2.6094,
      "theoretical_loss": 3.392905072779656,
      "tokens_seen": 2285699072
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.208473760231103e-05,
      "loss": 2.4284,
      "theoretical_loss": 3.392889273680106,
      "tokens_seen": 2285830144
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.20767132081528e-05,
      "loss": 2.6252,
      "theoretical_loss": 3.3928734757401138,
      "tokens_seen": 2285961216
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.206868881399455e-05,
      "loss": 2.5987,
      "theoretical_loss": 3.3928576789595275,
      "tokens_seen": 2286092288
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.20606644198363e-05,
      "loss": 2.5607,
      "theoretical_loss": 3.3928418833381957,
      "tokens_seen": 2286223360
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.205264002567807e-05,
      "loss": 2.545,
      "theoretical_loss": 3.392826088875967,
      "tokens_seen": 2286354432
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.204461563151982e-05,
      "loss": 2.576,
      "theoretical_loss": 3.3928102955726898,
      "tokens_seen": 2286485504
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.203659123736159e-05,
      "loss": 2.6459,
      "theoretical_loss": 3.3927945034282123,
      "tokens_seen": 2286616576
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.202856684320334e-05,
      "loss": 2.5031,
      "theoretical_loss": 3.3927787124423836,
      "tokens_seen": 2286747648
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.20205424490451e-05,
      "loss": 2.516,
      "theoretical_loss": 3.392762922615052,
      "tokens_seen": 2286878720
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.201251805488686e-05,
      "loss": 2.5389,
      "theoretical_loss": 3.3927471339460666,
      "tokens_seen": 2287009792
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.200449366072861e-05,
      "loss": 2.5867,
      "theoretical_loss": 3.392731346435275,
      "tokens_seen": 2287140864
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.199646926657038e-05,
      "loss": 2.6023,
      "theoretical_loss": 3.3927155600825265,
      "tokens_seen": 2287271936
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1262387,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.811039447784424,
      "objective/train/theoretical_loss": 3.3927076673403715,
      "objective/train/tokens_used": 657796576,
      "theoretical_loss": 3.3927076673403715,
      "tokens_seen": 2287337472
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.198844487241213e-05,
      "loss": 2.3361,
      "theoretical_loss": 3.39269977488767,
      "tokens_seen": 2287403008
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.19804204782539e-05,
      "loss": 2.3675,
      "theoretical_loss": 3.392683990850554,
      "tokens_seen": 2287534080
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.197239608409565e-05,
      "loss": 2.6312,
      "theoretical_loss": 3.3926682079710275,
      "tokens_seen": 2287665152
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.196437168993742e-05,
      "loss": 2.5818,
      "theoretical_loss": 3.3926524262489384,
      "tokens_seen": 2287796224
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.195634729577917e-05,
      "loss": 2.4892,
      "theoretical_loss": 3.392636645684137,
      "tokens_seen": 2287927296
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.194832290162093e-05,
      "loss": 2.4498,
      "theoretical_loss": 3.3926208662764705,
      "tokens_seen": 2288058368
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.194029850746269e-05,
      "loss": 2.5447,
      "theoretical_loss": 3.392605088025789,
      "tokens_seen": 2288189440
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.193227411330445e-05,
      "loss": 2.5413,
      "theoretical_loss": 3.3925893109319407,
      "tokens_seen": 2288320512
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.192424971914621e-05,
      "loss": 2.6062,
      "theoretical_loss": 3.392573534994775,
      "tokens_seen": 2288451584
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.191622532498797e-05,
      "loss": 2.5116,
      "theoretical_loss": 3.3925577602141406,
      "tokens_seen": 2288582656
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.190820093082973e-05,
      "loss": 2.5007,
      "theoretical_loss": 3.3925419865898867,
      "tokens_seen": 2288713728
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.190017653667148e-05,
      "loss": 2.4369,
      "theoretical_loss": 3.392526214121862,
      "tokens_seen": 2288844800
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1263579,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.539278745651245,
      "objective/train/theoretical_loss": 3.392510442809916,
      "objective/train/tokens_used": 659434976,
      "theoretical_loss": 3.392510442809916,
      "tokens_seen": 2288975872
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.189215214251324e-05,
      "loss": 2.6523,
      "theoretical_loss": 3.392510442809916,
      "tokens_seen": 2288975872
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.1884127748355e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.392494672653897,
      "tokens_seen": 2289106944
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.187610335419676e-05,
      "loss": 2.5141,
      "theoretical_loss": 3.392478903653655,
      "tokens_seen": 2289238016
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.186807896003852e-05,
      "loss": 2.6254,
      "theoretical_loss": 3.392463135809039,
      "tokens_seen": 2289369088
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.186005456588028e-05,
      "loss": 2.4007,
      "theoretical_loss": 3.392447369119897,
      "tokens_seen": 2289500160
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.185203017172204e-05,
      "loss": 2.4746,
      "theoretical_loss": 3.3924316035860795,
      "tokens_seen": 2289631232
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.18440057775638e-05,
      "loss": 2.6335,
      "theoretical_loss": 3.3924158392074353,
      "tokens_seen": 2289762304
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.183598138340556e-05,
      "loss": 2.5504,
      "theoretical_loss": 3.3924000759838133,
      "tokens_seen": 2289893376
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.182795698924732e-05,
      "loss": 2.5918,
      "theoretical_loss": 3.3923843139150636,
      "tokens_seen": 2290024448
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.181993259508907e-05,
      "loss": 2.541,
      "theoretical_loss": 3.3923685530010346,
      "tokens_seen": 2290155520
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.181190820093084e-05,
      "loss": 2.5115,
      "theoretical_loss": 3.3923527932415762,
      "tokens_seen": 2290286592
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.180388380677259e-05,
      "loss": 2.6476,
      "theoretical_loss": 3.392337034636537,
      "tokens_seen": 2290417664
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.179585941261436e-05,
      "loss": 2.6693,
      "theoretical_loss": 3.3923212771857676,
      "tokens_seen": 2290548736
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1264107,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.165374755859375,
      "objective/train/theoretical_loss": 3.3923133988931866,
      "objective/train/tokens_used": 661073376,
      "theoretical_loss": 3.3923133988931866,
      "tokens_seen": 2290614272
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.178783501845611e-05,
      "loss": 2.6599,
      "theoretical_loss": 3.392305520889116,
      "tokens_seen": 2290679808
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.177981062429788e-05,
      "loss": 2.5995,
      "theoretical_loss": 3.3922897657464333,
      "tokens_seen": 2290810880
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.177178623013963e-05,
      "loss": 2.5903,
      "theoretical_loss": 3.392274011757567,
      "tokens_seen": 2290941952
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.176376183598138e-05,
      "loss": 2.6104,
      "theoretical_loss": 3.3922582589223684,
      "tokens_seen": 2291073024
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.175573744182315e-05,
      "loss": 2.6337,
      "theoretical_loss": 3.392242507240686,
      "tokens_seen": 2291204096
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.17477130476649e-05,
      "loss": 2.5884,
      "theoretical_loss": 3.3922267567123696,
      "tokens_seen": 2291335168
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.173968865350667e-05,
      "loss": 2.495,
      "theoretical_loss": 3.3922110073372687,
      "tokens_seen": 2291466240
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.173166425934842e-05,
      "loss": 2.6121,
      "theoretical_loss": 3.392195259115233,
      "tokens_seen": 2291597312
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.172363986519019e-05,
      "loss": 2.612,
      "theoretical_loss": 3.3921795120461122,
      "tokens_seen": 2291728384
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.171561547103194e-05,
      "loss": 2.5074,
      "theoretical_loss": 3.392163766129756,
      "tokens_seen": 2291859456
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.17075910768737e-05,
      "loss": 2.4929,
      "theoretical_loss": 3.3921480213660136,
      "tokens_seen": 2291990528
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.169956668271546e-05,
      "loss": 2.528,
      "theoretical_loss": 3.3921322777547354,
      "tokens_seen": 2292121600
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1264661,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.496931314468384,
      "objective/train/theoretical_loss": 3.392116535295771,
      "objective/train/tokens_used": 662711776,
      "theoretical_loss": 3.392116535295771,
      "tokens_seen": 2292252672
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.169154228855722e-05,
      "loss": 2.5325,
      "theoretical_loss": 3.392116535295771,
      "tokens_seen": 2292252672
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.168351789439898e-05,
      "loss": 2.6184,
      "theoretical_loss": 3.39210079398897,
      "tokens_seen": 2292383744
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.167549350024074e-05,
      "loss": 2.4983,
      "theoretical_loss": 3.392085053834182,
      "tokens_seen": 2292514816
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.16674691060825e-05,
      "loss": 2.6247,
      "theoretical_loss": 3.3920693148312573,
      "tokens_seen": 2292645888
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.165944471192425e-05,
      "loss": 2.5484,
      "theoretical_loss": 3.392053576980045,
      "tokens_seen": 2292776960
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.165142031776601e-05,
      "loss": 2.5444,
      "theoretical_loss": 3.3920378402803957,
      "tokens_seen": 2292908032
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.164339592360777e-05,
      "loss": 2.4537,
      "theoretical_loss": 3.3920221047321597,
      "tokens_seen": 2293039104
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.163537152944953e-05,
      "loss": 2.4478,
      "theoretical_loss": 3.392006370335186,
      "tokens_seen": 2293170176
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.16273471352913e-05,
      "loss": 2.5498,
      "theoretical_loss": 3.3919906370893247,
      "tokens_seen": 2293301248
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.161932274113305e-05,
      "loss": 2.439,
      "theoretical_loss": 3.3919749049944263,
      "tokens_seen": 2293432320
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.161129834697481e-05,
      "loss": 2.4433,
      "theoretical_loss": 3.3919591740503408,
      "tokens_seen": 2293563392
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.160327395281657e-05,
      "loss": 2.5527,
      "theoretical_loss": 3.391943444256918,
      "tokens_seen": 2293694464
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.159524955865832e-05,
      "loss": 2.5234,
      "theoretical_loss": 3.3919277156140075,
      "tokens_seen": 2293825536
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1265892,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.169233560562134,
      "objective/train/theoretical_loss": 3.391919851723948,
      "objective/train/tokens_used": 664350176,
      "theoretical_loss": 3.391919851723948,
      "tokens_seen": 2293891072
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.158722516450009e-05,
      "loss": 2.5547,
      "theoretical_loss": 3.3919119881214606,
      "tokens_seen": 2293956608
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.157920077034184e-05,
      "loss": 2.4961,
      "theoretical_loss": 3.3918962617791264,
      "tokens_seen": 2294087680
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.157117637618361e-05,
      "loss": 2.6464,
      "theoretical_loss": 3.3918805365868554,
      "tokens_seen": 2294218752
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.156315198202536e-05,
      "loss": 2.5783,
      "theoretical_loss": 3.391864812544498,
      "tokens_seen": 2294349824
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.155512758786713e-05,
      "loss": 2.7335,
      "theoretical_loss": 3.3918490896519042,
      "tokens_seen": 2294480896
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.154710319370888e-05,
      "loss": 2.5165,
      "theoretical_loss": 3.3918333679089243,
      "tokens_seen": 2294611968
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.153907879955065e-05,
      "loss": 2.5589,
      "theoretical_loss": 3.391817647315409,
      "tokens_seen": 2294743040
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.15310544053924e-05,
      "loss": 2.4536,
      "theoretical_loss": 3.391801927871208,
      "tokens_seen": 2294874112
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.152303001123415e-05,
      "loss": 2.4815,
      "theoretical_loss": 3.3917862095761717,
      "tokens_seen": 2295005184
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.151500561707592e-05,
      "loss": 2.6744,
      "theoretical_loss": 3.3917704924301506,
      "tokens_seen": 2295136256
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.150698122291767e-05,
      "loss": 2.4871,
      "theoretical_loss": 3.3917547764329954,
      "tokens_seen": 2295267328
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.149895682875944e-05,
      "loss": 2.6185,
      "theoretical_loss": 3.3917390615845564,
      "tokens_seen": 2295398400
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1267119,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5357112884521484,
      "objective/train/theoretical_loss": 3.3917233478846835,
      "objective/train/tokens_used": 665988576,
      "theoretical_loss": 3.3917233478846835,
      "tokens_seen": 2295529472
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.149093243460119e-05,
      "loss": 2.4802,
      "theoretical_loss": 3.3917233478846835,
      "tokens_seen": 2295529472
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.148290804044296e-05,
      "loss": 2.473,
      "theoretical_loss": 3.3917076353332276,
      "tokens_seen": 2295660544
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.147488364628471e-05,
      "loss": 2.5155,
      "theoretical_loss": 3.39169192393004,
      "tokens_seen": 2295791616
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.146685925212647e-05,
      "loss": 2.5867,
      "theoretical_loss": 3.391676213674969,
      "tokens_seen": 2295922688
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.145883485796823e-05,
      "loss": 2.5223,
      "theoretical_loss": 3.391660504567868,
      "tokens_seen": 2296053760
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.145081046380999e-05,
      "loss": 2.5643,
      "theoretical_loss": 3.391644796608585,
      "tokens_seen": 2296184832
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.144278606965175e-05,
      "loss": 2.6168,
      "theoretical_loss": 3.3916290897969725,
      "tokens_seen": 2296315904
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.14347616754935e-05,
      "loss": 2.5584,
      "theoretical_loss": 3.39161338413288,
      "tokens_seen": 2296446976
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.142673728133527e-05,
      "loss": 2.5964,
      "theoretical_loss": 3.3915976796161593,
      "tokens_seen": 2296578048
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.141871288717702e-05,
      "loss": 2.495,
      "theoretical_loss": 3.39158197624666,
      "tokens_seen": 2296709120
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.141068849301878e-05,
      "loss": 2.5428,
      "theoretical_loss": 3.391566274024233,
      "tokens_seen": 2296840192
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.140266409886054e-05,
      "loss": 2.6739,
      "theoretical_loss": 3.39155057294873,
      "tokens_seen": 2296971264
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.13946397047023e-05,
      "loss": 2.479,
      "theoretical_loss": 3.3915348730200003,
      "tokens_seen": 2297102336
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1267727,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8644251823425293,
      "objective/train/theoretical_loss": 3.3915270234856294,
      "objective/train/tokens_used": 667626976,
      "theoretical_loss": 3.3915270234856294,
      "tokens_seen": 2297167872
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.138661531054406e-05,
      "loss": 2.6772,
      "theoretical_loss": 3.3915191742378963,
      "tokens_seen": 2297233408
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.137859091638582e-05,
      "loss": 2.6287,
      "theoretical_loss": 3.3915034766022676,
      "tokens_seen": 2297364480
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.137056652222758e-05,
      "loss": 2.5341,
      "theoretical_loss": 3.3914877801129655,
      "tokens_seen": 2297495552
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.136254212806934e-05,
      "loss": 2.6252,
      "theoretical_loss": 3.391472084769841,
      "tokens_seen": 2297626624
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.135451773391109e-05,
      "loss": 2.434,
      "theoretical_loss": 3.391456390572745,
      "tokens_seen": 2297757696
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.134649333975286e-05,
      "loss": 2.5343,
      "theoretical_loss": 3.3914406975215283,
      "tokens_seen": 2297888768
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.133846894559461e-05,
      "loss": 2.6846,
      "theoretical_loss": 3.391425005616042,
      "tokens_seen": 2298019840
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.133044455143638e-05,
      "loss": 2.6056,
      "theoretical_loss": 3.3914093148561375,
      "tokens_seen": 2298150912
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.132242015727813e-05,
      "loss": 2.6407,
      "theoretical_loss": 3.391393625241665,
      "tokens_seen": 2298281984
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.13143957631199e-05,
      "loss": 2.563,
      "theoretical_loss": 3.391377936772476,
      "tokens_seen": 2298413056
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.130637136896165e-05,
      "loss": 2.4517,
      "theoretical_loss": 3.391362249448422,
      "tokens_seen": 2298544128
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.12983469748034e-05,
      "loss": 2.4836,
      "theoretical_loss": 3.3913465632693534,
      "tokens_seen": 2298675200
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1268805,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6418638229370117,
      "objective/train/theoretical_loss": 3.3913308782351215,
      "objective/train/tokens_used": 669265376,
      "theoretical_loss": 3.3913308782351215,
      "tokens_seen": 2298806272
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.129032258064517e-05,
      "loss": 2.4528,
      "theoretical_loss": 3.3913308782351215,
      "tokens_seen": 2298806272
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.128229818648692e-05,
      "loss": 2.4908,
      "theoretical_loss": 3.391315194345578,
      "tokens_seen": 2298937344
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.127427379232869e-05,
      "loss": 2.5891,
      "theoretical_loss": 3.3912995116005735,
      "tokens_seen": 2299068416
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.126624939817044e-05,
      "loss": 2.568,
      "theoretical_loss": 3.3912838299999595,
      "tokens_seen": 2299199488
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.125822500401221e-05,
      "loss": 2.4487,
      "theoretical_loss": 3.391268149543587,
      "tokens_seen": 2299330560
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.125020060985396e-05,
      "loss": 2.4322,
      "theoretical_loss": 3.391252470231308,
      "tokens_seen": 2299461632
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.124217621569572e-05,
      "loss": 2.5052,
      "theoretical_loss": 3.391236792062973,
      "tokens_seen": 2299592704
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.123415182153748e-05,
      "loss": 2.5633,
      "theoretical_loss": 3.391221115038434,
      "tokens_seen": 2299723776
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.122612742737924e-05,
      "loss": 2.6071,
      "theoretical_loss": 3.3912054391575417,
      "tokens_seen": 2299854848
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.1218103033221e-05,
      "loss": 2.3632,
      "theoretical_loss": 3.3911897644201483,
      "tokens_seen": 2299985920
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.121007863906276e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.3911740908261043,
      "tokens_seen": 2300116992
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.120205424490452e-05,
      "loss": 2.4519,
      "theoretical_loss": 3.3911584183752623,
      "tokens_seen": 2300248064
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.119402985074628e-05,
      "loss": 2.6335,
      "theoretical_loss": 3.3911427470674727,
      "tokens_seen": 2300379136
    },
    {
      "epoch": 0.39,
      "objective/train/docs_used": 1269425,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5342440605163574,
      "objective/train/theoretical_loss": 3.391134911842176,
      "objective/train/tokens_used": 670903776,
      "theoretical_loss": 3.391134911842176,
      "tokens_seen": 2300444672
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.118600545658804e-05,
      "loss": 2.5486,
      "theoretical_loss": 3.3911270769025874,
      "tokens_seen": 2300510208
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.11779810624298e-05,
      "loss": 2.5324,
      "theoretical_loss": 3.391111407880458,
      "tokens_seen": 2300641280
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.116995666827155e-05,
      "loss": 2.4238,
      "theoretical_loss": 3.391095740000936,
      "tokens_seen": 2300772352
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.116193227411331e-05,
      "loss": 2.5696,
      "theoretical_loss": 3.3910800732638733,
      "tokens_seen": 2300903424
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.115390787995507e-05,
      "loss": 2.4227,
      "theoretical_loss": 3.3910644076691208,
      "tokens_seen": 2301034496
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.114588348579683e-05,
      "loss": 2.512,
      "theoretical_loss": 3.391048743216531,
      "tokens_seen": 2301165568
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.113785909163859e-05,
      "loss": 2.5283,
      "theoretical_loss": 3.391033079905955,
      "tokens_seen": 2301296640
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.112983469748035e-05,
      "loss": 2.4977,
      "theoretical_loss": 3.3910174177372445,
      "tokens_seen": 2301427712
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.112181030332211e-05,
      "loss": 2.4592,
      "theoretical_loss": 3.391001756710252,
      "tokens_seen": 2301558784
    },
    {
      "epoch": 0.39,
      "learning_rate": 6.111378590916386e-05,
      "loss": 2.5168,
      "theoretical_loss": 3.390986096824828,
      "tokens_seen": 2301689856
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.110576151500563e-05,
      "loss": 2.524,
      "theoretical_loss": 3.390970438080825,
      "tokens_seen": 2301820928
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.109773712084738e-05,
      "loss": 2.5812,
      "theoretical_loss": 3.3909547804780953,
      "tokens_seen": 2301952000
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1270779,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4318535327911377,
      "objective/train/theoretical_loss": 3.39093912401649,
      "objective/train/tokens_used": 672542176,
      "theoretical_loss": 3.39093912401649,
      "tokens_seen": 2302083072
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.108971272668915e-05,
      "loss": 2.5938,
      "theoretical_loss": 3.39093912401649,
      "tokens_seen": 2302083072
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.10816883325309e-05,
      "loss": 2.4497,
      "theoretical_loss": 3.3909234686958616,
      "tokens_seen": 2302214144
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.107366393837267e-05,
      "loss": 2.3694,
      "theoretical_loss": 3.390907814516061,
      "tokens_seen": 2302345216
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.106563954421442e-05,
      "loss": 2.5692,
      "theoretical_loss": 3.3908921614769407,
      "tokens_seen": 2302476288
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.105761515005617e-05,
      "loss": 2.4478,
      "theoretical_loss": 3.390876509578353,
      "tokens_seen": 2302607360
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.104959075589794e-05,
      "loss": 2.4334,
      "theoretical_loss": 3.390860858820149,
      "tokens_seen": 2302738432
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.104156636173969e-05,
      "loss": 2.475,
      "theoretical_loss": 3.390845209202182,
      "tokens_seen": 2302869504
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.103354196758145e-05,
      "loss": 2.6671,
      "theoretical_loss": 3.390829560724303,
      "tokens_seen": 2303000576
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.1025517573423206e-05,
      "loss": 2.51,
      "theoretical_loss": 3.3908139133863644,
      "tokens_seen": 2303131648
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.101749317926497e-05,
      "loss": 2.5443,
      "theoretical_loss": 3.390798267188218,
      "tokens_seen": 2303262720
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.1009468785106726e-05,
      "loss": 2.4232,
      "theoretical_loss": 3.3907826221297164,
      "tokens_seen": 2303393792
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.100144439094848e-05,
      "loss": 2.5159,
      "theoretical_loss": 3.3907669782107113,
      "tokens_seen": 2303524864
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0993419996790246e-05,
      "loss": 2.6357,
      "theoretical_loss": 3.3907513354310557,
      "tokens_seen": 2303655936
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1271371,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2652809619903564,
      "objective/train/theoretical_loss": 3.3907435144684372,
      "objective/train/tokens_used": 674180576,
      "theoretical_loss": 3.3907435144684372,
      "tokens_seen": 2303721472
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0985395602632e-05,
      "loss": 2.4737,
      "theoretical_loss": 3.3907356937906004,
      "tokens_seen": 2303787008
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0977371208473766e-05,
      "loss": 2.565,
      "theoretical_loss": 3.390720053289199,
      "tokens_seen": 2303918080
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.096934681431552e-05,
      "loss": 2.257,
      "theoretical_loss": 3.390704413926703,
      "tokens_seen": 2304049152
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0961322420157285e-05,
      "loss": 2.5282,
      "theoretical_loss": 3.390688775702965,
      "tokens_seen": 2304180224
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.095329802599904e-05,
      "loss": 2.5341,
      "theoretical_loss": 3.390673138617837,
      "tokens_seen": 2304311296
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.094527363184079e-05,
      "loss": 2.5535,
      "theoretical_loss": 3.3906575026711714,
      "tokens_seen": 2304442368
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.093724923768256e-05,
      "loss": 2.5071,
      "theoretical_loss": 3.3906418678628207,
      "tokens_seen": 2304573440
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.092922484352431e-05,
      "loss": 2.5005,
      "theoretical_loss": 3.3906262341926374,
      "tokens_seen": 2304704512
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.092120044936608e-05,
      "loss": 2.4322,
      "theoretical_loss": 3.390610601660474,
      "tokens_seen": 2304835584
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.091317605520783e-05,
      "loss": 2.4756,
      "theoretical_loss": 3.3905949702661826,
      "tokens_seen": 2304966656
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.09051516610496e-05,
      "loss": 2.3177,
      "theoretical_loss": 3.390579340009616,
      "tokens_seen": 2305097728
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.089712726689135e-05,
      "loss": 2.4599,
      "theoretical_loss": 3.390563710890626,
      "tokens_seen": 2305228800
    },
    {
      "debugging/Self-BLEU-5": 0.43298113666112864,
      "debugging/distinct-1-grams": 0.7776999941413213,
      "debugging/distinct-2-grams": 0.9535998885349362,
      "debugging/entropy-1-grams": 5.6285201862472185,
      "debugging/entropy-2-grams": 6.29126139825565,
      "debugging/length": 545.75,
      "debugging/num_segments": 8,
      "debugging/score": 0.006776578879058141,
      "debugging/score_std": 0.003879118421535622,
      "epoch": 0.4,
      "objective/train/docs_used": 1272482,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.490351676940918,
      "objective/train/theoretical_loss": 3.390548082909066,
      "objective/train/tokens_used": 675818976,
      "theoretical_loss": 3.390548082909066,
      "tokens_seen": 2305359872
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.088910287273312e-05,
      "loss": 2.5077,
      "theoretical_loss": 3.390548082909066,
      "tokens_seen": 2305359872
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.088107847857487e-05,
      "loss": 2.4906,
      "theoretical_loss": 3.3905324560647885,
      "tokens_seen": 2305490944
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0873054084416624e-05,
      "loss": 2.5881,
      "theoretical_loss": 3.3905168303576456,
      "tokens_seen": 2305622016
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.086502969025839e-05,
      "loss": 2.6183,
      "theoretical_loss": 3.39050120578749,
      "tokens_seen": 2305753088
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0857005296100144e-05,
      "loss": 2.5572,
      "theoretical_loss": 3.3904855823541746,
      "tokens_seen": 2305884160
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.084898090194191e-05,
      "loss": 2.5107,
      "theoretical_loss": 3.390469960057552,
      "tokens_seen": 2306015232
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0840956507783664e-05,
      "loss": 2.6371,
      "theoretical_loss": 3.3904543388974746,
      "tokens_seen": 2306146304
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.083293211362543e-05,
      "loss": 2.6085,
      "theoretical_loss": 3.3904387188737957,
      "tokens_seen": 2306277376
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0824907719467183e-05,
      "loss": 2.4177,
      "theoretical_loss": 3.390423099986368,
      "tokens_seen": 2306408448
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0816883325308937e-05,
      "loss": 2.5985,
      "theoretical_loss": 3.390407482235043,
      "tokens_seen": 2306539520
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.08088589311507e-05,
      "loss": 2.5685,
      "theoretical_loss": 3.390391865619675,
      "tokens_seen": 2306670592
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0800834536992456e-05,
      "loss": 2.7307,
      "theoretical_loss": 3.3903762501401165,
      "tokens_seen": 2306801664
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.079281014283422e-05,
      "loss": 2.4756,
      "theoretical_loss": 3.3903606357962204,
      "tokens_seen": 2306932736
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1273129,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.518545627593994,
      "objective/train/theoretical_loss": 3.3903528290500993,
      "objective/train/tokens_used": 677457376,
      "theoretical_loss": 3.3903528290500993,
      "tokens_seen": 2306998272
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0784785748675976e-05,
      "loss": 2.3769,
      "theoretical_loss": 3.390345022587839,
      "tokens_seen": 2307063808
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.077676135451774e-05,
      "loss": 2.5298,
      "theoretical_loss": 3.390329410514826,
      "tokens_seen": 2307194880
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0768736960359496e-05,
      "loss": 2.4769,
      "theoretical_loss": 3.3903137995770334,
      "tokens_seen": 2307325952
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.076071256620125e-05,
      "loss": 2.7667,
      "theoretical_loss": 3.390298189774315,
      "tokens_seen": 2307457024
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0752688172043016e-05,
      "loss": 2.5325,
      "theoretical_loss": 3.390282581106524,
      "tokens_seen": 2307588096
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.074466377788477e-05,
      "loss": 2.5,
      "theoretical_loss": 3.3902669735735125,
      "tokens_seen": 2307719168
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0736639383726536e-05,
      "loss": 2.5798,
      "theoretical_loss": 3.390251367175134,
      "tokens_seen": 2307850240
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.072861498956829e-05,
      "loss": 2.5994,
      "theoretical_loss": 3.3902357619112413,
      "tokens_seen": 2307981312
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0720590595410055e-05,
      "loss": 2.4868,
      "theoretical_loss": 3.3902201577816884,
      "tokens_seen": 2308112384
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.071256620125181e-05,
      "loss": 2.5244,
      "theoretical_loss": 3.3902045547863273,
      "tokens_seen": 2308243456
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.070454180709356e-05,
      "loss": 2.4328,
      "theoretical_loss": 3.390188952925012,
      "tokens_seen": 2308374528
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.069651741293533e-05,
      "loss": 2.6042,
      "theoretical_loss": 3.3901733521975954,
      "tokens_seen": 2308505600
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1273675,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2433297634124756,
      "objective/train/theoretical_loss": 3.3901577526039306,
      "objective/train/tokens_used": 679095776,
      "theoretical_loss": 3.3901577526039306,
      "tokens_seen": 2308636672
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.068849301877708e-05,
      "loss": 2.455,
      "theoretical_loss": 3.3901577526039306,
      "tokens_seen": 2308636672
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.068046862461885e-05,
      "loss": 2.4181,
      "theoretical_loss": 3.390142154143871,
      "tokens_seen": 2308767744
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.06724442304606e-05,
      "loss": 2.4947,
      "theoretical_loss": 3.39012655681727,
      "tokens_seen": 2308898816
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.066441983630237e-05,
      "loss": 2.3941,
      "theoretical_loss": 3.3901109606239808,
      "tokens_seen": 2309029888
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.065639544214412e-05,
      "loss": 2.4661,
      "theoretical_loss": 3.390095365563856,
      "tokens_seen": 2309160960
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0648371047985874e-05,
      "loss": 2.3421,
      "theoretical_loss": 3.3900797716367506,
      "tokens_seen": 2309292032
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.064034665382764e-05,
      "loss": 2.5064,
      "theoretical_loss": 3.390064178842516,
      "tokens_seen": 2309423104
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0632322259669394e-05,
      "loss": 2.6014,
      "theoretical_loss": 3.390048587181007,
      "tokens_seen": 2309554176
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.062429786551116e-05,
      "loss": 2.5406,
      "theoretical_loss": 3.390032996652077,
      "tokens_seen": 2309685248
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0616273471352914e-05,
      "loss": 2.6872,
      "theoretical_loss": 3.390017407255579,
      "tokens_seen": 2309816320
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.060824907719468e-05,
      "loss": 2.5322,
      "theoretical_loss": 3.390001818991366,
      "tokens_seen": 2309947392
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0600224683036434e-05,
      "loss": 2.4703,
      "theoretical_loss": 3.3899862318592926,
      "tokens_seen": 2310078464
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.05922002888782e-05,
      "loss": 2.5021,
      "theoretical_loss": 3.389970645859212,
      "tokens_seen": 2310209536
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1274788,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4272327423095703,
      "objective/train/theoretical_loss": 3.389962853283623,
      "objective/train/tokens_used": 680734176,
      "theoretical_loss": 3.389962853283623,
      "tokens_seen": 2310275072
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0584175894719954e-05,
      "loss": 2.4588,
      "theoretical_loss": 3.389955060990977,
      "tokens_seen": 2310340608
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.057615150056171e-05,
      "loss": 2.6382,
      "theoretical_loss": 3.3899394772544422,
      "tokens_seen": 2310471680
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.056812710640347e-05,
      "loss": 2.4812,
      "theoretical_loss": 3.3899238946494608,
      "tokens_seen": 2310602752
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0560102712245226e-05,
      "loss": 2.6145,
      "theoretical_loss": 3.3899083131758867,
      "tokens_seen": 2310733824
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.055207831808699e-05,
      "loss": 2.4977,
      "theoretical_loss": 3.389892732833573,
      "tokens_seen": 2310864896
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0544053923928746e-05,
      "loss": 2.4079,
      "theoretical_loss": 3.389877153622374,
      "tokens_seen": 2310995968
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.053602952977051e-05,
      "loss": 2.6061,
      "theoretical_loss": 3.3898615755421435,
      "tokens_seen": 2311127040
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0528005135612266e-05,
      "loss": 2.4405,
      "theoretical_loss": 3.389845998592735,
      "tokens_seen": 2311258112
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.051998074145402e-05,
      "loss": 2.5403,
      "theoretical_loss": 3.3898304227740024,
      "tokens_seen": 2311389184
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0511956347295786e-05,
      "loss": 2.5089,
      "theoretical_loss": 3.389814848085799,
      "tokens_seen": 2311520256
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.050393195313754e-05,
      "loss": 2.525,
      "theoretical_loss": 3.389799274527979,
      "tokens_seen": 2311651328
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0495907558979306e-05,
      "loss": 2.4141,
      "theoretical_loss": 3.3897837021003965,
      "tokens_seen": 2311782400
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1275747,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.075777053833008,
      "objective/train/theoretical_loss": 3.3897681308029055,
      "objective/train/tokens_used": 682372576,
      "theoretical_loss": 3.3897681308029055,
      "tokens_seen": 2311913472
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.048788316482106e-05,
      "loss": 2.4512,
      "theoretical_loss": 3.3897681308029055,
      "tokens_seen": 2311913472
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0479858770662826e-05,
      "loss": 2.5678,
      "theoretical_loss": 3.389752560635359,
      "tokens_seen": 2312044544
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.047183437650458e-05,
      "loss": 2.6394,
      "theoretical_loss": 3.389736991597612,
      "tokens_seen": 2312175616
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.046380998234633e-05,
      "loss": 2.6075,
      "theoretical_loss": 3.3897214236895183,
      "tokens_seen": 2312306688
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.04557855881881e-05,
      "loss": 2.518,
      "theoretical_loss": 3.3897058569109317,
      "tokens_seen": 2312437760
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.044776119402985e-05,
      "loss": 2.4914,
      "theoretical_loss": 3.3896902912617057,
      "tokens_seen": 2312568832
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.043973679987162e-05,
      "loss": 2.6291,
      "theoretical_loss": 3.3896747267416956,
      "tokens_seen": 2312699904
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.043171240571337e-05,
      "loss": 2.5727,
      "theoretical_loss": 3.3896591633507542,
      "tokens_seen": 2312830976
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.042368801155514e-05,
      "loss": 2.4357,
      "theoretical_loss": 3.3896436010887365,
      "tokens_seen": 2312962048
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.041566361739689e-05,
      "loss": 2.6064,
      "theoretical_loss": 3.3896280399554963,
      "tokens_seen": 2313093120
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0407639223238644e-05,
      "loss": 2.4785,
      "theoretical_loss": 3.3896124799508875,
      "tokens_seen": 2313224192
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.039961482908041e-05,
      "loss": 2.5568,
      "theoretical_loss": 3.389596921074765,
      "tokens_seen": 2313355264
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0391590434922164e-05,
      "loss": 2.6242,
      "theoretical_loss": 3.3895813633269825,
      "tokens_seen": 2313486336
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1276293,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2193737030029297,
      "objective/train/theoretical_loss": 3.389573584876173,
      "objective/train/tokens_used": 684010976,
      "theoretical_loss": 3.389573584876173,
      "tokens_seen": 2313551872
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.038356604076393e-05,
      "loss": 2.4727,
      "theoretical_loss": 3.3895658067073944,
      "tokens_seen": 2313617408
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0375541646605684e-05,
      "loss": 2.6057,
      "theoretical_loss": 3.389550251215855,
      "tokens_seen": 2313748480
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.036751725244745e-05,
      "loss": 2.5193,
      "theoretical_loss": 3.389534696852219,
      "tokens_seen": 2313879552
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0359492858289204e-05,
      "loss": 2.7326,
      "theoretical_loss": 3.3895191436163397,
      "tokens_seen": 2314010624
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.035146846413096e-05,
      "loss": 2.7271,
      "theoretical_loss": 3.3895035915080722,
      "tokens_seen": 2314141696
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0343444069972724e-05,
      "loss": 2.7267,
      "theoretical_loss": 3.389488040527271,
      "tokens_seen": 2314272768
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.033541967581448e-05,
      "loss": 2.6486,
      "theoretical_loss": 3.38947249067379,
      "tokens_seen": 2314403840
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0327395281656243e-05,
      "loss": 2.4973,
      "theoretical_loss": 3.3894569419474836,
      "tokens_seen": 2314534912
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0319370887497997e-05,
      "loss": 2.4632,
      "theoretical_loss": 3.389441394348207,
      "tokens_seen": 2314665984
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0311346493339756e-05,
      "loss": 2.6751,
      "theoretical_loss": 3.3894258478758146,
      "tokens_seen": 2314797056
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0303322099181516e-05,
      "loss": 2.436,
      "theoretical_loss": 3.3894103025301603,
      "tokens_seen": 2314928128
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.029529770502327e-05,
      "loss": 2.5694,
      "theoretical_loss": 3.3893947583110986,
      "tokens_seen": 2315059200
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1277268,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.424570322036743,
      "objective/train/theoretical_loss": 3.3893792152184847,
      "objective/train/tokens_used": 685649376,
      "theoretical_loss": 3.3893792152184847,
      "tokens_seen": 2315190272
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0287273310865036e-05,
      "loss": 2.5064,
      "theoretical_loss": 3.3893792152184847,
      "tokens_seen": 2315190272
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.027924891670679e-05,
      "loss": 2.6438,
      "theoretical_loss": 3.3893636732521735,
      "tokens_seen": 2315321344
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0271224522548556e-05,
      "loss": 2.7128,
      "theoretical_loss": 3.3893481324120183,
      "tokens_seen": 2315452416
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.026320012839031e-05,
      "loss": 2.7182,
      "theoretical_loss": 3.389332592697875,
      "tokens_seen": 2315583488
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.025517573423207e-05,
      "loss": 2.5202,
      "theoretical_loss": 3.3893170541095974,
      "tokens_seen": 2315714560
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.024715134007383e-05,
      "loss": 2.644,
      "theoretical_loss": 3.389301516647041,
      "tokens_seen": 2315845632
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.023912694591559e-05,
      "loss": 2.4893,
      "theoretical_loss": 3.38928598031006,
      "tokens_seen": 2315976704
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.023110255175735e-05,
      "loss": 2.4382,
      "theoretical_loss": 3.389270445098509,
      "tokens_seen": 2316107776
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.02230781575991e-05,
      "loss": 2.5753,
      "theoretical_loss": 3.3892549110122436,
      "tokens_seen": 2316238848
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.021505376344086e-05,
      "loss": 2.4659,
      "theoretical_loss": 3.3892393780511174,
      "tokens_seen": 2316369920
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.020702936928262e-05,
      "loss": 2.6214,
      "theoretical_loss": 3.389223846214987,
      "tokens_seen": 2316500992
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.019900497512438e-05,
      "loss": 2.6186,
      "theoretical_loss": 3.389208315503706,
      "tokens_seen": 2316632064
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.019098058096614e-05,
      "loss": 2.5551,
      "theoretical_loss": 3.3891927859171287,
      "tokens_seen": 2316763136
    },
    {
      "epoch": 0.4,
      "objective/train/docs_used": 1277883,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.049539804458618,
      "objective/train/theoretical_loss": 3.3891850215455595,
      "objective/train/tokens_used": 687287776,
      "theoretical_loss": 3.3891850215455595,
      "tokens_seen": 2316828672
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.01829561868079e-05,
      "loss": 2.7491,
      "theoretical_loss": 3.3891772574551116,
      "tokens_seen": 2316894208
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.017493179264966e-05,
      "loss": 2.587,
      "theoretical_loss": 3.389161730117509,
      "tokens_seen": 2317025280
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0166907398491414e-05,
      "loss": 2.7186,
      "theoretical_loss": 3.3891462039041755,
      "tokens_seen": 2317156352
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0158883004333174e-05,
      "loss": 2.5124,
      "theoretical_loss": 3.3891306788149667,
      "tokens_seen": 2317287424
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0150858610174934e-05,
      "loss": 2.5957,
      "theoretical_loss": 3.389115154849737,
      "tokens_seen": 2317418496
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0142834216016694e-05,
      "loss": 2.7088,
      "theoretical_loss": 3.3890996320083424,
      "tokens_seen": 2317549568
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0134809821858454e-05,
      "loss": 2.7321,
      "theoretical_loss": 3.3890841102906366,
      "tokens_seen": 2317680640
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.0126785427700214e-05,
      "loss": 2.5725,
      "theoretical_loss": 3.3890685896964765,
      "tokens_seen": 2317811712
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.011876103354197e-05,
      "loss": 2.5568,
      "theoretical_loss": 3.3890530702257156,
      "tokens_seen": 2317942784
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.011073663938373e-05,
      "loss": 2.5002,
      "theoretical_loss": 3.38903755187821,
      "tokens_seen": 2318073856
    },
    {
      "epoch": 0.4,
      "learning_rate": 6.010271224522549e-05,
      "loss": 2.5189,
      "theoretical_loss": 3.3890220346538142,
      "tokens_seen": 2318204928
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.009468785106725e-05,
      "loss": 2.5884,
      "theoretical_loss": 3.3890065185523843,
      "tokens_seen": 2318336000
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1279166,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8162686824798584,
      "objective/train/theoretical_loss": 3.388991003573775,
      "objective/train/tokens_used": 688926176,
      "theoretical_loss": 3.388991003573775,
      "tokens_seen": 2318467072
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.008666345690901e-05,
      "loss": 2.434,
      "theoretical_loss": 3.388991003573775,
      "tokens_seen": 2318467072
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.007863906275077e-05,
      "loss": 2.5111,
      "theoretical_loss": 3.3889754897178417,
      "tokens_seen": 2318598144
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.0070614668592527e-05,
      "loss": 2.6265,
      "theoretical_loss": 3.38895997698444,
      "tokens_seen": 2318729216
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.006259027443428e-05,
      "loss": 2.4992,
      "theoretical_loss": 3.388944465373424,
      "tokens_seen": 2318860288
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.005456588027604e-05,
      "loss": 2.5315,
      "theoretical_loss": 3.388928954884651,
      "tokens_seen": 2318991360
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.00465414861178e-05,
      "loss": 2.5491,
      "theoretical_loss": 3.3889134455179746,
      "tokens_seen": 2319122432
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.003851709195956e-05,
      "loss": 2.6084,
      "theoretical_loss": 3.388897937273251,
      "tokens_seen": 2319253504
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.003049269780132e-05,
      "loss": 2.5187,
      "theoretical_loss": 3.388882430150336,
      "tokens_seen": 2319384576
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.002246830364307e-05,
      "loss": 2.6039,
      "theoretical_loss": 3.3888669241490845,
      "tokens_seen": 2319515648
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.001444390948484e-05,
      "loss": 2.5828,
      "theoretical_loss": 3.3888514192693524,
      "tokens_seen": 2319646720
    },
    {
      "epoch": 0.41,
      "learning_rate": 6.000641951532659e-05,
      "loss": 2.4489,
      "theoretical_loss": 3.388835915510995,
      "tokens_seen": 2319777792
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.999839512116835e-05,
      "loss": 2.533,
      "theoretical_loss": 3.3888204128738675,
      "tokens_seen": 2319908864
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.999037072701011e-05,
      "loss": 2.6468,
      "theoretical_loss": 3.388804911357826,
      "tokens_seen": 2320039936
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1279806,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8292062282562256,
      "objective/train/theoretical_loss": 3.3887971610201673,
      "objective/train/tokens_used": 690564576,
      "theoretical_loss": 3.3887971610201673,
      "tokens_seen": 2320105472
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.998234633285187e-05,
      "loss": 2.5787,
      "theoretical_loss": 3.3887894109627257,
      "tokens_seen": 2320171008
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.997432193869363e-05,
      "loss": 2.7014,
      "theoretical_loss": 3.388773911688422,
      "tokens_seen": 2320302080
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9966297544535385e-05,
      "loss": 2.5897,
      "theoretical_loss": 3.388758413534772,
      "tokens_seen": 2320433152
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.995827315037715e-05,
      "loss": 2.7845,
      "theoretical_loss": 3.38874291650163,
      "tokens_seen": 2320564224
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9950248756218905e-05,
      "loss": 2.5022,
      "theoretical_loss": 3.3887274205888516,
      "tokens_seen": 2320695296
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.994222436206067e-05,
      "loss": 2.4866,
      "theoretical_loss": 3.388711925796293,
      "tokens_seen": 2320826368
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9934199967902425e-05,
      "loss": 2.4957,
      "theoretical_loss": 3.3886964321238104,
      "tokens_seen": 2320957440
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.992617557374418e-05,
      "loss": 2.6397,
      "theoretical_loss": 3.388680939571259,
      "tokens_seen": 2321088512
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9918151179585944e-05,
      "loss": 2.5842,
      "theoretical_loss": 3.388665448138495,
      "tokens_seen": 2321219584
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.99101267854277e-05,
      "loss": 2.6929,
      "theoretical_loss": 3.3886499578253737,
      "tokens_seen": 2321350656
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9902102391269464e-05,
      "loss": 2.5263,
      "theoretical_loss": 3.388634468631751,
      "tokens_seen": 2321481728
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.989407799711122e-05,
      "loss": 2.5219,
      "theoretical_loss": 3.3886189805574833,
      "tokens_seen": 2321612800
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1281164,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.399672269821167,
      "objective/train/theoretical_loss": 3.3886034936024263,
      "objective/train/tokens_used": 692202976,
      "theoretical_loss": 3.3886034936024263,
      "tokens_seen": 2321743872
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9886053602952984e-05,
      "loss": 2.6074,
      "theoretical_loss": 3.3886034936024263,
      "tokens_seen": 2321743872
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.987802920879474e-05,
      "loss": 2.5763,
      "theoretical_loss": 3.388588007766436,
      "tokens_seen": 2321874944
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.987000481463649e-05,
      "loss": 2.4471,
      "theoretical_loss": 3.3885725230493677,
      "tokens_seen": 2322006016
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.986198042047826e-05,
      "loss": 2.5937,
      "theoretical_loss": 3.3885570394510784,
      "tokens_seen": 2322137088
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.985395602632001e-05,
      "loss": 2.6915,
      "theoretical_loss": 3.3885415569714237,
      "tokens_seen": 2322268160
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.984593163216178e-05,
      "loss": 2.4943,
      "theoretical_loss": 3.3885260756102595,
      "tokens_seen": 2322399232
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.983790723800353e-05,
      "loss": 2.6171,
      "theoretical_loss": 3.388510595367442,
      "tokens_seen": 2322530304
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.98298828438453e-05,
      "loss": 2.5476,
      "theoretical_loss": 3.3884951162428276,
      "tokens_seen": 2322661376
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.982185844968705e-05,
      "loss": 2.6127,
      "theoretical_loss": 3.388479638236272,
      "tokens_seen": 2322792448
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.98138340555288e-05,
      "loss": 2.611,
      "theoretical_loss": 3.388464161347632,
      "tokens_seen": 2322923520
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.980580966137057e-05,
      "loss": 2.6217,
      "theoretical_loss": 3.388448685576763,
      "tokens_seen": 2323054592
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.979778526721232e-05,
      "loss": 2.6289,
      "theoretical_loss": 3.3884332109235213,
      "tokens_seen": 2323185664
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.978976087305409e-05,
      "loss": 2.552,
      "theoretical_loss": 3.3884177373877633,
      "tokens_seen": 2323316736
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1281765,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3108794689178467,
      "objective/train/theoretical_loss": 3.388410001038896,
      "objective/train/tokens_used": 693841376,
      "theoretical_loss": 3.388410001038896,
      "tokens_seen": 2323382272
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.978173647889584e-05,
      "loss": 2.5039,
      "theoretical_loss": 3.3884022649693453,
      "tokens_seen": 2323447808
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.977371208473761e-05,
      "loss": 2.577,
      "theoretical_loss": 3.3883867936681242,
      "tokens_seen": 2323578880
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.976568769057936e-05,
      "loss": 2.6303,
      "theoretical_loss": 3.3883713234839554,
      "tokens_seen": 2323709952
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9757663296421115e-05,
      "loss": 2.555,
      "theoretical_loss": 3.3883558544166954,
      "tokens_seen": 2323841024
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.974963890226288e-05,
      "loss": 2.6508,
      "theoretical_loss": 3.388340386466201,
      "tokens_seen": 2323972096
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9741614508104635e-05,
      "loss": 2.4878,
      "theoretical_loss": 3.3883249196323284,
      "tokens_seen": 2324103168
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.97335901139464e-05,
      "loss": 2.3204,
      "theoretical_loss": 3.3883094539149337,
      "tokens_seen": 2324234240
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9725565719788155e-05,
      "loss": 2.4961,
      "theoretical_loss": 3.388293989313874,
      "tokens_seen": 2324365312
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.971754132562992e-05,
      "loss": 2.6842,
      "theoretical_loss": 3.388278525829005,
      "tokens_seen": 2324496384
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9709516931471675e-05,
      "loss": 2.6049,
      "theoretical_loss": 3.388263063460184,
      "tokens_seen": 2324627456
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.970149253731343e-05,
      "loss": 2.5084,
      "theoretical_loss": 3.388247602207267,
      "tokens_seen": 2324758528
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9693468143155195e-05,
      "loss": 2.577,
      "theoretical_loss": 3.3882321420701103,
      "tokens_seen": 2324889600
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1282973,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2994158267974854,
      "objective/train/theoretical_loss": 3.3882166830485714,
      "objective/train/tokens_used": 695479776,
      "theoretical_loss": 3.3882166830485714,
      "tokens_seen": 2325020672
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.968544374899695e-05,
      "loss": 2.524,
      "theoretical_loss": 3.3882166830485714,
      "tokens_seen": 2325020672
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9677419354838715e-05,
      "loss": 2.5124,
      "theoretical_loss": 3.388201225142506,
      "tokens_seen": 2325151744
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.966939496068047e-05,
      "loss": 2.4966,
      "theoretical_loss": 3.3881857683517715,
      "tokens_seen": 2325282816
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9661370566522234e-05,
      "loss": 2.4446,
      "theoretical_loss": 3.388170312676224,
      "tokens_seen": 2325413888
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.965334617236399e-05,
      "loss": 2.5587,
      "theoretical_loss": 3.3881548581157204,
      "tokens_seen": 2325544960
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9645321778205754e-05,
      "loss": 2.4818,
      "theoretical_loss": 3.3881394046701176,
      "tokens_seen": 2325676032
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.963729738404751e-05,
      "loss": 2.6061,
      "theoretical_loss": 3.3881239523392717,
      "tokens_seen": 2325807104
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.962927298988926e-05,
      "loss": 2.6662,
      "theoretical_loss": 3.3881085011230403,
      "tokens_seen": 2325938176
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.962124859573103e-05,
      "loss": 2.4937,
      "theoretical_loss": 3.3880930510212797,
      "tokens_seen": 2326069248
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.961322420157278e-05,
      "loss": 2.6491,
      "theoretical_loss": 3.388077602033847,
      "tokens_seen": 2326200320
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.960519980741455e-05,
      "loss": 2.508,
      "theoretical_loss": 3.3880621541605986,
      "tokens_seen": 2326331392
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.95971754132563e-05,
      "loss": 2.3725,
      "theoretical_loss": 3.388046707401392,
      "tokens_seen": 2326462464
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.958915101909807e-05,
      "loss": 2.4341,
      "theoretical_loss": 3.3880312617560833,
      "tokens_seen": 2326593536
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1283390,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1181111335754395,
      "objective/train/theoretical_loss": 3.3880235393510962,
      "objective/train/tokens_used": 697118176,
      "theoretical_loss": 3.3880235393510962,
      "tokens_seen": 2326659072
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.958112662493982e-05,
      "loss": 2.5477,
      "theoretical_loss": 3.38801581722453,
      "tokens_seen": 2326724608
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.957310223078157e-05,
      "loss": 2.4414,
      "theoretical_loss": 3.388000373806589,
      "tokens_seen": 2326855680
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.956507783662334e-05,
      "loss": 2.527,
      "theoretical_loss": 3.3879849315021175,
      "tokens_seen": 2326986752
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.955705344246509e-05,
      "loss": 2.3674,
      "theoretical_loss": 3.387969490310972,
      "tokens_seen": 2327117824
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.954902904830686e-05,
      "loss": 2.7315,
      "theoretical_loss": 3.3879540502330094,
      "tokens_seen": 2327248896
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.954100465414861e-05,
      "loss": 2.6365,
      "theoretical_loss": 3.387938611268088,
      "tokens_seen": 2327379968
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.953298025999038e-05,
      "loss": 2.5415,
      "theoretical_loss": 3.387923173416063,
      "tokens_seen": 2327511040
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.952495586583213e-05,
      "loss": 2.5746,
      "theoretical_loss": 3.387907736676793,
      "tokens_seen": 2327642112
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9516931471673886e-05,
      "loss": 2.5115,
      "theoretical_loss": 3.387892301050135,
      "tokens_seen": 2327773184
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.950890707751565e-05,
      "loss": 2.6609,
      "theoretical_loss": 3.3878768665359456,
      "tokens_seen": 2327904256
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9500882683357405e-05,
      "loss": 2.5681,
      "theoretical_loss": 3.387861433134082,
      "tokens_seen": 2328035328
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.949285828919917e-05,
      "loss": 2.4529,
      "theoretical_loss": 3.3878460008444016,
      "tokens_seen": 2328166400
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1284624,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.565575122833252,
      "objective/train/theoretical_loss": 3.387830569666762,
      "objective/train/tokens_used": 698756576,
      "theoretical_loss": 3.387830569666762,
      "tokens_seen": 2328297472
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9484833895040925e-05,
      "loss": 2.5543,
      "theoretical_loss": 3.387830569666762,
      "tokens_seen": 2328297472
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.947680950088269e-05,
      "loss": 2.5037,
      "theoretical_loss": 3.38781513960102,
      "tokens_seen": 2328428544
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9468785106724445e-05,
      "loss": 2.5511,
      "theoretical_loss": 3.387799710647033,
      "tokens_seen": 2328559616
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.94607607125662e-05,
      "loss": 2.4309,
      "theoretical_loss": 3.3877842828046587,
      "tokens_seen": 2328690688
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9452736318407965e-05,
      "loss": 2.4482,
      "theoretical_loss": 3.3877688560737536,
      "tokens_seen": 2328821760
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.944471192424972e-05,
      "loss": 2.5951,
      "theoretical_loss": 3.387753430454176,
      "tokens_seen": 2328952832
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9436687530091485e-05,
      "loss": 2.5926,
      "theoretical_loss": 3.3877380059457827,
      "tokens_seen": 2329083904
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.942866313593324e-05,
      "loss": 2.61,
      "theoretical_loss": 3.3877225825484314,
      "tokens_seen": 2329214976
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9420638741775004e-05,
      "loss": 2.6907,
      "theoretical_loss": 3.3877071602619795,
      "tokens_seen": 2329346048
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.941261434761676e-05,
      "loss": 2.4477,
      "theoretical_loss": 3.3876917390862844,
      "tokens_seen": 2329477120
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.940458995345851e-05,
      "loss": 2.3911,
      "theoretical_loss": 3.387676319021204,
      "tokens_seen": 2329608192
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.939656555930028e-05,
      "loss": 2.5841,
      "theoretical_loss": 3.387660900066595,
      "tokens_seen": 2329739264
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.938854116514203e-05,
      "loss": 2.5772,
      "theoretical_loss": 3.387645482222316,
      "tokens_seen": 2329870336
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1285083,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 1.9569989442825317,
      "objective/train/theoretical_loss": 3.3876377737165053,
      "objective/train/tokens_used": 700394976,
      "theoretical_loss": 3.3876377737165053,
      "tokens_seen": 2329935872
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.93805167709838e-05,
      "loss": 2.332,
      "theoretical_loss": 3.387630065488224,
      "tokens_seen": 2330001408
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.937249237682555e-05,
      "loss": 2.6089,
      "theoretical_loss": 3.387614649864176,
      "tokens_seen": 2330132480
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.936446798266732e-05,
      "loss": 2.6178,
      "theoretical_loss": 3.387599235350031,
      "tokens_seen": 2330263552
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.935644358850907e-05,
      "loss": 2.5833,
      "theoretical_loss": 3.387583821945646,
      "tokens_seen": 2330394624
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.934841919435082e-05,
      "loss": 2.5008,
      "theoretical_loss": 3.3875684096508785,
      "tokens_seen": 2330525696
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.934039480019259e-05,
      "loss": 2.417,
      "theoretical_loss": 3.387552998465586,
      "tokens_seen": 2330656768
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.933237040603434e-05,
      "loss": 2.4472,
      "theoretical_loss": 3.3875375883896273,
      "tokens_seen": 2330787840
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.932434601187611e-05,
      "loss": 2.6171,
      "theoretical_loss": 3.3875221794228594,
      "tokens_seen": 2330918912
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.931632161771786e-05,
      "loss": 2.5458,
      "theoretical_loss": 3.38750677156514,
      "tokens_seen": 2331049984
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.930829722355963e-05,
      "loss": 2.6473,
      "theoretical_loss": 3.3874913648163267,
      "tokens_seen": 2331181056
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.930027282940138e-05,
      "loss": 2.5647,
      "theoretical_loss": 3.387475959176278,
      "tokens_seen": 2331312128
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.929224843524315e-05,
      "loss": 2.5548,
      "theoretical_loss": 3.387460554644852,
      "tokens_seen": 2331443200
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1286380,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.882779836654663,
      "objective/train/theoretical_loss": 3.387445151221906,
      "objective/train/tokens_used": 702033376,
      "theoretical_loss": 3.387445151221906,
      "tokens_seen": 2331574272
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.92842240410849e-05,
      "loss": 2.5951,
      "theoretical_loss": 3.387445151221906,
      "tokens_seen": 2331574272
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9276199646926656e-05,
      "loss": 2.5724,
      "theoretical_loss": 3.387429748907298,
      "tokens_seen": 2331705344
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.926817525276842e-05,
      "loss": 2.5311,
      "theoretical_loss": 3.3874143477008865,
      "tokens_seen": 2331836416
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9260150858610175e-05,
      "loss": 2.446,
      "theoretical_loss": 3.387398947602528,
      "tokens_seen": 2331967488
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.925212646445194e-05,
      "loss": 2.4216,
      "theoretical_loss": 3.3873835486120822,
      "tokens_seen": 2332098560
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9244102070293695e-05,
      "loss": 2.4455,
      "theoretical_loss": 3.3873681507294067,
      "tokens_seen": 2332229632
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.923607767613546e-05,
      "loss": 2.4227,
      "theoretical_loss": 3.387352753954359,
      "tokens_seen": 2332360704
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9228053281977215e-05,
      "loss": 2.4882,
      "theoretical_loss": 3.387337358286797,
      "tokens_seen": 2332491776
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.922002888781897e-05,
      "loss": 2.4296,
      "theoretical_loss": 3.3873219637265803,
      "tokens_seen": 2332622848
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9212004493660735e-05,
      "loss": 2.5089,
      "theoretical_loss": 3.3873065702735654,
      "tokens_seen": 2332753920
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.920398009950249e-05,
      "loss": 2.5498,
      "theoretical_loss": 3.3872911779276116,
      "tokens_seen": 2332884992
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9195955705344255e-05,
      "loss": 2.6462,
      "theoretical_loss": 3.3872757866885763,
      "tokens_seen": 2333016064
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.918793131118601e-05,
      "loss": 2.6157,
      "theoretical_loss": 3.387260396556318,
      "tokens_seen": 2333147136
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1286803,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8880960941314697,
      "objective/train/theoretical_loss": 3.3872527019051857,
      "objective/train/tokens_used": 703671776,
      "theoretical_loss": 3.3872527019051857,
      "tokens_seen": 2333212672
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.9179906917027774e-05,
      "loss": 2.6258,
      "theoretical_loss": 3.387245007530695,
      "tokens_seen": 2333278208
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.917188252286953e-05,
      "loss": 2.638,
      "theoretical_loss": 3.387229619611565,
      "tokens_seen": 2333409280
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.916385812871128e-05,
      "loss": 2.6047,
      "theoretical_loss": 3.3872142327987875,
      "tokens_seen": 2333540352
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.915583373455305e-05,
      "loss": 2.5849,
      "theoretical_loss": 3.38719884709222,
      "tokens_seen": 2333671424
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.91478093403948e-05,
      "loss": 2.4521,
      "theoretical_loss": 3.387183462491721,
      "tokens_seen": 2333802496
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.913978494623657e-05,
      "loss": 2.657,
      "theoretical_loss": 3.3871680789971492,
      "tokens_seen": 2333933568
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.913176055207832e-05,
      "loss": 2.4622,
      "theoretical_loss": 3.3871526966083625,
      "tokens_seen": 2334064640
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.912373615792009e-05,
      "loss": 2.4988,
      "theoretical_loss": 3.3871373153252193,
      "tokens_seen": 2334195712
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.911571176376184e-05,
      "loss": 2.5101,
      "theoretical_loss": 3.387121935147578,
      "tokens_seen": 2334326784
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.910768736960359e-05,
      "loss": 2.4301,
      "theoretical_loss": 3.3871065560752975,
      "tokens_seen": 2334457856
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.909966297544536e-05,
      "loss": 2.6256,
      "theoretical_loss": 3.3870911781082365,
      "tokens_seen": 2334588928
    },
    {
      "epoch": 0.41,
      "learning_rate": 5.909163858128711e-05,
      "loss": 2.524,
      "theoretical_loss": 3.3870758012462527,
      "tokens_seen": 2334720000
    },
    {
      "epoch": 0.41,
      "objective/train/docs_used": 1288265,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5533502101898193,
      "objective/train/theoretical_loss": 3.387060425489205,
      "objective/train/tokens_used": 705310176,
      "theoretical_loss": 3.387060425489205,
      "tokens_seen": 2334851072
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.908361418712888e-05,
      "loss": 2.5422,
      "theoretical_loss": 3.387060425489205,
      "tokens_seen": 2334851072
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.907558979297063e-05,
      "loss": 2.7267,
      "theoretical_loss": 3.3870450508369525,
      "tokens_seen": 2334982144
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.90675653988124e-05,
      "loss": 2.5964,
      "theoretical_loss": 3.3870296772893536,
      "tokens_seen": 2335113216
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.905954100465415e-05,
      "loss": 2.5346,
      "theoretical_loss": 3.387014304846266,
      "tokens_seen": 2335244288
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.9051516610495906e-05,
      "loss": 2.4027,
      "theoretical_loss": 3.3869989335075497,
      "tokens_seen": 2335375360
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.904349221633767e-05,
      "loss": 2.465,
      "theoretical_loss": 3.3869835632730627,
      "tokens_seen": 2335506432
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.9035467822179426e-05,
      "loss": 2.563,
      "theoretical_loss": 3.3869681941426633,
      "tokens_seen": 2335637504
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.902744342802119e-05,
      "loss": 2.4427,
      "theoretical_loss": 3.386952826116211,
      "tokens_seen": 2335768576
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.9019419033862945e-05,
      "loss": 2.6736,
      "theoretical_loss": 3.3869374591935646,
      "tokens_seen": 2335899648
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.901139463970471e-05,
      "loss": 2.6175,
      "theoretical_loss": 3.386922093374582,
      "tokens_seen": 2336030720
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.9003370245546465e-05,
      "loss": 2.6195,
      "theoretical_loss": 3.386906728659123,
      "tokens_seen": 2336161792
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8995345851388225e-05,
      "loss": 2.6358,
      "theoretical_loss": 3.3868913650470462,
      "tokens_seen": 2336292864
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8987321457229985e-05,
      "loss": 2.6297,
      "theoretical_loss": 3.3868760025382096,
      "tokens_seen": 2336423936
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1288802,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3036727905273438,
      "objective/train/theoretical_loss": 3.386868321697463,
      "objective/train/tokens_used": 706948576,
      "theoretical_loss": 3.386868321697463,
      "tokens_seen": 2336489472
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.897929706307174e-05,
      "loss": 2.4331,
      "theoretical_loss": 3.3868606411324738,
      "tokens_seen": 2336555008
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8971272668913505e-05,
      "loss": 2.4585,
      "theoretical_loss": 3.386845280829696,
      "tokens_seen": 2336686080
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.896324827475526e-05,
      "loss": 2.5723,
      "theoretical_loss": 3.3868299216297357,
      "tokens_seen": 2336817152
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8955223880597025e-05,
      "loss": 2.661,
      "theoretical_loss": 3.3868145635324525,
      "tokens_seen": 2336948224
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.894719948643878e-05,
      "loss": 2.6349,
      "theoretical_loss": 3.3867992065377046,
      "tokens_seen": 2337079296
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.893917509228054e-05,
      "loss": 2.4407,
      "theoretical_loss": 3.386783850645352,
      "tokens_seen": 2337210368
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.89311506981223e-05,
      "loss": 2.4847,
      "theoretical_loss": 3.386768495855252,
      "tokens_seen": 2337341440
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.892312630396405e-05,
      "loss": 2.599,
      "theoretical_loss": 3.3867531421672656,
      "tokens_seen": 2337472512
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.891510190980582e-05,
      "loss": 2.486,
      "theoretical_loss": 3.3867377895812507,
      "tokens_seen": 2337603584
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.890707751564757e-05,
      "loss": 2.4696,
      "theoretical_loss": 3.386722438097067,
      "tokens_seen": 2337734656
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.889905312148933e-05,
      "loss": 2.5101,
      "theoretical_loss": 3.3867070877145733,
      "tokens_seen": 2337865728
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.889102872733109e-05,
      "loss": 2.6604,
      "theoretical_loss": 3.3866917384336293,
      "tokens_seen": 2337996800
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1289730,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0127034187316895,
      "objective/train/theoretical_loss": 3.3866763902540935,
      "objective/train/tokens_used": 708586976,
      "theoretical_loss": 3.3866763902540935,
      "tokens_seen": 2338127872
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.888300433317285e-05,
      "loss": 2.5908,
      "theoretical_loss": 3.3866763902540935,
      "tokens_seen": 2338127872
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.887497993901461e-05,
      "loss": 2.6727,
      "theoretical_loss": 3.3866610431758253,
      "tokens_seen": 2338258944
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.886695554485636e-05,
      "loss": 2.6052,
      "theoretical_loss": 3.3866456971986842,
      "tokens_seen": 2338390016
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.885893115069813e-05,
      "loss": 2.5066,
      "theoretical_loss": 3.3866303523225296,
      "tokens_seen": 2338521088
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.885090675653988e-05,
      "loss": 2.5048,
      "theoretical_loss": 3.3866150085472206,
      "tokens_seen": 2338652160
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.884288236238164e-05,
      "loss": 2.5927,
      "theoretical_loss": 3.3865996658726165,
      "tokens_seen": 2338783232
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.88348579682234e-05,
      "loss": 2.5956,
      "theoretical_loss": 3.3865843242985765,
      "tokens_seen": 2338914304
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.882683357406516e-05,
      "loss": 2.5094,
      "theoretical_loss": 3.3865689838249606,
      "tokens_seen": 2339045376
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.881880917990692e-05,
      "loss": 2.5411,
      "theoretical_loss": 3.3865536444516273,
      "tokens_seen": 2339176448
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8810784785748676e-05,
      "loss": 2.4394,
      "theoretical_loss": 3.386538306178437,
      "tokens_seen": 2339307520
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8802760391590436e-05,
      "loss": 2.3762,
      "theoretical_loss": 3.3865229690052487,
      "tokens_seen": 2339438592
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8794735997432196e-05,
      "loss": 2.5426,
      "theoretical_loss": 3.3865076329319215,
      "tokens_seen": 2339569664
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8786711603273956e-05,
      "loss": 2.681,
      "theoretical_loss": 3.3864922979583154,
      "tokens_seen": 2339700736
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1290793,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1198654174804688,
      "objective/train/theoretical_loss": 3.386484630883864,
      "objective/train/tokens_used": 710225376,
      "theoretical_loss": 3.386484630883864,
      "tokens_seen": 2339766272
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8778687209115716e-05,
      "loss": 2.5137,
      "theoretical_loss": 3.3864769640842898,
      "tokens_seen": 2339831808
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8770662814957475e-05,
      "loss": 2.5042,
      "theoretical_loss": 3.3864616313097047,
      "tokens_seen": 2339962880
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8762638420799235e-05,
      "loss": 2.3994,
      "theoretical_loss": 3.386446299634419,
      "tokens_seen": 2340093952
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.875461402664099e-05,
      "loss": 2.5447,
      "theoretical_loss": 3.386430969058293,
      "tokens_seen": 2340225024
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.874658963248275e-05,
      "loss": 2.3838,
      "theoretical_loss": 3.3864156395811853,
      "tokens_seen": 2340356096
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.873856523832451e-05,
      "loss": 2.5524,
      "theoretical_loss": 3.386400311202957,
      "tokens_seen": 2340487168
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.873054084416627e-05,
      "loss": 2.5065,
      "theoretical_loss": 3.3863849839234663,
      "tokens_seen": 2340618240
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.872251645000803e-05,
      "loss": 2.4252,
      "theoretical_loss": 3.386369657742574,
      "tokens_seen": 2340749312
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.871449205584979e-05,
      "loss": 2.5425,
      "theoretical_loss": 3.3863543326601393,
      "tokens_seen": 2340880384
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.870646766169154e-05,
      "loss": 2.5635,
      "theoretical_loss": 3.386339008676022,
      "tokens_seen": 2341011456
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.869844326753331e-05,
      "loss": 2.6626,
      "theoretical_loss": 3.3863236857900825,
      "tokens_seen": 2341142528
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.869041887337506e-05,
      "loss": 2.3922,
      "theoretical_loss": 3.38630836400218,
      "tokens_seen": 2341273600
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1291340,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.626932382583618,
      "objective/train/theoretical_loss": 3.386293043312175,
      "objective/train/tokens_used": 711863776,
      "theoretical_loss": 3.386293043312175,
      "tokens_seen": 2341404672
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.868239447921682e-05,
      "loss": 2.6208,
      "theoretical_loss": 3.386293043312175,
      "tokens_seen": 2341404672
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.867437008505858e-05,
      "loss": 2.4149,
      "theoretical_loss": 3.3862777237199264,
      "tokens_seen": 2341535744
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8666345690900334e-05,
      "loss": 2.5403,
      "theoretical_loss": 3.3862624052252945,
      "tokens_seen": 2341666816
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.86583212967421e-05,
      "loss": 2.5613,
      "theoretical_loss": 3.38624708782814,
      "tokens_seen": 2341797888
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8650296902583854e-05,
      "loss": 2.4958,
      "theoretical_loss": 3.3862317715283217,
      "tokens_seen": 2341928960
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.864227250842562e-05,
      "loss": 2.5425,
      "theoretical_loss": 3.3862164563257,
      "tokens_seen": 2342060032
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8634248114267374e-05,
      "loss": 2.6168,
      "theoretical_loss": 3.386201142220135,
      "tokens_seen": 2342191104
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8626223720109133e-05,
      "loss": 2.4386,
      "theoretical_loss": 3.386185829211487,
      "tokens_seen": 2342322176
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.861819932595089e-05,
      "loss": 2.4753,
      "theoretical_loss": 3.3861705172996155,
      "tokens_seen": 2342453248
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8610174931792646e-05,
      "loss": 2.4767,
      "theoretical_loss": 3.386155206484381,
      "tokens_seen": 2342584320
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.860215053763441e-05,
      "loss": 2.5488,
      "theoretical_loss": 3.3861398967656435,
      "tokens_seen": 2342715392
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8594126143476166e-05,
      "loss": 2.393,
      "theoretical_loss": 3.3861245881432627,
      "tokens_seen": 2342846464
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.858610174931793e-05,
      "loss": 2.5483,
      "theoretical_loss": 3.3861092806170996,
      "tokens_seen": 2342977536
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1292270,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6421544551849365,
      "objective/train/theoretical_loss": 3.386101627265056,
      "objective/train/tokens_used": 713502176,
      "theoretical_loss": 3.386101627265056,
      "tokens_seen": 2343043072
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8578077355159686e-05,
      "loss": 2.5855,
      "theoretical_loss": 3.3860939741870135,
      "tokens_seen": 2343108608
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.857005296100144e-05,
      "loss": 2.3921,
      "theoretical_loss": 3.3860786688528655,
      "tokens_seen": 2343239680
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8562028566843206e-05,
      "loss": 2.587,
      "theoretical_loss": 3.386063364614515,
      "tokens_seen": 2343370752
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.855400417268496e-05,
      "loss": 2.6345,
      "theoretical_loss": 3.386048061471823,
      "tokens_seen": 2343501824
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8545979778526726e-05,
      "loss": 2.5607,
      "theoretical_loss": 3.386032759424649,
      "tokens_seen": 2343632896
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.853795538436848e-05,
      "loss": 2.3979,
      "theoretical_loss": 3.386017458472854,
      "tokens_seen": 2343763968
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8529930990210246e-05,
      "loss": 2.6633,
      "theoretical_loss": 3.386002158616298,
      "tokens_seen": 2343895040
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8521906596052e-05,
      "loss": 2.489,
      "theoretical_loss": 3.3859868598548415,
      "tokens_seen": 2344026112
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.851388220189375e-05,
      "loss": 2.4694,
      "theoretical_loss": 3.3859715621883444,
      "tokens_seen": 2344157184
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.850585780773552e-05,
      "loss": 2.5831,
      "theoretical_loss": 3.385956265616668,
      "tokens_seen": 2344288256
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.849783341357727e-05,
      "loss": 2.6124,
      "theoretical_loss": 3.3859409701396723,
      "tokens_seen": 2344419328
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.848980901941904e-05,
      "loss": 2.6426,
      "theoretical_loss": 3.3859256757572176,
      "tokens_seen": 2344550400
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1292697,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3175153732299805,
      "objective/train/theoretical_loss": 3.3859103824691643,
      "objective/train/tokens_used": 715140576,
      "theoretical_loss": 3.3859103824691643,
      "tokens_seen": 2344681472
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.848178462526079e-05,
      "loss": 2.6733,
      "theoretical_loss": 3.3859103824691643,
      "tokens_seen": 2344681472
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.847376023110256e-05,
      "loss": 2.5185,
      "theoretical_loss": 3.3858950902753735,
      "tokens_seen": 2344812544
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.846573583694431e-05,
      "loss": 2.6166,
      "theoretical_loss": 3.3858797991757053,
      "tokens_seen": 2344943616
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8457711442786064e-05,
      "loss": 2.5241,
      "theoretical_loss": 3.38586450917002,
      "tokens_seen": 2345074688
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.844968704862783e-05,
      "loss": 2.5079,
      "theoretical_loss": 3.385849220258179,
      "tokens_seen": 2345205760
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8441662654469584e-05,
      "loss": 2.4976,
      "theoretical_loss": 3.3858339324400424,
      "tokens_seen": 2345336832
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.843363826031135e-05,
      "loss": 2.5434,
      "theoretical_loss": 3.3858186457154704,
      "tokens_seen": 2345467904
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8425613866153104e-05,
      "loss": 2.5645,
      "theoretical_loss": 3.3858033600843247,
      "tokens_seen": 2345598976
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.841758947199487e-05,
      "loss": 2.6008,
      "theoretical_loss": 3.3857880755464658,
      "tokens_seen": 2345730048
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8409565077836624e-05,
      "loss": 2.608,
      "theoretical_loss": 3.3857727921017533,
      "tokens_seen": 2345861120
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.840154068367838e-05,
      "loss": 2.5562,
      "theoretical_loss": 3.385757509750049,
      "tokens_seen": 2345992192
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8393516289520144e-05,
      "loss": 2.6444,
      "theoretical_loss": 3.385742228491214,
      "tokens_seen": 2346123264
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.83854918953619e-05,
      "loss": 2.6761,
      "theoretical_loss": 3.385726948325108,
      "tokens_seen": 2346254336
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1293722,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2620153427124023,
      "objective/train/theoretical_loss": 3.385719308651785,
      "objective/train/tokens_used": 716778976,
      "theoretical_loss": 3.385719308651785,
      "tokens_seen": 2346319872
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8377467501203663e-05,
      "loss": 2.4176,
      "theoretical_loss": 3.3857116692515925,
      "tokens_seen": 2346385408
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8369443107045417e-05,
      "loss": 2.6602,
      "theoretical_loss": 3.385696391270528,
      "tokens_seen": 2346516480
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.836141871288718e-05,
      "loss": 2.535,
      "theoretical_loss": 3.3856811143817755,
      "tokens_seen": 2346647552
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8353394318728936e-05,
      "loss": 2.5504,
      "theoretical_loss": 3.385665838585196,
      "tokens_seen": 2346778624
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.83453699245707e-05,
      "loss": 2.6594,
      "theoretical_loss": 3.3856505638806507,
      "tokens_seen": 2346909696
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8337345530412456e-05,
      "loss": 2.616,
      "theoretical_loss": 3.3856352902679996,
      "tokens_seen": 2347040768
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.832932113625421e-05,
      "loss": 2.6099,
      "theoretical_loss": 3.385620017747105,
      "tokens_seen": 2347171840
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8321296742095976e-05,
      "loss": 2.6064,
      "theoretical_loss": 3.385604746317827,
      "tokens_seen": 2347302912
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.831327234793773e-05,
      "loss": 2.5122,
      "theoretical_loss": 3.3855894759800265,
      "tokens_seen": 2347433984
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8305247953779496e-05,
      "loss": 2.3559,
      "theoretical_loss": 3.3855742067335655,
      "tokens_seen": 2347565056
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.829722355962125e-05,
      "loss": 2.5931,
      "theoretical_loss": 3.385558938578304,
      "tokens_seen": 2347696128
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8289199165463016e-05,
      "loss": 2.6417,
      "theoretical_loss": 3.385543671514104,
      "tokens_seen": 2347827200
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1294165,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2179014682769775,
      "objective/train/theoretical_loss": 3.3855284055408257,
      "objective/train/tokens_used": 718417376,
      "theoretical_loss": 3.3855284055408257,
      "tokens_seen": 2347958272
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.828117477130477e-05,
      "loss": 2.4197,
      "theoretical_loss": 3.3855284055408257,
      "tokens_seen": 2347958272
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.827315037714652e-05,
      "loss": 2.5461,
      "theoretical_loss": 3.385513140658331,
      "tokens_seen": 2348089344
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.826512598298829e-05,
      "loss": 2.6113,
      "theoretical_loss": 3.385497876866481,
      "tokens_seen": 2348220416
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.825710158883004e-05,
      "loss": 2.6432,
      "theoretical_loss": 3.3854826141651366,
      "tokens_seen": 2348351488
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.824907719467181e-05,
      "loss": 2.5889,
      "theoretical_loss": 3.385467352554159,
      "tokens_seen": 2348482560
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.824105280051356e-05,
      "loss": 2.5039,
      "theoretical_loss": 3.38545209203341,
      "tokens_seen": 2348613632
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.823302840635533e-05,
      "loss": 2.6764,
      "theoretical_loss": 3.38543683260275,
      "tokens_seen": 2348744704
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.822500401219708e-05,
      "loss": 2.6093,
      "theoretical_loss": 3.3854215742620415,
      "tokens_seen": 2348875776
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8216979618038834e-05,
      "loss": 2.4995,
      "theoretical_loss": 3.3854063170111446,
      "tokens_seen": 2349006848
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.82089552238806e-05,
      "loss": 2.7344,
      "theoretical_loss": 3.3853910608499214,
      "tokens_seen": 2349137920
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8200930829722354e-05,
      "loss": 2.7268,
      "theoretical_loss": 3.3853758057782333,
      "tokens_seen": 2349268992
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.819290643556412e-05,
      "loss": 2.4094,
      "theoretical_loss": 3.385360551795941,
      "tokens_seen": 2349400064
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8184882041405874e-05,
      "loss": 2.6371,
      "theoretical_loss": 3.385345298902907,
      "tokens_seen": 2349531136
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1295470,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5095014572143555,
      "objective/train/theoretical_loss": 3.385337672864818,
      "objective/train/tokens_used": 720055776,
      "theoretical_loss": 3.385337672864818,
      "tokens_seen": 2349596672
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.817685764724764e-05,
      "loss": 2.7028,
      "theoretical_loss": 3.385330047098992,
      "tokens_seen": 2349662208
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8168833253089394e-05,
      "loss": 2.3716,
      "theoretical_loss": 3.3853147963840575,
      "tokens_seen": 2349793280
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.816080885893115e-05,
      "loss": 2.5393,
      "theoretical_loss": 3.385299546757965,
      "tokens_seen": 2349924352
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8152784464772914e-05,
      "loss": 2.6145,
      "theoretical_loss": 3.3852842982205766,
      "tokens_seen": 2350055424
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.814476007061467e-05,
      "loss": 2.5938,
      "theoretical_loss": 3.385269050771753,
      "tokens_seen": 2350186496
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8136735676456433e-05,
      "loss": 2.6389,
      "theoretical_loss": 3.3852538044113567,
      "tokens_seen": 2350317568
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.812871128229819e-05,
      "loss": 2.5708,
      "theoretical_loss": 3.385238559139249,
      "tokens_seen": 2350448640
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.812068688813995e-05,
      "loss": 2.6075,
      "theoretical_loss": 3.385223314955291,
      "tokens_seen": 2350579712
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8112662493981706e-05,
      "loss": 2.5073,
      "theoretical_loss": 3.385208071859345,
      "tokens_seen": 2350710784
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.810463809982346e-05,
      "loss": 2.4663,
      "theoretical_loss": 3.3851928298512726,
      "tokens_seen": 2350841856
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.8096613705665226e-05,
      "loss": 2.573,
      "theoretical_loss": 3.3851775889309352,
      "tokens_seen": 2350972928
    },
    {
      "epoch": 0.42,
      "learning_rate": 5.808858931150698e-05,
      "loss": 2.533,
      "theoretical_loss": 3.385162349098195,
      "tokens_seen": 2351104000
    },
    {
      "epoch": 0.42,
      "objective/train/docs_used": 1295872,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.463392734527588,
      "objective/train/theoretical_loss": 3.385147110352913,
      "objective/train/tokens_used": 721694176,
      "theoretical_loss": 3.385147110352913,
      "tokens_seen": 2351235072
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.8080564917348746e-05,
      "loss": 2.4912,
      "theoretical_loss": 3.385147110352913,
      "tokens_seen": 2351235072
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.80725405231905e-05,
      "loss": 2.6252,
      "theoretical_loss": 3.3851318726949513,
      "tokens_seen": 2351366144
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.8064516129032266e-05,
      "loss": 2.4905,
      "theoretical_loss": 3.3851166361241725,
      "tokens_seen": 2351497216
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.805649173487402e-05,
      "loss": 2.5806,
      "theoretical_loss": 3.385101400640438,
      "tokens_seen": 2351628288
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.8048467340715786e-05,
      "loss": 2.649,
      "theoretical_loss": 3.3850861662436085,
      "tokens_seen": 2351759360
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.804044294655754e-05,
      "loss": 2.514,
      "theoretical_loss": 3.3850709329335475,
      "tokens_seen": 2351890432
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.803241855239929e-05,
      "loss": 2.4705,
      "theoretical_loss": 3.385055700710116,
      "tokens_seen": 2352021504
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.802439415824106e-05,
      "loss": 2.5074,
      "theoretical_loss": 3.3850404695731764,
      "tokens_seen": 2352152576
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.801636976408281e-05,
      "loss": 2.5268,
      "theoretical_loss": 3.385025239522591,
      "tokens_seen": 2352283648
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.800834536992458e-05,
      "loss": 2.5474,
      "theoretical_loss": 3.3850100105582204,
      "tokens_seen": 2352414720
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.800032097576633e-05,
      "loss": 2.5474,
      "theoretical_loss": 3.384994782679928,
      "tokens_seen": 2352545792
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.79922965816081e-05,
      "loss": 2.6361,
      "theoretical_loss": 3.384979555887575,
      "tokens_seen": 2352676864
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.798427218744985e-05,
      "loss": 2.5841,
      "theoretical_loss": 3.384964330181024,
      "tokens_seen": 2352807936
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1296233,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5529563426971436,
      "objective/train/theoretical_loss": 3.3849567177348807,
      "objective/train/tokens_used": 723332576,
      "theoretical_loss": 3.3849567177348807,
      "tokens_seen": 2352873472
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7976247793291605e-05,
      "loss": 2.6446,
      "theoretical_loss": 3.3849491055601364,
      "tokens_seen": 2352939008
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.796822339913337e-05,
      "loss": 2.4279,
      "theoretical_loss": 3.3849338820247756,
      "tokens_seen": 2353070080
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7960199004975124e-05,
      "loss": 2.6022,
      "theoretical_loss": 3.3849186595748026,
      "tokens_seen": 2353201152
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.795217461081689e-05,
      "loss": 2.6832,
      "theoretical_loss": 3.3849034382100798,
      "tokens_seen": 2353332224
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7944150216658644e-05,
      "loss": 2.6591,
      "theoretical_loss": 3.3848882179304693,
      "tokens_seen": 2353463296
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.793612582250041e-05,
      "loss": 2.6853,
      "theoretical_loss": 3.384872998735834,
      "tokens_seen": 2353594368
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7928101428342164e-05,
      "loss": 2.6657,
      "theoretical_loss": 3.3848577806260356,
      "tokens_seen": 2353725440
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.792007703418392e-05,
      "loss": 2.6369,
      "theoretical_loss": 3.3848425636009365,
      "tokens_seen": 2353856512
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7912052640025684e-05,
      "loss": 2.5669,
      "theoretical_loss": 3.3848273476603987,
      "tokens_seen": 2353987584
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.790402824586744e-05,
      "loss": 2.6338,
      "theoretical_loss": 3.3848121328042846,
      "tokens_seen": 2354118656
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7896003851709204e-05,
      "loss": 2.6512,
      "theoretical_loss": 3.384796919032457,
      "tokens_seen": 2354249728
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.788797945755096e-05,
      "loss": 2.6298,
      "theoretical_loss": 3.3847817063447776,
      "tokens_seen": 2354380800
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1296233,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4254140853881836,
      "objective/train/theoretical_loss": 3.3847664947411094,
      "objective/train/tokens_used": 724970976,
      "theoretical_loss": 3.3847664947411094,
      "tokens_seen": 2354511872
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.787995506339272e-05,
      "loss": 2.586,
      "theoretical_loss": 3.3847664947411094,
      "tokens_seen": 2354511872
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7871930669234477e-05,
      "loss": 2.6616,
      "theoretical_loss": 3.3847512842213145,
      "tokens_seen": 2354642944
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.786390627507623e-05,
      "loss": 2.7663,
      "theoretical_loss": 3.384736074785255,
      "tokens_seen": 2354774016
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7855881880917996e-05,
      "loss": 2.7011,
      "theoretical_loss": 3.384720866432794,
      "tokens_seen": 2354905088
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.784785748675975e-05,
      "loss": 2.8609,
      "theoretical_loss": 3.384705659163794,
      "tokens_seen": 2355036160
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7839833092601516e-05,
      "loss": 2.6335,
      "theoretical_loss": 3.384690452978117,
      "tokens_seen": 2355167232
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.783180869844327e-05,
      "loss": 2.7139,
      "theoretical_loss": 3.3846752478756263,
      "tokens_seen": 2355298304
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7823784304285036e-05,
      "loss": 2.5951,
      "theoretical_loss": 3.384660043856184,
      "tokens_seen": 2355429376
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.781575991012679e-05,
      "loss": 2.7079,
      "theoretical_loss": 3.384644840919652,
      "tokens_seen": 2355560448
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.780773551596854e-05,
      "loss": 2.6649,
      "theoretical_loss": 3.384629639065894,
      "tokens_seen": 2355691520
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.779971112181031e-05,
      "loss": 2.8242,
      "theoretical_loss": 3.384614438294772,
      "tokens_seen": 2355822592
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.779168672765206e-05,
      "loss": 2.6926,
      "theoretical_loss": 3.384599238606149,
      "tokens_seen": 2355953664
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.778366233349383e-05,
      "loss": 2.7872,
      "theoretical_loss": 3.384584039999888,
      "tokens_seen": 2356084736
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1296233,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6371724605560303,
      "objective/train/theoretical_loss": 3.3845764411026,
      "objective/train/tokens_used": 726609376,
      "theoretical_loss": 3.3845764411026,
      "tokens_seen": 2356150272
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.777563793933558e-05,
      "loss": 2.6989,
      "theoretical_loss": 3.384568842475851,
      "tokens_seen": 2356215808
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.776761354517735e-05,
      "loss": 2.7725,
      "theoretical_loss": 3.3845536460339014,
      "tokens_seen": 2356346880
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.77595891510191e-05,
      "loss": 2.8416,
      "theoretical_loss": 3.384538450673901,
      "tokens_seen": 2356477952
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.775156475686087e-05,
      "loss": 2.7238,
      "theoretical_loss": 3.3845232563957137,
      "tokens_seen": 2356609024
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.774354036270262e-05,
      "loss": 2.6543,
      "theoretical_loss": 3.3845080631992017,
      "tokens_seen": 2356740096
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7735515968544375e-05,
      "loss": 2.7049,
      "theoretical_loss": 3.384492871084228,
      "tokens_seen": 2356871168
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.772749157438614e-05,
      "loss": 2.7103,
      "theoretical_loss": 3.3844776800506557,
      "tokens_seen": 2357002240
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7719467180227894e-05,
      "loss": 2.6335,
      "theoretical_loss": 3.3844624900983473,
      "tokens_seen": 2357133312
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.771144278606966e-05,
      "loss": 2.5968,
      "theoretical_loss": 3.3844473012271656,
      "tokens_seen": 2357264384
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7703418391911414e-05,
      "loss": 2.8002,
      "theoretical_loss": 3.3844321134369744,
      "tokens_seen": 2357395456
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.769539399775318e-05,
      "loss": 2.7501,
      "theoretical_loss": 3.384416926727636,
      "tokens_seen": 2357526528
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7687369603594934e-05,
      "loss": 2.7608,
      "theoretical_loss": 3.384401741099013,
      "tokens_seen": 2357657600
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1296233,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.529184103012085,
      "objective/train/theoretical_loss": 3.384386556550969,
      "objective/train/tokens_used": 728247776,
      "theoretical_loss": 3.384386556550969,
      "tokens_seen": 2357788672
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.767934520943669e-05,
      "loss": 2.7595,
      "theoretical_loss": 3.384386556550969,
      "tokens_seen": 2357788672
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7671320815278454e-05,
      "loss": 2.7837,
      "theoretical_loss": 3.384371373083367,
      "tokens_seen": 2357919744
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.766329642112021e-05,
      "loss": 2.7426,
      "theoretical_loss": 3.3843561906960704,
      "tokens_seen": 2358050816
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7655272026961974e-05,
      "loss": 2.8049,
      "theoretical_loss": 3.3843410093889417,
      "tokens_seen": 2358181888
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.764724763280373e-05,
      "loss": 2.836,
      "theoretical_loss": 3.384325829161844,
      "tokens_seen": 2358312960
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.763922323864549e-05,
      "loss": 2.7711,
      "theoretical_loss": 3.384310650014641,
      "tokens_seen": 2358444032
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7631198844487247e-05,
      "loss": 2.8929,
      "theoretical_loss": 3.3842954719471954,
      "tokens_seen": 2358575104
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7623174450329e-05,
      "loss": 2.7552,
      "theoretical_loss": 3.3842802949593707,
      "tokens_seen": 2358706176
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7615150056170766e-05,
      "loss": 2.6488,
      "theoretical_loss": 3.3842651190510296,
      "tokens_seen": 2358837248
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.760712566201252e-05,
      "loss": 2.7407,
      "theoretical_loss": 3.384249944222036,
      "tokens_seen": 2358968320
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7599101267854286e-05,
      "loss": 2.8939,
      "theoretical_loss": 3.384234770472253,
      "tokens_seen": 2359099392
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.759107687369604e-05,
      "loss": 2.8017,
      "theoretical_loss": 3.3842195978015432,
      "tokens_seen": 2359230464
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.75830524795378e-05,
      "loss": 2.7878,
      "theoretical_loss": 3.384204426209771,
      "tokens_seen": 2359361536
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1296233,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9862067699432373,
      "objective/train/theoretical_loss": 3.3841968408184435,
      "objective/train/tokens_used": 729886176,
      "theoretical_loss": 3.3841968408184435,
      "tokens_seen": 2359427072
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.757502808537956e-05,
      "loss": 2.6802,
      "theoretical_loss": 3.384189255696799,
      "tokens_seen": 2359492608
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.756700369122131e-05,
      "loss": 2.6695,
      "theoretical_loss": 3.3841740862624907,
      "tokens_seen": 2359623680
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.755897929706308e-05,
      "loss": 2.6718,
      "theoretical_loss": 3.38415891790671,
      "tokens_seen": 2359754752
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.755095490290483e-05,
      "loss": 2.9466,
      "theoretical_loss": 3.3841437506293195,
      "tokens_seen": 2359885824
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.754293050874659e-05,
      "loss": 2.8172,
      "theoretical_loss": 3.384128584430183,
      "tokens_seen": 2360016896
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.753490611458835e-05,
      "loss": 2.8624,
      "theoretical_loss": 3.3841134193091644,
      "tokens_seen": 2360147968
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.752688172043011e-05,
      "loss": 2.8597,
      "theoretical_loss": 3.3840982552661263,
      "tokens_seen": 2360279040
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.751885732627187e-05,
      "loss": 2.8153,
      "theoretical_loss": 3.3840830923009335,
      "tokens_seen": 2360410112
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7510832932113625e-05,
      "loss": 2.8208,
      "theoretical_loss": 3.384067930413448,
      "tokens_seen": 2360541184
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.750280853795539e-05,
      "loss": 2.8885,
      "theoretical_loss": 3.3840527696035343,
      "tokens_seen": 2360672256
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7494784143797145e-05,
      "loss": 2.9305,
      "theoretical_loss": 3.384037609871056,
      "tokens_seen": 2360803328
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7486759749638905e-05,
      "loss": 2.7255,
      "theoretical_loss": 3.3840224512158765,
      "tokens_seen": 2360934400
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1296233,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.547431468963623,
      "objective/train/theoretical_loss": 3.3840072936378593,
      "objective/train/tokens_used": 731524576,
      "theoretical_loss": 3.3840072936378593,
      "tokens_seen": 2361065472
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7478735355480664e-05,
      "loss": 2.6954,
      "theoretical_loss": 3.3840072936378593,
      "tokens_seen": 2361065472
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7470710961322424e-05,
      "loss": 2.9092,
      "theoretical_loss": 3.3839921371368686,
      "tokens_seen": 2361196544
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7462686567164184e-05,
      "loss": 2.8024,
      "theoretical_loss": 3.3839769817127676,
      "tokens_seen": 2361327616
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.745466217300594e-05,
      "loss": 2.7412,
      "theoretical_loss": 3.38396182736542,
      "tokens_seen": 2361458688
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.74466377788477e-05,
      "loss": 2.9804,
      "theoretical_loss": 3.38394667409469,
      "tokens_seen": 2361589760
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.743861338468946e-05,
      "loss": 2.8076,
      "theoretical_loss": 3.3839315219004407,
      "tokens_seen": 2361720832
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.743058899053122e-05,
      "loss": 2.873,
      "theoretical_loss": 3.3839163707825364,
      "tokens_seen": 2361851904
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.742256459637298e-05,
      "loss": 2.8507,
      "theoretical_loss": 3.383901220740841,
      "tokens_seen": 2361982976
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.741454020221474e-05,
      "loss": 2.8076,
      "theoretical_loss": 3.383886071775218,
      "tokens_seen": 2362114048
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.74065158080565e-05,
      "loss": 2.9191,
      "theoretical_loss": 3.383870923885531,
      "tokens_seen": 2362245120
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.739849141389826e-05,
      "loss": 2.8189,
      "theoretical_loss": 3.383855777071645,
      "tokens_seen": 2362376192
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.739046701974001e-05,
      "loss": 2.7455,
      "theoretical_loss": 3.3838406313334226,
      "tokens_seen": 2362507264
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.738244262558177e-05,
      "loss": 2.7955,
      "theoretical_loss": 3.3838254866707285,
      "tokens_seen": 2362638336
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1296897,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8064303398132324,
      "objective/train/theoretical_loss": 3.3838179147426617,
      "objective/train/tokens_used": 733162976,
      "theoretical_loss": 3.3838179147426617,
      "tokens_seen": 2362703872
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.737441823142353e-05,
      "loss": 2.7695,
      "theoretical_loss": 3.383810343083426,
      "tokens_seen": 2362769408
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.736639383726529e-05,
      "loss": 2.8729,
      "theoretical_loss": 3.3837952005713805,
      "tokens_seen": 2362900480
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.735836944310705e-05,
      "loss": 2.8079,
      "theoretical_loss": 3.3837800591344545,
      "tokens_seen": 2363031552
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.73503450489488e-05,
      "loss": 2.8238,
      "theoretical_loss": 3.3837649187725125,
      "tokens_seen": 2363162624
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.734232065479057e-05,
      "loss": 2.8355,
      "theoretical_loss": 3.383749779485419,
      "tokens_seen": 2363293696
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.733429626063232e-05,
      "loss": 2.707,
      "theoretical_loss": 3.383734641273038,
      "tokens_seen": 2363424768
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.732627186647408e-05,
      "loss": 2.8741,
      "theoretical_loss": 3.383719504135233,
      "tokens_seen": 2363555840
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.731824747231584e-05,
      "loss": 2.6753,
      "theoretical_loss": 3.383704368071869,
      "tokens_seen": 2363686912
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.73102230781576e-05,
      "loss": 2.7772,
      "theoretical_loss": 3.383689233082809,
      "tokens_seen": 2363817984
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.730219868399936e-05,
      "loss": 2.6709,
      "theoretical_loss": 3.383674099167918,
      "tokens_seen": 2363949056
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7294174289841115e-05,
      "loss": 2.8573,
      "theoretical_loss": 3.38365896632706,
      "tokens_seen": 2364080128
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.728614989568288e-05,
      "loss": 2.8733,
      "theoretical_loss": 3.3836438345601,
      "tokens_seen": 2364211200
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1297792,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8505661487579346,
      "objective/train/theoretical_loss": 3.383628703866901,
      "objective/train/tokens_used": 734801376,
      "theoretical_loss": 3.383628703866901,
      "tokens_seen": 2364342272
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7278125501524635e-05,
      "loss": 2.7348,
      "theoretical_loss": 3.383628703866901,
      "tokens_seen": 2364342272
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7270101107366395e-05,
      "loss": 2.7654,
      "theoretical_loss": 3.3836135742473283,
      "tokens_seen": 2364473344
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7262076713208155e-05,
      "loss": 2.7622,
      "theoretical_loss": 3.3835984457012453,
      "tokens_seen": 2364604416
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.725405231904991e-05,
      "loss": 2.7652,
      "theoretical_loss": 3.383583318228517,
      "tokens_seen": 2364735488
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7246027924891675e-05,
      "loss": 2.7463,
      "theoretical_loss": 3.3835681918290077,
      "tokens_seen": 2364866560
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.723800353073343e-05,
      "loss": 2.7084,
      "theoretical_loss": 3.3835530665025813,
      "tokens_seen": 2364997632
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7229979136575194e-05,
      "loss": 2.8003,
      "theoretical_loss": 3.3835379422491028,
      "tokens_seen": 2365128704
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.722195474241695e-05,
      "loss": 2.7724,
      "theoretical_loss": 3.3835228190684363,
      "tokens_seen": 2365259776
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.721393034825871e-05,
      "loss": 2.8615,
      "theoretical_loss": 3.3835076969604465,
      "tokens_seen": 2365390848
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.720590595410047e-05,
      "loss": 2.7432,
      "theoretical_loss": 3.383492575924998,
      "tokens_seen": 2365521920
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.719788155994222e-05,
      "loss": 2.8073,
      "theoretical_loss": 3.3834774559619545,
      "tokens_seen": 2365652992
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.718985716578399e-05,
      "loss": 2.8,
      "theoretical_loss": 3.3834623370711814,
      "tokens_seen": 2365784064
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.718183277162574e-05,
      "loss": 2.8586,
      "theoretical_loss": 3.3834472192525427,
      "tokens_seen": 2365915136
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1298994,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.880993604660034,
      "objective/train/theoretical_loss": 3.3834396607452315,
      "objective/train/tokens_used": 736439776,
      "theoretical_loss": 3.3834396607452315,
      "tokens_seen": 2365980672
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.717380837746751e-05,
      "loss": 2.748,
      "theoretical_loss": 3.3834321025059033,
      "tokens_seen": 2366046208
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.716578398330926e-05,
      "loss": 2.7962,
      "theoretical_loss": 3.383416986831128,
      "tokens_seen": 2366177280
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.715775958915101e-05,
      "loss": 2.9175,
      "theoretical_loss": 3.383401872228081,
      "tokens_seen": 2366308352
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.714973519499278e-05,
      "loss": 2.7201,
      "theoretical_loss": 3.3833867586966266,
      "tokens_seen": 2366439424
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.714171080083453e-05,
      "loss": 2.8739,
      "theoretical_loss": 3.383371646236631,
      "tokens_seen": 2366570496
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.71336864066763e-05,
      "loss": 2.7995,
      "theoretical_loss": 3.383356534847957,
      "tokens_seen": 2366701568
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.712566201251805e-05,
      "loss": 2.7746,
      "theoretical_loss": 3.3833414245304705,
      "tokens_seen": 2366832640
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.711763761835982e-05,
      "loss": 2.7344,
      "theoretical_loss": 3.3833263152840356,
      "tokens_seen": 2366963712
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.710961322420157e-05,
      "loss": 2.722,
      "theoretical_loss": 3.383311207108518,
      "tokens_seen": 2367094784
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.710158883004334e-05,
      "loss": 2.5951,
      "theoretical_loss": 3.383296100003782,
      "tokens_seen": 2367225856
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.709356443588509e-05,
      "loss": 2.7145,
      "theoretical_loss": 3.383280993969692,
      "tokens_seen": 2367356928
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.7085540041726846e-05,
      "loss": 2.7715,
      "theoretical_loss": 3.3832658890061134,
      "tokens_seen": 2367488000
    },
    {
      "epoch": 0.43,
      "objective/train/docs_used": 1299472,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0345065593719482,
      "objective/train/theoretical_loss": 3.383250785112911,
      "objective/train/tokens_used": 738078176,
      "theoretical_loss": 3.383250785112911,
      "tokens_seen": 2367619072
    },
    {
      "epoch": 0.43,
      "learning_rate": 5.707751564756861e-05,
      "loss": 2.784,
      "theoretical_loss": 3.383250785112911,
      "tokens_seen": 2367619072
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.7069491253410365e-05,
      "loss": 2.6148,
      "theoretical_loss": 3.383235682289949,
      "tokens_seen": 2367750144
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.706146685925213e-05,
      "loss": 2.7107,
      "theoretical_loss": 3.3832205805370936,
      "tokens_seen": 2367881216
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.7053442465093885e-05,
      "loss": 2.8164,
      "theoretical_loss": 3.383205479854209,
      "tokens_seen": 2368012288
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.704541807093565e-05,
      "loss": 2.7386,
      "theoretical_loss": 3.3831903802411603,
      "tokens_seen": 2368143360
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.7037393676777405e-05,
      "loss": 2.8022,
      "theoretical_loss": 3.3831752816978122,
      "tokens_seen": 2368274432
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.702936928261916e-05,
      "loss": 2.6544,
      "theoretical_loss": 3.38316018422403,
      "tokens_seen": 2368405504
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.7021344888460925e-05,
      "loss": 2.7692,
      "theoretical_loss": 3.3831450878196785,
      "tokens_seen": 2368536576
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.701332049430268e-05,
      "loss": 2.8365,
      "theoretical_loss": 3.383129992484623,
      "tokens_seen": 2368667648
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.7005296100144445e-05,
      "loss": 2.803,
      "theoretical_loss": 3.383114898218729,
      "tokens_seen": 2368798720
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.69972717059862e-05,
      "loss": 2.807,
      "theoretical_loss": 3.3830998050218613,
      "tokens_seen": 2368929792
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6989247311827965e-05,
      "loss": 2.9609,
      "theoretical_loss": 3.3830847128938846,
      "tokens_seen": 2369060864
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.698122291766972e-05,
      "loss": 2.7113,
      "theoretical_loss": 3.3830696218346645,
      "tokens_seen": 2369191936
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1300592,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.462731122970581,
      "objective/train/theoretical_loss": 3.383062076705796,
      "objective/train/tokens_used": 739716576,
      "theoretical_loss": 3.383062076705796,
      "tokens_seen": 2369257472
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.697319852351147e-05,
      "loss": 2.6522,
      "theoretical_loss": 3.383054531844066,
      "tokens_seen": 2369323008
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.696517412935324e-05,
      "loss": 2.8167,
      "theoretical_loss": 3.383039442921955,
      "tokens_seen": 2369454080
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.695714973519499e-05,
      "loss": 2.8221,
      "theoretical_loss": 3.3830243550681955,
      "tokens_seen": 2369585152
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.694912534103676e-05,
      "loss": 2.7618,
      "theoretical_loss": 3.383009268282654,
      "tokens_seen": 2369716224
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.694110094687851e-05,
      "loss": 2.6241,
      "theoretical_loss": 3.382994182565195,
      "tokens_seen": 2369847296
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.693307655272028e-05,
      "loss": 2.8554,
      "theoretical_loss": 3.3829790979156837,
      "tokens_seen": 2369978368
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.692505215856203e-05,
      "loss": 2.6757,
      "theoretical_loss": 3.3829640143339863,
      "tokens_seen": 2370109440
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.691702776440378e-05,
      "loss": 2.6053,
      "theoretical_loss": 3.3829489318199673,
      "tokens_seen": 2370240512
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.690900337024555e-05,
      "loss": 2.7019,
      "theoretical_loss": 3.382933850373493,
      "tokens_seen": 2370371584
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.69009789760873e-05,
      "loss": 2.7461,
      "theoretical_loss": 3.3829187699944274,
      "tokens_seen": 2370502656
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.689295458192907e-05,
      "loss": 2.795,
      "theoretical_loss": 3.3829036906826375,
      "tokens_seen": 2370633728
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.688493018777082e-05,
      "loss": 2.7016,
      "theoretical_loss": 3.3828886124379878,
      "tokens_seen": 2370764800
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1301282,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8106579780578613,
      "objective/train/theoretical_loss": 3.3828735352603445,
      "objective/train/tokens_used": 741354976,
      "theoretical_loss": 3.3828735352603445,
      "tokens_seen": 2370895872
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.687690579361259e-05,
      "loss": 2.6631,
      "theoretical_loss": 3.3828735352603445,
      "tokens_seen": 2370895872
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.686888139945434e-05,
      "loss": 2.7346,
      "theoretical_loss": 3.3828584591495723,
      "tokens_seen": 2371026944
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6860857005296096e-05,
      "loss": 2.7316,
      "theoretical_loss": 3.382843384105537,
      "tokens_seen": 2371158016
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.685283261113786e-05,
      "loss": 2.646,
      "theoretical_loss": 3.3828283101281045,
      "tokens_seen": 2371289088
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6844808216979616e-05,
      "loss": 2.6804,
      "theoretical_loss": 3.38281323721714,
      "tokens_seen": 2371420160
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.683678382282138e-05,
      "loss": 2.7237,
      "theoretical_loss": 3.382798165372509,
      "tokens_seen": 2371551232
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6828759428663136e-05,
      "loss": 2.7476,
      "theoretical_loss": 3.382783094594078,
      "tokens_seen": 2371682304
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.68207350345049e-05,
      "loss": 2.8014,
      "theoretical_loss": 3.382768024881712,
      "tokens_seen": 2371813376
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6812710640346655e-05,
      "loss": 2.8276,
      "theoretical_loss": 3.3827529562352763,
      "tokens_seen": 2371944448
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.680468624618841e-05,
      "loss": 2.8166,
      "theoretical_loss": 3.382737888654637,
      "tokens_seen": 2372075520
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6796661852030175e-05,
      "loss": 2.68,
      "theoretical_loss": 3.38272282213966,
      "tokens_seen": 2372206592
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.678863745787193e-05,
      "loss": 2.7083,
      "theoretical_loss": 3.3827077566902113,
      "tokens_seen": 2372337664
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6780613063713695e-05,
      "loss": 2.7768,
      "theoretical_loss": 3.382692692306156,
      "tokens_seen": 2372468736
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1302157,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.031487226486206,
      "objective/train/theoretical_loss": 3.382685160513609,
      "objective/train/tokens_used": 742993376,
      "theoretical_loss": 3.382685160513609,
      "tokens_seen": 2372534272
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.677258866955545e-05,
      "loss": 2.7334,
      "theoretical_loss": 3.38267762898736,
      "tokens_seen": 2372599808
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6764564275397215e-05,
      "loss": 2.6291,
      "theoretical_loss": 3.38266256673369,
      "tokens_seen": 2372730880
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.675653988123897e-05,
      "loss": 2.7014,
      "theoretical_loss": 3.3826475055450107,
      "tokens_seen": 2372861952
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6748515487080735e-05,
      "loss": 2.7827,
      "theoretical_loss": 3.3826324454211885,
      "tokens_seen": 2372993024
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.674049109292249e-05,
      "loss": 2.7505,
      "theoretical_loss": 3.3826173863620896,
      "tokens_seen": 2373124096
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.673246669876424e-05,
      "loss": 2.6665,
      "theoretical_loss": 3.3826023283675792,
      "tokens_seen": 2373255168
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.672444230460601e-05,
      "loss": 2.6076,
      "theoretical_loss": 3.382587271437524,
      "tokens_seen": 2373386240
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.671641791044776e-05,
      "loss": 2.8538,
      "theoretical_loss": 3.3825722155717894,
      "tokens_seen": 2373517312
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.670839351628953e-05,
      "loss": 2.6044,
      "theoretical_loss": 3.3825571607702414,
      "tokens_seen": 2373648384
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.670036912213128e-05,
      "loss": 2.7014,
      "theoretical_loss": 3.3825421070327466,
      "tokens_seen": 2373779456
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.669234472797305e-05,
      "loss": 2.6982,
      "theoretical_loss": 3.3825270543591706,
      "tokens_seen": 2373910528
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.66843203338148e-05,
      "loss": 2.7067,
      "theoretical_loss": 3.3825120027493796,
      "tokens_seen": 2374041600
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1302854,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0018539428710938,
      "objective/train/theoretical_loss": 3.3824969522032395,
      "objective/train/tokens_used": 744631776,
      "theoretical_loss": 3.3824969522032395,
      "tokens_seen": 2374172672
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6676295939656553e-05,
      "loss": 2.8434,
      "theoretical_loss": 3.3824969522032395,
      "tokens_seen": 2374172672
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.666827154549832e-05,
      "loss": 2.6915,
      "theoretical_loss": 3.3824819027206168,
      "tokens_seen": 2374303744
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.666024715134007e-05,
      "loss": 2.7146,
      "theoretical_loss": 3.382466854301377,
      "tokens_seen": 2374434816
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.665222275718184e-05,
      "loss": 2.7207,
      "theoretical_loss": 3.382451806945387,
      "tokens_seen": 2374565888
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.664419836302359e-05,
      "loss": 2.6916,
      "theoretical_loss": 3.3824367606525128,
      "tokens_seen": 2374696960
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.663617396886536e-05,
      "loss": 2.5943,
      "theoretical_loss": 3.3824217154226206,
      "tokens_seen": 2374828032
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.662814957470711e-05,
      "loss": 2.7399,
      "theoretical_loss": 3.382406671255576,
      "tokens_seen": 2374959104
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6620125180548866e-05,
      "loss": 2.7384,
      "theoretical_loss": 3.3823916281512463,
      "tokens_seen": 2375090176
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.661210078639063e-05,
      "loss": 2.6197,
      "theoretical_loss": 3.382376586109497,
      "tokens_seen": 2375221248
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6604076392232386e-05,
      "loss": 2.7029,
      "theoretical_loss": 3.3823615451301947,
      "tokens_seen": 2375352320
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.659605199807415e-05,
      "loss": 2.7693,
      "theoretical_loss": 3.382346505213206,
      "tokens_seen": 2375483392
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6588027603915906e-05,
      "loss": 2.5935,
      "theoretical_loss": 3.382331466358397,
      "tokens_seen": 2375614464
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.658000320975767e-05,
      "loss": 2.7039,
      "theoretical_loss": 3.382316428565634,
      "tokens_seen": 2375745536
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1303970,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.590346336364746,
      "objective/train/theoretical_loss": 3.382308910067478,
      "objective/train/tokens_used": 746270176,
      "theoretical_loss": 3.382308910067478,
      "tokens_seen": 2375811072
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6571978815599425e-05,
      "loss": 2.6357,
      "theoretical_loss": 3.3823013918347833,
      "tokens_seen": 2375876608
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.656395442144118e-05,
      "loss": 2.6932,
      "theoretical_loss": 3.3822863561657117,
      "tokens_seen": 2376007680
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6555930027282945e-05,
      "loss": 2.716,
      "theoretical_loss": 3.382271321558286,
      "tokens_seen": 2376138752
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.65479056331247e-05,
      "loss": 2.6002,
      "theoretical_loss": 3.3822562880123717,
      "tokens_seen": 2376269824
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6539881238966465e-05,
      "loss": 2.6741,
      "theoretical_loss": 3.382241255527836,
      "tokens_seen": 2376400896
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.653185684480822e-05,
      "loss": 2.8283,
      "theoretical_loss": 3.3822262241045444,
      "tokens_seen": 2376531968
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6523832450649985e-05,
      "loss": 2.7868,
      "theoretical_loss": 3.382211193742365,
      "tokens_seen": 2376663040
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.651580805649174e-05,
      "loss": 2.7779,
      "theoretical_loss": 3.3821961644411633,
      "tokens_seen": 2376794112
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.650778366233349e-05,
      "loss": 2.6668,
      "theoretical_loss": 3.3821811362008067,
      "tokens_seen": 2376925184
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.649975926817526e-05,
      "loss": 2.8336,
      "theoretical_loss": 3.382166109021161,
      "tokens_seen": 2377056256
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.649173487401701e-05,
      "loss": 2.7752,
      "theoretical_loss": 3.3821510829020935,
      "tokens_seen": 2377187328
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.648371047985878e-05,
      "loss": 2.5915,
      "theoretical_loss": 3.3821360578434705,
      "tokens_seen": 2377318400
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1304587,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7148685455322266,
      "objective/train/theoretical_loss": 3.3821210338451593,
      "objective/train/tokens_used": 747908576,
      "theoretical_loss": 3.3821210338451593,
      "tokens_seen": 2377449472
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.647568608570053e-05,
      "loss": 2.7112,
      "theoretical_loss": 3.3821210338451593,
      "tokens_seen": 2377449472
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.64676616915423e-05,
      "loss": 2.5038,
      "theoretical_loss": 3.382106010907026,
      "tokens_seen": 2377580544
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.645963729738405e-05,
      "loss": 2.5681,
      "theoretical_loss": 3.382090989028937,
      "tokens_seen": 2377711616
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.645161290322582e-05,
      "loss": 2.6103,
      "theoretical_loss": 3.3820759682107595,
      "tokens_seen": 2377842688
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.644358850906757e-05,
      "loss": 2.6683,
      "theoretical_loss": 3.3820609484523603,
      "tokens_seen": 2377973760
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6435564114909324e-05,
      "loss": 2.7189,
      "theoretical_loss": 3.3820459297536067,
      "tokens_seen": 2378104832
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.642753972075109e-05,
      "loss": 2.8319,
      "theoretical_loss": 3.3820309121143652,
      "tokens_seen": 2378235904
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.641951532659284e-05,
      "loss": 2.578,
      "theoretical_loss": 3.3820158955345025,
      "tokens_seen": 2378366976
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.641149093243461e-05,
      "loss": 2.5791,
      "theoretical_loss": 3.3820008800138854,
      "tokens_seen": 2378498048
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.640346653827636e-05,
      "loss": 2.7067,
      "theoretical_loss": 3.381985865552381,
      "tokens_seen": 2378629120
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.639544214411813e-05,
      "loss": 2.6642,
      "theoretical_loss": 3.3819708521498564,
      "tokens_seen": 2378760192
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.638741774995988e-05,
      "loss": 2.4669,
      "theoretical_loss": 3.381955839806178,
      "tokens_seen": 2378891264
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6379393355801636e-05,
      "loss": 2.6951,
      "theoretical_loss": 3.3819408285212136,
      "tokens_seen": 2379022336
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1305970,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3961341381073,
      "objective/train/theoretical_loss": 3.3819333232757076,
      "objective/train/tokens_used": 749546976,
      "theoretical_loss": 3.3819333232757076,
      "tokens_seen": 2379087872
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.63713689616434e-05,
      "loss": 2.5991,
      "theoretical_loss": 3.3819258182948295,
      "tokens_seen": 2379153408
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6363344567485156e-05,
      "loss": 2.516,
      "theoretical_loss": 3.3819108091268935,
      "tokens_seen": 2379284480
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.635532017332692e-05,
      "loss": 2.6829,
      "theoretical_loss": 3.3818958010172717,
      "tokens_seen": 2379415552
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6347295779168676e-05,
      "loss": 2.7288,
      "theoretical_loss": 3.381880793965832,
      "tokens_seen": 2379546624
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.633927138501044e-05,
      "loss": 2.5515,
      "theoretical_loss": 3.3818657879724414,
      "tokens_seen": 2379677696
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6331246990852196e-05,
      "loss": 2.617,
      "theoretical_loss": 3.3818507830369664,
      "tokens_seen": 2379808768
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.632322259669395e-05,
      "loss": 2.7448,
      "theoretical_loss": 3.3818357791592746,
      "tokens_seen": 2379939840
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6315198202535715e-05,
      "loss": 2.7401,
      "theoretical_loss": 3.381820776339233,
      "tokens_seen": 2380070912
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.630717380837747e-05,
      "loss": 2.5827,
      "theoretical_loss": 3.38180577457671,
      "tokens_seen": 2380201984
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6299149414219235e-05,
      "loss": 2.735,
      "theoretical_loss": 3.381790773871571,
      "tokens_seen": 2380333056
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.629112502006099e-05,
      "loss": 2.6706,
      "theoretical_loss": 3.381775774223684,
      "tokens_seen": 2380464128
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6283100625902755e-05,
      "loss": 2.7461,
      "theoretical_loss": 3.3817607756329164,
      "tokens_seen": 2380595200
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1307290,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9881303310394287,
      "objective/train/theoretical_loss": 3.3817457780991353,
      "objective/train/tokens_used": 751185376,
      "theoretical_loss": 3.3817457780991353,
      "tokens_seen": 2380726272
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.627507623174451e-05,
      "loss": 2.6021,
      "theoretical_loss": 3.3817457780991353,
      "tokens_seen": 2380726272
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.626705183758626e-05,
      "loss": 2.6797,
      "theoretical_loss": 3.3817307816222084,
      "tokens_seen": 2380857344
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.625902744342803e-05,
      "loss": 2.6479,
      "theoretical_loss": 3.381715786202003,
      "tokens_seen": 2380988416
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.625100304926978e-05,
      "loss": 2.54,
      "theoretical_loss": 3.381700791838386,
      "tokens_seen": 2381119488
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.624297865511155e-05,
      "loss": 2.5575,
      "theoretical_loss": 3.381685798531225,
      "tokens_seen": 2381250560
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.62349542609533e-05,
      "loss": 2.6102,
      "theoretical_loss": 3.3816708062803875,
      "tokens_seen": 2381381632
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.622692986679506e-05,
      "loss": 2.6952,
      "theoretical_loss": 3.3816558150857414,
      "tokens_seen": 2381512704
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.621890547263682e-05,
      "loss": 2.497,
      "theoretical_loss": 3.3816408249471532,
      "tokens_seen": 2381643776
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6210881078478574e-05,
      "loss": 2.6587,
      "theoretical_loss": 3.381625835864491,
      "tokens_seen": 2381774848
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.620285668432034e-05,
      "loss": 2.6085,
      "theoretical_loss": 3.3816108478376226,
      "tokens_seen": 2381905920
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6194832290162094e-05,
      "loss": 2.6701,
      "theoretical_loss": 3.3815958608664145,
      "tokens_seen": 2382036992
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6186807896003853e-05,
      "loss": 2.6523,
      "theoretical_loss": 3.3815808749507354,
      "tokens_seen": 2382168064
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6178783501845613e-05,
      "loss": 2.6501,
      "theoretical_loss": 3.381565890090452,
      "tokens_seen": 2382299136
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1307945,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9193787574768066,
      "objective/train/theoretical_loss": 3.381558398056043,
      "objective/train/tokens_used": 752823776,
      "theoretical_loss": 3.381558398056043,
      "tokens_seen": 2382364672
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.617075910768737e-05,
      "loss": 2.7432,
      "theoretical_loss": 3.3815509062854328,
      "tokens_seen": 2382430208
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.616273471352913e-05,
      "loss": 2.5401,
      "theoretical_loss": 3.381535923535545,
      "tokens_seen": 2382561280
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.615471031937089e-05,
      "loss": 2.6953,
      "theoretical_loss": 3.3815209418406558,
      "tokens_seen": 2382692352
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.614668592521265e-05,
      "loss": 2.5979,
      "theoretical_loss": 3.3815059612006335,
      "tokens_seen": 2382823424
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6138661531054406e-05,
      "loss": 2.6828,
      "theoretical_loss": 3.381490981615346,
      "tokens_seen": 2382954496
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6130637136896166e-05,
      "loss": 2.7251,
      "theoretical_loss": 3.38147600308466,
      "tokens_seen": 2383085568
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6122612742737926e-05,
      "loss": 2.4586,
      "theoretical_loss": 3.381461025608444,
      "tokens_seen": 2383216640
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6114588348579686e-05,
      "loss": 2.5998,
      "theoretical_loss": 3.381446049186566,
      "tokens_seen": 2383347712
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6106563954421446e-05,
      "loss": 2.6438,
      "theoretical_loss": 3.3814310738188933,
      "tokens_seen": 2383478784
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.6098539560263206e-05,
      "loss": 2.6546,
      "theoretical_loss": 3.3814160995052935,
      "tokens_seen": 2383609856
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.609051516610496e-05,
      "loss": 2.6753,
      "theoretical_loss": 3.3814011262456356,
      "tokens_seen": 2383740928
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.608249077194672e-05,
      "loss": 2.5646,
      "theoretical_loss": 3.3813861540397863,
      "tokens_seen": 2383872000
    },
    {
      "epoch": 0.44,
      "objective/train/docs_used": 1309221,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7739500999450684,
      "objective/train/theoretical_loss": 3.381371182887614,
      "objective/train/tokens_used": 754462176,
      "theoretical_loss": 3.381371182887614,
      "tokens_seen": 2384003072
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.607446637778848e-05,
      "loss": 2.7357,
      "theoretical_loss": 3.381371182887614,
      "tokens_seen": 2384003072
    },
    {
      "epoch": 0.44,
      "learning_rate": 5.606644198363024e-05,
      "loss": 2.6646,
      "theoretical_loss": 3.3813562127889867,
      "tokens_seen": 2384134144
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.6058417589472e-05,
      "loss": 2.5678,
      "theoretical_loss": 3.3813412437437718,
      "tokens_seen": 2384265216
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.605039319531376e-05,
      "loss": 2.5437,
      "theoretical_loss": 3.381326275751838,
      "tokens_seen": 2384396288
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.604236880115552e-05,
      "loss": 2.7734,
      "theoretical_loss": 3.3813113088130526,
      "tokens_seen": 2384527360
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.603434440699727e-05,
      "loss": 2.6464,
      "theoretical_loss": 3.3812963429272846,
      "tokens_seen": 2384658432
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.602632001283903e-05,
      "loss": 2.514,
      "theoretical_loss": 3.381281378094401,
      "tokens_seen": 2384789504
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.601829561868079e-05,
      "loss": 2.7409,
      "theoretical_loss": 3.38126641431427,
      "tokens_seen": 2384920576
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.601027122452255e-05,
      "loss": 2.7185,
      "theoretical_loss": 3.3812514515867607,
      "tokens_seen": 2385051648
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.600224683036431e-05,
      "loss": 2.5876,
      "theoretical_loss": 3.38123648991174,
      "tokens_seen": 2385182720
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5994222436206064e-05,
      "loss": 2.5251,
      "theoretical_loss": 3.3812215292890766,
      "tokens_seen": 2385313792
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.598619804204783e-05,
      "loss": 2.5594,
      "theoretical_loss": 3.3812065697186386,
      "tokens_seen": 2385444864
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5978173647889584e-05,
      "loss": 2.6084,
      "theoretical_loss": 3.381191611200294,
      "tokens_seen": 2385575936
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1309756,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.809346914291382,
      "objective/train/theoretical_loss": 3.381184132335616,
      "objective/train/tokens_used": 756100576,
      "theoretical_loss": 3.381184132335616,
      "tokens_seen": 2385641472
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5970149253731344e-05,
      "loss": 2.6377,
      "theoretical_loss": 3.381176653733912,
      "tokens_seen": 2385707008
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5962124859573104e-05,
      "loss": 2.6447,
      "theoretical_loss": 3.3811616973193592,
      "tokens_seen": 2385838080
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5954100465414864e-05,
      "loss": 2.6085,
      "theoretical_loss": 3.3811467419565044,
      "tokens_seen": 2385969152
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5946076071256624e-05,
      "loss": 2.5808,
      "theoretical_loss": 3.381131787645217,
      "tokens_seen": 2386100224
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.593805167709838e-05,
      "loss": 2.6756,
      "theoretical_loss": 3.381116834385364,
      "tokens_seen": 2386231296
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.593002728294014e-05,
      "loss": 2.4824,
      "theoretical_loss": 3.381101882176814,
      "tokens_seen": 2386362368
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5922002888781897e-05,
      "loss": 2.7075,
      "theoretical_loss": 3.3810869310194356,
      "tokens_seen": 2386493440
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5913978494623656e-05,
      "loss": 2.6192,
      "theoretical_loss": 3.3810719809130974,
      "tokens_seen": 2386624512
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5905954100465416e-05,
      "loss": 2.4395,
      "theoretical_loss": 3.381057031857667,
      "tokens_seen": 2386755584
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.589792970630717e-05,
      "loss": 2.5051,
      "theoretical_loss": 3.3810420838530137,
      "tokens_seen": 2386886656
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5889905312148936e-05,
      "loss": 2.5776,
      "theoretical_loss": 3.3810271368990055,
      "tokens_seen": 2387017728
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.588188091799069e-05,
      "loss": 2.6031,
      "theoretical_loss": 3.3810121909955106,
      "tokens_seen": 2387148800
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1311026,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2848219871520996,
      "objective/train/theoretical_loss": 3.3809972461423983,
      "objective/train/tokens_used": 757738976,
      "theoretical_loss": 3.3809972461423983,
      "tokens_seen": 2387279872
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5873856523832456e-05,
      "loss": 2.6379,
      "theoretical_loss": 3.3809972461423983,
      "tokens_seen": 2387279872
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.586583212967421e-05,
      "loss": 2.5815,
      "theoretical_loss": 3.3809823023395364,
      "tokens_seen": 2387410944
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.585780773551597e-05,
      "loss": 2.5957,
      "theoretical_loss": 3.3809673595867933,
      "tokens_seen": 2387542016
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.584978334135773e-05,
      "loss": 2.4613,
      "theoretical_loss": 3.3809524178840387,
      "tokens_seen": 2387673088
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.584175894719948e-05,
      "loss": 2.6596,
      "theoretical_loss": 3.38093747723114,
      "tokens_seen": 2387804160
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.583373455304125e-05,
      "loss": 2.6319,
      "theoretical_loss": 3.380922537627966,
      "tokens_seen": 2387935232
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5825710158883e-05,
      "loss": 2.4635,
      "theoretical_loss": 3.3809075990743858,
      "tokens_seen": 2388066304
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.581768576472477e-05,
      "loss": 2.5646,
      "theoretical_loss": 3.3808926615702677,
      "tokens_seen": 2388197376
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.580966137056652e-05,
      "loss": 2.5483,
      "theoretical_loss": 3.380877725115481,
      "tokens_seen": 2388328448
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.580163697640829e-05,
      "loss": 2.5691,
      "theoretical_loss": 3.380862789709893,
      "tokens_seen": 2388459520
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.579361258225004e-05,
      "loss": 2.6872,
      "theoretical_loss": 3.380847855353374,
      "tokens_seen": 2388590592
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5785588188091795e-05,
      "loss": 2.6173,
      "theoretical_loss": 3.380832922045792,
      "tokens_seen": 2388721664
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.577756379393356e-05,
      "loss": 2.5909,
      "theoretical_loss": 3.380817989787016,
      "tokens_seen": 2388852736
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1311624,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7213802337646484,
      "objective/train/theoretical_loss": 3.3808105240508888,
      "objective/train/tokens_used": 759377376,
      "theoretical_loss": 3.3808105240508888,
      "tokens_seen": 2388918272
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5769539399775314e-05,
      "loss": 2.4739,
      "theoretical_loss": 3.380803058576914,
      "tokens_seen": 2388983808
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.576151500561708e-05,
      "loss": 2.6132,
      "theoretical_loss": 3.3807881284153565,
      "tokens_seen": 2389114880
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5753490611458834e-05,
      "loss": 2.701,
      "theoretical_loss": 3.3807731993022108,
      "tokens_seen": 2389245952
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.57454662173006e-05,
      "loss": 2.6447,
      "theoretical_loss": 3.380758271237346,
      "tokens_seen": 2389377024
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5737441823142354e-05,
      "loss": 2.6681,
      "theoretical_loss": 3.3807433442206314,
      "tokens_seen": 2389508096
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.572941742898411e-05,
      "loss": 2.7169,
      "theoretical_loss": 3.380728418251936,
      "tokens_seen": 2389639168
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5721393034825874e-05,
      "loss": 2.6528,
      "theoretical_loss": 3.380713493331129,
      "tokens_seen": 2389770240
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.571336864066763e-05,
      "loss": 2.7172,
      "theoretical_loss": 3.3806985694580782,
      "tokens_seen": 2389901312
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5705344246509394e-05,
      "loss": 2.6488,
      "theoretical_loss": 3.3806836466326535,
      "tokens_seen": 2390032384
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.569731985235115e-05,
      "loss": 2.6782,
      "theoretical_loss": 3.3806687248547242,
      "tokens_seen": 2390163456
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5689295458192913e-05,
      "loss": 2.5,
      "theoretical_loss": 3.380653804124158,
      "tokens_seen": 2390294528
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5681271064034667e-05,
      "loss": 2.6104,
      "theoretical_loss": 3.3806388844408257,
      "tokens_seen": 2390425600
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1312746,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.828706979751587,
      "objective/train/theoretical_loss": 3.380623965804595,
      "objective/train/tokens_used": 761015776,
      "theoretical_loss": 3.380623965804595,
      "tokens_seen": 2390556672
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.567324666987642e-05,
      "loss": 2.5251,
      "theoretical_loss": 3.380623965804595,
      "tokens_seen": 2390556672
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5665222275718186e-05,
      "loss": 2.5664,
      "theoretical_loss": 3.3806090482153355,
      "tokens_seen": 2390687744
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.565719788155994e-05,
      "loss": 2.5467,
      "theoretical_loss": 3.3805941316729164,
      "tokens_seen": 2390818816
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5649173487401706e-05,
      "loss": 2.5371,
      "theoretical_loss": 3.380579216177207,
      "tokens_seen": 2390949888
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.564114909324346e-05,
      "loss": 2.6228,
      "theoretical_loss": 3.380564301728076,
      "tokens_seen": 2391080960
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5633124699085226e-05,
      "loss": 2.6312,
      "theoretical_loss": 3.380549388325393,
      "tokens_seen": 2391212032
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.562510030492698e-05,
      "loss": 2.5165,
      "theoretical_loss": 3.3805344759690272,
      "tokens_seen": 2391343104
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.561707591076873e-05,
      "loss": 2.5614,
      "theoretical_loss": 3.3805195646588473,
      "tokens_seen": 2391474176
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.56090515166105e-05,
      "loss": 2.6058,
      "theoretical_loss": 3.3805046543947235,
      "tokens_seen": 2391605248
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.560102712245225e-05,
      "loss": 2.5771,
      "theoretical_loss": 3.3804897451765243,
      "tokens_seen": 2391736320
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.559300272829402e-05,
      "loss": 2.573,
      "theoretical_loss": 3.3804748370041193,
      "tokens_seen": 2391867392
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.558497833413577e-05,
      "loss": 2.6682,
      "theoretical_loss": 3.380459929877378,
      "tokens_seen": 2391998464
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.557695393997754e-05,
      "loss": 2.5988,
      "theoretical_loss": 3.3804450237961694,
      "tokens_seen": 2392129536
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1313476,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7581470012664795,
      "objective/train/theoretical_loss": 3.3804375711475996,
      "objective/train/tokens_used": 762654176,
      "theoretical_loss": 3.3804375711475996,
      "tokens_seen": 2392195072
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.556892954581929e-05,
      "loss": 2.7053,
      "theoretical_loss": 3.380430118760364,
      "tokens_seen": 2392260608
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5560905151661045e-05,
      "loss": 2.6768,
      "theoretical_loss": 3.380415214769829,
      "tokens_seen": 2392391680
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.555288075750281e-05,
      "loss": 2.6145,
      "theoretical_loss": 3.380400311824436,
      "tokens_seen": 2392522752
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5544856363344565e-05,
      "loss": 2.4482,
      "theoretical_loss": 3.3803854099240533,
      "tokens_seen": 2392653824
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.553683196918633e-05,
      "loss": 2.6114,
      "theoretical_loss": 3.380370509068551,
      "tokens_seen": 2392784896
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5528807575028084e-05,
      "loss": 2.494,
      "theoretical_loss": 3.3803556092577978,
      "tokens_seen": 2392915968
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.552078318086985e-05,
      "loss": 2.5101,
      "theoretical_loss": 3.380340710491664,
      "tokens_seen": 2393047040
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5512758786711604e-05,
      "loss": 2.5249,
      "theoretical_loss": 3.3803258127700193,
      "tokens_seen": 2393178112
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.550473439255337e-05,
      "loss": 2.7193,
      "theoretical_loss": 3.380310916092732,
      "tokens_seen": 2393309184
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5496709998395124e-05,
      "loss": 2.6344,
      "theoretical_loss": 3.3802960204596735,
      "tokens_seen": 2393440256
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.548868560423688e-05,
      "loss": 2.6615,
      "theoretical_loss": 3.380281125870712,
      "tokens_seen": 2393571328
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5480661210078644e-05,
      "loss": 2.5943,
      "theoretical_loss": 3.3802662323257175,
      "tokens_seen": 2393702400
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1314763,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.905195951461792,
      "objective/train/theoretical_loss": 3.38025133982456,
      "objective/train/tokens_used": 764292576,
      "theoretical_loss": 3.38025133982456,
      "tokens_seen": 2393833472
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.54726368159204e-05,
      "loss": 2.6042,
      "theoretical_loss": 3.38025133982456,
      "tokens_seen": 2393833472
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5464612421762164e-05,
      "loss": 2.5882,
      "theoretical_loss": 3.380236448367109,
      "tokens_seen": 2393964544
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.545658802760392e-05,
      "loss": 2.44,
      "theoretical_loss": 3.380221557953234,
      "tokens_seen": 2394095616
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5448563633445684e-05,
      "loss": 2.714,
      "theoretical_loss": 3.3802066685828054,
      "tokens_seen": 2394226688
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.544053923928744e-05,
      "loss": 2.4515,
      "theoretical_loss": 3.380191780255692,
      "tokens_seen": 2394357760
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.543251484512919e-05,
      "loss": 2.6101,
      "theoretical_loss": 3.380176892971764,
      "tokens_seen": 2394488832
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5424490450970956e-05,
      "loss": 2.5848,
      "theoretical_loss": 3.3801620067308917,
      "tokens_seen": 2394619904
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.541646605681271e-05,
      "loss": 2.5574,
      "theoretical_loss": 3.3801471215329446,
      "tokens_seen": 2394750976
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5408441662654476e-05,
      "loss": 2.5288,
      "theoretical_loss": 3.3801322373777927,
      "tokens_seen": 2394882048
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.540041726849623e-05,
      "loss": 2.5644,
      "theoretical_loss": 3.3801173542653054,
      "tokens_seen": 2395013120
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5392392874337996e-05,
      "loss": 2.5046,
      "theoretical_loss": 3.3801024721953525,
      "tokens_seen": 2395144192
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.538436848017975e-05,
      "loss": 2.4909,
      "theoretical_loss": 3.3800875911678046,
      "tokens_seen": 2395275264
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.53763440860215e-05,
      "loss": 2.4923,
      "theoretical_loss": 3.3800727111825317,
      "tokens_seen": 2395406336
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1315396,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6342225074768066,
      "objective/train/theoretical_loss": 3.3800652715807074,
      "objective/train/tokens_used": 765930976,
      "theoretical_loss": 3.3800652715807074,
      "tokens_seen": 2395471872
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.536831969186327e-05,
      "loss": 2.4363,
      "theoretical_loss": 3.3800578322394026,
      "tokens_seen": 2395537408
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.536029529770502e-05,
      "loss": 2.5553,
      "theoretical_loss": 3.380042954338289,
      "tokens_seen": 2395668480
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.535227090354679e-05,
      "loss": 2.6008,
      "theoretical_loss": 3.3800280774790594,
      "tokens_seen": 2395799552
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.534424650938854e-05,
      "loss": 2.6577,
      "theoretical_loss": 3.380013201661585,
      "tokens_seen": 2395930624
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.533622211523031e-05,
      "loss": 2.4719,
      "theoretical_loss": 3.379998326885735,
      "tokens_seen": 2396061696
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.532819772107206e-05,
      "loss": 2.4111,
      "theoretical_loss": 3.3799834531513797,
      "tokens_seen": 2396192768
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5320173326913815e-05,
      "loss": 2.5345,
      "theoretical_loss": 3.37996858045839,
      "tokens_seen": 2396323840
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.531214893275558e-05,
      "loss": 2.4507,
      "theoretical_loss": 3.379953708806635,
      "tokens_seen": 2396454912
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5304124538597335e-05,
      "loss": 2.5533,
      "theoretical_loss": 3.379938838195985,
      "tokens_seen": 2396585984
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.52961001444391e-05,
      "loss": 2.6565,
      "theoretical_loss": 3.379923968626311,
      "tokens_seen": 2396717056
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5288075750280855e-05,
      "loss": 2.5447,
      "theoretical_loss": 3.3799091000974824,
      "tokens_seen": 2396848128
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.528005135612262e-05,
      "loss": 2.5323,
      "theoretical_loss": 3.3798942326093697,
      "tokens_seen": 2396979200
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1316445,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.639875650405884,
      "objective/train/theoretical_loss": 3.379879366161843,
      "objective/train/tokens_used": 767569376,
      "theoretical_loss": 3.379879366161843,
      "tokens_seen": 2397110272
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5272026961964374e-05,
      "loss": 2.5263,
      "theoretical_loss": 3.379879366161843,
      "tokens_seen": 2397110272
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.526400256780613e-05,
      "loss": 2.569,
      "theoretical_loss": 3.3798645007547727,
      "tokens_seen": 2397241344
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5255978173647894e-05,
      "loss": 2.5214,
      "theoretical_loss": 3.3798496363880295,
      "tokens_seen": 2397372416
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.524795377948965e-05,
      "loss": 2.3692,
      "theoretical_loss": 3.3798347730614826,
      "tokens_seen": 2397503488
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5239929385331414e-05,
      "loss": 2.5421,
      "theoretical_loss": 3.3798199107750038,
      "tokens_seen": 2397634560
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.523190499117317e-05,
      "loss": 2.5866,
      "theoretical_loss": 3.3798050495284624,
      "tokens_seen": 2397765632
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5223880597014934e-05,
      "loss": 2.4599,
      "theoretical_loss": 3.379790189321729,
      "tokens_seen": 2397896704
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.521585620285669e-05,
      "loss": 2.4592,
      "theoretical_loss": 3.3797753301546742,
      "tokens_seen": 2398027776
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5207831808698454e-05,
      "loss": 2.4998,
      "theoretical_loss": 3.3797604720271686,
      "tokens_seen": 2398158848
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.519980741454021e-05,
      "loss": 2.4942,
      "theoretical_loss": 3.3797456149390825,
      "tokens_seen": 2398289920
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.519178302038196e-05,
      "loss": 2.6444,
      "theoretical_loss": 3.379730758890286,
      "tokens_seen": 2398420992
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5183758626223727e-05,
      "loss": 2.4578,
      "theoretical_loss": 3.3797159038806504,
      "tokens_seen": 2398552064
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.517573423206548e-05,
      "loss": 2.6303,
      "theoretical_loss": 3.3797010499100453,
      "tokens_seen": 2398683136
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1317092,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3850042819976807,
      "objective/train/theoretical_loss": 3.379693623314339,
      "objective/train/tokens_used": 769207776,
      "theoretical_loss": 3.379693623314339,
      "tokens_seen": 2398748672
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5167709837907246e-05,
      "loss": 2.4892,
      "theoretical_loss": 3.379686196978342,
      "tokens_seen": 2398814208
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5159685443749e-05,
      "loss": 2.6844,
      "theoretical_loss": 3.3796713450854106,
      "tokens_seen": 2398945280
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5151661049590766e-05,
      "loss": 2.5773,
      "theoretical_loss": 3.379656494231122,
      "tokens_seen": 2399076352
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.514363665543252e-05,
      "loss": 2.5601,
      "theoretical_loss": 3.379641644415347,
      "tokens_seen": 2399207424
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.513561226127427e-05,
      "loss": 2.4204,
      "theoretical_loss": 3.3796267956379555,
      "tokens_seen": 2399338496
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.512758786711604e-05,
      "loss": 2.4937,
      "theoretical_loss": 3.379611947898819,
      "tokens_seen": 2399469568
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.511956347295779e-05,
      "loss": 2.5614,
      "theoretical_loss": 3.379597101197808,
      "tokens_seen": 2399600640
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.511153907879956e-05,
      "loss": 2.4579,
      "theoretical_loss": 3.3795822555347925,
      "tokens_seen": 2399731712
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.510351468464131e-05,
      "loss": 2.4595,
      "theoretical_loss": 3.3795674109096443,
      "tokens_seen": 2399862784
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.509549029048308e-05,
      "loss": 2.5017,
      "theoretical_loss": 3.3795525673222335,
      "tokens_seen": 2399993856
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.508746589632483e-05,
      "loss": 2.5535,
      "theoretical_loss": 3.3795377247724314,
      "tokens_seen": 2400124928
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5079441502166585e-05,
      "loss": 2.5604,
      "theoretical_loss": 3.379522883260108,
      "tokens_seen": 2400256000
    },
    {
      "epoch": 0.45,
      "objective/train/docs_used": 1318422,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.598881959915161,
      "objective/train/theoretical_loss": 3.379508042785135,
      "objective/train/tokens_used": 770846176,
      "theoretical_loss": 3.379508042785135,
      "tokens_seen": 2400387072
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.507141710800835e-05,
      "loss": 2.5778,
      "theoretical_loss": 3.379508042785135,
      "tokens_seen": 2400387072
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.5063392713850105e-05,
      "loss": 2.4412,
      "theoretical_loss": 3.3794932033473826,
      "tokens_seen": 2400518144
    },
    {
      "epoch": 0.45,
      "learning_rate": 5.505536831969187e-05,
      "loss": 2.5274,
      "theoretical_loss": 3.3794783649467215,
      "tokens_seen": 2400649216
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.5047343925533625e-05,
      "loss": 2.5056,
      "theoretical_loss": 3.379463527583024,
      "tokens_seen": 2400780288
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.503931953137539e-05,
      "loss": 2.5522,
      "theoretical_loss": 3.379448691256159,
      "tokens_seen": 2400911360
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.5031295137217144e-05,
      "loss": 2.5819,
      "theoretical_loss": 3.379433855965999,
      "tokens_seen": 2401042432
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.50232707430589e-05,
      "loss": 2.529,
      "theoretical_loss": 3.379419021712415,
      "tokens_seen": 2401173504
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.5015246348900664e-05,
      "loss": 2.3629,
      "theoretical_loss": 3.379404188495277,
      "tokens_seen": 2401304576
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.500722195474242e-05,
      "loss": 2.5947,
      "theoretical_loss": 3.3793893563144564,
      "tokens_seen": 2401435648
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4999197560584184e-05,
      "loss": 2.5389,
      "theoretical_loss": 3.3793745251698244,
      "tokens_seen": 2401566720
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.499117316642594e-05,
      "loss": 2.4569,
      "theoretical_loss": 3.379359695061252,
      "tokens_seen": 2401697792
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4983148772267704e-05,
      "loss": 2.5019,
      "theoretical_loss": 3.3793448659886103,
      "tokens_seen": 2401828864
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.497512437810946e-05,
      "loss": 2.567,
      "theoretical_loss": 3.3793300379517706,
      "tokens_seen": 2401959936
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1319687,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7546088695526123,
      "objective/train/theoretical_loss": 3.379322624321736,
      "objective/train/tokens_used": 772484576,
      "theoretical_loss": 3.379322624321736,
      "tokens_seen": 2402025472
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.496709998395121e-05,
      "loss": 2.4976,
      "theoretical_loss": 3.3793152109506037,
      "tokens_seen": 2402091008
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.495907558979298e-05,
      "loss": 2.5438,
      "theoretical_loss": 3.3793003849849805,
      "tokens_seen": 2402222080
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.495105119563473e-05,
      "loss": 2.4978,
      "theoretical_loss": 3.379285560054773,
      "tokens_seen": 2402353152
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.49430268014765e-05,
      "loss": 2.5106,
      "theoretical_loss": 3.379270736159852,
      "tokens_seen": 2402484224
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.493500240731825e-05,
      "loss": 2.5004,
      "theoretical_loss": 3.3792559133000886,
      "tokens_seen": 2402615296
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4926978013160016e-05,
      "loss": 2.6882,
      "theoretical_loss": 3.3792410914753543,
      "tokens_seen": 2402746368
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.491895361900177e-05,
      "loss": 2.4467,
      "theoretical_loss": 3.37922627068552,
      "tokens_seen": 2402877440
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.491092922484352e-05,
      "loss": 2.5897,
      "theoretical_loss": 3.379211450930457,
      "tokens_seen": 2403008512
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.490290483068529e-05,
      "loss": 2.5128,
      "theoretical_loss": 3.3791966322100375,
      "tokens_seen": 2403139584
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.489488043652704e-05,
      "loss": 2.5733,
      "theoretical_loss": 3.3791818145241317,
      "tokens_seen": 2403270656
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.488685604236881e-05,
      "loss": 2.7445,
      "theoretical_loss": 3.3791669978726113,
      "tokens_seen": 2403401728
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.487883164821056e-05,
      "loss": 2.4769,
      "theoretical_loss": 3.379152182255348,
      "tokens_seen": 2403532800
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1320162,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.617546796798706,
      "objective/train/theoretical_loss": 3.379137367672213,
      "objective/train/tokens_used": 774122976,
      "theoretical_loss": 3.379137367672213,
      "tokens_seen": 2403663872
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.487080725405232e-05,
      "loss": 2.3921,
      "theoretical_loss": 3.379137367672213,
      "tokens_seen": 2403663872
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.486278285989408e-05,
      "loss": 2.5964,
      "theoretical_loss": 3.3791225541230783,
      "tokens_seen": 2403794944
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.485475846573584e-05,
      "loss": 2.4279,
      "theoretical_loss": 3.379107741607814,
      "tokens_seen": 2403926016
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.48467340715776e-05,
      "loss": 2.5559,
      "theoretical_loss": 3.379092930126293,
      "tokens_seen": 2404057088
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4838709677419355e-05,
      "loss": 2.4422,
      "theoretical_loss": 3.379078119678386,
      "tokens_seen": 2404188160
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.483068528326112e-05,
      "loss": 2.4967,
      "theoretical_loss": 3.3790633102639647,
      "tokens_seen": 2404319232
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4822660889102875e-05,
      "loss": 2.5944,
      "theoretical_loss": 3.3790485018829006,
      "tokens_seen": 2404450304
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4814636494944635e-05,
      "loss": 2.6128,
      "theoretical_loss": 3.379033694535065,
      "tokens_seen": 2404581376
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4806612100786395e-05,
      "loss": 2.5823,
      "theoretical_loss": 3.3790188882203305,
      "tokens_seen": 2404712448
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4798587706628155e-05,
      "loss": 2.4524,
      "theoretical_loss": 3.3790040829385677,
      "tokens_seen": 2404843520
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4790563312469914e-05,
      "loss": 2.5909,
      "theoretical_loss": 3.378989278689649,
      "tokens_seen": 2404974592
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.478253891831167e-05,
      "loss": 2.513,
      "theoretical_loss": 3.378974475473445,
      "tokens_seen": 2405105664
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.477451452415343e-05,
      "loss": 2.5035,
      "theoretical_loss": 3.3789596732898284,
      "tokens_seen": 2405236736
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1321303,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4370169639587402,
      "objective/train/theoretical_loss": 3.3789522725852,
      "objective/train/tokens_used": 775761376,
      "theoretical_loss": 3.3789522725852,
      "tokens_seen": 2405302272
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.476649012999519e-05,
      "loss": 2.4061,
      "theoretical_loss": 3.3789448721386703,
      "tokens_seen": 2405367808
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.475846573583695e-05,
      "loss": 2.5969,
      "theoretical_loss": 3.378930072019843,
      "tokens_seen": 2405498880
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.475044134167871e-05,
      "loss": 2.5381,
      "theoretical_loss": 3.378915272933218,
      "tokens_seen": 2405629952
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.474241694752047e-05,
      "loss": 2.405,
      "theoretical_loss": 3.378900474878667,
      "tokens_seen": 2405761024
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.473439255336223e-05,
      "loss": 2.4685,
      "theoretical_loss": 3.3788856778560614,
      "tokens_seen": 2405892096
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.472636815920398e-05,
      "loss": 2.5649,
      "theoretical_loss": 3.3788708818652737,
      "tokens_seen": 2406023168
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.471834376504574e-05,
      "loss": 2.5203,
      "theoretical_loss": 3.3788560869061754,
      "tokens_seen": 2406154240
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.47103193708875e-05,
      "loss": 2.6004,
      "theoretical_loss": 3.3788412929786382,
      "tokens_seen": 2406285312
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.470229497672926e-05,
      "loss": 2.5926,
      "theoretical_loss": 3.3788265000825346,
      "tokens_seen": 2406416384
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.469427058257102e-05,
      "loss": 2.6392,
      "theoretical_loss": 3.3788117082177362,
      "tokens_seen": 2406547456
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.468624618841278e-05,
      "loss": 2.5661,
      "theoretical_loss": 3.378796917384115,
      "tokens_seen": 2406678528
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.467822179425453e-05,
      "loss": 2.5163,
      "theoretical_loss": 3.3787821275815424,
      "tokens_seen": 2406809600
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1321925,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.268547296524048,
      "objective/train/theoretical_loss": 3.378767338809891,
      "objective/train/tokens_used": 777399776,
      "theoretical_loss": 3.378767338809891,
      "tokens_seen": 2406940672
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.467019740009629e-05,
      "loss": 2.435,
      "theoretical_loss": 3.378767338809891,
      "tokens_seen": 2406940672
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.466217300593805e-05,
      "loss": 2.5572,
      "theoretical_loss": 3.378752551069033,
      "tokens_seen": 2407071744
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.465414861177981e-05,
      "loss": 2.6102,
      "theoretical_loss": 3.37873776435884,
      "tokens_seen": 2407202816
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.464612421762157e-05,
      "loss": 2.704,
      "theoretical_loss": 3.3787229786791837,
      "tokens_seen": 2407333888
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.463809982346333e-05,
      "loss": 2.401,
      "theoretical_loss": 3.3787081940299366,
      "tokens_seen": 2407464960
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.463007542930509e-05,
      "loss": 2.4335,
      "theoretical_loss": 3.378693410410971,
      "tokens_seen": 2407596032
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4622051035146845e-05,
      "loss": 2.628,
      "theoretical_loss": 3.378678627822159,
      "tokens_seen": 2407727104
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4614026640988605e-05,
      "loss": 2.5319,
      "theoretical_loss": 3.3786638462633722,
      "tokens_seen": 2407858176
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4606002246830365e-05,
      "loss": 2.4947,
      "theoretical_loss": 3.3786490657344834,
      "tokens_seen": 2407989248
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4597977852672125e-05,
      "loss": 2.468,
      "theoretical_loss": 3.378634286235364,
      "tokens_seen": 2408120320
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4589953458513885e-05,
      "loss": 2.516,
      "theoretical_loss": 3.3786195077658876,
      "tokens_seen": 2408251392
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.458192906435564e-05,
      "loss": 2.4975,
      "theoretical_loss": 3.3786047303259252,
      "tokens_seen": 2408382464
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4573904670197405e-05,
      "loss": 2.597,
      "theoretical_loss": 3.378589953915349,
      "tokens_seen": 2408513536
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1323064,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5613491535186768,
      "objective/train/theoretical_loss": 3.3785825660960414,
      "objective/train/tokens_used": 779038176,
      "theoretical_loss": 3.3785825660960414,
      "tokens_seen": 2408579072
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.456588027603916e-05,
      "loss": 2.6614,
      "theoretical_loss": 3.378575178534032,
      "tokens_seen": 2408644608
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4557855881880925e-05,
      "loss": 2.5185,
      "theoretical_loss": 3.3785604041818464,
      "tokens_seen": 2408775680
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.454983148772268e-05,
      "loss": 2.5571,
      "theoretical_loss": 3.3785456308586643,
      "tokens_seen": 2408906752
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.454180709356443e-05,
      "loss": 2.5294,
      "theoretical_loss": 3.378530858564358,
      "tokens_seen": 2409037824
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.45337826994062e-05,
      "loss": 2.5203,
      "theoretical_loss": 3.3785160872987996,
      "tokens_seen": 2409168896
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.452575830524795e-05,
      "loss": 2.6502,
      "theoretical_loss": 3.378501317061862,
      "tokens_seen": 2409299968
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.451773391108972e-05,
      "loss": 2.5972,
      "theoretical_loss": 3.3784865478534174,
      "tokens_seen": 2409431040
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.450970951693147e-05,
      "loss": 2.5656,
      "theoretical_loss": 3.3784717796733386,
      "tokens_seen": 2409562112
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.450168512277324e-05,
      "loss": 2.5192,
      "theoretical_loss": 3.3784570125214977,
      "tokens_seen": 2409693184
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.449366072861499e-05,
      "loss": 2.5197,
      "theoretical_loss": 3.378442246397767,
      "tokens_seen": 2409824256
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4485636334456743e-05,
      "loss": 2.428,
      "theoretical_loss": 3.3784274813020194,
      "tokens_seen": 2409955328
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.447761194029851e-05,
      "loss": 2.5974,
      "theoretical_loss": 3.378412717234127,
      "tokens_seen": 2410086400
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1323777,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4356136322021484,
      "objective/train/theoretical_loss": 3.378397954193963,
      "objective/train/tokens_used": 780676576,
      "theoretical_loss": 3.378397954193963,
      "tokens_seen": 2410217472
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.446958754614026e-05,
      "loss": 2.5048,
      "theoretical_loss": 3.378397954193963,
      "tokens_seen": 2410217472
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.446156315198203e-05,
      "loss": 2.5731,
      "theoretical_loss": 3.3783831921813996,
      "tokens_seen": 2410348544
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.445353875782378e-05,
      "loss": 2.3714,
      "theoretical_loss": 3.378368431196309,
      "tokens_seen": 2410479616
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.444551436366555e-05,
      "loss": 2.4744,
      "theoretical_loss": 3.3783536712385644,
      "tokens_seen": 2410610688
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.44374899695073e-05,
      "loss": 2.7018,
      "theoretical_loss": 3.3783389123080387,
      "tokens_seen": 2410741760
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4429465575349056e-05,
      "loss": 2.5208,
      "theoretical_loss": 3.378324154404604,
      "tokens_seen": 2410872832
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.442144118119082e-05,
      "loss": 2.5483,
      "theoretical_loss": 3.378309397528133,
      "tokens_seen": 2411003904
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4413416787032576e-05,
      "loss": 2.4754,
      "theoretical_loss": 3.3782946416784982,
      "tokens_seen": 2411134976
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.440539239287434e-05,
      "loss": 2.5531,
      "theoretical_loss": 3.378279886855573,
      "tokens_seen": 2411266048
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4397367998716096e-05,
      "loss": 2.5559,
      "theoretical_loss": 3.3782651330592297,
      "tokens_seen": 2411397120
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.438934360455786e-05,
      "loss": 2.3269,
      "theoretical_loss": 3.3782503802893413,
      "tokens_seen": 2411528192
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4381319210399615e-05,
      "loss": 2.5977,
      "theoretical_loss": 3.3782356285457804,
      "tokens_seen": 2411659264
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.437329481624137e-05,
      "loss": 2.4914,
      "theoretical_loss": 3.37822087782842,
      "tokens_seen": 2411790336
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1325032,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5119264125823975,
      "objective/train/theoretical_loss": 3.3782135028545253,
      "objective/train/tokens_used": 782314976,
      "theoretical_loss": 3.3782135028545253,
      "tokens_seen": 2411855872
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4365270422083135e-05,
      "loss": 2.5239,
      "theoretical_loss": 3.378206128137133,
      "tokens_seen": 2411921408
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.435724602792489e-05,
      "loss": 2.569,
      "theoretical_loss": 3.378191379471792,
      "tokens_seen": 2412052480
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4349221633766655e-05,
      "loss": 2.5473,
      "theoretical_loss": 3.3781766318322703,
      "tokens_seen": 2412183552
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.434119723960841e-05,
      "loss": 2.6283,
      "theoretical_loss": 3.3781618852184403,
      "tokens_seen": 2412314624
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4333172845450175e-05,
      "loss": 2.442,
      "theoretical_loss": 3.378147139630175,
      "tokens_seen": 2412445696
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.432514845129193e-05,
      "loss": 2.5497,
      "theoretical_loss": 3.3781323950673485,
      "tokens_seen": 2412576768
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.431712405713368e-05,
      "loss": 2.587,
      "theoretical_loss": 3.378117651529832,
      "tokens_seen": 2412707840
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.430909966297545e-05,
      "loss": 2.7423,
      "theoretical_loss": 3.3781029090174997,
      "tokens_seen": 2412838912
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.43010752688172e-05,
      "loss": 2.6118,
      "theoretical_loss": 3.3780881675302243,
      "tokens_seen": 2412969984
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.429305087465897e-05,
      "loss": 2.324,
      "theoretical_loss": 3.3780734270678785,
      "tokens_seen": 2413101056
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.428502648050072e-05,
      "loss": 2.5258,
      "theoretical_loss": 3.378058687630336,
      "tokens_seen": 2413232128
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.427700208634249e-05,
      "loss": 2.4511,
      "theoretical_loss": 3.3780439492174694,
      "tokens_seen": 2413363200
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1325688,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5455198287963867,
      "objective/train/theoretical_loss": 3.378029211829152,
      "objective/train/tokens_used": 783953376,
      "theoretical_loss": 3.378029211829152,
      "tokens_seen": 2413494272
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.426897769218424e-05,
      "loss": 2.595,
      "theoretical_loss": 3.378029211829152,
      "tokens_seen": 2413494272
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.426095329802601e-05,
      "loss": 2.5726,
      "theoretical_loss": 3.3780144754652577,
      "tokens_seen": 2413625344
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.425292890386776e-05,
      "loss": 2.4572,
      "theoretical_loss": 3.3779997401256585,
      "tokens_seen": 2413756416
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4244904509709514e-05,
      "loss": 2.5213,
      "theoretical_loss": 3.3779850058102276,
      "tokens_seen": 2413887488
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.423688011555128e-05,
      "loss": 2.6122,
      "theoretical_loss": 3.377970272518839,
      "tokens_seen": 2414018560
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.422885572139303e-05,
      "loss": 2.4838,
      "theoretical_loss": 3.3779555402513655,
      "tokens_seen": 2414149632
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.42208313272348e-05,
      "loss": 2.4408,
      "theoretical_loss": 3.3779408090076806,
      "tokens_seen": 2414280704
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.421280693307655e-05,
      "loss": 2.5687,
      "theoretical_loss": 3.3779260787876573,
      "tokens_seen": 2414411776
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.420478253891832e-05,
      "loss": 2.5043,
      "theoretical_loss": 3.377911349591169,
      "tokens_seen": 2414542848
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.419675814476007e-05,
      "loss": 2.4312,
      "theoretical_loss": 3.377896621418089,
      "tokens_seen": 2414673920
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4188733750601826e-05,
      "loss": 2.487,
      "theoretical_loss": 3.377881894268291,
      "tokens_seen": 2414804992
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.418070935644359e-05,
      "loss": 2.4378,
      "theoretical_loss": 3.3778671681416474,
      "tokens_seen": 2414936064
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4172684962285346e-05,
      "loss": 2.5642,
      "theoretical_loss": 3.377852443038033,
      "tokens_seen": 2415067136
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1326614,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4216179847717285,
      "objective/train/theoretical_loss": 3.3778450808698213,
      "objective/train/tokens_used": 785591776,
      "theoretical_loss": 3.3778450808698213,
      "tokens_seen": 2415132672
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.416466056812711e-05,
      "loss": 2.5758,
      "theoretical_loss": 3.37783771895732,
      "tokens_seen": 2415198208
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4156636173968866e-05,
      "loss": 2.5595,
      "theoretical_loss": 3.377822995899382,
      "tokens_seen": 2415329280
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.414861177981063e-05,
      "loss": 2.4036,
      "theoretical_loss": 3.377808273864093,
      "tokens_seen": 2415460352
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4140587385652386e-05,
      "loss": 2.6588,
      "theoretical_loss": 3.377793552851326,
      "tokens_seen": 2415591424
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.413256299149414e-05,
      "loss": 2.5935,
      "theoretical_loss": 3.3777788328609555,
      "tokens_seen": 2415722496
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4124538597335905e-05,
      "loss": 2.5657,
      "theoretical_loss": 3.3777641138928534,
      "tokens_seen": 2415853568
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.411651420317766e-05,
      "loss": 2.4348,
      "theoretical_loss": 3.3777493959468945,
      "tokens_seen": 2415984640
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4108489809019425e-05,
      "loss": 2.6172,
      "theoretical_loss": 3.377734679022952,
      "tokens_seen": 2416115712
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.410046541486118e-05,
      "loss": 2.4744,
      "theoretical_loss": 3.3777199631208994,
      "tokens_seen": 2416246784
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.4092441020702945e-05,
      "loss": 2.4383,
      "theoretical_loss": 3.3777052482406105,
      "tokens_seen": 2416377856
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.40844166265447e-05,
      "loss": 2.4488,
      "theoretical_loss": 3.3776905343819585,
      "tokens_seen": 2416508928
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.407639223238645e-05,
      "loss": 2.5904,
      "theoretical_loss": 3.377675821544818,
      "tokens_seen": 2416640000
    },
    {
      "epoch": 0.46,
      "objective/train/docs_used": 1327216,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4877939224243164,
      "objective/train/theoretical_loss": 3.377661109729061,
      "objective/train/tokens_used": 787230176,
      "theoretical_loss": 3.377661109729061,
      "tokens_seen": 2416771072
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.406836783822822e-05,
      "loss": 2.3429,
      "theoretical_loss": 3.377661109729061,
      "tokens_seen": 2416771072
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.406034344406997e-05,
      "loss": 2.419,
      "theoretical_loss": 3.377646398934563,
      "tokens_seen": 2416902144
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.405231904991174e-05,
      "loss": 2.6701,
      "theoretical_loss": 3.377631689161197,
      "tokens_seen": 2417033216
    },
    {
      "epoch": 0.46,
      "learning_rate": 5.404429465575349e-05,
      "loss": 2.5824,
      "theoretical_loss": 3.3776169804088365,
      "tokens_seen": 2417164288
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.403627026159526e-05,
      "loss": 2.6254,
      "theoretical_loss": 3.377602272677356,
      "tokens_seen": 2417295360
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.402824586743701e-05,
      "loss": 2.4818,
      "theoretical_loss": 3.3775875659666283,
      "tokens_seen": 2417426432
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.4020221473278764e-05,
      "loss": 2.6733,
      "theoretical_loss": 3.377572860276528,
      "tokens_seen": 2417557504
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.401219707912053e-05,
      "loss": 2.6771,
      "theoretical_loss": 3.3775581556069283,
      "tokens_seen": 2417688576
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.4004172684962284e-05,
      "loss": 2.6009,
      "theoretical_loss": 3.377543451957704,
      "tokens_seen": 2417819648
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.399614829080405e-05,
      "loss": 2.5779,
      "theoretical_loss": 3.377528749328728,
      "tokens_seen": 2417950720
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3988123896645803e-05,
      "loss": 2.5507,
      "theoretical_loss": 3.377514047719875,
      "tokens_seen": 2418081792
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.398009950248757e-05,
      "loss": 2.5918,
      "theoretical_loss": 3.377499347131018,
      "tokens_seen": 2418212864
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.397207510832932e-05,
      "loss": 2.5341,
      "theoretical_loss": 3.377484647562032,
      "tokens_seen": 2418343936
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1328499,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.503131151199341,
      "objective/train/theoretical_loss": 3.377477298159951,
      "objective/train/tokens_used": 788868576,
      "theoretical_loss": 3.377477298159951,
      "tokens_seen": 2418409472
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3964050714171076e-05,
      "loss": 2.5173,
      "theoretical_loss": 3.37746994901279,
      "tokens_seen": 2418475008
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.395602632001284e-05,
      "loss": 2.5265,
      "theoretical_loss": 3.377455251483167,
      "tokens_seen": 2418606080
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3948001925854596e-05,
      "loss": 2.6263,
      "theoretical_loss": 3.3774405549730364,
      "tokens_seen": 2418737152
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.393997753169636e-05,
      "loss": 2.3933,
      "theoretical_loss": 3.377425859482272,
      "tokens_seen": 2418868224
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3931953137538116e-05,
      "loss": 2.4687,
      "theoretical_loss": 3.3774111650107486,
      "tokens_seen": 2418999296
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.392392874337988e-05,
      "loss": 2.5433,
      "theoretical_loss": 3.37739647155834,
      "tokens_seen": 2419130368
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3915904349221636e-05,
      "loss": 2.6307,
      "theoretical_loss": 3.37738177912492,
      "tokens_seen": 2419261440
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.39078799550634e-05,
      "loss": 2.7126,
      "theoretical_loss": 3.377367087710363,
      "tokens_seen": 2419392512
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3899855560905156e-05,
      "loss": 2.6034,
      "theoretical_loss": 3.3773523973145427,
      "tokens_seen": 2419523584
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.389183116674691e-05,
      "loss": 2.5136,
      "theoretical_loss": 3.377337707937334,
      "tokens_seen": 2419654656
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3883806772588675e-05,
      "loss": 2.5819,
      "theoretical_loss": 3.377323019578611,
      "tokens_seen": 2419785728
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.387578237843043e-05,
      "loss": 2.4908,
      "theoretical_loss": 3.3773083322382473,
      "tokens_seen": 2419916800
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1329844,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.938718557357788,
      "objective/train/theoretical_loss": 3.377293645916118,
      "objective/train/tokens_used": 790506976,
      "theoretical_loss": 3.377293645916118,
      "tokens_seen": 2420047872
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3867757984272195e-05,
      "loss": 2.5834,
      "theoretical_loss": 3.377293645916118,
      "tokens_seen": 2420047872
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.385973359011395e-05,
      "loss": 2.514,
      "theoretical_loss": 3.377278960612096,
      "tokens_seen": 2420178944
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3851709195955715e-05,
      "loss": 2.5113,
      "theoretical_loss": 3.3772642763260574,
      "tokens_seen": 2420310016
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.384368480179747e-05,
      "loss": 2.5993,
      "theoretical_loss": 3.377249593057875,
      "tokens_seen": 2420441088
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.383566040763922e-05,
      "loss": 2.4977,
      "theoretical_loss": 3.377234910807424,
      "tokens_seen": 2420572160
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.382763601348099e-05,
      "loss": 2.6783,
      "theoretical_loss": 3.3772202295745783,
      "tokens_seen": 2420703232
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.381961161932274e-05,
      "loss": 2.4175,
      "theoretical_loss": 3.377205549359213,
      "tokens_seen": 2420834304
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.381158722516451e-05,
      "loss": 2.3895,
      "theoretical_loss": 3.3771908701612015,
      "tokens_seen": 2420965376
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.380356283100626e-05,
      "loss": 2.3957,
      "theoretical_loss": 3.3771761919804186,
      "tokens_seen": 2421096448
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.379553843684803e-05,
      "loss": 2.2962,
      "theoretical_loss": 3.377161514816739,
      "tokens_seen": 2421227520
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.378751404268978e-05,
      "loss": 2.6406,
      "theoretical_loss": 3.377146838670037,
      "tokens_seen": 2421358592
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3779489648531534e-05,
      "loss": 2.4753,
      "theoretical_loss": 3.377132163540187,
      "tokens_seen": 2421489664
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.37714652543733e-05,
      "loss": 2.5184,
      "theoretical_loss": 3.377117489427064,
      "tokens_seen": 2421620736
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1330351,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.552644968032837,
      "objective/train/theoretical_loss": 3.3771101527517358,
      "objective/train/tokens_used": 792145376,
      "theoretical_loss": 3.3771101527517358,
      "tokens_seen": 2421686272
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3763440860215054e-05,
      "loss": 2.6453,
      "theoretical_loss": 3.377102816330542,
      "tokens_seen": 2421751808
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.375541646605682e-05,
      "loss": 2.49,
      "theoretical_loss": 3.3770881442504956,
      "tokens_seen": 2421882880
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3747392071898574e-05,
      "loss": 2.4389,
      "theoretical_loss": 3.3770734731867993,
      "tokens_seen": 2422013952
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.373936767774034e-05,
      "loss": 2.4152,
      "theoretical_loss": 3.377058803139328,
      "tokens_seen": 2422145024
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.373134328358209e-05,
      "loss": 2.529,
      "theoretical_loss": 3.377044134107956,
      "tokens_seen": 2422276096
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3723318889423846e-05,
      "loss": 2.4047,
      "theoretical_loss": 3.377029466092558,
      "tokens_seen": 2422407168
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.371529449526561e-05,
      "loss": 2.5738,
      "theoretical_loss": 3.3770147990930095,
      "tokens_seen": 2422538240
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3707270101107366e-05,
      "loss": 2.6262,
      "theoretical_loss": 3.377000133109184,
      "tokens_seen": 2422669312
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.369924570694913e-05,
      "loss": 2.5234,
      "theoretical_loss": 3.3769854681409566,
      "tokens_seen": 2422800384
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3691221312790886e-05,
      "loss": 2.4704,
      "theoretical_loss": 3.3769708041882023,
      "tokens_seen": 2422931456
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.368319691863265e-05,
      "loss": 2.4136,
      "theoretical_loss": 3.376956141250796,
      "tokens_seen": 2423062528
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3675172524474406e-05,
      "loss": 2.3786,
      "theoretical_loss": 3.3769414793286114,
      "tokens_seen": 2423193600
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1331600,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.512211799621582,
      "objective/train/theoretical_loss": 3.3769268184215244,
      "objective/train/tokens_used": 793783776,
      "theoretical_loss": 3.3769268184215244,
      "tokens_seen": 2423324672
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.366714813031616e-05,
      "loss": 2.4773,
      "theoretical_loss": 3.3769268184215244,
      "tokens_seen": 2423324672
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3659123736157926e-05,
      "loss": 2.5779,
      "theoretical_loss": 3.37691215852941,
      "tokens_seen": 2423455744
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.365109934199968e-05,
      "loss": 2.3919,
      "theoretical_loss": 3.376897499652142,
      "tokens_seen": 2423586816
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3643074947841446e-05,
      "loss": 2.5697,
      "theoretical_loss": 3.376882841789596,
      "tokens_seen": 2423717888
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.36350505536832e-05,
      "loss": 2.4966,
      "theoretical_loss": 3.3768681849416464,
      "tokens_seen": 2423848960
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3627026159524965e-05,
      "loss": 2.5197,
      "theoretical_loss": 3.3768535291081685,
      "tokens_seen": 2423980032
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.361900176536672e-05,
      "loss": 2.5537,
      "theoretical_loss": 3.3768388742890374,
      "tokens_seen": 2424111104
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.361097737120848e-05,
      "loss": 2.4198,
      "theoretical_loss": 3.376824220484128,
      "tokens_seen": 2424242176
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.360295297705024e-05,
      "loss": 2.4866,
      "theoretical_loss": 3.3768095676933143,
      "tokens_seen": 2424373248
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.359492858289199e-05,
      "loss": 2.6334,
      "theoretical_loss": 3.3767949159164727,
      "tokens_seen": 2424504320
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.358690418873376e-05,
      "loss": 2.6085,
      "theoretical_loss": 3.3767802651534775,
      "tokens_seen": 2424635392
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.357887979457551e-05,
      "loss": 2.4126,
      "theoretical_loss": 3.376765615404204,
      "tokens_seen": 2424766464
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.357085540041728e-05,
      "loss": 2.3968,
      "theoretical_loss": 3.3767509666685265,
      "tokens_seen": 2424897536
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1332182,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9074771404266357,
      "objective/train/theoretical_loss": 3.3767436426807476,
      "objective/train/tokens_used": 795422176,
      "theoretical_loss": 3.3767436426807476,
      "tokens_seen": 2424963072
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.356283100625903e-05,
      "loss": 2.3835,
      "theoretical_loss": 3.376736318946321,
      "tokens_seen": 2425028608
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.355480661210079e-05,
      "loss": 2.3949,
      "theoretical_loss": 3.376721672237462,
      "tokens_seen": 2425159680
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.354678221794255e-05,
      "loss": 2.6908,
      "theoretical_loss": 3.3767070265418253,
      "tokens_seen": 2425290752
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3538757823784304e-05,
      "loss": 2.4389,
      "theoretical_loss": 3.3766923818592858,
      "tokens_seen": 2425421824
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.353073342962607e-05,
      "loss": 2.6217,
      "theoretical_loss": 3.3766777381897186,
      "tokens_seen": 2425552896
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3522709035467824e-05,
      "loss": 2.5104,
      "theoretical_loss": 3.3766630955329986,
      "tokens_seen": 2425683968
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3514684641309584e-05,
      "loss": 2.5317,
      "theoretical_loss": 3.3766484538890014,
      "tokens_seen": 2425815040
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3506660247151344e-05,
      "loss": 2.7119,
      "theoretical_loss": 3.376633813257602,
      "tokens_seen": 2425946112
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3498635852993104e-05,
      "loss": 2.4551,
      "theoretical_loss": 3.376619173638676,
      "tokens_seen": 2426077184
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3490611458834863e-05,
      "loss": 2.4489,
      "theoretical_loss": 3.3766045350320986,
      "tokens_seen": 2426208256
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3482587064676617e-05,
      "loss": 2.4335,
      "theoretical_loss": 3.3765898974377446,
      "tokens_seen": 2426339328
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.347456267051838e-05,
      "loss": 2.5277,
      "theoretical_loss": 3.37657526085549,
      "tokens_seen": 2426470400
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1333317,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.659700393676758,
      "objective/train/theoretical_loss": 3.37656062528521,
      "objective/train/tokens_used": 797060576,
      "theoretical_loss": 3.37656062528521,
      "tokens_seen": 2426601472
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3466538276360136e-05,
      "loss": 2.5915,
      "theoretical_loss": 3.37656062528521,
      "tokens_seen": 2426601472
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3458513882201896e-05,
      "loss": 2.4816,
      "theoretical_loss": 3.3765459907267794,
      "tokens_seen": 2426732544
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3450489488043656e-05,
      "loss": 2.4914,
      "theoretical_loss": 3.3765313571800744,
      "tokens_seen": 2426863616
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3442465093885416e-05,
      "loss": 2.5979,
      "theoretical_loss": 3.37651672464497,
      "tokens_seen": 2426994688
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3434440699727176e-05,
      "loss": 2.6746,
      "theoretical_loss": 3.3765020931213416,
      "tokens_seen": 2427125760
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.342641630556893e-05,
      "loss": 2.379,
      "theoretical_loss": 3.3764874626090653,
      "tokens_seen": 2427256832
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.341839191141069e-05,
      "loss": 2.562,
      "theoretical_loss": 3.3764728331080156,
      "tokens_seen": 2427387904
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.341036751725245e-05,
      "loss": 2.3364,
      "theoretical_loss": 3.3764582046180687,
      "tokens_seen": 2427518976
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.340234312309421e-05,
      "loss": 2.4422,
      "theoretical_loss": 3.3764435771390997,
      "tokens_seen": 2427650048
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.339431872893597e-05,
      "loss": 2.389,
      "theoretical_loss": 3.3764289506709844,
      "tokens_seen": 2427781120
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.338629433477773e-05,
      "loss": 2.5068,
      "theoretical_loss": 3.3764143252135983,
      "tokens_seen": 2427912192
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.337826994061949e-05,
      "loss": 2.532,
      "theoretical_loss": 3.376399700766817,
      "tokens_seen": 2428043264
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.337024554646124e-05,
      "loss": 2.4549,
      "theoretical_loss": 3.376385077330516,
      "tokens_seen": 2428174336
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1333920,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.964534044265747,
      "objective/train/theoretical_loss": 3.3763777659912573,
      "objective/train/tokens_used": 798698976,
      "theoretical_loss": 3.3763777659912573,
      "tokens_seen": 2428239872
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3362221152303e-05,
      "loss": 2.5831,
      "theoretical_loss": 3.376370454904572,
      "tokens_seen": 2428305408
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.335419675814476e-05,
      "loss": 2.5318,
      "theoretical_loss": 3.376355833488859,
      "tokens_seen": 2428436480
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.334617236398652e-05,
      "loss": 2.6454,
      "theoretical_loss": 3.3763412130832537,
      "tokens_seen": 2428567552
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.333814796982828e-05,
      "loss": 2.4917,
      "theoretical_loss": 3.376326593687631,
      "tokens_seen": 2428698624
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.333012357567004e-05,
      "loss": 2.4694,
      "theoretical_loss": 3.376311975301868,
      "tokens_seen": 2428829696
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3322099181511794e-05,
      "loss": 2.6382,
      "theoretical_loss": 3.3762973579258393,
      "tokens_seen": 2428960768
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.331407478735356e-05,
      "loss": 2.5724,
      "theoretical_loss": 3.3762827415594208,
      "tokens_seen": 2429091840
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3306050393195314e-05,
      "loss": 2.5351,
      "theoretical_loss": 3.3762681262024885,
      "tokens_seen": 2429222912
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3298025999037074e-05,
      "loss": 2.5725,
      "theoretical_loss": 3.3762535118549186,
      "tokens_seen": 2429353984
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3290001604878834e-05,
      "loss": 2.453,
      "theoretical_loss": 3.3762388985165863,
      "tokens_seen": 2429485056
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3281977210720594e-05,
      "loss": 2.6346,
      "theoretical_loss": 3.3762242861873677,
      "tokens_seen": 2429616128
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3273952816562354e-05,
      "loss": 2.5161,
      "theoretical_loss": 3.3762096748671384,
      "tokens_seen": 2429747200
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1334728,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.1242477893829346,
      "objective/train/theoretical_loss": 3.376195064555775,
      "objective/train/tokens_used": 800337376,
      "theoretical_loss": 3.376195064555775,
      "tokens_seen": 2429878272
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.326592842240411e-05,
      "loss": 2.5918,
      "theoretical_loss": 3.376195064555775,
      "tokens_seen": 2429878272
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3257904028245874e-05,
      "loss": 2.305,
      "theoretical_loss": 3.376180455253153,
      "tokens_seen": 2430009344
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.324987963408763e-05,
      "loss": 2.5245,
      "theoretical_loss": 3.3761658469591485,
      "tokens_seen": 2430140416
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.324185523992939e-05,
      "loss": 2.5522,
      "theoretical_loss": 3.376151239673637,
      "tokens_seen": 2430271488
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3233830845771147e-05,
      "loss": 2.5314,
      "theoretical_loss": 3.376136633396495,
      "tokens_seen": 2430402560
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.32258064516129e-05,
      "loss": 2.5198,
      "theoretical_loss": 3.3761220281275985,
      "tokens_seen": 2430533632
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3217782057454666e-05,
      "loss": 2.6067,
      "theoretical_loss": 3.376107423866823,
      "tokens_seen": 2430664704
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.320975766329642e-05,
      "loss": 2.5802,
      "theoretical_loss": 3.376092820614045,
      "tokens_seen": 2430795776
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3201733269138186e-05,
      "loss": 2.4849,
      "theoretical_loss": 3.376078218369141,
      "tokens_seen": 2430926848
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.319370887497994e-05,
      "loss": 2.461,
      "theoretical_loss": 3.3760636171319867,
      "tokens_seen": 2431057920
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.31856844808217e-05,
      "loss": 2.5747,
      "theoretical_loss": 3.3760490169024573,
      "tokens_seen": 2431188992
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.317766008666346e-05,
      "loss": 2.5342,
      "theoretical_loss": 3.3760344176804304,
      "tokens_seen": 2431320064
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.316963569250521e-05,
      "loss": 2.5691,
      "theoretical_loss": 3.376019819465782,
      "tokens_seen": 2431451136
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1335789,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4593210220336914,
      "objective/train/theoretical_loss": 3.376012520736185,
      "objective/train/tokens_used": 801975776,
      "theoretical_loss": 3.376012520736185,
      "tokens_seen": 2431516672
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.316161129834698e-05,
      "loss": 2.578,
      "theoretical_loss": 3.376005222258387,
      "tokens_seen": 2431582208
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.315358690418873e-05,
      "loss": 2.4974,
      "theoretical_loss": 3.3759906260581225,
      "tokens_seen": 2431713280
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.31455625100305e-05,
      "loss": 2.4921,
      "theoretical_loss": 3.375976030864865,
      "tokens_seen": 2431844352
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.313753811587225e-05,
      "loss": 2.5215,
      "theoretical_loss": 3.375961436678491,
      "tokens_seen": 2431975424
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3129513721714005e-05,
      "loss": 2.5972,
      "theoretical_loss": 3.3759468434988755,
      "tokens_seen": 2432106496
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.312148932755577e-05,
      "loss": 2.458,
      "theoretical_loss": 3.375932251325896,
      "tokens_seen": 2432237568
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3113464933397525e-05,
      "loss": 2.5595,
      "theoretical_loss": 3.375917660159428,
      "tokens_seen": 2432368640
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.310544053923929e-05,
      "loss": 2.4036,
      "theoretical_loss": 3.375903069999348,
      "tokens_seen": 2432499712
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3097416145081045e-05,
      "loss": 2.5762,
      "theoretical_loss": 3.375888480845533,
      "tokens_seen": 2432630784
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.308939175092281e-05,
      "loss": 2.57,
      "theoretical_loss": 3.3758738926978586,
      "tokens_seen": 2432761856
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3081367356764564e-05,
      "loss": 2.4884,
      "theoretical_loss": 3.375859305556202,
      "tokens_seen": 2432892928
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.307334296260632e-05,
      "loss": 2.6351,
      "theoretical_loss": 3.375844719420439,
      "tokens_seen": 2433024000
    },
    {
      "epoch": 0.47,
      "objective/train/docs_used": 1336369,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7755563259124756,
      "objective/train/theoretical_loss": 3.375830134290446,
      "objective/train/tokens_used": 803614176,
      "theoretical_loss": 3.375830134290446,
      "tokens_seen": 2433155072
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3065318568448084e-05,
      "loss": 2.4941,
      "theoretical_loss": 3.375830134290446,
      "tokens_seen": 2433155072
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.305729417428984e-05,
      "loss": 2.5475,
      "theoretical_loss": 3.3758155501661,
      "tokens_seen": 2433286144
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3049269780131604e-05,
      "loss": 2.6054,
      "theoretical_loss": 3.3758009670472773,
      "tokens_seen": 2433417216
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.304124538597336e-05,
      "loss": 2.7243,
      "theoretical_loss": 3.375786384933854,
      "tokens_seen": 2433548288
    },
    {
      "epoch": 0.47,
      "learning_rate": 5.3033220991815124e-05,
      "loss": 2.7134,
      "theoretical_loss": 3.375771803825707,
      "tokens_seen": 2433679360
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.302519659765688e-05,
      "loss": 2.63,
      "theoretical_loss": 3.375757223722713,
      "tokens_seen": 2433810432
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.301717220349863e-05,
      "loss": 2.4773,
      "theoretical_loss": 3.3757426446247485,
      "tokens_seen": 2433941504
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.30091478093404e-05,
      "loss": 2.6269,
      "theoretical_loss": 3.3757280665316896,
      "tokens_seen": 2434072576
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.300112341518215e-05,
      "loss": 2.5212,
      "theoretical_loss": 3.375713489443414,
      "tokens_seen": 2434203648
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2993099021023917e-05,
      "loss": 2.4834,
      "theoretical_loss": 3.375698913359797,
      "tokens_seen": 2434334720
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.298507462686567e-05,
      "loss": 2.5918,
      "theoretical_loss": 3.375684338280716,
      "tokens_seen": 2434465792
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2977050232707436e-05,
      "loss": 2.4533,
      "theoretical_loss": 3.375669764206048,
      "tokens_seen": 2434596864
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.296902583854919e-05,
      "loss": 2.5103,
      "theoretical_loss": 3.3756551911356696,
      "tokens_seen": 2434727936
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1337478,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.477384567260742,
      "objective/train/theoretical_loss": 3.37564790497705,
      "objective/train/tokens_used": 805252576,
      "theoretical_loss": 3.37564790497705,
      "tokens_seen": 2434793472
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2961001444390956e-05,
      "loss": 2.6785,
      "theoretical_loss": 3.375640619069457,
      "tokens_seen": 2434859008
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.295297705023271e-05,
      "loss": 2.5723,
      "theoretical_loss": 3.375626048007287,
      "tokens_seen": 2434990080
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.294495265607446e-05,
      "loss": 2.5132,
      "theoretical_loss": 3.375611477949037,
      "tokens_seen": 2435121152
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.293692826191623e-05,
      "loss": 2.525,
      "theoretical_loss": 3.3755969088945834,
      "tokens_seen": 2435252224
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.292890386775798e-05,
      "loss": 2.6585,
      "theoretical_loss": 3.375582340843803,
      "tokens_seen": 2435383296
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.292087947359975e-05,
      "loss": 2.547,
      "theoretical_loss": 3.375567773796573,
      "tokens_seen": 2435514368
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.29128550794415e-05,
      "loss": 2.5946,
      "theoretical_loss": 3.3755532077527692,
      "tokens_seen": 2435645440
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.290483068528327e-05,
      "loss": 2.4455,
      "theoretical_loss": 3.3755386427122698,
      "tokens_seen": 2435776512
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.289680629112502e-05,
      "loss": 2.5117,
      "theoretical_loss": 3.375524078674951,
      "tokens_seen": 2435907584
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2888781896966775e-05,
      "loss": 2.6521,
      "theoretical_loss": 3.3755095156406902,
      "tokens_seen": 2436038656
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.288075750280854e-05,
      "loss": 2.5638,
      "theoretical_loss": 3.3754949536093637,
      "tokens_seen": 2436169728
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2872733108650295e-05,
      "loss": 2.4175,
      "theoretical_loss": 3.375480392580849,
      "tokens_seen": 2436300800
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1338213,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.685955047607422,
      "objective/train/theoretical_loss": 3.375465832555023,
      "objective/train/tokens_used": 806890976,
      "theoretical_loss": 3.375465832555023,
      "tokens_seen": 2436431872
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.286470871449206e-05,
      "loss": 2.7174,
      "theoretical_loss": 3.375465832555023,
      "tokens_seen": 2436431872
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2856684320333815e-05,
      "loss": 2.5334,
      "theoretical_loss": 3.3754512735317626,
      "tokens_seen": 2436562944
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.284865992617558e-05,
      "loss": 2.485,
      "theoretical_loss": 3.3754367155109444,
      "tokens_seen": 2436694016
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2840635532017334e-05,
      "loss": 2.6672,
      "theoretical_loss": 3.3754221584924466,
      "tokens_seen": 2436825088
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.283261113785909e-05,
      "loss": 2.6874,
      "theoretical_loss": 3.3754076024761455,
      "tokens_seen": 2436956160
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2824586743700854e-05,
      "loss": 2.5222,
      "theoretical_loss": 3.375393047461918,
      "tokens_seen": 2437087232
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.281656234954261e-05,
      "loss": 2.6489,
      "theoretical_loss": 3.3753784934496416,
      "tokens_seen": 2437218304
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2808537955384374e-05,
      "loss": 2.5962,
      "theoretical_loss": 3.3753639404391937,
      "tokens_seen": 2437349376
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.280051356122613e-05,
      "loss": 2.5768,
      "theoretical_loss": 3.3753493884304513,
      "tokens_seen": 2437480448
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2792489167067894e-05,
      "loss": 2.5981,
      "theoretical_loss": 3.375334837423291,
      "tokens_seen": 2437611520
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.278446477290965e-05,
      "loss": 2.492,
      "theoretical_loss": 3.3753202874175905,
      "tokens_seen": 2437742592
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.27764403787514e-05,
      "loss": 2.4221,
      "theoretical_loss": 3.3753057384132275,
      "tokens_seen": 2437873664
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.276841598459317e-05,
      "loss": 2.5098,
      "theoretical_loss": 3.3752911904100786,
      "tokens_seen": 2438004736
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1339345,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3139004707336426,
      "objective/train/theoretical_loss": 3.375283916783921,
      "objective/train/tokens_used": 808529376,
      "theoretical_loss": 3.375283916783921,
      "tokens_seen": 2438070272
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.276039159043492e-05,
      "loss": 2.4769,
      "theoretical_loss": 3.375276643408021,
      "tokens_seen": 2438135808
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.275236719627669e-05,
      "loss": 2.6318,
      "theoretical_loss": 3.3752620974069325,
      "tokens_seen": 2438266880
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.274434280211844e-05,
      "loss": 2.6198,
      "theoretical_loss": 3.37524755240669,
      "tokens_seen": 2438397952
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2736318407960206e-05,
      "loss": 2.4233,
      "theoretical_loss": 3.3752330084071707,
      "tokens_seen": 2438529024
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.272829401380196e-05,
      "loss": 2.7931,
      "theoretical_loss": 3.3752184654082527,
      "tokens_seen": 2438660096
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.272026961964371e-05,
      "loss": 2.4301,
      "theoretical_loss": 3.375203923409813,
      "tokens_seen": 2438791168
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.271224522548548e-05,
      "loss": 2.5547,
      "theoretical_loss": 3.3751893824117287,
      "tokens_seen": 2438922240
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.270422083132723e-05,
      "loss": 2.467,
      "theoretical_loss": 3.3751748424138777,
      "tokens_seen": 2439053312
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2696196437169e-05,
      "loss": 2.4588,
      "theoretical_loss": 3.3751603034161373,
      "tokens_seen": 2439184384
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.268817204301075e-05,
      "loss": 2.5843,
      "theoretical_loss": 3.375145765418385,
      "tokens_seen": 2439315456
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.268014764885252e-05,
      "loss": 2.4959,
      "theoretical_loss": 3.3751312284204977,
      "tokens_seen": 2439446528
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.267212325469427e-05,
      "loss": 2.6227,
      "theoretical_loss": 3.375116692422354,
      "tokens_seen": 2439577600
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1339900,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3154780864715576,
      "objective/train/theoretical_loss": 3.3751021574238305,
      "objective/train/tokens_used": 810167776,
      "theoretical_loss": 3.3751021574238305,
      "tokens_seen": 2439708672
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.266409886053604e-05,
      "loss": 2.5463,
      "theoretical_loss": 3.3751021574238305,
      "tokens_seen": 2439708672
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.265607446637779e-05,
      "loss": 2.6162,
      "theoretical_loss": 3.3750876234248053,
      "tokens_seen": 2439839744
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2648050072219545e-05,
      "loss": 2.4569,
      "theoretical_loss": 3.375073090425156,
      "tokens_seen": 2439970816
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.264002567806131e-05,
      "loss": 2.7436,
      "theoretical_loss": 3.37505855842476,
      "tokens_seen": 2440101888
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2632001283903065e-05,
      "loss": 2.5632,
      "theoretical_loss": 3.3750440274234945,
      "tokens_seen": 2440232960
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.262397688974483e-05,
      "loss": 2.5628,
      "theoretical_loss": 3.375029497421238,
      "tokens_seen": 2440364032
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2615952495586585e-05,
      "loss": 2.5599,
      "theoretical_loss": 3.3750149684178674,
      "tokens_seen": 2440495104
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.260792810142835e-05,
      "loss": 2.456,
      "theoretical_loss": 3.375000440413261,
      "tokens_seen": 2440626176
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2599903707270105e-05,
      "loss": 2.4967,
      "theoretical_loss": 3.374985913407296,
      "tokens_seen": 2440757248
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.259187931311186e-05,
      "loss": 2.4109,
      "theoretical_loss": 3.374971387399851,
      "tokens_seen": 2440888320
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2583854918953624e-05,
      "loss": 2.4852,
      "theoretical_loss": 3.3749568623908024,
      "tokens_seen": 2441019392
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.257583052479538e-05,
      "loss": 2.5694,
      "theoretical_loss": 3.3749423383800288,
      "tokens_seen": 2441150464
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2567806130637144e-05,
      "loss": 2.4533,
      "theoretical_loss": 3.374927815367408,
      "tokens_seen": 2441281536
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1340505,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.749099016189575,
      "objective/train/theoretical_loss": 3.374920554235367,
      "objective/train/tokens_used": 811806176,
      "theoretical_loss": 3.374920554235367,
      "tokens_seen": 2441347072
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.25597817364789e-05,
      "loss": 2.4812,
      "theoretical_loss": 3.374913293352818,
      "tokens_seen": 2441412608
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2551757342320664e-05,
      "loss": 2.735,
      "theoretical_loss": 3.3748987723361363,
      "tokens_seen": 2441543680
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.254373294816242e-05,
      "loss": 2.6154,
      "theoretical_loss": 3.3748842523172407,
      "tokens_seen": 2441674752
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.253570855400417e-05,
      "loss": 2.5746,
      "theoretical_loss": 3.3748697332960087,
      "tokens_seen": 2441805824
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.252768415984594e-05,
      "loss": 2.4751,
      "theoretical_loss": 3.3748552152723192,
      "tokens_seen": 2441936896
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.251965976568769e-05,
      "loss": 2.4473,
      "theoretical_loss": 3.3748406982460497,
      "tokens_seen": 2442067968
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.251163537152946e-05,
      "loss": 2.6456,
      "theoretical_loss": 3.374826182217078,
      "tokens_seen": 2442199040
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.250361097737121e-05,
      "loss": 2.5668,
      "theoretical_loss": 3.3748116671852824,
      "tokens_seen": 2442330112
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2495586583212977e-05,
      "loss": 2.4327,
      "theoretical_loss": 3.3747971531505403,
      "tokens_seen": 2442461184
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.248756218905473e-05,
      "loss": 2.6369,
      "theoretical_loss": 3.3747826401127297,
      "tokens_seen": 2442592256
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.247953779489648e-05,
      "loss": 2.5001,
      "theoretical_loss": 3.3747681280717297,
      "tokens_seen": 2442723328
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.247151340073825e-05,
      "loss": 2.4715,
      "theoretical_loss": 3.374753617027417,
      "tokens_seen": 2442854400
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1341788,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.345665454864502,
      "objective/train/theoretical_loss": 3.374739106979671,
      "objective/train/tokens_used": 813444576,
      "theoretical_loss": 3.374739106979671,
      "tokens_seen": 2442985472
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.246348900658e-05,
      "loss": 2.5201,
      "theoretical_loss": 3.374739106979671,
      "tokens_seen": 2442985472
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.245546461242177e-05,
      "loss": 2.4456,
      "theoretical_loss": 3.3747245979283687,
      "tokens_seen": 2443116544
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.244744021826352e-05,
      "loss": 2.5736,
      "theoretical_loss": 3.374710089873388,
      "tokens_seen": 2443247616
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.243941582410529e-05,
      "loss": 2.5839,
      "theoretical_loss": 3.3746955828146086,
      "tokens_seen": 2443378688
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.243139142994704e-05,
      "loss": 2.5527,
      "theoretical_loss": 3.3746810767519073,
      "tokens_seen": 2443509760
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2423367035788795e-05,
      "loss": 2.6065,
      "theoretical_loss": 3.374666571685163,
      "tokens_seen": 2443640832
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.241534264163056e-05,
      "loss": 2.63,
      "theoretical_loss": 3.3746520676142535,
      "tokens_seen": 2443771904
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2407318247472315e-05,
      "loss": 2.6752,
      "theoretical_loss": 3.374637564539057,
      "tokens_seen": 2443902976
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.239929385331408e-05,
      "loss": 2.5175,
      "theoretical_loss": 3.374623062459452,
      "tokens_seen": 2444034048
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2391269459155835e-05,
      "loss": 2.5524,
      "theoretical_loss": 3.3746085613753163,
      "tokens_seen": 2444165120
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.23832450649976e-05,
      "loss": 2.5963,
      "theoretical_loss": 3.374594061286529,
      "tokens_seen": 2444296192
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2375220670839355e-05,
      "loss": 2.5523,
      "theoretical_loss": 3.3745795621929675,
      "tokens_seen": 2444427264
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.236719627668112e-05,
      "loss": 2.5234,
      "theoretical_loss": 3.374565064094511,
      "tokens_seen": 2444558336
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1342804,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.517542839050293,
      "objective/train/theoretical_loss": 3.374557815418409,
      "objective/train/tokens_used": 815082976,
      "theoretical_loss": 3.374557815418409,
      "tokens_seen": 2444623872
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2359171882522875e-05,
      "loss": 2.6831,
      "theoretical_loss": 3.3745505669910374,
      "tokens_seen": 2444689408
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.235114748836463e-05,
      "loss": 2.6545,
      "theoretical_loss": 3.374536070882425,
      "tokens_seen": 2444820480
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2343123094206394e-05,
      "loss": 2.7343,
      "theoretical_loss": 3.3745215757685525,
      "tokens_seen": 2444951552
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.233509870004815e-05,
      "loss": 2.5024,
      "theoretical_loss": 3.3745070816492975,
      "tokens_seen": 2445082624
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2327074305889914e-05,
      "loss": 2.6012,
      "theoretical_loss": 3.37449258852454,
      "tokens_seen": 2445213696
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.231904991173167e-05,
      "loss": 2.5065,
      "theoretical_loss": 3.374478096394157,
      "tokens_seen": 2445344768
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2311025517573434e-05,
      "loss": 2.7124,
      "theoretical_loss": 3.3744636052580272,
      "tokens_seen": 2445475840
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.230300112341519e-05,
      "loss": 2.4797,
      "theoretical_loss": 3.37444911511603,
      "tokens_seen": 2445606912
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.229497672925694e-05,
      "loss": 2.624,
      "theoretical_loss": 3.3744346259680427,
      "tokens_seen": 2445737984
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.228695233509871e-05,
      "loss": 2.6437,
      "theoretical_loss": 3.374420137813945,
      "tokens_seen": 2445869056
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.227892794094046e-05,
      "loss": 2.591,
      "theoretical_loss": 3.374405650653615,
      "tokens_seen": 2446000128
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.227090354678223e-05,
      "loss": 2.6138,
      "theoretical_loss": 3.374391164486931,
      "tokens_seen": 2446131200
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1343168,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7890498638153076,
      "objective/train/theoretical_loss": 3.3743766793137717,
      "objective/train/tokens_used": 816721376,
      "theoretical_loss": 3.3743766793137717,
      "tokens_seen": 2446262272
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.226287915262398e-05,
      "loss": 2.6727,
      "theoretical_loss": 3.3743766793137717,
      "tokens_seen": 2446262272
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.225485475846575e-05,
      "loss": 2.5655,
      "theoretical_loss": 3.374362195134016,
      "tokens_seen": 2446393344
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.22468303643075e-05,
      "loss": 2.6566,
      "theoretical_loss": 3.374347711947543,
      "tokens_seen": 2446524416
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.223880597014925e-05,
      "loss": 2.4497,
      "theoretical_loss": 3.37433322975423,
      "tokens_seen": 2446655488
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.223078157599102e-05,
      "loss": 2.5973,
      "theoretical_loss": 3.3743187485539567,
      "tokens_seen": 2446786560
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.222275718183277e-05,
      "loss": 2.5613,
      "theoretical_loss": 3.374304268346602,
      "tokens_seen": 2446917632
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.221473278767454e-05,
      "loss": 2.5539,
      "theoretical_loss": 3.374289789132044,
      "tokens_seen": 2447048704
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.220670839351629e-05,
      "loss": 2.4292,
      "theoretical_loss": 3.3742753109101615,
      "tokens_seen": 2447179776
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.219868399935805e-05,
      "loss": 2.3909,
      "theoretical_loss": 3.3742608336808337,
      "tokens_seen": 2447310848
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.219065960519981e-05,
      "loss": 2.5961,
      "theoretical_loss": 3.3742463574439396,
      "tokens_seen": 2447441920
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2182635211041565e-05,
      "loss": 2.5862,
      "theoretical_loss": 3.374231882199357,
      "tokens_seen": 2447572992
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.217461081688333e-05,
      "loss": 2.5475,
      "theoretical_loss": 3.374217407946966,
      "tokens_seen": 2447704064
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2166586422725085e-05,
      "loss": 2.6481,
      "theoretical_loss": 3.374202934686644,
      "tokens_seen": 2447835136
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1344177,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.358229398727417,
      "objective/train/theoretical_loss": 3.374195698428472,
      "objective/train/tokens_used": 818359776,
      "theoretical_loss": 3.374195698428472,
      "tokens_seen": 2447900672
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.215856202856685e-05,
      "loss": 2.5777,
      "theoretical_loss": 3.3741884624182714,
      "tokens_seen": 2447966208
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2150537634408605e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.3741739911417263,
      "tokens_seen": 2448097280
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2142513240250365e-05,
      "loss": 2.6933,
      "theoretical_loss": 3.3741595208568875,
      "tokens_seen": 2448228352
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2134488846092125e-05,
      "loss": 2.4876,
      "theoretical_loss": 3.3741450515636346,
      "tokens_seen": 2448359424
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.212646445193388e-05,
      "loss": 2.5319,
      "theoretical_loss": 3.374130583261846,
      "tokens_seen": 2448490496
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.2118440057775645e-05,
      "loss": 2.6603,
      "theoretical_loss": 3.374116115951401,
      "tokens_seen": 2448621568
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.21104156636174e-05,
      "loss": 2.7219,
      "theoretical_loss": 3.374101649632178,
      "tokens_seen": 2448752640
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.210239126945916e-05,
      "loss": 2.4685,
      "theoretical_loss": 3.3740871843040567,
      "tokens_seen": 2448883712
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.209436687530092e-05,
      "loss": 2.6344,
      "theoretical_loss": 3.3740727199669163,
      "tokens_seen": 2449014784
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.208634248114268e-05,
      "loss": 2.7244,
      "theoretical_loss": 3.3740582566206356,
      "tokens_seen": 2449145856
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.207831808698444e-05,
      "loss": 2.5339,
      "theoretical_loss": 3.374043794265093,
      "tokens_seen": 2449276928
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.207029369282619e-05,
      "loss": 2.4953,
      "theoretical_loss": 3.374029332900169,
      "tokens_seen": 2449408000
    },
    {
      "epoch": 0.48,
      "objective/train/docs_used": 1344624,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.644493818283081,
      "objective/train/theoretical_loss": 3.3740148725257417,
      "objective/train/tokens_used": 819998176,
      "theoretical_loss": 3.3740148725257417,
      "tokens_seen": 2449539072
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.206226929866795e-05,
      "loss": 2.6183,
      "theoretical_loss": 3.3740148725257417,
      "tokens_seen": 2449539072
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.205424490450971e-05,
      "loss": 2.7257,
      "theoretical_loss": 3.3740004131416907,
      "tokens_seen": 2449670144
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.204622051035147e-05,
      "loss": 2.6306,
      "theoretical_loss": 3.373985954747895,
      "tokens_seen": 2449801216
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.203819611619323e-05,
      "loss": 2.7159,
      "theoretical_loss": 3.3739714973442343,
      "tokens_seen": 2449932288
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.203017172203499e-05,
      "loss": 2.5434,
      "theoretical_loss": 3.373957040930587,
      "tokens_seen": 2450063360
    },
    {
      "epoch": 0.48,
      "learning_rate": 5.202214732787675e-05,
      "loss": 2.5135,
      "theoretical_loss": 3.373942585506833,
      "tokens_seen": 2450194432
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.201412293371851e-05,
      "loss": 2.6426,
      "theoretical_loss": 3.3739281310728506,
      "tokens_seen": 2450325504
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.200609853956026e-05,
      "loss": 2.5788,
      "theoretical_loss": 3.3739136776285203,
      "tokens_seen": 2450456576
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.199807414540202e-05,
      "loss": 2.5713,
      "theoretical_loss": 3.373899225173721,
      "tokens_seen": 2450587648
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.199004975124378e-05,
      "loss": 2.6554,
      "theoretical_loss": 3.3738847737083324,
      "tokens_seen": 2450718720
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.198202535708554e-05,
      "loss": 2.6084,
      "theoretical_loss": 3.3738703232322327,
      "tokens_seen": 2450849792
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.19740009629273e-05,
      "loss": 2.5222,
      "theoretical_loss": 3.373855873745302,
      "tokens_seen": 2450980864
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1965976568769056e-05,
      "loss": 2.5954,
      "theoretical_loss": 3.37384142524742,
      "tokens_seen": 2451111936
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1344624,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7756404876708984,
      "objective/train/theoretical_loss": 3.3738342013693345,
      "objective/train/tokens_used": 821636576,
      "theoretical_loss": 3.3738342013693345,
      "tokens_seen": 2451177472
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.195795217461082e-05,
      "loss": 2.524,
      "theoretical_loss": 3.373826977738466,
      "tokens_seen": 2451243008
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1949927780452576e-05,
      "loss": 2.6562,
      "theoretical_loss": 3.3738125312183187,
      "tokens_seen": 2451374080
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1941903386294336e-05,
      "loss": 2.6662,
      "theoretical_loss": 3.3737980856868584,
      "tokens_seen": 2451505152
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1933878992136095e-05,
      "loss": 2.8164,
      "theoretical_loss": 3.3737836411439646,
      "tokens_seen": 2451636224
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1925854597977855e-05,
      "loss": 2.7042,
      "theoretical_loss": 3.3737691975895157,
      "tokens_seen": 2451767296
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1917830203819615e-05,
      "loss": 2.764,
      "theoretical_loss": 3.3737547550233926,
      "tokens_seen": 2451898368
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.190980580966137e-05,
      "loss": 2.8597,
      "theoretical_loss": 3.373740313445474,
      "tokens_seen": 2452029440
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1901781415503135e-05,
      "loss": 2.7401,
      "theoretical_loss": 3.37372587285564,
      "tokens_seen": 2452160512
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.189375702134489e-05,
      "loss": 2.8246,
      "theoretical_loss": 3.37371143325377,
      "tokens_seen": 2452291584
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.188573262718665e-05,
      "loss": 2.8532,
      "theoretical_loss": 3.373696994639743,
      "tokens_seen": 2452422656
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.187770823302841e-05,
      "loss": 2.7613,
      "theoretical_loss": 3.3736825570134394,
      "tokens_seen": 2452553728
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.186968383887016e-05,
      "loss": 2.7873,
      "theoretical_loss": 3.3736681203747385,
      "tokens_seen": 2452684800
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1345207,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.519474983215332,
      "objective/train/theoretical_loss": 3.37365368472352,
      "objective/train/tokens_used": 823274976,
      "theoretical_loss": 3.37365368472352,
      "tokens_seen": 2452815872
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.186165944471193e-05,
      "loss": 2.7948,
      "theoretical_loss": 3.37365368472352,
      "tokens_seen": 2452815872
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.185363505055368e-05,
      "loss": 2.6965,
      "theoretical_loss": 3.373639250059664,
      "tokens_seen": 2452946944
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.184561065639545e-05,
      "loss": 2.8823,
      "theoretical_loss": 3.3736248163830496,
      "tokens_seen": 2453078016
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.18375862622372e-05,
      "loss": 2.8049,
      "theoretical_loss": 3.3736103836935567,
      "tokens_seen": 2453209088
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.182956186807896e-05,
      "loss": 2.7712,
      "theoretical_loss": 3.3735959519910654,
      "tokens_seen": 2453340160
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.182153747392072e-05,
      "loss": 2.7904,
      "theoretical_loss": 3.373581521275455,
      "tokens_seen": 2453471232
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1813513079762474e-05,
      "loss": 2.7418,
      "theoretical_loss": 3.3735670915466054,
      "tokens_seen": 2453602304
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.180548868560424e-05,
      "loss": 2.7428,
      "theoretical_loss": 3.3735526628043973,
      "tokens_seen": 2453733376
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1797464291445994e-05,
      "loss": 2.6637,
      "theoretical_loss": 3.3735382350487093,
      "tokens_seen": 2453864448
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.178943989728776e-05,
      "loss": 2.7932,
      "theoretical_loss": 3.3735238082794217,
      "tokens_seen": 2453995520
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.178141550312951e-05,
      "loss": 2.7252,
      "theoretical_loss": 3.3735093824964144,
      "tokens_seen": 2454126592
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1773391108971266e-05,
      "loss": 2.7983,
      "theoretical_loss": 3.373494957699567,
      "tokens_seen": 2454257664
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.176536671481303e-05,
      "loss": 2.5646,
      "theoretical_loss": 3.3734805338887606,
      "tokens_seen": 2454388736
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1345665,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9347848892211914,
      "objective/train/theoretical_loss": 3.3734733223530844,
      "objective/train/tokens_used": 824913376,
      "theoretical_loss": 3.3734733223530844,
      "tokens_seen": 2454454272
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1757342320654786e-05,
      "loss": 2.6224,
      "theoretical_loss": 3.3734661110638737,
      "tokens_seen": 2454519808
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.174931792649655e-05,
      "loss": 2.7118,
      "theoretical_loss": 3.3734516892247868,
      "tokens_seen": 2454650880
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1741293532338306e-05,
      "loss": 2.6218,
      "theoretical_loss": 3.3734372683713802,
      "tokens_seen": 2454781952
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.173326913818007e-05,
      "loss": 2.6646,
      "theoretical_loss": 3.373422848503534,
      "tokens_seen": 2454913024
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1725244744021826e-05,
      "loss": 2.5645,
      "theoretical_loss": 3.3734084296211275,
      "tokens_seen": 2455044096
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.171722034986359e-05,
      "loss": 2.5378,
      "theoretical_loss": 3.373394011724041,
      "tokens_seen": 2455175168
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1709195955705346e-05,
      "loss": 2.9382,
      "theoretical_loss": 3.373379594812155,
      "tokens_seen": 2455306240
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.17011715615471e-05,
      "loss": 2.6852,
      "theoretical_loss": 3.373365178885349,
      "tokens_seen": 2455437312
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1693147167388866e-05,
      "loss": 2.7406,
      "theoretical_loss": 3.373350763943504,
      "tokens_seen": 2455568384
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.168512277323062e-05,
      "loss": 2.6722,
      "theoretical_loss": 3.373336349986499,
      "tokens_seen": 2455699456
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1677098379072385e-05,
      "loss": 2.7807,
      "theoretical_loss": 3.373321937014215,
      "tokens_seen": 2455830528
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.166907398491414e-05,
      "loss": 2.6833,
      "theoretical_loss": 3.3733075250265316,
      "tokens_seen": 2455961600
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1346863,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5424647331237793,
      "objective/train/theoretical_loss": 3.3732931140233293,
      "objective/train/tokens_used": 826551776,
      "theoretical_loss": 3.3732931140233293,
      "tokens_seen": 2456092672
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1661049590755905e-05,
      "loss": 2.6451,
      "theoretical_loss": 3.3732931140233293,
      "tokens_seen": 2456092672
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.165302519659766e-05,
      "loss": 2.662,
      "theoretical_loss": 3.3732787040044885,
      "tokens_seen": 2456223744
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.164500080243941e-05,
      "loss": 2.6761,
      "theoretical_loss": 3.373264294969889,
      "tokens_seen": 2456354816
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.163697640828118e-05,
      "loss": 2.7512,
      "theoretical_loss": 3.3732498869194116,
      "tokens_seen": 2456485888
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.162895201412293e-05,
      "loss": 2.6402,
      "theoretical_loss": 3.373235479852936,
      "tokens_seen": 2456616960
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.16209276199647e-05,
      "loss": 2.7339,
      "theoretical_loss": 3.3732210737703427,
      "tokens_seen": 2456748032
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.161290322580645e-05,
      "loss": 2.5177,
      "theoretical_loss": 3.3732066686715125,
      "tokens_seen": 2456879104
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.160487883164822e-05,
      "loss": 2.7342,
      "theoretical_loss": 3.373192264556325,
      "tokens_seen": 2457010176
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.159685443748997e-05,
      "loss": 2.7059,
      "theoretical_loss": 3.373177861424661,
      "tokens_seen": 2457141248
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1588830043331724e-05,
      "loss": 2.7184,
      "theoretical_loss": 3.373163459276401,
      "tokens_seen": 2457272320
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.158080564917349e-05,
      "loss": 2.7066,
      "theoretical_loss": 3.3731490581114247,
      "tokens_seen": 2457403392
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1572781255015244e-05,
      "loss": 2.6591,
      "theoretical_loss": 3.3731346579296133,
      "tokens_seen": 2457534464
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.156475686085701e-05,
      "loss": 2.8136,
      "theoretical_loss": 3.3731202587308475,
      "tokens_seen": 2457665536
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1347964,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6252315044403076,
      "objective/train/theoretical_loss": 3.373113059500069,
      "objective/train/tokens_used": 828190176,
      "theoretical_loss": 3.373113059500069,
      "tokens_seen": 2457731072
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1556732466698764e-05,
      "loss": 2.7172,
      "theoretical_loss": 3.3731058605150066,
      "tokens_seen": 2457796608
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.154870807254053e-05,
      "loss": 2.7197,
      "theoretical_loss": 3.3730914632819715,
      "tokens_seen": 2457927680
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1540683678382283e-05,
      "loss": 2.7076,
      "theoretical_loss": 3.3730770670316237,
      "tokens_seen": 2458058752
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1532659284224037e-05,
      "loss": 2.5872,
      "theoretical_loss": 3.3730626717638423,
      "tokens_seen": 2458189824
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.15246348900658e-05,
      "loss": 2.5952,
      "theoretical_loss": 3.373048277478509,
      "tokens_seen": 2458320896
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1516610495907556e-05,
      "loss": 2.6224,
      "theoretical_loss": 3.373033884175504,
      "tokens_seen": 2458451968
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.150858610174932e-05,
      "loss": 2.8066,
      "theoretical_loss": 3.3730194918547074,
      "tokens_seen": 2458583040
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1500561707591076e-05,
      "loss": 2.6329,
      "theoretical_loss": 3.373005100516,
      "tokens_seen": 2458714112
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.149253731343284e-05,
      "loss": 2.6037,
      "theoretical_loss": 3.372990710159263,
      "tokens_seen": 2458845184
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1484512919274596e-05,
      "loss": 2.7772,
      "theoretical_loss": 3.372976320784377,
      "tokens_seen": 2458976256
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.147648852511635e-05,
      "loss": 2.5924,
      "theoretical_loss": 3.3729619323912217,
      "tokens_seen": 2459107328
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1468464130958116e-05,
      "loss": 2.5872,
      "theoretical_loss": 3.372947544979679,
      "tokens_seen": 2459238400
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1348513,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.667315721511841,
      "objective/train/theoretical_loss": 3.372933158549629,
      "objective/train/tokens_used": 829828576,
      "theoretical_loss": 3.372933158549629,
      "tokens_seen": 2459369472
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.146043973679987e-05,
      "loss": 2.55,
      "theoretical_loss": 3.372933158549629,
      "tokens_seen": 2459369472
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1452415342641636e-05,
      "loss": 2.8019,
      "theoretical_loss": 3.3729187731009524,
      "tokens_seen": 2459500544
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.144439094848339e-05,
      "loss": 2.6017,
      "theoretical_loss": 3.37290438863353,
      "tokens_seen": 2459631616
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1436366554325155e-05,
      "loss": 2.7036,
      "theoretical_loss": 3.372890005147243,
      "tokens_seen": 2459762688
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.142834216016691e-05,
      "loss": 2.6547,
      "theoretical_loss": 3.372875622641972,
      "tokens_seen": 2459893760
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1420317766008675e-05,
      "loss": 2.5111,
      "theoretical_loss": 3.3728612411175973,
      "tokens_seen": 2460024832
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.141229337185043e-05,
      "loss": 2.7321,
      "theoretical_loss": 3.372846860574,
      "tokens_seen": 2460155904
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.140426897769218e-05,
      "loss": 2.6469,
      "theoretical_loss": 3.3728324810110615,
      "tokens_seen": 2460286976
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.139624458353395e-05,
      "loss": 2.6475,
      "theoretical_loss": 3.3728181024286625,
      "tokens_seen": 2460418048
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.13882201893757e-05,
      "loss": 2.7767,
      "theoretical_loss": 3.3728037248266833,
      "tokens_seen": 2460549120
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.138019579521747e-05,
      "loss": 2.7516,
      "theoretical_loss": 3.3727893482050053,
      "tokens_seen": 2460680192
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.137217140105922e-05,
      "loss": 2.7426,
      "theoretical_loss": 3.3727749725635094,
      "tokens_seen": 2460811264
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.136414700690099e-05,
      "loss": 2.5846,
      "theoretical_loss": 3.3727605979020767,
      "tokens_seen": 2460942336
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1349898,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7990500926971436,
      "objective/train/theoretical_loss": 3.372753410938847,
      "objective/train/tokens_used": 831466976,
      "theoretical_loss": 3.372753410938847,
      "tokens_seen": 2461007872
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.135612261274274e-05,
      "loss": 2.608,
      "theoretical_loss": 3.372746224220588,
      "tokens_seen": 2461073408
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1348098218584494e-05,
      "loss": 2.6354,
      "theoretical_loss": 3.372731851518924,
      "tokens_seen": 2461204480
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.134007382442626e-05,
      "loss": 2.7203,
      "theoretical_loss": 3.3727174797969663,
      "tokens_seen": 2461335552
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1332049430268014e-05,
      "loss": 2.4329,
      "theoretical_loss": 3.372703109054596,
      "tokens_seen": 2461466624
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.132402503610978e-05,
      "loss": 2.7455,
      "theoretical_loss": 3.3726887392916938,
      "tokens_seen": 2461597696
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1316000641951534e-05,
      "loss": 2.5633,
      "theoretical_loss": 3.372674370508141,
      "tokens_seen": 2461728768
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.13079762477933e-05,
      "loss": 2.7246,
      "theoretical_loss": 3.3726600027038183,
      "tokens_seen": 2461859840
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1299951853635053e-05,
      "loss": 2.6832,
      "theoretical_loss": 3.3726456358786074,
      "tokens_seen": 2461990912
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1291927459476807e-05,
      "loss": 2.6317,
      "theoretical_loss": 3.3726312700323895,
      "tokens_seen": 2462121984
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.128390306531857e-05,
      "loss": 2.6313,
      "theoretical_loss": 3.372616905165045,
      "tokens_seen": 2462253056
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1275878671160326e-05,
      "loss": 2.7807,
      "theoretical_loss": 3.3726025412764558,
      "tokens_seen": 2462384128
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.126785427700209e-05,
      "loss": 2.6791,
      "theoretical_loss": 3.372588178366503,
      "tokens_seen": 2462515200
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1350590,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.718848466873169,
      "objective/train/theoretical_loss": 3.3725738164350676,
      "objective/train/tokens_used": 833105376,
      "theoretical_loss": 3.3725738164350676,
      "tokens_seen": 2462646272
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1259829882843846e-05,
      "loss": 2.5802,
      "theoretical_loss": 3.3725738164350676,
      "tokens_seen": 2462646272
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.125180548868561e-05,
      "loss": 2.6466,
      "theoretical_loss": 3.372559455482031,
      "tokens_seen": 2462777344
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1243781094527366e-05,
      "loss": 2.6131,
      "theoretical_loss": 3.3725450955072747,
      "tokens_seen": 2462908416
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.123575670036912e-05,
      "loss": 2.4635,
      "theoretical_loss": 3.3725307365106794,
      "tokens_seen": 2463039488
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1227732306210886e-05,
      "loss": 2.6381,
      "theoretical_loss": 3.3725163784921275,
      "tokens_seen": 2463170560
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.121970791205264e-05,
      "loss": 2.6174,
      "theoretical_loss": 3.372502021451499,
      "tokens_seen": 2463301632
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1211683517894406e-05,
      "loss": 2.6745,
      "theoretical_loss": 3.372487665388676,
      "tokens_seen": 2463432704
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.120365912373616e-05,
      "loss": 2.5964,
      "theoretical_loss": 3.37247331030354,
      "tokens_seen": 2463563776
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1195634729577925e-05,
      "loss": 2.7596,
      "theoretical_loss": 3.372458956195972,
      "tokens_seen": 2463694848
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.118761033541968e-05,
      "loss": 2.7028,
      "theoretical_loss": 3.372444603065854,
      "tokens_seen": 2463825920
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.117958594126143e-05,
      "loss": 2.637,
      "theoretical_loss": 3.372430250913067,
      "tokens_seen": 2463956992
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.11715615471032e-05,
      "loss": 2.6179,
      "theoretical_loss": 3.3724158997374922,
      "tokens_seen": 2464088064
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.116353715294495e-05,
      "loss": 2.6899,
      "theoretical_loss": 3.3724015495390116,
      "tokens_seen": 2464219136
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1351948,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8566057682037354,
      "objective/train/theoretical_loss": 3.3723943748061447,
      "objective/train/tokens_used": 834743776,
      "theoretical_loss": 3.3723943748061447,
      "tokens_seen": 2464284672
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.115551275878672e-05,
      "loss": 2.5858,
      "theoretical_loss": 3.372387200317507,
      "tokens_seen": 2464350208
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.114748836462847e-05,
      "loss": 2.6211,
      "theoretical_loss": 3.372372852072859,
      "tokens_seen": 2464481280
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.113946397047024e-05,
      "loss": 2.7022,
      "theoretical_loss": 3.3723585048049496,
      "tokens_seen": 2464612352
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.113143957631199e-05,
      "loss": 2.6312,
      "theoretical_loss": 3.3723441585136604,
      "tokens_seen": 2464743424
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1123415182153744e-05,
      "loss": 2.5399,
      "theoretical_loss": 3.3723298131988733,
      "tokens_seen": 2464874496
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.111539078799551e-05,
      "loss": 2.6586,
      "theoretical_loss": 3.3723154688604695,
      "tokens_seen": 2465005568
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1107366393837264e-05,
      "loss": 2.5115,
      "theoretical_loss": 3.3723011254983306,
      "tokens_seen": 2465136640
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.109934199967903e-05,
      "loss": 2.615,
      "theoretical_loss": 3.3722867831123384,
      "tokens_seen": 2465267712
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1091317605520784e-05,
      "loss": 2.7011,
      "theoretical_loss": 3.372272441702375,
      "tokens_seen": 2465398784
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.108329321136255e-05,
      "loss": 2.4709,
      "theoretical_loss": 3.372258101268321,
      "tokens_seen": 2465529856
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1075268817204304e-05,
      "loss": 2.7755,
      "theoretical_loss": 3.3722437618100587,
      "tokens_seen": 2465660928
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.106724442304607e-05,
      "loss": 2.5329,
      "theoretical_loss": 3.3722294233274703,
      "tokens_seen": 2465792000
    },
    {
      "epoch": 0.49,
      "objective/train/docs_used": 1352648,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8698949813842773,
      "objective/train/theoretical_loss": 3.372215085820437,
      "objective/train/tokens_used": 836382176,
      "theoretical_loss": 3.372215085820437,
      "tokens_seen": 2465923072
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1059220028887824e-05,
      "loss": 2.5483,
      "theoretical_loss": 3.372215085820437,
      "tokens_seen": 2465923072
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.105119563472958e-05,
      "loss": 2.7256,
      "theoretical_loss": 3.372200749288841,
      "tokens_seen": 2466054144
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.104317124057134e-05,
      "loss": 2.5776,
      "theoretical_loss": 3.372186413732564,
      "tokens_seen": 2466185216
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1035146846413096e-05,
      "loss": 2.6369,
      "theoretical_loss": 3.372172079151487,
      "tokens_seen": 2466316288
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.102712245225486e-05,
      "loss": 2.7825,
      "theoretical_loss": 3.3721577455454925,
      "tokens_seen": 2466447360
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.1019098058096616e-05,
      "loss": 2.7237,
      "theoretical_loss": 3.372143412914463,
      "tokens_seen": 2466578432
    },
    {
      "epoch": 0.49,
      "learning_rate": 5.101107366393838e-05,
      "loss": 2.6892,
      "theoretical_loss": 3.3721290812582794,
      "tokens_seen": 2466709504
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.1003049269780136e-05,
      "loss": 2.639,
      "theoretical_loss": 3.3721147505768236,
      "tokens_seen": 2466840576
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.099502487562189e-05,
      "loss": 2.5936,
      "theoretical_loss": 3.3721004208699785,
      "tokens_seen": 2466971648
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0987000481463656e-05,
      "loss": 2.5451,
      "theoretical_loss": 3.372086092137625,
      "tokens_seen": 2467102720
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.097897608730541e-05,
      "loss": 2.6,
      "theoretical_loss": 3.3720717643796454,
      "tokens_seen": 2467233792
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0970951693147176e-05,
      "loss": 2.64,
      "theoretical_loss": 3.372057437595922,
      "tokens_seen": 2467364864
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.096292729898893e-05,
      "loss": 2.5725,
      "theoretical_loss": 3.3720431117863363,
      "tokens_seen": 2467495936
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1353854,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4819204807281494,
      "objective/train/theoretical_loss": 3.3720359492468086,
      "objective/train/tokens_used": 838020576,
      "theoretical_loss": 3.3720359492468086,
      "tokens_seen": 2467561472
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0954902904830696e-05,
      "loss": 2.598,
      "theoretical_loss": 3.3720287869507706,
      "tokens_seen": 2467627008
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.094687851067245e-05,
      "loss": 2.5879,
      "theoretical_loss": 3.3720144630891076,
      "tokens_seen": 2467758080
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.09388541165142e-05,
      "loss": 2.6088,
      "theoretical_loss": 3.372000140201228,
      "tokens_seen": 2467889152
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.093082972235597e-05,
      "loss": 2.7607,
      "theoretical_loss": 3.3719858182870146,
      "tokens_seen": 2468020224
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.092280532819772e-05,
      "loss": 2.5092,
      "theoretical_loss": 3.3719714973463493,
      "tokens_seen": 2468151296
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.091478093403949e-05,
      "loss": 2.7098,
      "theoretical_loss": 3.3719571773791146,
      "tokens_seen": 2468282368
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.090675653988124e-05,
      "loss": 2.4803,
      "theoretical_loss": 3.3719428583851925,
      "tokens_seen": 2468413440
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.089873214572301e-05,
      "loss": 2.5423,
      "theoretical_loss": 3.3719285403644648,
      "tokens_seen": 2468544512
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.089070775156476e-05,
      "loss": 2.5517,
      "theoretical_loss": 3.3719142233168142,
      "tokens_seen": 2468675584
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0882683357406514e-05,
      "loss": 2.6932,
      "theoretical_loss": 3.371899907242123,
      "tokens_seen": 2468806656
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.087465896324828e-05,
      "loss": 2.5645,
      "theoretical_loss": 3.371885592140273,
      "tokens_seen": 2468937728
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0866634569090034e-05,
      "loss": 2.5671,
      "theoretical_loss": 3.371871278011146,
      "tokens_seen": 2469068800
    },
    {
      "debugging/Self-BLEU-5": 0.42407658532325987,
      "debugging/distinct-1-grams": 0.7177325581395348,
      "debugging/distinct-2-grams": 0.8906535491482647,
      "debugging/entropy-1-grams": 5.181652846266418,
      "debugging/entropy-2-grams": 5.834764426176351,
      "debugging/length": 560.8,
      "debugging/num_segments": 5,
      "debugging/score": 0.0042737600834809476,
      "debugging/score_std": 0.004323690737910133,
      "epoch": 0.5,
      "objective/train/docs_used": 1354527,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7488014698028564,
      "objective/train/theoretical_loss": 3.3718569648546257,
      "objective/train/tokens_used": 839658976,
      "theoretical_loss": 3.3718569648546257,
      "tokens_seen": 2469199872
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.08586101749318e-05,
      "loss": 2.6643,
      "theoretical_loss": 3.3718569648546257,
      "tokens_seen": 2469199872
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0850585780773554e-05,
      "loss": 2.5209,
      "theoretical_loss": 3.371842652670593,
      "tokens_seen": 2469330944
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0842561386615314e-05,
      "loss": 2.6081,
      "theoretical_loss": 3.371828341458931,
      "tokens_seen": 2469462016
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0834536992457074e-05,
      "loss": 2.7511,
      "theoretical_loss": 3.371814031219522,
      "tokens_seen": 2469593088
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.082651259829883e-05,
      "loss": 2.635,
      "theoretical_loss": 3.371799721952248,
      "tokens_seen": 2469724160
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0818488204140594e-05,
      "loss": 2.5922,
      "theoretical_loss": 3.3717854136569914,
      "tokens_seen": 2469855232
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.081046380998235e-05,
      "loss": 2.7167,
      "theoretical_loss": 3.3717711063336346,
      "tokens_seen": 2469986304
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0802439415824113e-05,
      "loss": 2.4931,
      "theoretical_loss": 3.37175679998206,
      "tokens_seen": 2470117376
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0794415021665867e-05,
      "loss": 2.5988,
      "theoretical_loss": 3.3717424946021506,
      "tokens_seen": 2470248448
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0786390627507626e-05,
      "loss": 2.4903,
      "theoretical_loss": 3.371728190193788,
      "tokens_seen": 2470379520
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0778366233349386e-05,
      "loss": 2.5597,
      "theoretical_loss": 3.3717138867568552,
      "tokens_seen": 2470510592
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0770341839191146e-05,
      "loss": 2.3949,
      "theoretical_loss": 3.371699584291235,
      "tokens_seen": 2470641664
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0762317445032906e-05,
      "loss": 2.6598,
      "theoretical_loss": 3.371685282796809,
      "tokens_seen": 2470772736
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1355685,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7235753536224365,
      "objective/train/theoretical_loss": 3.3716781324137575,
      "objective/train/tokens_used": 841297376,
      "theoretical_loss": 3.3716781324137575,
      "tokens_seen": 2470838272
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.075429305087466e-05,
      "loss": 2.6547,
      "theoretical_loss": 3.3716709822734603,
      "tokens_seen": 2470903808
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.074626865671642e-05,
      "loss": 2.5513,
      "theoretical_loss": 3.3716566827210714,
      "tokens_seen": 2471034880
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.073824426255818e-05,
      "loss": 2.4664,
      "theoretical_loss": 3.371642384139525,
      "tokens_seen": 2471165952
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.073021986839994e-05,
      "loss": 2.4618,
      "theoretical_loss": 3.3716280865287036,
      "tokens_seen": 2471297024
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.07221954742417e-05,
      "loss": 2.7009,
      "theoretical_loss": 3.3716137898884897,
      "tokens_seen": 2471428096
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.071417108008346e-05,
      "loss": 2.6315,
      "theoretical_loss": 3.371599494218766,
      "tokens_seen": 2471559168
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.070614668592522e-05,
      "loss": 2.4532,
      "theoretical_loss": 3.3715851995194153,
      "tokens_seen": 2471690240
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.069812229176697e-05,
      "loss": 2.6446,
      "theoretical_loss": 3.3715709057903203,
      "tokens_seen": 2471821312
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.069009789760873e-05,
      "loss": 2.6,
      "theoretical_loss": 3.371556613031363,
      "tokens_seen": 2471952384
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.068207350345049e-05,
      "loss": 2.4892,
      "theoretical_loss": 3.371542321242427,
      "tokens_seen": 2472083456
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.067404910929225e-05,
      "loss": 2.608,
      "theoretical_loss": 3.371528030423395,
      "tokens_seen": 2472214528
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.066602471513401e-05,
      "loss": 2.5077,
      "theoretical_loss": 3.371513740574149,
      "tokens_seen": 2472345600
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1356735,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5951905250549316,
      "objective/train/theoretical_loss": 3.3714994516945724,
      "objective/train/tokens_used": 842935776,
      "theoretical_loss": 3.3714994516945724,
      "tokens_seen": 2472476672
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.065800032097577e-05,
      "loss": 2.6154,
      "theoretical_loss": 3.3714994516945724,
      "tokens_seen": 2472476672
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0649975926817525e-05,
      "loss": 2.5361,
      "theoretical_loss": 3.371485163784548,
      "tokens_seen": 2472607744
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0641951532659284e-05,
      "loss": 2.5984,
      "theoretical_loss": 3.3714708768439583,
      "tokens_seen": 2472738816
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0633927138501044e-05,
      "loss": 2.6947,
      "theoretical_loss": 3.3714565908726866,
      "tokens_seen": 2472869888
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0625902744342804e-05,
      "loss": 2.6804,
      "theoretical_loss": 3.371442305870615,
      "tokens_seen": 2473000960
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0617878350184564e-05,
      "loss": 2.5793,
      "theoretical_loss": 3.3714280218376276,
      "tokens_seen": 2473132032
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0609853956026324e-05,
      "loss": 2.7422,
      "theoretical_loss": 3.371413738773606,
      "tokens_seen": 2473263104
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0601829561868084e-05,
      "loss": 2.7063,
      "theoretical_loss": 3.371399456678434,
      "tokens_seen": 2473394176
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.059380516770984e-05,
      "loss": 2.6895,
      "theoretical_loss": 3.371385175551994,
      "tokens_seen": 2473525248
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.05857807735516e-05,
      "loss": 2.7181,
      "theoretical_loss": 3.3713708953941692,
      "tokens_seen": 2473656320
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.057775637939336e-05,
      "loss": 2.4096,
      "theoretical_loss": 3.3713566162048427,
      "tokens_seen": 2473787392
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.056973198523512e-05,
      "loss": 2.6133,
      "theoretical_loss": 3.3713423379838976,
      "tokens_seen": 2473918464
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.056170759107688e-05,
      "loss": 2.6098,
      "theoretical_loss": 3.3713280607312166,
      "tokens_seen": 2474049536
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1357343,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.478666067123413,
      "objective/train/theoretical_loss": 3.3713209224679384,
      "objective/train/tokens_used": 844574176,
      "theoretical_loss": 3.3713209224679384,
      "tokens_seen": 2474115072
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.055368319691863e-05,
      "loss": 2.6532,
      "theoretical_loss": 3.3713137844466825,
      "tokens_seen": 2474180608
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0545658802760397e-05,
      "loss": 2.6792,
      "theoretical_loss": 3.3712995091301794,
      "tokens_seen": 2474311680
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.053763440860215e-05,
      "loss": 2.7248,
      "theoretical_loss": 3.371285234781589,
      "tokens_seen": 2474442752
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.052961001444391e-05,
      "loss": 2.5951,
      "theoretical_loss": 3.371270961400796,
      "tokens_seen": 2474573824
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.052158562028567e-05,
      "loss": 2.6659,
      "theoretical_loss": 3.371256688987682,
      "tokens_seen": 2474704896
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.051356122612743e-05,
      "loss": 2.6503,
      "theoretical_loss": 3.371242417542131,
      "tokens_seen": 2474835968
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.050553683196919e-05,
      "loss": 2.7322,
      "theoretical_loss": 3.3712281470640257,
      "tokens_seen": 2474967040
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.049751243781094e-05,
      "loss": 2.7879,
      "theoretical_loss": 3.37121387755325,
      "tokens_seen": 2475098112
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.048948804365271e-05,
      "loss": 2.6329,
      "theoretical_loss": 3.3711996090096865,
      "tokens_seen": 2475229184
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.048146364949446e-05,
      "loss": 2.4507,
      "theoretical_loss": 3.3711853414332187,
      "tokens_seen": 2475360256
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.047343925533623e-05,
      "loss": 2.6312,
      "theoretical_loss": 3.3711710748237294,
      "tokens_seen": 2475491328
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.046541486117798e-05,
      "loss": 2.7004,
      "theoretical_loss": 3.3711568091811026,
      "tokens_seen": 2475622400
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1358753,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6069936752319336,
      "objective/train/theoretical_loss": 3.371142544505221,
      "objective/train/tokens_used": 846212576,
      "theoretical_loss": 3.371142544505221,
      "tokens_seen": 2475753472
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0457390467019735e-05,
      "loss": 2.4947,
      "theoretical_loss": 3.371142544505221,
      "tokens_seen": 2475753472
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.04493660728615e-05,
      "loss": 2.4908,
      "theoretical_loss": 3.3711282807959684,
      "tokens_seen": 2475884544
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0441341678703255e-05,
      "loss": 2.5781,
      "theoretical_loss": 3.3711140180532273,
      "tokens_seen": 2476015616
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.043331728454502e-05,
      "loss": 2.5672,
      "theoretical_loss": 3.371099756276882,
      "tokens_seen": 2476146688
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0425292890386775e-05,
      "loss": 2.4306,
      "theoretical_loss": 3.3710854954668155,
      "tokens_seen": 2476277760
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.041726849622854e-05,
      "loss": 2.6745,
      "theoretical_loss": 3.371071235622911,
      "tokens_seen": 2476408832
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0409244102070295e-05,
      "loss": 2.6364,
      "theoretical_loss": 3.3710569767450522,
      "tokens_seen": 2476539904
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.040121970791205e-05,
      "loss": 2.6201,
      "theoretical_loss": 3.371042718833122,
      "tokens_seen": 2476670976
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0393195313753814e-05,
      "loss": 2.7649,
      "theoretical_loss": 3.3710284618870046,
      "tokens_seen": 2476802048
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.038517091959557e-05,
      "loss": 2.4941,
      "theoretical_loss": 3.371014205906583,
      "tokens_seen": 2476933120
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0377146525437334e-05,
      "loss": 2.5364,
      "theoretical_loss": 3.3709999508917408,
      "tokens_seen": 2477064192
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.036912213127909e-05,
      "loss": 2.7514,
      "theoretical_loss": 3.3709856968423617,
      "tokens_seen": 2477195264
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0361097737120854e-05,
      "loss": 2.6732,
      "theoretical_loss": 3.370971443758329,
      "tokens_seen": 2477326336
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1359369,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6083905696868896,
      "objective/train/theoretical_loss": 3.3709643175782813,
      "objective/train/tokens_used": 847850976,
      "theoretical_loss": 3.3709643175782813,
      "tokens_seen": 2477391872
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.035307334296261e-05,
      "loss": 2.5511,
      "theoretical_loss": 3.3709571916395262,
      "tokens_seen": 2477457408
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.034504894880436e-05,
      "loss": 2.6522,
      "theoretical_loss": 3.370942940485837,
      "tokens_seen": 2477588480
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.033702455464613e-05,
      "loss": 2.6202,
      "theoretical_loss": 3.3709286902971454,
      "tokens_seen": 2477719552
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.032900016048788e-05,
      "loss": 2.3975,
      "theoretical_loss": 3.3709144410733343,
      "tokens_seen": 2477850624
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.032097576632965e-05,
      "loss": 2.4792,
      "theoretical_loss": 3.3709001928142874,
      "tokens_seen": 2477981696
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.03129513721714e-05,
      "loss": 2.5769,
      "theoretical_loss": 3.3708859455198885,
      "tokens_seen": 2478112768
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.030492697801317e-05,
      "loss": 2.6005,
      "theoretical_loss": 3.3708716991900216,
      "tokens_seen": 2478243840
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.029690258385492e-05,
      "loss": 2.6088,
      "theoretical_loss": 3.3708574538245704,
      "tokens_seen": 2478374912
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.028887818969667e-05,
      "loss": 2.366,
      "theoretical_loss": 3.370843209423418,
      "tokens_seen": 2478505984
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.028085379553844e-05,
      "loss": 2.6869,
      "theoretical_loss": 3.3708289659864485,
      "tokens_seen": 2478637056
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.027282940138019e-05,
      "loss": 2.5781,
      "theoretical_loss": 3.3708147235135457,
      "tokens_seen": 2478768128
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.026480500722196e-05,
      "loss": 2.5796,
      "theoretical_loss": 3.3708004820045936,
      "tokens_seen": 2478899200
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1360741,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.478170871734619,
      "objective/train/theoretical_loss": 3.3707862414594754,
      "objective/train/tokens_used": 849489376,
      "theoretical_loss": 3.3707862414594754,
      "tokens_seen": 2479030272
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.025678061306371e-05,
      "loss": 2.6031,
      "theoretical_loss": 3.3707862414594754,
      "tokens_seen": 2479030272
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.024875621890548e-05,
      "loss": 2.7964,
      "theoretical_loss": 3.370772001878075,
      "tokens_seen": 2479161344
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.024073182474723e-05,
      "loss": 2.5007,
      "theoretical_loss": 3.370757763260277,
      "tokens_seen": 2479292416
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0232707430588985e-05,
      "loss": 2.5323,
      "theoretical_loss": 3.3707435256059646,
      "tokens_seen": 2479423488
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.022468303643075e-05,
      "loss": 2.6202,
      "theoretical_loss": 3.3707292889150215,
      "tokens_seen": 2479554560
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0216658642272505e-05,
      "loss": 2.6986,
      "theoretical_loss": 3.370715053187332,
      "tokens_seen": 2479685632
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.020863424811427e-05,
      "loss": 2.7548,
      "theoretical_loss": 3.37070081842278,
      "tokens_seen": 2479816704
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0200609853956025e-05,
      "loss": 2.5603,
      "theoretical_loss": 3.3706865846212493,
      "tokens_seen": 2479947776
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.019258545979779e-05,
      "loss": 2.6339,
      "theoretical_loss": 3.3706723517826243,
      "tokens_seen": 2480078848
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0184561065639545e-05,
      "loss": 2.6331,
      "theoretical_loss": 3.3706581199067878,
      "tokens_seen": 2480209920
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.01765366714813e-05,
      "loss": 2.4755,
      "theoretical_loss": 3.370643888993625,
      "tokens_seen": 2480340992
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0168512277323065e-05,
      "loss": 2.5282,
      "theoretical_loss": 3.3706296590430194,
      "tokens_seen": 2480472064
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.016048788316482e-05,
      "loss": 2.5725,
      "theoretical_loss": 3.370615430054855,
      "tokens_seen": 2480603136
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1361264,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.49733829498291,
      "objective/train/theoretical_loss": 3.370608315921652,
      "objective/train/tokens_used": 851127776,
      "theoretical_loss": 3.370608315921652,
      "tokens_seen": 2480668672
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0152463489006585e-05,
      "loss": 2.5991,
      "theoretical_loss": 3.3706012020290164,
      "tokens_seen": 2480734208
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.014443909484834e-05,
      "loss": 2.4878,
      "theoretical_loss": 3.370586974965387,
      "tokens_seen": 2480865280
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0136414700690104e-05,
      "loss": 2.5512,
      "theoretical_loss": 3.370572748863851,
      "tokens_seen": 2480996352
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.012839030653186e-05,
      "loss": 2.4598,
      "theoretical_loss": 3.370558523724293,
      "tokens_seen": 2481127424
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.0120365912373624e-05,
      "loss": 2.5487,
      "theoretical_loss": 3.3705442995465966,
      "tokens_seen": 2481258496
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.011234151821538e-05,
      "loss": 2.5929,
      "theoretical_loss": 3.3705300763306463,
      "tokens_seen": 2481389568
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.010431712405713e-05,
      "loss": 2.6349,
      "theoretical_loss": 3.3705158540763263,
      "tokens_seen": 2481520640
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.00962927298989e-05,
      "loss": 2.5983,
      "theoretical_loss": 3.3705016327835198,
      "tokens_seen": 2481651712
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.008826833574065e-05,
      "loss": 2.4785,
      "theoretical_loss": 3.3704874124521127,
      "tokens_seen": 2481782784
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.008024394158242e-05,
      "loss": 2.5442,
      "theoretical_loss": 3.3704731930819882,
      "tokens_seen": 2481913856
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.007221954742417e-05,
      "loss": 2.6366,
      "theoretical_loss": 3.3704589746730305,
      "tokens_seen": 2482044928
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.006419515326594e-05,
      "loss": 2.5095,
      "theoretical_loss": 3.3704447572251244,
      "tokens_seen": 2482176000
    },
    {
      "epoch": 0.5,
      "objective/train/docs_used": 1362376,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.822953462600708,
      "objective/train/theoretical_loss": 3.3704305407381536,
      "objective/train/tokens_used": 852766176,
      "theoretical_loss": 3.3704305407381536,
      "tokens_seen": 2482307072
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.005617075910769e-05,
      "loss": 2.5868,
      "theoretical_loss": 3.3704305407381536,
      "tokens_seen": 2482307072
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.004814636494944e-05,
      "loss": 2.5643,
      "theoretical_loss": 3.370416325212003,
      "tokens_seen": 2482438144
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.004012197079121e-05,
      "loss": 2.6704,
      "theoretical_loss": 3.3704021106465567,
      "tokens_seen": 2482569216
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.003209757663296e-05,
      "loss": 2.4693,
      "theoretical_loss": 3.370387897041699,
      "tokens_seen": 2482700288
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.002407318247473e-05,
      "loss": 2.4997,
      "theoretical_loss": 3.370373684397314,
      "tokens_seen": 2482831360
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.001604878831648e-05,
      "loss": 2.5091,
      "theoretical_loss": 3.370359472713287,
      "tokens_seen": 2482962432
    },
    {
      "epoch": 0.5,
      "learning_rate": 5.000802439415825e-05,
      "loss": 2.5122,
      "theoretical_loss": 3.3703452619895016,
      "tokens_seen": 2483093504
    },
    {
      "epoch": 0.51,
      "learning_rate": 5e-05,
      "loss": 2.6438,
      "theoretical_loss": 3.370331052225842,
      "tokens_seen": 2483224576
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.999197560584176e-05,
      "loss": 2.5141,
      "theoretical_loss": 3.3703168434221937,
      "tokens_seen": 2483355648
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.998395121168352e-05,
      "loss": 2.5419,
      "theoretical_loss": 3.3703026355784402,
      "tokens_seen": 2483486720
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.997592681752528e-05,
      "loss": 2.556,
      "theoretical_loss": 3.370288428694467,
      "tokens_seen": 2483617792
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9967902423367035e-05,
      "loss": 2.6535,
      "theoretical_loss": 3.370274222770157,
      "tokens_seen": 2483748864
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9959878029208795e-05,
      "loss": 2.5,
      "theoretical_loss": 3.3702600178053967,
      "tokens_seen": 2483879936
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1362743,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5020596981048584,
      "objective/train/theoretical_loss": 3.370252915682811,
      "objective/train/tokens_used": 854404576,
      "theoretical_loss": 3.370252915682811,
      "tokens_seen": 2483945472
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9951853635050555e-05,
      "loss": 2.6361,
      "theoretical_loss": 3.3702458138000693,
      "tokens_seen": 2484011008
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9943829240892315e-05,
      "loss": 2.5663,
      "theoretical_loss": 3.3702316107540597,
      "tokens_seen": 2484142080
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9935804846734075e-05,
      "loss": 2.427,
      "theoretical_loss": 3.3702174086672527,
      "tokens_seen": 2484273152
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9927780452575835e-05,
      "loss": 2.6123,
      "theoretical_loss": 3.3702032075395327,
      "tokens_seen": 2484404224
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9919756058417595e-05,
      "loss": 2.5062,
      "theoretical_loss": 3.3701890073707848,
      "tokens_seen": 2484535296
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9911731664259355e-05,
      "loss": 2.5672,
      "theoretical_loss": 3.370174808160893,
      "tokens_seen": 2484666368
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.990370727010111e-05,
      "loss": 2.4534,
      "theoretical_loss": 3.370160609909742,
      "tokens_seen": 2484797440
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.989568287594287e-05,
      "loss": 2.5751,
      "theoretical_loss": 3.3701464126172174,
      "tokens_seen": 2484928512
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.988765848178463e-05,
      "loss": 2.4986,
      "theoretical_loss": 3.370132216283203,
      "tokens_seen": 2485059584
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.987963408762639e-05,
      "loss": 2.4899,
      "theoretical_loss": 3.370118020907584,
      "tokens_seen": 2485190656
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.987160969346815e-05,
      "loss": 2.5682,
      "theoretical_loss": 3.3701038264902445,
      "tokens_seen": 2485321728
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.986358529930991e-05,
      "loss": 2.556,
      "theoretical_loss": 3.37008963303107,
      "tokens_seen": 2485452800
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1363961,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8076930046081543,
      "objective/train/theoretical_loss": 3.370075440529945,
      "objective/train/tokens_used": 856042976,
      "theoretical_loss": 3.370075440529945,
      "tokens_seen": 2485583872
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.985556090515167e-05,
      "loss": 2.649,
      "theoretical_loss": 3.370075440529945,
      "tokens_seen": 2485583872
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.984753651099342e-05,
      "loss": 2.4928,
      "theoretical_loss": 3.370061248986754,
      "tokens_seen": 2485714944
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.983951211683518e-05,
      "loss": 2.5232,
      "theoretical_loss": 3.370047058401383,
      "tokens_seen": 2485846016
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.983148772267694e-05,
      "loss": 2.6823,
      "theoretical_loss": 3.3700328687737153,
      "tokens_seen": 2485977088
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.98234633285187e-05,
      "loss": 2.4137,
      "theoretical_loss": 3.3700186801036374,
      "tokens_seen": 2486108160
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.981543893436046e-05,
      "loss": 2.5207,
      "theoretical_loss": 3.3700044923910326,
      "tokens_seen": 2486239232
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.980741454020222e-05,
      "loss": 2.5952,
      "theoretical_loss": 3.3699903056357865,
      "tokens_seen": 2486370304
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.979939014604398e-05,
      "loss": 2.5019,
      "theoretical_loss": 3.3699761198377844,
      "tokens_seen": 2486501376
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.979136575188573e-05,
      "loss": 2.6931,
      "theoretical_loss": 3.369961934996911,
      "tokens_seen": 2486632448
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.978334135772749e-05,
      "loss": 2.6033,
      "theoretical_loss": 3.3699477511130516,
      "tokens_seen": 2486763520
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.977531696356925e-05,
      "loss": 2.5466,
      "theoretical_loss": 3.36993356818609,
      "tokens_seen": 2486894592
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.976729256941101e-05,
      "loss": 2.4649,
      "theoretical_loss": 3.3699193862159125,
      "tokens_seen": 2487025664
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.975926817525277e-05,
      "loss": 2.491,
      "theoretical_loss": 3.3699052052024037,
      "tokens_seen": 2487156736
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1365358,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3158061504364014,
      "objective/train/theoretical_loss": 3.369898115054364,
      "objective/train/tokens_used": 857681376,
      "theoretical_loss": 3.369898115054364,
      "tokens_seen": 2487222272
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.975124378109453e-05,
      "loss": 2.5351,
      "theoretical_loss": 3.3698910251454484,
      "tokens_seen": 2487287808
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.974321938693629e-05,
      "loss": 2.5235,
      "theoretical_loss": 3.369876846044932,
      "tokens_seen": 2487418880
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.973519499277805e-05,
      "loss": 2.5791,
      "theoretical_loss": 3.3698626679007395,
      "tokens_seen": 2487549952
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9727170598619805e-05,
      "loss": 2.5045,
      "theoretical_loss": 3.369848490712756,
      "tokens_seen": 2487681024
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9719146204461565e-05,
      "loss": 2.5234,
      "theoretical_loss": 3.3698343144808662,
      "tokens_seen": 2487812096
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9711121810303325e-05,
      "loss": 2.6729,
      "theoretical_loss": 3.3698201392049563,
      "tokens_seen": 2487943168
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9703097416145085e-05,
      "loss": 2.6268,
      "theoretical_loss": 3.3698059648849106,
      "tokens_seen": 2488074240
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9695073021986845e-05,
      "loss": 2.5327,
      "theoretical_loss": 3.3697917915206146,
      "tokens_seen": 2488205312
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9687048627828605e-05,
      "loss": 2.6413,
      "theoretical_loss": 3.3697776191119537,
      "tokens_seen": 2488336384
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9679024233670365e-05,
      "loss": 2.6266,
      "theoretical_loss": 3.3697634476588125,
      "tokens_seen": 2488467456
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.967099983951212e-05,
      "loss": 2.4853,
      "theoretical_loss": 3.369749277161077,
      "tokens_seen": 2488598528
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.966297544535388e-05,
      "loss": 2.5342,
      "theoretical_loss": 3.3697351076186317,
      "tokens_seen": 2488729600
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1366123,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4803953170776367,
      "objective/train/theoretical_loss": 3.369720939031363,
      "objective/train/tokens_used": 859319776,
      "theoretical_loss": 3.369720939031363,
      "tokens_seen": 2488860672
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.965495105119564e-05,
      "loss": 2.5624,
      "theoretical_loss": 3.369720939031363,
      "tokens_seen": 2488860672
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.96469266570374e-05,
      "loss": 2.7138,
      "theoretical_loss": 3.369706771399155,
      "tokens_seen": 2488991744
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.963890226287916e-05,
      "loss": 2.4183,
      "theoretical_loss": 3.3696926047218936,
      "tokens_seen": 2489122816
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.963087786872092e-05,
      "loss": 2.5153,
      "theoretical_loss": 3.3696784389994643,
      "tokens_seen": 2489253888
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.962285347456268e-05,
      "loss": 2.5115,
      "theoretical_loss": 3.369664274231752,
      "tokens_seen": 2489384960
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.961482908040443e-05,
      "loss": 2.6792,
      "theoretical_loss": 3.369650110418643,
      "tokens_seen": 2489516032
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.960680468624619e-05,
      "loss": 2.614,
      "theoretical_loss": 3.369635947560022,
      "tokens_seen": 2489647104
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.959878029208795e-05,
      "loss": 2.6546,
      "theoretical_loss": 3.369621785655774,
      "tokens_seen": 2489778176
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.959075589792971e-05,
      "loss": 2.5551,
      "theoretical_loss": 3.3696076247057856,
      "tokens_seen": 2489909248
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.958273150377147e-05,
      "loss": 2.4587,
      "theoretical_loss": 3.3695934647099413,
      "tokens_seen": 2490040320
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.957470710961323e-05,
      "loss": 2.4979,
      "theoretical_loss": 3.369579305668127,
      "tokens_seen": 2490171392
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.956668271545499e-05,
      "loss": 2.5172,
      "theoretical_loss": 3.3695651475802286,
      "tokens_seen": 2490302464
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.955865832129675e-05,
      "loss": 2.6105,
      "theoretical_loss": 3.3695509904461307,
      "tokens_seen": 2490433536
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1367242,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.245980978012085,
      "objective/train/theoretical_loss": 3.3695439122367215,
      "objective/train/tokens_used": 860958176,
      "theoretical_loss": 3.3695439122367215,
      "tokens_seen": 2490499072
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.95506339271385e-05,
      "loss": 2.4907,
      "theoretical_loss": 3.3695368342657197,
      "tokens_seen": 2490564608
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.954260953298026e-05,
      "loss": 2.5153,
      "theoretical_loss": 3.36952267903888,
      "tokens_seen": 2490695680
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.953458513882202e-05,
      "loss": 2.6679,
      "theoretical_loss": 3.3695085247654992,
      "tokens_seen": 2490826752
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.952656074466378e-05,
      "loss": 2.5825,
      "theoretical_loss": 3.369494371445461,
      "tokens_seen": 2490957824
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.951853635050554e-05,
      "loss": 2.6121,
      "theoretical_loss": 3.369480219078652,
      "tokens_seen": 2491088896
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.95105119563473e-05,
      "loss": 2.6157,
      "theoretical_loss": 3.3694660676649573,
      "tokens_seen": 2491219968
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.950248756218906e-05,
      "loss": 2.5463,
      "theoretical_loss": 3.369451917204263,
      "tokens_seen": 2491351040
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9494463168030815e-05,
      "loss": 2.6447,
      "theoretical_loss": 3.369437767696455,
      "tokens_seen": 2491482112
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9486438773872575e-05,
      "loss": 2.5978,
      "theoretical_loss": 3.3694236191414184,
      "tokens_seen": 2491613184
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9478414379714335e-05,
      "loss": 2.6127,
      "theoretical_loss": 3.369409471539039,
      "tokens_seen": 2491744256
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9470389985556095e-05,
      "loss": 2.6995,
      "theoretical_loss": 3.369395324889203,
      "tokens_seen": 2491875328
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9462365591397855e-05,
      "loss": 2.3666,
      "theoretical_loss": 3.369381179191796,
      "tokens_seen": 2492006400
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1367729,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2709412574768066,
      "objective/train/theoretical_loss": 3.3693670344467033,
      "objective/train/tokens_used": 862596576,
      "theoretical_loss": 3.3693670344467033,
      "tokens_seen": 2492137472
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9454341197239615e-05,
      "loss": 2.4568,
      "theoretical_loss": 3.3693670344467033,
      "tokens_seen": 2492137472
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9446316803081375e-05,
      "loss": 2.6108,
      "theoretical_loss": 3.3693528906538113,
      "tokens_seen": 2492268544
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.943829240892313e-05,
      "loss": 2.7159,
      "theoretical_loss": 3.3693387478130057,
      "tokens_seen": 2492399616
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.943026801476489e-05,
      "loss": 2.3573,
      "theoretical_loss": 3.3693246059241724,
      "tokens_seen": 2492530688
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.942224362060665e-05,
      "loss": 2.4629,
      "theoretical_loss": 3.3693104649871968,
      "tokens_seen": 2492661760
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.941421922644841e-05,
      "loss": 2.4896,
      "theoretical_loss": 3.369296325001965,
      "tokens_seen": 2492792832
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.940619483229017e-05,
      "loss": 2.6108,
      "theoretical_loss": 3.369282185968364,
      "tokens_seen": 2492923904
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.939817043813193e-05,
      "loss": 2.4589,
      "theoretical_loss": 3.369268047886278,
      "tokens_seen": 2493054976
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.939014604397368e-05,
      "loss": 2.439,
      "theoretical_loss": 3.3692539107555937,
      "tokens_seen": 2493186048
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.938212164981544e-05,
      "loss": 2.7177,
      "theoretical_loss": 3.3692397745761973,
      "tokens_seen": 2493317120
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.93740972556572e-05,
      "loss": 2.6196,
      "theoretical_loss": 3.3692256393479747,
      "tokens_seen": 2493448192
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.936607286149896e-05,
      "loss": 2.5573,
      "theoretical_loss": 3.369211505070811,
      "tokens_seen": 2493579264
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.935804846734072e-05,
      "loss": 2.6251,
      "theoretical_loss": 3.3691973717445935,
      "tokens_seen": 2493710336
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1368772,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5175697803497314,
      "objective/train/theoretical_loss": 3.369190305438054,
      "objective/train/tokens_used": 864234976,
      "theoretical_loss": 3.369190305438054,
      "tokens_seen": 2493775872
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.935002407318248e-05,
      "loss": 2.5824,
      "theoretical_loss": 3.369183239369208,
      "tokens_seen": 2493841408
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.934199967902423e-05,
      "loss": 2.4809,
      "theoretical_loss": 3.36916910794454,
      "tokens_seen": 2493972480
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.933397528486599e-05,
      "loss": 2.5429,
      "theoretical_loss": 3.369154977470476,
      "tokens_seen": 2494103552
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.932595089070775e-05,
      "loss": 2.5156,
      "theoretical_loss": 3.369140847946902,
      "tokens_seen": 2494234624
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.931792649654951e-05,
      "loss": 2.5766,
      "theoretical_loss": 3.369126719373704,
      "tokens_seen": 2494365696
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.930990210239127e-05,
      "loss": 2.4916,
      "theoretical_loss": 3.369112591750768,
      "tokens_seen": 2494496768
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.930187770823303e-05,
      "loss": 2.6154,
      "theoretical_loss": 3.3690984650779807,
      "tokens_seen": 2494627840
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9293853314074786e-05,
      "loss": 2.4216,
      "theoretical_loss": 3.3690843393552274,
      "tokens_seen": 2494758912
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9285828919916546e-05,
      "loss": 2.5462,
      "theoretical_loss": 3.3690702145823956,
      "tokens_seen": 2494889984
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9277804525758306e-05,
      "loss": 2.5539,
      "theoretical_loss": 3.3690560907593707,
      "tokens_seen": 2495021056
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9269780131600066e-05,
      "loss": 2.6297,
      "theoretical_loss": 3.369041967886039,
      "tokens_seen": 2495152128
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9261755737441826e-05,
      "loss": 2.3965,
      "theoretical_loss": 3.3690278459622864,
      "tokens_seen": 2495283200
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1369331,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3660659790039062,
      "objective/train/theoretical_loss": 3.369013724988,
      "objective/train/tokens_used": 865873376,
      "theoretical_loss": 3.369013724988,
      "tokens_seen": 2495414272
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9253731343283586e-05,
      "loss": 2.6804,
      "theoretical_loss": 3.369013724988,
      "tokens_seen": 2495414272
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.924570694912534e-05,
      "loss": 2.5133,
      "theoretical_loss": 3.368999604963066,
      "tokens_seen": 2495545344
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.92376825549671e-05,
      "loss": 2.7273,
      "theoretical_loss": 3.36898548588737,
      "tokens_seen": 2495676416
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.922965816080886e-05,
      "loss": 2.5223,
      "theoretical_loss": 3.3689713677607984,
      "tokens_seen": 2495807488
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.922163376665062e-05,
      "loss": 2.6208,
      "theoretical_loss": 3.368957250583238,
      "tokens_seen": 2495938560
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.921360937249238e-05,
      "loss": 2.5849,
      "theoretical_loss": 3.3689431343545753,
      "tokens_seen": 2496069632
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.920558497833414e-05,
      "loss": 2.4412,
      "theoretical_loss": 3.3689290190746966,
      "tokens_seen": 2496200704
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.919756058417589e-05,
      "loss": 2.5758,
      "theoretical_loss": 3.368914904743488,
      "tokens_seen": 2496331776
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.918953619001765e-05,
      "loss": 2.4951,
      "theoretical_loss": 3.368900791360836,
      "tokens_seen": 2496462848
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.918151179585941e-05,
      "loss": 2.608,
      "theoretical_loss": 3.368886678926627,
      "tokens_seen": 2496593920
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.917348740170117e-05,
      "loss": 2.4136,
      "theoretical_loss": 3.3688725674407483,
      "tokens_seen": 2496724992
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.916546300754293e-05,
      "loss": 2.5014,
      "theoretical_loss": 3.3688584569030855,
      "tokens_seen": 2496856064
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.915743861338469e-05,
      "loss": 2.3839,
      "theoretical_loss": 3.368844347313525,
      "tokens_seen": 2496987136
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1369989,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.0298242568969727,
      "objective/train/theoretical_loss": 3.3688372928742476,
      "objective/train/tokens_used": 867511776,
      "theoretical_loss": 3.3688372928742476,
      "tokens_seen": 2497052672
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.914941421922645e-05,
      "loss": 2.5819,
      "theoretical_loss": 3.3688302386719537,
      "tokens_seen": 2497118208
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9141389825068204e-05,
      "loss": 2.5843,
      "theoretical_loss": 3.3688161309782583,
      "tokens_seen": 2497249280
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9133365430909964e-05,
      "loss": 2.507,
      "theoretical_loss": 3.368802024232325,
      "tokens_seen": 2497380352
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9125341036751724e-05,
      "loss": 2.5364,
      "theoretical_loss": 3.368787918434041,
      "tokens_seen": 2497511424
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9117316642593484e-05,
      "loss": 2.3699,
      "theoretical_loss": 3.3687738135832923,
      "tokens_seen": 2497642496
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9109292248435244e-05,
      "loss": 2.4503,
      "theoretical_loss": 3.368759709679966,
      "tokens_seen": 2497773568
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9101267854277003e-05,
      "loss": 2.5373,
      "theoretical_loss": 3.368745606723948,
      "tokens_seen": 2497904640
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.909324346011876e-05,
      "loss": 2.7115,
      "theoretical_loss": 3.368731504715126,
      "tokens_seen": 2498035712
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.908521906596052e-05,
      "loss": 2.6725,
      "theoretical_loss": 3.3687174036533856,
      "tokens_seen": 2498166784
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9077194671802276e-05,
      "loss": 2.5665,
      "theoretical_loss": 3.3687033035386147,
      "tokens_seen": 2498297856
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9069170277644036e-05,
      "loss": 2.4985,
      "theoretical_loss": 3.368689204370699,
      "tokens_seen": 2498428928
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9061145883485796e-05,
      "loss": 2.5538,
      "theoretical_loss": 3.3686751061495253,
      "tokens_seen": 2498560000
    },
    {
      "epoch": 0.51,
      "objective/train/docs_used": 1371286,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.572655200958252,
      "objective/train/theoretical_loss": 3.368661008874981,
      "objective/train/tokens_used": 869150176,
      "theoretical_loss": 3.368661008874981,
      "tokens_seen": 2498691072
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9053121489327556e-05,
      "loss": 2.5129,
      "theoretical_loss": 3.368661008874981,
      "tokens_seen": 2498691072
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9045097095169316e-05,
      "loss": 2.5709,
      "theoretical_loss": 3.3686469125469527,
      "tokens_seen": 2498822144
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9037072701011076e-05,
      "loss": 2.6015,
      "theoretical_loss": 3.3686328171653273,
      "tokens_seen": 2498953216
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.9029048306852836e-05,
      "loss": 2.4381,
      "theoretical_loss": 3.368618722729991,
      "tokens_seen": 2499084288
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.902102391269459e-05,
      "loss": 2.3638,
      "theoretical_loss": 3.3686046292408314,
      "tokens_seen": 2499215360
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.901299951853635e-05,
      "loss": 2.5115,
      "theoretical_loss": 3.368590536697735,
      "tokens_seen": 2499346432
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.900497512437811e-05,
      "loss": 2.5566,
      "theoretical_loss": 3.3685764451005884,
      "tokens_seen": 2499477504
    },
    {
      "epoch": 0.51,
      "learning_rate": 4.899695073021987e-05,
      "loss": 2.5392,
      "theoretical_loss": 3.368562354449279,
      "tokens_seen": 2499608576
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.898892633606163e-05,
      "loss": 2.4501,
      "theoretical_loss": 3.3685482647436937,
      "tokens_seen": 2499739648
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.898090194190339e-05,
      "loss": 2.6578,
      "theoretical_loss": 3.368534175983719,
      "tokens_seen": 2499870720
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.897287754774515e-05,
      "loss": 2.5015,
      "theoretical_loss": 3.3685200881692423,
      "tokens_seen": 2500001792
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.896485315358691e-05,
      "loss": 2.5423,
      "theoretical_loss": 3.36850600130015,
      "tokens_seen": 2500132864
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.895682875942866e-05,
      "loss": 2.6089,
      "theoretical_loss": 3.3684919153763304,
      "tokens_seen": 2500263936
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1371802,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.660773277282715,
      "objective/train/theoretical_loss": 3.368484872768862,
      "objective/train/tokens_used": 870788576,
      "theoretical_loss": 3.368484872768862,
      "tokens_seen": 2500329472
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.894880436527042e-05,
      "loss": 2.5976,
      "theoretical_loss": 3.3684778303976692,
      "tokens_seen": 2500395008
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.894077997111218e-05,
      "loss": 2.5346,
      "theoretical_loss": 3.3684637463640534,
      "tokens_seen": 2500526080
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.893275557695394e-05,
      "loss": 2.532,
      "theoretical_loss": 3.3684496632753715,
      "tokens_seen": 2500657152
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.89247311827957e-05,
      "loss": 2.6926,
      "theoretical_loss": 3.368435581131509,
      "tokens_seen": 2500788224
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.891670678863746e-05,
      "loss": 2.7204,
      "theoretical_loss": 3.368421499932354,
      "tokens_seen": 2500919296
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.890868239447922e-05,
      "loss": 2.5059,
      "theoretical_loss": 3.3684074196777924,
      "tokens_seen": 2501050368
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8900658000320974e-05,
      "loss": 2.6287,
      "theoretical_loss": 3.368393340367713,
      "tokens_seen": 2501181440
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8892633606162734e-05,
      "loss": 2.6112,
      "theoretical_loss": 3.368379262002002,
      "tokens_seen": 2501312512
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8884609212004494e-05,
      "loss": 2.6782,
      "theoretical_loss": 3.3683651845805467,
      "tokens_seen": 2501443584
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8876584817846254e-05,
      "loss": 2.6049,
      "theoretical_loss": 3.368351108103234,
      "tokens_seen": 2501574656
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8868560423688014e-05,
      "loss": 2.5635,
      "theoretical_loss": 3.3683370325699515,
      "tokens_seen": 2501705728
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8860536029529774e-05,
      "loss": 2.571,
      "theoretical_loss": 3.3683229579805865,
      "tokens_seen": 2501836800
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1373004,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.812018871307373,
      "objective/train/theoretical_loss": 3.3683088843350264,
      "objective/train/tokens_used": 872426976,
      "theoretical_loss": 3.3683088843350264,
      "tokens_seen": 2501967872
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8852511635371533e-05,
      "loss": 2.5624,
      "theoretical_loss": 3.3683088843350264,
      "tokens_seen": 2501967872
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8844487241213287e-05,
      "loss": 2.5664,
      "theoretical_loss": 3.3682948116331577,
      "tokens_seen": 2502098944
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8836462847055046e-05,
      "loss": 2.5155,
      "theoretical_loss": 3.3682807398748684,
      "tokens_seen": 2502230016
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8828438452896806e-05,
      "loss": 2.6867,
      "theoretical_loss": 3.3682666690600453,
      "tokens_seen": 2502361088
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8820414058738566e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.368252599188576,
      "tokens_seen": 2502492160
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8812389664580326e-05,
      "loss": 2.638,
      "theoretical_loss": 3.3682385302603484,
      "tokens_seen": 2502623232
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8804365270422086e-05,
      "loss": 2.5934,
      "theoretical_loss": 3.368224462275249,
      "tokens_seen": 2502754304
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8796340876263846e-05,
      "loss": 2.518,
      "theoretical_loss": 3.3682103952331652,
      "tokens_seen": 2502885376
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8788316482105606e-05,
      "loss": 2.4324,
      "theoretical_loss": 3.368196329133985,
      "tokens_seen": 2503016448
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.878029208794736e-05,
      "loss": 2.5402,
      "theoretical_loss": 3.368182263977596,
      "tokens_seen": 2503147520
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.877226769378912e-05,
      "loss": 2.557,
      "theoretical_loss": 3.3681681997638844,
      "tokens_seen": 2503278592
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.876424329963088e-05,
      "loss": 2.4608,
      "theoretical_loss": 3.368154136492739,
      "tokens_seen": 2503409664
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.875621890547264e-05,
      "loss": 2.6075,
      "theoretical_loss": 3.368140074164047,
      "tokens_seen": 2503540736
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1374296,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.103992462158203,
      "objective/train/theoretical_loss": 3.3681330433530854,
      "objective/train/tokens_used": 874065376,
      "theoretical_loss": 3.3681330433530854,
      "tokens_seen": 2503606272
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.87481945113144e-05,
      "loss": 2.4213,
      "theoretical_loss": 3.3681260127776946,
      "tokens_seen": 2503671808
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.874017011715616e-05,
      "loss": 2.4916,
      "theoretical_loss": 3.3681119523335714,
      "tokens_seen": 2503802880
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.873214572299792e-05,
      "loss": 2.5939,
      "theoretical_loss": 3.3680978928315635,
      "tokens_seen": 2503933952
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.872412132883967e-05,
      "loss": 2.573,
      "theoretical_loss": 3.368083834271559,
      "tokens_seen": 2504065024
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.871609693468143e-05,
      "loss": 2.6522,
      "theoretical_loss": 3.3680697766534453,
      "tokens_seen": 2504196096
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.870807254052319e-05,
      "loss": 2.5498,
      "theoretical_loss": 3.36805571997711,
      "tokens_seen": 2504327168
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.870004814636495e-05,
      "loss": 2.4721,
      "theoretical_loss": 3.368041664242441,
      "tokens_seen": 2504458240
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.869202375220671e-05,
      "loss": 2.4782,
      "theoretical_loss": 3.3680276094493253,
      "tokens_seen": 2504589312
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.868399935804847e-05,
      "loss": 2.6176,
      "theoretical_loss": 3.3680135555976514,
      "tokens_seen": 2504720384
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.867597496389023e-05,
      "loss": 2.464,
      "theoretical_loss": 3.367999502687306,
      "tokens_seen": 2504851456
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8667950569731984e-05,
      "loss": 2.5257,
      "theoretical_loss": 3.367985450718178,
      "tokens_seen": 2504982528
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8659926175573744e-05,
      "loss": 2.6085,
      "theoretical_loss": 3.367971399690154,
      "tokens_seen": 2505113600
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1374892,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3439693450927734,
      "objective/train/theoretical_loss": 3.367957349603123,
      "objective/train/tokens_used": 875703776,
      "theoretical_loss": 3.367957349603123,
      "tokens_seen": 2505244672
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8651901781415504e-05,
      "loss": 2.5946,
      "theoretical_loss": 3.367957349603123,
      "tokens_seen": 2505244672
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8643877387257264e-05,
      "loss": 2.6487,
      "theoretical_loss": 3.367943300456971,
      "tokens_seen": 2505375744
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8635852993099024e-05,
      "loss": 2.5105,
      "theoretical_loss": 3.3679292522515873,
      "tokens_seen": 2505506816
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8627828598940784e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.367915204986859,
      "tokens_seen": 2505637888
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8619804204782544e-05,
      "loss": 2.5829,
      "theoretical_loss": 3.367901158662674,
      "tokens_seen": 2505768960
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8611779810624304e-05,
      "loss": 2.3615,
      "theoretical_loss": 3.36788711327892,
      "tokens_seen": 2505900032
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.860375541646606e-05,
      "loss": 2.5354,
      "theoretical_loss": 3.367873068835485,
      "tokens_seen": 2506031104
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8595731022307817e-05,
      "loss": 2.4848,
      "theoretical_loss": 3.367859025332258,
      "tokens_seen": 2506162176
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8587706628149576e-05,
      "loss": 2.5954,
      "theoretical_loss": 3.3678449827691246,
      "tokens_seen": 2506293248
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8579682233991336e-05,
      "loss": 2.5928,
      "theoretical_loss": 3.367830941145974,
      "tokens_seen": 2506424320
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8571657839833096e-05,
      "loss": 2.5398,
      "theoretical_loss": 3.3678169004626946,
      "tokens_seen": 2506555392
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8563633445674856e-05,
      "loss": 2.5305,
      "theoretical_loss": 3.3678028607191735,
      "tokens_seen": 2506686464
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8555609051516616e-05,
      "loss": 2.6508,
      "theoretical_loss": 3.3677888219152985,
      "tokens_seen": 2506817536
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1376192,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 3.142772674560547,
      "objective/train/theoretical_loss": 3.367781802865694,
      "objective/train/tokens_used": 877342176,
      "theoretical_loss": 3.367781802865694,
      "tokens_seen": 2506883072
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.854758465735837e-05,
      "loss": 2.5421,
      "theoretical_loss": 3.3677747840509586,
      "tokens_seen": 2506948608
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.853956026320013e-05,
      "loss": 2.517,
      "theoretical_loss": 3.367760747126041,
      "tokens_seen": 2507079680
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.853153586904189e-05,
      "loss": 2.48,
      "theoretical_loss": 3.367746711140434,
      "tokens_seen": 2507210752
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.852351147488365e-05,
      "loss": 2.5435,
      "theoretical_loss": 3.367732676094026,
      "tokens_seen": 2507341824
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.851548708072541e-05,
      "loss": 2.43,
      "theoretical_loss": 3.3677186419867042,
      "tokens_seen": 2507472896
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.850746268656717e-05,
      "loss": 2.4617,
      "theoretical_loss": 3.3677046088183573,
      "tokens_seen": 2507603968
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.849943829240893e-05,
      "loss": 2.5015,
      "theoretical_loss": 3.367690576588873,
      "tokens_seen": 2507735040
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.849141389825069e-05,
      "loss": 2.4922,
      "theoretical_loss": 3.36767654529814,
      "tokens_seen": 2507866112
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.848338950409244e-05,
      "loss": 2.5681,
      "theoretical_loss": 3.367662514946046,
      "tokens_seen": 2507997184
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.84753651099342e-05,
      "loss": 2.6439,
      "theoretical_loss": 3.367648485532479,
      "tokens_seen": 2508128256
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.846734071577596e-05,
      "loss": 2.6526,
      "theoretical_loss": 3.367634457057328,
      "tokens_seen": 2508259328
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.845931632161772e-05,
      "loss": 2.5571,
      "theoretical_loss": 3.36762042952048,
      "tokens_seen": 2508390400
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1376516,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5870392322540283,
      "objective/train/theoretical_loss": 3.3676064029218242,
      "objective/train/tokens_used": 878980576,
      "theoretical_loss": 3.3676064029218242,
      "tokens_seen": 2508521472
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.845129192745948e-05,
      "loss": 2.6055,
      "theoretical_loss": 3.3676064029218242,
      "tokens_seen": 2508521472
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.844326753330124e-05,
      "loss": 2.5119,
      "theoretical_loss": 3.3675923772612486,
      "tokens_seen": 2508652544
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8435243139143e-05,
      "loss": 2.5264,
      "theoretical_loss": 3.3675783525386414,
      "tokens_seen": 2508783616
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8427218744984754e-05,
      "loss": 2.5302,
      "theoretical_loss": 3.3675643287538906,
      "tokens_seen": 2508914688
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8419194350826514e-05,
      "loss": 2.72,
      "theoretical_loss": 3.3675503059068848,
      "tokens_seen": 2509045760
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8411169956668274e-05,
      "loss": 2.4637,
      "theoretical_loss": 3.367536283997512,
      "tokens_seen": 2509176832
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8403145562510034e-05,
      "loss": 2.571,
      "theoretical_loss": 3.367522263025661,
      "tokens_seen": 2509307904
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8395121168351794e-05,
      "loss": 2.4807,
      "theoretical_loss": 3.36750824299122,
      "tokens_seen": 2509438976
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8387096774193554e-05,
      "loss": 2.5401,
      "theoretical_loss": 3.3674942238940773,
      "tokens_seen": 2509570048
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8379072380035314e-05,
      "loss": 2.4567,
      "theoretical_loss": 3.367480205734121,
      "tokens_seen": 2509701120
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.837104798587707e-05,
      "loss": 2.6387,
      "theoretical_loss": 3.36746618851124,
      "tokens_seen": 2509832192
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.836302359171883e-05,
      "loss": 2.6431,
      "theoretical_loss": 3.367452172225322,
      "tokens_seen": 2509963264
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.835499919756059e-05,
      "loss": 2.6223,
      "theoretical_loss": 3.367438156876257,
      "tokens_seen": 2510094336
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1376516,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8122689723968506,
      "objective/train/theoretical_loss": 3.3674311495530085,
      "objective/train/tokens_used": 880618976,
      "theoretical_loss": 3.3674311495530085,
      "tokens_seen": 2510159872
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8346974803402347e-05,
      "loss": 2.5765,
      "theoretical_loss": 3.3674241424639315,
      "tokens_seen": 2510225408
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8338950409244106e-05,
      "loss": 2.6367,
      "theoretical_loss": 3.367410128988235,
      "tokens_seen": 2510356480
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8330926015085866e-05,
      "loss": 2.6433,
      "theoretical_loss": 3.367396116449056,
      "tokens_seen": 2510487552
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8322901620927626e-05,
      "loss": 2.6982,
      "theoretical_loss": 3.3673821048462833,
      "tokens_seen": 2510618624
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8314877226769386e-05,
      "loss": 2.7463,
      "theoretical_loss": 3.3673680941798048,
      "tokens_seen": 2510749696
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.830685283261114e-05,
      "loss": 2.6165,
      "theoretical_loss": 3.3673540844495093,
      "tokens_seen": 2510880768
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.82988284384529e-05,
      "loss": 2.7569,
      "theoretical_loss": 3.367340075655285,
      "tokens_seen": 2511011840
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.829080404429466e-05,
      "loss": 2.5475,
      "theoretical_loss": 3.3673260677970216,
      "tokens_seen": 2511142912
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.828277965013642e-05,
      "loss": 2.746,
      "theoretical_loss": 3.3673120608746068,
      "tokens_seen": 2511273984
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.827475525597818e-05,
      "loss": 2.7353,
      "theoretical_loss": 3.367298054887929,
      "tokens_seen": 2511405056
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.826673086181994e-05,
      "loss": 2.8524,
      "theoretical_loss": 3.367284049836878,
      "tokens_seen": 2511536128
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.82587064676617e-05,
      "loss": 2.5463,
      "theoretical_loss": 3.3672700457213414,
      "tokens_seen": 2511667200
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1377425,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6701934337615967,
      "objective/train/theoretical_loss": 3.367256042541208,
      "objective/train/tokens_used": 882257376,
      "theoretical_loss": 3.367256042541208,
      "tokens_seen": 2511798272
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.825068207350345e-05,
      "loss": 2.6103,
      "theoretical_loss": 3.367256042541208,
      "tokens_seen": 2511798272
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.824265767934521e-05,
      "loss": 2.4983,
      "theoretical_loss": 3.3672420402963676,
      "tokens_seen": 2511929344
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.823463328518697e-05,
      "loss": 2.7509,
      "theoretical_loss": 3.3672280389867075,
      "tokens_seen": 2512060416
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.822660889102873e-05,
      "loss": 2.8019,
      "theoretical_loss": 3.3672140386121177,
      "tokens_seen": 2512191488
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.821858449687049e-05,
      "loss": 2.614,
      "theoretical_loss": 3.3672000391724857,
      "tokens_seen": 2512322560
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.821056010271225e-05,
      "loss": 2.6892,
      "theoretical_loss": 3.367186040667701,
      "tokens_seen": 2512453632
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.820253570855401e-05,
      "loss": 2.5277,
      "theoretical_loss": 3.3671720430976526,
      "tokens_seen": 2512584704
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8194511314395764e-05,
      "loss": 2.757,
      "theoretical_loss": 3.367158046462229,
      "tokens_seen": 2512715776
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8186486920237524e-05,
      "loss": 2.6371,
      "theoretical_loss": 3.367144050761319,
      "tokens_seen": 2512846848
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8178462526079284e-05,
      "loss": 2.6286,
      "theoretical_loss": 3.367130055994812,
      "tokens_seen": 2512977920
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8170438131921044e-05,
      "loss": 2.6977,
      "theoretical_loss": 3.3671160621625957,
      "tokens_seen": 2513108992
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8162413737762804e-05,
      "loss": 2.5703,
      "theoretical_loss": 3.3671020692645603,
      "tokens_seen": 2513240064
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8154389343604564e-05,
      "loss": 2.4807,
      "theoretical_loss": 3.3670880773005942,
      "tokens_seen": 2513371136
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1378070,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3798375129699707,
      "objective/train/theoretical_loss": 3.3670810816688523,
      "objective/train/tokens_used": 883895776,
      "theoretical_loss": 3.3670810816688523,
      "tokens_seen": 2513436672
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8146364949446324e-05,
      "loss": 2.6042,
      "theoretical_loss": 3.367074086270586,
      "tokens_seen": 2513502208
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8138340555288084e-05,
      "loss": 2.4991,
      "theoretical_loss": 3.3670600961744253,
      "tokens_seen": 2513633280
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.813031616112984e-05,
      "loss": 2.6307,
      "theoretical_loss": 3.367046107012001,
      "tokens_seen": 2513764352
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.81222917669716e-05,
      "loss": 2.5312,
      "theoretical_loss": 3.3670321187832015,
      "tokens_seen": 2513895424
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.811426737281336e-05,
      "loss": 2.5389,
      "theoretical_loss": 3.3670181314879164,
      "tokens_seen": 2514026496
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8106242978655117e-05,
      "loss": 2.6553,
      "theoretical_loss": 3.3670041451260344,
      "tokens_seen": 2514157568
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8098218584496877e-05,
      "loss": 2.5299,
      "theoretical_loss": 3.3669901596974445,
      "tokens_seen": 2514288640
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8090194190338636e-05,
      "loss": 2.5993,
      "theoretical_loss": 3.366976175202036,
      "tokens_seen": 2514419712
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.8082169796180396e-05,
      "loss": 2.5919,
      "theoretical_loss": 3.3669621916396983,
      "tokens_seen": 2514550784
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.807414540202215e-05,
      "loss": 2.6763,
      "theoretical_loss": 3.36694820901032,
      "tokens_seen": 2514681856
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.806612100786391e-05,
      "loss": 2.5359,
      "theoretical_loss": 3.36693422731379,
      "tokens_seen": 2514812928
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.805809661370567e-05,
      "loss": 2.6382,
      "theoretical_loss": 3.3669202465499986,
      "tokens_seen": 2514944000
    },
    {
      "epoch": 0.52,
      "objective/train/docs_used": 1379318,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.843836545944214,
      "objective/train/theoretical_loss": 3.366906266718834,
      "objective/train/tokens_used": 885534176,
      "theoretical_loss": 3.366906266718834,
      "tokens_seen": 2515075072
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.805007221954743e-05,
      "loss": 2.7247,
      "theoretical_loss": 3.366906266718834,
      "tokens_seen": 2515075072
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.804204782538919e-05,
      "loss": 2.6126,
      "theoretical_loss": 3.366892287820185,
      "tokens_seen": 2515206144
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.803402343123095e-05,
      "loss": 2.5857,
      "theoretical_loss": 3.366878309853942,
      "tokens_seen": 2515337216
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.80259990370727e-05,
      "loss": 2.6734,
      "theoretical_loss": 3.3668643328199934,
      "tokens_seen": 2515468288
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.801797464291446e-05,
      "loss": 2.6715,
      "theoretical_loss": 3.3668503567182286,
      "tokens_seen": 2515599360
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.800995024875622e-05,
      "loss": 2.7025,
      "theoretical_loss": 3.366836381548537,
      "tokens_seen": 2515730432
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.800192585459798e-05,
      "loss": 2.6618,
      "theoretical_loss": 3.3668224073108077,
      "tokens_seen": 2515861504
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.799390146043974e-05,
      "loss": 2.6227,
      "theoretical_loss": 3.3668084340049305,
      "tokens_seen": 2515992576
    },
    {
      "epoch": 0.52,
      "learning_rate": 4.79858770662815e-05,
      "loss": 2.6876,
      "theoretical_loss": 3.366794461630794,
      "tokens_seen": 2516123648
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7977852672123255e-05,
      "loss": 2.6487,
      "theoretical_loss": 3.366780490188288,
      "tokens_seen": 2516254720
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7969828277965015e-05,
      "loss": 2.6247,
      "theoretical_loss": 3.366766519677302,
      "tokens_seen": 2516385792
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7961803883806775e-05,
      "loss": 2.6998,
      "theoretical_loss": 3.3667525500977247,
      "tokens_seen": 2516516864
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7953779489648534e-05,
      "loss": 2.6981,
      "theoretical_loss": 3.366738581449446,
      "tokens_seen": 2516647936
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1379956,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.391172409057617,
      "objective/train/theoretical_loss": 3.3667315974745096,
      "objective/train/tokens_used": 887172576,
      "theoretical_loss": 3.3667315974745096,
      "tokens_seen": 2516713472
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7945755095490294e-05,
      "loss": 2.4727,
      "theoretical_loss": 3.3667246137323557,
      "tokens_seen": 2516779008
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7937730701332054e-05,
      "loss": 2.6718,
      "theoretical_loss": 3.366710646946342,
      "tokens_seen": 2516910080
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.792970630717381e-05,
      "loss": 2.679,
      "theoretical_loss": 3.366696681091296,
      "tokens_seen": 2517041152
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.792168191301557e-05,
      "loss": 2.6756,
      "theoretical_loss": 3.3666827161671056,
      "tokens_seen": 2517172224
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.791365751885733e-05,
      "loss": 2.6579,
      "theoretical_loss": 3.366668752173661,
      "tokens_seen": 2517303296
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.790563312469909e-05,
      "loss": 2.7533,
      "theoretical_loss": 3.366654789110852,
      "tokens_seen": 2517434368
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.789760873054085e-05,
      "loss": 2.4818,
      "theoretical_loss": 3.3666408269785677,
      "tokens_seen": 2517565440
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.78895843363826e-05,
      "loss": 2.4124,
      "theoretical_loss": 3.3666268657766976,
      "tokens_seen": 2517696512
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.788155994222436e-05,
      "loss": 2.6119,
      "theoretical_loss": 3.3666129055051317,
      "tokens_seen": 2517827584
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.787353554806612e-05,
      "loss": 2.576,
      "theoretical_loss": 3.366598946163759,
      "tokens_seen": 2517958656
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.786551115390788e-05,
      "loss": 2.5626,
      "theoretical_loss": 3.3665849877524696,
      "tokens_seen": 2518089728
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.785748675974964e-05,
      "loss": 2.6858,
      "theoretical_loss": 3.366571030271153,
      "tokens_seen": 2518220800
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1381079,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.684238910675049,
      "objective/train/theoretical_loss": 3.366557073719699,
      "objective/train/tokens_used": 888810976,
      "theoretical_loss": 3.366557073719699,
      "tokens_seen": 2518351872
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.78494623655914e-05,
      "loss": 2.6355,
      "theoretical_loss": 3.366557073719699,
      "tokens_seen": 2518351872
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.784143797143316e-05,
      "loss": 2.6237,
      "theoretical_loss": 3.366543118097997,
      "tokens_seen": 2518482944
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.783341357727491e-05,
      "loss": 2.4905,
      "theoretical_loss": 3.366529163405936,
      "tokens_seen": 2518614016
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.782538918311667e-05,
      "loss": 2.5384,
      "theoretical_loss": 3.3665152096434072,
      "tokens_seen": 2518745088
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.781736478895843e-05,
      "loss": 2.5284,
      "theoretical_loss": 3.366501256810299,
      "tokens_seen": 2518876160
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.780934039480019e-05,
      "loss": 2.7665,
      "theoretical_loss": 3.366487304906502,
      "tokens_seen": 2519007232
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.780131600064195e-05,
      "loss": 2.3966,
      "theoretical_loss": 3.366473353931905,
      "tokens_seen": 2519138304
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.779329160648371e-05,
      "loss": 2.5817,
      "theoretical_loss": 3.3664594038863993,
      "tokens_seen": 2519269376
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.778526721232547e-05,
      "loss": 2.8254,
      "theoretical_loss": 3.3664454547698734,
      "tokens_seen": 2519400448
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7777242818167225e-05,
      "loss": 2.492,
      "theoretical_loss": 3.3664315065822175,
      "tokens_seen": 2519531520
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7769218424008985e-05,
      "loss": 2.5963,
      "theoretical_loss": 3.3664175593233217,
      "tokens_seen": 2519662592
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7761194029850745e-05,
      "loss": 2.6447,
      "theoretical_loss": 3.366403612993075,
      "tokens_seen": 2519793664
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7753169635692505e-05,
      "loss": 2.6689,
      "theoretical_loss": 3.3663896675913683,
      "tokens_seen": 2519924736
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1381616,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9621894359588623,
      "objective/train/theoretical_loss": 3.366382695238683,
      "objective/train/tokens_used": 890449376,
      "theoretical_loss": 3.366382695238683,
      "tokens_seen": 2519990272
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7745145241534265e-05,
      "loss": 2.556,
      "theoretical_loss": 3.366375723118091,
      "tokens_seen": 2520055808
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7737120847376025e-05,
      "loss": 2.6218,
      "theoretical_loss": 3.3663617795731327,
      "tokens_seen": 2520186880
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7729096453217785e-05,
      "loss": 2.559,
      "theoretical_loss": 3.366347836956384,
      "tokens_seen": 2520317952
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.772107205905954e-05,
      "loss": 2.5452,
      "theoretical_loss": 3.3663338952677346,
      "tokens_seen": 2520449024
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.77130476649013e-05,
      "loss": 2.5247,
      "theoretical_loss": 3.366319954507074,
      "tokens_seen": 2520580096
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.770502327074306e-05,
      "loss": 2.5345,
      "theoretical_loss": 3.3663060146742927,
      "tokens_seen": 2520711168
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.769699887658482e-05,
      "loss": 2.4603,
      "theoretical_loss": 3.3662920757692802,
      "tokens_seen": 2520842240
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.768897448242658e-05,
      "loss": 2.7226,
      "theoretical_loss": 3.3662781377919275,
      "tokens_seen": 2520973312
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.768095008826834e-05,
      "loss": 2.5504,
      "theoretical_loss": 3.3662642007421235,
      "tokens_seen": 2521104384
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.76729256941101e-05,
      "loss": 2.5523,
      "theoretical_loss": 3.366250264619759,
      "tokens_seen": 2521235456
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.766490129995186e-05,
      "loss": 2.5654,
      "theoretical_loss": 3.3662363294247237,
      "tokens_seen": 2521366528
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.765687690579361e-05,
      "loss": 2.6242,
      "theoretical_loss": 3.3662223951569077,
      "tokens_seen": 2521497600
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1382692,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.633779764175415,
      "objective/train/theoretical_loss": 3.3662084618162016,
      "objective/train/tokens_used": 892087776,
      "theoretical_loss": 3.3662084618162016,
      "tokens_seen": 2521628672
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.764885251163537e-05,
      "loss": 2.4958,
      "theoretical_loss": 3.3662084618162016,
      "tokens_seen": 2521628672
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.764082811747713e-05,
      "loss": 2.5144,
      "theoretical_loss": 3.366194529402495,
      "tokens_seen": 2521759744
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.763280372331889e-05,
      "loss": 2.5313,
      "theoretical_loss": 3.366180597915678,
      "tokens_seen": 2521890816
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.762477932916065e-05,
      "loss": 2.7044,
      "theoretical_loss": 3.366166667355641,
      "tokens_seen": 2522021888
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.761675493500241e-05,
      "loss": 2.6387,
      "theoretical_loss": 3.366152737722274,
      "tokens_seen": 2522152960
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.760873054084417e-05,
      "loss": 2.6255,
      "theoretical_loss": 3.366138809015468,
      "tokens_seen": 2522284032
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.760070614668592e-05,
      "loss": 2.6242,
      "theoretical_loss": 3.366124881235112,
      "tokens_seen": 2522415104
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.759268175252768e-05,
      "loss": 2.6112,
      "theoretical_loss": 3.3661109543810968,
      "tokens_seen": 2522546176
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.758465735836944e-05,
      "loss": 2.7586,
      "theoretical_loss": 3.3660970284533134,
      "tokens_seen": 2522677248
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.75766329642112e-05,
      "loss": 2.604,
      "theoretical_loss": 3.3660831034516505,
      "tokens_seen": 2522808320
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.756860857005296e-05,
      "loss": 2.5642,
      "theoretical_loss": 3.366069179376,
      "tokens_seen": 2522939392
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.756058417589472e-05,
      "loss": 2.5398,
      "theoretical_loss": 3.366055256226251,
      "tokens_seen": 2523070464
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.755255978173648e-05,
      "loss": 2.6155,
      "theoretical_loss": 3.3660413340022943,
      "tokens_seen": 2523201536
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1383183,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.505521297454834,
      "objective/train/theoretical_loss": 3.366034373237454,
      "objective/train/tokens_used": 893726176,
      "theoretical_loss": 3.366034373237454,
      "tokens_seen": 2523267072
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.754453538757824e-05,
      "loss": 2.6113,
      "theoretical_loss": 3.36602741270402,
      "tokens_seen": 2523332608
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7536510993419995e-05,
      "loss": 2.7356,
      "theoretical_loss": 3.3660134923313194,
      "tokens_seen": 2523463680
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7528486599261755e-05,
      "loss": 2.5268,
      "theoretical_loss": 3.365999572884082,
      "tokens_seen": 2523594752
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7520462205103515e-05,
      "loss": 2.7341,
      "theoretical_loss": 3.3659856543621984,
      "tokens_seen": 2523725824
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7512437810945275e-05,
      "loss": 2.6552,
      "theoretical_loss": 3.3659717367655593,
      "tokens_seen": 2523856896
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7504413416787035e-05,
      "loss": 2.5299,
      "theoretical_loss": 3.3659578200940548,
      "tokens_seen": 2523987968
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7496389022628795e-05,
      "loss": 2.5459,
      "theoretical_loss": 3.365943904347575,
      "tokens_seen": 2524119040
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7488364628470555e-05,
      "loss": 2.5231,
      "theoretical_loss": 3.3659299895260113,
      "tokens_seen": 2524250112
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.748034023431231e-05,
      "loss": 2.6234,
      "theoretical_loss": 3.365916075629254,
      "tokens_seen": 2524381184
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.747231584015407e-05,
      "loss": 2.544,
      "theoretical_loss": 3.3659021626571928,
      "tokens_seen": 2524512256
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.746429144599583e-05,
      "loss": 2.4326,
      "theoretical_loss": 3.3658882506097196,
      "tokens_seen": 2524643328
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.745626705183759e-05,
      "loss": 2.5907,
      "theoretical_loss": 3.365874339486724,
      "tokens_seen": 2524774400
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1384240,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3049447536468506,
      "objective/train/theoretical_loss": 3.3658604292880963,
      "objective/train/tokens_used": 895364576,
      "theoretical_loss": 3.3658604292880963,
      "tokens_seen": 2524905472
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.744824265767935e-05,
      "loss": 2.4003,
      "theoretical_loss": 3.3658604292880963,
      "tokens_seen": 2524905472
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.744021826352111e-05,
      "loss": 2.6488,
      "theoretical_loss": 3.3658465200137275,
      "tokens_seen": 2525036544
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.743219386936287e-05,
      "loss": 2.4416,
      "theoretical_loss": 3.365832611663509,
      "tokens_seen": 2525167616
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.742416947520462e-05,
      "loss": 2.5324,
      "theoretical_loss": 3.36581870423733,
      "tokens_seen": 2525298688
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.741614508104638e-05,
      "loss": 2.6176,
      "theoretical_loss": 3.3658047977350822,
      "tokens_seen": 2525429760
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.740812068688814e-05,
      "loss": 2.5067,
      "theoretical_loss": 3.365790892156656,
      "tokens_seen": 2525560832
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.74000962927299e-05,
      "loss": 2.7227,
      "theoretical_loss": 3.365776987501942,
      "tokens_seen": 2525691904
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.739207189857166e-05,
      "loss": 2.5388,
      "theoretical_loss": 3.365763083770831,
      "tokens_seen": 2525822976
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.738404750441342e-05,
      "loss": 2.5103,
      "theoretical_loss": 3.3657491809632134,
      "tokens_seen": 2525954048
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.737602311025518e-05,
      "loss": 2.5524,
      "theoretical_loss": 3.3657352790789807,
      "tokens_seen": 2526085120
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.736799871609694e-05,
      "loss": 2.6606,
      "theoretical_loss": 3.365721378118023,
      "tokens_seen": 2526216192
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.735997432193869e-05,
      "loss": 2.4384,
      "theoretical_loss": 3.365707478080231,
      "tokens_seen": 2526347264
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.735194992778045e-05,
      "loss": 2.5257,
      "theoretical_loss": 3.3656935789654963,
      "tokens_seen": 2526478336
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1385479,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.296886444091797,
      "objective/train/theoretical_loss": 3.3656866297542405,
      "objective/train/tokens_used": 897002976,
      "theoretical_loss": 3.3656866297542405,
      "tokens_seen": 2526543872
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.734392553362221e-05,
      "loss": 2.6146,
      "theoretical_loss": 3.3656796807737086,
      "tokens_seen": 2526609408
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.733590113946397e-05,
      "loss": 2.6998,
      "theoretical_loss": 3.3656657835047596,
      "tokens_seen": 2526740480
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.732787674530573e-05,
      "loss": 2.6655,
      "theoretical_loss": 3.3656518871585397,
      "tokens_seen": 2526871552
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.731985235114749e-05,
      "loss": 2.4147,
      "theoretical_loss": 3.36563799173494,
      "tokens_seen": 2527002624
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.731182795698925e-05,
      "loss": 2.3529,
      "theoretical_loss": 3.3656240972338516,
      "tokens_seen": 2527133696
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7303803562831006e-05,
      "loss": 2.6953,
      "theoretical_loss": 3.365610203655165,
      "tokens_seen": 2527264768
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7295779168672765e-05,
      "loss": 2.5245,
      "theoretical_loss": 3.365596310998771,
      "tokens_seen": 2527395840
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7287754774514525e-05,
      "loss": 2.4851,
      "theoretical_loss": 3.3655824192645616,
      "tokens_seen": 2527526912
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7279730380356285e-05,
      "loss": 2.5641,
      "theoretical_loss": 3.3655685284524264,
      "tokens_seen": 2527657984
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7271705986198045e-05,
      "loss": 2.6068,
      "theoretical_loss": 3.365554638562257,
      "tokens_seen": 2527789056
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7263681592039805e-05,
      "loss": 2.6198,
      "theoretical_loss": 3.3655407495939444,
      "tokens_seen": 2527920128
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7255657197881565e-05,
      "loss": 2.5761,
      "theoretical_loss": 3.36552686154738,
      "tokens_seen": 2528051200
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1385945,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3946783542633057,
      "objective/train/theoretical_loss": 3.365512974422454,
      "objective/train/tokens_used": 898641376,
      "theoretical_loss": 3.365512974422454,
      "tokens_seen": 2528182272
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.724763280372332e-05,
      "loss": 2.6438,
      "theoretical_loss": 3.365512974422454,
      "tokens_seen": 2528182272
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.723960840956508e-05,
      "loss": 2.593,
      "theoretical_loss": 3.3654990882190585,
      "tokens_seen": 2528313344
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.723158401540684e-05,
      "loss": 2.7042,
      "theoretical_loss": 3.3654852029370836,
      "tokens_seen": 2528444416
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.72235596212486e-05,
      "loss": 2.6143,
      "theoretical_loss": 3.365471318576421,
      "tokens_seen": 2528575488
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.721553522709036e-05,
      "loss": 2.5997,
      "theoretical_loss": 3.3654574351369617,
      "tokens_seen": 2528706560
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.720751083293212e-05,
      "loss": 2.4652,
      "theoretical_loss": 3.365443552618596,
      "tokens_seen": 2528837632
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.719948643877388e-05,
      "loss": 2.5552,
      "theoretical_loss": 3.365429671021217,
      "tokens_seen": 2528968704
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.719146204461564e-05,
      "loss": 2.6855,
      "theoretical_loss": 3.365415790344714,
      "tokens_seen": 2529099776
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.718343765045739e-05,
      "loss": 2.6556,
      "theoretical_loss": 3.365401910588979,
      "tokens_seen": 2529230848
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.717541325629915e-05,
      "loss": 2.4773,
      "theoretical_loss": 3.365388031753903,
      "tokens_seen": 2529361920
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.716738886214091e-05,
      "loss": 2.7238,
      "theoretical_loss": 3.3653741538393778,
      "tokens_seen": 2529492992
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.715936446798267e-05,
      "loss": 2.6008,
      "theoretical_loss": 3.3653602768452937,
      "tokens_seen": 2529624064
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.715134007382443e-05,
      "loss": 2.5201,
      "theoretical_loss": 3.3653464007715423,
      "tokens_seen": 2529755136
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1387323,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5731987953186035,
      "objective/train/theoretical_loss": 3.3653394630797577,
      "objective/train/tokens_used": 900279776,
      "theoretical_loss": 3.3653394630797577,
      "tokens_seen": 2529820672
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.714331567966619e-05,
      "loss": 2.6717,
      "theoretical_loss": 3.3653325256180153,
      "tokens_seen": 2529886208
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.713529128550795e-05,
      "loss": 2.4919,
      "theoretical_loss": 3.3653186513846034,
      "tokens_seen": 2530017280
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.71272668913497e-05,
      "loss": 2.6858,
      "theoretical_loss": 3.3653047780711987,
      "tokens_seen": 2530148352
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.711924249719146e-05,
      "loss": 2.5927,
      "theoretical_loss": 3.365290905677692,
      "tokens_seen": 2530279424
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.711121810303322e-05,
      "loss": 2.4124,
      "theoretical_loss": 3.365277034203974,
      "tokens_seen": 2530410496
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.710319370887498e-05,
      "loss": 2.5957,
      "theoretical_loss": 3.3652631636499377,
      "tokens_seen": 2530541568
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.709516931471674e-05,
      "loss": 2.4652,
      "theoretical_loss": 3.3652492940154732,
      "tokens_seen": 2530672640
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.70871449205585e-05,
      "loss": 2.722,
      "theoretical_loss": 3.3652354253004724,
      "tokens_seen": 2530803712
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.707912052640026e-05,
      "loss": 2.7066,
      "theoretical_loss": 3.3652215575048268,
      "tokens_seen": 2530934784
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.707109613224202e-05,
      "loss": 2.5831,
      "theoretical_loss": 3.365207690628427,
      "tokens_seen": 2531065856
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7063071738083776e-05,
      "loss": 2.5865,
      "theoretical_loss": 3.365193824671166,
      "tokens_seen": 2531196928
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7055047343925536e-05,
      "loss": 2.7067,
      "theoretical_loss": 3.3651799596329344,
      "tokens_seen": 2531328000
    },
    {
      "epoch": 0.53,
      "objective/train/docs_used": 1387695,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5774648189544678,
      "objective/train/theoretical_loss": 3.365166095513623,
      "objective/train/tokens_used": 901918176,
      "theoretical_loss": 3.365166095513623,
      "tokens_seen": 2531459072
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7047022949767295e-05,
      "loss": 2.4986,
      "theoretical_loss": 3.365166095513623,
      "tokens_seen": 2531459072
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7038998555609055e-05,
      "loss": 2.5257,
      "theoretical_loss": 3.3651522323131244,
      "tokens_seen": 2531590144
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7030974161450815e-05,
      "loss": 2.597,
      "theoretical_loss": 3.36513837003133,
      "tokens_seen": 2531721216
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7022949767292575e-05,
      "loss": 2.7001,
      "theoretical_loss": 3.3651245086681314,
      "tokens_seen": 2531852288
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.7014925373134335e-05,
      "loss": 2.5374,
      "theoretical_loss": 3.3651106482234194,
      "tokens_seen": 2531983360
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.700690097897609e-05,
      "loss": 2.5377,
      "theoretical_loss": 3.3650967886970866,
      "tokens_seen": 2532114432
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.699887658481785e-05,
      "loss": 2.5266,
      "theoretical_loss": 3.365082930089024,
      "tokens_seen": 2532245504
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.699085219065961e-05,
      "loss": 2.6136,
      "theoretical_loss": 3.3650690723991232,
      "tokens_seen": 2532376576
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.698282779650137e-05,
      "loss": 2.4857,
      "theoretical_loss": 3.3650552156272764,
      "tokens_seen": 2532507648
    },
    {
      "epoch": 0.53,
      "learning_rate": 4.697480340234313e-05,
      "loss": 2.7073,
      "theoretical_loss": 3.365041359773375,
      "tokens_seen": 2532638720
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.696677900818489e-05,
      "loss": 2.5181,
      "theoretical_loss": 3.36502750483731,
      "tokens_seen": 2532769792
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.695875461402665e-05,
      "loss": 2.4032,
      "theoretical_loss": 3.365013650818974,
      "tokens_seen": 2532900864
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.69507302198684e-05,
      "loss": 2.5306,
      "theoretical_loss": 3.3649997977182586,
      "tokens_seen": 2533031936
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1388584,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8600099086761475,
      "objective/train/theoretical_loss": 3.3649928715119746,
      "objective/train/tokens_used": 903556576,
      "theoretical_loss": 3.3649928715119746,
      "tokens_seen": 2533097472
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.694270582571016e-05,
      "loss": 2.7793,
      "theoretical_loss": 3.3649859455350555,
      "tokens_seen": 2533163008
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.693468143155192e-05,
      "loss": 2.5522,
      "theoretical_loss": 3.3649720942692563,
      "tokens_seen": 2533294080
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.692665703739368e-05,
      "loss": 2.6216,
      "theoretical_loss": 3.3649582439207526,
      "tokens_seen": 2533425152
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.691863264323544e-05,
      "loss": 2.5232,
      "theoretical_loss": 3.364944394489436,
      "tokens_seen": 2533556224
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.69106082490772e-05,
      "loss": 2.5141,
      "theoretical_loss": 3.3649305459751995,
      "tokens_seen": 2533687296
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.690258385491896e-05,
      "loss": 2.6482,
      "theoretical_loss": 3.364916698377934,
      "tokens_seen": 2533818368
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.689455946076072e-05,
      "loss": 2.6283,
      "theoretical_loss": 3.3649028516975314,
      "tokens_seen": 2533949440
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.688653506660247e-05,
      "loss": 2.6347,
      "theoretical_loss": 3.364889005933884,
      "tokens_seen": 2534080512
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.687851067244423e-05,
      "loss": 2.5845,
      "theoretical_loss": 3.364875161086883,
      "tokens_seen": 2534211584
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.687048627828599e-05,
      "loss": 2.5044,
      "theoretical_loss": 3.3648613171564206,
      "tokens_seen": 2534342656
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.686246188412775e-05,
      "loss": 2.5999,
      "theoretical_loss": 3.3648474741423895,
      "tokens_seen": 2534473728
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.685443748996951e-05,
      "loss": 2.4732,
      "theoretical_loss": 3.36483363204468,
      "tokens_seen": 2534604800
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1389704,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.414130449295044,
      "objective/train/theoretical_loss": 3.364819790863186,
      "objective/train/tokens_used": 905194976,
      "theoretical_loss": 3.364819790863186,
      "tokens_seen": 2534735872
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.684641309581127e-05,
      "loss": 2.4403,
      "theoretical_loss": 3.364819790863186,
      "tokens_seen": 2534735872
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.683838870165303e-05,
      "loss": 2.5613,
      "theoretical_loss": 3.364805950597798,
      "tokens_seen": 2534866944
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6830364307494786e-05,
      "loss": 2.6744,
      "theoretical_loss": 3.3647921112484087,
      "tokens_seen": 2534998016
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6822339913336546e-05,
      "loss": 2.6223,
      "theoretical_loss": 3.36477827281491,
      "tokens_seen": 2535129088
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6814315519178306e-05,
      "loss": 2.5623,
      "theoretical_loss": 3.364764435297194,
      "tokens_seen": 2535260160
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6806291125020066e-05,
      "loss": 2.5392,
      "theoretical_loss": 3.364750598695152,
      "tokens_seen": 2535391232
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6798266730861825e-05,
      "loss": 2.4866,
      "theoretical_loss": 3.3647367630086773,
      "tokens_seen": 2535522304
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6790242336703585e-05,
      "loss": 2.5354,
      "theoretical_loss": 3.364722928237661,
      "tokens_seen": 2535653376
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6782217942545345e-05,
      "loss": 2.5069,
      "theoretical_loss": 3.364709094381996,
      "tokens_seen": 2535784448
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.67741935483871e-05,
      "loss": 2.5704,
      "theoretical_loss": 3.364695261441574,
      "tokens_seen": 2535915520
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.676616915422886e-05,
      "loss": 2.5766,
      "theoretical_loss": 3.364681429416287,
      "tokens_seen": 2536046592
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.675814476007062e-05,
      "loss": 2.6808,
      "theoretical_loss": 3.3646675983060272,
      "tokens_seen": 2536177664
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.675012036591238e-05,
      "loss": 2.5677,
      "theoretical_loss": 3.364653768110687,
      "tokens_seen": 2536308736
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1390171,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4923295974731445,
      "objective/train/theoretical_loss": 3.3646468533560787,
      "objective/train/tokens_used": 906833376,
      "theoretical_loss": 3.3646468533560787,
      "tokens_seen": 2536374272
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.674209597175414e-05,
      "loss": 2.5076,
      "theoretical_loss": 3.364639938830159,
      "tokens_seen": 2536439808
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.67340715775959e-05,
      "loss": 2.5555,
      "theoretical_loss": 3.364626110464335,
      "tokens_seen": 2536570880
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.672604718343766e-05,
      "loss": 2.5306,
      "theoretical_loss": 3.3646122830131064,
      "tokens_seen": 2536701952
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.671802278927941e-05,
      "loss": 2.6071,
      "theoretical_loss": 3.364598456476367,
      "tokens_seen": 2536833024
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.670999839512117e-05,
      "loss": 2.7333,
      "theoretical_loss": 3.364584630854008,
      "tokens_seen": 2536964096
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.670197400096293e-05,
      "loss": 2.4534,
      "theoretical_loss": 3.364570806145922,
      "tokens_seen": 2537095168
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.669394960680469e-05,
      "loss": 2.6157,
      "theoretical_loss": 3.3645569823520014,
      "tokens_seen": 2537226240
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.668592521264645e-05,
      "loss": 2.6054,
      "theoretical_loss": 3.3645431594721384,
      "tokens_seen": 2537357312
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.667790081848821e-05,
      "loss": 2.6694,
      "theoretical_loss": 3.3645293375062253,
      "tokens_seen": 2537488384
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6669876424329964e-05,
      "loss": 2.7058,
      "theoretical_loss": 3.364515516454155,
      "tokens_seen": 2537619456
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6661852030171723e-05,
      "loss": 2.6628,
      "theoretical_loss": 3.364501696315819,
      "tokens_seen": 2537750528
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6653827636013483e-05,
      "loss": 2.5902,
      "theoretical_loss": 3.3644878770911104,
      "tokens_seen": 2537881600
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1391196,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.740980863571167,
      "objective/train/theoretical_loss": 3.364474058779922,
      "objective/train/tokens_used": 908471776,
      "theoretical_loss": 3.364474058779922,
      "tokens_seen": 2538012672
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.664580324185524e-05,
      "loss": 2.6695,
      "theoretical_loss": 3.364474058779922,
      "tokens_seen": 2538012672
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6637778847697e-05,
      "loss": 2.6864,
      "theoretical_loss": 3.3644602413821447,
      "tokens_seen": 2538143744
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.662975445353876e-05,
      "loss": 2.5227,
      "theoretical_loss": 3.3644464248976718,
      "tokens_seen": 2538274816
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6621730059380516e-05,
      "loss": 2.672,
      "theoretical_loss": 3.3644326093263963,
      "tokens_seen": 2538405888
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6613705665222276e-05,
      "loss": 2.7277,
      "theoretical_loss": 3.36441879466821,
      "tokens_seen": 2538536960
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6605681271064036e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.364404980923006,
      "tokens_seen": 2538668032
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6597656876905796e-05,
      "loss": 2.6118,
      "theoretical_loss": 3.3643911680906764,
      "tokens_seen": 2538799104
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6589632482747556e-05,
      "loss": 2.5991,
      "theoretical_loss": 3.3643773561711137,
      "tokens_seen": 2538930176
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6581608088589316e-05,
      "loss": 2.5461,
      "theoretical_loss": 3.3643635451642107,
      "tokens_seen": 2539061248
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.657358369443107e-05,
      "loss": 2.5967,
      "theoretical_loss": 3.36434973506986,
      "tokens_seen": 2539192320
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.656555930027283e-05,
      "loss": 2.4883,
      "theoretical_loss": 3.364335925887954,
      "tokens_seen": 2539323392
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.655753490611459e-05,
      "loss": 2.6744,
      "theoretical_loss": 3.3643221176183853,
      "tokens_seen": 2539454464
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.654951051195635e-05,
      "loss": 2.6774,
      "theoretical_loss": 3.364308310261047,
      "tokens_seen": 2539585536
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1391838,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.731227159500122,
      "objective/train/theoretical_loss": 3.3643014069244304,
      "objective/train/tokens_used": 910110176,
      "theoretical_loss": 3.3643014069244304,
      "tokens_seen": 2539651072
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.654148611779811e-05,
      "loss": 2.7158,
      "theoretical_loss": 3.364294503815831,
      "tokens_seen": 2539716608
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.653346172363987e-05,
      "loss": 2.706,
      "theoretical_loss": 3.3642806982826308,
      "tokens_seen": 2539847680
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.652543732948162e-05,
      "loss": 2.513,
      "theoretical_loss": 3.364266893661338,
      "tokens_seen": 2539978752
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.651741293532338e-05,
      "loss": 2.7028,
      "theoretical_loss": 3.3642530899518466,
      "tokens_seen": 2540109824
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.650938854116514e-05,
      "loss": 2.5307,
      "theoretical_loss": 3.3642392871540485,
      "tokens_seen": 2540240896
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.65013641470069e-05,
      "loss": 2.4628,
      "theoretical_loss": 3.364225485267837,
      "tokens_seen": 2540371968
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.649333975284866e-05,
      "loss": 2.5531,
      "theoretical_loss": 3.364211684293104,
      "tokens_seen": 2540503040
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.648531535869042e-05,
      "loss": 2.5976,
      "theoretical_loss": 3.364197884229743,
      "tokens_seen": 2540634112
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6477290964532174e-05,
      "loss": 2.6477,
      "theoretical_loss": 3.364184085077647,
      "tokens_seen": 2540765184
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6469266570373934e-05,
      "loss": 2.387,
      "theoretical_loss": 3.364170286836708,
      "tokens_seen": 2540896256
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6461242176215694e-05,
      "loss": 2.5084,
      "theoretical_loss": 3.3641564895068194,
      "tokens_seen": 2541027328
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6453217782057454e-05,
      "loss": 2.5988,
      "theoretical_loss": 3.3641426930878744,
      "tokens_seen": 2541158400
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1392953,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.693942070007324,
      "objective/train/theoretical_loss": 3.3641288975797647,
      "objective/train/tokens_used": 911748576,
      "theoretical_loss": 3.3641288975797647,
      "tokens_seen": 2541289472
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6445193387899214e-05,
      "loss": 2.4564,
      "theoretical_loss": 3.3641288975797647,
      "tokens_seen": 2541289472
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6437168993740974e-05,
      "loss": 2.6742,
      "theoretical_loss": 3.3641151029823844,
      "tokens_seen": 2541420544
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6429144599582734e-05,
      "loss": 2.6573,
      "theoretical_loss": 3.3641013092956253,
      "tokens_seen": 2541551616
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6421120205424494e-05,
      "loss": 2.612,
      "theoretical_loss": 3.3640875165193815,
      "tokens_seen": 2541682688
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.641309581126625e-05,
      "loss": 2.5469,
      "theoretical_loss": 3.3640737246535455,
      "tokens_seen": 2541813760
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6405071417108007e-05,
      "loss": 2.6249,
      "theoretical_loss": 3.36405993369801,
      "tokens_seen": 2541944832
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6397047022949767e-05,
      "loss": 2.5959,
      "theoretical_loss": 3.364046143652668,
      "tokens_seen": 2542075904
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6389022628791526e-05,
      "loss": 2.6936,
      "theoretical_loss": 3.3640323545174127,
      "tokens_seen": 2542206976
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6380998234633286e-05,
      "loss": 2.7513,
      "theoretical_loss": 3.3640185662921374,
      "tokens_seen": 2542338048
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6372973840475046e-05,
      "loss": 2.5647,
      "theoretical_loss": 3.3640047789767342,
      "tokens_seen": 2542469120
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6364949446316806e-05,
      "loss": 2.5039,
      "theoretical_loss": 3.3639909925710976,
      "tokens_seen": 2542600192
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.635692505215856e-05,
      "loss": 2.5581,
      "theoretical_loss": 3.363977207075119,
      "tokens_seen": 2542731264
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.634890065800032e-05,
      "loss": 2.6156,
      "theoretical_loss": 3.363963422488693,
      "tokens_seen": 2542862336
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1393377,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.728156805038452,
      "objective/train/theoretical_loss": 3.3639565305365284,
      "objective/train/tokens_used": 913386976,
      "theoretical_loss": 3.3639565305365284,
      "tokens_seen": 2542927872
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.634087626384208e-05,
      "loss": 2.6067,
      "theoretical_loss": 3.363949638811712,
      "tokens_seen": 2542993408
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.633285186968384e-05,
      "loss": 2.6934,
      "theoretical_loss": 3.3639358560440686,
      "tokens_seen": 2543124480
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.63248274755256e-05,
      "loss": 2.6777,
      "theoretical_loss": 3.363922074185657,
      "tokens_seen": 2543255552
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.631680308136736e-05,
      "loss": 2.6308,
      "theoretical_loss": 3.3639082932363698,
      "tokens_seen": 2543386624
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.630877868720912e-05,
      "loss": 2.7084,
      "theoretical_loss": 3.3638945131961,
      "tokens_seen": 2543517696
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.630075429305087e-05,
      "loss": 2.5315,
      "theoretical_loss": 3.3638807340647414,
      "tokens_seen": 2543648768
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.629272989889263e-05,
      "loss": 2.5836,
      "theoretical_loss": 3.3638669558421865,
      "tokens_seen": 2543779840
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.628470550473439e-05,
      "loss": 2.5142,
      "theoretical_loss": 3.363853178528329,
      "tokens_seen": 2543910912
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.627668111057615e-05,
      "loss": 2.656,
      "theoretical_loss": 3.3638394021230624,
      "tokens_seen": 2544041984
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.626865671641791e-05,
      "loss": 2.5832,
      "theoretical_loss": 3.3638256266262796,
      "tokens_seen": 2544173056
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.626063232225967e-05,
      "loss": 2.5888,
      "theoretical_loss": 3.363811852037874,
      "tokens_seen": 2544304128
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.625260792810143e-05,
      "loss": 2.5318,
      "theoretical_loss": 3.3637980783577386,
      "tokens_seen": 2544435200
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1394146,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2629482746124268,
      "objective/train/theoretical_loss": 3.3637843055857672,
      "objective/train/tokens_used": 915025376,
      "theoretical_loss": 3.3637843055857672,
      "tokens_seen": 2544566272
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.624458353394319e-05,
      "loss": 2.6897,
      "theoretical_loss": 3.3637843055857672,
      "tokens_seen": 2544566272
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6236559139784944e-05,
      "loss": 2.5242,
      "theoretical_loss": 3.3637705337218526,
      "tokens_seen": 2544697344
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6228534745626704e-05,
      "loss": 2.6539,
      "theoretical_loss": 3.3637567627658886,
      "tokens_seen": 2544828416
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6220510351468464e-05,
      "loss": 2.5891,
      "theoretical_loss": 3.3637429927177687,
      "tokens_seen": 2544959488
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6212485957310224e-05,
      "loss": 2.5884,
      "theoretical_loss": 3.363729223577386,
      "tokens_seen": 2545090560
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6204461563151984e-05,
      "loss": 2.5462,
      "theoretical_loss": 3.363715455344634,
      "tokens_seen": 2545221632
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6196437168993744e-05,
      "loss": 2.7389,
      "theoretical_loss": 3.3637016880194057,
      "tokens_seen": 2545352704
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6188412774835504e-05,
      "loss": 2.6143,
      "theoretical_loss": 3.3636879216015956,
      "tokens_seen": 2545483776
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.618038838067726e-05,
      "loss": 2.8025,
      "theoretical_loss": 3.363674156091096,
      "tokens_seen": 2545614848
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.617236398651902e-05,
      "loss": 2.5315,
      "theoretical_loss": 3.363660391487801,
      "tokens_seen": 2545745920
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.616433959236078e-05,
      "loss": 2.5752,
      "theoretical_loss": 3.3636466277916046,
      "tokens_seen": 2545876992
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6156315198202537e-05,
      "loss": 2.5858,
      "theoretical_loss": 3.3636328650023994,
      "tokens_seen": 2546008064
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6148290804044296e-05,
      "loss": 2.6586,
      "theoretical_loss": 3.363619103120079,
      "tokens_seen": 2546139136
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1395124,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2976861000061035,
      "objective/train/theoretical_loss": 3.3636122225189675,
      "objective/train/tokens_used": 916663776,
      "theoretical_loss": 3.3636122225189675,
      "tokens_seen": 2546204672
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6140266409886056e-05,
      "loss": 2.5335,
      "theoretical_loss": 3.3636053421445378,
      "tokens_seen": 2546270208
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6132242015727816e-05,
      "loss": 2.5371,
      "theoretical_loss": 3.3635915820756686,
      "tokens_seen": 2546401280
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.612421762156957e-05,
      "loss": 2.6153,
      "theoretical_loss": 3.3635778229133653,
      "tokens_seen": 2546532352
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.611619322741133e-05,
      "loss": 2.6593,
      "theoretical_loss": 3.3635640646575213,
      "tokens_seen": 2546663424
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.610816883325309e-05,
      "loss": 2.552,
      "theoretical_loss": 3.3635503073080306,
      "tokens_seen": 2546794496
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.610014443909485e-05,
      "loss": 2.5235,
      "theoretical_loss": 3.3635365508647865,
      "tokens_seen": 2546925568
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.609212004493661e-05,
      "loss": 2.6038,
      "theoretical_loss": 3.363522795327683,
      "tokens_seen": 2547056640
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.608409565077837e-05,
      "loss": 2.529,
      "theoretical_loss": 3.3635090406966133,
      "tokens_seen": 2547187712
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.607607125662013e-05,
      "loss": 2.6483,
      "theoretical_loss": 3.363495286971472,
      "tokens_seen": 2547318784
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.606804686246189e-05,
      "loss": 2.4785,
      "theoretical_loss": 3.3634815341521516,
      "tokens_seen": 2547449856
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.606002246830364e-05,
      "loss": 2.5247,
      "theoretical_loss": 3.3634677822385464,
      "tokens_seen": 2547580928
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.60519980741454e-05,
      "loss": 2.6911,
      "theoretical_loss": 3.3634540312305505,
      "tokens_seen": 2547712000
    },
    {
      "epoch": 0.54,
      "objective/train/docs_used": 1395711,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6512081623077393,
      "objective/train/theoretical_loss": 3.3634402811280575,
      "objective/train/tokens_used": 918302176,
      "theoretical_loss": 3.3634402811280575,
      "tokens_seen": 2547843072
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.604397367998716e-05,
      "loss": 2.5705,
      "theoretical_loss": 3.3634402811280575,
      "tokens_seen": 2547843072
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.603594928582892e-05,
      "loss": 2.6375,
      "theoretical_loss": 3.363426531930961,
      "tokens_seen": 2547974144
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.602792489167068e-05,
      "loss": 2.6553,
      "theoretical_loss": 3.3634127836391547,
      "tokens_seen": 2548105216
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.601990049751244e-05,
      "loss": 2.6957,
      "theoretical_loss": 3.363399036252533,
      "tokens_seen": 2548236288
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.60118761033542e-05,
      "loss": 2.5482,
      "theoretical_loss": 3.363385289770989,
      "tokens_seen": 2548367360
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.6003851709195954e-05,
      "loss": 2.5083,
      "theoretical_loss": 3.3633715441944174,
      "tokens_seen": 2548498432
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.5995827315037714e-05,
      "loss": 2.4978,
      "theoretical_loss": 3.3633577995227113,
      "tokens_seen": 2548629504
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.5987802920879474e-05,
      "loss": 2.5004,
      "theoretical_loss": 3.363344055755765,
      "tokens_seen": 2548760576
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.5979778526721234e-05,
      "loss": 2.5514,
      "theoretical_loss": 3.3633303128934724,
      "tokens_seen": 2548891648
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.5971754132562994e-05,
      "loss": 2.5959,
      "theoretical_loss": 3.3633165709357273,
      "tokens_seen": 2549022720
    },
    {
      "epoch": 0.54,
      "learning_rate": 4.5963729738404754e-05,
      "loss": 2.6378,
      "theoretical_loss": 3.363302829882424,
      "tokens_seen": 2549153792
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5955705344246514e-05,
      "loss": 2.6907,
      "theoretical_loss": 3.363289089733456,
      "tokens_seen": 2549284864
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5947680950088274e-05,
      "loss": 2.6071,
      "theoretical_loss": 3.363275350488718,
      "tokens_seen": 2549415936
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1397015,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.686042308807373,
      "objective/train/theoretical_loss": 3.3632684812054015,
      "objective/train/tokens_used": 919940576,
      "theoretical_loss": 3.3632684812054015,
      "tokens_seen": 2549481472
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.593965655593003e-05,
      "loss": 2.6513,
      "theoretical_loss": 3.363261612148103,
      "tokens_seen": 2549547008
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.593163216177179e-05,
      "loss": 2.634,
      "theoretical_loss": 3.3632478747115053,
      "tokens_seen": 2549678080
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.592360776761355e-05,
      "loss": 2.5412,
      "theoretical_loss": 3.3632341381788198,
      "tokens_seen": 2549809152
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.591558337345531e-05,
      "loss": 2.6273,
      "theoretical_loss": 3.3632204025499397,
      "tokens_seen": 2549940224
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5907558979297067e-05,
      "loss": 2.496,
      "theoretical_loss": 3.3632066678247594,
      "tokens_seen": 2550071296
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5899534585138826e-05,
      "loss": 2.6088,
      "theoretical_loss": 3.3631929340031728,
      "tokens_seen": 2550202368
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5891510190980586e-05,
      "loss": 2.6715,
      "theoretical_loss": 3.3631792010850745,
      "tokens_seen": 2550333440
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.588348579682234e-05,
      "loss": 2.4074,
      "theoretical_loss": 3.363165469070358,
      "tokens_seen": 2550464512
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.58754614026641e-05,
      "loss": 2.6167,
      "theoretical_loss": 3.3631517379589178,
      "tokens_seen": 2550595584
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.586743700850586e-05,
      "loss": 2.5815,
      "theoretical_loss": 3.363138007750648,
      "tokens_seen": 2550726656
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.585941261434762e-05,
      "loss": 2.5765,
      "theoretical_loss": 3.3631242784454427,
      "tokens_seen": 2550857728
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.585138822018938e-05,
      "loss": 2.5372,
      "theoretical_loss": 3.3631105500431966,
      "tokens_seen": 2550988800
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1397571,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.595946788787842,
      "objective/train/theoretical_loss": 3.363096822543803,
      "objective/train/tokens_used": 921578976,
      "theoretical_loss": 3.363096822543803,
      "tokens_seen": 2551119872
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.584336382603114e-05,
      "loss": 2.5753,
      "theoretical_loss": 3.363096822543803,
      "tokens_seen": 2551119872
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.58353394318729e-05,
      "loss": 2.4736,
      "theoretical_loss": 3.363083095947157,
      "tokens_seen": 2551250944
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.582731503771465e-05,
      "loss": 2.5798,
      "theoretical_loss": 3.3630693702531524,
      "tokens_seen": 2551382016
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.581929064355641e-05,
      "loss": 2.6599,
      "theoretical_loss": 3.3630556454616833,
      "tokens_seen": 2551513088
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.581126624939817e-05,
      "loss": 2.5256,
      "theoretical_loss": 3.363041921572645,
      "tokens_seen": 2551644160
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.580324185523993e-05,
      "loss": 2.6855,
      "theoretical_loss": 3.3630281985859307,
      "tokens_seen": 2551775232
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.579521746108169e-05,
      "loss": 2.5716,
      "theoretical_loss": 3.3630144765014354,
      "tokens_seen": 2551906304
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.578719306692345e-05,
      "loss": 2.5468,
      "theoretical_loss": 3.363000755319053,
      "tokens_seen": 2552037376
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.577916867276521e-05,
      "loss": 2.5389,
      "theoretical_loss": 3.3629870350386777,
      "tokens_seen": 2552168448
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.577114427860697e-05,
      "loss": 2.4739,
      "theoretical_loss": 3.362973315660205,
      "tokens_seen": 2552299520
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5763119884448725e-05,
      "loss": 2.6023,
      "theoretical_loss": 3.362959597183528,
      "tokens_seen": 2552430592
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5755095490290484e-05,
      "loss": 2.6172,
      "theoretical_loss": 3.3629458796085414,
      "tokens_seen": 2552561664
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5747071096132244e-05,
      "loss": 2.5386,
      "theoretical_loss": 3.3629321629351407,
      "tokens_seen": 2552692736
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1398421,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.1858367919921875,
      "objective/train/theoretical_loss": 3.3629253049365015,
      "objective/train/tokens_used": 923217376,
      "theoretical_loss": 3.3629253049365015,
      "tokens_seen": 2552758272
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5739046701974004e-05,
      "loss": 2.5921,
      "theoretical_loss": 3.362918447163219,
      "tokens_seen": 2552823808
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5731022307815764e-05,
      "loss": 2.5813,
      "theoretical_loss": 3.3629047322926713,
      "tokens_seen": 2552954880
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5722997913657524e-05,
      "loss": 2.4869,
      "theoretical_loss": 3.3628910183233924,
      "tokens_seen": 2553085952
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5714973519499284e-05,
      "loss": 2.7103,
      "theoretical_loss": 3.3628773052552763,
      "tokens_seen": 2553217024
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.570694912534104e-05,
      "loss": 2.5565,
      "theoretical_loss": 3.362863593088218,
      "tokens_seen": 2553348096
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.56989247311828e-05,
      "loss": 2.5588,
      "theoretical_loss": 3.3628498818221115,
      "tokens_seen": 2553479168
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.569090033702456e-05,
      "loss": 2.6041,
      "theoretical_loss": 3.362836171456852,
      "tokens_seen": 2553610240
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.568287594286632e-05,
      "loss": 2.5203,
      "theoretical_loss": 3.3628224619923333,
      "tokens_seen": 2553741312
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.567485154870808e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.3628087534284505,
      "tokens_seen": 2553872384
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.566682715454984e-05,
      "loss": 2.5169,
      "theoretical_loss": 3.3627950457650986,
      "tokens_seen": 2554003456
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5658802760391597e-05,
      "loss": 2.5348,
      "theoretical_loss": 3.3627813390021712,
      "tokens_seen": 2554134528
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.565077836623335e-05,
      "loss": 2.5583,
      "theoretical_loss": 3.3627676331395637,
      "tokens_seen": 2554265600
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1399303,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.303102493286133,
      "objective/train/theoretical_loss": 3.3627539281771712,
      "objective/train/tokens_used": 924855776,
      "theoretical_loss": 3.3627539281771712,
      "tokens_seen": 2554396672
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.564275397207511e-05,
      "loss": 2.3812,
      "theoretical_loss": 3.3627539281771712,
      "tokens_seen": 2554396672
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.563472957791687e-05,
      "loss": 2.6224,
      "theoretical_loss": 3.362740224114887,
      "tokens_seen": 2554527744
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.562670518375863e-05,
      "loss": 2.487,
      "theoretical_loss": 3.3627265209526067,
      "tokens_seen": 2554658816
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.561868078960039e-05,
      "loss": 2.4577,
      "theoretical_loss": 3.362712818690225,
      "tokens_seen": 2554789888
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.561065639544215e-05,
      "loss": 2.5577,
      "theoretical_loss": 3.362699117327637,
      "tokens_seen": 2554920960
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.560263200128391e-05,
      "loss": 2.4576,
      "theoretical_loss": 3.3626854168647364,
      "tokens_seen": 2555052032
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.559460760712567e-05,
      "loss": 2.6145,
      "theoretical_loss": 3.362671717301419,
      "tokens_seen": 2555183104
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.558658321296742e-05,
      "loss": 2.5011,
      "theoretical_loss": 3.3626580186375787,
      "tokens_seen": 2555314176
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.557855881880918e-05,
      "loss": 2.4978,
      "theoretical_loss": 3.362644320873111,
      "tokens_seen": 2555445248
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.557053442465094e-05,
      "loss": 2.6062,
      "theoretical_loss": 3.3626306240079105,
      "tokens_seen": 2555576320
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.55625100304927e-05,
      "loss": 2.614,
      "theoretical_loss": 3.3626169280418723,
      "tokens_seen": 2555707392
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.555448563633446e-05,
      "loss": 2.4572,
      "theoretical_loss": 3.3626032329748905,
      "tokens_seen": 2555838464
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.554646124217622e-05,
      "loss": 2.5684,
      "theoretical_loss": 3.362589538806861,
      "tokens_seen": 2555969536
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1399979,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.480012893676758,
      "objective/train/theoretical_loss": 3.36258269205992,
      "objective/train/tokens_used": 926494176,
      "theoretical_loss": 3.36258269205992,
      "tokens_seen": 2556035072
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.553843684801798e-05,
      "loss": 2.5772,
      "theoretical_loss": 3.3625758455376777,
      "tokens_seen": 2556100608
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5530412453859735e-05,
      "loss": 2.5696,
      "theoretical_loss": 3.3625621531672367,
      "tokens_seen": 2556231680
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5522388059701495e-05,
      "loss": 2.5449,
      "theoretical_loss": 3.3625484616954315,
      "tokens_seen": 2556362752
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5514363665543255e-05,
      "loss": 2.3782,
      "theoretical_loss": 3.3625347711221583,
      "tokens_seen": 2556493824
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5506339271385014e-05,
      "loss": 2.6605,
      "theoretical_loss": 3.362521081447311,
      "tokens_seen": 2556624896
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5498314877226774e-05,
      "loss": 2.6123,
      "theoretical_loss": 3.3625073926707856,
      "tokens_seen": 2556755968
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5490290483068534e-05,
      "loss": 2.5033,
      "theoretical_loss": 3.3624937047924766,
      "tokens_seen": 2556887040
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5482266088910294e-05,
      "loss": 2.5175,
      "theoretical_loss": 3.362480017812279,
      "tokens_seen": 2557018112
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5474241694752054e-05,
      "loss": 2.5411,
      "theoretical_loss": 3.362466331730088,
      "tokens_seen": 2557149184
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.546621730059381e-05,
      "loss": 2.6008,
      "theoretical_loss": 3.3624526465457985,
      "tokens_seen": 2557280256
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.545819290643557e-05,
      "loss": 2.6209,
      "theoretical_loss": 3.362438962259305,
      "tokens_seen": 2557411328
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.545016851227733e-05,
      "loss": 2.4899,
      "theoretical_loss": 3.362425278870504,
      "tokens_seen": 2557542400
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1401394,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3670003414154053,
      "objective/train/theoretical_loss": 3.3624115963792898,
      "objective/train/tokens_used": 928132576,
      "theoretical_loss": 3.3624115963792898,
      "tokens_seen": 2557673472
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.544214411811909e-05,
      "loss": 2.5437,
      "theoretical_loss": 3.3624115963792898,
      "tokens_seen": 2557673472
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.543411972396085e-05,
      "loss": 2.5497,
      "theoretical_loss": 3.362397914785557,
      "tokens_seen": 2557804544
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.542609532980261e-05,
      "loss": 2.4735,
      "theoretical_loss": 3.3623842340892023,
      "tokens_seen": 2557935616
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.541807093564437e-05,
      "loss": 2.6774,
      "theoretical_loss": 3.362370554290119,
      "tokens_seen": 2558066688
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.541004654148612e-05,
      "loss": 2.5931,
      "theoretical_loss": 3.362356875388204,
      "tokens_seen": 2558197760
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.540202214732788e-05,
      "loss": 2.6006,
      "theoretical_loss": 3.3623431973833506,
      "tokens_seen": 2558328832
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.539399775316964e-05,
      "loss": 2.5531,
      "theoretical_loss": 3.362329520275456,
      "tokens_seen": 2558459904
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.53859733590114e-05,
      "loss": 2.5637,
      "theoretical_loss": 3.3623158440644136,
      "tokens_seen": 2558590976
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.537794896485316e-05,
      "loss": 2.545,
      "theoretical_loss": 3.3623021687501202,
      "tokens_seen": 2558722048
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.536992457069492e-05,
      "loss": 2.7055,
      "theoretical_loss": 3.3622884943324705,
      "tokens_seen": 2558853120
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.536190017653667e-05,
      "loss": 2.6548,
      "theoretical_loss": 3.3622748208113595,
      "tokens_seen": 2558984192
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.535387578237843e-05,
      "loss": 2.4317,
      "theoretical_loss": 3.3622611481866826,
      "tokens_seen": 2559115264
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.534585138822019e-05,
      "loss": 2.4451,
      "theoretical_loss": 3.3622474764583354,
      "tokens_seen": 2559246336
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1402751,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5578465461730957,
      "objective/train/theoretical_loss": 3.3622406409302528,
      "objective/train/tokens_used": 929770976,
      "theoretical_loss": 3.3622406409302528,
      "tokens_seen": 2559311872
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.533782699406195e-05,
      "loss": 2.7087,
      "theoretical_loss": 3.3622338056262127,
      "tokens_seen": 2559377408
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.532980259990371e-05,
      "loss": 2.5805,
      "theoretical_loss": 3.362220135690211,
      "tokens_seen": 2559508480
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.532177820574547e-05,
      "loss": 2.5726,
      "theoretical_loss": 3.362206466650224,
      "tokens_seen": 2559639552
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5313753811587225e-05,
      "loss": 2.6407,
      "theoretical_loss": 3.3621927985061486,
      "tokens_seen": 2559770624
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5305729417428985e-05,
      "loss": 2.4324,
      "theoretical_loss": 3.3621791312578795,
      "tokens_seen": 2559901696
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5297705023270745e-05,
      "loss": 2.5759,
      "theoretical_loss": 3.362165464905312,
      "tokens_seen": 2560032768
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5289680629112505e-05,
      "loss": 2.3916,
      "theoretical_loss": 3.362151799448342,
      "tokens_seen": 2560163840
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5281656234954265e-05,
      "loss": 2.663,
      "theoretical_loss": 3.3621381348868646,
      "tokens_seen": 2560294912
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5273631840796025e-05,
      "loss": 2.5597,
      "theoretical_loss": 3.3621244712207754,
      "tokens_seen": 2560425984
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.526560744663778e-05,
      "loss": 2.4081,
      "theoretical_loss": 3.36211080844997,
      "tokens_seen": 2560557056
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.525758305247954e-05,
      "loss": 2.6559,
      "theoretical_loss": 3.3620971465743437,
      "tokens_seen": 2560688128
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.52495586583213e-05,
      "loss": 2.606,
      "theoretical_loss": 3.3620834855937924,
      "tokens_seen": 2560819200
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1403315,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7906315326690674,
      "objective/train/theoretical_loss": 3.362069825508211,
      "objective/train/tokens_used": 931409376,
      "theoretical_loss": 3.362069825508211,
      "tokens_seen": 2560950272
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.524153426416306e-05,
      "loss": 2.6635,
      "theoretical_loss": 3.362069825508211,
      "tokens_seen": 2560950272
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.523350987000482e-05,
      "loss": 2.4842,
      "theoretical_loss": 3.3620561663174957,
      "tokens_seen": 2561081344
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.522548547584658e-05,
      "loss": 2.5513,
      "theoretical_loss": 3.362042508021542,
      "tokens_seen": 2561212416
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.521746108168833e-05,
      "loss": 2.6291,
      "theoretical_loss": 3.3620288506202454,
      "tokens_seen": 2561343488
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.520943668753009e-05,
      "loss": 2.6285,
      "theoretical_loss": 3.3620151941135012,
      "tokens_seen": 2561474560
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.520141229337185e-05,
      "loss": 2.5342,
      "theoretical_loss": 3.3620015385012056,
      "tokens_seen": 2561605632
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.519338789921361e-05,
      "loss": 2.6027,
      "theoretical_loss": 3.3619878837832537,
      "tokens_seen": 2561736704
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.518536350505537e-05,
      "loss": 2.6192,
      "theoretical_loss": 3.361974229959541,
      "tokens_seen": 2561867776
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.517733911089713e-05,
      "loss": 2.5033,
      "theoretical_loss": 3.3619605770299645,
      "tokens_seen": 2561998848
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.516931471673888e-05,
      "loss": 2.5271,
      "theoretical_loss": 3.361946924994419,
      "tokens_seen": 2562129920
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.516129032258064e-05,
      "loss": 2.3692,
      "theoretical_loss": 3.3619332738527996,
      "tokens_seen": 2562260992
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.51532659284224e-05,
      "loss": 2.5969,
      "theoretical_loss": 3.3619196236050035,
      "tokens_seen": 2562392064
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.514524153426416e-05,
      "loss": 2.5732,
      "theoretical_loss": 3.361905974250925,
      "tokens_seen": 2562523136
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1404572,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8795578479766846,
      "objective/train/theoretical_loss": 3.3618991499089974,
      "objective/train/tokens_used": 933047776,
      "theoretical_loss": 3.3618991499089974,
      "tokens_seen": 2562588672
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.513721714010592e-05,
      "loss": 2.6146,
      "theoretical_loss": 3.3618923257904605,
      "tokens_seen": 2562654208
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.512919274594768e-05,
      "loss": 2.5252,
      "theoretical_loss": 3.3618786782235066,
      "tokens_seen": 2562785280
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.512116835178944e-05,
      "loss": 2.5647,
      "theoretical_loss": 3.3618650315499576,
      "tokens_seen": 2562916352
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5113143957631196e-05,
      "loss": 2.5046,
      "theoretical_loss": 3.3618513857697105,
      "tokens_seen": 2563047424
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5105119563472956e-05,
      "loss": 2.6008,
      "theoretical_loss": 3.3618377408826605,
      "tokens_seen": 2563178496
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5097095169314715e-05,
      "loss": 2.4981,
      "theoretical_loss": 3.3618240968887037,
      "tokens_seen": 2563309568
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5089070775156475e-05,
      "loss": 2.5218,
      "theoretical_loss": 3.361810453787736,
      "tokens_seen": 2563440640
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5081046380998235e-05,
      "loss": 2.4183,
      "theoretical_loss": 3.3617968115796533,
      "tokens_seen": 2563571712
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5073021986839995e-05,
      "loss": 2.5733,
      "theoretical_loss": 3.3617831702643515,
      "tokens_seen": 2563702784
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.5064997592681755e-05,
      "loss": 2.507,
      "theoretical_loss": 3.3617695298417267,
      "tokens_seen": 2563833856
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.505697319852351e-05,
      "loss": 2.6407,
      "theoretical_loss": 3.361755890311674,
      "tokens_seen": 2563964928
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.504894880436527e-05,
      "loss": 2.7645,
      "theoretical_loss": 3.361742251674091,
      "tokens_seen": 2564096000
    },
    {
      "epoch": 0.55,
      "objective/train/docs_used": 1405290,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.719271659851074,
      "objective/train/theoretical_loss": 3.361728613928872,
      "objective/train/tokens_used": 934686176,
      "theoretical_loss": 3.361728613928872,
      "tokens_seen": 2564227072
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.504092441020703e-05,
      "loss": 2.5601,
      "theoretical_loss": 3.361728613928872,
      "tokens_seen": 2564227072
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.503290001604879e-05,
      "loss": 2.4802,
      "theoretical_loss": 3.361714977075914,
      "tokens_seen": 2564358144
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.502487562189055e-05,
      "loss": 2.6267,
      "theoretical_loss": 3.3617013411151127,
      "tokens_seen": 2564489216
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.501685122773231e-05,
      "loss": 2.431,
      "theoretical_loss": 3.3616877060463644,
      "tokens_seen": 2564620288
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.500882683357407e-05,
      "loss": 2.2944,
      "theoretical_loss": 3.3616740718695644,
      "tokens_seen": 2564751360
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.500080243941583e-05,
      "loss": 2.3951,
      "theoretical_loss": 3.36166043858461,
      "tokens_seen": 2564882432
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.499277804525758e-05,
      "loss": 2.5859,
      "theoretical_loss": 3.361646806191396,
      "tokens_seen": 2565013504
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.498475365109934e-05,
      "loss": 2.4016,
      "theoretical_loss": 3.36163317468982,
      "tokens_seen": 2565144576
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.49767292569411e-05,
      "loss": 2.3147,
      "theoretical_loss": 3.3616195440797765,
      "tokens_seen": 2565275648
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.496870486278286e-05,
      "loss": 2.4702,
      "theoretical_loss": 3.3616059143611623,
      "tokens_seen": 2565406720
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.496068046862462e-05,
      "loss": 2.4628,
      "theoretical_loss": 3.3615922855338742,
      "tokens_seen": 2565537792
    },
    {
      "epoch": 0.55,
      "learning_rate": 4.495265607446638e-05,
      "loss": 2.5872,
      "theoretical_loss": 3.3615786575978075,
      "tokens_seen": 2565668864
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.494463168030814e-05,
      "loss": 2.4588,
      "theoretical_loss": 3.3615650305528586,
      "tokens_seen": 2565799936
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1405994,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.2271580696105957,
      "objective/train/theoretical_loss": 3.361558217364521,
      "objective/train/tokens_used": 936324576,
      "theoretical_loss": 3.361558217364521,
      "tokens_seen": 2565865472
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.493660728614989e-05,
      "loss": 2.3935,
      "theoretical_loss": 3.361551404398924,
      "tokens_seen": 2565931008
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.492858289199165e-05,
      "loss": 2.6389,
      "theoretical_loss": 3.3615377791359,
      "tokens_seen": 2566062080
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.492055849783341e-05,
      "loss": 2.5881,
      "theoretical_loss": 3.361524154763682,
      "tokens_seen": 2566193152
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.491253410367517e-05,
      "loss": 2.589,
      "theoretical_loss": 3.3615105312821676,
      "tokens_seen": 2566324224
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.490450970951693e-05,
      "loss": 2.5712,
      "theoretical_loss": 3.3614969086912523,
      "tokens_seen": 2566455296
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.489648531535869e-05,
      "loss": 2.7065,
      "theoretical_loss": 3.361483286990832,
      "tokens_seen": 2566586368
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.488846092120045e-05,
      "loss": 2.3785,
      "theoretical_loss": 3.361469666180804,
      "tokens_seen": 2566717440
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4880436527042206e-05,
      "loss": 2.61,
      "theoretical_loss": 3.361456046261064,
      "tokens_seen": 2566848512
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4872412132883966e-05,
      "loss": 2.6727,
      "theoretical_loss": 3.3614424272315078,
      "tokens_seen": 2566979584
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4864387738725726e-05,
      "loss": 2.6467,
      "theoretical_loss": 3.361428809092033,
      "tokens_seen": 2567110656
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4856363344567486e-05,
      "loss": 2.4898,
      "theoretical_loss": 3.3614151918425357,
      "tokens_seen": 2567241728
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4848338950409245e-05,
      "loss": 2.5376,
      "theoretical_loss": 3.3614015754829114,
      "tokens_seen": 2567372800
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1406689,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3093464374542236,
      "objective/train/theoretical_loss": 3.361387960013057,
      "objective/train/tokens_used": 937962976,
      "theoretical_loss": 3.361387960013057,
      "tokens_seen": 2567503872
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4840314556251005e-05,
      "loss": 2.5131,
      "theoretical_loss": 3.361387960013057,
      "tokens_seen": 2567503872
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4832290162092765e-05,
      "loss": 2.371,
      "theoretical_loss": 3.36137434543287,
      "tokens_seen": 2567634944
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4824265767934525e-05,
      "loss": 2.5093,
      "theoretical_loss": 3.3613607317422454,
      "tokens_seen": 2567766016
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.481624137377628e-05,
      "loss": 2.4777,
      "theoretical_loss": 3.3613471189410804,
      "tokens_seen": 2567897088
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.480821697961804e-05,
      "loss": 2.6384,
      "theoretical_loss": 3.361333507029271,
      "tokens_seen": 2568028160
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.48001925854598e-05,
      "loss": 2.5472,
      "theoretical_loss": 3.361319896006714,
      "tokens_seen": 2568159232
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.479216819130156e-05,
      "loss": 2.4633,
      "theoretical_loss": 3.3613062858733063,
      "tokens_seen": 2568290304
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.478414379714332e-05,
      "loss": 2.5846,
      "theoretical_loss": 3.361292676628944,
      "tokens_seen": 2568421376
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.477611940298508e-05,
      "loss": 2.4566,
      "theoretical_loss": 3.361279068273524,
      "tokens_seen": 2568552448
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.476809500882684e-05,
      "loss": 2.4543,
      "theoretical_loss": 3.3612654608069423,
      "tokens_seen": 2568683520
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.476007061466859e-05,
      "loss": 2.4805,
      "theoretical_loss": 3.3612518542290957,
      "tokens_seen": 2568814592
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.475204622051035e-05,
      "loss": 2.5021,
      "theoretical_loss": 3.361238248539881,
      "tokens_seen": 2568945664
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.474402182635211e-05,
      "loss": 2.53,
      "theoretical_loss": 3.361224643739195,
      "tokens_seen": 2569076736
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1408046,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.762916088104248,
      "objective/train/theoretical_loss": 3.3612178416720178,
      "objective/train/tokens_used": 939601376,
      "theoretical_loss": 3.3612178416720178,
      "tokens_seen": 2569142272
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.473599743219387e-05,
      "loss": 2.4548,
      "theoretical_loss": 3.3612110398269337,
      "tokens_seen": 2569207808
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.472797303803563e-05,
      "loss": 2.5672,
      "theoretical_loss": 3.3611974368029944,
      "tokens_seen": 2569338880
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.471994864387739e-05,
      "loss": 2.548,
      "theoretical_loss": 3.3611838346672736,
      "tokens_seen": 2569469952
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.471192424971915e-05,
      "loss": 2.5908,
      "theoretical_loss": 3.361170233419668,
      "tokens_seen": 2569601024
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.47038998555609e-05,
      "loss": 2.4423,
      "theoretical_loss": 3.361156633060074,
      "tokens_seen": 2569732096
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.469587546140266e-05,
      "loss": 2.5423,
      "theoretical_loss": 3.361143033588389,
      "tokens_seen": 2569863168
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.468785106724442e-05,
      "loss": 2.552,
      "theoretical_loss": 3.361129435004509,
      "tokens_seen": 2569994240
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.467982667308618e-05,
      "loss": 2.5089,
      "theoretical_loss": 3.3611158373083314,
      "tokens_seen": 2570125312
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.467180227892794e-05,
      "loss": 2.4429,
      "theoretical_loss": 3.3611022404997524,
      "tokens_seen": 2570256384
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.46637778847697e-05,
      "loss": 2.5397,
      "theoretical_loss": 3.361088644578669,
      "tokens_seen": 2570387456
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.465575349061146e-05,
      "loss": 2.4347,
      "theoretical_loss": 3.3610750495449784,
      "tokens_seen": 2570518528
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.464772909645322e-05,
      "loss": 2.5479,
      "theoretical_loss": 3.3610614553985774,
      "tokens_seen": 2570649600
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1408629,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.9207708835601807,
      "objective/train/theoretical_loss": 3.3610478621393622,
      "objective/train/tokens_used": 941239776,
      "theoretical_loss": 3.3610478621393622,
      "tokens_seen": 2570780672
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4639704702294976e-05,
      "loss": 2.7436,
      "theoretical_loss": 3.3610478621393622,
      "tokens_seen": 2570780672
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4631680308136736e-05,
      "loss": 2.4666,
      "theoretical_loss": 3.3610342697672304,
      "tokens_seen": 2570911744
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4623655913978496e-05,
      "loss": 2.4001,
      "theoretical_loss": 3.3610206782820784,
      "tokens_seen": 2571042816
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4615631519820256e-05,
      "loss": 2.5874,
      "theoretical_loss": 3.361007087683803,
      "tokens_seen": 2571173888
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4607607125662015e-05,
      "loss": 2.4978,
      "theoretical_loss": 3.3609934979723017,
      "tokens_seen": 2571304960
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4599582731503775e-05,
      "loss": 2.5588,
      "theoretical_loss": 3.360979909147471,
      "tokens_seen": 2571436032
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4591558337345535e-05,
      "loss": 2.5659,
      "theoretical_loss": 3.3609663212092085,
      "tokens_seen": 2571567104
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.458353394318729e-05,
      "loss": 2.4862,
      "theoretical_loss": 3.36095273415741,
      "tokens_seen": 2571698176
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.457550954902905e-05,
      "loss": 2.5765,
      "theoretical_loss": 3.3609391479919735,
      "tokens_seen": 2571829248
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.456748515487081e-05,
      "loss": 2.4857,
      "theoretical_loss": 3.360925562712796,
      "tokens_seen": 2571960320
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.455946076071257e-05,
      "loss": 2.4668,
      "theoretical_loss": 3.3609119783197734,
      "tokens_seen": 2572091392
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.455143636655433e-05,
      "loss": 2.5625,
      "theoretical_loss": 3.360898394812804,
      "tokens_seen": 2572222464
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.454341197239609e-05,
      "loss": 2.5671,
      "theoretical_loss": 3.3608848121917845,
      "tokens_seen": 2572353536
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1409650,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6852149963378906,
      "objective/train/theoretical_loss": 3.3608780212134737,
      "objective/train/tokens_used": 942878176,
      "theoretical_loss": 3.3608780212134737,
      "tokens_seen": 2572419072
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.453538757823785e-05,
      "loss": 2.667,
      "theoretical_loss": 3.3608712304566115,
      "tokens_seen": 2572484608
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.452736318407961e-05,
      "loss": 2.4249,
      "theoretical_loss": 3.360857649607183,
      "tokens_seen": 2572615680
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.451933878992136e-05,
      "loss": 2.5772,
      "theoretical_loss": 3.360844069643395,
      "tokens_seen": 2572746752
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.451131439576312e-05,
      "loss": 2.7505,
      "theoretical_loss": 3.360830490565146,
      "tokens_seen": 2572877824
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.450329000160488e-05,
      "loss": 2.5656,
      "theoretical_loss": 3.3608169123723317,
      "tokens_seen": 2573008896
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.449526560744664e-05,
      "loss": 2.5588,
      "theoretical_loss": 3.3608033350648503,
      "tokens_seen": 2573139968
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.44872412132884e-05,
      "loss": 2.4582,
      "theoretical_loss": 3.360789758642598,
      "tokens_seen": 2573271040
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.447921681913016e-05,
      "loss": 2.5461,
      "theoretical_loss": 3.3607761831054734,
      "tokens_seen": 2573402112
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.447119242497192e-05,
      "loss": 2.5004,
      "theoretical_loss": 3.3607626084533724,
      "tokens_seen": 2573533184
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4463168030813673e-05,
      "loss": 2.5159,
      "theoretical_loss": 3.360749034686193,
      "tokens_seen": 2573664256
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.445514363665543e-05,
      "loss": 2.577,
      "theoretical_loss": 3.360735461803832,
      "tokens_seen": 2573795328
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.444711924249719e-05,
      "loss": 2.6183,
      "theoretical_loss": 3.360721889806187,
      "tokens_seen": 2573926400
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1410556,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3823792934417725,
      "objective/train/theoretical_loss": 3.3607083186931552,
      "objective/train/tokens_used": 944516576,
      "theoretical_loss": 3.3607083186931552,
      "tokens_seen": 2574057472
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.443909484833895e-05,
      "loss": 2.5098,
      "theoretical_loss": 3.3607083186931552,
      "tokens_seen": 2574057472
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.443107045418071e-05,
      "loss": 2.5404,
      "theoretical_loss": 3.360694748464634,
      "tokens_seen": 2574188544
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.442304606002247e-05,
      "loss": 2.4381,
      "theoretical_loss": 3.3606811791205202,
      "tokens_seen": 2574319616
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.441502166586423e-05,
      "loss": 2.5789,
      "theoretical_loss": 3.3606676106607116,
      "tokens_seen": 2574450688
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4406997271705986e-05,
      "loss": 2.6221,
      "theoretical_loss": 3.3606540430851055,
      "tokens_seen": 2574581760
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4398972877547746e-05,
      "loss": 2.4964,
      "theoretical_loss": 3.3606404763935993,
      "tokens_seen": 2574712832
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4390948483389506e-05,
      "loss": 2.4961,
      "theoretical_loss": 3.3606269105860904,
      "tokens_seen": 2574843904
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4382924089231266e-05,
      "loss": 2.295,
      "theoretical_loss": 3.3606133456624763,
      "tokens_seen": 2574974976
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4374899695073026e-05,
      "loss": 2.5428,
      "theoretical_loss": 3.3605997816226534,
      "tokens_seen": 2575106048
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4366875300914786e-05,
      "loss": 2.6581,
      "theoretical_loss": 3.360586218466521,
      "tokens_seen": 2575237120
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4358850906756545e-05,
      "loss": 2.4116,
      "theoretical_loss": 3.360572656193975,
      "tokens_seen": 2575368192
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4350826512598305e-05,
      "loss": 2.4559,
      "theoretical_loss": 3.3605590948049135,
      "tokens_seen": 2575499264
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.434280211844006e-05,
      "loss": 2.4946,
      "theoretical_loss": 3.360545534299234,
      "tokens_seen": 2575630336
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1411184,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.4256932735443115,
      "objective/train/theoretical_loss": 3.36053875437763,
      "objective/train/tokens_used": 946154976,
      "theoretical_loss": 3.36053875437763,
      "tokens_seen": 2575695872
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.433477772428182e-05,
      "loss": 2.5491,
      "theoretical_loss": 3.3605319746768334,
      "tokens_seen": 2575761408
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.432675333012358e-05,
      "loss": 2.6144,
      "theoretical_loss": 3.3605184159376105,
      "tokens_seen": 2575892480
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.431872893596534e-05,
      "loss": 2.5346,
      "theoretical_loss": 3.3605048580814616,
      "tokens_seen": 2576023552
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.43107045418071e-05,
      "loss": 2.5056,
      "theoretical_loss": 3.3604913011082846,
      "tokens_seen": 2576154624
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.430268014764886e-05,
      "loss": 2.5873,
      "theoretical_loss": 3.3604777450179775,
      "tokens_seen": 2576285696
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.429465575349062e-05,
      "loss": 2.536,
      "theoretical_loss": 3.360464189810438,
      "tokens_seen": 2576416768
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.428663135933237e-05,
      "loss": 2.5434,
      "theoretical_loss": 3.3604506354855626,
      "tokens_seen": 2576547840
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.427860696517413e-05,
      "loss": 2.6844,
      "theoretical_loss": 3.3604370820432496,
      "tokens_seen": 2576678912
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.427058257101589e-05,
      "loss": 2.748,
      "theoretical_loss": 3.360423529483397,
      "tokens_seen": 2576809984
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.426255817685765e-05,
      "loss": 2.6688,
      "theoretical_loss": 3.3604099778059022,
      "tokens_seen": 2576941056
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.425453378269941e-05,
      "loss": 2.4215,
      "theoretical_loss": 3.3603964270106625,
      "tokens_seen": 2577072128
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.424650938854117e-05,
      "loss": 2.6003,
      "theoretical_loss": 3.3603828770975763,
      "tokens_seen": 2577203200
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1412403,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.7557308673858643,
      "objective/train/theoretical_loss": 3.3603693280665405,
      "objective/train/tokens_used": 947793376,
      "theoretical_loss": 3.3603693280665405,
      "tokens_seen": 2577334272
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.423848499438293e-05,
      "loss": 2.5315,
      "theoretical_loss": 3.3603693280665405,
      "tokens_seen": 2577334272
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4230460600224684e-05,
      "loss": 2.4856,
      "theoretical_loss": 3.3603557799174535,
      "tokens_seen": 2577465344
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4222436206066444e-05,
      "loss": 2.4088,
      "theoretical_loss": 3.360342232650213,
      "tokens_seen": 2577596416
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4214411811908203e-05,
      "loss": 2.5578,
      "theoretical_loss": 3.360328686264716,
      "tokens_seen": 2577727488
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.420638741774996e-05,
      "loss": 2.5137,
      "theoretical_loss": 3.3603151407608616,
      "tokens_seen": 2577858560
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.419836302359172e-05,
      "loss": 2.6998,
      "theoretical_loss": 3.3603015961385463,
      "tokens_seen": 2577989632
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.419033862943348e-05,
      "loss": 2.4891,
      "theoretical_loss": 3.3602880523976686,
      "tokens_seen": 2578120704
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.418231423527524e-05,
      "loss": 2.71,
      "theoretical_loss": 3.360274509538126,
      "tokens_seen": 2578251776
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4174289841117e-05,
      "loss": 2.6158,
      "theoretical_loss": 3.3602609675598165,
      "tokens_seen": 2578382848
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4166265446958756e-05,
      "loss": 2.5242,
      "theoretical_loss": 3.3602474264626383,
      "tokens_seen": 2578513920
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4158241052800516e-05,
      "loss": 2.4932,
      "theoretical_loss": 3.3602338862464887,
      "tokens_seen": 2578644992
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4150216658642276e-05,
      "loss": 2.6617,
      "theoretical_loss": 3.360220346911266,
      "tokens_seen": 2578776064
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4142192264484036e-05,
      "loss": 2.4997,
      "theoretical_loss": 3.3602068084568684,
      "tokens_seen": 2578907136
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1412775,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.234215259552002,
      "objective/train/theoretical_loss": 3.3602000395599463,
      "objective/train/tokens_used": 949431776,
      "theoretical_loss": 3.3602000395599463,
      "tokens_seen": 2578972672
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4134167870325796e-05,
      "loss": 2.5342,
      "theoretical_loss": 3.360193270883193,
      "tokens_seen": 2579038208
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4126143476167556e-05,
      "loss": 2.564,
      "theoretical_loss": 3.360179734190138,
      "tokens_seen": 2579169280
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.4118119082009316e-05,
      "loss": 2.6898,
      "theoretical_loss": 3.360166198377602,
      "tokens_seen": 2579300352
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.411009468785107e-05,
      "loss": 2.4675,
      "theoretical_loss": 3.360152663445482,
      "tokens_seen": 2579431424
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.410207029369283e-05,
      "loss": 2.5835,
      "theoretical_loss": 3.3601391293936764,
      "tokens_seen": 2579562496
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.409404589953459e-05,
      "loss": 2.5584,
      "theoretical_loss": 3.360125596222084,
      "tokens_seen": 2579693568
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.408602150537635e-05,
      "loss": 2.5646,
      "theoretical_loss": 3.360112063930602,
      "tokens_seen": 2579824640
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.407799711121811e-05,
      "loss": 2.6259,
      "theoretical_loss": 3.3600985325191286,
      "tokens_seen": 2579955712
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.406997271705987e-05,
      "loss": 2.5303,
      "theoretical_loss": 3.360085001987562,
      "tokens_seen": 2580086784
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.406194832290163e-05,
      "loss": 2.7238,
      "theoretical_loss": 3.3600714723358,
      "tokens_seen": 2580217856
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.405392392874339e-05,
      "loss": 2.5707,
      "theoretical_loss": 3.3600579435637408,
      "tokens_seen": 2580348928
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.404589953458514e-05,
      "loss": 2.7141,
      "theoretical_loss": 3.3600444156712825,
      "tokens_seen": 2580480000
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1413702,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.833010673522949,
      "objective/train/theoretical_loss": 3.3600308886583234,
      "objective/train/tokens_used": 951070176,
      "theoretical_loss": 3.3600308886583234,
      "tokens_seen": 2580611072
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.40378751404269e-05,
      "loss": 2.6625,
      "theoretical_loss": 3.3600308886583234,
      "tokens_seen": 2580611072
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.402985074626866e-05,
      "loss": 2.4369,
      "theoretical_loss": 3.3600173625247622,
      "tokens_seen": 2580742144
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.402182635211042e-05,
      "loss": 2.6426,
      "theoretical_loss": 3.360003837270496,
      "tokens_seen": 2580873216
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.401380195795218e-05,
      "loss": 2.4289,
      "theoretical_loss": 3.359990312895423,
      "tokens_seen": 2581004288
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.400577756379394e-05,
      "loss": 2.5973,
      "theoretical_loss": 3.3599767893994428,
      "tokens_seen": 2581135360
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.3997753169635694e-05,
      "loss": 2.4835,
      "theoretical_loss": 3.359963266782452,
      "tokens_seen": 2581266432
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.3989728775477454e-05,
      "loss": 2.5595,
      "theoretical_loss": 3.35994974504435,
      "tokens_seen": 2581397504
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.3981704381319214e-05,
      "loss": 2.6413,
      "theoretical_loss": 3.3599362241850343,
      "tokens_seen": 2581528576
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.3973679987160974e-05,
      "loss": 2.5503,
      "theoretical_loss": 3.3599227042044033,
      "tokens_seen": 2581659648
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.3965655593002733e-05,
      "loss": 2.5183,
      "theoretical_loss": 3.3599091851023557,
      "tokens_seen": 2581790720
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.395763119884449e-05,
      "loss": 2.4992,
      "theoretical_loss": 3.35989566687879,
      "tokens_seen": 2581921792
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.3949606804686246e-05,
      "loss": 2.5265,
      "theoretical_loss": 3.359882149533603,
      "tokens_seen": 2582052864
    },
    {
      "epoch": 0.56,
      "learning_rate": 4.3941582410528006e-05,
      "loss": 2.6401,
      "theoretical_loss": 3.359868633066695,
      "tokens_seen": 2582183936
    },
    {
      "epoch": 0.56,
      "objective/train/docs_used": 1414384,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.684948682785034,
      "objective/train/theoretical_loss": 3.3598618751625633,
      "objective/train/tokens_used": 952708576,
      "theoretical_loss": 3.3598618751625633,
      "tokens_seen": 2582249472
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3933558016369766e-05,
      "loss": 2.6089,
      "theoretical_loss": 3.359855117477963,
      "tokens_seen": 2582315008
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3925533622211526e-05,
      "loss": 2.366,
      "theoretical_loss": 3.359841602767306,
      "tokens_seen": 2582446080
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3917509228053286e-05,
      "loss": 2.4965,
      "theoretical_loss": 3.3598280889346226,
      "tokens_seen": 2582577152
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3909484833895046e-05,
      "loss": 2.5821,
      "theoretical_loss": 3.3598145759798106,
      "tokens_seen": 2582708224
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.39014604397368e-05,
      "loss": 2.6362,
      "theoretical_loss": 3.3598010639027684,
      "tokens_seen": 2582839296
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.389343604557856e-05,
      "loss": 2.6019,
      "theoretical_loss": 3.359787552703395,
      "tokens_seen": 2582970368
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.388541165142032e-05,
      "loss": 2.5663,
      "theoretical_loss": 3.3597740423815883,
      "tokens_seen": 2583101440
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.387738725726208e-05,
      "loss": 2.5881,
      "theoretical_loss": 3.3597605329372473,
      "tokens_seen": 2583232512
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.386936286310384e-05,
      "loss": 2.6147,
      "theoretical_loss": 3.3597470243702703,
      "tokens_seen": 2583363584
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.38613384689456e-05,
      "loss": 2.4081,
      "theoretical_loss": 3.3597335166805555,
      "tokens_seen": 2583494656
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.385331407478735e-05,
      "loss": 2.5283,
      "theoretical_loss": 3.359720009868002,
      "tokens_seen": 2583625728
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.384528968062911e-05,
      "loss": 2.4299,
      "theoretical_loss": 3.3597065039325074,
      "tokens_seen": 2583756800
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1415698,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.223771810531616,
      "objective/train/theoretical_loss": 3.3596929988739714,
      "objective/train/tokens_used": 954346976,
      "theoretical_loss": 3.3596929988739714,
      "tokens_seen": 2583887872
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.383726528647087e-05,
      "loss": 2.4919,
      "theoretical_loss": 3.3596929988739714,
      "tokens_seen": 2583887872
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.382924089231263e-05,
      "loss": 2.638,
      "theoretical_loss": 3.3596794946922923,
      "tokens_seen": 2584018944
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.382121649815439e-05,
      "loss": 2.598,
      "theoretical_loss": 3.359665991387368,
      "tokens_seen": 2584150016
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.381319210399615e-05,
      "loss": 2.5097,
      "theoretical_loss": 3.3596524889590977,
      "tokens_seen": 2584281088
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3805167709837904e-05,
      "loss": 2.5365,
      "theoretical_loss": 3.3596389874073798,
      "tokens_seen": 2584412160
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3797143315679664e-05,
      "loss": 2.4864,
      "theoretical_loss": 3.359625486732113,
      "tokens_seen": 2584543232
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3789118921521424e-05,
      "loss": 2.51,
      "theoretical_loss": 3.359611986933196,
      "tokens_seen": 2584674304
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3781094527363184e-05,
      "loss": 2.5867,
      "theoretical_loss": 3.3595984880105276,
      "tokens_seen": 2584805376
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3773070133204944e-05,
      "loss": 2.4277,
      "theoretical_loss": 3.3595849899640067,
      "tokens_seen": 2584936448
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3765045739046704e-05,
      "loss": 2.6637,
      "theoretical_loss": 3.359571492793531,
      "tokens_seen": 2585067520
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.375702134488846e-05,
      "loss": 2.4325,
      "theoretical_loss": 3.3595579964990003,
      "tokens_seen": 2585198592
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.374899695073022e-05,
      "loss": 2.6014,
      "theoretical_loss": 3.359544501080313,
      "tokens_seen": 2585329664
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.374097255657198e-05,
      "loss": 2.5262,
      "theoretical_loss": 3.3595310065373676,
      "tokens_seen": 2585460736
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1416170,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.412297010421753,
      "objective/train/theoretical_loss": 3.3595242595942665,
      "objective/train/tokens_used": 955985376,
      "theoretical_loss": 3.3595242595942665,
      "tokens_seen": 2585526272
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.373294816241374e-05,
      "loss": 2.5099,
      "theoretical_loss": 3.359517512870063,
      "tokens_seen": 2585591808
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.37249237682555e-05,
      "loss": 2.5199,
      "theoretical_loss": 3.3595040200782984,
      "tokens_seen": 2585722880
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.371689937409726e-05,
      "loss": 2.4225,
      "theoretical_loss": 3.3594905281619725,
      "tokens_seen": 2585853952
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3708874979939017e-05,
      "loss": 2.3979,
      "theoretical_loss": 3.3594770371209837,
      "tokens_seen": 2585985024
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3700850585780776e-05,
      "loss": 2.5816,
      "theoretical_loss": 3.359463546955231,
      "tokens_seen": 2586116096
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.369282619162253e-05,
      "loss": 2.4065,
      "theoretical_loss": 3.3594500576646134,
      "tokens_seen": 2586247168
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.368480179746429e-05,
      "loss": 2.4668,
      "theoretical_loss": 3.35943656924903,
      "tokens_seen": 2586378240
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.367677740330605e-05,
      "loss": 2.5211,
      "theoretical_loss": 3.3594230817083788,
      "tokens_seen": 2586509312
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.366875300914781e-05,
      "loss": 2.5312,
      "theoretical_loss": 3.3594095950425595,
      "tokens_seen": 2586640384
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.366072861498957e-05,
      "loss": 2.3878,
      "theoretical_loss": 3.3593961092514713,
      "tokens_seen": 2586771456
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.365270422083133e-05,
      "loss": 2.4908,
      "theoretical_loss": 3.3593826243350127,
      "tokens_seen": 2586902528
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.364467982667309e-05,
      "loss": 2.6214,
      "theoretical_loss": 3.359369140293082,
      "tokens_seen": 2587033600
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1417329,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5255184173583984,
      "objective/train/theoretical_loss": 3.3593556571255796,
      "objective/train/tokens_used": 957623776,
      "theoretical_loss": 3.3593556571255796,
      "tokens_seen": 2587164672
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.363665543251484e-05,
      "loss": 2.6045,
      "theoretical_loss": 3.3593556571255796,
      "tokens_seen": 2587164672
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.36286310383566e-05,
      "loss": 2.4828,
      "theoretical_loss": 3.3593421748324035,
      "tokens_seen": 2587295744
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.362060664419836e-05,
      "loss": 2.5751,
      "theoretical_loss": 3.359328693413453,
      "tokens_seen": 2587426816
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.361258225004012e-05,
      "loss": 2.4211,
      "theoretical_loss": 3.359315212868627,
      "tokens_seen": 2587557888
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.360455785588188e-05,
      "loss": 2.6507,
      "theoretical_loss": 3.359301733197825,
      "tokens_seen": 2587688960
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.359653346172364e-05,
      "loss": 2.5545,
      "theoretical_loss": 3.3592882544009455,
      "tokens_seen": 2587820032
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.35885090675654e-05,
      "loss": 2.5015,
      "theoretical_loss": 3.3592747764778874,
      "tokens_seen": 2587951104
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.358048467340716e-05,
      "loss": 2.5675,
      "theoretical_loss": 3.3592612994285505,
      "tokens_seen": 2588082176
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3572460279248915e-05,
      "loss": 2.5841,
      "theoretical_loss": 3.3592478232528338,
      "tokens_seen": 2588213248
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3564435885090675e-05,
      "loss": 2.5749,
      "theoretical_loss": 3.359234347950636,
      "tokens_seen": 2588344320
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3556411490932434e-05,
      "loss": 2.5033,
      "theoretical_loss": 3.3592208735218567,
      "tokens_seen": 2588475392
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3548387096774194e-05,
      "loss": 2.4545,
      "theoretical_loss": 3.3592073999663947,
      "tokens_seen": 2588606464
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3540362702615954e-05,
      "loss": 2.6047,
      "theoretical_loss": 3.3591939272841493,
      "tokens_seen": 2588737536
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1417937,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.3127598762512207,
      "objective/train/theoretical_loss": 3.359187191270452,
      "objective/train/tokens_used": 959262176,
      "theoretical_loss": 3.359187191270452,
      "tokens_seen": 2588803072
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3532338308457714e-05,
      "loss": 2.472,
      "theoretical_loss": 3.35918045547502,
      "tokens_seen": 2588868608
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3524313914299474e-05,
      "loss": 2.6133,
      "theoretical_loss": 3.359166984538906,
      "tokens_seen": 2588999680
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.351628952014123e-05,
      "loss": 2.2738,
      "theoretical_loss": 3.3591535144757056,
      "tokens_seen": 2589130752
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.350826512598299e-05,
      "loss": 2.6542,
      "theoretical_loss": 3.3591400452853195,
      "tokens_seen": 2589261824
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.350024073182475e-05,
      "loss": 2.4462,
      "theoretical_loss": 3.3591265769676455,
      "tokens_seen": 2589392896
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.349221633766651e-05,
      "loss": 2.5233,
      "theoretical_loss": 3.359113109522584,
      "tokens_seen": 2589523968
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.348419194350827e-05,
      "loss": 2.5961,
      "theoretical_loss": 3.3590996429500337,
      "tokens_seen": 2589655040
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.347616754935003e-05,
      "loss": 2.3866,
      "theoretical_loss": 3.3590861772498943,
      "tokens_seen": 2589786112
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3468143155191787e-05,
      "loss": 2.6735,
      "theoretical_loss": 3.3590727124220647,
      "tokens_seen": 2589917184
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.346011876103354e-05,
      "loss": 2.5495,
      "theoretical_loss": 3.3590592484664445,
      "tokens_seen": 2590048256
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.34520943668753e-05,
      "loss": 2.5255,
      "theoretical_loss": 3.359045785382933,
      "tokens_seen": 2590179328
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.344406997271706e-05,
      "loss": 2.5271,
      "theoretical_loss": 3.35903232317143,
      "tokens_seen": 2590310400
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1418406,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.977363109588623,
      "objective/train/theoretical_loss": 3.359018861831834,
      "objective/train/tokens_used": 960900576,
      "theoretical_loss": 3.359018861831834,
      "tokens_seen": 2590441472
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.343604557855882e-05,
      "loss": 2.6362,
      "theoretical_loss": 3.359018861831834,
      "tokens_seen": 2590441472
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.342802118440058e-05,
      "loss": 2.4466,
      "theoretical_loss": 3.359005401364046,
      "tokens_seen": 2590572544
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.341999679024234e-05,
      "loss": 2.5624,
      "theoretical_loss": 3.3589919417679632,
      "tokens_seen": 2590703616
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.34119723960841e-05,
      "loss": 2.4344,
      "theoretical_loss": 3.3589784830434866,
      "tokens_seen": 2590834688
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.340394800192586e-05,
      "loss": 2.6116,
      "theoretical_loss": 3.3589650251905154,
      "tokens_seen": 2590965760
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.339592360776761e-05,
      "loss": 2.5694,
      "theoretical_loss": 3.358951568208949,
      "tokens_seen": 2591096832
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.338789921360937e-05,
      "loss": 2.5525,
      "theoretical_loss": 3.358938112098687,
      "tokens_seen": 2591227904
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.337987481945113e-05,
      "loss": 2.5562,
      "theoretical_loss": 3.358924656859629,
      "tokens_seen": 2591358976
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.337185042529289e-05,
      "loss": 2.3901,
      "theoretical_loss": 3.3589112024916736,
      "tokens_seen": 2591490048
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.336382603113465e-05,
      "loss": 2.3919,
      "theoretical_loss": 3.3588977489947216,
      "tokens_seen": 2591621120
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.335580163697641e-05,
      "loss": 2.5913,
      "theoretical_loss": 3.358884296368672,
      "tokens_seen": 2591752192
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.334777724281817e-05,
      "loss": 2.5002,
      "theoretical_loss": 3.358870844613424,
      "tokens_seen": 2591883264
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3339752848659925e-05,
      "loss": 2.4822,
      "theoretical_loss": 3.3588573937288784,
      "tokens_seen": 2592014336
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1419620,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.6387507915496826,
      "objective/train/theoretical_loss": 3.358850668613087,
      "objective/train/tokens_used": 962538976,
      "theoretical_loss": 3.358850668613087,
      "tokens_seen": 2592079872
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3331728454501685e-05,
      "loss": 2.5499,
      "theoretical_loss": 3.3588439437149336,
      "tokens_seen": 2592145408
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3323704060343445e-05,
      "loss": 2.481,
      "theoretical_loss": 3.3588304945714897,
      "tokens_seen": 2592276480
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3315679666185204e-05,
      "loss": 2.4989,
      "theoretical_loss": 3.3588170462984466,
      "tokens_seen": 2592407552
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3307655272026964e-05,
      "loss": 2.61,
      "theoretical_loss": 3.3588035988957032,
      "tokens_seen": 2592538624
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3299630877868724e-05,
      "loss": 2.4469,
      "theoretical_loss": 3.3587901523631603,
      "tokens_seen": 2592669696
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3291606483710484e-05,
      "loss": 2.5985,
      "theoretical_loss": 3.3587767067007164,
      "tokens_seen": 2592800768
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.328358208955224e-05,
      "loss": 2.4818,
      "theoretical_loss": 3.358763261908272,
      "tokens_seen": 2592931840
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3275557695394e-05,
      "loss": 2.5826,
      "theoretical_loss": 3.3587498179857267,
      "tokens_seen": 2593062912
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.326753330123576e-05,
      "loss": 2.4678,
      "theoretical_loss": 3.35873637493298,
      "tokens_seen": 2593193984
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.325950890707752e-05,
      "loss": 2.5931,
      "theoretical_loss": 3.3587229327499326,
      "tokens_seen": 2593325056
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.325148451291928e-05,
      "loss": 2.5536,
      "theoretical_loss": 3.358709491436483,
      "tokens_seen": 2593456128
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.324346011876104e-05,
      "loss": 2.4584,
      "theoretical_loss": 3.358696050992531,
      "tokens_seen": 2593587200
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1420064,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.8486673831939697,
      "objective/train/theoretical_loss": 3.3586826114179775,
      "objective/train/tokens_used": 964177376,
      "theoretical_loss": 3.3586826114179775,
      "tokens_seen": 2593718272
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.32354357246028e-05,
      "loss": 2.4882,
      "theoretical_loss": 3.3586826114179775,
      "tokens_seen": 2593718272
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.322741133044456e-05,
      "loss": 2.4858,
      "theoretical_loss": 3.358669172712722,
      "tokens_seen": 2593849344
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.321938693628631e-05,
      "loss": 2.4799,
      "theoretical_loss": 3.3586557348766637,
      "tokens_seen": 2593980416
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.321136254212807e-05,
      "loss": 2.4642,
      "theoretical_loss": 3.358642297909703,
      "tokens_seen": 2594111488
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.320333814796983e-05,
      "loss": 2.3856,
      "theoretical_loss": 3.3586288618117397,
      "tokens_seen": 2594242560
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.319531375381159e-05,
      "loss": 2.554,
      "theoretical_loss": 3.3586154265826735,
      "tokens_seen": 2594373632
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.318728935965335e-05,
      "loss": 2.3107,
      "theoretical_loss": 3.358601992222405,
      "tokens_seen": 2594504704
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.317926496549511e-05,
      "loss": 2.5941,
      "theoretical_loss": 3.3585885587308333,
      "tokens_seen": 2594635776
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.317124057133687e-05,
      "loss": 2.4145,
      "theoretical_loss": 3.3585751261078585,
      "tokens_seen": 2594766848
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.316321617717862e-05,
      "loss": 2.5545,
      "theoretical_loss": 3.358561694353381,
      "tokens_seen": 2594897920
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.315519178302038e-05,
      "loss": 2.4887,
      "theoretical_loss": 3.3585482634673003,
      "tokens_seen": 2595028992
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.314716738886214e-05,
      "loss": 2.8372,
      "theoretical_loss": 3.3585348334495166,
      "tokens_seen": 2595160064
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.31391429947039e-05,
      "loss": 2.4657,
      "theoretical_loss": 3.35852140429993,
      "tokens_seen": 2595291136
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1420982,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5949935913085938,
      "objective/train/theoretical_loss": 3.3585146900506793,
      "objective/train/tokens_used": 965815776,
      "theoretical_loss": 3.3585146900506793,
      "tokens_seen": 2595356672
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.313111860054566e-05,
      "loss": 2.6028,
      "theoretical_loss": 3.3585079760184406,
      "tokens_seen": 2595422208
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.312309420638742e-05,
      "loss": 2.5677,
      "theoretical_loss": 3.3584945486049484,
      "tokens_seen": 2595553280
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.311506981222918e-05,
      "loss": 2.5317,
      "theoretical_loss": 3.358481122059353,
      "tokens_seen": 2595684352
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.310704541807094e-05,
      "loss": 2.55,
      "theoretical_loss": 3.358467696381555,
      "tokens_seen": 2595815424
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3099021023912695e-05,
      "loss": 2.5328,
      "theoretical_loss": 3.3584542715714543,
      "tokens_seen": 2595946496
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3090996629754455e-05,
      "loss": 2.4475,
      "theoretical_loss": 3.3584408476289513,
      "tokens_seen": 2596077568
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3082972235596215e-05,
      "loss": 2.4731,
      "theoretical_loss": 3.358427424553946,
      "tokens_seen": 2596208640
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3074947841437975e-05,
      "loss": 2.5151,
      "theoretical_loss": 3.358414002346338,
      "tokens_seen": 2596339712
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3066923447279734e-05,
      "loss": 2.5764,
      "theoretical_loss": 3.3584005810060282,
      "tokens_seen": 2596470784
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3058899053121494e-05,
      "loss": 2.5897,
      "theoretical_loss": 3.3583871605329163,
      "tokens_seen": 2596601856
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.3050874658963254e-05,
      "loss": 2.5039,
      "theoretical_loss": 3.358373740926903,
      "tokens_seen": 2596732928
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.304285026480501e-05,
      "loss": 2.5007,
      "theoretical_loss": 3.358360322187888,
      "tokens_seen": 2596864000
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1421571,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.679337978363037,
      "objective/train/theoretical_loss": 3.3583469043157717,
      "objective/train/tokens_used": 967454176,
      "theoretical_loss": 3.3583469043157717,
      "tokens_seen": 2596995072
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.303482587064677e-05,
      "loss": 2.5579,
      "theoretical_loss": 3.3583469043157717,
      "tokens_seen": 2596995072
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.302680147648853e-05,
      "loss": 2.4033,
      "theoretical_loss": 3.3583334873104547,
      "tokens_seen": 2597126144
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.301877708233029e-05,
      "loss": 2.65,
      "theoretical_loss": 3.3583200711718364,
      "tokens_seen": 2597257216
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.301075268817205e-05,
      "loss": 2.4555,
      "theoretical_loss": 3.358306655899818,
      "tokens_seen": 2597388288
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.300272829401381e-05,
      "loss": 2.3422,
      "theoretical_loss": 3.3582932414942994,
      "tokens_seen": 2597519360
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.299470389985557e-05,
      "loss": 2.4833,
      "theoretical_loss": 3.358279827955181,
      "tokens_seen": 2597650432
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.298667950569732e-05,
      "loss": 2.432,
      "theoretical_loss": 3.3582664152823627,
      "tokens_seen": 2597781504
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.297865511153908e-05,
      "loss": 2.5715,
      "theoretical_loss": 3.3582530034757454,
      "tokens_seen": 2597912576
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.297063071738084e-05,
      "loss": 2.4453,
      "theoretical_loss": 3.3582395925352295,
      "tokens_seen": 2598043648
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.29626063232226e-05,
      "loss": 2.5452,
      "theoretical_loss": 3.3582261824607147,
      "tokens_seen": 2598174720
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.295458192906436e-05,
      "loss": 2.5757,
      "theoretical_loss": 3.3582127732521023,
      "tokens_seen": 2598305792
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.294655753490612e-05,
      "loss": 2.4751,
      "theoretical_loss": 3.358199364909292,
      "tokens_seen": 2598436864
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.293853314074788e-05,
      "loss": 2.3885,
      "theoretical_loss": 3.3581859574321844,
      "tokens_seen": 2598567936
    },
    {
      "epoch": 0.57,
      "objective/train/docs_used": 1422608,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.032858371734619,
      "objective/train/theoretical_loss": 3.3581792540182382,
      "objective/train/tokens_used": 969092576,
      "theoretical_loss": 3.3581792540182382,
      "tokens_seen": 2598633472
    },
    {
      "epoch": 0.57,
      "learning_rate": 4.293050874658964e-05,
      "loss": 2.425,
      "theoretical_loss": 3.35817255082068,
      "tokens_seen": 2598699008
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.292248435243139e-05,
      "loss": 2.3444,
      "theoretical_loss": 3.3581591450746795,
      "tokens_seen": 2598830080
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.291445995827315e-05,
      "loss": 2.442,
      "theoretical_loss": 3.358145740194083,
      "tokens_seen": 2598961152
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.290643556411491e-05,
      "loss": 2.5509,
      "theoretical_loss": 3.3581323361787914,
      "tokens_seen": 2599092224
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.289841116995667e-05,
      "loss": 2.4825,
      "theoretical_loss": 3.3581189330287047,
      "tokens_seen": 2599223296
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.289038677579843e-05,
      "loss": 2.6839,
      "theoretical_loss": 3.3581055307437238,
      "tokens_seen": 2599354368
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.288236238164019e-05,
      "loss": 2.681,
      "theoretical_loss": 3.358092129323749,
      "tokens_seen": 2599485440
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.287433798748195e-05,
      "loss": 2.578,
      "theoretical_loss": 3.358078728768681,
      "tokens_seen": 2599616512
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2866313593323705e-05,
      "loss": 2.5397,
      "theoretical_loss": 3.358065329078421,
      "tokens_seen": 2599747584
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2858289199165465e-05,
      "loss": 2.7148,
      "theoretical_loss": 3.358051930252868,
      "tokens_seen": 2599878656
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2850264805007225e-05,
      "loss": 2.5636,
      "theoretical_loss": 3.358038532291924,
      "tokens_seen": 2600009728
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2842240410848985e-05,
      "loss": 2.4291,
      "theoretical_loss": 3.358025135195489,
      "tokens_seen": 2600140800
    },
    {
      "epoch": 0.58,
      "objective/train/docs_used": 1423091,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.546130657196045,
      "objective/train/theoretical_loss": 3.3580117389634645,
      "objective/train/tokens_used": 970730976,
      "theoretical_loss": 3.3580117389634645,
      "tokens_seen": 2600271872
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2834216016690745e-05,
      "loss": 2.5889,
      "theoretical_loss": 3.3580117389634645,
      "tokens_seen": 2600271872
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2826191622532505e-05,
      "loss": 2.4526,
      "theoretical_loss": 3.35799834359575,
      "tokens_seen": 2600402944
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2818167228374264e-05,
      "loss": 2.5043,
      "theoretical_loss": 3.3579849490922467,
      "tokens_seen": 2600534016
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.281014283421602e-05,
      "loss": 2.4791,
      "theoretical_loss": 3.3579715554528553,
      "tokens_seen": 2600665088
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.280211844005778e-05,
      "loss": 2.5038,
      "theoretical_loss": 3.3579581626774764,
      "tokens_seen": 2600796160
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.279409404589954e-05,
      "loss": 2.4278,
      "theoretical_loss": 3.357944770766011,
      "tokens_seen": 2600927232
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.27860696517413e-05,
      "loss": 2.5213,
      "theoretical_loss": 3.3579313797183596,
      "tokens_seen": 2601058304
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.277804525758306e-05,
      "loss": 2.5781,
      "theoretical_loss": 3.357917989534423,
      "tokens_seen": 2601189376
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.277002086342482e-05,
      "loss": 2.515,
      "theoretical_loss": 3.3579046002141015,
      "tokens_seen": 2601320448
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.276199646926658e-05,
      "loss": 2.5409,
      "theoretical_loss": 3.3578912117572965,
      "tokens_seen": 2601451520
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.275397207510834e-05,
      "loss": 2.4656,
      "theoretical_loss": 3.3578778241639093,
      "tokens_seen": 2601582592
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.274594768095009e-05,
      "loss": 2.5142,
      "theoretical_loss": 3.3578644374338396,
      "tokens_seen": 2601713664
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.273792328679185e-05,
      "loss": 2.6067,
      "theoretical_loss": 3.357851051566989,
      "tokens_seen": 2601844736
    },
    {
      "epoch": 0.58,
      "objective/train/docs_used": 1424292,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.5811870098114014,
      "objective/train/theoretical_loss": 3.3578443589572395,
      "objective/train/tokens_used": 972369376,
      "theoretical_loss": 3.3578443589572395,
      "tokens_seen": 2601910272
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.272989889263361e-05,
      "loss": 2.5605,
      "theoretical_loss": 3.357837666563258,
      "tokens_seen": 2601975808
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.272187449847537e-05,
      "loss": 2.4923,
      "theoretical_loss": 3.357824282422547,
      "tokens_seen": 2602106880
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.271385010431713e-05,
      "loss": 2.449,
      "theoretical_loss": 3.357810899144758,
      "tokens_seen": 2602237952
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.270582571015889e-05,
      "loss": 2.6668,
      "theoretical_loss": 3.357797516729791,
      "tokens_seen": 2602369024
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.269780131600065e-05,
      "loss": 2.5838,
      "theoretical_loss": 3.357784135177548,
      "tokens_seen": 2602500096
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.26897769218424e-05,
      "loss": 2.7115,
      "theoretical_loss": 3.3577707544879285,
      "tokens_seen": 2602631168
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.268175252768416e-05,
      "loss": 2.5673,
      "theoretical_loss": 3.3577573746608342,
      "tokens_seen": 2602762240
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.267372813352592e-05,
      "loss": 2.5383,
      "theoretical_loss": 3.357743995696166,
      "tokens_seen": 2602893312
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.266570373936768e-05,
      "loss": 2.5499,
      "theoretical_loss": 3.357730617593825,
      "tokens_seen": 2603024384
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.265767934520944e-05,
      "loss": 2.5093,
      "theoretical_loss": 3.357717240353712,
      "tokens_seen": 2603155456
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.26496549510512e-05,
      "loss": 2.663,
      "theoretical_loss": 3.3577038639757286,
      "tokens_seen": 2603286528
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2641630556892955e-05,
      "loss": 2.7217,
      "theoretical_loss": 3.357690488459775,
      "tokens_seen": 2603417600
    },
    {
      "epoch": 0.58,
      "objective/train/docs_used": 1424790,
      "objective/train/instantaneous_batch_size": 16,
      "objective/train/instantaneous_microbatch_size": 16384,
      "objective/train/original_loss": 2.476832866668701,
      "objective/train/theoretical_loss": 3.3576771138057526,
      "objective/train/tokens_used": 974007776,
      "theoretical_loss": 3.3576771138057526,
      "tokens_seen": 2603548672
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2633606162734715e-05,
      "loss": 2.5643,
      "theoretical_loss": 3.3576771138057526,
      "tokens_seen": 2603548672
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2625581768576475e-05,
      "loss": 2.4756,
      "theoretical_loss": 3.357663740013563,
      "tokens_seen": 2603679744
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2617557374418235e-05,
      "loss": 2.5981,
      "theoretical_loss": 3.3576503670831057,
      "tokens_seen": 2603810816
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2609532980259995e-05,
      "loss": 2.5278,
      "theoretical_loss": 3.3576369950142837,
      "tokens_seen": 2603941888
    },
    {
      "epoch": 0.58,
      "learning_rate": 4.2601508586101755e-05,
      "loss": 2.5158,
      "theoretical_loss": 3.3576236238069974,
      "tokens_seen": 2604072960
    }
  ],
  "max_steps": 12588,
  "num_train_epochs": 9223372036854775807,
  "total_flos": 4.86897843437568e+17,
  "trial_name": null,
  "trial_params": null
}