{
  "best_metric": 4.204440593719482,
  "best_model_checkpoint": "/mmfs1/gscratch/stf/abhinavp/corpus-filtering/outputs/existential-there-quantifier/lstm/0/checkpoints/checkpoint-381600",
  "epoch": 0.025000606015738065,
  "eval_steps": 10,
  "global_step": 381600,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0,
      "learning_rate": 4.999998362119627e-05,
      "loss": 10.8202,
      "step": 1
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.999161405248948e-05,
      "loss": 7.5501,
      "step": 512
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.998322810497896e-05,
      "loss": 7.0601,
      "step": 1024
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.997484215746844e-05,
      "loss": 6.9987,
      "step": 1536
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.996645620995792e-05,
      "loss": 6.9435,
      "step": 2048
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.99580702624474e-05,
      "loss": 6.8551,
      "step": 2560
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.994968431493688e-05,
      "loss": 6.716,
      "step": 3072
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.994129836742636e-05,
      "loss": 6.6125,
      "step": 3584
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.993291241991584e-05,
      "loss": 6.5199,
      "step": 4096
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.992452647240532e-05,
      "loss": 6.4481,
      "step": 4608
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.99161405248948e-05,
      "loss": 6.382,
      "step": 5120
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.990775457738428e-05,
      "loss": 6.318,
      "step": 5632
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.989936862987376e-05,
      "loss": 6.248,
      "step": 6144
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.989098268236324e-05,
      "loss": 6.1881,
      "step": 6656
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.988259673485272e-05,
      "loss": 6.1322,
      "step": 7168
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.98742107873422e-05,
      "loss": 6.0831,
      "step": 7680
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.986582483983168e-05,
      "loss": 6.0396,
      "step": 8192
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.985743889232116e-05,
      "loss": 6.0082,
      "step": 8704
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.984905294481064e-05,
      "loss": 5.9595,
      "step": 9216
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.984066699730012e-05,
      "loss": 5.9276,
      "step": 9728
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.983229742859333e-05,
      "loss": 5.8946,
      "step": 10240
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.982391148108281e-05,
      "loss": 5.8473,
      "step": 10752
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.981552553357229e-05,
      "loss": 5.8181,
      "step": 11264
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.980713958606178e-05,
      "loss": 5.7865,
      "step": 11776
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9798770017354986e-05,
      "loss": 5.7688,
      "step": 12288
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9790384069844466e-05,
      "loss": 5.7329,
      "step": 12800
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9781998122333946e-05,
      "loss": 5.7049,
      "step": 13312
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9773612174823426e-05,
      "loss": 5.686,
      "step": 13824
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9765258984920364e-05,
      "loss": 5.6602,
      "step": 14336
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9756873037409844e-05,
      "loss": 5.6315,
      "step": 14848
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9748487089899324e-05,
      "loss": 5.6079,
      "step": 15360
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9740117521192533e-05,
      "loss": 5.5989,
      "step": 15872
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9731731573682013e-05,
      "loss": 5.5766,
      "step": 16384
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.972336200497523e-05,
      "loss": 5.5671,
      "step": 16896
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.971497605746471e-05,
      "loss": 5.5462,
      "step": 17408
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.970659010995419e-05,
      "loss": 5.5259,
      "step": 17920
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.969820416244367e-05,
      "loss": 5.5215,
      "step": 18432
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.968981821493315e-05,
      "loss": 5.4825,
      "step": 18944
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.968143226742263e-05,
      "loss": 5.4722,
      "step": 19456
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.967304631991211e-05,
      "loss": 5.4539,
      "step": 19968
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.966466037240159e-05,
      "loss": 5.4479,
      "step": 20480
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.965627442489107e-05,
      "loss": 5.4242,
      "step": 20992
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.964788847738054e-05,
      "loss": 5.4303,
      "step": 21504
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.963950252987002e-05,
      "loss": 5.3955,
      "step": 22016
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.96311165823595e-05,
      "loss": 5.3949,
      "step": 22528
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.962273063484898e-05,
      "loss": 5.3894,
      "step": 23040
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.96143610661422e-05,
      "loss": 5.3811,
      "step": 23552
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.960597511863168e-05,
      "loss": 5.3721,
      "step": 24064
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.959758917112116e-05,
      "loss": 5.3486,
      "step": 24576
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.958920322361064e-05,
      "loss": 5.329,
      "step": 25088
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.958081727610012e-05,
      "loss": 5.3417,
      "step": 25600
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.957244770739333e-05,
      "loss": 5.3217,
      "step": 26112
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.956406175988281e-05,
      "loss": 5.3079,
      "step": 26624
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.955567581237229e-05,
      "loss": 5.2908,
      "step": 27136
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9547306243665496e-05,
      "loss": 5.2977,
      "step": 27648
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9538920296154976e-05,
      "loss": 5.2752,
      "step": 28160
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9530534348644456e-05,
      "loss": 5.29,
      "step": 28672
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9522148401133936e-05,
      "loss": 5.2597,
      "step": 29184
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.951376245362342e-05,
      "loss": 5.2517,
      "step": 29696
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.95053765061129e-05,
      "loss": 5.2515,
      "step": 30208
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.949699055860238e-05,
      "loss": 5.2362,
      "step": 30720
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.948860461109186e-05,
      "loss": 5.214,
      "step": 31232
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.948021866358134e-05,
      "loss": 5.2197,
      "step": 31744
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.947184909487455e-05,
      "loss": 5.196,
      "step": 32256
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.946346314736403e-05,
      "loss": 5.1979,
      "step": 32768
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.945507719985351e-05,
      "loss": 5.2075,
      "step": 33280
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.944669125234299e-05,
      "loss": 5.1997,
      "step": 33792
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.943830530483247e-05,
      "loss": 5.1805,
      "step": 34304
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.942991935732195e-05,
      "loss": 5.1615,
      "step": 34816
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.942153340981143e-05,
      "loss": 5.1557,
      "step": 35328
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9413147462300904e-05,
      "loss": 5.1659,
      "step": 35840
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.940477789359412e-05,
      "loss": 5.1552,
      "step": 36352
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9396408324887336e-05,
      "loss": 5.1506,
      "step": 36864
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9388022377376816e-05,
      "loss": 5.1489,
      "step": 37376
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9379636429866296e-05,
      "loss": 5.1484,
      "step": 37888
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9371250482355776e-05,
      "loss": 5.1404,
      "step": 38400
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9362864534845256e-05,
      "loss": 5.1126,
      "step": 38912
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9354494966138465e-05,
      "loss": 5.1165,
      "step": 39424
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9346109018627945e-05,
      "loss": 5.1105,
      "step": 39936
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9337723071117425e-05,
      "loss": 5.0958,
      "step": 40448
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9329337123606905e-05,
      "loss": 5.1147,
      "step": 40960
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.932095117609638e-05,
      "loss": 5.0939,
      "step": 41472
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9312581607389594e-05,
      "loss": 5.0995,
      "step": 41984
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9304195659879074e-05,
      "loss": 5.0842,
      "step": 42496
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9295809712368554e-05,
      "loss": 5.0604,
      "step": 43008
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.928744014366177e-05,
      "loss": 5.0646,
      "step": 43520
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.927905419615125e-05,
      "loss": 5.0693,
      "step": 44032
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.927066824864073e-05,
      "loss": 5.0626,
      "step": 44544
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.92622823011302e-05,
      "loss": 5.0564,
      "step": 45056
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.925389635361968e-05,
      "loss": 5.0435,
      "step": 45568
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.924551040610916e-05,
      "loss": 5.0419,
      "step": 46080
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.923712445859864e-05,
      "loss": 5.0303,
      "step": 46592
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.922873851108812e-05,
      "loss": 5.0242,
      "step": 47104
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.92203525635776e-05,
      "loss": 5.0214,
      "step": 47616
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.921198299487081e-05,
      "loss": 5.014,
      "step": 48128
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.920359704736029e-05,
      "loss": 5.0222,
      "step": 48640
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.919521109984978e-05,
      "loss": 5.0035,
      "step": 49152
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.918682515233926e-05,
      "loss": 4.997,
      "step": 49664
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.917843920482874e-05,
      "loss": 4.9919,
      "step": 50176
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.917006963612195e-05,
      "loss": 4.9932,
      "step": 50688
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.916168368861143e-05,
      "loss": 4.991,
      "step": 51200
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.915329774110091e-05,
      "loss": 4.9799,
      "step": 51712
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.914491179359039e-05,
      "loss": 4.973,
      "step": 52224
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.913652584607987e-05,
      "loss": 4.9576,
      "step": 52736
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.912813989856935e-05,
      "loss": 4.9654,
      "step": 53248
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9119770329862556e-05,
      "loss": 4.9519,
      "step": 53760
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9111384382352036e-05,
      "loss": 4.9468,
      "step": 54272
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9102998434841516e-05,
      "loss": 4.948,
      "step": 54784
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9094612487330996e-05,
      "loss": 4.947,
      "step": 55296
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.908624291862421e-05,
      "loss": 4.941,
      "step": 55808
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.907785697111369e-05,
      "loss": 4.9361,
      "step": 56320
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.906947102360317e-05,
      "loss": 4.926,
      "step": 56832
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.906108507609265e-05,
      "loss": 4.9266,
      "step": 57344
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.905271550738586e-05,
      "loss": 4.9163,
      "step": 57856
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.904432955987534e-05,
      "loss": 4.9113,
      "step": 58368
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.903594361236482e-05,
      "loss": 4.9211,
      "step": 58880
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.90275576648543e-05,
      "loss": 4.9099,
      "step": 59392
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.901917171734378e-05,
      "loss": 4.91,
      "step": 59904
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.901078576983326e-05,
      "loss": 4.9001,
      "step": 60416
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.900239982232274e-05,
      "loss": 4.9076,
      "step": 60928
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8994013874812214e-05,
      "loss": 4.9026,
      "step": 61440
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.89856279273017e-05,
      "loss": 4.8829,
      "step": 61952
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.897724197979118e-05,
      "loss": 4.898,
      "step": 62464
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8968872411084396e-05,
      "loss": 4.8867,
      "step": 62976
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.896048646357387e-05,
      "loss": 4.8873,
      "step": 63488
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.895210051606335e-05,
      "loss": 4.8669,
      "step": 64000
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.894371456855283e-05,
      "loss": 4.8607,
      "step": 64512
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.893532862104231e-05,
      "loss": 4.8558,
      "step": 65024
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.892694267353179e-05,
      "loss": 4.8612,
      "step": 65536
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8918573104825e-05,
      "loss": 4.8478,
      "step": 66048
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.891018715731448e-05,
      "loss": 4.8666,
      "step": 66560
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.890180120980396e-05,
      "loss": 4.8621,
      "step": 67072
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.889341526229344e-05,
      "loss": 4.8444,
      "step": 67584
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.888502931478292e-05,
      "loss": 4.8491,
      "step": 68096
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.88766433672724e-05,
      "loss": 4.8443,
      "step": 68608
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8868273798565614e-05,
      "loss": 4.8437,
      "step": 69120
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8859887851055094e-05,
      "loss": 4.8443,
      "step": 69632
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8851501903544574e-05,
      "loss": 4.8446,
      "step": 70144
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8843115956034054e-05,
      "loss": 4.827,
      "step": 70656
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8834730008523534e-05,
      "loss": 4.8284,
      "step": 71168
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.882636043981674e-05,
      "loss": 4.822,
      "step": 71680
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.881797449230622e-05,
      "loss": 4.8139,
      "step": 72192
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.880960492359943e-05,
      "loss": 4.8269,
      "step": 72704
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.880121897608891e-05,
      "loss": 4.8123,
      "step": 73216
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.879283302857839e-05,
      "loss": 4.8029,
      "step": 73728
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.878444708106787e-05,
      "loss": 4.8172,
      "step": 74240
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.877606113355735e-05,
      "loss": 4.8,
      "step": 74752
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.876767518604683e-05,
      "loss": 4.7971,
      "step": 75264
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.875930561734005e-05,
      "loss": 4.8003,
      "step": 75776
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.875091966982953e-05,
      "loss": 4.7923,
      "step": 76288
    },
    {
      "epoch": 0.03,
      "eval_loss": 4.756740570068359,
      "eval_runtime": 300.3167,
      "eval_samples_per_second": 1270.629,
      "eval_steps_per_second": 39.708,
      "step": 76320
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.874253372231901e-05,
      "loss": 4.7773,
      "step": 76800
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.873414777480849e-05,
      "loss": 4.7893,
      "step": 77312
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.872576182729797e-05,
      "loss": 4.8041,
      "step": 77824
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.871737587978745e-05,
      "loss": 4.7779,
      "step": 78336
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.870898993227693e-05,
      "loss": 4.7911,
      "step": 78848
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.87006039847664e-05,
      "loss": 4.7642,
      "step": 79360
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.869221803725588e-05,
      "loss": 4.7809,
      "step": 79872
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8683848468549096e-05,
      "loss": 4.7578,
      "step": 80384
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8675478899842306e-05,
      "loss": 4.7691,
      "step": 80896
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8667092952331785e-05,
      "loss": 4.7673,
      "step": 81408
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.865870700482127e-05,
      "loss": 4.7694,
      "step": 81920
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.865032105731075e-05,
      "loss": 4.7651,
      "step": 82432
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8641935109800225e-05,
      "loss": 4.7487,
      "step": 82944
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8633549162289705e-05,
      "loss": 4.746,
      "step": 83456
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8625163214779185e-05,
      "loss": 4.7399,
      "step": 83968
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8616777267268665e-05,
      "loss": 4.7446,
      "step": 84480
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8608391319758145e-05,
      "loss": 4.7577,
      "step": 84992
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8600021751051354e-05,
      "loss": 4.7405,
      "step": 85504
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8591635803540834e-05,
      "loss": 4.7471,
      "step": 86016
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8583249856030314e-05,
      "loss": 4.7582,
      "step": 86528
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8574863908519794e-05,
      "loss": 4.7336,
      "step": 87040
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8566477961009274e-05,
      "loss": 4.7289,
      "step": 87552
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8558092013498754e-05,
      "loss": 4.7291,
      "step": 88064
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.854970606598824e-05,
      "loss": 4.7406,
      "step": 88576
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.854132011847772e-05,
      "loss": 4.7201,
      "step": 89088
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.853295054977093e-05,
      "loss": 4.7226,
      "step": 89600
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.852456460226041e-05,
      "loss": 4.721,
      "step": 90112
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.851617865474989e-05,
      "loss": 4.7188,
      "step": 90624
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.85078090860431e-05,
      "loss": 4.7018,
      "step": 91136
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.849942313853258e-05,
      "loss": 4.7066,
      "step": 91648
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.849103719102206e-05,
      "loss": 4.7041,
      "step": 92160
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.848265124351154e-05,
      "loss": 4.7117,
      "step": 92672
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.847426529600102e-05,
      "loss": 4.7157,
      "step": 93184
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.84658793484905e-05,
      "loss": 4.7016,
      "step": 93696
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.845749340097998e-05,
      "loss": 4.6973,
      "step": 94208
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.844910745346946e-05,
      "loss": 4.7105,
      "step": 94720
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8440737884762674e-05,
      "loss": 4.6809,
      "step": 95232
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8432351937252154e-05,
      "loss": 4.6888,
      "step": 95744
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8423982368545363e-05,
      "loss": 4.6772,
      "step": 96256
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8415596421034843e-05,
      "loss": 4.6838,
      "step": 96768
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.840721047352432e-05,
      "loss": 4.676,
      "step": 97280
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.83988245260138e-05,
      "loss": 4.6909,
      "step": 97792
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.839043857850328e-05,
      "loss": 4.6736,
      "step": 98304
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.838205263099276e-05,
      "loss": 4.679,
      "step": 98816
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8373666683482236e-05,
      "loss": 4.6725,
      "step": 99328
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8365280735971716e-05,
      "loss": 4.6782,
      "step": 99840
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8356894788461196e-05,
      "loss": 4.674,
      "step": 100352
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.834852521975441e-05,
      "loss": 4.6646,
      "step": 100864
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.834013927224389e-05,
      "loss": 4.6506,
      "step": 101376
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.833175332473337e-05,
      "loss": 4.6763,
      "step": 101888
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.832338375602659e-05,
      "loss": 4.6616,
      "step": 102400
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.831499780851606e-05,
      "loss": 4.6537,
      "step": 102912
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.830661186100554e-05,
      "loss": 4.6484,
      "step": 103424
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.829822591349502e-05,
      "loss": 4.6549,
      "step": 103936
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.82898399659845e-05,
      "loss": 4.6469,
      "step": 104448
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.828145401847398e-05,
      "loss": 4.6644,
      "step": 104960
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.827306807096346e-05,
      "loss": 4.6449,
      "step": 105472
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.826469850225667e-05,
      "loss": 4.6415,
      "step": 105984
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.825631255474615e-05,
      "loss": 4.654,
      "step": 106496
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.824792660723563e-05,
      "loss": 4.633,
      "step": 107008
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.823954065972511e-05,
      "loss": 4.6278,
      "step": 107520
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8231171091018326e-05,
      "loss": 4.6374,
      "step": 108032
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8222785143507806e-05,
      "loss": 4.6182,
      "step": 108544
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8214399195997286e-05,
      "loss": 4.6263,
      "step": 109056
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8206013248486766e-05,
      "loss": 4.6357,
      "step": 109568
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8197627300976246e-05,
      "loss": 4.6418,
      "step": 110080
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8189241353465726e-05,
      "loss": 4.6251,
      "step": 110592
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8180855405955206e-05,
      "loss": 4.6167,
      "step": 111104
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8172469458444685e-05,
      "loss": 4.6109,
      "step": 111616
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8164083510934165e-05,
      "loss": 4.6306,
      "step": 112128
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8155697563423645e-05,
      "loss": 4.6227,
      "step": 112640
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8147327994716854e-05,
      "loss": 4.6242,
      "step": 113152
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8138942047206334e-05,
      "loss": 4.6239,
      "step": 113664
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8130556099695814e-05,
      "loss": 4.629,
      "step": 114176
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8122170152185294e-05,
      "loss": 4.6262,
      "step": 114688
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8113784204674774e-05,
      "loss": 4.6046,
      "step": 115200
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.810541463596799e-05,
      "loss": 4.6147,
      "step": 115712
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.809702868845747e-05,
      "loss": 4.6143,
      "step": 116224
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.808865911975068e-05,
      "loss": 4.5957,
      "step": 116736
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.808027317224016e-05,
      "loss": 4.6247,
      "step": 117248
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.807188722472964e-05,
      "loss": 4.6107,
      "step": 117760
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.806350127721912e-05,
      "loss": 4.6145,
      "step": 118272
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.80551153297086e-05,
      "loss": 4.6116,
      "step": 118784
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.804672938219807e-05,
      "loss": 4.5847,
      "step": 119296
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.803834343468755e-05,
      "loss": 4.6022,
      "step": 119808
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.802995748717703e-05,
      "loss": 4.5995,
      "step": 120320
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.802157153966652e-05,
      "loss": 4.6023,
      "step": 120832
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.801320197095973e-05,
      "loss": 4.599,
      "step": 121344
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.800481602344921e-05,
      "loss": 4.5956,
      "step": 121856
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.799643007593869e-05,
      "loss": 4.5954,
      "step": 122368
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.798804412842817e-05,
      "loss": 4.5916,
      "step": 122880
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.797967455972138e-05,
      "loss": 4.5778,
      "step": 123392
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.797128861221086e-05,
      "loss": 4.5839,
      "step": 123904
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.796291904350407e-05,
      "loss": 4.5879,
      "step": 124416
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7954533095993546e-05,
      "loss": 4.591,
      "step": 124928
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7946147148483026e-05,
      "loss": 4.5792,
      "step": 125440
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7937761200972506e-05,
      "loss": 4.5706,
      "step": 125952
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7929375253461986e-05,
      "loss": 4.5736,
      "step": 126464
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.792098930595147e-05,
      "loss": 4.5834,
      "step": 126976
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.791260335844095e-05,
      "loss": 4.5804,
      "step": 127488
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.790421741093043e-05,
      "loss": 4.5739,
      "step": 128000
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.789584784222364e-05,
      "loss": 4.5652,
      "step": 128512
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.788746189471312e-05,
      "loss": 4.5608,
      "step": 129024
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.78790759472026e-05,
      "loss": 4.5686,
      "step": 129536
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.787068999969208e-05,
      "loss": 4.5555,
      "step": 130048
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.786230405218156e-05,
      "loss": 4.5597,
      "step": 130560
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.785391810467104e-05,
      "loss": 4.5615,
      "step": 131072
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.784553215716052e-05,
      "loss": 4.5624,
      "step": 131584
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.783714620965e-05,
      "loss": 4.5598,
      "step": 132096
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.782877664094321e-05,
      "loss": 4.562,
      "step": 132608
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.782039069343269e-05,
      "loss": 4.5508,
      "step": 133120
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.781200474592217e-05,
      "loss": 4.5536,
      "step": 133632
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.780361879841165e-05,
      "loss": 4.55,
      "step": 134144
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.779523285090114e-05,
      "loss": 4.5439,
      "step": 134656
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7786863282194346e-05,
      "loss": 4.5604,
      "step": 135168
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.777852647109502e-05,
      "loss": 4.5531,
      "step": 135680
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.777014052358449e-05,
      "loss": 4.5534,
      "step": 136192
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.776175457607397e-05,
      "loss": 4.546,
      "step": 136704
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.775336862856345e-05,
      "loss": 4.5518,
      "step": 137216
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.774498268105293e-05,
      "loss": 4.5496,
      "step": 137728
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.773659673354241e-05,
      "loss": 4.5374,
      "step": 138240
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.772821078603189e-05,
      "loss": 4.5556,
      "step": 138752
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.771982483852137e-05,
      "loss": 4.5423,
      "step": 139264
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.771143889101086e-05,
      "loss": 4.5452,
      "step": 139776
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.770305294350034e-05,
      "loss": 4.5301,
      "step": 140288
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.769466699598982e-05,
      "loss": 4.53,
      "step": 140800
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.76862810484793e-05,
      "loss": 4.5232,
      "step": 141312
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.767789510096878e-05,
      "loss": 4.5281,
      "step": 141824
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.766950915345826e-05,
      "loss": 4.5231,
      "step": 142336
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.766112320594773e-05,
      "loss": 4.5391,
      "step": 142848
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.765273725843721e-05,
      "loss": 4.5407,
      "step": 143360
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.764436768973043e-05,
      "loss": 4.5295,
      "step": 143872
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.76359817422199e-05,
      "loss": 4.5231,
      "step": 144384
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.762759579470938e-05,
      "loss": 4.5294,
      "step": 144896
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.761920984719886e-05,
      "loss": 4.5292,
      "step": 145408
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.761082389968834e-05,
      "loss": 4.5367,
      "step": 145920
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.760243795217783e-05,
      "loss": 4.533,
      "step": 146432
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.759405200466731e-05,
      "loss": 4.5246,
      "step": 146944
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.758568243596052e-05,
      "loss": 4.5164,
      "step": 147456
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.757729648845e-05,
      "loss": 4.5243,
      "step": 147968
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.756891054093948e-05,
      "loss": 4.5168,
      "step": 148480
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.756052459342896e-05,
      "loss": 4.5222,
      "step": 148992
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.755213864591844e-05,
      "loss": 4.518,
      "step": 149504
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.754375269840792e-05,
      "loss": 4.513,
      "step": 150016
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7535383129701126e-05,
      "loss": 4.5244,
      "step": 150528
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7526997182190606e-05,
      "loss": 4.511,
      "step": 151040
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7518627613483815e-05,
      "loss": 4.5067,
      "step": 151552
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7510241665973295e-05,
      "loss": 4.5103,
      "step": 152064
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.750185571846278e-05,
      "loss": 4.5062,
      "step": 152576
    },
    {
      "epoch": 1.03,
      "eval_loss": 4.478397369384766,
      "eval_runtime": 296.4911,
      "eval_samples_per_second": 1287.023,
      "eval_steps_per_second": 40.22,
      "step": 152640
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.749346977095226e-05,
      "loss": 4.4912,
      "step": 153088
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.748508382344174e-05,
      "loss": 4.5065,
      "step": 153600
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.747669787593122e-05,
      "loss": 4.5225,
      "step": 154112
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.74683119284207e-05,
      "loss": 4.5012,
      "step": 154624
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.745992598091018e-05,
      "loss": 4.5129,
      "step": 155136
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.745154003339966e-05,
      "loss": 4.4928,
      "step": 155648
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.744315408588914e-05,
      "loss": 4.5087,
      "step": 156160
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.743476813837862e-05,
      "loss": 4.4842,
      "step": 156672
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7426382190868095e-05,
      "loss": 4.5026,
      "step": 157184
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7417996243357575e-05,
      "loss": 4.5018,
      "step": 157696
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7409610295847055e-05,
      "loss": 4.4994,
      "step": 158208
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7401224348336535e-05,
      "loss": 4.5066,
      "step": 158720
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.739285477962975e-05,
      "loss": 4.4824,
      "step": 159232
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.738446883211923e-05,
      "loss": 4.4834,
      "step": 159744
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.737608288460871e-05,
      "loss": 4.4806,
      "step": 160256
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.736771331590192e-05,
      "loss": 4.4898,
      "step": 160768
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7359343747195135e-05,
      "loss": 4.4956,
      "step": 161280
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7350957799684615e-05,
      "loss": 4.4867,
      "step": 161792
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7342571852174095e-05,
      "loss": 4.4905,
      "step": 162304
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.733418590466357e-05,
      "loss": 4.5062,
      "step": 162816
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.732579995715305e-05,
      "loss": 4.4828,
      "step": 163328
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.731741400964253e-05,
      "loss": 4.4773,
      "step": 163840
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.730902806213201e-05,
      "loss": 4.4838,
      "step": 164352
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.730064211462149e-05,
      "loss": 4.4942,
      "step": 164864
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.729225616711097e-05,
      "loss": 4.4757,
      "step": 165376
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.728387021960045e-05,
      "loss": 4.4787,
      "step": 165888
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.727548427208993e-05,
      "loss": 4.4781,
      "step": 166400
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7267098324579415e-05,
      "loss": 4.4769,
      "step": 166912
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.7258712377068895e-05,
      "loss": 4.4617,
      "step": 167424
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7250342808362104e-05,
      "loss": 4.4694,
      "step": 167936
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7241956860851584e-05,
      "loss": 4.4703,
      "step": 168448
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7233570913341064e-05,
      "loss": 4.4712,
      "step": 168960
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7225184965830544e-05,
      "loss": 4.4839,
      "step": 169472
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7216799018320024e-05,
      "loss": 4.4641,
      "step": 169984
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7208413070809504e-05,
      "loss": 4.4669,
      "step": 170496
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7200027123298984e-05,
      "loss": 4.4764,
      "step": 171008
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.719164117578846e-05,
      "loss": 4.4579,
      "step": 171520
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.718327160708167e-05,
      "loss": 4.456,
      "step": 172032
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.717488565957115e-05,
      "loss": 4.4518,
      "step": 172544
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.716649971206063e-05,
      "loss": 4.46,
      "step": 173056
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.715811376455011e-05,
      "loss": 4.4503,
      "step": 173568
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.714972781703959e-05,
      "loss": 4.468,
      "step": 174080
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.714135824833281e-05,
      "loss": 4.4485,
      "step": 174592
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.713297230082228e-05,
      "loss": 4.4637,
      "step": 175104
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.712458635331176e-05,
      "loss": 4.456,
      "step": 175616
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.711620040580124e-05,
      "loss": 4.453,
      "step": 176128
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.710783083709446e-05,
      "loss": 4.4592,
      "step": 176640
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.709944488958393e-05,
      "loss": 4.4446,
      "step": 177152
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.709105894207341e-05,
      "loss": 4.4353,
      "step": 177664
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.708267299456289e-05,
      "loss": 4.4604,
      "step": 178176
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7074303425856106e-05,
      "loss": 4.4495,
      "step": 178688
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7065917478345586e-05,
      "loss": 4.4409,
      "step": 179200
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7057531530835066e-05,
      "loss": 4.443,
      "step": 179712
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7049145583324546e-05,
      "loss": 4.4478,
      "step": 180224
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7040776014617755e-05,
      "loss": 4.433,
      "step": 180736
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7032390067107235e-05,
      "loss": 4.4539,
      "step": 181248
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7024004119596715e-05,
      "loss": 4.4387,
      "step": 181760
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7015618172086195e-05,
      "loss": 4.4359,
      "step": 182272
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.7007232224575675e-05,
      "loss": 4.4506,
      "step": 182784
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6998846277065155e-05,
      "loss": 4.4285,
      "step": 183296
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6990460329554635e-05,
      "loss": 4.4261,
      "step": 183808
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6982074382044115e-05,
      "loss": 4.4382,
      "step": 184320
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6973704813337324e-05,
      "loss": 4.4216,
      "step": 184832
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6965318865826804e-05,
      "loss": 4.4245,
      "step": 185344
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.695693291831629e-05,
      "loss": 4.4377,
      "step": 185856
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.694854697080577e-05,
      "loss": 4.442,
      "step": 186368
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.694016102329525e-05,
      "loss": 4.424,
      "step": 186880
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.693177507578473e-05,
      "loss": 4.4251,
      "step": 187392
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.692338912827421e-05,
      "loss": 4.4129,
      "step": 187904
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.691500318076369e-05,
      "loss": 4.4365,
      "step": 188416
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.69066336120569e-05,
      "loss": 4.4284,
      "step": 188928
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.689824766454638e-05,
      "loss": 4.4324,
      "step": 189440
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.688986171703586e-05,
      "loss": 4.429,
      "step": 189952
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.688147576952534e-05,
      "loss": 4.4371,
      "step": 190464
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.687312257962228e-05,
      "loss": 4.4411,
      "step": 190976
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.686473663211176e-05,
      "loss": 4.4152,
      "step": 191488
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6856350684601245e-05,
      "loss": 4.4253,
      "step": 192000
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6847964737090725e-05,
      "loss": 4.4306,
      "step": 192512
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6839578789580204e-05,
      "loss": 4.4093,
      "step": 193024
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6831192842069684e-05,
      "loss": 4.4366,
      "step": 193536
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6822806894559164e-05,
      "loss": 4.4259,
      "step": 194048
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6814420947048644e-05,
      "loss": 4.4302,
      "step": 194560
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6806051378341853e-05,
      "loss": 4.4294,
      "step": 195072
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6797665430831333e-05,
      "loss": 4.4034,
      "step": 195584
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.678927948332081e-05,
      "loss": 4.4209,
      "step": 196096
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6780893535810287e-05,
      "loss": 4.4174,
      "step": 196608
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.6772507588299766e-05,
      "loss": 4.425,
      "step": 197120
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.676413801959298e-05,
      "loss": 4.4168,
      "step": 197632
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.675575207208246e-05,
      "loss": 4.4203,
      "step": 198144
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.674736612457194e-05,
      "loss": 4.4165,
      "step": 198656
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.673898017706142e-05,
      "loss": 4.4137,
      "step": 199168
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.673061060835464e-05,
      "loss": 4.4032,
      "step": 199680
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.672224103964785e-05,
      "loss": 4.4081,
      "step": 200192
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.671385509213733e-05,
      "loss": 4.4141,
      "step": 200704
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.670546914462681e-05,
      "loss": 4.4154,
      "step": 201216
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.669708319711629e-05,
      "loss": 4.4087,
      "step": 201728
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.668869724960576e-05,
      "loss": 4.3968,
      "step": 202240
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.668031130209524e-05,
      "loss": 4.4011,
      "step": 202752
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.667192535458472e-05,
      "loss": 4.4122,
      "step": 203264
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6663555785877936e-05,
      "loss": 4.4073,
      "step": 203776
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6655169838367416e-05,
      "loss": 4.4043,
      "step": 204288
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6646783890856896e-05,
      "loss": 4.399,
      "step": 204800
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6638397943346376e-05,
      "loss": 4.3892,
      "step": 205312
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6630011995835856e-05,
      "loss": 4.4055,
      "step": 205824
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6621626048325336e-05,
      "loss": 4.3834,
      "step": 206336
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6613240100814816e-05,
      "loss": 4.3961,
      "step": 206848
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6604870532108025e-05,
      "loss": 4.3896,
      "step": 207360
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6596484584597505e-05,
      "loss": 4.4034,
      "step": 207872
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6588098637086985e-05,
      "loss": 4.3908,
      "step": 208384
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6579712689576465e-05,
      "loss": 4.4009,
      "step": 208896
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6571326742065945e-05,
      "loss": 4.389,
      "step": 209408
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6562940794555425e-05,
      "loss": 4.388,
      "step": 209920
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6554554847044905e-05,
      "loss": 4.3919,
      "step": 210432
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6546168899534385e-05,
      "loss": 4.3795,
      "step": 210944
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.653781570963133e-05,
      "loss": 4.4004,
      "step": 211456
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.652942976212081e-05,
      "loss": 4.3958,
      "step": 211968
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.652104381461029e-05,
      "loss": 4.394,
      "step": 212480
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.651265786709977e-05,
      "loss": 4.3849,
      "step": 212992
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.650427191958925e-05,
      "loss": 4.3917,
      "step": 213504
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.649588597207873e-05,
      "loss": 4.3964,
      "step": 214016
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.648750002456821e-05,
      "loss": 4.377,
      "step": 214528
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.647911407705769e-05,
      "loss": 4.3994,
      "step": 215040
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.647072812954717e-05,
      "loss": 4.388,
      "step": 215552
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.646234218203665e-05,
      "loss": 4.3887,
      "step": 216064
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.645395623452612e-05,
      "loss": 4.3793,
      "step": 216576
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.64455702870156e-05,
      "loss": 4.3811,
      "step": 217088
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.643720071830882e-05,
      "loss": 4.3658,
      "step": 217600
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.64288147707983e-05,
      "loss": 4.3772,
      "step": 218112
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.642042882328778e-05,
      "loss": 4.3729,
      "step": 218624
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.641204287577726e-05,
      "loss": 4.3865,
      "step": 219136
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6403673307070474e-05,
      "loss": 4.39,
      "step": 219648
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.639528735955995e-05,
      "loss": 4.3798,
      "step": 220160
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.638690141204943e-05,
      "loss": 4.3731,
      "step": 220672
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.637851546453891e-05,
      "loss": 4.3779,
      "step": 221184
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.637014589583212e-05,
      "loss": 4.3823,
      "step": 221696
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6361759948321596e-05,
      "loss": 4.3871,
      "step": 222208
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6353374000811076e-05,
      "loss": 4.3831,
      "step": 222720
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6344988053300556e-05,
      "loss": 4.3734,
      "step": 223232
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.633661848459377e-05,
      "loss": 4.3778,
      "step": 223744
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.632823253708325e-05,
      "loss": 4.3812,
      "step": 224256
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.631984658957273e-05,
      "loss": 4.3667,
      "step": 224768
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.631146064206221e-05,
      "loss": 4.3771,
      "step": 225280
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.630307469455169e-05,
      "loss": 4.3725,
      "step": 225792
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.629468874704117e-05,
      "loss": 4.3685,
      "step": 226304
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.628633555713812e-05,
      "loss": 4.3809,
      "step": 226816
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.6277949609627597e-05,
      "loss": 4.3688,
      "step": 227328
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.626956366211707e-05,
      "loss": 4.3638,
      "step": 227840
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.626117771460655e-05,
      "loss": 4.3666,
      "step": 228352
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.625279176709603e-05,
      "loss": 4.3611,
      "step": 228864
    },
    {
      "epoch": 0.03,
      "eval_loss": 4.343437194824219,
      "eval_runtime": 297.7634,
      "eval_samples_per_second": 1281.524,
      "eval_steps_per_second": 40.049,
      "step": 228960
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.624440581958551e-05,
      "loss": 4.3582,
      "step": 229376
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.623601987207499e-05,
      "loss": 4.3609,
      "step": 229888
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6227633924564476e-05,
      "loss": 4.3815,
      "step": 230400
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6219264355857685e-05,
      "loss": 4.3601,
      "step": 230912
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6210878408347165e-05,
      "loss": 4.3788,
      "step": 231424
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6202508839640374e-05,
      "loss": 4.3517,
      "step": 231936
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6194122892129854e-05,
      "loss": 4.3713,
      "step": 232448
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6185736944619334e-05,
      "loss": 4.3467,
      "step": 232960
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6177350997108814e-05,
      "loss": 4.3646,
      "step": 233472
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6168965049598294e-05,
      "loss": 4.3672,
      "step": 233984
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6160579102087774e-05,
      "loss": 4.3627,
      "step": 234496
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6152193154577254e-05,
      "loss": 4.3677,
      "step": 235008
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.614382358587046e-05,
      "loss": 4.3492,
      "step": 235520
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.613543763835994e-05,
      "loss": 4.3496,
      "step": 236032
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.612705169084942e-05,
      "loss": 4.348,
      "step": 236544
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.611866574333891e-05,
      "loss": 4.3517,
      "step": 237056
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.611027979582839e-05,
      "loss": 4.3615,
      "step": 237568
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.610189384831787e-05,
      "loss": 4.3571,
      "step": 238080
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.609350790080735e-05,
      "loss": 4.3563,
      "step": 238592
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.608512195329683e-05,
      "loss": 4.3713,
      "step": 239104
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.607673600578631e-05,
      "loss": 4.3547,
      "step": 239616
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.606836643707952e-05,
      "loss": 4.3496,
      "step": 240128
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.6059980489569e-05,
      "loss": 4.3542,
      "step": 240640
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.605159454205848e-05,
      "loss": 4.3591,
      "step": 241152
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.604320859454796e-05,
      "loss": 4.3449,
      "step": 241664
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.603482264703743e-05,
      "loss": 4.3496,
      "step": 242176
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.602643669952691e-05,
      "loss": 4.35,
      "step": 242688
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.601806713082013e-05,
      "loss": 4.3423,
      "step": 243200
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.600968118330961e-05,
      "loss": 4.3369,
      "step": 243712
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.600129523579909e-05,
      "loss": 4.3383,
      "step": 244224
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.599290928828857e-05,
      "loss": 4.3442,
      "step": 244736
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.598452334077805e-05,
      "loss": 4.344,
      "step": 245248
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.597613739326753e-05,
      "loss": 4.3555,
      "step": 245760
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.596778420336447e-05,
      "loss": 4.3458,
      "step": 246272
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.595939825585395e-05,
      "loss": 4.334,
      "step": 246784
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.595101230834343e-05,
      "loss": 4.3471,
      "step": 247296
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5942626360832906e-05,
      "loss": 4.3386,
      "step": 247808
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5934240413322386e-05,
      "loss": 4.328,
      "step": 248320
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5925854465811865e-05,
      "loss": 4.3319,
      "step": 248832
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5917468518301345e-05,
      "loss": 4.3331,
      "step": 249344
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.590908257079083e-05,
      "loss": 4.3301,
      "step": 249856
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.590071300208404e-05,
      "loss": 4.3446,
      "step": 250368
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.589232705457352e-05,
      "loss": 4.3291,
      "step": 250880
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5883941107063e-05,
      "loss": 4.3395,
      "step": 251392
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.587555515955248e-05,
      "loss": 4.3367,
      "step": 251904
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.586716921204196e-05,
      "loss": 4.3297,
      "step": 252416
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.585878326453144e-05,
      "loss": 4.3405,
      "step": 252928
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.585041369582465e-05,
      "loss": 4.3236,
      "step": 253440
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.584202774831413e-05,
      "loss": 4.3099,
      "step": 253952
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.583364180080361e-05,
      "loss": 4.3445,
      "step": 254464
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.582525585329309e-05,
      "loss": 4.3302,
      "step": 254976
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.581686990578257e-05,
      "loss": 4.3261,
      "step": 255488
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.580848395827205e-05,
      "loss": 4.3215,
      "step": 256000
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.580009801076153e-05,
      "loss": 4.3261,
      "step": 256512
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5791712063251017e-05,
      "loss": 4.3167,
      "step": 257024
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5783342494544226e-05,
      "loss": 4.3321,
      "step": 257536
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5774989304641164e-05,
      "loss": 4.3262,
      "step": 258048
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5766603357130644e-05,
      "loss": 4.3208,
      "step": 258560
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5758217409620124e-05,
      "loss": 4.3309,
      "step": 259072
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5749831462109604e-05,
      "loss": 4.3111,
      "step": 259584
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5741445514599084e-05,
      "loss": 4.31,
      "step": 260096
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5733059567088564e-05,
      "loss": 4.3269,
      "step": 260608
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5724673619578044e-05,
      "loss": 4.3027,
      "step": 261120
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5716287672067524e-05,
      "loss": 4.3076,
      "step": 261632
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5707901724557004e-05,
      "loss": 4.3256,
      "step": 262144
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5699515777046484e-05,
      "loss": 4.3275,
      "step": 262656
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5691129829535963e-05,
      "loss": 4.3092,
      "step": 263168
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.568276026082918e-05,
      "loss": 4.3162,
      "step": 263680
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.567437431331866e-05,
      "loss": 4.2988,
      "step": 264192
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.566598836580814e-05,
      "loss": 4.3244,
      "step": 264704
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.565761879710135e-05,
      "loss": 4.3151,
      "step": 265216
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.564923284959083e-05,
      "loss": 4.3251,
      "step": 265728
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.564084690208031e-05,
      "loss": 4.3151,
      "step": 266240
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.563246095456979e-05,
      "loss": 4.3253,
      "step": 266752
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5624091385863e-05,
      "loss": 4.333,
      "step": 267264
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.561570543835248e-05,
      "loss": 4.3016,
      "step": 267776
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.560731949084196e-05,
      "loss": 4.3177,
      "step": 268288
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.559893354333144e-05,
      "loss": 4.3211,
      "step": 268800
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.559054759582092e-05,
      "loss": 4.2951,
      "step": 269312
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.55821616483104e-05,
      "loss": 4.3292,
      "step": 269824
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.557377570079988e-05,
      "loss": 4.3176,
      "step": 270336
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.556538975328936e-05,
      "loss": 4.3184,
      "step": 270848
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5557020184582566e-05,
      "loss": 4.3261,
      "step": 271360
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5548634237072046e-05,
      "loss": 4.3004,
      "step": 271872
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5540248289561526e-05,
      "loss": 4.3068,
      "step": 272384
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.553187872085474e-05,
      "loss": 4.3107,
      "step": 272896
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5523492773344215e-05,
      "loss": 4.3164,
      "step": 273408
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5515106825833695e-05,
      "loss": 4.3101,
      "step": 273920
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.5506720878323175e-05,
      "loss": 4.3137,
      "step": 274432
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5498334930812655e-05,
      "loss": 4.3057,
      "step": 274944
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.548994898330214e-05,
      "loss": 4.3135,
      "step": 275456
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.548156303579162e-05,
      "loss": 4.2972,
      "step": 275968
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.54731770882811e-05,
      "loss": 4.3029,
      "step": 276480
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.546480751957431e-05,
      "loss": 4.3098,
      "step": 276992
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.545642157206379e-05,
      "loss": 4.3125,
      "step": 277504
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5448052003357e-05,
      "loss": 4.3059,
      "step": 278016
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.543966605584648e-05,
      "loss": 4.2926,
      "step": 278528
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.543128010833596e-05,
      "loss": 4.2964,
      "step": 279040
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.542291053962917e-05,
      "loss": 4.3056,
      "step": 279552
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.541452459211865e-05,
      "loss": 4.3054,
      "step": 280064
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.540613864460813e-05,
      "loss": 4.3049,
      "step": 280576
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.539775269709761e-05,
      "loss": 4.2945,
      "step": 281088
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5389366749587095e-05,
      "loss": 4.2893,
      "step": 281600
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5380980802076575e-05,
      "loss": 4.3025,
      "step": 282112
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5372594854566055e-05,
      "loss": 4.2838,
      "step": 282624
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5364208907055535e-05,
      "loss": 4.2928,
      "step": 283136
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5355822959545015e-05,
      "loss": 4.2863,
      "step": 283648
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5347437012034495e-05,
      "loss": 4.3034,
      "step": 284160
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5339051064523975e-05,
      "loss": 4.2917,
      "step": 284672
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5330665117013455e-05,
      "loss": 4.3063,
      "step": 285184
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.532227916950293e-05,
      "loss": 4.2851,
      "step": 285696
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.531389322199241e-05,
      "loss": 4.2893,
      "step": 286208
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5305523653285624e-05,
      "loss": 4.2913,
      "step": 286720
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5297137705775104e-05,
      "loss": 4.2846,
      "step": 287232
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.528875175826458e-05,
      "loss": 4.2971,
      "step": 287744
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5280365810754064e-05,
      "loss": 4.3022,
      "step": 288256
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5271979863243544e-05,
      "loss": 4.2955,
      "step": 288768
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.526361029453675e-05,
      "loss": 4.2871,
      "step": 289280
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.525522434702623e-05,
      "loss": 4.2918,
      "step": 289792
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.524683839951571e-05,
      "loss": 4.297,
      "step": 290304
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.523845245200519e-05,
      "loss": 4.2803,
      "step": 290816
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.523006650449467e-05,
      "loss": 4.3031,
      "step": 291328
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.522168055698415e-05,
      "loss": 4.2871,
      "step": 291840
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.521331098827736e-05,
      "loss": 4.2938,
      "step": 292352
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.520492504076684e-05,
      "loss": 4.2854,
      "step": 292864
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.519653909325632e-05,
      "loss": 4.2858,
      "step": 293376
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.518816952454953e-05,
      "loss": 4.2682,
      "step": 293888
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.517978357703902e-05,
      "loss": 4.2861,
      "step": 294400
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.51713976295285e-05,
      "loss": 4.2779,
      "step": 294912
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.516301168201798e-05,
      "loss": 4.2884,
      "step": 295424
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.515462573450746e-05,
      "loss": 4.2916,
      "step": 295936
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.514623978699694e-05,
      "loss": 4.2901,
      "step": 296448
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.513785383948642e-05,
      "loss": 4.2743,
      "step": 296960
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.51294678919759e-05,
      "loss": 4.2877,
      "step": 297472
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.512108194446538e-05,
      "loss": 4.2901,
      "step": 297984
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.511269599695486e-05,
      "loss": 4.2909,
      "step": 298496
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.510431004944434e-05,
      "loss": 4.2889,
      "step": 299008
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.509592410193382e-05,
      "loss": 4.2823,
      "step": 299520
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5087554533227026e-05,
      "loss": 4.2797,
      "step": 300032
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5079168585716506e-05,
      "loss": 4.2889,
      "step": 300544
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5070782638205986e-05,
      "loss": 4.274,
      "step": 301056
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5062396690695466e-05,
      "loss": 4.286,
      "step": 301568
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.5054010743184946e-05,
      "loss": 4.2758,
      "step": 302080
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.504565755328189e-05,
      "loss": 4.2809,
      "step": 302592
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.503727160577137e-05,
      "loss": 4.2898,
      "step": 303104
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.502888565826085e-05,
      "loss": 4.276,
      "step": 303616
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.502049971075033e-05,
      "loss": 4.2706,
      "step": 304128
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.501211376323981e-05,
      "loss": 4.2738,
      "step": 304640
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.500374419453302e-05,
      "loss": 4.2754,
      "step": 305152
    },
    {
      "epoch": 1.03,
      "eval_loss": 4.261149883270264,
      "eval_runtime": 298.1026,
      "eval_samples_per_second": 1280.066,
      "eval_steps_per_second": 40.003,
      "step": 305280
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.49953582470225e-05,
      "loss": 4.2618,
      "step": 305664
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.498697229951198e-05,
      "loss": 4.2725,
      "step": 306176
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.497858635200146e-05,
      "loss": 4.2881,
      "step": 306688
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.497020040449093e-05,
      "loss": 4.2725,
      "step": 307200
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.496181445698042e-05,
      "loss": 4.289,
      "step": 307712
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.49534285094699e-05,
      "loss": 4.2694,
      "step": 308224
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.494504256195938e-05,
      "loss": 4.2754,
      "step": 308736
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.493665661444886e-05,
      "loss": 4.2637,
      "step": 309248
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.492827066693834e-05,
      "loss": 4.2725,
      "step": 309760
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.491988471942782e-05,
      "loss": 4.2777,
      "step": 310272
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.49114987719173e-05,
      "loss": 4.2757,
      "step": 310784
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.490311282440678e-05,
      "loss": 4.2819,
      "step": 311296
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.489474325569999e-05,
      "loss": 4.2605,
      "step": 311808
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.488635730818947e-05,
      "loss": 4.2646,
      "step": 312320
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.487797136067895e-05,
      "loss": 4.2626,
      "step": 312832
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.486958541316843e-05,
      "loss": 4.2636,
      "step": 313344
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.486119946565791e-05,
      "loss": 4.274,
      "step": 313856
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.485281351814739e-05,
      "loss": 4.2685,
      "step": 314368
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.484442757063687e-05,
      "loss": 4.2712,
      "step": 314880
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4836058001930084e-05,
      "loss": 4.2847,
      "step": 315392
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4827672054419564e-05,
      "loss": 4.2716,
      "step": 315904
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4819286106909044e-05,
      "loss": 4.2632,
      "step": 316416
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4810900159398524e-05,
      "loss": 4.2665,
      "step": 316928
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4802514211888004e-05,
      "loss": 4.2717,
      "step": 317440
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.479412826437748e-05,
      "loss": 4.2604,
      "step": 317952
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.478575869567069e-05,
      "loss": 4.2666,
      "step": 318464
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.477737274816017e-05,
      "loss": 4.2628,
      "step": 318976
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.476898680064965e-05,
      "loss": 4.2602,
      "step": 319488
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4760600853139126e-05,
      "loss": 4.2555,
      "step": 320000
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.4752214905628606e-05,
      "loss": 4.2567,
      "step": 320512
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4743828958118086e-05,
      "loss": 4.2574,
      "step": 321024
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.473544301060757e-05,
      "loss": 4.2621,
      "step": 321536
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.472705706309705e-05,
      "loss": 4.2716,
      "step": 322048
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.471867111558653e-05,
      "loss": 4.2612,
      "step": 322560
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.471030154687974e-05,
      "loss": 4.2512,
      "step": 323072
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.470193197817295e-05,
      "loss": 4.2655,
      "step": 323584
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.469354603066243e-05,
      "loss": 4.2631,
      "step": 324096
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.468516008315191e-05,
      "loss": 4.2419,
      "step": 324608
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.467677413564139e-05,
      "loss": 4.2484,
      "step": 325120
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.466838818813087e-05,
      "loss": 4.2526,
      "step": 325632
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.466000224062035e-05,
      "loss": 4.2473,
      "step": 326144
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.465161629310983e-05,
      "loss": 4.2664,
      "step": 326656
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.464323034559931e-05,
      "loss": 4.2471,
      "step": 327168
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4634860776892526e-05,
      "loss": 4.256,
      "step": 327680
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4626474829382006e-05,
      "loss": 4.2567,
      "step": 328192
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4618088881871486e-05,
      "loss": 4.2526,
      "step": 328704
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4609702934360966e-05,
      "loss": 4.2592,
      "step": 329216
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4601316986850446e-05,
      "loss": 4.2451,
      "step": 329728
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4592931039339926e-05,
      "loss": 4.2264,
      "step": 330240
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4584545091829406e-05,
      "loss": 4.2682,
      "step": 330752
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4576159144318886e-05,
      "loss": 4.247,
      "step": 331264
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4567789575612095e-05,
      "loss": 4.249,
      "step": 331776
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4559403628101575e-05,
      "loss": 4.2459,
      "step": 332288
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4551034059394784e-05,
      "loss": 4.2404,
      "step": 332800
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4542648111884264e-05,
      "loss": 4.2412,
      "step": 333312
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4534262164373744e-05,
      "loss": 4.2502,
      "step": 333824
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4525876216863224e-05,
      "loss": 4.2452,
      "step": 334336
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4517490269352704e-05,
      "loss": 4.2421,
      "step": 334848
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.450910432184219e-05,
      "loss": 4.2567,
      "step": 335360
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.45007347531354e-05,
      "loss": 4.2304,
      "step": 335872
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.449236518442861e-05,
      "loss": 4.2292,
      "step": 336384
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.448397923691809e-05,
      "loss": 4.2546,
      "step": 336896
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.447559328940757e-05,
      "loss": 4.2205,
      "step": 337408
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.446720734189705e-05,
      "loss": 4.2297,
      "step": 337920
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.445882139438653e-05,
      "loss": 4.2505,
      "step": 338432
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.445043544687601e-05,
      "loss": 4.2514,
      "step": 338944
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.444204949936548e-05,
      "loss": 4.234,
      "step": 339456
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.443366355185496e-05,
      "loss": 4.2429,
      "step": 339968
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.442527760434444e-05,
      "loss": 4.2256,
      "step": 340480
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.441689165683393e-05,
      "loss": 4.2465,
      "step": 340992
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.440850570932341e-05,
      "loss": 4.2385,
      "step": 341504
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.440011976181289e-05,
      "loss": 4.247,
      "step": 342016
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.43917501931061e-05,
      "loss": 4.2389,
      "step": 342528
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.438336424559558e-05,
      "loss": 4.25,
      "step": 343040
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.437497829808506e-05,
      "loss": 4.2559,
      "step": 343552
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.436659235057454e-05,
      "loss": 4.2299,
      "step": 344064
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.435820640306402e-05,
      "loss": 4.2372,
      "step": 344576
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.43498204555535e-05,
      "loss": 4.2468,
      "step": 345088
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4341450886846707e-05,
      "loss": 4.2222,
      "step": 345600
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4333064939336186e-05,
      "loss": 4.2492,
      "step": 346112
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4324678991825666e-05,
      "loss": 4.2463,
      "step": 346624
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4316293044315146e-05,
      "loss": 4.2458,
      "step": 347136
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.4307907096804626e-05,
      "loss": 4.251,
      "step": 347648
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.429953752809784e-05,
      "loss": 4.2277,
      "step": 348160
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.429115158058732e-05,
      "loss": 4.2303,
      "step": 348672
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.42827656330768e-05,
      "loss": 4.2388,
      "step": 349184
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.427439606437001e-05,
      "loss": 4.2461,
      "step": 349696
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.426601011685949e-05,
      "loss": 4.2374,
      "step": 350208
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.425762416934897e-05,
      "loss": 4.2378,
      "step": 350720
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.424923822183845e-05,
      "loss": 4.2296,
      "step": 351232
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.424085227432793e-05,
      "loss": 4.2424,
      "step": 351744
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.423248270562114e-05,
      "loss": 4.2279,
      "step": 352256
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.422409675811062e-05,
      "loss": 4.2274,
      "step": 352768
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.42157108106001e-05,
      "loss": 4.2349,
      "step": 353280
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.420732486308958e-05,
      "loss": 4.2411,
      "step": 353792
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.419893891557907e-05,
      "loss": 4.2328,
      "step": 354304
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.419055296806855e-05,
      "loss": 4.2234,
      "step": 354816
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.418216702055803e-05,
      "loss": 4.2233,
      "step": 355328
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4173797451851236e-05,
      "loss": 4.2328,
      "step": 355840
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4165411504340716e-05,
      "loss": 4.2358,
      "step": 356352
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4157025556830196e-05,
      "loss": 4.2303,
      "step": 356864
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4148639609319676e-05,
      "loss": 4.2271,
      "step": 357376
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.414025366180915e-05,
      "loss": 4.218,
      "step": 357888
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.413186771429863e-05,
      "loss": 4.2282,
      "step": 358400
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.412348176678811e-05,
      "loss": 4.2133,
      "step": 358912
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.411509581927759e-05,
      "loss": 4.2274,
      "step": 359424
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.410670987176707e-05,
      "loss": 4.2089,
      "step": 359936
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4098340303060284e-05,
      "loss": 4.24,
      "step": 360448
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4089954355549764e-05,
      "loss": 4.2226,
      "step": 360960
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4081568408039244e-05,
      "loss": 4.2286,
      "step": 361472
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4073182460528724e-05,
      "loss": 4.2194,
      "step": 361984
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4064796513018204e-05,
      "loss": 4.2196,
      "step": 362496
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4056410565507684e-05,
      "loss": 4.2234,
      "step": 363008
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.4048024617997164e-05,
      "loss": 4.2174,
      "step": 363520
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.403965504929037e-05,
      "loss": 4.2274,
      "step": 364032
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.403126910177985e-05,
      "loss": 4.2307,
      "step": 364544
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.402288315426933e-05,
      "loss": 4.226,
      "step": 365056
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.401449720675881e-05,
      "loss": 4.2211,
      "step": 365568
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.400611125924829e-05,
      "loss": 4.2228,
      "step": 366080
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.399772531173777e-05,
      "loss": 4.2322,
      "step": 366592
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.398935574303098e-05,
      "loss": 4.2083,
      "step": 367104
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.398096979552047e-05,
      "loss": 4.2345,
      "step": 367616
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.397258384800995e-05,
      "loss": 4.2174,
      "step": 368128
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.396419790049943e-05,
      "loss": 4.2277,
      "step": 368640
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.395581195298891e-05,
      "loss": 4.2172,
      "step": 369152
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.394742600547839e-05,
      "loss": 4.2197,
      "step": 369664
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.393904005796786e-05,
      "loss": 4.2037,
      "step": 370176
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.393065411045734e-05,
      "loss": 4.2162,
      "step": 370688
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.392228454175056e-05,
      "loss": 4.2146,
      "step": 371200
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.391389859424003e-05,
      "loss": 4.2188,
      "step": 371712
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.390551264672951e-05,
      "loss": 4.2199,
      "step": 372224
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.389712669921899e-05,
      "loss": 4.2272,
      "step": 372736
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.388874075170847e-05,
      "loss": 4.2076,
      "step": 373248
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.388035480419795e-05,
      "loss": 4.2205,
      "step": 373760
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.387196885668744e-05,
      "loss": 4.2263,
      "step": 374272
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.386358290917692e-05,
      "loss": 4.2241,
      "step": 374784
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.385522971927386e-05,
      "loss": 4.2206,
      "step": 375296
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.3846843771763336e-05,
      "loss": 4.2227,
      "step": 375808
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.383847420305655e-05,
      "loss": 4.2133,
      "step": 376320
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.383008825554603e-05,
      "loss": 4.2234,
      "step": 376832
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.3821702308035505e-05,
      "loss": 4.2067,
      "step": 377344
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.3813316360524985e-05,
      "loss": 4.2225,
      "step": 377856
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.3804930413014465e-05,
      "loss": 4.2106,
      "step": 378368
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.3796544465503944e-05,
      "loss": 4.2133,
      "step": 378880
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.3788158517993424e-05,
      "loss": 4.2243,
      "step": 379392
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.3779772570482904e-05,
      "loss": 4.2142,
      "step": 379904
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.377140300177612e-05,
      "loss": 4.2082,
      "step": 380416
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.37630170542656e-05,
      "loss": 4.204,
      "step": 380928
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.375464748555881e-05,
      "loss": 4.2127,
      "step": 381440
    },
    {
      "epoch": 0.03,
      "eval_loss": 4.204440593719482,
      "eval_runtime": 294.7319,
      "eval_samples_per_second": 1294.705,
      "eval_steps_per_second": 40.46,
      "step": 381600
    }
  ],
  "logging_steps": 512,
  "max_steps": 3052726,
  "num_train_epochs": 9223372036854775807,
  "save_steps": 10,
  "total_flos": 1.570037074412695e+17,
  "trial_name": null,
  "trial_params": null
}