{ "best_metric": null, "best_model_checkpoint": null, "epoch": 20.217435897435898, "eval_steps": 500, "global_step": 4928, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.020512820512820513, "grad_norm": 0.875, "learning_rate": 2.9999969395283144e-06, "loss": 0.6637, "num_input_tokens_seen": 75456, "step": 5 }, { "epoch": 0.041025641025641026, "grad_norm": 1.4140625, "learning_rate": 2.9999877581257458e-06, "loss": 0.6963, "num_input_tokens_seen": 160992, "step": 10 }, { "epoch": 0.06153846153846154, "grad_norm": 1.59375, "learning_rate": 2.9999724558297605e-06, "loss": 0.6919, "num_input_tokens_seen": 253856, "step": 15 }, { "epoch": 0.08205128205128205, "grad_norm": 2.40625, "learning_rate": 2.999951032702801e-06, "loss": 0.6605, "num_input_tokens_seen": 328640, "step": 20 }, { "epoch": 0.10256410256410256, "grad_norm": 0.75390625, "learning_rate": 2.9999234888322877e-06, "loss": 0.5888, "num_input_tokens_seen": 401504, "step": 25 }, { "epoch": 0.12307692307692308, "grad_norm": 1.578125, "learning_rate": 2.9998898243306162e-06, "loss": 0.6087, "num_input_tokens_seen": 475168, "step": 30 }, { "epoch": 0.14358974358974358, "grad_norm": 0.89453125, "learning_rate": 2.99985003933516e-06, "loss": 0.6405, "num_input_tokens_seen": 547168, "step": 35 }, { "epoch": 0.1641025641025641, "grad_norm": 1.1015625, "learning_rate": 2.999804134008266e-06, "loss": 0.6443, "num_input_tokens_seen": 618176, "step": 40 }, { "epoch": 0.18461538461538463, "grad_norm": 0.69921875, "learning_rate": 2.9997521085372565e-06, "loss": 0.6554, "num_input_tokens_seen": 701184, "step": 45 }, { "epoch": 0.20512820512820512, "grad_norm": 1.0234375, "learning_rate": 2.999693963134429e-06, "loss": 0.6903, "num_input_tokens_seen": 782656, "step": 50 }, { "epoch": 0.22564102564102564, "grad_norm": 0.796875, "learning_rate": 2.9996296980370526e-06, "loss": 0.6915, "num_input_tokens_seen": 869280, "step": 55 }, { "epoch": 0.24615384615384617, "grad_norm": 0.8515625, "learning_rate": 2.99955931350737e-06, "loss": 0.584, "num_input_tokens_seen": 944032, "step": 60 }, { "epoch": 0.26666666666666666, "grad_norm": 0.91796875, "learning_rate": 2.999482809832594e-06, "loss": 0.5908, "num_input_tokens_seen": 1025568, "step": 65 }, { "epoch": 0.28717948717948716, "grad_norm": 0.84375, "learning_rate": 2.9994001873249074e-06, "loss": 0.5811, "num_input_tokens_seen": 1105344, "step": 70 }, { "epoch": 0.3076923076923077, "grad_norm": 0.6953125, "learning_rate": 2.999311446321462e-06, "loss": 0.5792, "num_input_tokens_seen": 1182016, "step": 75 }, { "epoch": 0.3282051282051282, "grad_norm": 0.99609375, "learning_rate": 2.999216587184378e-06, "loss": 0.5923, "num_input_tokens_seen": 1265440, "step": 80 }, { "epoch": 0.3487179487179487, "grad_norm": 0.91015625, "learning_rate": 2.9991156103007394e-06, "loss": 0.5896, "num_input_tokens_seen": 1347296, "step": 85 }, { "epoch": 0.36923076923076925, "grad_norm": 0.90234375, "learning_rate": 2.9990085160825954e-06, "loss": 0.6283, "num_input_tokens_seen": 1424064, "step": 90 }, { "epoch": 0.38974358974358975, "grad_norm": 0.67578125, "learning_rate": 2.9988953049669577e-06, "loss": 0.5329, "num_input_tokens_seen": 1492416, "step": 95 }, { "epoch": 0.41025641025641024, "grad_norm": 0.44921875, "learning_rate": 2.998775977415799e-06, "loss": 0.5582, "num_input_tokens_seen": 1572192, "step": 100 }, { "epoch": 0.4307692307692308, "grad_norm": 0.59375, "learning_rate": 2.998650533916051e-06, "loss": 0.6529, "num_input_tokens_seen": 1650592, "step": 105 }, { "epoch": 0.4512820512820513, "grad_norm": 0.515625, "learning_rate": 2.998518974979602e-06, "loss": 0.5449, "num_input_tokens_seen": 1728128, "step": 110 }, { "epoch": 0.4717948717948718, "grad_norm": 0.494140625, "learning_rate": 2.998381301143295e-06, "loss": 0.573, "num_input_tokens_seen": 1806080, "step": 115 }, { "epoch": 0.49230769230769234, "grad_norm": 0.38671875, "learning_rate": 2.9982375129689253e-06, "loss": 0.6065, "num_input_tokens_seen": 1884544, "step": 120 }, { "epoch": 0.5128205128205128, "grad_norm": 0.396484375, "learning_rate": 2.9980876110432404e-06, "loss": 0.6232, "num_input_tokens_seen": 1961376, "step": 125 }, { "epoch": 0.5333333333333333, "grad_norm": 0.408203125, "learning_rate": 2.9979315959779335e-06, "loss": 0.5061, "num_input_tokens_seen": 2036064, "step": 130 }, { "epoch": 0.5538461538461539, "grad_norm": 0.271484375, "learning_rate": 2.9977694684096447e-06, "loss": 0.5657, "num_input_tokens_seen": 2109376, "step": 135 }, { "epoch": 0.5743589743589743, "grad_norm": 0.27734375, "learning_rate": 2.997601228999956e-06, "loss": 0.5597, "num_input_tokens_seen": 2181344, "step": 140 }, { "epoch": 0.5948717948717949, "grad_norm": 0.2451171875, "learning_rate": 2.99742687843539e-06, "loss": 0.5514, "num_input_tokens_seen": 2256768, "step": 145 }, { "epoch": 0.6153846153846154, "grad_norm": 0.2265625, "learning_rate": 2.997246417427407e-06, "loss": 0.53, "num_input_tokens_seen": 2330144, "step": 150 }, { "epoch": 0.6358974358974359, "grad_norm": 0.19921875, "learning_rate": 2.9970598467124008e-06, "loss": 0.5365, "num_input_tokens_seen": 2402688, "step": 155 }, { "epoch": 0.6564102564102564, "grad_norm": 0.1923828125, "learning_rate": 2.9968671670516983e-06, "loss": 0.5578, "num_input_tokens_seen": 2476800, "step": 160 }, { "epoch": 0.676923076923077, "grad_norm": 0.166015625, "learning_rate": 2.9966683792315528e-06, "loss": 0.5492, "num_input_tokens_seen": 2565792, "step": 165 }, { "epoch": 0.6974358974358974, "grad_norm": 0.1533203125, "learning_rate": 2.9964634840631435e-06, "loss": 0.5144, "num_input_tokens_seen": 2637792, "step": 170 }, { "epoch": 0.717948717948718, "grad_norm": 0.154296875, "learning_rate": 2.9962524823825724e-06, "loss": 0.5741, "num_input_tokens_seen": 2718944, "step": 175 }, { "epoch": 0.7384615384615385, "grad_norm": 0.1513671875, "learning_rate": 2.9960353750508583e-06, "loss": 0.5846, "num_input_tokens_seen": 2802240, "step": 180 }, { "epoch": 0.7589743589743589, "grad_norm": 0.1474609375, "learning_rate": 2.995812162953936e-06, "loss": 0.5834, "num_input_tokens_seen": 2884672, "step": 185 }, { "epoch": 0.7794871794871795, "grad_norm": 0.154296875, "learning_rate": 2.9955828470026515e-06, "loss": 0.5493, "num_input_tokens_seen": 2953536, "step": 190 }, { "epoch": 0.8, "grad_norm": 0.138671875, "learning_rate": 2.9953474281327576e-06, "loss": 0.5503, "num_input_tokens_seen": 3026496, "step": 195 }, { "epoch": 0.8205128205128205, "grad_norm": 0.244140625, "learning_rate": 2.995105907304912e-06, "loss": 0.5835, "num_input_tokens_seen": 3105376, "step": 200 }, { "epoch": 0.841025641025641, "grad_norm": 0.349609375, "learning_rate": 2.9948582855046704e-06, "loss": 0.5583, "num_input_tokens_seen": 3179776, "step": 205 }, { "epoch": 0.8615384615384616, "grad_norm": 0.1826171875, "learning_rate": 2.9946045637424864e-06, "loss": 0.6392, "num_input_tokens_seen": 3262336, "step": 210 }, { "epoch": 0.882051282051282, "grad_norm": 0.1640625, "learning_rate": 2.994344743053704e-06, "loss": 0.5658, "num_input_tokens_seen": 3343904, "step": 215 }, { "epoch": 0.9025641025641026, "grad_norm": 0.50390625, "learning_rate": 2.9940788244985557e-06, "loss": 0.5609, "num_input_tokens_seen": 3419264, "step": 220 }, { "epoch": 0.9230769230769231, "grad_norm": 0.458984375, "learning_rate": 2.9938068091621556e-06, "loss": 0.5305, "num_input_tokens_seen": 3496032, "step": 225 }, { "epoch": 0.9435897435897436, "grad_norm": 0.41015625, "learning_rate": 2.9935286981544975e-06, "loss": 0.5652, "num_input_tokens_seen": 3570880, "step": 230 }, { "epoch": 0.9641025641025641, "grad_norm": 0.7734375, "learning_rate": 2.9932444926104495e-06, "loss": 0.5348, "num_input_tokens_seen": 3644096, "step": 235 }, { "epoch": 0.9846153846153847, "grad_norm": 0.86328125, "learning_rate": 2.992954193689749e-06, "loss": 0.5344, "num_input_tokens_seen": 3717088, "step": 240 }, { "epoch": 1.005128205128205, "grad_norm": 0.435546875, "learning_rate": 2.9926578025769978e-06, "loss": 0.5977, "num_input_tokens_seen": 3783648, "step": 245 }, { "epoch": 1.0256410256410255, "grad_norm": 0.408203125, "learning_rate": 2.992355320481658e-06, "loss": 0.5304, "num_input_tokens_seen": 3852160, "step": 250 }, { "epoch": 1.0461538461538462, "grad_norm": 0.75390625, "learning_rate": 2.9920467486380475e-06, "loss": 0.5617, "num_input_tokens_seen": 3924416, "step": 255 }, { "epoch": 1.0666666666666667, "grad_norm": 0.46875, "learning_rate": 2.991732088305333e-06, "loss": 0.5367, "num_input_tokens_seen": 4006432, "step": 260 }, { "epoch": 1.087179487179487, "grad_norm": 0.5234375, "learning_rate": 2.991411340767526e-06, "loss": 0.5207, "num_input_tokens_seen": 4081888, "step": 265 }, { "epoch": 1.1076923076923078, "grad_norm": 0.41015625, "learning_rate": 2.9910845073334793e-06, "loss": 0.556, "num_input_tokens_seen": 4155968, "step": 270 }, { "epoch": 1.1282051282051282, "grad_norm": 0.421875, "learning_rate": 2.9907515893368784e-06, "loss": 0.537, "num_input_tokens_seen": 4234272, "step": 275 }, { "epoch": 1.1487179487179486, "grad_norm": 0.40234375, "learning_rate": 2.9904125881362378e-06, "loss": 0.5305, "num_input_tokens_seen": 4302368, "step": 280 }, { "epoch": 1.1692307692307693, "grad_norm": 0.5625, "learning_rate": 2.990067505114896e-06, "loss": 0.5337, "num_input_tokens_seen": 4376640, "step": 285 }, { "epoch": 1.1897435897435897, "grad_norm": 0.451171875, "learning_rate": 2.9897163416810084e-06, "loss": 0.592, "num_input_tokens_seen": 4458208, "step": 290 }, { "epoch": 1.2102564102564102, "grad_norm": 0.40625, "learning_rate": 2.9893590992675427e-06, "loss": 0.5808, "num_input_tokens_seen": 4537920, "step": 295 }, { "epoch": 1.2307692307692308, "grad_norm": 0.58984375, "learning_rate": 2.988995779332273e-06, "loss": 0.5569, "num_input_tokens_seen": 4606880, "step": 300 }, { "epoch": 1.2512820512820513, "grad_norm": 0.431640625, "learning_rate": 2.9886263833577725e-06, "loss": 0.5422, "num_input_tokens_seen": 4682816, "step": 305 }, { "epoch": 1.2717948717948717, "grad_norm": 0.5234375, "learning_rate": 2.98825091285141e-06, "loss": 0.5296, "num_input_tokens_seen": 4756032, "step": 310 }, { "epoch": 1.2923076923076924, "grad_norm": 0.4453125, "learning_rate": 2.987869369345341e-06, "loss": 0.5012, "num_input_tokens_seen": 4827232, "step": 315 }, { "epoch": 1.3128205128205128, "grad_norm": 0.392578125, "learning_rate": 2.987481754396502e-06, "loss": 0.5289, "num_input_tokens_seen": 4902368, "step": 320 }, { "epoch": 1.3333333333333333, "grad_norm": 0.326171875, "learning_rate": 2.9870880695866067e-06, "loss": 0.5245, "num_input_tokens_seen": 4978080, "step": 325 }, { "epoch": 1.353846153846154, "grad_norm": 0.62109375, "learning_rate": 2.986688316522136e-06, "loss": 0.5325, "num_input_tokens_seen": 5047456, "step": 330 }, { "epoch": 1.3743589743589744, "grad_norm": 0.390625, "learning_rate": 2.9862824968343352e-06, "loss": 0.5068, "num_input_tokens_seen": 5118720, "step": 335 }, { "epoch": 1.3948717948717948, "grad_norm": 0.345703125, "learning_rate": 2.9858706121792036e-06, "loss": 0.5165, "num_input_tokens_seen": 5196288, "step": 340 }, { "epoch": 1.4153846153846155, "grad_norm": 0.251953125, "learning_rate": 2.985452664237488e-06, "loss": 0.5025, "num_input_tokens_seen": 5272480, "step": 345 }, { "epoch": 1.435897435897436, "grad_norm": 0.302734375, "learning_rate": 2.98502865471468e-06, "loss": 0.5285, "num_input_tokens_seen": 5343296, "step": 350 }, { "epoch": 1.4564102564102563, "grad_norm": 0.314453125, "learning_rate": 2.9845985853410053e-06, "loss": 0.4983, "num_input_tokens_seen": 5415904, "step": 355 }, { "epoch": 1.476923076923077, "grad_norm": 0.236328125, "learning_rate": 2.9841624578714167e-06, "loss": 0.5789, "num_input_tokens_seen": 5502176, "step": 360 }, { "epoch": 1.4974358974358974, "grad_norm": 0.2060546875, "learning_rate": 2.9837202740855897e-06, "loss": 0.5394, "num_input_tokens_seen": 5580352, "step": 365 }, { "epoch": 1.5179487179487179, "grad_norm": 0.267578125, "learning_rate": 2.9832720357879107e-06, "loss": 0.5664, "num_input_tokens_seen": 5662912, "step": 370 }, { "epoch": 1.5384615384615383, "grad_norm": 0.22265625, "learning_rate": 2.9828177448074753e-06, "loss": 0.5546, "num_input_tokens_seen": 5743776, "step": 375 }, { "epoch": 1.558974358974359, "grad_norm": 0.20703125, "learning_rate": 2.9823574029980757e-06, "loss": 0.5412, "num_input_tokens_seen": 5812384, "step": 380 }, { "epoch": 1.5794871794871796, "grad_norm": 0.1630859375, "learning_rate": 2.981891012238196e-06, "loss": 0.5587, "num_input_tokens_seen": 5892768, "step": 385 }, { "epoch": 1.6, "grad_norm": 0.1640625, "learning_rate": 2.9814185744310036e-06, "loss": 0.535, "num_input_tokens_seen": 5974592, "step": 390 }, { "epoch": 1.6205128205128205, "grad_norm": 0.1533203125, "learning_rate": 2.9809400915043424e-06, "loss": 0.512, "num_input_tokens_seen": 6050080, "step": 395 }, { "epoch": 1.641025641025641, "grad_norm": 0.142578125, "learning_rate": 2.9804555654107243e-06, "loss": 0.5392, "num_input_tokens_seen": 6137248, "step": 400 }, { "epoch": 1.6615384615384614, "grad_norm": 0.1376953125, "learning_rate": 2.9799649981273185e-06, "loss": 0.6444, "num_input_tokens_seen": 6232704, "step": 405 }, { "epoch": 1.682051282051282, "grad_norm": 0.1435546875, "learning_rate": 2.9794683916559493e-06, "loss": 0.5202, "num_input_tokens_seen": 6304064, "step": 410 }, { "epoch": 1.7025641025641025, "grad_norm": 0.1455078125, "learning_rate": 2.9789657480230842e-06, "loss": 0.5344, "num_input_tokens_seen": 6392928, "step": 415 }, { "epoch": 1.7230769230769232, "grad_norm": 0.16015625, "learning_rate": 2.9784570692798236e-06, "loss": 0.4614, "num_input_tokens_seen": 6473120, "step": 420 }, { "epoch": 1.7435897435897436, "grad_norm": 0.19921875, "learning_rate": 2.977942357501898e-06, "loss": 0.5036, "num_input_tokens_seen": 6545312, "step": 425 }, { "epoch": 1.764102564102564, "grad_norm": 0.41796875, "learning_rate": 2.977421614789655e-06, "loss": 0.5308, "num_input_tokens_seen": 6629984, "step": 430 }, { "epoch": 1.7846153846153845, "grad_norm": 0.1767578125, "learning_rate": 2.976894843268051e-06, "loss": 0.5475, "num_input_tokens_seen": 6715936, "step": 435 }, { "epoch": 1.8051282051282052, "grad_norm": 0.150390625, "learning_rate": 2.976362045086647e-06, "loss": 0.5704, "num_input_tokens_seen": 6797472, "step": 440 }, { "epoch": 1.8256410256410256, "grad_norm": 0.416015625, "learning_rate": 2.975823222419594e-06, "loss": 0.4781, "num_input_tokens_seen": 6867808, "step": 445 }, { "epoch": 1.8461538461538463, "grad_norm": 0.25, "learning_rate": 2.9752783774656267e-06, "loss": 0.5247, "num_input_tokens_seen": 6944480, "step": 450 }, { "epoch": 1.8666666666666667, "grad_norm": 0.703125, "learning_rate": 2.974727512448056e-06, "loss": 0.5161, "num_input_tokens_seen": 7024064, "step": 455 }, { "epoch": 1.8871794871794871, "grad_norm": 0.65234375, "learning_rate": 2.974170629614757e-06, "loss": 0.5049, "num_input_tokens_seen": 7103488, "step": 460 }, { "epoch": 1.9076923076923076, "grad_norm": 0.734375, "learning_rate": 2.9736077312381624e-06, "loss": 0.5712, "num_input_tokens_seen": 7190304, "step": 465 }, { "epoch": 1.9282051282051282, "grad_norm": 0.40234375, "learning_rate": 2.9730388196152513e-06, "loss": 0.5222, "num_input_tokens_seen": 7265056, "step": 470 }, { "epoch": 1.9487179487179487, "grad_norm": 0.28125, "learning_rate": 2.972463897067541e-06, "loss": 0.4829, "num_input_tokens_seen": 7336992, "step": 475 }, { "epoch": 1.9692307692307693, "grad_norm": 0.56640625, "learning_rate": 2.971882965941077e-06, "loss": 0.5218, "num_input_tokens_seen": 7413984, "step": 480 }, { "epoch": 1.9897435897435898, "grad_norm": 0.357421875, "learning_rate": 2.9712960286064237e-06, "loss": 0.5543, "num_input_tokens_seen": 7492768, "step": 485 }, { "epoch": 2.01025641025641, "grad_norm": 0.26171875, "learning_rate": 2.970703087458655e-06, "loss": 0.5036, "num_input_tokens_seen": 7563904, "step": 490 }, { "epoch": 2.0307692307692307, "grad_norm": 0.38671875, "learning_rate": 2.9701041449173426e-06, "loss": 0.556, "num_input_tokens_seen": 7634464, "step": 495 }, { "epoch": 2.051282051282051, "grad_norm": 0.375, "learning_rate": 2.96949920342655e-06, "loss": 0.5742, "num_input_tokens_seen": 7711168, "step": 500 }, { "epoch": 2.071794871794872, "grad_norm": 0.365234375, "learning_rate": 2.968888265454818e-06, "loss": 0.4905, "num_input_tokens_seen": 7786656, "step": 505 }, { "epoch": 2.0923076923076924, "grad_norm": 0.271484375, "learning_rate": 2.968271333495158e-06, "loss": 0.5134, "num_input_tokens_seen": 7858240, "step": 510 }, { "epoch": 2.112820512820513, "grad_norm": 0.373046875, "learning_rate": 2.967648410065041e-06, "loss": 0.6211, "num_input_tokens_seen": 7952064, "step": 515 }, { "epoch": 2.1333333333333333, "grad_norm": 0.45703125, "learning_rate": 2.9670194977063857e-06, "loss": 0.515, "num_input_tokens_seen": 8026688, "step": 520 }, { "epoch": 2.1538461538461537, "grad_norm": 0.400390625, "learning_rate": 2.96638459898555e-06, "loss": 0.5464, "num_input_tokens_seen": 8104064, "step": 525 }, { "epoch": 2.174358974358974, "grad_norm": 0.421875, "learning_rate": 2.9657437164933205e-06, "loss": 0.5331, "num_input_tokens_seen": 8187968, "step": 530 }, { "epoch": 2.194871794871795, "grad_norm": 0.396484375, "learning_rate": 2.9650968528449e-06, "loss": 0.5236, "num_input_tokens_seen": 8261312, "step": 535 }, { "epoch": 2.2153846153846155, "grad_norm": 0.34765625, "learning_rate": 2.9644440106799e-06, "loss": 0.4345, "num_input_tokens_seen": 8336032, "step": 540 }, { "epoch": 2.235897435897436, "grad_norm": 0.353515625, "learning_rate": 2.963785192662327e-06, "loss": 0.4853, "num_input_tokens_seen": 8405024, "step": 545 }, { "epoch": 2.2564102564102564, "grad_norm": 0.365234375, "learning_rate": 2.9631204014805716e-06, "loss": 0.5, "num_input_tokens_seen": 8483456, "step": 550 }, { "epoch": 2.276923076923077, "grad_norm": 0.38671875, "learning_rate": 2.9624496398474014e-06, "loss": 0.4863, "num_input_tokens_seen": 8560768, "step": 555 }, { "epoch": 2.2974358974358973, "grad_norm": 0.32421875, "learning_rate": 2.961772910499945e-06, "loss": 0.48, "num_input_tokens_seen": 8637888, "step": 560 }, { "epoch": 2.3179487179487177, "grad_norm": 0.376953125, "learning_rate": 2.9610902161996838e-06, "loss": 0.5768, "num_input_tokens_seen": 8715360, "step": 565 }, { "epoch": 2.3384615384615386, "grad_norm": 0.265625, "learning_rate": 2.96040155973244e-06, "loss": 0.4751, "num_input_tokens_seen": 8795200, "step": 570 }, { "epoch": 2.358974358974359, "grad_norm": 0.283203125, "learning_rate": 2.959706943908365e-06, "loss": 0.5161, "num_input_tokens_seen": 8883136, "step": 575 }, { "epoch": 2.3794871794871795, "grad_norm": 0.271484375, "learning_rate": 2.9590063715619287e-06, "loss": 0.5588, "num_input_tokens_seen": 8961568, "step": 580 }, { "epoch": 2.4, "grad_norm": 0.2373046875, "learning_rate": 2.9582998455519062e-06, "loss": 0.5527, "num_input_tokens_seen": 9043360, "step": 585 }, { "epoch": 2.4205128205128204, "grad_norm": 0.21875, "learning_rate": 2.9575873687613676e-06, "loss": 0.4897, "num_input_tokens_seen": 9116448, "step": 590 }, { "epoch": 2.4410256410256412, "grad_norm": 0.21484375, "learning_rate": 2.9568689440976676e-06, "loss": 0.5359, "num_input_tokens_seen": 9193120, "step": 595 }, { "epoch": 2.4615384615384617, "grad_norm": 0.205078125, "learning_rate": 2.95614457449243e-06, "loss": 0.5763, "num_input_tokens_seen": 9269920, "step": 600 }, { "epoch": 2.482051282051282, "grad_norm": 0.1845703125, "learning_rate": 2.9554142629015382e-06, "loss": 0.4631, "num_input_tokens_seen": 9339968, "step": 605 }, { "epoch": 2.5025641025641026, "grad_norm": 0.1650390625, "learning_rate": 2.954678012305123e-06, "loss": 0.5349, "num_input_tokens_seen": 9426976, "step": 610 }, { "epoch": 2.523076923076923, "grad_norm": 0.1708984375, "learning_rate": 2.9539358257075495e-06, "loss": 0.5532, "num_input_tokens_seen": 9509056, "step": 615 }, { "epoch": 2.5435897435897434, "grad_norm": 0.1611328125, "learning_rate": 2.9531877061374066e-06, "loss": 0.4748, "num_input_tokens_seen": 9590720, "step": 620 }, { "epoch": 2.564102564102564, "grad_norm": 0.15234375, "learning_rate": 2.9524336566474915e-06, "loss": 0.5022, "num_input_tokens_seen": 9667648, "step": 625 }, { "epoch": 2.5846153846153848, "grad_norm": 0.1337890625, "learning_rate": 2.9516736803148014e-06, "loss": 0.5005, "num_input_tokens_seen": 9738016, "step": 630 }, { "epoch": 2.605128205128205, "grad_norm": 0.1337890625, "learning_rate": 2.9509077802405174e-06, "loss": 0.5297, "num_input_tokens_seen": 9816224, "step": 635 }, { "epoch": 2.6256410256410256, "grad_norm": 0.13671875, "learning_rate": 2.9501359595499933e-06, "loss": 0.5399, "num_input_tokens_seen": 9891104, "step": 640 }, { "epoch": 2.646153846153846, "grad_norm": 0.177734375, "learning_rate": 2.9493582213927425e-06, "loss": 0.4901, "num_input_tokens_seen": 9969792, "step": 645 }, { "epoch": 2.6666666666666665, "grad_norm": 0.142578125, "learning_rate": 2.9485745689424267e-06, "loss": 0.4591, "num_input_tokens_seen": 10044608, "step": 650 }, { "epoch": 2.6871794871794874, "grad_norm": 0.2451171875, "learning_rate": 2.9477850053968405e-06, "loss": 0.5729, "num_input_tokens_seen": 10132640, "step": 655 }, { "epoch": 2.707692307692308, "grad_norm": 0.150390625, "learning_rate": 2.9469895339778995e-06, "loss": 0.5405, "num_input_tokens_seen": 10207968, "step": 660 }, { "epoch": 2.7282051282051283, "grad_norm": 0.30859375, "learning_rate": 2.946188157931627e-06, "loss": 0.4786, "num_input_tokens_seen": 10277408, "step": 665 }, { "epoch": 2.7487179487179487, "grad_norm": 0.314453125, "learning_rate": 2.9453808805281423e-06, "loss": 0.5035, "num_input_tokens_seen": 10349184, "step": 670 }, { "epoch": 2.769230769230769, "grad_norm": 0.2490234375, "learning_rate": 2.944567705061644e-06, "loss": 0.4719, "num_input_tokens_seen": 10434112, "step": 675 }, { "epoch": 2.7897435897435896, "grad_norm": 0.65234375, "learning_rate": 2.9437486348504e-06, "loss": 0.5118, "num_input_tokens_seen": 10506208, "step": 680 }, { "epoch": 2.81025641025641, "grad_norm": 0.68359375, "learning_rate": 2.9429236732367318e-06, "loss": 0.5014, "num_input_tokens_seen": 10577696, "step": 685 }, { "epoch": 2.830769230769231, "grad_norm": 0.71484375, "learning_rate": 2.942092823587001e-06, "loss": 0.4827, "num_input_tokens_seen": 10657984, "step": 690 }, { "epoch": 2.8512820512820514, "grad_norm": 0.71484375, "learning_rate": 2.941256089291597e-06, "loss": 0.5177, "num_input_tokens_seen": 10734688, "step": 695 }, { "epoch": 2.871794871794872, "grad_norm": 0.6328125, "learning_rate": 2.940413473764923e-06, "loss": 0.4517, "num_input_tokens_seen": 10812640, "step": 700 }, { "epoch": 2.8923076923076922, "grad_norm": 0.70703125, "learning_rate": 2.9395649804453786e-06, "loss": 0.4574, "num_input_tokens_seen": 10884800, "step": 705 }, { "epoch": 2.9128205128205127, "grad_norm": 0.27734375, "learning_rate": 2.9387106127953515e-06, "loss": 0.5092, "num_input_tokens_seen": 10962016, "step": 710 }, { "epoch": 2.9333333333333336, "grad_norm": 0.34375, "learning_rate": 2.937850374301198e-06, "loss": 0.4888, "num_input_tokens_seen": 11033280, "step": 715 }, { "epoch": 2.953846153846154, "grad_norm": 0.375, "learning_rate": 2.9369842684732336e-06, "loss": 0.5447, "num_input_tokens_seen": 11113696, "step": 720 }, { "epoch": 2.9743589743589745, "grad_norm": 0.396484375, "learning_rate": 2.936112298845713e-06, "loss": 0.5438, "num_input_tokens_seen": 11195104, "step": 725 }, { "epoch": 2.994871794871795, "grad_norm": 0.2470703125, "learning_rate": 2.935234468976822e-06, "loss": 0.46, "num_input_tokens_seen": 11270304, "step": 730 }, { "epoch": 3.0153846153846153, "grad_norm": 0.392578125, "learning_rate": 2.934350782448658e-06, "loss": 0.569, "num_input_tokens_seen": 11350784, "step": 735 }, { "epoch": 3.0358974358974358, "grad_norm": 0.318359375, "learning_rate": 2.9334612428672175e-06, "loss": 0.5246, "num_input_tokens_seen": 11429568, "step": 740 }, { "epoch": 3.056410256410256, "grad_norm": 0.35546875, "learning_rate": 2.9325658538623822e-06, "loss": 0.4587, "num_input_tokens_seen": 11502784, "step": 745 }, { "epoch": 3.076923076923077, "grad_norm": 0.3671875, "learning_rate": 2.931664619087902e-06, "loss": 0.5095, "num_input_tokens_seen": 11575680, "step": 750 }, { "epoch": 3.0974358974358975, "grad_norm": 0.35546875, "learning_rate": 2.9307575422213813e-06, "loss": 0.4916, "num_input_tokens_seen": 11649856, "step": 755 }, { "epoch": 3.117948717948718, "grad_norm": 0.404296875, "learning_rate": 2.929844626964265e-06, "loss": 0.5647, "num_input_tokens_seen": 11727616, "step": 760 }, { "epoch": 3.1384615384615384, "grad_norm": 0.421875, "learning_rate": 2.9289258770418208e-06, "loss": 0.4448, "num_input_tokens_seen": 11806208, "step": 765 }, { "epoch": 3.158974358974359, "grad_norm": 0.39453125, "learning_rate": 2.9280012962031263e-06, "loss": 0.5086, "num_input_tokens_seen": 11884096, "step": 770 }, { "epoch": 3.1794871794871793, "grad_norm": 0.29296875, "learning_rate": 2.9270708882210525e-06, "loss": 0.4796, "num_input_tokens_seen": 11956416, "step": 775 }, { "epoch": 3.2, "grad_norm": 0.33203125, "learning_rate": 2.92613465689225e-06, "loss": 0.4797, "num_input_tokens_seen": 12032384, "step": 780 }, { "epoch": 3.2205128205128206, "grad_norm": 0.287109375, "learning_rate": 2.92519260603713e-06, "loss": 0.4523, "num_input_tokens_seen": 12107328, "step": 785 }, { "epoch": 3.241025641025641, "grad_norm": 0.326171875, "learning_rate": 2.9242447394998545e-06, "loss": 0.4795, "num_input_tokens_seen": 12178848, "step": 790 }, { "epoch": 3.2615384615384615, "grad_norm": 0.2734375, "learning_rate": 2.923291061148314e-06, "loss": 0.5164, "num_input_tokens_seen": 12252160, "step": 795 }, { "epoch": 3.282051282051282, "grad_norm": 0.271484375, "learning_rate": 2.9223315748741146e-06, "loss": 0.4949, "num_input_tokens_seen": 12325120, "step": 800 }, { "epoch": 3.3025641025641024, "grad_norm": 0.244140625, "learning_rate": 2.9213662845925662e-06, "loss": 0.4848, "num_input_tokens_seen": 12398144, "step": 805 }, { "epoch": 3.3230769230769233, "grad_norm": 0.2265625, "learning_rate": 2.9203951942426586e-06, "loss": 0.5114, "num_input_tokens_seen": 12475008, "step": 810 }, { "epoch": 3.3435897435897437, "grad_norm": 0.2109375, "learning_rate": 2.9194183077870516e-06, "loss": 0.6022, "num_input_tokens_seen": 12562336, "step": 815 }, { "epoch": 3.364102564102564, "grad_norm": 0.1923828125, "learning_rate": 2.9184356292120562e-06, "loss": 0.4922, "num_input_tokens_seen": 12646560, "step": 820 }, { "epoch": 3.3846153846153846, "grad_norm": 0.1962890625, "learning_rate": 2.9174471625276198e-06, "loss": 0.5707, "num_input_tokens_seen": 12718848, "step": 825 }, { "epoch": 3.405128205128205, "grad_norm": 0.1650390625, "learning_rate": 2.916452911767307e-06, "loss": 0.4784, "num_input_tokens_seen": 12798240, "step": 830 }, { "epoch": 3.4256410256410255, "grad_norm": 0.1552734375, "learning_rate": 2.915452880988287e-06, "loss": 0.4423, "num_input_tokens_seen": 12872608, "step": 835 }, { "epoch": 3.4461538461538463, "grad_norm": 0.1396484375, "learning_rate": 2.914447074271314e-06, "loss": 0.4809, "num_input_tokens_seen": 12952896, "step": 840 }, { "epoch": 3.466666666666667, "grad_norm": 0.1650390625, "learning_rate": 2.913435495720712e-06, "loss": 0.5316, "num_input_tokens_seen": 13036768, "step": 845 }, { "epoch": 3.4871794871794872, "grad_norm": 0.140625, "learning_rate": 2.9124181494643574e-06, "loss": 0.4592, "num_input_tokens_seen": 13114784, "step": 850 }, { "epoch": 3.5076923076923077, "grad_norm": 0.138671875, "learning_rate": 2.911395039653663e-06, "loss": 0.4878, "num_input_tokens_seen": 13188448, "step": 855 }, { "epoch": 3.528205128205128, "grad_norm": 0.1298828125, "learning_rate": 2.9103661704635604e-06, "loss": 0.5066, "num_input_tokens_seen": 13262592, "step": 860 }, { "epoch": 3.5487179487179485, "grad_norm": 0.11962890625, "learning_rate": 2.909331546092483e-06, "loss": 0.4649, "num_input_tokens_seen": 13339936, "step": 865 }, { "epoch": 3.569230769230769, "grad_norm": 0.14453125, "learning_rate": 2.908291170762349e-06, "loss": 0.5233, "num_input_tokens_seen": 13416256, "step": 870 }, { "epoch": 3.58974358974359, "grad_norm": 0.306640625, "learning_rate": 2.9072450487185434e-06, "loss": 0.5018, "num_input_tokens_seen": 13487392, "step": 875 }, { "epoch": 3.6102564102564103, "grad_norm": 0.337890625, "learning_rate": 2.9061931842299026e-06, "loss": 0.4602, "num_input_tokens_seen": 13569984, "step": 880 }, { "epoch": 3.6307692307692307, "grad_norm": 0.1640625, "learning_rate": 2.9051355815886952e-06, "loss": 0.5309, "num_input_tokens_seen": 13650944, "step": 885 }, { "epoch": 3.651282051282051, "grad_norm": 0.30859375, "learning_rate": 2.904072245110605e-06, "loss": 0.5186, "num_input_tokens_seen": 13720736, "step": 890 }, { "epoch": 3.6717948717948716, "grad_norm": 0.333984375, "learning_rate": 2.9030031791347136e-06, "loss": 0.4839, "num_input_tokens_seen": 13791616, "step": 895 }, { "epoch": 3.6923076923076925, "grad_norm": 0.25, "learning_rate": 2.901928388023483e-06, "loss": 0.5199, "num_input_tokens_seen": 13867488, "step": 900 }, { "epoch": 3.712820512820513, "grad_norm": 0.69921875, "learning_rate": 2.900847876162736e-06, "loss": 0.5414, "num_input_tokens_seen": 13954848, "step": 905 }, { "epoch": 3.7333333333333334, "grad_norm": 0.80859375, "learning_rate": 2.899761647961641e-06, "loss": 0.5451, "num_input_tokens_seen": 14037792, "step": 910 }, { "epoch": 3.753846153846154, "grad_norm": 1.0, "learning_rate": 2.898669707852692e-06, "loss": 0.4448, "num_input_tokens_seen": 14107520, "step": 915 }, { "epoch": 3.7743589743589743, "grad_norm": 0.3125, "learning_rate": 2.897572060291692e-06, "loss": 0.5213, "num_input_tokens_seen": 14193888, "step": 920 }, { "epoch": 3.7948717948717947, "grad_norm": 0.5, "learning_rate": 2.896468709757733e-06, "loss": 0.4968, "num_input_tokens_seen": 14270976, "step": 925 }, { "epoch": 3.815384615384615, "grad_norm": 0.58203125, "learning_rate": 2.8953596607531788e-06, "loss": 0.5769, "num_input_tokens_seen": 14351232, "step": 930 }, { "epoch": 3.835897435897436, "grad_norm": 0.26171875, "learning_rate": 2.894244917803647e-06, "loss": 0.4925, "num_input_tokens_seen": 14426912, "step": 935 }, { "epoch": 3.8564102564102565, "grad_norm": 0.333984375, "learning_rate": 2.8931244854579904e-06, "loss": 0.481, "num_input_tokens_seen": 14515776, "step": 940 }, { "epoch": 3.876923076923077, "grad_norm": 0.322265625, "learning_rate": 2.891998368288277e-06, "loss": 0.4699, "num_input_tokens_seen": 14587104, "step": 945 }, { "epoch": 3.8974358974358974, "grad_norm": 0.341796875, "learning_rate": 2.890866570889773e-06, "loss": 0.5206, "num_input_tokens_seen": 14663680, "step": 950 }, { "epoch": 3.917948717948718, "grad_norm": 0.35546875, "learning_rate": 2.8897290978809245e-06, "loss": 0.5117, "num_input_tokens_seen": 14747360, "step": 955 }, { "epoch": 3.9384615384615387, "grad_norm": 0.3671875, "learning_rate": 2.888585953903336e-06, "loss": 0.4891, "num_input_tokens_seen": 14826464, "step": 960 }, { "epoch": 3.958974358974359, "grad_norm": 0.376953125, "learning_rate": 2.8874371436217534e-06, "loss": 0.4943, "num_input_tokens_seen": 14916416, "step": 965 }, { "epoch": 3.9794871794871796, "grad_norm": 0.32421875, "learning_rate": 2.8862826717240464e-06, "loss": 0.5222, "num_input_tokens_seen": 14995072, "step": 970 }, { "epoch": 4.0, "grad_norm": 0.337890625, "learning_rate": 2.8851225429211855e-06, "loss": 0.5197, "num_input_tokens_seen": 15070304, "step": 975 }, { "epoch": 4.02051282051282, "grad_norm": 0.380859375, "learning_rate": 2.883956761947226e-06, "loss": 0.501, "num_input_tokens_seen": 15152480, "step": 980 }, { "epoch": 4.041025641025641, "grad_norm": 0.478515625, "learning_rate": 2.8827853335592876e-06, "loss": 0.5142, "num_input_tokens_seen": 15229184, "step": 985 }, { "epoch": 4.061538461538461, "grad_norm": 0.376953125, "learning_rate": 2.8816082625375353e-06, "loss": 0.5239, "num_input_tokens_seen": 15311072, "step": 990 }, { "epoch": 4.082051282051282, "grad_norm": 0.380859375, "learning_rate": 2.8804255536851584e-06, "loss": 0.4823, "num_input_tokens_seen": 15383232, "step": 995 }, { "epoch": 4.102564102564102, "grad_norm": 0.34375, "learning_rate": 2.8792372118283528e-06, "loss": 0.5416, "num_input_tokens_seen": 15464064, "step": 1000 }, { "epoch": 4.123076923076923, "grad_norm": 0.416015625, "learning_rate": 2.878043241816301e-06, "loss": 0.4889, "num_input_tokens_seen": 15536480, "step": 1005 }, { "epoch": 4.143589743589744, "grad_norm": 0.3828125, "learning_rate": 2.876843648521152e-06, "loss": 0.5338, "num_input_tokens_seen": 15618816, "step": 1010 }, { "epoch": 4.164102564102564, "grad_norm": 0.2578125, "learning_rate": 2.8756384368380003e-06, "loss": 0.5101, "num_input_tokens_seen": 15694304, "step": 1015 }, { "epoch": 4.184615384615385, "grad_norm": 0.25390625, "learning_rate": 2.874427611684867e-06, "loss": 0.4792, "num_input_tokens_seen": 15770784, "step": 1020 }, { "epoch": 4.205128205128205, "grad_norm": 0.27734375, "learning_rate": 2.8732111780026813e-06, "loss": 0.4959, "num_input_tokens_seen": 15837312, "step": 1025 }, { "epoch": 4.225641025641026, "grad_norm": 0.22265625, "learning_rate": 2.871989140755257e-06, "loss": 0.5227, "num_input_tokens_seen": 15908096, "step": 1030 }, { "epoch": 4.246153846153846, "grad_norm": 0.2392578125, "learning_rate": 2.870761504929275e-06, "loss": 0.473, "num_input_tokens_seen": 15982720, "step": 1035 }, { "epoch": 4.266666666666667, "grad_norm": 0.1943359375, "learning_rate": 2.869528275534261e-06, "loss": 0.4911, "num_input_tokens_seen": 16056256, "step": 1040 }, { "epoch": 4.287179487179487, "grad_norm": 0.1787109375, "learning_rate": 2.8682894576025677e-06, "loss": 0.4678, "num_input_tokens_seen": 16128256, "step": 1045 }, { "epoch": 4.3076923076923075, "grad_norm": 0.1650390625, "learning_rate": 2.8670450561893498e-06, "loss": 0.4534, "num_input_tokens_seen": 16203808, "step": 1050 }, { "epoch": 4.328205128205128, "grad_norm": 0.1796875, "learning_rate": 2.865795076372549e-06, "loss": 0.5788, "num_input_tokens_seen": 16290464, "step": 1055 }, { "epoch": 4.348717948717948, "grad_norm": 0.1845703125, "learning_rate": 2.8645395232528683e-06, "loss": 0.4744, "num_input_tokens_seen": 16362688, "step": 1060 }, { "epoch": 4.36923076923077, "grad_norm": 0.146484375, "learning_rate": 2.863278401953754e-06, "loss": 0.4801, "num_input_tokens_seen": 16434976, "step": 1065 }, { "epoch": 4.38974358974359, "grad_norm": 0.134765625, "learning_rate": 2.862011717621375e-06, "loss": 0.5035, "num_input_tokens_seen": 16514880, "step": 1070 }, { "epoch": 4.410256410256411, "grad_norm": 0.15625, "learning_rate": 2.860739475424599e-06, "loss": 0.5456, "num_input_tokens_seen": 16592544, "step": 1075 }, { "epoch": 4.430769230769231, "grad_norm": 0.125, "learning_rate": 2.859461680554975e-06, "loss": 0.4773, "num_input_tokens_seen": 16675360, "step": 1080 }, { "epoch": 4.4512820512820515, "grad_norm": 0.1103515625, "learning_rate": 2.858178338226709e-06, "loss": 0.4793, "num_input_tokens_seen": 16753728, "step": 1085 }, { "epoch": 4.471794871794872, "grad_norm": 0.11572265625, "learning_rate": 2.8568894536766462e-06, "loss": 0.4698, "num_input_tokens_seen": 16829696, "step": 1090 }, { "epoch": 4.492307692307692, "grad_norm": 0.19140625, "learning_rate": 2.8555950321642444e-06, "loss": 0.4648, "num_input_tokens_seen": 16908128, "step": 1095 }, { "epoch": 4.512820512820513, "grad_norm": 0.2333984375, "learning_rate": 2.8542950789715587e-06, "loss": 0.4473, "num_input_tokens_seen": 16979136, "step": 1100 }, { "epoch": 4.533333333333333, "grad_norm": 0.216796875, "learning_rate": 2.8529895994032153e-06, "loss": 0.5128, "num_input_tokens_seen": 17063296, "step": 1105 }, { "epoch": 4.553846153846154, "grad_norm": 0.15234375, "learning_rate": 2.851678598786392e-06, "loss": 0.4834, "num_input_tokens_seen": 17145984, "step": 1110 }, { "epoch": 4.574358974358974, "grad_norm": 0.296875, "learning_rate": 2.8503620824707946e-06, "loss": 0.4581, "num_input_tokens_seen": 17221152, "step": 1115 }, { "epoch": 4.5948717948717945, "grad_norm": 0.267578125, "learning_rate": 2.8490400558286395e-06, "loss": 0.5189, "num_input_tokens_seen": 17298688, "step": 1120 }, { "epoch": 4.615384615384615, "grad_norm": 0.302734375, "learning_rate": 2.847712524254626e-06, "loss": 0.4739, "num_input_tokens_seen": 17374240, "step": 1125 }, { "epoch": 4.635897435897435, "grad_norm": 0.6640625, "learning_rate": 2.846379493165918e-06, "loss": 0.5376, "num_input_tokens_seen": 17451360, "step": 1130 }, { "epoch": 4.656410256410257, "grad_norm": 0.82421875, "learning_rate": 2.8450409680021204e-06, "loss": 0.5152, "num_input_tokens_seen": 17535776, "step": 1135 }, { "epoch": 4.676923076923077, "grad_norm": 0.90234375, "learning_rate": 2.8436969542252576e-06, "loss": 0.4957, "num_input_tokens_seen": 17616256, "step": 1140 }, { "epoch": 4.697435897435898, "grad_norm": 0.31640625, "learning_rate": 2.842347457319752e-06, "loss": 0.4963, "num_input_tokens_seen": 17691616, "step": 1145 }, { "epoch": 4.717948717948718, "grad_norm": 0.5078125, "learning_rate": 2.8409924827923985e-06, "loss": 0.4868, "num_input_tokens_seen": 17767264, "step": 1150 }, { "epoch": 4.7384615384615385, "grad_norm": 0.36328125, "learning_rate": 2.839632036172346e-06, "loss": 0.5595, "num_input_tokens_seen": 17847648, "step": 1155 }, { "epoch": 4.758974358974359, "grad_norm": 0.296875, "learning_rate": 2.8382661230110716e-06, "loss": 0.6248, "num_input_tokens_seen": 17932992, "step": 1160 }, { "epoch": 4.779487179487179, "grad_norm": 0.326171875, "learning_rate": 2.8368947488823613e-06, "loss": 0.4935, "num_input_tokens_seen": 18008544, "step": 1165 }, { "epoch": 4.8, "grad_norm": 0.3203125, "learning_rate": 2.8355179193822834e-06, "loss": 0.472, "num_input_tokens_seen": 18091904, "step": 1170 }, { "epoch": 4.82051282051282, "grad_norm": 0.396484375, "learning_rate": 2.834135640129168e-06, "loss": 0.4618, "num_input_tokens_seen": 18169760, "step": 1175 }, { "epoch": 4.841025641025641, "grad_norm": 0.2890625, "learning_rate": 2.8327479167635834e-06, "loss": 0.4375, "num_input_tokens_seen": 18242016, "step": 1180 }, { "epoch": 4.861538461538462, "grad_norm": 0.314453125, "learning_rate": 2.831354754948315e-06, "loss": 0.4954, "num_input_tokens_seen": 18316192, "step": 1185 }, { "epoch": 4.8820512820512825, "grad_norm": 0.330078125, "learning_rate": 2.829956160368338e-06, "loss": 0.4885, "num_input_tokens_seen": 18402720, "step": 1190 }, { "epoch": 4.902564102564103, "grad_norm": 0.482421875, "learning_rate": 2.828552138730798e-06, "loss": 0.4452, "num_input_tokens_seen": 18472768, "step": 1195 }, { "epoch": 4.923076923076923, "grad_norm": 0.33984375, "learning_rate": 2.8271426957649868e-06, "loss": 0.4602, "num_input_tokens_seen": 18548128, "step": 1200 }, { "epoch": 4.943589743589744, "grad_norm": 0.412109375, "learning_rate": 2.8257278372223177e-06, "loss": 0.4391, "num_input_tokens_seen": 18622112, "step": 1205 }, { "epoch": 4.964102564102564, "grad_norm": 0.484375, "learning_rate": 2.824307568876304e-06, "loss": 0.4614, "num_input_tokens_seen": 18704288, "step": 1210 }, { "epoch": 4.984615384615385, "grad_norm": 0.35546875, "learning_rate": 2.8228818965225326e-06, "loss": 0.5284, "num_input_tokens_seen": 18780128, "step": 1215 }, { "epoch": 5.005128205128205, "grad_norm": 0.326171875, "learning_rate": 2.8214508259786443e-06, "loss": 0.5213, "num_input_tokens_seen": 18850496, "step": 1220 }, { "epoch": 5.0256410256410255, "grad_norm": 0.333984375, "learning_rate": 2.820014363084307e-06, "loss": 0.5071, "num_input_tokens_seen": 18926816, "step": 1225 }, { "epoch": 5.046153846153846, "grad_norm": 0.341796875, "learning_rate": 2.8185725137011922e-06, "loss": 0.4964, "num_input_tokens_seen": 19002624, "step": 1230 }, { "epoch": 5.066666666666666, "grad_norm": 0.3359375, "learning_rate": 2.8171252837129523e-06, "loss": 0.5196, "num_input_tokens_seen": 19083296, "step": 1235 }, { "epoch": 5.087179487179487, "grad_norm": 0.28125, "learning_rate": 2.815672679025196e-06, "loss": 0.5272, "num_input_tokens_seen": 19158048, "step": 1240 }, { "epoch": 5.107692307692307, "grad_norm": 0.26171875, "learning_rate": 2.814214705565464e-06, "loss": 0.5034, "num_input_tokens_seen": 19233888, "step": 1245 }, { "epoch": 5.128205128205128, "grad_norm": 0.2890625, "learning_rate": 2.8127513692832047e-06, "loss": 0.5069, "num_input_tokens_seen": 19317472, "step": 1250 }, { "epoch": 5.148717948717949, "grad_norm": 0.2265625, "learning_rate": 2.8112826761497507e-06, "loss": 0.5116, "num_input_tokens_seen": 19398496, "step": 1255 }, { "epoch": 5.1692307692307695, "grad_norm": 0.21484375, "learning_rate": 2.8098086321582937e-06, "loss": 0.4286, "num_input_tokens_seen": 19466400, "step": 1260 }, { "epoch": 5.18974358974359, "grad_norm": 0.2119140625, "learning_rate": 2.8083292433238602e-06, "loss": 0.5058, "num_input_tokens_seen": 19550336, "step": 1265 }, { "epoch": 5.21025641025641, "grad_norm": 0.2177734375, "learning_rate": 2.8068445156832864e-06, "loss": 0.4587, "num_input_tokens_seen": 19625792, "step": 1270 }, { "epoch": 5.230769230769231, "grad_norm": 0.1943359375, "learning_rate": 2.805354455295196e-06, "loss": 0.4901, "num_input_tokens_seen": 19711776, "step": 1275 }, { "epoch": 5.251282051282051, "grad_norm": 0.173828125, "learning_rate": 2.8038590682399718e-06, "loss": 0.5474, "num_input_tokens_seen": 19794208, "step": 1280 }, { "epoch": 5.271794871794872, "grad_norm": 0.1611328125, "learning_rate": 2.8023583606197336e-06, "loss": 0.4452, "num_input_tokens_seen": 19872128, "step": 1285 }, { "epoch": 5.292307692307692, "grad_norm": 0.1357421875, "learning_rate": 2.800852338558312e-06, "loss": 0.5081, "num_input_tokens_seen": 19953856, "step": 1290 }, { "epoch": 5.312820512820513, "grad_norm": 0.1376953125, "learning_rate": 2.7993410082012247e-06, "loss": 0.4863, "num_input_tokens_seen": 20026848, "step": 1295 }, { "epoch": 5.333333333333333, "grad_norm": 0.138671875, "learning_rate": 2.79782437571565e-06, "loss": 0.4979, "num_input_tokens_seen": 20100928, "step": 1300 }, { "epoch": 5.3538461538461535, "grad_norm": 0.1318359375, "learning_rate": 2.7963024472904013e-06, "loss": 0.4676, "num_input_tokens_seen": 20173504, "step": 1305 }, { "epoch": 5.374358974358975, "grad_norm": 0.12451171875, "learning_rate": 2.7947752291359053e-06, "loss": 0.4256, "num_input_tokens_seen": 20244608, "step": 1310 }, { "epoch": 5.394871794871795, "grad_norm": 0.130859375, "learning_rate": 2.7932427274841715e-06, "loss": 0.4576, "num_input_tokens_seen": 20324992, "step": 1315 }, { "epoch": 5.415384615384616, "grad_norm": 0.1376953125, "learning_rate": 2.7917049485887705e-06, "loss": 0.5155, "num_input_tokens_seen": 20402304, "step": 1320 }, { "epoch": 5.435897435897436, "grad_norm": 0.20703125, "learning_rate": 2.790161898724808e-06, "loss": 0.4304, "num_input_tokens_seen": 20480800, "step": 1325 }, { "epoch": 5.456410256410257, "grad_norm": 0.138671875, "learning_rate": 2.7886135841888973e-06, "loss": 0.4759, "num_input_tokens_seen": 20560096, "step": 1330 }, { "epoch": 5.476923076923077, "grad_norm": 0.15234375, "learning_rate": 2.787060011299135e-06, "loss": 0.4974, "num_input_tokens_seen": 20645216, "step": 1335 }, { "epoch": 5.4974358974358974, "grad_norm": 0.341796875, "learning_rate": 2.785501186395077e-06, "loss": 0.5174, "num_input_tokens_seen": 20738688, "step": 1340 }, { "epoch": 5.517948717948718, "grad_norm": 0.271484375, "learning_rate": 2.7839371158377077e-06, "loss": 0.5272, "num_input_tokens_seen": 20812928, "step": 1345 }, { "epoch": 5.538461538461538, "grad_norm": 0.296875, "learning_rate": 2.78236780600942e-06, "loss": 0.5129, "num_input_tokens_seen": 20890592, "step": 1350 }, { "epoch": 5.558974358974359, "grad_norm": 0.57421875, "learning_rate": 2.780793263313984e-06, "loss": 0.484, "num_input_tokens_seen": 20961984, "step": 1355 }, { "epoch": 5.579487179487179, "grad_norm": 0.8671875, "learning_rate": 2.7792134941765247e-06, "loss": 0.4793, "num_input_tokens_seen": 21030784, "step": 1360 }, { "epoch": 5.6, "grad_norm": 0.267578125, "learning_rate": 2.7776285050434937e-06, "loss": 0.4521, "num_input_tokens_seen": 21108960, "step": 1365 }, { "epoch": 5.62051282051282, "grad_norm": 0.267578125, "learning_rate": 2.7760383023826425e-06, "loss": 0.4192, "num_input_tokens_seen": 21181728, "step": 1370 }, { "epoch": 5.641025641025641, "grad_norm": 0.64453125, "learning_rate": 2.7744428926829993e-06, "loss": 0.5131, "num_input_tokens_seen": 21255328, "step": 1375 }, { "epoch": 5.661538461538462, "grad_norm": 0.271484375, "learning_rate": 2.7728422824548387e-06, "loss": 0.483, "num_input_tokens_seen": 21324064, "step": 1380 }, { "epoch": 5.682051282051282, "grad_norm": 0.341796875, "learning_rate": 2.7712364782296567e-06, "loss": 0.4858, "num_input_tokens_seen": 21399040, "step": 1385 }, { "epoch": 5.702564102564103, "grad_norm": 0.318359375, "learning_rate": 2.769625486560145e-06, "loss": 0.4629, "num_input_tokens_seen": 21472640, "step": 1390 }, { "epoch": 5.723076923076923, "grad_norm": 0.39453125, "learning_rate": 2.7680093140201625e-06, "loss": 0.5023, "num_input_tokens_seen": 21544448, "step": 1395 }, { "epoch": 5.743589743589744, "grad_norm": 0.28515625, "learning_rate": 2.766387967204709e-06, "loss": 0.4903, "num_input_tokens_seen": 21611136, "step": 1400 }, { "epoch": 5.764102564102564, "grad_norm": 0.2578125, "learning_rate": 2.7647614527299007e-06, "loss": 0.5558, "num_input_tokens_seen": 21703040, "step": 1405 }, { "epoch": 5.7846153846153845, "grad_norm": 0.45703125, "learning_rate": 2.763129777232938e-06, "loss": 0.5612, "num_input_tokens_seen": 21784096, "step": 1410 }, { "epoch": 5.805128205128205, "grad_norm": 0.333984375, "learning_rate": 2.7614929473720847e-06, "loss": 0.4683, "num_input_tokens_seen": 21855072, "step": 1415 }, { "epoch": 5.825641025641025, "grad_norm": 0.38671875, "learning_rate": 2.7598509698266346e-06, "loss": 0.5171, "num_input_tokens_seen": 21933312, "step": 1420 }, { "epoch": 5.846153846153846, "grad_norm": 0.359375, "learning_rate": 2.758203851296889e-06, "loss": 0.5214, "num_input_tokens_seen": 22019008, "step": 1425 }, { "epoch": 5.866666666666667, "grad_norm": 0.326171875, "learning_rate": 2.756551598504128e-06, "loss": 0.4975, "num_input_tokens_seen": 22092864, "step": 1430 }, { "epoch": 5.887179487179488, "grad_norm": 0.392578125, "learning_rate": 2.7548942181905816e-06, "loss": 0.4853, "num_input_tokens_seen": 22171584, "step": 1435 }, { "epoch": 5.907692307692308, "grad_norm": 0.375, "learning_rate": 2.753231717119405e-06, "loss": 0.483, "num_input_tokens_seen": 22241376, "step": 1440 }, { "epoch": 5.9282051282051285, "grad_norm": 0.373046875, "learning_rate": 2.751564102074646e-06, "loss": 0.4965, "num_input_tokens_seen": 22313664, "step": 1445 }, { "epoch": 5.948717948717949, "grad_norm": 0.34765625, "learning_rate": 2.749891379861225e-06, "loss": 0.5342, "num_input_tokens_seen": 22397408, "step": 1450 }, { "epoch": 5.969230769230769, "grad_norm": 0.3359375, "learning_rate": 2.748213557304899e-06, "loss": 0.4233, "num_input_tokens_seen": 22473664, "step": 1455 }, { "epoch": 5.98974358974359, "grad_norm": 0.349609375, "learning_rate": 2.74653064125224e-06, "loss": 0.5244, "num_input_tokens_seen": 22553760, "step": 1460 }, { "epoch": 6.01025641025641, "grad_norm": 0.298828125, "learning_rate": 2.7448426385706036e-06, "loss": 0.5211, "num_input_tokens_seen": 22624608, "step": 1465 }, { "epoch": 6.030769230769231, "grad_norm": 0.271484375, "learning_rate": 2.7431495561481027e-06, "loss": 0.5618, "num_input_tokens_seen": 22710048, "step": 1470 }, { "epoch": 6.051282051282051, "grad_norm": 0.2734375, "learning_rate": 2.741451400893578e-06, "loss": 0.5172, "num_input_tokens_seen": 22787392, "step": 1475 }, { "epoch": 6.0717948717948715, "grad_norm": 0.22265625, "learning_rate": 2.739748179736571e-06, "loss": 0.5035, "num_input_tokens_seen": 22865120, "step": 1480 }, { "epoch": 6.092307692307692, "grad_norm": 0.205078125, "learning_rate": 2.7380398996272955e-06, "loss": 0.519, "num_input_tokens_seen": 22952832, "step": 1485 }, { "epoch": 6.112820512820512, "grad_norm": 0.2138671875, "learning_rate": 2.736326567536609e-06, "loss": 0.4438, "num_input_tokens_seen": 23028544, "step": 1490 }, { "epoch": 6.133333333333334, "grad_norm": 0.2021484375, "learning_rate": 2.7346081904559827e-06, "loss": 0.4669, "num_input_tokens_seen": 23100096, "step": 1495 }, { "epoch": 6.153846153846154, "grad_norm": 0.1533203125, "learning_rate": 2.732884775397477e-06, "loss": 0.4702, "num_input_tokens_seen": 23183392, "step": 1500 }, { "epoch": 6.174358974358975, "grad_norm": 0.1533203125, "learning_rate": 2.731156329393709e-06, "loss": 0.5031, "num_input_tokens_seen": 23266208, "step": 1505 }, { "epoch": 6.194871794871795, "grad_norm": 0.1484375, "learning_rate": 2.729422859497825e-06, "loss": 0.5005, "num_input_tokens_seen": 23348064, "step": 1510 }, { "epoch": 6.2153846153846155, "grad_norm": 0.1552734375, "learning_rate": 2.7276843727834727e-06, "loss": 0.4798, "num_input_tokens_seen": 23420128, "step": 1515 }, { "epoch": 6.235897435897436, "grad_norm": 0.1337890625, "learning_rate": 2.725940876344771e-06, "loss": 0.5059, "num_input_tokens_seen": 23497056, "step": 1520 }, { "epoch": 6.256410256410256, "grad_norm": 0.1376953125, "learning_rate": 2.7241923772962823e-06, "loss": 0.4582, "num_input_tokens_seen": 23564928, "step": 1525 }, { "epoch": 6.276923076923077, "grad_norm": 0.1220703125, "learning_rate": 2.722438882772982e-06, "loss": 0.4295, "num_input_tokens_seen": 23646624, "step": 1530 }, { "epoch": 6.297435897435897, "grad_norm": 0.1484375, "learning_rate": 2.720680399930231e-06, "loss": 0.4682, "num_input_tokens_seen": 23716960, "step": 1535 }, { "epoch": 6.317948717948718, "grad_norm": 0.1611328125, "learning_rate": 2.7189169359437443e-06, "loss": 0.4944, "num_input_tokens_seen": 23796032, "step": 1540 }, { "epoch": 6.338461538461538, "grad_norm": 0.1318359375, "learning_rate": 2.7171484980095653e-06, "loss": 0.4405, "num_input_tokens_seen": 23868768, "step": 1545 }, { "epoch": 6.358974358974359, "grad_norm": 0.1953125, "learning_rate": 2.715375093344032e-06, "loss": 0.4742, "num_input_tokens_seen": 23937824, "step": 1550 }, { "epoch": 6.37948717948718, "grad_norm": 0.146484375, "learning_rate": 2.713596729183751e-06, "loss": 0.4654, "num_input_tokens_seen": 24009472, "step": 1555 }, { "epoch": 6.4, "grad_norm": 0.146484375, "learning_rate": 2.7118134127855667e-06, "loss": 0.4686, "num_input_tokens_seen": 24096256, "step": 1560 }, { "epoch": 6.420512820512821, "grad_norm": 0.3203125, "learning_rate": 2.7100251514265317e-06, "loss": 0.5152, "num_input_tokens_seen": 24180640, "step": 1565 }, { "epoch": 6.441025641025641, "grad_norm": 0.392578125, "learning_rate": 2.7082319524038764e-06, "loss": 0.4762, "num_input_tokens_seen": 24251296, "step": 1570 }, { "epoch": 6.461538461538462, "grad_norm": 0.74609375, "learning_rate": 2.706433823034981e-06, "loss": 0.5113, "num_input_tokens_seen": 24329760, "step": 1575 }, { "epoch": 6.482051282051282, "grad_norm": 0.55078125, "learning_rate": 2.7046307706573445e-06, "loss": 0.4942, "num_input_tokens_seen": 24420896, "step": 1580 }, { "epoch": 6.5025641025641026, "grad_norm": 0.703125, "learning_rate": 2.702822802628554e-06, "loss": 0.475, "num_input_tokens_seen": 24495360, "step": 1585 }, { "epoch": 6.523076923076923, "grad_norm": 0.2216796875, "learning_rate": 2.701009926326256e-06, "loss": 0.4801, "num_input_tokens_seen": 24572192, "step": 1590 }, { "epoch": 6.543589743589743, "grad_norm": 0.267578125, "learning_rate": 2.6991921491481267e-06, "loss": 0.4776, "num_input_tokens_seen": 24647552, "step": 1595 }, { "epoch": 6.564102564102564, "grad_norm": 0.578125, "learning_rate": 2.6973694785118394e-06, "loss": 0.4878, "num_input_tokens_seen": 24719136, "step": 1600 }, { "epoch": 6.584615384615384, "grad_norm": 0.37109375, "learning_rate": 2.695541921855037e-06, "loss": 0.5138, "num_input_tokens_seen": 24800320, "step": 1605 }, { "epoch": 6.605128205128205, "grad_norm": 0.310546875, "learning_rate": 2.6937094866353006e-06, "loss": 0.4782, "num_input_tokens_seen": 24877088, "step": 1610 }, { "epoch": 6.625641025641025, "grad_norm": 0.498046875, "learning_rate": 2.6918721803301174e-06, "loss": 0.5043, "num_input_tokens_seen": 24954272, "step": 1615 }, { "epoch": 6.6461538461538465, "grad_norm": 0.294921875, "learning_rate": 2.690030010436853e-06, "loss": 0.4237, "num_input_tokens_seen": 25023744, "step": 1620 }, { "epoch": 6.666666666666667, "grad_norm": 0.64453125, "learning_rate": 2.688182984472719e-06, "loss": 0.5302, "num_input_tokens_seen": 25105664, "step": 1625 }, { "epoch": 6.687179487179487, "grad_norm": 0.3515625, "learning_rate": 2.686331109974743e-06, "loss": 0.4991, "num_input_tokens_seen": 25183680, "step": 1630 }, { "epoch": 6.707692307692308, "grad_norm": 0.443359375, "learning_rate": 2.684474394499738e-06, "loss": 0.5142, "num_input_tokens_seen": 25265920, "step": 1635 }, { "epoch": 6.728205128205128, "grad_norm": 0.4609375, "learning_rate": 2.6826128456242708e-06, "loss": 0.4651, "num_input_tokens_seen": 25343648, "step": 1640 }, { "epoch": 6.748717948717949, "grad_norm": 0.361328125, "learning_rate": 2.680746470944631e-06, "loss": 0.5633, "num_input_tokens_seen": 25418176, "step": 1645 }, { "epoch": 6.769230769230769, "grad_norm": 0.375, "learning_rate": 2.6788752780768007e-06, "loss": 0.5124, "num_input_tokens_seen": 25504832, "step": 1650 }, { "epoch": 6.78974358974359, "grad_norm": 0.62109375, "learning_rate": 2.6769992746564256e-06, "loss": 0.5046, "num_input_tokens_seen": 25582112, "step": 1655 }, { "epoch": 6.81025641025641, "grad_norm": 0.35546875, "learning_rate": 2.6751184683387777e-06, "loss": 0.484, "num_input_tokens_seen": 25656992, "step": 1660 }, { "epoch": 6.8307692307692305, "grad_norm": 0.3203125, "learning_rate": 2.67323286679873e-06, "loss": 0.4526, "num_input_tokens_seen": 25729600, "step": 1665 }, { "epoch": 6.851282051282051, "grad_norm": 0.365234375, "learning_rate": 2.671342477730723e-06, "loss": 0.4563, "num_input_tokens_seen": 25801536, "step": 1670 }, { "epoch": 6.871794871794872, "grad_norm": 0.380859375, "learning_rate": 2.6694473088487324e-06, "loss": 0.4951, "num_input_tokens_seen": 25882912, "step": 1675 }, { "epoch": 6.892307692307693, "grad_norm": 0.32421875, "learning_rate": 2.6675473678862403e-06, "loss": 0.5223, "num_input_tokens_seen": 25957952, "step": 1680 }, { "epoch": 6.912820512820513, "grad_norm": 0.353515625, "learning_rate": 2.6656426625961993e-06, "loss": 0.5471, "num_input_tokens_seen": 26034432, "step": 1685 }, { "epoch": 6.933333333333334, "grad_norm": 0.251953125, "learning_rate": 2.6637332007510063e-06, "loss": 0.4252, "num_input_tokens_seen": 26106656, "step": 1690 }, { "epoch": 6.953846153846154, "grad_norm": 0.31640625, "learning_rate": 2.661818990142465e-06, "loss": 0.5269, "num_input_tokens_seen": 26186976, "step": 1695 }, { "epoch": 6.9743589743589745, "grad_norm": 0.25, "learning_rate": 2.65990003858176e-06, "loss": 0.4487, "num_input_tokens_seen": 26259264, "step": 1700 }, { "epoch": 6.994871794871795, "grad_norm": 0.2353515625, "learning_rate": 2.6579763538994197e-06, "loss": 0.4705, "num_input_tokens_seen": 26333248, "step": 1705 }, { "epoch": 7.015384615384615, "grad_norm": 0.205078125, "learning_rate": 2.656047943945287e-06, "loss": 0.4443, "num_input_tokens_seen": 26404832, "step": 1710 }, { "epoch": 7.035897435897436, "grad_norm": 0.1865234375, "learning_rate": 2.6541148165884885e-06, "loss": 0.4615, "num_input_tokens_seen": 26484608, "step": 1715 }, { "epoch": 7.056410256410256, "grad_norm": 0.1943359375, "learning_rate": 2.652176979717399e-06, "loss": 0.5042, "num_input_tokens_seen": 26556224, "step": 1720 }, { "epoch": 7.076923076923077, "grad_norm": 0.16796875, "learning_rate": 2.6502344412396116e-06, "loss": 0.4601, "num_input_tokens_seen": 26629632, "step": 1725 }, { "epoch": 7.097435897435897, "grad_norm": 0.146484375, "learning_rate": 2.6482872090819053e-06, "loss": 0.534, "num_input_tokens_seen": 26709568, "step": 1730 }, { "epoch": 7.1179487179487175, "grad_norm": 0.201171875, "learning_rate": 2.646335291190211e-06, "loss": 0.4875, "num_input_tokens_seen": 26785728, "step": 1735 }, { "epoch": 7.138461538461539, "grad_norm": 0.150390625, "learning_rate": 2.6443786955295827e-06, "loss": 0.5223, "num_input_tokens_seen": 26865024, "step": 1740 }, { "epoch": 7.158974358974359, "grad_norm": 0.140625, "learning_rate": 2.6424174300841606e-06, "loss": 0.4365, "num_input_tokens_seen": 26934720, "step": 1745 }, { "epoch": 7.17948717948718, "grad_norm": 0.138671875, "learning_rate": 2.6404515028571406e-06, "loss": 0.4951, "num_input_tokens_seen": 27008192, "step": 1750 }, { "epoch": 7.2, "grad_norm": 0.11962890625, "learning_rate": 2.638480921870743e-06, "loss": 0.5132, "num_input_tokens_seen": 27092000, "step": 1755 }, { "epoch": 7.220512820512821, "grad_norm": 0.1298828125, "learning_rate": 2.636505695166177e-06, "loss": 0.4713, "num_input_tokens_seen": 27172160, "step": 1760 }, { "epoch": 7.241025641025641, "grad_norm": 0.16015625, "learning_rate": 2.63452583080361e-06, "loss": 0.479, "num_input_tokens_seen": 27255712, "step": 1765 }, { "epoch": 7.2615384615384615, "grad_norm": 0.1396484375, "learning_rate": 2.6325413368621337e-06, "loss": 0.4967, "num_input_tokens_seen": 27343136, "step": 1770 }, { "epoch": 7.282051282051282, "grad_norm": 0.2470703125, "learning_rate": 2.630552221439732e-06, "loss": 0.4843, "num_input_tokens_seen": 27417312, "step": 1775 }, { "epoch": 7.302564102564102, "grad_norm": 0.1396484375, "learning_rate": 2.6285584926532465e-06, "loss": 0.4738, "num_input_tokens_seen": 27505824, "step": 1780 }, { "epoch": 7.323076923076923, "grad_norm": 0.2890625, "learning_rate": 2.626560158638344e-06, "loss": 0.5716, "num_input_tokens_seen": 27583776, "step": 1785 }, { "epoch": 7.343589743589743, "grad_norm": 0.357421875, "learning_rate": 2.6245572275494845e-06, "loss": 0.515, "num_input_tokens_seen": 27658912, "step": 1790 }, { "epoch": 7.364102564102564, "grad_norm": 0.2314453125, "learning_rate": 2.6225497075598865e-06, "loss": 0.47, "num_input_tokens_seen": 27733472, "step": 1795 }, { "epoch": 7.384615384615385, "grad_norm": 0.51953125, "learning_rate": 2.6205376068614943e-06, "loss": 0.4749, "num_input_tokens_seen": 27812160, "step": 1800 }, { "epoch": 7.4051282051282055, "grad_norm": 0.55078125, "learning_rate": 2.6185209336649438e-06, "loss": 0.4727, "num_input_tokens_seen": 27885024, "step": 1805 }, { "epoch": 7.425641025641026, "grad_norm": 0.75, "learning_rate": 2.61649969619953e-06, "loss": 0.478, "num_input_tokens_seen": 27956480, "step": 1810 }, { "epoch": 7.446153846153846, "grad_norm": 0.28125, "learning_rate": 2.614473902713173e-06, "loss": 0.4778, "num_input_tokens_seen": 28028032, "step": 1815 }, { "epoch": 7.466666666666667, "grad_norm": 0.49609375, "learning_rate": 2.612443561472385e-06, "loss": 0.4443, "num_input_tokens_seen": 28112992, "step": 1820 }, { "epoch": 7.487179487179487, "grad_norm": 0.5, "learning_rate": 2.610408680762234e-06, "loss": 0.5186, "num_input_tokens_seen": 28191520, "step": 1825 }, { "epoch": 7.507692307692308, "grad_norm": 0.314453125, "learning_rate": 2.6083692688863135e-06, "loss": 0.5152, "num_input_tokens_seen": 28277440, "step": 1830 }, { "epoch": 7.528205128205128, "grad_norm": 0.29296875, "learning_rate": 2.6063253341667064e-06, "loss": 0.5173, "num_input_tokens_seen": 28357440, "step": 1835 }, { "epoch": 7.5487179487179485, "grad_norm": 0.326171875, "learning_rate": 2.604276884943953e-06, "loss": 0.4585, "num_input_tokens_seen": 28426656, "step": 1840 }, { "epoch": 7.569230769230769, "grad_norm": 0.279296875, "learning_rate": 2.602223929577013e-06, "loss": 0.4611, "num_input_tokens_seen": 28499968, "step": 1845 }, { "epoch": 7.589743589743589, "grad_norm": 0.310546875, "learning_rate": 2.6001664764432363e-06, "loss": 0.4929, "num_input_tokens_seen": 28573664, "step": 1850 }, { "epoch": 7.61025641025641, "grad_norm": 0.318359375, "learning_rate": 2.5981045339383244e-06, "loss": 0.5018, "num_input_tokens_seen": 28658144, "step": 1855 }, { "epoch": 7.63076923076923, "grad_norm": 0.328125, "learning_rate": 2.596038110476301e-06, "loss": 0.483, "num_input_tokens_seen": 28730944, "step": 1860 }, { "epoch": 7.651282051282052, "grad_norm": 0.52734375, "learning_rate": 2.593967214489473e-06, "loss": 0.5111, "num_input_tokens_seen": 28816384, "step": 1865 }, { "epoch": 7.671794871794872, "grad_norm": 0.37109375, "learning_rate": 2.591891854428398e-06, "loss": 0.4689, "num_input_tokens_seen": 28891616, "step": 1870 }, { "epoch": 7.6923076923076925, "grad_norm": 0.33203125, "learning_rate": 2.5898120387618507e-06, "loss": 0.4917, "num_input_tokens_seen": 28970400, "step": 1875 }, { "epoch": 7.712820512820513, "grad_norm": 0.33203125, "learning_rate": 2.587727775976787e-06, "loss": 0.4956, "num_input_tokens_seen": 29051520, "step": 1880 }, { "epoch": 7.733333333333333, "grad_norm": 0.3046875, "learning_rate": 2.585639074578309e-06, "loss": 0.438, "num_input_tokens_seen": 29128544, "step": 1885 }, { "epoch": 7.753846153846154, "grad_norm": 0.306640625, "learning_rate": 2.5835459430896333e-06, "loss": 0.4644, "num_input_tokens_seen": 29210496, "step": 1890 }, { "epoch": 7.774358974358974, "grad_norm": 0.365234375, "learning_rate": 2.5814483900520522e-06, "loss": 0.4901, "num_input_tokens_seen": 29282400, "step": 1895 }, { "epoch": 7.794871794871795, "grad_norm": 0.4140625, "learning_rate": 2.5793464240249014e-06, "loss": 0.4879, "num_input_tokens_seen": 29352256, "step": 1900 }, { "epoch": 7.815384615384615, "grad_norm": 0.330078125, "learning_rate": 2.5772400535855242e-06, "loss": 0.4552, "num_input_tokens_seen": 29426336, "step": 1905 }, { "epoch": 7.835897435897436, "grad_norm": 0.353515625, "learning_rate": 2.575129287329237e-06, "loss": 0.5417, "num_input_tokens_seen": 29512224, "step": 1910 }, { "epoch": 7.856410256410256, "grad_norm": 0.27734375, "learning_rate": 2.5730141338692926e-06, "loss": 0.4637, "num_input_tokens_seen": 29590112, "step": 1915 }, { "epoch": 7.876923076923077, "grad_norm": 0.31640625, "learning_rate": 2.5708946018368487e-06, "loss": 0.4486, "num_input_tokens_seen": 29672608, "step": 1920 }, { "epoch": 7.897435897435898, "grad_norm": 0.228515625, "learning_rate": 2.568770699880928e-06, "loss": 0.5094, "num_input_tokens_seen": 29755520, "step": 1925 }, { "epoch": 7.917948717948718, "grad_norm": 0.2197265625, "learning_rate": 2.566642436668387e-06, "loss": 0.5111, "num_input_tokens_seen": 29833344, "step": 1930 }, { "epoch": 7.938461538461539, "grad_norm": 0.1923828125, "learning_rate": 2.5645098208838774e-06, "loss": 0.4737, "num_input_tokens_seen": 29904800, "step": 1935 }, { "epoch": 7.958974358974359, "grad_norm": 0.2080078125, "learning_rate": 2.562372861229813e-06, "loss": 0.4384, "num_input_tokens_seen": 29975488, "step": 1940 }, { "epoch": 7.97948717948718, "grad_norm": 0.197265625, "learning_rate": 2.5602315664263337e-06, "loss": 0.4383, "num_input_tokens_seen": 30046496, "step": 1945 }, { "epoch": 8.0, "grad_norm": 0.1611328125, "learning_rate": 2.5580859452112685e-06, "loss": 0.4782, "num_input_tokens_seen": 30119840, "step": 1950 }, { "epoch": 8.02051282051282, "grad_norm": 0.1484375, "learning_rate": 2.555936006340101e-06, "loss": 0.5371, "num_input_tokens_seen": 30207040, "step": 1955 }, { "epoch": 8.04102564102564, "grad_norm": 0.1376953125, "learning_rate": 2.553781758585935e-06, "loss": 0.4867, "num_input_tokens_seen": 30283968, "step": 1960 }, { "epoch": 8.061538461538461, "grad_norm": 0.1376953125, "learning_rate": 2.551623210739455e-06, "loss": 0.4309, "num_input_tokens_seen": 30355552, "step": 1965 }, { "epoch": 8.082051282051282, "grad_norm": 0.12109375, "learning_rate": 2.549460371608895e-06, "loss": 0.5087, "num_input_tokens_seen": 30435776, "step": 1970 }, { "epoch": 8.102564102564102, "grad_norm": 0.1328125, "learning_rate": 2.5472932500199976e-06, "loss": 0.4746, "num_input_tokens_seen": 30507616, "step": 1975 }, { "epoch": 8.123076923076923, "grad_norm": 0.1396484375, "learning_rate": 2.5451218548159823e-06, "loss": 0.4833, "num_input_tokens_seen": 30583456, "step": 1980 }, { "epoch": 8.143589743589743, "grad_norm": 0.12890625, "learning_rate": 2.5429461948575077e-06, "loss": 0.4849, "num_input_tokens_seen": 30654176, "step": 1985 }, { "epoch": 8.164102564102564, "grad_norm": 0.1416015625, "learning_rate": 2.540766279022634e-06, "loss": 0.4812, "num_input_tokens_seen": 30721920, "step": 1990 }, { "epoch": 8.184615384615384, "grad_norm": 0.216796875, "learning_rate": 2.53858211620679e-06, "loss": 0.4976, "num_input_tokens_seen": 30791264, "step": 1995 }, { "epoch": 8.205128205128204, "grad_norm": 0.20703125, "learning_rate": 2.536393715322732e-06, "loss": 0.4556, "num_input_tokens_seen": 30862336, "step": 2000 }, { "epoch": 8.225641025641025, "grad_norm": 0.150390625, "learning_rate": 2.5342010853005127e-06, "loss": 0.4496, "num_input_tokens_seen": 30940064, "step": 2005 }, { "epoch": 8.246153846153845, "grad_norm": 0.318359375, "learning_rate": 2.532004235087441e-06, "loss": 0.4722, "num_input_tokens_seen": 31013248, "step": 2010 }, { "epoch": 8.266666666666667, "grad_norm": 0.26171875, "learning_rate": 2.529803173648049e-06, "loss": 0.4875, "num_input_tokens_seen": 31094496, "step": 2015 }, { "epoch": 8.287179487179488, "grad_norm": 0.2578125, "learning_rate": 2.52759790996405e-06, "loss": 0.4598, "num_input_tokens_seen": 31171680, "step": 2020 }, { "epoch": 8.307692307692308, "grad_norm": 0.77734375, "learning_rate": 2.525388453034307e-06, "loss": 0.5069, "num_input_tokens_seen": 31252064, "step": 2025 }, { "epoch": 8.328205128205129, "grad_norm": 0.6875, "learning_rate": 2.5231748118747945e-06, "loss": 0.5155, "num_input_tokens_seen": 31329696, "step": 2030 }, { "epoch": 8.34871794871795, "grad_norm": 0.96484375, "learning_rate": 2.5209569955185604e-06, "loss": 0.5436, "num_input_tokens_seen": 31407648, "step": 2035 }, { "epoch": 8.36923076923077, "grad_norm": 0.30078125, "learning_rate": 2.51873501301569e-06, "loss": 0.4953, "num_input_tokens_seen": 31475200, "step": 2040 }, { "epoch": 8.38974358974359, "grad_norm": 0.6328125, "learning_rate": 2.5165088734332695e-06, "loss": 0.4804, "num_input_tokens_seen": 31547104, "step": 2045 }, { "epoch": 8.41025641025641, "grad_norm": 0.52734375, "learning_rate": 2.5142785858553486e-06, "loss": 0.5533, "num_input_tokens_seen": 31629440, "step": 2050 }, { "epoch": 8.430769230769231, "grad_norm": 0.310546875, "learning_rate": 2.512044159382903e-06, "loss": 0.541, "num_input_tokens_seen": 31713024, "step": 2055 }, { "epoch": 8.451282051282051, "grad_norm": 0.71484375, "learning_rate": 2.5098056031337975e-06, "loss": 0.4444, "num_input_tokens_seen": 31790432, "step": 2060 }, { "epoch": 8.471794871794872, "grad_norm": 0.3046875, "learning_rate": 2.5075629262427507e-06, "loss": 0.4869, "num_input_tokens_seen": 31870592, "step": 2065 }, { "epoch": 8.492307692307692, "grad_norm": 0.6015625, "learning_rate": 2.505316137861294e-06, "loss": 0.4855, "num_input_tokens_seen": 31945344, "step": 2070 }, { "epoch": 8.512820512820513, "grad_norm": 0.283203125, "learning_rate": 2.503065247157737e-06, "loss": 0.5027, "num_input_tokens_seen": 32030016, "step": 2075 }, { "epoch": 8.533333333333333, "grad_norm": 0.3515625, "learning_rate": 2.500810263317129e-06, "loss": 0.4885, "num_input_tokens_seen": 32108160, "step": 2080 }, { "epoch": 8.553846153846154, "grad_norm": 0.33203125, "learning_rate": 2.4985511955412238e-06, "loss": 0.4451, "num_input_tokens_seen": 32188288, "step": 2085 }, { "epoch": 8.574358974358974, "grad_norm": 0.35546875, "learning_rate": 2.4962880530484375e-06, "loss": 0.4899, "num_input_tokens_seen": 32266656, "step": 2090 }, { "epoch": 8.594871794871795, "grad_norm": 0.3515625, "learning_rate": 2.4940208450738146e-06, "loss": 0.5083, "num_input_tokens_seen": 32356544, "step": 2095 }, { "epoch": 8.615384615384615, "grad_norm": 0.37890625, "learning_rate": 2.49174958086899e-06, "loss": 0.4839, "num_input_tokens_seen": 32434720, "step": 2100 }, { "epoch": 8.635897435897435, "grad_norm": 0.373046875, "learning_rate": 2.48947426970215e-06, "loss": 0.437, "num_input_tokens_seen": 32507712, "step": 2105 }, { "epoch": 8.656410256410256, "grad_norm": 0.361328125, "learning_rate": 2.487194920857995e-06, "loss": 0.435, "num_input_tokens_seen": 32577216, "step": 2110 }, { "epoch": 8.676923076923076, "grad_norm": 0.3359375, "learning_rate": 2.484911543637702e-06, "loss": 0.4768, "num_input_tokens_seen": 32647552, "step": 2115 }, { "epoch": 8.697435897435897, "grad_norm": 0.34765625, "learning_rate": 2.4826241473588855e-06, "loss": 0.4578, "num_input_tokens_seen": 32727520, "step": 2120 }, { "epoch": 8.717948717948717, "grad_norm": 0.365234375, "learning_rate": 2.4803327413555623e-06, "loss": 0.5142, "num_input_tokens_seen": 32805440, "step": 2125 }, { "epoch": 8.73846153846154, "grad_norm": 0.35546875, "learning_rate": 2.4780373349781083e-06, "loss": 0.4013, "num_input_tokens_seen": 32880480, "step": 2130 }, { "epoch": 8.75897435897436, "grad_norm": 0.30078125, "learning_rate": 2.4757379375932265e-06, "loss": 0.4616, "num_input_tokens_seen": 32951936, "step": 2135 }, { "epoch": 8.77948717948718, "grad_norm": 0.287109375, "learning_rate": 2.473434558583903e-06, "loss": 0.4791, "num_input_tokens_seen": 33027104, "step": 2140 }, { "epoch": 8.8, "grad_norm": 0.2451171875, "learning_rate": 2.4711272073493745e-06, "loss": 0.5163, "num_input_tokens_seen": 33111040, "step": 2145 }, { "epoch": 8.820512820512821, "grad_norm": 0.2470703125, "learning_rate": 2.468815893305084e-06, "loss": 0.4761, "num_input_tokens_seen": 33202304, "step": 2150 }, { "epoch": 8.841025641025642, "grad_norm": 0.2275390625, "learning_rate": 2.466500625882646e-06, "loss": 0.4405, "num_input_tokens_seen": 33281376, "step": 2155 }, { "epoch": 8.861538461538462, "grad_norm": 0.21484375, "learning_rate": 2.464181414529809e-06, "loss": 0.4538, "num_input_tokens_seen": 33352640, "step": 2160 }, { "epoch": 8.882051282051282, "grad_norm": 0.19140625, "learning_rate": 2.4618582687104132e-06, "loss": 0.4598, "num_input_tokens_seen": 33423232, "step": 2165 }, { "epoch": 8.902564102564103, "grad_norm": 0.169921875, "learning_rate": 2.4595311979043545e-06, "loss": 0.4556, "num_input_tokens_seen": 33503744, "step": 2170 }, { "epoch": 8.923076923076923, "grad_norm": 0.16796875, "learning_rate": 2.4572002116075454e-06, "loss": 0.4665, "num_input_tokens_seen": 33576800, "step": 2175 }, { "epoch": 8.943589743589744, "grad_norm": 0.1748046875, "learning_rate": 2.454865319331876e-06, "loss": 0.4683, "num_input_tokens_seen": 33661120, "step": 2180 }, { "epoch": 8.964102564102564, "grad_norm": 0.1484375, "learning_rate": 2.4525265306051755e-06, "loss": 0.5183, "num_input_tokens_seen": 33733568, "step": 2185 }, { "epoch": 8.984615384615385, "grad_norm": 0.15234375, "learning_rate": 2.4501838549711723e-06, "loss": 0.4671, "num_input_tokens_seen": 33802592, "step": 2190 }, { "epoch": 9.005128205128205, "grad_norm": 0.126953125, "learning_rate": 2.447837301989457e-06, "loss": 0.4858, "num_input_tokens_seen": 33882272, "step": 2195 }, { "epoch": 9.025641025641026, "grad_norm": 0.138671875, "learning_rate": 2.4454868812354403e-06, "loss": 0.4574, "num_input_tokens_seen": 33953920, "step": 2200 }, { "epoch": 9.046153846153846, "grad_norm": 0.11669921875, "learning_rate": 2.4431326023003188e-06, "loss": 0.4419, "num_input_tokens_seen": 34027552, "step": 2205 }, { "epoch": 9.066666666666666, "grad_norm": 0.11669921875, "learning_rate": 2.44077447479103e-06, "loss": 0.5126, "num_input_tokens_seen": 34112480, "step": 2210 }, { "epoch": 9.087179487179487, "grad_norm": 0.1357421875, "learning_rate": 2.4384125083302178e-06, "loss": 0.4517, "num_input_tokens_seen": 34183840, "step": 2215 }, { "epoch": 9.107692307692307, "grad_norm": 0.251953125, "learning_rate": 2.4360467125561907e-06, "loss": 0.5161, "num_input_tokens_seen": 34258912, "step": 2220 }, { "epoch": 9.128205128205128, "grad_norm": 0.240234375, "learning_rate": 2.433677097122883e-06, "loss": 0.4486, "num_input_tokens_seen": 34329248, "step": 2225 }, { "epoch": 9.148717948717948, "grad_norm": 0.166015625, "learning_rate": 2.4313036716998154e-06, "loss": 0.5191, "num_input_tokens_seen": 34411232, "step": 2230 }, { "epoch": 9.169230769230769, "grad_norm": 0.3125, "learning_rate": 2.428926445972058e-06, "loss": 0.5117, "num_input_tokens_seen": 34486144, "step": 2235 }, { "epoch": 9.189743589743589, "grad_norm": 0.2734375, "learning_rate": 2.4265454296401857e-06, "loss": 0.4739, "num_input_tokens_seen": 34564864, "step": 2240 }, { "epoch": 9.21025641025641, "grad_norm": 0.404296875, "learning_rate": 2.4241606324202426e-06, "loss": 0.4468, "num_input_tokens_seen": 34640704, "step": 2245 }, { "epoch": 9.23076923076923, "grad_norm": 0.53125, "learning_rate": 2.4217720640437015e-06, "loss": 0.457, "num_input_tokens_seen": 34715744, "step": 2250 }, { "epoch": 9.25128205128205, "grad_norm": 0.64453125, "learning_rate": 2.4193797342574235e-06, "loss": 0.4915, "num_input_tokens_seen": 34798144, "step": 2255 }, { "epoch": 9.271794871794873, "grad_norm": 0.71875, "learning_rate": 2.4169836528236187e-06, "loss": 0.4417, "num_input_tokens_seen": 34873440, "step": 2260 }, { "epoch": 9.292307692307693, "grad_norm": 0.26953125, "learning_rate": 2.4145838295198066e-06, "loss": 0.4999, "num_input_tokens_seen": 34951552, "step": 2265 }, { "epoch": 9.312820512820513, "grad_norm": 1.203125, "learning_rate": 2.4121802741387743e-06, "loss": 0.453, "num_input_tokens_seen": 35021184, "step": 2270 }, { "epoch": 9.333333333333334, "grad_norm": 0.298828125, "learning_rate": 2.4097729964885407e-06, "loss": 0.4473, "num_input_tokens_seen": 35098080, "step": 2275 }, { "epoch": 9.353846153846154, "grad_norm": 0.53515625, "learning_rate": 2.4073620063923123e-06, "loss": 0.4749, "num_input_tokens_seen": 35170336, "step": 2280 }, { "epoch": 9.374358974358975, "grad_norm": 0.314453125, "learning_rate": 2.404947313688445e-06, "loss": 0.5229, "num_input_tokens_seen": 35255200, "step": 2285 }, { "epoch": 9.394871794871795, "grad_norm": 0.4140625, "learning_rate": 2.4025289282304037e-06, "loss": 0.5158, "num_input_tokens_seen": 35328928, "step": 2290 }, { "epoch": 9.415384615384616, "grad_norm": 0.34375, "learning_rate": 2.4001068598867216e-06, "loss": 0.4548, "num_input_tokens_seen": 35402976, "step": 2295 }, { "epoch": 9.435897435897436, "grad_norm": 0.2373046875, "learning_rate": 2.397681118540961e-06, "loss": 0.4313, "num_input_tokens_seen": 35481344, "step": 2300 }, { "epoch": 9.456410256410257, "grad_norm": 0.369140625, "learning_rate": 2.3952517140916724e-06, "loss": 0.4664, "num_input_tokens_seen": 35553664, "step": 2305 }, { "epoch": 9.476923076923077, "grad_norm": 0.37890625, "learning_rate": 2.392818656452354e-06, "loss": 0.4948, "num_input_tokens_seen": 35630592, "step": 2310 }, { "epoch": 9.497435897435897, "grad_norm": 0.3046875, "learning_rate": 2.39038195555141e-06, "loss": 0.5035, "num_input_tokens_seen": 35712864, "step": 2315 }, { "epoch": 9.517948717948718, "grad_norm": 0.34765625, "learning_rate": 2.387941621332114e-06, "loss": 0.4964, "num_input_tokens_seen": 35790784, "step": 2320 }, { "epoch": 9.538461538461538, "grad_norm": 0.33203125, "learning_rate": 2.3854976637525637e-06, "loss": 0.4684, "num_input_tokens_seen": 35868960, "step": 2325 }, { "epoch": 9.558974358974359, "grad_norm": 0.345703125, "learning_rate": 2.3830500927856433e-06, "loss": 0.5117, "num_input_tokens_seen": 35956832, "step": 2330 }, { "epoch": 9.57948717948718, "grad_norm": 0.345703125, "learning_rate": 2.3805989184189813e-06, "loss": 0.4574, "num_input_tokens_seen": 36027520, "step": 2335 }, { "epoch": 9.6, "grad_norm": 0.30859375, "learning_rate": 2.378144150654911e-06, "loss": 0.5291, "num_input_tokens_seen": 36109248, "step": 2340 }, { "epoch": 9.62051282051282, "grad_norm": 0.357421875, "learning_rate": 2.3756857995104286e-06, "loss": 0.4528, "num_input_tokens_seen": 36179584, "step": 2345 }, { "epoch": 9.64102564102564, "grad_norm": 0.3671875, "learning_rate": 2.3732238750171527e-06, "loss": 0.4976, "num_input_tokens_seen": 36257216, "step": 2350 }, { "epoch": 9.661538461538461, "grad_norm": 0.30859375, "learning_rate": 2.3707583872212837e-06, "loss": 0.4576, "num_input_tokens_seen": 36326272, "step": 2355 }, { "epoch": 9.682051282051281, "grad_norm": 0.2294921875, "learning_rate": 2.3682893461835626e-06, "loss": 0.5315, "num_input_tokens_seen": 36405504, "step": 2360 }, { "epoch": 9.702564102564102, "grad_norm": 0.28515625, "learning_rate": 2.3658167619792294e-06, "loss": 0.5407, "num_input_tokens_seen": 36481536, "step": 2365 }, { "epoch": 9.723076923076922, "grad_norm": 0.267578125, "learning_rate": 2.363340644697983e-06, "loss": 0.4525, "num_input_tokens_seen": 36563200, "step": 2370 }, { "epoch": 9.743589743589745, "grad_norm": 0.205078125, "learning_rate": 2.360861004443939e-06, "loss": 0.5628, "num_input_tokens_seen": 36661120, "step": 2375 }, { "epoch": 9.764102564102565, "grad_norm": 0.2119140625, "learning_rate": 2.358377851335589e-06, "loss": 0.4633, "num_input_tokens_seen": 36738880, "step": 2380 }, { "epoch": 9.784615384615385, "grad_norm": 0.21484375, "learning_rate": 2.3558911955057592e-06, "loss": 0.4438, "num_input_tokens_seen": 36820128, "step": 2385 }, { "epoch": 9.805128205128206, "grad_norm": 0.1806640625, "learning_rate": 2.35340104710157e-06, "loss": 0.4609, "num_input_tokens_seen": 36900128, "step": 2390 }, { "epoch": 9.825641025641026, "grad_norm": 0.1708984375, "learning_rate": 2.350907416284392e-06, "loss": 0.4843, "num_input_tokens_seen": 36969024, "step": 2395 }, { "epoch": 9.846153846153847, "grad_norm": 0.162109375, "learning_rate": 2.348410313229808e-06, "loss": 0.4607, "num_input_tokens_seen": 37053440, "step": 2400 }, { "epoch": 9.866666666666667, "grad_norm": 0.146484375, "learning_rate": 2.3459097481275687e-06, "loss": 0.5134, "num_input_tokens_seen": 37132128, "step": 2405 }, { "epoch": 9.887179487179488, "grad_norm": 0.1396484375, "learning_rate": 2.343405731181552e-06, "loss": 0.478, "num_input_tokens_seen": 37209664, "step": 2410 }, { "epoch": 9.907692307692308, "grad_norm": 0.146484375, "learning_rate": 2.3408982726097227e-06, "loss": 0.4864, "num_input_tokens_seen": 37283936, "step": 2415 }, { "epoch": 9.928205128205128, "grad_norm": 0.1455078125, "learning_rate": 2.3383873826440878e-06, "loss": 0.4876, "num_input_tokens_seen": 37359552, "step": 2420 }, { "epoch": 9.948717948717949, "grad_norm": 0.13671875, "learning_rate": 2.3358730715306574e-06, "loss": 0.5265, "num_input_tokens_seen": 37434112, "step": 2425 }, { "epoch": 9.96923076923077, "grad_norm": 0.12353515625, "learning_rate": 2.3333553495294033e-06, "loss": 0.4759, "num_input_tokens_seen": 37511456, "step": 2430 }, { "epoch": 9.98974358974359, "grad_norm": 0.11865234375, "learning_rate": 2.330834226914214e-06, "loss": 0.4336, "num_input_tokens_seen": 37584320, "step": 2435 }, { "epoch": 10.01025641025641, "grad_norm": 0.123046875, "learning_rate": 2.3283097139728557e-06, "loss": 0.5338, "num_input_tokens_seen": 37672864, "step": 2440 }, { "epoch": 10.03076923076923, "grad_norm": 0.220703125, "learning_rate": 2.3257818210069277e-06, "loss": 0.4542, "num_input_tokens_seen": 37743488, "step": 2445 }, { "epoch": 10.051282051282051, "grad_norm": 0.14453125, "learning_rate": 2.3232505583318246e-06, "loss": 0.5379, "num_input_tokens_seen": 37832256, "step": 2450 }, { "epoch": 10.071794871794872, "grad_norm": 0.1533203125, "learning_rate": 2.320715936276689e-06, "loss": 0.4418, "num_input_tokens_seen": 37908832, "step": 2455 }, { "epoch": 10.092307692307692, "grad_norm": 0.3984375, "learning_rate": 2.3181779651843738e-06, "loss": 0.5294, "num_input_tokens_seen": 37989088, "step": 2460 }, { "epoch": 10.112820512820512, "grad_norm": 0.26953125, "learning_rate": 2.3156366554113967e-06, "loss": 0.4438, "num_input_tokens_seen": 38061024, "step": 2465 }, { "epoch": 10.133333333333333, "grad_norm": 0.318359375, "learning_rate": 2.3130920173278997e-06, "loss": 0.4544, "num_input_tokens_seen": 38136480, "step": 2470 }, { "epoch": 10.153846153846153, "grad_norm": 0.55078125, "learning_rate": 2.310544061317607e-06, "loss": 0.4918, "num_input_tokens_seen": 38213280, "step": 2475 }, { "epoch": 10.174358974358974, "grad_norm": 0.8984375, "learning_rate": 2.307992797777782e-06, "loss": 0.4821, "num_input_tokens_seen": 38283296, "step": 2480 }, { "epoch": 10.194871794871794, "grad_norm": 0.27734375, "learning_rate": 2.3054382371191836e-06, "loss": 0.464, "num_input_tokens_seen": 38360320, "step": 2485 }, { "epoch": 10.215384615384615, "grad_norm": 0.251953125, "learning_rate": 2.3028803897660256e-06, "loss": 0.4829, "num_input_tokens_seen": 38440000, "step": 2490 }, { "epoch": 10.235897435897435, "grad_norm": 0.53125, "learning_rate": 2.3003192661559346e-06, "loss": 0.4495, "num_input_tokens_seen": 38510208, "step": 2495 }, { "epoch": 10.256410256410255, "grad_norm": 0.30859375, "learning_rate": 2.297754876739905e-06, "loss": 0.5503, "num_input_tokens_seen": 38600256, "step": 2500 }, { "epoch": 10.276923076923078, "grad_norm": 0.384765625, "learning_rate": 2.2951872319822597e-06, "loss": 0.5341, "num_input_tokens_seen": 38679168, "step": 2505 }, { "epoch": 10.297435897435898, "grad_norm": 0.3046875, "learning_rate": 2.2926163423606027e-06, "loss": 0.4903, "num_input_tokens_seen": 38752064, "step": 2510 }, { "epoch": 10.317948717948719, "grad_norm": 0.353515625, "learning_rate": 2.2900422183657816e-06, "loss": 0.4543, "num_input_tokens_seen": 38829504, "step": 2515 }, { "epoch": 10.338461538461539, "grad_norm": 0.330078125, "learning_rate": 2.2874648705018403e-06, "loss": 0.5428, "num_input_tokens_seen": 38915904, "step": 2520 }, { "epoch": 10.35897435897436, "grad_norm": 0.349609375, "learning_rate": 2.28488430928598e-06, "loss": 0.4588, "num_input_tokens_seen": 38997760, "step": 2525 }, { "epoch": 10.37948717948718, "grad_norm": 0.37109375, "learning_rate": 2.282300545248512e-06, "loss": 0.4441, "num_input_tokens_seen": 39068000, "step": 2530 }, { "epoch": 10.4, "grad_norm": 0.375, "learning_rate": 2.27971358893282e-06, "loss": 0.4441, "num_input_tokens_seen": 39136480, "step": 2535 }, { "epoch": 10.42051282051282, "grad_norm": 0.421875, "learning_rate": 2.2771234508953116e-06, "loss": 0.442, "num_input_tokens_seen": 39219488, "step": 2540 }, { "epoch": 10.441025641025641, "grad_norm": 0.625, "learning_rate": 2.27453014170538e-06, "loss": 0.5199, "num_input_tokens_seen": 39295712, "step": 2545 }, { "epoch": 10.461538461538462, "grad_norm": 0.345703125, "learning_rate": 2.271933671945357e-06, "loss": 0.5147, "num_input_tokens_seen": 39371008, "step": 2550 }, { "epoch": 10.482051282051282, "grad_norm": 0.359375, "learning_rate": 2.2693340522104727e-06, "loss": 0.5296, "num_input_tokens_seen": 39450944, "step": 2555 }, { "epoch": 10.502564102564103, "grad_norm": 0.310546875, "learning_rate": 2.2667312931088096e-06, "loss": 0.4707, "num_input_tokens_seen": 39530912, "step": 2560 }, { "epoch": 10.523076923076923, "grad_norm": 0.34375, "learning_rate": 2.264125405261263e-06, "loss": 0.4726, "num_input_tokens_seen": 39602400, "step": 2565 }, { "epoch": 10.543589743589743, "grad_norm": 0.37890625, "learning_rate": 2.261516399301493e-06, "loss": 0.4344, "num_input_tokens_seen": 39679232, "step": 2570 }, { "epoch": 10.564102564102564, "grad_norm": 0.302734375, "learning_rate": 2.2589042858758853e-06, "loss": 0.4427, "num_input_tokens_seen": 39755904, "step": 2575 }, { "epoch": 10.584615384615384, "grad_norm": 0.326171875, "learning_rate": 2.256289075643506e-06, "loss": 0.4975, "num_input_tokens_seen": 39826368, "step": 2580 }, { "epoch": 10.605128205128205, "grad_norm": 0.287109375, "learning_rate": 2.2536707792760566e-06, "loss": 0.5045, "num_input_tokens_seen": 39906464, "step": 2585 }, { "epoch": 10.625641025641025, "grad_norm": 0.27734375, "learning_rate": 2.251049407457833e-06, "loss": 0.4833, "num_input_tokens_seen": 39975712, "step": 2590 }, { "epoch": 10.646153846153846, "grad_norm": 0.263671875, "learning_rate": 2.2484249708856823e-06, "loss": 0.4984, "num_input_tokens_seen": 40060032, "step": 2595 }, { "epoch": 10.666666666666666, "grad_norm": 0.2333984375, "learning_rate": 2.2457974802689545e-06, "loss": 0.4186, "num_input_tokens_seen": 40131520, "step": 2600 }, { "epoch": 10.687179487179487, "grad_norm": 0.1962890625, "learning_rate": 2.2431669463294646e-06, "loss": 0.4441, "num_input_tokens_seen": 40205760, "step": 2605 }, { "epoch": 10.707692307692307, "grad_norm": 0.2119140625, "learning_rate": 2.2405333798014453e-06, "loss": 0.5337, "num_input_tokens_seen": 40288992, "step": 2610 }, { "epoch": 10.728205128205127, "grad_norm": 0.208984375, "learning_rate": 2.237896791431505e-06, "loss": 0.4703, "num_input_tokens_seen": 40369440, "step": 2615 }, { "epoch": 10.74871794871795, "grad_norm": 0.1630859375, "learning_rate": 2.2352571919785812e-06, "loss": 0.5217, "num_input_tokens_seen": 40452288, "step": 2620 }, { "epoch": 10.76923076923077, "grad_norm": 0.1572265625, "learning_rate": 2.2326145922139004e-06, "loss": 0.4475, "num_input_tokens_seen": 40523808, "step": 2625 }, { "epoch": 10.78974358974359, "grad_norm": 0.1884765625, "learning_rate": 2.2299690029209313e-06, "loss": 0.4734, "num_input_tokens_seen": 40606496, "step": 2630 }, { "epoch": 10.810256410256411, "grad_norm": 0.1259765625, "learning_rate": 2.227320434895343e-06, "loss": 0.4686, "num_input_tokens_seen": 40684672, "step": 2635 }, { "epoch": 10.830769230769231, "grad_norm": 0.1279296875, "learning_rate": 2.2246688989449577e-06, "loss": 0.5027, "num_input_tokens_seen": 40762752, "step": 2640 }, { "epoch": 10.851282051282052, "grad_norm": 0.12109375, "learning_rate": 2.2220144058897104e-06, "loss": 0.4582, "num_input_tokens_seen": 40851776, "step": 2645 }, { "epoch": 10.871794871794872, "grad_norm": 0.1318359375, "learning_rate": 2.2193569665616017e-06, "loss": 0.4516, "num_input_tokens_seen": 40922304, "step": 2650 }, { "epoch": 10.892307692307693, "grad_norm": 0.12109375, "learning_rate": 2.2166965918046554e-06, "loss": 0.4346, "num_input_tokens_seen": 40998784, "step": 2655 }, { "epoch": 10.912820512820513, "grad_norm": 0.16015625, "learning_rate": 2.214033292474874e-06, "loss": 0.5093, "num_input_tokens_seen": 41075872, "step": 2660 }, { "epoch": 10.933333333333334, "grad_norm": 0.154296875, "learning_rate": 2.2113670794401935e-06, "loss": 0.4853, "num_input_tokens_seen": 41149728, "step": 2665 }, { "epoch": 10.953846153846154, "grad_norm": 0.2333984375, "learning_rate": 2.20869796358044e-06, "loss": 0.5042, "num_input_tokens_seen": 41220160, "step": 2670 }, { "epoch": 10.974358974358974, "grad_norm": 0.1630859375, "learning_rate": 2.2060259557872845e-06, "loss": 0.4601, "num_input_tokens_seen": 41287712, "step": 2675 }, { "epoch": 10.994871794871795, "grad_norm": 0.1533203125, "learning_rate": 2.2033510669641997e-06, "loss": 0.4138, "num_input_tokens_seen": 41363264, "step": 2680 }, { "epoch": 11.015384615384615, "grad_norm": 0.306640625, "learning_rate": 2.2006733080264144e-06, "loss": 0.4724, "num_input_tokens_seen": 41437152, "step": 2685 }, { "epoch": 11.035897435897436, "grad_norm": 0.26953125, "learning_rate": 2.197992689900869e-06, "loss": 0.4932, "num_input_tokens_seen": 41515520, "step": 2690 }, { "epoch": 11.056410256410256, "grad_norm": 0.5859375, "learning_rate": 2.195309223526171e-06, "loss": 0.4299, "num_input_tokens_seen": 41587200, "step": 2695 }, { "epoch": 11.076923076923077, "grad_norm": 0.59375, "learning_rate": 2.192622919852551e-06, "loss": 0.4774, "num_input_tokens_seen": 41663008, "step": 2700 }, { "epoch": 11.097435897435897, "grad_norm": 0.6640625, "learning_rate": 2.1899337898418174e-06, "loss": 0.5241, "num_input_tokens_seen": 41743264, "step": 2705 }, { "epoch": 11.117948717948718, "grad_norm": 0.29296875, "learning_rate": 2.187241844467313e-06, "loss": 0.498, "num_input_tokens_seen": 41823264, "step": 2710 }, { "epoch": 11.138461538461538, "grad_norm": 0.2890625, "learning_rate": 2.1845470947138658e-06, "loss": 0.5311, "num_input_tokens_seen": 41914560, "step": 2715 }, { "epoch": 11.158974358974358, "grad_norm": 0.58984375, "learning_rate": 2.181849551577751e-06, "loss": 0.4464, "num_input_tokens_seen": 41988288, "step": 2720 }, { "epoch": 11.179487179487179, "grad_norm": 0.34375, "learning_rate": 2.179149226066641e-06, "loss": 0.4905, "num_input_tokens_seen": 42068416, "step": 2725 }, { "epoch": 11.2, "grad_norm": 0.275390625, "learning_rate": 2.1764461291995618e-06, "loss": 0.4629, "num_input_tokens_seen": 42139744, "step": 2730 }, { "epoch": 11.22051282051282, "grad_norm": 0.2578125, "learning_rate": 2.173740272006849e-06, "loss": 0.5278, "num_input_tokens_seen": 42228672, "step": 2735 }, { "epoch": 11.24102564102564, "grad_norm": 0.6640625, "learning_rate": 2.1710316655301016e-06, "loss": 0.4513, "num_input_tokens_seen": 42306528, "step": 2740 }, { "epoch": 11.261538461538462, "grad_norm": 0.2470703125, "learning_rate": 2.1683203208221375e-06, "loss": 0.4917, "num_input_tokens_seen": 42389024, "step": 2745 }, { "epoch": 11.282051282051283, "grad_norm": 0.333984375, "learning_rate": 2.165606248946948e-06, "loss": 0.4159, "num_input_tokens_seen": 42473088, "step": 2750 }, { "epoch": 11.302564102564103, "grad_norm": 0.291015625, "learning_rate": 2.1628894609796533e-06, "loss": 0.4917, "num_input_tokens_seen": 42551360, "step": 2755 }, { "epoch": 11.323076923076924, "grad_norm": 0.390625, "learning_rate": 2.1601699680064573e-06, "loss": 0.5037, "num_input_tokens_seen": 42626688, "step": 2760 }, { "epoch": 11.343589743589744, "grad_norm": 0.369140625, "learning_rate": 2.1574477811246014e-06, "loss": 0.4756, "num_input_tokens_seen": 42705056, "step": 2765 }, { "epoch": 11.364102564102565, "grad_norm": 0.421875, "learning_rate": 2.1547229114423207e-06, "loss": 0.4985, "num_input_tokens_seen": 42777632, "step": 2770 }, { "epoch": 11.384615384615385, "grad_norm": 0.326171875, "learning_rate": 2.1519953700787963e-06, "loss": 0.4561, "num_input_tokens_seen": 42845888, "step": 2775 }, { "epoch": 11.405128205128205, "grad_norm": 0.36328125, "learning_rate": 2.149265168164113e-06, "loss": 0.5091, "num_input_tokens_seen": 42922976, "step": 2780 }, { "epoch": 11.425641025641026, "grad_norm": 0.353515625, "learning_rate": 2.146532316839211e-06, "loss": 0.4711, "num_input_tokens_seen": 42996000, "step": 2785 }, { "epoch": 11.446153846153846, "grad_norm": 0.38671875, "learning_rate": 2.1437968272558435e-06, "loss": 0.457, "num_input_tokens_seen": 43074688, "step": 2790 }, { "epoch": 11.466666666666667, "grad_norm": 0.31640625, "learning_rate": 2.1410587105765275e-06, "loss": 0.541, "num_input_tokens_seen": 43157280, "step": 2795 }, { "epoch": 11.487179487179487, "grad_norm": 0.361328125, "learning_rate": 2.138317977974501e-06, "loss": 0.4279, "num_input_tokens_seen": 43234016, "step": 2800 }, { "epoch": 11.507692307692308, "grad_norm": 0.306640625, "learning_rate": 2.135574640633678e-06, "loss": 0.5213, "num_input_tokens_seen": 43310816, "step": 2805 }, { "epoch": 11.528205128205128, "grad_norm": 0.275390625, "learning_rate": 2.132828709748598e-06, "loss": 0.4444, "num_input_tokens_seen": 43382976, "step": 2810 }, { "epoch": 11.548717948717949, "grad_norm": 0.279296875, "learning_rate": 2.130080196524388e-06, "loss": 0.4768, "num_input_tokens_seen": 43462944, "step": 2815 }, { "epoch": 11.569230769230769, "grad_norm": 0.283203125, "learning_rate": 2.1273291121767094e-06, "loss": 0.4376, "num_input_tokens_seen": 43535232, "step": 2820 }, { "epoch": 11.58974358974359, "grad_norm": 0.23828125, "learning_rate": 2.124575467931717e-06, "loss": 0.4342, "num_input_tokens_seen": 43618528, "step": 2825 }, { "epoch": 11.61025641025641, "grad_norm": 0.1982421875, "learning_rate": 2.1218192750260114e-06, "loss": 0.4596, "num_input_tokens_seen": 43691904, "step": 2830 }, { "epoch": 11.63076923076923, "grad_norm": 0.1904296875, "learning_rate": 2.119060544706592e-06, "loss": 0.4811, "num_input_tokens_seen": 43760480, "step": 2835 }, { "epoch": 11.65128205128205, "grad_norm": 0.1923828125, "learning_rate": 2.1162992882308147e-06, "loss": 0.4864, "num_input_tokens_seen": 43829984, "step": 2840 }, { "epoch": 11.671794871794871, "grad_norm": 0.1875, "learning_rate": 2.1135355168663417e-06, "loss": 0.4678, "num_input_tokens_seen": 43906816, "step": 2845 }, { "epoch": 11.692307692307692, "grad_norm": 0.1484375, "learning_rate": 2.1107692418910985e-06, "loss": 0.5001, "num_input_tokens_seen": 43984960, "step": 2850 }, { "epoch": 11.712820512820512, "grad_norm": 0.154296875, "learning_rate": 2.1080004745932274e-06, "loss": 0.4662, "num_input_tokens_seen": 44061440, "step": 2855 }, { "epoch": 11.733333333333333, "grad_norm": 0.1611328125, "learning_rate": 2.1052292262710392e-06, "loss": 0.526, "num_input_tokens_seen": 44147008, "step": 2860 }, { "epoch": 11.753846153846155, "grad_norm": 0.1337890625, "learning_rate": 2.102455508232971e-06, "loss": 0.4821, "num_input_tokens_seen": 44224224, "step": 2865 }, { "epoch": 11.774358974358975, "grad_norm": 0.146484375, "learning_rate": 2.099679331797536e-06, "loss": 0.4923, "num_input_tokens_seen": 44302816, "step": 2870 }, { "epoch": 11.794871794871796, "grad_norm": 0.1650390625, "learning_rate": 2.0969007082932803e-06, "loss": 0.521, "num_input_tokens_seen": 44376160, "step": 2875 }, { "epoch": 11.815384615384616, "grad_norm": 0.125, "learning_rate": 2.0941196490587354e-06, "loss": 0.4932, "num_input_tokens_seen": 44459200, "step": 2880 }, { "epoch": 11.835897435897436, "grad_norm": 0.140625, "learning_rate": 2.0913361654423723e-06, "loss": 0.4866, "num_input_tokens_seen": 44536128, "step": 2885 }, { "epoch": 11.856410256410257, "grad_norm": 0.1357421875, "learning_rate": 2.0885502688025538e-06, "loss": 0.4826, "num_input_tokens_seen": 44605088, "step": 2890 }, { "epoch": 11.876923076923077, "grad_norm": 0.2177734375, "learning_rate": 2.0857619705074912e-06, "loss": 0.4433, "num_input_tokens_seen": 44677984, "step": 2895 }, { "epoch": 11.897435897435898, "grad_norm": 0.166015625, "learning_rate": 2.082971281935195e-06, "loss": 0.4122, "num_input_tokens_seen": 44751200, "step": 2900 }, { "epoch": 11.917948717948718, "grad_norm": 0.28515625, "learning_rate": 2.0801782144734295e-06, "loss": 0.4266, "num_input_tokens_seen": 44824672, "step": 2905 }, { "epoch": 11.938461538461539, "grad_norm": 0.341796875, "learning_rate": 2.0773827795196667e-06, "loss": 0.4574, "num_input_tokens_seen": 44912768, "step": 2910 }, { "epoch": 11.95897435897436, "grad_norm": 0.2578125, "learning_rate": 2.074584988481039e-06, "loss": 0.5026, "num_input_tokens_seen": 44993632, "step": 2915 }, { "epoch": 11.97948717948718, "grad_norm": 0.79296875, "learning_rate": 2.0717848527742935e-06, "loss": 0.5444, "num_input_tokens_seen": 45070848, "step": 2920 }, { "epoch": 12.0, "grad_norm": 0.546875, "learning_rate": 2.0689823838257455e-06, "loss": 0.4509, "num_input_tokens_seen": 45150496, "step": 2925 }, { "epoch": 12.02051282051282, "grad_norm": 0.7421875, "learning_rate": 2.0661775930712297e-06, "loss": 0.4534, "num_input_tokens_seen": 45225440, "step": 2930 }, { "epoch": 12.04102564102564, "grad_norm": 0.3984375, "learning_rate": 2.0633704919560573e-06, "loss": 0.5264, "num_input_tokens_seen": 45300992, "step": 2935 }, { "epoch": 12.061538461538461, "grad_norm": 0.84765625, "learning_rate": 2.0605610919349658e-06, "loss": 0.4373, "num_input_tokens_seen": 45378944, "step": 2940 }, { "epoch": 12.082051282051282, "grad_norm": 0.53515625, "learning_rate": 2.0577494044720746e-06, "loss": 0.4779, "num_input_tokens_seen": 45453504, "step": 2945 }, { "epoch": 12.102564102564102, "grad_norm": 0.265625, "learning_rate": 2.0549354410408364e-06, "loss": 0.556, "num_input_tokens_seen": 45532992, "step": 2950 }, { "epoch": 12.123076923076923, "grad_norm": 0.2890625, "learning_rate": 2.052119213123992e-06, "loss": 0.5152, "num_input_tokens_seen": 45609120, "step": 2955 }, { "epoch": 12.143589743589743, "grad_norm": 0.318359375, "learning_rate": 2.049300732213522e-06, "loss": 0.4412, "num_input_tokens_seen": 45690624, "step": 2960 }, { "epoch": 12.164102564102564, "grad_norm": 0.46875, "learning_rate": 2.046480009810602e-06, "loss": 0.4553, "num_input_tokens_seen": 45763264, "step": 2965 }, { "epoch": 12.184615384615384, "grad_norm": 0.275390625, "learning_rate": 2.0436570574255523e-06, "loss": 0.4913, "num_input_tokens_seen": 45849472, "step": 2970 }, { "epoch": 12.205128205128204, "grad_norm": 0.41015625, "learning_rate": 2.0408318865777953e-06, "loss": 0.5487, "num_input_tokens_seen": 45927552, "step": 2975 }, { "epoch": 12.225641025641025, "grad_norm": 0.291015625, "learning_rate": 2.0380045087958036e-06, "loss": 0.465, "num_input_tokens_seen": 46002656, "step": 2980 }, { "epoch": 12.246153846153845, "grad_norm": 0.404296875, "learning_rate": 2.0351749356170574e-06, "loss": 0.4854, "num_input_tokens_seen": 46087904, "step": 2985 }, { "epoch": 12.266666666666667, "grad_norm": 0.37109375, "learning_rate": 2.032343178587995e-06, "loss": 0.4568, "num_input_tokens_seen": 46165408, "step": 2990 }, { "epoch": 12.287179487179488, "grad_norm": 0.3515625, "learning_rate": 2.0295092492639657e-06, "loss": 0.4926, "num_input_tokens_seen": 46237344, "step": 2995 }, { "epoch": 12.307692307692308, "grad_norm": 0.5, "learning_rate": 2.0266731592091834e-06, "loss": 0.5093, "num_input_tokens_seen": 46307456, "step": 3000 }, { "epoch": 12.328205128205129, "grad_norm": 0.33984375, "learning_rate": 2.0238349199966793e-06, "loss": 0.5077, "num_input_tokens_seen": 46390688, "step": 3005 }, { "epoch": 12.34871794871795, "grad_norm": 0.32421875, "learning_rate": 2.020994543208254e-06, "loss": 0.4541, "num_input_tokens_seen": 46467904, "step": 3010 }, { "epoch": 12.36923076923077, "grad_norm": 0.33203125, "learning_rate": 2.018152040434432e-06, "loss": 0.4975, "num_input_tokens_seen": 46545632, "step": 3015 }, { "epoch": 12.38974358974359, "grad_norm": 0.40234375, "learning_rate": 2.015307423274411e-06, "loss": 0.4988, "num_input_tokens_seen": 46624768, "step": 3020 }, { "epoch": 12.41025641025641, "grad_norm": 0.298828125, "learning_rate": 2.0124607033360193e-06, "loss": 0.4877, "num_input_tokens_seen": 46694528, "step": 3025 }, { "epoch": 12.430769230769231, "grad_norm": 0.3046875, "learning_rate": 2.0096118922356646e-06, "loss": 0.4621, "num_input_tokens_seen": 46767520, "step": 3030 }, { "epoch": 12.451282051282051, "grad_norm": 0.255859375, "learning_rate": 2.0067610015982868e-06, "loss": 0.4742, "num_input_tokens_seen": 46843616, "step": 3035 }, { "epoch": 12.471794871794872, "grad_norm": 0.27734375, "learning_rate": 2.0039080430573133e-06, "loss": 0.3993, "num_input_tokens_seen": 46915136, "step": 3040 }, { "epoch": 12.492307692307692, "grad_norm": 0.232421875, "learning_rate": 2.0010530282546093e-06, "loss": 0.464, "num_input_tokens_seen": 46992000, "step": 3045 }, { "epoch": 12.512820512820513, "grad_norm": 0.2060546875, "learning_rate": 1.9981959688404303e-06, "loss": 0.4744, "num_input_tokens_seen": 47075360, "step": 3050 }, { "epoch": 12.533333333333333, "grad_norm": 0.1953125, "learning_rate": 1.9953368764733763e-06, "loss": 0.4788, "num_input_tokens_seen": 47152704, "step": 3055 }, { "epoch": 12.553846153846154, "grad_norm": 0.1923828125, "learning_rate": 1.992475762820342e-06, "loss": 0.4704, "num_input_tokens_seen": 47229696, "step": 3060 }, { "epoch": 12.574358974358974, "grad_norm": 0.193359375, "learning_rate": 1.9896126395564695e-06, "loss": 0.4645, "num_input_tokens_seen": 47307360, "step": 3065 }, { "epoch": 12.594871794871795, "grad_norm": 0.1728515625, "learning_rate": 1.986747518365104e-06, "loss": 0.5485, "num_input_tokens_seen": 47393824, "step": 3070 }, { "epoch": 12.615384615384615, "grad_norm": 0.154296875, "learning_rate": 1.9838804109377405e-06, "loss": 0.511, "num_input_tokens_seen": 47475104, "step": 3075 }, { "epoch": 12.635897435897435, "grad_norm": 0.1630859375, "learning_rate": 1.9810113289739818e-06, "loss": 0.4624, "num_input_tokens_seen": 47548704, "step": 3080 }, { "epoch": 12.656410256410256, "grad_norm": 0.1435546875, "learning_rate": 1.9781402841814855e-06, "loss": 0.5197, "num_input_tokens_seen": 47638880, "step": 3085 }, { "epoch": 12.676923076923076, "grad_norm": 0.134765625, "learning_rate": 1.9752672882759204e-06, "loss": 0.4277, "num_input_tokens_seen": 47711456, "step": 3090 }, { "epoch": 12.697435897435897, "grad_norm": 0.138671875, "learning_rate": 1.972392352980917e-06, "loss": 0.5166, "num_input_tokens_seen": 47787776, "step": 3095 }, { "epoch": 12.717948717948717, "grad_norm": 0.1357421875, "learning_rate": 1.969515490028019e-06, "loss": 0.4201, "num_input_tokens_seen": 47860480, "step": 3100 }, { "epoch": 12.73846153846154, "grad_norm": 0.1064453125, "learning_rate": 1.966636711156636e-06, "loss": 0.4425, "num_input_tokens_seen": 47936704, "step": 3105 }, { "epoch": 12.75897435897436, "grad_norm": 0.154296875, "learning_rate": 1.9637560281139982e-06, "loss": 0.5056, "num_input_tokens_seen": 48021792, "step": 3110 }, { "epoch": 12.77948717948718, "grad_norm": 0.1953125, "learning_rate": 1.960873452655102e-06, "loss": 0.4929, "num_input_tokens_seen": 48101984, "step": 3115 }, { "epoch": 12.8, "grad_norm": 0.2392578125, "learning_rate": 1.95798899654267e-06, "loss": 0.4868, "num_input_tokens_seen": 48186688, "step": 3120 }, { "epoch": 12.820512820512821, "grad_norm": 0.134765625, "learning_rate": 1.9551026715470954e-06, "loss": 0.4668, "num_input_tokens_seen": 48259872, "step": 3125 }, { "epoch": 12.841025641025642, "grad_norm": 0.6953125, "learning_rate": 1.952214489446401e-06, "loss": 0.53, "num_input_tokens_seen": 48347232, "step": 3130 }, { "epoch": 12.861538461538462, "grad_norm": 0.296875, "learning_rate": 1.949324462026185e-06, "loss": 0.4388, "num_input_tokens_seen": 48422624, "step": 3135 }, { "epoch": 12.882051282051282, "grad_norm": 0.28125, "learning_rate": 1.9464326010795776e-06, "loss": 0.4246, "num_input_tokens_seen": 48492288, "step": 3140 }, { "epoch": 12.902564102564103, "grad_norm": 0.72265625, "learning_rate": 1.9435389184071895e-06, "loss": 0.5186, "num_input_tokens_seen": 48568224, "step": 3145 }, { "epoch": 12.923076923076923, "grad_norm": 0.7734375, "learning_rate": 1.9406434258170666e-06, "loss": 0.4351, "num_input_tokens_seen": 48637280, "step": 3150 }, { "epoch": 12.943589743589744, "grad_norm": 2.03125, "learning_rate": 1.9377461351246395e-06, "loss": 0.5281, "num_input_tokens_seen": 48717088, "step": 3155 }, { "epoch": 12.964102564102564, "grad_norm": 0.236328125, "learning_rate": 1.9348470581526763e-06, "loss": 0.4308, "num_input_tokens_seen": 48787584, "step": 3160 }, { "epoch": 12.984615384615385, "grad_norm": 0.52734375, "learning_rate": 1.9319462067312344e-06, "loss": 0.4133, "num_input_tokens_seen": 48864640, "step": 3165 }, { "epoch": 13.005128205128205, "grad_norm": 0.55859375, "learning_rate": 1.929043592697612e-06, "loss": 0.4802, "num_input_tokens_seen": 48938176, "step": 3170 }, { "epoch": 13.025641025641026, "grad_norm": 0.306640625, "learning_rate": 1.926139227896299e-06, "loss": 0.4504, "num_input_tokens_seen": 49015040, "step": 3175 }, { "epoch": 13.046153846153846, "grad_norm": 0.4140625, "learning_rate": 1.923233124178932e-06, "loss": 0.494, "num_input_tokens_seen": 49091680, "step": 3180 }, { "epoch": 13.066666666666666, "grad_norm": 0.3671875, "learning_rate": 1.9203252934042403e-06, "loss": 0.4628, "num_input_tokens_seen": 49165600, "step": 3185 }, { "epoch": 13.087179487179487, "grad_norm": 0.318359375, "learning_rate": 1.9174157474380034e-06, "loss": 0.4431, "num_input_tokens_seen": 49247136, "step": 3190 }, { "epoch": 13.107692307692307, "grad_norm": 0.27734375, "learning_rate": 1.914504498152998e-06, "loss": 0.4479, "num_input_tokens_seen": 49319648, "step": 3195 }, { "epoch": 13.128205128205128, "grad_norm": 0.40625, "learning_rate": 1.9115915574289525e-06, "loss": 0.4278, "num_input_tokens_seen": 49393216, "step": 3200 }, { "epoch": 13.148717948717948, "grad_norm": 0.302734375, "learning_rate": 1.9086769371524966e-06, "loss": 0.5124, "num_input_tokens_seen": 49480864, "step": 3205 }, { "epoch": 13.169230769230769, "grad_norm": 0.3515625, "learning_rate": 1.9057606492171144e-06, "loss": 0.4438, "num_input_tokens_seen": 49559904, "step": 3210 }, { "epoch": 13.189743589743589, "grad_norm": 0.412109375, "learning_rate": 1.9028427055230948e-06, "loss": 0.4312, "num_input_tokens_seen": 49630240, "step": 3215 }, { "epoch": 13.21025641025641, "grad_norm": 0.443359375, "learning_rate": 1.8999231179774833e-06, "loss": 0.5431, "num_input_tokens_seen": 49709888, "step": 3220 }, { "epoch": 13.23076923076923, "grad_norm": 0.447265625, "learning_rate": 1.897001898494033e-06, "loss": 0.4567, "num_input_tokens_seen": 49778848, "step": 3225 }, { "epoch": 13.25128205128205, "grad_norm": 0.396484375, "learning_rate": 1.8940790589931568e-06, "loss": 0.4227, "num_input_tokens_seen": 49849024, "step": 3230 }, { "epoch": 13.271794871794873, "grad_norm": 0.306640625, "learning_rate": 1.8911546114018775e-06, "loss": 0.4738, "num_input_tokens_seen": 49917952, "step": 3235 }, { "epoch": 13.292307692307693, "grad_norm": 0.3203125, "learning_rate": 1.888228567653781e-06, "loss": 0.4661, "num_input_tokens_seen": 50003392, "step": 3240 }, { "epoch": 13.312820512820513, "grad_norm": 0.330078125, "learning_rate": 1.8853009396889665e-06, "loss": 0.5039, "num_input_tokens_seen": 50085280, "step": 3245 }, { "epoch": 13.333333333333334, "grad_norm": 0.33984375, "learning_rate": 1.8823717394539966e-06, "loss": 0.4442, "num_input_tokens_seen": 50166304, "step": 3250 }, { "epoch": 13.353846153846154, "grad_norm": 0.271484375, "learning_rate": 1.8794409789018507e-06, "loss": 0.5048, "num_input_tokens_seen": 50247200, "step": 3255 }, { "epoch": 13.374358974358975, "grad_norm": 0.275390625, "learning_rate": 1.8765086699918747e-06, "loss": 0.4615, "num_input_tokens_seen": 50321120, "step": 3260 }, { "epoch": 13.394871794871795, "grad_norm": 0.275390625, "learning_rate": 1.8735748246897337e-06, "loss": 0.4838, "num_input_tokens_seen": 50403680, "step": 3265 }, { "epoch": 13.415384615384616, "grad_norm": 0.203125, "learning_rate": 1.8706394549673615e-06, "loss": 0.4454, "num_input_tokens_seen": 50481504, "step": 3270 }, { "epoch": 13.435897435897436, "grad_norm": 0.220703125, "learning_rate": 1.8677025728029122e-06, "loss": 0.4806, "num_input_tokens_seen": 50556448, "step": 3275 }, { "epoch": 13.456410256410257, "grad_norm": 0.2060546875, "learning_rate": 1.8647641901807126e-06, "loss": 0.4601, "num_input_tokens_seen": 50630304, "step": 3280 }, { "epoch": 13.476923076923077, "grad_norm": 0.2236328125, "learning_rate": 1.861824319091212e-06, "loss": 0.4513, "num_input_tokens_seen": 50704704, "step": 3285 }, { "epoch": 13.497435897435897, "grad_norm": 0.1572265625, "learning_rate": 1.8588829715309324e-06, "loss": 0.4704, "num_input_tokens_seen": 50782464, "step": 3290 }, { "epoch": 13.517948717948718, "grad_norm": 0.166015625, "learning_rate": 1.855940159502423e-06, "loss": 0.4959, "num_input_tokens_seen": 50861696, "step": 3295 }, { "epoch": 13.538461538461538, "grad_norm": 0.1591796875, "learning_rate": 1.8529958950142066e-06, "loss": 0.5215, "num_input_tokens_seen": 50949568, "step": 3300 }, { "epoch": 13.558974358974359, "grad_norm": 0.14453125, "learning_rate": 1.8500501900807345e-06, "loss": 0.4692, "num_input_tokens_seen": 51018432, "step": 3305 }, { "epoch": 13.57948717948718, "grad_norm": 0.1435546875, "learning_rate": 1.847103056722335e-06, "loss": 0.5015, "num_input_tokens_seen": 51100352, "step": 3310 }, { "epoch": 13.6, "grad_norm": 0.1337890625, "learning_rate": 1.8441545069651665e-06, "loss": 0.4779, "num_input_tokens_seen": 51178144, "step": 3315 }, { "epoch": 13.62051282051282, "grad_norm": 0.146484375, "learning_rate": 1.8412045528411652e-06, "loss": 0.4971, "num_input_tokens_seen": 51262432, "step": 3320 }, { "epoch": 13.64102564102564, "grad_norm": 0.11279296875, "learning_rate": 1.8382532063880005e-06, "loss": 0.5063, "num_input_tokens_seen": 51342624, "step": 3325 }, { "epoch": 13.661538461538461, "grad_norm": 0.1328125, "learning_rate": 1.8353004796490212e-06, "loss": 0.4273, "num_input_tokens_seen": 51413920, "step": 3330 }, { "epoch": 13.682051282051281, "grad_norm": 0.14453125, "learning_rate": 1.8323463846732099e-06, "loss": 0.5491, "num_input_tokens_seen": 51501184, "step": 3335 }, { "epoch": 13.702564102564102, "grad_norm": 0.2109375, "learning_rate": 1.8293909335151316e-06, "loss": 0.4199, "num_input_tokens_seen": 51570944, "step": 3340 }, { "epoch": 13.723076923076922, "grad_norm": 0.236328125, "learning_rate": 1.826434138234886e-06, "loss": 0.4468, "num_input_tokens_seen": 51650784, "step": 3345 }, { "epoch": 13.743589743589745, "grad_norm": 0.150390625, "learning_rate": 1.8234760108980572e-06, "loss": 0.5225, "num_input_tokens_seen": 51731008, "step": 3350 }, { "epoch": 13.764102564102565, "grad_norm": 0.33203125, "learning_rate": 1.820516563575665e-06, "loss": 0.4634, "num_input_tokens_seen": 51809056, "step": 3355 }, { "epoch": 13.784615384615385, "grad_norm": 0.298828125, "learning_rate": 1.8175558083441164e-06, "loss": 0.5321, "num_input_tokens_seen": 51886688, "step": 3360 }, { "epoch": 13.805128205128206, "grad_norm": 0.2119140625, "learning_rate": 1.814593757285154e-06, "loss": 0.4633, "num_input_tokens_seen": 51959488, "step": 3365 }, { "epoch": 13.825641025641026, "grad_norm": 0.59765625, "learning_rate": 1.8116304224858092e-06, "loss": 0.4812, "num_input_tokens_seen": 52035424, "step": 3370 }, { "epoch": 13.846153846153847, "grad_norm": 0.859375, "learning_rate": 1.8086658160383524e-06, "loss": 0.506, "num_input_tokens_seen": 52112576, "step": 3375 }, { "epoch": 13.866666666666667, "grad_norm": 0.83984375, "learning_rate": 1.8056999500402414e-06, "loss": 0.5091, "num_input_tokens_seen": 52200480, "step": 3380 }, { "epoch": 13.887179487179488, "grad_norm": 0.296875, "learning_rate": 1.8027328365940755e-06, "loss": 0.4861, "num_input_tokens_seen": 52274368, "step": 3385 }, { "epoch": 13.907692307692308, "grad_norm": 0.462890625, "learning_rate": 1.799764487807543e-06, "loss": 0.5051, "num_input_tokens_seen": 52350656, "step": 3390 }, { "epoch": 13.928205128205128, "grad_norm": 0.279296875, "learning_rate": 1.7967949157933742e-06, "loss": 0.4608, "num_input_tokens_seen": 52427328, "step": 3395 }, { "epoch": 13.948717948717949, "grad_norm": 0.3203125, "learning_rate": 1.7938241326692907e-06, "loss": 0.4719, "num_input_tokens_seen": 52503840, "step": 3400 }, { "epoch": 13.96923076923077, "grad_norm": 0.291015625, "learning_rate": 1.7908521505579554e-06, "loss": 0.5062, "num_input_tokens_seen": 52577056, "step": 3405 }, { "epoch": 13.98974358974359, "grad_norm": 0.310546875, "learning_rate": 1.7878789815869247e-06, "loss": 0.5096, "num_input_tokens_seen": 52656000, "step": 3410 }, { "epoch": 14.01025641025641, "grad_norm": 0.291015625, "learning_rate": 1.7849046378885977e-06, "loss": 0.4476, "num_input_tokens_seen": 52725312, "step": 3415 }, { "epoch": 14.03076923076923, "grad_norm": 0.361328125, "learning_rate": 1.7819291316001679e-06, "loss": 0.436, "num_input_tokens_seen": 52804736, "step": 3420 }, { "epoch": 14.051282051282051, "grad_norm": 0.353515625, "learning_rate": 1.7789524748635717e-06, "loss": 0.4761, "num_input_tokens_seen": 52885056, "step": 3425 }, { "epoch": 14.071794871794872, "grad_norm": 0.36328125, "learning_rate": 1.775974679825441e-06, "loss": 0.5039, "num_input_tokens_seen": 52963136, "step": 3430 }, { "epoch": 14.092307692307692, "grad_norm": 0.470703125, "learning_rate": 1.7729957586370525e-06, "loss": 0.5039, "num_input_tokens_seen": 53049856, "step": 3435 }, { "epoch": 14.112820512820512, "grad_norm": 0.45703125, "learning_rate": 1.7700157234542773e-06, "loss": 0.5397, "num_input_tokens_seen": 53127616, "step": 3440 }, { "epoch": 14.133333333333333, "grad_norm": 0.34765625, "learning_rate": 1.7670345864375339e-06, "loss": 0.4747, "num_input_tokens_seen": 53201600, "step": 3445 }, { "epoch": 14.153846153846153, "grad_norm": 0.37890625, "learning_rate": 1.7640523597517357e-06, "loss": 0.4582, "num_input_tokens_seen": 53275520, "step": 3450 }, { "epoch": 14.174358974358974, "grad_norm": 0.328125, "learning_rate": 1.7610690555662435e-06, "loss": 0.4747, "num_input_tokens_seen": 53346720, "step": 3455 }, { "epoch": 14.194871794871794, "grad_norm": 0.337890625, "learning_rate": 1.7580846860548146e-06, "loss": 0.4758, "num_input_tokens_seen": 53420864, "step": 3460 }, { "epoch": 14.215384615384615, "grad_norm": 0.380859375, "learning_rate": 1.7550992633955535e-06, "loss": 0.4305, "num_input_tokens_seen": 53490976, "step": 3465 }, { "epoch": 14.235897435897435, "grad_norm": 0.357421875, "learning_rate": 1.7521127997708621e-06, "loss": 0.4533, "num_input_tokens_seen": 53570816, "step": 3470 }, { "epoch": 14.256410256410255, "grad_norm": 0.30078125, "learning_rate": 1.7491253073673903e-06, "loss": 0.5019, "num_input_tokens_seen": 53647072, "step": 3475 }, { "epoch": 14.276923076923078, "grad_norm": 0.28515625, "learning_rate": 1.7461367983759862e-06, "loss": 0.4575, "num_input_tokens_seen": 53727648, "step": 3480 }, { "epoch": 14.297435897435898, "grad_norm": 0.28515625, "learning_rate": 1.7431472849916455e-06, "loss": 0.4728, "num_input_tokens_seen": 53800544, "step": 3485 }, { "epoch": 14.317948717948719, "grad_norm": 0.298828125, "learning_rate": 1.7401567794134636e-06, "loss": 0.4677, "num_input_tokens_seen": 53875552, "step": 3490 }, { "epoch": 14.338461538461539, "grad_norm": 0.216796875, "learning_rate": 1.737165293844583e-06, "loss": 0.456, "num_input_tokens_seen": 53954208, "step": 3495 }, { "epoch": 14.35897435897436, "grad_norm": 0.205078125, "learning_rate": 1.7341728404921471e-06, "loss": 0.4798, "num_input_tokens_seen": 54033248, "step": 3500 }, { "epoch": 14.37948717948718, "grad_norm": 0.208984375, "learning_rate": 1.7311794315672477e-06, "loss": 0.4166, "num_input_tokens_seen": 54109440, "step": 3505 }, { "epoch": 14.4, "grad_norm": 0.1953125, "learning_rate": 1.7281850792848752e-06, "loss": 0.4663, "num_input_tokens_seen": 54186368, "step": 3510 }, { "epoch": 14.42051282051282, "grad_norm": 0.1591796875, "learning_rate": 1.7251897958638704e-06, "loss": 0.4053, "num_input_tokens_seen": 54267648, "step": 3515 }, { "epoch": 14.441025641025641, "grad_norm": 0.193359375, "learning_rate": 1.7221935935268735e-06, "loss": 0.5342, "num_input_tokens_seen": 54342304, "step": 3520 }, { "epoch": 14.461538461538462, "grad_norm": 0.1455078125, "learning_rate": 1.719196484500274e-06, "loss": 0.4992, "num_input_tokens_seen": 54423808, "step": 3525 }, { "epoch": 14.482051282051282, "grad_norm": 0.1494140625, "learning_rate": 1.7161984810141625e-06, "loss": 0.5067, "num_input_tokens_seen": 54501760, "step": 3530 }, { "epoch": 14.502564102564103, "grad_norm": 0.126953125, "learning_rate": 1.7131995953022776e-06, "loss": 0.4286, "num_input_tokens_seen": 54576512, "step": 3535 }, { "epoch": 14.523076923076923, "grad_norm": 0.1611328125, "learning_rate": 1.7101998396019593e-06, "loss": 0.436, "num_input_tokens_seen": 54649696, "step": 3540 }, { "epoch": 14.543589743589743, "grad_norm": 0.138671875, "learning_rate": 1.7071992261540983e-06, "loss": 0.4889, "num_input_tokens_seen": 54728864, "step": 3545 }, { "epoch": 14.564102564102564, "grad_norm": 0.1328125, "learning_rate": 1.7041977672030842e-06, "loss": 0.4585, "num_input_tokens_seen": 54797856, "step": 3550 }, { "epoch": 14.584615384615384, "grad_norm": 0.1064453125, "learning_rate": 1.7011954749967564e-06, "loss": 0.4646, "num_input_tokens_seen": 54875168, "step": 3555 }, { "epoch": 14.605128205128205, "grad_norm": 0.1474609375, "learning_rate": 1.6981923617863566e-06, "loss": 0.4894, "num_input_tokens_seen": 54949312, "step": 3560 }, { "epoch": 14.625641025641025, "grad_norm": 0.2255859375, "learning_rate": 1.6951884398264742e-06, "loss": 0.5159, "num_input_tokens_seen": 55024768, "step": 3565 }, { "epoch": 14.646153846153846, "grad_norm": 0.150390625, "learning_rate": 1.692183721375001e-06, "loss": 0.4221, "num_input_tokens_seen": 55100128, "step": 3570 }, { "epoch": 14.666666666666666, "grad_norm": 0.1435546875, "learning_rate": 1.6891782186930767e-06, "loss": 0.5417, "num_input_tokens_seen": 55182080, "step": 3575 }, { "epoch": 14.687179487179487, "grad_norm": 0.28515625, "learning_rate": 1.6861719440450437e-06, "loss": 0.5072, "num_input_tokens_seen": 55274048, "step": 3580 }, { "epoch": 14.707692307692307, "grad_norm": 0.267578125, "learning_rate": 1.6831649096983923e-06, "loss": 0.5008, "num_input_tokens_seen": 55347648, "step": 3585 }, { "epoch": 14.728205128205127, "grad_norm": 0.33203125, "learning_rate": 1.680157127923715e-06, "loss": 0.4411, "num_input_tokens_seen": 55418368, "step": 3590 }, { "epoch": 14.74871794871795, "grad_norm": 0.74609375, "learning_rate": 1.677148610994652e-06, "loss": 0.4654, "num_input_tokens_seen": 55490176, "step": 3595 }, { "epoch": 14.76923076923077, "grad_norm": 0.70703125, "learning_rate": 1.6741393711878454e-06, "loss": 0.4576, "num_input_tokens_seen": 55570784, "step": 3600 }, { "epoch": 14.78974358974359, "grad_norm": 0.2353515625, "learning_rate": 1.6711294207828852e-06, "loss": 0.4488, "num_input_tokens_seen": 55643936, "step": 3605 }, { "epoch": 14.810256410256411, "grad_norm": 0.349609375, "learning_rate": 1.6681187720622627e-06, "loss": 0.5175, "num_input_tokens_seen": 55727936, "step": 3610 }, { "epoch": 14.830769230769231, "grad_norm": 0.484375, "learning_rate": 1.6651074373113176e-06, "loss": 0.4271, "num_input_tokens_seen": 55804992, "step": 3615 }, { "epoch": 14.851282051282052, "grad_norm": 0.3203125, "learning_rate": 1.66209542881819e-06, "loss": 0.4631, "num_input_tokens_seen": 55878976, "step": 3620 }, { "epoch": 14.871794871794872, "grad_norm": 0.26171875, "learning_rate": 1.6590827588737685e-06, "loss": 0.5453, "num_input_tokens_seen": 55969376, "step": 3625 }, { "epoch": 14.892307692307693, "grad_norm": 0.34765625, "learning_rate": 1.6560694397716412e-06, "loss": 0.5086, "num_input_tokens_seen": 56041024, "step": 3630 }, { "epoch": 14.912820512820513, "grad_norm": 0.28515625, "learning_rate": 1.6530554838080458e-06, "loss": 0.4602, "num_input_tokens_seen": 56117152, "step": 3635 }, { "epoch": 14.933333333333334, "grad_norm": 0.251953125, "learning_rate": 1.6500409032818175e-06, "loss": 0.5046, "num_input_tokens_seen": 56191264, "step": 3640 }, { "epoch": 14.953846153846154, "grad_norm": 0.271484375, "learning_rate": 1.6470257104943413e-06, "loss": 0.53, "num_input_tokens_seen": 56279264, "step": 3645 }, { "epoch": 14.974358974358974, "grad_norm": 0.341796875, "learning_rate": 1.6440099177494991e-06, "loss": 0.4639, "num_input_tokens_seen": 56350784, "step": 3650 }, { "epoch": 14.994871794871795, "grad_norm": 0.32421875, "learning_rate": 1.6409935373536227e-06, "loss": 0.5015, "num_input_tokens_seen": 56430432, "step": 3655 }, { "epoch": 15.015384615384615, "grad_norm": 0.33203125, "learning_rate": 1.6379765816154413e-06, "loss": 0.4503, "num_input_tokens_seen": 56509056, "step": 3660 }, { "epoch": 15.035897435897436, "grad_norm": 0.322265625, "learning_rate": 1.634959062846031e-06, "loss": 0.4688, "num_input_tokens_seen": 56578016, "step": 3665 }, { "epoch": 15.056410256410256, "grad_norm": 0.34375, "learning_rate": 1.631940993358767e-06, "loss": 0.4761, "num_input_tokens_seen": 56653536, "step": 3670 }, { "epoch": 15.076923076923077, "grad_norm": 0.328125, "learning_rate": 1.6289223854692708e-06, "loss": 0.4433, "num_input_tokens_seen": 56730560, "step": 3675 }, { "epoch": 15.097435897435897, "grad_norm": 0.3046875, "learning_rate": 1.6259032514953601e-06, "loss": 0.4315, "num_input_tokens_seen": 56813312, "step": 3680 }, { "epoch": 15.117948717948718, "grad_norm": 0.283203125, "learning_rate": 1.6228836037570015e-06, "loss": 0.4885, "num_input_tokens_seen": 56891264, "step": 3685 }, { "epoch": 15.138461538461538, "grad_norm": 0.400390625, "learning_rate": 1.619863454576256e-06, "loss": 0.4703, "num_input_tokens_seen": 56964864, "step": 3690 }, { "epoch": 15.158974358974358, "grad_norm": 0.322265625, "learning_rate": 1.6168428162772322e-06, "loss": 0.4518, "num_input_tokens_seen": 57050144, "step": 3695 }, { "epoch": 15.179487179487179, "grad_norm": 0.328125, "learning_rate": 1.6138217011860336e-06, "loss": 0.4858, "num_input_tokens_seen": 57125632, "step": 3700 }, { "epoch": 15.2, "grad_norm": 0.296875, "learning_rate": 1.6108001216307107e-06, "loss": 0.459, "num_input_tokens_seen": 57204128, "step": 3705 }, { "epoch": 15.22051282051282, "grad_norm": 0.33984375, "learning_rate": 1.6077780899412068e-06, "loss": 0.4639, "num_input_tokens_seen": 57281536, "step": 3710 }, { "epoch": 15.24102564102564, "grad_norm": 0.27734375, "learning_rate": 1.6047556184493133e-06, "loss": 0.4496, "num_input_tokens_seen": 57354816, "step": 3715 }, { "epoch": 15.261538461538462, "grad_norm": 0.22265625, "learning_rate": 1.601732719488614e-06, "loss": 0.5043, "num_input_tokens_seen": 57435328, "step": 3720 }, { "epoch": 15.282051282051283, "grad_norm": 0.2109375, "learning_rate": 1.5987094053944384e-06, "loss": 0.4357, "num_input_tokens_seen": 57507424, "step": 3725 }, { "epoch": 15.302564102564103, "grad_norm": 0.228515625, "learning_rate": 1.5956856885038086e-06, "loss": 0.5211, "num_input_tokens_seen": 57586880, "step": 3730 }, { "epoch": 15.323076923076924, "grad_norm": 0.1982421875, "learning_rate": 1.592661581155392e-06, "loss": 0.5036, "num_input_tokens_seen": 57669600, "step": 3735 }, { "epoch": 15.343589743589744, "grad_norm": 0.1708984375, "learning_rate": 1.5896370956894477e-06, "loss": 0.4751, "num_input_tokens_seen": 57744608, "step": 3740 }, { "epoch": 15.364102564102565, "grad_norm": 0.1533203125, "learning_rate": 1.5866122444477794e-06, "loss": 0.4764, "num_input_tokens_seen": 57818496, "step": 3745 }, { "epoch": 15.384615384615385, "grad_norm": 0.1376953125, "learning_rate": 1.5835870397736817e-06, "loss": 0.4551, "num_input_tokens_seen": 57898432, "step": 3750 }, { "epoch": 15.405128205128205, "grad_norm": 0.166015625, "learning_rate": 1.5805614940118928e-06, "loss": 0.5198, "num_input_tokens_seen": 57974624, "step": 3755 }, { "epoch": 15.425641025641026, "grad_norm": 0.1337890625, "learning_rate": 1.577535619508542e-06, "loss": 0.5164, "num_input_tokens_seen": 58052576, "step": 3760 }, { "epoch": 15.446153846153846, "grad_norm": 0.130859375, "learning_rate": 1.5745094286111004e-06, "loss": 0.4699, "num_input_tokens_seen": 58126592, "step": 3765 }, { "epoch": 15.466666666666667, "grad_norm": 0.125, "learning_rate": 1.5714829336683297e-06, "loss": 0.4804, "num_input_tokens_seen": 58214784, "step": 3770 }, { "epoch": 15.487179487179487, "grad_norm": 0.1259765625, "learning_rate": 1.5684561470302337e-06, "loss": 0.4731, "num_input_tokens_seen": 58284480, "step": 3775 }, { "epoch": 15.507692307692308, "grad_norm": 0.1279296875, "learning_rate": 1.5654290810480041e-06, "loss": 0.5115, "num_input_tokens_seen": 58359488, "step": 3780 }, { "epoch": 15.528205128205128, "grad_norm": 0.1201171875, "learning_rate": 1.562401748073975e-06, "loss": 0.4713, "num_input_tokens_seen": 58433632, "step": 3785 }, { "epoch": 15.548717948717949, "grad_norm": 0.275390625, "learning_rate": 1.5593741604615679e-06, "loss": 0.5218, "num_input_tokens_seen": 58526048, "step": 3790 }, { "epoch": 15.569230769230769, "grad_norm": 0.15234375, "learning_rate": 1.5563463305652454e-06, "loss": 0.4432, "num_input_tokens_seen": 58593440, "step": 3795 }, { "epoch": 15.58974358974359, "grad_norm": 0.1552734375, "learning_rate": 1.5533182707404563e-06, "loss": 0.4793, "num_input_tokens_seen": 58664832, "step": 3800 }, { "epoch": 15.61025641025641, "grad_norm": 0.29296875, "learning_rate": 1.550289993343591e-06, "loss": 0.4287, "num_input_tokens_seen": 58734976, "step": 3805 }, { "epoch": 15.63076923076923, "grad_norm": 0.37109375, "learning_rate": 1.547261510731924e-06, "loss": 0.5701, "num_input_tokens_seen": 58826304, "step": 3810 }, { "epoch": 15.65128205128205, "grad_norm": 0.515625, "learning_rate": 1.5442328352635706e-06, "loss": 0.4858, "num_input_tokens_seen": 58897216, "step": 3815 }, { "epoch": 15.671794871794871, "grad_norm": 0.63671875, "learning_rate": 1.54120397929743e-06, "loss": 0.4541, "num_input_tokens_seen": 58970880, "step": 3820 }, { "epoch": 15.692307692307692, "grad_norm": 0.91015625, "learning_rate": 1.5381749551931405e-06, "loss": 0.5178, "num_input_tokens_seen": 59049408, "step": 3825 }, { "epoch": 15.712820512820512, "grad_norm": 0.298828125, "learning_rate": 1.5351457753110244e-06, "loss": 0.4799, "num_input_tokens_seen": 59122336, "step": 3830 }, { "epoch": 15.733333333333333, "grad_norm": 0.2333984375, "learning_rate": 1.5321164520120421e-06, "loss": 0.4284, "num_input_tokens_seen": 59197600, "step": 3835 }, { "epoch": 15.753846153846155, "grad_norm": 0.640625, "learning_rate": 1.5290869976577366e-06, "loss": 0.4343, "num_input_tokens_seen": 59268224, "step": 3840 }, { "epoch": 15.774358974358975, "grad_norm": 0.279296875, "learning_rate": 1.5260574246101875e-06, "loss": 0.4371, "num_input_tokens_seen": 59341408, "step": 3845 }, { "epoch": 15.794871794871796, "grad_norm": 0.27734375, "learning_rate": 1.5230277452319585e-06, "loss": 0.5017, "num_input_tokens_seen": 59423648, "step": 3850 }, { "epoch": 15.815384615384616, "grad_norm": 0.35546875, "learning_rate": 1.519997971886046e-06, "loss": 0.4473, "num_input_tokens_seen": 59495872, "step": 3855 }, { "epoch": 15.835897435897436, "grad_norm": 0.31640625, "learning_rate": 1.5169681169358314e-06, "loss": 0.4693, "num_input_tokens_seen": 59574880, "step": 3860 }, { "epoch": 15.856410256410257, "grad_norm": 0.3046875, "learning_rate": 1.513938192745028e-06, "loss": 0.4955, "num_input_tokens_seen": 59657472, "step": 3865 }, { "epoch": 15.876923076923077, "grad_norm": 0.337890625, "learning_rate": 1.5109082116776328e-06, "loss": 0.4219, "num_input_tokens_seen": 59733888, "step": 3870 }, { "epoch": 15.897435897435898, "grad_norm": 0.423828125, "learning_rate": 1.5078781860978732e-06, "loss": 0.4462, "num_input_tokens_seen": 59811392, "step": 3875 }, { "epoch": 15.917948717948718, "grad_norm": 0.359375, "learning_rate": 1.5048481283701594e-06, "loss": 0.4938, "num_input_tokens_seen": 59886784, "step": 3880 }, { "epoch": 15.938461538461539, "grad_norm": 0.390625, "learning_rate": 1.501818050859033e-06, "loss": 0.4465, "num_input_tokens_seen": 59959072, "step": 3885 }, { "epoch": 15.95897435897436, "grad_norm": 0.408203125, "learning_rate": 1.498787965929116e-06, "loss": 0.462, "num_input_tokens_seen": 60032832, "step": 3890 }, { "epoch": 15.97948717948718, "grad_norm": 0.28515625, "learning_rate": 1.49575788594506e-06, "loss": 0.5279, "num_input_tokens_seen": 60119456, "step": 3895 }, { "epoch": 16.0, "grad_norm": 0.30859375, "learning_rate": 1.4927278232714975e-06, "loss": 0.4908, "num_input_tokens_seen": 60204448, "step": 3900 }, { "epoch": 16.02051282051282, "grad_norm": 0.37109375, "learning_rate": 1.4896977902729893e-06, "loss": 0.4607, "num_input_tokens_seen": 60279104, "step": 3905 }, { "epoch": 16.04102564102564, "grad_norm": 0.30859375, "learning_rate": 1.4866677993139761e-06, "loss": 0.4405, "num_input_tokens_seen": 60349504, "step": 3910 }, { "epoch": 16.06153846153846, "grad_norm": 0.34375, "learning_rate": 1.4836378627587266e-06, "loss": 0.4744, "num_input_tokens_seen": 60420640, "step": 3915 }, { "epoch": 16.08205128205128, "grad_norm": 0.32421875, "learning_rate": 1.4806079929712874e-06, "loss": 0.4868, "num_input_tokens_seen": 60497728, "step": 3920 }, { "epoch": 16.102564102564102, "grad_norm": 0.341796875, "learning_rate": 1.477578202315433e-06, "loss": 0.5095, "num_input_tokens_seen": 60579008, "step": 3925 }, { "epoch": 16.123076923076923, "grad_norm": 0.271484375, "learning_rate": 1.4745485031546143e-06, "loss": 0.5101, "num_input_tokens_seen": 60657088, "step": 3930 }, { "epoch": 16.143589743589743, "grad_norm": 0.28515625, "learning_rate": 1.4715189078519094e-06, "loss": 0.4938, "num_input_tokens_seen": 60734560, "step": 3935 }, { "epoch": 16.164102564102564, "grad_norm": 0.265625, "learning_rate": 1.468489428769973e-06, "loss": 0.4785, "num_input_tokens_seen": 60812960, "step": 3940 }, { "epoch": 16.184615384615384, "grad_norm": 0.2138671875, "learning_rate": 1.4654600782709843e-06, "loss": 0.4953, "num_input_tokens_seen": 60892768, "step": 3945 }, { "epoch": 16.205128205128204, "grad_norm": 0.208984375, "learning_rate": 1.4624308687165985e-06, "loss": 0.4472, "num_input_tokens_seen": 60964096, "step": 3950 }, { "epoch": 16.225641025641025, "grad_norm": 0.1943359375, "learning_rate": 1.4594018124678965e-06, "loss": 0.4191, "num_input_tokens_seen": 61034304, "step": 3955 }, { "epoch": 16.246153846153845, "grad_norm": 0.181640625, "learning_rate": 1.4563729218853329e-06, "loss": 0.465, "num_input_tokens_seen": 61113024, "step": 3960 }, { "epoch": 16.266666666666666, "grad_norm": 0.1787109375, "learning_rate": 1.4533442093286853e-06, "loss": 0.4703, "num_input_tokens_seen": 61184256, "step": 3965 }, { "epoch": 16.287179487179486, "grad_norm": 0.1416015625, "learning_rate": 1.4503156871570062e-06, "loss": 0.4773, "num_input_tokens_seen": 61262208, "step": 3970 }, { "epoch": 16.307692307692307, "grad_norm": 0.1630859375, "learning_rate": 1.4472873677285706e-06, "loss": 0.4172, "num_input_tokens_seen": 61334560, "step": 3975 }, { "epoch": 16.328205128205127, "grad_norm": 0.1484375, "learning_rate": 1.4442592634008266e-06, "loss": 0.4326, "num_input_tokens_seen": 61405888, "step": 3980 }, { "epoch": 16.348717948717947, "grad_norm": 0.1240234375, "learning_rate": 1.4412313865303438e-06, "loss": 0.5303, "num_input_tokens_seen": 61489280, "step": 3985 }, { "epoch": 16.369230769230768, "grad_norm": 0.16015625, "learning_rate": 1.4382037494727649e-06, "loss": 0.4726, "num_input_tokens_seen": 61562912, "step": 3990 }, { "epoch": 16.38974358974359, "grad_norm": 0.1328125, "learning_rate": 1.435176364582752e-06, "loss": 0.458, "num_input_tokens_seen": 61652064, "step": 3995 }, { "epoch": 16.41025641025641, "grad_norm": 0.1171875, "learning_rate": 1.4321492442139405e-06, "loss": 0.5266, "num_input_tokens_seen": 61737536, "step": 4000 }, { "epoch": 16.43076923076923, "grad_norm": 0.158203125, "learning_rate": 1.4291224007188849e-06, "loss": 0.4394, "num_input_tokens_seen": 61811840, "step": 4005 }, { "epoch": 16.45128205128205, "grad_norm": 0.142578125, "learning_rate": 1.4260958464490104e-06, "loss": 0.4474, "num_input_tokens_seen": 61884288, "step": 4010 }, { "epoch": 16.47179487179487, "grad_norm": 0.22265625, "learning_rate": 1.4230695937545616e-06, "loss": 0.5, "num_input_tokens_seen": 61958944, "step": 4015 }, { "epoch": 16.49230769230769, "grad_norm": 0.1455078125, "learning_rate": 1.420043654984553e-06, "loss": 0.4766, "num_input_tokens_seen": 62038016, "step": 4020 }, { "epoch": 16.51282051282051, "grad_norm": 0.2890625, "learning_rate": 1.4170180424867176e-06, "loss": 0.4593, "num_input_tokens_seen": 62107616, "step": 4025 }, { "epoch": 16.533333333333335, "grad_norm": 0.359375, "learning_rate": 1.4139927686074577e-06, "loss": 0.4446, "num_input_tokens_seen": 62188352, "step": 4030 }, { "epoch": 16.553846153846155, "grad_norm": 0.259765625, "learning_rate": 1.4109678456917926e-06, "loss": 0.4658, "num_input_tokens_seen": 62259456, "step": 4035 }, { "epoch": 16.574358974358976, "grad_norm": 0.50390625, "learning_rate": 1.4079432860833104e-06, "loss": 0.5363, "num_input_tokens_seen": 62356448, "step": 4040 }, { "epoch": 16.594871794871796, "grad_norm": 0.78515625, "learning_rate": 1.4049191021241176e-06, "loss": 0.5006, "num_input_tokens_seen": 62429824, "step": 4045 }, { "epoch": 16.615384615384617, "grad_norm": 0.66796875, "learning_rate": 1.4018953061547853e-06, "loss": 0.4638, "num_input_tokens_seen": 62514528, "step": 4050 }, { "epoch": 16.635897435897437, "grad_norm": 0.267578125, "learning_rate": 1.3988719105143038e-06, "loss": 0.4413, "num_input_tokens_seen": 62593472, "step": 4055 }, { "epoch": 16.656410256410258, "grad_norm": 0.48046875, "learning_rate": 1.395848927540028e-06, "loss": 0.4506, "num_input_tokens_seen": 62671520, "step": 4060 }, { "epoch": 16.676923076923078, "grad_norm": 0.5078125, "learning_rate": 1.39282636956763e-06, "loss": 0.4616, "num_input_tokens_seen": 62749856, "step": 4065 }, { "epoch": 16.6974358974359, "grad_norm": 0.32421875, "learning_rate": 1.3898042489310471e-06, "loss": 0.456, "num_input_tokens_seen": 62823616, "step": 4070 }, { "epoch": 16.71794871794872, "grad_norm": 0.357421875, "learning_rate": 1.3867825779624325e-06, "loss": 0.45, "num_input_tokens_seen": 62908416, "step": 4075 }, { "epoch": 16.73846153846154, "grad_norm": 0.41015625, "learning_rate": 1.3837613689921037e-06, "loss": 0.4563, "num_input_tokens_seen": 62978016, "step": 4080 }, { "epoch": 16.75897435897436, "grad_norm": 0.259765625, "learning_rate": 1.380740634348494e-06, "loss": 0.4952, "num_input_tokens_seen": 63065952, "step": 4085 }, { "epoch": 16.77948717948718, "grad_norm": 0.3359375, "learning_rate": 1.3777203863580993e-06, "loss": 0.4327, "num_input_tokens_seen": 63141792, "step": 4090 }, { "epoch": 16.8, "grad_norm": 0.361328125, "learning_rate": 1.3747006373454321e-06, "loss": 0.5152, "num_input_tokens_seen": 63223520, "step": 4095 }, { "epoch": 16.82051282051282, "grad_norm": 0.326171875, "learning_rate": 1.3716813996329669e-06, "loss": 0.4952, "num_input_tokens_seen": 63297696, "step": 4100 }, { "epoch": 16.84102564102564, "grad_norm": 0.4609375, "learning_rate": 1.3686626855410929e-06, "loss": 0.4873, "num_input_tokens_seen": 63378304, "step": 4105 }, { "epoch": 16.861538461538462, "grad_norm": 0.388671875, "learning_rate": 1.3656445073880615e-06, "loss": 0.4461, "num_input_tokens_seen": 63454400, "step": 4110 }, { "epoch": 16.882051282051282, "grad_norm": 0.3671875, "learning_rate": 1.3626268774899381e-06, "loss": 0.4612, "num_input_tokens_seen": 63526400, "step": 4115 }, { "epoch": 16.902564102564103, "grad_norm": 0.33984375, "learning_rate": 1.3596098081605505e-06, "loss": 0.5061, "num_input_tokens_seen": 63612448, "step": 4120 }, { "epoch": 16.923076923076923, "grad_norm": 0.34765625, "learning_rate": 1.3565933117114385e-06, "loss": 0.486, "num_input_tokens_seen": 63693248, "step": 4125 }, { "epoch": 16.943589743589744, "grad_norm": 0.35546875, "learning_rate": 1.3535774004518057e-06, "loss": 0.4772, "num_input_tokens_seen": 63765504, "step": 4130 }, { "epoch": 16.964102564102564, "grad_norm": 0.330078125, "learning_rate": 1.3505620866884666e-06, "loss": 0.457, "num_input_tokens_seen": 63839616, "step": 4135 }, { "epoch": 16.984615384615385, "grad_norm": 0.341796875, "learning_rate": 1.347547382725797e-06, "loss": 0.4961, "num_input_tokens_seen": 63924736, "step": 4140 }, { "epoch": 17.005128205128205, "grad_norm": 0.349609375, "learning_rate": 1.344533300865686e-06, "loss": 0.5218, "num_input_tokens_seen": 64000256, "step": 4145 }, { "epoch": 17.025641025641026, "grad_norm": 0.44921875, "learning_rate": 1.3415198534074823e-06, "loss": 0.4501, "num_input_tokens_seen": 64077472, "step": 4150 }, { "epoch": 17.046153846153846, "grad_norm": 0.2578125, "learning_rate": 1.3385070526479475e-06, "loss": 0.4704, "num_input_tokens_seen": 64157024, "step": 4155 }, { "epoch": 17.066666666666666, "grad_norm": 0.2734375, "learning_rate": 1.3354949108812026e-06, "loss": 0.485, "num_input_tokens_seen": 64226656, "step": 4160 }, { "epoch": 17.087179487179487, "grad_norm": 0.2236328125, "learning_rate": 1.3324834403986815e-06, "loss": 0.4316, "num_input_tokens_seen": 64302816, "step": 4165 }, { "epoch": 17.107692307692307, "grad_norm": 0.2373046875, "learning_rate": 1.3294726534890766e-06, "loss": 0.4781, "num_input_tokens_seen": 64371840, "step": 4170 }, { "epoch": 17.128205128205128, "grad_norm": 0.189453125, "learning_rate": 1.326462562438293e-06, "loss": 0.4331, "num_input_tokens_seen": 64451776, "step": 4175 }, { "epoch": 17.148717948717948, "grad_norm": 0.2080078125, "learning_rate": 1.3234531795293945e-06, "loss": 0.4754, "num_input_tokens_seen": 64523712, "step": 4180 }, { "epoch": 17.16923076923077, "grad_norm": 0.1904296875, "learning_rate": 1.3204445170425565e-06, "loss": 0.4983, "num_input_tokens_seen": 64594272, "step": 4185 }, { "epoch": 17.18974358974359, "grad_norm": 0.1650390625, "learning_rate": 1.3174365872550138e-06, "loss": 0.4852, "num_input_tokens_seen": 64677280, "step": 4190 }, { "epoch": 17.21025641025641, "grad_norm": 0.1591796875, "learning_rate": 1.3144294024410122e-06, "loss": 0.4843, "num_input_tokens_seen": 64758080, "step": 4195 }, { "epoch": 17.23076923076923, "grad_norm": 0.1435546875, "learning_rate": 1.3114229748717563e-06, "loss": 0.4564, "num_input_tokens_seen": 64830848, "step": 4200 }, { "epoch": 17.25128205128205, "grad_norm": 0.1474609375, "learning_rate": 1.308417316815362e-06, "loss": 0.4432, "num_input_tokens_seen": 64907296, "step": 4205 }, { "epoch": 17.27179487179487, "grad_norm": 0.134765625, "learning_rate": 1.3054124405368036e-06, "loss": 0.4901, "num_input_tokens_seen": 64987392, "step": 4210 }, { "epoch": 17.29230769230769, "grad_norm": 0.1318359375, "learning_rate": 1.3024083582978668e-06, "loss": 0.5076, "num_input_tokens_seen": 65077408, "step": 4215 }, { "epoch": 17.31282051282051, "grad_norm": 0.12158203125, "learning_rate": 1.2994050823570968e-06, "loss": 0.4818, "num_input_tokens_seen": 65159104, "step": 4220 }, { "epoch": 17.333333333333332, "grad_norm": 0.099609375, "learning_rate": 1.2964026249697475e-06, "loss": 0.4525, "num_input_tokens_seen": 65242816, "step": 4225 }, { "epoch": 17.353846153846153, "grad_norm": 0.1337890625, "learning_rate": 1.293400998387734e-06, "loss": 0.4444, "num_input_tokens_seen": 65317376, "step": 4230 }, { "epoch": 17.374358974358973, "grad_norm": 0.2314453125, "learning_rate": 1.2904002148595797e-06, "loss": 0.4897, "num_input_tokens_seen": 65391264, "step": 4235 }, { "epoch": 17.394871794871793, "grad_norm": 0.208984375, "learning_rate": 1.2874002866303695e-06, "loss": 0.5116, "num_input_tokens_seen": 65467744, "step": 4240 }, { "epoch": 17.415384615384614, "grad_norm": 0.1298828125, "learning_rate": 1.2844012259416965e-06, "loss": 0.4498, "num_input_tokens_seen": 65535680, "step": 4245 }, { "epoch": 17.435897435897434, "grad_norm": 0.2890625, "learning_rate": 1.2814030450316151e-06, "loss": 0.4984, "num_input_tokens_seen": 65611680, "step": 4250 }, { "epoch": 17.456410256410255, "grad_norm": 0.318359375, "learning_rate": 1.2784057561345885e-06, "loss": 0.5276, "num_input_tokens_seen": 65684160, "step": 4255 }, { "epoch": 17.476923076923075, "grad_norm": 0.291015625, "learning_rate": 1.2754093714814407e-06, "loss": 0.4434, "num_input_tokens_seen": 65755104, "step": 4260 }, { "epoch": 17.4974358974359, "grad_norm": 0.59765625, "learning_rate": 1.2724139032993057e-06, "loss": 0.4743, "num_input_tokens_seen": 65835072, "step": 4265 }, { "epoch": 17.51794871794872, "grad_norm": 0.89453125, "learning_rate": 1.269419363811577e-06, "loss": 0.4691, "num_input_tokens_seen": 65916704, "step": 4270 }, { "epoch": 17.53846153846154, "grad_norm": 0.6484375, "learning_rate": 1.2664257652378598e-06, "loss": 0.5129, "num_input_tokens_seen": 65989696, "step": 4275 }, { "epoch": 17.55897435897436, "grad_norm": 0.259765625, "learning_rate": 1.2634331197939183e-06, "loss": 0.4868, "num_input_tokens_seen": 66071744, "step": 4280 }, { "epoch": 17.57948717948718, "grad_norm": 0.54296875, "learning_rate": 1.2604414396916286e-06, "loss": 0.489, "num_input_tokens_seen": 66153024, "step": 4285 }, { "epoch": 17.6, "grad_norm": 0.55078125, "learning_rate": 1.2574507371389267e-06, "loss": 0.4855, "num_input_tokens_seen": 66220384, "step": 4290 }, { "epoch": 17.620512820512822, "grad_norm": 0.298828125, "learning_rate": 1.25446102433976e-06, "loss": 0.4791, "num_input_tokens_seen": 66298912, "step": 4295 }, { "epoch": 17.641025641025642, "grad_norm": 0.263671875, "learning_rate": 1.2514723134940365e-06, "loss": 0.4563, "num_input_tokens_seen": 66373120, "step": 4300 }, { "epoch": 17.661538461538463, "grad_norm": 0.26171875, "learning_rate": 1.2484846167975767e-06, "loss": 0.4577, "num_input_tokens_seen": 66452096, "step": 4305 }, { "epoch": 17.682051282051283, "grad_norm": 0.302734375, "learning_rate": 1.2454979464420624e-06, "loss": 0.4875, "num_input_tokens_seen": 66530784, "step": 4310 }, { "epoch": 17.702564102564104, "grad_norm": 0.34375, "learning_rate": 1.2425123146149863e-06, "loss": 0.4551, "num_input_tokens_seen": 66607264, "step": 4315 }, { "epoch": 17.723076923076924, "grad_norm": 0.328125, "learning_rate": 1.2395277334996047e-06, "loss": 0.5378, "num_input_tokens_seen": 66692608, "step": 4320 }, { "epoch": 17.743589743589745, "grad_norm": 0.341796875, "learning_rate": 1.2365442152748846e-06, "loss": 0.4518, "num_input_tokens_seen": 66773504, "step": 4325 }, { "epoch": 17.764102564102565, "grad_norm": 0.625, "learning_rate": 1.2335617721154577e-06, "loss": 0.4412, "num_input_tokens_seen": 66844928, "step": 4330 }, { "epoch": 17.784615384615385, "grad_norm": 0.443359375, "learning_rate": 1.2305804161915671e-06, "loss": 0.5214, "num_input_tokens_seen": 66918080, "step": 4335 }, { "epoch": 17.805128205128206, "grad_norm": 0.32421875, "learning_rate": 1.2276001596690205e-06, "loss": 0.47, "num_input_tokens_seen": 66988992, "step": 4340 }, { "epoch": 17.825641025641026, "grad_norm": 0.419921875, "learning_rate": 1.2246210147091382e-06, "loss": 0.4545, "num_input_tokens_seen": 67068288, "step": 4345 }, { "epoch": 17.846153846153847, "grad_norm": 0.3125, "learning_rate": 1.2216429934687062e-06, "loss": 0.4574, "num_input_tokens_seen": 67142560, "step": 4350 }, { "epoch": 17.866666666666667, "grad_norm": 0.40234375, "learning_rate": 1.2186661080999234e-06, "loss": 0.4717, "num_input_tokens_seen": 67222560, "step": 4355 }, { "epoch": 17.887179487179488, "grad_norm": 0.318359375, "learning_rate": 1.2156903707503544e-06, "loss": 0.4662, "num_input_tokens_seen": 67301760, "step": 4360 }, { "epoch": 17.907692307692308, "grad_norm": 0.345703125, "learning_rate": 1.2127157935628789e-06, "loss": 0.4958, "num_input_tokens_seen": 67381856, "step": 4365 }, { "epoch": 17.92820512820513, "grad_norm": 0.318359375, "learning_rate": 1.2097423886756433e-06, "loss": 0.5327, "num_input_tokens_seen": 67467968, "step": 4370 }, { "epoch": 17.94871794871795, "grad_norm": 0.248046875, "learning_rate": 1.2067701682220084e-06, "loss": 0.4606, "num_input_tokens_seen": 67548672, "step": 4375 }, { "epoch": 17.96923076923077, "grad_norm": 0.255859375, "learning_rate": 1.2037991443305043e-06, "loss": 0.4451, "num_input_tokens_seen": 67625248, "step": 4380 }, { "epoch": 17.98974358974359, "grad_norm": 0.28125, "learning_rate": 1.2008293291247754e-06, "loss": 0.4416, "num_input_tokens_seen": 67695808, "step": 4385 }, { "epoch": 18.01025641025641, "grad_norm": 0.23046875, "learning_rate": 1.1978607347235367e-06, "loss": 0.4741, "num_input_tokens_seen": 67774400, "step": 4390 }, { "epoch": 18.03076923076923, "grad_norm": 0.236328125, "learning_rate": 1.1948933732405205e-06, "loss": 0.446, "num_input_tokens_seen": 67846848, "step": 4395 }, { "epoch": 18.05128205128205, "grad_norm": 0.2060546875, "learning_rate": 1.191927256784427e-06, "loss": 0.4315, "num_input_tokens_seen": 67917440, "step": 4400 }, { "epoch": 18.07179487179487, "grad_norm": 0.181640625, "learning_rate": 1.1889623974588772e-06, "loss": 0.4055, "num_input_tokens_seen": 67990656, "step": 4405 }, { "epoch": 18.092307692307692, "grad_norm": 0.1591796875, "learning_rate": 1.185998807362362e-06, "loss": 0.499, "num_input_tokens_seen": 68067872, "step": 4410 }, { "epoch": 18.112820512820512, "grad_norm": 0.162109375, "learning_rate": 1.1830364985881924e-06, "loss": 0.481, "num_input_tokens_seen": 68145632, "step": 4415 }, { "epoch": 18.133333333333333, "grad_norm": 0.1650390625, "learning_rate": 1.1800754832244515e-06, "loss": 0.4931, "num_input_tokens_seen": 68221472, "step": 4420 }, { "epoch": 18.153846153846153, "grad_norm": 0.1513671875, "learning_rate": 1.1771157733539442e-06, "loss": 0.4377, "num_input_tokens_seen": 68292352, "step": 4425 }, { "epoch": 18.174358974358974, "grad_norm": 0.150390625, "learning_rate": 1.174157381054148e-06, "loss": 0.5265, "num_input_tokens_seen": 68374880, "step": 4430 }, { "epoch": 18.194871794871794, "grad_norm": 0.1416015625, "learning_rate": 1.1712003183971644e-06, "loss": 0.5103, "num_input_tokens_seen": 68457280, "step": 4435 }, { "epoch": 18.215384615384615, "grad_norm": 0.119140625, "learning_rate": 1.1682445974496686e-06, "loss": 0.5133, "num_input_tokens_seen": 68540224, "step": 4440 }, { "epoch": 18.235897435897435, "grad_norm": 0.11865234375, "learning_rate": 1.1652902302728607e-06, "loss": 0.4805, "num_input_tokens_seen": 68621056, "step": 4445 }, { "epoch": 18.256410256410255, "grad_norm": 0.12353515625, "learning_rate": 1.1623372289224172e-06, "loss": 0.4863, "num_input_tokens_seen": 68697440, "step": 4450 }, { "epoch": 18.276923076923076, "grad_norm": 0.125, "learning_rate": 1.1593856054484403e-06, "loss": 0.4685, "num_input_tokens_seen": 68778944, "step": 4455 }, { "epoch": 18.297435897435896, "grad_norm": 0.1982421875, "learning_rate": 1.156435371895411e-06, "loss": 0.4934, "num_input_tokens_seen": 68859712, "step": 4460 }, { "epoch": 18.317948717948717, "grad_norm": 0.1962890625, "learning_rate": 1.1534865403021366e-06, "loss": 0.4439, "num_input_tokens_seen": 68932512, "step": 4465 }, { "epoch": 18.338461538461537, "grad_norm": 0.15625, "learning_rate": 1.1505391227017046e-06, "loss": 0.4295, "num_input_tokens_seen": 69006080, "step": 4470 }, { "epoch": 18.358974358974358, "grad_norm": 0.29296875, "learning_rate": 1.1475931311214338e-06, "loss": 0.4644, "num_input_tokens_seen": 69081984, "step": 4475 }, { "epoch": 18.379487179487178, "grad_norm": 0.2421875, "learning_rate": 1.144648577582821e-06, "loss": 0.4899, "num_input_tokens_seen": 69158240, "step": 4480 }, { "epoch": 18.4, "grad_norm": 0.3671875, "learning_rate": 1.141705474101498e-06, "loss": 0.4763, "num_input_tokens_seen": 69241184, "step": 4485 }, { "epoch": 18.42051282051282, "grad_norm": 0.5859375, "learning_rate": 1.138763832687177e-06, "loss": 0.4784, "num_input_tokens_seen": 69326112, "step": 4490 }, { "epoch": 18.44102564102564, "grad_norm": 0.765625, "learning_rate": 1.1358236653436052e-06, "loss": 0.4624, "num_input_tokens_seen": 69400832, "step": 4495 }, { "epoch": 18.46153846153846, "grad_norm": 0.8984375, "learning_rate": 1.1328849840685143e-06, "loss": 0.4523, "num_input_tokens_seen": 69478464, "step": 4500 }, { "epoch": 18.48205128205128, "grad_norm": 0.43359375, "learning_rate": 1.1299478008535726e-06, "loss": 0.5372, "num_input_tokens_seen": 69559712, "step": 4505 }, { "epoch": 18.5025641025641, "grad_norm": 0.458984375, "learning_rate": 1.1270121276843342e-06, "loss": 0.4224, "num_input_tokens_seen": 69638528, "step": 4510 }, { "epoch": 18.523076923076925, "grad_norm": 0.23828125, "learning_rate": 1.1240779765401926e-06, "loss": 0.4726, "num_input_tokens_seen": 69719264, "step": 4515 }, { "epoch": 18.543589743589745, "grad_norm": 0.30078125, "learning_rate": 1.1211453593943293e-06, "loss": 0.4407, "num_input_tokens_seen": 69798432, "step": 4520 }, { "epoch": 18.564102564102566, "grad_norm": 0.310546875, "learning_rate": 1.118214288213667e-06, "loss": 0.4594, "num_input_tokens_seen": 69869984, "step": 4525 }, { "epoch": 18.584615384615386, "grad_norm": 0.318359375, "learning_rate": 1.1152847749588186e-06, "loss": 0.5029, "num_input_tokens_seen": 69949088, "step": 4530 }, { "epoch": 18.605128205128207, "grad_norm": 0.6796875, "learning_rate": 1.1123568315840419e-06, "loss": 0.4386, "num_input_tokens_seen": 70023904, "step": 4535 }, { "epoch": 18.625641025641027, "grad_norm": 0.5625, "learning_rate": 1.1094304700371863e-06, "loss": 0.5118, "num_input_tokens_seen": 70098400, "step": 4540 }, { "epoch": 18.646153846153847, "grad_norm": 0.31640625, "learning_rate": 1.1065057022596483e-06, "loss": 0.4559, "num_input_tokens_seen": 70176576, "step": 4545 }, { "epoch": 18.666666666666668, "grad_norm": 0.345703125, "learning_rate": 1.1035825401863185e-06, "loss": 0.4418, "num_input_tokens_seen": 70255264, "step": 4550 }, { "epoch": 18.68717948717949, "grad_norm": 0.345703125, "learning_rate": 1.100660995745538e-06, "loss": 0.435, "num_input_tokens_seen": 70324160, "step": 4555 }, { "epoch": 18.70769230769231, "grad_norm": 0.365234375, "learning_rate": 1.0977410808590437e-06, "loss": 0.5499, "num_input_tokens_seen": 70411616, "step": 4560 }, { "epoch": 18.72820512820513, "grad_norm": 0.470703125, "learning_rate": 1.0948228074419269e-06, "loss": 0.4705, "num_input_tokens_seen": 70482592, "step": 4565 }, { "epoch": 18.74871794871795, "grad_norm": 0.44921875, "learning_rate": 1.0919061874025774e-06, "loss": 0.4466, "num_input_tokens_seen": 70565120, "step": 4570 }, { "epoch": 18.76923076923077, "grad_norm": 0.3984375, "learning_rate": 1.0889912326426393e-06, "loss": 0.4323, "num_input_tokens_seen": 70635168, "step": 4575 }, { "epoch": 18.78974358974359, "grad_norm": 0.310546875, "learning_rate": 1.0860779550569609e-06, "loss": 0.5235, "num_input_tokens_seen": 70720096, "step": 4580 }, { "epoch": 18.81025641025641, "grad_norm": 0.322265625, "learning_rate": 1.083166366533548e-06, "loss": 0.4659, "num_input_tokens_seen": 70807648, "step": 4585 }, { "epoch": 18.83076923076923, "grad_norm": 0.314453125, "learning_rate": 1.080256478953512e-06, "loss": 0.4968, "num_input_tokens_seen": 70888832, "step": 4590 }, { "epoch": 18.851282051282052, "grad_norm": 0.33203125, "learning_rate": 1.0773483041910247e-06, "loss": 0.5112, "num_input_tokens_seen": 70971136, "step": 4595 }, { "epoch": 18.871794871794872, "grad_norm": 0.29296875, "learning_rate": 1.0744418541132676e-06, "loss": 0.4414, "num_input_tokens_seen": 71038816, "step": 4600 }, { "epoch": 18.892307692307693, "grad_norm": 0.314453125, "learning_rate": 1.0715371405803858e-06, "loss": 0.4874, "num_input_tokens_seen": 71114720, "step": 4605 }, { "epoch": 18.912820512820513, "grad_norm": 0.294921875, "learning_rate": 1.0686341754454364e-06, "loss": 0.4571, "num_input_tokens_seen": 71189856, "step": 4610 }, { "epoch": 18.933333333333334, "grad_norm": 0.23046875, "learning_rate": 1.0657329705543439e-06, "loss": 0.4661, "num_input_tokens_seen": 71271136, "step": 4615 }, { "epoch": 18.953846153846154, "grad_norm": 0.2109375, "learning_rate": 1.0628335377458477e-06, "loss": 0.435, "num_input_tokens_seen": 71343488, "step": 4620 }, { "epoch": 18.974358974358974, "grad_norm": 0.21484375, "learning_rate": 1.0599358888514582e-06, "loss": 0.4845, "num_input_tokens_seen": 71419104, "step": 4625 }, { "epoch": 18.994871794871795, "grad_norm": 0.2021484375, "learning_rate": 1.0570400356954044e-06, "loss": 0.5111, "num_input_tokens_seen": 71491776, "step": 4630 }, { "epoch": 19.015384615384615, "grad_norm": 0.18359375, "learning_rate": 1.0541459900945892e-06, "loss": 0.5006, "num_input_tokens_seen": 71573184, "step": 4635 }, { "epoch": 19.035897435897436, "grad_norm": 0.1630859375, "learning_rate": 1.0512537638585379e-06, "loss": 0.4828, "num_input_tokens_seen": 71648768, "step": 4640 }, { "epoch": 19.056410256410256, "grad_norm": 0.1572265625, "learning_rate": 1.0483633687893526e-06, "loss": 0.4607, "num_input_tokens_seen": 71721824, "step": 4645 }, { "epoch": 19.076923076923077, "grad_norm": 0.1376953125, "learning_rate": 1.0454748166816645e-06, "loss": 0.497, "num_input_tokens_seen": 71797568, "step": 4650 }, { "epoch": 19.097435897435897, "grad_norm": 0.13671875, "learning_rate": 1.0425881193225808e-06, "loss": 0.4607, "num_input_tokens_seen": 71871136, "step": 4655 }, { "epoch": 19.117948717948718, "grad_norm": 0.1337890625, "learning_rate": 1.0397032884916438e-06, "loss": 0.4232, "num_input_tokens_seen": 71941984, "step": 4660 }, { "epoch": 19.138461538461538, "grad_norm": 0.123046875, "learning_rate": 1.0368203359607767e-06, "loss": 0.5216, "num_input_tokens_seen": 72024736, "step": 4665 }, { "epoch": 19.15897435897436, "grad_norm": 0.12109375, "learning_rate": 1.0339392734942393e-06, "loss": 0.4565, "num_input_tokens_seen": 72096064, "step": 4670 }, { "epoch": 19.17948717948718, "grad_norm": 0.1337890625, "learning_rate": 1.031060112848578e-06, "loss": 0.4488, "num_input_tokens_seen": 72171168, "step": 4675 }, { "epoch": 19.2, "grad_norm": 0.126953125, "learning_rate": 1.0281828657725798e-06, "loss": 0.4567, "num_input_tokens_seen": 72249056, "step": 4680 }, { "epoch": 19.22051282051282, "grad_norm": 0.2080078125, "learning_rate": 1.0253075440072212e-06, "loss": 0.4556, "num_input_tokens_seen": 72338944, "step": 4685 }, { "epoch": 19.24102564102564, "grad_norm": 0.1484375, "learning_rate": 1.0224341592856245e-06, "loss": 0.4686, "num_input_tokens_seen": 72412032, "step": 4690 }, { "epoch": 19.26153846153846, "grad_norm": 0.1435546875, "learning_rate": 1.0195627233330052e-06, "loss": 0.4087, "num_input_tokens_seen": 72485760, "step": 4695 }, { "epoch": 19.28205128205128, "grad_norm": 0.3046875, "learning_rate": 1.0166932478666292e-06, "loss": 0.4634, "num_input_tokens_seen": 72565344, "step": 4700 }, { "epoch": 19.3025641025641, "grad_norm": 0.275390625, "learning_rate": 1.0138257445957601e-06, "loss": 0.5316, "num_input_tokens_seen": 72644800, "step": 4705 }, { "epoch": 19.323076923076922, "grad_norm": 0.31640625, "learning_rate": 1.0109602252216153e-06, "loss": 0.4929, "num_input_tokens_seen": 72725440, "step": 4710 }, { "epoch": 19.343589743589742, "grad_norm": 0.640625, "learning_rate": 1.0080967014373152e-06, "loss": 0.442, "num_input_tokens_seen": 72796064, "step": 4715 }, { "epoch": 19.364102564102563, "grad_norm": 0.91015625, "learning_rate": 1.0052351849278385e-06, "loss": 0.4442, "num_input_tokens_seen": 72870080, "step": 4720 }, { "epoch": 19.384615384615383, "grad_norm": 0.2490234375, "learning_rate": 1.0023756873699723e-06, "loss": 0.4751, "num_input_tokens_seen": 72944288, "step": 4725 }, { "epoch": 19.405128205128204, "grad_norm": 0.333984375, "learning_rate": 9.995182204322637e-07, "loss": 0.5129, "num_input_tokens_seen": 73021664, "step": 4730 }, { "epoch": 19.425641025641024, "grad_norm": 0.81640625, "learning_rate": 9.966627957749767e-07, "loss": 0.4796, "num_input_tokens_seen": 73097920, "step": 4735 }, { "epoch": 19.446153846153845, "grad_norm": 0.333984375, "learning_rate": 9.93809425050039e-07, "loss": 0.4785, "num_input_tokens_seen": 73170112, "step": 4740 }, { "epoch": 19.466666666666665, "grad_norm": 0.2255859375, "learning_rate": 9.909581199009971e-07, "loss": 0.4379, "num_input_tokens_seen": 73241312, "step": 4745 }, { "epoch": 19.487179487179485, "grad_norm": 0.267578125, "learning_rate": 9.8810889196297e-07, "loss": 0.5408, "num_input_tokens_seen": 73319936, "step": 4750 }, { "epoch": 19.50769230769231, "grad_norm": 0.3046875, "learning_rate": 9.852617528625992e-07, "loss": 0.4878, "num_input_tokens_seen": 73397824, "step": 4755 }, { "epoch": 19.52820512820513, "grad_norm": 0.29296875, "learning_rate": 9.824167142180032e-07, "loss": 0.4627, "num_input_tokens_seen": 73470688, "step": 4760 }, { "epoch": 19.54871794871795, "grad_norm": 0.283203125, "learning_rate": 9.795737876387285e-07, "loss": 0.4799, "num_input_tokens_seen": 73545792, "step": 4765 }, { "epoch": 19.56923076923077, "grad_norm": 0.34765625, "learning_rate": 9.76732984725704e-07, "loss": 0.428, "num_input_tokens_seen": 73617760, "step": 4770 }, { "epoch": 19.58974358974359, "grad_norm": 0.3515625, "learning_rate": 9.738943170711916e-07, "loss": 0.4608, "num_input_tokens_seen": 73707328, "step": 4775 }, { "epoch": 19.61025641025641, "grad_norm": 0.515625, "learning_rate": 9.710577962587412e-07, "loss": 0.4951, "num_input_tokens_seen": 73782976, "step": 4780 }, { "epoch": 19.630769230769232, "grad_norm": 0.3828125, "learning_rate": 9.68223433863141e-07, "loss": 0.53, "num_input_tokens_seen": 73872256, "step": 4785 }, { "epoch": 19.651282051282053, "grad_norm": 0.341796875, "learning_rate": 9.653912414503731e-07, "loss": 0.45, "num_input_tokens_seen": 73947520, "step": 4790 }, { "epoch": 19.671794871794873, "grad_norm": 0.337890625, "learning_rate": 9.625612305775626e-07, "loss": 0.4218, "num_input_tokens_seen": 74023104, "step": 4795 }, { "epoch": 19.692307692307693, "grad_norm": 0.333984375, "learning_rate": 9.597334127929346e-07, "loss": 0.4531, "num_input_tokens_seen": 74105920, "step": 4800 }, { "epoch": 19.712820512820514, "grad_norm": 0.306640625, "learning_rate": 9.569077996357638e-07, "loss": 0.4142, "num_input_tokens_seen": 74182752, "step": 4805 }, { "epoch": 19.733333333333334, "grad_norm": 0.37109375, "learning_rate": 9.54084402636329e-07, "loss": 0.5121, "num_input_tokens_seen": 74265248, "step": 4810 }, { "epoch": 19.753846153846155, "grad_norm": 0.328125, "learning_rate": 9.512632333158653e-07, "loss": 0.523, "num_input_tokens_seen": 74363232, "step": 4815 }, { "epoch": 19.774358974358975, "grad_norm": 0.353515625, "learning_rate": 9.484443031865176e-07, "loss": 0.5103, "num_input_tokens_seen": 74442848, "step": 4820 }, { "epoch": 19.794871794871796, "grad_norm": 0.27734375, "learning_rate": 9.456276237512949e-07, "loss": 0.4725, "num_input_tokens_seen": 74513376, "step": 4825 }, { "epoch": 19.815384615384616, "grad_norm": 0.271484375, "learning_rate": 9.428132065040198e-07, "loss": 0.4921, "num_input_tokens_seen": 74588928, "step": 4830 }, { "epoch": 19.835897435897436, "grad_norm": 0.291015625, "learning_rate": 9.40001062929285e-07, "loss": 0.4331, "num_input_tokens_seen": 74659040, "step": 4835 }, { "epoch": 19.856410256410257, "grad_norm": 0.216796875, "learning_rate": 9.371912045024046e-07, "loss": 0.44, "num_input_tokens_seen": 74737472, "step": 4840 }, { "epoch": 19.876923076923077, "grad_norm": 0.2158203125, "learning_rate": 9.343836426893687e-07, "loss": 0.4831, "num_input_tokens_seen": 74810592, "step": 4845 }, { "epoch": 19.897435897435898, "grad_norm": 0.1943359375, "learning_rate": 9.315783889467943e-07, "loss": 0.4494, "num_input_tokens_seen": 74880096, "step": 4850 }, { "epoch": 19.91794871794872, "grad_norm": 0.1953125, "learning_rate": 9.287754547218821e-07, "loss": 0.5063, "num_input_tokens_seen": 74951168, "step": 4855 }, { "epoch": 19.93846153846154, "grad_norm": 0.1611328125, "learning_rate": 9.259748514523654e-07, "loss": 0.4975, "num_input_tokens_seen": 75034592, "step": 4860 }, { "epoch": 19.95897435897436, "grad_norm": 0.1728515625, "learning_rate": 9.231765905664677e-07, "loss": 0.4802, "num_input_tokens_seen": 75108992, "step": 4865 }, { "epoch": 19.97948717948718, "grad_norm": 0.166015625, "learning_rate": 9.20380683482853e-07, "loss": 0.4473, "num_input_tokens_seen": 75185824, "step": 4870 }, { "epoch": 20.0, "grad_norm": 0.140625, "learning_rate": 9.175871416105802e-07, "loss": 0.4971, "num_input_tokens_seen": 75263936, "step": 4875 }, { "epoch": 20.02051282051282, "grad_norm": 0.134765625, "learning_rate": 9.147959763490578e-07, "loss": 0.493, "num_input_tokens_seen": 75342752, "step": 4880 }, { "epoch": 20.04102564102564, "grad_norm": 0.11669921875, "learning_rate": 9.120071990879949e-07, "loss": 0.4709, "num_input_tokens_seen": 75420704, "step": 4885 }, { "epoch": 20.06153846153846, "grad_norm": 0.13671875, "learning_rate": 9.092208212073569e-07, "loss": 0.5248, "num_input_tokens_seen": 75496128, "step": 4890 }, { "epoch": 20.08205128205128, "grad_norm": 0.1220703125, "learning_rate": 9.064368540773177e-07, "loss": 0.525, "num_input_tokens_seen": 75585984, "step": 4895 }, { "epoch": 20.102564102564102, "grad_norm": 0.1357421875, "learning_rate": 9.036553090582145e-07, "loss": 0.5186, "num_input_tokens_seen": 75668416, "step": 4900 }, { "epoch": 20.123076923076923, "grad_norm": 0.15234375, "learning_rate": 9.008761975004994e-07, "loss": 0.456, "num_input_tokens_seen": 75744960, "step": 4905 }, { "epoch": 20.143589743589743, "grad_norm": 0.203125, "learning_rate": 8.98099530744696e-07, "loss": 0.4507, "num_input_tokens_seen": 75818688, "step": 4910 }, { "epoch": 20.164102564102564, "grad_norm": 0.150390625, "learning_rate": 8.953253201213517e-07, "loss": 0.4552, "num_input_tokens_seen": 75901824, "step": 4915 }, { "epoch": 20.184615384615384, "grad_norm": 0.146484375, "learning_rate": 8.925535769509895e-07, "loss": 0.4153, "num_input_tokens_seen": 75973248, "step": 4920 }, { "epoch": 20.205128205128204, "grad_norm": 0.310546875, "learning_rate": 8.897843125440653e-07, "loss": 0.4452, "num_input_tokens_seen": 76043840, "step": 4925 }, { "epoch": 20.217435897435898, "num_input_tokens_seen": 76105440, "step": 4928, "total_flos": 1.6998510508071322e+18, "train_loss": 0.4908539823365289, "train_runtime": 36017.1103, "train_samples_per_second": 3.465, "train_steps_per_second": 0.216 } ], "logging_steps": 5, "max_steps": 7776, "num_input_tokens_seen": 76105440, "num_train_epochs": 32, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.6998510508071322e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }