diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7923 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 20.217435897435898, + "eval_steps": 500, + "global_step": 4928, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.020512820512820513, + "grad_norm": 0.875, + "learning_rate": 2.9999969395283144e-06, + "loss": 0.6637, + "num_input_tokens_seen": 75456, + "step": 5 + }, + { + "epoch": 0.041025641025641026, + "grad_norm": 1.4140625, + "learning_rate": 2.9999877581257458e-06, + "loss": 0.6963, + "num_input_tokens_seen": 160992, + "step": 10 + }, + { + "epoch": 0.06153846153846154, + "grad_norm": 1.59375, + "learning_rate": 2.9999724558297605e-06, + "loss": 0.6919, + "num_input_tokens_seen": 253856, + "step": 15 + }, + { + "epoch": 0.08205128205128205, + "grad_norm": 2.40625, + "learning_rate": 2.999951032702801e-06, + "loss": 0.6605, + "num_input_tokens_seen": 328640, + "step": 20 + }, + { + "epoch": 0.10256410256410256, + "grad_norm": 0.75390625, + "learning_rate": 2.9999234888322877e-06, + "loss": 0.5888, + "num_input_tokens_seen": 401504, + "step": 25 + }, + { + "epoch": 0.12307692307692308, + "grad_norm": 1.578125, + "learning_rate": 2.9998898243306162e-06, + "loss": 0.6087, + "num_input_tokens_seen": 475168, + "step": 30 + }, + { + "epoch": 0.14358974358974358, + "grad_norm": 0.89453125, + "learning_rate": 2.99985003933516e-06, + "loss": 0.6405, + "num_input_tokens_seen": 547168, + "step": 35 + }, + { + "epoch": 0.1641025641025641, + "grad_norm": 1.1015625, + "learning_rate": 2.999804134008266e-06, + "loss": 0.6443, + "num_input_tokens_seen": 618176, + "step": 40 + }, + { + "epoch": 0.18461538461538463, + "grad_norm": 0.69921875, + "learning_rate": 2.9997521085372565e-06, + "loss": 0.6554, + "num_input_tokens_seen": 701184, + "step": 45 + }, + { + "epoch": 0.20512820512820512, + "grad_norm": 1.0234375, + "learning_rate": 2.999693963134429e-06, + "loss": 0.6903, + "num_input_tokens_seen": 782656, + "step": 50 + }, + { + "epoch": 0.22564102564102564, + "grad_norm": 0.796875, + "learning_rate": 2.9996296980370526e-06, + "loss": 0.6915, + "num_input_tokens_seen": 869280, + "step": 55 + }, + { + "epoch": 0.24615384615384617, + "grad_norm": 0.8515625, + "learning_rate": 2.99955931350737e-06, + "loss": 0.584, + "num_input_tokens_seen": 944032, + "step": 60 + }, + { + "epoch": 0.26666666666666666, + "grad_norm": 0.91796875, + "learning_rate": 2.999482809832594e-06, + "loss": 0.5908, + "num_input_tokens_seen": 1025568, + "step": 65 + }, + { + "epoch": 0.28717948717948716, + "grad_norm": 0.84375, + "learning_rate": 2.9994001873249074e-06, + "loss": 0.5811, + "num_input_tokens_seen": 1105344, + "step": 70 + }, + { + "epoch": 0.3076923076923077, + "grad_norm": 0.6953125, + "learning_rate": 2.999311446321462e-06, + "loss": 0.5792, + "num_input_tokens_seen": 1182016, + "step": 75 + }, + { + "epoch": 0.3282051282051282, + "grad_norm": 0.99609375, + "learning_rate": 2.999216587184378e-06, + "loss": 0.5923, + "num_input_tokens_seen": 1265440, + "step": 80 + }, + { + "epoch": 0.3487179487179487, + "grad_norm": 0.91015625, + "learning_rate": 2.9991156103007394e-06, + "loss": 0.5896, + "num_input_tokens_seen": 1347296, + "step": 85 + }, + { + "epoch": 0.36923076923076925, + "grad_norm": 0.90234375, + "learning_rate": 2.9990085160825954e-06, + "loss": 0.6283, + "num_input_tokens_seen": 1424064, + "step": 90 + }, + { + "epoch": 0.38974358974358975, + "grad_norm": 0.67578125, + "learning_rate": 2.9988953049669577e-06, + "loss": 0.5329, + "num_input_tokens_seen": 1492416, + "step": 95 + }, + { + "epoch": 0.41025641025641024, + "grad_norm": 0.44921875, + "learning_rate": 2.998775977415799e-06, + "loss": 0.5582, + "num_input_tokens_seen": 1572192, + "step": 100 + }, + { + "epoch": 0.4307692307692308, + "grad_norm": 0.59375, + "learning_rate": 2.998650533916051e-06, + "loss": 0.6529, + "num_input_tokens_seen": 1650592, + "step": 105 + }, + { + "epoch": 0.4512820512820513, + "grad_norm": 0.515625, + "learning_rate": 2.998518974979602e-06, + "loss": 0.5449, + "num_input_tokens_seen": 1728128, + "step": 110 + }, + { + "epoch": 0.4717948717948718, + "grad_norm": 0.494140625, + "learning_rate": 2.998381301143295e-06, + "loss": 0.573, + "num_input_tokens_seen": 1806080, + "step": 115 + }, + { + "epoch": 0.49230769230769234, + "grad_norm": 0.38671875, + "learning_rate": 2.9982375129689253e-06, + "loss": 0.6065, + "num_input_tokens_seen": 1884544, + "step": 120 + }, + { + "epoch": 0.5128205128205128, + "grad_norm": 0.396484375, + "learning_rate": 2.9980876110432404e-06, + "loss": 0.6232, + "num_input_tokens_seen": 1961376, + "step": 125 + }, + { + "epoch": 0.5333333333333333, + "grad_norm": 0.408203125, + "learning_rate": 2.9979315959779335e-06, + "loss": 0.5061, + "num_input_tokens_seen": 2036064, + "step": 130 + }, + { + "epoch": 0.5538461538461539, + "grad_norm": 0.271484375, + "learning_rate": 2.9977694684096447e-06, + "loss": 0.5657, + "num_input_tokens_seen": 2109376, + "step": 135 + }, + { + "epoch": 0.5743589743589743, + "grad_norm": 0.27734375, + "learning_rate": 2.997601228999956e-06, + "loss": 0.5597, + "num_input_tokens_seen": 2181344, + "step": 140 + }, + { + "epoch": 0.5948717948717949, + "grad_norm": 0.2451171875, + "learning_rate": 2.99742687843539e-06, + "loss": 0.5514, + "num_input_tokens_seen": 2256768, + "step": 145 + }, + { + "epoch": 0.6153846153846154, + "grad_norm": 0.2265625, + "learning_rate": 2.997246417427407e-06, + "loss": 0.53, + "num_input_tokens_seen": 2330144, + "step": 150 + }, + { + "epoch": 0.6358974358974359, + "grad_norm": 0.19921875, + "learning_rate": 2.9970598467124008e-06, + "loss": 0.5365, + "num_input_tokens_seen": 2402688, + "step": 155 + }, + { + "epoch": 0.6564102564102564, + "grad_norm": 0.1923828125, + "learning_rate": 2.9968671670516983e-06, + "loss": 0.5578, + "num_input_tokens_seen": 2476800, + "step": 160 + }, + { + "epoch": 0.676923076923077, + "grad_norm": 0.166015625, + "learning_rate": 2.9966683792315528e-06, + "loss": 0.5492, + "num_input_tokens_seen": 2565792, + "step": 165 + }, + { + "epoch": 0.6974358974358974, + "grad_norm": 0.1533203125, + "learning_rate": 2.9964634840631435e-06, + "loss": 0.5144, + "num_input_tokens_seen": 2637792, + "step": 170 + }, + { + "epoch": 0.717948717948718, + "grad_norm": 0.154296875, + "learning_rate": 2.9962524823825724e-06, + "loss": 0.5741, + "num_input_tokens_seen": 2718944, + "step": 175 + }, + { + "epoch": 0.7384615384615385, + "grad_norm": 0.1513671875, + "learning_rate": 2.9960353750508583e-06, + "loss": 0.5846, + "num_input_tokens_seen": 2802240, + "step": 180 + }, + { + "epoch": 0.7589743589743589, + "grad_norm": 0.1474609375, + "learning_rate": 2.995812162953936e-06, + "loss": 0.5834, + "num_input_tokens_seen": 2884672, + "step": 185 + }, + { + "epoch": 0.7794871794871795, + "grad_norm": 0.154296875, + "learning_rate": 2.9955828470026515e-06, + "loss": 0.5493, + "num_input_tokens_seen": 2953536, + "step": 190 + }, + { + "epoch": 0.8, + "grad_norm": 0.138671875, + "learning_rate": 2.9953474281327576e-06, + "loss": 0.5503, + "num_input_tokens_seen": 3026496, + "step": 195 + }, + { + "epoch": 0.8205128205128205, + "grad_norm": 0.244140625, + "learning_rate": 2.995105907304912e-06, + "loss": 0.5835, + "num_input_tokens_seen": 3105376, + "step": 200 + }, + { + "epoch": 0.841025641025641, + "grad_norm": 0.349609375, + "learning_rate": 2.9948582855046704e-06, + "loss": 0.5583, + "num_input_tokens_seen": 3179776, + "step": 205 + }, + { + "epoch": 0.8615384615384616, + "grad_norm": 0.1826171875, + "learning_rate": 2.9946045637424864e-06, + "loss": 0.6392, + "num_input_tokens_seen": 3262336, + "step": 210 + }, + { + "epoch": 0.882051282051282, + "grad_norm": 0.1640625, + "learning_rate": 2.994344743053704e-06, + "loss": 0.5658, + "num_input_tokens_seen": 3343904, + "step": 215 + }, + { + "epoch": 0.9025641025641026, + "grad_norm": 0.50390625, + "learning_rate": 2.9940788244985557e-06, + "loss": 0.5609, + "num_input_tokens_seen": 3419264, + "step": 220 + }, + { + "epoch": 0.9230769230769231, + "grad_norm": 0.458984375, + "learning_rate": 2.9938068091621556e-06, + "loss": 0.5305, + "num_input_tokens_seen": 3496032, + "step": 225 + }, + { + "epoch": 0.9435897435897436, + "grad_norm": 0.41015625, + "learning_rate": 2.9935286981544975e-06, + "loss": 0.5652, + "num_input_tokens_seen": 3570880, + "step": 230 + }, + { + "epoch": 0.9641025641025641, + "grad_norm": 0.7734375, + "learning_rate": 2.9932444926104495e-06, + "loss": 0.5348, + "num_input_tokens_seen": 3644096, + "step": 235 + }, + { + "epoch": 0.9846153846153847, + "grad_norm": 0.86328125, + "learning_rate": 2.992954193689749e-06, + "loss": 0.5344, + "num_input_tokens_seen": 3717088, + "step": 240 + }, + { + "epoch": 1.005128205128205, + "grad_norm": 0.435546875, + "learning_rate": 2.9926578025769978e-06, + "loss": 0.5977, + "num_input_tokens_seen": 3783648, + "step": 245 + }, + { + "epoch": 1.0256410256410255, + "grad_norm": 0.408203125, + "learning_rate": 2.992355320481658e-06, + "loss": 0.5304, + "num_input_tokens_seen": 3852160, + "step": 250 + }, + { + "epoch": 1.0461538461538462, + "grad_norm": 0.75390625, + "learning_rate": 2.9920467486380475e-06, + "loss": 0.5617, + "num_input_tokens_seen": 3924416, + "step": 255 + }, + { + "epoch": 1.0666666666666667, + "grad_norm": 0.46875, + "learning_rate": 2.991732088305333e-06, + "loss": 0.5367, + "num_input_tokens_seen": 4006432, + "step": 260 + }, + { + "epoch": 1.087179487179487, + "grad_norm": 0.5234375, + "learning_rate": 2.991411340767526e-06, + "loss": 0.5207, + "num_input_tokens_seen": 4081888, + "step": 265 + }, + { + "epoch": 1.1076923076923078, + "grad_norm": 0.41015625, + "learning_rate": 2.9910845073334793e-06, + "loss": 0.556, + "num_input_tokens_seen": 4155968, + "step": 270 + }, + { + "epoch": 1.1282051282051282, + "grad_norm": 0.421875, + "learning_rate": 2.9907515893368784e-06, + "loss": 0.537, + "num_input_tokens_seen": 4234272, + "step": 275 + }, + { + "epoch": 1.1487179487179486, + "grad_norm": 0.40234375, + "learning_rate": 2.9904125881362378e-06, + "loss": 0.5305, + "num_input_tokens_seen": 4302368, + "step": 280 + }, + { + "epoch": 1.1692307692307693, + "grad_norm": 0.5625, + "learning_rate": 2.990067505114896e-06, + "loss": 0.5337, + "num_input_tokens_seen": 4376640, + "step": 285 + }, + { + "epoch": 1.1897435897435897, + "grad_norm": 0.451171875, + "learning_rate": 2.9897163416810084e-06, + "loss": 0.592, + "num_input_tokens_seen": 4458208, + "step": 290 + }, + { + "epoch": 1.2102564102564102, + "grad_norm": 0.40625, + "learning_rate": 2.9893590992675427e-06, + "loss": 0.5808, + "num_input_tokens_seen": 4537920, + "step": 295 + }, + { + "epoch": 1.2307692307692308, + "grad_norm": 0.58984375, + "learning_rate": 2.988995779332273e-06, + "loss": 0.5569, + "num_input_tokens_seen": 4606880, + "step": 300 + }, + { + "epoch": 1.2512820512820513, + "grad_norm": 0.431640625, + "learning_rate": 2.9886263833577725e-06, + "loss": 0.5422, + "num_input_tokens_seen": 4682816, + "step": 305 + }, + { + "epoch": 1.2717948717948717, + "grad_norm": 0.5234375, + "learning_rate": 2.98825091285141e-06, + "loss": 0.5296, + "num_input_tokens_seen": 4756032, + "step": 310 + }, + { + "epoch": 1.2923076923076924, + "grad_norm": 0.4453125, + "learning_rate": 2.987869369345341e-06, + "loss": 0.5012, + "num_input_tokens_seen": 4827232, + "step": 315 + }, + { + "epoch": 1.3128205128205128, + "grad_norm": 0.392578125, + "learning_rate": 2.987481754396502e-06, + "loss": 0.5289, + "num_input_tokens_seen": 4902368, + "step": 320 + }, + { + "epoch": 1.3333333333333333, + "grad_norm": 0.326171875, + "learning_rate": 2.9870880695866067e-06, + "loss": 0.5245, + "num_input_tokens_seen": 4978080, + "step": 325 + }, + { + "epoch": 1.353846153846154, + "grad_norm": 0.62109375, + "learning_rate": 2.986688316522136e-06, + "loss": 0.5325, + "num_input_tokens_seen": 5047456, + "step": 330 + }, + { + "epoch": 1.3743589743589744, + "grad_norm": 0.390625, + "learning_rate": 2.9862824968343352e-06, + "loss": 0.5068, + "num_input_tokens_seen": 5118720, + "step": 335 + }, + { + "epoch": 1.3948717948717948, + "grad_norm": 0.345703125, + "learning_rate": 2.9858706121792036e-06, + "loss": 0.5165, + "num_input_tokens_seen": 5196288, + "step": 340 + }, + { + "epoch": 1.4153846153846155, + "grad_norm": 0.251953125, + "learning_rate": 2.985452664237488e-06, + "loss": 0.5025, + "num_input_tokens_seen": 5272480, + "step": 345 + }, + { + "epoch": 1.435897435897436, + "grad_norm": 0.302734375, + "learning_rate": 2.98502865471468e-06, + "loss": 0.5285, + "num_input_tokens_seen": 5343296, + "step": 350 + }, + { + "epoch": 1.4564102564102563, + "grad_norm": 0.314453125, + "learning_rate": 2.9845985853410053e-06, + "loss": 0.4983, + "num_input_tokens_seen": 5415904, + "step": 355 + }, + { + "epoch": 1.476923076923077, + "grad_norm": 0.236328125, + "learning_rate": 2.9841624578714167e-06, + "loss": 0.5789, + "num_input_tokens_seen": 5502176, + "step": 360 + }, + { + "epoch": 1.4974358974358974, + "grad_norm": 0.2060546875, + "learning_rate": 2.9837202740855897e-06, + "loss": 0.5394, + "num_input_tokens_seen": 5580352, + "step": 365 + }, + { + "epoch": 1.5179487179487179, + "grad_norm": 0.267578125, + "learning_rate": 2.9832720357879107e-06, + "loss": 0.5664, + "num_input_tokens_seen": 5662912, + "step": 370 + }, + { + "epoch": 1.5384615384615383, + "grad_norm": 0.22265625, + "learning_rate": 2.9828177448074753e-06, + "loss": 0.5546, + "num_input_tokens_seen": 5743776, + "step": 375 + }, + { + "epoch": 1.558974358974359, + "grad_norm": 0.20703125, + "learning_rate": 2.9823574029980757e-06, + "loss": 0.5412, + "num_input_tokens_seen": 5812384, + "step": 380 + }, + { + "epoch": 1.5794871794871796, + "grad_norm": 0.1630859375, + "learning_rate": 2.981891012238196e-06, + "loss": 0.5587, + "num_input_tokens_seen": 5892768, + "step": 385 + }, + { + "epoch": 1.6, + "grad_norm": 0.1640625, + "learning_rate": 2.9814185744310036e-06, + "loss": 0.535, + "num_input_tokens_seen": 5974592, + "step": 390 + }, + { + "epoch": 1.6205128205128205, + "grad_norm": 0.1533203125, + "learning_rate": 2.9809400915043424e-06, + "loss": 0.512, + "num_input_tokens_seen": 6050080, + "step": 395 + }, + { + "epoch": 1.641025641025641, + "grad_norm": 0.142578125, + "learning_rate": 2.9804555654107243e-06, + "loss": 0.5392, + "num_input_tokens_seen": 6137248, + "step": 400 + }, + { + "epoch": 1.6615384615384614, + "grad_norm": 0.1376953125, + "learning_rate": 2.9799649981273185e-06, + "loss": 0.6444, + "num_input_tokens_seen": 6232704, + "step": 405 + }, + { + "epoch": 1.682051282051282, + "grad_norm": 0.1435546875, + "learning_rate": 2.9794683916559493e-06, + "loss": 0.5202, + "num_input_tokens_seen": 6304064, + "step": 410 + }, + { + "epoch": 1.7025641025641025, + "grad_norm": 0.1455078125, + "learning_rate": 2.9789657480230842e-06, + "loss": 0.5344, + "num_input_tokens_seen": 6392928, + "step": 415 + }, + { + "epoch": 1.7230769230769232, + "grad_norm": 0.16015625, + "learning_rate": 2.9784570692798236e-06, + "loss": 0.4614, + "num_input_tokens_seen": 6473120, + "step": 420 + }, + { + "epoch": 1.7435897435897436, + "grad_norm": 0.19921875, + "learning_rate": 2.977942357501898e-06, + "loss": 0.5036, + "num_input_tokens_seen": 6545312, + "step": 425 + }, + { + "epoch": 1.764102564102564, + "grad_norm": 0.41796875, + "learning_rate": 2.977421614789655e-06, + "loss": 0.5308, + "num_input_tokens_seen": 6629984, + "step": 430 + }, + { + "epoch": 1.7846153846153845, + "grad_norm": 0.1767578125, + "learning_rate": 2.976894843268051e-06, + "loss": 0.5475, + "num_input_tokens_seen": 6715936, + "step": 435 + }, + { + "epoch": 1.8051282051282052, + "grad_norm": 0.150390625, + "learning_rate": 2.976362045086647e-06, + "loss": 0.5704, + "num_input_tokens_seen": 6797472, + "step": 440 + }, + { + "epoch": 1.8256410256410256, + "grad_norm": 0.416015625, + "learning_rate": 2.975823222419594e-06, + "loss": 0.4781, + "num_input_tokens_seen": 6867808, + "step": 445 + }, + { + "epoch": 1.8461538461538463, + "grad_norm": 0.25, + "learning_rate": 2.9752783774656267e-06, + "loss": 0.5247, + "num_input_tokens_seen": 6944480, + "step": 450 + }, + { + "epoch": 1.8666666666666667, + "grad_norm": 0.703125, + "learning_rate": 2.974727512448056e-06, + "loss": 0.5161, + "num_input_tokens_seen": 7024064, + "step": 455 + }, + { + "epoch": 1.8871794871794871, + "grad_norm": 0.65234375, + "learning_rate": 2.974170629614757e-06, + "loss": 0.5049, + "num_input_tokens_seen": 7103488, + "step": 460 + }, + { + "epoch": 1.9076923076923076, + "grad_norm": 0.734375, + "learning_rate": 2.9736077312381624e-06, + "loss": 0.5712, + "num_input_tokens_seen": 7190304, + "step": 465 + }, + { + "epoch": 1.9282051282051282, + "grad_norm": 0.40234375, + "learning_rate": 2.9730388196152513e-06, + "loss": 0.5222, + "num_input_tokens_seen": 7265056, + "step": 470 + }, + { + "epoch": 1.9487179487179487, + "grad_norm": 0.28125, + "learning_rate": 2.972463897067541e-06, + "loss": 0.4829, + "num_input_tokens_seen": 7336992, + "step": 475 + }, + { + "epoch": 1.9692307692307693, + "grad_norm": 0.56640625, + "learning_rate": 2.971882965941077e-06, + "loss": 0.5218, + "num_input_tokens_seen": 7413984, + "step": 480 + }, + { + "epoch": 1.9897435897435898, + "grad_norm": 0.357421875, + "learning_rate": 2.9712960286064237e-06, + "loss": 0.5543, + "num_input_tokens_seen": 7492768, + "step": 485 + }, + { + "epoch": 2.01025641025641, + "grad_norm": 0.26171875, + "learning_rate": 2.970703087458655e-06, + "loss": 0.5036, + "num_input_tokens_seen": 7563904, + "step": 490 + }, + { + "epoch": 2.0307692307692307, + "grad_norm": 0.38671875, + "learning_rate": 2.9701041449173426e-06, + "loss": 0.556, + "num_input_tokens_seen": 7634464, + "step": 495 + }, + { + "epoch": 2.051282051282051, + "grad_norm": 0.375, + "learning_rate": 2.96949920342655e-06, + "loss": 0.5742, + "num_input_tokens_seen": 7711168, + "step": 500 + }, + { + "epoch": 2.071794871794872, + "grad_norm": 0.365234375, + "learning_rate": 2.968888265454818e-06, + "loss": 0.4905, + "num_input_tokens_seen": 7786656, + "step": 505 + }, + { + "epoch": 2.0923076923076924, + "grad_norm": 0.271484375, + "learning_rate": 2.968271333495158e-06, + "loss": 0.5134, + "num_input_tokens_seen": 7858240, + "step": 510 + }, + { + "epoch": 2.112820512820513, + "grad_norm": 0.373046875, + "learning_rate": 2.967648410065041e-06, + "loss": 0.6211, + "num_input_tokens_seen": 7952064, + "step": 515 + }, + { + "epoch": 2.1333333333333333, + "grad_norm": 0.45703125, + "learning_rate": 2.9670194977063857e-06, + "loss": 0.515, + "num_input_tokens_seen": 8026688, + "step": 520 + }, + { + "epoch": 2.1538461538461537, + "grad_norm": 0.400390625, + "learning_rate": 2.96638459898555e-06, + "loss": 0.5464, + "num_input_tokens_seen": 8104064, + "step": 525 + }, + { + "epoch": 2.174358974358974, + "grad_norm": 0.421875, + "learning_rate": 2.9657437164933205e-06, + "loss": 0.5331, + "num_input_tokens_seen": 8187968, + "step": 530 + }, + { + "epoch": 2.194871794871795, + "grad_norm": 0.396484375, + "learning_rate": 2.9650968528449e-06, + "loss": 0.5236, + "num_input_tokens_seen": 8261312, + "step": 535 + }, + { + "epoch": 2.2153846153846155, + "grad_norm": 0.34765625, + "learning_rate": 2.9644440106799e-06, + "loss": 0.4345, + "num_input_tokens_seen": 8336032, + "step": 540 + }, + { + "epoch": 2.235897435897436, + "grad_norm": 0.353515625, + "learning_rate": 2.963785192662327e-06, + "loss": 0.4853, + "num_input_tokens_seen": 8405024, + "step": 545 + }, + { + "epoch": 2.2564102564102564, + "grad_norm": 0.365234375, + "learning_rate": 2.9631204014805716e-06, + "loss": 0.5, + "num_input_tokens_seen": 8483456, + "step": 550 + }, + { + "epoch": 2.276923076923077, + "grad_norm": 0.38671875, + "learning_rate": 2.9624496398474014e-06, + "loss": 0.4863, + "num_input_tokens_seen": 8560768, + "step": 555 + }, + { + "epoch": 2.2974358974358973, + "grad_norm": 0.32421875, + "learning_rate": 2.961772910499945e-06, + "loss": 0.48, + "num_input_tokens_seen": 8637888, + "step": 560 + }, + { + "epoch": 2.3179487179487177, + "grad_norm": 0.376953125, + "learning_rate": 2.9610902161996838e-06, + "loss": 0.5768, + "num_input_tokens_seen": 8715360, + "step": 565 + }, + { + "epoch": 2.3384615384615386, + "grad_norm": 0.265625, + "learning_rate": 2.96040155973244e-06, + "loss": 0.4751, + "num_input_tokens_seen": 8795200, + "step": 570 + }, + { + "epoch": 2.358974358974359, + "grad_norm": 0.283203125, + "learning_rate": 2.959706943908365e-06, + "loss": 0.5161, + "num_input_tokens_seen": 8883136, + "step": 575 + }, + { + "epoch": 2.3794871794871795, + "grad_norm": 0.271484375, + "learning_rate": 2.9590063715619287e-06, + "loss": 0.5588, + "num_input_tokens_seen": 8961568, + "step": 580 + }, + { + "epoch": 2.4, + "grad_norm": 0.2373046875, + "learning_rate": 2.9582998455519062e-06, + "loss": 0.5527, + "num_input_tokens_seen": 9043360, + "step": 585 + }, + { + "epoch": 2.4205128205128204, + "grad_norm": 0.21875, + "learning_rate": 2.9575873687613676e-06, + "loss": 0.4897, + "num_input_tokens_seen": 9116448, + "step": 590 + }, + { + "epoch": 2.4410256410256412, + "grad_norm": 0.21484375, + "learning_rate": 2.9568689440976676e-06, + "loss": 0.5359, + "num_input_tokens_seen": 9193120, + "step": 595 + }, + { + "epoch": 2.4615384615384617, + "grad_norm": 0.205078125, + "learning_rate": 2.95614457449243e-06, + "loss": 0.5763, + "num_input_tokens_seen": 9269920, + "step": 600 + }, + { + "epoch": 2.482051282051282, + "grad_norm": 0.1845703125, + "learning_rate": 2.9554142629015382e-06, + "loss": 0.4631, + "num_input_tokens_seen": 9339968, + "step": 605 + }, + { + "epoch": 2.5025641025641026, + "grad_norm": 0.1650390625, + "learning_rate": 2.954678012305123e-06, + "loss": 0.5349, + "num_input_tokens_seen": 9426976, + "step": 610 + }, + { + "epoch": 2.523076923076923, + "grad_norm": 0.1708984375, + "learning_rate": 2.9539358257075495e-06, + "loss": 0.5532, + "num_input_tokens_seen": 9509056, + "step": 615 + }, + { + "epoch": 2.5435897435897434, + "grad_norm": 0.1611328125, + "learning_rate": 2.9531877061374066e-06, + "loss": 0.4748, + "num_input_tokens_seen": 9590720, + "step": 620 + }, + { + "epoch": 2.564102564102564, + "grad_norm": 0.15234375, + "learning_rate": 2.9524336566474915e-06, + "loss": 0.5022, + "num_input_tokens_seen": 9667648, + "step": 625 + }, + { + "epoch": 2.5846153846153848, + "grad_norm": 0.1337890625, + "learning_rate": 2.9516736803148014e-06, + "loss": 0.5005, + "num_input_tokens_seen": 9738016, + "step": 630 + }, + { + "epoch": 2.605128205128205, + "grad_norm": 0.1337890625, + "learning_rate": 2.9509077802405174e-06, + "loss": 0.5297, + "num_input_tokens_seen": 9816224, + "step": 635 + }, + { + "epoch": 2.6256410256410256, + "grad_norm": 0.13671875, + "learning_rate": 2.9501359595499933e-06, + "loss": 0.5399, + "num_input_tokens_seen": 9891104, + "step": 640 + }, + { + "epoch": 2.646153846153846, + "grad_norm": 0.177734375, + "learning_rate": 2.9493582213927425e-06, + "loss": 0.4901, + "num_input_tokens_seen": 9969792, + "step": 645 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 0.142578125, + "learning_rate": 2.9485745689424267e-06, + "loss": 0.4591, + "num_input_tokens_seen": 10044608, + "step": 650 + }, + { + "epoch": 2.6871794871794874, + "grad_norm": 0.2451171875, + "learning_rate": 2.9477850053968405e-06, + "loss": 0.5729, + "num_input_tokens_seen": 10132640, + "step": 655 + }, + { + "epoch": 2.707692307692308, + "grad_norm": 0.150390625, + "learning_rate": 2.9469895339778995e-06, + "loss": 0.5405, + "num_input_tokens_seen": 10207968, + "step": 660 + }, + { + "epoch": 2.7282051282051283, + "grad_norm": 0.30859375, + "learning_rate": 2.946188157931627e-06, + "loss": 0.4786, + "num_input_tokens_seen": 10277408, + "step": 665 + }, + { + "epoch": 2.7487179487179487, + "grad_norm": 0.314453125, + "learning_rate": 2.9453808805281423e-06, + "loss": 0.5035, + "num_input_tokens_seen": 10349184, + "step": 670 + }, + { + "epoch": 2.769230769230769, + "grad_norm": 0.2490234375, + "learning_rate": 2.944567705061644e-06, + "loss": 0.4719, + "num_input_tokens_seen": 10434112, + "step": 675 + }, + { + "epoch": 2.7897435897435896, + "grad_norm": 0.65234375, + "learning_rate": 2.9437486348504e-06, + "loss": 0.5118, + "num_input_tokens_seen": 10506208, + "step": 680 + }, + { + "epoch": 2.81025641025641, + "grad_norm": 0.68359375, + "learning_rate": 2.9429236732367318e-06, + "loss": 0.5014, + "num_input_tokens_seen": 10577696, + "step": 685 + }, + { + "epoch": 2.830769230769231, + "grad_norm": 0.71484375, + "learning_rate": 2.942092823587001e-06, + "loss": 0.4827, + "num_input_tokens_seen": 10657984, + "step": 690 + }, + { + "epoch": 2.8512820512820514, + "grad_norm": 0.71484375, + "learning_rate": 2.941256089291597e-06, + "loss": 0.5177, + "num_input_tokens_seen": 10734688, + "step": 695 + }, + { + "epoch": 2.871794871794872, + "grad_norm": 0.6328125, + "learning_rate": 2.940413473764923e-06, + "loss": 0.4517, + "num_input_tokens_seen": 10812640, + "step": 700 + }, + { + "epoch": 2.8923076923076922, + "grad_norm": 0.70703125, + "learning_rate": 2.9395649804453786e-06, + "loss": 0.4574, + "num_input_tokens_seen": 10884800, + "step": 705 + }, + { + "epoch": 2.9128205128205127, + "grad_norm": 0.27734375, + "learning_rate": 2.9387106127953515e-06, + "loss": 0.5092, + "num_input_tokens_seen": 10962016, + "step": 710 + }, + { + "epoch": 2.9333333333333336, + "grad_norm": 0.34375, + "learning_rate": 2.937850374301198e-06, + "loss": 0.4888, + "num_input_tokens_seen": 11033280, + "step": 715 + }, + { + "epoch": 2.953846153846154, + "grad_norm": 0.375, + "learning_rate": 2.9369842684732336e-06, + "loss": 0.5447, + "num_input_tokens_seen": 11113696, + "step": 720 + }, + { + "epoch": 2.9743589743589745, + "grad_norm": 0.396484375, + "learning_rate": 2.936112298845713e-06, + "loss": 0.5438, + "num_input_tokens_seen": 11195104, + "step": 725 + }, + { + "epoch": 2.994871794871795, + "grad_norm": 0.2470703125, + "learning_rate": 2.935234468976822e-06, + "loss": 0.46, + "num_input_tokens_seen": 11270304, + "step": 730 + }, + { + "epoch": 3.0153846153846153, + "grad_norm": 0.392578125, + "learning_rate": 2.934350782448658e-06, + "loss": 0.569, + "num_input_tokens_seen": 11350784, + "step": 735 + }, + { + "epoch": 3.0358974358974358, + "grad_norm": 0.318359375, + "learning_rate": 2.9334612428672175e-06, + "loss": 0.5246, + "num_input_tokens_seen": 11429568, + "step": 740 + }, + { + "epoch": 3.056410256410256, + "grad_norm": 0.35546875, + "learning_rate": 2.9325658538623822e-06, + "loss": 0.4587, + "num_input_tokens_seen": 11502784, + "step": 745 + }, + { + "epoch": 3.076923076923077, + "grad_norm": 0.3671875, + "learning_rate": 2.931664619087902e-06, + "loss": 0.5095, + "num_input_tokens_seen": 11575680, + "step": 750 + }, + { + "epoch": 3.0974358974358975, + "grad_norm": 0.35546875, + "learning_rate": 2.9307575422213813e-06, + "loss": 0.4916, + "num_input_tokens_seen": 11649856, + "step": 755 + }, + { + "epoch": 3.117948717948718, + "grad_norm": 0.404296875, + "learning_rate": 2.929844626964265e-06, + "loss": 0.5647, + "num_input_tokens_seen": 11727616, + "step": 760 + }, + { + "epoch": 3.1384615384615384, + "grad_norm": 0.421875, + "learning_rate": 2.9289258770418208e-06, + "loss": 0.4448, + "num_input_tokens_seen": 11806208, + "step": 765 + }, + { + "epoch": 3.158974358974359, + "grad_norm": 0.39453125, + "learning_rate": 2.9280012962031263e-06, + "loss": 0.5086, + "num_input_tokens_seen": 11884096, + "step": 770 + }, + { + "epoch": 3.1794871794871793, + "grad_norm": 0.29296875, + "learning_rate": 2.9270708882210525e-06, + "loss": 0.4796, + "num_input_tokens_seen": 11956416, + "step": 775 + }, + { + "epoch": 3.2, + "grad_norm": 0.33203125, + "learning_rate": 2.92613465689225e-06, + "loss": 0.4797, + "num_input_tokens_seen": 12032384, + "step": 780 + }, + { + "epoch": 3.2205128205128206, + "grad_norm": 0.287109375, + "learning_rate": 2.92519260603713e-06, + "loss": 0.4523, + "num_input_tokens_seen": 12107328, + "step": 785 + }, + { + "epoch": 3.241025641025641, + "grad_norm": 0.326171875, + "learning_rate": 2.9242447394998545e-06, + "loss": 0.4795, + "num_input_tokens_seen": 12178848, + "step": 790 + }, + { + "epoch": 3.2615384615384615, + "grad_norm": 0.2734375, + "learning_rate": 2.923291061148314e-06, + "loss": 0.5164, + "num_input_tokens_seen": 12252160, + "step": 795 + }, + { + "epoch": 3.282051282051282, + "grad_norm": 0.271484375, + "learning_rate": 2.9223315748741146e-06, + "loss": 0.4949, + "num_input_tokens_seen": 12325120, + "step": 800 + }, + { + "epoch": 3.3025641025641024, + "grad_norm": 0.244140625, + "learning_rate": 2.9213662845925662e-06, + "loss": 0.4848, + "num_input_tokens_seen": 12398144, + "step": 805 + }, + { + "epoch": 3.3230769230769233, + "grad_norm": 0.2265625, + "learning_rate": 2.9203951942426586e-06, + "loss": 0.5114, + "num_input_tokens_seen": 12475008, + "step": 810 + }, + { + "epoch": 3.3435897435897437, + "grad_norm": 0.2109375, + "learning_rate": 2.9194183077870516e-06, + "loss": 0.6022, + "num_input_tokens_seen": 12562336, + "step": 815 + }, + { + "epoch": 3.364102564102564, + "grad_norm": 0.1923828125, + "learning_rate": 2.9184356292120562e-06, + "loss": 0.4922, + "num_input_tokens_seen": 12646560, + "step": 820 + }, + { + "epoch": 3.3846153846153846, + "grad_norm": 0.1962890625, + "learning_rate": 2.9174471625276198e-06, + "loss": 0.5707, + "num_input_tokens_seen": 12718848, + "step": 825 + }, + { + "epoch": 3.405128205128205, + "grad_norm": 0.1650390625, + "learning_rate": 2.916452911767307e-06, + "loss": 0.4784, + "num_input_tokens_seen": 12798240, + "step": 830 + }, + { + "epoch": 3.4256410256410255, + "grad_norm": 0.1552734375, + "learning_rate": 2.915452880988287e-06, + "loss": 0.4423, + "num_input_tokens_seen": 12872608, + "step": 835 + }, + { + "epoch": 3.4461538461538463, + "grad_norm": 0.1396484375, + "learning_rate": 2.914447074271314e-06, + "loss": 0.4809, + "num_input_tokens_seen": 12952896, + "step": 840 + }, + { + "epoch": 3.466666666666667, + "grad_norm": 0.1650390625, + "learning_rate": 2.913435495720712e-06, + "loss": 0.5316, + "num_input_tokens_seen": 13036768, + "step": 845 + }, + { + "epoch": 3.4871794871794872, + "grad_norm": 0.140625, + "learning_rate": 2.9124181494643574e-06, + "loss": 0.4592, + "num_input_tokens_seen": 13114784, + "step": 850 + }, + { + "epoch": 3.5076923076923077, + "grad_norm": 0.138671875, + "learning_rate": 2.911395039653663e-06, + "loss": 0.4878, + "num_input_tokens_seen": 13188448, + "step": 855 + }, + { + "epoch": 3.528205128205128, + "grad_norm": 0.1298828125, + "learning_rate": 2.9103661704635604e-06, + "loss": 0.5066, + "num_input_tokens_seen": 13262592, + "step": 860 + }, + { + "epoch": 3.5487179487179485, + "grad_norm": 0.11962890625, + "learning_rate": 2.909331546092483e-06, + "loss": 0.4649, + "num_input_tokens_seen": 13339936, + "step": 865 + }, + { + "epoch": 3.569230769230769, + "grad_norm": 0.14453125, + "learning_rate": 2.908291170762349e-06, + "loss": 0.5233, + "num_input_tokens_seen": 13416256, + "step": 870 + }, + { + "epoch": 3.58974358974359, + "grad_norm": 0.306640625, + "learning_rate": 2.9072450487185434e-06, + "loss": 0.5018, + "num_input_tokens_seen": 13487392, + "step": 875 + }, + { + "epoch": 3.6102564102564103, + "grad_norm": 0.337890625, + "learning_rate": 2.9061931842299026e-06, + "loss": 0.4602, + "num_input_tokens_seen": 13569984, + "step": 880 + }, + { + "epoch": 3.6307692307692307, + "grad_norm": 0.1640625, + "learning_rate": 2.9051355815886952e-06, + "loss": 0.5309, + "num_input_tokens_seen": 13650944, + "step": 885 + }, + { + "epoch": 3.651282051282051, + "grad_norm": 0.30859375, + "learning_rate": 2.904072245110605e-06, + "loss": 0.5186, + "num_input_tokens_seen": 13720736, + "step": 890 + }, + { + "epoch": 3.6717948717948716, + "grad_norm": 0.333984375, + "learning_rate": 2.9030031791347136e-06, + "loss": 0.4839, + "num_input_tokens_seen": 13791616, + "step": 895 + }, + { + "epoch": 3.6923076923076925, + "grad_norm": 0.25, + "learning_rate": 2.901928388023483e-06, + "loss": 0.5199, + "num_input_tokens_seen": 13867488, + "step": 900 + }, + { + "epoch": 3.712820512820513, + "grad_norm": 0.69921875, + "learning_rate": 2.900847876162736e-06, + "loss": 0.5414, + "num_input_tokens_seen": 13954848, + "step": 905 + }, + { + "epoch": 3.7333333333333334, + "grad_norm": 0.80859375, + "learning_rate": 2.899761647961641e-06, + "loss": 0.5451, + "num_input_tokens_seen": 14037792, + "step": 910 + }, + { + "epoch": 3.753846153846154, + "grad_norm": 1.0, + "learning_rate": 2.898669707852692e-06, + "loss": 0.4448, + "num_input_tokens_seen": 14107520, + "step": 915 + }, + { + "epoch": 3.7743589743589743, + "grad_norm": 0.3125, + "learning_rate": 2.897572060291692e-06, + "loss": 0.5213, + "num_input_tokens_seen": 14193888, + "step": 920 + }, + { + "epoch": 3.7948717948717947, + "grad_norm": 0.5, + "learning_rate": 2.896468709757733e-06, + "loss": 0.4968, + "num_input_tokens_seen": 14270976, + "step": 925 + }, + { + "epoch": 3.815384615384615, + "grad_norm": 0.58203125, + "learning_rate": 2.8953596607531788e-06, + "loss": 0.5769, + "num_input_tokens_seen": 14351232, + "step": 930 + }, + { + "epoch": 3.835897435897436, + "grad_norm": 0.26171875, + "learning_rate": 2.894244917803647e-06, + "loss": 0.4925, + "num_input_tokens_seen": 14426912, + "step": 935 + }, + { + "epoch": 3.8564102564102565, + "grad_norm": 0.333984375, + "learning_rate": 2.8931244854579904e-06, + "loss": 0.481, + "num_input_tokens_seen": 14515776, + "step": 940 + }, + { + "epoch": 3.876923076923077, + "grad_norm": 0.322265625, + "learning_rate": 2.891998368288277e-06, + "loss": 0.4699, + "num_input_tokens_seen": 14587104, + "step": 945 + }, + { + "epoch": 3.8974358974358974, + "grad_norm": 0.341796875, + "learning_rate": 2.890866570889773e-06, + "loss": 0.5206, + "num_input_tokens_seen": 14663680, + "step": 950 + }, + { + "epoch": 3.917948717948718, + "grad_norm": 0.35546875, + "learning_rate": 2.8897290978809245e-06, + "loss": 0.5117, + "num_input_tokens_seen": 14747360, + "step": 955 + }, + { + "epoch": 3.9384615384615387, + "grad_norm": 0.3671875, + "learning_rate": 2.888585953903336e-06, + "loss": 0.4891, + "num_input_tokens_seen": 14826464, + "step": 960 + }, + { + "epoch": 3.958974358974359, + "grad_norm": 0.376953125, + "learning_rate": 2.8874371436217534e-06, + "loss": 0.4943, + "num_input_tokens_seen": 14916416, + "step": 965 + }, + { + "epoch": 3.9794871794871796, + "grad_norm": 0.32421875, + "learning_rate": 2.8862826717240464e-06, + "loss": 0.5222, + "num_input_tokens_seen": 14995072, + "step": 970 + }, + { + "epoch": 4.0, + "grad_norm": 0.337890625, + "learning_rate": 2.8851225429211855e-06, + "loss": 0.5197, + "num_input_tokens_seen": 15070304, + "step": 975 + }, + { + "epoch": 4.02051282051282, + "grad_norm": 0.380859375, + "learning_rate": 2.883956761947226e-06, + "loss": 0.501, + "num_input_tokens_seen": 15152480, + "step": 980 + }, + { + "epoch": 4.041025641025641, + "grad_norm": 0.478515625, + "learning_rate": 2.8827853335592876e-06, + "loss": 0.5142, + "num_input_tokens_seen": 15229184, + "step": 985 + }, + { + "epoch": 4.061538461538461, + "grad_norm": 0.376953125, + "learning_rate": 2.8816082625375353e-06, + "loss": 0.5239, + "num_input_tokens_seen": 15311072, + "step": 990 + }, + { + "epoch": 4.082051282051282, + "grad_norm": 0.380859375, + "learning_rate": 2.8804255536851584e-06, + "loss": 0.4823, + "num_input_tokens_seen": 15383232, + "step": 995 + }, + { + "epoch": 4.102564102564102, + "grad_norm": 0.34375, + "learning_rate": 2.8792372118283528e-06, + "loss": 0.5416, + "num_input_tokens_seen": 15464064, + "step": 1000 + }, + { + "epoch": 4.123076923076923, + "grad_norm": 0.416015625, + "learning_rate": 2.878043241816301e-06, + "loss": 0.4889, + "num_input_tokens_seen": 15536480, + "step": 1005 + }, + { + "epoch": 4.143589743589744, + "grad_norm": 0.3828125, + "learning_rate": 2.876843648521152e-06, + "loss": 0.5338, + "num_input_tokens_seen": 15618816, + "step": 1010 + }, + { + "epoch": 4.164102564102564, + "grad_norm": 0.2578125, + "learning_rate": 2.8756384368380003e-06, + "loss": 0.5101, + "num_input_tokens_seen": 15694304, + "step": 1015 + }, + { + "epoch": 4.184615384615385, + "grad_norm": 0.25390625, + "learning_rate": 2.874427611684867e-06, + "loss": 0.4792, + "num_input_tokens_seen": 15770784, + "step": 1020 + }, + { + "epoch": 4.205128205128205, + "grad_norm": 0.27734375, + "learning_rate": 2.8732111780026813e-06, + "loss": 0.4959, + "num_input_tokens_seen": 15837312, + "step": 1025 + }, + { + "epoch": 4.225641025641026, + "grad_norm": 0.22265625, + "learning_rate": 2.871989140755257e-06, + "loss": 0.5227, + "num_input_tokens_seen": 15908096, + "step": 1030 + }, + { + "epoch": 4.246153846153846, + "grad_norm": 0.2392578125, + "learning_rate": 2.870761504929275e-06, + "loss": 0.473, + "num_input_tokens_seen": 15982720, + "step": 1035 + }, + { + "epoch": 4.266666666666667, + "grad_norm": 0.1943359375, + "learning_rate": 2.869528275534261e-06, + "loss": 0.4911, + "num_input_tokens_seen": 16056256, + "step": 1040 + }, + { + "epoch": 4.287179487179487, + "grad_norm": 0.1787109375, + "learning_rate": 2.8682894576025677e-06, + "loss": 0.4678, + "num_input_tokens_seen": 16128256, + "step": 1045 + }, + { + "epoch": 4.3076923076923075, + "grad_norm": 0.1650390625, + "learning_rate": 2.8670450561893498e-06, + "loss": 0.4534, + "num_input_tokens_seen": 16203808, + "step": 1050 + }, + { + "epoch": 4.328205128205128, + "grad_norm": 0.1796875, + "learning_rate": 2.865795076372549e-06, + "loss": 0.5788, + "num_input_tokens_seen": 16290464, + "step": 1055 + }, + { + "epoch": 4.348717948717948, + "grad_norm": 0.1845703125, + "learning_rate": 2.8645395232528683e-06, + "loss": 0.4744, + "num_input_tokens_seen": 16362688, + "step": 1060 + }, + { + "epoch": 4.36923076923077, + "grad_norm": 0.146484375, + "learning_rate": 2.863278401953754e-06, + "loss": 0.4801, + "num_input_tokens_seen": 16434976, + "step": 1065 + }, + { + "epoch": 4.38974358974359, + "grad_norm": 0.134765625, + "learning_rate": 2.862011717621375e-06, + "loss": 0.5035, + "num_input_tokens_seen": 16514880, + "step": 1070 + }, + { + "epoch": 4.410256410256411, + "grad_norm": 0.15625, + "learning_rate": 2.860739475424599e-06, + "loss": 0.5456, + "num_input_tokens_seen": 16592544, + "step": 1075 + }, + { + "epoch": 4.430769230769231, + "grad_norm": 0.125, + "learning_rate": 2.859461680554975e-06, + "loss": 0.4773, + "num_input_tokens_seen": 16675360, + "step": 1080 + }, + { + "epoch": 4.4512820512820515, + "grad_norm": 0.1103515625, + "learning_rate": 2.858178338226709e-06, + "loss": 0.4793, + "num_input_tokens_seen": 16753728, + "step": 1085 + }, + { + "epoch": 4.471794871794872, + "grad_norm": 0.11572265625, + "learning_rate": 2.8568894536766462e-06, + "loss": 0.4698, + "num_input_tokens_seen": 16829696, + "step": 1090 + }, + { + "epoch": 4.492307692307692, + "grad_norm": 0.19140625, + "learning_rate": 2.8555950321642444e-06, + "loss": 0.4648, + "num_input_tokens_seen": 16908128, + "step": 1095 + }, + { + "epoch": 4.512820512820513, + "grad_norm": 0.2333984375, + "learning_rate": 2.8542950789715587e-06, + "loss": 0.4473, + "num_input_tokens_seen": 16979136, + "step": 1100 + }, + { + "epoch": 4.533333333333333, + "grad_norm": 0.216796875, + "learning_rate": 2.8529895994032153e-06, + "loss": 0.5128, + "num_input_tokens_seen": 17063296, + "step": 1105 + }, + { + "epoch": 4.553846153846154, + "grad_norm": 0.15234375, + "learning_rate": 2.851678598786392e-06, + "loss": 0.4834, + "num_input_tokens_seen": 17145984, + "step": 1110 + }, + { + "epoch": 4.574358974358974, + "grad_norm": 0.296875, + "learning_rate": 2.8503620824707946e-06, + "loss": 0.4581, + "num_input_tokens_seen": 17221152, + "step": 1115 + }, + { + "epoch": 4.5948717948717945, + "grad_norm": 0.267578125, + "learning_rate": 2.8490400558286395e-06, + "loss": 0.5189, + "num_input_tokens_seen": 17298688, + "step": 1120 + }, + { + "epoch": 4.615384615384615, + "grad_norm": 0.302734375, + "learning_rate": 2.847712524254626e-06, + "loss": 0.4739, + "num_input_tokens_seen": 17374240, + "step": 1125 + }, + { + "epoch": 4.635897435897435, + "grad_norm": 0.6640625, + "learning_rate": 2.846379493165918e-06, + "loss": 0.5376, + "num_input_tokens_seen": 17451360, + "step": 1130 + }, + { + "epoch": 4.656410256410257, + "grad_norm": 0.82421875, + "learning_rate": 2.8450409680021204e-06, + "loss": 0.5152, + "num_input_tokens_seen": 17535776, + "step": 1135 + }, + { + "epoch": 4.676923076923077, + "grad_norm": 0.90234375, + "learning_rate": 2.8436969542252576e-06, + "loss": 0.4957, + "num_input_tokens_seen": 17616256, + "step": 1140 + }, + { + "epoch": 4.697435897435898, + "grad_norm": 0.31640625, + "learning_rate": 2.842347457319752e-06, + "loss": 0.4963, + "num_input_tokens_seen": 17691616, + "step": 1145 + }, + { + "epoch": 4.717948717948718, + "grad_norm": 0.5078125, + "learning_rate": 2.8409924827923985e-06, + "loss": 0.4868, + "num_input_tokens_seen": 17767264, + "step": 1150 + }, + { + "epoch": 4.7384615384615385, + "grad_norm": 0.36328125, + "learning_rate": 2.839632036172346e-06, + "loss": 0.5595, + "num_input_tokens_seen": 17847648, + "step": 1155 + }, + { + "epoch": 4.758974358974359, + "grad_norm": 0.296875, + "learning_rate": 2.8382661230110716e-06, + "loss": 0.6248, + "num_input_tokens_seen": 17932992, + "step": 1160 + }, + { + "epoch": 4.779487179487179, + "grad_norm": 0.326171875, + "learning_rate": 2.8368947488823613e-06, + "loss": 0.4935, + "num_input_tokens_seen": 18008544, + "step": 1165 + }, + { + "epoch": 4.8, + "grad_norm": 0.3203125, + "learning_rate": 2.8355179193822834e-06, + "loss": 0.472, + "num_input_tokens_seen": 18091904, + "step": 1170 + }, + { + "epoch": 4.82051282051282, + "grad_norm": 0.396484375, + "learning_rate": 2.834135640129168e-06, + "loss": 0.4618, + "num_input_tokens_seen": 18169760, + "step": 1175 + }, + { + "epoch": 4.841025641025641, + "grad_norm": 0.2890625, + "learning_rate": 2.8327479167635834e-06, + "loss": 0.4375, + "num_input_tokens_seen": 18242016, + "step": 1180 + }, + { + "epoch": 4.861538461538462, + "grad_norm": 0.314453125, + "learning_rate": 2.831354754948315e-06, + "loss": 0.4954, + "num_input_tokens_seen": 18316192, + "step": 1185 + }, + { + "epoch": 4.8820512820512825, + "grad_norm": 0.330078125, + "learning_rate": 2.829956160368338e-06, + "loss": 0.4885, + "num_input_tokens_seen": 18402720, + "step": 1190 + }, + { + "epoch": 4.902564102564103, + "grad_norm": 0.482421875, + "learning_rate": 2.828552138730798e-06, + "loss": 0.4452, + "num_input_tokens_seen": 18472768, + "step": 1195 + }, + { + "epoch": 4.923076923076923, + "grad_norm": 0.33984375, + "learning_rate": 2.8271426957649868e-06, + "loss": 0.4602, + "num_input_tokens_seen": 18548128, + "step": 1200 + }, + { + "epoch": 4.943589743589744, + "grad_norm": 0.412109375, + "learning_rate": 2.8257278372223177e-06, + "loss": 0.4391, + "num_input_tokens_seen": 18622112, + "step": 1205 + }, + { + "epoch": 4.964102564102564, + "grad_norm": 0.484375, + "learning_rate": 2.824307568876304e-06, + "loss": 0.4614, + "num_input_tokens_seen": 18704288, + "step": 1210 + }, + { + "epoch": 4.984615384615385, + "grad_norm": 0.35546875, + "learning_rate": 2.8228818965225326e-06, + "loss": 0.5284, + "num_input_tokens_seen": 18780128, + "step": 1215 + }, + { + "epoch": 5.005128205128205, + "grad_norm": 0.326171875, + "learning_rate": 2.8214508259786443e-06, + "loss": 0.5213, + "num_input_tokens_seen": 18850496, + "step": 1220 + }, + { + "epoch": 5.0256410256410255, + "grad_norm": 0.333984375, + "learning_rate": 2.820014363084307e-06, + "loss": 0.5071, + "num_input_tokens_seen": 18926816, + "step": 1225 + }, + { + "epoch": 5.046153846153846, + "grad_norm": 0.341796875, + "learning_rate": 2.8185725137011922e-06, + "loss": 0.4964, + "num_input_tokens_seen": 19002624, + "step": 1230 + }, + { + "epoch": 5.066666666666666, + "grad_norm": 0.3359375, + "learning_rate": 2.8171252837129523e-06, + "loss": 0.5196, + "num_input_tokens_seen": 19083296, + "step": 1235 + }, + { + "epoch": 5.087179487179487, + "grad_norm": 0.28125, + "learning_rate": 2.815672679025196e-06, + "loss": 0.5272, + "num_input_tokens_seen": 19158048, + "step": 1240 + }, + { + "epoch": 5.107692307692307, + "grad_norm": 0.26171875, + "learning_rate": 2.814214705565464e-06, + "loss": 0.5034, + "num_input_tokens_seen": 19233888, + "step": 1245 + }, + { + "epoch": 5.128205128205128, + "grad_norm": 0.2890625, + "learning_rate": 2.8127513692832047e-06, + "loss": 0.5069, + "num_input_tokens_seen": 19317472, + "step": 1250 + }, + { + "epoch": 5.148717948717949, + "grad_norm": 0.2265625, + "learning_rate": 2.8112826761497507e-06, + "loss": 0.5116, + "num_input_tokens_seen": 19398496, + "step": 1255 + }, + { + "epoch": 5.1692307692307695, + "grad_norm": 0.21484375, + "learning_rate": 2.8098086321582937e-06, + "loss": 0.4286, + "num_input_tokens_seen": 19466400, + "step": 1260 + }, + { + "epoch": 5.18974358974359, + "grad_norm": 0.2119140625, + "learning_rate": 2.8083292433238602e-06, + "loss": 0.5058, + "num_input_tokens_seen": 19550336, + "step": 1265 + }, + { + "epoch": 5.21025641025641, + "grad_norm": 0.2177734375, + "learning_rate": 2.8068445156832864e-06, + "loss": 0.4587, + "num_input_tokens_seen": 19625792, + "step": 1270 + }, + { + "epoch": 5.230769230769231, + "grad_norm": 0.1943359375, + "learning_rate": 2.805354455295196e-06, + "loss": 0.4901, + "num_input_tokens_seen": 19711776, + "step": 1275 + }, + { + "epoch": 5.251282051282051, + "grad_norm": 0.173828125, + "learning_rate": 2.8038590682399718e-06, + "loss": 0.5474, + "num_input_tokens_seen": 19794208, + "step": 1280 + }, + { + "epoch": 5.271794871794872, + "grad_norm": 0.1611328125, + "learning_rate": 2.8023583606197336e-06, + "loss": 0.4452, + "num_input_tokens_seen": 19872128, + "step": 1285 + }, + { + "epoch": 5.292307692307692, + "grad_norm": 0.1357421875, + "learning_rate": 2.800852338558312e-06, + "loss": 0.5081, + "num_input_tokens_seen": 19953856, + "step": 1290 + }, + { + "epoch": 5.312820512820513, + "grad_norm": 0.1376953125, + "learning_rate": 2.7993410082012247e-06, + "loss": 0.4863, + "num_input_tokens_seen": 20026848, + "step": 1295 + }, + { + "epoch": 5.333333333333333, + "grad_norm": 0.138671875, + "learning_rate": 2.79782437571565e-06, + "loss": 0.4979, + "num_input_tokens_seen": 20100928, + "step": 1300 + }, + { + "epoch": 5.3538461538461535, + "grad_norm": 0.1318359375, + "learning_rate": 2.7963024472904013e-06, + "loss": 0.4676, + "num_input_tokens_seen": 20173504, + "step": 1305 + }, + { + "epoch": 5.374358974358975, + "grad_norm": 0.12451171875, + "learning_rate": 2.7947752291359053e-06, + "loss": 0.4256, + "num_input_tokens_seen": 20244608, + "step": 1310 + }, + { + "epoch": 5.394871794871795, + "grad_norm": 0.130859375, + "learning_rate": 2.7932427274841715e-06, + "loss": 0.4576, + "num_input_tokens_seen": 20324992, + "step": 1315 + }, + { + "epoch": 5.415384615384616, + "grad_norm": 0.1376953125, + "learning_rate": 2.7917049485887705e-06, + "loss": 0.5155, + "num_input_tokens_seen": 20402304, + "step": 1320 + }, + { + "epoch": 5.435897435897436, + "grad_norm": 0.20703125, + "learning_rate": 2.790161898724808e-06, + "loss": 0.4304, + "num_input_tokens_seen": 20480800, + "step": 1325 + }, + { + "epoch": 5.456410256410257, + "grad_norm": 0.138671875, + "learning_rate": 2.7886135841888973e-06, + "loss": 0.4759, + "num_input_tokens_seen": 20560096, + "step": 1330 + }, + { + "epoch": 5.476923076923077, + "grad_norm": 0.15234375, + "learning_rate": 2.787060011299135e-06, + "loss": 0.4974, + "num_input_tokens_seen": 20645216, + "step": 1335 + }, + { + "epoch": 5.4974358974358974, + "grad_norm": 0.341796875, + "learning_rate": 2.785501186395077e-06, + "loss": 0.5174, + "num_input_tokens_seen": 20738688, + "step": 1340 + }, + { + "epoch": 5.517948717948718, + "grad_norm": 0.271484375, + "learning_rate": 2.7839371158377077e-06, + "loss": 0.5272, + "num_input_tokens_seen": 20812928, + "step": 1345 + }, + { + "epoch": 5.538461538461538, + "grad_norm": 0.296875, + "learning_rate": 2.78236780600942e-06, + "loss": 0.5129, + "num_input_tokens_seen": 20890592, + "step": 1350 + }, + { + "epoch": 5.558974358974359, + "grad_norm": 0.57421875, + "learning_rate": 2.780793263313984e-06, + "loss": 0.484, + "num_input_tokens_seen": 20961984, + "step": 1355 + }, + { + "epoch": 5.579487179487179, + "grad_norm": 0.8671875, + "learning_rate": 2.7792134941765247e-06, + "loss": 0.4793, + "num_input_tokens_seen": 21030784, + "step": 1360 + }, + { + "epoch": 5.6, + "grad_norm": 0.267578125, + "learning_rate": 2.7776285050434937e-06, + "loss": 0.4521, + "num_input_tokens_seen": 21108960, + "step": 1365 + }, + { + "epoch": 5.62051282051282, + "grad_norm": 0.267578125, + "learning_rate": 2.7760383023826425e-06, + "loss": 0.4192, + "num_input_tokens_seen": 21181728, + "step": 1370 + }, + { + "epoch": 5.641025641025641, + "grad_norm": 0.64453125, + "learning_rate": 2.7744428926829993e-06, + "loss": 0.5131, + "num_input_tokens_seen": 21255328, + "step": 1375 + }, + { + "epoch": 5.661538461538462, + "grad_norm": 0.271484375, + "learning_rate": 2.7728422824548387e-06, + "loss": 0.483, + "num_input_tokens_seen": 21324064, + "step": 1380 + }, + { + "epoch": 5.682051282051282, + "grad_norm": 0.341796875, + "learning_rate": 2.7712364782296567e-06, + "loss": 0.4858, + "num_input_tokens_seen": 21399040, + "step": 1385 + }, + { + "epoch": 5.702564102564103, + "grad_norm": 0.318359375, + "learning_rate": 2.769625486560145e-06, + "loss": 0.4629, + "num_input_tokens_seen": 21472640, + "step": 1390 + }, + { + "epoch": 5.723076923076923, + "grad_norm": 0.39453125, + "learning_rate": 2.7680093140201625e-06, + "loss": 0.5023, + "num_input_tokens_seen": 21544448, + "step": 1395 + }, + { + "epoch": 5.743589743589744, + "grad_norm": 0.28515625, + "learning_rate": 2.766387967204709e-06, + "loss": 0.4903, + "num_input_tokens_seen": 21611136, + "step": 1400 + }, + { + "epoch": 5.764102564102564, + "grad_norm": 0.2578125, + "learning_rate": 2.7647614527299007e-06, + "loss": 0.5558, + "num_input_tokens_seen": 21703040, + "step": 1405 + }, + { + "epoch": 5.7846153846153845, + "grad_norm": 0.45703125, + "learning_rate": 2.763129777232938e-06, + "loss": 0.5612, + "num_input_tokens_seen": 21784096, + "step": 1410 + }, + { + "epoch": 5.805128205128205, + "grad_norm": 0.333984375, + "learning_rate": 2.7614929473720847e-06, + "loss": 0.4683, + "num_input_tokens_seen": 21855072, + "step": 1415 + }, + { + "epoch": 5.825641025641025, + "grad_norm": 0.38671875, + "learning_rate": 2.7598509698266346e-06, + "loss": 0.5171, + "num_input_tokens_seen": 21933312, + "step": 1420 + }, + { + "epoch": 5.846153846153846, + "grad_norm": 0.359375, + "learning_rate": 2.758203851296889e-06, + "loss": 0.5214, + "num_input_tokens_seen": 22019008, + "step": 1425 + }, + { + "epoch": 5.866666666666667, + "grad_norm": 0.326171875, + "learning_rate": 2.756551598504128e-06, + "loss": 0.4975, + "num_input_tokens_seen": 22092864, + "step": 1430 + }, + { + "epoch": 5.887179487179488, + "grad_norm": 0.392578125, + "learning_rate": 2.7548942181905816e-06, + "loss": 0.4853, + "num_input_tokens_seen": 22171584, + "step": 1435 + }, + { + "epoch": 5.907692307692308, + "grad_norm": 0.375, + "learning_rate": 2.753231717119405e-06, + "loss": 0.483, + "num_input_tokens_seen": 22241376, + "step": 1440 + }, + { + "epoch": 5.9282051282051285, + "grad_norm": 0.373046875, + "learning_rate": 2.751564102074646e-06, + "loss": 0.4965, + "num_input_tokens_seen": 22313664, + "step": 1445 + }, + { + "epoch": 5.948717948717949, + "grad_norm": 0.34765625, + "learning_rate": 2.749891379861225e-06, + "loss": 0.5342, + "num_input_tokens_seen": 22397408, + "step": 1450 + }, + { + "epoch": 5.969230769230769, + "grad_norm": 0.3359375, + "learning_rate": 2.748213557304899e-06, + "loss": 0.4233, + "num_input_tokens_seen": 22473664, + "step": 1455 + }, + { + "epoch": 5.98974358974359, + "grad_norm": 0.349609375, + "learning_rate": 2.74653064125224e-06, + "loss": 0.5244, + "num_input_tokens_seen": 22553760, + "step": 1460 + }, + { + "epoch": 6.01025641025641, + "grad_norm": 0.298828125, + "learning_rate": 2.7448426385706036e-06, + "loss": 0.5211, + "num_input_tokens_seen": 22624608, + "step": 1465 + }, + { + "epoch": 6.030769230769231, + "grad_norm": 0.271484375, + "learning_rate": 2.7431495561481027e-06, + "loss": 0.5618, + "num_input_tokens_seen": 22710048, + "step": 1470 + }, + { + "epoch": 6.051282051282051, + "grad_norm": 0.2734375, + "learning_rate": 2.741451400893578e-06, + "loss": 0.5172, + "num_input_tokens_seen": 22787392, + "step": 1475 + }, + { + "epoch": 6.0717948717948715, + "grad_norm": 0.22265625, + "learning_rate": 2.739748179736571e-06, + "loss": 0.5035, + "num_input_tokens_seen": 22865120, + "step": 1480 + }, + { + "epoch": 6.092307692307692, + "grad_norm": 0.205078125, + "learning_rate": 2.7380398996272955e-06, + "loss": 0.519, + "num_input_tokens_seen": 22952832, + "step": 1485 + }, + { + "epoch": 6.112820512820512, + "grad_norm": 0.2138671875, + "learning_rate": 2.736326567536609e-06, + "loss": 0.4438, + "num_input_tokens_seen": 23028544, + "step": 1490 + }, + { + "epoch": 6.133333333333334, + "grad_norm": 0.2021484375, + "learning_rate": 2.7346081904559827e-06, + "loss": 0.4669, + "num_input_tokens_seen": 23100096, + "step": 1495 + }, + { + "epoch": 6.153846153846154, + "grad_norm": 0.1533203125, + "learning_rate": 2.732884775397477e-06, + "loss": 0.4702, + "num_input_tokens_seen": 23183392, + "step": 1500 + }, + { + "epoch": 6.174358974358975, + "grad_norm": 0.1533203125, + "learning_rate": 2.731156329393709e-06, + "loss": 0.5031, + "num_input_tokens_seen": 23266208, + "step": 1505 + }, + { + "epoch": 6.194871794871795, + "grad_norm": 0.1484375, + "learning_rate": 2.729422859497825e-06, + "loss": 0.5005, + "num_input_tokens_seen": 23348064, + "step": 1510 + }, + { + "epoch": 6.2153846153846155, + "grad_norm": 0.1552734375, + "learning_rate": 2.7276843727834727e-06, + "loss": 0.4798, + "num_input_tokens_seen": 23420128, + "step": 1515 + }, + { + "epoch": 6.235897435897436, + "grad_norm": 0.1337890625, + "learning_rate": 2.725940876344771e-06, + "loss": 0.5059, + "num_input_tokens_seen": 23497056, + "step": 1520 + }, + { + "epoch": 6.256410256410256, + "grad_norm": 0.1376953125, + "learning_rate": 2.7241923772962823e-06, + "loss": 0.4582, + "num_input_tokens_seen": 23564928, + "step": 1525 + }, + { + "epoch": 6.276923076923077, + "grad_norm": 0.1220703125, + "learning_rate": 2.722438882772982e-06, + "loss": 0.4295, + "num_input_tokens_seen": 23646624, + "step": 1530 + }, + { + "epoch": 6.297435897435897, + "grad_norm": 0.1484375, + "learning_rate": 2.720680399930231e-06, + "loss": 0.4682, + "num_input_tokens_seen": 23716960, + "step": 1535 + }, + { + "epoch": 6.317948717948718, + "grad_norm": 0.1611328125, + "learning_rate": 2.7189169359437443e-06, + "loss": 0.4944, + "num_input_tokens_seen": 23796032, + "step": 1540 + }, + { + "epoch": 6.338461538461538, + "grad_norm": 0.1318359375, + "learning_rate": 2.7171484980095653e-06, + "loss": 0.4405, + "num_input_tokens_seen": 23868768, + "step": 1545 + }, + { + "epoch": 6.358974358974359, + "grad_norm": 0.1953125, + "learning_rate": 2.715375093344032e-06, + "loss": 0.4742, + "num_input_tokens_seen": 23937824, + "step": 1550 + }, + { + "epoch": 6.37948717948718, + "grad_norm": 0.146484375, + "learning_rate": 2.713596729183751e-06, + "loss": 0.4654, + "num_input_tokens_seen": 24009472, + "step": 1555 + }, + { + "epoch": 6.4, + "grad_norm": 0.146484375, + "learning_rate": 2.7118134127855667e-06, + "loss": 0.4686, + "num_input_tokens_seen": 24096256, + "step": 1560 + }, + { + "epoch": 6.420512820512821, + "grad_norm": 0.3203125, + "learning_rate": 2.7100251514265317e-06, + "loss": 0.5152, + "num_input_tokens_seen": 24180640, + "step": 1565 + }, + { + "epoch": 6.441025641025641, + "grad_norm": 0.392578125, + "learning_rate": 2.7082319524038764e-06, + "loss": 0.4762, + "num_input_tokens_seen": 24251296, + "step": 1570 + }, + { + "epoch": 6.461538461538462, + "grad_norm": 0.74609375, + "learning_rate": 2.706433823034981e-06, + "loss": 0.5113, + "num_input_tokens_seen": 24329760, + "step": 1575 + }, + { + "epoch": 6.482051282051282, + "grad_norm": 0.55078125, + "learning_rate": 2.7046307706573445e-06, + "loss": 0.4942, + "num_input_tokens_seen": 24420896, + "step": 1580 + }, + { + "epoch": 6.5025641025641026, + "grad_norm": 0.703125, + "learning_rate": 2.702822802628554e-06, + "loss": 0.475, + "num_input_tokens_seen": 24495360, + "step": 1585 + }, + { + "epoch": 6.523076923076923, + "grad_norm": 0.2216796875, + "learning_rate": 2.701009926326256e-06, + "loss": 0.4801, + "num_input_tokens_seen": 24572192, + "step": 1590 + }, + { + "epoch": 6.543589743589743, + "grad_norm": 0.267578125, + "learning_rate": 2.6991921491481267e-06, + "loss": 0.4776, + "num_input_tokens_seen": 24647552, + "step": 1595 + }, + { + "epoch": 6.564102564102564, + "grad_norm": 0.578125, + "learning_rate": 2.6973694785118394e-06, + "loss": 0.4878, + "num_input_tokens_seen": 24719136, + "step": 1600 + }, + { + "epoch": 6.584615384615384, + "grad_norm": 0.37109375, + "learning_rate": 2.695541921855037e-06, + "loss": 0.5138, + "num_input_tokens_seen": 24800320, + "step": 1605 + }, + { + "epoch": 6.605128205128205, + "grad_norm": 0.310546875, + "learning_rate": 2.6937094866353006e-06, + "loss": 0.4782, + "num_input_tokens_seen": 24877088, + "step": 1610 + }, + { + "epoch": 6.625641025641025, + "grad_norm": 0.498046875, + "learning_rate": 2.6918721803301174e-06, + "loss": 0.5043, + "num_input_tokens_seen": 24954272, + "step": 1615 + }, + { + "epoch": 6.6461538461538465, + "grad_norm": 0.294921875, + "learning_rate": 2.690030010436853e-06, + "loss": 0.4237, + "num_input_tokens_seen": 25023744, + "step": 1620 + }, + { + "epoch": 6.666666666666667, + "grad_norm": 0.64453125, + "learning_rate": 2.688182984472719e-06, + "loss": 0.5302, + "num_input_tokens_seen": 25105664, + "step": 1625 + }, + { + "epoch": 6.687179487179487, + "grad_norm": 0.3515625, + "learning_rate": 2.686331109974743e-06, + "loss": 0.4991, + "num_input_tokens_seen": 25183680, + "step": 1630 + }, + { + "epoch": 6.707692307692308, + "grad_norm": 0.443359375, + "learning_rate": 2.684474394499738e-06, + "loss": 0.5142, + "num_input_tokens_seen": 25265920, + "step": 1635 + }, + { + "epoch": 6.728205128205128, + "grad_norm": 0.4609375, + "learning_rate": 2.6826128456242708e-06, + "loss": 0.4651, + "num_input_tokens_seen": 25343648, + "step": 1640 + }, + { + "epoch": 6.748717948717949, + "grad_norm": 0.361328125, + "learning_rate": 2.680746470944631e-06, + "loss": 0.5633, + "num_input_tokens_seen": 25418176, + "step": 1645 + }, + { + "epoch": 6.769230769230769, + "grad_norm": 0.375, + "learning_rate": 2.6788752780768007e-06, + "loss": 0.5124, + "num_input_tokens_seen": 25504832, + "step": 1650 + }, + { + "epoch": 6.78974358974359, + "grad_norm": 0.62109375, + "learning_rate": 2.6769992746564256e-06, + "loss": 0.5046, + "num_input_tokens_seen": 25582112, + "step": 1655 + }, + { + "epoch": 6.81025641025641, + "grad_norm": 0.35546875, + "learning_rate": 2.6751184683387777e-06, + "loss": 0.484, + "num_input_tokens_seen": 25656992, + "step": 1660 + }, + { + "epoch": 6.8307692307692305, + "grad_norm": 0.3203125, + "learning_rate": 2.67323286679873e-06, + "loss": 0.4526, + "num_input_tokens_seen": 25729600, + "step": 1665 + }, + { + "epoch": 6.851282051282051, + "grad_norm": 0.365234375, + "learning_rate": 2.671342477730723e-06, + "loss": 0.4563, + "num_input_tokens_seen": 25801536, + "step": 1670 + }, + { + "epoch": 6.871794871794872, + "grad_norm": 0.380859375, + "learning_rate": 2.6694473088487324e-06, + "loss": 0.4951, + "num_input_tokens_seen": 25882912, + "step": 1675 + }, + { + "epoch": 6.892307692307693, + "grad_norm": 0.32421875, + "learning_rate": 2.6675473678862403e-06, + "loss": 0.5223, + "num_input_tokens_seen": 25957952, + "step": 1680 + }, + { + "epoch": 6.912820512820513, + "grad_norm": 0.353515625, + "learning_rate": 2.6656426625961993e-06, + "loss": 0.5471, + "num_input_tokens_seen": 26034432, + "step": 1685 + }, + { + "epoch": 6.933333333333334, + "grad_norm": 0.251953125, + "learning_rate": 2.6637332007510063e-06, + "loss": 0.4252, + "num_input_tokens_seen": 26106656, + "step": 1690 + }, + { + "epoch": 6.953846153846154, + "grad_norm": 0.31640625, + "learning_rate": 2.661818990142465e-06, + "loss": 0.5269, + "num_input_tokens_seen": 26186976, + "step": 1695 + }, + { + "epoch": 6.9743589743589745, + "grad_norm": 0.25, + "learning_rate": 2.65990003858176e-06, + "loss": 0.4487, + "num_input_tokens_seen": 26259264, + "step": 1700 + }, + { + "epoch": 6.994871794871795, + "grad_norm": 0.2353515625, + "learning_rate": 2.6579763538994197e-06, + "loss": 0.4705, + "num_input_tokens_seen": 26333248, + "step": 1705 + }, + { + "epoch": 7.015384615384615, + "grad_norm": 0.205078125, + "learning_rate": 2.656047943945287e-06, + "loss": 0.4443, + "num_input_tokens_seen": 26404832, + "step": 1710 + }, + { + "epoch": 7.035897435897436, + "grad_norm": 0.1865234375, + "learning_rate": 2.6541148165884885e-06, + "loss": 0.4615, + "num_input_tokens_seen": 26484608, + "step": 1715 + }, + { + "epoch": 7.056410256410256, + "grad_norm": 0.1943359375, + "learning_rate": 2.652176979717399e-06, + "loss": 0.5042, + "num_input_tokens_seen": 26556224, + "step": 1720 + }, + { + "epoch": 7.076923076923077, + "grad_norm": 0.16796875, + "learning_rate": 2.6502344412396116e-06, + "loss": 0.4601, + "num_input_tokens_seen": 26629632, + "step": 1725 + }, + { + "epoch": 7.097435897435897, + "grad_norm": 0.146484375, + "learning_rate": 2.6482872090819053e-06, + "loss": 0.534, + "num_input_tokens_seen": 26709568, + "step": 1730 + }, + { + "epoch": 7.1179487179487175, + "grad_norm": 0.201171875, + "learning_rate": 2.646335291190211e-06, + "loss": 0.4875, + "num_input_tokens_seen": 26785728, + "step": 1735 + }, + { + "epoch": 7.138461538461539, + "grad_norm": 0.150390625, + "learning_rate": 2.6443786955295827e-06, + "loss": 0.5223, + "num_input_tokens_seen": 26865024, + "step": 1740 + }, + { + "epoch": 7.158974358974359, + "grad_norm": 0.140625, + "learning_rate": 2.6424174300841606e-06, + "loss": 0.4365, + "num_input_tokens_seen": 26934720, + "step": 1745 + }, + { + "epoch": 7.17948717948718, + "grad_norm": 0.138671875, + "learning_rate": 2.6404515028571406e-06, + "loss": 0.4951, + "num_input_tokens_seen": 27008192, + "step": 1750 + }, + { + "epoch": 7.2, + "grad_norm": 0.11962890625, + "learning_rate": 2.638480921870743e-06, + "loss": 0.5132, + "num_input_tokens_seen": 27092000, + "step": 1755 + }, + { + "epoch": 7.220512820512821, + "grad_norm": 0.1298828125, + "learning_rate": 2.636505695166177e-06, + "loss": 0.4713, + "num_input_tokens_seen": 27172160, + "step": 1760 + }, + { + "epoch": 7.241025641025641, + "grad_norm": 0.16015625, + "learning_rate": 2.63452583080361e-06, + "loss": 0.479, + "num_input_tokens_seen": 27255712, + "step": 1765 + }, + { + "epoch": 7.2615384615384615, + "grad_norm": 0.1396484375, + "learning_rate": 2.6325413368621337e-06, + "loss": 0.4967, + "num_input_tokens_seen": 27343136, + "step": 1770 + }, + { + "epoch": 7.282051282051282, + "grad_norm": 0.2470703125, + "learning_rate": 2.630552221439732e-06, + "loss": 0.4843, + "num_input_tokens_seen": 27417312, + "step": 1775 + }, + { + "epoch": 7.302564102564102, + "grad_norm": 0.1396484375, + "learning_rate": 2.6285584926532465e-06, + "loss": 0.4738, + "num_input_tokens_seen": 27505824, + "step": 1780 + }, + { + "epoch": 7.323076923076923, + "grad_norm": 0.2890625, + "learning_rate": 2.626560158638344e-06, + "loss": 0.5716, + "num_input_tokens_seen": 27583776, + "step": 1785 + }, + { + "epoch": 7.343589743589743, + "grad_norm": 0.357421875, + "learning_rate": 2.6245572275494845e-06, + "loss": 0.515, + "num_input_tokens_seen": 27658912, + "step": 1790 + }, + { + "epoch": 7.364102564102564, + "grad_norm": 0.2314453125, + "learning_rate": 2.6225497075598865e-06, + "loss": 0.47, + "num_input_tokens_seen": 27733472, + "step": 1795 + }, + { + "epoch": 7.384615384615385, + "grad_norm": 0.51953125, + "learning_rate": 2.6205376068614943e-06, + "loss": 0.4749, + "num_input_tokens_seen": 27812160, + "step": 1800 + }, + { + "epoch": 7.4051282051282055, + "grad_norm": 0.55078125, + "learning_rate": 2.6185209336649438e-06, + "loss": 0.4727, + "num_input_tokens_seen": 27885024, + "step": 1805 + }, + { + "epoch": 7.425641025641026, + "grad_norm": 0.75, + "learning_rate": 2.61649969619953e-06, + "loss": 0.478, + "num_input_tokens_seen": 27956480, + "step": 1810 + }, + { + "epoch": 7.446153846153846, + "grad_norm": 0.28125, + "learning_rate": 2.614473902713173e-06, + "loss": 0.4778, + "num_input_tokens_seen": 28028032, + "step": 1815 + }, + { + "epoch": 7.466666666666667, + "grad_norm": 0.49609375, + "learning_rate": 2.612443561472385e-06, + "loss": 0.4443, + "num_input_tokens_seen": 28112992, + "step": 1820 + }, + { + "epoch": 7.487179487179487, + "grad_norm": 0.5, + "learning_rate": 2.610408680762234e-06, + "loss": 0.5186, + "num_input_tokens_seen": 28191520, + "step": 1825 + }, + { + "epoch": 7.507692307692308, + "grad_norm": 0.314453125, + "learning_rate": 2.6083692688863135e-06, + "loss": 0.5152, + "num_input_tokens_seen": 28277440, + "step": 1830 + }, + { + "epoch": 7.528205128205128, + "grad_norm": 0.29296875, + "learning_rate": 2.6063253341667064e-06, + "loss": 0.5173, + "num_input_tokens_seen": 28357440, + "step": 1835 + }, + { + "epoch": 7.5487179487179485, + "grad_norm": 0.326171875, + "learning_rate": 2.604276884943953e-06, + "loss": 0.4585, + "num_input_tokens_seen": 28426656, + "step": 1840 + }, + { + "epoch": 7.569230769230769, + "grad_norm": 0.279296875, + "learning_rate": 2.602223929577013e-06, + "loss": 0.4611, + "num_input_tokens_seen": 28499968, + "step": 1845 + }, + { + "epoch": 7.589743589743589, + "grad_norm": 0.310546875, + "learning_rate": 2.6001664764432363e-06, + "loss": 0.4929, + "num_input_tokens_seen": 28573664, + "step": 1850 + }, + { + "epoch": 7.61025641025641, + "grad_norm": 0.318359375, + "learning_rate": 2.5981045339383244e-06, + "loss": 0.5018, + "num_input_tokens_seen": 28658144, + "step": 1855 + }, + { + "epoch": 7.63076923076923, + "grad_norm": 0.328125, + "learning_rate": 2.596038110476301e-06, + "loss": 0.483, + "num_input_tokens_seen": 28730944, + "step": 1860 + }, + { + "epoch": 7.651282051282052, + "grad_norm": 0.52734375, + "learning_rate": 2.593967214489473e-06, + "loss": 0.5111, + "num_input_tokens_seen": 28816384, + "step": 1865 + }, + { + "epoch": 7.671794871794872, + "grad_norm": 0.37109375, + "learning_rate": 2.591891854428398e-06, + "loss": 0.4689, + "num_input_tokens_seen": 28891616, + "step": 1870 + }, + { + "epoch": 7.6923076923076925, + "grad_norm": 0.33203125, + "learning_rate": 2.5898120387618507e-06, + "loss": 0.4917, + "num_input_tokens_seen": 28970400, + "step": 1875 + }, + { + "epoch": 7.712820512820513, + "grad_norm": 0.33203125, + "learning_rate": 2.587727775976787e-06, + "loss": 0.4956, + "num_input_tokens_seen": 29051520, + "step": 1880 + }, + { + "epoch": 7.733333333333333, + "grad_norm": 0.3046875, + "learning_rate": 2.585639074578309e-06, + "loss": 0.438, + "num_input_tokens_seen": 29128544, + "step": 1885 + }, + { + "epoch": 7.753846153846154, + "grad_norm": 0.306640625, + "learning_rate": 2.5835459430896333e-06, + "loss": 0.4644, + "num_input_tokens_seen": 29210496, + "step": 1890 + }, + { + "epoch": 7.774358974358974, + "grad_norm": 0.365234375, + "learning_rate": 2.5814483900520522e-06, + "loss": 0.4901, + "num_input_tokens_seen": 29282400, + "step": 1895 + }, + { + "epoch": 7.794871794871795, + "grad_norm": 0.4140625, + "learning_rate": 2.5793464240249014e-06, + "loss": 0.4879, + "num_input_tokens_seen": 29352256, + "step": 1900 + }, + { + "epoch": 7.815384615384615, + "grad_norm": 0.330078125, + "learning_rate": 2.5772400535855242e-06, + "loss": 0.4552, + "num_input_tokens_seen": 29426336, + "step": 1905 + }, + { + "epoch": 7.835897435897436, + "grad_norm": 0.353515625, + "learning_rate": 2.575129287329237e-06, + "loss": 0.5417, + "num_input_tokens_seen": 29512224, + "step": 1910 + }, + { + "epoch": 7.856410256410256, + "grad_norm": 0.27734375, + "learning_rate": 2.5730141338692926e-06, + "loss": 0.4637, + "num_input_tokens_seen": 29590112, + "step": 1915 + }, + { + "epoch": 7.876923076923077, + "grad_norm": 0.31640625, + "learning_rate": 2.5708946018368487e-06, + "loss": 0.4486, + "num_input_tokens_seen": 29672608, + "step": 1920 + }, + { + "epoch": 7.897435897435898, + "grad_norm": 0.228515625, + "learning_rate": 2.568770699880928e-06, + "loss": 0.5094, + "num_input_tokens_seen": 29755520, + "step": 1925 + }, + { + "epoch": 7.917948717948718, + "grad_norm": 0.2197265625, + "learning_rate": 2.566642436668387e-06, + "loss": 0.5111, + "num_input_tokens_seen": 29833344, + "step": 1930 + }, + { + "epoch": 7.938461538461539, + "grad_norm": 0.1923828125, + "learning_rate": 2.5645098208838774e-06, + "loss": 0.4737, + "num_input_tokens_seen": 29904800, + "step": 1935 + }, + { + "epoch": 7.958974358974359, + "grad_norm": 0.2080078125, + "learning_rate": 2.562372861229813e-06, + "loss": 0.4384, + "num_input_tokens_seen": 29975488, + "step": 1940 + }, + { + "epoch": 7.97948717948718, + "grad_norm": 0.197265625, + "learning_rate": 2.5602315664263337e-06, + "loss": 0.4383, + "num_input_tokens_seen": 30046496, + "step": 1945 + }, + { + "epoch": 8.0, + "grad_norm": 0.1611328125, + "learning_rate": 2.5580859452112685e-06, + "loss": 0.4782, + "num_input_tokens_seen": 30119840, + "step": 1950 + }, + { + "epoch": 8.02051282051282, + "grad_norm": 0.1484375, + "learning_rate": 2.555936006340101e-06, + "loss": 0.5371, + "num_input_tokens_seen": 30207040, + "step": 1955 + }, + { + "epoch": 8.04102564102564, + "grad_norm": 0.1376953125, + "learning_rate": 2.553781758585935e-06, + "loss": 0.4867, + "num_input_tokens_seen": 30283968, + "step": 1960 + }, + { + "epoch": 8.061538461538461, + "grad_norm": 0.1376953125, + "learning_rate": 2.551623210739455e-06, + "loss": 0.4309, + "num_input_tokens_seen": 30355552, + "step": 1965 + }, + { + "epoch": 8.082051282051282, + "grad_norm": 0.12109375, + "learning_rate": 2.549460371608895e-06, + "loss": 0.5087, + "num_input_tokens_seen": 30435776, + "step": 1970 + }, + { + "epoch": 8.102564102564102, + "grad_norm": 0.1328125, + "learning_rate": 2.5472932500199976e-06, + "loss": 0.4746, + "num_input_tokens_seen": 30507616, + "step": 1975 + }, + { + "epoch": 8.123076923076923, + "grad_norm": 0.1396484375, + "learning_rate": 2.5451218548159823e-06, + "loss": 0.4833, + "num_input_tokens_seen": 30583456, + "step": 1980 + }, + { + "epoch": 8.143589743589743, + "grad_norm": 0.12890625, + "learning_rate": 2.5429461948575077e-06, + "loss": 0.4849, + "num_input_tokens_seen": 30654176, + "step": 1985 + }, + { + "epoch": 8.164102564102564, + "grad_norm": 0.1416015625, + "learning_rate": 2.540766279022634e-06, + "loss": 0.4812, + "num_input_tokens_seen": 30721920, + "step": 1990 + }, + { + "epoch": 8.184615384615384, + "grad_norm": 0.216796875, + "learning_rate": 2.53858211620679e-06, + "loss": 0.4976, + "num_input_tokens_seen": 30791264, + "step": 1995 + }, + { + "epoch": 8.205128205128204, + "grad_norm": 0.20703125, + "learning_rate": 2.536393715322732e-06, + "loss": 0.4556, + "num_input_tokens_seen": 30862336, + "step": 2000 + }, + { + "epoch": 8.225641025641025, + "grad_norm": 0.150390625, + "learning_rate": 2.5342010853005127e-06, + "loss": 0.4496, + "num_input_tokens_seen": 30940064, + "step": 2005 + }, + { + "epoch": 8.246153846153845, + "grad_norm": 0.318359375, + "learning_rate": 2.532004235087441e-06, + "loss": 0.4722, + "num_input_tokens_seen": 31013248, + "step": 2010 + }, + { + "epoch": 8.266666666666667, + "grad_norm": 0.26171875, + "learning_rate": 2.529803173648049e-06, + "loss": 0.4875, + "num_input_tokens_seen": 31094496, + "step": 2015 + }, + { + "epoch": 8.287179487179488, + "grad_norm": 0.2578125, + "learning_rate": 2.52759790996405e-06, + "loss": 0.4598, + "num_input_tokens_seen": 31171680, + "step": 2020 + }, + { + "epoch": 8.307692307692308, + "grad_norm": 0.77734375, + "learning_rate": 2.525388453034307e-06, + "loss": 0.5069, + "num_input_tokens_seen": 31252064, + "step": 2025 + }, + { + "epoch": 8.328205128205129, + "grad_norm": 0.6875, + "learning_rate": 2.5231748118747945e-06, + "loss": 0.5155, + "num_input_tokens_seen": 31329696, + "step": 2030 + }, + { + "epoch": 8.34871794871795, + "grad_norm": 0.96484375, + "learning_rate": 2.5209569955185604e-06, + "loss": 0.5436, + "num_input_tokens_seen": 31407648, + "step": 2035 + }, + { + "epoch": 8.36923076923077, + "grad_norm": 0.30078125, + "learning_rate": 2.51873501301569e-06, + "loss": 0.4953, + "num_input_tokens_seen": 31475200, + "step": 2040 + }, + { + "epoch": 8.38974358974359, + "grad_norm": 0.6328125, + "learning_rate": 2.5165088734332695e-06, + "loss": 0.4804, + "num_input_tokens_seen": 31547104, + "step": 2045 + }, + { + "epoch": 8.41025641025641, + "grad_norm": 0.52734375, + "learning_rate": 2.5142785858553486e-06, + "loss": 0.5533, + "num_input_tokens_seen": 31629440, + "step": 2050 + }, + { + "epoch": 8.430769230769231, + "grad_norm": 0.310546875, + "learning_rate": 2.512044159382903e-06, + "loss": 0.541, + "num_input_tokens_seen": 31713024, + "step": 2055 + }, + { + "epoch": 8.451282051282051, + "grad_norm": 0.71484375, + "learning_rate": 2.5098056031337975e-06, + "loss": 0.4444, + "num_input_tokens_seen": 31790432, + "step": 2060 + }, + { + "epoch": 8.471794871794872, + "grad_norm": 0.3046875, + "learning_rate": 2.5075629262427507e-06, + "loss": 0.4869, + "num_input_tokens_seen": 31870592, + "step": 2065 + }, + { + "epoch": 8.492307692307692, + "grad_norm": 0.6015625, + "learning_rate": 2.505316137861294e-06, + "loss": 0.4855, + "num_input_tokens_seen": 31945344, + "step": 2070 + }, + { + "epoch": 8.512820512820513, + "grad_norm": 0.283203125, + "learning_rate": 2.503065247157737e-06, + "loss": 0.5027, + "num_input_tokens_seen": 32030016, + "step": 2075 + }, + { + "epoch": 8.533333333333333, + "grad_norm": 0.3515625, + "learning_rate": 2.500810263317129e-06, + "loss": 0.4885, + "num_input_tokens_seen": 32108160, + "step": 2080 + }, + { + "epoch": 8.553846153846154, + "grad_norm": 0.33203125, + "learning_rate": 2.4985511955412238e-06, + "loss": 0.4451, + "num_input_tokens_seen": 32188288, + "step": 2085 + }, + { + "epoch": 8.574358974358974, + "grad_norm": 0.35546875, + "learning_rate": 2.4962880530484375e-06, + "loss": 0.4899, + "num_input_tokens_seen": 32266656, + "step": 2090 + }, + { + "epoch": 8.594871794871795, + "grad_norm": 0.3515625, + "learning_rate": 2.4940208450738146e-06, + "loss": 0.5083, + "num_input_tokens_seen": 32356544, + "step": 2095 + }, + { + "epoch": 8.615384615384615, + "grad_norm": 0.37890625, + "learning_rate": 2.49174958086899e-06, + "loss": 0.4839, + "num_input_tokens_seen": 32434720, + "step": 2100 + }, + { + "epoch": 8.635897435897435, + "grad_norm": 0.373046875, + "learning_rate": 2.48947426970215e-06, + "loss": 0.437, + "num_input_tokens_seen": 32507712, + "step": 2105 + }, + { + "epoch": 8.656410256410256, + "grad_norm": 0.361328125, + "learning_rate": 2.487194920857995e-06, + "loss": 0.435, + "num_input_tokens_seen": 32577216, + "step": 2110 + }, + { + "epoch": 8.676923076923076, + "grad_norm": 0.3359375, + "learning_rate": 2.484911543637702e-06, + "loss": 0.4768, + "num_input_tokens_seen": 32647552, + "step": 2115 + }, + { + "epoch": 8.697435897435897, + "grad_norm": 0.34765625, + "learning_rate": 2.4826241473588855e-06, + "loss": 0.4578, + "num_input_tokens_seen": 32727520, + "step": 2120 + }, + { + "epoch": 8.717948717948717, + "grad_norm": 0.365234375, + "learning_rate": 2.4803327413555623e-06, + "loss": 0.5142, + "num_input_tokens_seen": 32805440, + "step": 2125 + }, + { + "epoch": 8.73846153846154, + "grad_norm": 0.35546875, + "learning_rate": 2.4780373349781083e-06, + "loss": 0.4013, + "num_input_tokens_seen": 32880480, + "step": 2130 + }, + { + "epoch": 8.75897435897436, + "grad_norm": 0.30078125, + "learning_rate": 2.4757379375932265e-06, + "loss": 0.4616, + "num_input_tokens_seen": 32951936, + "step": 2135 + }, + { + "epoch": 8.77948717948718, + "grad_norm": 0.287109375, + "learning_rate": 2.473434558583903e-06, + "loss": 0.4791, + "num_input_tokens_seen": 33027104, + "step": 2140 + }, + { + "epoch": 8.8, + "grad_norm": 0.2451171875, + "learning_rate": 2.4711272073493745e-06, + "loss": 0.5163, + "num_input_tokens_seen": 33111040, + "step": 2145 + }, + { + "epoch": 8.820512820512821, + "grad_norm": 0.2470703125, + "learning_rate": 2.468815893305084e-06, + "loss": 0.4761, + "num_input_tokens_seen": 33202304, + "step": 2150 + }, + { + "epoch": 8.841025641025642, + "grad_norm": 0.2275390625, + "learning_rate": 2.466500625882646e-06, + "loss": 0.4405, + "num_input_tokens_seen": 33281376, + "step": 2155 + }, + { + "epoch": 8.861538461538462, + "grad_norm": 0.21484375, + "learning_rate": 2.464181414529809e-06, + "loss": 0.4538, + "num_input_tokens_seen": 33352640, + "step": 2160 + }, + { + "epoch": 8.882051282051282, + "grad_norm": 0.19140625, + "learning_rate": 2.4618582687104132e-06, + "loss": 0.4598, + "num_input_tokens_seen": 33423232, + "step": 2165 + }, + { + "epoch": 8.902564102564103, + "grad_norm": 0.169921875, + "learning_rate": 2.4595311979043545e-06, + "loss": 0.4556, + "num_input_tokens_seen": 33503744, + "step": 2170 + }, + { + "epoch": 8.923076923076923, + "grad_norm": 0.16796875, + "learning_rate": 2.4572002116075454e-06, + "loss": 0.4665, + "num_input_tokens_seen": 33576800, + "step": 2175 + }, + { + "epoch": 8.943589743589744, + "grad_norm": 0.1748046875, + "learning_rate": 2.454865319331876e-06, + "loss": 0.4683, + "num_input_tokens_seen": 33661120, + "step": 2180 + }, + { + "epoch": 8.964102564102564, + "grad_norm": 0.1484375, + "learning_rate": 2.4525265306051755e-06, + "loss": 0.5183, + "num_input_tokens_seen": 33733568, + "step": 2185 + }, + { + "epoch": 8.984615384615385, + "grad_norm": 0.15234375, + "learning_rate": 2.4501838549711723e-06, + "loss": 0.4671, + "num_input_tokens_seen": 33802592, + "step": 2190 + }, + { + "epoch": 9.005128205128205, + "grad_norm": 0.126953125, + "learning_rate": 2.447837301989457e-06, + "loss": 0.4858, + "num_input_tokens_seen": 33882272, + "step": 2195 + }, + { + "epoch": 9.025641025641026, + "grad_norm": 0.138671875, + "learning_rate": 2.4454868812354403e-06, + "loss": 0.4574, + "num_input_tokens_seen": 33953920, + "step": 2200 + }, + { + "epoch": 9.046153846153846, + "grad_norm": 0.11669921875, + "learning_rate": 2.4431326023003188e-06, + "loss": 0.4419, + "num_input_tokens_seen": 34027552, + "step": 2205 + }, + { + "epoch": 9.066666666666666, + "grad_norm": 0.11669921875, + "learning_rate": 2.44077447479103e-06, + "loss": 0.5126, + "num_input_tokens_seen": 34112480, + "step": 2210 + }, + { + "epoch": 9.087179487179487, + "grad_norm": 0.1357421875, + "learning_rate": 2.4384125083302178e-06, + "loss": 0.4517, + "num_input_tokens_seen": 34183840, + "step": 2215 + }, + { + "epoch": 9.107692307692307, + "grad_norm": 0.251953125, + "learning_rate": 2.4360467125561907e-06, + "loss": 0.5161, + "num_input_tokens_seen": 34258912, + "step": 2220 + }, + { + "epoch": 9.128205128205128, + "grad_norm": 0.240234375, + "learning_rate": 2.433677097122883e-06, + "loss": 0.4486, + "num_input_tokens_seen": 34329248, + "step": 2225 + }, + { + "epoch": 9.148717948717948, + "grad_norm": 0.166015625, + "learning_rate": 2.4313036716998154e-06, + "loss": 0.5191, + "num_input_tokens_seen": 34411232, + "step": 2230 + }, + { + "epoch": 9.169230769230769, + "grad_norm": 0.3125, + "learning_rate": 2.428926445972058e-06, + "loss": 0.5117, + "num_input_tokens_seen": 34486144, + "step": 2235 + }, + { + "epoch": 9.189743589743589, + "grad_norm": 0.2734375, + "learning_rate": 2.4265454296401857e-06, + "loss": 0.4739, + "num_input_tokens_seen": 34564864, + "step": 2240 + }, + { + "epoch": 9.21025641025641, + "grad_norm": 0.404296875, + "learning_rate": 2.4241606324202426e-06, + "loss": 0.4468, + "num_input_tokens_seen": 34640704, + "step": 2245 + }, + { + "epoch": 9.23076923076923, + "grad_norm": 0.53125, + "learning_rate": 2.4217720640437015e-06, + "loss": 0.457, + "num_input_tokens_seen": 34715744, + "step": 2250 + }, + { + "epoch": 9.25128205128205, + "grad_norm": 0.64453125, + "learning_rate": 2.4193797342574235e-06, + "loss": 0.4915, + "num_input_tokens_seen": 34798144, + "step": 2255 + }, + { + "epoch": 9.271794871794873, + "grad_norm": 0.71875, + "learning_rate": 2.4169836528236187e-06, + "loss": 0.4417, + "num_input_tokens_seen": 34873440, + "step": 2260 + }, + { + "epoch": 9.292307692307693, + "grad_norm": 0.26953125, + "learning_rate": 2.4145838295198066e-06, + "loss": 0.4999, + "num_input_tokens_seen": 34951552, + "step": 2265 + }, + { + "epoch": 9.312820512820513, + "grad_norm": 1.203125, + "learning_rate": 2.4121802741387743e-06, + "loss": 0.453, + "num_input_tokens_seen": 35021184, + "step": 2270 + }, + { + "epoch": 9.333333333333334, + "grad_norm": 0.298828125, + "learning_rate": 2.4097729964885407e-06, + "loss": 0.4473, + "num_input_tokens_seen": 35098080, + "step": 2275 + }, + { + "epoch": 9.353846153846154, + "grad_norm": 0.53515625, + "learning_rate": 2.4073620063923123e-06, + "loss": 0.4749, + "num_input_tokens_seen": 35170336, + "step": 2280 + }, + { + "epoch": 9.374358974358975, + "grad_norm": 0.314453125, + "learning_rate": 2.404947313688445e-06, + "loss": 0.5229, + "num_input_tokens_seen": 35255200, + "step": 2285 + }, + { + "epoch": 9.394871794871795, + "grad_norm": 0.4140625, + "learning_rate": 2.4025289282304037e-06, + "loss": 0.5158, + "num_input_tokens_seen": 35328928, + "step": 2290 + }, + { + "epoch": 9.415384615384616, + "grad_norm": 0.34375, + "learning_rate": 2.4001068598867216e-06, + "loss": 0.4548, + "num_input_tokens_seen": 35402976, + "step": 2295 + }, + { + "epoch": 9.435897435897436, + "grad_norm": 0.2373046875, + "learning_rate": 2.397681118540961e-06, + "loss": 0.4313, + "num_input_tokens_seen": 35481344, + "step": 2300 + }, + { + "epoch": 9.456410256410257, + "grad_norm": 0.369140625, + "learning_rate": 2.3952517140916724e-06, + "loss": 0.4664, + "num_input_tokens_seen": 35553664, + "step": 2305 + }, + { + "epoch": 9.476923076923077, + "grad_norm": 0.37890625, + "learning_rate": 2.392818656452354e-06, + "loss": 0.4948, + "num_input_tokens_seen": 35630592, + "step": 2310 + }, + { + "epoch": 9.497435897435897, + "grad_norm": 0.3046875, + "learning_rate": 2.39038195555141e-06, + "loss": 0.5035, + "num_input_tokens_seen": 35712864, + "step": 2315 + }, + { + "epoch": 9.517948717948718, + "grad_norm": 0.34765625, + "learning_rate": 2.387941621332114e-06, + "loss": 0.4964, + "num_input_tokens_seen": 35790784, + "step": 2320 + }, + { + "epoch": 9.538461538461538, + "grad_norm": 0.33203125, + "learning_rate": 2.3854976637525637e-06, + "loss": 0.4684, + "num_input_tokens_seen": 35868960, + "step": 2325 + }, + { + "epoch": 9.558974358974359, + "grad_norm": 0.345703125, + "learning_rate": 2.3830500927856433e-06, + "loss": 0.5117, + "num_input_tokens_seen": 35956832, + "step": 2330 + }, + { + "epoch": 9.57948717948718, + "grad_norm": 0.345703125, + "learning_rate": 2.3805989184189813e-06, + "loss": 0.4574, + "num_input_tokens_seen": 36027520, + "step": 2335 + }, + { + "epoch": 9.6, + "grad_norm": 0.30859375, + "learning_rate": 2.378144150654911e-06, + "loss": 0.5291, + "num_input_tokens_seen": 36109248, + "step": 2340 + }, + { + "epoch": 9.62051282051282, + "grad_norm": 0.357421875, + "learning_rate": 2.3756857995104286e-06, + "loss": 0.4528, + "num_input_tokens_seen": 36179584, + "step": 2345 + }, + { + "epoch": 9.64102564102564, + "grad_norm": 0.3671875, + "learning_rate": 2.3732238750171527e-06, + "loss": 0.4976, + "num_input_tokens_seen": 36257216, + "step": 2350 + }, + { + "epoch": 9.661538461538461, + "grad_norm": 0.30859375, + "learning_rate": 2.3707583872212837e-06, + "loss": 0.4576, + "num_input_tokens_seen": 36326272, + "step": 2355 + }, + { + "epoch": 9.682051282051281, + "grad_norm": 0.2294921875, + "learning_rate": 2.3682893461835626e-06, + "loss": 0.5315, + "num_input_tokens_seen": 36405504, + "step": 2360 + }, + { + "epoch": 9.702564102564102, + "grad_norm": 0.28515625, + "learning_rate": 2.3658167619792294e-06, + "loss": 0.5407, + "num_input_tokens_seen": 36481536, + "step": 2365 + }, + { + "epoch": 9.723076923076922, + "grad_norm": 0.267578125, + "learning_rate": 2.363340644697983e-06, + "loss": 0.4525, + "num_input_tokens_seen": 36563200, + "step": 2370 + }, + { + "epoch": 9.743589743589745, + "grad_norm": 0.205078125, + "learning_rate": 2.360861004443939e-06, + "loss": 0.5628, + "num_input_tokens_seen": 36661120, + "step": 2375 + }, + { + "epoch": 9.764102564102565, + "grad_norm": 0.2119140625, + "learning_rate": 2.358377851335589e-06, + "loss": 0.4633, + "num_input_tokens_seen": 36738880, + "step": 2380 + }, + { + "epoch": 9.784615384615385, + "grad_norm": 0.21484375, + "learning_rate": 2.3558911955057592e-06, + "loss": 0.4438, + "num_input_tokens_seen": 36820128, + "step": 2385 + }, + { + "epoch": 9.805128205128206, + "grad_norm": 0.1806640625, + "learning_rate": 2.35340104710157e-06, + "loss": 0.4609, + "num_input_tokens_seen": 36900128, + "step": 2390 + }, + { + "epoch": 9.825641025641026, + "grad_norm": 0.1708984375, + "learning_rate": 2.350907416284392e-06, + "loss": 0.4843, + "num_input_tokens_seen": 36969024, + "step": 2395 + }, + { + "epoch": 9.846153846153847, + "grad_norm": 0.162109375, + "learning_rate": 2.348410313229808e-06, + "loss": 0.4607, + "num_input_tokens_seen": 37053440, + "step": 2400 + }, + { + "epoch": 9.866666666666667, + "grad_norm": 0.146484375, + "learning_rate": 2.3459097481275687e-06, + "loss": 0.5134, + "num_input_tokens_seen": 37132128, + "step": 2405 + }, + { + "epoch": 9.887179487179488, + "grad_norm": 0.1396484375, + "learning_rate": 2.343405731181552e-06, + "loss": 0.478, + "num_input_tokens_seen": 37209664, + "step": 2410 + }, + { + "epoch": 9.907692307692308, + "grad_norm": 0.146484375, + "learning_rate": 2.3408982726097227e-06, + "loss": 0.4864, + "num_input_tokens_seen": 37283936, + "step": 2415 + }, + { + "epoch": 9.928205128205128, + "grad_norm": 0.1455078125, + "learning_rate": 2.3383873826440878e-06, + "loss": 0.4876, + "num_input_tokens_seen": 37359552, + "step": 2420 + }, + { + "epoch": 9.948717948717949, + "grad_norm": 0.13671875, + "learning_rate": 2.3358730715306574e-06, + "loss": 0.5265, + "num_input_tokens_seen": 37434112, + "step": 2425 + }, + { + "epoch": 9.96923076923077, + "grad_norm": 0.12353515625, + "learning_rate": 2.3333553495294033e-06, + "loss": 0.4759, + "num_input_tokens_seen": 37511456, + "step": 2430 + }, + { + "epoch": 9.98974358974359, + "grad_norm": 0.11865234375, + "learning_rate": 2.330834226914214e-06, + "loss": 0.4336, + "num_input_tokens_seen": 37584320, + "step": 2435 + }, + { + "epoch": 10.01025641025641, + "grad_norm": 0.123046875, + "learning_rate": 2.3283097139728557e-06, + "loss": 0.5338, + "num_input_tokens_seen": 37672864, + "step": 2440 + }, + { + "epoch": 10.03076923076923, + "grad_norm": 0.220703125, + "learning_rate": 2.3257818210069277e-06, + "loss": 0.4542, + "num_input_tokens_seen": 37743488, + "step": 2445 + }, + { + "epoch": 10.051282051282051, + "grad_norm": 0.14453125, + "learning_rate": 2.3232505583318246e-06, + "loss": 0.5379, + "num_input_tokens_seen": 37832256, + "step": 2450 + }, + { + "epoch": 10.071794871794872, + "grad_norm": 0.1533203125, + "learning_rate": 2.320715936276689e-06, + "loss": 0.4418, + "num_input_tokens_seen": 37908832, + "step": 2455 + }, + { + "epoch": 10.092307692307692, + "grad_norm": 0.3984375, + "learning_rate": 2.3181779651843738e-06, + "loss": 0.5294, + "num_input_tokens_seen": 37989088, + "step": 2460 + }, + { + "epoch": 10.112820512820512, + "grad_norm": 0.26953125, + "learning_rate": 2.3156366554113967e-06, + "loss": 0.4438, + "num_input_tokens_seen": 38061024, + "step": 2465 + }, + { + "epoch": 10.133333333333333, + "grad_norm": 0.318359375, + "learning_rate": 2.3130920173278997e-06, + "loss": 0.4544, + "num_input_tokens_seen": 38136480, + "step": 2470 + }, + { + "epoch": 10.153846153846153, + "grad_norm": 0.55078125, + "learning_rate": 2.310544061317607e-06, + "loss": 0.4918, + "num_input_tokens_seen": 38213280, + "step": 2475 + }, + { + "epoch": 10.174358974358974, + "grad_norm": 0.8984375, + "learning_rate": 2.307992797777782e-06, + "loss": 0.4821, + "num_input_tokens_seen": 38283296, + "step": 2480 + }, + { + "epoch": 10.194871794871794, + "grad_norm": 0.27734375, + "learning_rate": 2.3054382371191836e-06, + "loss": 0.464, + "num_input_tokens_seen": 38360320, + "step": 2485 + }, + { + "epoch": 10.215384615384615, + "grad_norm": 0.251953125, + "learning_rate": 2.3028803897660256e-06, + "loss": 0.4829, + "num_input_tokens_seen": 38440000, + "step": 2490 + }, + { + "epoch": 10.235897435897435, + "grad_norm": 0.53125, + "learning_rate": 2.3003192661559346e-06, + "loss": 0.4495, + "num_input_tokens_seen": 38510208, + "step": 2495 + }, + { + "epoch": 10.256410256410255, + "grad_norm": 0.30859375, + "learning_rate": 2.297754876739905e-06, + "loss": 0.5503, + "num_input_tokens_seen": 38600256, + "step": 2500 + }, + { + "epoch": 10.276923076923078, + "grad_norm": 0.384765625, + "learning_rate": 2.2951872319822597e-06, + "loss": 0.5341, + "num_input_tokens_seen": 38679168, + "step": 2505 + }, + { + "epoch": 10.297435897435898, + "grad_norm": 0.3046875, + "learning_rate": 2.2926163423606027e-06, + "loss": 0.4903, + "num_input_tokens_seen": 38752064, + "step": 2510 + }, + { + "epoch": 10.317948717948719, + "grad_norm": 0.353515625, + "learning_rate": 2.2900422183657816e-06, + "loss": 0.4543, + "num_input_tokens_seen": 38829504, + "step": 2515 + }, + { + "epoch": 10.338461538461539, + "grad_norm": 0.330078125, + "learning_rate": 2.2874648705018403e-06, + "loss": 0.5428, + "num_input_tokens_seen": 38915904, + "step": 2520 + }, + { + "epoch": 10.35897435897436, + "grad_norm": 0.349609375, + "learning_rate": 2.28488430928598e-06, + "loss": 0.4588, + "num_input_tokens_seen": 38997760, + "step": 2525 + }, + { + "epoch": 10.37948717948718, + "grad_norm": 0.37109375, + "learning_rate": 2.282300545248512e-06, + "loss": 0.4441, + "num_input_tokens_seen": 39068000, + "step": 2530 + }, + { + "epoch": 10.4, + "grad_norm": 0.375, + "learning_rate": 2.27971358893282e-06, + "loss": 0.4441, + "num_input_tokens_seen": 39136480, + "step": 2535 + }, + { + "epoch": 10.42051282051282, + "grad_norm": 0.421875, + "learning_rate": 2.2771234508953116e-06, + "loss": 0.442, + "num_input_tokens_seen": 39219488, + "step": 2540 + }, + { + "epoch": 10.441025641025641, + "grad_norm": 0.625, + "learning_rate": 2.27453014170538e-06, + "loss": 0.5199, + "num_input_tokens_seen": 39295712, + "step": 2545 + }, + { + "epoch": 10.461538461538462, + "grad_norm": 0.345703125, + "learning_rate": 2.271933671945357e-06, + "loss": 0.5147, + "num_input_tokens_seen": 39371008, + "step": 2550 + }, + { + "epoch": 10.482051282051282, + "grad_norm": 0.359375, + "learning_rate": 2.2693340522104727e-06, + "loss": 0.5296, + "num_input_tokens_seen": 39450944, + "step": 2555 + }, + { + "epoch": 10.502564102564103, + "grad_norm": 0.310546875, + "learning_rate": 2.2667312931088096e-06, + "loss": 0.4707, + "num_input_tokens_seen": 39530912, + "step": 2560 + }, + { + "epoch": 10.523076923076923, + "grad_norm": 0.34375, + "learning_rate": 2.264125405261263e-06, + "loss": 0.4726, + "num_input_tokens_seen": 39602400, + "step": 2565 + }, + { + "epoch": 10.543589743589743, + "grad_norm": 0.37890625, + "learning_rate": 2.261516399301493e-06, + "loss": 0.4344, + "num_input_tokens_seen": 39679232, + "step": 2570 + }, + { + "epoch": 10.564102564102564, + "grad_norm": 0.302734375, + "learning_rate": 2.2589042858758853e-06, + "loss": 0.4427, + "num_input_tokens_seen": 39755904, + "step": 2575 + }, + { + "epoch": 10.584615384615384, + "grad_norm": 0.326171875, + "learning_rate": 2.256289075643506e-06, + "loss": 0.4975, + "num_input_tokens_seen": 39826368, + "step": 2580 + }, + { + "epoch": 10.605128205128205, + "grad_norm": 0.287109375, + "learning_rate": 2.2536707792760566e-06, + "loss": 0.5045, + "num_input_tokens_seen": 39906464, + "step": 2585 + }, + { + "epoch": 10.625641025641025, + "grad_norm": 0.27734375, + "learning_rate": 2.251049407457833e-06, + "loss": 0.4833, + "num_input_tokens_seen": 39975712, + "step": 2590 + }, + { + "epoch": 10.646153846153846, + "grad_norm": 0.263671875, + "learning_rate": 2.2484249708856823e-06, + "loss": 0.4984, + "num_input_tokens_seen": 40060032, + "step": 2595 + }, + { + "epoch": 10.666666666666666, + "grad_norm": 0.2333984375, + "learning_rate": 2.2457974802689545e-06, + "loss": 0.4186, + "num_input_tokens_seen": 40131520, + "step": 2600 + }, + { + "epoch": 10.687179487179487, + "grad_norm": 0.1962890625, + "learning_rate": 2.2431669463294646e-06, + "loss": 0.4441, + "num_input_tokens_seen": 40205760, + "step": 2605 + }, + { + "epoch": 10.707692307692307, + "grad_norm": 0.2119140625, + "learning_rate": 2.2405333798014453e-06, + "loss": 0.5337, + "num_input_tokens_seen": 40288992, + "step": 2610 + }, + { + "epoch": 10.728205128205127, + "grad_norm": 0.208984375, + "learning_rate": 2.237896791431505e-06, + "loss": 0.4703, + "num_input_tokens_seen": 40369440, + "step": 2615 + }, + { + "epoch": 10.74871794871795, + "grad_norm": 0.1630859375, + "learning_rate": 2.2352571919785812e-06, + "loss": 0.5217, + "num_input_tokens_seen": 40452288, + "step": 2620 + }, + { + "epoch": 10.76923076923077, + "grad_norm": 0.1572265625, + "learning_rate": 2.2326145922139004e-06, + "loss": 0.4475, + "num_input_tokens_seen": 40523808, + "step": 2625 + }, + { + "epoch": 10.78974358974359, + "grad_norm": 0.1884765625, + "learning_rate": 2.2299690029209313e-06, + "loss": 0.4734, + "num_input_tokens_seen": 40606496, + "step": 2630 + }, + { + "epoch": 10.810256410256411, + "grad_norm": 0.1259765625, + "learning_rate": 2.227320434895343e-06, + "loss": 0.4686, + "num_input_tokens_seen": 40684672, + "step": 2635 + }, + { + "epoch": 10.830769230769231, + "grad_norm": 0.1279296875, + "learning_rate": 2.2246688989449577e-06, + "loss": 0.5027, + "num_input_tokens_seen": 40762752, + "step": 2640 + }, + { + "epoch": 10.851282051282052, + "grad_norm": 0.12109375, + "learning_rate": 2.2220144058897104e-06, + "loss": 0.4582, + "num_input_tokens_seen": 40851776, + "step": 2645 + }, + { + "epoch": 10.871794871794872, + "grad_norm": 0.1318359375, + "learning_rate": 2.2193569665616017e-06, + "loss": 0.4516, + "num_input_tokens_seen": 40922304, + "step": 2650 + }, + { + "epoch": 10.892307692307693, + "grad_norm": 0.12109375, + "learning_rate": 2.2166965918046554e-06, + "loss": 0.4346, + "num_input_tokens_seen": 40998784, + "step": 2655 + }, + { + "epoch": 10.912820512820513, + "grad_norm": 0.16015625, + "learning_rate": 2.214033292474874e-06, + "loss": 0.5093, + "num_input_tokens_seen": 41075872, + "step": 2660 + }, + { + "epoch": 10.933333333333334, + "grad_norm": 0.154296875, + "learning_rate": 2.2113670794401935e-06, + "loss": 0.4853, + "num_input_tokens_seen": 41149728, + "step": 2665 + }, + { + "epoch": 10.953846153846154, + "grad_norm": 0.2333984375, + "learning_rate": 2.20869796358044e-06, + "loss": 0.5042, + "num_input_tokens_seen": 41220160, + "step": 2670 + }, + { + "epoch": 10.974358974358974, + "grad_norm": 0.1630859375, + "learning_rate": 2.2060259557872845e-06, + "loss": 0.4601, + "num_input_tokens_seen": 41287712, + "step": 2675 + }, + { + "epoch": 10.994871794871795, + "grad_norm": 0.1533203125, + "learning_rate": 2.2033510669641997e-06, + "loss": 0.4138, + "num_input_tokens_seen": 41363264, + "step": 2680 + }, + { + "epoch": 11.015384615384615, + "grad_norm": 0.306640625, + "learning_rate": 2.2006733080264144e-06, + "loss": 0.4724, + "num_input_tokens_seen": 41437152, + "step": 2685 + }, + { + "epoch": 11.035897435897436, + "grad_norm": 0.26953125, + "learning_rate": 2.197992689900869e-06, + "loss": 0.4932, + "num_input_tokens_seen": 41515520, + "step": 2690 + }, + { + "epoch": 11.056410256410256, + "grad_norm": 0.5859375, + "learning_rate": 2.195309223526171e-06, + "loss": 0.4299, + "num_input_tokens_seen": 41587200, + "step": 2695 + }, + { + "epoch": 11.076923076923077, + "grad_norm": 0.59375, + "learning_rate": 2.192622919852551e-06, + "loss": 0.4774, + "num_input_tokens_seen": 41663008, + "step": 2700 + }, + { + "epoch": 11.097435897435897, + "grad_norm": 0.6640625, + "learning_rate": 2.1899337898418174e-06, + "loss": 0.5241, + "num_input_tokens_seen": 41743264, + "step": 2705 + }, + { + "epoch": 11.117948717948718, + "grad_norm": 0.29296875, + "learning_rate": 2.187241844467313e-06, + "loss": 0.498, + "num_input_tokens_seen": 41823264, + "step": 2710 + }, + { + "epoch": 11.138461538461538, + "grad_norm": 0.2890625, + "learning_rate": 2.1845470947138658e-06, + "loss": 0.5311, + "num_input_tokens_seen": 41914560, + "step": 2715 + }, + { + "epoch": 11.158974358974358, + "grad_norm": 0.58984375, + "learning_rate": 2.181849551577751e-06, + "loss": 0.4464, + "num_input_tokens_seen": 41988288, + "step": 2720 + }, + { + "epoch": 11.179487179487179, + "grad_norm": 0.34375, + "learning_rate": 2.179149226066641e-06, + "loss": 0.4905, + "num_input_tokens_seen": 42068416, + "step": 2725 + }, + { + "epoch": 11.2, + "grad_norm": 0.275390625, + "learning_rate": 2.1764461291995618e-06, + "loss": 0.4629, + "num_input_tokens_seen": 42139744, + "step": 2730 + }, + { + "epoch": 11.22051282051282, + "grad_norm": 0.2578125, + "learning_rate": 2.173740272006849e-06, + "loss": 0.5278, + "num_input_tokens_seen": 42228672, + "step": 2735 + }, + { + "epoch": 11.24102564102564, + "grad_norm": 0.6640625, + "learning_rate": 2.1710316655301016e-06, + "loss": 0.4513, + "num_input_tokens_seen": 42306528, + "step": 2740 + }, + { + "epoch": 11.261538461538462, + "grad_norm": 0.2470703125, + "learning_rate": 2.1683203208221375e-06, + "loss": 0.4917, + "num_input_tokens_seen": 42389024, + "step": 2745 + }, + { + "epoch": 11.282051282051283, + "grad_norm": 0.333984375, + "learning_rate": 2.165606248946948e-06, + "loss": 0.4159, + "num_input_tokens_seen": 42473088, + "step": 2750 + }, + { + "epoch": 11.302564102564103, + "grad_norm": 0.291015625, + "learning_rate": 2.1628894609796533e-06, + "loss": 0.4917, + "num_input_tokens_seen": 42551360, + "step": 2755 + }, + { + "epoch": 11.323076923076924, + "grad_norm": 0.390625, + "learning_rate": 2.1601699680064573e-06, + "loss": 0.5037, + "num_input_tokens_seen": 42626688, + "step": 2760 + }, + { + "epoch": 11.343589743589744, + "grad_norm": 0.369140625, + "learning_rate": 2.1574477811246014e-06, + "loss": 0.4756, + "num_input_tokens_seen": 42705056, + "step": 2765 + }, + { + "epoch": 11.364102564102565, + "grad_norm": 0.421875, + "learning_rate": 2.1547229114423207e-06, + "loss": 0.4985, + "num_input_tokens_seen": 42777632, + "step": 2770 + }, + { + "epoch": 11.384615384615385, + "grad_norm": 0.326171875, + "learning_rate": 2.1519953700787963e-06, + "loss": 0.4561, + "num_input_tokens_seen": 42845888, + "step": 2775 + }, + { + "epoch": 11.405128205128205, + "grad_norm": 0.36328125, + "learning_rate": 2.149265168164113e-06, + "loss": 0.5091, + "num_input_tokens_seen": 42922976, + "step": 2780 + }, + { + "epoch": 11.425641025641026, + "grad_norm": 0.353515625, + "learning_rate": 2.146532316839211e-06, + "loss": 0.4711, + "num_input_tokens_seen": 42996000, + "step": 2785 + }, + { + "epoch": 11.446153846153846, + "grad_norm": 0.38671875, + "learning_rate": 2.1437968272558435e-06, + "loss": 0.457, + "num_input_tokens_seen": 43074688, + "step": 2790 + }, + { + "epoch": 11.466666666666667, + "grad_norm": 0.31640625, + "learning_rate": 2.1410587105765275e-06, + "loss": 0.541, + "num_input_tokens_seen": 43157280, + "step": 2795 + }, + { + "epoch": 11.487179487179487, + "grad_norm": 0.361328125, + "learning_rate": 2.138317977974501e-06, + "loss": 0.4279, + "num_input_tokens_seen": 43234016, + "step": 2800 + }, + { + "epoch": 11.507692307692308, + "grad_norm": 0.306640625, + "learning_rate": 2.135574640633678e-06, + "loss": 0.5213, + "num_input_tokens_seen": 43310816, + "step": 2805 + }, + { + "epoch": 11.528205128205128, + "grad_norm": 0.275390625, + "learning_rate": 2.132828709748598e-06, + "loss": 0.4444, + "num_input_tokens_seen": 43382976, + "step": 2810 + }, + { + "epoch": 11.548717948717949, + "grad_norm": 0.279296875, + "learning_rate": 2.130080196524388e-06, + "loss": 0.4768, + "num_input_tokens_seen": 43462944, + "step": 2815 + }, + { + "epoch": 11.569230769230769, + "grad_norm": 0.283203125, + "learning_rate": 2.1273291121767094e-06, + "loss": 0.4376, + "num_input_tokens_seen": 43535232, + "step": 2820 + }, + { + "epoch": 11.58974358974359, + "grad_norm": 0.23828125, + "learning_rate": 2.124575467931717e-06, + "loss": 0.4342, + "num_input_tokens_seen": 43618528, + "step": 2825 + }, + { + "epoch": 11.61025641025641, + "grad_norm": 0.1982421875, + "learning_rate": 2.1218192750260114e-06, + "loss": 0.4596, + "num_input_tokens_seen": 43691904, + "step": 2830 + }, + { + "epoch": 11.63076923076923, + "grad_norm": 0.1904296875, + "learning_rate": 2.119060544706592e-06, + "loss": 0.4811, + "num_input_tokens_seen": 43760480, + "step": 2835 + }, + { + "epoch": 11.65128205128205, + "grad_norm": 0.1923828125, + "learning_rate": 2.1162992882308147e-06, + "loss": 0.4864, + "num_input_tokens_seen": 43829984, + "step": 2840 + }, + { + "epoch": 11.671794871794871, + "grad_norm": 0.1875, + "learning_rate": 2.1135355168663417e-06, + "loss": 0.4678, + "num_input_tokens_seen": 43906816, + "step": 2845 + }, + { + "epoch": 11.692307692307692, + "grad_norm": 0.1484375, + "learning_rate": 2.1107692418910985e-06, + "loss": 0.5001, + "num_input_tokens_seen": 43984960, + "step": 2850 + }, + { + "epoch": 11.712820512820512, + "grad_norm": 0.154296875, + "learning_rate": 2.1080004745932274e-06, + "loss": 0.4662, + "num_input_tokens_seen": 44061440, + "step": 2855 + }, + { + "epoch": 11.733333333333333, + "grad_norm": 0.1611328125, + "learning_rate": 2.1052292262710392e-06, + "loss": 0.526, + "num_input_tokens_seen": 44147008, + "step": 2860 + }, + { + "epoch": 11.753846153846155, + "grad_norm": 0.1337890625, + "learning_rate": 2.102455508232971e-06, + "loss": 0.4821, + "num_input_tokens_seen": 44224224, + "step": 2865 + }, + { + "epoch": 11.774358974358975, + "grad_norm": 0.146484375, + "learning_rate": 2.099679331797536e-06, + "loss": 0.4923, + "num_input_tokens_seen": 44302816, + "step": 2870 + }, + { + "epoch": 11.794871794871796, + "grad_norm": 0.1650390625, + "learning_rate": 2.0969007082932803e-06, + "loss": 0.521, + "num_input_tokens_seen": 44376160, + "step": 2875 + }, + { + "epoch": 11.815384615384616, + "grad_norm": 0.125, + "learning_rate": 2.0941196490587354e-06, + "loss": 0.4932, + "num_input_tokens_seen": 44459200, + "step": 2880 + }, + { + "epoch": 11.835897435897436, + "grad_norm": 0.140625, + "learning_rate": 2.0913361654423723e-06, + "loss": 0.4866, + "num_input_tokens_seen": 44536128, + "step": 2885 + }, + { + "epoch": 11.856410256410257, + "grad_norm": 0.1357421875, + "learning_rate": 2.0885502688025538e-06, + "loss": 0.4826, + "num_input_tokens_seen": 44605088, + "step": 2890 + }, + { + "epoch": 11.876923076923077, + "grad_norm": 0.2177734375, + "learning_rate": 2.0857619705074912e-06, + "loss": 0.4433, + "num_input_tokens_seen": 44677984, + "step": 2895 + }, + { + "epoch": 11.897435897435898, + "grad_norm": 0.166015625, + "learning_rate": 2.082971281935195e-06, + "loss": 0.4122, + "num_input_tokens_seen": 44751200, + "step": 2900 + }, + { + "epoch": 11.917948717948718, + "grad_norm": 0.28515625, + "learning_rate": 2.0801782144734295e-06, + "loss": 0.4266, + "num_input_tokens_seen": 44824672, + "step": 2905 + }, + { + "epoch": 11.938461538461539, + "grad_norm": 0.341796875, + "learning_rate": 2.0773827795196667e-06, + "loss": 0.4574, + "num_input_tokens_seen": 44912768, + "step": 2910 + }, + { + "epoch": 11.95897435897436, + "grad_norm": 0.2578125, + "learning_rate": 2.074584988481039e-06, + "loss": 0.5026, + "num_input_tokens_seen": 44993632, + "step": 2915 + }, + { + "epoch": 11.97948717948718, + "grad_norm": 0.79296875, + "learning_rate": 2.0717848527742935e-06, + "loss": 0.5444, + "num_input_tokens_seen": 45070848, + "step": 2920 + }, + { + "epoch": 12.0, + "grad_norm": 0.546875, + "learning_rate": 2.0689823838257455e-06, + "loss": 0.4509, + "num_input_tokens_seen": 45150496, + "step": 2925 + }, + { + "epoch": 12.02051282051282, + "grad_norm": 0.7421875, + "learning_rate": 2.0661775930712297e-06, + "loss": 0.4534, + "num_input_tokens_seen": 45225440, + "step": 2930 + }, + { + "epoch": 12.04102564102564, + "grad_norm": 0.3984375, + "learning_rate": 2.0633704919560573e-06, + "loss": 0.5264, + "num_input_tokens_seen": 45300992, + "step": 2935 + }, + { + "epoch": 12.061538461538461, + "grad_norm": 0.84765625, + "learning_rate": 2.0605610919349658e-06, + "loss": 0.4373, + "num_input_tokens_seen": 45378944, + "step": 2940 + }, + { + "epoch": 12.082051282051282, + "grad_norm": 0.53515625, + "learning_rate": 2.0577494044720746e-06, + "loss": 0.4779, + "num_input_tokens_seen": 45453504, + "step": 2945 + }, + { + "epoch": 12.102564102564102, + "grad_norm": 0.265625, + "learning_rate": 2.0549354410408364e-06, + "loss": 0.556, + "num_input_tokens_seen": 45532992, + "step": 2950 + }, + { + "epoch": 12.123076923076923, + "grad_norm": 0.2890625, + "learning_rate": 2.052119213123992e-06, + "loss": 0.5152, + "num_input_tokens_seen": 45609120, + "step": 2955 + }, + { + "epoch": 12.143589743589743, + "grad_norm": 0.318359375, + "learning_rate": 2.049300732213522e-06, + "loss": 0.4412, + "num_input_tokens_seen": 45690624, + "step": 2960 + }, + { + "epoch": 12.164102564102564, + "grad_norm": 0.46875, + "learning_rate": 2.046480009810602e-06, + "loss": 0.4553, + "num_input_tokens_seen": 45763264, + "step": 2965 + }, + { + "epoch": 12.184615384615384, + "grad_norm": 0.275390625, + "learning_rate": 2.0436570574255523e-06, + "loss": 0.4913, + "num_input_tokens_seen": 45849472, + "step": 2970 + }, + { + "epoch": 12.205128205128204, + "grad_norm": 0.41015625, + "learning_rate": 2.0408318865777953e-06, + "loss": 0.5487, + "num_input_tokens_seen": 45927552, + "step": 2975 + }, + { + "epoch": 12.225641025641025, + "grad_norm": 0.291015625, + "learning_rate": 2.0380045087958036e-06, + "loss": 0.465, + "num_input_tokens_seen": 46002656, + "step": 2980 + }, + { + "epoch": 12.246153846153845, + "grad_norm": 0.404296875, + "learning_rate": 2.0351749356170574e-06, + "loss": 0.4854, + "num_input_tokens_seen": 46087904, + "step": 2985 + }, + { + "epoch": 12.266666666666667, + "grad_norm": 0.37109375, + "learning_rate": 2.032343178587995e-06, + "loss": 0.4568, + "num_input_tokens_seen": 46165408, + "step": 2990 + }, + { + "epoch": 12.287179487179488, + "grad_norm": 0.3515625, + "learning_rate": 2.0295092492639657e-06, + "loss": 0.4926, + "num_input_tokens_seen": 46237344, + "step": 2995 + }, + { + "epoch": 12.307692307692308, + "grad_norm": 0.5, + "learning_rate": 2.0266731592091834e-06, + "loss": 0.5093, + "num_input_tokens_seen": 46307456, + "step": 3000 + }, + { + "epoch": 12.328205128205129, + "grad_norm": 0.33984375, + "learning_rate": 2.0238349199966793e-06, + "loss": 0.5077, + "num_input_tokens_seen": 46390688, + "step": 3005 + }, + { + "epoch": 12.34871794871795, + "grad_norm": 0.32421875, + "learning_rate": 2.020994543208254e-06, + "loss": 0.4541, + "num_input_tokens_seen": 46467904, + "step": 3010 + }, + { + "epoch": 12.36923076923077, + "grad_norm": 0.33203125, + "learning_rate": 2.018152040434432e-06, + "loss": 0.4975, + "num_input_tokens_seen": 46545632, + "step": 3015 + }, + { + "epoch": 12.38974358974359, + "grad_norm": 0.40234375, + "learning_rate": 2.015307423274411e-06, + "loss": 0.4988, + "num_input_tokens_seen": 46624768, + "step": 3020 + }, + { + "epoch": 12.41025641025641, + "grad_norm": 0.298828125, + "learning_rate": 2.0124607033360193e-06, + "loss": 0.4877, + "num_input_tokens_seen": 46694528, + "step": 3025 + }, + { + "epoch": 12.430769230769231, + "grad_norm": 0.3046875, + "learning_rate": 2.0096118922356646e-06, + "loss": 0.4621, + "num_input_tokens_seen": 46767520, + "step": 3030 + }, + { + "epoch": 12.451282051282051, + "grad_norm": 0.255859375, + "learning_rate": 2.0067610015982868e-06, + "loss": 0.4742, + "num_input_tokens_seen": 46843616, + "step": 3035 + }, + { + "epoch": 12.471794871794872, + "grad_norm": 0.27734375, + "learning_rate": 2.0039080430573133e-06, + "loss": 0.3993, + "num_input_tokens_seen": 46915136, + "step": 3040 + }, + { + "epoch": 12.492307692307692, + "grad_norm": 0.232421875, + "learning_rate": 2.0010530282546093e-06, + "loss": 0.464, + "num_input_tokens_seen": 46992000, + "step": 3045 + }, + { + "epoch": 12.512820512820513, + "grad_norm": 0.2060546875, + "learning_rate": 1.9981959688404303e-06, + "loss": 0.4744, + "num_input_tokens_seen": 47075360, + "step": 3050 + }, + { + "epoch": 12.533333333333333, + "grad_norm": 0.1953125, + "learning_rate": 1.9953368764733763e-06, + "loss": 0.4788, + "num_input_tokens_seen": 47152704, + "step": 3055 + }, + { + "epoch": 12.553846153846154, + "grad_norm": 0.1923828125, + "learning_rate": 1.992475762820342e-06, + "loss": 0.4704, + "num_input_tokens_seen": 47229696, + "step": 3060 + }, + { + "epoch": 12.574358974358974, + "grad_norm": 0.193359375, + "learning_rate": 1.9896126395564695e-06, + "loss": 0.4645, + "num_input_tokens_seen": 47307360, + "step": 3065 + }, + { + "epoch": 12.594871794871795, + "grad_norm": 0.1728515625, + "learning_rate": 1.986747518365104e-06, + "loss": 0.5485, + "num_input_tokens_seen": 47393824, + "step": 3070 + }, + { + "epoch": 12.615384615384615, + "grad_norm": 0.154296875, + "learning_rate": 1.9838804109377405e-06, + "loss": 0.511, + "num_input_tokens_seen": 47475104, + "step": 3075 + }, + { + "epoch": 12.635897435897435, + "grad_norm": 0.1630859375, + "learning_rate": 1.9810113289739818e-06, + "loss": 0.4624, + "num_input_tokens_seen": 47548704, + "step": 3080 + }, + { + "epoch": 12.656410256410256, + "grad_norm": 0.1435546875, + "learning_rate": 1.9781402841814855e-06, + "loss": 0.5197, + "num_input_tokens_seen": 47638880, + "step": 3085 + }, + { + "epoch": 12.676923076923076, + "grad_norm": 0.134765625, + "learning_rate": 1.9752672882759204e-06, + "loss": 0.4277, + "num_input_tokens_seen": 47711456, + "step": 3090 + }, + { + "epoch": 12.697435897435897, + "grad_norm": 0.138671875, + "learning_rate": 1.972392352980917e-06, + "loss": 0.5166, + "num_input_tokens_seen": 47787776, + "step": 3095 + }, + { + "epoch": 12.717948717948717, + "grad_norm": 0.1357421875, + "learning_rate": 1.969515490028019e-06, + "loss": 0.4201, + "num_input_tokens_seen": 47860480, + "step": 3100 + }, + { + "epoch": 12.73846153846154, + "grad_norm": 0.1064453125, + "learning_rate": 1.966636711156636e-06, + "loss": 0.4425, + "num_input_tokens_seen": 47936704, + "step": 3105 + }, + { + "epoch": 12.75897435897436, + "grad_norm": 0.154296875, + "learning_rate": 1.9637560281139982e-06, + "loss": 0.5056, + "num_input_tokens_seen": 48021792, + "step": 3110 + }, + { + "epoch": 12.77948717948718, + "grad_norm": 0.1953125, + "learning_rate": 1.960873452655102e-06, + "loss": 0.4929, + "num_input_tokens_seen": 48101984, + "step": 3115 + }, + { + "epoch": 12.8, + "grad_norm": 0.2392578125, + "learning_rate": 1.95798899654267e-06, + "loss": 0.4868, + "num_input_tokens_seen": 48186688, + "step": 3120 + }, + { + "epoch": 12.820512820512821, + "grad_norm": 0.134765625, + "learning_rate": 1.9551026715470954e-06, + "loss": 0.4668, + "num_input_tokens_seen": 48259872, + "step": 3125 + }, + { + "epoch": 12.841025641025642, + "grad_norm": 0.6953125, + "learning_rate": 1.952214489446401e-06, + "loss": 0.53, + "num_input_tokens_seen": 48347232, + "step": 3130 + }, + { + "epoch": 12.861538461538462, + "grad_norm": 0.296875, + "learning_rate": 1.949324462026185e-06, + "loss": 0.4388, + "num_input_tokens_seen": 48422624, + "step": 3135 + }, + { + "epoch": 12.882051282051282, + "grad_norm": 0.28125, + "learning_rate": 1.9464326010795776e-06, + "loss": 0.4246, + "num_input_tokens_seen": 48492288, + "step": 3140 + }, + { + "epoch": 12.902564102564103, + "grad_norm": 0.72265625, + "learning_rate": 1.9435389184071895e-06, + "loss": 0.5186, + "num_input_tokens_seen": 48568224, + "step": 3145 + }, + { + "epoch": 12.923076923076923, + "grad_norm": 0.7734375, + "learning_rate": 1.9406434258170666e-06, + "loss": 0.4351, + "num_input_tokens_seen": 48637280, + "step": 3150 + }, + { + "epoch": 12.943589743589744, + "grad_norm": 2.03125, + "learning_rate": 1.9377461351246395e-06, + "loss": 0.5281, + "num_input_tokens_seen": 48717088, + "step": 3155 + }, + { + "epoch": 12.964102564102564, + "grad_norm": 0.236328125, + "learning_rate": 1.9348470581526763e-06, + "loss": 0.4308, + "num_input_tokens_seen": 48787584, + "step": 3160 + }, + { + "epoch": 12.984615384615385, + "grad_norm": 0.52734375, + "learning_rate": 1.9319462067312344e-06, + "loss": 0.4133, + "num_input_tokens_seen": 48864640, + "step": 3165 + }, + { + "epoch": 13.005128205128205, + "grad_norm": 0.55859375, + "learning_rate": 1.929043592697612e-06, + "loss": 0.4802, + "num_input_tokens_seen": 48938176, + "step": 3170 + }, + { + "epoch": 13.025641025641026, + "grad_norm": 0.306640625, + "learning_rate": 1.926139227896299e-06, + "loss": 0.4504, + "num_input_tokens_seen": 49015040, + "step": 3175 + }, + { + "epoch": 13.046153846153846, + "grad_norm": 0.4140625, + "learning_rate": 1.923233124178932e-06, + "loss": 0.494, + "num_input_tokens_seen": 49091680, + "step": 3180 + }, + { + "epoch": 13.066666666666666, + "grad_norm": 0.3671875, + "learning_rate": 1.9203252934042403e-06, + "loss": 0.4628, + "num_input_tokens_seen": 49165600, + "step": 3185 + }, + { + "epoch": 13.087179487179487, + "grad_norm": 0.318359375, + "learning_rate": 1.9174157474380034e-06, + "loss": 0.4431, + "num_input_tokens_seen": 49247136, + "step": 3190 + }, + { + "epoch": 13.107692307692307, + "grad_norm": 0.27734375, + "learning_rate": 1.914504498152998e-06, + "loss": 0.4479, + "num_input_tokens_seen": 49319648, + "step": 3195 + }, + { + "epoch": 13.128205128205128, + "grad_norm": 0.40625, + "learning_rate": 1.9115915574289525e-06, + "loss": 0.4278, + "num_input_tokens_seen": 49393216, + "step": 3200 + }, + { + "epoch": 13.148717948717948, + "grad_norm": 0.302734375, + "learning_rate": 1.9086769371524966e-06, + "loss": 0.5124, + "num_input_tokens_seen": 49480864, + "step": 3205 + }, + { + "epoch": 13.169230769230769, + "grad_norm": 0.3515625, + "learning_rate": 1.9057606492171144e-06, + "loss": 0.4438, + "num_input_tokens_seen": 49559904, + "step": 3210 + }, + { + "epoch": 13.189743589743589, + "grad_norm": 0.412109375, + "learning_rate": 1.9028427055230948e-06, + "loss": 0.4312, + "num_input_tokens_seen": 49630240, + "step": 3215 + }, + { + "epoch": 13.21025641025641, + "grad_norm": 0.443359375, + "learning_rate": 1.8999231179774833e-06, + "loss": 0.5431, + "num_input_tokens_seen": 49709888, + "step": 3220 + }, + { + "epoch": 13.23076923076923, + "grad_norm": 0.447265625, + "learning_rate": 1.897001898494033e-06, + "loss": 0.4567, + "num_input_tokens_seen": 49778848, + "step": 3225 + }, + { + "epoch": 13.25128205128205, + "grad_norm": 0.396484375, + "learning_rate": 1.8940790589931568e-06, + "loss": 0.4227, + "num_input_tokens_seen": 49849024, + "step": 3230 + }, + { + "epoch": 13.271794871794873, + "grad_norm": 0.306640625, + "learning_rate": 1.8911546114018775e-06, + "loss": 0.4738, + "num_input_tokens_seen": 49917952, + "step": 3235 + }, + { + "epoch": 13.292307692307693, + "grad_norm": 0.3203125, + "learning_rate": 1.888228567653781e-06, + "loss": 0.4661, + "num_input_tokens_seen": 50003392, + "step": 3240 + }, + { + "epoch": 13.312820512820513, + "grad_norm": 0.330078125, + "learning_rate": 1.8853009396889665e-06, + "loss": 0.5039, + "num_input_tokens_seen": 50085280, + "step": 3245 + }, + { + "epoch": 13.333333333333334, + "grad_norm": 0.33984375, + "learning_rate": 1.8823717394539966e-06, + "loss": 0.4442, + "num_input_tokens_seen": 50166304, + "step": 3250 + }, + { + "epoch": 13.353846153846154, + "grad_norm": 0.271484375, + "learning_rate": 1.8794409789018507e-06, + "loss": 0.5048, + "num_input_tokens_seen": 50247200, + "step": 3255 + }, + { + "epoch": 13.374358974358975, + "grad_norm": 0.275390625, + "learning_rate": 1.8765086699918747e-06, + "loss": 0.4615, + "num_input_tokens_seen": 50321120, + "step": 3260 + }, + { + "epoch": 13.394871794871795, + "grad_norm": 0.275390625, + "learning_rate": 1.8735748246897337e-06, + "loss": 0.4838, + "num_input_tokens_seen": 50403680, + "step": 3265 + }, + { + "epoch": 13.415384615384616, + "grad_norm": 0.203125, + "learning_rate": 1.8706394549673615e-06, + "loss": 0.4454, + "num_input_tokens_seen": 50481504, + "step": 3270 + }, + { + "epoch": 13.435897435897436, + "grad_norm": 0.220703125, + "learning_rate": 1.8677025728029122e-06, + "loss": 0.4806, + "num_input_tokens_seen": 50556448, + "step": 3275 + }, + { + "epoch": 13.456410256410257, + "grad_norm": 0.2060546875, + "learning_rate": 1.8647641901807126e-06, + "loss": 0.4601, + "num_input_tokens_seen": 50630304, + "step": 3280 + }, + { + "epoch": 13.476923076923077, + "grad_norm": 0.2236328125, + "learning_rate": 1.861824319091212e-06, + "loss": 0.4513, + "num_input_tokens_seen": 50704704, + "step": 3285 + }, + { + "epoch": 13.497435897435897, + "grad_norm": 0.1572265625, + "learning_rate": 1.8588829715309324e-06, + "loss": 0.4704, + "num_input_tokens_seen": 50782464, + "step": 3290 + }, + { + "epoch": 13.517948717948718, + "grad_norm": 0.166015625, + "learning_rate": 1.855940159502423e-06, + "loss": 0.4959, + "num_input_tokens_seen": 50861696, + "step": 3295 + }, + { + "epoch": 13.538461538461538, + "grad_norm": 0.1591796875, + "learning_rate": 1.8529958950142066e-06, + "loss": 0.5215, + "num_input_tokens_seen": 50949568, + "step": 3300 + }, + { + "epoch": 13.558974358974359, + "grad_norm": 0.14453125, + "learning_rate": 1.8500501900807345e-06, + "loss": 0.4692, + "num_input_tokens_seen": 51018432, + "step": 3305 + }, + { + "epoch": 13.57948717948718, + "grad_norm": 0.1435546875, + "learning_rate": 1.847103056722335e-06, + "loss": 0.5015, + "num_input_tokens_seen": 51100352, + "step": 3310 + }, + { + "epoch": 13.6, + "grad_norm": 0.1337890625, + "learning_rate": 1.8441545069651665e-06, + "loss": 0.4779, + "num_input_tokens_seen": 51178144, + "step": 3315 + }, + { + "epoch": 13.62051282051282, + "grad_norm": 0.146484375, + "learning_rate": 1.8412045528411652e-06, + "loss": 0.4971, + "num_input_tokens_seen": 51262432, + "step": 3320 + }, + { + "epoch": 13.64102564102564, + "grad_norm": 0.11279296875, + "learning_rate": 1.8382532063880005e-06, + "loss": 0.5063, + "num_input_tokens_seen": 51342624, + "step": 3325 + }, + { + "epoch": 13.661538461538461, + "grad_norm": 0.1328125, + "learning_rate": 1.8353004796490212e-06, + "loss": 0.4273, + "num_input_tokens_seen": 51413920, + "step": 3330 + }, + { + "epoch": 13.682051282051281, + "grad_norm": 0.14453125, + "learning_rate": 1.8323463846732099e-06, + "loss": 0.5491, + "num_input_tokens_seen": 51501184, + "step": 3335 + }, + { + "epoch": 13.702564102564102, + "grad_norm": 0.2109375, + "learning_rate": 1.8293909335151316e-06, + "loss": 0.4199, + "num_input_tokens_seen": 51570944, + "step": 3340 + }, + { + "epoch": 13.723076923076922, + "grad_norm": 0.236328125, + "learning_rate": 1.826434138234886e-06, + "loss": 0.4468, + "num_input_tokens_seen": 51650784, + "step": 3345 + }, + { + "epoch": 13.743589743589745, + "grad_norm": 0.150390625, + "learning_rate": 1.8234760108980572e-06, + "loss": 0.5225, + "num_input_tokens_seen": 51731008, + "step": 3350 + }, + { + "epoch": 13.764102564102565, + "grad_norm": 0.33203125, + "learning_rate": 1.820516563575665e-06, + "loss": 0.4634, + "num_input_tokens_seen": 51809056, + "step": 3355 + }, + { + "epoch": 13.784615384615385, + "grad_norm": 0.298828125, + "learning_rate": 1.8175558083441164e-06, + "loss": 0.5321, + "num_input_tokens_seen": 51886688, + "step": 3360 + }, + { + "epoch": 13.805128205128206, + "grad_norm": 0.2119140625, + "learning_rate": 1.814593757285154e-06, + "loss": 0.4633, + "num_input_tokens_seen": 51959488, + "step": 3365 + }, + { + "epoch": 13.825641025641026, + "grad_norm": 0.59765625, + "learning_rate": 1.8116304224858092e-06, + "loss": 0.4812, + "num_input_tokens_seen": 52035424, + "step": 3370 + }, + { + "epoch": 13.846153846153847, + "grad_norm": 0.859375, + "learning_rate": 1.8086658160383524e-06, + "loss": 0.506, + "num_input_tokens_seen": 52112576, + "step": 3375 + }, + { + "epoch": 13.866666666666667, + "grad_norm": 0.83984375, + "learning_rate": 1.8056999500402414e-06, + "loss": 0.5091, + "num_input_tokens_seen": 52200480, + "step": 3380 + }, + { + "epoch": 13.887179487179488, + "grad_norm": 0.296875, + "learning_rate": 1.8027328365940755e-06, + "loss": 0.4861, + "num_input_tokens_seen": 52274368, + "step": 3385 + }, + { + "epoch": 13.907692307692308, + "grad_norm": 0.462890625, + "learning_rate": 1.799764487807543e-06, + "loss": 0.5051, + "num_input_tokens_seen": 52350656, + "step": 3390 + }, + { + "epoch": 13.928205128205128, + "grad_norm": 0.279296875, + "learning_rate": 1.7967949157933742e-06, + "loss": 0.4608, + "num_input_tokens_seen": 52427328, + "step": 3395 + }, + { + "epoch": 13.948717948717949, + "grad_norm": 0.3203125, + "learning_rate": 1.7938241326692907e-06, + "loss": 0.4719, + "num_input_tokens_seen": 52503840, + "step": 3400 + }, + { + "epoch": 13.96923076923077, + "grad_norm": 0.291015625, + "learning_rate": 1.7908521505579554e-06, + "loss": 0.5062, + "num_input_tokens_seen": 52577056, + "step": 3405 + }, + { + "epoch": 13.98974358974359, + "grad_norm": 0.310546875, + "learning_rate": 1.7878789815869247e-06, + "loss": 0.5096, + "num_input_tokens_seen": 52656000, + "step": 3410 + }, + { + "epoch": 14.01025641025641, + "grad_norm": 0.291015625, + "learning_rate": 1.7849046378885977e-06, + "loss": 0.4476, + "num_input_tokens_seen": 52725312, + "step": 3415 + }, + { + "epoch": 14.03076923076923, + "grad_norm": 0.361328125, + "learning_rate": 1.7819291316001679e-06, + "loss": 0.436, + "num_input_tokens_seen": 52804736, + "step": 3420 + }, + { + "epoch": 14.051282051282051, + "grad_norm": 0.353515625, + "learning_rate": 1.7789524748635717e-06, + "loss": 0.4761, + "num_input_tokens_seen": 52885056, + "step": 3425 + }, + { + "epoch": 14.071794871794872, + "grad_norm": 0.36328125, + "learning_rate": 1.775974679825441e-06, + "loss": 0.5039, + "num_input_tokens_seen": 52963136, + "step": 3430 + }, + { + "epoch": 14.092307692307692, + "grad_norm": 0.470703125, + "learning_rate": 1.7729957586370525e-06, + "loss": 0.5039, + "num_input_tokens_seen": 53049856, + "step": 3435 + }, + { + "epoch": 14.112820512820512, + "grad_norm": 0.45703125, + "learning_rate": 1.7700157234542773e-06, + "loss": 0.5397, + "num_input_tokens_seen": 53127616, + "step": 3440 + }, + { + "epoch": 14.133333333333333, + "grad_norm": 0.34765625, + "learning_rate": 1.7670345864375339e-06, + "loss": 0.4747, + "num_input_tokens_seen": 53201600, + "step": 3445 + }, + { + "epoch": 14.153846153846153, + "grad_norm": 0.37890625, + "learning_rate": 1.7640523597517357e-06, + "loss": 0.4582, + "num_input_tokens_seen": 53275520, + "step": 3450 + }, + { + "epoch": 14.174358974358974, + "grad_norm": 0.328125, + "learning_rate": 1.7610690555662435e-06, + "loss": 0.4747, + "num_input_tokens_seen": 53346720, + "step": 3455 + }, + { + "epoch": 14.194871794871794, + "grad_norm": 0.337890625, + "learning_rate": 1.7580846860548146e-06, + "loss": 0.4758, + "num_input_tokens_seen": 53420864, + "step": 3460 + }, + { + "epoch": 14.215384615384615, + "grad_norm": 0.380859375, + "learning_rate": 1.7550992633955535e-06, + "loss": 0.4305, + "num_input_tokens_seen": 53490976, + "step": 3465 + }, + { + "epoch": 14.235897435897435, + "grad_norm": 0.357421875, + "learning_rate": 1.7521127997708621e-06, + "loss": 0.4533, + "num_input_tokens_seen": 53570816, + "step": 3470 + }, + { + "epoch": 14.256410256410255, + "grad_norm": 0.30078125, + "learning_rate": 1.7491253073673903e-06, + "loss": 0.5019, + "num_input_tokens_seen": 53647072, + "step": 3475 + }, + { + "epoch": 14.276923076923078, + "grad_norm": 0.28515625, + "learning_rate": 1.7461367983759862e-06, + "loss": 0.4575, + "num_input_tokens_seen": 53727648, + "step": 3480 + }, + { + "epoch": 14.297435897435898, + "grad_norm": 0.28515625, + "learning_rate": 1.7431472849916455e-06, + "loss": 0.4728, + "num_input_tokens_seen": 53800544, + "step": 3485 + }, + { + "epoch": 14.317948717948719, + "grad_norm": 0.298828125, + "learning_rate": 1.7401567794134636e-06, + "loss": 0.4677, + "num_input_tokens_seen": 53875552, + "step": 3490 + }, + { + "epoch": 14.338461538461539, + "grad_norm": 0.216796875, + "learning_rate": 1.737165293844583e-06, + "loss": 0.456, + "num_input_tokens_seen": 53954208, + "step": 3495 + }, + { + "epoch": 14.35897435897436, + "grad_norm": 0.205078125, + "learning_rate": 1.7341728404921471e-06, + "loss": 0.4798, + "num_input_tokens_seen": 54033248, + "step": 3500 + }, + { + "epoch": 14.37948717948718, + "grad_norm": 0.208984375, + "learning_rate": 1.7311794315672477e-06, + "loss": 0.4166, + "num_input_tokens_seen": 54109440, + "step": 3505 + }, + { + "epoch": 14.4, + "grad_norm": 0.1953125, + "learning_rate": 1.7281850792848752e-06, + "loss": 0.4663, + "num_input_tokens_seen": 54186368, + "step": 3510 + }, + { + "epoch": 14.42051282051282, + "grad_norm": 0.1591796875, + "learning_rate": 1.7251897958638704e-06, + "loss": 0.4053, + "num_input_tokens_seen": 54267648, + "step": 3515 + }, + { + "epoch": 14.441025641025641, + "grad_norm": 0.193359375, + "learning_rate": 1.7221935935268735e-06, + "loss": 0.5342, + "num_input_tokens_seen": 54342304, + "step": 3520 + }, + { + "epoch": 14.461538461538462, + "grad_norm": 0.1455078125, + "learning_rate": 1.719196484500274e-06, + "loss": 0.4992, + "num_input_tokens_seen": 54423808, + "step": 3525 + }, + { + "epoch": 14.482051282051282, + "grad_norm": 0.1494140625, + "learning_rate": 1.7161984810141625e-06, + "loss": 0.5067, + "num_input_tokens_seen": 54501760, + "step": 3530 + }, + { + "epoch": 14.502564102564103, + "grad_norm": 0.126953125, + "learning_rate": 1.7131995953022776e-06, + "loss": 0.4286, + "num_input_tokens_seen": 54576512, + "step": 3535 + }, + { + "epoch": 14.523076923076923, + "grad_norm": 0.1611328125, + "learning_rate": 1.7101998396019593e-06, + "loss": 0.436, + "num_input_tokens_seen": 54649696, + "step": 3540 + }, + { + "epoch": 14.543589743589743, + "grad_norm": 0.138671875, + "learning_rate": 1.7071992261540983e-06, + "loss": 0.4889, + "num_input_tokens_seen": 54728864, + "step": 3545 + }, + { + "epoch": 14.564102564102564, + "grad_norm": 0.1328125, + "learning_rate": 1.7041977672030842e-06, + "loss": 0.4585, + "num_input_tokens_seen": 54797856, + "step": 3550 + }, + { + "epoch": 14.584615384615384, + "grad_norm": 0.1064453125, + "learning_rate": 1.7011954749967564e-06, + "loss": 0.4646, + "num_input_tokens_seen": 54875168, + "step": 3555 + }, + { + "epoch": 14.605128205128205, + "grad_norm": 0.1474609375, + "learning_rate": 1.6981923617863566e-06, + "loss": 0.4894, + "num_input_tokens_seen": 54949312, + "step": 3560 + }, + { + "epoch": 14.625641025641025, + "grad_norm": 0.2255859375, + "learning_rate": 1.6951884398264742e-06, + "loss": 0.5159, + "num_input_tokens_seen": 55024768, + "step": 3565 + }, + { + "epoch": 14.646153846153846, + "grad_norm": 0.150390625, + "learning_rate": 1.692183721375001e-06, + "loss": 0.4221, + "num_input_tokens_seen": 55100128, + "step": 3570 + }, + { + "epoch": 14.666666666666666, + "grad_norm": 0.1435546875, + "learning_rate": 1.6891782186930767e-06, + "loss": 0.5417, + "num_input_tokens_seen": 55182080, + "step": 3575 + }, + { + "epoch": 14.687179487179487, + "grad_norm": 0.28515625, + "learning_rate": 1.6861719440450437e-06, + "loss": 0.5072, + "num_input_tokens_seen": 55274048, + "step": 3580 + }, + { + "epoch": 14.707692307692307, + "grad_norm": 0.267578125, + "learning_rate": 1.6831649096983923e-06, + "loss": 0.5008, + "num_input_tokens_seen": 55347648, + "step": 3585 + }, + { + "epoch": 14.728205128205127, + "grad_norm": 0.33203125, + "learning_rate": 1.680157127923715e-06, + "loss": 0.4411, + "num_input_tokens_seen": 55418368, + "step": 3590 + }, + { + "epoch": 14.74871794871795, + "grad_norm": 0.74609375, + "learning_rate": 1.677148610994652e-06, + "loss": 0.4654, + "num_input_tokens_seen": 55490176, + "step": 3595 + }, + { + "epoch": 14.76923076923077, + "grad_norm": 0.70703125, + "learning_rate": 1.6741393711878454e-06, + "loss": 0.4576, + "num_input_tokens_seen": 55570784, + "step": 3600 + }, + { + "epoch": 14.78974358974359, + "grad_norm": 0.2353515625, + "learning_rate": 1.6711294207828852e-06, + "loss": 0.4488, + "num_input_tokens_seen": 55643936, + "step": 3605 + }, + { + "epoch": 14.810256410256411, + "grad_norm": 0.349609375, + "learning_rate": 1.6681187720622627e-06, + "loss": 0.5175, + "num_input_tokens_seen": 55727936, + "step": 3610 + }, + { + "epoch": 14.830769230769231, + "grad_norm": 0.484375, + "learning_rate": 1.6651074373113176e-06, + "loss": 0.4271, + "num_input_tokens_seen": 55804992, + "step": 3615 + }, + { + "epoch": 14.851282051282052, + "grad_norm": 0.3203125, + "learning_rate": 1.66209542881819e-06, + "loss": 0.4631, + "num_input_tokens_seen": 55878976, + "step": 3620 + }, + { + "epoch": 14.871794871794872, + "grad_norm": 0.26171875, + "learning_rate": 1.6590827588737685e-06, + "loss": 0.5453, + "num_input_tokens_seen": 55969376, + "step": 3625 + }, + { + "epoch": 14.892307692307693, + "grad_norm": 0.34765625, + "learning_rate": 1.6560694397716412e-06, + "loss": 0.5086, + "num_input_tokens_seen": 56041024, + "step": 3630 + }, + { + "epoch": 14.912820512820513, + "grad_norm": 0.28515625, + "learning_rate": 1.6530554838080458e-06, + "loss": 0.4602, + "num_input_tokens_seen": 56117152, + "step": 3635 + }, + { + "epoch": 14.933333333333334, + "grad_norm": 0.251953125, + "learning_rate": 1.6500409032818175e-06, + "loss": 0.5046, + "num_input_tokens_seen": 56191264, + "step": 3640 + }, + { + "epoch": 14.953846153846154, + "grad_norm": 0.271484375, + "learning_rate": 1.6470257104943413e-06, + "loss": 0.53, + "num_input_tokens_seen": 56279264, + "step": 3645 + }, + { + "epoch": 14.974358974358974, + "grad_norm": 0.341796875, + "learning_rate": 1.6440099177494991e-06, + "loss": 0.4639, + "num_input_tokens_seen": 56350784, + "step": 3650 + }, + { + "epoch": 14.994871794871795, + "grad_norm": 0.32421875, + "learning_rate": 1.6409935373536227e-06, + "loss": 0.5015, + "num_input_tokens_seen": 56430432, + "step": 3655 + }, + { + "epoch": 15.015384615384615, + "grad_norm": 0.33203125, + "learning_rate": 1.6379765816154413e-06, + "loss": 0.4503, + "num_input_tokens_seen": 56509056, + "step": 3660 + }, + { + "epoch": 15.035897435897436, + "grad_norm": 0.322265625, + "learning_rate": 1.634959062846031e-06, + "loss": 0.4688, + "num_input_tokens_seen": 56578016, + "step": 3665 + }, + { + "epoch": 15.056410256410256, + "grad_norm": 0.34375, + "learning_rate": 1.631940993358767e-06, + "loss": 0.4761, + "num_input_tokens_seen": 56653536, + "step": 3670 + }, + { + "epoch": 15.076923076923077, + "grad_norm": 0.328125, + "learning_rate": 1.6289223854692708e-06, + "loss": 0.4433, + "num_input_tokens_seen": 56730560, + "step": 3675 + }, + { + "epoch": 15.097435897435897, + "grad_norm": 0.3046875, + "learning_rate": 1.6259032514953601e-06, + "loss": 0.4315, + "num_input_tokens_seen": 56813312, + "step": 3680 + }, + { + "epoch": 15.117948717948718, + "grad_norm": 0.283203125, + "learning_rate": 1.6228836037570015e-06, + "loss": 0.4885, + "num_input_tokens_seen": 56891264, + "step": 3685 + }, + { + "epoch": 15.138461538461538, + "grad_norm": 0.400390625, + "learning_rate": 1.619863454576256e-06, + "loss": 0.4703, + "num_input_tokens_seen": 56964864, + "step": 3690 + }, + { + "epoch": 15.158974358974358, + "grad_norm": 0.322265625, + "learning_rate": 1.6168428162772322e-06, + "loss": 0.4518, + "num_input_tokens_seen": 57050144, + "step": 3695 + }, + { + "epoch": 15.179487179487179, + "grad_norm": 0.328125, + "learning_rate": 1.6138217011860336e-06, + "loss": 0.4858, + "num_input_tokens_seen": 57125632, + "step": 3700 + }, + { + "epoch": 15.2, + "grad_norm": 0.296875, + "learning_rate": 1.6108001216307107e-06, + "loss": 0.459, + "num_input_tokens_seen": 57204128, + "step": 3705 + }, + { + "epoch": 15.22051282051282, + "grad_norm": 0.33984375, + "learning_rate": 1.6077780899412068e-06, + "loss": 0.4639, + "num_input_tokens_seen": 57281536, + "step": 3710 + }, + { + "epoch": 15.24102564102564, + "grad_norm": 0.27734375, + "learning_rate": 1.6047556184493133e-06, + "loss": 0.4496, + "num_input_tokens_seen": 57354816, + "step": 3715 + }, + { + "epoch": 15.261538461538462, + "grad_norm": 0.22265625, + "learning_rate": 1.601732719488614e-06, + "loss": 0.5043, + "num_input_tokens_seen": 57435328, + "step": 3720 + }, + { + "epoch": 15.282051282051283, + "grad_norm": 0.2109375, + "learning_rate": 1.5987094053944384e-06, + "loss": 0.4357, + "num_input_tokens_seen": 57507424, + "step": 3725 + }, + { + "epoch": 15.302564102564103, + "grad_norm": 0.228515625, + "learning_rate": 1.5956856885038086e-06, + "loss": 0.5211, + "num_input_tokens_seen": 57586880, + "step": 3730 + }, + { + "epoch": 15.323076923076924, + "grad_norm": 0.1982421875, + "learning_rate": 1.592661581155392e-06, + "loss": 0.5036, + "num_input_tokens_seen": 57669600, + "step": 3735 + }, + { + "epoch": 15.343589743589744, + "grad_norm": 0.1708984375, + "learning_rate": 1.5896370956894477e-06, + "loss": 0.4751, + "num_input_tokens_seen": 57744608, + "step": 3740 + }, + { + "epoch": 15.364102564102565, + "grad_norm": 0.1533203125, + "learning_rate": 1.5866122444477794e-06, + "loss": 0.4764, + "num_input_tokens_seen": 57818496, + "step": 3745 + }, + { + "epoch": 15.384615384615385, + "grad_norm": 0.1376953125, + "learning_rate": 1.5835870397736817e-06, + "loss": 0.4551, + "num_input_tokens_seen": 57898432, + "step": 3750 + }, + { + "epoch": 15.405128205128205, + "grad_norm": 0.166015625, + "learning_rate": 1.5805614940118928e-06, + "loss": 0.5198, + "num_input_tokens_seen": 57974624, + "step": 3755 + }, + { + "epoch": 15.425641025641026, + "grad_norm": 0.1337890625, + "learning_rate": 1.577535619508542e-06, + "loss": 0.5164, + "num_input_tokens_seen": 58052576, + "step": 3760 + }, + { + "epoch": 15.446153846153846, + "grad_norm": 0.130859375, + "learning_rate": 1.5745094286111004e-06, + "loss": 0.4699, + "num_input_tokens_seen": 58126592, + "step": 3765 + }, + { + "epoch": 15.466666666666667, + "grad_norm": 0.125, + "learning_rate": 1.5714829336683297e-06, + "loss": 0.4804, + "num_input_tokens_seen": 58214784, + "step": 3770 + }, + { + "epoch": 15.487179487179487, + "grad_norm": 0.1259765625, + "learning_rate": 1.5684561470302337e-06, + "loss": 0.4731, + "num_input_tokens_seen": 58284480, + "step": 3775 + }, + { + "epoch": 15.507692307692308, + "grad_norm": 0.1279296875, + "learning_rate": 1.5654290810480041e-06, + "loss": 0.5115, + "num_input_tokens_seen": 58359488, + "step": 3780 + }, + { + "epoch": 15.528205128205128, + "grad_norm": 0.1201171875, + "learning_rate": 1.562401748073975e-06, + "loss": 0.4713, + "num_input_tokens_seen": 58433632, + "step": 3785 + }, + { + "epoch": 15.548717948717949, + "grad_norm": 0.275390625, + "learning_rate": 1.5593741604615679e-06, + "loss": 0.5218, + "num_input_tokens_seen": 58526048, + "step": 3790 + }, + { + "epoch": 15.569230769230769, + "grad_norm": 0.15234375, + "learning_rate": 1.5563463305652454e-06, + "loss": 0.4432, + "num_input_tokens_seen": 58593440, + "step": 3795 + }, + { + "epoch": 15.58974358974359, + "grad_norm": 0.1552734375, + "learning_rate": 1.5533182707404563e-06, + "loss": 0.4793, + "num_input_tokens_seen": 58664832, + "step": 3800 + }, + { + "epoch": 15.61025641025641, + "grad_norm": 0.29296875, + "learning_rate": 1.550289993343591e-06, + "loss": 0.4287, + "num_input_tokens_seen": 58734976, + "step": 3805 + }, + { + "epoch": 15.63076923076923, + "grad_norm": 0.37109375, + "learning_rate": 1.547261510731924e-06, + "loss": 0.5701, + "num_input_tokens_seen": 58826304, + "step": 3810 + }, + { + "epoch": 15.65128205128205, + "grad_norm": 0.515625, + "learning_rate": 1.5442328352635706e-06, + "loss": 0.4858, + "num_input_tokens_seen": 58897216, + "step": 3815 + }, + { + "epoch": 15.671794871794871, + "grad_norm": 0.63671875, + "learning_rate": 1.54120397929743e-06, + "loss": 0.4541, + "num_input_tokens_seen": 58970880, + "step": 3820 + }, + { + "epoch": 15.692307692307692, + "grad_norm": 0.91015625, + "learning_rate": 1.5381749551931405e-06, + "loss": 0.5178, + "num_input_tokens_seen": 59049408, + "step": 3825 + }, + { + "epoch": 15.712820512820512, + "grad_norm": 0.298828125, + "learning_rate": 1.5351457753110244e-06, + "loss": 0.4799, + "num_input_tokens_seen": 59122336, + "step": 3830 + }, + { + "epoch": 15.733333333333333, + "grad_norm": 0.2333984375, + "learning_rate": 1.5321164520120421e-06, + "loss": 0.4284, + "num_input_tokens_seen": 59197600, + "step": 3835 + }, + { + "epoch": 15.753846153846155, + "grad_norm": 0.640625, + "learning_rate": 1.5290869976577366e-06, + "loss": 0.4343, + "num_input_tokens_seen": 59268224, + "step": 3840 + }, + { + "epoch": 15.774358974358975, + "grad_norm": 0.279296875, + "learning_rate": 1.5260574246101875e-06, + "loss": 0.4371, + "num_input_tokens_seen": 59341408, + "step": 3845 + }, + { + "epoch": 15.794871794871796, + "grad_norm": 0.27734375, + "learning_rate": 1.5230277452319585e-06, + "loss": 0.5017, + "num_input_tokens_seen": 59423648, + "step": 3850 + }, + { + "epoch": 15.815384615384616, + "grad_norm": 0.35546875, + "learning_rate": 1.519997971886046e-06, + "loss": 0.4473, + "num_input_tokens_seen": 59495872, + "step": 3855 + }, + { + "epoch": 15.835897435897436, + "grad_norm": 0.31640625, + "learning_rate": 1.5169681169358314e-06, + "loss": 0.4693, + "num_input_tokens_seen": 59574880, + "step": 3860 + }, + { + "epoch": 15.856410256410257, + "grad_norm": 0.3046875, + "learning_rate": 1.513938192745028e-06, + "loss": 0.4955, + "num_input_tokens_seen": 59657472, + "step": 3865 + }, + { + "epoch": 15.876923076923077, + "grad_norm": 0.337890625, + "learning_rate": 1.5109082116776328e-06, + "loss": 0.4219, + "num_input_tokens_seen": 59733888, + "step": 3870 + }, + { + "epoch": 15.897435897435898, + "grad_norm": 0.423828125, + "learning_rate": 1.5078781860978732e-06, + "loss": 0.4462, + "num_input_tokens_seen": 59811392, + "step": 3875 + }, + { + "epoch": 15.917948717948718, + "grad_norm": 0.359375, + "learning_rate": 1.5048481283701594e-06, + "loss": 0.4938, + "num_input_tokens_seen": 59886784, + "step": 3880 + }, + { + "epoch": 15.938461538461539, + "grad_norm": 0.390625, + "learning_rate": 1.501818050859033e-06, + "loss": 0.4465, + "num_input_tokens_seen": 59959072, + "step": 3885 + }, + { + "epoch": 15.95897435897436, + "grad_norm": 0.408203125, + "learning_rate": 1.498787965929116e-06, + "loss": 0.462, + "num_input_tokens_seen": 60032832, + "step": 3890 + }, + { + "epoch": 15.97948717948718, + "grad_norm": 0.28515625, + "learning_rate": 1.49575788594506e-06, + "loss": 0.5279, + "num_input_tokens_seen": 60119456, + "step": 3895 + }, + { + "epoch": 16.0, + "grad_norm": 0.30859375, + "learning_rate": 1.4927278232714975e-06, + "loss": 0.4908, + "num_input_tokens_seen": 60204448, + "step": 3900 + }, + { + "epoch": 16.02051282051282, + "grad_norm": 0.37109375, + "learning_rate": 1.4896977902729893e-06, + "loss": 0.4607, + "num_input_tokens_seen": 60279104, + "step": 3905 + }, + { + "epoch": 16.04102564102564, + "grad_norm": 0.30859375, + "learning_rate": 1.4866677993139761e-06, + "loss": 0.4405, + "num_input_tokens_seen": 60349504, + "step": 3910 + }, + { + "epoch": 16.06153846153846, + "grad_norm": 0.34375, + "learning_rate": 1.4836378627587266e-06, + "loss": 0.4744, + "num_input_tokens_seen": 60420640, + "step": 3915 + }, + { + "epoch": 16.08205128205128, + "grad_norm": 0.32421875, + "learning_rate": 1.4806079929712874e-06, + "loss": 0.4868, + "num_input_tokens_seen": 60497728, + "step": 3920 + }, + { + "epoch": 16.102564102564102, + "grad_norm": 0.341796875, + "learning_rate": 1.477578202315433e-06, + "loss": 0.5095, + "num_input_tokens_seen": 60579008, + "step": 3925 + }, + { + "epoch": 16.123076923076923, + "grad_norm": 0.271484375, + "learning_rate": 1.4745485031546143e-06, + "loss": 0.5101, + "num_input_tokens_seen": 60657088, + "step": 3930 + }, + { + "epoch": 16.143589743589743, + "grad_norm": 0.28515625, + "learning_rate": 1.4715189078519094e-06, + "loss": 0.4938, + "num_input_tokens_seen": 60734560, + "step": 3935 + }, + { + "epoch": 16.164102564102564, + "grad_norm": 0.265625, + "learning_rate": 1.468489428769973e-06, + "loss": 0.4785, + "num_input_tokens_seen": 60812960, + "step": 3940 + }, + { + "epoch": 16.184615384615384, + "grad_norm": 0.2138671875, + "learning_rate": 1.4654600782709843e-06, + "loss": 0.4953, + "num_input_tokens_seen": 60892768, + "step": 3945 + }, + { + "epoch": 16.205128205128204, + "grad_norm": 0.208984375, + "learning_rate": 1.4624308687165985e-06, + "loss": 0.4472, + "num_input_tokens_seen": 60964096, + "step": 3950 + }, + { + "epoch": 16.225641025641025, + "grad_norm": 0.1943359375, + "learning_rate": 1.4594018124678965e-06, + "loss": 0.4191, + "num_input_tokens_seen": 61034304, + "step": 3955 + }, + { + "epoch": 16.246153846153845, + "grad_norm": 0.181640625, + "learning_rate": 1.4563729218853329e-06, + "loss": 0.465, + "num_input_tokens_seen": 61113024, + "step": 3960 + }, + { + "epoch": 16.266666666666666, + "grad_norm": 0.1787109375, + "learning_rate": 1.4533442093286853e-06, + "loss": 0.4703, + "num_input_tokens_seen": 61184256, + "step": 3965 + }, + { + "epoch": 16.287179487179486, + "grad_norm": 0.1416015625, + "learning_rate": 1.4503156871570062e-06, + "loss": 0.4773, + "num_input_tokens_seen": 61262208, + "step": 3970 + }, + { + "epoch": 16.307692307692307, + "grad_norm": 0.1630859375, + "learning_rate": 1.4472873677285706e-06, + "loss": 0.4172, + "num_input_tokens_seen": 61334560, + "step": 3975 + }, + { + "epoch": 16.328205128205127, + "grad_norm": 0.1484375, + "learning_rate": 1.4442592634008266e-06, + "loss": 0.4326, + "num_input_tokens_seen": 61405888, + "step": 3980 + }, + { + "epoch": 16.348717948717947, + "grad_norm": 0.1240234375, + "learning_rate": 1.4412313865303438e-06, + "loss": 0.5303, + "num_input_tokens_seen": 61489280, + "step": 3985 + }, + { + "epoch": 16.369230769230768, + "grad_norm": 0.16015625, + "learning_rate": 1.4382037494727649e-06, + "loss": 0.4726, + "num_input_tokens_seen": 61562912, + "step": 3990 + }, + { + "epoch": 16.38974358974359, + "grad_norm": 0.1328125, + "learning_rate": 1.435176364582752e-06, + "loss": 0.458, + "num_input_tokens_seen": 61652064, + "step": 3995 + }, + { + "epoch": 16.41025641025641, + "grad_norm": 0.1171875, + "learning_rate": 1.4321492442139405e-06, + "loss": 0.5266, + "num_input_tokens_seen": 61737536, + "step": 4000 + }, + { + "epoch": 16.43076923076923, + "grad_norm": 0.158203125, + "learning_rate": 1.4291224007188849e-06, + "loss": 0.4394, + "num_input_tokens_seen": 61811840, + "step": 4005 + }, + { + "epoch": 16.45128205128205, + "grad_norm": 0.142578125, + "learning_rate": 1.4260958464490104e-06, + "loss": 0.4474, + "num_input_tokens_seen": 61884288, + "step": 4010 + }, + { + "epoch": 16.47179487179487, + "grad_norm": 0.22265625, + "learning_rate": 1.4230695937545616e-06, + "loss": 0.5, + "num_input_tokens_seen": 61958944, + "step": 4015 + }, + { + "epoch": 16.49230769230769, + "grad_norm": 0.1455078125, + "learning_rate": 1.420043654984553e-06, + "loss": 0.4766, + "num_input_tokens_seen": 62038016, + "step": 4020 + }, + { + "epoch": 16.51282051282051, + "grad_norm": 0.2890625, + "learning_rate": 1.4170180424867176e-06, + "loss": 0.4593, + "num_input_tokens_seen": 62107616, + "step": 4025 + }, + { + "epoch": 16.533333333333335, + "grad_norm": 0.359375, + "learning_rate": 1.4139927686074577e-06, + "loss": 0.4446, + "num_input_tokens_seen": 62188352, + "step": 4030 + }, + { + "epoch": 16.553846153846155, + "grad_norm": 0.259765625, + "learning_rate": 1.4109678456917926e-06, + "loss": 0.4658, + "num_input_tokens_seen": 62259456, + "step": 4035 + }, + { + "epoch": 16.574358974358976, + "grad_norm": 0.50390625, + "learning_rate": 1.4079432860833104e-06, + "loss": 0.5363, + "num_input_tokens_seen": 62356448, + "step": 4040 + }, + { + "epoch": 16.594871794871796, + "grad_norm": 0.78515625, + "learning_rate": 1.4049191021241176e-06, + "loss": 0.5006, + "num_input_tokens_seen": 62429824, + "step": 4045 + }, + { + "epoch": 16.615384615384617, + "grad_norm": 0.66796875, + "learning_rate": 1.4018953061547853e-06, + "loss": 0.4638, + "num_input_tokens_seen": 62514528, + "step": 4050 + }, + { + "epoch": 16.635897435897437, + "grad_norm": 0.267578125, + "learning_rate": 1.3988719105143038e-06, + "loss": 0.4413, + "num_input_tokens_seen": 62593472, + "step": 4055 + }, + { + "epoch": 16.656410256410258, + "grad_norm": 0.48046875, + "learning_rate": 1.395848927540028e-06, + "loss": 0.4506, + "num_input_tokens_seen": 62671520, + "step": 4060 + }, + { + "epoch": 16.676923076923078, + "grad_norm": 0.5078125, + "learning_rate": 1.39282636956763e-06, + "loss": 0.4616, + "num_input_tokens_seen": 62749856, + "step": 4065 + }, + { + "epoch": 16.6974358974359, + "grad_norm": 0.32421875, + "learning_rate": 1.3898042489310471e-06, + "loss": 0.456, + "num_input_tokens_seen": 62823616, + "step": 4070 + }, + { + "epoch": 16.71794871794872, + "grad_norm": 0.357421875, + "learning_rate": 1.3867825779624325e-06, + "loss": 0.45, + "num_input_tokens_seen": 62908416, + "step": 4075 + }, + { + "epoch": 16.73846153846154, + "grad_norm": 0.41015625, + "learning_rate": 1.3837613689921037e-06, + "loss": 0.4563, + "num_input_tokens_seen": 62978016, + "step": 4080 + }, + { + "epoch": 16.75897435897436, + "grad_norm": 0.259765625, + "learning_rate": 1.380740634348494e-06, + "loss": 0.4952, + "num_input_tokens_seen": 63065952, + "step": 4085 + }, + { + "epoch": 16.77948717948718, + "grad_norm": 0.3359375, + "learning_rate": 1.3777203863580993e-06, + "loss": 0.4327, + "num_input_tokens_seen": 63141792, + "step": 4090 + }, + { + "epoch": 16.8, + "grad_norm": 0.361328125, + "learning_rate": 1.3747006373454321e-06, + "loss": 0.5152, + "num_input_tokens_seen": 63223520, + "step": 4095 + }, + { + "epoch": 16.82051282051282, + "grad_norm": 0.326171875, + "learning_rate": 1.3716813996329669e-06, + "loss": 0.4952, + "num_input_tokens_seen": 63297696, + "step": 4100 + }, + { + "epoch": 16.84102564102564, + "grad_norm": 0.4609375, + "learning_rate": 1.3686626855410929e-06, + "loss": 0.4873, + "num_input_tokens_seen": 63378304, + "step": 4105 + }, + { + "epoch": 16.861538461538462, + "grad_norm": 0.388671875, + "learning_rate": 1.3656445073880615e-06, + "loss": 0.4461, + "num_input_tokens_seen": 63454400, + "step": 4110 + }, + { + "epoch": 16.882051282051282, + "grad_norm": 0.3671875, + "learning_rate": 1.3626268774899381e-06, + "loss": 0.4612, + "num_input_tokens_seen": 63526400, + "step": 4115 + }, + { + "epoch": 16.902564102564103, + "grad_norm": 0.33984375, + "learning_rate": 1.3596098081605505e-06, + "loss": 0.5061, + "num_input_tokens_seen": 63612448, + "step": 4120 + }, + { + "epoch": 16.923076923076923, + "grad_norm": 0.34765625, + "learning_rate": 1.3565933117114385e-06, + "loss": 0.486, + "num_input_tokens_seen": 63693248, + "step": 4125 + }, + { + "epoch": 16.943589743589744, + "grad_norm": 0.35546875, + "learning_rate": 1.3535774004518057e-06, + "loss": 0.4772, + "num_input_tokens_seen": 63765504, + "step": 4130 + }, + { + "epoch": 16.964102564102564, + "grad_norm": 0.330078125, + "learning_rate": 1.3505620866884666e-06, + "loss": 0.457, + "num_input_tokens_seen": 63839616, + "step": 4135 + }, + { + "epoch": 16.984615384615385, + "grad_norm": 0.341796875, + "learning_rate": 1.347547382725797e-06, + "loss": 0.4961, + "num_input_tokens_seen": 63924736, + "step": 4140 + }, + { + "epoch": 17.005128205128205, + "grad_norm": 0.349609375, + "learning_rate": 1.344533300865686e-06, + "loss": 0.5218, + "num_input_tokens_seen": 64000256, + "step": 4145 + }, + { + "epoch": 17.025641025641026, + "grad_norm": 0.44921875, + "learning_rate": 1.3415198534074823e-06, + "loss": 0.4501, + "num_input_tokens_seen": 64077472, + "step": 4150 + }, + { + "epoch": 17.046153846153846, + "grad_norm": 0.2578125, + "learning_rate": 1.3385070526479475e-06, + "loss": 0.4704, + "num_input_tokens_seen": 64157024, + "step": 4155 + }, + { + "epoch": 17.066666666666666, + "grad_norm": 0.2734375, + "learning_rate": 1.3354949108812026e-06, + "loss": 0.485, + "num_input_tokens_seen": 64226656, + "step": 4160 + }, + { + "epoch": 17.087179487179487, + "grad_norm": 0.2236328125, + "learning_rate": 1.3324834403986815e-06, + "loss": 0.4316, + "num_input_tokens_seen": 64302816, + "step": 4165 + }, + { + "epoch": 17.107692307692307, + "grad_norm": 0.2373046875, + "learning_rate": 1.3294726534890766e-06, + "loss": 0.4781, + "num_input_tokens_seen": 64371840, + "step": 4170 + }, + { + "epoch": 17.128205128205128, + "grad_norm": 0.189453125, + "learning_rate": 1.326462562438293e-06, + "loss": 0.4331, + "num_input_tokens_seen": 64451776, + "step": 4175 + }, + { + "epoch": 17.148717948717948, + "grad_norm": 0.2080078125, + "learning_rate": 1.3234531795293945e-06, + "loss": 0.4754, + "num_input_tokens_seen": 64523712, + "step": 4180 + }, + { + "epoch": 17.16923076923077, + "grad_norm": 0.1904296875, + "learning_rate": 1.3204445170425565e-06, + "loss": 0.4983, + "num_input_tokens_seen": 64594272, + "step": 4185 + }, + { + "epoch": 17.18974358974359, + "grad_norm": 0.1650390625, + "learning_rate": 1.3174365872550138e-06, + "loss": 0.4852, + "num_input_tokens_seen": 64677280, + "step": 4190 + }, + { + "epoch": 17.21025641025641, + "grad_norm": 0.1591796875, + "learning_rate": 1.3144294024410122e-06, + "loss": 0.4843, + "num_input_tokens_seen": 64758080, + "step": 4195 + }, + { + "epoch": 17.23076923076923, + "grad_norm": 0.1435546875, + "learning_rate": 1.3114229748717563e-06, + "loss": 0.4564, + "num_input_tokens_seen": 64830848, + "step": 4200 + }, + { + "epoch": 17.25128205128205, + "grad_norm": 0.1474609375, + "learning_rate": 1.308417316815362e-06, + "loss": 0.4432, + "num_input_tokens_seen": 64907296, + "step": 4205 + }, + { + "epoch": 17.27179487179487, + "grad_norm": 0.134765625, + "learning_rate": 1.3054124405368036e-06, + "loss": 0.4901, + "num_input_tokens_seen": 64987392, + "step": 4210 + }, + { + "epoch": 17.29230769230769, + "grad_norm": 0.1318359375, + "learning_rate": 1.3024083582978668e-06, + "loss": 0.5076, + "num_input_tokens_seen": 65077408, + "step": 4215 + }, + { + "epoch": 17.31282051282051, + "grad_norm": 0.12158203125, + "learning_rate": 1.2994050823570968e-06, + "loss": 0.4818, + "num_input_tokens_seen": 65159104, + "step": 4220 + }, + { + "epoch": 17.333333333333332, + "grad_norm": 0.099609375, + "learning_rate": 1.2964026249697475e-06, + "loss": 0.4525, + "num_input_tokens_seen": 65242816, + "step": 4225 + }, + { + "epoch": 17.353846153846153, + "grad_norm": 0.1337890625, + "learning_rate": 1.293400998387734e-06, + "loss": 0.4444, + "num_input_tokens_seen": 65317376, + "step": 4230 + }, + { + "epoch": 17.374358974358973, + "grad_norm": 0.2314453125, + "learning_rate": 1.2904002148595797e-06, + "loss": 0.4897, + "num_input_tokens_seen": 65391264, + "step": 4235 + }, + { + "epoch": 17.394871794871793, + "grad_norm": 0.208984375, + "learning_rate": 1.2874002866303695e-06, + "loss": 0.5116, + "num_input_tokens_seen": 65467744, + "step": 4240 + }, + { + "epoch": 17.415384615384614, + "grad_norm": 0.1298828125, + "learning_rate": 1.2844012259416965e-06, + "loss": 0.4498, + "num_input_tokens_seen": 65535680, + "step": 4245 + }, + { + "epoch": 17.435897435897434, + "grad_norm": 0.2890625, + "learning_rate": 1.2814030450316151e-06, + "loss": 0.4984, + "num_input_tokens_seen": 65611680, + "step": 4250 + }, + { + "epoch": 17.456410256410255, + "grad_norm": 0.318359375, + "learning_rate": 1.2784057561345885e-06, + "loss": 0.5276, + "num_input_tokens_seen": 65684160, + "step": 4255 + }, + { + "epoch": 17.476923076923075, + "grad_norm": 0.291015625, + "learning_rate": 1.2754093714814407e-06, + "loss": 0.4434, + "num_input_tokens_seen": 65755104, + "step": 4260 + }, + { + "epoch": 17.4974358974359, + "grad_norm": 0.59765625, + "learning_rate": 1.2724139032993057e-06, + "loss": 0.4743, + "num_input_tokens_seen": 65835072, + "step": 4265 + }, + { + "epoch": 17.51794871794872, + "grad_norm": 0.89453125, + "learning_rate": 1.269419363811577e-06, + "loss": 0.4691, + "num_input_tokens_seen": 65916704, + "step": 4270 + }, + { + "epoch": 17.53846153846154, + "grad_norm": 0.6484375, + "learning_rate": 1.2664257652378598e-06, + "loss": 0.5129, + "num_input_tokens_seen": 65989696, + "step": 4275 + }, + { + "epoch": 17.55897435897436, + "grad_norm": 0.259765625, + "learning_rate": 1.2634331197939183e-06, + "loss": 0.4868, + "num_input_tokens_seen": 66071744, + "step": 4280 + }, + { + "epoch": 17.57948717948718, + "grad_norm": 0.54296875, + "learning_rate": 1.2604414396916286e-06, + "loss": 0.489, + "num_input_tokens_seen": 66153024, + "step": 4285 + }, + { + "epoch": 17.6, + "grad_norm": 0.55078125, + "learning_rate": 1.2574507371389267e-06, + "loss": 0.4855, + "num_input_tokens_seen": 66220384, + "step": 4290 + }, + { + "epoch": 17.620512820512822, + "grad_norm": 0.298828125, + "learning_rate": 1.25446102433976e-06, + "loss": 0.4791, + "num_input_tokens_seen": 66298912, + "step": 4295 + }, + { + "epoch": 17.641025641025642, + "grad_norm": 0.263671875, + "learning_rate": 1.2514723134940365e-06, + "loss": 0.4563, + "num_input_tokens_seen": 66373120, + "step": 4300 + }, + { + "epoch": 17.661538461538463, + "grad_norm": 0.26171875, + "learning_rate": 1.2484846167975767e-06, + "loss": 0.4577, + "num_input_tokens_seen": 66452096, + "step": 4305 + }, + { + "epoch": 17.682051282051283, + "grad_norm": 0.302734375, + "learning_rate": 1.2454979464420624e-06, + "loss": 0.4875, + "num_input_tokens_seen": 66530784, + "step": 4310 + }, + { + "epoch": 17.702564102564104, + "grad_norm": 0.34375, + "learning_rate": 1.2425123146149863e-06, + "loss": 0.4551, + "num_input_tokens_seen": 66607264, + "step": 4315 + }, + { + "epoch": 17.723076923076924, + "grad_norm": 0.328125, + "learning_rate": 1.2395277334996047e-06, + "loss": 0.5378, + "num_input_tokens_seen": 66692608, + "step": 4320 + }, + { + "epoch": 17.743589743589745, + "grad_norm": 0.341796875, + "learning_rate": 1.2365442152748846e-06, + "loss": 0.4518, + "num_input_tokens_seen": 66773504, + "step": 4325 + }, + { + "epoch": 17.764102564102565, + "grad_norm": 0.625, + "learning_rate": 1.2335617721154577e-06, + "loss": 0.4412, + "num_input_tokens_seen": 66844928, + "step": 4330 + }, + { + "epoch": 17.784615384615385, + "grad_norm": 0.443359375, + "learning_rate": 1.2305804161915671e-06, + "loss": 0.5214, + "num_input_tokens_seen": 66918080, + "step": 4335 + }, + { + "epoch": 17.805128205128206, + "grad_norm": 0.32421875, + "learning_rate": 1.2276001596690205e-06, + "loss": 0.47, + "num_input_tokens_seen": 66988992, + "step": 4340 + }, + { + "epoch": 17.825641025641026, + "grad_norm": 0.419921875, + "learning_rate": 1.2246210147091382e-06, + "loss": 0.4545, + "num_input_tokens_seen": 67068288, + "step": 4345 + }, + { + "epoch": 17.846153846153847, + "grad_norm": 0.3125, + "learning_rate": 1.2216429934687062e-06, + "loss": 0.4574, + "num_input_tokens_seen": 67142560, + "step": 4350 + }, + { + "epoch": 17.866666666666667, + "grad_norm": 0.40234375, + "learning_rate": 1.2186661080999234e-06, + "loss": 0.4717, + "num_input_tokens_seen": 67222560, + "step": 4355 + }, + { + "epoch": 17.887179487179488, + "grad_norm": 0.318359375, + "learning_rate": 1.2156903707503544e-06, + "loss": 0.4662, + "num_input_tokens_seen": 67301760, + "step": 4360 + }, + { + "epoch": 17.907692307692308, + "grad_norm": 0.345703125, + "learning_rate": 1.2127157935628789e-06, + "loss": 0.4958, + "num_input_tokens_seen": 67381856, + "step": 4365 + }, + { + "epoch": 17.92820512820513, + "grad_norm": 0.318359375, + "learning_rate": 1.2097423886756433e-06, + "loss": 0.5327, + "num_input_tokens_seen": 67467968, + "step": 4370 + }, + { + "epoch": 17.94871794871795, + "grad_norm": 0.248046875, + "learning_rate": 1.2067701682220084e-06, + "loss": 0.4606, + "num_input_tokens_seen": 67548672, + "step": 4375 + }, + { + "epoch": 17.96923076923077, + "grad_norm": 0.255859375, + "learning_rate": 1.2037991443305043e-06, + "loss": 0.4451, + "num_input_tokens_seen": 67625248, + "step": 4380 + }, + { + "epoch": 17.98974358974359, + "grad_norm": 0.28125, + "learning_rate": 1.2008293291247754e-06, + "loss": 0.4416, + "num_input_tokens_seen": 67695808, + "step": 4385 + }, + { + "epoch": 18.01025641025641, + "grad_norm": 0.23046875, + "learning_rate": 1.1978607347235367e-06, + "loss": 0.4741, + "num_input_tokens_seen": 67774400, + "step": 4390 + }, + { + "epoch": 18.03076923076923, + "grad_norm": 0.236328125, + "learning_rate": 1.1948933732405205e-06, + "loss": 0.446, + "num_input_tokens_seen": 67846848, + "step": 4395 + }, + { + "epoch": 18.05128205128205, + "grad_norm": 0.2060546875, + "learning_rate": 1.191927256784427e-06, + "loss": 0.4315, + "num_input_tokens_seen": 67917440, + "step": 4400 + }, + { + "epoch": 18.07179487179487, + "grad_norm": 0.181640625, + "learning_rate": 1.1889623974588772e-06, + "loss": 0.4055, + "num_input_tokens_seen": 67990656, + "step": 4405 + }, + { + "epoch": 18.092307692307692, + "grad_norm": 0.1591796875, + "learning_rate": 1.185998807362362e-06, + "loss": 0.499, + "num_input_tokens_seen": 68067872, + "step": 4410 + }, + { + "epoch": 18.112820512820512, + "grad_norm": 0.162109375, + "learning_rate": 1.1830364985881924e-06, + "loss": 0.481, + "num_input_tokens_seen": 68145632, + "step": 4415 + }, + { + "epoch": 18.133333333333333, + "grad_norm": 0.1650390625, + "learning_rate": 1.1800754832244515e-06, + "loss": 0.4931, + "num_input_tokens_seen": 68221472, + "step": 4420 + }, + { + "epoch": 18.153846153846153, + "grad_norm": 0.1513671875, + "learning_rate": 1.1771157733539442e-06, + "loss": 0.4377, + "num_input_tokens_seen": 68292352, + "step": 4425 + }, + { + "epoch": 18.174358974358974, + "grad_norm": 0.150390625, + "learning_rate": 1.174157381054148e-06, + "loss": 0.5265, + "num_input_tokens_seen": 68374880, + "step": 4430 + }, + { + "epoch": 18.194871794871794, + "grad_norm": 0.1416015625, + "learning_rate": 1.1712003183971644e-06, + "loss": 0.5103, + "num_input_tokens_seen": 68457280, + "step": 4435 + }, + { + "epoch": 18.215384615384615, + "grad_norm": 0.119140625, + "learning_rate": 1.1682445974496686e-06, + "loss": 0.5133, + "num_input_tokens_seen": 68540224, + "step": 4440 + }, + { + "epoch": 18.235897435897435, + "grad_norm": 0.11865234375, + "learning_rate": 1.1652902302728607e-06, + "loss": 0.4805, + "num_input_tokens_seen": 68621056, + "step": 4445 + }, + { + "epoch": 18.256410256410255, + "grad_norm": 0.12353515625, + "learning_rate": 1.1623372289224172e-06, + "loss": 0.4863, + "num_input_tokens_seen": 68697440, + "step": 4450 + }, + { + "epoch": 18.276923076923076, + "grad_norm": 0.125, + "learning_rate": 1.1593856054484403e-06, + "loss": 0.4685, + "num_input_tokens_seen": 68778944, + "step": 4455 + }, + { + "epoch": 18.297435897435896, + "grad_norm": 0.1982421875, + "learning_rate": 1.156435371895411e-06, + "loss": 0.4934, + "num_input_tokens_seen": 68859712, + "step": 4460 + }, + { + "epoch": 18.317948717948717, + "grad_norm": 0.1962890625, + "learning_rate": 1.1534865403021366e-06, + "loss": 0.4439, + "num_input_tokens_seen": 68932512, + "step": 4465 + }, + { + "epoch": 18.338461538461537, + "grad_norm": 0.15625, + "learning_rate": 1.1505391227017046e-06, + "loss": 0.4295, + "num_input_tokens_seen": 69006080, + "step": 4470 + }, + { + "epoch": 18.358974358974358, + "grad_norm": 0.29296875, + "learning_rate": 1.1475931311214338e-06, + "loss": 0.4644, + "num_input_tokens_seen": 69081984, + "step": 4475 + }, + { + "epoch": 18.379487179487178, + "grad_norm": 0.2421875, + "learning_rate": 1.144648577582821e-06, + "loss": 0.4899, + "num_input_tokens_seen": 69158240, + "step": 4480 + }, + { + "epoch": 18.4, + "grad_norm": 0.3671875, + "learning_rate": 1.141705474101498e-06, + "loss": 0.4763, + "num_input_tokens_seen": 69241184, + "step": 4485 + }, + { + "epoch": 18.42051282051282, + "grad_norm": 0.5859375, + "learning_rate": 1.138763832687177e-06, + "loss": 0.4784, + "num_input_tokens_seen": 69326112, + "step": 4490 + }, + { + "epoch": 18.44102564102564, + "grad_norm": 0.765625, + "learning_rate": 1.1358236653436052e-06, + "loss": 0.4624, + "num_input_tokens_seen": 69400832, + "step": 4495 + }, + { + "epoch": 18.46153846153846, + "grad_norm": 0.8984375, + "learning_rate": 1.1328849840685143e-06, + "loss": 0.4523, + "num_input_tokens_seen": 69478464, + "step": 4500 + }, + { + "epoch": 18.48205128205128, + "grad_norm": 0.43359375, + "learning_rate": 1.1299478008535726e-06, + "loss": 0.5372, + "num_input_tokens_seen": 69559712, + "step": 4505 + }, + { + "epoch": 18.5025641025641, + "grad_norm": 0.458984375, + "learning_rate": 1.1270121276843342e-06, + "loss": 0.4224, + "num_input_tokens_seen": 69638528, + "step": 4510 + }, + { + "epoch": 18.523076923076925, + "grad_norm": 0.23828125, + "learning_rate": 1.1240779765401926e-06, + "loss": 0.4726, + "num_input_tokens_seen": 69719264, + "step": 4515 + }, + { + "epoch": 18.543589743589745, + "grad_norm": 0.30078125, + "learning_rate": 1.1211453593943293e-06, + "loss": 0.4407, + "num_input_tokens_seen": 69798432, + "step": 4520 + }, + { + "epoch": 18.564102564102566, + "grad_norm": 0.310546875, + "learning_rate": 1.118214288213667e-06, + "loss": 0.4594, + "num_input_tokens_seen": 69869984, + "step": 4525 + }, + { + "epoch": 18.584615384615386, + "grad_norm": 0.318359375, + "learning_rate": 1.1152847749588186e-06, + "loss": 0.5029, + "num_input_tokens_seen": 69949088, + "step": 4530 + }, + { + "epoch": 18.605128205128207, + "grad_norm": 0.6796875, + "learning_rate": 1.1123568315840419e-06, + "loss": 0.4386, + "num_input_tokens_seen": 70023904, + "step": 4535 + }, + { + "epoch": 18.625641025641027, + "grad_norm": 0.5625, + "learning_rate": 1.1094304700371863e-06, + "loss": 0.5118, + "num_input_tokens_seen": 70098400, + "step": 4540 + }, + { + "epoch": 18.646153846153847, + "grad_norm": 0.31640625, + "learning_rate": 1.1065057022596483e-06, + "loss": 0.4559, + "num_input_tokens_seen": 70176576, + "step": 4545 + }, + { + "epoch": 18.666666666666668, + "grad_norm": 0.345703125, + "learning_rate": 1.1035825401863185e-06, + "loss": 0.4418, + "num_input_tokens_seen": 70255264, + "step": 4550 + }, + { + "epoch": 18.68717948717949, + "grad_norm": 0.345703125, + "learning_rate": 1.100660995745538e-06, + "loss": 0.435, + "num_input_tokens_seen": 70324160, + "step": 4555 + }, + { + "epoch": 18.70769230769231, + "grad_norm": 0.365234375, + "learning_rate": 1.0977410808590437e-06, + "loss": 0.5499, + "num_input_tokens_seen": 70411616, + "step": 4560 + }, + { + "epoch": 18.72820512820513, + "grad_norm": 0.470703125, + "learning_rate": 1.0948228074419269e-06, + "loss": 0.4705, + "num_input_tokens_seen": 70482592, + "step": 4565 + }, + { + "epoch": 18.74871794871795, + "grad_norm": 0.44921875, + "learning_rate": 1.0919061874025774e-06, + "loss": 0.4466, + "num_input_tokens_seen": 70565120, + "step": 4570 + }, + { + "epoch": 18.76923076923077, + "grad_norm": 0.3984375, + "learning_rate": 1.0889912326426393e-06, + "loss": 0.4323, + "num_input_tokens_seen": 70635168, + "step": 4575 + }, + { + "epoch": 18.78974358974359, + "grad_norm": 0.310546875, + "learning_rate": 1.0860779550569609e-06, + "loss": 0.5235, + "num_input_tokens_seen": 70720096, + "step": 4580 + }, + { + "epoch": 18.81025641025641, + "grad_norm": 0.322265625, + "learning_rate": 1.083166366533548e-06, + "loss": 0.4659, + "num_input_tokens_seen": 70807648, + "step": 4585 + }, + { + "epoch": 18.83076923076923, + "grad_norm": 0.314453125, + "learning_rate": 1.080256478953512e-06, + "loss": 0.4968, + "num_input_tokens_seen": 70888832, + "step": 4590 + }, + { + "epoch": 18.851282051282052, + "grad_norm": 0.33203125, + "learning_rate": 1.0773483041910247e-06, + "loss": 0.5112, + "num_input_tokens_seen": 70971136, + "step": 4595 + }, + { + "epoch": 18.871794871794872, + "grad_norm": 0.29296875, + "learning_rate": 1.0744418541132676e-06, + "loss": 0.4414, + "num_input_tokens_seen": 71038816, + "step": 4600 + }, + { + "epoch": 18.892307692307693, + "grad_norm": 0.314453125, + "learning_rate": 1.0715371405803858e-06, + "loss": 0.4874, + "num_input_tokens_seen": 71114720, + "step": 4605 + }, + { + "epoch": 18.912820512820513, + "grad_norm": 0.294921875, + "learning_rate": 1.0686341754454364e-06, + "loss": 0.4571, + "num_input_tokens_seen": 71189856, + "step": 4610 + }, + { + "epoch": 18.933333333333334, + "grad_norm": 0.23046875, + "learning_rate": 1.0657329705543439e-06, + "loss": 0.4661, + "num_input_tokens_seen": 71271136, + "step": 4615 + }, + { + "epoch": 18.953846153846154, + "grad_norm": 0.2109375, + "learning_rate": 1.0628335377458477e-06, + "loss": 0.435, + "num_input_tokens_seen": 71343488, + "step": 4620 + }, + { + "epoch": 18.974358974358974, + "grad_norm": 0.21484375, + "learning_rate": 1.0599358888514582e-06, + "loss": 0.4845, + "num_input_tokens_seen": 71419104, + "step": 4625 + }, + { + "epoch": 18.994871794871795, + "grad_norm": 0.2021484375, + "learning_rate": 1.0570400356954044e-06, + "loss": 0.5111, + "num_input_tokens_seen": 71491776, + "step": 4630 + }, + { + "epoch": 19.015384615384615, + "grad_norm": 0.18359375, + "learning_rate": 1.0541459900945892e-06, + "loss": 0.5006, + "num_input_tokens_seen": 71573184, + "step": 4635 + }, + { + "epoch": 19.035897435897436, + "grad_norm": 0.1630859375, + "learning_rate": 1.0512537638585379e-06, + "loss": 0.4828, + "num_input_tokens_seen": 71648768, + "step": 4640 + }, + { + "epoch": 19.056410256410256, + "grad_norm": 0.1572265625, + "learning_rate": 1.0483633687893526e-06, + "loss": 0.4607, + "num_input_tokens_seen": 71721824, + "step": 4645 + }, + { + "epoch": 19.076923076923077, + "grad_norm": 0.1376953125, + "learning_rate": 1.0454748166816645e-06, + "loss": 0.497, + "num_input_tokens_seen": 71797568, + "step": 4650 + }, + { + "epoch": 19.097435897435897, + "grad_norm": 0.13671875, + "learning_rate": 1.0425881193225808e-06, + "loss": 0.4607, + "num_input_tokens_seen": 71871136, + "step": 4655 + }, + { + "epoch": 19.117948717948718, + "grad_norm": 0.1337890625, + "learning_rate": 1.0397032884916438e-06, + "loss": 0.4232, + "num_input_tokens_seen": 71941984, + "step": 4660 + }, + { + "epoch": 19.138461538461538, + "grad_norm": 0.123046875, + "learning_rate": 1.0368203359607767e-06, + "loss": 0.5216, + "num_input_tokens_seen": 72024736, + "step": 4665 + }, + { + "epoch": 19.15897435897436, + "grad_norm": 0.12109375, + "learning_rate": 1.0339392734942393e-06, + "loss": 0.4565, + "num_input_tokens_seen": 72096064, + "step": 4670 + }, + { + "epoch": 19.17948717948718, + "grad_norm": 0.1337890625, + "learning_rate": 1.031060112848578e-06, + "loss": 0.4488, + "num_input_tokens_seen": 72171168, + "step": 4675 + }, + { + "epoch": 19.2, + "grad_norm": 0.126953125, + "learning_rate": 1.0281828657725798e-06, + "loss": 0.4567, + "num_input_tokens_seen": 72249056, + "step": 4680 + }, + { + "epoch": 19.22051282051282, + "grad_norm": 0.2080078125, + "learning_rate": 1.0253075440072212e-06, + "loss": 0.4556, + "num_input_tokens_seen": 72338944, + "step": 4685 + }, + { + "epoch": 19.24102564102564, + "grad_norm": 0.1484375, + "learning_rate": 1.0224341592856245e-06, + "loss": 0.4686, + "num_input_tokens_seen": 72412032, + "step": 4690 + }, + { + "epoch": 19.26153846153846, + "grad_norm": 0.1435546875, + "learning_rate": 1.0195627233330052e-06, + "loss": 0.4087, + "num_input_tokens_seen": 72485760, + "step": 4695 + }, + { + "epoch": 19.28205128205128, + "grad_norm": 0.3046875, + "learning_rate": 1.0166932478666292e-06, + "loss": 0.4634, + "num_input_tokens_seen": 72565344, + "step": 4700 + }, + { + "epoch": 19.3025641025641, + "grad_norm": 0.275390625, + "learning_rate": 1.0138257445957601e-06, + "loss": 0.5316, + "num_input_tokens_seen": 72644800, + "step": 4705 + }, + { + "epoch": 19.323076923076922, + "grad_norm": 0.31640625, + "learning_rate": 1.0109602252216153e-06, + "loss": 0.4929, + "num_input_tokens_seen": 72725440, + "step": 4710 + }, + { + "epoch": 19.343589743589742, + "grad_norm": 0.640625, + "learning_rate": 1.0080967014373152e-06, + "loss": 0.442, + "num_input_tokens_seen": 72796064, + "step": 4715 + }, + { + "epoch": 19.364102564102563, + "grad_norm": 0.91015625, + "learning_rate": 1.0052351849278385e-06, + "loss": 0.4442, + "num_input_tokens_seen": 72870080, + "step": 4720 + }, + { + "epoch": 19.384615384615383, + "grad_norm": 0.2490234375, + "learning_rate": 1.0023756873699723e-06, + "loss": 0.4751, + "num_input_tokens_seen": 72944288, + "step": 4725 + }, + { + "epoch": 19.405128205128204, + "grad_norm": 0.333984375, + "learning_rate": 9.995182204322637e-07, + "loss": 0.5129, + "num_input_tokens_seen": 73021664, + "step": 4730 + }, + { + "epoch": 19.425641025641024, + "grad_norm": 0.81640625, + "learning_rate": 9.966627957749767e-07, + "loss": 0.4796, + "num_input_tokens_seen": 73097920, + "step": 4735 + }, + { + "epoch": 19.446153846153845, + "grad_norm": 0.333984375, + "learning_rate": 9.93809425050039e-07, + "loss": 0.4785, + "num_input_tokens_seen": 73170112, + "step": 4740 + }, + { + "epoch": 19.466666666666665, + "grad_norm": 0.2255859375, + "learning_rate": 9.909581199009971e-07, + "loss": 0.4379, + "num_input_tokens_seen": 73241312, + "step": 4745 + }, + { + "epoch": 19.487179487179485, + "grad_norm": 0.267578125, + "learning_rate": 9.8810889196297e-07, + "loss": 0.5408, + "num_input_tokens_seen": 73319936, + "step": 4750 + }, + { + "epoch": 19.50769230769231, + "grad_norm": 0.3046875, + "learning_rate": 9.852617528625992e-07, + "loss": 0.4878, + "num_input_tokens_seen": 73397824, + "step": 4755 + }, + { + "epoch": 19.52820512820513, + "grad_norm": 0.29296875, + "learning_rate": 9.824167142180032e-07, + "loss": 0.4627, + "num_input_tokens_seen": 73470688, + "step": 4760 + }, + { + "epoch": 19.54871794871795, + "grad_norm": 0.283203125, + "learning_rate": 9.795737876387285e-07, + "loss": 0.4799, + "num_input_tokens_seen": 73545792, + "step": 4765 + }, + { + "epoch": 19.56923076923077, + "grad_norm": 0.34765625, + "learning_rate": 9.76732984725704e-07, + "loss": 0.428, + "num_input_tokens_seen": 73617760, + "step": 4770 + }, + { + "epoch": 19.58974358974359, + "grad_norm": 0.3515625, + "learning_rate": 9.738943170711916e-07, + "loss": 0.4608, + "num_input_tokens_seen": 73707328, + "step": 4775 + }, + { + "epoch": 19.61025641025641, + "grad_norm": 0.515625, + "learning_rate": 9.710577962587412e-07, + "loss": 0.4951, + "num_input_tokens_seen": 73782976, + "step": 4780 + }, + { + "epoch": 19.630769230769232, + "grad_norm": 0.3828125, + "learning_rate": 9.68223433863141e-07, + "loss": 0.53, + "num_input_tokens_seen": 73872256, + "step": 4785 + }, + { + "epoch": 19.651282051282053, + "grad_norm": 0.341796875, + "learning_rate": 9.653912414503731e-07, + "loss": 0.45, + "num_input_tokens_seen": 73947520, + "step": 4790 + }, + { + "epoch": 19.671794871794873, + "grad_norm": 0.337890625, + "learning_rate": 9.625612305775626e-07, + "loss": 0.4218, + "num_input_tokens_seen": 74023104, + "step": 4795 + }, + { + "epoch": 19.692307692307693, + "grad_norm": 0.333984375, + "learning_rate": 9.597334127929346e-07, + "loss": 0.4531, + "num_input_tokens_seen": 74105920, + "step": 4800 + }, + { + "epoch": 19.712820512820514, + "grad_norm": 0.306640625, + "learning_rate": 9.569077996357638e-07, + "loss": 0.4142, + "num_input_tokens_seen": 74182752, + "step": 4805 + }, + { + "epoch": 19.733333333333334, + "grad_norm": 0.37109375, + "learning_rate": 9.54084402636329e-07, + "loss": 0.5121, + "num_input_tokens_seen": 74265248, + "step": 4810 + }, + { + "epoch": 19.753846153846155, + "grad_norm": 0.328125, + "learning_rate": 9.512632333158653e-07, + "loss": 0.523, + "num_input_tokens_seen": 74363232, + "step": 4815 + }, + { + "epoch": 19.774358974358975, + "grad_norm": 0.353515625, + "learning_rate": 9.484443031865176e-07, + "loss": 0.5103, + "num_input_tokens_seen": 74442848, + "step": 4820 + }, + { + "epoch": 19.794871794871796, + "grad_norm": 0.27734375, + "learning_rate": 9.456276237512949e-07, + "loss": 0.4725, + "num_input_tokens_seen": 74513376, + "step": 4825 + }, + { + "epoch": 19.815384615384616, + "grad_norm": 0.271484375, + "learning_rate": 9.428132065040198e-07, + "loss": 0.4921, + "num_input_tokens_seen": 74588928, + "step": 4830 + }, + { + "epoch": 19.835897435897436, + "grad_norm": 0.291015625, + "learning_rate": 9.40001062929285e-07, + "loss": 0.4331, + "num_input_tokens_seen": 74659040, + "step": 4835 + }, + { + "epoch": 19.856410256410257, + "grad_norm": 0.216796875, + "learning_rate": 9.371912045024046e-07, + "loss": 0.44, + "num_input_tokens_seen": 74737472, + "step": 4840 + }, + { + "epoch": 19.876923076923077, + "grad_norm": 0.2158203125, + "learning_rate": 9.343836426893687e-07, + "loss": 0.4831, + "num_input_tokens_seen": 74810592, + "step": 4845 + }, + { + "epoch": 19.897435897435898, + "grad_norm": 0.1943359375, + "learning_rate": 9.315783889467943e-07, + "loss": 0.4494, + "num_input_tokens_seen": 74880096, + "step": 4850 + }, + { + "epoch": 19.91794871794872, + "grad_norm": 0.1953125, + "learning_rate": 9.287754547218821e-07, + "loss": 0.5063, + "num_input_tokens_seen": 74951168, + "step": 4855 + }, + { + "epoch": 19.93846153846154, + "grad_norm": 0.1611328125, + "learning_rate": 9.259748514523654e-07, + "loss": 0.4975, + "num_input_tokens_seen": 75034592, + "step": 4860 + }, + { + "epoch": 19.95897435897436, + "grad_norm": 0.1728515625, + "learning_rate": 9.231765905664677e-07, + "loss": 0.4802, + "num_input_tokens_seen": 75108992, + "step": 4865 + }, + { + "epoch": 19.97948717948718, + "grad_norm": 0.166015625, + "learning_rate": 9.20380683482853e-07, + "loss": 0.4473, + "num_input_tokens_seen": 75185824, + "step": 4870 + }, + { + "epoch": 20.0, + "grad_norm": 0.140625, + "learning_rate": 9.175871416105802e-07, + "loss": 0.4971, + "num_input_tokens_seen": 75263936, + "step": 4875 + }, + { + "epoch": 20.02051282051282, + "grad_norm": 0.134765625, + "learning_rate": 9.147959763490578e-07, + "loss": 0.493, + "num_input_tokens_seen": 75342752, + "step": 4880 + }, + { + "epoch": 20.04102564102564, + "grad_norm": 0.11669921875, + "learning_rate": 9.120071990879949e-07, + "loss": 0.4709, + "num_input_tokens_seen": 75420704, + "step": 4885 + }, + { + "epoch": 20.06153846153846, + "grad_norm": 0.13671875, + "learning_rate": 9.092208212073569e-07, + "loss": 0.5248, + "num_input_tokens_seen": 75496128, + "step": 4890 + }, + { + "epoch": 20.08205128205128, + "grad_norm": 0.1220703125, + "learning_rate": 9.064368540773177e-07, + "loss": 0.525, + "num_input_tokens_seen": 75585984, + "step": 4895 + }, + { + "epoch": 20.102564102564102, + "grad_norm": 0.1357421875, + "learning_rate": 9.036553090582145e-07, + "loss": 0.5186, + "num_input_tokens_seen": 75668416, + "step": 4900 + }, + { + "epoch": 20.123076923076923, + "grad_norm": 0.15234375, + "learning_rate": 9.008761975004994e-07, + "loss": 0.456, + "num_input_tokens_seen": 75744960, + "step": 4905 + }, + { + "epoch": 20.143589743589743, + "grad_norm": 0.203125, + "learning_rate": 8.98099530744696e-07, + "loss": 0.4507, + "num_input_tokens_seen": 75818688, + "step": 4910 + }, + { + "epoch": 20.164102564102564, + "grad_norm": 0.150390625, + "learning_rate": 8.953253201213517e-07, + "loss": 0.4552, + "num_input_tokens_seen": 75901824, + "step": 4915 + }, + { + "epoch": 20.184615384615384, + "grad_norm": 0.146484375, + "learning_rate": 8.925535769509895e-07, + "loss": 0.4153, + "num_input_tokens_seen": 75973248, + "step": 4920 + }, + { + "epoch": 20.205128205128204, + "grad_norm": 0.310546875, + "learning_rate": 8.897843125440653e-07, + "loss": 0.4452, + "num_input_tokens_seen": 76043840, + "step": 4925 + }, + { + "epoch": 20.217435897435898, + "num_input_tokens_seen": 76105440, + "step": 4928, + "total_flos": 1.6998510508071322e+18, + "train_loss": 0.4908539823365289, + "train_runtime": 36017.1103, + "train_samples_per_second": 3.465, + "train_steps_per_second": 0.216 + } + ], + "logging_steps": 5, + "max_steps": 7776, + "num_input_tokens_seen": 76105440, + "num_train_epochs": 32, + "save_steps": 2000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.6998510508071322e+18, + "train_batch_size": 4, + "trial_name": null, + "trial_params": null +}