{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.9997197702115734, "eval_steps": 50, "global_step": 892, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.02241838307412078, "grad_norm": 11.57065531644996, "learning_rate": 1.9775784753363228e-06, "loss": 0.2054, "step": 10 }, { "epoch": 0.04483676614824156, "grad_norm": 10.25086765406488, "learning_rate": 1.9551569506726456e-06, "loss": 0.1492, "step": 20 }, { "epoch": 0.06725514922236234, "grad_norm": 7.822767237983853, "learning_rate": 1.9327354260089685e-06, "loss": 0.152, "step": 30 }, { "epoch": 0.08967353229648312, "grad_norm": 8.157708200242721, "learning_rate": 1.9103139013452914e-06, "loss": 0.1472, "step": 40 }, { "epoch": 0.1120919153706039, "grad_norm": 12.8257632132892, "learning_rate": 1.8878923766816142e-06, "loss": 0.1356, "step": 50 }, { "epoch": 0.1120919153706039, "eval_loss": 0.1209246814250946, "eval_runtime": 477.8616, "eval_samples_per_second": 6.288, "eval_steps_per_second": 0.787, "step": 50 }, { "epoch": 0.13451029844472467, "grad_norm": 8.90816614814061, "learning_rate": 1.865470852017937e-06, "loss": 0.1281, "step": 60 }, { "epoch": 0.15692868151884545, "grad_norm": 9.628880600516572, "learning_rate": 1.84304932735426e-06, "loss": 0.1381, "step": 70 }, { "epoch": 0.17934706459296623, "grad_norm": 8.899868902286775, "learning_rate": 1.8206278026905828e-06, "loss": 0.1324, "step": 80 }, { "epoch": 0.201765447667087, "grad_norm": 130.29750942761115, "learning_rate": 1.798206278026906e-06, "loss": 0.1277, "step": 90 }, { "epoch": 0.2241838307412078, "grad_norm": 6.4686996638845145, "learning_rate": 1.7757847533632286e-06, "loss": 0.1221, "step": 100 }, { "epoch": 0.2241838307412078, "eval_loss": 0.11111436039209366, "eval_runtime": 477.824, "eval_samples_per_second": 6.289, "eval_steps_per_second": 0.787, "step": 100 }, { "epoch": 0.24660221381532857, "grad_norm": 8.356076121379255, "learning_rate": 1.7533632286995514e-06, "loss": 0.1198, "step": 110 }, { "epoch": 0.26902059688944935, "grad_norm": 6.743428651631778, "learning_rate": 1.7309417040358743e-06, "loss": 0.1076, "step": 120 }, { "epoch": 0.29143897996357016, "grad_norm": 6.8645809938665465, "learning_rate": 1.7085201793721974e-06, "loss": 0.1187, "step": 130 }, { "epoch": 0.3138573630376909, "grad_norm": 5.999842993813071, "learning_rate": 1.68609865470852e-06, "loss": 0.1182, "step": 140 }, { "epoch": 0.3362757461118117, "grad_norm": 7.7478230858078305, "learning_rate": 1.6636771300448429e-06, "loss": 0.1127, "step": 150 }, { "epoch": 0.3362757461118117, "eval_loss": 0.10507839918136597, "eval_runtime": 477.8167, "eval_samples_per_second": 6.289, "eval_steps_per_second": 0.787, "step": 150 }, { "epoch": 0.35869412918593246, "grad_norm": 6.564504796683449, "learning_rate": 1.641255605381166e-06, "loss": 0.1144, "step": 160 }, { "epoch": 0.38111251226005327, "grad_norm": 20.75971255482363, "learning_rate": 1.6188340807174888e-06, "loss": 0.1078, "step": 170 }, { "epoch": 0.403530895334174, "grad_norm": 6.913685427209001, "learning_rate": 1.5964125560538115e-06, "loss": 0.1189, "step": 180 }, { "epoch": 0.42594927840829483, "grad_norm": 11.143255383687194, "learning_rate": 1.5739910313901345e-06, "loss": 0.1189, "step": 190 }, { "epoch": 0.4483676614824156, "grad_norm": 4.747347623410045, "learning_rate": 1.5515695067264574e-06, "loss": 0.1072, "step": 200 }, { "epoch": 0.4483676614824156, "eval_loss": 0.09710206091403961, "eval_runtime": 477.8045, "eval_samples_per_second": 6.289, "eval_steps_per_second": 0.787, "step": 200 }, { "epoch": 0.4707860445565364, "grad_norm": 5.512284850848945, "learning_rate": 1.5291479820627803e-06, "loss": 0.1136, "step": 210 }, { "epoch": 0.49320442763065714, "grad_norm": 17.278200474599828, "learning_rate": 1.506726457399103e-06, "loss": 0.1094, "step": 220 }, { "epoch": 0.5156228107047779, "grad_norm": 46.879816629413476, "learning_rate": 1.484304932735426e-06, "loss": 0.1091, "step": 230 }, { "epoch": 0.5380411937788987, "grad_norm": 8.493451669290872, "learning_rate": 1.4618834080717489e-06, "loss": 0.1019, "step": 240 }, { "epoch": 0.5604595768530195, "grad_norm": 9.286929395356434, "learning_rate": 1.4394618834080715e-06, "loss": 0.1075, "step": 250 }, { "epoch": 0.5604595768530195, "eval_loss": 0.09229769557714462, "eval_runtime": 477.6411, "eval_samples_per_second": 6.291, "eval_steps_per_second": 0.787, "step": 250 }, { "epoch": 0.5828779599271403, "grad_norm": 7.559299755990858, "learning_rate": 1.4170403587443946e-06, "loss": 0.0985, "step": 260 }, { "epoch": 0.605296343001261, "grad_norm": 3.9026812483393387, "learning_rate": 1.3946188340807175e-06, "loss": 0.0896, "step": 270 }, { "epoch": 0.6277147260753818, "grad_norm": 8.335653074155596, "learning_rate": 1.3721973094170403e-06, "loss": 0.0893, "step": 280 }, { "epoch": 0.6501331091495026, "grad_norm": 6.897239653871222, "learning_rate": 1.349775784753363e-06, "loss": 0.0914, "step": 290 }, { "epoch": 0.6725514922236234, "grad_norm": 29.648594106464095, "learning_rate": 1.327354260089686e-06, "loss": 0.0972, "step": 300 }, { "epoch": 0.6725514922236234, "eval_loss": 0.08544214069843292, "eval_runtime": 477.8106, "eval_samples_per_second": 6.289, "eval_steps_per_second": 0.787, "step": 300 }, { "epoch": 0.6949698752977441, "grad_norm": 3.969809179016744, "learning_rate": 1.304932735426009e-06, "loss": 0.089, "step": 310 }, { "epoch": 0.7173882583718649, "grad_norm": 5.34402617506164, "learning_rate": 1.2825112107623318e-06, "loss": 0.0867, "step": 320 }, { "epoch": 0.7398066414459857, "grad_norm": 5.506697389437645, "learning_rate": 1.2600896860986546e-06, "loss": 0.0803, "step": 330 }, { "epoch": 0.7622250245201065, "grad_norm": 8.056193416126446, "learning_rate": 1.2376681614349775e-06, "loss": 0.0917, "step": 340 }, { "epoch": 0.7846434075942272, "grad_norm": 4.7612681555615, "learning_rate": 1.2152466367713004e-06, "loss": 0.089, "step": 350 }, { "epoch": 0.7846434075942272, "eval_loss": 0.07997283339500427, "eval_runtime": 477.8824, "eval_samples_per_second": 6.288, "eval_steps_per_second": 0.787, "step": 350 }, { "epoch": 0.807061790668348, "grad_norm": 9.702063628974917, "learning_rate": 1.1928251121076232e-06, "loss": 0.0886, "step": 360 }, { "epoch": 0.8294801737424689, "grad_norm": 4.937196401368392, "learning_rate": 1.170403587443946e-06, "loss": 0.0805, "step": 370 }, { "epoch": 0.8518985568165897, "grad_norm": 6.8769697579900555, "learning_rate": 1.147982062780269e-06, "loss": 0.0754, "step": 380 }, { "epoch": 0.8743169398907104, "grad_norm": 10.270492248069893, "learning_rate": 1.1255605381165918e-06, "loss": 0.0855, "step": 390 }, { "epoch": 0.8967353229648312, "grad_norm": 8.129704483519474, "learning_rate": 1.103139013452915e-06, "loss": 0.0898, "step": 400 }, { "epoch": 0.8967353229648312, "eval_loss": 0.0756540298461914, "eval_runtime": 477.7284, "eval_samples_per_second": 6.29, "eval_steps_per_second": 0.787, "step": 400 }, { "epoch": 0.919153706038952, "grad_norm": 7.785886269800572, "learning_rate": 1.0807174887892376e-06, "loss": 0.0827, "step": 410 }, { "epoch": 0.9415720891130728, "grad_norm": 7.119394503978009, "learning_rate": 1.0582959641255604e-06, "loss": 0.0862, "step": 420 }, { "epoch": 0.9639904721871935, "grad_norm": 5.06264160032142, "learning_rate": 1.0358744394618835e-06, "loss": 0.078, "step": 430 }, { "epoch": 0.9864088552613143, "grad_norm": 6.034467278570911, "learning_rate": 1.0134529147982064e-06, "loss": 0.0745, "step": 440 }, { "epoch": 1.008827238335435, "grad_norm": 6.107563557049032, "learning_rate": 9.91031390134529e-07, "loss": 0.0635, "step": 450 }, { "epoch": 1.008827238335435, "eval_loss": 0.07114721089601517, "eval_runtime": 477.6398, "eval_samples_per_second": 6.291, "eval_steps_per_second": 0.787, "step": 450 }, { "epoch": 1.0312456214095558, "grad_norm": 8.472989007522791, "learning_rate": 9.68609865470852e-07, "loss": 0.0687, "step": 460 }, { "epoch": 1.0536640044836767, "grad_norm": 5.080789399491274, "learning_rate": 9.461883408071749e-07, "loss": 0.0635, "step": 470 }, { "epoch": 1.0760823875577974, "grad_norm": 8.06976721278842, "learning_rate": 9.237668161434977e-07, "loss": 0.0674, "step": 480 }, { "epoch": 1.098500770631918, "grad_norm": 9.00400557879772, "learning_rate": 9.013452914798207e-07, "loss": 0.0623, "step": 490 }, { "epoch": 1.120919153706039, "grad_norm": 10.419684143484417, "learning_rate": 8.789237668161434e-07, "loss": 0.0742, "step": 500 }, { "epoch": 1.120919153706039, "eval_loss": 0.06726241111755371, "eval_runtime": 478.2051, "eval_samples_per_second": 6.284, "eval_steps_per_second": 0.786, "step": 500 }, { "epoch": 1.1433375367801597, "grad_norm": 5.834457752004053, "learning_rate": 8.565022421524663e-07, "loss": 0.0698, "step": 510 }, { "epoch": 1.1657559198542806, "grad_norm": 5.920575213015667, "learning_rate": 8.340807174887892e-07, "loss": 0.061, "step": 520 }, { "epoch": 1.1881743029284013, "grad_norm": 11.4664445988588, "learning_rate": 8.11659192825112e-07, "loss": 0.0612, "step": 530 }, { "epoch": 1.210592686002522, "grad_norm": 5.074793088570693, "learning_rate": 7.892376681614349e-07, "loss": 0.0582, "step": 540 }, { "epoch": 1.233011069076643, "grad_norm": 4.964609158362768, "learning_rate": 7.668161434977578e-07, "loss": 0.0627, "step": 550 }, { "epoch": 1.233011069076643, "eval_loss": 0.06428983807563782, "eval_runtime": 477.8813, "eval_samples_per_second": 6.288, "eval_steps_per_second": 0.787, "step": 550 }, { "epoch": 1.2554294521507636, "grad_norm": 5.775433133659037, "learning_rate": 7.443946188340807e-07, "loss": 0.0613, "step": 560 }, { "epoch": 1.2778478352248843, "grad_norm": 7.504088139984216, "learning_rate": 7.219730941704035e-07, "loss": 0.0625, "step": 570 }, { "epoch": 1.3002662182990052, "grad_norm": 12.423874958836857, "learning_rate": 6.995515695067265e-07, "loss": 0.0541, "step": 580 }, { "epoch": 1.322684601373126, "grad_norm": 13.037612748873672, "learning_rate": 6.771300448430492e-07, "loss": 0.0638, "step": 590 }, { "epoch": 1.3451029844472466, "grad_norm": 7.643633731840427, "learning_rate": 6.547085201793722e-07, "loss": 0.0528, "step": 600 }, { "epoch": 1.3451029844472466, "eval_loss": 0.06060384213924408, "eval_runtime": 477.628, "eval_samples_per_second": 6.292, "eval_steps_per_second": 0.787, "step": 600 }, { "epoch": 1.3675213675213675, "grad_norm": 9.627024940339941, "learning_rate": 6.322869955156951e-07, "loss": 0.0569, "step": 610 }, { "epoch": 1.3899397505954882, "grad_norm": 14.780208722543843, "learning_rate": 6.098654708520179e-07, "loss": 0.0504, "step": 620 }, { "epoch": 1.4123581336696092, "grad_norm": 7.999021557044027, "learning_rate": 5.874439461883408e-07, "loss": 0.0609, "step": 630 }, { "epoch": 1.4347765167437299, "grad_norm": 8.443684543866313, "learning_rate": 5.650224215246636e-07, "loss": 0.0531, "step": 640 }, { "epoch": 1.4571948998178508, "grad_norm": 3.5781417864085086, "learning_rate": 5.426008968609865e-07, "loss": 0.0542, "step": 650 }, { "epoch": 1.4571948998178508, "eval_loss": 0.057715680450201035, "eval_runtime": 477.9099, "eval_samples_per_second": 6.288, "eval_steps_per_second": 0.787, "step": 650 }, { "epoch": 1.4796132828919715, "grad_norm": 6.394299603034078, "learning_rate": 5.201793721973094e-07, "loss": 0.0549, "step": 660 }, { "epoch": 1.5020316659660922, "grad_norm": 6.251887881398542, "learning_rate": 4.977578475336322e-07, "loss": 0.0536, "step": 670 }, { "epoch": 1.524450049040213, "grad_norm": 4.45363233623469, "learning_rate": 4.753363228699551e-07, "loss": 0.0519, "step": 680 }, { "epoch": 1.5468684321143338, "grad_norm": 7.236100371063919, "learning_rate": 4.5291479820627797e-07, "loss": 0.0516, "step": 690 }, { "epoch": 1.5692868151884545, "grad_norm": 11.475278894303377, "learning_rate": 4.304932735426009e-07, "loss": 0.0463, "step": 700 }, { "epoch": 1.5692868151884545, "eval_loss": 0.05437139794230461, "eval_runtime": 477.9054, "eval_samples_per_second": 6.288, "eval_steps_per_second": 0.787, "step": 700 }, { "epoch": 1.5917051982625754, "grad_norm": 6.676408323109306, "learning_rate": 4.0807174887892375e-07, "loss": 0.0532, "step": 710 }, { "epoch": 1.614123581336696, "grad_norm": 4.965760634983937, "learning_rate": 3.856502242152466e-07, "loss": 0.0545, "step": 720 }, { "epoch": 1.6365419644108168, "grad_norm": 27.900762279082834, "learning_rate": 3.632286995515695e-07, "loss": 0.0544, "step": 730 }, { "epoch": 1.6589603474849377, "grad_norm": 5.498723625213236, "learning_rate": 3.4080717488789235e-07, "loss": 0.0518, "step": 740 }, { "epoch": 1.6813787305590584, "grad_norm": 5.5744858185141775, "learning_rate": 3.183856502242152e-07, "loss": 0.0463, "step": 750 }, { "epoch": 1.6813787305590584, "eval_loss": 0.05197111889719963, "eval_runtime": 477.8827, "eval_samples_per_second": 6.288, "eval_steps_per_second": 0.787, "step": 750 }, { "epoch": 1.703797113633179, "grad_norm": 5.741468990313337, "learning_rate": 2.9596412556053813e-07, "loss": 0.0454, "step": 760 }, { "epoch": 1.7262154967073, "grad_norm": 9.331526709026493, "learning_rate": 2.73542600896861e-07, "loss": 0.0448, "step": 770 }, { "epoch": 1.748633879781421, "grad_norm": 8.801924259928679, "learning_rate": 2.5112107623318386e-07, "loss": 0.0491, "step": 780 }, { "epoch": 1.7710522628555414, "grad_norm": 3.9216347240361435, "learning_rate": 2.2869955156950672e-07, "loss": 0.0456, "step": 790 }, { "epoch": 1.7934706459296623, "grad_norm": 5.775711795911055, "learning_rate": 2.062780269058296e-07, "loss": 0.0434, "step": 800 }, { "epoch": 1.7934706459296623, "eval_loss": 0.04976892098784447, "eval_runtime": 477.8498, "eval_samples_per_second": 6.289, "eval_steps_per_second": 0.787, "step": 800 }, { "epoch": 1.8158890290037832, "grad_norm": 15.832420424237657, "learning_rate": 1.8385650224215245e-07, "loss": 0.0549, "step": 810 }, { "epoch": 1.838307412077904, "grad_norm": 11.331927257096979, "learning_rate": 1.6143497757847531e-07, "loss": 0.0479, "step": 820 }, { "epoch": 1.8607257951520246, "grad_norm": 3.4779552694261846, "learning_rate": 1.390134529147982e-07, "loss": 0.045, "step": 830 }, { "epoch": 1.8831441782261455, "grad_norm": 14.096117466941674, "learning_rate": 1.1659192825112107e-07, "loss": 0.0468, "step": 840 }, { "epoch": 1.9055625613002662, "grad_norm": 4.258023342595773, "learning_rate": 9.417040358744393e-08, "loss": 0.043, "step": 850 }, { "epoch": 1.9055625613002662, "eval_loss": 0.047411367297172546, "eval_runtime": 478.0538, "eval_samples_per_second": 6.286, "eval_steps_per_second": 0.787, "step": 850 }, { "epoch": 1.927980944374387, "grad_norm": 4.472323710821738, "learning_rate": 7.174887892376681e-08, "loss": 0.0485, "step": 860 }, { "epoch": 1.9503993274485079, "grad_norm": 8.238550394871776, "learning_rate": 4.932735426008968e-08, "loss": 0.0461, "step": 870 }, { "epoch": 1.9728177105226286, "grad_norm": 3.661996756017377, "learning_rate": 2.6905829596412556e-08, "loss": 0.0414, "step": 880 }, { "epoch": 1.9952360935967492, "grad_norm": 4.240752961770031, "learning_rate": 4.484304932735426e-09, "loss": 0.0496, "step": 890 } ], "logging_steps": 10, "max_steps": 892, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 360, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.1364495346169283e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }