|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9907692307692306, |
|
"eval_steps": 40, |
|
"global_step": 243, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.06153846153846154, |
|
"grad_norm": 67.0723563361413, |
|
"learning_rate": 5e-07, |
|
"logits/chosen": -2.709871530532837, |
|
"logits/rejected": -2.7169337272644043, |
|
"logps/chosen": -301.914306640625, |
|
"logps/rejected": -224.25662231445312, |
|
"loss": 0.6893, |
|
"rewards/accuracies": 0.33125001192092896, |
|
"rewards/chosen": 0.021815577521920204, |
|
"rewards/margins": 0.006514790467917919, |
|
"rewards/rejected": 0.015300785191357136, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.12307692307692308, |
|
"grad_norm": 63.50419416702775, |
|
"learning_rate": 1e-06, |
|
"logits/chosen": -2.673720598220825, |
|
"logits/rejected": -2.6661205291748047, |
|
"logps/chosen": -289.4712829589844, |
|
"logps/rejected": -227.0055694580078, |
|
"loss": 0.6229, |
|
"rewards/accuracies": 0.675000011920929, |
|
"rewards/chosen": 0.7337898015975952, |
|
"rewards/margins": 0.23366117477416992, |
|
"rewards/rejected": 0.5001285672187805, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.18461538461538463, |
|
"grad_norm": 50.38404433953605, |
|
"learning_rate": 9.98864195911451e-07, |
|
"logits/chosen": -2.5014748573303223, |
|
"logits/rejected": -2.4808177947998047, |
|
"logps/chosen": -269.33612060546875, |
|
"logps/rejected": -221.9925079345703, |
|
"loss": 0.5587, |
|
"rewards/accuracies": 0.675000011920929, |
|
"rewards/chosen": 1.8688875436782837, |
|
"rewards/margins": 0.7238284945487976, |
|
"rewards/rejected": 1.1450591087341309, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.24615384615384617, |
|
"grad_norm": 46.1003816476988, |
|
"learning_rate": 9.95461943849514e-07, |
|
"logits/chosen": -2.3574154376983643, |
|
"logits/rejected": -2.3489270210266113, |
|
"logps/chosen": -279.2587585449219, |
|
"logps/rejected": -240.3167724609375, |
|
"loss": 0.5578, |
|
"rewards/accuracies": 0.6937500238418579, |
|
"rewards/chosen": 2.3487281799316406, |
|
"rewards/margins": 1.1014046669006348, |
|
"rewards/rejected": 1.2473232746124268, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.3076923076923077, |
|
"grad_norm": 46.94795150121206, |
|
"learning_rate": 9.898087009813985e-07, |
|
"logits/chosen": -2.298891544342041, |
|
"logits/rejected": -2.2741196155548096, |
|
"logps/chosen": -274.5340576171875, |
|
"logps/rejected": -243.38784790039062, |
|
"loss": 0.5498, |
|
"rewards/accuracies": 0.7437499761581421, |
|
"rewards/chosen": 2.5186121463775635, |
|
"rewards/margins": 1.7126624584197998, |
|
"rewards/rejected": 0.8059496879577637, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.36923076923076925, |
|
"grad_norm": 46.91929359569832, |
|
"learning_rate": 9.819301512125564e-07, |
|
"logits/chosen": -2.2155728340148926, |
|
"logits/rejected": -2.1926393508911133, |
|
"logps/chosen": -303.2185363769531, |
|
"logps/rejected": -230.5345001220703, |
|
"loss": 0.5365, |
|
"rewards/accuracies": 0.768750011920929, |
|
"rewards/chosen": 2.6670188903808594, |
|
"rewards/margins": 1.5836530923843384, |
|
"rewards/rejected": 1.0833656787872314, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.4307692307692308, |
|
"grad_norm": 39.966430884552075, |
|
"learning_rate": 9.718620884991454e-07, |
|
"logits/chosen": -2.1086268424987793, |
|
"logits/rejected": -2.077179193496704, |
|
"logps/chosen": -293.8847351074219, |
|
"logps/rejected": -250.50009155273438, |
|
"loss": 0.539, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": 2.3622381687164307, |
|
"rewards/margins": 1.3279569149017334, |
|
"rewards/rejected": 1.0342811346054077, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.49230769230769234, |
|
"grad_norm": 39.51653677279534, |
|
"learning_rate": 9.596502542283398e-07, |
|
"logits/chosen": -2.0062007904052734, |
|
"logits/rejected": -1.9648046493530273, |
|
"logps/chosen": -271.65533447265625, |
|
"logps/rejected": -209.97476196289062, |
|
"loss": 0.5436, |
|
"rewards/accuracies": 0.7562500238418579, |
|
"rewards/chosen": 2.328826665878296, |
|
"rewards/margins": 1.5009214878082275, |
|
"rewards/rejected": 0.8279051780700684, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.49230769230769234, |
|
"eval_logits/chosen": -2.035125732421875, |
|
"eval_logits/rejected": -2.0087647438049316, |
|
"eval_logps/chosen": -303.9334411621094, |
|
"eval_logps/rejected": -223.16908264160156, |
|
"eval_loss": 0.5011464357376099, |
|
"eval_rewards/accuracies": 0.7736486196517944, |
|
"eval_rewards/chosen": 2.5953011512756348, |
|
"eval_rewards/margins": 1.6505608558654785, |
|
"eval_rewards/rejected": 0.9447402954101562, |
|
"eval_runtime": 156.8895, |
|
"eval_samples_per_second": 14.692, |
|
"eval_steps_per_second": 0.236, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.5538461538461539, |
|
"grad_norm": 39.53692648423416, |
|
"learning_rate": 9.453501294053137e-07, |
|
"logits/chosen": -2.023144245147705, |
|
"logits/rejected": -2.0024471282958984, |
|
"logps/chosen": -288.55859375, |
|
"logps/rejected": -237.53579711914062, |
|
"loss": 0.5107, |
|
"rewards/accuracies": 0.7875000238418579, |
|
"rewards/chosen": 2.411022663116455, |
|
"rewards/margins": 1.7200405597686768, |
|
"rewards/rejected": 0.6909819841384888, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.6153846153846154, |
|
"grad_norm": 45.18002841742545, |
|
"learning_rate": 9.2902668259103e-07, |
|
"logits/chosen": -2.0483999252319336, |
|
"logits/rejected": -1.9713916778564453, |
|
"logps/chosen": -285.7639465332031, |
|
"logps/rejected": -219.0935516357422, |
|
"loss": 0.5436, |
|
"rewards/accuracies": 0.8374999761581421, |
|
"rewards/chosen": 2.238405466079712, |
|
"rewards/margins": 1.7145936489105225, |
|
"rewards/rejected": 0.5238116979598999, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.676923076923077, |
|
"grad_norm": 36.94795750119324, |
|
"learning_rate": 9.107540747360123e-07, |
|
"logits/chosen": -2.019026756286621, |
|
"logits/rejected": -1.9596664905548096, |
|
"logps/chosen": -288.3727111816406, |
|
"logps/rejected": -236.1953582763672, |
|
"loss": 0.5107, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": 2.0629780292510986, |
|
"rewards/margins": 1.6129209995269775, |
|
"rewards/rejected": 0.45005717873573303, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.7384615384615385, |
|
"grad_norm": 45.823356835075316, |
|
"learning_rate": 8.906153222511012e-07, |
|
"logits/chosen": -2.1011242866516113, |
|
"logits/rejected": -2.058206081390381, |
|
"logps/chosen": -267.6994934082031, |
|
"logps/rejected": -240.79833984375, |
|
"loss": 0.5475, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 2.305724620819092, |
|
"rewards/margins": 1.3877848386764526, |
|
"rewards/rejected": 0.9179398417472839, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"grad_norm": 41.349964597831146, |
|
"learning_rate": 8.687019198459393e-07, |
|
"logits/chosen": -2.26322078704834, |
|
"logits/rejected": -2.2332780361175537, |
|
"logps/chosen": -284.81500244140625, |
|
"logps/rejected": -243.53466796875, |
|
"loss": 0.5145, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": 2.149193525314331, |
|
"rewards/margins": 1.4158048629760742, |
|
"rewards/rejected": 0.7333890199661255, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.8615384615384616, |
|
"grad_norm": 35.03982177026574, |
|
"learning_rate": 8.451134248487099e-07, |
|
"logits/chosen": -2.3586983680725098, |
|
"logits/rejected": -2.338914394378662, |
|
"logps/chosen": -301.6893005371094, |
|
"logps/rejected": -228.70260620117188, |
|
"loss": 0.4817, |
|
"rewards/accuracies": 0.8187500238418579, |
|
"rewards/chosen": 2.474177360534668, |
|
"rewards/margins": 2.0998575687408447, |
|
"rewards/rejected": 0.3743199408054352, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.9230769230769231, |
|
"grad_norm": 36.788375659444604, |
|
"learning_rate": 8.199570048956553e-07, |
|
"logits/chosen": -2.4227261543273926, |
|
"logits/rejected": -2.4093117713928223, |
|
"logps/chosen": -297.9197082519531, |
|
"logps/rejected": -216.85360717773438, |
|
"loss": 0.4727, |
|
"rewards/accuracies": 0.8500000238418579, |
|
"rewards/chosen": 1.9344415664672852, |
|
"rewards/margins": 2.026484727859497, |
|
"rewards/rejected": -0.09204334020614624, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.9846153846153847, |
|
"grad_norm": 29.272816983661464, |
|
"learning_rate": 7.933469510453187e-07, |
|
"logits/chosen": -2.3556580543518066, |
|
"logits/rejected": -2.3152148723602295, |
|
"logps/chosen": -286.50604248046875, |
|
"logps/rejected": -264.03253173828125, |
|
"loss": 0.4689, |
|
"rewards/accuracies": 0.8187500238418579, |
|
"rewards/chosen": 1.6535011529922485, |
|
"rewards/margins": 2.2574427127838135, |
|
"rewards/rejected": -0.6039413213729858, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.9846153846153847, |
|
"eval_logits/chosen": -2.2517545223236084, |
|
"eval_logits/rejected": -2.229193925857544, |
|
"eval_logps/chosen": -311.55841064453125, |
|
"eval_logps/rejected": -235.54119873046875, |
|
"eval_loss": 0.47693783044815063, |
|
"eval_rewards/accuracies": 0.8074324131011963, |
|
"eval_rewards/chosen": 1.8328040838241577, |
|
"eval_rewards/margins": 2.1252739429473877, |
|
"eval_rewards/rejected": -0.2924700975418091, |
|
"eval_runtime": 154.8855, |
|
"eval_samples_per_second": 14.882, |
|
"eval_steps_per_second": 0.239, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.0461538461538462, |
|
"grad_norm": 15.165513534162447, |
|
"learning_rate": 7.654041585295399e-07, |
|
"logits/chosen": -2.1545214653015137, |
|
"logits/rejected": -2.1319217681884766, |
|
"logps/chosen": -286.4499816894531, |
|
"logps/rejected": -262.167724609375, |
|
"loss": 0.2094, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 2.428976058959961, |
|
"rewards/margins": 3.554943084716797, |
|
"rewards/rejected": -1.125967264175415, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.1076923076923078, |
|
"grad_norm": 22.133936868649922, |
|
"learning_rate": 7.362555775002579e-07, |
|
"logits/chosen": -2.101616859436035, |
|
"logits/rejected": -2.094041347503662, |
|
"logps/chosen": -280.25897216796875, |
|
"logps/rejected": -252.6195526123047, |
|
"loss": 0.1923, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": 3.1136231422424316, |
|
"rewards/margins": 3.7958176136016846, |
|
"rewards/rejected": -0.6821939945220947, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.1692307692307693, |
|
"grad_norm": 21.235888542941137, |
|
"learning_rate": 7.060336362675068e-07, |
|
"logits/chosen": -2.1554484367370605, |
|
"logits/rejected": -2.087667226791382, |
|
"logps/chosen": -284.35028076171875, |
|
"logps/rejected": -240.01785278320312, |
|
"loss": 0.1724, |
|
"rewards/accuracies": 0.9437500238418579, |
|
"rewards/chosen": 3.559795379638672, |
|
"rewards/margins": 4.128841400146484, |
|
"rewards/rejected": -0.569046139717102, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.2307692307692308, |
|
"grad_norm": 19.651603274408373, |
|
"learning_rate": 6.748756396489505e-07, |
|
"logits/chosen": -2.1568312644958496, |
|
"logits/rejected": -2.117281436920166, |
|
"logps/chosen": -269.5829162597656, |
|
"logps/rejected": -248.9884033203125, |
|
"loss": 0.1757, |
|
"rewards/accuracies": 0.9437500238418579, |
|
"rewards/chosen": 2.9166793823242188, |
|
"rewards/margins": 4.183014869689941, |
|
"rewards/rejected": -1.2663352489471436, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.2923076923076924, |
|
"grad_norm": 21.2287191738782, |
|
"learning_rate": 6.429231451643906e-07, |
|
"logits/chosen": -2.1700432300567627, |
|
"logits/rejected": -2.1068949699401855, |
|
"logps/chosen": -274.2152404785156, |
|
"logps/rejected": -229.031005859375, |
|
"loss": 0.1806, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": 3.072680711746216, |
|
"rewards/margins": 4.2831034660339355, |
|
"rewards/rejected": -1.210423231124878, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.353846153846154, |
|
"grad_norm": 17.119914873342427, |
|
"learning_rate": 6.103213199093267e-07, |
|
"logits/chosen": -2.1095035076141357, |
|
"logits/rejected": -2.0814878940582275, |
|
"logps/chosen": -289.4352111816406, |
|
"logps/rejected": -246.68759155273438, |
|
"loss": 0.1544, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 3.430999755859375, |
|
"rewards/margins": 4.2461419105529785, |
|
"rewards/rejected": -0.8151422739028931, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.4153846153846155, |
|
"grad_norm": 22.13412271462813, |
|
"learning_rate": 5.772182810294344e-07, |
|
"logits/chosen": -2.121804714202881, |
|
"logits/rejected": -2.097852945327759, |
|
"logps/chosen": -265.60552978515625, |
|
"logps/rejected": -243.84390258789062, |
|
"loss": 0.1949, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 2.722719192504883, |
|
"rewards/margins": 4.1747965812683105, |
|
"rewards/rejected": -1.452077031135559, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.476923076923077, |
|
"grad_norm": 19.045988439656412, |
|
"learning_rate": 5.43764422792326e-07, |
|
"logits/chosen": -2.1834075450897217, |
|
"logits/rejected": -2.1550724506378174, |
|
"logps/chosen": -287.23614501953125, |
|
"logps/rejected": -262.32855224609375, |
|
"loss": 0.1825, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": 3.010335683822632, |
|
"rewards/margins": 4.589625358581543, |
|
"rewards/rejected": -1.5792896747589111, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.476923076923077, |
|
"eval_logits/chosen": -2.2158141136169434, |
|
"eval_logits/rejected": -2.195758819580078, |
|
"eval_logps/chosen": -310.03826904296875, |
|
"eval_logps/rejected": -240.60853576660156, |
|
"eval_loss": 0.5121302604675293, |
|
"eval_rewards/accuracies": 0.8141891956329346, |
|
"eval_rewards/chosen": 1.9848158359527588, |
|
"eval_rewards/margins": 2.784022092819214, |
|
"eval_rewards/rejected": -0.7992062568664551, |
|
"eval_runtime": 154.8152, |
|
"eval_samples_per_second": 14.889, |
|
"eval_steps_per_second": 0.239, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.5384615384615383, |
|
"grad_norm": 18.888758026934507, |
|
"learning_rate": 5.101117333138557e-07, |
|
"logits/chosen": -2.2510101795196533, |
|
"logits/rejected": -2.2272305488586426, |
|
"logps/chosen": -291.73980712890625, |
|
"logps/rejected": -265.9889221191406, |
|
"loss": 0.1577, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 3.0048184394836426, |
|
"rewards/margins": 4.716927528381348, |
|
"rewards/rejected": -1.712108850479126, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 1.6, |
|
"grad_norm": 19.29495102896881, |
|
"learning_rate": 4.764131040432247e-07, |
|
"logits/chosen": -2.225930690765381, |
|
"logits/rejected": -2.253312349319458, |
|
"logps/chosen": -274.7572021484375, |
|
"logps/rejected": -255.572998046875, |
|
"loss": 0.1913, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": 2.918527126312256, |
|
"rewards/margins": 4.673203945159912, |
|
"rewards/rejected": -1.754677414894104, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.6615384615384614, |
|
"grad_norm": 25.5621828915234, |
|
"learning_rate": 4.428216351440491e-07, |
|
"logits/chosen": -2.251063585281372, |
|
"logits/rejected": -2.2444169521331787, |
|
"logps/chosen": -276.546630859375, |
|
"logps/rejected": -242.7362823486328, |
|
"loss": 0.2192, |
|
"rewards/accuracies": 0.96875, |
|
"rewards/chosen": 3.2366764545440674, |
|
"rewards/margins": 4.48799991607666, |
|
"rewards/rejected": -1.2513238191604614, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 1.7230769230769232, |
|
"grad_norm": 22.24001791940726, |
|
"learning_rate": 4.0948993992719343e-07, |
|
"logits/chosen": -2.2279458045959473, |
|
"logits/rejected": -2.211503744125366, |
|
"logps/chosen": -269.0787658691406, |
|
"logps/rejected": -245.9545440673828, |
|
"loss": 0.2094, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": 3.218578338623047, |
|
"rewards/margins": 4.087203502655029, |
|
"rewards/rejected": -0.8686248660087585, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.7846153846153845, |
|
"grad_norm": 20.544231797693058, |
|
"learning_rate": 3.765694514954795e-07, |
|
"logits/chosen": -2.15586256980896, |
|
"logits/rejected": -2.1282958984375, |
|
"logps/chosen": -286.4849548339844, |
|
"logps/rejected": -253.61160278320312, |
|
"loss": 0.2194, |
|
"rewards/accuracies": 0.8812500238418579, |
|
"rewards/chosen": 3.1202824115753174, |
|
"rewards/margins": 4.100513458251953, |
|
"rewards/rejected": -0.9802314043045044, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 1.8461538461538463, |
|
"grad_norm": 19.256578235605794, |
|
"learning_rate": 3.4420973475033887e-07, |
|
"logits/chosen": -2.118685722351074, |
|
"logits/rejected": -2.0710177421569824, |
|
"logps/chosen": -275.2586364746094, |
|
"logps/rejected": -253.2464141845703, |
|
"loss": 0.2143, |
|
"rewards/accuracies": 0.9312499761581421, |
|
"rewards/chosen": 3.2161738872528076, |
|
"rewards/margins": 4.078971862792969, |
|
"rewards/rejected": -0.862797737121582, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.9076923076923076, |
|
"grad_norm": 23.98963369524157, |
|
"learning_rate": 3.1255780688610506e-07, |
|
"logits/chosen": -2.166189432144165, |
|
"logits/rejected": -2.098736047744751, |
|
"logps/chosen": -276.6828308105469, |
|
"logps/rejected": -248.3531036376953, |
|
"loss": 0.1708, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 3.080463409423828, |
|
"rewards/margins": 4.5233073234558105, |
|
"rewards/rejected": -1.4428437948226929, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 1.9692307692307693, |
|
"grad_norm": 19.206785735628465, |
|
"learning_rate": 2.8175746945909274e-07, |
|
"logits/chosen": -2.137047290802002, |
|
"logits/rejected": -2.1248533725738525, |
|
"logps/chosen": -272.5418395996094, |
|
"logps/rejected": -239.5931854248047, |
|
"loss": 0.2112, |
|
"rewards/accuracies": 0.9437500238418579, |
|
"rewards/chosen": 3.2736434936523438, |
|
"rewards/margins": 4.470524787902832, |
|
"rewards/rejected": -1.1968815326690674, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.9692307692307693, |
|
"eval_logits/chosen": -2.181283712387085, |
|
"eval_logits/rejected": -2.160254716873169, |
|
"eval_logps/chosen": -305.28289794921875, |
|
"eval_logps/rejected": -235.87991333007812, |
|
"eval_loss": 0.48850271105766296, |
|
"eval_rewards/accuracies": 0.8175675868988037, |
|
"eval_rewards/chosen": 2.460352897644043, |
|
"eval_rewards/margins": 2.78669810295105, |
|
"eval_rewards/rejected": -0.32634544372558594, |
|
"eval_runtime": 154.8615, |
|
"eval_samples_per_second": 14.884, |
|
"eval_steps_per_second": 0.239, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.0307692307692307, |
|
"grad_norm": 13.040095905596985, |
|
"learning_rate": 2.51948655066015e-07, |
|
"logits/chosen": -2.1852829456329346, |
|
"logits/rejected": -2.1388115882873535, |
|
"logps/chosen": -276.78057861328125, |
|
"logps/rejected": -241.5438690185547, |
|
"loss": 0.1728, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 3.1819839477539062, |
|
"rewards/margins": 4.5904693603515625, |
|
"rewards/rejected": -1.4084855318069458, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.0923076923076924, |
|
"grad_norm": 9.499778107585676, |
|
"learning_rate": 2.2326679159992156e-07, |
|
"logits/chosen": -2.2025036811828613, |
|
"logits/rejected": -2.180441379547119, |
|
"logps/chosen": -282.98736572265625, |
|
"logps/rejected": -248.19552612304688, |
|
"loss": 0.0866, |
|
"rewards/accuracies": 0.981249988079071, |
|
"rewards/chosen": 3.148271083831787, |
|
"rewards/margins": 4.650267601013184, |
|
"rewards/rejected": -1.5019972324371338, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.1538461538461537, |
|
"grad_norm": 15.892804029114844, |
|
"learning_rate": 1.9584218697198068e-07, |
|
"logits/chosen": -2.2108588218688965, |
|
"logits/rejected": -2.2013297080993652, |
|
"logps/chosen": -273.5797119140625, |
|
"logps/rejected": -248.348876953125, |
|
"loss": 0.1031, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": 3.147243022918701, |
|
"rewards/margins": 4.491833209991455, |
|
"rewards/rejected": -1.3445903062820435, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.2153846153846155, |
|
"grad_norm": 14.22570343292715, |
|
"learning_rate": 1.6979943709444517e-07, |
|
"logits/chosen": -2.190584421157837, |
|
"logits/rejected": -2.174020290374756, |
|
"logps/chosen": -262.9644470214844, |
|
"logps/rejected": -239.3780975341797, |
|
"loss": 0.0932, |
|
"rewards/accuracies": 0.96875, |
|
"rewards/chosen": 2.893162965774536, |
|
"rewards/margins": 4.69353723526001, |
|
"rewards/rejected": -1.8003742694854736, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 2.276923076923077, |
|
"grad_norm": 12.806415359522136, |
|
"learning_rate": 1.4525685981446679e-07, |
|
"logits/chosen": -2.1847140789031982, |
|
"logits/rejected": -2.2051777839660645, |
|
"logps/chosen": -252.59158325195312, |
|
"logps/rejected": -246.86373901367188, |
|
"loss": 0.0929, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": 2.550929307937622, |
|
"rewards/margins": 4.282719612121582, |
|
"rewards/rejected": -1.7317907810211182, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 2.3384615384615386, |
|
"grad_norm": 10.861908687655884, |
|
"learning_rate": 1.2232595737051837e-07, |
|
"logits/chosen": -2.1893796920776367, |
|
"logits/rejected": -2.158172369003296, |
|
"logps/chosen": -281.5319519042969, |
|
"logps/rejected": -253.09927368164062, |
|
"loss": 0.0853, |
|
"rewards/accuracies": 0.981249988079071, |
|
"rewards/chosen": 3.348022937774658, |
|
"rewards/margins": 4.836338996887207, |
|
"rewards/rejected": -1.488316297531128, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 13.046588546578379, |
|
"learning_rate": 1.011109098135996e-07, |
|
"logits/chosen": -2.18019437789917, |
|
"logits/rejected": -2.1145331859588623, |
|
"logps/chosen": -299.452880859375, |
|
"logps/rejected": -278.765869140625, |
|
"loss": 0.0841, |
|
"rewards/accuracies": 0.981249988079071, |
|
"rewards/chosen": 3.1544060707092285, |
|
"rewards/margins": 5.230926990509033, |
|
"rewards/rejected": -2.0765209197998047, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 2.4615384615384617, |
|
"grad_norm": 17.569771715724876, |
|
"learning_rate": 8.170810169472592e-08, |
|
"logits/chosen": -2.1487338542938232, |
|
"logits/rejected": -2.103668689727783, |
|
"logps/chosen": -285.69757080078125, |
|
"logps/rejected": -238.96499633789062, |
|
"loss": 0.1059, |
|
"rewards/accuracies": 0.956250011920929, |
|
"rewards/chosen": 3.010953187942505, |
|
"rewards/margins": 4.7436604499816895, |
|
"rewards/rejected": -1.7327070236206055, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.4615384615384617, |
|
"eval_logits/chosen": -2.1317200660705566, |
|
"eval_logits/rejected": -2.1061103343963623, |
|
"eval_logps/chosen": -306.38006591796875, |
|
"eval_logps/rejected": -239.89051818847656, |
|
"eval_loss": 0.49472159147262573, |
|
"eval_rewards/accuracies": 0.8108108043670654, |
|
"eval_rewards/chosen": 2.350633144378662, |
|
"eval_rewards/margins": 3.0780346393585205, |
|
"eval_rewards/rejected": -0.7274015545845032, |
|
"eval_runtime": 154.9725, |
|
"eval_samples_per_second": 14.874, |
|
"eval_steps_per_second": 0.239, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 2.523076923076923, |
|
"grad_norm": 12.526662899067567, |
|
"learning_rate": 6.420568416906058e-08, |
|
"logits/chosen": -2.110624074935913, |
|
"logits/rejected": -2.0670769214630127, |
|
"logps/chosen": -267.602294921875, |
|
"logps/rejected": -255.283203125, |
|
"loss": 0.0798, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": 3.103508472442627, |
|
"rewards/margins": 5.05588436126709, |
|
"rewards/rejected": -1.9523770809173584, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 2.5846153846153848, |
|
"grad_norm": 15.586951889499693, |
|
"learning_rate": 4.8683174506144064e-08, |
|
"logits/chosen": -2.1358416080474854, |
|
"logits/rejected": -2.0804855823516846, |
|
"logps/chosen": -267.86932373046875, |
|
"logps/rejected": -266.5166320800781, |
|
"loss": 0.0799, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": 3.441650390625, |
|
"rewards/margins": 5.49423885345459, |
|
"rewards/rejected": -2.0525882244110107, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 2.646153846153846, |
|
"grad_norm": 12.921163363324, |
|
"learning_rate": 3.5211094825735145e-08, |
|
"logits/chosen": -2.1344826221466064, |
|
"logits/rejected": -2.095686435699463, |
|
"logps/chosen": -273.34466552734375, |
|
"logps/rejected": -259.7888488769531, |
|
"loss": 0.0906, |
|
"rewards/accuracies": 0.9937499761581421, |
|
"rewards/chosen": 3.337371826171875, |
|
"rewards/margins": 5.457432270050049, |
|
"rewards/rejected": -2.1200602054595947, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 2.707692307692308, |
|
"grad_norm": 11.698637983314368, |
|
"learning_rate": 2.385065170056283e-08, |
|
"logits/chosen": -2.1433401107788086, |
|
"logits/rejected": -2.081160545349121, |
|
"logps/chosen": -273.98004150390625, |
|
"logps/rejected": -262.38897705078125, |
|
"loss": 0.0834, |
|
"rewards/accuracies": 0.981249988079071, |
|
"rewards/chosen": 3.1992526054382324, |
|
"rewards/margins": 5.575494766235352, |
|
"rewards/rejected": -2.376242160797119, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 2.769230769230769, |
|
"grad_norm": 12.58060499267709, |
|
"learning_rate": 1.465345808162427e-08, |
|
"logits/chosen": -2.1173667907714844, |
|
"logits/rejected": -2.0692567825317383, |
|
"logps/chosen": -273.2080383300781, |
|
"logps/rejected": -258.080810546875, |
|
"loss": 0.0774, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": 3.0734853744506836, |
|
"rewards/margins": 5.315826892852783, |
|
"rewards/rejected": -2.2423415184020996, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 2.830769230769231, |
|
"grad_norm": 11.786260598928802, |
|
"learning_rate": 7.661298809381877e-09, |
|
"logits/chosen": -2.126652717590332, |
|
"logits/rejected": -2.076254367828369, |
|
"logps/chosen": -289.486083984375, |
|
"logps/rejected": -265.6861877441406, |
|
"loss": 0.1024, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": 3.2749016284942627, |
|
"rewards/margins": 5.374606132507324, |
|
"rewards/rejected": -2.0997047424316406, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 2.8923076923076922, |
|
"grad_norm": 17.82006078758849, |
|
"learning_rate": 2.9059407761923836e-09, |
|
"logits/chosen": -2.1355929374694824, |
|
"logits/rejected": -2.1044511795043945, |
|
"logps/chosen": -269.39984130859375, |
|
"logps/rejected": -247.22671508789062, |
|
"loss": 0.0907, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": 3.20941424369812, |
|
"rewards/margins": 4.841259956359863, |
|
"rewards/rejected": -1.6318458318710327, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 2.953846153846154, |
|
"grad_norm": 17.458154920989255, |
|
"learning_rate": 4.0898860244215074e-10, |
|
"logits/chosen": -2.1104207038879395, |
|
"logits/rejected": -2.0696866512298584, |
|
"logps/chosen": -259.77471923828125, |
|
"logps/rejected": -241.05380249023438, |
|
"loss": 0.1134, |
|
"rewards/accuracies": 0.96875, |
|
"rewards/chosen": 2.9383902549743652, |
|
"rewards/margins": 4.824470520019531, |
|
"rewards/rejected": -1.8860807418823242, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.953846153846154, |
|
"eval_logits/chosen": -2.1210827827453613, |
|
"eval_logits/rejected": -2.094482660293579, |
|
"eval_logps/chosen": -306.5203552246094, |
|
"eval_logps/rejected": -240.8570556640625, |
|
"eval_loss": 0.4984860122203827, |
|
"eval_rewards/accuracies": 0.8108108043670654, |
|
"eval_rewards/chosen": 2.3366057872772217, |
|
"eval_rewards/margins": 3.1606650352478027, |
|
"eval_rewards/rejected": -0.8240591287612915, |
|
"eval_runtime": 154.8776, |
|
"eval_samples_per_second": 14.883, |
|
"eval_steps_per_second": 0.239, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.9907692307692306, |
|
"step": 243, |
|
"total_flos": 2865208385077248.0, |
|
"train_loss": 0.2735099794933335, |
|
"train_runtime": 9352.9566, |
|
"train_samples_per_second": 6.652, |
|
"train_steps_per_second": 0.026 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 243, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 40, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 2865208385077248.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|