|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.2488646684831971, |
|
"eval_steps": 500, |
|
"global_step": 137, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0018165304268846503, |
|
"grad_norm": 0.7654335498809814, |
|
"learning_rate": 9.99999541185824e-06, |
|
"logits/chosen": 0.10746052116155624, |
|
"logits/rejected": 0.16293610632419586, |
|
"logps/chosen": -51.70099639892578, |
|
"logps/rejected": -56.85424041748047, |
|
"loss": 2.7726, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0036330608537693005, |
|
"grad_norm": 0.8493374586105347, |
|
"learning_rate": 9.999981647442309e-06, |
|
"logits/chosen": 0.12589210271835327, |
|
"logits/rejected": 0.18243370950222015, |
|
"logps/chosen": -61.092987060546875, |
|
"logps/rejected": -67.26304626464844, |
|
"loss": 2.7516, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": 0.005276113282889128, |
|
"rewards/margins": 0.010965153574943542, |
|
"rewards/rejected": -0.00568903936073184, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.005449591280653951, |
|
"grad_norm": 0.739442765712738, |
|
"learning_rate": 9.999958706780277e-06, |
|
"logits/chosen": 0.0766456350684166, |
|
"logits/rejected": 0.1422545611858368, |
|
"logps/chosen": -55.764503479003906, |
|
"logps/rejected": -57.38707733154297, |
|
"loss": 2.7902, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.0027439936529845, |
|
"rewards/margins": -0.008403200656175613, |
|
"rewards/rejected": 0.005659207701683044, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.007266121707538601, |
|
"grad_norm": 0.7261953949928284, |
|
"learning_rate": 9.999926589918927e-06, |
|
"logits/chosen": 0.1812177300453186, |
|
"logits/rejected": 0.13825736939907074, |
|
"logps/chosen": -56.98471450805664, |
|
"logps/rejected": -55.47361373901367, |
|
"loss": 2.7537, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": 0.006436044350266457, |
|
"rewards/margins": 0.009899044409394264, |
|
"rewards/rejected": -0.003463000524789095, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.009082652134423252, |
|
"grad_norm": 0.850686252117157, |
|
"learning_rate": 9.999885296923748e-06, |
|
"logits/chosen": 0.17714114487171173, |
|
"logits/rejected": 0.12910494208335876, |
|
"logps/chosen": -71.72655487060547, |
|
"logps/rejected": -70.75139617919922, |
|
"loss": 2.79, |
|
"rewards/accuracies": 0.453125, |
|
"rewards/chosen": -0.008413795381784439, |
|
"rewards/margins": -0.007964953780174255, |
|
"rewards/rejected": -0.0004488405538722873, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.010899182561307902, |
|
"grad_norm": 0.6627741456031799, |
|
"learning_rate": 9.999834827878942e-06, |
|
"logits/chosen": 0.13277305662631989, |
|
"logits/rejected": 0.20749863982200623, |
|
"logps/chosen": -47.56520080566406, |
|
"logps/rejected": -53.18217468261719, |
|
"loss": 2.7732, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.001662571681663394, |
|
"rewards/margins": 0.00014669005759060383, |
|
"rewards/rejected": 0.0015158820897340775, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.012715712988192553, |
|
"grad_norm": 0.8335347175598145, |
|
"learning_rate": 9.999775182887429e-06, |
|
"logits/chosen": 0.07216031849384308, |
|
"logits/rejected": 0.13223227858543396, |
|
"logps/chosen": -73.06997680664062, |
|
"logps/rejected": -75.58594512939453, |
|
"loss": 2.7716, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.004659444559365511, |
|
"rewards/margins": 0.0013355333358049393, |
|
"rewards/rejected": -0.005994977429509163, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.014532243415077202, |
|
"grad_norm": 0.7128610014915466, |
|
"learning_rate": 9.999706362070833e-06, |
|
"logits/chosen": 0.11368857324123383, |
|
"logits/rejected": 0.15805509686470032, |
|
"logps/chosen": -56.45945358276367, |
|
"logps/rejected": -56.9697151184082, |
|
"loss": 2.7544, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.0038012717850506306, |
|
"rewards/margins": 0.009716177359223366, |
|
"rewards/rejected": -0.013517449609935284, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01634877384196185, |
|
"grad_norm": 0.77040034532547, |
|
"learning_rate": 9.99962836556949e-06, |
|
"logits/chosen": 0.1818978637456894, |
|
"logits/rejected": 0.20750640332698822, |
|
"logps/chosen": -59.184532165527344, |
|
"logps/rejected": -61.59263610839844, |
|
"loss": 2.7688, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.009461460635066032, |
|
"rewards/margins": 0.002394758863374591, |
|
"rewards/rejected": -0.011856218799948692, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.018165304268846504, |
|
"grad_norm": 0.7393640875816345, |
|
"learning_rate": 9.99954119354245e-06, |
|
"logits/chosen": 0.14076584577560425, |
|
"logits/rejected": 0.22711633145809174, |
|
"logps/chosen": -56.34330749511719, |
|
"logps/rejected": -60.51960372924805, |
|
"loss": 2.7827, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.004167273174971342, |
|
"rewards/margins": -0.00429012393578887, |
|
"rewards/rejected": 0.00012284982949495316, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.019981834695731154, |
|
"grad_norm": 0.7271163463592529, |
|
"learning_rate": 9.999444846167473e-06, |
|
"logits/chosen": 0.20141208171844482, |
|
"logits/rejected": 0.2038215547800064, |
|
"logps/chosen": -57.712646484375, |
|
"logps/rejected": -54.343231201171875, |
|
"loss": 2.7565, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.0009536671568639576, |
|
"rewards/margins": 0.00866839662194252, |
|
"rewards/rejected": -0.007714730221778154, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.021798365122615803, |
|
"grad_norm": 0.9451335668563843, |
|
"learning_rate": 9.999339323641027e-06, |
|
"logits/chosen": 0.16046015918254852, |
|
"logits/rejected": 0.1963367462158203, |
|
"logps/chosen": -51.579856872558594, |
|
"logps/rejected": -54.97206115722656, |
|
"loss": 2.7732, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.005096627864986658, |
|
"rewards/margins": 0.00017735245637595654, |
|
"rewards/rejected": -0.005273980088531971, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.023614895549500452, |
|
"grad_norm": 0.8279874324798584, |
|
"learning_rate": 9.99922462617829e-06, |
|
"logits/chosen": 0.12593263387680054, |
|
"logits/rejected": 0.11708654463291168, |
|
"logps/chosen": -62.25929641723633, |
|
"logps/rejected": -60.633296966552734, |
|
"loss": 2.7693, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.013154934160411358, |
|
"rewards/margins": 0.002298446139320731, |
|
"rewards/rejected": -0.015453380532562733, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.025431425976385105, |
|
"grad_norm": 0.7863137125968933, |
|
"learning_rate": 9.999100754013152e-06, |
|
"logits/chosen": 0.12714676558971405, |
|
"logits/rejected": 0.18583568930625916, |
|
"logps/chosen": -59.119529724121094, |
|
"logps/rejected": -65.85352325439453, |
|
"loss": 2.7715, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.01445689145475626, |
|
"rewards/margins": 0.001161783467978239, |
|
"rewards/rejected": -0.015618674457073212, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.027247956403269755, |
|
"grad_norm": 0.7623695731163025, |
|
"learning_rate": 9.998967707398207e-06, |
|
"logits/chosen": 0.09560943394899368, |
|
"logits/rejected": 0.14794519543647766, |
|
"logps/chosen": -47.649497985839844, |
|
"logps/rejected": -57.21004867553711, |
|
"loss": 2.7405, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.0010593307670205832, |
|
"rewards/margins": 0.0166107639670372, |
|
"rewards/rejected": -0.017670094966888428, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.029064486830154404, |
|
"grad_norm": 0.7020726203918457, |
|
"learning_rate": 9.998825486604765e-06, |
|
"logits/chosen": 0.19466270506381989, |
|
"logits/rejected": 0.21006342768669128, |
|
"logps/chosen": -50.8302116394043, |
|
"logps/rejected": -58.202754974365234, |
|
"loss": 2.7628, |
|
"rewards/accuracies": 0.484375, |
|
"rewards/chosen": -0.009497416205704212, |
|
"rewards/margins": 0.005540419369935989, |
|
"rewards/rejected": -0.015037836506962776, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.030881017257039057, |
|
"grad_norm": 0.7085703611373901, |
|
"learning_rate": 9.998674091922833e-06, |
|
"logits/chosen": 0.06517614424228668, |
|
"logits/rejected": 0.15166127681732178, |
|
"logps/chosen": -45.203216552734375, |
|
"logps/rejected": -51.454952239990234, |
|
"loss": 2.7568, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.009879402816295624, |
|
"rewards/margins": 0.00827767699956894, |
|
"rewards/rejected": -0.018157079815864563, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0326975476839237, |
|
"grad_norm": 0.8115389347076416, |
|
"learning_rate": 9.998513523661136e-06, |
|
"logits/chosen": 0.09045977890491486, |
|
"logits/rejected": 0.1498938351869583, |
|
"logps/chosen": -60.35250473022461, |
|
"logps/rejected": -62.11174774169922, |
|
"loss": 2.7888, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.01872912421822548, |
|
"rewards/margins": -0.0073302448727190495, |
|
"rewards/rejected": -0.011398878879845142, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03451407811080836, |
|
"grad_norm": 0.8222002983093262, |
|
"learning_rate": 9.998343782147098e-06, |
|
"logits/chosen": 0.08091418445110321, |
|
"logits/rejected": 0.1225643903017044, |
|
"logps/chosen": -54.25471115112305, |
|
"logps/rejected": -57.91140365600586, |
|
"loss": 2.7652, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.018515082076191902, |
|
"rewards/margins": 0.004185312893241644, |
|
"rewards/rejected": -0.022700395435094833, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.03633060853769301, |
|
"grad_norm": 0.862041711807251, |
|
"learning_rate": 9.998164867726853e-06, |
|
"logits/chosen": 0.04738985374569893, |
|
"logits/rejected": 0.11436720937490463, |
|
"logps/chosen": -62.22935104370117, |
|
"logps/rejected": -67.67117309570312, |
|
"loss": 2.7607, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.01734255626797676, |
|
"rewards/margins": 0.006629224866628647, |
|
"rewards/rejected": -0.023971781134605408, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03814713896457766, |
|
"grad_norm": 0.8169613480567932, |
|
"learning_rate": 9.997976780765237e-06, |
|
"logits/chosen": 0.14641642570495605, |
|
"logits/rejected": 0.17384907603263855, |
|
"logps/chosen": -61.21760559082031, |
|
"logps/rejected": -61.094486236572266, |
|
"loss": 2.7752, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.021789425984025, |
|
"rewards/margins": -0.000787546974606812, |
|
"rewards/rejected": -0.02100187912583351, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.03996366939146231, |
|
"grad_norm": 0.7457932829856873, |
|
"learning_rate": 9.997779521645793e-06, |
|
"logits/chosen": 0.16773203015327454, |
|
"logits/rejected": 0.16435235738754272, |
|
"logps/chosen": -55.192623138427734, |
|
"logps/rejected": -53.05280303955078, |
|
"loss": 2.7661, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.021701142191886902, |
|
"rewards/margins": 0.0036970973014831543, |
|
"rewards/rejected": -0.025398239493370056, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04178019981834696, |
|
"grad_norm": 0.8263576030731201, |
|
"learning_rate": 9.997573090770766e-06, |
|
"logits/chosen": 0.10863066464662552, |
|
"logits/rejected": 0.1277829110622406, |
|
"logps/chosen": -61.80084228515625, |
|
"logps/rejected": -64.94878387451172, |
|
"loss": 2.7526, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.019380860030651093, |
|
"rewards/margins": 0.010501865297555923, |
|
"rewards/rejected": -0.029882723465561867, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.043596730245231606, |
|
"grad_norm": 0.7587623000144958, |
|
"learning_rate": 9.997357488561107e-06, |
|
"logits/chosen": 0.17992480099201202, |
|
"logits/rejected": 0.16478824615478516, |
|
"logps/chosen": -62.946449279785156, |
|
"logps/rejected": -61.98965072631836, |
|
"loss": 2.7534, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.02564322203397751, |
|
"rewards/margins": 0.010079940780997276, |
|
"rewards/rejected": -0.035723160952329636, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.045413260672116255, |
|
"grad_norm": 0.7733523845672607, |
|
"learning_rate": 9.997132715456464e-06, |
|
"logits/chosen": 0.1268949955701828, |
|
"logits/rejected": 0.14968058466911316, |
|
"logps/chosen": -53.57306671142578, |
|
"logps/rejected": -58.07795715332031, |
|
"loss": 2.7602, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.030423898249864578, |
|
"rewards/margins": 0.006796791218221188, |
|
"rewards/rejected": -0.03722068667411804, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.047229791099000905, |
|
"grad_norm": 0.8851524591445923, |
|
"learning_rate": 9.99689877191519e-06, |
|
"logits/chosen": 0.17443379759788513, |
|
"logits/rejected": 0.1787741333246231, |
|
"logps/chosen": -58.031558990478516, |
|
"logps/rejected": -59.206520080566406, |
|
"loss": 2.762, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.02750026062130928, |
|
"rewards/margins": 0.006129855290055275, |
|
"rewards/rejected": -0.033630117774009705, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.04904632152588556, |
|
"grad_norm": 0.8524475693702698, |
|
"learning_rate": 9.996655658414331e-06, |
|
"logits/chosen": 0.18196682631969452, |
|
"logits/rejected": 0.237998366355896, |
|
"logps/chosen": -63.18816375732422, |
|
"logps/rejected": -67.31481170654297, |
|
"loss": 2.7519, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.03693533316254616, |
|
"rewards/margins": 0.011266030371189117, |
|
"rewards/rejected": -0.048201363533735275, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05086285195277021, |
|
"grad_norm": 0.8785350918769836, |
|
"learning_rate": 9.996403375449647e-06, |
|
"logits/chosen": 0.12056512385606766, |
|
"logits/rejected": 0.17535904049873352, |
|
"logps/chosen": -62.035221099853516, |
|
"logps/rejected": -69.88546752929688, |
|
"loss": 2.743, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.03874595835804939, |
|
"rewards/margins": 0.015791552141308784, |
|
"rewards/rejected": -0.054537512362003326, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.05267938237965486, |
|
"grad_norm": 0.9492508172988892, |
|
"learning_rate": 9.996141923535582e-06, |
|
"logits/chosen": 0.12200314551591873, |
|
"logits/rejected": 0.13907021284103394, |
|
"logps/chosen": -65.31945037841797, |
|
"logps/rejected": -66.59127807617188, |
|
"loss": 2.7641, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.0383562371134758, |
|
"rewards/margins": 0.005150892771780491, |
|
"rewards/rejected": -0.043507132679224014, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.05449591280653951, |
|
"grad_norm": 0.8084173202514648, |
|
"learning_rate": 9.995871303205282e-06, |
|
"logits/chosen": 0.07616369426250458, |
|
"logits/rejected": 0.2001558393239975, |
|
"logps/chosen": -51.103309631347656, |
|
"logps/rejected": -62.28483581542969, |
|
"loss": 2.7519, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.04615113511681557, |
|
"rewards/margins": 0.011578184552490711, |
|
"rewards/rejected": -0.057729318737983704, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.05631244323342416, |
|
"grad_norm": 0.943800151348114, |
|
"learning_rate": 9.995591515010589e-06, |
|
"logits/chosen": 0.04337490350008011, |
|
"logits/rejected": 0.15334263443946838, |
|
"logps/chosen": -57.265567779541016, |
|
"logps/rejected": -68.5506820678711, |
|
"loss": 2.7253, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.04982424154877663, |
|
"rewards/margins": 0.024712499231100082, |
|
"rewards/rejected": -0.07453674077987671, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.05812897366030881, |
|
"grad_norm": 0.9372840523719788, |
|
"learning_rate": 9.99530255952204e-06, |
|
"logits/chosen": 0.13595089316368103, |
|
"logits/rejected": 0.14287832379341125, |
|
"logps/chosen": -61.515377044677734, |
|
"logps/rejected": -63.1092529296875, |
|
"loss": 2.7571, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.05759892612695694, |
|
"rewards/margins": 0.009150844067335129, |
|
"rewards/rejected": -0.06674977391958237, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.05994550408719346, |
|
"grad_norm": 0.8408157825469971, |
|
"learning_rate": 9.995004437328866e-06, |
|
"logits/chosen": 0.18553151190280914, |
|
"logits/rejected": 0.14024314284324646, |
|
"logps/chosen": -61.73942565917969, |
|
"logps/rejected": -53.5772819519043, |
|
"loss": 2.7627, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.06591347604990005, |
|
"rewards/margins": 0.006169732194393873, |
|
"rewards/rejected": -0.07208321243524551, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.061762034514078114, |
|
"grad_norm": 0.8856032490730286, |
|
"learning_rate": 9.994697149038989e-06, |
|
"logits/chosen": 0.08410881459712982, |
|
"logits/rejected": 0.13991469144821167, |
|
"logps/chosen": -60.161231994628906, |
|
"logps/rejected": -60.03232192993164, |
|
"loss": 2.7437, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.06631935387849808, |
|
"rewards/margins": 0.016240764409303665, |
|
"rewards/rejected": -0.08256012946367264, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.06357856494096276, |
|
"grad_norm": 0.9037399291992188, |
|
"learning_rate": 9.994380695279025e-06, |
|
"logits/chosen": 0.06669525057077408, |
|
"logits/rejected": 0.16171438992023468, |
|
"logps/chosen": -59.874202728271484, |
|
"logps/rejected": -67.9597396850586, |
|
"loss": 2.7221, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.06999661773443222, |
|
"rewards/margins": 0.027104372158646584, |
|
"rewards/rejected": -0.09710099548101425, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0653950953678474, |
|
"grad_norm": 0.8819396495819092, |
|
"learning_rate": 9.994055076694276e-06, |
|
"logits/chosen": 0.18232542276382446, |
|
"logits/rejected": 0.2102084755897522, |
|
"logps/chosen": -59.298095703125, |
|
"logps/rejected": -63.70309066772461, |
|
"loss": 2.7179, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.08072888106107712, |
|
"rewards/margins": 0.029247857630252838, |
|
"rewards/rejected": -0.10997673869132996, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.06721162579473206, |
|
"grad_norm": 0.9046475291252136, |
|
"learning_rate": 9.993720293948739e-06, |
|
"logits/chosen": 0.13181668519973755, |
|
"logits/rejected": 0.1316901445388794, |
|
"logps/chosen": -71.46318054199219, |
|
"logps/rejected": -76.40208435058594, |
|
"loss": 2.7505, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.08365033566951752, |
|
"rewards/margins": 0.01496690884232521, |
|
"rewards/rejected": -0.09861725568771362, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.06902815622161672, |
|
"grad_norm": 0.9267721772193909, |
|
"learning_rate": 9.993376347725091e-06, |
|
"logits/chosen": 0.14555476605892181, |
|
"logits/rejected": 0.21660488843917847, |
|
"logps/chosen": -61.28938293457031, |
|
"logps/rejected": -67.28329467773438, |
|
"loss": 2.7007, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.07440298050642014, |
|
"rewards/margins": 0.03913095220923424, |
|
"rewards/rejected": -0.11353392899036407, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07084468664850137, |
|
"grad_norm": 0.8545824885368347, |
|
"learning_rate": 9.993023238724696e-06, |
|
"logits/chosen": 0.09474823623895645, |
|
"logits/rejected": 0.09304340183734894, |
|
"logps/chosen": -61.27484130859375, |
|
"logps/rejected": -62.003597259521484, |
|
"loss": 2.7406, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.09001085162162781, |
|
"rewards/margins": 0.017711879685521126, |
|
"rewards/rejected": -0.10772272944450378, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.07266121707538602, |
|
"grad_norm": 0.9086014628410339, |
|
"learning_rate": 9.99266096766761e-06, |
|
"logits/chosen": 0.17910577356815338, |
|
"logits/rejected": 0.14881345629692078, |
|
"logps/chosen": -59.22669219970703, |
|
"logps/rejected": -55.103187561035156, |
|
"loss": 2.7422, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.1086670309305191, |
|
"rewards/margins": 0.01710200309753418, |
|
"rewards/rejected": -0.12576903402805328, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07447774750227067, |
|
"grad_norm": 0.9802166223526001, |
|
"learning_rate": 9.992289535292565e-06, |
|
"logits/chosen": 0.18225271999835968, |
|
"logits/rejected": 0.20810523629188538, |
|
"logps/chosen": -55.57660675048828, |
|
"logps/rejected": -59.129417419433594, |
|
"loss": 2.7505, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.13455544412136078, |
|
"rewards/margins": 0.012644169852137566, |
|
"rewards/rejected": -0.1471996009349823, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.07629427792915532, |
|
"grad_norm": 0.919211208820343, |
|
"learning_rate": 9.991908942356977e-06, |
|
"logits/chosen": 0.07866425067186356, |
|
"logits/rejected": 0.07479682564735413, |
|
"logps/chosen": -66.11737060546875, |
|
"logps/rejected": -64.02122497558594, |
|
"loss": 2.7855, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.15568655729293823, |
|
"rewards/margins": -0.0022164471447467804, |
|
"rewards/rejected": -0.15347009897232056, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.07811080835603997, |
|
"grad_norm": 0.9612410068511963, |
|
"learning_rate": 9.991519189636937e-06, |
|
"logits/chosen": 0.11137107014656067, |
|
"logits/rejected": 0.14462286233901978, |
|
"logps/chosen": -67.43673706054688, |
|
"logps/rejected": -74.29210662841797, |
|
"loss": 2.7203, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.16468609869480133, |
|
"rewards/margins": 0.032481495290994644, |
|
"rewards/rejected": -0.19716759026050568, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.07992733878292461, |
|
"grad_norm": 0.8044790625572205, |
|
"learning_rate": 9.991120277927224e-06, |
|
"logits/chosen": 0.08509679138660431, |
|
"logits/rejected": 0.11362393945455551, |
|
"logps/chosen": -56.867591857910156, |
|
"logps/rejected": -55.66028594970703, |
|
"loss": 2.7302, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.15751589834690094, |
|
"rewards/margins": 0.02398722618818283, |
|
"rewards/rejected": -0.18150311708450317, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.08174386920980926, |
|
"grad_norm": 0.8599545955657959, |
|
"learning_rate": 9.990712208041284e-06, |
|
"logits/chosen": 0.1391746699810028, |
|
"logits/rejected": 0.1746281236410141, |
|
"logps/chosen": -62.22108459472656, |
|
"logps/rejected": -58.50814437866211, |
|
"loss": 2.7288, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.16528920829296112, |
|
"rewards/margins": 0.02449135296046734, |
|
"rewards/rejected": -0.1897805631160736, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.08356039963669391, |
|
"grad_norm": 0.9419183731079102, |
|
"learning_rate": 9.990294980811244e-06, |
|
"logits/chosen": 0.1693490892648697, |
|
"logits/rejected": 0.18133510649204254, |
|
"logps/chosen": -60.769744873046875, |
|
"logps/rejected": -65.61067199707031, |
|
"loss": 2.7003, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.16450145840644836, |
|
"rewards/margins": 0.04409575089812279, |
|
"rewards/rejected": -0.20859721302986145, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.08537693006357856, |
|
"grad_norm": 0.9839057326316833, |
|
"learning_rate": 9.989868597087907e-06, |
|
"logits/chosen": 0.1363808959722519, |
|
"logits/rejected": 0.22261787950992584, |
|
"logps/chosen": -65.37456512451172, |
|
"logps/rejected": -72.04705047607422, |
|
"loss": 2.6636, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.18397849798202515, |
|
"rewards/margins": 0.06703396141529083, |
|
"rewards/rejected": -0.2510124444961548, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.08719346049046321, |
|
"grad_norm": 0.8818926811218262, |
|
"learning_rate": 9.989433057740738e-06, |
|
"logits/chosen": 0.1580447256565094, |
|
"logits/rejected": 0.18051129579544067, |
|
"logps/chosen": -56.6561279296875, |
|
"logps/rejected": -64.66617584228516, |
|
"loss": 2.6658, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2094140350818634, |
|
"rewards/margins": 0.05971873924136162, |
|
"rewards/rejected": -0.2691327929496765, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.08900999091734786, |
|
"grad_norm": 0.8838489651679993, |
|
"learning_rate": 9.98898836365788e-06, |
|
"logits/chosen": 0.11484000086784363, |
|
"logits/rejected": 0.12583914399147034, |
|
"logps/chosen": -65.76107788085938, |
|
"logps/rejected": -75.4742660522461, |
|
"loss": 2.6777, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.19882163405418396, |
|
"rewards/margins": 0.0690067932009697, |
|
"rewards/rejected": -0.26782843470573425, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.09082652134423251, |
|
"grad_norm": 0.9035817980766296, |
|
"learning_rate": 9.988534515746141e-06, |
|
"logits/chosen": 0.12592823803424835, |
|
"logits/rejected": 0.18080386519432068, |
|
"logps/chosen": -60.42765808105469, |
|
"logps/rejected": -65.03330993652344, |
|
"loss": 2.68, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.25037306547164917, |
|
"rewards/margins": 0.05723651498556137, |
|
"rewards/rejected": -0.30760958790779114, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09264305177111716, |
|
"grad_norm": 1.0728137493133545, |
|
"learning_rate": 9.988071514930998e-06, |
|
"logits/chosen": 0.0955006331205368, |
|
"logits/rejected": 0.15643228590488434, |
|
"logps/chosen": -56.28706359863281, |
|
"logps/rejected": -70.3702392578125, |
|
"loss": 2.6104, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.2517072558403015, |
|
"rewards/margins": 0.11176257580518723, |
|
"rewards/rejected": -0.36346977949142456, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.09445958219800181, |
|
"grad_norm": 0.9143723845481873, |
|
"learning_rate": 9.987599362156587e-06, |
|
"logits/chosen": 0.1018611341714859, |
|
"logits/rejected": 0.21036753058433533, |
|
"logps/chosen": -55.815738677978516, |
|
"logps/rejected": -66.23766326904297, |
|
"loss": 2.6207, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.2496231347322464, |
|
"rewards/margins": 0.09438168257474899, |
|
"rewards/rejected": -0.3440048098564148, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.09627611262488647, |
|
"grad_norm": 0.8731315732002258, |
|
"learning_rate": 9.987118058385712e-06, |
|
"logits/chosen": 0.12294681370258331, |
|
"logits/rejected": 0.12897568941116333, |
|
"logps/chosen": -63.27684020996094, |
|
"logps/rejected": -65.21257019042969, |
|
"loss": 2.7093, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.28397974371910095, |
|
"rewards/margins": 0.05316353589296341, |
|
"rewards/rejected": -0.33714330196380615, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.09809264305177112, |
|
"grad_norm": 0.951884388923645, |
|
"learning_rate": 9.986627604599835e-06, |
|
"logits/chosen": 0.1588897705078125, |
|
"logits/rejected": 0.10302369296550751, |
|
"logps/chosen": -62.071083068847656, |
|
"logps/rejected": -59.575416564941406, |
|
"loss": 2.7725, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.31246036291122437, |
|
"rewards/margins": 0.011688929051160812, |
|
"rewards/rejected": -0.3241492807865143, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.09990917347865577, |
|
"grad_norm": 1.0379350185394287, |
|
"learning_rate": 9.986128001799077e-06, |
|
"logits/chosen": 0.15824833512306213, |
|
"logits/rejected": 0.16385456919670105, |
|
"logps/chosen": -79.65828704833984, |
|
"logps/rejected": -82.06480407714844, |
|
"loss": 2.6826, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.36005648970603943, |
|
"rewards/margins": 0.07091177999973297, |
|
"rewards/rejected": -0.4309682846069336, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.10172570390554042, |
|
"grad_norm": 0.8740183115005493, |
|
"learning_rate": 9.985619251002214e-06, |
|
"logits/chosen": 0.0996306911110878, |
|
"logits/rejected": 0.14288735389709473, |
|
"logps/chosen": -59.8499755859375, |
|
"logps/rejected": -63.10956573486328, |
|
"loss": 2.664, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.3242360055446625, |
|
"rewards/margins": 0.06764136999845505, |
|
"rewards/rejected": -0.39187735319137573, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.10354223433242507, |
|
"grad_norm": 0.9942138195037842, |
|
"learning_rate": 9.985101353246676e-06, |
|
"logits/chosen": 0.12070289999246597, |
|
"logits/rejected": 0.1866157352924347, |
|
"logps/chosen": -70.04698181152344, |
|
"logps/rejected": -76.2674560546875, |
|
"loss": 2.6165, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.4166935980319977, |
|
"rewards/margins": 0.10780875384807587, |
|
"rewards/rejected": -0.5245023965835571, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.10535876475930972, |
|
"grad_norm": 0.9078152775764465, |
|
"learning_rate": 9.98457430958855e-06, |
|
"logits/chosen": 0.16699416935443878, |
|
"logits/rejected": 0.216594398021698, |
|
"logps/chosen": -59.81795883178711, |
|
"logps/rejected": -64.79395294189453, |
|
"loss": 2.551, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.39265742897987366, |
|
"rewards/margins": 0.14742323756217957, |
|
"rewards/rejected": -0.5400806665420532, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.10717529518619437, |
|
"grad_norm": 0.9305158257484436, |
|
"learning_rate": 9.984038121102569e-06, |
|
"logits/chosen": 0.16001635789871216, |
|
"logits/rejected": 0.18797104060649872, |
|
"logps/chosen": -58.087158203125, |
|
"logps/rejected": -59.003414154052734, |
|
"loss": 2.6345, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.3626652956008911, |
|
"rewards/margins": 0.08553728461265564, |
|
"rewards/rejected": -0.44820258021354675, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.10899182561307902, |
|
"grad_norm": 0.9690614342689514, |
|
"learning_rate": 9.983492788882113e-06, |
|
"logits/chosen": 0.161406010389328, |
|
"logits/rejected": 0.14847029745578766, |
|
"logps/chosen": -61.01350021362305, |
|
"logps/rejected": -60.993080139160156, |
|
"loss": 2.7099, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.4803884029388428, |
|
"rewards/margins": 0.0831567794084549, |
|
"rewards/rejected": -0.5635451078414917, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11080835603996367, |
|
"grad_norm": 0.971172034740448, |
|
"learning_rate": 9.98293831403921e-06, |
|
"logits/chosen": 0.14095918834209442, |
|
"logits/rejected": 0.16142162680625916, |
|
"logps/chosen": -62.31782150268555, |
|
"logps/rejected": -64.44837188720703, |
|
"loss": 2.6373, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.47784310579299927, |
|
"rewards/margins": 0.09006088227033615, |
|
"rewards/rejected": -0.567903995513916, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.11262488646684832, |
|
"grad_norm": 0.9984462857246399, |
|
"learning_rate": 9.982374697704532e-06, |
|
"logits/chosen": 0.08945554494857788, |
|
"logits/rejected": 0.17623476684093475, |
|
"logps/chosen": -58.84153366088867, |
|
"logps/rejected": -72.59091186523438, |
|
"loss": 2.5465, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.41726160049438477, |
|
"rewards/margins": 0.16527526080608368, |
|
"rewards/rejected": -0.5825368165969849, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.11444141689373297, |
|
"grad_norm": 1.007622241973877, |
|
"learning_rate": 9.981801941027388e-06, |
|
"logits/chosen": 0.16932496428489685, |
|
"logits/rejected": 0.24257409572601318, |
|
"logps/chosen": -61.62455749511719, |
|
"logps/rejected": -69.40538024902344, |
|
"loss": 2.5876, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.4484747052192688, |
|
"rewards/margins": 0.11513397097587585, |
|
"rewards/rejected": -0.563608705997467, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.11625794732061762, |
|
"grad_norm": 1.021275281906128, |
|
"learning_rate": 9.981220045175731e-06, |
|
"logits/chosen": 0.13468854129314423, |
|
"logits/rejected": 0.10648790001869202, |
|
"logps/chosen": -56.91660690307617, |
|
"logps/rejected": -54.84038162231445, |
|
"loss": 2.7096, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.47126585245132446, |
|
"rewards/margins": 0.05656271427869797, |
|
"rewards/rejected": -0.527828574180603, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.11807447774750227, |
|
"grad_norm": 0.9580955505371094, |
|
"learning_rate": 9.980629011336149e-06, |
|
"logits/chosen": 0.10542559623718262, |
|
"logits/rejected": 0.18956655263900757, |
|
"logps/chosen": -63.21677017211914, |
|
"logps/rejected": -68.92955017089844, |
|
"loss": 2.5447, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.4920822083950043, |
|
"rewards/margins": 0.1548418551683426, |
|
"rewards/rejected": -0.6469241380691528, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.11989100817438691, |
|
"grad_norm": 0.9601484537124634, |
|
"learning_rate": 9.98002884071386e-06, |
|
"logits/chosen": 0.1211993545293808, |
|
"logits/rejected": 0.15335297584533691, |
|
"logps/chosen": -65.79328918457031, |
|
"logps/rejected": -71.8165283203125, |
|
"loss": 2.6013, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.4890082776546478, |
|
"rewards/margins": 0.16498278081417084, |
|
"rewards/rejected": -0.6539911031723022, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.12170753860127158, |
|
"grad_norm": 0.9475951194763184, |
|
"learning_rate": 9.97941953453272e-06, |
|
"logits/chosen": 0.129757359623909, |
|
"logits/rejected": 0.19597335159778595, |
|
"logps/chosen": -61.52772903442383, |
|
"logps/rejected": -62.8316764831543, |
|
"loss": 2.6008, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.6176788210868835, |
|
"rewards/margins": 0.11390677094459534, |
|
"rewards/rejected": -0.7315855622291565, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.12352406902815623, |
|
"grad_norm": 1.0577987432479858, |
|
"learning_rate": 9.978801094035207e-06, |
|
"logits/chosen": 0.1363590657711029, |
|
"logits/rejected": 0.1551840454339981, |
|
"logps/chosen": -71.82952117919922, |
|
"logps/rejected": -78.26856994628906, |
|
"loss": 2.6714, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.6073123216629028, |
|
"rewards/margins": 0.08929078280925751, |
|
"rewards/rejected": -0.6966031193733215, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.12534059945504086, |
|
"grad_norm": 1.2286291122436523, |
|
"learning_rate": 9.978173520482429e-06, |
|
"logits/chosen": 0.12522992491722107, |
|
"logits/rejected": 0.1818617582321167, |
|
"logps/chosen": -60.64283752441406, |
|
"logps/rejected": -67.53813934326172, |
|
"loss": 2.6884, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.5357421040534973, |
|
"rewards/margins": 0.0799265205860138, |
|
"rewards/rejected": -0.6156685948371887, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1271571298819255, |
|
"grad_norm": 1.0266982316970825, |
|
"learning_rate": 9.97753681515412e-06, |
|
"logits/chosen": 0.17851999402046204, |
|
"logits/rejected": 0.25888925790786743, |
|
"logps/chosen": -71.65196990966797, |
|
"logps/rejected": -82.34913635253906, |
|
"loss": 2.5226, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.7787651419639587, |
|
"rewards/margins": 0.1905478835105896, |
|
"rewards/rejected": -0.9693130254745483, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12897366030881016, |
|
"grad_norm": 0.999660313129425, |
|
"learning_rate": 9.976890979348628e-06, |
|
"logits/chosen": 0.18746232986450195, |
|
"logits/rejected": 0.2187044620513916, |
|
"logps/chosen": -69.84027099609375, |
|
"logps/rejected": -73.04399108886719, |
|
"loss": 2.6945, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.6969642639160156, |
|
"rewards/margins": 0.08156725764274597, |
|
"rewards/rejected": -0.778531551361084, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.1307901907356948, |
|
"grad_norm": 1.0596684217453003, |
|
"learning_rate": 9.976236014382934e-06, |
|
"logits/chosen": 0.17854923009872437, |
|
"logits/rejected": 0.18649883568286896, |
|
"logps/chosen": -63.992881774902344, |
|
"logps/rejected": -64.36544799804688, |
|
"loss": 2.7689, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.7436442375183105, |
|
"rewards/margins": 0.06337633728981018, |
|
"rewards/rejected": -0.8070206046104431, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.13260672116257946, |
|
"grad_norm": 1.0307612419128418, |
|
"learning_rate": 9.975571921592618e-06, |
|
"logits/chosen": 0.0964425653219223, |
|
"logits/rejected": 0.14382749795913696, |
|
"logps/chosen": -66.61788940429688, |
|
"logps/rejected": -70.80413055419922, |
|
"loss": 2.5859, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.6208910942077637, |
|
"rewards/margins": 0.17540723085403442, |
|
"rewards/rejected": -0.7962983250617981, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1344232515894641, |
|
"grad_norm": 0.9487422108650208, |
|
"learning_rate": 9.97489870233188e-06, |
|
"logits/chosen": 0.06293762475252151, |
|
"logits/rejected": 0.19688080251216888, |
|
"logps/chosen": -55.884613037109375, |
|
"logps/rejected": -64.30046081542969, |
|
"loss": 2.4062, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6415266990661621, |
|
"rewards/margins": 0.2741628885269165, |
|
"rewards/rejected": -0.9156895875930786, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.1362397820163488, |
|
"grad_norm": 1.223663330078125, |
|
"learning_rate": 9.974216357973539e-06, |
|
"logits/chosen": 0.10017126798629761, |
|
"logits/rejected": 0.10504551976919174, |
|
"logps/chosen": -83.83055114746094, |
|
"logps/rejected": -81.8952407836914, |
|
"loss": 2.8556, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8103373050689697, |
|
"rewards/margins": 0.027550537139177322, |
|
"rewards/rejected": -0.8378878831863403, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.13805631244323344, |
|
"grad_norm": 1.0121136903762817, |
|
"learning_rate": 9.973524889909007e-06, |
|
"logits/chosen": 0.07612155377864838, |
|
"logits/rejected": 0.15954618155956268, |
|
"logps/chosen": -65.6831283569336, |
|
"logps/rejected": -75.42146301269531, |
|
"loss": 2.5576, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6988077759742737, |
|
"rewards/margins": 0.1909356415271759, |
|
"rewards/rejected": -0.889743447303772, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.13987284287011809, |
|
"grad_norm": 1.0312719345092773, |
|
"learning_rate": 9.97282429954831e-06, |
|
"logits/chosen": 0.07015375047922134, |
|
"logits/rejected": 0.09143385291099548, |
|
"logps/chosen": -59.444358825683594, |
|
"logps/rejected": -61.26897430419922, |
|
"loss": 2.7845, |
|
"rewards/accuracies": 0.453125, |
|
"rewards/chosen": -0.7886172533035278, |
|
"rewards/margins": 0.038636498153209686, |
|
"rewards/rejected": -0.8272536993026733, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.14168937329700274, |
|
"grad_norm": 1.0625450611114502, |
|
"learning_rate": 9.972114588320073e-06, |
|
"logits/chosen": 0.07196499407291412, |
|
"logits/rejected": 0.12630172073841095, |
|
"logps/chosen": -75.5164566040039, |
|
"logps/rejected": -80.74308776855469, |
|
"loss": 2.5834, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.6821954250335693, |
|
"rewards/margins": 0.16858401894569397, |
|
"rewards/rejected": -0.8507794737815857, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.14350590372388738, |
|
"grad_norm": 0.9069045186042786, |
|
"learning_rate": 9.97139575767152e-06, |
|
"logits/chosen": 0.10926744341850281, |
|
"logits/rejected": 0.14418405294418335, |
|
"logps/chosen": -53.721046447753906, |
|
"logps/rejected": -58.17475891113281, |
|
"loss": 2.4955, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6156648993492126, |
|
"rewards/margins": 0.19551396369934082, |
|
"rewards/rejected": -0.8111788034439087, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.14532243415077203, |
|
"grad_norm": 1.1463656425476074, |
|
"learning_rate": 9.970667809068476e-06, |
|
"logits/chosen": 0.09196805953979492, |
|
"logits/rejected": 0.14393256604671478, |
|
"logps/chosen": -77.11197662353516, |
|
"logps/rejected": -87.39771270751953, |
|
"loss": 2.5763, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8568902611732483, |
|
"rewards/margins": 0.15524335205554962, |
|
"rewards/rejected": -1.0121335983276367, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.14713896457765668, |
|
"grad_norm": 1.0046662092208862, |
|
"learning_rate": 9.969930743995351e-06, |
|
"logits/chosen": 0.2063535749912262, |
|
"logits/rejected": 0.20602768659591675, |
|
"logps/chosen": -59.890933990478516, |
|
"logps/rejected": -59.117794036865234, |
|
"loss": 2.6829, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.7329096794128418, |
|
"rewards/margins": 0.0983649268746376, |
|
"rewards/rejected": -0.8312745094299316, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.14895549500454133, |
|
"grad_norm": 1.0280438661575317, |
|
"learning_rate": 9.969184563955152e-06, |
|
"logits/chosen": 0.17393875122070312, |
|
"logits/rejected": 0.16783203184604645, |
|
"logps/chosen": -72.82754516601562, |
|
"logps/rejected": -70.6113510131836, |
|
"loss": 2.6277, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.7506997585296631, |
|
"rewards/margins": 0.13392671942710876, |
|
"rewards/rejected": -0.8846263885498047, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.15077202543142598, |
|
"grad_norm": 0.9356668591499329, |
|
"learning_rate": 9.968429270469467e-06, |
|
"logits/chosen": 0.09279462695121765, |
|
"logits/rejected": 0.1681107133626938, |
|
"logps/chosen": -62.300445556640625, |
|
"logps/rejected": -68.70709991455078, |
|
"loss": 2.4821, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.7132639288902283, |
|
"rewards/margins": 0.2519262731075287, |
|
"rewards/rejected": -0.9651902318000793, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.15258855585831063, |
|
"grad_norm": 1.0442771911621094, |
|
"learning_rate": 9.967664865078472e-06, |
|
"logits/chosen": 0.17844007909297943, |
|
"logits/rejected": 0.21383100748062134, |
|
"logps/chosen": -61.43730163574219, |
|
"logps/rejected": -64.16680908203125, |
|
"loss": 2.6289, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.7574411034584045, |
|
"rewards/margins": 0.13140378892421722, |
|
"rewards/rejected": -0.888844907283783, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.15440508628519528, |
|
"grad_norm": 1.137267827987671, |
|
"learning_rate": 9.966891349340922e-06, |
|
"logits/chosen": 0.19703873991966248, |
|
"logits/rejected": 0.17027492821216583, |
|
"logps/chosen": -68.07546997070312, |
|
"logps/rejected": -63.9476203918457, |
|
"loss": 2.7702, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.8362983465194702, |
|
"rewards/margins": 0.05476854741573334, |
|
"rewards/rejected": -0.8910670280456543, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.15622161671207993, |
|
"grad_norm": 1.0395056009292603, |
|
"learning_rate": 9.966108724834151e-06, |
|
"logits/chosen": 0.10182631760835648, |
|
"logits/rejected": 0.11173731088638306, |
|
"logps/chosen": -64.47720336914062, |
|
"logps/rejected": -75.30056762695312, |
|
"loss": 2.5889, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8160837292671204, |
|
"rewards/margins": 0.2798304557800293, |
|
"rewards/rejected": -1.0959142446517944, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.15803814713896458, |
|
"grad_norm": 1.1256901025772095, |
|
"learning_rate": 9.965316993154069e-06, |
|
"logits/chosen": 0.1232331395149231, |
|
"logits/rejected": 0.19064000248908997, |
|
"logps/chosen": -64.3778305053711, |
|
"logps/rejected": -76.91246032714844, |
|
"loss": 2.5522, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8341800570487976, |
|
"rewards/margins": 0.2616519033908844, |
|
"rewards/rejected": -1.0958319902420044, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.15985467756584923, |
|
"grad_norm": 1.20338773727417, |
|
"learning_rate": 9.964516155915152e-06, |
|
"logits/chosen": 0.12199988961219788, |
|
"logits/rejected": 0.12264345586299896, |
|
"logps/chosen": -81.56141662597656, |
|
"logps/rejected": -80.49986267089844, |
|
"loss": 2.7432, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.7930806875228882, |
|
"rewards/margins": 0.06459490954875946, |
|
"rewards/rejected": -0.8576756119728088, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.16167120799273388, |
|
"grad_norm": 1.005759835243225, |
|
"learning_rate": 9.963706214750446e-06, |
|
"logits/chosen": 0.07040335237979889, |
|
"logits/rejected": 0.10741756111383438, |
|
"logps/chosen": -65.43824768066406, |
|
"logps/rejected": -74.47370147705078, |
|
"loss": 2.5472, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.8203200697898865, |
|
"rewards/margins": 0.24586233496665955, |
|
"rewards/rejected": -1.0661823749542236, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.16348773841961853, |
|
"grad_norm": 0.9761494994163513, |
|
"learning_rate": 9.962887171311563e-06, |
|
"logits/chosen": 0.1630188375711441, |
|
"logits/rejected": 0.190489262342453, |
|
"logps/chosen": -57.84222412109375, |
|
"logps/rejected": -64.116455078125, |
|
"loss": 2.5339, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8240211009979248, |
|
"rewards/margins": 0.21757929027080536, |
|
"rewards/rejected": -1.0416003465652466, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.16530426884650318, |
|
"grad_norm": 1.002192497253418, |
|
"learning_rate": 9.962059027268676e-06, |
|
"logits/chosen": 0.08669686317443848, |
|
"logits/rejected": 0.13606388866901398, |
|
"logps/chosen": -63.68013000488281, |
|
"logps/rejected": -72.30770111083984, |
|
"loss": 2.5165, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.7919780611991882, |
|
"rewards/margins": 0.27081233263015747, |
|
"rewards/rejected": -1.0627902746200562, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.16712079927338783, |
|
"grad_norm": 0.9397479295730591, |
|
"learning_rate": 9.961221784310514e-06, |
|
"logits/chosen": 0.14396092295646667, |
|
"logits/rejected": 0.1820589005947113, |
|
"logps/chosen": -57.551353454589844, |
|
"logps/rejected": -66.83124542236328, |
|
"loss": 2.4254, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.7751535177230835, |
|
"rewards/margins": 0.26002568006515503, |
|
"rewards/rejected": -1.0351792573928833, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.16893732970027248, |
|
"grad_norm": 1.2053430080413818, |
|
"learning_rate": 9.96037544414436e-06, |
|
"logits/chosen": 0.16130733489990234, |
|
"logits/rejected": 0.1852559745311737, |
|
"logps/chosen": -73.8154067993164, |
|
"logps/rejected": -81.3624267578125, |
|
"loss": 2.7543, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.8622196316719055, |
|
"rewards/margins": 0.11067891120910645, |
|
"rewards/rejected": -0.9728984236717224, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.17075386012715713, |
|
"grad_norm": 1.1681262254714966, |
|
"learning_rate": 9.959520008496054e-06, |
|
"logits/chosen": 0.17866984009742737, |
|
"logits/rejected": 0.20266617834568024, |
|
"logps/chosen": -61.40083312988281, |
|
"logps/rejected": -65.13737487792969, |
|
"loss": 2.7134, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9223482012748718, |
|
"rewards/margins": 0.08481010049581528, |
|
"rewards/rejected": -1.0071581602096558, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.17257039055404177, |
|
"grad_norm": 1.0269131660461426, |
|
"learning_rate": 9.95865547910997e-06, |
|
"logits/chosen": 0.11446906626224518, |
|
"logits/rejected": 0.17379909753799438, |
|
"logps/chosen": -62.823944091796875, |
|
"logps/rejected": -69.5682601928711, |
|
"loss": 2.5031, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8894773125648499, |
|
"rewards/margins": 0.24369436502456665, |
|
"rewards/rejected": -1.1331716775894165, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.17438692098092642, |
|
"grad_norm": 1.0400865077972412, |
|
"learning_rate": 9.957781857749043e-06, |
|
"logits/chosen": 0.16556067764759064, |
|
"logits/rejected": 0.23536017537117004, |
|
"logps/chosen": -61.09260940551758, |
|
"logps/rejected": -63.444427490234375, |
|
"loss": 2.7239, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.0037989616394043, |
|
"rewards/margins": 0.07835513353347778, |
|
"rewards/rejected": -1.0821542739868164, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.17620345140781107, |
|
"grad_norm": 1.1138463020324707, |
|
"learning_rate": 9.956899146194732e-06, |
|
"logits/chosen": 0.13205701112747192, |
|
"logits/rejected": 0.1702471375465393, |
|
"logps/chosen": -63.70584487915039, |
|
"logps/rejected": -69.10890197753906, |
|
"loss": 2.7076, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9300363659858704, |
|
"rewards/margins": 0.17512086033821106, |
|
"rewards/rejected": -1.1051572561264038, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.17801998183469572, |
|
"grad_norm": 0.9917119741439819, |
|
"learning_rate": 9.956007346247043e-06, |
|
"logits/chosen": 0.12265195697546005, |
|
"logits/rejected": 0.18780440092086792, |
|
"logps/chosen": -62.316497802734375, |
|
"logps/rejected": -69.26319885253906, |
|
"loss": 2.4996, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8957506418228149, |
|
"rewards/margins": 0.2764662802219391, |
|
"rewards/rejected": -1.1722170114517212, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.17983651226158037, |
|
"grad_norm": 1.00100576877594, |
|
"learning_rate": 9.95510645972451e-06, |
|
"logits/chosen": 0.16969357430934906, |
|
"logits/rejected": 0.16122889518737793, |
|
"logps/chosen": -60.42213439941406, |
|
"logps/rejected": -63.85418701171875, |
|
"loss": 2.5677, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8840566873550415, |
|
"rewards/margins": 0.1611585170030594, |
|
"rewards/rejected": -1.0452152490615845, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.18165304268846502, |
|
"grad_norm": 1.1503106355667114, |
|
"learning_rate": 9.954196488464198e-06, |
|
"logits/chosen": 0.19831174612045288, |
|
"logits/rejected": 0.20889577269554138, |
|
"logps/chosen": -65.17144775390625, |
|
"logps/rejected": -69.9671859741211, |
|
"loss": 2.7065, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.9322465062141418, |
|
"rewards/margins": 0.1645345687866211, |
|
"rewards/rejected": -1.0967810153961182, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.18346957311534967, |
|
"grad_norm": 1.416610836982727, |
|
"learning_rate": 9.953277434321696e-06, |
|
"logits/chosen": 0.08757522702217102, |
|
"logits/rejected": 0.1615956872701645, |
|
"logps/chosen": -72.49248504638672, |
|
"logps/rejected": -80.52176666259766, |
|
"loss": 2.5346, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8279029130935669, |
|
"rewards/margins": 0.26418614387512207, |
|
"rewards/rejected": -1.092089056968689, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.18528610354223432, |
|
"grad_norm": 0.9907792210578918, |
|
"learning_rate": 9.952349299171117e-06, |
|
"logits/chosen": 0.07832024991512299, |
|
"logits/rejected": 0.16370174288749695, |
|
"logps/chosen": -67.5047836303711, |
|
"logps/rejected": -73.78684997558594, |
|
"loss": 2.4325, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8129651546478271, |
|
"rewards/margins": 0.3001910448074341, |
|
"rewards/rejected": -1.1131561994552612, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.18710263396911897, |
|
"grad_norm": 1.1051058769226074, |
|
"learning_rate": 9.95141208490509e-06, |
|
"logits/chosen": 0.10540622472763062, |
|
"logits/rejected": 0.2087182104587555, |
|
"logps/chosen": -55.68672180175781, |
|
"logps/rejected": -72.7021484375, |
|
"loss": 2.3004, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8526840209960938, |
|
"rewards/margins": 0.3691489100456238, |
|
"rewards/rejected": -1.2218331098556519, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.18891916439600362, |
|
"grad_norm": 1.0593968629837036, |
|
"learning_rate": 9.950465793434759e-06, |
|
"logits/chosen": 0.10488969832658768, |
|
"logits/rejected": 0.13292263448238373, |
|
"logps/chosen": -69.47488403320312, |
|
"logps/rejected": -73.08143615722656, |
|
"loss": 2.5388, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.8357015252113342, |
|
"rewards/margins": 0.2204672396183014, |
|
"rewards/rejected": -1.0561686754226685, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.1907356948228883, |
|
"grad_norm": 1.119372010231018, |
|
"learning_rate": 9.949510426689773e-06, |
|
"logits/chosen": 0.06970994919538498, |
|
"logits/rejected": 0.0831708014011383, |
|
"logps/chosen": -66.97635650634766, |
|
"logps/rejected": -67.95066833496094, |
|
"loss": 2.6443, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.9023284912109375, |
|
"rewards/margins": 0.1277586966753006, |
|
"rewards/rejected": -1.0300871133804321, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.19255222524977295, |
|
"grad_norm": 1.2842986583709717, |
|
"learning_rate": 9.948545986618295e-06, |
|
"logits/chosen": 0.17146825790405273, |
|
"logits/rejected": 0.1910206824541092, |
|
"logps/chosen": -69.1142578125, |
|
"logps/rejected": -70.71878814697266, |
|
"loss": 2.8494, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.0275481939315796, |
|
"rewards/margins": 0.09755454212427139, |
|
"rewards/rejected": -1.1251027584075928, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.1943687556766576, |
|
"grad_norm": 1.044800877571106, |
|
"learning_rate": 9.947572475186984e-06, |
|
"logits/chosen": 0.1206701397895813, |
|
"logits/rejected": 0.17635390162467957, |
|
"logps/chosen": -67.00174713134766, |
|
"logps/rejected": -71.5347900390625, |
|
"loss": 2.46, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.889815628528595, |
|
"rewards/margins": 0.255011647939682, |
|
"rewards/rejected": -1.1448272466659546, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.19618528610354224, |
|
"grad_norm": 1.1496058702468872, |
|
"learning_rate": 9.946589894381002e-06, |
|
"logits/chosen": 0.18972846865653992, |
|
"logits/rejected": 0.12482471019029617, |
|
"logps/chosen": -66.56938934326172, |
|
"logps/rejected": -73.7935791015625, |
|
"loss": 2.6472, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.8382111191749573, |
|
"rewards/margins": 0.16597144305706024, |
|
"rewards/rejected": -1.0041825771331787, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.1980018165304269, |
|
"grad_norm": 1.1357221603393555, |
|
"learning_rate": 9.945598246204e-06, |
|
"logits/chosen": 0.1766098588705063, |
|
"logits/rejected": 0.16134579479694366, |
|
"logps/chosen": -72.30979919433594, |
|
"logps/rejected": -69.26260375976562, |
|
"loss": 2.7037, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.9316724538803101, |
|
"rewards/margins": 0.0998368114233017, |
|
"rewards/rejected": -1.0315091609954834, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.19981834695731154, |
|
"grad_norm": 1.1147487163543701, |
|
"learning_rate": 9.94459753267812e-06, |
|
"logits/chosen": 0.10100046545267105, |
|
"logits/rejected": 0.12630510330200195, |
|
"logps/chosen": -66.10484313964844, |
|
"logps/rejected": -68.13362121582031, |
|
"loss": 2.6224, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8918489217758179, |
|
"rewards/margins": 0.165731742978096, |
|
"rewards/rejected": -1.0575807094573975, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2016348773841962, |
|
"grad_norm": 1.3498740196228027, |
|
"learning_rate": 9.943587755843996e-06, |
|
"logits/chosen": 0.19711939990520477, |
|
"logits/rejected": 0.18357205390930176, |
|
"logps/chosen": -64.5146484375, |
|
"logps/rejected": -65.80925750732422, |
|
"loss": 2.6102, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.9744794368743896, |
|
"rewards/margins": 0.16535742580890656, |
|
"rewards/rejected": -1.1398367881774902, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.20345140781108084, |
|
"grad_norm": 1.0210968255996704, |
|
"learning_rate": 9.942568917760733e-06, |
|
"logits/chosen": 0.16077302396297455, |
|
"logits/rejected": 0.22646722197532654, |
|
"logps/chosen": -59.73281478881836, |
|
"logps/rejected": -64.54032897949219, |
|
"loss": 2.4897, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8548685312271118, |
|
"rewards/margins": 0.19910478591918945, |
|
"rewards/rejected": -1.0539731979370117, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.2052679382379655, |
|
"grad_norm": 1.160152554512024, |
|
"learning_rate": 9.941541020505924e-06, |
|
"logits/chosen": 0.10292509198188782, |
|
"logits/rejected": 0.09458380937576294, |
|
"logps/chosen": -77.44420623779297, |
|
"logps/rejected": -85.09677124023438, |
|
"loss": 2.5037, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.9961792230606079, |
|
"rewards/margins": 0.2911871373653412, |
|
"rewards/rejected": -1.287366509437561, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.20708446866485014, |
|
"grad_norm": 1.1742796897888184, |
|
"learning_rate": 9.940504066175626e-06, |
|
"logits/chosen": 0.10538715869188309, |
|
"logits/rejected": 0.07242526113986969, |
|
"logps/chosen": -70.37040710449219, |
|
"logps/rejected": -68.23387908935547, |
|
"loss": 2.6328, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.9170363545417786, |
|
"rewards/margins": 0.14501458406448364, |
|
"rewards/rejected": -1.0620509386062622, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.2089009990917348, |
|
"grad_norm": 1.1259719133377075, |
|
"learning_rate": 9.939458056884375e-06, |
|
"logits/chosen": 0.12264375388622284, |
|
"logits/rejected": 0.16776585578918457, |
|
"logps/chosen": -59.847049713134766, |
|
"logps/rejected": -65.45285034179688, |
|
"loss": 2.5752, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.8872510194778442, |
|
"rewards/margins": 0.2041575014591217, |
|
"rewards/rejected": -1.0914084911346436, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.21071752951861944, |
|
"grad_norm": 2.1809544563293457, |
|
"learning_rate": 9.938402994765163e-06, |
|
"logits/chosen": 0.13642175495624542, |
|
"logits/rejected": 0.11773751676082611, |
|
"logps/chosen": -65.34564971923828, |
|
"logps/rejected": -66.55570983886719, |
|
"loss": 2.6792, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9432386755943298, |
|
"rewards/margins": 0.12600603699684143, |
|
"rewards/rejected": -1.0692447423934937, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.2125340599455041, |
|
"grad_norm": 1.4895256757736206, |
|
"learning_rate": 9.937338881969444e-06, |
|
"logits/chosen": 0.12349803745746613, |
|
"logits/rejected": 0.1182522252202034, |
|
"logps/chosen": -66.9234619140625, |
|
"logps/rejected": -69.5535888671875, |
|
"loss": 2.6029, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.9096715450286865, |
|
"rewards/margins": 0.23732726275920868, |
|
"rewards/rejected": -1.1469988822937012, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.21435059037238874, |
|
"grad_norm": 1.0333324670791626, |
|
"learning_rate": 9.93626572066713e-06, |
|
"logits/chosen": 0.16722331941127777, |
|
"logits/rejected": 0.21501797437667847, |
|
"logps/chosen": -66.49773406982422, |
|
"logps/rejected": -72.3947982788086, |
|
"loss": 2.4687, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.9357748627662659, |
|
"rewards/margins": 0.2686734199523926, |
|
"rewards/rejected": -1.2044482231140137, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2161671207992734, |
|
"grad_norm": 1.2127379179000854, |
|
"learning_rate": 9.935183513046585e-06, |
|
"logits/chosen": 0.10065922141075134, |
|
"logits/rejected": 0.12764661014080048, |
|
"logps/chosen": -73.05583953857422, |
|
"logps/rejected": -77.18297576904297, |
|
"loss": 2.6288, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8347401022911072, |
|
"rewards/margins": 0.2145780473947525, |
|
"rewards/rejected": -1.0493181943893433, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.21798365122615804, |
|
"grad_norm": 1.4050469398498535, |
|
"learning_rate": 9.93409226131462e-06, |
|
"logits/chosen": 0.08468589186668396, |
|
"logits/rejected": 0.179460808634758, |
|
"logps/chosen": -66.12811279296875, |
|
"logps/rejected": -74.2131118774414, |
|
"loss": 2.4658, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8611673712730408, |
|
"rewards/margins": 0.20363186299800873, |
|
"rewards/rejected": -1.064799189567566, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.2198001816530427, |
|
"grad_norm": 1.18356192111969, |
|
"learning_rate": 9.932991967696484e-06, |
|
"logits/chosen": 0.18308167159557343, |
|
"logits/rejected": 0.12839025259017944, |
|
"logps/chosen": -63.325035095214844, |
|
"logps/rejected": -63.72959518432617, |
|
"loss": 2.579, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9539028406143188, |
|
"rewards/margins": 0.16736355423927307, |
|
"rewards/rejected": -1.1212663650512695, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.22161671207992734, |
|
"grad_norm": 1.1829110383987427, |
|
"learning_rate": 9.931882634435868e-06, |
|
"logits/chosen": 0.10780126601457596, |
|
"logits/rejected": 0.109224833548069, |
|
"logps/chosen": -71.1589126586914, |
|
"logps/rejected": -74.609375, |
|
"loss": 2.6368, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.9908429384231567, |
|
"rewards/margins": 0.12795159220695496, |
|
"rewards/rejected": -1.118794560432434, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.22343324250681199, |
|
"grad_norm": 1.1116541624069214, |
|
"learning_rate": 9.930764263794898e-06, |
|
"logits/chosen": 0.10946042090654373, |
|
"logits/rejected": 0.11234283447265625, |
|
"logps/chosen": -67.98091888427734, |
|
"logps/rejected": -70.58726501464844, |
|
"loss": 2.4417, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.871981143951416, |
|
"rewards/margins": 0.22934575378894806, |
|
"rewards/rejected": -1.1013269424438477, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.22524977293369663, |
|
"grad_norm": 1.1564345359802246, |
|
"learning_rate": 9.929636858054124e-06, |
|
"logits/chosen": 0.11665979772806168, |
|
"logits/rejected": 0.15129488706588745, |
|
"logps/chosen": -62.33778381347656, |
|
"logps/rejected": -67.58541870117188, |
|
"loss": 2.6023, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8282152414321899, |
|
"rewards/margins": 0.206033393740654, |
|
"rewards/rejected": -1.0342485904693604, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.22706630336058128, |
|
"grad_norm": 1.149032711982727, |
|
"learning_rate": 9.928500419512522e-06, |
|
"logits/chosen": 0.02143971435725689, |
|
"logits/rejected": 0.10582254827022552, |
|
"logps/chosen": -66.31533813476562, |
|
"logps/rejected": -76.4206771850586, |
|
"loss": 2.4589, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.8935360908508301, |
|
"rewards/margins": 0.3176972270011902, |
|
"rewards/rejected": -1.2112332582473755, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.22888283378746593, |
|
"grad_norm": 1.1076140403747559, |
|
"learning_rate": 9.927354950487491e-06, |
|
"logits/chosen": 0.13772962987422943, |
|
"logits/rejected": 0.12653110921382904, |
|
"logps/chosen": -62.97186279296875, |
|
"logps/rejected": -65.31454467773438, |
|
"loss": 2.4839, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9201483130455017, |
|
"rewards/margins": 0.2287607342004776, |
|
"rewards/rejected": -1.1489089727401733, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.23069936421435058, |
|
"grad_norm": 1.1476696729660034, |
|
"learning_rate": 9.926200453314842e-06, |
|
"logits/chosen": 0.11984287202358246, |
|
"logits/rejected": 0.1570628136396408, |
|
"logps/chosen": -72.74269104003906, |
|
"logps/rejected": -77.52953338623047, |
|
"loss": 2.5198, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9817119836807251, |
|
"rewards/margins": 0.2599087059497833, |
|
"rewards/rejected": -1.241620659828186, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.23251589464123523, |
|
"grad_norm": 1.0800604820251465, |
|
"learning_rate": 9.925036930348793e-06, |
|
"logits/chosen": 0.03499937057495117, |
|
"logits/rejected": 0.16788874566555023, |
|
"logps/chosen": -64.4096450805664, |
|
"logps/rejected": -81.16387939453125, |
|
"loss": 2.3547, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.040608286857605, |
|
"rewards/margins": 0.48978835344314575, |
|
"rewards/rejected": -1.530396580696106, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.23433242506811988, |
|
"grad_norm": 1.2677640914916992, |
|
"learning_rate": 9.92386438396197e-06, |
|
"logits/chosen": 0.1200772225856781, |
|
"logits/rejected": 0.11345633864402771, |
|
"logps/chosen": -68.35733032226562, |
|
"logps/rejected": -71.72528839111328, |
|
"loss": 2.631, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.9437850117683411, |
|
"rewards/margins": 0.13758014142513275, |
|
"rewards/rejected": -1.0813652276992798, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.23614895549500453, |
|
"grad_norm": 1.2041019201278687, |
|
"learning_rate": 9.9226828165454e-06, |
|
"logits/chosen": 0.07035915553569794, |
|
"logits/rejected": 0.16790318489074707, |
|
"logps/chosen": -73.99917602539062, |
|
"logps/rejected": -83.07939147949219, |
|
"loss": 2.4497, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0043660402297974, |
|
"rewards/margins": 0.36712244153022766, |
|
"rewards/rejected": -1.3714885711669922, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.23796548592188918, |
|
"grad_norm": 1.112176537513733, |
|
"learning_rate": 9.921492230508507e-06, |
|
"logits/chosen": 0.18427804112434387, |
|
"logits/rejected": 0.17424902319908142, |
|
"logps/chosen": -63.21034240722656, |
|
"logps/rejected": -65.23414611816406, |
|
"loss": 2.4612, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.9832345843315125, |
|
"rewards/margins": 0.21530815958976746, |
|
"rewards/rejected": -1.1985427141189575, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.23978201634877383, |
|
"grad_norm": 1.1540772914886475, |
|
"learning_rate": 9.9202926282791e-06, |
|
"logits/chosen": 0.14970409870147705, |
|
"logits/rejected": 0.17991000413894653, |
|
"logps/chosen": -76.66849517822266, |
|
"logps/rejected": -85.59910583496094, |
|
"loss": 2.4164, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0053367614746094, |
|
"rewards/margins": 0.3513883352279663, |
|
"rewards/rejected": -1.3567250967025757, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.24159854677565848, |
|
"grad_norm": 1.32212495803833, |
|
"learning_rate": 9.919084012303378e-06, |
|
"logits/chosen": 0.11897419393062592, |
|
"logits/rejected": 0.15737299621105194, |
|
"logps/chosen": -75.7364501953125, |
|
"logps/rejected": -77.97785186767578, |
|
"loss": 2.7243, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.139754295349121, |
|
"rewards/margins": 0.17446394264698029, |
|
"rewards/rejected": -1.3142181634902954, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.24341507720254316, |
|
"grad_norm": 1.4282020330429077, |
|
"learning_rate": 9.917866385045918e-06, |
|
"logits/chosen": 0.13694903254508972, |
|
"logits/rejected": 0.2048657089471817, |
|
"logps/chosen": -70.1098403930664, |
|
"logps/rejected": -78.7418212890625, |
|
"loss": 2.8086, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.2937737703323364, |
|
"rewards/margins": 0.2787961959838867, |
|
"rewards/rejected": -1.5725698471069336, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2452316076294278, |
|
"grad_norm": 1.1548171043395996, |
|
"learning_rate": 9.916639748989677e-06, |
|
"logits/chosen": 0.050287194550037384, |
|
"logits/rejected": 0.10146909952163696, |
|
"logps/chosen": -72.51569366455078, |
|
"logps/rejected": -80.42554473876953, |
|
"loss": 2.4774, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0650720596313477, |
|
"rewards/margins": 0.2093387395143509, |
|
"rewards/rejected": -1.2744107246398926, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.24704813805631246, |
|
"grad_norm": 1.1347655057907104, |
|
"learning_rate": 9.915404106635979e-06, |
|
"logits/chosen": 0.13506914675235748, |
|
"logits/rejected": 0.18150334060192108, |
|
"logps/chosen": -71.03006744384766, |
|
"logps/rejected": -75.59716033935547, |
|
"loss": 2.4399, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1068118810653687, |
|
"rewards/margins": 0.3031606674194336, |
|
"rewards/rejected": -1.4099724292755127, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2488646684831971, |
|
"grad_norm": 1.1634560823440552, |
|
"learning_rate": 9.914159460504512e-06, |
|
"logits/chosen": 0.17681393027305603, |
|
"logits/rejected": 0.2249586582183838, |
|
"logps/chosen": -66.17890167236328, |
|
"logps/rejected": -71.03226470947266, |
|
"loss": 2.5281, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0759899616241455, |
|
"rewards/margins": 0.21175377070903778, |
|
"rewards/rejected": -1.2877436876296997, |
|
"step": 137 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 2200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 137, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 6.440843988697088e+16, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|