|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.4977293369663942, |
|
"eval_steps": 500, |
|
"global_step": 274, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0018165304268846503, |
|
"grad_norm": 0.7654335498809814, |
|
"learning_rate": 9.99999541185824e-06, |
|
"logits/chosen": 0.10746052116155624, |
|
"logits/rejected": 0.16293610632419586, |
|
"logps/chosen": -51.70099639892578, |
|
"logps/rejected": -56.85424041748047, |
|
"loss": 2.7726, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0036330608537693005, |
|
"grad_norm": 0.8493374586105347, |
|
"learning_rate": 9.999981647442309e-06, |
|
"logits/chosen": 0.12589210271835327, |
|
"logits/rejected": 0.18243370950222015, |
|
"logps/chosen": -61.092987060546875, |
|
"logps/rejected": -67.26304626464844, |
|
"loss": 2.7516, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": 0.005276113282889128, |
|
"rewards/margins": 0.010965153574943542, |
|
"rewards/rejected": -0.00568903936073184, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.005449591280653951, |
|
"grad_norm": 0.739442765712738, |
|
"learning_rate": 9.999958706780277e-06, |
|
"logits/chosen": 0.0766456350684166, |
|
"logits/rejected": 0.1422545611858368, |
|
"logps/chosen": -55.764503479003906, |
|
"logps/rejected": -57.38707733154297, |
|
"loss": 2.7902, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.0027439936529845, |
|
"rewards/margins": -0.008403200656175613, |
|
"rewards/rejected": 0.005659207701683044, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.007266121707538601, |
|
"grad_norm": 0.7261953949928284, |
|
"learning_rate": 9.999926589918927e-06, |
|
"logits/chosen": 0.1812177300453186, |
|
"logits/rejected": 0.13825736939907074, |
|
"logps/chosen": -56.98471450805664, |
|
"logps/rejected": -55.47361373901367, |
|
"loss": 2.7537, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": 0.006436044350266457, |
|
"rewards/margins": 0.009899044409394264, |
|
"rewards/rejected": -0.003463000524789095, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.009082652134423252, |
|
"grad_norm": 0.850686252117157, |
|
"learning_rate": 9.999885296923748e-06, |
|
"logits/chosen": 0.17714114487171173, |
|
"logits/rejected": 0.12910494208335876, |
|
"logps/chosen": -71.72655487060547, |
|
"logps/rejected": -70.75139617919922, |
|
"loss": 2.79, |
|
"rewards/accuracies": 0.453125, |
|
"rewards/chosen": -0.008413795381784439, |
|
"rewards/margins": -0.007964953780174255, |
|
"rewards/rejected": -0.0004488405538722873, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.010899182561307902, |
|
"grad_norm": 0.6627741456031799, |
|
"learning_rate": 9.999834827878942e-06, |
|
"logits/chosen": 0.13277305662631989, |
|
"logits/rejected": 0.20749863982200623, |
|
"logps/chosen": -47.56520080566406, |
|
"logps/rejected": -53.18217468261719, |
|
"loss": 2.7732, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.001662571681663394, |
|
"rewards/margins": 0.00014669005759060383, |
|
"rewards/rejected": 0.0015158820897340775, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.012715712988192553, |
|
"grad_norm": 0.8335347175598145, |
|
"learning_rate": 9.999775182887429e-06, |
|
"logits/chosen": 0.07216031849384308, |
|
"logits/rejected": 0.13223227858543396, |
|
"logps/chosen": -73.06997680664062, |
|
"logps/rejected": -75.58594512939453, |
|
"loss": 2.7716, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.004659444559365511, |
|
"rewards/margins": 0.0013355333358049393, |
|
"rewards/rejected": -0.005994977429509163, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.014532243415077202, |
|
"grad_norm": 0.7128610014915466, |
|
"learning_rate": 9.999706362070833e-06, |
|
"logits/chosen": 0.11368857324123383, |
|
"logits/rejected": 0.15805509686470032, |
|
"logps/chosen": -56.45945358276367, |
|
"logps/rejected": -56.9697151184082, |
|
"loss": 2.7544, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.0038012717850506306, |
|
"rewards/margins": 0.009716177359223366, |
|
"rewards/rejected": -0.013517449609935284, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01634877384196185, |
|
"grad_norm": 0.77040034532547, |
|
"learning_rate": 9.99962836556949e-06, |
|
"logits/chosen": 0.1818978637456894, |
|
"logits/rejected": 0.20750640332698822, |
|
"logps/chosen": -59.184532165527344, |
|
"logps/rejected": -61.59263610839844, |
|
"loss": 2.7688, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.009461460635066032, |
|
"rewards/margins": 0.002394758863374591, |
|
"rewards/rejected": -0.011856218799948692, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.018165304268846504, |
|
"grad_norm": 0.7393640875816345, |
|
"learning_rate": 9.99954119354245e-06, |
|
"logits/chosen": 0.14076584577560425, |
|
"logits/rejected": 0.22711633145809174, |
|
"logps/chosen": -56.34330749511719, |
|
"logps/rejected": -60.51960372924805, |
|
"loss": 2.7827, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.004167273174971342, |
|
"rewards/margins": -0.00429012393578887, |
|
"rewards/rejected": 0.00012284982949495316, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.019981834695731154, |
|
"grad_norm": 0.7271163463592529, |
|
"learning_rate": 9.999444846167473e-06, |
|
"logits/chosen": 0.20141208171844482, |
|
"logits/rejected": 0.2038215547800064, |
|
"logps/chosen": -57.712646484375, |
|
"logps/rejected": -54.343231201171875, |
|
"loss": 2.7565, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.0009536671568639576, |
|
"rewards/margins": 0.00866839662194252, |
|
"rewards/rejected": -0.007714730221778154, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.021798365122615803, |
|
"grad_norm": 0.9451335668563843, |
|
"learning_rate": 9.999339323641027e-06, |
|
"logits/chosen": 0.16046015918254852, |
|
"logits/rejected": 0.1963367462158203, |
|
"logps/chosen": -51.579856872558594, |
|
"logps/rejected": -54.97206115722656, |
|
"loss": 2.7732, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.005096627864986658, |
|
"rewards/margins": 0.00017735245637595654, |
|
"rewards/rejected": -0.005273980088531971, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.023614895549500452, |
|
"grad_norm": 0.8279874324798584, |
|
"learning_rate": 9.99922462617829e-06, |
|
"logits/chosen": 0.12593263387680054, |
|
"logits/rejected": 0.11708654463291168, |
|
"logps/chosen": -62.25929641723633, |
|
"logps/rejected": -60.633296966552734, |
|
"loss": 2.7693, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.013154934160411358, |
|
"rewards/margins": 0.002298446139320731, |
|
"rewards/rejected": -0.015453380532562733, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.025431425976385105, |
|
"grad_norm": 0.7863137125968933, |
|
"learning_rate": 9.999100754013152e-06, |
|
"logits/chosen": 0.12714676558971405, |
|
"logits/rejected": 0.18583568930625916, |
|
"logps/chosen": -59.119529724121094, |
|
"logps/rejected": -65.85352325439453, |
|
"loss": 2.7715, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.01445689145475626, |
|
"rewards/margins": 0.001161783467978239, |
|
"rewards/rejected": -0.015618674457073212, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.027247956403269755, |
|
"grad_norm": 0.7623695731163025, |
|
"learning_rate": 9.998967707398207e-06, |
|
"logits/chosen": 0.09560943394899368, |
|
"logits/rejected": 0.14794519543647766, |
|
"logps/chosen": -47.649497985839844, |
|
"logps/rejected": -57.21004867553711, |
|
"loss": 2.7405, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.0010593307670205832, |
|
"rewards/margins": 0.0166107639670372, |
|
"rewards/rejected": -0.017670094966888428, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.029064486830154404, |
|
"grad_norm": 0.7020726203918457, |
|
"learning_rate": 9.998825486604765e-06, |
|
"logits/chosen": 0.19466270506381989, |
|
"logits/rejected": 0.21006342768669128, |
|
"logps/chosen": -50.8302116394043, |
|
"logps/rejected": -58.202754974365234, |
|
"loss": 2.7628, |
|
"rewards/accuracies": 0.484375, |
|
"rewards/chosen": -0.009497416205704212, |
|
"rewards/margins": 0.005540419369935989, |
|
"rewards/rejected": -0.015037836506962776, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.030881017257039057, |
|
"grad_norm": 0.7085703611373901, |
|
"learning_rate": 9.998674091922833e-06, |
|
"logits/chosen": 0.06517614424228668, |
|
"logits/rejected": 0.15166127681732178, |
|
"logps/chosen": -45.203216552734375, |
|
"logps/rejected": -51.454952239990234, |
|
"loss": 2.7568, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.009879402816295624, |
|
"rewards/margins": 0.00827767699956894, |
|
"rewards/rejected": -0.018157079815864563, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0326975476839237, |
|
"grad_norm": 0.8115389347076416, |
|
"learning_rate": 9.998513523661136e-06, |
|
"logits/chosen": 0.09045977890491486, |
|
"logits/rejected": 0.1498938351869583, |
|
"logps/chosen": -60.35250473022461, |
|
"logps/rejected": -62.11174774169922, |
|
"loss": 2.7888, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.01872912421822548, |
|
"rewards/margins": -0.0073302448727190495, |
|
"rewards/rejected": -0.011398878879845142, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03451407811080836, |
|
"grad_norm": 0.8222002983093262, |
|
"learning_rate": 9.998343782147098e-06, |
|
"logits/chosen": 0.08091418445110321, |
|
"logits/rejected": 0.1225643903017044, |
|
"logps/chosen": -54.25471115112305, |
|
"logps/rejected": -57.91140365600586, |
|
"loss": 2.7652, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.018515082076191902, |
|
"rewards/margins": 0.004185312893241644, |
|
"rewards/rejected": -0.022700395435094833, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.03633060853769301, |
|
"grad_norm": 0.862041711807251, |
|
"learning_rate": 9.998164867726853e-06, |
|
"logits/chosen": 0.04738985374569893, |
|
"logits/rejected": 0.11436720937490463, |
|
"logps/chosen": -62.22935104370117, |
|
"logps/rejected": -67.67117309570312, |
|
"loss": 2.7607, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.01734255626797676, |
|
"rewards/margins": 0.006629224866628647, |
|
"rewards/rejected": -0.023971781134605408, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03814713896457766, |
|
"grad_norm": 0.8169613480567932, |
|
"learning_rate": 9.997976780765237e-06, |
|
"logits/chosen": 0.14641642570495605, |
|
"logits/rejected": 0.17384907603263855, |
|
"logps/chosen": -61.21760559082031, |
|
"logps/rejected": -61.094486236572266, |
|
"loss": 2.7752, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.021789425984025, |
|
"rewards/margins": -0.000787546974606812, |
|
"rewards/rejected": -0.02100187912583351, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.03996366939146231, |
|
"grad_norm": 0.7457932829856873, |
|
"learning_rate": 9.997779521645793e-06, |
|
"logits/chosen": 0.16773203015327454, |
|
"logits/rejected": 0.16435235738754272, |
|
"logps/chosen": -55.192623138427734, |
|
"logps/rejected": -53.05280303955078, |
|
"loss": 2.7661, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.021701142191886902, |
|
"rewards/margins": 0.0036970973014831543, |
|
"rewards/rejected": -0.025398239493370056, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04178019981834696, |
|
"grad_norm": 0.8263576030731201, |
|
"learning_rate": 9.997573090770766e-06, |
|
"logits/chosen": 0.10863066464662552, |
|
"logits/rejected": 0.1277829110622406, |
|
"logps/chosen": -61.80084228515625, |
|
"logps/rejected": -64.94878387451172, |
|
"loss": 2.7526, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.019380860030651093, |
|
"rewards/margins": 0.010501865297555923, |
|
"rewards/rejected": -0.029882723465561867, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.043596730245231606, |
|
"grad_norm": 0.7587623000144958, |
|
"learning_rate": 9.997357488561107e-06, |
|
"logits/chosen": 0.17992480099201202, |
|
"logits/rejected": 0.16478824615478516, |
|
"logps/chosen": -62.946449279785156, |
|
"logps/rejected": -61.98965072631836, |
|
"loss": 2.7534, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.02564322203397751, |
|
"rewards/margins": 0.010079940780997276, |
|
"rewards/rejected": -0.035723160952329636, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.045413260672116255, |
|
"grad_norm": 0.7733523845672607, |
|
"learning_rate": 9.997132715456464e-06, |
|
"logits/chosen": 0.1268949955701828, |
|
"logits/rejected": 0.14968058466911316, |
|
"logps/chosen": -53.57306671142578, |
|
"logps/rejected": -58.07795715332031, |
|
"loss": 2.7602, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.030423898249864578, |
|
"rewards/margins": 0.006796791218221188, |
|
"rewards/rejected": -0.03722068667411804, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.047229791099000905, |
|
"grad_norm": 0.8851524591445923, |
|
"learning_rate": 9.99689877191519e-06, |
|
"logits/chosen": 0.17443379759788513, |
|
"logits/rejected": 0.1787741333246231, |
|
"logps/chosen": -58.031558990478516, |
|
"logps/rejected": -59.206520080566406, |
|
"loss": 2.762, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.02750026062130928, |
|
"rewards/margins": 0.006129855290055275, |
|
"rewards/rejected": -0.033630117774009705, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.04904632152588556, |
|
"grad_norm": 0.8524475693702698, |
|
"learning_rate": 9.996655658414331e-06, |
|
"logits/chosen": 0.18196682631969452, |
|
"logits/rejected": 0.237998366355896, |
|
"logps/chosen": -63.18816375732422, |
|
"logps/rejected": -67.31481170654297, |
|
"loss": 2.7519, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.03693533316254616, |
|
"rewards/margins": 0.011266030371189117, |
|
"rewards/rejected": -0.048201363533735275, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05086285195277021, |
|
"grad_norm": 0.8785350918769836, |
|
"learning_rate": 9.996403375449647e-06, |
|
"logits/chosen": 0.12056512385606766, |
|
"logits/rejected": 0.17535904049873352, |
|
"logps/chosen": -62.035221099853516, |
|
"logps/rejected": -69.88546752929688, |
|
"loss": 2.743, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.03874595835804939, |
|
"rewards/margins": 0.015791552141308784, |
|
"rewards/rejected": -0.054537512362003326, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.05267938237965486, |
|
"grad_norm": 0.9492508172988892, |
|
"learning_rate": 9.996141923535582e-06, |
|
"logits/chosen": 0.12200314551591873, |
|
"logits/rejected": 0.13907021284103394, |
|
"logps/chosen": -65.31945037841797, |
|
"logps/rejected": -66.59127807617188, |
|
"loss": 2.7641, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.0383562371134758, |
|
"rewards/margins": 0.005150892771780491, |
|
"rewards/rejected": -0.043507132679224014, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.05449591280653951, |
|
"grad_norm": 0.8084173202514648, |
|
"learning_rate": 9.995871303205282e-06, |
|
"logits/chosen": 0.07616369426250458, |
|
"logits/rejected": 0.2001558393239975, |
|
"logps/chosen": -51.103309631347656, |
|
"logps/rejected": -62.28483581542969, |
|
"loss": 2.7519, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.04615113511681557, |
|
"rewards/margins": 0.011578184552490711, |
|
"rewards/rejected": -0.057729318737983704, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.05631244323342416, |
|
"grad_norm": 0.943800151348114, |
|
"learning_rate": 9.995591515010589e-06, |
|
"logits/chosen": 0.04337490350008011, |
|
"logits/rejected": 0.15334263443946838, |
|
"logps/chosen": -57.265567779541016, |
|
"logps/rejected": -68.5506820678711, |
|
"loss": 2.7253, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.04982424154877663, |
|
"rewards/margins": 0.024712499231100082, |
|
"rewards/rejected": -0.07453674077987671, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.05812897366030881, |
|
"grad_norm": 0.9372840523719788, |
|
"learning_rate": 9.99530255952204e-06, |
|
"logits/chosen": 0.13595089316368103, |
|
"logits/rejected": 0.14287832379341125, |
|
"logps/chosen": -61.515377044677734, |
|
"logps/rejected": -63.1092529296875, |
|
"loss": 2.7571, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.05759892612695694, |
|
"rewards/margins": 0.009150844067335129, |
|
"rewards/rejected": -0.06674977391958237, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.05994550408719346, |
|
"grad_norm": 0.8408157825469971, |
|
"learning_rate": 9.995004437328866e-06, |
|
"logits/chosen": 0.18553151190280914, |
|
"logits/rejected": 0.14024314284324646, |
|
"logps/chosen": -61.73942565917969, |
|
"logps/rejected": -53.5772819519043, |
|
"loss": 2.7627, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.06591347604990005, |
|
"rewards/margins": 0.006169732194393873, |
|
"rewards/rejected": -0.07208321243524551, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.061762034514078114, |
|
"grad_norm": 0.8856032490730286, |
|
"learning_rate": 9.994697149038989e-06, |
|
"logits/chosen": 0.08410881459712982, |
|
"logits/rejected": 0.13991469144821167, |
|
"logps/chosen": -60.161231994628906, |
|
"logps/rejected": -60.03232192993164, |
|
"loss": 2.7437, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.06631935387849808, |
|
"rewards/margins": 0.016240764409303665, |
|
"rewards/rejected": -0.08256012946367264, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.06357856494096276, |
|
"grad_norm": 0.9037399291992188, |
|
"learning_rate": 9.994380695279025e-06, |
|
"logits/chosen": 0.06669525057077408, |
|
"logits/rejected": 0.16171438992023468, |
|
"logps/chosen": -59.874202728271484, |
|
"logps/rejected": -67.9597396850586, |
|
"loss": 2.7221, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.06999661773443222, |
|
"rewards/margins": 0.027104372158646584, |
|
"rewards/rejected": -0.09710099548101425, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0653950953678474, |
|
"grad_norm": 0.8819396495819092, |
|
"learning_rate": 9.994055076694276e-06, |
|
"logits/chosen": 0.18232542276382446, |
|
"logits/rejected": 0.2102084755897522, |
|
"logps/chosen": -59.298095703125, |
|
"logps/rejected": -63.70309066772461, |
|
"loss": 2.7179, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.08072888106107712, |
|
"rewards/margins": 0.029247857630252838, |
|
"rewards/rejected": -0.10997673869132996, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.06721162579473206, |
|
"grad_norm": 0.9046475291252136, |
|
"learning_rate": 9.993720293948739e-06, |
|
"logits/chosen": 0.13181668519973755, |
|
"logits/rejected": 0.1316901445388794, |
|
"logps/chosen": -71.46318054199219, |
|
"logps/rejected": -76.40208435058594, |
|
"loss": 2.7505, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.08365033566951752, |
|
"rewards/margins": 0.01496690884232521, |
|
"rewards/rejected": -0.09861725568771362, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.06902815622161672, |
|
"grad_norm": 0.9267721772193909, |
|
"learning_rate": 9.993376347725091e-06, |
|
"logits/chosen": 0.14555476605892181, |
|
"logits/rejected": 0.21660488843917847, |
|
"logps/chosen": -61.28938293457031, |
|
"logps/rejected": -67.28329467773438, |
|
"loss": 2.7007, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.07440298050642014, |
|
"rewards/margins": 0.03913095220923424, |
|
"rewards/rejected": -0.11353392899036407, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07084468664850137, |
|
"grad_norm": 0.8545824885368347, |
|
"learning_rate": 9.993023238724696e-06, |
|
"logits/chosen": 0.09474823623895645, |
|
"logits/rejected": 0.09304340183734894, |
|
"logps/chosen": -61.27484130859375, |
|
"logps/rejected": -62.003597259521484, |
|
"loss": 2.7406, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.09001085162162781, |
|
"rewards/margins": 0.017711879685521126, |
|
"rewards/rejected": -0.10772272944450378, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.07266121707538602, |
|
"grad_norm": 0.9086014628410339, |
|
"learning_rate": 9.99266096766761e-06, |
|
"logits/chosen": 0.17910577356815338, |
|
"logits/rejected": 0.14881345629692078, |
|
"logps/chosen": -59.22669219970703, |
|
"logps/rejected": -55.103187561035156, |
|
"loss": 2.7422, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.1086670309305191, |
|
"rewards/margins": 0.01710200309753418, |
|
"rewards/rejected": -0.12576903402805328, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07447774750227067, |
|
"grad_norm": 0.9802166223526001, |
|
"learning_rate": 9.992289535292565e-06, |
|
"logits/chosen": 0.18225271999835968, |
|
"logits/rejected": 0.20810523629188538, |
|
"logps/chosen": -55.57660675048828, |
|
"logps/rejected": -59.129417419433594, |
|
"loss": 2.7505, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.13455544412136078, |
|
"rewards/margins": 0.012644169852137566, |
|
"rewards/rejected": -0.1471996009349823, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.07629427792915532, |
|
"grad_norm": 0.919211208820343, |
|
"learning_rate": 9.991908942356977e-06, |
|
"logits/chosen": 0.07866425067186356, |
|
"logits/rejected": 0.07479682564735413, |
|
"logps/chosen": -66.11737060546875, |
|
"logps/rejected": -64.02122497558594, |
|
"loss": 2.7855, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.15568655729293823, |
|
"rewards/margins": -0.0022164471447467804, |
|
"rewards/rejected": -0.15347009897232056, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.07811080835603997, |
|
"grad_norm": 0.9612410068511963, |
|
"learning_rate": 9.991519189636937e-06, |
|
"logits/chosen": 0.11137107014656067, |
|
"logits/rejected": 0.14462286233901978, |
|
"logps/chosen": -67.43673706054688, |
|
"logps/rejected": -74.29210662841797, |
|
"loss": 2.7203, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.16468609869480133, |
|
"rewards/margins": 0.032481495290994644, |
|
"rewards/rejected": -0.19716759026050568, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.07992733878292461, |
|
"grad_norm": 0.8044790625572205, |
|
"learning_rate": 9.991120277927224e-06, |
|
"logits/chosen": 0.08509679138660431, |
|
"logits/rejected": 0.11362393945455551, |
|
"logps/chosen": -56.867591857910156, |
|
"logps/rejected": -55.66028594970703, |
|
"loss": 2.7302, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.15751589834690094, |
|
"rewards/margins": 0.02398722618818283, |
|
"rewards/rejected": -0.18150311708450317, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.08174386920980926, |
|
"grad_norm": 0.8599545955657959, |
|
"learning_rate": 9.990712208041284e-06, |
|
"logits/chosen": 0.1391746699810028, |
|
"logits/rejected": 0.1746281236410141, |
|
"logps/chosen": -62.22108459472656, |
|
"logps/rejected": -58.50814437866211, |
|
"loss": 2.7288, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.16528920829296112, |
|
"rewards/margins": 0.02449135296046734, |
|
"rewards/rejected": -0.1897805631160736, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.08356039963669391, |
|
"grad_norm": 0.9419183731079102, |
|
"learning_rate": 9.990294980811244e-06, |
|
"logits/chosen": 0.1693490892648697, |
|
"logits/rejected": 0.18133510649204254, |
|
"logps/chosen": -60.769744873046875, |
|
"logps/rejected": -65.61067199707031, |
|
"loss": 2.7003, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.16450145840644836, |
|
"rewards/margins": 0.04409575089812279, |
|
"rewards/rejected": -0.20859721302986145, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.08537693006357856, |
|
"grad_norm": 0.9839057326316833, |
|
"learning_rate": 9.989868597087907e-06, |
|
"logits/chosen": 0.1363808959722519, |
|
"logits/rejected": 0.22261787950992584, |
|
"logps/chosen": -65.37456512451172, |
|
"logps/rejected": -72.04705047607422, |
|
"loss": 2.6636, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.18397849798202515, |
|
"rewards/margins": 0.06703396141529083, |
|
"rewards/rejected": -0.2510124444961548, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.08719346049046321, |
|
"grad_norm": 0.8818926811218262, |
|
"learning_rate": 9.989433057740738e-06, |
|
"logits/chosen": 0.1580447256565094, |
|
"logits/rejected": 0.18051129579544067, |
|
"logps/chosen": -56.6561279296875, |
|
"logps/rejected": -64.66617584228516, |
|
"loss": 2.6658, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2094140350818634, |
|
"rewards/margins": 0.05971873924136162, |
|
"rewards/rejected": -0.2691327929496765, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.08900999091734786, |
|
"grad_norm": 0.8838489651679993, |
|
"learning_rate": 9.98898836365788e-06, |
|
"logits/chosen": 0.11484000086784363, |
|
"logits/rejected": 0.12583914399147034, |
|
"logps/chosen": -65.76107788085938, |
|
"logps/rejected": -75.4742660522461, |
|
"loss": 2.6777, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.19882163405418396, |
|
"rewards/margins": 0.0690067932009697, |
|
"rewards/rejected": -0.26782843470573425, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.09082652134423251, |
|
"grad_norm": 0.9035817980766296, |
|
"learning_rate": 9.988534515746141e-06, |
|
"logits/chosen": 0.12592823803424835, |
|
"logits/rejected": 0.18080386519432068, |
|
"logps/chosen": -60.42765808105469, |
|
"logps/rejected": -65.03330993652344, |
|
"loss": 2.68, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.25037306547164917, |
|
"rewards/margins": 0.05723651498556137, |
|
"rewards/rejected": -0.30760958790779114, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09264305177111716, |
|
"grad_norm": 1.0728137493133545, |
|
"learning_rate": 9.988071514930998e-06, |
|
"logits/chosen": 0.0955006331205368, |
|
"logits/rejected": 0.15643228590488434, |
|
"logps/chosen": -56.28706359863281, |
|
"logps/rejected": -70.3702392578125, |
|
"loss": 2.6104, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.2517072558403015, |
|
"rewards/margins": 0.11176257580518723, |
|
"rewards/rejected": -0.36346977949142456, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.09445958219800181, |
|
"grad_norm": 0.9143723845481873, |
|
"learning_rate": 9.987599362156587e-06, |
|
"logits/chosen": 0.1018611341714859, |
|
"logits/rejected": 0.21036753058433533, |
|
"logps/chosen": -55.815738677978516, |
|
"logps/rejected": -66.23766326904297, |
|
"loss": 2.6207, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.2496231347322464, |
|
"rewards/margins": 0.09438168257474899, |
|
"rewards/rejected": -0.3440048098564148, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.09627611262488647, |
|
"grad_norm": 0.8731315732002258, |
|
"learning_rate": 9.987118058385712e-06, |
|
"logits/chosen": 0.12294681370258331, |
|
"logits/rejected": 0.12897568941116333, |
|
"logps/chosen": -63.27684020996094, |
|
"logps/rejected": -65.21257019042969, |
|
"loss": 2.7093, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.28397974371910095, |
|
"rewards/margins": 0.05316353589296341, |
|
"rewards/rejected": -0.33714330196380615, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.09809264305177112, |
|
"grad_norm": 0.951884388923645, |
|
"learning_rate": 9.986627604599835e-06, |
|
"logits/chosen": 0.1588897705078125, |
|
"logits/rejected": 0.10302369296550751, |
|
"logps/chosen": -62.071083068847656, |
|
"logps/rejected": -59.575416564941406, |
|
"loss": 2.7725, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.31246036291122437, |
|
"rewards/margins": 0.011688929051160812, |
|
"rewards/rejected": -0.3241492807865143, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.09990917347865577, |
|
"grad_norm": 1.0379350185394287, |
|
"learning_rate": 9.986128001799077e-06, |
|
"logits/chosen": 0.15824833512306213, |
|
"logits/rejected": 0.16385456919670105, |
|
"logps/chosen": -79.65828704833984, |
|
"logps/rejected": -82.06480407714844, |
|
"loss": 2.6826, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.36005648970603943, |
|
"rewards/margins": 0.07091177999973297, |
|
"rewards/rejected": -0.4309682846069336, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.10172570390554042, |
|
"grad_norm": 0.8740183115005493, |
|
"learning_rate": 9.985619251002214e-06, |
|
"logits/chosen": 0.0996306911110878, |
|
"logits/rejected": 0.14288735389709473, |
|
"logps/chosen": -59.8499755859375, |
|
"logps/rejected": -63.10956573486328, |
|
"loss": 2.664, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.3242360055446625, |
|
"rewards/margins": 0.06764136999845505, |
|
"rewards/rejected": -0.39187735319137573, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.10354223433242507, |
|
"grad_norm": 0.9942138195037842, |
|
"learning_rate": 9.985101353246676e-06, |
|
"logits/chosen": 0.12070289999246597, |
|
"logits/rejected": 0.1866157352924347, |
|
"logps/chosen": -70.04698181152344, |
|
"logps/rejected": -76.2674560546875, |
|
"loss": 2.6165, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.4166935980319977, |
|
"rewards/margins": 0.10780875384807587, |
|
"rewards/rejected": -0.5245023965835571, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.10535876475930972, |
|
"grad_norm": 0.9078152775764465, |
|
"learning_rate": 9.98457430958855e-06, |
|
"logits/chosen": 0.16699416935443878, |
|
"logits/rejected": 0.216594398021698, |
|
"logps/chosen": -59.81795883178711, |
|
"logps/rejected": -64.79395294189453, |
|
"loss": 2.551, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.39265742897987366, |
|
"rewards/margins": 0.14742323756217957, |
|
"rewards/rejected": -0.5400806665420532, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.10717529518619437, |
|
"grad_norm": 0.9305158257484436, |
|
"learning_rate": 9.984038121102569e-06, |
|
"logits/chosen": 0.16001635789871216, |
|
"logits/rejected": 0.18797104060649872, |
|
"logps/chosen": -58.087158203125, |
|
"logps/rejected": -59.003414154052734, |
|
"loss": 2.6345, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.3626652956008911, |
|
"rewards/margins": 0.08553728461265564, |
|
"rewards/rejected": -0.44820258021354675, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.10899182561307902, |
|
"grad_norm": 0.9690614342689514, |
|
"learning_rate": 9.983492788882113e-06, |
|
"logits/chosen": 0.161406010389328, |
|
"logits/rejected": 0.14847029745578766, |
|
"logps/chosen": -61.01350021362305, |
|
"logps/rejected": -60.993080139160156, |
|
"loss": 2.7099, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.4803884029388428, |
|
"rewards/margins": 0.0831567794084549, |
|
"rewards/rejected": -0.5635451078414917, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11080835603996367, |
|
"grad_norm": 0.971172034740448, |
|
"learning_rate": 9.98293831403921e-06, |
|
"logits/chosen": 0.14095918834209442, |
|
"logits/rejected": 0.16142162680625916, |
|
"logps/chosen": -62.31782150268555, |
|
"logps/rejected": -64.44837188720703, |
|
"loss": 2.6373, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.47784310579299927, |
|
"rewards/margins": 0.09006088227033615, |
|
"rewards/rejected": -0.567903995513916, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.11262488646684832, |
|
"grad_norm": 0.9984462857246399, |
|
"learning_rate": 9.982374697704532e-06, |
|
"logits/chosen": 0.08945554494857788, |
|
"logits/rejected": 0.17623476684093475, |
|
"logps/chosen": -58.84153366088867, |
|
"logps/rejected": -72.59091186523438, |
|
"loss": 2.5465, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.41726160049438477, |
|
"rewards/margins": 0.16527526080608368, |
|
"rewards/rejected": -0.5825368165969849, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.11444141689373297, |
|
"grad_norm": 1.007622241973877, |
|
"learning_rate": 9.981801941027388e-06, |
|
"logits/chosen": 0.16932496428489685, |
|
"logits/rejected": 0.24257409572601318, |
|
"logps/chosen": -61.62455749511719, |
|
"logps/rejected": -69.40538024902344, |
|
"loss": 2.5876, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.4484747052192688, |
|
"rewards/margins": 0.11513397097587585, |
|
"rewards/rejected": -0.563608705997467, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.11625794732061762, |
|
"grad_norm": 1.021275281906128, |
|
"learning_rate": 9.981220045175731e-06, |
|
"logits/chosen": 0.13468854129314423, |
|
"logits/rejected": 0.10648790001869202, |
|
"logps/chosen": -56.91660690307617, |
|
"logps/rejected": -54.84038162231445, |
|
"loss": 2.7096, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.47126585245132446, |
|
"rewards/margins": 0.05656271427869797, |
|
"rewards/rejected": -0.527828574180603, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.11807447774750227, |
|
"grad_norm": 0.9580955505371094, |
|
"learning_rate": 9.980629011336149e-06, |
|
"logits/chosen": 0.10542559623718262, |
|
"logits/rejected": 0.18956655263900757, |
|
"logps/chosen": -63.21677017211914, |
|
"logps/rejected": -68.92955017089844, |
|
"loss": 2.5447, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.4920822083950043, |
|
"rewards/margins": 0.1548418551683426, |
|
"rewards/rejected": -0.6469241380691528, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.11989100817438691, |
|
"grad_norm": 0.9601484537124634, |
|
"learning_rate": 9.98002884071386e-06, |
|
"logits/chosen": 0.1211993545293808, |
|
"logits/rejected": 0.15335297584533691, |
|
"logps/chosen": -65.79328918457031, |
|
"logps/rejected": -71.8165283203125, |
|
"loss": 2.6013, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.4890082776546478, |
|
"rewards/margins": 0.16498278081417084, |
|
"rewards/rejected": -0.6539911031723022, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.12170753860127158, |
|
"grad_norm": 0.9475951194763184, |
|
"learning_rate": 9.97941953453272e-06, |
|
"logits/chosen": 0.129757359623909, |
|
"logits/rejected": 0.19597335159778595, |
|
"logps/chosen": -61.52772903442383, |
|
"logps/rejected": -62.8316764831543, |
|
"loss": 2.6008, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.6176788210868835, |
|
"rewards/margins": 0.11390677094459534, |
|
"rewards/rejected": -0.7315855622291565, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.12352406902815623, |
|
"grad_norm": 1.0577987432479858, |
|
"learning_rate": 9.978801094035207e-06, |
|
"logits/chosen": 0.1363590657711029, |
|
"logits/rejected": 0.1551840454339981, |
|
"logps/chosen": -71.82952117919922, |
|
"logps/rejected": -78.26856994628906, |
|
"loss": 2.6714, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.6073123216629028, |
|
"rewards/margins": 0.08929078280925751, |
|
"rewards/rejected": -0.6966031193733215, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.12534059945504086, |
|
"grad_norm": 1.2286291122436523, |
|
"learning_rate": 9.978173520482429e-06, |
|
"logits/chosen": 0.12522992491722107, |
|
"logits/rejected": 0.1818617582321167, |
|
"logps/chosen": -60.64283752441406, |
|
"logps/rejected": -67.53813934326172, |
|
"loss": 2.6884, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.5357421040534973, |
|
"rewards/margins": 0.0799265205860138, |
|
"rewards/rejected": -0.6156685948371887, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1271571298819255, |
|
"grad_norm": 1.0266982316970825, |
|
"learning_rate": 9.97753681515412e-06, |
|
"logits/chosen": 0.17851999402046204, |
|
"logits/rejected": 0.25888925790786743, |
|
"logps/chosen": -71.65196990966797, |
|
"logps/rejected": -82.34913635253906, |
|
"loss": 2.5226, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.7787651419639587, |
|
"rewards/margins": 0.1905478835105896, |
|
"rewards/rejected": -0.9693130254745483, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12897366030881016, |
|
"grad_norm": 0.999660313129425, |
|
"learning_rate": 9.976890979348628e-06, |
|
"logits/chosen": 0.18746232986450195, |
|
"logits/rejected": 0.2187044620513916, |
|
"logps/chosen": -69.84027099609375, |
|
"logps/rejected": -73.04399108886719, |
|
"loss": 2.6945, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.6969642639160156, |
|
"rewards/margins": 0.08156725764274597, |
|
"rewards/rejected": -0.778531551361084, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.1307901907356948, |
|
"grad_norm": 1.0596684217453003, |
|
"learning_rate": 9.976236014382934e-06, |
|
"logits/chosen": 0.17854923009872437, |
|
"logits/rejected": 0.18649883568286896, |
|
"logps/chosen": -63.992881774902344, |
|
"logps/rejected": -64.36544799804688, |
|
"loss": 2.7689, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.7436442375183105, |
|
"rewards/margins": 0.06337633728981018, |
|
"rewards/rejected": -0.8070206046104431, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.13260672116257946, |
|
"grad_norm": 1.0307612419128418, |
|
"learning_rate": 9.975571921592618e-06, |
|
"logits/chosen": 0.0964425653219223, |
|
"logits/rejected": 0.14382749795913696, |
|
"logps/chosen": -66.61788940429688, |
|
"logps/rejected": -70.80413055419922, |
|
"loss": 2.5859, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.6208910942077637, |
|
"rewards/margins": 0.17540723085403442, |
|
"rewards/rejected": -0.7962983250617981, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1344232515894641, |
|
"grad_norm": 0.9487422108650208, |
|
"learning_rate": 9.97489870233188e-06, |
|
"logits/chosen": 0.06293762475252151, |
|
"logits/rejected": 0.19688080251216888, |
|
"logps/chosen": -55.884613037109375, |
|
"logps/rejected": -64.30046081542969, |
|
"loss": 2.4062, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6415266990661621, |
|
"rewards/margins": 0.2741628885269165, |
|
"rewards/rejected": -0.9156895875930786, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.1362397820163488, |
|
"grad_norm": 1.223663330078125, |
|
"learning_rate": 9.974216357973539e-06, |
|
"logits/chosen": 0.10017126798629761, |
|
"logits/rejected": 0.10504551976919174, |
|
"logps/chosen": -83.83055114746094, |
|
"logps/rejected": -81.8952407836914, |
|
"loss": 2.8556, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8103373050689697, |
|
"rewards/margins": 0.027550537139177322, |
|
"rewards/rejected": -0.8378878831863403, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.13805631244323344, |
|
"grad_norm": 1.0121136903762817, |
|
"learning_rate": 9.973524889909007e-06, |
|
"logits/chosen": 0.07612155377864838, |
|
"logits/rejected": 0.15954618155956268, |
|
"logps/chosen": -65.6831283569336, |
|
"logps/rejected": -75.42146301269531, |
|
"loss": 2.5576, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6988077759742737, |
|
"rewards/margins": 0.1909356415271759, |
|
"rewards/rejected": -0.889743447303772, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.13987284287011809, |
|
"grad_norm": 1.0312719345092773, |
|
"learning_rate": 9.97282429954831e-06, |
|
"logits/chosen": 0.07015375047922134, |
|
"logits/rejected": 0.09143385291099548, |
|
"logps/chosen": -59.444358825683594, |
|
"logps/rejected": -61.26897430419922, |
|
"loss": 2.7845, |
|
"rewards/accuracies": 0.453125, |
|
"rewards/chosen": -0.7886172533035278, |
|
"rewards/margins": 0.038636498153209686, |
|
"rewards/rejected": -0.8272536993026733, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.14168937329700274, |
|
"grad_norm": 1.0625450611114502, |
|
"learning_rate": 9.972114588320073e-06, |
|
"logits/chosen": 0.07196499407291412, |
|
"logits/rejected": 0.12630172073841095, |
|
"logps/chosen": -75.5164566040039, |
|
"logps/rejected": -80.74308776855469, |
|
"loss": 2.5834, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.6821954250335693, |
|
"rewards/margins": 0.16858401894569397, |
|
"rewards/rejected": -0.8507794737815857, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.14350590372388738, |
|
"grad_norm": 0.9069045186042786, |
|
"learning_rate": 9.97139575767152e-06, |
|
"logits/chosen": 0.10926744341850281, |
|
"logits/rejected": 0.14418405294418335, |
|
"logps/chosen": -53.721046447753906, |
|
"logps/rejected": -58.17475891113281, |
|
"loss": 2.4955, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6156648993492126, |
|
"rewards/margins": 0.19551396369934082, |
|
"rewards/rejected": -0.8111788034439087, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.14532243415077203, |
|
"grad_norm": 1.1463656425476074, |
|
"learning_rate": 9.970667809068476e-06, |
|
"logits/chosen": 0.09196805953979492, |
|
"logits/rejected": 0.14393256604671478, |
|
"logps/chosen": -77.11197662353516, |
|
"logps/rejected": -87.39771270751953, |
|
"loss": 2.5763, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8568902611732483, |
|
"rewards/margins": 0.15524335205554962, |
|
"rewards/rejected": -1.0121335983276367, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.14713896457765668, |
|
"grad_norm": 1.0046662092208862, |
|
"learning_rate": 9.969930743995351e-06, |
|
"logits/chosen": 0.2063535749912262, |
|
"logits/rejected": 0.20602768659591675, |
|
"logps/chosen": -59.890933990478516, |
|
"logps/rejected": -59.117794036865234, |
|
"loss": 2.6829, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.7329096794128418, |
|
"rewards/margins": 0.0983649268746376, |
|
"rewards/rejected": -0.8312745094299316, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.14895549500454133, |
|
"grad_norm": 1.0280438661575317, |
|
"learning_rate": 9.969184563955152e-06, |
|
"logits/chosen": 0.17393875122070312, |
|
"logits/rejected": 0.16783203184604645, |
|
"logps/chosen": -72.82754516601562, |
|
"logps/rejected": -70.6113510131836, |
|
"loss": 2.6277, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.7506997585296631, |
|
"rewards/margins": 0.13392671942710876, |
|
"rewards/rejected": -0.8846263885498047, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.15077202543142598, |
|
"grad_norm": 0.9356668591499329, |
|
"learning_rate": 9.968429270469467e-06, |
|
"logits/chosen": 0.09279462695121765, |
|
"logits/rejected": 0.1681107133626938, |
|
"logps/chosen": -62.300445556640625, |
|
"logps/rejected": -68.70709991455078, |
|
"loss": 2.4821, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.7132639288902283, |
|
"rewards/margins": 0.2519262731075287, |
|
"rewards/rejected": -0.9651902318000793, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.15258855585831063, |
|
"grad_norm": 1.0442771911621094, |
|
"learning_rate": 9.967664865078472e-06, |
|
"logits/chosen": 0.17844007909297943, |
|
"logits/rejected": 0.21383100748062134, |
|
"logps/chosen": -61.43730163574219, |
|
"logps/rejected": -64.16680908203125, |
|
"loss": 2.6289, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.7574411034584045, |
|
"rewards/margins": 0.13140378892421722, |
|
"rewards/rejected": -0.888844907283783, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.15440508628519528, |
|
"grad_norm": 1.137267827987671, |
|
"learning_rate": 9.966891349340922e-06, |
|
"logits/chosen": 0.19703873991966248, |
|
"logits/rejected": 0.17027492821216583, |
|
"logps/chosen": -68.07546997070312, |
|
"logps/rejected": -63.9476203918457, |
|
"loss": 2.7702, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.8362983465194702, |
|
"rewards/margins": 0.05476854741573334, |
|
"rewards/rejected": -0.8910670280456543, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.15622161671207993, |
|
"grad_norm": 1.0395056009292603, |
|
"learning_rate": 9.966108724834151e-06, |
|
"logits/chosen": 0.10182631760835648, |
|
"logits/rejected": 0.11173731088638306, |
|
"logps/chosen": -64.47720336914062, |
|
"logps/rejected": -75.30056762695312, |
|
"loss": 2.5889, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8160837292671204, |
|
"rewards/margins": 0.2798304557800293, |
|
"rewards/rejected": -1.0959142446517944, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.15803814713896458, |
|
"grad_norm": 1.1256901025772095, |
|
"learning_rate": 9.965316993154069e-06, |
|
"logits/chosen": 0.1232331395149231, |
|
"logits/rejected": 0.19064000248908997, |
|
"logps/chosen": -64.3778305053711, |
|
"logps/rejected": -76.91246032714844, |
|
"loss": 2.5522, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8341800570487976, |
|
"rewards/margins": 0.2616519033908844, |
|
"rewards/rejected": -1.0958319902420044, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.15985467756584923, |
|
"grad_norm": 1.20338773727417, |
|
"learning_rate": 9.964516155915152e-06, |
|
"logits/chosen": 0.12199988961219788, |
|
"logits/rejected": 0.12264345586299896, |
|
"logps/chosen": -81.56141662597656, |
|
"logps/rejected": -80.49986267089844, |
|
"loss": 2.7432, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.7930806875228882, |
|
"rewards/margins": 0.06459490954875946, |
|
"rewards/rejected": -0.8576756119728088, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.16167120799273388, |
|
"grad_norm": 1.005759835243225, |
|
"learning_rate": 9.963706214750446e-06, |
|
"logits/chosen": 0.07040335237979889, |
|
"logits/rejected": 0.10741756111383438, |
|
"logps/chosen": -65.43824768066406, |
|
"logps/rejected": -74.47370147705078, |
|
"loss": 2.5472, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.8203200697898865, |
|
"rewards/margins": 0.24586233496665955, |
|
"rewards/rejected": -1.0661823749542236, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.16348773841961853, |
|
"grad_norm": 0.9761494994163513, |
|
"learning_rate": 9.962887171311563e-06, |
|
"logits/chosen": 0.1630188375711441, |
|
"logits/rejected": 0.190489262342453, |
|
"logps/chosen": -57.84222412109375, |
|
"logps/rejected": -64.116455078125, |
|
"loss": 2.5339, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8240211009979248, |
|
"rewards/margins": 0.21757929027080536, |
|
"rewards/rejected": -1.0416003465652466, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.16530426884650318, |
|
"grad_norm": 1.002192497253418, |
|
"learning_rate": 9.962059027268676e-06, |
|
"logits/chosen": 0.08669686317443848, |
|
"logits/rejected": 0.13606388866901398, |
|
"logps/chosen": -63.68013000488281, |
|
"logps/rejected": -72.30770111083984, |
|
"loss": 2.5165, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.7919780611991882, |
|
"rewards/margins": 0.27081233263015747, |
|
"rewards/rejected": -1.0627902746200562, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.16712079927338783, |
|
"grad_norm": 0.9397479295730591, |
|
"learning_rate": 9.961221784310514e-06, |
|
"logits/chosen": 0.14396092295646667, |
|
"logits/rejected": 0.1820589005947113, |
|
"logps/chosen": -57.551353454589844, |
|
"logps/rejected": -66.83124542236328, |
|
"loss": 2.4254, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.7751535177230835, |
|
"rewards/margins": 0.26002568006515503, |
|
"rewards/rejected": -1.0351792573928833, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.16893732970027248, |
|
"grad_norm": 1.2053430080413818, |
|
"learning_rate": 9.96037544414436e-06, |
|
"logits/chosen": 0.16130733489990234, |
|
"logits/rejected": 0.1852559745311737, |
|
"logps/chosen": -73.8154067993164, |
|
"logps/rejected": -81.3624267578125, |
|
"loss": 2.7543, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.8622196316719055, |
|
"rewards/margins": 0.11067891120910645, |
|
"rewards/rejected": -0.9728984236717224, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.17075386012715713, |
|
"grad_norm": 1.1681262254714966, |
|
"learning_rate": 9.959520008496054e-06, |
|
"logits/chosen": 0.17866984009742737, |
|
"logits/rejected": 0.20266617834568024, |
|
"logps/chosen": -61.40083312988281, |
|
"logps/rejected": -65.13737487792969, |
|
"loss": 2.7134, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9223482012748718, |
|
"rewards/margins": 0.08481010049581528, |
|
"rewards/rejected": -1.0071581602096558, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.17257039055404177, |
|
"grad_norm": 1.0269131660461426, |
|
"learning_rate": 9.95865547910997e-06, |
|
"logits/chosen": 0.11446906626224518, |
|
"logits/rejected": 0.17379909753799438, |
|
"logps/chosen": -62.823944091796875, |
|
"logps/rejected": -69.5682601928711, |
|
"loss": 2.5031, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8894773125648499, |
|
"rewards/margins": 0.24369436502456665, |
|
"rewards/rejected": -1.1331716775894165, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.17438692098092642, |
|
"grad_norm": 1.0400865077972412, |
|
"learning_rate": 9.957781857749043e-06, |
|
"logits/chosen": 0.16556067764759064, |
|
"logits/rejected": 0.23536017537117004, |
|
"logps/chosen": -61.09260940551758, |
|
"logps/rejected": -63.444427490234375, |
|
"loss": 2.7239, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.0037989616394043, |
|
"rewards/margins": 0.07835513353347778, |
|
"rewards/rejected": -1.0821542739868164, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.17620345140781107, |
|
"grad_norm": 1.1138463020324707, |
|
"learning_rate": 9.956899146194732e-06, |
|
"logits/chosen": 0.13205701112747192, |
|
"logits/rejected": 0.1702471375465393, |
|
"logps/chosen": -63.70584487915039, |
|
"logps/rejected": -69.10890197753906, |
|
"loss": 2.7076, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9300363659858704, |
|
"rewards/margins": 0.17512086033821106, |
|
"rewards/rejected": -1.1051572561264038, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.17801998183469572, |
|
"grad_norm": 0.9917119741439819, |
|
"learning_rate": 9.956007346247043e-06, |
|
"logits/chosen": 0.12265195697546005, |
|
"logits/rejected": 0.18780440092086792, |
|
"logps/chosen": -62.316497802734375, |
|
"logps/rejected": -69.26319885253906, |
|
"loss": 2.4996, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8957506418228149, |
|
"rewards/margins": 0.2764662802219391, |
|
"rewards/rejected": -1.1722170114517212, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.17983651226158037, |
|
"grad_norm": 1.00100576877594, |
|
"learning_rate": 9.95510645972451e-06, |
|
"logits/chosen": 0.16969357430934906, |
|
"logits/rejected": 0.16122889518737793, |
|
"logps/chosen": -60.42213439941406, |
|
"logps/rejected": -63.85418701171875, |
|
"loss": 2.5677, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8840566873550415, |
|
"rewards/margins": 0.1611585170030594, |
|
"rewards/rejected": -1.0452152490615845, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.18165304268846502, |
|
"grad_norm": 1.1503106355667114, |
|
"learning_rate": 9.954196488464198e-06, |
|
"logits/chosen": 0.19831174612045288, |
|
"logits/rejected": 0.20889577269554138, |
|
"logps/chosen": -65.17144775390625, |
|
"logps/rejected": -69.9671859741211, |
|
"loss": 2.7065, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.9322465062141418, |
|
"rewards/margins": 0.1645345687866211, |
|
"rewards/rejected": -1.0967810153961182, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.18346957311534967, |
|
"grad_norm": 1.416610836982727, |
|
"learning_rate": 9.953277434321696e-06, |
|
"logits/chosen": 0.08757522702217102, |
|
"logits/rejected": 0.1615956872701645, |
|
"logps/chosen": -72.49248504638672, |
|
"logps/rejected": -80.52176666259766, |
|
"loss": 2.5346, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8279029130935669, |
|
"rewards/margins": 0.26418614387512207, |
|
"rewards/rejected": -1.092089056968689, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.18528610354223432, |
|
"grad_norm": 0.9907792210578918, |
|
"learning_rate": 9.952349299171117e-06, |
|
"logits/chosen": 0.07832024991512299, |
|
"logits/rejected": 0.16370174288749695, |
|
"logps/chosen": -67.5047836303711, |
|
"logps/rejected": -73.78684997558594, |
|
"loss": 2.4325, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8129651546478271, |
|
"rewards/margins": 0.3001910448074341, |
|
"rewards/rejected": -1.1131561994552612, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.18710263396911897, |
|
"grad_norm": 1.1051058769226074, |
|
"learning_rate": 9.95141208490509e-06, |
|
"logits/chosen": 0.10540622472763062, |
|
"logits/rejected": 0.2087182104587555, |
|
"logps/chosen": -55.68672180175781, |
|
"logps/rejected": -72.7021484375, |
|
"loss": 2.3004, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8526840209960938, |
|
"rewards/margins": 0.3691489100456238, |
|
"rewards/rejected": -1.2218331098556519, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.18891916439600362, |
|
"grad_norm": 1.0593968629837036, |
|
"learning_rate": 9.950465793434759e-06, |
|
"logits/chosen": 0.10488969832658768, |
|
"logits/rejected": 0.13292263448238373, |
|
"logps/chosen": -69.47488403320312, |
|
"logps/rejected": -73.08143615722656, |
|
"loss": 2.5388, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.8357015252113342, |
|
"rewards/margins": 0.2204672396183014, |
|
"rewards/rejected": -1.0561686754226685, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.1907356948228883, |
|
"grad_norm": 1.119372010231018, |
|
"learning_rate": 9.949510426689773e-06, |
|
"logits/chosen": 0.06970994919538498, |
|
"logits/rejected": 0.0831708014011383, |
|
"logps/chosen": -66.97635650634766, |
|
"logps/rejected": -67.95066833496094, |
|
"loss": 2.6443, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.9023284912109375, |
|
"rewards/margins": 0.1277586966753006, |
|
"rewards/rejected": -1.0300871133804321, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.19255222524977295, |
|
"grad_norm": 1.2842986583709717, |
|
"learning_rate": 9.948545986618295e-06, |
|
"logits/chosen": 0.17146825790405273, |
|
"logits/rejected": 0.1910206824541092, |
|
"logps/chosen": -69.1142578125, |
|
"logps/rejected": -70.71878814697266, |
|
"loss": 2.8494, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.0275481939315796, |
|
"rewards/margins": 0.09755454212427139, |
|
"rewards/rejected": -1.1251027584075928, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.1943687556766576, |
|
"grad_norm": 1.044800877571106, |
|
"learning_rate": 9.947572475186984e-06, |
|
"logits/chosen": 0.1206701397895813, |
|
"logits/rejected": 0.17635390162467957, |
|
"logps/chosen": -67.00174713134766, |
|
"logps/rejected": -71.5347900390625, |
|
"loss": 2.46, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.889815628528595, |
|
"rewards/margins": 0.255011647939682, |
|
"rewards/rejected": -1.1448272466659546, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.19618528610354224, |
|
"grad_norm": 1.1496058702468872, |
|
"learning_rate": 9.946589894381002e-06, |
|
"logits/chosen": 0.18972846865653992, |
|
"logits/rejected": 0.12482471019029617, |
|
"logps/chosen": -66.56938934326172, |
|
"logps/rejected": -73.7935791015625, |
|
"loss": 2.6472, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.8382111191749573, |
|
"rewards/margins": 0.16597144305706024, |
|
"rewards/rejected": -1.0041825771331787, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.1980018165304269, |
|
"grad_norm": 1.1357221603393555, |
|
"learning_rate": 9.945598246204e-06, |
|
"logits/chosen": 0.1766098588705063, |
|
"logits/rejected": 0.16134579479694366, |
|
"logps/chosen": -72.30979919433594, |
|
"logps/rejected": -69.26260375976562, |
|
"loss": 2.7037, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.9316724538803101, |
|
"rewards/margins": 0.0998368114233017, |
|
"rewards/rejected": -1.0315091609954834, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.19981834695731154, |
|
"grad_norm": 1.1147487163543701, |
|
"learning_rate": 9.94459753267812e-06, |
|
"logits/chosen": 0.10100046545267105, |
|
"logits/rejected": 0.12630510330200195, |
|
"logps/chosen": -66.10484313964844, |
|
"logps/rejected": -68.13362121582031, |
|
"loss": 2.6224, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8918489217758179, |
|
"rewards/margins": 0.165731742978096, |
|
"rewards/rejected": -1.0575807094573975, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2016348773841962, |
|
"grad_norm": 1.3498740196228027, |
|
"learning_rate": 9.943587755843996e-06, |
|
"logits/chosen": 0.19711939990520477, |
|
"logits/rejected": 0.18357205390930176, |
|
"logps/chosen": -64.5146484375, |
|
"logps/rejected": -65.80925750732422, |
|
"loss": 2.6102, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.9744794368743896, |
|
"rewards/margins": 0.16535742580890656, |
|
"rewards/rejected": -1.1398367881774902, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.20345140781108084, |
|
"grad_norm": 1.0210968255996704, |
|
"learning_rate": 9.942568917760733e-06, |
|
"logits/chosen": 0.16077302396297455, |
|
"logits/rejected": 0.22646722197532654, |
|
"logps/chosen": -59.73281478881836, |
|
"logps/rejected": -64.54032897949219, |
|
"loss": 2.4897, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8548685312271118, |
|
"rewards/margins": 0.19910478591918945, |
|
"rewards/rejected": -1.0539731979370117, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.2052679382379655, |
|
"grad_norm": 1.160152554512024, |
|
"learning_rate": 9.941541020505924e-06, |
|
"logits/chosen": 0.10292509198188782, |
|
"logits/rejected": 0.09458380937576294, |
|
"logps/chosen": -77.44420623779297, |
|
"logps/rejected": -85.09677124023438, |
|
"loss": 2.5037, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.9961792230606079, |
|
"rewards/margins": 0.2911871373653412, |
|
"rewards/rejected": -1.287366509437561, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.20708446866485014, |
|
"grad_norm": 1.1742796897888184, |
|
"learning_rate": 9.940504066175626e-06, |
|
"logits/chosen": 0.10538715869188309, |
|
"logits/rejected": 0.07242526113986969, |
|
"logps/chosen": -70.37040710449219, |
|
"logps/rejected": -68.23387908935547, |
|
"loss": 2.6328, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.9170363545417786, |
|
"rewards/margins": 0.14501458406448364, |
|
"rewards/rejected": -1.0620509386062622, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.2089009990917348, |
|
"grad_norm": 1.1259719133377075, |
|
"learning_rate": 9.939458056884375e-06, |
|
"logits/chosen": 0.12264375388622284, |
|
"logits/rejected": 0.16776585578918457, |
|
"logps/chosen": -59.847049713134766, |
|
"logps/rejected": -65.45285034179688, |
|
"loss": 2.5752, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.8872510194778442, |
|
"rewards/margins": 0.2041575014591217, |
|
"rewards/rejected": -1.0914084911346436, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.21071752951861944, |
|
"grad_norm": 2.1809544563293457, |
|
"learning_rate": 9.938402994765163e-06, |
|
"logits/chosen": 0.13642175495624542, |
|
"logits/rejected": 0.11773751676082611, |
|
"logps/chosen": -65.34564971923828, |
|
"logps/rejected": -66.55570983886719, |
|
"loss": 2.6792, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9432386755943298, |
|
"rewards/margins": 0.12600603699684143, |
|
"rewards/rejected": -1.0692447423934937, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.2125340599455041, |
|
"grad_norm": 1.4895256757736206, |
|
"learning_rate": 9.937338881969444e-06, |
|
"logits/chosen": 0.12349803745746613, |
|
"logits/rejected": 0.1182522252202034, |
|
"logps/chosen": -66.9234619140625, |
|
"logps/rejected": -69.5535888671875, |
|
"loss": 2.6029, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.9096715450286865, |
|
"rewards/margins": 0.23732726275920868, |
|
"rewards/rejected": -1.1469988822937012, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.21435059037238874, |
|
"grad_norm": 1.0333324670791626, |
|
"learning_rate": 9.93626572066713e-06, |
|
"logits/chosen": 0.16722331941127777, |
|
"logits/rejected": 0.21501797437667847, |
|
"logps/chosen": -66.49773406982422, |
|
"logps/rejected": -72.3947982788086, |
|
"loss": 2.4687, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.9357748627662659, |
|
"rewards/margins": 0.2686734199523926, |
|
"rewards/rejected": -1.2044482231140137, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2161671207992734, |
|
"grad_norm": 1.2127379179000854, |
|
"learning_rate": 9.935183513046585e-06, |
|
"logits/chosen": 0.10065922141075134, |
|
"logits/rejected": 0.12764661014080048, |
|
"logps/chosen": -73.05583953857422, |
|
"logps/rejected": -77.18297576904297, |
|
"loss": 2.6288, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8347401022911072, |
|
"rewards/margins": 0.2145780473947525, |
|
"rewards/rejected": -1.0493181943893433, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.21798365122615804, |
|
"grad_norm": 1.4050469398498535, |
|
"learning_rate": 9.93409226131462e-06, |
|
"logits/chosen": 0.08468589186668396, |
|
"logits/rejected": 0.179460808634758, |
|
"logps/chosen": -66.12811279296875, |
|
"logps/rejected": -74.2131118774414, |
|
"loss": 2.4658, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8611673712730408, |
|
"rewards/margins": 0.20363186299800873, |
|
"rewards/rejected": -1.064799189567566, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.2198001816530427, |
|
"grad_norm": 1.18356192111969, |
|
"learning_rate": 9.932991967696484e-06, |
|
"logits/chosen": 0.18308167159557343, |
|
"logits/rejected": 0.12839025259017944, |
|
"logps/chosen": -63.325035095214844, |
|
"logps/rejected": -63.72959518432617, |
|
"loss": 2.579, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9539028406143188, |
|
"rewards/margins": 0.16736355423927307, |
|
"rewards/rejected": -1.1212663650512695, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.22161671207992734, |
|
"grad_norm": 1.1829110383987427, |
|
"learning_rate": 9.931882634435868e-06, |
|
"logits/chosen": 0.10780126601457596, |
|
"logits/rejected": 0.109224833548069, |
|
"logps/chosen": -71.1589126586914, |
|
"logps/rejected": -74.609375, |
|
"loss": 2.6368, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.9908429384231567, |
|
"rewards/margins": 0.12795159220695496, |
|
"rewards/rejected": -1.118794560432434, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.22343324250681199, |
|
"grad_norm": 1.1116541624069214, |
|
"learning_rate": 9.930764263794898e-06, |
|
"logits/chosen": 0.10946042090654373, |
|
"logits/rejected": 0.11234283447265625, |
|
"logps/chosen": -67.98091888427734, |
|
"logps/rejected": -70.58726501464844, |
|
"loss": 2.4417, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.871981143951416, |
|
"rewards/margins": 0.22934575378894806, |
|
"rewards/rejected": -1.1013269424438477, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.22524977293369663, |
|
"grad_norm": 1.1564345359802246, |
|
"learning_rate": 9.929636858054124e-06, |
|
"logits/chosen": 0.11665979772806168, |
|
"logits/rejected": 0.15129488706588745, |
|
"logps/chosen": -62.33778381347656, |
|
"logps/rejected": -67.58541870117188, |
|
"loss": 2.6023, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8282152414321899, |
|
"rewards/margins": 0.206033393740654, |
|
"rewards/rejected": -1.0342485904693604, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.22706630336058128, |
|
"grad_norm": 1.149032711982727, |
|
"learning_rate": 9.928500419512522e-06, |
|
"logits/chosen": 0.02143971435725689, |
|
"logits/rejected": 0.10582254827022552, |
|
"logps/chosen": -66.31533813476562, |
|
"logps/rejected": -76.4206771850586, |
|
"loss": 2.4589, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.8935360908508301, |
|
"rewards/margins": 0.3176972270011902, |
|
"rewards/rejected": -1.2112332582473755, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.22888283378746593, |
|
"grad_norm": 1.1076140403747559, |
|
"learning_rate": 9.927354950487491e-06, |
|
"logits/chosen": 0.13772962987422943, |
|
"logits/rejected": 0.12653110921382904, |
|
"logps/chosen": -62.97186279296875, |
|
"logps/rejected": -65.31454467773438, |
|
"loss": 2.4839, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9201483130455017, |
|
"rewards/margins": 0.2287607342004776, |
|
"rewards/rejected": -1.1489089727401733, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.23069936421435058, |
|
"grad_norm": 1.1476696729660034, |
|
"learning_rate": 9.926200453314842e-06, |
|
"logits/chosen": 0.11984287202358246, |
|
"logits/rejected": 0.1570628136396408, |
|
"logps/chosen": -72.74269104003906, |
|
"logps/rejected": -77.52953338623047, |
|
"loss": 2.5198, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9817119836807251, |
|
"rewards/margins": 0.2599087059497833, |
|
"rewards/rejected": -1.241620659828186, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.23251589464123523, |
|
"grad_norm": 1.0800604820251465, |
|
"learning_rate": 9.925036930348793e-06, |
|
"logits/chosen": 0.03499937057495117, |
|
"logits/rejected": 0.16788874566555023, |
|
"logps/chosen": -64.4096450805664, |
|
"logps/rejected": -81.16387939453125, |
|
"loss": 2.3547, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.040608286857605, |
|
"rewards/margins": 0.48978835344314575, |
|
"rewards/rejected": -1.530396580696106, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.23433242506811988, |
|
"grad_norm": 1.2677640914916992, |
|
"learning_rate": 9.92386438396197e-06, |
|
"logits/chosen": 0.1200772225856781, |
|
"logits/rejected": 0.11345633864402771, |
|
"logps/chosen": -68.35733032226562, |
|
"logps/rejected": -71.72528839111328, |
|
"loss": 2.631, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.9437850117683411, |
|
"rewards/margins": 0.13758014142513275, |
|
"rewards/rejected": -1.0813652276992798, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.23614895549500453, |
|
"grad_norm": 1.2041019201278687, |
|
"learning_rate": 9.9226828165454e-06, |
|
"logits/chosen": 0.07035915553569794, |
|
"logits/rejected": 0.16790318489074707, |
|
"logps/chosen": -73.99917602539062, |
|
"logps/rejected": -83.07939147949219, |
|
"loss": 2.4497, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0043660402297974, |
|
"rewards/margins": 0.36712244153022766, |
|
"rewards/rejected": -1.3714885711669922, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.23796548592188918, |
|
"grad_norm": 1.112176537513733, |
|
"learning_rate": 9.921492230508507e-06, |
|
"logits/chosen": 0.18427804112434387, |
|
"logits/rejected": 0.17424902319908142, |
|
"logps/chosen": -63.21034240722656, |
|
"logps/rejected": -65.23414611816406, |
|
"loss": 2.4612, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.9832345843315125, |
|
"rewards/margins": 0.21530815958976746, |
|
"rewards/rejected": -1.1985427141189575, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.23978201634877383, |
|
"grad_norm": 1.1540772914886475, |
|
"learning_rate": 9.9202926282791e-06, |
|
"logits/chosen": 0.14970409870147705, |
|
"logits/rejected": 0.17991000413894653, |
|
"logps/chosen": -76.66849517822266, |
|
"logps/rejected": -85.59910583496094, |
|
"loss": 2.4164, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0053367614746094, |
|
"rewards/margins": 0.3513883352279663, |
|
"rewards/rejected": -1.3567250967025757, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.24159854677565848, |
|
"grad_norm": 1.32212495803833, |
|
"learning_rate": 9.919084012303378e-06, |
|
"logits/chosen": 0.11897419393062592, |
|
"logits/rejected": 0.15737299621105194, |
|
"logps/chosen": -75.7364501953125, |
|
"logps/rejected": -77.97785186767578, |
|
"loss": 2.7243, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.139754295349121, |
|
"rewards/margins": 0.17446394264698029, |
|
"rewards/rejected": -1.3142181634902954, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.24341507720254316, |
|
"grad_norm": 1.4282020330429077, |
|
"learning_rate": 9.917866385045918e-06, |
|
"logits/chosen": 0.13694903254508972, |
|
"logits/rejected": 0.2048657089471817, |
|
"logps/chosen": -70.1098403930664, |
|
"logps/rejected": -78.7418212890625, |
|
"loss": 2.8086, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.2937737703323364, |
|
"rewards/margins": 0.2787961959838867, |
|
"rewards/rejected": -1.5725698471069336, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2452316076294278, |
|
"grad_norm": 1.1548171043395996, |
|
"learning_rate": 9.916639748989677e-06, |
|
"logits/chosen": 0.050287194550037384, |
|
"logits/rejected": 0.10146909952163696, |
|
"logps/chosen": -72.51569366455078, |
|
"logps/rejected": -80.42554473876953, |
|
"loss": 2.4774, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0650720596313477, |
|
"rewards/margins": 0.2093387395143509, |
|
"rewards/rejected": -1.2744107246398926, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.24704813805631246, |
|
"grad_norm": 1.1347655057907104, |
|
"learning_rate": 9.915404106635979e-06, |
|
"logits/chosen": 0.13506914675235748, |
|
"logits/rejected": 0.18150334060192108, |
|
"logps/chosen": -71.03006744384766, |
|
"logps/rejected": -75.59716033935547, |
|
"loss": 2.4399, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1068118810653687, |
|
"rewards/margins": 0.3031606674194336, |
|
"rewards/rejected": -1.4099724292755127, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2488646684831971, |
|
"grad_norm": 1.1634560823440552, |
|
"learning_rate": 9.914159460504512e-06, |
|
"logits/chosen": 0.17681393027305603, |
|
"logits/rejected": 0.2249586582183838, |
|
"logps/chosen": -66.17890167236328, |
|
"logps/rejected": -71.03226470947266, |
|
"loss": 2.5281, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0759899616241455, |
|
"rewards/margins": 0.21175377070903778, |
|
"rewards/rejected": -1.2877436876296997, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.2506811989100817, |
|
"grad_norm": 2.4183638095855713, |
|
"learning_rate": 9.912905813133325e-06, |
|
"logits/chosen": 0.053306616842746735, |
|
"logits/rejected": 0.08808214217424393, |
|
"logps/chosen": -75.88124084472656, |
|
"logps/rejected": -86.6006088256836, |
|
"loss": 2.5718, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.1421475410461426, |
|
"rewards/margins": 0.36426350474357605, |
|
"rewards/rejected": -1.506411075592041, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.2524977293369664, |
|
"grad_norm": 1.1398063898086548, |
|
"learning_rate": 9.911643167078827e-06, |
|
"logits/chosen": 0.046535998582839966, |
|
"logits/rejected": 0.11657831072807312, |
|
"logps/chosen": -62.68219757080078, |
|
"logps/rejected": -74.52103424072266, |
|
"loss": 2.5334, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.0171339511871338, |
|
"rewards/margins": 0.3033108115196228, |
|
"rewards/rejected": -1.3204445838928223, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.254314259763851, |
|
"grad_norm": 1.2559521198272705, |
|
"learning_rate": 9.91037152491577e-06, |
|
"logits/chosen": 0.07763661444187164, |
|
"logits/rejected": 0.11103180050849915, |
|
"logps/chosen": -68.87091064453125, |
|
"logps/rejected": -79.08797454833984, |
|
"loss": 2.454, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.017935872077942, |
|
"rewards/margins": 0.2824122905731201, |
|
"rewards/rejected": -1.300348162651062, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2561307901907357, |
|
"grad_norm": 1.1818301677703857, |
|
"learning_rate": 9.909090889237257e-06, |
|
"logits/chosen": 0.07567673176527023, |
|
"logits/rejected": 0.1396160125732422, |
|
"logps/chosen": -62.44426345825195, |
|
"logps/rejected": -61.77680206298828, |
|
"loss": 2.4511, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.9497895240783691, |
|
"rewards/margins": 0.23719964921474457, |
|
"rewards/rejected": -1.1869890689849854, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.2579473206176203, |
|
"grad_norm": 1.1855460405349731, |
|
"learning_rate": 9.907801262654725e-06, |
|
"logits/chosen": 0.08448053896427155, |
|
"logits/rejected": 0.1705108880996704, |
|
"logps/chosen": -66.66730499267578, |
|
"logps/rejected": -76.75447845458984, |
|
"loss": 2.4476, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.105279803276062, |
|
"rewards/margins": 0.2967475950717926, |
|
"rewards/rejected": -1.4020274877548218, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.259763851044505, |
|
"grad_norm": 1.3652665615081787, |
|
"learning_rate": 9.906502647797946e-06, |
|
"logits/chosen": 0.09915038198232651, |
|
"logits/rejected": 0.08110683411359787, |
|
"logps/chosen": -66.32193756103516, |
|
"logps/rejected": -80.20841217041016, |
|
"loss": 2.5982, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -1.1550779342651367, |
|
"rewards/margins": 0.36828702688217163, |
|
"rewards/rejected": -1.5233650207519531, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.2615803814713896, |
|
"grad_norm": 1.4188799858093262, |
|
"learning_rate": 9.905195047315024e-06, |
|
"logits/chosen": 0.12019304931163788, |
|
"logits/rejected": 0.11188551783561707, |
|
"logps/chosen": -85.44612884521484, |
|
"logps/rejected": -91.55411529541016, |
|
"loss": 2.8244, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -1.171909213066101, |
|
"rewards/margins": 0.16801074147224426, |
|
"rewards/rejected": -1.3399198055267334, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.2633969118982743, |
|
"grad_norm": 1.307726502418518, |
|
"learning_rate": 9.903878463872384e-06, |
|
"logits/chosen": 0.12220380455255508, |
|
"logits/rejected": 0.08115807920694351, |
|
"logps/chosen": -71.56598663330078, |
|
"logps/rejected": -71.203857421875, |
|
"loss": 2.5832, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.1082667112350464, |
|
"rewards/margins": 0.18824435770511627, |
|
"rewards/rejected": -1.296510934829712, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.2652134423251589, |
|
"grad_norm": 1.0912556648254395, |
|
"learning_rate": 9.902552900154769e-06, |
|
"logits/chosen": 0.14343701303005219, |
|
"logits/rejected": 0.17181995511054993, |
|
"logps/chosen": -65.52398681640625, |
|
"logps/rejected": -72.67620086669922, |
|
"loss": 2.3888, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0277646780014038, |
|
"rewards/margins": 0.34770333766937256, |
|
"rewards/rejected": -1.375468134880066, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.2670299727520436, |
|
"grad_norm": 1.375835657119751, |
|
"learning_rate": 9.90121835886523e-06, |
|
"logits/chosen": 0.06410901993513107, |
|
"logits/rejected": 0.09152361750602722, |
|
"logps/chosen": -78.27774047851562, |
|
"logps/rejected": -82.27354431152344, |
|
"loss": 2.5421, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.0331169366836548, |
|
"rewards/margins": 0.20058104395866394, |
|
"rewards/rejected": -1.2336980104446411, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.2688465031789282, |
|
"grad_norm": 1.1775563955307007, |
|
"learning_rate": 9.899874842725136e-06, |
|
"logits/chosen": 0.15871602296829224, |
|
"logits/rejected": 0.12121336162090302, |
|
"logps/chosen": -68.25074005126953, |
|
"logps/rejected": -69.08786010742188, |
|
"loss": 2.5082, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.1035441160202026, |
|
"rewards/margins": 0.2747136950492859, |
|
"rewards/rejected": -1.3782578706741333, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.2706630336058129, |
|
"grad_norm": 1.1836682558059692, |
|
"learning_rate": 9.898522354474144e-06, |
|
"logits/chosen": 0.10703336447477341, |
|
"logits/rejected": 0.09559071063995361, |
|
"logps/chosen": -67.44327545166016, |
|
"logps/rejected": -66.1326904296875, |
|
"loss": 2.4807, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0860165357589722, |
|
"rewards/margins": 0.24197286367416382, |
|
"rewards/rejected": -1.3279893398284912, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.2724795640326976, |
|
"grad_norm": 1.4430557489395142, |
|
"learning_rate": 9.897160896870217e-06, |
|
"logits/chosen": 0.1853700429201126, |
|
"logits/rejected": 0.2392357438802719, |
|
"logps/chosen": -69.39933776855469, |
|
"logps/rejected": -76.33921813964844, |
|
"loss": 2.6237, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.2587306499481201, |
|
"rewards/margins": 0.1618267297744751, |
|
"rewards/rejected": -1.4205572605133057, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2742960944595822, |
|
"grad_norm": 1.3623749017715454, |
|
"learning_rate": 9.895790472689605e-06, |
|
"logits/chosen": 0.15992893278598785, |
|
"logits/rejected": 0.09628183394670486, |
|
"logps/chosen": -70.02740478515625, |
|
"logps/rejected": -66.36286926269531, |
|
"loss": 2.7214, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.1345252990722656, |
|
"rewards/margins": 0.1051594465970993, |
|
"rewards/rejected": -1.2396849393844604, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.2761126248864669, |
|
"grad_norm": 1.3764369487762451, |
|
"learning_rate": 9.894411084726837e-06, |
|
"logits/chosen": 0.14009161293506622, |
|
"logits/rejected": 0.13651950657367706, |
|
"logps/chosen": -74.57782745361328, |
|
"logps/rejected": -79.09630584716797, |
|
"loss": 2.5094, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.1602882146835327, |
|
"rewards/margins": 0.3822028338909149, |
|
"rewards/rejected": -1.54249107837677, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.2779291553133515, |
|
"grad_norm": 1.3273645639419556, |
|
"learning_rate": 9.893022735794728e-06, |
|
"logits/chosen": 0.03166107460856438, |
|
"logits/rejected": 0.059291813522577286, |
|
"logps/chosen": -75.47296142578125, |
|
"logps/rejected": -88.63102722167969, |
|
"loss": 2.4789, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -1.10808265209198, |
|
"rewards/margins": 0.3505283296108246, |
|
"rewards/rejected": -1.4586111307144165, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.27974568574023617, |
|
"grad_norm": 1.1496751308441162, |
|
"learning_rate": 9.891625428724365e-06, |
|
"logits/chosen": 0.13897705078125, |
|
"logits/rejected": 0.13928522169589996, |
|
"logps/chosen": -65.34259796142578, |
|
"logps/rejected": -68.99885559082031, |
|
"loss": 2.2804, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.1391716003417969, |
|
"rewards/margins": 0.3395899534225464, |
|
"rewards/rejected": -1.4787613153457642, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.2815622161671208, |
|
"grad_norm": 1.3669211864471436, |
|
"learning_rate": 9.890219166365097e-06, |
|
"logits/chosen": 0.08065556734800339, |
|
"logits/rejected": 0.14106576144695282, |
|
"logps/chosen": -72.62825012207031, |
|
"logps/rejected": -77.62796020507812, |
|
"loss": 2.5929, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0997321605682373, |
|
"rewards/margins": 0.20812571048736572, |
|
"rewards/rejected": -1.3078577518463135, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.28337874659400547, |
|
"grad_norm": 1.2757997512817383, |
|
"learning_rate": 9.888803951584537e-06, |
|
"logits/chosen": 0.0465204194188118, |
|
"logits/rejected": 0.06447532027959824, |
|
"logps/chosen": -66.84536743164062, |
|
"logps/rejected": -78.97770690917969, |
|
"loss": 2.3491, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.1228134632110596, |
|
"rewards/margins": 0.4900767207145691, |
|
"rewards/rejected": -1.612890362739563, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.2851952770208901, |
|
"grad_norm": 1.391892910003662, |
|
"learning_rate": 9.887379787268558e-06, |
|
"logits/chosen": 0.08740498870611191, |
|
"logits/rejected": 0.11772032082080841, |
|
"logps/chosen": -67.31288146972656, |
|
"logps/rejected": -70.06088256835938, |
|
"loss": 2.6179, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.2256510257720947, |
|
"rewards/margins": 0.19449593126773834, |
|
"rewards/rejected": -1.4201467037200928, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.28701180744777477, |
|
"grad_norm": 1.445900321006775, |
|
"learning_rate": 9.885946676321279e-06, |
|
"logits/chosen": 0.1524508148431778, |
|
"logits/rejected": 0.21413244307041168, |
|
"logps/chosen": -74.10218811035156, |
|
"logps/rejected": -76.91876220703125, |
|
"loss": 2.5262, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.1993122100830078, |
|
"rewards/margins": 0.33456599712371826, |
|
"rewards/rejected": -1.5338780879974365, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.2888283378746594, |
|
"grad_norm": 1.361208438873291, |
|
"learning_rate": 9.884504621665059e-06, |
|
"logits/chosen": 0.1192079707980156, |
|
"logits/rejected": 0.21478833258152008, |
|
"logps/chosen": -75.69478607177734, |
|
"logps/rejected": -84.40799713134766, |
|
"loss": 2.3984, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.2067720890045166, |
|
"rewards/margins": 0.34302443265914917, |
|
"rewards/rejected": -1.5497965812683105, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.29064486830154407, |
|
"grad_norm": 3.4042046070098877, |
|
"learning_rate": 9.883053626240503e-06, |
|
"logits/chosen": 0.13823604583740234, |
|
"logits/rejected": 0.0957983061671257, |
|
"logps/chosen": -85.02832794189453, |
|
"logps/rejected": -86.334716796875, |
|
"loss": 3.0326, |
|
"rewards/accuracies": 0.484375, |
|
"rewards/chosen": -1.1793922185897827, |
|
"rewards/margins": 0.013544075191020966, |
|
"rewards/rejected": -1.1929364204406738, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.2924613987284287, |
|
"grad_norm": 1.7824169397354126, |
|
"learning_rate": 9.881593693006438e-06, |
|
"logits/chosen": 0.09778247773647308, |
|
"logits/rejected": 0.12810076773166656, |
|
"logps/chosen": -78.78877258300781, |
|
"logps/rejected": -78.63519287109375, |
|
"loss": 3.1665, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -1.360141396522522, |
|
"rewards/margins": 0.148757204413414, |
|
"rewards/rejected": -1.5088986158370972, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.29427792915531337, |
|
"grad_norm": 1.4275903701782227, |
|
"learning_rate": 9.880124824939927e-06, |
|
"logits/chosen": 0.17759747803211212, |
|
"logits/rejected": 0.14611241221427917, |
|
"logps/chosen": -70.13676452636719, |
|
"logps/rejected": -67.08554077148438, |
|
"loss": 2.8382, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.1756031513214111, |
|
"rewards/margins": 0.14519944787025452, |
|
"rewards/rejected": -1.3208026885986328, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.296094459582198, |
|
"grad_norm": 1.3242970705032349, |
|
"learning_rate": 9.878647025036245e-06, |
|
"logits/chosen": 0.07917390763759613, |
|
"logits/rejected": 0.16502070426940918, |
|
"logps/chosen": -77.08013916015625, |
|
"logps/rejected": -92.47212982177734, |
|
"loss": 2.3183, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.2245774269104004, |
|
"rewards/margins": 0.573800802230835, |
|
"rewards/rejected": -1.7983782291412354, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.29791099000908267, |
|
"grad_norm": 1.160335659980774, |
|
"learning_rate": 9.877160296308886e-06, |
|
"logits/chosen": 0.06782057881355286, |
|
"logits/rejected": 0.09083382785320282, |
|
"logps/chosen": -61.769866943359375, |
|
"logps/rejected": -66.22183990478516, |
|
"loss": 2.3323, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.2067499160766602, |
|
"rewards/margins": 0.36826351284980774, |
|
"rewards/rejected": -1.575013279914856, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.2997275204359673, |
|
"grad_norm": 1.2561142444610596, |
|
"learning_rate": 9.875664641789545e-06, |
|
"logits/chosen": 0.10073137283325195, |
|
"logits/rejected": 0.12745651602745056, |
|
"logps/chosen": -68.94499206542969, |
|
"logps/rejected": -76.13021850585938, |
|
"loss": 2.4058, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.2001702785491943, |
|
"rewards/margins": 0.31210076808929443, |
|
"rewards/rejected": -1.5122709274291992, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.30154405086285196, |
|
"grad_norm": 1.2163783311843872, |
|
"learning_rate": 9.874160064528124e-06, |
|
"logits/chosen": 0.1380203366279602, |
|
"logits/rejected": 0.20783495903015137, |
|
"logps/chosen": -62.44160461425781, |
|
"logps/rejected": -70.53887176513672, |
|
"loss": 2.2759, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.243611216545105, |
|
"rewards/margins": 0.46271997690200806, |
|
"rewards/rejected": -1.7063312530517578, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3033605812897366, |
|
"grad_norm": 1.2518328428268433, |
|
"learning_rate": 9.872646567592719e-06, |
|
"logits/chosen": 0.13933810591697693, |
|
"logits/rejected": 0.1436997652053833, |
|
"logps/chosen": -69.45280456542969, |
|
"logps/rejected": -78.73098754882812, |
|
"loss": 2.368, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1905428171157837, |
|
"rewards/margins": 0.42389774322509766, |
|
"rewards/rejected": -1.6144405603408813, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.30517711171662126, |
|
"grad_norm": 1.2762094736099243, |
|
"learning_rate": 9.871124154069613e-06, |
|
"logits/chosen": 0.11822449415922165, |
|
"logits/rejected": 0.13434451818466187, |
|
"logps/chosen": -72.77944946289062, |
|
"logps/rejected": -75.60210418701172, |
|
"loss": 2.5178, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.304598331451416, |
|
"rewards/margins": 0.21065138280391693, |
|
"rewards/rejected": -1.5152498483657837, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.3069936421435059, |
|
"grad_norm": 1.7200556993484497, |
|
"learning_rate": 9.86959282706327e-06, |
|
"logits/chosen": 0.12234638631343842, |
|
"logits/rejected": 0.123184435069561, |
|
"logps/chosen": -84.78328704833984, |
|
"logps/rejected": -82.90792083740234, |
|
"loss": 2.7762, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.4255175590515137, |
|
"rewards/margins": 0.3239808976650238, |
|
"rewards/rejected": -1.7494984865188599, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.30881017257039056, |
|
"grad_norm": 1.2458773851394653, |
|
"learning_rate": 9.868052589696337e-06, |
|
"logits/chosen": 0.14292597770690918, |
|
"logits/rejected": 0.1542571634054184, |
|
"logps/chosen": -66.72957611083984, |
|
"logps/rejected": -76.97203826904297, |
|
"loss": 2.419, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.2561144828796387, |
|
"rewards/margins": 0.39490899443626404, |
|
"rewards/rejected": -1.6510233879089355, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.3106267029972752, |
|
"grad_norm": 1.192762017250061, |
|
"learning_rate": 9.866503445109621e-06, |
|
"logits/chosen": 0.12845008075237274, |
|
"logits/rejected": 0.10619282722473145, |
|
"logps/chosen": -64.939453125, |
|
"logps/rejected": -70.89356231689453, |
|
"loss": 2.3475, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.2126998901367188, |
|
"rewards/margins": 0.39672625064849854, |
|
"rewards/rejected": -1.6094262599945068, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.31244323342415986, |
|
"grad_norm": 1.6527279615402222, |
|
"learning_rate": 9.864945396462101e-06, |
|
"logits/chosen": 0.03876817971467972, |
|
"logits/rejected": 0.01723310723900795, |
|
"logps/chosen": -80.94947052001953, |
|
"logps/rejected": -83.43637084960938, |
|
"loss": 2.6719, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -1.203782320022583, |
|
"rewards/margins": 0.2551845610141754, |
|
"rewards/rejected": -1.458966851234436, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.3142597638510445, |
|
"grad_norm": 1.427216649055481, |
|
"learning_rate": 9.86337844693091e-06, |
|
"logits/chosen": 0.08040126413106918, |
|
"logits/rejected": 0.13297931849956512, |
|
"logps/chosen": -67.49502563476562, |
|
"logps/rejected": -77.2923812866211, |
|
"loss": 2.4931, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.3284540176391602, |
|
"rewards/margins": 0.4811919927597046, |
|
"rewards/rejected": -1.8096460103988647, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.31607629427792916, |
|
"grad_norm": 1.3090578317642212, |
|
"learning_rate": 9.861802599711329e-06, |
|
"logits/chosen": 0.109119713306427, |
|
"logits/rejected": 0.07613471150398254, |
|
"logps/chosen": -71.72999572753906, |
|
"logps/rejected": -73.84215545654297, |
|
"loss": 2.4972, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.2370662689208984, |
|
"rewards/margins": 0.27422747015953064, |
|
"rewards/rejected": -1.5112937688827515, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.3178928247048138, |
|
"grad_norm": 1.351342797279358, |
|
"learning_rate": 9.860217858016783e-06, |
|
"logits/chosen": 0.10673967003822327, |
|
"logits/rejected": 0.1354019045829773, |
|
"logps/chosen": -70.85772705078125, |
|
"logps/rejected": -81.281982421875, |
|
"loss": 2.4456, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.2566276788711548, |
|
"rewards/margins": 0.34767431020736694, |
|
"rewards/rejected": -1.604301929473877, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.31970935513169846, |
|
"grad_norm": 1.3725523948669434, |
|
"learning_rate": 9.858624225078841e-06, |
|
"logits/chosen": 0.12088489532470703, |
|
"logits/rejected": 0.0771353617310524, |
|
"logps/chosen": -74.0596694946289, |
|
"logps/rejected": -71.7733154296875, |
|
"loss": 2.5003, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1832196712493896, |
|
"rewards/margins": 0.31638142466545105, |
|
"rewards/rejected": -1.4996010065078735, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.3215258855585831, |
|
"grad_norm": 1.3968805074691772, |
|
"learning_rate": 9.857021704147195e-06, |
|
"logits/chosen": 0.11635589599609375, |
|
"logits/rejected": 0.09500478953123093, |
|
"logps/chosen": -75.86662292480469, |
|
"logps/rejected": -75.28397369384766, |
|
"loss": 2.5589, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.3443347215652466, |
|
"rewards/margins": 0.22420868277549744, |
|
"rewards/rejected": -1.5685434341430664, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.32334241598546776, |
|
"grad_norm": 1.3207515478134155, |
|
"learning_rate": 9.855410298489663e-06, |
|
"logits/chosen": 0.032826680690050125, |
|
"logits/rejected": 0.0877794623374939, |
|
"logps/chosen": -67.0349349975586, |
|
"logps/rejected": -75.06578063964844, |
|
"loss": 2.4595, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.2243653535842896, |
|
"rewards/margins": 0.36731529235839844, |
|
"rewards/rejected": -1.591680645942688, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.32515894641235243, |
|
"grad_norm": 1.3098173141479492, |
|
"learning_rate": 9.853790011392186e-06, |
|
"logits/chosen": 0.08737780898809433, |
|
"logits/rejected": 0.08714289963245392, |
|
"logps/chosen": -71.3626708984375, |
|
"logps/rejected": -81.07239532470703, |
|
"loss": 2.2606, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1984796524047852, |
|
"rewards/margins": 0.4800388514995575, |
|
"rewards/rejected": -1.678518533706665, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.32697547683923706, |
|
"grad_norm": 1.3656742572784424, |
|
"learning_rate": 9.852160846158808e-06, |
|
"logits/chosen": 0.08435464650392532, |
|
"logits/rejected": 0.13674329221248627, |
|
"logps/chosen": -69.021484375, |
|
"logps/rejected": -79.02738952636719, |
|
"loss": 2.4039, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.4128899574279785, |
|
"rewards/margins": 0.3882848620414734, |
|
"rewards/rejected": -1.8011748790740967, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.32879200726612173, |
|
"grad_norm": 1.2837380170822144, |
|
"learning_rate": 9.850522806111681e-06, |
|
"logits/chosen": 0.20875662565231323, |
|
"logits/rejected": 0.1668507307767868, |
|
"logps/chosen": -68.29991149902344, |
|
"logps/rejected": -67.59306335449219, |
|
"loss": 2.5716, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -1.3281925916671753, |
|
"rewards/margins": 0.21396151185035706, |
|
"rewards/rejected": -1.54215407371521, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.33060853769300635, |
|
"grad_norm": 1.3804494142532349, |
|
"learning_rate": 9.848875894591055e-06, |
|
"logits/chosen": 0.12790340185165405, |
|
"logits/rejected": 0.1156705766916275, |
|
"logps/chosen": -75.76934814453125, |
|
"logps/rejected": -76.44409942626953, |
|
"loss": 2.2623, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.260763168334961, |
|
"rewards/margins": 0.45292162895202637, |
|
"rewards/rejected": -1.7136849164962769, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.33242506811989103, |
|
"grad_norm": 1.287442922592163, |
|
"learning_rate": 9.847220114955269e-06, |
|
"logits/chosen": 0.1627904772758484, |
|
"logits/rejected": 0.1957186460494995, |
|
"logps/chosen": -67.52108001708984, |
|
"logps/rejected": -79.52376556396484, |
|
"loss": 2.348, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.2697941064834595, |
|
"rewards/margins": 0.5318787097930908, |
|
"rewards/rejected": -1.8016728162765503, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.33424159854677565, |
|
"grad_norm": 1.3078449964523315, |
|
"learning_rate": 9.845555470580746e-06, |
|
"logits/chosen": 0.08075303584337234, |
|
"logits/rejected": 0.08640636503696442, |
|
"logps/chosen": -64.12767791748047, |
|
"logps/rejected": -69.04762268066406, |
|
"loss": 2.4969, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -1.2283066511154175, |
|
"rewards/margins": 0.31595146656036377, |
|
"rewards/rejected": -1.5442581176757812, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.33605812897366033, |
|
"grad_norm": 1.649794578552246, |
|
"learning_rate": 9.843881964861985e-06, |
|
"logits/chosen": 0.07760760188102722, |
|
"logits/rejected": 0.11186876147985458, |
|
"logps/chosen": -75.8207015991211, |
|
"logps/rejected": -81.5796890258789, |
|
"loss": 2.5595, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.4034444093704224, |
|
"rewards/margins": 0.33963268995285034, |
|
"rewards/rejected": -1.743077039718628, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.33787465940054495, |
|
"grad_norm": 1.2912336587905884, |
|
"learning_rate": 9.842199601211556e-06, |
|
"logits/chosen": 0.133261039853096, |
|
"logits/rejected": 0.1473626047372818, |
|
"logps/chosen": -69.50003051757812, |
|
"logps/rejected": -77.0198745727539, |
|
"loss": 2.3544, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.2854596376419067, |
|
"rewards/margins": 0.4699360728263855, |
|
"rewards/rejected": -1.7553956508636475, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.33969118982742963, |
|
"grad_norm": 1.4907958507537842, |
|
"learning_rate": 9.840508383060092e-06, |
|
"logits/chosen": 0.06240752339363098, |
|
"logits/rejected": 0.15645891427993774, |
|
"logps/chosen": -66.49507141113281, |
|
"logps/rejected": -72.68167114257812, |
|
"loss": 2.5754, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.3408069610595703, |
|
"rewards/margins": 0.2825961112976074, |
|
"rewards/rejected": -1.6234029531478882, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.34150772025431425, |
|
"grad_norm": 1.528688669204712, |
|
"learning_rate": 9.838808313856281e-06, |
|
"logits/chosen": -0.010553614236414433, |
|
"logits/rejected": -0.0162151250988245, |
|
"logps/chosen": -87.83446502685547, |
|
"logps/rejected": -82.62303161621094, |
|
"loss": 2.6229, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.2400531768798828, |
|
"rewards/margins": 0.21957488358020782, |
|
"rewards/rejected": -1.4596279859542847, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.34332425068119893, |
|
"grad_norm": 1.5255874395370483, |
|
"learning_rate": 9.83709939706686e-06, |
|
"logits/chosen": 0.12232109159231186, |
|
"logits/rejected": 0.08726370334625244, |
|
"logps/chosen": -66.86070251464844, |
|
"logps/rejected": -70.5438232421875, |
|
"loss": 2.7017, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.433100700378418, |
|
"rewards/margins": 0.20580635964870453, |
|
"rewards/rejected": -1.6389069557189941, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.34514078110808355, |
|
"grad_norm": 1.4504551887512207, |
|
"learning_rate": 9.835381636176604e-06, |
|
"logits/chosen": 0.1462351530790329, |
|
"logits/rejected": 0.13504080474376678, |
|
"logps/chosen": -76.45362091064453, |
|
"logps/rejected": -78.51449584960938, |
|
"loss": 2.5936, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -1.356174349784851, |
|
"rewards/margins": 0.21241840720176697, |
|
"rewards/rejected": -1.5685927867889404, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.3469573115349682, |
|
"grad_norm": 1.4351036548614502, |
|
"learning_rate": 9.833655034688336e-06, |
|
"logits/chosen": 0.1534399539232254, |
|
"logits/rejected": 0.178826704621315, |
|
"logps/chosen": -66.26270294189453, |
|
"logps/rejected": -70.225830078125, |
|
"loss": 2.6442, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.3860033750534058, |
|
"rewards/margins": 0.16981087625026703, |
|
"rewards/rejected": -1.555814266204834, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.34877384196185285, |
|
"grad_norm": 1.295323133468628, |
|
"learning_rate": 9.831919596122888e-06, |
|
"logits/chosen": 0.11593925207853317, |
|
"logits/rejected": 0.19400468468666077, |
|
"logps/chosen": -66.74591064453125, |
|
"logps/rejected": -72.82708740234375, |
|
"loss": 2.3599, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.2114710807800293, |
|
"rewards/margins": 0.34326374530792236, |
|
"rewards/rejected": -1.5547348260879517, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.3505903723887375, |
|
"grad_norm": 1.452673077583313, |
|
"learning_rate": 9.830175324019125e-06, |
|
"logits/chosen": 0.13779595494270325, |
|
"logits/rejected": 0.15601256489753723, |
|
"logps/chosen": -73.34432220458984, |
|
"logps/rejected": -76.34349822998047, |
|
"loss": 2.4369, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.409874677658081, |
|
"rewards/margins": 0.3434732258319855, |
|
"rewards/rejected": -1.7533478736877441, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.35240690281562215, |
|
"grad_norm": 1.3285003900527954, |
|
"learning_rate": 9.828422221933924e-06, |
|
"logits/chosen": 0.020087052136659622, |
|
"logits/rejected": 0.07995946705341339, |
|
"logps/chosen": -72.9058837890625, |
|
"logps/rejected": -81.98945617675781, |
|
"loss": 2.3612, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.2450451850891113, |
|
"rewards/margins": 0.40142494440078735, |
|
"rewards/rejected": -1.646470069885254, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.3542234332425068, |
|
"grad_norm": 1.2274519205093384, |
|
"learning_rate": 9.826660293442158e-06, |
|
"logits/chosen": 0.0241906326264143, |
|
"logits/rejected": 0.09953958541154861, |
|
"logps/chosen": -66.5189437866211, |
|
"logps/rejected": -75.99092102050781, |
|
"loss": 2.2288, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.3973838090896606, |
|
"rewards/margins": 0.5354034304618835, |
|
"rewards/rejected": -1.9327871799468994, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.35603996366939145, |
|
"grad_norm": 1.3833938837051392, |
|
"learning_rate": 9.824889542136714e-06, |
|
"logits/chosen": 0.086525097489357, |
|
"logits/rejected": 0.08451628684997559, |
|
"logps/chosen": -74.15232849121094, |
|
"logps/rejected": -79.37950897216797, |
|
"loss": 2.4671, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.4214783906936646, |
|
"rewards/margins": 0.3752206265926361, |
|
"rewards/rejected": -1.7966989278793335, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.3578564940962761, |
|
"grad_norm": 1.6680957078933716, |
|
"learning_rate": 9.823109971628459e-06, |
|
"logits/chosen": 0.06370481848716736, |
|
"logits/rejected": 0.12739142775535583, |
|
"logps/chosen": -75.81684875488281, |
|
"logps/rejected": -76.65689086914062, |
|
"loss": 2.3705, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.3144075870513916, |
|
"rewards/margins": 0.36334556341171265, |
|
"rewards/rejected": -1.6777533292770386, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.35967302452316074, |
|
"grad_norm": 1.4894323348999023, |
|
"learning_rate": 9.821321585546244e-06, |
|
"logits/chosen": 0.10617184638977051, |
|
"logits/rejected": 0.11465627700090408, |
|
"logps/chosen": -73.0533218383789, |
|
"logps/rejected": -83.94466400146484, |
|
"loss": 2.2915, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.4909334182739258, |
|
"rewards/margins": 0.5887378454208374, |
|
"rewards/rejected": -2.0796711444854736, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.3614895549500454, |
|
"grad_norm": 1.422853946685791, |
|
"learning_rate": 9.819524387536905e-06, |
|
"logits/chosen": 0.069038525223732, |
|
"logits/rejected": 0.09677774459123611, |
|
"logps/chosen": -85.49876403808594, |
|
"logps/rejected": -89.743408203125, |
|
"loss": 2.3, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.3645334243774414, |
|
"rewards/margins": 0.5102630853652954, |
|
"rewards/rejected": -1.8747965097427368, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.36330608537693004, |
|
"grad_norm": 1.3350121974945068, |
|
"learning_rate": 9.81771838126524e-06, |
|
"logits/chosen": 0.0018447795882821083, |
|
"logits/rejected": 0.055721428245306015, |
|
"logps/chosen": -71.9021224975586, |
|
"logps/rejected": -81.6389389038086, |
|
"loss": 2.2872, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.3360776901245117, |
|
"rewards/margins": 0.46150827407836914, |
|
"rewards/rejected": -1.7975859642028809, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3651226158038147, |
|
"grad_norm": 1.458803653717041, |
|
"learning_rate": 9.815903570414006e-06, |
|
"logits/chosen": 0.059184275567531586, |
|
"logits/rejected": 0.07923795282840729, |
|
"logps/chosen": -80.88011932373047, |
|
"logps/rejected": -86.74174499511719, |
|
"loss": 2.4279, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.6620349884033203, |
|
"rewards/margins": 0.38666611909866333, |
|
"rewards/rejected": -2.048701047897339, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.36693914623069934, |
|
"grad_norm": 1.3583872318267822, |
|
"learning_rate": 9.814079958683925e-06, |
|
"logits/chosen": 0.11471173167228699, |
|
"logits/rejected": 0.1523369550704956, |
|
"logps/chosen": -71.39076232910156, |
|
"logps/rejected": -78.50016021728516, |
|
"loss": 2.4603, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.4875473976135254, |
|
"rewards/margins": 0.3805108070373535, |
|
"rewards/rejected": -1.8680580854415894, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.368755676657584, |
|
"grad_norm": 1.6755985021591187, |
|
"learning_rate": 9.812247549793656e-06, |
|
"logits/chosen": 0.14959998428821564, |
|
"logits/rejected": 0.1801329255104065, |
|
"logps/chosen": -76.0824203491211, |
|
"logps/rejected": -85.81806945800781, |
|
"loss": 2.8134, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.7176380157470703, |
|
"rewards/margins": 0.28729167580604553, |
|
"rewards/rejected": -2.004929780960083, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.37057220708446864, |
|
"grad_norm": 1.4551233053207397, |
|
"learning_rate": 9.810406347479798e-06, |
|
"logits/chosen": 0.08063512295484543, |
|
"logits/rejected": 0.03579093888401985, |
|
"logps/chosen": -87.19414520263672, |
|
"logps/rejected": -89.355224609375, |
|
"loss": 2.4092, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.5363514423370361, |
|
"rewards/margins": 0.4680227041244507, |
|
"rewards/rejected": -2.0043740272521973, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.3723887375113533, |
|
"grad_norm": 1.4248968362808228, |
|
"learning_rate": 9.808556355496885e-06, |
|
"logits/chosen": 0.06655821204185486, |
|
"logits/rejected": 0.050458114594221115, |
|
"logps/chosen": -92.04095458984375, |
|
"logps/rejected": -95.39706420898438, |
|
"loss": 2.3507, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.5707283020019531, |
|
"rewards/margins": 0.388569712638855, |
|
"rewards/rejected": -1.959298014640808, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.37420526793823794, |
|
"grad_norm": 1.3779215812683105, |
|
"learning_rate": 9.806697577617371e-06, |
|
"logits/chosen": 0.09702017903327942, |
|
"logits/rejected": 0.13923662900924683, |
|
"logps/chosen": -78.27027893066406, |
|
"logps/rejected": -85.93509674072266, |
|
"loss": 2.1888, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.4591946601867676, |
|
"rewards/margins": 0.5624303817749023, |
|
"rewards/rejected": -2.02162504196167, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.3760217983651226, |
|
"grad_norm": 1.5306010246276855, |
|
"learning_rate": 9.804830017631631e-06, |
|
"logits/chosen": 0.038323137909173965, |
|
"logits/rejected": 0.04337020218372345, |
|
"logps/chosen": -75.31045532226562, |
|
"logps/rejected": -83.26338958740234, |
|
"loss": 2.7312, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.4613063335418701, |
|
"rewards/margins": 0.3355714678764343, |
|
"rewards/rejected": -1.7968778610229492, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.37783832879200724, |
|
"grad_norm": 1.629341959953308, |
|
"learning_rate": 9.802953679347943e-06, |
|
"logits/chosen": 0.1168309897184372, |
|
"logits/rejected": 0.21053184568881989, |
|
"logps/chosen": -70.42684936523438, |
|
"logps/rejected": -87.77594757080078, |
|
"loss": 2.7324, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.8266733884811401, |
|
"rewards/margins": 0.5579056739807129, |
|
"rewards/rejected": -2.3845791816711426, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.3796548592188919, |
|
"grad_norm": 1.5751092433929443, |
|
"learning_rate": 9.801068566592486e-06, |
|
"logits/chosen": 0.11355097591876984, |
|
"logits/rejected": 0.11962890625, |
|
"logps/chosen": -87.75645446777344, |
|
"logps/rejected": -96.15601348876953, |
|
"loss": 2.5237, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.4331413507461548, |
|
"rewards/margins": 0.38925477862358093, |
|
"rewards/rejected": -1.8223960399627686, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.3814713896457766, |
|
"grad_norm": 1.6638984680175781, |
|
"learning_rate": 9.799174683209336e-06, |
|
"logits/chosen": 0.0970507487654686, |
|
"logits/rejected": 0.07422082126140594, |
|
"logps/chosen": -85.44343566894531, |
|
"logps/rejected": -90.49578857421875, |
|
"loss": 2.5179, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.6639773845672607, |
|
"rewards/margins": 0.4513780176639557, |
|
"rewards/rejected": -2.1153552532196045, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3832879200726612, |
|
"grad_norm": 1.389356017112732, |
|
"learning_rate": 9.79727203306045e-06, |
|
"logits/chosen": 0.027732742950320244, |
|
"logits/rejected": 0.0795917734503746, |
|
"logps/chosen": -85.95429229736328, |
|
"logps/rejected": -92.91676330566406, |
|
"loss": 2.3892, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.4551334381103516, |
|
"rewards/margins": 0.4000782370567322, |
|
"rewards/rejected": -1.8552117347717285, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.3851044504995459, |
|
"grad_norm": 1.438284993171692, |
|
"learning_rate": 9.79536062002566e-06, |
|
"logits/chosen": 0.04227686673402786, |
|
"logits/rejected": 0.09655077010393143, |
|
"logps/chosen": -77.67405700683594, |
|
"logps/rejected": -88.0655746459961, |
|
"loss": 2.3246, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.5603740215301514, |
|
"rewards/margins": 0.5005100965499878, |
|
"rewards/rejected": -2.060884475708008, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.3869209809264305, |
|
"grad_norm": 1.35500168800354, |
|
"learning_rate": 9.793440448002676e-06, |
|
"logits/chosen": 0.10956872254610062, |
|
"logits/rejected": 0.10562983900308609, |
|
"logps/chosen": -74.20452117919922, |
|
"logps/rejected": -73.23489379882812, |
|
"loss": 2.4585, |
|
"rewards/accuracies": 0.796875, |
|
"rewards/chosen": -1.399983286857605, |
|
"rewards/margins": 0.2682442367076874, |
|
"rewards/rejected": -1.6682274341583252, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.3887375113533152, |
|
"grad_norm": 1.3036773204803467, |
|
"learning_rate": 9.791511520907056e-06, |
|
"logits/chosen": 0.06179399788379669, |
|
"logits/rejected": 0.052528850734233856, |
|
"logps/chosen": -72.4264144897461, |
|
"logps/rejected": -72.77079010009766, |
|
"loss": 2.3441, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.391230821609497, |
|
"rewards/margins": 0.3969228267669678, |
|
"rewards/rejected": -1.7881536483764648, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.3905540417801998, |
|
"grad_norm": 1.4877872467041016, |
|
"learning_rate": 9.789573842672223e-06, |
|
"logits/chosen": 0.05735350027680397, |
|
"logits/rejected": 0.09979183971881866, |
|
"logps/chosen": -82.134033203125, |
|
"logps/rejected": -96.38580322265625, |
|
"loss": 2.2192, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7035160064697266, |
|
"rewards/margins": 0.6622112989425659, |
|
"rewards/rejected": -2.365727424621582, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.3923705722070845, |
|
"grad_norm": 1.247817039489746, |
|
"learning_rate": 9.787627417249441e-06, |
|
"logits/chosen": 0.0014538783580064774, |
|
"logits/rejected": 0.03942735865712166, |
|
"logps/chosen": -73.8500747680664, |
|
"logps/rejected": -95.03350067138672, |
|
"loss": 1.9498, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.4928035736083984, |
|
"rewards/margins": 0.9987993836402893, |
|
"rewards/rejected": -2.491603374481201, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.3941871026339691, |
|
"grad_norm": 1.3884788751602173, |
|
"learning_rate": 9.785672248607807e-06, |
|
"logits/chosen": 0.029594585299491882, |
|
"logits/rejected": 0.11805769056081772, |
|
"logps/chosen": -66.9087905883789, |
|
"logps/rejected": -80.82840728759766, |
|
"loss": 2.2455, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.6260508298873901, |
|
"rewards/margins": 0.5117952227592468, |
|
"rewards/rejected": -2.137845993041992, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.3960036330608538, |
|
"grad_norm": 2.9690237045288086, |
|
"learning_rate": 9.78370834073425e-06, |
|
"logits/chosen": 0.11608768254518509, |
|
"logits/rejected": 0.11099248379468918, |
|
"logps/chosen": -73.23339080810547, |
|
"logps/rejected": -73.60865783691406, |
|
"loss": 2.6526, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.5958513021469116, |
|
"rewards/margins": 0.20439797639846802, |
|
"rewards/rejected": -1.8002492189407349, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.3978201634877384, |
|
"grad_norm": 1.4823459386825562, |
|
"learning_rate": 9.781735697633526e-06, |
|
"logits/chosen": 0.07910319417715073, |
|
"logits/rejected": 0.1617601215839386, |
|
"logps/chosen": -72.02169799804688, |
|
"logps/rejected": -79.70098114013672, |
|
"loss": 2.3689, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.722070336341858, |
|
"rewards/margins": 0.4829583168029785, |
|
"rewards/rejected": -2.205028772354126, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.3996366939146231, |
|
"grad_norm": 2.249466896057129, |
|
"learning_rate": 9.779754323328192e-06, |
|
"logits/chosen": 0.17587795853614807, |
|
"logits/rejected": 0.13541430234909058, |
|
"logps/chosen": -75.69686889648438, |
|
"logps/rejected": -77.53881072998047, |
|
"loss": 3.2204, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -2.03781795501709, |
|
"rewards/margins": 0.1416795551776886, |
|
"rewards/rejected": -2.179497480392456, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4014532243415077, |
|
"grad_norm": 1.4283430576324463, |
|
"learning_rate": 9.777764221858616e-06, |
|
"logits/chosen": 0.1529032289981842, |
|
"logits/rejected": 0.1623322069644928, |
|
"logps/chosen": -69.98622131347656, |
|
"logps/rejected": -77.31781005859375, |
|
"loss": 2.3358, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.6996090412139893, |
|
"rewards/margins": 0.568400502204895, |
|
"rewards/rejected": -2.2680094242095947, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.4032697547683924, |
|
"grad_norm": 1.4971157312393188, |
|
"learning_rate": 9.775765397282963e-06, |
|
"logits/chosen": 0.13248610496520996, |
|
"logits/rejected": 0.13485054671764374, |
|
"logps/chosen": -70.01846313476562, |
|
"logps/rejected": -75.87899017333984, |
|
"loss": 2.4037, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.6077332496643066, |
|
"rewards/margins": 0.39039328694343567, |
|
"rewards/rejected": -1.99812650680542, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.405086285195277, |
|
"grad_norm": 1.590364694595337, |
|
"learning_rate": 9.773757853677182e-06, |
|
"logits/chosen": 0.08200166374444962, |
|
"logits/rejected": 0.06919535249471664, |
|
"logps/chosen": -85.51278686523438, |
|
"logps/rejected": -89.3447265625, |
|
"loss": 2.6149, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.5626994371414185, |
|
"rewards/margins": 0.30814388394355774, |
|
"rewards/rejected": -1.8708434104919434, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.4069028156221617, |
|
"grad_norm": 1.41611647605896, |
|
"learning_rate": 9.771741595135009e-06, |
|
"logits/chosen": 0.057254984974861145, |
|
"logits/rejected": 0.13574014604091644, |
|
"logps/chosen": -73.97720336914062, |
|
"logps/rejected": -84.72975158691406, |
|
"loss": 2.3215, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.6877132654190063, |
|
"rewards/margins": 0.4473133981227875, |
|
"rewards/rejected": -2.135026693344116, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.4087193460490463, |
|
"grad_norm": 1.7008063793182373, |
|
"learning_rate": 9.769716625767939e-06, |
|
"logits/chosen": 0.05822606012225151, |
|
"logits/rejected": 0.06510132551193237, |
|
"logps/chosen": -81.26387023925781, |
|
"logps/rejected": -82.76637268066406, |
|
"loss": 2.5366, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.739980697631836, |
|
"rewards/margins": 0.3888433277606964, |
|
"rewards/rejected": -2.12882399559021, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.410535876475931, |
|
"grad_norm": 1.3915003538131714, |
|
"learning_rate": 9.767682949705243e-06, |
|
"logits/chosen": 0.08782866597175598, |
|
"logits/rejected": 0.17832686007022858, |
|
"logps/chosen": -67.25590515136719, |
|
"logps/rejected": -78.19799041748047, |
|
"loss": 2.3992, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.6588337421417236, |
|
"rewards/margins": 0.4392687976360321, |
|
"rewards/rejected": -2.098102331161499, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.4123524069028156, |
|
"grad_norm": 1.4107364416122437, |
|
"learning_rate": 9.765640571093938e-06, |
|
"logits/chosen": 0.14615394175052643, |
|
"logits/rejected": 0.14398689568042755, |
|
"logps/chosen": -66.720703125, |
|
"logps/rejected": -72.72846221923828, |
|
"loss": 2.488, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.740675449371338, |
|
"rewards/margins": 0.34203463792800903, |
|
"rewards/rejected": -2.0827102661132812, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.4141689373297003, |
|
"grad_norm": 1.438272476196289, |
|
"learning_rate": 9.76358949409879e-06, |
|
"logits/chosen": 0.1331941783428192, |
|
"logits/rejected": 0.18831086158752441, |
|
"logps/chosen": -77.92586517333984, |
|
"logps/rejected": -81.89257049560547, |
|
"loss": 2.466, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.7596431970596313, |
|
"rewards/margins": 0.3160095512866974, |
|
"rewards/rejected": -2.075652837753296, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.4159854677565849, |
|
"grad_norm": 1.3581331968307495, |
|
"learning_rate": 9.7615297229023e-06, |
|
"logits/chosen": 0.13822412490844727, |
|
"logits/rejected": 0.14220967888832092, |
|
"logps/chosen": -64.66896057128906, |
|
"logps/rejected": -76.72779846191406, |
|
"loss": 2.282, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.5948115587234497, |
|
"rewards/margins": 0.5508276224136353, |
|
"rewards/rejected": -2.145639181137085, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.4178019981834696, |
|
"grad_norm": 1.4766261577606201, |
|
"learning_rate": 9.759461261704705e-06, |
|
"logits/chosen": 0.06772036850452423, |
|
"logits/rejected": 0.1212601587176323, |
|
"logps/chosen": -75.17322540283203, |
|
"logps/rejected": -86.30448913574219, |
|
"loss": 2.1345, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.6930228471755981, |
|
"rewards/margins": 0.607439398765564, |
|
"rewards/rejected": -2.300462245941162, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4196185286103542, |
|
"grad_norm": 1.4598506689071655, |
|
"learning_rate": 9.757384114723954e-06, |
|
"logits/chosen": 0.11245124042034149, |
|
"logits/rejected": 0.17101560533046722, |
|
"logps/chosen": -75.50772094726562, |
|
"logps/rejected": -83.95561218261719, |
|
"loss": 2.3606, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.7730398178100586, |
|
"rewards/margins": 0.4058808386325836, |
|
"rewards/rejected": -2.1789209842681885, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.4214350590372389, |
|
"grad_norm": 1.5139139890670776, |
|
"learning_rate": 9.755298286195712e-06, |
|
"logits/chosen": 0.05479501932859421, |
|
"logits/rejected": 0.12558911740779877, |
|
"logps/chosen": -81.28207397460938, |
|
"logps/rejected": -86.58873748779297, |
|
"loss": 2.4171, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.6220098733901978, |
|
"rewards/margins": 0.49352455139160156, |
|
"rewards/rejected": -2.1155343055725098, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.4232515894641235, |
|
"grad_norm": 1.6987231969833374, |
|
"learning_rate": 9.753203780373348e-06, |
|
"logits/chosen": 0.07917524874210358, |
|
"logits/rejected": 0.14443910121917725, |
|
"logps/chosen": -81.83525848388672, |
|
"logps/rejected": -77.00749206542969, |
|
"loss": 2.7548, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7713696956634521, |
|
"rewards/margins": 0.20764464139938354, |
|
"rewards/rejected": -1.9790143966674805, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.4250681198910082, |
|
"grad_norm": 1.6056586503982544, |
|
"learning_rate": 9.751100601527922e-06, |
|
"logits/chosen": 0.12120751291513443, |
|
"logits/rejected": 0.21893832087516785, |
|
"logps/chosen": -73.1818618774414, |
|
"logps/rejected": -84.82020568847656, |
|
"loss": 2.4369, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.8145970106124878, |
|
"rewards/margins": 0.5701674818992615, |
|
"rewards/rejected": -2.3847644329071045, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.4268846503178928, |
|
"grad_norm": 1.4865500926971436, |
|
"learning_rate": 9.748988753948183e-06, |
|
"logits/chosen": 0.0659053698182106, |
|
"logits/rejected": 0.09985598176717758, |
|
"logps/chosen": -83.79960632324219, |
|
"logps/rejected": -87.37390899658203, |
|
"loss": 2.3437, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8071691989898682, |
|
"rewards/margins": 0.4156024754047394, |
|
"rewards/rejected": -2.2227721214294434, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.4287011807447775, |
|
"grad_norm": 1.7585958242416382, |
|
"learning_rate": 9.746868241940554e-06, |
|
"logits/chosen": 0.13636741042137146, |
|
"logits/rejected": 0.16040681302547455, |
|
"logps/chosen": -70.46460723876953, |
|
"logps/rejected": -71.19532012939453, |
|
"loss": 2.6303, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.7820823192596436, |
|
"rewards/margins": 0.37900543212890625, |
|
"rewards/rejected": -2.16108775138855, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.4305177111716621, |
|
"grad_norm": 1.4949400424957275, |
|
"learning_rate": 9.744739069829132e-06, |
|
"logits/chosen": 0.16385243833065033, |
|
"logits/rejected": 0.13090217113494873, |
|
"logps/chosen": -74.92865753173828, |
|
"logps/rejected": -78.68392181396484, |
|
"loss": 2.3038, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.847611904144287, |
|
"rewards/margins": 0.561863899230957, |
|
"rewards/rejected": -2.409475564956665, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.4323342415985468, |
|
"grad_norm": 1.6149885654449463, |
|
"learning_rate": 9.742601241955666e-06, |
|
"logits/chosen": 0.10731178522109985, |
|
"logits/rejected": 0.11118797957897186, |
|
"logps/chosen": -87.08879089355469, |
|
"logps/rejected": -90.8418960571289, |
|
"loss": 2.2981, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.7146490812301636, |
|
"rewards/margins": 0.4193970859050751, |
|
"rewards/rejected": -2.1340463161468506, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.43415077202543145, |
|
"grad_norm": 1.4300076961517334, |
|
"learning_rate": 9.740454762679562e-06, |
|
"logits/chosen": 0.031154140830039978, |
|
"logits/rejected": 0.13953763246536255, |
|
"logps/chosen": -68.21051025390625, |
|
"logps/rejected": -87.66563415527344, |
|
"loss": 2.0885, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.6824418306350708, |
|
"rewards/margins": 0.6189282536506653, |
|
"rewards/rejected": -2.3013699054718018, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.4359673024523161, |
|
"grad_norm": 1.3545743227005005, |
|
"learning_rate": 9.738299636377863e-06, |
|
"logits/chosen": 0.1132105141878128, |
|
"logits/rejected": 0.1149899885058403, |
|
"logps/chosen": -75.41773223876953, |
|
"logps/rejected": -80.03020477294922, |
|
"loss": 2.3774, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.5836522579193115, |
|
"rewards/margins": 0.40351229906082153, |
|
"rewards/rejected": -1.9871646165847778, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.43778383287920075, |
|
"grad_norm": 1.2866510152816772, |
|
"learning_rate": 9.736135867445246e-06, |
|
"logits/chosen": 0.08788580447435379, |
|
"logits/rejected": 0.15736152231693268, |
|
"logps/chosen": -70.54780578613281, |
|
"logps/rejected": -86.70913696289062, |
|
"loss": 1.9408, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.7565429210662842, |
|
"rewards/margins": 0.7674389481544495, |
|
"rewards/rejected": -2.523982048034668, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.4396003633060854, |
|
"grad_norm": 1.5900717973709106, |
|
"learning_rate": 9.733963460294016e-06, |
|
"logits/chosen": 0.08901432901620865, |
|
"logits/rejected": 0.09095099568367004, |
|
"logps/chosen": -77.32083129882812, |
|
"logps/rejected": -83.0787124633789, |
|
"loss": 2.4779, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.9288290739059448, |
|
"rewards/margins": 0.3023950755596161, |
|
"rewards/rejected": -2.2312240600585938, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.44141689373297005, |
|
"grad_norm": 1.664976716041565, |
|
"learning_rate": 9.731782419354087e-06, |
|
"logits/chosen": 0.0038250258658081293, |
|
"logits/rejected": 0.05805446207523346, |
|
"logps/chosen": -75.7310562133789, |
|
"logps/rejected": -81.24979400634766, |
|
"loss": 2.4426, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.8681721687316895, |
|
"rewards/margins": 0.41048479080200195, |
|
"rewards/rejected": -2.2786567211151123, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.44323342415985467, |
|
"grad_norm": 1.688614845275879, |
|
"learning_rate": 9.729592749072981e-06, |
|
"logits/chosen": 0.10514964163303375, |
|
"logits/rejected": 0.08623237907886505, |
|
"logps/chosen": -78.1123046875, |
|
"logps/rejected": -85.77177429199219, |
|
"loss": 2.4137, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.709516167640686, |
|
"rewards/margins": 0.477593332529068, |
|
"rewards/rejected": -2.1871094703674316, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.44504995458673935, |
|
"grad_norm": 1.603507399559021, |
|
"learning_rate": 9.727394453915817e-06, |
|
"logits/chosen": 0.06938113272190094, |
|
"logits/rejected": 0.10225434601306915, |
|
"logps/chosen": -72.41216278076172, |
|
"logps/rejected": -82.71170043945312, |
|
"loss": 2.3143, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.8322726488113403, |
|
"rewards/margins": 0.3731868863105774, |
|
"rewards/rejected": -2.2054593563079834, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.44686648501362397, |
|
"grad_norm": 1.6047879457473755, |
|
"learning_rate": 9.725187538365304e-06, |
|
"logits/chosen": 0.11169447004795074, |
|
"logits/rejected": 0.14944276213645935, |
|
"logps/chosen": -75.95654296875, |
|
"logps/rejected": -87.93280029296875, |
|
"loss": 2.3278, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.9458414316177368, |
|
"rewards/margins": 0.5952720046043396, |
|
"rewards/rejected": -2.5411133766174316, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.44868301544050865, |
|
"grad_norm": 1.5822384357452393, |
|
"learning_rate": 9.722972006921725e-06, |
|
"logits/chosen": 0.07633841782808304, |
|
"logits/rejected": 0.13307756185531616, |
|
"logps/chosen": -82.36216735839844, |
|
"logps/rejected": -89.472900390625, |
|
"loss": 2.3939, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -2.026167154312134, |
|
"rewards/margins": 0.4467831552028656, |
|
"rewards/rejected": -2.4729504585266113, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.45049954586739327, |
|
"grad_norm": 1.636837363243103, |
|
"learning_rate": 9.720747864102935e-06, |
|
"logits/chosen": 0.013166696764528751, |
|
"logits/rejected": 0.10926786065101624, |
|
"logps/chosen": -81.95793151855469, |
|
"logps/rejected": -95.96204376220703, |
|
"loss": 2.4935, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -2.1589713096618652, |
|
"rewards/margins": 0.4789046049118042, |
|
"rewards/rejected": -2.637876033782959, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.45231607629427795, |
|
"grad_norm": 1.6743932962417603, |
|
"learning_rate": 9.718515114444347e-06, |
|
"logits/chosen": 0.09870442003011703, |
|
"logits/rejected": 0.14206278324127197, |
|
"logps/chosen": -81.16842651367188, |
|
"logps/rejected": -95.32533264160156, |
|
"loss": 2.2988, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.291147470474243, |
|
"rewards/margins": 0.7547516822814941, |
|
"rewards/rejected": -3.0458991527557373, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.45413260672116257, |
|
"grad_norm": 1.4136468172073364, |
|
"learning_rate": 9.716273762498929e-06, |
|
"logits/chosen": 0.11840160191059113, |
|
"logits/rejected": 0.14921030402183533, |
|
"logps/chosen": -70.83485412597656, |
|
"logps/rejected": -73.80975341796875, |
|
"loss": 2.4975, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.8664369583129883, |
|
"rewards/margins": 0.32539018988609314, |
|
"rewards/rejected": -2.1918272972106934, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.45594913714804725, |
|
"grad_norm": 1.771864652633667, |
|
"learning_rate": 9.714023812837185e-06, |
|
"logits/chosen": 0.04426509141921997, |
|
"logits/rejected": 0.08872814476490021, |
|
"logps/chosen": -82.2516098022461, |
|
"logps/rejected": -87.78474426269531, |
|
"loss": 2.627, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.2238872051239014, |
|
"rewards/margins": 0.48393067717552185, |
|
"rewards/rejected": -2.707818031311035, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.45776566757493187, |
|
"grad_norm": 1.4489670991897583, |
|
"learning_rate": 9.711765270047155e-06, |
|
"logits/chosen": 0.025218207389116287, |
|
"logits/rejected": 0.09768272191286087, |
|
"logps/chosen": -70.79837036132812, |
|
"logps/rejected": -86.96686553955078, |
|
"loss": 2.1641, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8761096000671387, |
|
"rewards/margins": 0.8277947306632996, |
|
"rewards/rejected": -2.703904151916504, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.45958219800181654, |
|
"grad_norm": 1.7034775018692017, |
|
"learning_rate": 9.709498138734405e-06, |
|
"logits/chosen": 0.04030866175889969, |
|
"logits/rejected": 0.08729197829961777, |
|
"logps/chosen": -81.70675659179688, |
|
"logps/rejected": -85.92189025878906, |
|
"loss": 2.58, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -2.0707144737243652, |
|
"rewards/margins": 0.3897002637386322, |
|
"rewards/rejected": -2.4604148864746094, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.46139872842870117, |
|
"grad_norm": 1.8713371753692627, |
|
"learning_rate": 9.707222423522004e-06, |
|
"logits/chosen": 0.047953542321920395, |
|
"logits/rejected": 0.02678016573190689, |
|
"logps/chosen": -84.65204620361328, |
|
"logps/rejected": -90.09396362304688, |
|
"loss": 2.7899, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -2.103985071182251, |
|
"rewards/margins": 0.22625797986984253, |
|
"rewards/rejected": -2.330242872238159, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.46321525885558584, |
|
"grad_norm": 1.4706422090530396, |
|
"learning_rate": 9.704938129050535e-06, |
|
"logits/chosen": 0.04734738916158676, |
|
"logits/rejected": 0.11658424139022827, |
|
"logps/chosen": -78.08146667480469, |
|
"logps/rejected": -96.80207061767578, |
|
"loss": 2.1796, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9245433807373047, |
|
"rewards/margins": 0.7155373096466064, |
|
"rewards/rejected": -2.640080451965332, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.46503178928247046, |
|
"grad_norm": 1.4902199506759644, |
|
"learning_rate": 9.702645259978072e-06, |
|
"logits/chosen": 0.09310627728700638, |
|
"logits/rejected": 0.1795577108860016, |
|
"logps/chosen": -78.96179962158203, |
|
"logps/rejected": -84.86495208740234, |
|
"loss": 2.2113, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -2.0413970947265625, |
|
"rewards/margins": 0.4318653345108032, |
|
"rewards/rejected": -2.4732625484466553, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.46684831970935514, |
|
"grad_norm": 1.3563650846481323, |
|
"learning_rate": 9.700343820980172e-06, |
|
"logits/chosen": 0.08617695420980453, |
|
"logits/rejected": 0.1092975065112114, |
|
"logps/chosen": -76.31070709228516, |
|
"logps/rejected": -82.52798461914062, |
|
"loss": 2.3793, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.7843788862228394, |
|
"rewards/margins": 0.5503235459327698, |
|
"rewards/rejected": -2.334702491760254, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.46866485013623976, |
|
"grad_norm": 1.353379249572754, |
|
"learning_rate": 9.698033816749874e-06, |
|
"logits/chosen": 0.07846446335315704, |
|
"logits/rejected": 0.15949462354183197, |
|
"logps/chosen": -70.9232177734375, |
|
"logps/rejected": -83.97347259521484, |
|
"loss": 2.0527, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": -1.7502186298370361, |
|
"rewards/margins": 0.6754422783851624, |
|
"rewards/rejected": -2.4256608486175537, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.47048138056312444, |
|
"grad_norm": 1.2879610061645508, |
|
"learning_rate": 9.695715251997676e-06, |
|
"logits/chosen": 0.041364431381225586, |
|
"logits/rejected": 0.2137874960899353, |
|
"logps/chosen": -74.40650177001953, |
|
"logps/rejected": -99.26273345947266, |
|
"loss": 1.8381, |
|
"rewards/accuracies": 0.859375, |
|
"rewards/chosen": -1.7944194078445435, |
|
"rewards/margins": 1.1633189916610718, |
|
"rewards/rejected": -2.9577386379241943, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.47229791099000906, |
|
"grad_norm": 1.5336003303527832, |
|
"learning_rate": 9.693388131451536e-06, |
|
"logits/chosen": 0.10320600867271423, |
|
"logits/rejected": 0.20043231546878815, |
|
"logps/chosen": -70.24287414550781, |
|
"logps/rejected": -87.45311737060547, |
|
"loss": 2.2716, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8973060846328735, |
|
"rewards/margins": 0.7614614963531494, |
|
"rewards/rejected": -2.6587674617767334, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.47411444141689374, |
|
"grad_norm": 1.50228750705719, |
|
"learning_rate": 9.691052459856858e-06, |
|
"logits/chosen": 0.10719013214111328, |
|
"logits/rejected": 0.11425416171550751, |
|
"logps/chosen": -77.69039154052734, |
|
"logps/rejected": -83.77989196777344, |
|
"loss": 2.3255, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7352136373519897, |
|
"rewards/margins": 0.588873565196991, |
|
"rewards/rejected": -2.324087142944336, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.47593097184377836, |
|
"grad_norm": 1.6516278982162476, |
|
"learning_rate": 9.688708241976484e-06, |
|
"logits/chosen": 0.0962933823466301, |
|
"logits/rejected": 0.09487868845462799, |
|
"logps/chosen": -83.95497131347656, |
|
"logps/rejected": -86.93241882324219, |
|
"loss": 2.5126, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.0139267444610596, |
|
"rewards/margins": 0.4219791889190674, |
|
"rewards/rejected": -2.435905694961548, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.47774750227066304, |
|
"grad_norm": 1.4588408470153809, |
|
"learning_rate": 9.686355482590679e-06, |
|
"logits/chosen": 0.04918690025806427, |
|
"logits/rejected": 0.1662694364786148, |
|
"logps/chosen": -76.76988983154297, |
|
"logps/rejected": -98.94943237304688, |
|
"loss": 2.0243, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": -1.8614740371704102, |
|
"rewards/margins": 0.9306913614273071, |
|
"rewards/rejected": -2.7921652793884277, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.47956403269754766, |
|
"grad_norm": 1.8992609977722168, |
|
"learning_rate": 9.683994186497132e-06, |
|
"logits/chosen": 0.039469510316848755, |
|
"logits/rejected": 0.11771678924560547, |
|
"logps/chosen": -78.4907455444336, |
|
"logps/rejected": -97.9975814819336, |
|
"loss": 2.6106, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": -2.1690123081207275, |
|
"rewards/margins": 0.9005274772644043, |
|
"rewards/rejected": -3.069540023803711, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.48138056312443234, |
|
"grad_norm": 2.196244239807129, |
|
"learning_rate": 9.681624358510936e-06, |
|
"logits/chosen": 0.11924441158771515, |
|
"logits/rejected": 0.08125054091215134, |
|
"logps/chosen": -67.94137573242188, |
|
"logps/rejected": -70.6710433959961, |
|
"loss": 2.1509, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.787060022354126, |
|
"rewards/margins": 0.5431486368179321, |
|
"rewards/rejected": -2.3302085399627686, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.48319709355131696, |
|
"grad_norm": 2.009992837905884, |
|
"learning_rate": 9.679246003464585e-06, |
|
"logits/chosen": 0.032514430582523346, |
|
"logits/rejected": 0.04692292958498001, |
|
"logps/chosen": -85.35857391357422, |
|
"logps/rejected": -90.84698486328125, |
|
"loss": 2.8119, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.1302294731140137, |
|
"rewards/margins": 0.3329467177391052, |
|
"rewards/rejected": -2.4631760120391846, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.48501362397820164, |
|
"grad_norm": 1.535208821296692, |
|
"learning_rate": 9.676859126207957e-06, |
|
"logits/chosen": 0.03791799396276474, |
|
"logits/rejected": 0.0628521591424942, |
|
"logps/chosen": -73.64868927001953, |
|
"logps/rejected": -80.9854965209961, |
|
"loss": 2.2386, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.9628286361694336, |
|
"rewards/margins": 0.6735512018203735, |
|
"rewards/rejected": -2.6363797187805176, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.4868301544050863, |
|
"grad_norm": 1.9440377950668335, |
|
"learning_rate": 9.674463731608309e-06, |
|
"logits/chosen": 0.0922878235578537, |
|
"logits/rejected": 0.08414338529109955, |
|
"logps/chosen": -80.34213256835938, |
|
"logps/rejected": -87.45000457763672, |
|
"loss": 2.8564, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.194643020629883, |
|
"rewards/margins": 0.5086687207221985, |
|
"rewards/rejected": -2.7033114433288574, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.48864668483197093, |
|
"grad_norm": 1.57882559299469, |
|
"learning_rate": 9.672059824550268e-06, |
|
"logits/chosen": 0.08647017180919647, |
|
"logits/rejected": 0.07438144087791443, |
|
"logps/chosen": -89.79737091064453, |
|
"logps/rejected": -95.89274597167969, |
|
"loss": 2.1801, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9004266262054443, |
|
"rewards/margins": 0.5936228632926941, |
|
"rewards/rejected": -2.494049549102783, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.4904632152588556, |
|
"grad_norm": 1.47967529296875, |
|
"learning_rate": 9.669647409935822e-06, |
|
"logits/chosen": 0.0663951188325882, |
|
"logits/rejected": 0.12090058624744415, |
|
"logps/chosen": -86.87987518310547, |
|
"logps/rejected": -90.49571228027344, |
|
"loss": 2.342, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -2.043788194656372, |
|
"rewards/margins": 0.4767158329486847, |
|
"rewards/rejected": -2.5205039978027344, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.49227974568574023, |
|
"grad_norm": 1.5860496759414673, |
|
"learning_rate": 9.667226492684302e-06, |
|
"logits/chosen": 0.06286406517028809, |
|
"logits/rejected": 0.1226339116692543, |
|
"logps/chosen": -91.13972473144531, |
|
"logps/rejected": -105.41304016113281, |
|
"loss": 2.2264, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.8733948469161987, |
|
"rewards/margins": 0.6329715251922607, |
|
"rewards/rejected": -2.50636625289917, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.4940962761126249, |
|
"grad_norm": 1.6157667636871338, |
|
"learning_rate": 9.66479707773238e-06, |
|
"logits/chosen": 0.07893703132867813, |
|
"logits/rejected": 0.1179615929722786, |
|
"logps/chosen": -77.89990234375, |
|
"logps/rejected": -87.70220947265625, |
|
"loss": 2.4723, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -2.136741876602173, |
|
"rewards/margins": 0.4580468237400055, |
|
"rewards/rejected": -2.5947885513305664, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.49591280653950953, |
|
"grad_norm": 1.5408835411071777, |
|
"learning_rate": 9.662359170034058e-06, |
|
"logits/chosen": 0.0976145789027214, |
|
"logits/rejected": 0.12423861026763916, |
|
"logps/chosen": -76.74501037597656, |
|
"logps/rejected": -82.18941497802734, |
|
"loss": 2.5565, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.039276361465454, |
|
"rewards/margins": 0.3934023678302765, |
|
"rewards/rejected": -2.432678461074829, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.4977293369663942, |
|
"grad_norm": 1.44657564163208, |
|
"learning_rate": 9.659912774560654e-06, |
|
"logits/chosen": 0.04748811572790146, |
|
"logits/rejected": 0.12317924201488495, |
|
"logps/chosen": -75.47578430175781, |
|
"logps/rejected": -88.56098937988281, |
|
"loss": 2.1908, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -2.050949811935425, |
|
"rewards/margins": 0.564163863658905, |
|
"rewards/rejected": -2.6151139736175537, |
|
"step": 274 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 2200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 137, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.299919692365824e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|