|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.7465940054495913, |
|
"eval_steps": 500, |
|
"global_step": 411, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0018165304268846503, |
|
"grad_norm": 0.7654335498809814, |
|
"learning_rate": 9.99999541185824e-06, |
|
"logits/chosen": 0.10746052116155624, |
|
"logits/rejected": 0.16293610632419586, |
|
"logps/chosen": -51.70099639892578, |
|
"logps/rejected": -56.85424041748047, |
|
"loss": 2.7726, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0036330608537693005, |
|
"grad_norm": 0.8493374586105347, |
|
"learning_rate": 9.999981647442309e-06, |
|
"logits/chosen": 0.12589210271835327, |
|
"logits/rejected": 0.18243370950222015, |
|
"logps/chosen": -61.092987060546875, |
|
"logps/rejected": -67.26304626464844, |
|
"loss": 2.7516, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": 0.005276113282889128, |
|
"rewards/margins": 0.010965153574943542, |
|
"rewards/rejected": -0.00568903936073184, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.005449591280653951, |
|
"grad_norm": 0.739442765712738, |
|
"learning_rate": 9.999958706780277e-06, |
|
"logits/chosen": 0.0766456350684166, |
|
"logits/rejected": 0.1422545611858368, |
|
"logps/chosen": -55.764503479003906, |
|
"logps/rejected": -57.38707733154297, |
|
"loss": 2.7902, |
|
"rewards/accuracies": 0.4375, |
|
"rewards/chosen": -0.0027439936529845, |
|
"rewards/margins": -0.008403200656175613, |
|
"rewards/rejected": 0.005659207701683044, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.007266121707538601, |
|
"grad_norm": 0.7261953949928284, |
|
"learning_rate": 9.999926589918927e-06, |
|
"logits/chosen": 0.1812177300453186, |
|
"logits/rejected": 0.13825736939907074, |
|
"logps/chosen": -56.98471450805664, |
|
"logps/rejected": -55.47361373901367, |
|
"loss": 2.7537, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": 0.006436044350266457, |
|
"rewards/margins": 0.009899044409394264, |
|
"rewards/rejected": -0.003463000524789095, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.009082652134423252, |
|
"grad_norm": 0.850686252117157, |
|
"learning_rate": 9.999885296923748e-06, |
|
"logits/chosen": 0.17714114487171173, |
|
"logits/rejected": 0.12910494208335876, |
|
"logps/chosen": -71.72655487060547, |
|
"logps/rejected": -70.75139617919922, |
|
"loss": 2.79, |
|
"rewards/accuracies": 0.453125, |
|
"rewards/chosen": -0.008413795381784439, |
|
"rewards/margins": -0.007964953780174255, |
|
"rewards/rejected": -0.0004488405538722873, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.010899182561307902, |
|
"grad_norm": 0.6627741456031799, |
|
"learning_rate": 9.999834827878942e-06, |
|
"logits/chosen": 0.13277305662631989, |
|
"logits/rejected": 0.20749863982200623, |
|
"logps/chosen": -47.56520080566406, |
|
"logps/rejected": -53.18217468261719, |
|
"loss": 2.7732, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": 0.001662571681663394, |
|
"rewards/margins": 0.00014669005759060383, |
|
"rewards/rejected": 0.0015158820897340775, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.012715712988192553, |
|
"grad_norm": 0.8335347175598145, |
|
"learning_rate": 9.999775182887429e-06, |
|
"logits/chosen": 0.07216031849384308, |
|
"logits/rejected": 0.13223227858543396, |
|
"logps/chosen": -73.06997680664062, |
|
"logps/rejected": -75.58594512939453, |
|
"loss": 2.7716, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.004659444559365511, |
|
"rewards/margins": 0.0013355333358049393, |
|
"rewards/rejected": -0.005994977429509163, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.014532243415077202, |
|
"grad_norm": 0.7128610014915466, |
|
"learning_rate": 9.999706362070833e-06, |
|
"logits/chosen": 0.11368857324123383, |
|
"logits/rejected": 0.15805509686470032, |
|
"logps/chosen": -56.45945358276367, |
|
"logps/rejected": -56.9697151184082, |
|
"loss": 2.7544, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.0038012717850506306, |
|
"rewards/margins": 0.009716177359223366, |
|
"rewards/rejected": -0.013517449609935284, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01634877384196185, |
|
"grad_norm": 0.77040034532547, |
|
"learning_rate": 9.99962836556949e-06, |
|
"logits/chosen": 0.1818978637456894, |
|
"logits/rejected": 0.20750640332698822, |
|
"logps/chosen": -59.184532165527344, |
|
"logps/rejected": -61.59263610839844, |
|
"loss": 2.7688, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.009461460635066032, |
|
"rewards/margins": 0.002394758863374591, |
|
"rewards/rejected": -0.011856218799948692, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.018165304268846504, |
|
"grad_norm": 0.7393640875816345, |
|
"learning_rate": 9.99954119354245e-06, |
|
"logits/chosen": 0.14076584577560425, |
|
"logits/rejected": 0.22711633145809174, |
|
"logps/chosen": -56.34330749511719, |
|
"logps/rejected": -60.51960372924805, |
|
"loss": 2.7827, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.004167273174971342, |
|
"rewards/margins": -0.00429012393578887, |
|
"rewards/rejected": 0.00012284982949495316, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.019981834695731154, |
|
"grad_norm": 0.7271163463592529, |
|
"learning_rate": 9.999444846167473e-06, |
|
"logits/chosen": 0.20141208171844482, |
|
"logits/rejected": 0.2038215547800064, |
|
"logps/chosen": -57.712646484375, |
|
"logps/rejected": -54.343231201171875, |
|
"loss": 2.7565, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": 0.0009536671568639576, |
|
"rewards/margins": 0.00866839662194252, |
|
"rewards/rejected": -0.007714730221778154, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.021798365122615803, |
|
"grad_norm": 0.9451335668563843, |
|
"learning_rate": 9.999339323641027e-06, |
|
"logits/chosen": 0.16046015918254852, |
|
"logits/rejected": 0.1963367462158203, |
|
"logps/chosen": -51.579856872558594, |
|
"logps/rejected": -54.97206115722656, |
|
"loss": 2.7732, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.005096627864986658, |
|
"rewards/margins": 0.00017735245637595654, |
|
"rewards/rejected": -0.005273980088531971, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.023614895549500452, |
|
"grad_norm": 0.8279874324798584, |
|
"learning_rate": 9.99922462617829e-06, |
|
"logits/chosen": 0.12593263387680054, |
|
"logits/rejected": 0.11708654463291168, |
|
"logps/chosen": -62.25929641723633, |
|
"logps/rejected": -60.633296966552734, |
|
"loss": 2.7693, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.013154934160411358, |
|
"rewards/margins": 0.002298446139320731, |
|
"rewards/rejected": -0.015453380532562733, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.025431425976385105, |
|
"grad_norm": 0.7863137125968933, |
|
"learning_rate": 9.999100754013152e-06, |
|
"logits/chosen": 0.12714676558971405, |
|
"logits/rejected": 0.18583568930625916, |
|
"logps/chosen": -59.119529724121094, |
|
"logps/rejected": -65.85352325439453, |
|
"loss": 2.7715, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.01445689145475626, |
|
"rewards/margins": 0.001161783467978239, |
|
"rewards/rejected": -0.015618674457073212, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.027247956403269755, |
|
"grad_norm": 0.7623695731163025, |
|
"learning_rate": 9.998967707398207e-06, |
|
"logits/chosen": 0.09560943394899368, |
|
"logits/rejected": 0.14794519543647766, |
|
"logps/chosen": -47.649497985839844, |
|
"logps/rejected": -57.21004867553711, |
|
"loss": 2.7405, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.0010593307670205832, |
|
"rewards/margins": 0.0166107639670372, |
|
"rewards/rejected": -0.017670094966888428, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.029064486830154404, |
|
"grad_norm": 0.7020726203918457, |
|
"learning_rate": 9.998825486604765e-06, |
|
"logits/chosen": 0.19466270506381989, |
|
"logits/rejected": 0.21006342768669128, |
|
"logps/chosen": -50.8302116394043, |
|
"logps/rejected": -58.202754974365234, |
|
"loss": 2.7628, |
|
"rewards/accuracies": 0.484375, |
|
"rewards/chosen": -0.009497416205704212, |
|
"rewards/margins": 0.005540419369935989, |
|
"rewards/rejected": -0.015037836506962776, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.030881017257039057, |
|
"grad_norm": 0.7085703611373901, |
|
"learning_rate": 9.998674091922833e-06, |
|
"logits/chosen": 0.06517614424228668, |
|
"logits/rejected": 0.15166127681732178, |
|
"logps/chosen": -45.203216552734375, |
|
"logps/rejected": -51.454952239990234, |
|
"loss": 2.7568, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.009879402816295624, |
|
"rewards/margins": 0.00827767699956894, |
|
"rewards/rejected": -0.018157079815864563, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.0326975476839237, |
|
"grad_norm": 0.8115389347076416, |
|
"learning_rate": 9.998513523661136e-06, |
|
"logits/chosen": 0.09045977890491486, |
|
"logits/rejected": 0.1498938351869583, |
|
"logps/chosen": -60.35250473022461, |
|
"logps/rejected": -62.11174774169922, |
|
"loss": 2.7888, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.01872912421822548, |
|
"rewards/margins": -0.0073302448727190495, |
|
"rewards/rejected": -0.011398878879845142, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03451407811080836, |
|
"grad_norm": 0.8222002983093262, |
|
"learning_rate": 9.998343782147098e-06, |
|
"logits/chosen": 0.08091418445110321, |
|
"logits/rejected": 0.1225643903017044, |
|
"logps/chosen": -54.25471115112305, |
|
"logps/rejected": -57.91140365600586, |
|
"loss": 2.7652, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.018515082076191902, |
|
"rewards/margins": 0.004185312893241644, |
|
"rewards/rejected": -0.022700395435094833, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.03633060853769301, |
|
"grad_norm": 0.862041711807251, |
|
"learning_rate": 9.998164867726853e-06, |
|
"logits/chosen": 0.04738985374569893, |
|
"logits/rejected": 0.11436720937490463, |
|
"logps/chosen": -62.22935104370117, |
|
"logps/rejected": -67.67117309570312, |
|
"loss": 2.7607, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.01734255626797676, |
|
"rewards/margins": 0.006629224866628647, |
|
"rewards/rejected": -0.023971781134605408, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03814713896457766, |
|
"grad_norm": 0.8169613480567932, |
|
"learning_rate": 9.997976780765237e-06, |
|
"logits/chosen": 0.14641642570495605, |
|
"logits/rejected": 0.17384907603263855, |
|
"logps/chosen": -61.21760559082031, |
|
"logps/rejected": -61.094486236572266, |
|
"loss": 2.7752, |
|
"rewards/accuracies": 0.46875, |
|
"rewards/chosen": -0.021789425984025, |
|
"rewards/margins": -0.000787546974606812, |
|
"rewards/rejected": -0.02100187912583351, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.03996366939146231, |
|
"grad_norm": 0.7457932829856873, |
|
"learning_rate": 9.997779521645793e-06, |
|
"logits/chosen": 0.16773203015327454, |
|
"logits/rejected": 0.16435235738754272, |
|
"logps/chosen": -55.192623138427734, |
|
"logps/rejected": -53.05280303955078, |
|
"loss": 2.7661, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.021701142191886902, |
|
"rewards/margins": 0.0036970973014831543, |
|
"rewards/rejected": -0.025398239493370056, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04178019981834696, |
|
"grad_norm": 0.8263576030731201, |
|
"learning_rate": 9.997573090770766e-06, |
|
"logits/chosen": 0.10863066464662552, |
|
"logits/rejected": 0.1277829110622406, |
|
"logps/chosen": -61.80084228515625, |
|
"logps/rejected": -64.94878387451172, |
|
"loss": 2.7526, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.019380860030651093, |
|
"rewards/margins": 0.010501865297555923, |
|
"rewards/rejected": -0.029882723465561867, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.043596730245231606, |
|
"grad_norm": 0.7587623000144958, |
|
"learning_rate": 9.997357488561107e-06, |
|
"logits/chosen": 0.17992480099201202, |
|
"logits/rejected": 0.16478824615478516, |
|
"logps/chosen": -62.946449279785156, |
|
"logps/rejected": -61.98965072631836, |
|
"loss": 2.7534, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.02564322203397751, |
|
"rewards/margins": 0.010079940780997276, |
|
"rewards/rejected": -0.035723160952329636, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.045413260672116255, |
|
"grad_norm": 0.7733523845672607, |
|
"learning_rate": 9.997132715456464e-06, |
|
"logits/chosen": 0.1268949955701828, |
|
"logits/rejected": 0.14968058466911316, |
|
"logps/chosen": -53.57306671142578, |
|
"logps/rejected": -58.07795715332031, |
|
"loss": 2.7602, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.030423898249864578, |
|
"rewards/margins": 0.006796791218221188, |
|
"rewards/rejected": -0.03722068667411804, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.047229791099000905, |
|
"grad_norm": 0.8851524591445923, |
|
"learning_rate": 9.99689877191519e-06, |
|
"logits/chosen": 0.17443379759788513, |
|
"logits/rejected": 0.1787741333246231, |
|
"logps/chosen": -58.031558990478516, |
|
"logps/rejected": -59.206520080566406, |
|
"loss": 2.762, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.02750026062130928, |
|
"rewards/margins": 0.006129855290055275, |
|
"rewards/rejected": -0.033630117774009705, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.04904632152588556, |
|
"grad_norm": 0.8524475693702698, |
|
"learning_rate": 9.996655658414331e-06, |
|
"logits/chosen": 0.18196682631969452, |
|
"logits/rejected": 0.237998366355896, |
|
"logps/chosen": -63.18816375732422, |
|
"logps/rejected": -67.31481170654297, |
|
"loss": 2.7519, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.03693533316254616, |
|
"rewards/margins": 0.011266030371189117, |
|
"rewards/rejected": -0.048201363533735275, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05086285195277021, |
|
"grad_norm": 0.8785350918769836, |
|
"learning_rate": 9.996403375449647e-06, |
|
"logits/chosen": 0.12056512385606766, |
|
"logits/rejected": 0.17535904049873352, |
|
"logps/chosen": -62.035221099853516, |
|
"logps/rejected": -69.88546752929688, |
|
"loss": 2.743, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.03874595835804939, |
|
"rewards/margins": 0.015791552141308784, |
|
"rewards/rejected": -0.054537512362003326, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.05267938237965486, |
|
"grad_norm": 0.9492508172988892, |
|
"learning_rate": 9.996141923535582e-06, |
|
"logits/chosen": 0.12200314551591873, |
|
"logits/rejected": 0.13907021284103394, |
|
"logps/chosen": -65.31945037841797, |
|
"logps/rejected": -66.59127807617188, |
|
"loss": 2.7641, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.0383562371134758, |
|
"rewards/margins": 0.005150892771780491, |
|
"rewards/rejected": -0.043507132679224014, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.05449591280653951, |
|
"grad_norm": 0.8084173202514648, |
|
"learning_rate": 9.995871303205282e-06, |
|
"logits/chosen": 0.07616369426250458, |
|
"logits/rejected": 0.2001558393239975, |
|
"logps/chosen": -51.103309631347656, |
|
"logps/rejected": -62.28483581542969, |
|
"loss": 2.7519, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.04615113511681557, |
|
"rewards/margins": 0.011578184552490711, |
|
"rewards/rejected": -0.057729318737983704, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.05631244323342416, |
|
"grad_norm": 0.943800151348114, |
|
"learning_rate": 9.995591515010589e-06, |
|
"logits/chosen": 0.04337490350008011, |
|
"logits/rejected": 0.15334263443946838, |
|
"logps/chosen": -57.265567779541016, |
|
"logps/rejected": -68.5506820678711, |
|
"loss": 2.7253, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.04982424154877663, |
|
"rewards/margins": 0.024712499231100082, |
|
"rewards/rejected": -0.07453674077987671, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.05812897366030881, |
|
"grad_norm": 0.9372840523719788, |
|
"learning_rate": 9.99530255952204e-06, |
|
"logits/chosen": 0.13595089316368103, |
|
"logits/rejected": 0.14287832379341125, |
|
"logps/chosen": -61.515377044677734, |
|
"logps/rejected": -63.1092529296875, |
|
"loss": 2.7571, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.05759892612695694, |
|
"rewards/margins": 0.009150844067335129, |
|
"rewards/rejected": -0.06674977391958237, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.05994550408719346, |
|
"grad_norm": 0.8408157825469971, |
|
"learning_rate": 9.995004437328866e-06, |
|
"logits/chosen": 0.18553151190280914, |
|
"logits/rejected": 0.14024314284324646, |
|
"logps/chosen": -61.73942565917969, |
|
"logps/rejected": -53.5772819519043, |
|
"loss": 2.7627, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.06591347604990005, |
|
"rewards/margins": 0.006169732194393873, |
|
"rewards/rejected": -0.07208321243524551, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.061762034514078114, |
|
"grad_norm": 0.8856032490730286, |
|
"learning_rate": 9.994697149038989e-06, |
|
"logits/chosen": 0.08410881459712982, |
|
"logits/rejected": 0.13991469144821167, |
|
"logps/chosen": -60.161231994628906, |
|
"logps/rejected": -60.03232192993164, |
|
"loss": 2.7437, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.06631935387849808, |
|
"rewards/margins": 0.016240764409303665, |
|
"rewards/rejected": -0.08256012946367264, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.06357856494096276, |
|
"grad_norm": 0.9037399291992188, |
|
"learning_rate": 9.994380695279025e-06, |
|
"logits/chosen": 0.06669525057077408, |
|
"logits/rejected": 0.16171438992023468, |
|
"logps/chosen": -59.874202728271484, |
|
"logps/rejected": -67.9597396850586, |
|
"loss": 2.7221, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.06999661773443222, |
|
"rewards/margins": 0.027104372158646584, |
|
"rewards/rejected": -0.09710099548101425, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.0653950953678474, |
|
"grad_norm": 0.8819396495819092, |
|
"learning_rate": 9.994055076694276e-06, |
|
"logits/chosen": 0.18232542276382446, |
|
"logits/rejected": 0.2102084755897522, |
|
"logps/chosen": -59.298095703125, |
|
"logps/rejected": -63.70309066772461, |
|
"loss": 2.7179, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.08072888106107712, |
|
"rewards/margins": 0.029247857630252838, |
|
"rewards/rejected": -0.10997673869132996, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.06721162579473206, |
|
"grad_norm": 0.9046475291252136, |
|
"learning_rate": 9.993720293948739e-06, |
|
"logits/chosen": 0.13181668519973755, |
|
"logits/rejected": 0.1316901445388794, |
|
"logps/chosen": -71.46318054199219, |
|
"logps/rejected": -76.40208435058594, |
|
"loss": 2.7505, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.08365033566951752, |
|
"rewards/margins": 0.01496690884232521, |
|
"rewards/rejected": -0.09861725568771362, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.06902815622161672, |
|
"grad_norm": 0.9267721772193909, |
|
"learning_rate": 9.993376347725091e-06, |
|
"logits/chosen": 0.14555476605892181, |
|
"logits/rejected": 0.21660488843917847, |
|
"logps/chosen": -61.28938293457031, |
|
"logps/rejected": -67.28329467773438, |
|
"loss": 2.7007, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.07440298050642014, |
|
"rewards/margins": 0.03913095220923424, |
|
"rewards/rejected": -0.11353392899036407, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.07084468664850137, |
|
"grad_norm": 0.8545824885368347, |
|
"learning_rate": 9.993023238724696e-06, |
|
"logits/chosen": 0.09474823623895645, |
|
"logits/rejected": 0.09304340183734894, |
|
"logps/chosen": -61.27484130859375, |
|
"logps/rejected": -62.003597259521484, |
|
"loss": 2.7406, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.09001085162162781, |
|
"rewards/margins": 0.017711879685521126, |
|
"rewards/rejected": -0.10772272944450378, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.07266121707538602, |
|
"grad_norm": 0.9086014628410339, |
|
"learning_rate": 9.99266096766761e-06, |
|
"logits/chosen": 0.17910577356815338, |
|
"logits/rejected": 0.14881345629692078, |
|
"logps/chosen": -59.22669219970703, |
|
"logps/rejected": -55.103187561035156, |
|
"loss": 2.7422, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.1086670309305191, |
|
"rewards/margins": 0.01710200309753418, |
|
"rewards/rejected": -0.12576903402805328, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.07447774750227067, |
|
"grad_norm": 0.9802166223526001, |
|
"learning_rate": 9.992289535292565e-06, |
|
"logits/chosen": 0.18225271999835968, |
|
"logits/rejected": 0.20810523629188538, |
|
"logps/chosen": -55.57660675048828, |
|
"logps/rejected": -59.129417419433594, |
|
"loss": 2.7505, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.13455544412136078, |
|
"rewards/margins": 0.012644169852137566, |
|
"rewards/rejected": -0.1471996009349823, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.07629427792915532, |
|
"grad_norm": 0.919211208820343, |
|
"learning_rate": 9.991908942356977e-06, |
|
"logits/chosen": 0.07866425067186356, |
|
"logits/rejected": 0.07479682564735413, |
|
"logps/chosen": -66.11737060546875, |
|
"logps/rejected": -64.02122497558594, |
|
"loss": 2.7855, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.15568655729293823, |
|
"rewards/margins": -0.0022164471447467804, |
|
"rewards/rejected": -0.15347009897232056, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.07811080835603997, |
|
"grad_norm": 0.9612410068511963, |
|
"learning_rate": 9.991519189636937e-06, |
|
"logits/chosen": 0.11137107014656067, |
|
"logits/rejected": 0.14462286233901978, |
|
"logps/chosen": -67.43673706054688, |
|
"logps/rejected": -74.29210662841797, |
|
"loss": 2.7203, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.16468609869480133, |
|
"rewards/margins": 0.032481495290994644, |
|
"rewards/rejected": -0.19716759026050568, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.07992733878292461, |
|
"grad_norm": 0.8044790625572205, |
|
"learning_rate": 9.991120277927224e-06, |
|
"logits/chosen": 0.08509679138660431, |
|
"logits/rejected": 0.11362393945455551, |
|
"logps/chosen": -56.867591857910156, |
|
"logps/rejected": -55.66028594970703, |
|
"loss": 2.7302, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.15751589834690094, |
|
"rewards/margins": 0.02398722618818283, |
|
"rewards/rejected": -0.18150311708450317, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.08174386920980926, |
|
"grad_norm": 0.8599545955657959, |
|
"learning_rate": 9.990712208041284e-06, |
|
"logits/chosen": 0.1391746699810028, |
|
"logits/rejected": 0.1746281236410141, |
|
"logps/chosen": -62.22108459472656, |
|
"logps/rejected": -58.50814437866211, |
|
"loss": 2.7288, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.16528920829296112, |
|
"rewards/margins": 0.02449135296046734, |
|
"rewards/rejected": -0.1897805631160736, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.08356039963669391, |
|
"grad_norm": 0.9419183731079102, |
|
"learning_rate": 9.990294980811244e-06, |
|
"logits/chosen": 0.1693490892648697, |
|
"logits/rejected": 0.18133510649204254, |
|
"logps/chosen": -60.769744873046875, |
|
"logps/rejected": -65.61067199707031, |
|
"loss": 2.7003, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.16450145840644836, |
|
"rewards/margins": 0.04409575089812279, |
|
"rewards/rejected": -0.20859721302986145, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.08537693006357856, |
|
"grad_norm": 0.9839057326316833, |
|
"learning_rate": 9.989868597087907e-06, |
|
"logits/chosen": 0.1363808959722519, |
|
"logits/rejected": 0.22261787950992584, |
|
"logps/chosen": -65.37456512451172, |
|
"logps/rejected": -72.04705047607422, |
|
"loss": 2.6636, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.18397849798202515, |
|
"rewards/margins": 0.06703396141529083, |
|
"rewards/rejected": -0.2510124444961548, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.08719346049046321, |
|
"grad_norm": 0.8818926811218262, |
|
"learning_rate": 9.989433057740738e-06, |
|
"logits/chosen": 0.1580447256565094, |
|
"logits/rejected": 0.18051129579544067, |
|
"logps/chosen": -56.6561279296875, |
|
"logps/rejected": -64.66617584228516, |
|
"loss": 2.6658, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.2094140350818634, |
|
"rewards/margins": 0.05971873924136162, |
|
"rewards/rejected": -0.2691327929496765, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.08900999091734786, |
|
"grad_norm": 0.8838489651679993, |
|
"learning_rate": 9.98898836365788e-06, |
|
"logits/chosen": 0.11484000086784363, |
|
"logits/rejected": 0.12583914399147034, |
|
"logps/chosen": -65.76107788085938, |
|
"logps/rejected": -75.4742660522461, |
|
"loss": 2.6777, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.19882163405418396, |
|
"rewards/margins": 0.0690067932009697, |
|
"rewards/rejected": -0.26782843470573425, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.09082652134423251, |
|
"grad_norm": 0.9035817980766296, |
|
"learning_rate": 9.988534515746141e-06, |
|
"logits/chosen": 0.12592823803424835, |
|
"logits/rejected": 0.18080386519432068, |
|
"logps/chosen": -60.42765808105469, |
|
"logps/rejected": -65.03330993652344, |
|
"loss": 2.68, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.25037306547164917, |
|
"rewards/margins": 0.05723651498556137, |
|
"rewards/rejected": -0.30760958790779114, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.09264305177111716, |
|
"grad_norm": 1.0728137493133545, |
|
"learning_rate": 9.988071514930998e-06, |
|
"logits/chosen": 0.0955006331205368, |
|
"logits/rejected": 0.15643228590488434, |
|
"logps/chosen": -56.28706359863281, |
|
"logps/rejected": -70.3702392578125, |
|
"loss": 2.6104, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.2517072558403015, |
|
"rewards/margins": 0.11176257580518723, |
|
"rewards/rejected": -0.36346977949142456, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.09445958219800181, |
|
"grad_norm": 0.9143723845481873, |
|
"learning_rate": 9.987599362156587e-06, |
|
"logits/chosen": 0.1018611341714859, |
|
"logits/rejected": 0.21036753058433533, |
|
"logps/chosen": -55.815738677978516, |
|
"logps/rejected": -66.23766326904297, |
|
"loss": 2.6207, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.2496231347322464, |
|
"rewards/margins": 0.09438168257474899, |
|
"rewards/rejected": -0.3440048098564148, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.09627611262488647, |
|
"grad_norm": 0.8731315732002258, |
|
"learning_rate": 9.987118058385712e-06, |
|
"logits/chosen": 0.12294681370258331, |
|
"logits/rejected": 0.12897568941116333, |
|
"logps/chosen": -63.27684020996094, |
|
"logps/rejected": -65.21257019042969, |
|
"loss": 2.7093, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.28397974371910095, |
|
"rewards/margins": 0.05316353589296341, |
|
"rewards/rejected": -0.33714330196380615, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.09809264305177112, |
|
"grad_norm": 0.951884388923645, |
|
"learning_rate": 9.986627604599835e-06, |
|
"logits/chosen": 0.1588897705078125, |
|
"logits/rejected": 0.10302369296550751, |
|
"logps/chosen": -62.071083068847656, |
|
"logps/rejected": -59.575416564941406, |
|
"loss": 2.7725, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.31246036291122437, |
|
"rewards/margins": 0.011688929051160812, |
|
"rewards/rejected": -0.3241492807865143, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.09990917347865577, |
|
"grad_norm": 1.0379350185394287, |
|
"learning_rate": 9.986128001799077e-06, |
|
"logits/chosen": 0.15824833512306213, |
|
"logits/rejected": 0.16385456919670105, |
|
"logps/chosen": -79.65828704833984, |
|
"logps/rejected": -82.06480407714844, |
|
"loss": 2.6826, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.36005648970603943, |
|
"rewards/margins": 0.07091177999973297, |
|
"rewards/rejected": -0.4309682846069336, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.10172570390554042, |
|
"grad_norm": 0.8740183115005493, |
|
"learning_rate": 9.985619251002214e-06, |
|
"logits/chosen": 0.0996306911110878, |
|
"logits/rejected": 0.14288735389709473, |
|
"logps/chosen": -59.8499755859375, |
|
"logps/rejected": -63.10956573486328, |
|
"loss": 2.664, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.3242360055446625, |
|
"rewards/margins": 0.06764136999845505, |
|
"rewards/rejected": -0.39187735319137573, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.10354223433242507, |
|
"grad_norm": 0.9942138195037842, |
|
"learning_rate": 9.985101353246676e-06, |
|
"logits/chosen": 0.12070289999246597, |
|
"logits/rejected": 0.1866157352924347, |
|
"logps/chosen": -70.04698181152344, |
|
"logps/rejected": -76.2674560546875, |
|
"loss": 2.6165, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.4166935980319977, |
|
"rewards/margins": 0.10780875384807587, |
|
"rewards/rejected": -0.5245023965835571, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.10535876475930972, |
|
"grad_norm": 0.9078152775764465, |
|
"learning_rate": 9.98457430958855e-06, |
|
"logits/chosen": 0.16699416935443878, |
|
"logits/rejected": 0.216594398021698, |
|
"logps/chosen": -59.81795883178711, |
|
"logps/rejected": -64.79395294189453, |
|
"loss": 2.551, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.39265742897987366, |
|
"rewards/margins": 0.14742323756217957, |
|
"rewards/rejected": -0.5400806665420532, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.10717529518619437, |
|
"grad_norm": 0.9305158257484436, |
|
"learning_rate": 9.984038121102569e-06, |
|
"logits/chosen": 0.16001635789871216, |
|
"logits/rejected": 0.18797104060649872, |
|
"logps/chosen": -58.087158203125, |
|
"logps/rejected": -59.003414154052734, |
|
"loss": 2.6345, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.3626652956008911, |
|
"rewards/margins": 0.08553728461265564, |
|
"rewards/rejected": -0.44820258021354675, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.10899182561307902, |
|
"grad_norm": 0.9690614342689514, |
|
"learning_rate": 9.983492788882113e-06, |
|
"logits/chosen": 0.161406010389328, |
|
"logits/rejected": 0.14847029745578766, |
|
"logps/chosen": -61.01350021362305, |
|
"logps/rejected": -60.993080139160156, |
|
"loss": 2.7099, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.4803884029388428, |
|
"rewards/margins": 0.0831567794084549, |
|
"rewards/rejected": -0.5635451078414917, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.11080835603996367, |
|
"grad_norm": 0.971172034740448, |
|
"learning_rate": 9.98293831403921e-06, |
|
"logits/chosen": 0.14095918834209442, |
|
"logits/rejected": 0.16142162680625916, |
|
"logps/chosen": -62.31782150268555, |
|
"logps/rejected": -64.44837188720703, |
|
"loss": 2.6373, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.47784310579299927, |
|
"rewards/margins": 0.09006088227033615, |
|
"rewards/rejected": -0.567903995513916, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.11262488646684832, |
|
"grad_norm": 0.9984462857246399, |
|
"learning_rate": 9.982374697704532e-06, |
|
"logits/chosen": 0.08945554494857788, |
|
"logits/rejected": 0.17623476684093475, |
|
"logps/chosen": -58.84153366088867, |
|
"logps/rejected": -72.59091186523438, |
|
"loss": 2.5465, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.41726160049438477, |
|
"rewards/margins": 0.16527526080608368, |
|
"rewards/rejected": -0.5825368165969849, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.11444141689373297, |
|
"grad_norm": 1.007622241973877, |
|
"learning_rate": 9.981801941027388e-06, |
|
"logits/chosen": 0.16932496428489685, |
|
"logits/rejected": 0.24257409572601318, |
|
"logps/chosen": -61.62455749511719, |
|
"logps/rejected": -69.40538024902344, |
|
"loss": 2.5876, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.4484747052192688, |
|
"rewards/margins": 0.11513397097587585, |
|
"rewards/rejected": -0.563608705997467, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.11625794732061762, |
|
"grad_norm": 1.021275281906128, |
|
"learning_rate": 9.981220045175731e-06, |
|
"logits/chosen": 0.13468854129314423, |
|
"logits/rejected": 0.10648790001869202, |
|
"logps/chosen": -56.91660690307617, |
|
"logps/rejected": -54.84038162231445, |
|
"loss": 2.7096, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.47126585245132446, |
|
"rewards/margins": 0.05656271427869797, |
|
"rewards/rejected": -0.527828574180603, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.11807447774750227, |
|
"grad_norm": 0.9580955505371094, |
|
"learning_rate": 9.980629011336149e-06, |
|
"logits/chosen": 0.10542559623718262, |
|
"logits/rejected": 0.18956655263900757, |
|
"logps/chosen": -63.21677017211914, |
|
"logps/rejected": -68.92955017089844, |
|
"loss": 2.5447, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.4920822083950043, |
|
"rewards/margins": 0.1548418551683426, |
|
"rewards/rejected": -0.6469241380691528, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.11989100817438691, |
|
"grad_norm": 0.9601484537124634, |
|
"learning_rate": 9.98002884071386e-06, |
|
"logits/chosen": 0.1211993545293808, |
|
"logits/rejected": 0.15335297584533691, |
|
"logps/chosen": -65.79328918457031, |
|
"logps/rejected": -71.8165283203125, |
|
"loss": 2.6013, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.4890082776546478, |
|
"rewards/margins": 0.16498278081417084, |
|
"rewards/rejected": -0.6539911031723022, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.12170753860127158, |
|
"grad_norm": 0.9475951194763184, |
|
"learning_rate": 9.97941953453272e-06, |
|
"logits/chosen": 0.129757359623909, |
|
"logits/rejected": 0.19597335159778595, |
|
"logps/chosen": -61.52772903442383, |
|
"logps/rejected": -62.8316764831543, |
|
"loss": 2.6008, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.6176788210868835, |
|
"rewards/margins": 0.11390677094459534, |
|
"rewards/rejected": -0.7315855622291565, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.12352406902815623, |
|
"grad_norm": 1.0577987432479858, |
|
"learning_rate": 9.978801094035207e-06, |
|
"logits/chosen": 0.1363590657711029, |
|
"logits/rejected": 0.1551840454339981, |
|
"logps/chosen": -71.82952117919922, |
|
"logps/rejected": -78.26856994628906, |
|
"loss": 2.6714, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.6073123216629028, |
|
"rewards/margins": 0.08929078280925751, |
|
"rewards/rejected": -0.6966031193733215, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.12534059945504086, |
|
"grad_norm": 1.2286291122436523, |
|
"learning_rate": 9.978173520482429e-06, |
|
"logits/chosen": 0.12522992491722107, |
|
"logits/rejected": 0.1818617582321167, |
|
"logps/chosen": -60.64283752441406, |
|
"logps/rejected": -67.53813934326172, |
|
"loss": 2.6884, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -0.5357421040534973, |
|
"rewards/margins": 0.0799265205860138, |
|
"rewards/rejected": -0.6156685948371887, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.1271571298819255, |
|
"grad_norm": 1.0266982316970825, |
|
"learning_rate": 9.97753681515412e-06, |
|
"logits/chosen": 0.17851999402046204, |
|
"logits/rejected": 0.25888925790786743, |
|
"logps/chosen": -71.65196990966797, |
|
"logps/rejected": -82.34913635253906, |
|
"loss": 2.5226, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.7787651419639587, |
|
"rewards/margins": 0.1905478835105896, |
|
"rewards/rejected": -0.9693130254745483, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.12897366030881016, |
|
"grad_norm": 0.999660313129425, |
|
"learning_rate": 9.976890979348628e-06, |
|
"logits/chosen": 0.18746232986450195, |
|
"logits/rejected": 0.2187044620513916, |
|
"logps/chosen": -69.84027099609375, |
|
"logps/rejected": -73.04399108886719, |
|
"loss": 2.6945, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.6969642639160156, |
|
"rewards/margins": 0.08156725764274597, |
|
"rewards/rejected": -0.778531551361084, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.1307901907356948, |
|
"grad_norm": 1.0596684217453003, |
|
"learning_rate": 9.976236014382934e-06, |
|
"logits/chosen": 0.17854923009872437, |
|
"logits/rejected": 0.18649883568286896, |
|
"logps/chosen": -63.992881774902344, |
|
"logps/rejected": -64.36544799804688, |
|
"loss": 2.7689, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.7436442375183105, |
|
"rewards/margins": 0.06337633728981018, |
|
"rewards/rejected": -0.8070206046104431, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.13260672116257946, |
|
"grad_norm": 1.0307612419128418, |
|
"learning_rate": 9.975571921592618e-06, |
|
"logits/chosen": 0.0964425653219223, |
|
"logits/rejected": 0.14382749795913696, |
|
"logps/chosen": -66.61788940429688, |
|
"logps/rejected": -70.80413055419922, |
|
"loss": 2.5859, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.6208910942077637, |
|
"rewards/margins": 0.17540723085403442, |
|
"rewards/rejected": -0.7962983250617981, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1344232515894641, |
|
"grad_norm": 0.9487422108650208, |
|
"learning_rate": 9.97489870233188e-06, |
|
"logits/chosen": 0.06293762475252151, |
|
"logits/rejected": 0.19688080251216888, |
|
"logps/chosen": -55.884613037109375, |
|
"logps/rejected": -64.30046081542969, |
|
"loss": 2.4062, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6415266990661621, |
|
"rewards/margins": 0.2741628885269165, |
|
"rewards/rejected": -0.9156895875930786, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.1362397820163488, |
|
"grad_norm": 1.223663330078125, |
|
"learning_rate": 9.974216357973539e-06, |
|
"logits/chosen": 0.10017126798629761, |
|
"logits/rejected": 0.10504551976919174, |
|
"logps/chosen": -83.83055114746094, |
|
"logps/rejected": -81.8952407836914, |
|
"loss": 2.8556, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8103373050689697, |
|
"rewards/margins": 0.027550537139177322, |
|
"rewards/rejected": -0.8378878831863403, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.13805631244323344, |
|
"grad_norm": 1.0121136903762817, |
|
"learning_rate": 9.973524889909007e-06, |
|
"logits/chosen": 0.07612155377864838, |
|
"logits/rejected": 0.15954618155956268, |
|
"logps/chosen": -65.6831283569336, |
|
"logps/rejected": -75.42146301269531, |
|
"loss": 2.5576, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6988077759742737, |
|
"rewards/margins": 0.1909356415271759, |
|
"rewards/rejected": -0.889743447303772, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.13987284287011809, |
|
"grad_norm": 1.0312719345092773, |
|
"learning_rate": 9.97282429954831e-06, |
|
"logits/chosen": 0.07015375047922134, |
|
"logits/rejected": 0.09143385291099548, |
|
"logps/chosen": -59.444358825683594, |
|
"logps/rejected": -61.26897430419922, |
|
"loss": 2.7845, |
|
"rewards/accuracies": 0.453125, |
|
"rewards/chosen": -0.7886172533035278, |
|
"rewards/margins": 0.038636498153209686, |
|
"rewards/rejected": -0.8272536993026733, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.14168937329700274, |
|
"grad_norm": 1.0625450611114502, |
|
"learning_rate": 9.972114588320073e-06, |
|
"logits/chosen": 0.07196499407291412, |
|
"logits/rejected": 0.12630172073841095, |
|
"logps/chosen": -75.5164566040039, |
|
"logps/rejected": -80.74308776855469, |
|
"loss": 2.5834, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.6821954250335693, |
|
"rewards/margins": 0.16858401894569397, |
|
"rewards/rejected": -0.8507794737815857, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.14350590372388738, |
|
"grad_norm": 0.9069045186042786, |
|
"learning_rate": 9.97139575767152e-06, |
|
"logits/chosen": 0.10926744341850281, |
|
"logits/rejected": 0.14418405294418335, |
|
"logps/chosen": -53.721046447753906, |
|
"logps/rejected": -58.17475891113281, |
|
"loss": 2.4955, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.6156648993492126, |
|
"rewards/margins": 0.19551396369934082, |
|
"rewards/rejected": -0.8111788034439087, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.14532243415077203, |
|
"grad_norm": 1.1463656425476074, |
|
"learning_rate": 9.970667809068476e-06, |
|
"logits/chosen": 0.09196805953979492, |
|
"logits/rejected": 0.14393256604671478, |
|
"logps/chosen": -77.11197662353516, |
|
"logps/rejected": -87.39771270751953, |
|
"loss": 2.5763, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8568902611732483, |
|
"rewards/margins": 0.15524335205554962, |
|
"rewards/rejected": -1.0121335983276367, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.14713896457765668, |
|
"grad_norm": 1.0046662092208862, |
|
"learning_rate": 9.969930743995351e-06, |
|
"logits/chosen": 0.2063535749912262, |
|
"logits/rejected": 0.20602768659591675, |
|
"logps/chosen": -59.890933990478516, |
|
"logps/rejected": -59.117794036865234, |
|
"loss": 2.6829, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.7329096794128418, |
|
"rewards/margins": 0.0983649268746376, |
|
"rewards/rejected": -0.8312745094299316, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.14895549500454133, |
|
"grad_norm": 1.0280438661575317, |
|
"learning_rate": 9.969184563955152e-06, |
|
"logits/chosen": 0.17393875122070312, |
|
"logits/rejected": 0.16783203184604645, |
|
"logps/chosen": -72.82754516601562, |
|
"logps/rejected": -70.6113510131836, |
|
"loss": 2.6277, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.7506997585296631, |
|
"rewards/margins": 0.13392671942710876, |
|
"rewards/rejected": -0.8846263885498047, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.15077202543142598, |
|
"grad_norm": 0.9356668591499329, |
|
"learning_rate": 9.968429270469467e-06, |
|
"logits/chosen": 0.09279462695121765, |
|
"logits/rejected": 0.1681107133626938, |
|
"logps/chosen": -62.300445556640625, |
|
"logps/rejected": -68.70709991455078, |
|
"loss": 2.4821, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.7132639288902283, |
|
"rewards/margins": 0.2519262731075287, |
|
"rewards/rejected": -0.9651902318000793, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.15258855585831063, |
|
"grad_norm": 1.0442771911621094, |
|
"learning_rate": 9.967664865078472e-06, |
|
"logits/chosen": 0.17844007909297943, |
|
"logits/rejected": 0.21383100748062134, |
|
"logps/chosen": -61.43730163574219, |
|
"logps/rejected": -64.16680908203125, |
|
"loss": 2.6289, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.7574411034584045, |
|
"rewards/margins": 0.13140378892421722, |
|
"rewards/rejected": -0.888844907283783, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.15440508628519528, |
|
"grad_norm": 1.137267827987671, |
|
"learning_rate": 9.966891349340922e-06, |
|
"logits/chosen": 0.19703873991966248, |
|
"logits/rejected": 0.17027492821216583, |
|
"logps/chosen": -68.07546997070312, |
|
"logps/rejected": -63.9476203918457, |
|
"loss": 2.7702, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -0.8362983465194702, |
|
"rewards/margins": 0.05476854741573334, |
|
"rewards/rejected": -0.8910670280456543, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.15622161671207993, |
|
"grad_norm": 1.0395056009292603, |
|
"learning_rate": 9.966108724834151e-06, |
|
"logits/chosen": 0.10182631760835648, |
|
"logits/rejected": 0.11173731088638306, |
|
"logps/chosen": -64.47720336914062, |
|
"logps/rejected": -75.30056762695312, |
|
"loss": 2.5889, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8160837292671204, |
|
"rewards/margins": 0.2798304557800293, |
|
"rewards/rejected": -1.0959142446517944, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.15803814713896458, |
|
"grad_norm": 1.1256901025772095, |
|
"learning_rate": 9.965316993154069e-06, |
|
"logits/chosen": 0.1232331395149231, |
|
"logits/rejected": 0.19064000248908997, |
|
"logps/chosen": -64.3778305053711, |
|
"logps/rejected": -76.91246032714844, |
|
"loss": 2.5522, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8341800570487976, |
|
"rewards/margins": 0.2616519033908844, |
|
"rewards/rejected": -1.0958319902420044, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.15985467756584923, |
|
"grad_norm": 1.20338773727417, |
|
"learning_rate": 9.964516155915152e-06, |
|
"logits/chosen": 0.12199988961219788, |
|
"logits/rejected": 0.12264345586299896, |
|
"logps/chosen": -81.56141662597656, |
|
"logps/rejected": -80.49986267089844, |
|
"loss": 2.7432, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -0.7930806875228882, |
|
"rewards/margins": 0.06459490954875946, |
|
"rewards/rejected": -0.8576756119728088, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.16167120799273388, |
|
"grad_norm": 1.005759835243225, |
|
"learning_rate": 9.963706214750446e-06, |
|
"logits/chosen": 0.07040335237979889, |
|
"logits/rejected": 0.10741756111383438, |
|
"logps/chosen": -65.43824768066406, |
|
"logps/rejected": -74.47370147705078, |
|
"loss": 2.5472, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.8203200697898865, |
|
"rewards/margins": 0.24586233496665955, |
|
"rewards/rejected": -1.0661823749542236, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.16348773841961853, |
|
"grad_norm": 0.9761494994163513, |
|
"learning_rate": 9.962887171311563e-06, |
|
"logits/chosen": 0.1630188375711441, |
|
"logits/rejected": 0.190489262342453, |
|
"logps/chosen": -57.84222412109375, |
|
"logps/rejected": -64.116455078125, |
|
"loss": 2.5339, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8240211009979248, |
|
"rewards/margins": 0.21757929027080536, |
|
"rewards/rejected": -1.0416003465652466, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.16530426884650318, |
|
"grad_norm": 1.002192497253418, |
|
"learning_rate": 9.962059027268676e-06, |
|
"logits/chosen": 0.08669686317443848, |
|
"logits/rejected": 0.13606388866901398, |
|
"logps/chosen": -63.68013000488281, |
|
"logps/rejected": -72.30770111083984, |
|
"loss": 2.5165, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.7919780611991882, |
|
"rewards/margins": 0.27081233263015747, |
|
"rewards/rejected": -1.0627902746200562, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.16712079927338783, |
|
"grad_norm": 0.9397479295730591, |
|
"learning_rate": 9.961221784310514e-06, |
|
"logits/chosen": 0.14396092295646667, |
|
"logits/rejected": 0.1820589005947113, |
|
"logps/chosen": -57.551353454589844, |
|
"logps/rejected": -66.83124542236328, |
|
"loss": 2.4254, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.7751535177230835, |
|
"rewards/margins": 0.26002568006515503, |
|
"rewards/rejected": -1.0351792573928833, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.16893732970027248, |
|
"grad_norm": 1.2053430080413818, |
|
"learning_rate": 9.96037544414436e-06, |
|
"logits/chosen": 0.16130733489990234, |
|
"logits/rejected": 0.1852559745311737, |
|
"logps/chosen": -73.8154067993164, |
|
"logps/rejected": -81.3624267578125, |
|
"loss": 2.7543, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.8622196316719055, |
|
"rewards/margins": 0.11067891120910645, |
|
"rewards/rejected": -0.9728984236717224, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.17075386012715713, |
|
"grad_norm": 1.1681262254714966, |
|
"learning_rate": 9.959520008496054e-06, |
|
"logits/chosen": 0.17866984009742737, |
|
"logits/rejected": 0.20266617834568024, |
|
"logps/chosen": -61.40083312988281, |
|
"logps/rejected": -65.13737487792969, |
|
"loss": 2.7134, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9223482012748718, |
|
"rewards/margins": 0.08481010049581528, |
|
"rewards/rejected": -1.0071581602096558, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.17257039055404177, |
|
"grad_norm": 1.0269131660461426, |
|
"learning_rate": 9.95865547910997e-06, |
|
"logits/chosen": 0.11446906626224518, |
|
"logits/rejected": 0.17379909753799438, |
|
"logps/chosen": -62.823944091796875, |
|
"logps/rejected": -69.5682601928711, |
|
"loss": 2.5031, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8894773125648499, |
|
"rewards/margins": 0.24369436502456665, |
|
"rewards/rejected": -1.1331716775894165, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.17438692098092642, |
|
"grad_norm": 1.0400865077972412, |
|
"learning_rate": 9.957781857749043e-06, |
|
"logits/chosen": 0.16556067764759064, |
|
"logits/rejected": 0.23536017537117004, |
|
"logps/chosen": -61.09260940551758, |
|
"logps/rejected": -63.444427490234375, |
|
"loss": 2.7239, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.0037989616394043, |
|
"rewards/margins": 0.07835513353347778, |
|
"rewards/rejected": -1.0821542739868164, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.17620345140781107, |
|
"grad_norm": 1.1138463020324707, |
|
"learning_rate": 9.956899146194732e-06, |
|
"logits/chosen": 0.13205701112747192, |
|
"logits/rejected": 0.1702471375465393, |
|
"logps/chosen": -63.70584487915039, |
|
"logps/rejected": -69.10890197753906, |
|
"loss": 2.7076, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9300363659858704, |
|
"rewards/margins": 0.17512086033821106, |
|
"rewards/rejected": -1.1051572561264038, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.17801998183469572, |
|
"grad_norm": 0.9917119741439819, |
|
"learning_rate": 9.956007346247043e-06, |
|
"logits/chosen": 0.12265195697546005, |
|
"logits/rejected": 0.18780440092086792, |
|
"logps/chosen": -62.316497802734375, |
|
"logps/rejected": -69.26319885253906, |
|
"loss": 2.4996, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8957506418228149, |
|
"rewards/margins": 0.2764662802219391, |
|
"rewards/rejected": -1.1722170114517212, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.17983651226158037, |
|
"grad_norm": 1.00100576877594, |
|
"learning_rate": 9.95510645972451e-06, |
|
"logits/chosen": 0.16969357430934906, |
|
"logits/rejected": 0.16122889518737793, |
|
"logps/chosen": -60.42213439941406, |
|
"logps/rejected": -63.85418701171875, |
|
"loss": 2.5677, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8840566873550415, |
|
"rewards/margins": 0.1611585170030594, |
|
"rewards/rejected": -1.0452152490615845, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.18165304268846502, |
|
"grad_norm": 1.1503106355667114, |
|
"learning_rate": 9.954196488464198e-06, |
|
"logits/chosen": 0.19831174612045288, |
|
"logits/rejected": 0.20889577269554138, |
|
"logps/chosen": -65.17144775390625, |
|
"logps/rejected": -69.9671859741211, |
|
"loss": 2.7065, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.9322465062141418, |
|
"rewards/margins": 0.1645345687866211, |
|
"rewards/rejected": -1.0967810153961182, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.18346957311534967, |
|
"grad_norm": 1.416610836982727, |
|
"learning_rate": 9.953277434321696e-06, |
|
"logits/chosen": 0.08757522702217102, |
|
"logits/rejected": 0.1615956872701645, |
|
"logps/chosen": -72.49248504638672, |
|
"logps/rejected": -80.52176666259766, |
|
"loss": 2.5346, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8279029130935669, |
|
"rewards/margins": 0.26418614387512207, |
|
"rewards/rejected": -1.092089056968689, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.18528610354223432, |
|
"grad_norm": 0.9907792210578918, |
|
"learning_rate": 9.952349299171117e-06, |
|
"logits/chosen": 0.07832024991512299, |
|
"logits/rejected": 0.16370174288749695, |
|
"logps/chosen": -67.5047836303711, |
|
"logps/rejected": -73.78684997558594, |
|
"loss": 2.4325, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8129651546478271, |
|
"rewards/margins": 0.3001910448074341, |
|
"rewards/rejected": -1.1131561994552612, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.18710263396911897, |
|
"grad_norm": 1.1051058769226074, |
|
"learning_rate": 9.95141208490509e-06, |
|
"logits/chosen": 0.10540622472763062, |
|
"logits/rejected": 0.2087182104587555, |
|
"logps/chosen": -55.68672180175781, |
|
"logps/rejected": -72.7021484375, |
|
"loss": 2.3004, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -0.8526840209960938, |
|
"rewards/margins": 0.3691489100456238, |
|
"rewards/rejected": -1.2218331098556519, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.18891916439600362, |
|
"grad_norm": 1.0593968629837036, |
|
"learning_rate": 9.950465793434759e-06, |
|
"logits/chosen": 0.10488969832658768, |
|
"logits/rejected": 0.13292263448238373, |
|
"logps/chosen": -69.47488403320312, |
|
"logps/rejected": -73.08143615722656, |
|
"loss": 2.5388, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.8357015252113342, |
|
"rewards/margins": 0.2204672396183014, |
|
"rewards/rejected": -1.0561686754226685, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.1907356948228883, |
|
"grad_norm": 1.119372010231018, |
|
"learning_rate": 9.949510426689773e-06, |
|
"logits/chosen": 0.06970994919538498, |
|
"logits/rejected": 0.0831708014011383, |
|
"logps/chosen": -66.97635650634766, |
|
"logps/rejected": -67.95066833496094, |
|
"loss": 2.6443, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -0.9023284912109375, |
|
"rewards/margins": 0.1277586966753006, |
|
"rewards/rejected": -1.0300871133804321, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.19255222524977295, |
|
"grad_norm": 1.2842986583709717, |
|
"learning_rate": 9.948545986618295e-06, |
|
"logits/chosen": 0.17146825790405273, |
|
"logits/rejected": 0.1910206824541092, |
|
"logps/chosen": -69.1142578125, |
|
"logps/rejected": -70.71878814697266, |
|
"loss": 2.8494, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.0275481939315796, |
|
"rewards/margins": 0.09755454212427139, |
|
"rewards/rejected": -1.1251027584075928, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.1943687556766576, |
|
"grad_norm": 1.044800877571106, |
|
"learning_rate": 9.947572475186984e-06, |
|
"logits/chosen": 0.1206701397895813, |
|
"logits/rejected": 0.17635390162467957, |
|
"logps/chosen": -67.00174713134766, |
|
"logps/rejected": -71.5347900390625, |
|
"loss": 2.46, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.889815628528595, |
|
"rewards/margins": 0.255011647939682, |
|
"rewards/rejected": -1.1448272466659546, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.19618528610354224, |
|
"grad_norm": 1.1496058702468872, |
|
"learning_rate": 9.946589894381002e-06, |
|
"logits/chosen": 0.18972846865653992, |
|
"logits/rejected": 0.12482471019029617, |
|
"logps/chosen": -66.56938934326172, |
|
"logps/rejected": -73.7935791015625, |
|
"loss": 2.6472, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.8382111191749573, |
|
"rewards/margins": 0.16597144305706024, |
|
"rewards/rejected": -1.0041825771331787, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.1980018165304269, |
|
"grad_norm": 1.1357221603393555, |
|
"learning_rate": 9.945598246204e-06, |
|
"logits/chosen": 0.1766098588705063, |
|
"logits/rejected": 0.16134579479694366, |
|
"logps/chosen": -72.30979919433594, |
|
"logps/rejected": -69.26260375976562, |
|
"loss": 2.7037, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.9316724538803101, |
|
"rewards/margins": 0.0998368114233017, |
|
"rewards/rejected": -1.0315091609954834, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.19981834695731154, |
|
"grad_norm": 1.1147487163543701, |
|
"learning_rate": 9.94459753267812e-06, |
|
"logits/chosen": 0.10100046545267105, |
|
"logits/rejected": 0.12630510330200195, |
|
"logps/chosen": -66.10484313964844, |
|
"logps/rejected": -68.13362121582031, |
|
"loss": 2.6224, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8918489217758179, |
|
"rewards/margins": 0.165731742978096, |
|
"rewards/rejected": -1.0575807094573975, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.2016348773841962, |
|
"grad_norm": 1.3498740196228027, |
|
"learning_rate": 9.943587755843996e-06, |
|
"logits/chosen": 0.19711939990520477, |
|
"logits/rejected": 0.18357205390930176, |
|
"logps/chosen": -64.5146484375, |
|
"logps/rejected": -65.80925750732422, |
|
"loss": 2.6102, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.9744794368743896, |
|
"rewards/margins": 0.16535742580890656, |
|
"rewards/rejected": -1.1398367881774902, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.20345140781108084, |
|
"grad_norm": 1.0210968255996704, |
|
"learning_rate": 9.942568917760733e-06, |
|
"logits/chosen": 0.16077302396297455, |
|
"logits/rejected": 0.22646722197532654, |
|
"logps/chosen": -59.73281478881836, |
|
"logps/rejected": -64.54032897949219, |
|
"loss": 2.4897, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8548685312271118, |
|
"rewards/margins": 0.19910478591918945, |
|
"rewards/rejected": -1.0539731979370117, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.2052679382379655, |
|
"grad_norm": 1.160152554512024, |
|
"learning_rate": 9.941541020505924e-06, |
|
"logits/chosen": 0.10292509198188782, |
|
"logits/rejected": 0.09458380937576294, |
|
"logps/chosen": -77.44420623779297, |
|
"logps/rejected": -85.09677124023438, |
|
"loss": 2.5037, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.9961792230606079, |
|
"rewards/margins": 0.2911871373653412, |
|
"rewards/rejected": -1.287366509437561, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.20708446866485014, |
|
"grad_norm": 1.1742796897888184, |
|
"learning_rate": 9.940504066175626e-06, |
|
"logits/chosen": 0.10538715869188309, |
|
"logits/rejected": 0.07242526113986969, |
|
"logps/chosen": -70.37040710449219, |
|
"logps/rejected": -68.23387908935547, |
|
"loss": 2.6328, |
|
"rewards/accuracies": 0.5, |
|
"rewards/chosen": -0.9170363545417786, |
|
"rewards/margins": 0.14501458406448364, |
|
"rewards/rejected": -1.0620509386062622, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.2089009990917348, |
|
"grad_norm": 1.1259719133377075, |
|
"learning_rate": 9.939458056884375e-06, |
|
"logits/chosen": 0.12264375388622284, |
|
"logits/rejected": 0.16776585578918457, |
|
"logps/chosen": -59.847049713134766, |
|
"logps/rejected": -65.45285034179688, |
|
"loss": 2.5752, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.8872510194778442, |
|
"rewards/margins": 0.2041575014591217, |
|
"rewards/rejected": -1.0914084911346436, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.21071752951861944, |
|
"grad_norm": 2.1809544563293457, |
|
"learning_rate": 9.938402994765163e-06, |
|
"logits/chosen": 0.13642175495624542, |
|
"logits/rejected": 0.11773751676082611, |
|
"logps/chosen": -65.34564971923828, |
|
"logps/rejected": -66.55570983886719, |
|
"loss": 2.6792, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9432386755943298, |
|
"rewards/margins": 0.12600603699684143, |
|
"rewards/rejected": -1.0692447423934937, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.2125340599455041, |
|
"grad_norm": 1.4895256757736206, |
|
"learning_rate": 9.937338881969444e-06, |
|
"logits/chosen": 0.12349803745746613, |
|
"logits/rejected": 0.1182522252202034, |
|
"logps/chosen": -66.9234619140625, |
|
"logps/rejected": -69.5535888671875, |
|
"loss": 2.6029, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.9096715450286865, |
|
"rewards/margins": 0.23732726275920868, |
|
"rewards/rejected": -1.1469988822937012, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.21435059037238874, |
|
"grad_norm": 1.0333324670791626, |
|
"learning_rate": 9.93626572066713e-06, |
|
"logits/chosen": 0.16722331941127777, |
|
"logits/rejected": 0.21501797437667847, |
|
"logps/chosen": -66.49773406982422, |
|
"logps/rejected": -72.3947982788086, |
|
"loss": 2.4687, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -0.9357748627662659, |
|
"rewards/margins": 0.2686734199523926, |
|
"rewards/rejected": -1.2044482231140137, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.2161671207992734, |
|
"grad_norm": 1.2127379179000854, |
|
"learning_rate": 9.935183513046585e-06, |
|
"logits/chosen": 0.10065922141075134, |
|
"logits/rejected": 0.12764661014080048, |
|
"logps/chosen": -73.05583953857422, |
|
"logps/rejected": -77.18297576904297, |
|
"loss": 2.6288, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -0.8347401022911072, |
|
"rewards/margins": 0.2145780473947525, |
|
"rewards/rejected": -1.0493181943893433, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.21798365122615804, |
|
"grad_norm": 1.4050469398498535, |
|
"learning_rate": 9.93409226131462e-06, |
|
"logits/chosen": 0.08468589186668396, |
|
"logits/rejected": 0.179460808634758, |
|
"logps/chosen": -66.12811279296875, |
|
"logps/rejected": -74.2131118774414, |
|
"loss": 2.4658, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -0.8611673712730408, |
|
"rewards/margins": 0.20363186299800873, |
|
"rewards/rejected": -1.064799189567566, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.2198001816530427, |
|
"grad_norm": 1.18356192111969, |
|
"learning_rate": 9.932991967696484e-06, |
|
"logits/chosen": 0.18308167159557343, |
|
"logits/rejected": 0.12839025259017944, |
|
"logps/chosen": -63.325035095214844, |
|
"logps/rejected": -63.72959518432617, |
|
"loss": 2.579, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9539028406143188, |
|
"rewards/margins": 0.16736355423927307, |
|
"rewards/rejected": -1.1212663650512695, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.22161671207992734, |
|
"grad_norm": 1.1829110383987427, |
|
"learning_rate": 9.931882634435868e-06, |
|
"logits/chosen": 0.10780126601457596, |
|
"logits/rejected": 0.109224833548069, |
|
"logps/chosen": -71.1589126586914, |
|
"logps/rejected": -74.609375, |
|
"loss": 2.6368, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.9908429384231567, |
|
"rewards/margins": 0.12795159220695496, |
|
"rewards/rejected": -1.118794560432434, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.22343324250681199, |
|
"grad_norm": 1.1116541624069214, |
|
"learning_rate": 9.930764263794898e-06, |
|
"logits/chosen": 0.10946042090654373, |
|
"logits/rejected": 0.11234283447265625, |
|
"logps/chosen": -67.98091888427734, |
|
"logps/rejected": -70.58726501464844, |
|
"loss": 2.4417, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -0.871981143951416, |
|
"rewards/margins": 0.22934575378894806, |
|
"rewards/rejected": -1.1013269424438477, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.22524977293369663, |
|
"grad_norm": 1.1564345359802246, |
|
"learning_rate": 9.929636858054124e-06, |
|
"logits/chosen": 0.11665979772806168, |
|
"logits/rejected": 0.15129488706588745, |
|
"logps/chosen": -62.33778381347656, |
|
"logps/rejected": -67.58541870117188, |
|
"loss": 2.6023, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.8282152414321899, |
|
"rewards/margins": 0.206033393740654, |
|
"rewards/rejected": -1.0342485904693604, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.22706630336058128, |
|
"grad_norm": 1.149032711982727, |
|
"learning_rate": 9.928500419512522e-06, |
|
"logits/chosen": 0.02143971435725689, |
|
"logits/rejected": 0.10582254827022552, |
|
"logps/chosen": -66.31533813476562, |
|
"logps/rejected": -76.4206771850586, |
|
"loss": 2.4589, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -0.8935360908508301, |
|
"rewards/margins": 0.3176972270011902, |
|
"rewards/rejected": -1.2112332582473755, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.22888283378746593, |
|
"grad_norm": 1.1076140403747559, |
|
"learning_rate": 9.927354950487491e-06, |
|
"logits/chosen": 0.13772962987422943, |
|
"logits/rejected": 0.12653110921382904, |
|
"logps/chosen": -62.97186279296875, |
|
"logps/rejected": -65.31454467773438, |
|
"loss": 2.4839, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -0.9201483130455017, |
|
"rewards/margins": 0.2287607342004776, |
|
"rewards/rejected": -1.1489089727401733, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.23069936421435058, |
|
"grad_norm": 1.1476696729660034, |
|
"learning_rate": 9.926200453314842e-06, |
|
"logits/chosen": 0.11984287202358246, |
|
"logits/rejected": 0.1570628136396408, |
|
"logps/chosen": -72.74269104003906, |
|
"logps/rejected": -77.52953338623047, |
|
"loss": 2.5198, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -0.9817119836807251, |
|
"rewards/margins": 0.2599087059497833, |
|
"rewards/rejected": -1.241620659828186, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.23251589464123523, |
|
"grad_norm": 1.0800604820251465, |
|
"learning_rate": 9.925036930348793e-06, |
|
"logits/chosen": 0.03499937057495117, |
|
"logits/rejected": 0.16788874566555023, |
|
"logps/chosen": -64.4096450805664, |
|
"logps/rejected": -81.16387939453125, |
|
"loss": 2.3547, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.040608286857605, |
|
"rewards/margins": 0.48978835344314575, |
|
"rewards/rejected": -1.530396580696106, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.23433242506811988, |
|
"grad_norm": 1.2677640914916992, |
|
"learning_rate": 9.92386438396197e-06, |
|
"logits/chosen": 0.1200772225856781, |
|
"logits/rejected": 0.11345633864402771, |
|
"logps/chosen": -68.35733032226562, |
|
"logps/rejected": -71.72528839111328, |
|
"loss": 2.631, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -0.9437850117683411, |
|
"rewards/margins": 0.13758014142513275, |
|
"rewards/rejected": -1.0813652276992798, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.23614895549500453, |
|
"grad_norm": 1.2041019201278687, |
|
"learning_rate": 9.9226828165454e-06, |
|
"logits/chosen": 0.07035915553569794, |
|
"logits/rejected": 0.16790318489074707, |
|
"logps/chosen": -73.99917602539062, |
|
"logps/rejected": -83.07939147949219, |
|
"loss": 2.4497, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0043660402297974, |
|
"rewards/margins": 0.36712244153022766, |
|
"rewards/rejected": -1.3714885711669922, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.23796548592188918, |
|
"grad_norm": 1.112176537513733, |
|
"learning_rate": 9.921492230508507e-06, |
|
"logits/chosen": 0.18427804112434387, |
|
"logits/rejected": 0.17424902319908142, |
|
"logps/chosen": -63.21034240722656, |
|
"logps/rejected": -65.23414611816406, |
|
"loss": 2.4612, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -0.9832345843315125, |
|
"rewards/margins": 0.21530815958976746, |
|
"rewards/rejected": -1.1985427141189575, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.23978201634877383, |
|
"grad_norm": 1.1540772914886475, |
|
"learning_rate": 9.9202926282791e-06, |
|
"logits/chosen": 0.14970409870147705, |
|
"logits/rejected": 0.17991000413894653, |
|
"logps/chosen": -76.66849517822266, |
|
"logps/rejected": -85.59910583496094, |
|
"loss": 2.4164, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0053367614746094, |
|
"rewards/margins": 0.3513883352279663, |
|
"rewards/rejected": -1.3567250967025757, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.24159854677565848, |
|
"grad_norm": 1.32212495803833, |
|
"learning_rate": 9.919084012303378e-06, |
|
"logits/chosen": 0.11897419393062592, |
|
"logits/rejected": 0.15737299621105194, |
|
"logps/chosen": -75.7364501953125, |
|
"logps/rejected": -77.97785186767578, |
|
"loss": 2.7243, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.139754295349121, |
|
"rewards/margins": 0.17446394264698029, |
|
"rewards/rejected": -1.3142181634902954, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.24341507720254316, |
|
"grad_norm": 1.4282020330429077, |
|
"learning_rate": 9.917866385045918e-06, |
|
"logits/chosen": 0.13694903254508972, |
|
"logits/rejected": 0.2048657089471817, |
|
"logps/chosen": -70.1098403930664, |
|
"logps/rejected": -78.7418212890625, |
|
"loss": 2.8086, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.2937737703323364, |
|
"rewards/margins": 0.2787961959838867, |
|
"rewards/rejected": -1.5725698471069336, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2452316076294278, |
|
"grad_norm": 1.1548171043395996, |
|
"learning_rate": 9.916639748989677e-06, |
|
"logits/chosen": 0.050287194550037384, |
|
"logits/rejected": 0.10146909952163696, |
|
"logps/chosen": -72.51569366455078, |
|
"logps/rejected": -80.42554473876953, |
|
"loss": 2.4774, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0650720596313477, |
|
"rewards/margins": 0.2093387395143509, |
|
"rewards/rejected": -1.2744107246398926, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.24704813805631246, |
|
"grad_norm": 1.1347655057907104, |
|
"learning_rate": 9.915404106635979e-06, |
|
"logits/chosen": 0.13506914675235748, |
|
"logits/rejected": 0.18150334060192108, |
|
"logps/chosen": -71.03006744384766, |
|
"logps/rejected": -75.59716033935547, |
|
"loss": 2.4399, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1068118810653687, |
|
"rewards/margins": 0.3031606674194336, |
|
"rewards/rejected": -1.4099724292755127, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2488646684831971, |
|
"grad_norm": 1.1634560823440552, |
|
"learning_rate": 9.914159460504512e-06, |
|
"logits/chosen": 0.17681393027305603, |
|
"logits/rejected": 0.2249586582183838, |
|
"logps/chosen": -66.17890167236328, |
|
"logps/rejected": -71.03226470947266, |
|
"loss": 2.5281, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0759899616241455, |
|
"rewards/margins": 0.21175377070903778, |
|
"rewards/rejected": -1.2877436876296997, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.2506811989100817, |
|
"grad_norm": 2.4183638095855713, |
|
"learning_rate": 9.912905813133325e-06, |
|
"logits/chosen": 0.053306616842746735, |
|
"logits/rejected": 0.08808214217424393, |
|
"logps/chosen": -75.88124084472656, |
|
"logps/rejected": -86.6006088256836, |
|
"loss": 2.5718, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.1421475410461426, |
|
"rewards/margins": 0.36426350474357605, |
|
"rewards/rejected": -1.506411075592041, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.2524977293369664, |
|
"grad_norm": 1.1398063898086548, |
|
"learning_rate": 9.911643167078827e-06, |
|
"logits/chosen": 0.046535998582839966, |
|
"logits/rejected": 0.11657831072807312, |
|
"logps/chosen": -62.68219757080078, |
|
"logps/rejected": -74.52103424072266, |
|
"loss": 2.5334, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.0171339511871338, |
|
"rewards/margins": 0.3033108115196228, |
|
"rewards/rejected": -1.3204445838928223, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.254314259763851, |
|
"grad_norm": 1.2559521198272705, |
|
"learning_rate": 9.91037152491577e-06, |
|
"logits/chosen": 0.07763661444187164, |
|
"logits/rejected": 0.11103180050849915, |
|
"logps/chosen": -68.87091064453125, |
|
"logps/rejected": -79.08797454833984, |
|
"loss": 2.454, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.017935872077942, |
|
"rewards/margins": 0.2824122905731201, |
|
"rewards/rejected": -1.300348162651062, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.2561307901907357, |
|
"grad_norm": 1.1818301677703857, |
|
"learning_rate": 9.909090889237257e-06, |
|
"logits/chosen": 0.07567673176527023, |
|
"logits/rejected": 0.1396160125732422, |
|
"logps/chosen": -62.44426345825195, |
|
"logps/rejected": -61.77680206298828, |
|
"loss": 2.4511, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -0.9497895240783691, |
|
"rewards/margins": 0.23719964921474457, |
|
"rewards/rejected": -1.1869890689849854, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.2579473206176203, |
|
"grad_norm": 1.1855460405349731, |
|
"learning_rate": 9.907801262654725e-06, |
|
"logits/chosen": 0.08448053896427155, |
|
"logits/rejected": 0.1705108880996704, |
|
"logps/chosen": -66.66730499267578, |
|
"logps/rejected": -76.75447845458984, |
|
"loss": 2.4476, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.105279803276062, |
|
"rewards/margins": 0.2967475950717926, |
|
"rewards/rejected": -1.4020274877548218, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.259763851044505, |
|
"grad_norm": 1.3652665615081787, |
|
"learning_rate": 9.906502647797946e-06, |
|
"logits/chosen": 0.09915038198232651, |
|
"logits/rejected": 0.08110683411359787, |
|
"logps/chosen": -66.32193756103516, |
|
"logps/rejected": -80.20841217041016, |
|
"loss": 2.5982, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -1.1550779342651367, |
|
"rewards/margins": 0.36828702688217163, |
|
"rewards/rejected": -1.5233650207519531, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.2615803814713896, |
|
"grad_norm": 1.4188799858093262, |
|
"learning_rate": 9.905195047315024e-06, |
|
"logits/chosen": 0.12019304931163788, |
|
"logits/rejected": 0.11188551783561707, |
|
"logps/chosen": -85.44612884521484, |
|
"logps/rejected": -91.55411529541016, |
|
"loss": 2.8244, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -1.171909213066101, |
|
"rewards/margins": 0.16801074147224426, |
|
"rewards/rejected": -1.3399198055267334, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.2633969118982743, |
|
"grad_norm": 1.307726502418518, |
|
"learning_rate": 9.903878463872384e-06, |
|
"logits/chosen": 0.12220380455255508, |
|
"logits/rejected": 0.08115807920694351, |
|
"logps/chosen": -71.56598663330078, |
|
"logps/rejected": -71.203857421875, |
|
"loss": 2.5832, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.1082667112350464, |
|
"rewards/margins": 0.18824435770511627, |
|
"rewards/rejected": -1.296510934829712, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.2652134423251589, |
|
"grad_norm": 1.0912556648254395, |
|
"learning_rate": 9.902552900154769e-06, |
|
"logits/chosen": 0.14343701303005219, |
|
"logits/rejected": 0.17181995511054993, |
|
"logps/chosen": -65.52398681640625, |
|
"logps/rejected": -72.67620086669922, |
|
"loss": 2.3888, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0277646780014038, |
|
"rewards/margins": 0.34770333766937256, |
|
"rewards/rejected": -1.375468134880066, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.2670299727520436, |
|
"grad_norm": 1.375835657119751, |
|
"learning_rate": 9.90121835886523e-06, |
|
"logits/chosen": 0.06410901993513107, |
|
"logits/rejected": 0.09152361750602722, |
|
"logps/chosen": -78.27774047851562, |
|
"logps/rejected": -82.27354431152344, |
|
"loss": 2.5421, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.0331169366836548, |
|
"rewards/margins": 0.20058104395866394, |
|
"rewards/rejected": -1.2336980104446411, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.2688465031789282, |
|
"grad_norm": 1.1775563955307007, |
|
"learning_rate": 9.899874842725136e-06, |
|
"logits/chosen": 0.15871602296829224, |
|
"logits/rejected": 0.12121336162090302, |
|
"logps/chosen": -68.25074005126953, |
|
"logps/rejected": -69.08786010742188, |
|
"loss": 2.5082, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.1035441160202026, |
|
"rewards/margins": 0.2747136950492859, |
|
"rewards/rejected": -1.3782578706741333, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.2706630336058129, |
|
"grad_norm": 1.1836682558059692, |
|
"learning_rate": 9.898522354474144e-06, |
|
"logits/chosen": 0.10703336447477341, |
|
"logits/rejected": 0.09559071063995361, |
|
"logps/chosen": -67.44327545166016, |
|
"logps/rejected": -66.1326904296875, |
|
"loss": 2.4807, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0860165357589722, |
|
"rewards/margins": 0.24197286367416382, |
|
"rewards/rejected": -1.3279893398284912, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.2724795640326976, |
|
"grad_norm": 1.4430557489395142, |
|
"learning_rate": 9.897160896870217e-06, |
|
"logits/chosen": 0.1853700429201126, |
|
"logits/rejected": 0.2392357438802719, |
|
"logps/chosen": -69.39933776855469, |
|
"logps/rejected": -76.33921813964844, |
|
"loss": 2.6237, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.2587306499481201, |
|
"rewards/margins": 0.1618267297744751, |
|
"rewards/rejected": -1.4205572605133057, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2742960944595822, |
|
"grad_norm": 1.3623749017715454, |
|
"learning_rate": 9.895790472689605e-06, |
|
"logits/chosen": 0.15992893278598785, |
|
"logits/rejected": 0.09628183394670486, |
|
"logps/chosen": -70.02740478515625, |
|
"logps/rejected": -66.36286926269531, |
|
"loss": 2.7214, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.1345252990722656, |
|
"rewards/margins": 0.1051594465970993, |
|
"rewards/rejected": -1.2396849393844604, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.2761126248864669, |
|
"grad_norm": 1.3764369487762451, |
|
"learning_rate": 9.894411084726837e-06, |
|
"logits/chosen": 0.14009161293506622, |
|
"logits/rejected": 0.13651950657367706, |
|
"logps/chosen": -74.57782745361328, |
|
"logps/rejected": -79.09630584716797, |
|
"loss": 2.5094, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.1602882146835327, |
|
"rewards/margins": 0.3822028338909149, |
|
"rewards/rejected": -1.54249107837677, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.2779291553133515, |
|
"grad_norm": 1.3273645639419556, |
|
"learning_rate": 9.893022735794728e-06, |
|
"logits/chosen": 0.03166107460856438, |
|
"logits/rejected": 0.059291813522577286, |
|
"logps/chosen": -75.47296142578125, |
|
"logps/rejected": -88.63102722167969, |
|
"loss": 2.4789, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -1.10808265209198, |
|
"rewards/margins": 0.3505283296108246, |
|
"rewards/rejected": -1.4586111307144165, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.27974568574023617, |
|
"grad_norm": 1.1496751308441162, |
|
"learning_rate": 9.891625428724365e-06, |
|
"logits/chosen": 0.13897705078125, |
|
"logits/rejected": 0.13928522169589996, |
|
"logps/chosen": -65.34259796142578, |
|
"logps/rejected": -68.99885559082031, |
|
"loss": 2.2804, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.1391716003417969, |
|
"rewards/margins": 0.3395899534225464, |
|
"rewards/rejected": -1.4787613153457642, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.2815622161671208, |
|
"grad_norm": 1.3669211864471436, |
|
"learning_rate": 9.890219166365097e-06, |
|
"logits/chosen": 0.08065556734800339, |
|
"logits/rejected": 0.14106576144695282, |
|
"logps/chosen": -72.62825012207031, |
|
"logps/rejected": -77.62796020507812, |
|
"loss": 2.5929, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.0997321605682373, |
|
"rewards/margins": 0.20812571048736572, |
|
"rewards/rejected": -1.3078577518463135, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.28337874659400547, |
|
"grad_norm": 1.2757997512817383, |
|
"learning_rate": 9.888803951584537e-06, |
|
"logits/chosen": 0.0465204194188118, |
|
"logits/rejected": 0.06447532027959824, |
|
"logps/chosen": -66.84536743164062, |
|
"logps/rejected": -78.97770690917969, |
|
"loss": 2.3491, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.1228134632110596, |
|
"rewards/margins": 0.4900767207145691, |
|
"rewards/rejected": -1.612890362739563, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.2851952770208901, |
|
"grad_norm": 1.391892910003662, |
|
"learning_rate": 9.887379787268558e-06, |
|
"logits/chosen": 0.08740498870611191, |
|
"logits/rejected": 0.11772032082080841, |
|
"logps/chosen": -67.31288146972656, |
|
"logps/rejected": -70.06088256835938, |
|
"loss": 2.6179, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.2256510257720947, |
|
"rewards/margins": 0.19449593126773834, |
|
"rewards/rejected": -1.4201467037200928, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.28701180744777477, |
|
"grad_norm": 1.445900321006775, |
|
"learning_rate": 9.885946676321279e-06, |
|
"logits/chosen": 0.1524508148431778, |
|
"logits/rejected": 0.21413244307041168, |
|
"logps/chosen": -74.10218811035156, |
|
"logps/rejected": -76.91876220703125, |
|
"loss": 2.5262, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.1993122100830078, |
|
"rewards/margins": 0.33456599712371826, |
|
"rewards/rejected": -1.5338780879974365, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.2888283378746594, |
|
"grad_norm": 1.361208438873291, |
|
"learning_rate": 9.884504621665059e-06, |
|
"logits/chosen": 0.1192079707980156, |
|
"logits/rejected": 0.21478833258152008, |
|
"logps/chosen": -75.69478607177734, |
|
"logps/rejected": -84.40799713134766, |
|
"loss": 2.3984, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.2067720890045166, |
|
"rewards/margins": 0.34302443265914917, |
|
"rewards/rejected": -1.5497965812683105, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.29064486830154407, |
|
"grad_norm": 3.4042046070098877, |
|
"learning_rate": 9.883053626240503e-06, |
|
"logits/chosen": 0.13823604583740234, |
|
"logits/rejected": 0.0957983061671257, |
|
"logps/chosen": -85.02832794189453, |
|
"logps/rejected": -86.334716796875, |
|
"loss": 3.0326, |
|
"rewards/accuracies": 0.484375, |
|
"rewards/chosen": -1.1793922185897827, |
|
"rewards/margins": 0.013544075191020966, |
|
"rewards/rejected": -1.1929364204406738, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.2924613987284287, |
|
"grad_norm": 1.7824169397354126, |
|
"learning_rate": 9.881593693006438e-06, |
|
"logits/chosen": 0.09778247773647308, |
|
"logits/rejected": 0.12810076773166656, |
|
"logps/chosen": -78.78877258300781, |
|
"logps/rejected": -78.63519287109375, |
|
"loss": 3.1665, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -1.360141396522522, |
|
"rewards/margins": 0.148757204413414, |
|
"rewards/rejected": -1.5088986158370972, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.29427792915531337, |
|
"grad_norm": 1.4275903701782227, |
|
"learning_rate": 9.880124824939927e-06, |
|
"logits/chosen": 0.17759747803211212, |
|
"logits/rejected": 0.14611241221427917, |
|
"logps/chosen": -70.13676452636719, |
|
"logps/rejected": -67.08554077148438, |
|
"loss": 2.8382, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.1756031513214111, |
|
"rewards/margins": 0.14519944787025452, |
|
"rewards/rejected": -1.3208026885986328, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.296094459582198, |
|
"grad_norm": 1.3242970705032349, |
|
"learning_rate": 9.878647025036245e-06, |
|
"logits/chosen": 0.07917390763759613, |
|
"logits/rejected": 0.16502070426940918, |
|
"logps/chosen": -77.08013916015625, |
|
"logps/rejected": -92.47212982177734, |
|
"loss": 2.3183, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.2245774269104004, |
|
"rewards/margins": 0.573800802230835, |
|
"rewards/rejected": -1.7983782291412354, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.29791099000908267, |
|
"grad_norm": 1.160335659980774, |
|
"learning_rate": 9.877160296308886e-06, |
|
"logits/chosen": 0.06782057881355286, |
|
"logits/rejected": 0.09083382785320282, |
|
"logps/chosen": -61.769866943359375, |
|
"logps/rejected": -66.22183990478516, |
|
"loss": 2.3323, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.2067499160766602, |
|
"rewards/margins": 0.36826351284980774, |
|
"rewards/rejected": -1.575013279914856, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.2997275204359673, |
|
"grad_norm": 1.2561142444610596, |
|
"learning_rate": 9.875664641789545e-06, |
|
"logits/chosen": 0.10073137283325195, |
|
"logits/rejected": 0.12745651602745056, |
|
"logps/chosen": -68.94499206542969, |
|
"logps/rejected": -76.13021850585938, |
|
"loss": 2.4058, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.2001702785491943, |
|
"rewards/margins": 0.31210076808929443, |
|
"rewards/rejected": -1.5122709274291992, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.30154405086285196, |
|
"grad_norm": 1.2163783311843872, |
|
"learning_rate": 9.874160064528124e-06, |
|
"logits/chosen": 0.1380203366279602, |
|
"logits/rejected": 0.20783495903015137, |
|
"logps/chosen": -62.44160461425781, |
|
"logps/rejected": -70.53887176513672, |
|
"loss": 2.2759, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.243611216545105, |
|
"rewards/margins": 0.46271997690200806, |
|
"rewards/rejected": -1.7063312530517578, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.3033605812897366, |
|
"grad_norm": 1.2518328428268433, |
|
"learning_rate": 9.872646567592719e-06, |
|
"logits/chosen": 0.13933810591697693, |
|
"logits/rejected": 0.1436997652053833, |
|
"logps/chosen": -69.45280456542969, |
|
"logps/rejected": -78.73098754882812, |
|
"loss": 2.368, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1905428171157837, |
|
"rewards/margins": 0.42389774322509766, |
|
"rewards/rejected": -1.6144405603408813, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.30517711171662126, |
|
"grad_norm": 1.2762094736099243, |
|
"learning_rate": 9.871124154069613e-06, |
|
"logits/chosen": 0.11822449415922165, |
|
"logits/rejected": 0.13434451818466187, |
|
"logps/chosen": -72.77944946289062, |
|
"logps/rejected": -75.60210418701172, |
|
"loss": 2.5178, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.304598331451416, |
|
"rewards/margins": 0.21065138280391693, |
|
"rewards/rejected": -1.5152498483657837, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.3069936421435059, |
|
"grad_norm": 1.7200556993484497, |
|
"learning_rate": 9.86959282706327e-06, |
|
"logits/chosen": 0.12234638631343842, |
|
"logits/rejected": 0.123184435069561, |
|
"logps/chosen": -84.78328704833984, |
|
"logps/rejected": -82.90792083740234, |
|
"loss": 2.7762, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.4255175590515137, |
|
"rewards/margins": 0.3239808976650238, |
|
"rewards/rejected": -1.7494984865188599, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.30881017257039056, |
|
"grad_norm": 1.2458773851394653, |
|
"learning_rate": 9.868052589696337e-06, |
|
"logits/chosen": 0.14292597770690918, |
|
"logits/rejected": 0.1542571634054184, |
|
"logps/chosen": -66.72957611083984, |
|
"logps/rejected": -76.97203826904297, |
|
"loss": 2.419, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.2561144828796387, |
|
"rewards/margins": 0.39490899443626404, |
|
"rewards/rejected": -1.6510233879089355, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.3106267029972752, |
|
"grad_norm": 1.192762017250061, |
|
"learning_rate": 9.866503445109621e-06, |
|
"logits/chosen": 0.12845008075237274, |
|
"logits/rejected": 0.10619282722473145, |
|
"logps/chosen": -64.939453125, |
|
"logps/rejected": -70.89356231689453, |
|
"loss": 2.3475, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.2126998901367188, |
|
"rewards/margins": 0.39672625064849854, |
|
"rewards/rejected": -1.6094262599945068, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.31244323342415986, |
|
"grad_norm": 1.6527279615402222, |
|
"learning_rate": 9.864945396462101e-06, |
|
"logits/chosen": 0.03876817971467972, |
|
"logits/rejected": 0.01723310723900795, |
|
"logps/chosen": -80.94947052001953, |
|
"logps/rejected": -83.43637084960938, |
|
"loss": 2.6719, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -1.203782320022583, |
|
"rewards/margins": 0.2551845610141754, |
|
"rewards/rejected": -1.458966851234436, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.3142597638510445, |
|
"grad_norm": 1.427216649055481, |
|
"learning_rate": 9.86337844693091e-06, |
|
"logits/chosen": 0.08040126413106918, |
|
"logits/rejected": 0.13297931849956512, |
|
"logps/chosen": -67.49502563476562, |
|
"logps/rejected": -77.2923812866211, |
|
"loss": 2.4931, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.3284540176391602, |
|
"rewards/margins": 0.4811919927597046, |
|
"rewards/rejected": -1.8096460103988647, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.31607629427792916, |
|
"grad_norm": 1.3090578317642212, |
|
"learning_rate": 9.861802599711329e-06, |
|
"logits/chosen": 0.109119713306427, |
|
"logits/rejected": 0.07613471150398254, |
|
"logps/chosen": -71.72999572753906, |
|
"logps/rejected": -73.84215545654297, |
|
"loss": 2.4972, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.2370662689208984, |
|
"rewards/margins": 0.27422747015953064, |
|
"rewards/rejected": -1.5112937688827515, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.3178928247048138, |
|
"grad_norm": 1.351342797279358, |
|
"learning_rate": 9.860217858016783e-06, |
|
"logits/chosen": 0.10673967003822327, |
|
"logits/rejected": 0.1354019045829773, |
|
"logps/chosen": -70.85772705078125, |
|
"logps/rejected": -81.281982421875, |
|
"loss": 2.4456, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.2566276788711548, |
|
"rewards/margins": 0.34767431020736694, |
|
"rewards/rejected": -1.604301929473877, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.31970935513169846, |
|
"grad_norm": 1.3725523948669434, |
|
"learning_rate": 9.858624225078841e-06, |
|
"logits/chosen": 0.12088489532470703, |
|
"logits/rejected": 0.0771353617310524, |
|
"logps/chosen": -74.0596694946289, |
|
"logps/rejected": -71.7733154296875, |
|
"loss": 2.5003, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1832196712493896, |
|
"rewards/margins": 0.31638142466545105, |
|
"rewards/rejected": -1.4996010065078735, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.3215258855585831, |
|
"grad_norm": 1.3968805074691772, |
|
"learning_rate": 9.857021704147195e-06, |
|
"logits/chosen": 0.11635589599609375, |
|
"logits/rejected": 0.09500478953123093, |
|
"logps/chosen": -75.86662292480469, |
|
"logps/rejected": -75.28397369384766, |
|
"loss": 2.5589, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.3443347215652466, |
|
"rewards/margins": 0.22420868277549744, |
|
"rewards/rejected": -1.5685434341430664, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.32334241598546776, |
|
"grad_norm": 1.3207515478134155, |
|
"learning_rate": 9.855410298489663e-06, |
|
"logits/chosen": 0.032826680690050125, |
|
"logits/rejected": 0.0877794623374939, |
|
"logps/chosen": -67.0349349975586, |
|
"logps/rejected": -75.06578063964844, |
|
"loss": 2.4595, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.2243653535842896, |
|
"rewards/margins": 0.36731529235839844, |
|
"rewards/rejected": -1.591680645942688, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.32515894641235243, |
|
"grad_norm": 1.3098173141479492, |
|
"learning_rate": 9.853790011392186e-06, |
|
"logits/chosen": 0.08737780898809433, |
|
"logits/rejected": 0.08714289963245392, |
|
"logps/chosen": -71.3626708984375, |
|
"logps/rejected": -81.07239532470703, |
|
"loss": 2.2606, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.1984796524047852, |
|
"rewards/margins": 0.4800388514995575, |
|
"rewards/rejected": -1.678518533706665, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.32697547683923706, |
|
"grad_norm": 1.3656742572784424, |
|
"learning_rate": 9.852160846158808e-06, |
|
"logits/chosen": 0.08435464650392532, |
|
"logits/rejected": 0.13674329221248627, |
|
"logps/chosen": -69.021484375, |
|
"logps/rejected": -79.02738952636719, |
|
"loss": 2.4039, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.4128899574279785, |
|
"rewards/margins": 0.3882848620414734, |
|
"rewards/rejected": -1.8011748790740967, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.32879200726612173, |
|
"grad_norm": 1.2837380170822144, |
|
"learning_rate": 9.850522806111681e-06, |
|
"logits/chosen": 0.20875662565231323, |
|
"logits/rejected": 0.1668507307767868, |
|
"logps/chosen": -68.29991149902344, |
|
"logps/rejected": -67.59306335449219, |
|
"loss": 2.5716, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -1.3281925916671753, |
|
"rewards/margins": 0.21396151185035706, |
|
"rewards/rejected": -1.54215407371521, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.33060853769300635, |
|
"grad_norm": 1.3804494142532349, |
|
"learning_rate": 9.848875894591055e-06, |
|
"logits/chosen": 0.12790340185165405, |
|
"logits/rejected": 0.1156705766916275, |
|
"logps/chosen": -75.76934814453125, |
|
"logps/rejected": -76.44409942626953, |
|
"loss": 2.2623, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.260763168334961, |
|
"rewards/margins": 0.45292162895202637, |
|
"rewards/rejected": -1.7136849164962769, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.33242506811989103, |
|
"grad_norm": 1.287442922592163, |
|
"learning_rate": 9.847220114955269e-06, |
|
"logits/chosen": 0.1627904772758484, |
|
"logits/rejected": 0.1957186460494995, |
|
"logps/chosen": -67.52108001708984, |
|
"logps/rejected": -79.52376556396484, |
|
"loss": 2.348, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.2697941064834595, |
|
"rewards/margins": 0.5318787097930908, |
|
"rewards/rejected": -1.8016728162765503, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.33424159854677565, |
|
"grad_norm": 1.3078449964523315, |
|
"learning_rate": 9.845555470580746e-06, |
|
"logits/chosen": 0.08075303584337234, |
|
"logits/rejected": 0.08640636503696442, |
|
"logps/chosen": -64.12767791748047, |
|
"logps/rejected": -69.04762268066406, |
|
"loss": 2.4969, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -1.2283066511154175, |
|
"rewards/margins": 0.31595146656036377, |
|
"rewards/rejected": -1.5442581176757812, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.33605812897366033, |
|
"grad_norm": 1.649794578552246, |
|
"learning_rate": 9.843881964861985e-06, |
|
"logits/chosen": 0.07760760188102722, |
|
"logits/rejected": 0.11186876147985458, |
|
"logps/chosen": -75.8207015991211, |
|
"logps/rejected": -81.5796890258789, |
|
"loss": 2.5595, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.4034444093704224, |
|
"rewards/margins": 0.33963268995285034, |
|
"rewards/rejected": -1.743077039718628, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.33787465940054495, |
|
"grad_norm": 1.2912336587905884, |
|
"learning_rate": 9.842199601211556e-06, |
|
"logits/chosen": 0.133261039853096, |
|
"logits/rejected": 0.1473626047372818, |
|
"logps/chosen": -69.50003051757812, |
|
"logps/rejected": -77.0198745727539, |
|
"loss": 2.3544, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.2854596376419067, |
|
"rewards/margins": 0.4699360728263855, |
|
"rewards/rejected": -1.7553956508636475, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.33969118982742963, |
|
"grad_norm": 1.4907958507537842, |
|
"learning_rate": 9.840508383060092e-06, |
|
"logits/chosen": 0.06240752339363098, |
|
"logits/rejected": 0.15645891427993774, |
|
"logps/chosen": -66.49507141113281, |
|
"logps/rejected": -72.68167114257812, |
|
"loss": 2.5754, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.3408069610595703, |
|
"rewards/margins": 0.2825961112976074, |
|
"rewards/rejected": -1.6234029531478882, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.34150772025431425, |
|
"grad_norm": 1.528688669204712, |
|
"learning_rate": 9.838808313856281e-06, |
|
"logits/chosen": -0.010553614236414433, |
|
"logits/rejected": -0.0162151250988245, |
|
"logps/chosen": -87.83446502685547, |
|
"logps/rejected": -82.62303161621094, |
|
"loss": 2.6229, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.2400531768798828, |
|
"rewards/margins": 0.21957488358020782, |
|
"rewards/rejected": -1.4596279859542847, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.34332425068119893, |
|
"grad_norm": 1.5255874395370483, |
|
"learning_rate": 9.83709939706686e-06, |
|
"logits/chosen": 0.12232109159231186, |
|
"logits/rejected": 0.08726370334625244, |
|
"logps/chosen": -66.86070251464844, |
|
"logps/rejected": -70.5438232421875, |
|
"loss": 2.7017, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.433100700378418, |
|
"rewards/margins": 0.20580635964870453, |
|
"rewards/rejected": -1.6389069557189941, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.34514078110808355, |
|
"grad_norm": 1.4504551887512207, |
|
"learning_rate": 9.835381636176604e-06, |
|
"logits/chosen": 0.1462351530790329, |
|
"logits/rejected": 0.13504080474376678, |
|
"logps/chosen": -76.45362091064453, |
|
"logps/rejected": -78.51449584960938, |
|
"loss": 2.5936, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -1.356174349784851, |
|
"rewards/margins": 0.21241840720176697, |
|
"rewards/rejected": -1.5685927867889404, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.3469573115349682, |
|
"grad_norm": 1.4351036548614502, |
|
"learning_rate": 9.833655034688336e-06, |
|
"logits/chosen": 0.1534399539232254, |
|
"logits/rejected": 0.178826704621315, |
|
"logps/chosen": -66.26270294189453, |
|
"logps/rejected": -70.225830078125, |
|
"loss": 2.6442, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.3860033750534058, |
|
"rewards/margins": 0.16981087625026703, |
|
"rewards/rejected": -1.555814266204834, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.34877384196185285, |
|
"grad_norm": 1.295323133468628, |
|
"learning_rate": 9.831919596122888e-06, |
|
"logits/chosen": 0.11593925207853317, |
|
"logits/rejected": 0.19400468468666077, |
|
"logps/chosen": -66.74591064453125, |
|
"logps/rejected": -72.82708740234375, |
|
"loss": 2.3599, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.2114710807800293, |
|
"rewards/margins": 0.34326374530792236, |
|
"rewards/rejected": -1.5547348260879517, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.3505903723887375, |
|
"grad_norm": 1.452673077583313, |
|
"learning_rate": 9.830175324019125e-06, |
|
"logits/chosen": 0.13779595494270325, |
|
"logits/rejected": 0.15601256489753723, |
|
"logps/chosen": -73.34432220458984, |
|
"logps/rejected": -76.34349822998047, |
|
"loss": 2.4369, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.409874677658081, |
|
"rewards/margins": 0.3434732258319855, |
|
"rewards/rejected": -1.7533478736877441, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.35240690281562215, |
|
"grad_norm": 1.3285003900527954, |
|
"learning_rate": 9.828422221933924e-06, |
|
"logits/chosen": 0.020087052136659622, |
|
"logits/rejected": 0.07995946705341339, |
|
"logps/chosen": -72.9058837890625, |
|
"logps/rejected": -81.98945617675781, |
|
"loss": 2.3612, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.2450451850891113, |
|
"rewards/margins": 0.40142494440078735, |
|
"rewards/rejected": -1.646470069885254, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.3542234332425068, |
|
"grad_norm": 1.2274519205093384, |
|
"learning_rate": 9.826660293442158e-06, |
|
"logits/chosen": 0.0241906326264143, |
|
"logits/rejected": 0.09953958541154861, |
|
"logps/chosen": -66.5189437866211, |
|
"logps/rejected": -75.99092102050781, |
|
"loss": 2.2288, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.3973838090896606, |
|
"rewards/margins": 0.5354034304618835, |
|
"rewards/rejected": -1.9327871799468994, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.35603996366939145, |
|
"grad_norm": 1.3833938837051392, |
|
"learning_rate": 9.824889542136714e-06, |
|
"logits/chosen": 0.086525097489357, |
|
"logits/rejected": 0.08451628684997559, |
|
"logps/chosen": -74.15232849121094, |
|
"logps/rejected": -79.37950897216797, |
|
"loss": 2.4671, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.4214783906936646, |
|
"rewards/margins": 0.3752206265926361, |
|
"rewards/rejected": -1.7966989278793335, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.3578564940962761, |
|
"grad_norm": 1.6680957078933716, |
|
"learning_rate": 9.823109971628459e-06, |
|
"logits/chosen": 0.06370481848716736, |
|
"logits/rejected": 0.12739142775535583, |
|
"logps/chosen": -75.81684875488281, |
|
"logps/rejected": -76.65689086914062, |
|
"loss": 2.3705, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.3144075870513916, |
|
"rewards/margins": 0.36334556341171265, |
|
"rewards/rejected": -1.6777533292770386, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.35967302452316074, |
|
"grad_norm": 1.4894323348999023, |
|
"learning_rate": 9.821321585546244e-06, |
|
"logits/chosen": 0.10617184638977051, |
|
"logits/rejected": 0.11465627700090408, |
|
"logps/chosen": -73.0533218383789, |
|
"logps/rejected": -83.94466400146484, |
|
"loss": 2.2915, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.4909334182739258, |
|
"rewards/margins": 0.5887378454208374, |
|
"rewards/rejected": -2.0796711444854736, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.3614895549500454, |
|
"grad_norm": 1.422853946685791, |
|
"learning_rate": 9.819524387536905e-06, |
|
"logits/chosen": 0.069038525223732, |
|
"logits/rejected": 0.09677774459123611, |
|
"logps/chosen": -85.49876403808594, |
|
"logps/rejected": -89.743408203125, |
|
"loss": 2.3, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.3645334243774414, |
|
"rewards/margins": 0.5102630853652954, |
|
"rewards/rejected": -1.8747965097427368, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.36330608537693004, |
|
"grad_norm": 1.3350121974945068, |
|
"learning_rate": 9.81771838126524e-06, |
|
"logits/chosen": 0.0018447795882821083, |
|
"logits/rejected": 0.055721428245306015, |
|
"logps/chosen": -71.9021224975586, |
|
"logps/rejected": -81.6389389038086, |
|
"loss": 2.2872, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.3360776901245117, |
|
"rewards/margins": 0.46150827407836914, |
|
"rewards/rejected": -1.7975859642028809, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.3651226158038147, |
|
"grad_norm": 1.458803653717041, |
|
"learning_rate": 9.815903570414006e-06, |
|
"logits/chosen": 0.059184275567531586, |
|
"logits/rejected": 0.07923795282840729, |
|
"logps/chosen": -80.88011932373047, |
|
"logps/rejected": -86.74174499511719, |
|
"loss": 2.4279, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.6620349884033203, |
|
"rewards/margins": 0.38666611909866333, |
|
"rewards/rejected": -2.048701047897339, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.36693914623069934, |
|
"grad_norm": 1.3583872318267822, |
|
"learning_rate": 9.814079958683925e-06, |
|
"logits/chosen": 0.11471173167228699, |
|
"logits/rejected": 0.1523369550704956, |
|
"logps/chosen": -71.39076232910156, |
|
"logps/rejected": -78.50016021728516, |
|
"loss": 2.4603, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.4875473976135254, |
|
"rewards/margins": 0.3805108070373535, |
|
"rewards/rejected": -1.8680580854415894, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.368755676657584, |
|
"grad_norm": 1.6755985021591187, |
|
"learning_rate": 9.812247549793656e-06, |
|
"logits/chosen": 0.14959998428821564, |
|
"logits/rejected": 0.1801329255104065, |
|
"logps/chosen": -76.0824203491211, |
|
"logps/rejected": -85.81806945800781, |
|
"loss": 2.8134, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.7176380157470703, |
|
"rewards/margins": 0.28729167580604553, |
|
"rewards/rejected": -2.004929780960083, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.37057220708446864, |
|
"grad_norm": 1.4551233053207397, |
|
"learning_rate": 9.810406347479798e-06, |
|
"logits/chosen": 0.08063512295484543, |
|
"logits/rejected": 0.03579093888401985, |
|
"logps/chosen": -87.19414520263672, |
|
"logps/rejected": -89.355224609375, |
|
"loss": 2.4092, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.5363514423370361, |
|
"rewards/margins": 0.4680227041244507, |
|
"rewards/rejected": -2.0043740272521973, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.3723887375113533, |
|
"grad_norm": 1.4248968362808228, |
|
"learning_rate": 9.808556355496885e-06, |
|
"logits/chosen": 0.06655821204185486, |
|
"logits/rejected": 0.050458114594221115, |
|
"logps/chosen": -92.04095458984375, |
|
"logps/rejected": -95.39706420898438, |
|
"loss": 2.3507, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.5707283020019531, |
|
"rewards/margins": 0.388569712638855, |
|
"rewards/rejected": -1.959298014640808, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.37420526793823794, |
|
"grad_norm": 1.3779215812683105, |
|
"learning_rate": 9.806697577617371e-06, |
|
"logits/chosen": 0.09702017903327942, |
|
"logits/rejected": 0.13923662900924683, |
|
"logps/chosen": -78.27027893066406, |
|
"logps/rejected": -85.93509674072266, |
|
"loss": 2.1888, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.4591946601867676, |
|
"rewards/margins": 0.5624303817749023, |
|
"rewards/rejected": -2.02162504196167, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.3760217983651226, |
|
"grad_norm": 1.5306010246276855, |
|
"learning_rate": 9.804830017631631e-06, |
|
"logits/chosen": 0.038323137909173965, |
|
"logits/rejected": 0.04337020218372345, |
|
"logps/chosen": -75.31045532226562, |
|
"logps/rejected": -83.26338958740234, |
|
"loss": 2.7312, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.4613063335418701, |
|
"rewards/margins": 0.3355714678764343, |
|
"rewards/rejected": -1.7968778610229492, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.37783832879200724, |
|
"grad_norm": 1.629341959953308, |
|
"learning_rate": 9.802953679347943e-06, |
|
"logits/chosen": 0.1168309897184372, |
|
"logits/rejected": 0.21053184568881989, |
|
"logps/chosen": -70.42684936523438, |
|
"logps/rejected": -87.77594757080078, |
|
"loss": 2.7324, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.8266733884811401, |
|
"rewards/margins": 0.5579056739807129, |
|
"rewards/rejected": -2.3845791816711426, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.3796548592188919, |
|
"grad_norm": 1.5751092433929443, |
|
"learning_rate": 9.801068566592486e-06, |
|
"logits/chosen": 0.11355097591876984, |
|
"logits/rejected": 0.11962890625, |
|
"logps/chosen": -87.75645446777344, |
|
"logps/rejected": -96.15601348876953, |
|
"loss": 2.5237, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.4331413507461548, |
|
"rewards/margins": 0.38925477862358093, |
|
"rewards/rejected": -1.8223960399627686, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.3814713896457766, |
|
"grad_norm": 1.6638984680175781, |
|
"learning_rate": 9.799174683209336e-06, |
|
"logits/chosen": 0.0970507487654686, |
|
"logits/rejected": 0.07422082126140594, |
|
"logps/chosen": -85.44343566894531, |
|
"logps/rejected": -90.49578857421875, |
|
"loss": 2.5179, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.6639773845672607, |
|
"rewards/margins": 0.4513780176639557, |
|
"rewards/rejected": -2.1153552532196045, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.3832879200726612, |
|
"grad_norm": 1.389356017112732, |
|
"learning_rate": 9.79727203306045e-06, |
|
"logits/chosen": 0.027732742950320244, |
|
"logits/rejected": 0.0795917734503746, |
|
"logps/chosen": -85.95429229736328, |
|
"logps/rejected": -92.91676330566406, |
|
"loss": 2.3892, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.4551334381103516, |
|
"rewards/margins": 0.4000782370567322, |
|
"rewards/rejected": -1.8552117347717285, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.3851044504995459, |
|
"grad_norm": 1.438284993171692, |
|
"learning_rate": 9.79536062002566e-06, |
|
"logits/chosen": 0.04227686673402786, |
|
"logits/rejected": 0.09655077010393143, |
|
"logps/chosen": -77.67405700683594, |
|
"logps/rejected": -88.0655746459961, |
|
"loss": 2.3246, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.5603740215301514, |
|
"rewards/margins": 0.5005100965499878, |
|
"rewards/rejected": -2.060884475708008, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.3869209809264305, |
|
"grad_norm": 1.35500168800354, |
|
"learning_rate": 9.793440448002676e-06, |
|
"logits/chosen": 0.10956872254610062, |
|
"logits/rejected": 0.10562983900308609, |
|
"logps/chosen": -74.20452117919922, |
|
"logps/rejected": -73.23489379882812, |
|
"loss": 2.4585, |
|
"rewards/accuracies": 0.796875, |
|
"rewards/chosen": -1.399983286857605, |
|
"rewards/margins": 0.2682442367076874, |
|
"rewards/rejected": -1.6682274341583252, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.3887375113533152, |
|
"grad_norm": 1.3036773204803467, |
|
"learning_rate": 9.791511520907056e-06, |
|
"logits/chosen": 0.06179399788379669, |
|
"logits/rejected": 0.052528850734233856, |
|
"logps/chosen": -72.4264144897461, |
|
"logps/rejected": -72.77079010009766, |
|
"loss": 2.3441, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.391230821609497, |
|
"rewards/margins": 0.3969228267669678, |
|
"rewards/rejected": -1.7881536483764648, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.3905540417801998, |
|
"grad_norm": 1.4877872467041016, |
|
"learning_rate": 9.789573842672223e-06, |
|
"logits/chosen": 0.05735350027680397, |
|
"logits/rejected": 0.09979183971881866, |
|
"logps/chosen": -82.134033203125, |
|
"logps/rejected": -96.38580322265625, |
|
"loss": 2.2192, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7035160064697266, |
|
"rewards/margins": 0.6622112989425659, |
|
"rewards/rejected": -2.365727424621582, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.3923705722070845, |
|
"grad_norm": 1.247817039489746, |
|
"learning_rate": 9.787627417249441e-06, |
|
"logits/chosen": 0.0014538783580064774, |
|
"logits/rejected": 0.03942735865712166, |
|
"logps/chosen": -73.8500747680664, |
|
"logps/rejected": -95.03350067138672, |
|
"loss": 1.9498, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.4928035736083984, |
|
"rewards/margins": 0.9987993836402893, |
|
"rewards/rejected": -2.491603374481201, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.3941871026339691, |
|
"grad_norm": 1.3884788751602173, |
|
"learning_rate": 9.785672248607807e-06, |
|
"logits/chosen": 0.029594585299491882, |
|
"logits/rejected": 0.11805769056081772, |
|
"logps/chosen": -66.9087905883789, |
|
"logps/rejected": -80.82840728759766, |
|
"loss": 2.2455, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.6260508298873901, |
|
"rewards/margins": 0.5117952227592468, |
|
"rewards/rejected": -2.137845993041992, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.3960036330608538, |
|
"grad_norm": 2.9690237045288086, |
|
"learning_rate": 9.78370834073425e-06, |
|
"logits/chosen": 0.11608768254518509, |
|
"logits/rejected": 0.11099248379468918, |
|
"logps/chosen": -73.23339080810547, |
|
"logps/rejected": -73.60865783691406, |
|
"loss": 2.6526, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.5958513021469116, |
|
"rewards/margins": 0.20439797639846802, |
|
"rewards/rejected": -1.8002492189407349, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.3978201634877384, |
|
"grad_norm": 1.4823459386825562, |
|
"learning_rate": 9.781735697633526e-06, |
|
"logits/chosen": 0.07910319417715073, |
|
"logits/rejected": 0.1617601215839386, |
|
"logps/chosen": -72.02169799804688, |
|
"logps/rejected": -79.70098114013672, |
|
"loss": 2.3689, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.722070336341858, |
|
"rewards/margins": 0.4829583168029785, |
|
"rewards/rejected": -2.205028772354126, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.3996366939146231, |
|
"grad_norm": 2.249466896057129, |
|
"learning_rate": 9.779754323328192e-06, |
|
"logits/chosen": 0.17587795853614807, |
|
"logits/rejected": 0.13541430234909058, |
|
"logps/chosen": -75.69686889648438, |
|
"logps/rejected": -77.53881072998047, |
|
"loss": 3.2204, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -2.03781795501709, |
|
"rewards/margins": 0.1416795551776886, |
|
"rewards/rejected": -2.179497480392456, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4014532243415077, |
|
"grad_norm": 1.4283430576324463, |
|
"learning_rate": 9.777764221858616e-06, |
|
"logits/chosen": 0.1529032289981842, |
|
"logits/rejected": 0.1623322069644928, |
|
"logps/chosen": -69.98622131347656, |
|
"logps/rejected": -77.31781005859375, |
|
"loss": 2.3358, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.6996090412139893, |
|
"rewards/margins": 0.568400502204895, |
|
"rewards/rejected": -2.2680094242095947, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.4032697547683924, |
|
"grad_norm": 1.4971157312393188, |
|
"learning_rate": 9.775765397282963e-06, |
|
"logits/chosen": 0.13248610496520996, |
|
"logits/rejected": 0.13485054671764374, |
|
"logps/chosen": -70.01846313476562, |
|
"logps/rejected": -75.87899017333984, |
|
"loss": 2.4037, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.6077332496643066, |
|
"rewards/margins": 0.39039328694343567, |
|
"rewards/rejected": -1.99812650680542, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.405086285195277, |
|
"grad_norm": 1.590364694595337, |
|
"learning_rate": 9.773757853677182e-06, |
|
"logits/chosen": 0.08200166374444962, |
|
"logits/rejected": 0.06919535249471664, |
|
"logps/chosen": -85.51278686523438, |
|
"logps/rejected": -89.3447265625, |
|
"loss": 2.6149, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.5626994371414185, |
|
"rewards/margins": 0.30814388394355774, |
|
"rewards/rejected": -1.8708434104919434, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.4069028156221617, |
|
"grad_norm": 1.41611647605896, |
|
"learning_rate": 9.771741595135009e-06, |
|
"logits/chosen": 0.057254984974861145, |
|
"logits/rejected": 0.13574014604091644, |
|
"logps/chosen": -73.97720336914062, |
|
"logps/rejected": -84.72975158691406, |
|
"loss": 2.3215, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.6877132654190063, |
|
"rewards/margins": 0.4473133981227875, |
|
"rewards/rejected": -2.135026693344116, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.4087193460490463, |
|
"grad_norm": 1.7008063793182373, |
|
"learning_rate": 9.769716625767939e-06, |
|
"logits/chosen": 0.05822606012225151, |
|
"logits/rejected": 0.06510132551193237, |
|
"logps/chosen": -81.26387023925781, |
|
"logps/rejected": -82.76637268066406, |
|
"loss": 2.5366, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.739980697631836, |
|
"rewards/margins": 0.3888433277606964, |
|
"rewards/rejected": -2.12882399559021, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.410535876475931, |
|
"grad_norm": 1.3915003538131714, |
|
"learning_rate": 9.767682949705243e-06, |
|
"logits/chosen": 0.08782866597175598, |
|
"logits/rejected": 0.17832686007022858, |
|
"logps/chosen": -67.25590515136719, |
|
"logps/rejected": -78.19799041748047, |
|
"loss": 2.3992, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.6588337421417236, |
|
"rewards/margins": 0.4392687976360321, |
|
"rewards/rejected": -2.098102331161499, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.4123524069028156, |
|
"grad_norm": 1.4107364416122437, |
|
"learning_rate": 9.765640571093938e-06, |
|
"logits/chosen": 0.14615394175052643, |
|
"logits/rejected": 0.14398689568042755, |
|
"logps/chosen": -66.720703125, |
|
"logps/rejected": -72.72846221923828, |
|
"loss": 2.488, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.740675449371338, |
|
"rewards/margins": 0.34203463792800903, |
|
"rewards/rejected": -2.0827102661132812, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.4141689373297003, |
|
"grad_norm": 1.438272476196289, |
|
"learning_rate": 9.76358949409879e-06, |
|
"logits/chosen": 0.1331941783428192, |
|
"logits/rejected": 0.18831086158752441, |
|
"logps/chosen": -77.92586517333984, |
|
"logps/rejected": -81.89257049560547, |
|
"loss": 2.466, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.7596431970596313, |
|
"rewards/margins": 0.3160095512866974, |
|
"rewards/rejected": -2.075652837753296, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.4159854677565849, |
|
"grad_norm": 1.3581331968307495, |
|
"learning_rate": 9.7615297229023e-06, |
|
"logits/chosen": 0.13822412490844727, |
|
"logits/rejected": 0.14220967888832092, |
|
"logps/chosen": -64.66896057128906, |
|
"logps/rejected": -76.72779846191406, |
|
"loss": 2.282, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.5948115587234497, |
|
"rewards/margins": 0.5508276224136353, |
|
"rewards/rejected": -2.145639181137085, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.4178019981834696, |
|
"grad_norm": 1.4766261577606201, |
|
"learning_rate": 9.759461261704705e-06, |
|
"logits/chosen": 0.06772036850452423, |
|
"logits/rejected": 0.1212601587176323, |
|
"logps/chosen": -75.17322540283203, |
|
"logps/rejected": -86.30448913574219, |
|
"loss": 2.1345, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.6930228471755981, |
|
"rewards/margins": 0.607439398765564, |
|
"rewards/rejected": -2.300462245941162, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4196185286103542, |
|
"grad_norm": 1.4598506689071655, |
|
"learning_rate": 9.757384114723954e-06, |
|
"logits/chosen": 0.11245124042034149, |
|
"logits/rejected": 0.17101560533046722, |
|
"logps/chosen": -75.50772094726562, |
|
"logps/rejected": -83.95561218261719, |
|
"loss": 2.3606, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.7730398178100586, |
|
"rewards/margins": 0.4058808386325836, |
|
"rewards/rejected": -2.1789209842681885, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.4214350590372389, |
|
"grad_norm": 1.5139139890670776, |
|
"learning_rate": 9.755298286195712e-06, |
|
"logits/chosen": 0.05479501932859421, |
|
"logits/rejected": 0.12558911740779877, |
|
"logps/chosen": -81.28207397460938, |
|
"logps/rejected": -86.58873748779297, |
|
"loss": 2.4171, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.6220098733901978, |
|
"rewards/margins": 0.49352455139160156, |
|
"rewards/rejected": -2.1155343055725098, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.4232515894641235, |
|
"grad_norm": 1.6987231969833374, |
|
"learning_rate": 9.753203780373348e-06, |
|
"logits/chosen": 0.07917524874210358, |
|
"logits/rejected": 0.14443910121917725, |
|
"logps/chosen": -81.83525848388672, |
|
"logps/rejected": -77.00749206542969, |
|
"loss": 2.7548, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7713696956634521, |
|
"rewards/margins": 0.20764464139938354, |
|
"rewards/rejected": -1.9790143966674805, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.4250681198910082, |
|
"grad_norm": 1.6056586503982544, |
|
"learning_rate": 9.751100601527922e-06, |
|
"logits/chosen": 0.12120751291513443, |
|
"logits/rejected": 0.21893832087516785, |
|
"logps/chosen": -73.1818618774414, |
|
"logps/rejected": -84.82020568847656, |
|
"loss": 2.4369, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.8145970106124878, |
|
"rewards/margins": 0.5701674818992615, |
|
"rewards/rejected": -2.3847644329071045, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.4268846503178928, |
|
"grad_norm": 1.4865500926971436, |
|
"learning_rate": 9.748988753948183e-06, |
|
"logits/chosen": 0.0659053698182106, |
|
"logits/rejected": 0.09985598176717758, |
|
"logps/chosen": -83.79960632324219, |
|
"logps/rejected": -87.37390899658203, |
|
"loss": 2.3437, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8071691989898682, |
|
"rewards/margins": 0.4156024754047394, |
|
"rewards/rejected": -2.2227721214294434, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.4287011807447775, |
|
"grad_norm": 1.7585958242416382, |
|
"learning_rate": 9.746868241940554e-06, |
|
"logits/chosen": 0.13636741042137146, |
|
"logits/rejected": 0.16040681302547455, |
|
"logps/chosen": -70.46460723876953, |
|
"logps/rejected": -71.19532012939453, |
|
"loss": 2.6303, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.7820823192596436, |
|
"rewards/margins": 0.37900543212890625, |
|
"rewards/rejected": -2.16108775138855, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.4305177111716621, |
|
"grad_norm": 1.4949400424957275, |
|
"learning_rate": 9.744739069829132e-06, |
|
"logits/chosen": 0.16385243833065033, |
|
"logits/rejected": 0.13090217113494873, |
|
"logps/chosen": -74.92865753173828, |
|
"logps/rejected": -78.68392181396484, |
|
"loss": 2.3038, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.847611904144287, |
|
"rewards/margins": 0.561863899230957, |
|
"rewards/rejected": -2.409475564956665, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.4323342415985468, |
|
"grad_norm": 1.6149885654449463, |
|
"learning_rate": 9.742601241955666e-06, |
|
"logits/chosen": 0.10731178522109985, |
|
"logits/rejected": 0.11118797957897186, |
|
"logps/chosen": -87.08879089355469, |
|
"logps/rejected": -90.8418960571289, |
|
"loss": 2.2981, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.7146490812301636, |
|
"rewards/margins": 0.4193970859050751, |
|
"rewards/rejected": -2.1340463161468506, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.43415077202543145, |
|
"grad_norm": 1.4300076961517334, |
|
"learning_rate": 9.740454762679562e-06, |
|
"logits/chosen": 0.031154140830039978, |
|
"logits/rejected": 0.13953763246536255, |
|
"logps/chosen": -68.21051025390625, |
|
"logps/rejected": -87.66563415527344, |
|
"loss": 2.0885, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.6824418306350708, |
|
"rewards/margins": 0.6189282536506653, |
|
"rewards/rejected": -2.3013699054718018, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.4359673024523161, |
|
"grad_norm": 1.3545743227005005, |
|
"learning_rate": 9.738299636377863e-06, |
|
"logits/chosen": 0.1132105141878128, |
|
"logits/rejected": 0.1149899885058403, |
|
"logps/chosen": -75.41773223876953, |
|
"logps/rejected": -80.03020477294922, |
|
"loss": 2.3774, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.5836522579193115, |
|
"rewards/margins": 0.40351229906082153, |
|
"rewards/rejected": -1.9871646165847778, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.43778383287920075, |
|
"grad_norm": 1.2866510152816772, |
|
"learning_rate": 9.736135867445246e-06, |
|
"logits/chosen": 0.08788580447435379, |
|
"logits/rejected": 0.15736152231693268, |
|
"logps/chosen": -70.54780578613281, |
|
"logps/rejected": -86.70913696289062, |
|
"loss": 1.9408, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.7565429210662842, |
|
"rewards/margins": 0.7674389481544495, |
|
"rewards/rejected": -2.523982048034668, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.4396003633060854, |
|
"grad_norm": 1.5900717973709106, |
|
"learning_rate": 9.733963460294016e-06, |
|
"logits/chosen": 0.08901432901620865, |
|
"logits/rejected": 0.09095099568367004, |
|
"logps/chosen": -77.32083129882812, |
|
"logps/rejected": -83.0787124633789, |
|
"loss": 2.4779, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.9288290739059448, |
|
"rewards/margins": 0.3023950755596161, |
|
"rewards/rejected": -2.2312240600585938, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.44141689373297005, |
|
"grad_norm": 1.664976716041565, |
|
"learning_rate": 9.731782419354087e-06, |
|
"logits/chosen": 0.0038250258658081293, |
|
"logits/rejected": 0.05805446207523346, |
|
"logps/chosen": -75.7310562133789, |
|
"logps/rejected": -81.24979400634766, |
|
"loss": 2.4426, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.8681721687316895, |
|
"rewards/margins": 0.41048479080200195, |
|
"rewards/rejected": -2.2786567211151123, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.44323342415985467, |
|
"grad_norm": 1.688614845275879, |
|
"learning_rate": 9.729592749072981e-06, |
|
"logits/chosen": 0.10514964163303375, |
|
"logits/rejected": 0.08623237907886505, |
|
"logps/chosen": -78.1123046875, |
|
"logps/rejected": -85.77177429199219, |
|
"loss": 2.4137, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.709516167640686, |
|
"rewards/margins": 0.477593332529068, |
|
"rewards/rejected": -2.1871094703674316, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.44504995458673935, |
|
"grad_norm": 1.603507399559021, |
|
"learning_rate": 9.727394453915817e-06, |
|
"logits/chosen": 0.06938113272190094, |
|
"logits/rejected": 0.10225434601306915, |
|
"logps/chosen": -72.41216278076172, |
|
"logps/rejected": -82.71170043945312, |
|
"loss": 2.3143, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.8322726488113403, |
|
"rewards/margins": 0.3731868863105774, |
|
"rewards/rejected": -2.2054593563079834, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.44686648501362397, |
|
"grad_norm": 1.6047879457473755, |
|
"learning_rate": 9.725187538365304e-06, |
|
"logits/chosen": 0.11169447004795074, |
|
"logits/rejected": 0.14944276213645935, |
|
"logps/chosen": -75.95654296875, |
|
"logps/rejected": -87.93280029296875, |
|
"loss": 2.3278, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.9458414316177368, |
|
"rewards/margins": 0.5952720046043396, |
|
"rewards/rejected": -2.5411133766174316, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.44868301544050865, |
|
"grad_norm": 1.5822384357452393, |
|
"learning_rate": 9.722972006921725e-06, |
|
"logits/chosen": 0.07633841782808304, |
|
"logits/rejected": 0.13307756185531616, |
|
"logps/chosen": -82.36216735839844, |
|
"logps/rejected": -89.472900390625, |
|
"loss": 2.3939, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -2.026167154312134, |
|
"rewards/margins": 0.4467831552028656, |
|
"rewards/rejected": -2.4729504585266113, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.45049954586739327, |
|
"grad_norm": 1.636837363243103, |
|
"learning_rate": 9.720747864102935e-06, |
|
"logits/chosen": 0.013166696764528751, |
|
"logits/rejected": 0.10926786065101624, |
|
"logps/chosen": -81.95793151855469, |
|
"logps/rejected": -95.96204376220703, |
|
"loss": 2.4935, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -2.1589713096618652, |
|
"rewards/margins": 0.4789046049118042, |
|
"rewards/rejected": -2.637876033782959, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.45231607629427795, |
|
"grad_norm": 1.6743932962417603, |
|
"learning_rate": 9.718515114444347e-06, |
|
"logits/chosen": 0.09870442003011703, |
|
"logits/rejected": 0.14206278324127197, |
|
"logps/chosen": -81.16842651367188, |
|
"logps/rejected": -95.32533264160156, |
|
"loss": 2.2988, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.291147470474243, |
|
"rewards/margins": 0.7547516822814941, |
|
"rewards/rejected": -3.0458991527557373, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.45413260672116257, |
|
"grad_norm": 1.4136468172073364, |
|
"learning_rate": 9.716273762498929e-06, |
|
"logits/chosen": 0.11840160191059113, |
|
"logits/rejected": 0.14921030402183533, |
|
"logps/chosen": -70.83485412597656, |
|
"logps/rejected": -73.80975341796875, |
|
"loss": 2.4975, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.8664369583129883, |
|
"rewards/margins": 0.32539018988609314, |
|
"rewards/rejected": -2.1918272972106934, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.45594913714804725, |
|
"grad_norm": 1.771864652633667, |
|
"learning_rate": 9.714023812837185e-06, |
|
"logits/chosen": 0.04426509141921997, |
|
"logits/rejected": 0.08872814476490021, |
|
"logps/chosen": -82.2516098022461, |
|
"logps/rejected": -87.78474426269531, |
|
"loss": 2.627, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.2238872051239014, |
|
"rewards/margins": 0.48393067717552185, |
|
"rewards/rejected": -2.707818031311035, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.45776566757493187, |
|
"grad_norm": 1.4489670991897583, |
|
"learning_rate": 9.711765270047155e-06, |
|
"logits/chosen": 0.025218207389116287, |
|
"logits/rejected": 0.09768272191286087, |
|
"logps/chosen": -70.79837036132812, |
|
"logps/rejected": -86.96686553955078, |
|
"loss": 2.1641, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8761096000671387, |
|
"rewards/margins": 0.8277947306632996, |
|
"rewards/rejected": -2.703904151916504, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.45958219800181654, |
|
"grad_norm": 1.7034775018692017, |
|
"learning_rate": 9.709498138734405e-06, |
|
"logits/chosen": 0.04030866175889969, |
|
"logits/rejected": 0.08729197829961777, |
|
"logps/chosen": -81.70675659179688, |
|
"logps/rejected": -85.92189025878906, |
|
"loss": 2.58, |
|
"rewards/accuracies": 0.546875, |
|
"rewards/chosen": -2.0707144737243652, |
|
"rewards/margins": 0.3897002637386322, |
|
"rewards/rejected": -2.4604148864746094, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.46139872842870117, |
|
"grad_norm": 1.8713371753692627, |
|
"learning_rate": 9.707222423522004e-06, |
|
"logits/chosen": 0.047953542321920395, |
|
"logits/rejected": 0.02678016573190689, |
|
"logps/chosen": -84.65204620361328, |
|
"logps/rejected": -90.09396362304688, |
|
"loss": 2.7899, |
|
"rewards/accuracies": 0.515625, |
|
"rewards/chosen": -2.103985071182251, |
|
"rewards/margins": 0.22625797986984253, |
|
"rewards/rejected": -2.330242872238159, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.46321525885558584, |
|
"grad_norm": 1.4706422090530396, |
|
"learning_rate": 9.704938129050535e-06, |
|
"logits/chosen": 0.04734738916158676, |
|
"logits/rejected": 0.11658424139022827, |
|
"logps/chosen": -78.08146667480469, |
|
"logps/rejected": -96.80207061767578, |
|
"loss": 2.1796, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9245433807373047, |
|
"rewards/margins": 0.7155373096466064, |
|
"rewards/rejected": -2.640080451965332, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.46503178928247046, |
|
"grad_norm": 1.4902199506759644, |
|
"learning_rate": 9.702645259978072e-06, |
|
"logits/chosen": 0.09310627728700638, |
|
"logits/rejected": 0.1795577108860016, |
|
"logps/chosen": -78.96179962158203, |
|
"logps/rejected": -84.86495208740234, |
|
"loss": 2.2113, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -2.0413970947265625, |
|
"rewards/margins": 0.4318653345108032, |
|
"rewards/rejected": -2.4732625484466553, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.46684831970935514, |
|
"grad_norm": 1.3563650846481323, |
|
"learning_rate": 9.700343820980172e-06, |
|
"logits/chosen": 0.08617695420980453, |
|
"logits/rejected": 0.1092975065112114, |
|
"logps/chosen": -76.31070709228516, |
|
"logps/rejected": -82.52798461914062, |
|
"loss": 2.3793, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.7843788862228394, |
|
"rewards/margins": 0.5503235459327698, |
|
"rewards/rejected": -2.334702491760254, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.46866485013623976, |
|
"grad_norm": 1.353379249572754, |
|
"learning_rate": 9.698033816749874e-06, |
|
"logits/chosen": 0.07846446335315704, |
|
"logits/rejected": 0.15949462354183197, |
|
"logps/chosen": -70.9232177734375, |
|
"logps/rejected": -83.97347259521484, |
|
"loss": 2.0527, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": -1.7502186298370361, |
|
"rewards/margins": 0.6754422783851624, |
|
"rewards/rejected": -2.4256608486175537, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.47048138056312444, |
|
"grad_norm": 1.2879610061645508, |
|
"learning_rate": 9.695715251997676e-06, |
|
"logits/chosen": 0.041364431381225586, |
|
"logits/rejected": 0.2137874960899353, |
|
"logps/chosen": -74.40650177001953, |
|
"logps/rejected": -99.26273345947266, |
|
"loss": 1.8381, |
|
"rewards/accuracies": 0.859375, |
|
"rewards/chosen": -1.7944194078445435, |
|
"rewards/margins": 1.1633189916610718, |
|
"rewards/rejected": -2.9577386379241943, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.47229791099000906, |
|
"grad_norm": 1.5336003303527832, |
|
"learning_rate": 9.693388131451536e-06, |
|
"logits/chosen": 0.10320600867271423, |
|
"logits/rejected": 0.20043231546878815, |
|
"logps/chosen": -70.24287414550781, |
|
"logps/rejected": -87.45311737060547, |
|
"loss": 2.2716, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8973060846328735, |
|
"rewards/margins": 0.7614614963531494, |
|
"rewards/rejected": -2.6587674617767334, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.47411444141689374, |
|
"grad_norm": 1.50228750705719, |
|
"learning_rate": 9.691052459856858e-06, |
|
"logits/chosen": 0.10719013214111328, |
|
"logits/rejected": 0.11425416171550751, |
|
"logps/chosen": -77.69039154052734, |
|
"logps/rejected": -83.77989196777344, |
|
"loss": 2.3255, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7352136373519897, |
|
"rewards/margins": 0.588873565196991, |
|
"rewards/rejected": -2.324087142944336, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.47593097184377836, |
|
"grad_norm": 1.6516278982162476, |
|
"learning_rate": 9.688708241976484e-06, |
|
"logits/chosen": 0.0962933823466301, |
|
"logits/rejected": 0.09487868845462799, |
|
"logps/chosen": -83.95497131347656, |
|
"logps/rejected": -86.93241882324219, |
|
"loss": 2.5126, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.0139267444610596, |
|
"rewards/margins": 0.4219791889190674, |
|
"rewards/rejected": -2.435905694961548, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.47774750227066304, |
|
"grad_norm": 1.4588408470153809, |
|
"learning_rate": 9.686355482590679e-06, |
|
"logits/chosen": 0.04918690025806427, |
|
"logits/rejected": 0.1662694364786148, |
|
"logps/chosen": -76.76988983154297, |
|
"logps/rejected": -98.94943237304688, |
|
"loss": 2.0243, |
|
"rewards/accuracies": 0.84375, |
|
"rewards/chosen": -1.8614740371704102, |
|
"rewards/margins": 0.9306913614273071, |
|
"rewards/rejected": -2.7921652793884277, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.47956403269754766, |
|
"grad_norm": 1.8992609977722168, |
|
"learning_rate": 9.683994186497132e-06, |
|
"logits/chosen": 0.039469510316848755, |
|
"logits/rejected": 0.11771678924560547, |
|
"logps/chosen": -78.4907455444336, |
|
"logps/rejected": -97.9975814819336, |
|
"loss": 2.6106, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": -2.1690123081207275, |
|
"rewards/margins": 0.9005274772644043, |
|
"rewards/rejected": -3.069540023803711, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.48138056312443234, |
|
"grad_norm": 2.196244239807129, |
|
"learning_rate": 9.681624358510936e-06, |
|
"logits/chosen": 0.11924441158771515, |
|
"logits/rejected": 0.08125054091215134, |
|
"logps/chosen": -67.94137573242188, |
|
"logps/rejected": -70.6710433959961, |
|
"loss": 2.1509, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.787060022354126, |
|
"rewards/margins": 0.5431486368179321, |
|
"rewards/rejected": -2.3302085399627686, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.48319709355131696, |
|
"grad_norm": 2.009992837905884, |
|
"learning_rate": 9.679246003464585e-06, |
|
"logits/chosen": 0.032514430582523346, |
|
"logits/rejected": 0.04692292958498001, |
|
"logps/chosen": -85.35857391357422, |
|
"logps/rejected": -90.84698486328125, |
|
"loss": 2.8119, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.1302294731140137, |
|
"rewards/margins": 0.3329467177391052, |
|
"rewards/rejected": -2.4631760120391846, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.48501362397820164, |
|
"grad_norm": 1.535208821296692, |
|
"learning_rate": 9.676859126207957e-06, |
|
"logits/chosen": 0.03791799396276474, |
|
"logits/rejected": 0.0628521591424942, |
|
"logps/chosen": -73.64868927001953, |
|
"logps/rejected": -80.9854965209961, |
|
"loss": 2.2386, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.9628286361694336, |
|
"rewards/margins": 0.6735512018203735, |
|
"rewards/rejected": -2.6363797187805176, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.4868301544050863, |
|
"grad_norm": 1.9440377950668335, |
|
"learning_rate": 9.674463731608309e-06, |
|
"logits/chosen": 0.0922878235578537, |
|
"logits/rejected": 0.08414338529109955, |
|
"logps/chosen": -80.34213256835938, |
|
"logps/rejected": -87.45000457763672, |
|
"loss": 2.8564, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.194643020629883, |
|
"rewards/margins": 0.5086687207221985, |
|
"rewards/rejected": -2.7033114433288574, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.48864668483197093, |
|
"grad_norm": 1.57882559299469, |
|
"learning_rate": 9.672059824550268e-06, |
|
"logits/chosen": 0.08647017180919647, |
|
"logits/rejected": 0.07438144087791443, |
|
"logps/chosen": -89.79737091064453, |
|
"logps/rejected": -95.89274597167969, |
|
"loss": 2.1801, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9004266262054443, |
|
"rewards/margins": 0.5936228632926941, |
|
"rewards/rejected": -2.494049549102783, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.4904632152588556, |
|
"grad_norm": 1.47967529296875, |
|
"learning_rate": 9.669647409935822e-06, |
|
"logits/chosen": 0.0663951188325882, |
|
"logits/rejected": 0.12090058624744415, |
|
"logps/chosen": -86.87987518310547, |
|
"logps/rejected": -90.49571228027344, |
|
"loss": 2.342, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -2.043788194656372, |
|
"rewards/margins": 0.4767158329486847, |
|
"rewards/rejected": -2.5205039978027344, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.49227974568574023, |
|
"grad_norm": 1.5860496759414673, |
|
"learning_rate": 9.667226492684302e-06, |
|
"logits/chosen": 0.06286406517028809, |
|
"logits/rejected": 0.1226339116692543, |
|
"logps/chosen": -91.13972473144531, |
|
"logps/rejected": -105.41304016113281, |
|
"loss": 2.2264, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.8733948469161987, |
|
"rewards/margins": 0.6329715251922607, |
|
"rewards/rejected": -2.50636625289917, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.4940962761126249, |
|
"grad_norm": 1.6157667636871338, |
|
"learning_rate": 9.66479707773238e-06, |
|
"logits/chosen": 0.07893703132867813, |
|
"logits/rejected": 0.1179615929722786, |
|
"logps/chosen": -77.89990234375, |
|
"logps/rejected": -87.70220947265625, |
|
"loss": 2.4723, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -2.136741876602173, |
|
"rewards/margins": 0.4580468237400055, |
|
"rewards/rejected": -2.5947885513305664, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.49591280653950953, |
|
"grad_norm": 1.5408835411071777, |
|
"learning_rate": 9.662359170034058e-06, |
|
"logits/chosen": 0.0976145789027214, |
|
"logits/rejected": 0.12423861026763916, |
|
"logps/chosen": -76.74501037597656, |
|
"logps/rejected": -82.18941497802734, |
|
"loss": 2.5565, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.039276361465454, |
|
"rewards/margins": 0.3934023678302765, |
|
"rewards/rejected": -2.432678461074829, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.4977293369663942, |
|
"grad_norm": 1.44657564163208, |
|
"learning_rate": 9.659912774560654e-06, |
|
"logits/chosen": 0.04748811572790146, |
|
"logits/rejected": 0.12317924201488495, |
|
"logps/chosen": -75.47578430175781, |
|
"logps/rejected": -88.56098937988281, |
|
"loss": 2.1908, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -2.050949811935425, |
|
"rewards/margins": 0.564163863658905, |
|
"rewards/rejected": -2.6151139736175537, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.49954586739327883, |
|
"grad_norm": 2.1769776344299316, |
|
"learning_rate": 9.65745789630079e-06, |
|
"logits/chosen": 0.11112834513187408, |
|
"logits/rejected": 0.10816515237092972, |
|
"logps/chosen": -83.5405044555664, |
|
"logps/rejected": -83.06329345703125, |
|
"loss": 3.0939, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -2.142024040222168, |
|
"rewards/margins": 0.1820124089717865, |
|
"rewards/rejected": -2.3240363597869873, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5013623978201635, |
|
"grad_norm": 1.7329221963882446, |
|
"learning_rate": 9.654994540260396e-06, |
|
"logits/chosen": 0.0653618574142456, |
|
"logits/rejected": 0.08004368096590042, |
|
"logps/chosen": -80.83209228515625, |
|
"logps/rejected": -82.97142028808594, |
|
"loss": 2.7161, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -1.851841688156128, |
|
"rewards/margins": 0.2948168218135834, |
|
"rewards/rejected": -2.146658420562744, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.5031789282470481, |
|
"grad_norm": 1.592657208442688, |
|
"learning_rate": 9.65252271146268e-06, |
|
"logits/chosen": 0.09880789369344711, |
|
"logits/rejected": 0.14229761064052582, |
|
"logps/chosen": -67.1727294921875, |
|
"logps/rejected": -75.03417205810547, |
|
"loss": 2.5366, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.8714643716812134, |
|
"rewards/margins": 0.37937116622924805, |
|
"rewards/rejected": -2.250835657119751, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.5049954586739328, |
|
"grad_norm": 1.6487712860107422, |
|
"learning_rate": 9.650042414948133e-06, |
|
"logits/chosen": 0.13465353846549988, |
|
"logits/rejected": 0.12865689396858215, |
|
"logps/chosen": -76.4417724609375, |
|
"logps/rejected": -78.6947021484375, |
|
"loss": 2.6085, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.875291347503662, |
|
"rewards/margins": 0.35175251960754395, |
|
"rewards/rejected": -2.227043867111206, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.5068119891008175, |
|
"grad_norm": 1.6523009538650513, |
|
"learning_rate": 9.64755365577451e-06, |
|
"logits/chosen": 0.04238567873835564, |
|
"logits/rejected": 0.07994347810745239, |
|
"logps/chosen": -76.92097473144531, |
|
"logps/rejected": -83.20886993408203, |
|
"loss": 2.3156, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.9759610891342163, |
|
"rewards/margins": 0.5398927330970764, |
|
"rewards/rejected": -2.5158536434173584, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.508628519527702, |
|
"grad_norm": 1.7999261617660522, |
|
"learning_rate": 9.645056439016827e-06, |
|
"logits/chosen": 0.07349395751953125, |
|
"logits/rejected": 0.07518415153026581, |
|
"logps/chosen": -77.56079864501953, |
|
"logps/rejected": -84.97645568847656, |
|
"loss": 2.8085, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.9408115148544312, |
|
"rewards/margins": 0.3470652997493744, |
|
"rewards/rejected": -2.287877082824707, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.5104450499545867, |
|
"grad_norm": 1.5684200525283813, |
|
"learning_rate": 9.642550769767342e-06, |
|
"logits/chosen": 0.16188879311084747, |
|
"logits/rejected": 0.12772323191165924, |
|
"logps/chosen": -89.29315185546875, |
|
"logps/rejected": -94.35065460205078, |
|
"loss": 2.2314, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.8958841562271118, |
|
"rewards/margins": 0.5102108716964722, |
|
"rewards/rejected": -2.406095027923584, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.5122615803814714, |
|
"grad_norm": 1.7878178358078003, |
|
"learning_rate": 9.640036653135548e-06, |
|
"logits/chosen": 0.060573749244213104, |
|
"logits/rejected": 0.13457715511322021, |
|
"logps/chosen": -68.9404525756836, |
|
"logps/rejected": -74.77693176269531, |
|
"loss": 2.4359, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.889530897140503, |
|
"rewards/margins": 0.45911547541618347, |
|
"rewards/rejected": -2.348646402359009, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.5140781108083561, |
|
"grad_norm": 1.4741288423538208, |
|
"learning_rate": 9.637514094248172e-06, |
|
"logits/chosen": 0.10433132946491241, |
|
"logits/rejected": 0.1439315229654312, |
|
"logps/chosen": -74.89447021484375, |
|
"logps/rejected": -86.58551788330078, |
|
"loss": 2.2478, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.9090807437896729, |
|
"rewards/margins": 0.6733560562133789, |
|
"rewards/rejected": -2.5824368000030518, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.5158946412352406, |
|
"grad_norm": 1.2864018678665161, |
|
"learning_rate": 9.634983098249146e-06, |
|
"logits/chosen": 0.10891089588403702, |
|
"logits/rejected": 0.11755162477493286, |
|
"logps/chosen": -66.90985107421875, |
|
"logps/rejected": -71.8125228881836, |
|
"loss": 2.1538, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.7779250144958496, |
|
"rewards/margins": 0.5625147223472595, |
|
"rewards/rejected": -2.340439796447754, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.5177111716621253, |
|
"grad_norm": 1.465747594833374, |
|
"learning_rate": 9.632443670299616e-06, |
|
"logits/chosen": 0.08224496245384216, |
|
"logits/rejected": 0.12130744755268097, |
|
"logps/chosen": -75.4281997680664, |
|
"logps/rejected": -85.0781021118164, |
|
"loss": 2.2988, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.7217226028442383, |
|
"rewards/margins": 0.4750482439994812, |
|
"rewards/rejected": -2.196770668029785, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.51952770208901, |
|
"grad_norm": 1.492859959602356, |
|
"learning_rate": 9.629895815577915e-06, |
|
"logits/chosen": 0.06619664281606674, |
|
"logits/rejected": 0.13152630627155304, |
|
"logps/chosen": -96.65383911132812, |
|
"logps/rejected": -111.93521881103516, |
|
"loss": 2.2831, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.7611618041992188, |
|
"rewards/margins": 0.5886337757110596, |
|
"rewards/rejected": -2.349795341491699, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.5213442325158947, |
|
"grad_norm": 1.5534065961837769, |
|
"learning_rate": 9.627339539279564e-06, |
|
"logits/chosen": 0.06637927144765854, |
|
"logits/rejected": 0.09107412397861481, |
|
"logps/chosen": -71.92534637451172, |
|
"logps/rejected": -82.98391723632812, |
|
"loss": 2.5101, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.7867075204849243, |
|
"rewards/margins": 0.39493298530578613, |
|
"rewards/rejected": -2.181640625, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.5231607629427792, |
|
"grad_norm": 1.6778221130371094, |
|
"learning_rate": 9.624774846617254e-06, |
|
"logits/chosen": 0.14700329303741455, |
|
"logits/rejected": 0.12778782844543457, |
|
"logps/chosen": -65.2364730834961, |
|
"logps/rejected": -74.39017486572266, |
|
"loss": 2.5236, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.6942007541656494, |
|
"rewards/margins": 0.35334131121635437, |
|
"rewards/rejected": -2.047542095184326, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.5249772933696639, |
|
"grad_norm": 1.462215542793274, |
|
"learning_rate": 9.622201742820839e-06, |
|
"logits/chosen": 0.11467991769313812, |
|
"logits/rejected": 0.1028795838356018, |
|
"logps/chosen": -67.85939025878906, |
|
"logps/rejected": -74.3462142944336, |
|
"loss": 2.3891, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7194863557815552, |
|
"rewards/margins": 0.3846975564956665, |
|
"rewards/rejected": -2.1041836738586426, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.5267938237965486, |
|
"grad_norm": 1.7591735124588013, |
|
"learning_rate": 9.619620233137328e-06, |
|
"logits/chosen": 0.08407986164093018, |
|
"logits/rejected": 0.12201236933469772, |
|
"logps/chosen": -82.57340240478516, |
|
"logps/rejected": -86.94192504882812, |
|
"loss": 2.4569, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8536537885665894, |
|
"rewards/margins": 0.38806283473968506, |
|
"rewards/rejected": -2.2417166233062744, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.5286103542234333, |
|
"grad_norm": 1.5584774017333984, |
|
"learning_rate": 9.617030322830868e-06, |
|
"logits/chosen": 0.1126309484243393, |
|
"logits/rejected": 0.08907752484083176, |
|
"logps/chosen": -73.36901092529297, |
|
"logps/rejected": -76.65506744384766, |
|
"loss": 2.4021, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.8091729879379272, |
|
"rewards/margins": 0.3603411316871643, |
|
"rewards/rejected": -2.1695141792297363, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.5304268846503178, |
|
"grad_norm": 1.487938642501831, |
|
"learning_rate": 9.614432017182736e-06, |
|
"logits/chosen": 0.09620118141174316, |
|
"logits/rejected": 0.08514149487018585, |
|
"logps/chosen": -88.66265106201172, |
|
"logps/rejected": -92.26091766357422, |
|
"loss": 2.3636, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.5980674028396606, |
|
"rewards/margins": 0.4888036549091339, |
|
"rewards/rejected": -2.0868711471557617, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.5322434150772025, |
|
"grad_norm": 1.445408582687378, |
|
"learning_rate": 9.611825321491331e-06, |
|
"logits/chosen": 0.10014252364635468, |
|
"logits/rejected": 0.12857215106487274, |
|
"logps/chosen": -82.46649932861328, |
|
"logps/rejected": -92.84205627441406, |
|
"loss": 2.343, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.4620139598846436, |
|
"rewards/margins": 0.5616316199302673, |
|
"rewards/rejected": -2.0236456394195557, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.5340599455040872, |
|
"grad_norm": 1.262193202972412, |
|
"learning_rate": 9.609210241072158e-06, |
|
"logits/chosen": 0.09868282079696655, |
|
"logits/rejected": 0.16074486076831818, |
|
"logps/chosen": -67.0407485961914, |
|
"logps/rejected": -82.36201477050781, |
|
"loss": 2.1163, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.5362765789031982, |
|
"rewards/margins": 0.7091963887214661, |
|
"rewards/rejected": -2.2454731464385986, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.5358764759309719, |
|
"grad_norm": 1.601891279220581, |
|
"learning_rate": 9.606586781257822e-06, |
|
"logits/chosen": 0.1236480325460434, |
|
"logits/rejected": 0.12187935411930084, |
|
"logps/chosen": -78.2726058959961, |
|
"logps/rejected": -84.49217987060547, |
|
"loss": 2.5075, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.779215931892395, |
|
"rewards/margins": 0.38987410068511963, |
|
"rewards/rejected": -2.1690900325775146, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.5376930063578564, |
|
"grad_norm": 1.4211548566818237, |
|
"learning_rate": 9.603954947398016e-06, |
|
"logits/chosen": 0.18375667929649353, |
|
"logits/rejected": 0.21945703029632568, |
|
"logps/chosen": -70.29829406738281, |
|
"logps/rejected": -78.42284393310547, |
|
"loss": 2.1959, |
|
"rewards/accuracies": 0.828125, |
|
"rewards/chosen": -1.7448753118515015, |
|
"rewards/margins": 0.5106840133666992, |
|
"rewards/rejected": -2.2555594444274902, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.5395095367847411, |
|
"grad_norm": 1.458617925643921, |
|
"learning_rate": 9.601314744859504e-06, |
|
"logits/chosen": 0.07360847294330597, |
|
"logits/rejected": 0.1596718430519104, |
|
"logps/chosen": -72.28546142578125, |
|
"logps/rejected": -87.13935089111328, |
|
"loss": 2.3538, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.8007893562316895, |
|
"rewards/margins": 0.5433144569396973, |
|
"rewards/rejected": -2.3441038131713867, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.5413260672116258, |
|
"grad_norm": 1.6526339054107666, |
|
"learning_rate": 9.598666179026123e-06, |
|
"logits/chosen": 0.12013350427150726, |
|
"logits/rejected": 0.10237178206443787, |
|
"logps/chosen": -83.06564331054688, |
|
"logps/rejected": -85.60771942138672, |
|
"loss": 2.4531, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -1.911988615989685, |
|
"rewards/margins": 0.35885563492774963, |
|
"rewards/rejected": -2.2708444595336914, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.5431425976385105, |
|
"grad_norm": 1.9808340072631836, |
|
"learning_rate": 9.596009255298755e-06, |
|
"logits/chosen": 0.062342576682567596, |
|
"logits/rejected": 0.09053834527730942, |
|
"logps/chosen": -96.33489990234375, |
|
"logps/rejected": -93.43024444580078, |
|
"loss": 2.8549, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.845227599143982, |
|
"rewards/margins": 0.22291362285614014, |
|
"rewards/rejected": -2.068141222000122, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.5449591280653951, |
|
"grad_norm": 1.6276236772537231, |
|
"learning_rate": 9.593343979095334e-06, |
|
"logits/chosen": 0.2073422521352768, |
|
"logits/rejected": 0.13107003271579742, |
|
"logps/chosen": -72.45758056640625, |
|
"logps/rejected": -77.42770385742188, |
|
"loss": 2.4644, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.9407453536987305, |
|
"rewards/margins": 0.3134301006793976, |
|
"rewards/rejected": -2.2541751861572266, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.5467756584922797, |
|
"grad_norm": 1.5001753568649292, |
|
"learning_rate": 9.590670355850819e-06, |
|
"logits/chosen": 0.15097060799598694, |
|
"logits/rejected": 0.16254279017448425, |
|
"logps/chosen": -70.44524383544922, |
|
"logps/rejected": -70.64158630371094, |
|
"loss": 2.4159, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.6921097040176392, |
|
"rewards/margins": 0.3336741626262665, |
|
"rewards/rejected": -2.0257837772369385, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.5485921889191644, |
|
"grad_norm": 1.8360233306884766, |
|
"learning_rate": 9.587988391017198e-06, |
|
"logits/chosen": 0.14593760669231415, |
|
"logits/rejected": 0.12311654537916183, |
|
"logps/chosen": -78.32576751708984, |
|
"logps/rejected": -88.25840759277344, |
|
"loss": 2.5826, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.6591644287109375, |
|
"rewards/margins": 0.5071319341659546, |
|
"rewards/rejected": -2.1662964820861816, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.5504087193460491, |
|
"grad_norm": 1.4822838306427002, |
|
"learning_rate": 9.585298090063459e-06, |
|
"logits/chosen": 0.20818498730659485, |
|
"logits/rejected": 0.16436657309532166, |
|
"logps/chosen": -67.59429931640625, |
|
"logps/rejected": -77.14763641357422, |
|
"loss": 2.2064, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.6798239946365356, |
|
"rewards/margins": 0.5954271554946899, |
|
"rewards/rejected": -2.2752511501312256, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.5522252497729337, |
|
"grad_norm": 1.6118305921554565, |
|
"learning_rate": 9.582599458475598e-06, |
|
"logits/chosen": 0.0827561467885971, |
|
"logits/rejected": 0.09151773154735565, |
|
"logps/chosen": -73.28964233398438, |
|
"logps/rejected": -76.90730285644531, |
|
"loss": 2.3792, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.8056440353393555, |
|
"rewards/margins": 0.39551618695259094, |
|
"rewards/rejected": -2.201160192489624, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.5540417801998183, |
|
"grad_norm": 1.4609856605529785, |
|
"learning_rate": 9.579892501756593e-06, |
|
"logits/chosen": 0.05332394689321518, |
|
"logits/rejected": 0.1590057611465454, |
|
"logps/chosen": -77.51653289794922, |
|
"logps/rejected": -95.51261138916016, |
|
"loss": 2.2105, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.740310549736023, |
|
"rewards/margins": 0.6771562695503235, |
|
"rewards/rejected": -2.4174671173095703, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.555858310626703, |
|
"grad_norm": 1.7181960344314575, |
|
"learning_rate": 9.5771772254264e-06, |
|
"logits/chosen": 0.05865276977419853, |
|
"logits/rejected": 0.03500773757696152, |
|
"logps/chosen": -83.72650146484375, |
|
"logps/rejected": -93.69804382324219, |
|
"loss": 2.1687, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.6025928258895874, |
|
"rewards/margins": 0.7339029312133789, |
|
"rewards/rejected": -2.3364956378936768, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.5576748410535877, |
|
"grad_norm": 1.497698187828064, |
|
"learning_rate": 9.57445363502194e-06, |
|
"logits/chosen": 0.1225174218416214, |
|
"logits/rejected": 0.09580346196889877, |
|
"logps/chosen": -76.27323150634766, |
|
"logps/rejected": -76.90716552734375, |
|
"loss": 2.2116, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.715336799621582, |
|
"rewards/margins": 0.4648086428642273, |
|
"rewards/rejected": -2.180145263671875, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.5594913714804723, |
|
"grad_norm": 1.3957064151763916, |
|
"learning_rate": 9.571721736097089e-06, |
|
"logits/chosen": 0.07596514374017715, |
|
"logits/rejected": 0.14832191169261932, |
|
"logps/chosen": -84.65274810791016, |
|
"logps/rejected": -105.76422119140625, |
|
"loss": 1.9942, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.6643587350845337, |
|
"rewards/margins": 0.7884883284568787, |
|
"rewards/rejected": -2.4528470039367676, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.5613079019073569, |
|
"grad_norm": 1.6824349164962769, |
|
"learning_rate": 9.568981534222664e-06, |
|
"logits/chosen": 0.049865882843732834, |
|
"logits/rejected": 0.023348212242126465, |
|
"logps/chosen": -80.3697509765625, |
|
"logps/rejected": -86.07074737548828, |
|
"loss": 2.4018, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -2.009049654006958, |
|
"rewards/margins": 0.5470253825187683, |
|
"rewards/rejected": -2.556075096130371, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.5631244323342416, |
|
"grad_norm": 1.8752086162567139, |
|
"learning_rate": 9.566233034986413e-06, |
|
"logits/chosen": 0.07250591367483139, |
|
"logits/rejected": 0.12809628248214722, |
|
"logps/chosen": -74.84794616699219, |
|
"logps/rejected": -81.79740905761719, |
|
"loss": 2.6286, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.75887131690979, |
|
"rewards/margins": 0.33979111909866333, |
|
"rewards/rejected": -2.0986623764038086, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.5649409627611263, |
|
"grad_norm": 1.7387233972549438, |
|
"learning_rate": 9.563476243993008e-06, |
|
"logits/chosen": 0.130618155002594, |
|
"logits/rejected": 0.12952059507369995, |
|
"logps/chosen": -80.75495910644531, |
|
"logps/rejected": -90.4281234741211, |
|
"loss": 2.4426, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.836004614830017, |
|
"rewards/margins": 0.4418962597846985, |
|
"rewards/rejected": -2.2779006958007812, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.5667574931880109, |
|
"grad_norm": 1.5147444009780884, |
|
"learning_rate": 9.56071116686402e-06, |
|
"logits/chosen": 0.10193713754415512, |
|
"logits/rejected": 0.22481802105903625, |
|
"logps/chosen": -73.21926879882812, |
|
"logps/rejected": -81.27864837646484, |
|
"loss": 2.578, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.6358015537261963, |
|
"rewards/margins": 0.38279739022254944, |
|
"rewards/rejected": -2.018598794937134, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.5685740236148955, |
|
"grad_norm": 1.6253665685653687, |
|
"learning_rate": 9.557937809237927e-06, |
|
"logits/chosen": 0.09468917548656464, |
|
"logits/rejected": 0.09415112435817719, |
|
"logps/chosen": -80.62995147705078, |
|
"logps/rejected": -86.65946197509766, |
|
"loss": 2.3854, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.8635404109954834, |
|
"rewards/margins": 0.4438764154911041, |
|
"rewards/rejected": -2.3074169158935547, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.5703905540417802, |
|
"grad_norm": 1.7893344163894653, |
|
"learning_rate": 9.555156176770087e-06, |
|
"logits/chosen": 0.15863659977912903, |
|
"logits/rejected": 0.09485571831464767, |
|
"logps/chosen": -76.10442352294922, |
|
"logps/rejected": -72.73162078857422, |
|
"loss": 2.6795, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.935206413269043, |
|
"rewards/margins": 0.22100940346717834, |
|
"rewards/rejected": -2.1562156677246094, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.5722070844686649, |
|
"grad_norm": 1.696327805519104, |
|
"learning_rate": 9.552366275132733e-06, |
|
"logits/chosen": 0.07012113183736801, |
|
"logits/rejected": 0.13891686499118805, |
|
"logps/chosen": -78.08012390136719, |
|
"logps/rejected": -83.05044555664062, |
|
"loss": 2.6181, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.7869051694869995, |
|
"rewards/margins": 0.23424415290355682, |
|
"rewards/rejected": -2.0211493968963623, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.5740236148955495, |
|
"grad_norm": 1.6044729948043823, |
|
"learning_rate": 9.54956811001496e-06, |
|
"logits/chosen": 0.11298641562461853, |
|
"logits/rejected": 0.2595204710960388, |
|
"logps/chosen": -75.5998306274414, |
|
"logps/rejected": -85.31849670410156, |
|
"loss": 2.5849, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.5757699012756348, |
|
"rewards/margins": 0.35041162371635437, |
|
"rewards/rejected": -1.9261815547943115, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.5758401453224341, |
|
"grad_norm": 1.7399548292160034, |
|
"learning_rate": 9.546761687122715e-06, |
|
"logits/chosen": 0.14933931827545166, |
|
"logits/rejected": 0.1868455857038498, |
|
"logps/chosen": -68.05965423583984, |
|
"logps/rejected": -82.37442779541016, |
|
"loss": 2.3208, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.8474600315093994, |
|
"rewards/margins": 0.5766161680221558, |
|
"rewards/rejected": -2.4240763187408447, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.5776566757493188, |
|
"grad_norm": 1.597495675086975, |
|
"learning_rate": 9.54394701217878e-06, |
|
"logits/chosen": 0.04729313403367996, |
|
"logits/rejected": 0.05815067142248154, |
|
"logps/chosen": -97.36541748046875, |
|
"logps/rejected": -104.02650451660156, |
|
"loss": 2.2106, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8631618022918701, |
|
"rewards/margins": 0.6027428507804871, |
|
"rewards/rejected": -2.465904474258423, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.5794732061762035, |
|
"grad_norm": 1.8646149635314941, |
|
"learning_rate": 9.541124090922771e-06, |
|
"logits/chosen": 0.1399674415588379, |
|
"logits/rejected": 0.11107950657606125, |
|
"logps/chosen": -82.39740753173828, |
|
"logps/rejected": -76.80902099609375, |
|
"loss": 2.6435, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.7670300006866455, |
|
"rewards/margins": 0.15345275402069092, |
|
"rewards/rejected": -1.920482873916626, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.5812897366030881, |
|
"grad_norm": 1.502447247505188, |
|
"learning_rate": 9.538292929111114e-06, |
|
"logits/chosen": 0.10672347247600555, |
|
"logits/rejected": 0.08290571719408035, |
|
"logps/chosen": -74.72683715820312, |
|
"logps/rejected": -86.37158203125, |
|
"loss": 2.2137, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.8041073083877563, |
|
"rewards/margins": 0.596272885799408, |
|
"rewards/rejected": -2.4003803730010986, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.5831062670299727, |
|
"grad_norm": 1.7287745475769043, |
|
"learning_rate": 9.535453532517039e-06, |
|
"logits/chosen": 0.13742896914482117, |
|
"logits/rejected": 0.07607734203338623, |
|
"logps/chosen": -86.82475280761719, |
|
"logps/rejected": -77.8934555053711, |
|
"loss": 2.6036, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.7918696403503418, |
|
"rewards/margins": 0.2591831088066101, |
|
"rewards/rejected": -2.0510525703430176, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.5849227974568574, |
|
"grad_norm": 1.5914596319198608, |
|
"learning_rate": 9.532605906930575e-06, |
|
"logits/chosen": 0.11301672458648682, |
|
"logits/rejected": 0.1944851577281952, |
|
"logps/chosen": -68.9898452758789, |
|
"logps/rejected": -73.17451477050781, |
|
"loss": 2.4565, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.6999026536941528, |
|
"rewards/margins": 0.30462026596069336, |
|
"rewards/rejected": -2.0045228004455566, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.5867393278837421, |
|
"grad_norm": 1.6072031259536743, |
|
"learning_rate": 9.529750058158522e-06, |
|
"logits/chosen": 0.07092760503292084, |
|
"logits/rejected": 0.08035591244697571, |
|
"logps/chosen": -81.5555419921875, |
|
"logps/rejected": -81.93560028076172, |
|
"loss": 2.2729, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -1.6834638118743896, |
|
"rewards/margins": 0.48882579803466797, |
|
"rewards/rejected": -2.1722893714904785, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.5885558583106267, |
|
"grad_norm": 1.6433221101760864, |
|
"learning_rate": 9.526885992024453e-06, |
|
"logits/chosen": 0.13823899626731873, |
|
"logits/rejected": 0.13610433042049408, |
|
"logps/chosen": -78.9625244140625, |
|
"logps/rejected": -83.2666015625, |
|
"loss": 2.2075, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.6939194202423096, |
|
"rewards/margins": 0.5755473971366882, |
|
"rewards/rejected": -2.2694668769836426, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.5903723887375113, |
|
"grad_norm": 1.7674918174743652, |
|
"learning_rate": 9.524013714368702e-06, |
|
"logits/chosen": 0.1704932153224945, |
|
"logits/rejected": 0.1073535829782486, |
|
"logps/chosen": -72.47895050048828, |
|
"logps/rejected": -73.0809326171875, |
|
"loss": 2.5473, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.8390110731124878, |
|
"rewards/margins": 0.34467947483062744, |
|
"rewards/rejected": -2.1836905479431152, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.592188919164396, |
|
"grad_norm": 1.8219250440597534, |
|
"learning_rate": 9.521133231048338e-06, |
|
"logits/chosen": 0.07941028475761414, |
|
"logits/rejected": 0.13045310974121094, |
|
"logps/chosen": -80.73494720458984, |
|
"logps/rejected": -91.58990478515625, |
|
"loss": 2.4866, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.8594518899917603, |
|
"rewards/margins": 0.4554288685321808, |
|
"rewards/rejected": -2.314880847930908, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.5940054495912807, |
|
"grad_norm": 1.3753328323364258, |
|
"learning_rate": 9.51824454793717e-06, |
|
"logits/chosen": 0.08879546821117401, |
|
"logits/rejected": 0.04693777486681938, |
|
"logps/chosen": -82.9569091796875, |
|
"logps/rejected": -91.44571685791016, |
|
"loss": 2.2602, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.549119472503662, |
|
"rewards/margins": 0.5145683288574219, |
|
"rewards/rejected": -2.063688039779663, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.5958219800181653, |
|
"grad_norm": 1.4296562671661377, |
|
"learning_rate": 9.515347670925728e-06, |
|
"logits/chosen": 0.15614314377307892, |
|
"logits/rejected": 0.1598319411277771, |
|
"logps/chosen": -75.60345458984375, |
|
"logps/rejected": -80.81770324707031, |
|
"loss": 2.3946, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.556883454322815, |
|
"rewards/margins": 0.4525718092918396, |
|
"rewards/rejected": -2.0094552040100098, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.59763851044505, |
|
"grad_norm": 1.7932195663452148, |
|
"learning_rate": 9.512442605921245e-06, |
|
"logits/chosen": 0.05863601714372635, |
|
"logits/rejected": 0.0788806602358818, |
|
"logps/chosen": -75.95397186279297, |
|
"logps/rejected": -84.50177001953125, |
|
"loss": 2.5085, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.9159862995147705, |
|
"rewards/margins": 0.4600725769996643, |
|
"rewards/rejected": -2.376059055328369, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.5994550408719346, |
|
"grad_norm": 1.6558383703231812, |
|
"learning_rate": 9.509529358847655e-06, |
|
"logits/chosen": 0.08205496519804001, |
|
"logits/rejected": 0.13091425597667694, |
|
"logps/chosen": -82.73474884033203, |
|
"logps/rejected": -94.02273559570312, |
|
"loss": 2.4419, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.083285093307495, |
|
"rewards/margins": 0.5300815105438232, |
|
"rewards/rejected": -2.6133666038513184, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6012715712988193, |
|
"grad_norm": 1.7888520956039429, |
|
"learning_rate": 9.506607935645579e-06, |
|
"logits/chosen": 0.11793217062950134, |
|
"logits/rejected": 0.19568441808223724, |
|
"logps/chosen": -79.86200714111328, |
|
"logps/rejected": -89.25471496582031, |
|
"loss": 2.4273, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.9273384809494019, |
|
"rewards/margins": 0.5115458965301514, |
|
"rewards/rejected": -2.4388844966888428, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.6030881017257039, |
|
"grad_norm": 2.0436160564422607, |
|
"learning_rate": 9.503678342272306e-06, |
|
"logits/chosen": 0.043473344296216965, |
|
"logits/rejected": 0.15554016828536987, |
|
"logps/chosen": -75.2901611328125, |
|
"logps/rejected": -83.20555877685547, |
|
"loss": 2.6941, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.725769281387329, |
|
"rewards/margins": 0.2789250314235687, |
|
"rewards/rejected": -2.0046942234039307, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.6049046321525886, |
|
"grad_norm": 1.665578842163086, |
|
"learning_rate": 9.500740584701785e-06, |
|
"logits/chosen": 0.17119848728179932, |
|
"logits/rejected": 0.14128939807415009, |
|
"logps/chosen": -83.1322021484375, |
|
"logps/rejected": -93.38603210449219, |
|
"loss": 2.0931, |
|
"rewards/accuracies": 0.765625, |
|
"rewards/chosen": -1.8975954055786133, |
|
"rewards/margins": 0.7810046672821045, |
|
"rewards/rejected": -2.6786000728607178, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.6067211625794732, |
|
"grad_norm": 1.9258419275283813, |
|
"learning_rate": 9.497794668924617e-06, |
|
"logits/chosen": 0.028591612353920937, |
|
"logits/rejected": 0.08894480764865875, |
|
"logps/chosen": -78.12230682373047, |
|
"logps/rejected": -90.00184631347656, |
|
"loss": 2.4006, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.8113291263580322, |
|
"rewards/margins": 0.5277000665664673, |
|
"rewards/rejected": -2.339029550552368, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.6085376930063578, |
|
"grad_norm": 1.4878523349761963, |
|
"learning_rate": 9.494840600948038e-06, |
|
"logits/chosen": 0.03548199310898781, |
|
"logits/rejected": 0.09170147776603699, |
|
"logps/chosen": -73.15607452392578, |
|
"logps/rejected": -83.77317810058594, |
|
"loss": 2.3301, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.6781296730041504, |
|
"rewards/margins": 0.4735565185546875, |
|
"rewards/rejected": -2.151685953140259, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.6103542234332425, |
|
"grad_norm": 1.8042774200439453, |
|
"learning_rate": 9.491878386795906e-06, |
|
"logits/chosen": 0.046132348477840424, |
|
"logits/rejected": 0.0721711814403534, |
|
"logps/chosen": -81.01045227050781, |
|
"logps/rejected": -89.28679656982422, |
|
"loss": 2.5593, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.012852191925049, |
|
"rewards/margins": 0.44549623131752014, |
|
"rewards/rejected": -2.458348512649536, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.6121707538601272, |
|
"grad_norm": 2.048952102661133, |
|
"learning_rate": 9.488908032508691e-06, |
|
"logits/chosen": 0.10774732381105423, |
|
"logits/rejected": 0.11935572326183319, |
|
"logps/chosen": -91.25210571289062, |
|
"logps/rejected": -90.42224884033203, |
|
"loss": 2.8282, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -2.0127718448638916, |
|
"rewards/margins": 0.16889013350009918, |
|
"rewards/rejected": -2.18166184425354, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.6139872842870118, |
|
"grad_norm": 1.6447219848632812, |
|
"learning_rate": 9.485929544143462e-06, |
|
"logits/chosen": 0.12260966747999191, |
|
"logits/rejected": 0.1517799198627472, |
|
"logps/chosen": -72.49166870117188, |
|
"logps/rejected": -82.3083724975586, |
|
"loss": 2.4029, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.624878168106079, |
|
"rewards/margins": 0.5487539768218994, |
|
"rewards/rejected": -2.1736321449279785, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.6158038147138964, |
|
"grad_norm": 1.599530577659607, |
|
"learning_rate": 9.482942927773876e-06, |
|
"logits/chosen": 0.09051798284053802, |
|
"logits/rejected": 0.11202570050954819, |
|
"logps/chosen": -84.42019653320312, |
|
"logps/rejected": -92.61677551269531, |
|
"loss": 2.2269, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.6424546241760254, |
|
"rewards/margins": 0.5715836882591248, |
|
"rewards/rejected": -2.214038133621216, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.6176203451407811, |
|
"grad_norm": 1.5411081314086914, |
|
"learning_rate": 9.479948189490164e-06, |
|
"logits/chosen": 0.06563656777143478, |
|
"logits/rejected": 0.17349205911159515, |
|
"logps/chosen": -70.98219299316406, |
|
"logps/rejected": -78.88846588134766, |
|
"loss": 2.3328, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.824373722076416, |
|
"rewards/margins": 0.46550822257995605, |
|
"rewards/rejected": -2.289881706237793, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.6194368755676658, |
|
"grad_norm": 1.8855030536651611, |
|
"learning_rate": 9.476945335399122e-06, |
|
"logits/chosen": 0.11399642378091812, |
|
"logits/rejected": 0.12690946459770203, |
|
"logps/chosen": -85.86933898925781, |
|
"logps/rejected": -89.75601196289062, |
|
"loss": 2.4269, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.9560539722442627, |
|
"rewards/margins": 0.38427016139030457, |
|
"rewards/rejected": -2.3403239250183105, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.6212534059945504, |
|
"grad_norm": 1.9973480701446533, |
|
"learning_rate": 9.473934371624087e-06, |
|
"logits/chosen": 0.029105912894010544, |
|
"logits/rejected": 0.12763622403144836, |
|
"logps/chosen": -80.68119049072266, |
|
"logps/rejected": -93.90747833251953, |
|
"loss": 2.4709, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.906477928161621, |
|
"rewards/margins": 0.6422742009162903, |
|
"rewards/rejected": -2.5487518310546875, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.623069936421435, |
|
"grad_norm": 1.7752224206924438, |
|
"learning_rate": 9.47091530430494e-06, |
|
"logits/chosen": 0.14096824824810028, |
|
"logits/rejected": 0.1774113029241562, |
|
"logps/chosen": -71.80216217041016, |
|
"logps/rejected": -79.65301513671875, |
|
"loss": 2.2678, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.5735952854156494, |
|
"rewards/margins": 0.5823659300804138, |
|
"rewards/rejected": -2.155961036682129, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.6248864668483197, |
|
"grad_norm": 1.855377197265625, |
|
"learning_rate": 9.467888139598086e-06, |
|
"logits/chosen": 0.10637074708938599, |
|
"logits/rejected": 0.0798158049583435, |
|
"logps/chosen": -79.01347351074219, |
|
"logps/rejected": -80.77239990234375, |
|
"loss": 2.7867, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.9598251581192017, |
|
"rewards/margins": 0.3184160590171814, |
|
"rewards/rejected": -2.2782411575317383, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.6267029972752044, |
|
"grad_norm": 1.423040747642517, |
|
"learning_rate": 9.464852883676441e-06, |
|
"logits/chosen": 0.1304859220981598, |
|
"logits/rejected": 0.1759231686592102, |
|
"logps/chosen": -74.09733581542969, |
|
"logps/rejected": -89.27589416503906, |
|
"loss": 2.2777, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.7697277069091797, |
|
"rewards/margins": 0.635611355304718, |
|
"rewards/rejected": -2.405339002609253, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.628519527702089, |
|
"grad_norm": 1.6246038675308228, |
|
"learning_rate": 9.461809542729421e-06, |
|
"logits/chosen": 0.03750050812959671, |
|
"logits/rejected": 0.10337980091571808, |
|
"logps/chosen": -81.89110565185547, |
|
"logps/rejected": -95.36811828613281, |
|
"loss": 2.0438, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.7299752235412598, |
|
"rewards/margins": 0.7404756546020508, |
|
"rewards/rejected": -2.4704508781433105, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.6303360581289736, |
|
"grad_norm": 1.6618752479553223, |
|
"learning_rate": 9.458758122962926e-06, |
|
"logits/chosen": 0.05359608680009842, |
|
"logits/rejected": 0.10455545783042908, |
|
"logps/chosen": -85.34078216552734, |
|
"logps/rejected": -90.46200561523438, |
|
"loss": 2.3299, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.9331045150756836, |
|
"rewards/margins": 0.4755082130432129, |
|
"rewards/rejected": -2.4086129665374756, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.6321525885558583, |
|
"grad_norm": 1.5805696249008179, |
|
"learning_rate": 9.455698630599332e-06, |
|
"logits/chosen": 0.10048776119947433, |
|
"logits/rejected": 0.1140337884426117, |
|
"logps/chosen": -80.25875854492188, |
|
"logps/rejected": -94.44998168945312, |
|
"loss": 2.3439, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.8421945571899414, |
|
"rewards/margins": 0.5977468490600586, |
|
"rewards/rejected": -2.43994140625, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.633969118982743, |
|
"grad_norm": 1.37093985080719, |
|
"learning_rate": 9.452631071877478e-06, |
|
"logits/chosen": 0.11764326691627502, |
|
"logits/rejected": 0.10735289752483368, |
|
"logps/chosen": -72.02367401123047, |
|
"logps/rejected": -77.9522705078125, |
|
"loss": 2.1232, |
|
"rewards/accuracies": 0.796875, |
|
"rewards/chosen": -1.6527085304260254, |
|
"rewards/margins": 0.6065118908882141, |
|
"rewards/rejected": -2.259220600128174, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.6357856494096276, |
|
"grad_norm": 2.0214192867279053, |
|
"learning_rate": 9.449555453052652e-06, |
|
"logits/chosen": 0.13177426159381866, |
|
"logits/rejected": 0.12408209592103958, |
|
"logps/chosen": -76.62931823730469, |
|
"logps/rejected": -81.16517639160156, |
|
"loss": 2.9175, |
|
"rewards/accuracies": 0.53125, |
|
"rewards/chosen": -1.9575048685073853, |
|
"rewards/margins": 0.21441945433616638, |
|
"rewards/rejected": -2.171924352645874, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.6376021798365122, |
|
"grad_norm": 1.6041687726974487, |
|
"learning_rate": 9.446471780396573e-06, |
|
"logits/chosen": 0.18574532866477966, |
|
"logits/rejected": 0.22683225572109222, |
|
"logps/chosen": -73.95652770996094, |
|
"logps/rejected": -80.29524993896484, |
|
"loss": 2.3833, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.8684748411178589, |
|
"rewards/margins": 0.37268272042274475, |
|
"rewards/rejected": -2.2411575317382812, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.6394187102633969, |
|
"grad_norm": 1.7410838603973389, |
|
"learning_rate": 9.443380060197387e-06, |
|
"logits/chosen": 0.07876043766736984, |
|
"logits/rejected": 0.14103996753692627, |
|
"logps/chosen": -74.64432525634766, |
|
"logps/rejected": -83.6999740600586, |
|
"loss": 2.5558, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -2.029916763305664, |
|
"rewards/margins": 0.4088842272758484, |
|
"rewards/rejected": -2.438800811767578, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.6412352406902816, |
|
"grad_norm": 1.87971031665802, |
|
"learning_rate": 9.440280298759653e-06, |
|
"logits/chosen": 0.13997013866901398, |
|
"logits/rejected": 0.1322249174118042, |
|
"logps/chosen": -88.78776550292969, |
|
"logps/rejected": -88.22732543945312, |
|
"loss": 2.6547, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -1.9983258247375488, |
|
"rewards/margins": 0.2458382397890091, |
|
"rewards/rejected": -2.244164228439331, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.6430517711171662, |
|
"grad_norm": 1.7056363821029663, |
|
"learning_rate": 9.437172502404318e-06, |
|
"logits/chosen": 0.07248476892709732, |
|
"logits/rejected": 0.13019773364067078, |
|
"logps/chosen": -78.4591064453125, |
|
"logps/rejected": -80.49126434326172, |
|
"loss": 2.3809, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.9740041494369507, |
|
"rewards/margins": 0.4853326082229614, |
|
"rewards/rejected": -2.459336757659912, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.6448683015440508, |
|
"grad_norm": 1.435718059539795, |
|
"learning_rate": 9.434056677468726e-06, |
|
"logits/chosen": 0.09164869040250778, |
|
"logits/rejected": 0.14243285357952118, |
|
"logps/chosen": -77.83367156982422, |
|
"logps/rejected": -86.18836212158203, |
|
"loss": 2.006, |
|
"rewards/accuracies": 0.796875, |
|
"rewards/chosen": -1.7355122566223145, |
|
"rewards/margins": 0.6869419813156128, |
|
"rewards/rejected": -2.422454357147217, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.6466848319709355, |
|
"grad_norm": 1.6553188562393188, |
|
"learning_rate": 9.430932830306587e-06, |
|
"logits/chosen": 0.04967673122882843, |
|
"logits/rejected": 0.1456151008605957, |
|
"logps/chosen": -73.50204467773438, |
|
"logps/rejected": -85.31363677978516, |
|
"loss": 2.4388, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.978384256362915, |
|
"rewards/margins": 0.48918718099594116, |
|
"rewards/rejected": -2.467571496963501, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.6485013623978202, |
|
"grad_norm": 1.8443480730056763, |
|
"learning_rate": 9.427800967287963e-06, |
|
"logits/chosen": 0.06455090641975403, |
|
"logits/rejected": 0.12608012557029724, |
|
"logps/chosen": -75.49232482910156, |
|
"logps/rejected": -77.49136352539062, |
|
"loss": 2.706, |
|
"rewards/accuracies": 0.5625, |
|
"rewards/chosen": -1.8730812072753906, |
|
"rewards/margins": 0.2989741861820221, |
|
"rewards/rejected": -2.17205548286438, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.6503178928247049, |
|
"grad_norm": 1.6912306547164917, |
|
"learning_rate": 9.424661094799273e-06, |
|
"logits/chosen": 0.11878645420074463, |
|
"logits/rejected": 0.13628609478473663, |
|
"logps/chosen": -73.09883880615234, |
|
"logps/rejected": -84.19624328613281, |
|
"loss": 2.3605, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.928421139717102, |
|
"rewards/margins": 0.5017052292823792, |
|
"rewards/rejected": -2.430126190185547, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.6521344232515894, |
|
"grad_norm": 1.3643461465835571, |
|
"learning_rate": 9.421513219243262e-06, |
|
"logits/chosen": 0.07683826237916946, |
|
"logits/rejected": 0.15765298902988434, |
|
"logps/chosen": -76.0871353149414, |
|
"logps/rejected": -97.47781372070312, |
|
"loss": 1.7844, |
|
"rewards/accuracies": 0.828125, |
|
"rewards/chosen": -1.8300602436065674, |
|
"rewards/margins": 0.911116361618042, |
|
"rewards/rejected": -2.7411766052246094, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.6539509536784741, |
|
"grad_norm": 1.6327749490737915, |
|
"learning_rate": 9.418357347038999e-06, |
|
"logits/chosen": 0.1078951433300972, |
|
"logits/rejected": 0.12233921140432358, |
|
"logps/chosen": -75.89913177490234, |
|
"logps/rejected": -78.4587173461914, |
|
"loss": 2.4976, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.7992844581604004, |
|
"rewards/margins": 0.311392605304718, |
|
"rewards/rejected": -2.1106772422790527, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.6557674841053588, |
|
"grad_norm": 1.463025689125061, |
|
"learning_rate": 9.415193484621852e-06, |
|
"logits/chosen": 0.057331383228302, |
|
"logits/rejected": 0.1563551127910614, |
|
"logps/chosen": -80.7889633178711, |
|
"logps/rejected": -93.21602630615234, |
|
"loss": 2.1398, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.744189739227295, |
|
"rewards/margins": 0.576555073261261, |
|
"rewards/rejected": -2.320744752883911, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.6575840145322435, |
|
"grad_norm": 1.743695616722107, |
|
"learning_rate": 9.412021638443491e-06, |
|
"logits/chosen": 0.03781123086810112, |
|
"logits/rejected": 0.154897540807724, |
|
"logps/chosen": -77.72047424316406, |
|
"logps/rejected": -88.80599975585938, |
|
"loss": 2.2361, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9849369525909424, |
|
"rewards/margins": 0.6562294363975525, |
|
"rewards/rejected": -2.6411664485931396, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.659400544959128, |
|
"grad_norm": 1.5648279190063477, |
|
"learning_rate": 9.408841814971862e-06, |
|
"logits/chosen": 0.10401102900505066, |
|
"logits/rejected": 0.12699122726917267, |
|
"logps/chosen": -83.10671997070312, |
|
"logps/rejected": -86.8663101196289, |
|
"loss": 2.4009, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7048468589782715, |
|
"rewards/margins": 0.4090174734592438, |
|
"rewards/rejected": -2.1138644218444824, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.6612170753860127, |
|
"grad_norm": 1.3385239839553833, |
|
"learning_rate": 9.405654020691178e-06, |
|
"logits/chosen": 0.06100422143936157, |
|
"logits/rejected": 0.10872650146484375, |
|
"logps/chosen": -76.72837829589844, |
|
"logps/rejected": -87.63470458984375, |
|
"loss": 2.141, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.747226595878601, |
|
"rewards/margins": 0.615075945854187, |
|
"rewards/rejected": -2.362302780151367, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.6630336058128974, |
|
"grad_norm": 1.4060734510421753, |
|
"learning_rate": 9.402458262101906e-06, |
|
"logits/chosen": 0.10638861358165741, |
|
"logits/rejected": 0.16951681673526764, |
|
"logps/chosen": -77.35757446289062, |
|
"logps/rejected": -92.03012084960938, |
|
"loss": 2.0511, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": -1.8359463214874268, |
|
"rewards/margins": 0.7444422841072083, |
|
"rewards/rejected": -2.5803885459899902, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.6648501362397821, |
|
"grad_norm": 1.4371014833450317, |
|
"learning_rate": 9.399254545720757e-06, |
|
"logits/chosen": 0.0383220911026001, |
|
"logits/rejected": 0.05856206640601158, |
|
"logps/chosen": -82.40064239501953, |
|
"logps/rejected": -92.1117935180664, |
|
"loss": 2.1756, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.8580785989761353, |
|
"rewards/margins": 0.6475991010665894, |
|
"rewards/rejected": -2.5056777000427246, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.6666666666666666, |
|
"grad_norm": 1.6673945188522339, |
|
"learning_rate": 9.396042878080661e-06, |
|
"logits/chosen": 0.14657820761203766, |
|
"logits/rejected": 0.19021111726760864, |
|
"logps/chosen": -73.56724548339844, |
|
"logps/rejected": -78.85279846191406, |
|
"loss": 2.4018, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.9079951047897339, |
|
"rewards/margins": 0.3709837794303894, |
|
"rewards/rejected": -2.2789790630340576, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.6684831970935513, |
|
"grad_norm": 1.7402448654174805, |
|
"learning_rate": 9.392823265730775e-06, |
|
"logits/chosen": 0.16515754163265228, |
|
"logits/rejected": 0.12460774928331375, |
|
"logps/chosen": -69.89266204833984, |
|
"logps/rejected": -74.53837585449219, |
|
"loss": 2.4262, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -2.0627622604370117, |
|
"rewards/margins": 0.5317320823669434, |
|
"rewards/rejected": -2.594494342803955, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.670299727520436, |
|
"grad_norm": 1.5290364027023315, |
|
"learning_rate": 9.389595715236446e-06, |
|
"logits/chosen": 0.0954434722661972, |
|
"logits/rejected": 0.16226956248283386, |
|
"logps/chosen": -77.17019653320312, |
|
"logps/rejected": -91.07398986816406, |
|
"loss": 2.2511, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -2.241170644760132, |
|
"rewards/margins": 0.6319360136985779, |
|
"rewards/rejected": -2.8731067180633545, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.6721162579473207, |
|
"grad_norm": 1.4633573293685913, |
|
"learning_rate": 9.386360233179206e-06, |
|
"logits/chosen": 0.06517557799816132, |
|
"logits/rejected": 0.07250035554170609, |
|
"logps/chosen": -79.23770141601562, |
|
"logps/rejected": -88.50403594970703, |
|
"loss": 2.3858, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.751698613166809, |
|
"rewards/margins": 0.49529415369033813, |
|
"rewards/rejected": -2.246993064880371, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.6739327883742052, |
|
"grad_norm": 2.007215976715088, |
|
"learning_rate": 9.383116826156775e-06, |
|
"logits/chosen": 0.13584929704666138, |
|
"logits/rejected": 0.10950647294521332, |
|
"logps/chosen": -80.75408935546875, |
|
"logps/rejected": -79.3903579711914, |
|
"loss": 2.6967, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -2.2421724796295166, |
|
"rewards/margins": 0.34328368306159973, |
|
"rewards/rejected": -2.585456132888794, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.6757493188010899, |
|
"grad_norm": 1.7323477268218994, |
|
"learning_rate": 9.37986550078302e-06, |
|
"logits/chosen": 0.011626070365309715, |
|
"logits/rejected": 0.06207559257745743, |
|
"logps/chosen": -75.51494598388672, |
|
"logps/rejected": -85.83218383789062, |
|
"loss": 2.4608, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.9569413661956787, |
|
"rewards/margins": 0.5160585641860962, |
|
"rewards/rejected": -2.4729998111724854, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.6775658492279746, |
|
"grad_norm": 1.6365752220153809, |
|
"learning_rate": 9.376606263687959e-06, |
|
"logits/chosen": 0.10213632136583328, |
|
"logits/rejected": 0.08355780690908432, |
|
"logps/chosen": -79.01370239257812, |
|
"logps/rejected": -81.9417953491211, |
|
"loss": 2.4701, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.064467668533325, |
|
"rewards/margins": 0.4157037138938904, |
|
"rewards/rejected": -2.4801712036132812, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.6793823796548593, |
|
"grad_norm": 1.5918922424316406, |
|
"learning_rate": 9.373339121517748e-06, |
|
"logits/chosen": 0.09486684203147888, |
|
"logits/rejected": 0.07311725616455078, |
|
"logps/chosen": -82.15351867675781, |
|
"logps/rejected": -91.7690658569336, |
|
"loss": 2.1297, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -2.0153114795684814, |
|
"rewards/margins": 0.669613242149353, |
|
"rewards/rejected": -2.684924840927124, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.6811989100817438, |
|
"grad_norm": 1.6511566638946533, |
|
"learning_rate": 9.370064080934654e-06, |
|
"logits/chosen": 0.1406637728214264, |
|
"logits/rejected": 0.20172299444675446, |
|
"logps/chosen": -69.41023254394531, |
|
"logps/rejected": -74.65138244628906, |
|
"loss": 2.4704, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -2.061182737350464, |
|
"rewards/margins": 0.3090137243270874, |
|
"rewards/rejected": -2.3701963424682617, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.6830154405086285, |
|
"grad_norm": 1.770624041557312, |
|
"learning_rate": 9.366781148617056e-06, |
|
"logits/chosen": 0.10267248749732971, |
|
"logits/rejected": 0.07900385558605194, |
|
"logps/chosen": -77.27003479003906, |
|
"logps/rejected": -88.09040832519531, |
|
"loss": 2.2813, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -2.132927656173706, |
|
"rewards/margins": 0.6629016399383545, |
|
"rewards/rejected": -2.7958290576934814, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.6848319709355132, |
|
"grad_norm": 1.5172028541564941, |
|
"learning_rate": 9.363490331259426e-06, |
|
"logits/chosen": 0.05240853130817413, |
|
"logits/rejected": 0.10445387661457062, |
|
"logps/chosen": -75.5218276977539, |
|
"logps/rejected": -85.49366760253906, |
|
"loss": 2.1442, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9302880764007568, |
|
"rewards/margins": 0.5776917338371277, |
|
"rewards/rejected": -2.5079798698425293, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.6866485013623979, |
|
"grad_norm": 1.6540950536727905, |
|
"learning_rate": 9.360191635572313e-06, |
|
"logits/chosen": 0.14195458590984344, |
|
"logits/rejected": 0.0978107899427414, |
|
"logps/chosen": -85.00608825683594, |
|
"logps/rejected": -85.1456298828125, |
|
"loss": 2.423, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.8765841722488403, |
|
"rewards/margins": 0.49441370368003845, |
|
"rewards/rejected": -2.370997905731201, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.6884650317892824, |
|
"grad_norm": 1.6463801860809326, |
|
"learning_rate": 9.356885068282334e-06, |
|
"logits/chosen": 0.13197994232177734, |
|
"logits/rejected": 0.09924699366092682, |
|
"logps/chosen": -86.94219970703125, |
|
"logps/rejected": -86.85704803466797, |
|
"loss": 2.4801, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.9674016237258911, |
|
"rewards/margins": 0.4530283510684967, |
|
"rewards/rejected": -2.4204299449920654, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.6902815622161671, |
|
"grad_norm": 1.5428895950317383, |
|
"learning_rate": 9.353570636132151e-06, |
|
"logits/chosen": 0.10434838384389877, |
|
"logits/rejected": 0.12619757652282715, |
|
"logps/chosen": -78.56607818603516, |
|
"logps/rejected": -86.33320617675781, |
|
"loss": 2.2711, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.9667280912399292, |
|
"rewards/margins": 0.6350463628768921, |
|
"rewards/rejected": -2.601774215698242, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.6920980926430518, |
|
"grad_norm": 1.6211251020431519, |
|
"learning_rate": 9.350248345880471e-06, |
|
"logits/chosen": 0.14081251621246338, |
|
"logits/rejected": 0.1453506350517273, |
|
"logps/chosen": -74.38184356689453, |
|
"logps/rejected": -80.61524200439453, |
|
"loss": 2.3863, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -2.0189671516418457, |
|
"rewards/margins": 0.4528267979621887, |
|
"rewards/rejected": -2.4717938899993896, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.6939146230699365, |
|
"grad_norm": 2.6346240043640137, |
|
"learning_rate": 9.346918204302022e-06, |
|
"logits/chosen": 0.0519830696284771, |
|
"logits/rejected": 0.04284125566482544, |
|
"logps/chosen": -82.38153839111328, |
|
"logps/rejected": -86.35488891601562, |
|
"loss": 2.5434, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.9824391603469849, |
|
"rewards/margins": 0.4156208634376526, |
|
"rewards/rejected": -2.3980600833892822, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.695731153496821, |
|
"grad_norm": 1.4762026071548462, |
|
"learning_rate": 9.343580218187544e-06, |
|
"logits/chosen": 0.02196469157934189, |
|
"logits/rejected": 0.04618150740861893, |
|
"logps/chosen": -72.66182708740234, |
|
"logps/rejected": -88.68983459472656, |
|
"loss": 2.1699, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.9227774143218994, |
|
"rewards/margins": 0.9200201630592346, |
|
"rewards/rejected": -2.8427975177764893, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.6975476839237057, |
|
"grad_norm": 1.9675711393356323, |
|
"learning_rate": 9.340234394343768e-06, |
|
"logits/chosen": 0.12956203520298004, |
|
"logits/rejected": 0.16079677641391754, |
|
"logps/chosen": -87.38492584228516, |
|
"logps/rejected": -87.80662536621094, |
|
"loss": 2.3138, |
|
"rewards/accuracies": 0.734375, |
|
"rewards/chosen": -2.144991397857666, |
|
"rewards/margins": 0.4893474578857422, |
|
"rewards/rejected": -2.634338855743408, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.6993642143505904, |
|
"grad_norm": 1.9116860628128052, |
|
"learning_rate": 9.336880739593415e-06, |
|
"logits/chosen": 0.06013559550046921, |
|
"logits/rejected": 0.09889352321624756, |
|
"logps/chosen": -86.55718231201172, |
|
"logps/rejected": -91.01994323730469, |
|
"loss": 2.4864, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.9528815746307373, |
|
"rewards/margins": 0.44100096821784973, |
|
"rewards/rejected": -2.3938825130462646, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.701180744777475, |
|
"grad_norm": 2.0582289695739746, |
|
"learning_rate": 9.33351926077517e-06, |
|
"logits/chosen": 0.11538423597812653, |
|
"logits/rejected": 0.17568480968475342, |
|
"logps/chosen": -76.24241638183594, |
|
"logps/rejected": -87.1357650756836, |
|
"loss": 2.6878, |
|
"rewards/accuracies": 0.578125, |
|
"rewards/chosen": -2.102504253387451, |
|
"rewards/margins": 0.3671773076057434, |
|
"rewards/rejected": -2.469681739807129, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.7029972752043597, |
|
"grad_norm": 1.6962640285491943, |
|
"learning_rate": 9.330149964743674e-06, |
|
"logits/chosen": 0.12425235658884048, |
|
"logits/rejected": 0.1852879822254181, |
|
"logps/chosen": -82.67887115478516, |
|
"logps/rejected": -98.66133117675781, |
|
"loss": 2.343, |
|
"rewards/accuracies": 0.796875, |
|
"rewards/chosen": -2.182525634765625, |
|
"rewards/margins": 0.6602460145950317, |
|
"rewards/rejected": -2.842771530151367, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.7048138056312443, |
|
"grad_norm": 1.6582266092300415, |
|
"learning_rate": 9.326772858369506e-06, |
|
"logits/chosen": 0.14438432455062866, |
|
"logits/rejected": 0.0922938883304596, |
|
"logps/chosen": -78.87794494628906, |
|
"logps/rejected": -84.929443359375, |
|
"loss": 2.3036, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.988234281539917, |
|
"rewards/margins": 0.4737466275691986, |
|
"rewards/rejected": -2.4619810581207275, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.706630336058129, |
|
"grad_norm": 1.6689919233322144, |
|
"learning_rate": 9.323387948539176e-06, |
|
"logits/chosen": 0.1282673329114914, |
|
"logits/rejected": 0.13633723556995392, |
|
"logps/chosen": -70.22606658935547, |
|
"logps/rejected": -79.26126861572266, |
|
"loss": 2.241, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9162399768829346, |
|
"rewards/margins": 0.5345122814178467, |
|
"rewards/rejected": -2.4507524967193604, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.7084468664850136, |
|
"grad_norm": 1.5861207246780396, |
|
"learning_rate": 9.319995242155102e-06, |
|
"logits/chosen": 0.13024169206619263, |
|
"logits/rejected": 0.14390775561332703, |
|
"logps/chosen": -87.00923919677734, |
|
"logps/rejected": -91.24594116210938, |
|
"loss": 2.284, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.9127002954483032, |
|
"rewards/margins": 0.5265774130821228, |
|
"rewards/rejected": -2.4392776489257812, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.7102633969118983, |
|
"grad_norm": 2.0118792057037354, |
|
"learning_rate": 9.316594746135608e-06, |
|
"logits/chosen": 0.14648675918579102, |
|
"logits/rejected": 0.11786539107561111, |
|
"logps/chosen": -82.68962097167969, |
|
"logps/rejected": -88.26126861572266, |
|
"loss": 2.7281, |
|
"rewards/accuracies": 0.625, |
|
"rewards/chosen": -1.9463748931884766, |
|
"rewards/margins": 0.20282456278800964, |
|
"rewards/rejected": -2.1491994857788086, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.7120799273387829, |
|
"grad_norm": 1.8176332712173462, |
|
"learning_rate": 9.313186467414892e-06, |
|
"logits/chosen": 0.12084021419286728, |
|
"logits/rejected": 0.1433698982000351, |
|
"logps/chosen": -73.46707153320312, |
|
"logps/rejected": -82.46024322509766, |
|
"loss": 2.5247, |
|
"rewards/accuracies": 0.59375, |
|
"rewards/chosen": -2.0802624225616455, |
|
"rewards/margins": 0.41601985692977905, |
|
"rewards/rejected": -2.4962823390960693, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.7138964577656676, |
|
"grad_norm": 1.7207351922988892, |
|
"learning_rate": 9.30977041294303e-06, |
|
"logits/chosen": 0.11983273923397064, |
|
"logits/rejected": 0.16106371581554413, |
|
"logps/chosen": -76.09751892089844, |
|
"logps/rejected": -85.20323944091797, |
|
"loss": 2.5284, |
|
"rewards/accuracies": 0.609375, |
|
"rewards/chosen": -1.9018669128417969, |
|
"rewards/margins": 0.4894491136074066, |
|
"rewards/rejected": -2.3913159370422363, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.7157129881925522, |
|
"grad_norm": 1.6489367485046387, |
|
"learning_rate": 9.306346589685956e-06, |
|
"logits/chosen": 0.14845696091651917, |
|
"logits/rejected": 0.10356368124485016, |
|
"logps/chosen": -83.94526672363281, |
|
"logps/rejected": -86.53378295898438, |
|
"loss": 2.3394, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.9089899063110352, |
|
"rewards/margins": 0.4619474411010742, |
|
"rewards/rejected": -2.3709375858306885, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.7175295186194369, |
|
"grad_norm": 1.792077898979187, |
|
"learning_rate": 9.302915004625435e-06, |
|
"logits/chosen": 0.06936248391866684, |
|
"logits/rejected": 0.13349927961826324, |
|
"logps/chosen": -91.82075500488281, |
|
"logps/rejected": -98.7391357421875, |
|
"loss": 2.3875, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.9371806383132935, |
|
"rewards/margins": 0.6781354546546936, |
|
"rewards/rejected": -2.6153156757354736, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.7193460490463215, |
|
"grad_norm": 2.049694538116455, |
|
"learning_rate": 9.29947566475907e-06, |
|
"logits/chosen": 0.08474650233983994, |
|
"logits/rejected": 0.13380834460258484, |
|
"logps/chosen": -84.59528350830078, |
|
"logps/rejected": -96.24267578125, |
|
"loss": 2.7001, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.9375882148742676, |
|
"rewards/margins": 0.4607384204864502, |
|
"rewards/rejected": -2.3983266353607178, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.7211625794732062, |
|
"grad_norm": 1.9235875606536865, |
|
"learning_rate": 9.296028577100271e-06, |
|
"logits/chosen": 0.06255945563316345, |
|
"logits/rejected": 0.17074811458587646, |
|
"logps/chosen": -68.99498748779297, |
|
"logps/rejected": -84.45899963378906, |
|
"loss": 2.347, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.9344629049301147, |
|
"rewards/margins": 0.688963770866394, |
|
"rewards/rejected": -2.6234264373779297, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.7229791099000908, |
|
"grad_norm": 1.6617276668548584, |
|
"learning_rate": 9.292573748678254e-06, |
|
"logits/chosen": 0.11962493509054184, |
|
"logits/rejected": 0.11647717654705048, |
|
"logps/chosen": -85.3626708984375, |
|
"logps/rejected": -95.70562744140625, |
|
"loss": 2.2229, |
|
"rewards/accuracies": 0.78125, |
|
"rewards/chosen": -1.8968594074249268, |
|
"rewards/margins": 0.6918852925300598, |
|
"rewards/rejected": -2.588744640350342, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.7247956403269755, |
|
"grad_norm": 1.4650444984436035, |
|
"learning_rate": 9.289111186538013e-06, |
|
"logits/chosen": 0.06390775740146637, |
|
"logits/rejected": 0.11076060682535172, |
|
"logps/chosen": -69.9189453125, |
|
"logps/rejected": -81.5618667602539, |
|
"loss": 2.2271, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.857581377029419, |
|
"rewards/margins": 0.5540103316307068, |
|
"rewards/rejected": -2.4115917682647705, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.7266121707538601, |
|
"grad_norm": 1.5728726387023926, |
|
"learning_rate": 9.285640897740316e-06, |
|
"logits/chosen": 0.08816932141780853, |
|
"logits/rejected": 0.1330798715353012, |
|
"logps/chosen": -78.04735565185547, |
|
"logps/rejected": -87.89193725585938, |
|
"loss": 2.3595, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.823161244392395, |
|
"rewards/margins": 0.4568220376968384, |
|
"rewards/rejected": -2.2799830436706543, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.7284287011807448, |
|
"grad_norm": 1.6291779279708862, |
|
"learning_rate": 9.282162889361686e-06, |
|
"logits/chosen": 0.17718347907066345, |
|
"logits/rejected": 0.20603135228157043, |
|
"logps/chosen": -67.99884033203125, |
|
"logps/rejected": -75.93153381347656, |
|
"loss": 2.3002, |
|
"rewards/accuracies": 0.71875, |
|
"rewards/chosen": -1.671633243560791, |
|
"rewards/margins": 0.45543500781059265, |
|
"rewards/rejected": -2.127068281173706, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.7302452316076294, |
|
"grad_norm": 1.5066276788711548, |
|
"learning_rate": 9.278677168494388e-06, |
|
"logits/chosen": 0.16840124130249023, |
|
"logits/rejected": 0.17102479934692383, |
|
"logps/chosen": -78.1556625366211, |
|
"logps/rejected": -84.91495513916016, |
|
"loss": 2.1543, |
|
"rewards/accuracies": 0.75, |
|
"rewards/chosen": -1.9279460906982422, |
|
"rewards/margins": 0.6659258008003235, |
|
"rewards/rejected": -2.593871831893921, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.7320617620345141, |
|
"grad_norm": 1.71995210647583, |
|
"learning_rate": 9.275183742246412e-06, |
|
"logits/chosen": -0.0005522281862795353, |
|
"logits/rejected": 0.1412215530872345, |
|
"logps/chosen": -73.37832641601562, |
|
"logps/rejected": -92.71226501464844, |
|
"loss": 2.3315, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.6880759000778198, |
|
"rewards/margins": 0.5844776630401611, |
|
"rewards/rejected": -2.2725534439086914, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.7338782924613987, |
|
"grad_norm": 1.676321029663086, |
|
"learning_rate": 9.271682617741466e-06, |
|
"logits/chosen": 0.15319044888019562, |
|
"logits/rejected": 0.14503052830696106, |
|
"logps/chosen": -75.85136413574219, |
|
"logps/rejected": -82.61698150634766, |
|
"loss": 2.3528, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.819403052330017, |
|
"rewards/margins": 0.3919827342033386, |
|
"rewards/rejected": -2.211385726928711, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.7356948228882834, |
|
"grad_norm": 1.4811294078826904, |
|
"learning_rate": 9.268173802118949e-06, |
|
"logits/chosen": 0.07122528553009033, |
|
"logits/rejected": 0.12146291136741638, |
|
"logps/chosen": -76.78916931152344, |
|
"logps/rejected": -85.33805084228516, |
|
"loss": 2.1505, |
|
"rewards/accuracies": 0.65625, |
|
"rewards/chosen": -1.9886473417282104, |
|
"rewards/margins": 0.5911746025085449, |
|
"rewards/rejected": -2.579822301864624, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.737511353315168, |
|
"grad_norm": 1.629601001739502, |
|
"learning_rate": 9.264657302533947e-06, |
|
"logits/chosen": 0.11274963617324829, |
|
"logits/rejected": 0.0651293620467186, |
|
"logps/chosen": -82.86985778808594, |
|
"logps/rejected": -88.68803405761719, |
|
"loss": 2.1873, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.8538885116577148, |
|
"rewards/margins": 0.6726438403129578, |
|
"rewards/rejected": -2.5265324115753174, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.7393278837420527, |
|
"grad_norm": 1.5687483549118042, |
|
"learning_rate": 9.261133126157218e-06, |
|
"logits/chosen": 0.06054290384054184, |
|
"logits/rejected": 0.1543959081172943, |
|
"logps/chosen": -68.5821304321289, |
|
"logps/rejected": -80.544921875, |
|
"loss": 2.3587, |
|
"rewards/accuracies": 0.703125, |
|
"rewards/chosen": -1.7550634145736694, |
|
"rewards/margins": 0.46927812695503235, |
|
"rewards/rejected": -2.224341630935669, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.7411444141689373, |
|
"grad_norm": 1.849908471107483, |
|
"learning_rate": 9.257601280175167e-06, |
|
"logits/chosen": 0.15177126228809357, |
|
"logits/rejected": 0.1485670655965805, |
|
"logps/chosen": -72.81680297851562, |
|
"logps/rejected": -80.58684539794922, |
|
"loss": 2.444, |
|
"rewards/accuracies": 0.6875, |
|
"rewards/chosen": -1.9739688634872437, |
|
"rewards/margins": 0.6166737079620361, |
|
"rewards/rejected": -2.5906424522399902, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.742960944595822, |
|
"grad_norm": 1.779342770576477, |
|
"learning_rate": 9.254061771789847e-06, |
|
"logits/chosen": 0.05093669891357422, |
|
"logits/rejected": 0.1048060953617096, |
|
"logps/chosen": -84.48141479492188, |
|
"logps/rejected": -91.49285125732422, |
|
"loss": 2.5324, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.9952929019927979, |
|
"rewards/margins": 0.3900759220123291, |
|
"rewards/rejected": -2.385368585586548, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.7447774750227066, |
|
"grad_norm": 1.71133291721344, |
|
"learning_rate": 9.25051460821893e-06, |
|
"logits/chosen": 0.1308833658695221, |
|
"logits/rejected": 0.08994461596012115, |
|
"logps/chosen": -74.09046936035156, |
|
"logps/rejected": -80.80438995361328, |
|
"loss": 2.1895, |
|
"rewards/accuracies": 0.640625, |
|
"rewards/chosen": -1.8457674980163574, |
|
"rewards/margins": 0.5731581449508667, |
|
"rewards/rejected": -2.4189257621765137, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.7465940054495913, |
|
"grad_norm": 1.6980112791061401, |
|
"learning_rate": 9.2469597966957e-06, |
|
"logits/chosen": 0.05856658145785332, |
|
"logits/rejected": 0.08371179550886154, |
|
"logps/chosen": -78.58245849609375, |
|
"logps/rejected": -88.30054473876953, |
|
"loss": 2.46, |
|
"rewards/accuracies": 0.671875, |
|
"rewards/chosen": -1.9803493022918701, |
|
"rewards/margins": 0.6115507483482361, |
|
"rewards/rejected": -2.591899871826172, |
|
"step": 411 |
|
} |
|
], |
|
"logging_steps": 1, |
|
"max_steps": 2200, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 4, |
|
"save_steps": 137, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.9534665846725018e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|