File size: 24,383 Bytes
cd83cf6
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.6837606837606838,
  "eval_steps": 50,
  "global_step": 200,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.017094017094017096,
      "grad_norm": 35.00202204303521,
      "learning_rate": 5e-07,
      "logits/chosen": -2.7455849647521973,
      "logits/rejected": -2.7442612648010254,
      "logps/chosen": -164.2725830078125,
      "logps/rejected": -170.57113647460938,
      "loss": 0.6934,
      "rewards/accuracies": 0.23749999701976776,
      "rewards/chosen": 0.0026612328365445137,
      "rewards/margins": -0.001539617427624762,
      "rewards/rejected": 0.004200850613415241,
      "step": 5
    },
    {
      "epoch": 0.03418803418803419,
      "grad_norm": 36.29266486314593,
      "learning_rate": 1e-06,
      "logits/chosen": -2.709902763366699,
      "logits/rejected": -2.7155404090881348,
      "logps/chosen": -171.80032348632812,
      "logps/rejected": -165.20169067382812,
      "loss": 0.6879,
      "rewards/accuracies": 0.5062500238418579,
      "rewards/chosen": 0.012009668163955212,
      "rewards/margins": 0.0021203968208283186,
      "rewards/rejected": 0.009889272041618824,
      "step": 10
    },
    {
      "epoch": 0.05128205128205128,
      "grad_norm": 33.83921269470837,
      "learning_rate": 9.999177507263144e-07,
      "logits/chosen": -2.6502068042755127,
      "logits/rejected": -2.628007411956787,
      "logps/chosen": -174.082275390625,
      "logps/rejected": -174.13429260253906,
      "loss": 0.6698,
      "rewards/accuracies": 0.643750011920929,
      "rewards/chosen": 0.23495244979858398,
      "rewards/margins": 0.1125468835234642,
      "rewards/rejected": 0.12240554392337799,
      "step": 15
    },
    {
      "epoch": 0.06837606837606838,
      "grad_norm": 34.14427373918799,
      "learning_rate": 9.996710299650301e-07,
      "logits/chosen": -2.473665714263916,
      "logits/rejected": -2.4469008445739746,
      "logps/chosen": -158.2163848876953,
      "logps/rejected": -158.0710906982422,
      "loss": 0.661,
      "rewards/accuracies": 0.581250011920929,
      "rewards/chosen": 0.4233472943305969,
      "rewards/margins": 0.1434161365032196,
      "rewards/rejected": 0.2799311578273773,
      "step": 20
    },
    {
      "epoch": 0.08547008547008547,
      "grad_norm": 33.2696083475879,
      "learning_rate": 9.992599188865604e-07,
      "logits/chosen": -2.314507007598877,
      "logits/rejected": -2.3168132305145264,
      "logps/chosen": -150.67019653320312,
      "logps/rejected": -156.8417510986328,
      "loss": 0.6501,
      "rewards/accuracies": 0.543749988079071,
      "rewards/chosen": 0.4975205063819885,
      "rewards/margins": 0.15743504464626312,
      "rewards/rejected": 0.3400854766368866,
      "step": 25
    },
    {
      "epoch": 0.10256410256410256,
      "grad_norm": 34.42253361988952,
      "learning_rate": 9.98684552745256e-07,
      "logits/chosen": -2.243194103240967,
      "logits/rejected": -2.251340866088867,
      "logps/chosen": -161.2266845703125,
      "logps/rejected": -161.32298278808594,
      "loss": 0.6289,
      "rewards/accuracies": 0.6187499761581421,
      "rewards/chosen": 0.4243805408477783,
      "rewards/margins": 0.2635195851325989,
      "rewards/rejected": 0.16086098551750183,
      "step": 30
    },
    {
      "epoch": 0.11965811965811966,
      "grad_norm": 31.414296706456245,
      "learning_rate": 9.979451208349055e-07,
      "logits/chosen": -2.30315899848938,
      "logits/rejected": -2.289762496948242,
      "logps/chosen": -171.71713256835938,
      "logps/rejected": -174.50900268554688,
      "loss": 0.6296,
      "rewards/accuracies": 0.65625,
      "rewards/chosen": -0.019384615123271942,
      "rewards/margins": 0.318477988243103,
      "rewards/rejected": -0.3378625512123108,
      "step": 35
    },
    {
      "epoch": 0.13675213675213677,
      "grad_norm": 32.071830655862556,
      "learning_rate": 9.970418664264595e-07,
      "logits/chosen": -2.3935599327087402,
      "logits/rejected": -2.3812546730041504,
      "logps/chosen": -171.0698699951172,
      "logps/rejected": -176.58578491210938,
      "loss": 0.5991,
      "rewards/accuracies": 0.675000011920929,
      "rewards/chosen": -0.26089853048324585,
      "rewards/margins": 0.5235068202018738,
      "rewards/rejected": -0.7844053506851196,
      "step": 40
    },
    {
      "epoch": 0.15384615384615385,
      "grad_norm": 36.19466541168301,
      "learning_rate": 9.95975086687994e-07,
      "logits/chosen": -2.4914021492004395,
      "logits/rejected": -2.4973323345184326,
      "logps/chosen": -163.68099975585938,
      "logps/rejected": -167.174072265625,
      "loss": 0.6141,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -0.2962096929550171,
      "rewards/margins": 0.4588828682899475,
      "rewards/rejected": -0.7550925016403198,
      "step": 45
    },
    {
      "epoch": 0.17094017094017094,
      "grad_norm": 31.16276115760231,
      "learning_rate": 9.947451325869439e-07,
      "logits/chosen": -2.5575203895568848,
      "logits/rejected": -2.557717800140381,
      "logps/chosen": -172.04318237304688,
      "logps/rejected": -177.67672729492188,
      "loss": 0.5777,
      "rewards/accuracies": 0.731249988079071,
      "rewards/chosen": -0.2069791853427887,
      "rewards/margins": 0.6018465757369995,
      "rewards/rejected": -0.808825671672821,
      "step": 50
    },
    {
      "epoch": 0.17094017094017094,
      "eval_logits/chosen": -2.5221025943756104,
      "eval_logits/rejected": -2.5152711868286133,
      "eval_logps/chosen": -163.01820373535156,
      "eval_logps/rejected": -169.54832458496094,
      "eval_loss": 0.5812540650367737,
      "eval_rewards/accuracies": 0.6682692170143127,
      "eval_rewards/chosen": -0.45408713817596436,
      "eval_rewards/margins": 0.6127156615257263,
      "eval_rewards/rejected": -1.0668028593063354,
      "eval_runtime": 510.3361,
      "eval_samples_per_second": 16.291,
      "eval_steps_per_second": 0.255,
      "step": 50
    },
    {
      "epoch": 0.18803418803418803,
      "grad_norm": 31.575578721339145,
      "learning_rate": 9.933524087746347e-07,
      "logits/chosen": -2.490377426147461,
      "logits/rejected": -2.4825081825256348,
      "logps/chosen": -168.06161499023438,
      "logps/rejected": -175.0494384765625,
      "loss": 0.5706,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -0.5060762763023376,
      "rewards/margins": 0.7589826583862305,
      "rewards/rejected": -1.2650587558746338,
      "step": 55
    },
    {
      "epoch": 0.20512820512820512,
      "grad_norm": 30.171745273288415,
      "learning_rate": 9.917973734531549e-07,
      "logits/chosen": -2.48228120803833,
      "logits/rejected": -2.4833157062530518,
      "logps/chosen": -159.47142028808594,
      "logps/rejected": -170.63671875,
      "loss": 0.5753,
      "rewards/accuracies": 0.675000011920929,
      "rewards/chosen": -0.35752761363983154,
      "rewards/margins": 0.5991309881210327,
      "rewards/rejected": -0.9566585421562195,
      "step": 60
    },
    {
      "epoch": 0.2222222222222222,
      "grad_norm": 32.13878319029882,
      "learning_rate": 9.90080538224607e-07,
      "logits/chosen": -2.585407018661499,
      "logits/rejected": -2.5767769813537598,
      "logps/chosen": -157.43936157226562,
      "logps/rejected": -166.13589477539062,
      "loss": 0.566,
      "rewards/accuracies": 0.6937500238418579,
      "rewards/chosen": -0.02057185396552086,
      "rewards/margins": 0.47568243741989136,
      "rewards/rejected": -0.4962543547153473,
      "step": 65
    },
    {
      "epoch": 0.23931623931623933,
      "grad_norm": 29.494674721856043,
      "learning_rate": 9.882024679227938e-07,
      "logits/chosen": -2.6504979133605957,
      "logits/rejected": -2.6398470401763916,
      "logps/chosen": -178.0801239013672,
      "logps/rejected": -179.46328735351562,
      "loss": 0.5444,
      "rewards/accuracies": 0.706250011920929,
      "rewards/chosen": -0.43436694145202637,
      "rewards/margins": 0.8427752256393433,
      "rewards/rejected": -1.27714204788208,
      "step": 70
    },
    {
      "epoch": 0.2564102564102564,
      "grad_norm": 28.856733948308104,
      "learning_rate": 9.861637804273881e-07,
      "logits/chosen": -2.660489082336426,
      "logits/rejected": -2.655539035797119,
      "logps/chosen": -162.1233673095703,
      "logps/rejected": -170.16131591796875,
      "loss": 0.5568,
      "rewards/accuracies": 0.699999988079071,
      "rewards/chosen": -0.4032784402370453,
      "rewards/margins": 0.6959114074707031,
      "rewards/rejected": -1.0991899967193604,
      "step": 75
    },
    {
      "epoch": 0.27350427350427353,
      "grad_norm": 26.646061534818323,
      "learning_rate": 9.83965146460653e-07,
      "logits/chosen": -2.6391615867614746,
      "logits/rejected": -2.628577709197998,
      "logps/chosen": -168.58099365234375,
      "logps/rejected": -179.22805786132812,
      "loss": 0.5448,
      "rewards/accuracies": 0.6812499761581421,
      "rewards/chosen": -0.6665827035903931,
      "rewards/margins": 0.8240470886230469,
      "rewards/rejected": -1.4906299114227295,
      "step": 80
    },
    {
      "epoch": 0.2905982905982906,
      "grad_norm": 36.04159750418885,
      "learning_rate": 9.816072893667758e-07,
      "logits/chosen": -2.6322970390319824,
      "logits/rejected": -2.6053385734558105,
      "logps/chosen": -174.82640075683594,
      "logps/rejected": -186.0735626220703,
      "loss": 0.5579,
      "rewards/accuracies": 0.7562500238418579,
      "rewards/chosen": -1.0639268159866333,
      "rewards/margins": 1.0258175134658813,
      "rewards/rejected": -2.0897443294525146,
      "step": 85
    },
    {
      "epoch": 0.3076923076923077,
      "grad_norm": 26.922939193632168,
      "learning_rate": 9.790909848738904e-07,
      "logits/chosen": -2.60801362991333,
      "logits/rejected": -2.6101624965667725,
      "logps/chosen": -176.20538330078125,
      "logps/rejected": -184.7812957763672,
      "loss": 0.5215,
      "rewards/accuracies": 0.6312500238418579,
      "rewards/chosen": -0.9929834604263306,
      "rewards/margins": 0.8646324276924133,
      "rewards/rejected": -1.8576160669326782,
      "step": 90
    },
    {
      "epoch": 0.3247863247863248,
      "grad_norm": 30.3564450245371,
      "learning_rate": 9.764170608388647e-07,
      "logits/chosen": -2.6054036617279053,
      "logits/rejected": -2.5733799934387207,
      "logps/chosen": -168.037109375,
      "logps/rejected": -174.51144409179688,
      "loss": 0.5197,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -0.6652337312698364,
      "rewards/margins": 1.060430884361267,
      "rewards/rejected": -1.725664734840393,
      "step": 95
    },
    {
      "epoch": 0.3418803418803419,
      "grad_norm": 28.936164680674203,
      "learning_rate": 9.735863969749371e-07,
      "logits/chosen": -2.5255179405212402,
      "logits/rejected": -2.4874520301818848,
      "logps/chosen": -177.73861694335938,
      "logps/rejected": -189.82369995117188,
      "loss": 0.4982,
      "rewards/accuracies": 0.706250011920929,
      "rewards/chosen": -0.8511013984680176,
      "rewards/margins": 1.1354777812957764,
      "rewards/rejected": -1.986579179763794,
      "step": 100
    },
    {
      "epoch": 0.3418803418803419,
      "eval_logits/chosen": -2.484687328338623,
      "eval_logits/rejected": -2.460559368133545,
      "eval_logps/chosen": -168.28323364257812,
      "eval_logps/rejected": -180.8539276123047,
      "eval_loss": 0.5161151885986328,
      "eval_rewards/accuracies": 0.7211538553237915,
      "eval_rewards/chosen": -0.9805887937545776,
      "eval_rewards/margins": 1.2167747020721436,
      "eval_rewards/rejected": -2.1973636150360107,
      "eval_runtime": 510.3447,
      "eval_samples_per_second": 16.291,
      "eval_steps_per_second": 0.255,
      "step": 100
    },
    {
      "epoch": 0.358974358974359,
      "grad_norm": 31.089971589067016,
      "learning_rate": 9.705999245622956e-07,
      "logits/chosen": -2.4702706336975098,
      "logits/rejected": -2.4523651599884033,
      "logps/chosen": -170.59246826171875,
      "logps/rejected": -182.99813842773438,
      "loss": 0.4991,
      "rewards/accuracies": 0.6937500238418579,
      "rewards/chosen": -0.8999192118644714,
      "rewards/margins": 0.8702341318130493,
      "rewards/rejected": -1.770153284072876,
      "step": 105
    },
    {
      "epoch": 0.37606837606837606,
      "grad_norm": 27.339023914835686,
      "learning_rate": 9.674586261416873e-07,
      "logits/chosen": -2.4866347312927246,
      "logits/rejected": -2.4518179893493652,
      "logps/chosen": -179.46290588378906,
      "logps/rejected": -188.7920379638672,
      "loss": 0.5213,
      "rewards/accuracies": 0.706250011920929,
      "rewards/chosen": -0.6938365697860718,
      "rewards/margins": 1.0765600204467773,
      "rewards/rejected": -1.7703965902328491,
      "step": 110
    },
    {
      "epoch": 0.39316239316239315,
      "grad_norm": 31.312902469600562,
      "learning_rate": 9.641635351911664e-07,
      "logits/chosen": -2.4456398487091064,
      "logits/rejected": -2.426159381866455,
      "logps/chosen": -170.3855438232422,
      "logps/rejected": -181.9676513671875,
      "loss": 0.4823,
      "rewards/accuracies": 0.71875,
      "rewards/chosen": -0.8490931391716003,
      "rewards/margins": 1.2224478721618652,
      "rewards/rejected": -2.0715408325195312,
      "step": 115
    },
    {
      "epoch": 0.41025641025641024,
      "grad_norm": 25.966469642807997,
      "learning_rate": 9.607157357860821e-07,
      "logits/chosen": -2.4072113037109375,
      "logits/rejected": -2.3874144554138184,
      "logps/chosen": -187.41197204589844,
      "logps/rejected": -201.69454956054688,
      "loss": 0.5037,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -1.0798847675323486,
      "rewards/margins": 1.3365159034729004,
      "rewards/rejected": -2.41640043258667,
      "step": 120
    },
    {
      "epoch": 0.42735042735042733,
      "grad_norm": 32.18242375190423,
      "learning_rate": 9.571163622424225e-07,
      "logits/chosen": -2.2766659259796143,
      "logits/rejected": -2.252072811126709,
      "logps/chosen": -174.78514099121094,
      "logps/rejected": -187.40646362304688,
      "loss": 0.5019,
      "rewards/accuracies": 0.706250011920929,
      "rewards/chosen": -1.5247443914413452,
      "rewards/margins": 1.2177503108978271,
      "rewards/rejected": -2.742494821548462,
      "step": 125
    },
    {
      "epoch": 0.4444444444444444,
      "grad_norm": 30.579550576640443,
      "learning_rate": 9.533665987436261e-07,
      "logits/chosen": -2.182610034942627,
      "logits/rejected": -2.128113269805908,
      "logps/chosen": -178.1033477783203,
      "logps/rejected": -197.4458465576172,
      "loss": 0.4957,
      "rewards/accuracies": 0.6875,
      "rewards/chosen": -1.5623412132263184,
      "rewards/margins": 1.2608497142791748,
      "rewards/rejected": -2.823190689086914,
      "step": 130
    },
    {
      "epoch": 0.46153846153846156,
      "grad_norm": 29.287644373971865,
      "learning_rate": 9.494676789509899e-07,
      "logits/chosen": -2.1067867279052734,
      "logits/rejected": -2.0683400630950928,
      "logps/chosen": -176.67918395996094,
      "logps/rejected": -193.65371704101562,
      "loss": 0.4939,
      "rewards/accuracies": 0.762499988079071,
      "rewards/chosen": -1.1053364276885986,
      "rewards/margins": 1.3829355239868164,
      "rewards/rejected": -2.488272190093994,
      "step": 135
    },
    {
      "epoch": 0.47863247863247865,
      "grad_norm": 27.563555703636343,
      "learning_rate": 9.454208855977985e-07,
      "logits/chosen": -2.0855822563171387,
      "logits/rejected": -2.013296127319336,
      "logps/chosen": -178.40390014648438,
      "logps/rejected": -196.03305053710938,
      "loss": 0.4715,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -1.443866491317749,
      "rewards/margins": 1.600778341293335,
      "rewards/rejected": -3.044644832611084,
      "step": 140
    },
    {
      "epoch": 0.49572649572649574,
      "grad_norm": 30.234814125811326,
      "learning_rate": 9.41227550067308e-07,
      "logits/chosen": -2.0734238624572754,
      "logits/rejected": -2.0634400844573975,
      "logps/chosen": -179.51080322265625,
      "logps/rejected": -191.87046813964844,
      "loss": 0.4798,
      "rewards/accuracies": 0.7749999761581421,
      "rewards/chosen": -1.6011661291122437,
      "rewards/margins": 1.445229172706604,
      "rewards/rejected": -3.0463955402374268,
      "step": 145
    },
    {
      "epoch": 0.5128205128205128,
      "grad_norm": 31.371346339775513,
      "learning_rate": 9.36889051954725e-07,
      "logits/chosen": -2.127821683883667,
      "logits/rejected": -2.080082416534424,
      "logps/chosen": -180.66383361816406,
      "logps/rejected": -196.1031494140625,
      "loss": 0.4954,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -1.6335647106170654,
      "rewards/margins": 1.5422546863555908,
      "rewards/rejected": -3.1758196353912354,
      "step": 150
    },
    {
      "epoch": 0.5128205128205128,
      "eval_logits/chosen": -2.1472573280334473,
      "eval_logits/rejected": -2.0990829467773438,
      "eval_logps/chosen": -173.8290557861328,
      "eval_logps/rejected": -191.683349609375,
      "eval_loss": 0.47699737548828125,
      "eval_rewards/accuracies": 0.754807710647583,
      "eval_rewards/chosen": -1.5351712703704834,
      "eval_rewards/margins": 1.7451337575912476,
      "eval_rewards/rejected": -3.2803049087524414,
      "eval_runtime": 510.7048,
      "eval_samples_per_second": 16.279,
      "eval_steps_per_second": 0.255,
      "step": 150
    },
    {
      "epoch": 0.5299145299145299,
      "grad_norm": 27.790777356361556,
      "learning_rate": 9.324068186133245e-07,
      "logits/chosen": -2.1372084617614746,
      "logits/rejected": -2.124948024749756,
      "logps/chosen": -172.3369598388672,
      "logps/rejected": -186.5850372314453,
      "loss": 0.4644,
      "rewards/accuracies": 0.7875000238418579,
      "rewards/chosen": -1.3583369255065918,
      "rewards/margins": 1.7439367771148682,
      "rewards/rejected": -3.102273464202881,
      "step": 155
    },
    {
      "epoch": 0.5470085470085471,
      "grad_norm": 27.612583401785376,
      "learning_rate": 9.277823246848536e-07,
      "logits/chosen": -2.2635793685913086,
      "logits/rejected": -2.2123026847839355,
      "logps/chosen": -186.25137329101562,
      "logps/rejected": -196.69564819335938,
      "loss": 0.4547,
      "rewards/accuracies": 0.71875,
      "rewards/chosen": -1.2192834615707397,
      "rewards/margins": 1.371063470840454,
      "rewards/rejected": -2.5903468132019043,
      "step": 160
    },
    {
      "epoch": 0.5641025641025641,
      "grad_norm": 28.168886287584876,
      "learning_rate": 9.230170916143793e-07,
      "logits/chosen": -2.3309006690979004,
      "logits/rejected": -2.2978808879852295,
      "logps/chosen": -174.7559814453125,
      "logps/rejected": -195.68280029296875,
      "loss": 0.4967,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -1.0855658054351807,
      "rewards/margins": 1.6249233484268188,
      "rewards/rejected": -2.710489273071289,
      "step": 165
    },
    {
      "epoch": 0.5811965811965812,
      "grad_norm": 28.93959851544435,
      "learning_rate": 9.181126871497378e-07,
      "logits/chosen": -2.376833915710449,
      "logits/rejected": -2.340681552886963,
      "logps/chosen": -175.3675537109375,
      "logps/rejected": -194.9619903564453,
      "loss": 0.4651,
      "rewards/accuracies": 0.7875000238418579,
      "rewards/chosen": -0.9624043703079224,
      "rewards/margins": 1.7745708227157593,
      "rewards/rejected": -2.7369751930236816,
      "step": 170
    },
    {
      "epoch": 0.5982905982905983,
      "grad_norm": 30.43477724579486,
      "learning_rate": 9.130707248257491e-07,
      "logits/chosen": -2.458378553390503,
      "logits/rejected": -2.4171223640441895,
      "logps/chosen": -168.79849243164062,
      "logps/rejected": -178.6556396484375,
      "loss": 0.4728,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -0.9254748225212097,
      "rewards/margins": 1.5751961469650269,
      "rewards/rejected": -2.500671148300171,
      "step": 175
    },
    {
      "epoch": 0.6153846153846154,
      "grad_norm": 25.220318056395065,
      "learning_rate": 9.078928634333698e-07,
      "logits/chosen": -2.4454641342163086,
      "logits/rejected": -2.4170265197753906,
      "logps/chosen": -181.41317749023438,
      "logps/rejected": -199.88668823242188,
      "loss": 0.4526,
      "rewards/accuracies": 0.824999988079071,
      "rewards/chosen": -0.84205561876297,
      "rewards/margins": 1.7803510427474976,
      "rewards/rejected": -2.622406482696533,
      "step": 180
    },
    {
      "epoch": 0.6324786324786325,
      "grad_norm": 29.414031929374275,
      "learning_rate": 9.025808064739549e-07,
      "logits/chosen": -2.4103400707244873,
      "logits/rejected": -2.370731830596924,
      "logps/chosen": -178.70916748046875,
      "logps/rejected": -193.0004119873047,
      "loss": 0.4891,
      "rewards/accuracies": 0.768750011920929,
      "rewards/chosen": -1.174070119857788,
      "rewards/margins": 1.535239338874817,
      "rewards/rejected": -2.7093093395233154,
      "step": 185
    },
    {
      "epoch": 0.6495726495726496,
      "grad_norm": 26.0372223221703,
      "learning_rate": 8.971363015988113e-07,
      "logits/chosen": -2.3428735733032227,
      "logits/rejected": -2.2986531257629395,
      "logps/chosen": -173.8651580810547,
      "logps/rejected": -194.9317626953125,
      "loss": 0.4643,
      "rewards/accuracies": 0.78125,
      "rewards/chosen": -1.1425771713256836,
      "rewards/margins": 1.611919641494751,
      "rewards/rejected": -2.7544968128204346,
      "step": 190
    },
    {
      "epoch": 0.6666666666666666,
      "grad_norm": 33.87434178682573,
      "learning_rate": 8.91561140034225e-07,
      "logits/chosen": -2.2664923667907715,
      "logits/rejected": -2.2088184356689453,
      "logps/chosen": -172.7240753173828,
      "logps/rejected": -193.1275177001953,
      "loss": 0.5029,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -1.3134868144989014,
      "rewards/margins": 1.5050963163375854,
      "rewards/rejected": -2.8185834884643555,
      "step": 195
    },
    {
      "epoch": 0.6837606837606838,
      "grad_norm": 25.21313391058931,
      "learning_rate": 8.858571559921537e-07,
      "logits/chosen": -2.191737174987793,
      "logits/rejected": -2.1188113689422607,
      "logps/chosen": -174.46722412109375,
      "logps/rejected": -188.4197540283203,
      "loss": 0.4567,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -1.026064157485962,
      "rewards/margins": 1.5865710973739624,
      "rewards/rejected": -2.612635374069214,
      "step": 200
    },
    {
      "epoch": 0.6837606837606838,
      "eval_logits/chosen": -2.1586899757385254,
      "eval_logits/rejected": -2.1090493202209473,
      "eval_logps/chosen": -170.4287872314453,
      "eval_logps/rejected": -187.2865447998047,
      "eval_loss": 0.45979756116867065,
      "eval_rewards/accuracies": 0.7596153616905212,
      "eval_rewards/chosen": -1.1951465606689453,
      "eval_rewards/margins": 1.64547860622406,
      "eval_rewards/rejected": -2.840625047683716,
      "eval_runtime": 510.4854,
      "eval_samples_per_second": 16.286,
      "eval_steps_per_second": 0.255,
      "step": 200
    }
  ],
  "logging_steps": 5,
  "max_steps": 876,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 3,
  "save_steps": 100,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": false
      },
      "attributes": {}
    }
  },
  "total_flos": 2358113407598592.0,
  "train_batch_size": 8,
  "trial_name": null,
  "trial_params": null
}