File size: 34,316 Bytes
cea8f12
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.9996020692399522,
  "eval_steps": 1000,
  "global_step": 628,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0015917230401910067,
      "grad_norm": 3.414436464235707,
      "learning_rate": 7.936507936507936e-09,
      "logits/chosen": -2.458916187286377,
      "logits/rejected": -2.50296688079834,
      "logps/chosen": -112.41877746582031,
      "logps/rejected": -153.82745361328125,
      "loss": 0.6931,
      "rewards/accuracies": 0.0,
      "rewards/chosen": 0.0,
      "rewards/margins": 0.0,
      "rewards/rejected": 0.0,
      "step": 1
    },
    {
      "epoch": 0.01591723040191007,
      "grad_norm": 3.47500960837261,
      "learning_rate": 7.936507936507936e-08,
      "logits/chosen": -2.5214052200317383,
      "logits/rejected": -2.508702039718628,
      "logps/chosen": -143.01979064941406,
      "logps/rejected": -133.09210205078125,
      "loss": 0.6932,
      "rewards/accuracies": 0.4166666567325592,
      "rewards/chosen": -0.00031594105530530214,
      "rewards/margins": -0.00045727533870376647,
      "rewards/rejected": 0.00014133438526187092,
      "step": 10
    },
    {
      "epoch": 0.03183446080382014,
      "grad_norm": 3.343704534649061,
      "learning_rate": 1.5873015873015872e-07,
      "logits/chosen": -2.4521286487579346,
      "logits/rejected": -2.448809862136841,
      "logps/chosen": -151.37095642089844,
      "logps/rejected": -162.80328369140625,
      "loss": 0.6931,
      "rewards/accuracies": 0.43437498807907104,
      "rewards/chosen": 4.1505660192342475e-05,
      "rewards/margins": -0.00036320873186923563,
      "rewards/rejected": 0.000404714432079345,
      "step": 20
    },
    {
      "epoch": 0.0477516912057302,
      "grad_norm": 3.366996896296212,
      "learning_rate": 2.3809523809523806e-07,
      "logits/chosen": -2.4904391765594482,
      "logits/rejected": -2.479889392852783,
      "logps/chosen": -147.56060791015625,
      "logps/rejected": -150.96810913085938,
      "loss": 0.6921,
      "rewards/accuracies": 0.550000011920929,
      "rewards/chosen": 0.003838698612526059,
      "rewards/margins": 0.002006606664508581,
      "rewards/rejected": 0.001832091948017478,
      "step": 30
    },
    {
      "epoch": 0.06366892160764027,
      "grad_norm": 3.3714607115747497,
      "learning_rate": 3.1746031746031743e-07,
      "logits/chosen": -2.503981590270996,
      "logits/rejected": -2.486074447631836,
      "logps/chosen": -153.0246124267578,
      "logps/rejected": -145.2753448486328,
      "loss": 0.6899,
      "rewards/accuracies": 0.53125,
      "rewards/chosen": 0.007592098321765661,
      "rewards/margins": 0.00429057702422142,
      "rewards/rejected": 0.0033015217632055283,
      "step": 40
    },
    {
      "epoch": 0.07958615200955034,
      "grad_norm": 4.06423931986909,
      "learning_rate": 3.968253968253968e-07,
      "logits/chosen": -2.4848971366882324,
      "logits/rejected": -2.4830713272094727,
      "logps/chosen": -146.40438842773438,
      "logps/rejected": -152.903564453125,
      "loss": 0.6849,
      "rewards/accuracies": 0.5531250238418579,
      "rewards/chosen": -0.01682356372475624,
      "rewards/margins": 0.0171933826059103,
      "rewards/rejected": -0.03401694819331169,
      "step": 50
    },
    {
      "epoch": 0.0955033824114604,
      "grad_norm": 3.933215748835345,
      "learning_rate": 4.761904761904761e-07,
      "logits/chosen": -2.4494073390960693,
      "logits/rejected": -2.425264835357666,
      "logps/chosen": -164.67039489746094,
      "logps/rejected": -159.42379760742188,
      "loss": 0.6811,
      "rewards/accuracies": 0.528124988079071,
      "rewards/chosen": -0.13071545958518982,
      "rewards/margins": 0.01678677648305893,
      "rewards/rejected": -0.14750224351882935,
      "step": 60
    },
    {
      "epoch": 0.11142061281337047,
      "grad_norm": 6.465696654064857,
      "learning_rate": 4.998106548810311e-07,
      "logits/chosen": -2.4344656467437744,
      "logits/rejected": -2.3922276496887207,
      "logps/chosen": -171.0955352783203,
      "logps/rejected": -158.62640380859375,
      "loss": 0.6666,
      "rewards/accuracies": 0.596875011920929,
      "rewards/chosen": -0.17716214060783386,
      "rewards/margins": 0.07567773759365082,
      "rewards/rejected": -0.2528398633003235,
      "step": 70
    },
    {
      "epoch": 0.12733784321528055,
      "grad_norm": 9.086502750868021,
      "learning_rate": 4.988839406031596e-07,
      "logits/chosen": -2.384453058242798,
      "logits/rejected": -2.3677101135253906,
      "logps/chosen": -153.10414123535156,
      "logps/rejected": -182.54360961914062,
      "loss": 0.6529,
      "rewards/accuracies": 0.690625011920929,
      "rewards/chosen": -0.20188739895820618,
      "rewards/margins": 0.12667986750602722,
      "rewards/rejected": -0.3285672962665558,
      "step": 80
    },
    {
      "epoch": 0.14325507361719061,
      "grad_norm": 14.480905623348013,
      "learning_rate": 4.971879403278432e-07,
      "logits/chosen": -2.2432703971862793,
      "logits/rejected": -2.2313826084136963,
      "logps/chosen": -185.1287384033203,
      "logps/rejected": -201.3049774169922,
      "loss": 0.6398,
      "rewards/accuracies": 0.643750011920929,
      "rewards/chosen": -0.39208894968032837,
      "rewards/margins": 0.144916832447052,
      "rewards/rejected": -0.5370057821273804,
      "step": 90
    },
    {
      "epoch": 0.15917230401910068,
      "grad_norm": 12.626265646656822,
      "learning_rate": 4.947278962947386e-07,
      "logits/chosen": -1.7865173816680908,
      "logits/rejected": -1.7619308233261108,
      "logps/chosen": -234.49526977539062,
      "logps/rejected": -266.22235107421875,
      "loss": 0.6186,
      "rewards/accuracies": 0.659375011920929,
      "rewards/chosen": -0.8486677408218384,
      "rewards/margins": 0.2785636782646179,
      "rewards/rejected": -1.127231478691101,
      "step": 100
    },
    {
      "epoch": 0.17508953442101075,
      "grad_norm": 11.440443186979932,
      "learning_rate": 4.915114123589732e-07,
      "logits/chosen": -1.477772831916809,
      "logits/rejected": -1.394840955734253,
      "logps/chosen": -206.17166137695312,
      "logps/rejected": -243.156005859375,
      "loss": 0.6156,
      "rewards/accuracies": 0.684374988079071,
      "rewards/chosen": -0.719870388507843,
      "rewards/margins": 0.34280186891555786,
      "rewards/rejected": -1.0626722574234009,
      "step": 110
    },
    {
      "epoch": 0.1910067648229208,
      "grad_norm": 14.957458897670607,
      "learning_rate": 4.875484304880629e-07,
      "logits/chosen": -0.8948208093643188,
      "logits/rejected": -0.773229718208313,
      "logps/chosen": -234.4495086669922,
      "logps/rejected": -270.4801330566406,
      "loss": 0.5931,
      "rewards/accuracies": 0.699999988079071,
      "rewards/chosen": -0.8048633337020874,
      "rewards/margins": 0.34693923592567444,
      "rewards/rejected": -1.151802659034729,
      "step": 120
    },
    {
      "epoch": 0.20692399522483088,
      "grad_norm": 13.673290760971415,
      "learning_rate": 4.828512000318616e-07,
      "logits/chosen": -0.20325860381126404,
      "logits/rejected": -0.022473735734820366,
      "logps/chosen": -233.56399536132812,
      "logps/rejected": -278.225830078125,
      "loss": 0.5842,
      "rewards/accuracies": 0.7093750238418579,
      "rewards/chosen": -0.7881739139556885,
      "rewards/margins": 0.4328451156616211,
      "rewards/rejected": -1.2210190296173096,
      "step": 130
    },
    {
      "epoch": 0.22284122562674094,
      "grad_norm": 15.252755388662429,
      "learning_rate": 4.774342398605221e-07,
      "logits/chosen": -0.4290715754032135,
      "logits/rejected": -0.26301848888397217,
      "logps/chosen": -245.796142578125,
      "logps/rejected": -278.6325988769531,
      "loss": 0.5803,
      "rewards/accuracies": 0.6781250238418579,
      "rewards/chosen": -0.9634370803833008,
      "rewards/margins": 0.42050686478614807,
      "rewards/rejected": -1.383944034576416,
      "step": 140
    },
    {
      "epoch": 0.238758456028651,
      "grad_norm": 15.98257839847235,
      "learning_rate": 4.713142934875005e-07,
      "logits/chosen": 0.04155537486076355,
      "logits/rejected": 0.4105464518070221,
      "logps/chosen": -294.55078125,
      "logps/rejected": -322.9659118652344,
      "loss": 0.5928,
      "rewards/accuracies": 0.684374988079071,
      "rewards/chosen": -1.3449639081954956,
      "rewards/margins": 0.41515955328941345,
      "rewards/rejected": -1.7601234912872314,
      "step": 150
    },
    {
      "epoch": 0.2546756864305611,
      "grad_norm": 16.296020720117806,
      "learning_rate": 4.64510277316316e-07,
      "logits/chosen": 0.07806523889303207,
      "logits/rejected": 0.4625118672847748,
      "logps/chosen": -242.23171997070312,
      "logps/rejected": -306.7259216308594,
      "loss": 0.5619,
      "rewards/accuracies": 0.7281249761581421,
      "rewards/chosen": -1.0831429958343506,
      "rewards/margins": 0.6153541803359985,
      "rewards/rejected": -1.6984970569610596,
      "step": 160
    },
    {
      "epoch": 0.27059291683247116,
      "grad_norm": 15.29787072223291,
      "learning_rate": 4.570432221710314e-07,
      "logits/chosen": 0.8908351063728333,
      "logits/rejected": 1.1234996318817139,
      "logps/chosen": -285.02685546875,
      "logps/rejected": -332.22576904296875,
      "loss": 0.5968,
      "rewards/accuracies": 0.6499999761581421,
      "rewards/chosen": -1.369902491569519,
      "rewards/margins": 0.42481645941734314,
      "rewards/rejected": -1.7947190999984741,
      "step": 170
    },
    {
      "epoch": 0.28651014723438123,
      "grad_norm": 12.440638905265972,
      "learning_rate": 4.4893620829118124e-07,
      "logits/chosen": 0.45191067457199097,
      "logits/rejected": 0.8097447156906128,
      "logps/chosen": -267.55511474609375,
      "logps/rejected": -303.0255432128906,
      "loss": 0.5782,
      "rewards/accuracies": 0.690625011920929,
      "rewards/chosen": -1.130518913269043,
      "rewards/margins": 0.47403573989868164,
      "rewards/rejected": -1.6045547723770142,
      "step": 180
    },
    {
      "epoch": 0.3024273776362913,
      "grad_norm": 14.662664269581565,
      "learning_rate": 4.40214293992074e-07,
      "logits/chosen": 0.07046165317296982,
      "logits/rejected": 0.5015997886657715,
      "logps/chosen": -276.8005676269531,
      "logps/rejected": -317.6512451171875,
      "loss": 0.5684,
      "rewards/accuracies": 0.699999988079071,
      "rewards/chosen": -1.328792929649353,
      "rewards/margins": 0.5068601369857788,
      "rewards/rejected": -1.835653305053711,
      "step": 190
    },
    {
      "epoch": 0.31834460803820136,
      "grad_norm": 13.757436126080059,
      "learning_rate": 4.3090443821097566e-07,
      "logits/chosen": 0.2187742292881012,
      "logits/rejected": 0.5610963106155396,
      "logps/chosen": -268.94757080078125,
      "logps/rejected": -333.0284729003906,
      "loss": 0.5596,
      "rewards/accuracies": 0.7124999761581421,
      "rewards/chosen": -1.2326629161834717,
      "rewards/margins": 0.5554476976394653,
      "rewards/rejected": -1.7881107330322266,
      "step": 200
    },
    {
      "epoch": 0.3342618384401114,
      "grad_norm": 12.638546225207325,
      "learning_rate": 4.210354171785795e-07,
      "logits/chosen": 0.5555538535118103,
      "logits/rejected": 0.9611243009567261,
      "logps/chosen": -258.8009338378906,
      "logps/rejected": -315.17242431640625,
      "loss": 0.5627,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -1.1692590713500977,
      "rewards/margins": 0.5407182574272156,
      "rewards/rejected": -1.709977388381958,
      "step": 210
    },
    {
      "epoch": 0.3501790688420215,
      "grad_norm": 12.533508219187622,
      "learning_rate": 4.1063773547332584e-07,
      "logits/chosen": 0.4853738248348236,
      "logits/rejected": 0.7825923562049866,
      "logps/chosen": -255.5019073486328,
      "logps/rejected": -307.07843017578125,
      "loss": 0.5684,
      "rewards/accuracies": 0.6656249761581421,
      "rewards/chosen": -1.0966681241989136,
      "rewards/margins": 0.5020471215248108,
      "rewards/rejected": -1.59871506690979,
      "step": 220
    },
    {
      "epoch": 0.36609629924393156,
      "grad_norm": 13.14652988888945,
      "learning_rate": 3.997435317334988e-07,
      "logits/chosen": 1.3910820484161377,
      "logits/rejected": 1.7489607334136963,
      "logps/chosen": -285.7388916015625,
      "logps/rejected": -342.02557373046875,
      "loss": 0.5556,
      "rewards/accuracies": 0.731249988079071,
      "rewards/chosen": -1.430323600769043,
      "rewards/margins": 0.5691004991531372,
      "rewards/rejected": -1.9994239807128906,
      "step": 230
    },
    {
      "epoch": 0.3820135296458416,
      "grad_norm": 16.893312636051345,
      "learning_rate": 3.8838647931853684e-07,
      "logits/chosen": 0.9681981801986694,
      "logits/rejected": 1.3831474781036377,
      "logps/chosen": -274.5780029296875,
      "logps/rejected": -334.71624755859375,
      "loss": 0.565,
      "rewards/accuracies": 0.684374988079071,
      "rewards/chosen": -1.331866979598999,
      "rewards/margins": 0.6004709005355835,
      "rewards/rejected": -1.932337999343872,
      "step": 240
    },
    {
      "epoch": 0.3979307600477517,
      "grad_norm": 16.93758119869443,
      "learning_rate": 3.7660168222660824e-07,
      "logits/chosen": -0.0617893747985363,
      "logits/rejected": 0.3982541561126709,
      "logps/chosen": -291.5325012207031,
      "logps/rejected": -349.10357666015625,
      "loss": 0.546,
      "rewards/accuracies": 0.7281249761581421,
      "rewards/chosen": -1.3192743062973022,
      "rewards/margins": 0.6018632650375366,
      "rewards/rejected": -1.9211375713348389,
      "step": 250
    },
    {
      "epoch": 0.41384799044966175,
      "grad_norm": 13.594747823748454,
      "learning_rate": 3.6442556659016475e-07,
      "logits/chosen": -0.1596679985523224,
      "logits/rejected": 0.12459886074066162,
      "logps/chosen": -283.2134704589844,
      "logps/rejected": -335.0664367675781,
      "loss": 0.5677,
      "rewards/accuracies": 0.703125,
      "rewards/chosen": -1.3012964725494385,
      "rewards/margins": 0.5175098180770874,
      "rewards/rejected": -1.8188062906265259,
      "step": 260
    },
    {
      "epoch": 0.4297652208515718,
      "grad_norm": 13.029006886811343,
      "learning_rate": 3.5189576808485404e-07,
      "logits/chosen": 0.6624835729598999,
      "logits/rejected": 0.867287278175354,
      "logps/chosen": -290.3371887207031,
      "logps/rejected": -335.45745849609375,
      "loss": 0.5603,
      "rewards/accuracies": 0.7093750238418579,
      "rewards/chosen": -1.3652312755584717,
      "rewards/margins": 0.44845834374427795,
      "rewards/rejected": -1.8136895895004272,
      "step": 270
    },
    {
      "epoch": 0.4456824512534819,
      "grad_norm": 14.229518064110062,
      "learning_rate": 3.390510155998023e-07,
      "logits/chosen": 1.0853196382522583,
      "logits/rejected": 1.5426385402679443,
      "logps/chosen": -301.244873046875,
      "logps/rejected": -352.26214599609375,
      "loss": 0.5576,
      "rewards/accuracies": 0.721875011920929,
      "rewards/chosen": -1.4498573541641235,
      "rewards/margins": 0.5846196413040161,
      "rewards/rejected": -2.0344767570495605,
      "step": 280
    },
    {
      "epoch": 0.46159968165539195,
      "grad_norm": 13.247814891025222,
      "learning_rate": 3.2593101152883795e-07,
      "logits/chosen": 0.6425064206123352,
      "logits/rejected": 1.0470354557037354,
      "logps/chosen": -304.13165283203125,
      "logps/rejected": -356.2009582519531,
      "loss": 0.5699,
      "rewards/accuracies": 0.6781250238418579,
      "rewards/chosen": -1.4935743808746338,
      "rewards/margins": 0.543701708316803,
      "rewards/rejected": -2.037276029586792,
      "step": 290
    },
    {
      "epoch": 0.477516912057302,
      "grad_norm": 14.136317593633118,
      "learning_rate": 3.125763090526674e-07,
      "logits/chosen": 1.010498285293579,
      "logits/rejected": 1.5458686351776123,
      "logps/chosen": -312.71893310546875,
      "logps/rejected": -367.11956787109375,
      "loss": 0.5526,
      "rewards/accuracies": 0.6968749761581421,
      "rewards/chosen": -1.607622742652893,
      "rewards/margins": 0.5898184180259705,
      "rewards/rejected": -2.1974411010742188,
      "step": 300
    },
    {
      "epoch": 0.4934341424592121,
      "grad_norm": 13.599580845693339,
      "learning_rate": 2.9902818679131775e-07,
      "logits/chosen": 1.556601881980896,
      "logits/rejected": 2.024049758911133,
      "logps/chosen": -317.42938232421875,
      "logps/rejected": -371.38623046875,
      "loss": 0.5571,
      "rewards/accuracies": 0.6937500238418579,
      "rewards/chosen": -1.6657898426055908,
      "rewards/margins": 0.5584100484848022,
      "rewards/rejected": -2.2241997718811035,
      "step": 310
    },
    {
      "epoch": 0.5093513728611222,
      "grad_norm": 15.486325517579422,
      "learning_rate": 2.8532852121428733e-07,
      "logits/chosen": 2.424159526824951,
      "logits/rejected": 2.8681328296661377,
      "logps/chosen": -326.8099670410156,
      "logps/rejected": -373.8744201660156,
      "loss": 0.5411,
      "rewards/accuracies": 0.6656249761581421,
      "rewards/chosen": -1.7966935634613037,
      "rewards/margins": 0.5875706076622009,
      "rewards/rejected": -2.3842644691467285,
      "step": 320
    },
    {
      "epoch": 0.5252686032630323,
      "grad_norm": 13.394275013231457,
      "learning_rate": 2.715196572027789e-07,
      "logits/chosen": 3.215583324432373,
      "logits/rejected": 3.725792646408081,
      "logps/chosen": -343.72344970703125,
      "logps/rejected": -432.32916259765625,
      "loss": 0.5414,
      "rewards/accuracies": 0.75,
      "rewards/chosen": -1.9976279735565186,
      "rewards/margins": 0.806243896484375,
      "rewards/rejected": -2.8038716316223145,
      "step": 330
    },
    {
      "epoch": 0.5411858336649423,
      "grad_norm": 13.373618178706888,
      "learning_rate": 2.5764427716409815e-07,
      "logits/chosen": 3.000930070877075,
      "logits/rejected": 3.6059279441833496,
      "logps/chosen": -347.54022216796875,
      "logps/rejected": -404.14654541015625,
      "loss": 0.5507,
      "rewards/accuracies": 0.71875,
      "rewards/chosen": -1.9271198511123657,
      "rewards/margins": 0.6116258502006531,
      "rewards/rejected": -2.538745880126953,
      "step": 340
    },
    {
      "epoch": 0.5571030640668524,
      "grad_norm": 14.475581092351593,
      "learning_rate": 2.4374526910277886e-07,
      "logits/chosen": 2.816096782684326,
      "logits/rejected": 3.264390230178833,
      "logps/chosen": -353.548828125,
      "logps/rejected": -420.4794006347656,
      "loss": 0.5564,
      "rewards/accuracies": 0.671875,
      "rewards/chosen": -2.0463950634002686,
      "rewards/margins": 0.6025586128234863,
      "rewards/rejected": -2.648953914642334,
      "step": 350
    },
    {
      "epoch": 0.5730202944687625,
      "grad_norm": 14.682353004178683,
      "learning_rate": 2.2986559405621886e-07,
      "logits/chosen": 2.8994641304016113,
      "logits/rejected": 3.608001232147217,
      "logps/chosen": -360.53143310546875,
      "logps/rejected": -417.3299865722656,
      "loss": 0.5525,
      "rewards/accuracies": 0.7124999761581421,
      "rewards/chosen": -1.981393814086914,
      "rewards/margins": 0.6567780375480652,
      "rewards/rejected": -2.638171911239624,
      "step": 360
    },
    {
      "epoch": 0.5889375248706725,
      "grad_norm": 18.068831076710342,
      "learning_rate": 2.160481533045751e-07,
      "logits/chosen": 2.5152053833007812,
      "logits/rejected": 3.3655197620391846,
      "logps/chosen": -339.1783752441406,
      "logps/rejected": -407.2680358886719,
      "loss": 0.5377,
      "rewards/accuracies": 0.7437499761581421,
      "rewards/chosen": -1.932498574256897,
      "rewards/margins": 0.7526292204856873,
      "rewards/rejected": -2.6851277351379395,
      "step": 370
    },
    {
      "epoch": 0.6048547552725826,
      "grad_norm": 13.903302227889153,
      "learning_rate": 2.0233565576536564e-07,
      "logits/chosen": 2.7033469676971436,
      "logits/rejected": 3.3517203330993652,
      "logps/chosen": -352.88543701171875,
      "logps/rejected": -414.46807861328125,
      "loss": 0.5464,
      "rewards/accuracies": 0.671875,
      "rewards/chosen": -2.0058979988098145,
      "rewards/margins": 0.6991748809814453,
      "rewards/rejected": -2.7050728797912598,
      "step": 380
    },
    {
      "epoch": 0.6207719856744927,
      "grad_norm": 14.437387596932032,
      "learning_rate": 1.887704859826528e-07,
      "logits/chosen": 2.677304744720459,
      "logits/rejected": 3.3016350269317627,
      "logps/chosen": -345.10186767578125,
      "logps/rejected": -422.6465759277344,
      "loss": 0.5441,
      "rewards/accuracies": 0.753125011920929,
      "rewards/chosen": -1.9631824493408203,
      "rewards/margins": 0.7392277717590332,
      "rewards/rejected": -2.7024102210998535,
      "step": 390
    },
    {
      "epoch": 0.6366892160764027,
      "grad_norm": 15.504689068952228,
      "learning_rate": 1.7539457311884675e-07,
      "logits/chosen": 2.6285808086395264,
      "logits/rejected": 3.402756452560425,
      "logps/chosen": -350.1034240722656,
      "logps/rejected": -420.3218688964844,
      "loss": 0.5387,
      "rewards/accuracies": 0.737500011920929,
      "rewards/chosen": -1.9105831384658813,
      "rewards/margins": 0.7946080565452576,
      "rewards/rejected": -2.705191135406494,
      "step": 400
    },
    {
      "epoch": 0.6526064464783128,
      "grad_norm": 12.730811962712194,
      "learning_rate": 1.6224926135406693e-07,
      "logits/chosen": 2.7364511489868164,
      "logits/rejected": 3.1644163131713867,
      "logps/chosen": -364.25469970703125,
      "logps/rejected": -419.5701599121094,
      "loss": 0.5618,
      "rewards/accuracies": 0.690625011920929,
      "rewards/chosen": -2.067072629928589,
      "rewards/margins": 0.5998597741127014,
      "rewards/rejected": -2.6669325828552246,
      "step": 410
    },
    {
      "epoch": 0.6685236768802229,
      "grad_norm": 14.2356920381672,
      "learning_rate": 1.4937518209365108e-07,
      "logits/chosen": 2.5111005306243896,
      "logits/rejected": 3.2514469623565674,
      "logps/chosen": -381.5763244628906,
      "logps/rejected": -433.4600524902344,
      "loss": 0.5494,
      "rewards/accuracies": 0.778124988079071,
      "rewards/chosen": -2.099307060241699,
      "rewards/margins": 0.6568456888198853,
      "rewards/rejected": -2.756152629852295,
      "step": 420
    },
    {
      "epoch": 0.6844409072821329,
      "grad_norm": 15.52832700397283,
      "learning_rate": 1.3681212837880977e-07,
      "logits/chosen": 2.709282398223877,
      "logits/rejected": 3.0862138271331787,
      "logps/chosen": -374.52288818359375,
      "logps/rejected": -447.3203125,
      "loss": 0.5539,
      "rewards/accuracies": 0.71875,
      "rewards/chosen": -2.311601161956787,
      "rewards/margins": 0.6582018136978149,
      "rewards/rejected": -2.9698030948638916,
      "step": 430
    },
    {
      "epoch": 0.700358137684043,
      "grad_norm": 13.300949643909883,
      "learning_rate": 1.2459893188861613e-07,
      "logits/chosen": 2.896282196044922,
      "logits/rejected": 3.553002119064331,
      "logps/chosen": -382.75054931640625,
      "logps/rejected": -458.756591796875,
      "loss": 0.5426,
      "rewards/accuracies": 0.765625,
      "rewards/chosen": -2.2905333042144775,
      "rewards/margins": 0.7688990831375122,
      "rewards/rejected": -3.0594325065612793,
      "step": 440
    },
    {
      "epoch": 0.716275368085953,
      "grad_norm": 14.064737088004772,
      "learning_rate": 1.1277334291351145e-07,
      "logits/chosen": 3.348231077194214,
      "logits/rejected": 4.010863780975342,
      "logps/chosen": -383.77569580078125,
      "logps/rejected": -465.36767578125,
      "loss": 0.5372,
      "rewards/accuracies": 0.7281249761581421,
      "rewards/chosen": -2.3260040283203125,
      "rewards/margins": 0.8284826278686523,
      "rewards/rejected": -3.1544864177703857,
      "step": 450
    },
    {
      "epoch": 0.7321925984878631,
      "grad_norm": 13.977439772366749,
      "learning_rate": 1.0137191367132078e-07,
      "logits/chosen": 3.446108341217041,
      "logits/rejected": 3.9590706825256348,
      "logps/chosen": -375.79705810546875,
      "logps/rejected": -461.6592712402344,
      "loss": 0.5282,
      "rewards/accuracies": 0.7250000238418579,
      "rewards/chosen": -2.228874683380127,
      "rewards/margins": 0.7711628675460815,
      "rewards/rejected": -3.000037670135498,
      "step": 460
    },
    {
      "epoch": 0.7481098288897732,
      "grad_norm": 15.466730145904025,
      "learning_rate": 9.042988532644249e-08,
      "logits/chosen": 3.792773485183716,
      "logits/rejected": 4.424340724945068,
      "logps/chosen": -395.2272033691406,
      "logps/rejected": -461.55352783203125,
      "loss": 0.5425,
      "rewards/accuracies": 0.690625011920929,
      "rewards/chosen": -2.369027614593506,
      "rewards/margins": 0.6728731393814087,
      "rewards/rejected": -3.041900157928467,
      "step": 470
    },
    {
      "epoch": 0.7640270592916832,
      "grad_norm": 14.83331875533172,
      "learning_rate": 7.998107906142839e-08,
      "logits/chosen": 3.4053382873535156,
      "logits/rejected": 3.8911843299865723,
      "logps/chosen": -388.53155517578125,
      "logps/rejected": -455.40185546875,
      "loss": 0.5425,
      "rewards/accuracies": 0.7281249761581421,
      "rewards/chosen": -2.44258189201355,
      "rewards/margins": 0.6656660437583923,
      "rewards/rejected": -3.108247756958008,
      "step": 480
    },
    {
      "epoch": 0.7799442896935933,
      "grad_norm": 15.620806793742632,
      "learning_rate": 7.005779153764682e-08,
      "logits/chosen": 3.4802298545837402,
      "logits/rejected": 3.920941114425659,
      "logps/chosen": -378.8732604980469,
      "logps/rejected": -466.740478515625,
      "loss": 0.5309,
      "rewards/accuracies": 0.731249988079071,
      "rewards/chosen": -2.4023540019989014,
      "rewards/margins": 0.7670234441757202,
      "rewards/rejected": -3.169377565383911,
      "step": 490
    },
    {
      "epoch": 0.7958615200955034,
      "grad_norm": 17.188173872991122,
      "learning_rate": 6.069069506815325e-08,
      "logits/chosen": 3.2549641132354736,
      "logits/rejected": 4.011484146118164,
      "logps/chosen": -376.5075988769531,
      "logps/rejected": -464.5885314941406,
      "loss": 0.5443,
      "rewards/accuracies": 0.731249988079071,
      "rewards/chosen": -2.2434165477752686,
      "rewards/margins": 0.8864636421203613,
      "rewards/rejected": -3.129879951477051,
      "step": 500
    },
    {
      "epoch": 0.8117787504974134,
      "grad_norm": 15.502718885907624,
      "learning_rate": 5.190874281132851e-08,
      "logits/chosen": 3.2985496520996094,
      "logits/rejected": 3.910979747772217,
      "logps/chosen": -390.97381591796875,
      "logps/rejected": -462.67816162109375,
      "loss": 0.5418,
      "rewards/accuracies": 0.7093750238418579,
      "rewards/chosen": -2.4295220375061035,
      "rewards/margins": 0.7420149445533752,
      "rewards/rejected": -3.171536922454834,
      "step": 510
    },
    {
      "epoch": 0.8276959808993235,
      "grad_norm": 14.242739712121452,
      "learning_rate": 4.373907927832513e-08,
      "logits/chosen": 3.5766899585723877,
      "logits/rejected": 4.207441806793213,
      "logps/chosen": -382.2040710449219,
      "logps/rejected": -460.5889587402344,
      "loss": 0.5461,
      "rewards/accuracies": 0.6968749761581421,
      "rewards/chosen": -2.4501075744628906,
      "rewards/margins": 0.7706823348999023,
      "rewards/rejected": -3.220790147781372,
      "step": 520
    },
    {
      "epoch": 0.8436132113012336,
      "grad_norm": 12.998296781925998,
      "learning_rate": 3.620695643093924e-08,
      "logits/chosen": 3.5845916271209717,
      "logits/rejected": 4.0032148361206055,
      "logps/chosen": -372.14898681640625,
      "logps/rejected": -455.68212890625,
      "loss": 0.5414,
      "rewards/accuracies": 0.734375,
      "rewards/chosen": -2.2561497688293457,
      "rewards/margins": 0.7765234112739563,
      "rewards/rejected": -3.032672643661499,
      "step": 530
    },
    {
      "epoch": 0.8595304417031436,
      "grad_norm": 17.292649632838344,
      "learning_rate": 2.9335655629243645e-08,
      "logits/chosen": 3.523092269897461,
      "logits/rejected": 3.9805259704589844,
      "logps/chosen": -388.76800537109375,
      "logps/rejected": -462.15625,
      "loss": 0.5454,
      "rewards/accuracies": 0.721875011920929,
      "rewards/chosen": -2.4221224784851074,
      "rewards/margins": 0.6886903643608093,
      "rewards/rejected": -3.1108126640319824,
      "step": 540
    },
    {
      "epoch": 0.8754476721050537,
      "grad_norm": 13.631324448665103,
      "learning_rate": 2.31464156702382e-08,
      "logits/chosen": 3.492096424102783,
      "logits/rejected": 4.266209125518799,
      "logps/chosen": -393.3036804199219,
      "logps/rejected": -445.7330627441406,
      "loss": 0.5519,
      "rewards/accuracies": 0.671875,
      "rewards/chosen": -2.4094600677490234,
      "rewards/margins": 0.6608883738517761,
      "rewards/rejected": -3.0703485012054443,
      "step": 550
    },
    {
      "epoch": 0.8913649025069638,
      "grad_norm": 15.338497216975412,
      "learning_rate": 1.7658367139945228e-08,
      "logits/chosen": 3.498994827270508,
      "logits/rejected": 3.9349846839904785,
      "logps/chosen": -387.2663269042969,
      "logps/rejected": -449.8990173339844,
      "loss": 0.5364,
      "rewards/accuracies": 0.659375011920929,
      "rewards/chosen": -2.4634361267089844,
      "rewards/margins": 0.5876127481460571,
      "rewards/rejected": -3.051048755645752,
      "step": 560
    },
    {
      "epoch": 0.9072821329088738,
      "grad_norm": 13.324839693428412,
      "learning_rate": 1.2888473281864597e-08,
      "logits/chosen": 3.275439500808716,
      "logits/rejected": 3.773510694503784,
      "logps/chosen": -398.3496398925781,
      "logps/rejected": -472.8216247558594,
      "loss": 0.5281,
      "rewards/accuracies": 0.6968749761581421,
      "rewards/chosen": -2.447859287261963,
      "rewards/margins": 0.6992862820625305,
      "rewards/rejected": -3.1471457481384277,
      "step": 570
    },
    {
      "epoch": 0.9231993633107839,
      "grad_norm": 16.191483693504896,
      "learning_rate": 8.851477564560061e-09,
      "logits/chosen": 3.820493221282959,
      "logits/rejected": 4.351908206939697,
      "logps/chosen": -389.7720947265625,
      "logps/rejected": -459.3812561035156,
      "loss": 0.55,
      "rewards/accuracies": 0.6968749761581421,
      "rewards/chosen": -2.404622793197632,
      "rewards/margins": 0.6956027746200562,
      "rewards/rejected": -3.1002252101898193,
      "step": 580
    },
    {
      "epoch": 0.939116593712694,
      "grad_norm": 12.583560577719238,
      "learning_rate": 5.559858110443016e-09,
      "logits/chosen": 3.1959145069122314,
      "logits/rejected": 3.7669143676757812,
      "logps/chosen": -405.40618896484375,
      "logps/rejected": -479.1631774902344,
      "loss": 0.5251,
      "rewards/accuracies": 0.7124999761581421,
      "rewards/chosen": -2.4700753688812256,
      "rewards/margins": 0.7442026138305664,
      "rewards/rejected": -3.214277982711792,
      "step": 590
    },
    {
      "epoch": 0.955033824114604,
      "grad_norm": 15.78612631007661,
      "learning_rate": 3.023789126611137e-09,
      "logits/chosen": 3.5097384452819824,
      "logits/rejected": 4.086105823516846,
      "logps/chosen": -395.22027587890625,
      "logps/rejected": -475.81787109375,
      "loss": 0.5376,
      "rewards/accuracies": 0.706250011920929,
      "rewards/chosen": -2.4702706336975098,
      "rewards/margins": 0.8460418581962585,
      "rewards/rejected": -3.316312313079834,
      "step": 600
    },
    {
      "epoch": 0.9709510545165141,
      "grad_norm": 15.913083563945913,
      "learning_rate": 1.2511094569571668e-09,
      "logits/chosen": 3.64544939994812,
      "logits/rejected": 4.29144811630249,
      "logps/chosen": -404.07501220703125,
      "logps/rejected": -467.01153564453125,
      "loss": 0.5567,
      "rewards/accuracies": 0.659375011920929,
      "rewards/chosen": -2.5605225563049316,
      "rewards/margins": 0.6661086082458496,
      "rewards/rejected": -3.2266311645507812,
      "step": 610
    },
    {
      "epoch": 0.9868682849184242,
      "grad_norm": 15.29692820991952,
      "learning_rate": 2.4729835275189016e-10,
      "logits/chosen": 3.532379150390625,
      "logits/rejected": 3.938063859939575,
      "logps/chosen": -392.1960754394531,
      "logps/rejected": -481.100830078125,
      "loss": 0.5331,
      "rewards/accuracies": 0.7093750238418579,
      "rewards/chosen": -2.4987053871154785,
      "rewards/margins": 0.7970455288887024,
      "rewards/rejected": -3.295750379562378,
      "step": 620
    },
    {
      "epoch": 0.9996020692399522,
      "step": 628,
      "total_flos": 0.0,
      "train_loss": 0.5721431234080321,
      "train_runtime": 15529.5131,
      "train_samples_per_second": 10.354,
      "train_steps_per_second": 0.04
    }
  ],
  "logging_steps": 10,
  "max_steps": 628,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 100,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 0.0,
  "train_batch_size": 8,
  "trial_name": null,
  "trial_params": null
}