{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 936, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.003205128205128205, "grad_norm": 38.020048422719306, "learning_rate": 5.3191489361702125e-09, "logits/chosen": -1.484375, "logits/rejected": -1.4140625, "logps/chosen": -172.0, "logps/rejected": -160.0, "loss": 0.6914, "rewards/accuracies": 0.0, "rewards/chosen": 0.0, "rewards/margins": 0.0, "rewards/rejected": 0.0, "step": 1 }, { "epoch": 0.03205128205128205, "grad_norm": 34.89707580881474, "learning_rate": 5.3191489361702123e-08, "logits/chosen": -1.484375, "logits/rejected": -1.4375, "logps/chosen": -140.0, "logps/rejected": -181.0, "loss": 0.691, "rewards/accuracies": 0.25, "rewards/chosen": -0.00592041015625, "rewards/margins": -0.003997802734375, "rewards/rejected": -0.0019073486328125, "step": 10 }, { "epoch": 0.0641025641025641, "grad_norm": 35.22039582592668, "learning_rate": 1.0638297872340425e-07, "logits/chosen": -1.4375, "logits/rejected": -1.40625, "logps/chosen": -116.0, "logps/rejected": -171.0, "loss": 0.6797, "rewards/accuracies": 0.48750001192092896, "rewards/chosen": 0.0260009765625, "rewards/margins": 0.0264892578125, "rewards/rejected": -0.0004749298095703125, "step": 20 }, { "epoch": 0.09615384615384616, "grad_norm": 40.58215068588761, "learning_rate": 1.5957446808510638e-07, "logits/chosen": -1.46875, "logits/rejected": -1.4140625, "logps/chosen": -122.5, "logps/rejected": -180.0, "loss": 0.6562, "rewards/accuracies": 0.7250000238418579, "rewards/chosen": 0.11474609375, "rewards/margins": 0.08203125, "rewards/rejected": 0.03271484375, "step": 30 }, { "epoch": 0.1282051282051282, "grad_norm": 35.55857804315714, "learning_rate": 2.127659574468085e-07, "logits/chosen": -1.53125, "logits/rejected": -1.4609375, "logps/chosen": -127.5, "logps/rejected": -180.0, "loss": 0.6092, "rewards/accuracies": 0.8125, "rewards/chosen": 0.259765625, "rewards/margins": 0.21484375, "rewards/rejected": 0.046142578125, "step": 40 }, { "epoch": 0.16025641025641027, "grad_norm": 23.579694487139133, "learning_rate": 2.659574468085106e-07, "logits/chosen": -1.515625, "logits/rejected": -1.453125, "logps/chosen": -116.5, "logps/rejected": -174.0, "loss": 0.5579, "rewards/accuracies": 0.762499988079071, "rewards/chosen": 0.26953125, "rewards/margins": 0.365234375, "rewards/rejected": -0.095703125, "step": 50 }, { "epoch": 0.19230769230769232, "grad_norm": 23.19132387624345, "learning_rate": 3.1914893617021275e-07, "logits/chosen": -1.453125, "logits/rejected": -1.40625, "logps/chosen": -99.5, "logps/rejected": -155.0, "loss": 0.5053, "rewards/accuracies": 0.8500000238418579, "rewards/chosen": 0.298828125, "rewards/margins": 0.6015625, "rewards/rejected": -0.30078125, "step": 60 }, { "epoch": 0.22435897435897437, "grad_norm": 21.818865607569602, "learning_rate": 3.7234042553191484e-07, "logits/chosen": -1.5625, "logits/rejected": -1.484375, "logps/chosen": -100.5, "logps/rejected": -162.0, "loss": 0.3894, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 0.3984375, "rewards/margins": 1.0703125, "rewards/rejected": -0.66796875, "step": 70 }, { "epoch": 0.2564102564102564, "grad_norm": 26.315987904129013, "learning_rate": 4.25531914893617e-07, "logits/chosen": -1.5625, "logits/rejected": -1.4921875, "logps/chosen": -117.5, "logps/rejected": -170.0, "loss": 0.3909, "rewards/accuracies": 0.8500000238418579, "rewards/chosen": 0.026611328125, "rewards/margins": 1.1875, "rewards/rejected": -1.1640625, "step": 80 }, { "epoch": 0.28846153846153844, "grad_norm": 28.51531224072289, "learning_rate": 4.787234042553192e-07, "logits/chosen": -1.5625, "logits/rejected": -1.515625, "logps/chosen": -114.5, "logps/rejected": -176.0, "loss": 0.387, "rewards/accuracies": 0.800000011920929, "rewards/chosen": -0.1240234375, "rewards/margins": 1.375, "rewards/rejected": -1.5, "step": 90 }, { "epoch": 0.32051282051282054, "grad_norm": 17.39940529070616, "learning_rate": 4.96437054631829e-07, "logits/chosen": -1.59375, "logits/rejected": -1.5234375, "logps/chosen": -105.5, "logps/rejected": -169.0, "loss": 0.324, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": 0.390625, "rewards/margins": 1.6640625, "rewards/rejected": -1.2734375, "step": 100 }, { "epoch": 0.3525641025641026, "grad_norm": 28.443023292937838, "learning_rate": 4.904988123515439e-07, "logits/chosen": -1.6015625, "logits/rejected": -1.5546875, "logps/chosen": -118.0, "logps/rejected": -199.0, "loss": 0.3077, "rewards/accuracies": 0.875, "rewards/chosen": 0.052490234375, "rewards/margins": 1.75, "rewards/rejected": -1.703125, "step": 110 }, { "epoch": 0.38461538461538464, "grad_norm": 19.86248937078093, "learning_rate": 4.845605700712589e-07, "logits/chosen": -1.640625, "logits/rejected": -1.515625, "logps/chosen": -118.0, "logps/rejected": -182.0, "loss": 0.3212, "rewards/accuracies": 0.8374999761581421, "rewards/chosen": -0.41015625, "rewards/margins": 2.0625, "rewards/rejected": -2.46875, "step": 120 }, { "epoch": 0.4166666666666667, "grad_norm": 15.34168051684897, "learning_rate": 4.786223277909738e-07, "logits/chosen": -1.625, "logits/rejected": -1.515625, "logps/chosen": -127.0, "logps/rejected": -196.0, "loss": 0.3436, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.73046875, "rewards/margins": 2.25, "rewards/rejected": -2.984375, "step": 130 }, { "epoch": 0.44871794871794873, "grad_norm": 31.34040320177804, "learning_rate": 4.7268408551068883e-07, "logits/chosen": -1.609375, "logits/rejected": -1.515625, "logps/chosen": -138.0, "logps/rejected": -193.0, "loss": 0.335, "rewards/accuracies": 0.862500011920929, "rewards/chosen": -0.453125, "rewards/margins": 1.7578125, "rewards/rejected": -2.21875, "step": 140 }, { "epoch": 0.4807692307692308, "grad_norm": 39.09253912133659, "learning_rate": 4.667458432304038e-07, "logits/chosen": -1.640625, "logits/rejected": -1.6328125, "logps/chosen": -128.0, "logps/rejected": -249.0, "loss": 0.2853, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.08056640625, "rewards/margins": 2.53125, "rewards/rejected": -2.609375, "step": 150 }, { "epoch": 0.5128205128205128, "grad_norm": 17.194231471019577, "learning_rate": 4.6080760095011875e-07, "logits/chosen": -1.6015625, "logits/rejected": -1.5859375, "logps/chosen": -107.0, "logps/rejected": -191.0, "loss": 0.2666, "rewards/accuracies": 0.925000011920929, "rewards/chosen": -0.310546875, "rewards/margins": 2.71875, "rewards/rejected": -3.015625, "step": 160 }, { "epoch": 0.5448717948717948, "grad_norm": 18.722209877304238, "learning_rate": 4.548693586698337e-07, "logits/chosen": -1.578125, "logits/rejected": -1.59375, "logps/chosen": -115.0, "logps/rejected": -175.0, "loss": 0.2696, "rewards/accuracies": 0.875, "rewards/chosen": -0.150390625, "rewards/margins": 2.234375, "rewards/rejected": -2.390625, "step": 170 }, { "epoch": 0.5769230769230769, "grad_norm": 21.42273705026191, "learning_rate": 4.4893111638954866e-07, "logits/chosen": -1.5859375, "logits/rejected": -1.6171875, "logps/chosen": -121.5, "logps/rejected": -203.0, "loss": 0.2899, "rewards/accuracies": 0.9125000238418579, "rewards/chosen": -0.248046875, "rewards/margins": 2.390625, "rewards/rejected": -2.640625, "step": 180 }, { "epoch": 0.6089743589743589, "grad_norm": 35.105792045016955, "learning_rate": 4.429928741092636e-07, "logits/chosen": -1.625, "logits/rejected": -1.546875, "logps/chosen": -126.5, "logps/rejected": -199.0, "loss": 0.2834, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.38671875, "rewards/margins": 2.609375, "rewards/rejected": -3.0, "step": 190 }, { "epoch": 0.6410256410256411, "grad_norm": 25.037051509827627, "learning_rate": 4.3705463182897863e-07, "logits/chosen": -1.59375, "logits/rejected": -1.5078125, "logps/chosen": -116.5, "logps/rejected": -171.0, "loss": 0.3022, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.59375, "rewards/margins": 2.15625, "rewards/rejected": -2.75, "step": 200 }, { "epoch": 0.6730769230769231, "grad_norm": 25.3426239738337, "learning_rate": 4.311163895486936e-07, "logits/chosen": -1.5859375, "logits/rejected": -1.5546875, "logps/chosen": -118.5, "logps/rejected": -185.0, "loss": 0.3209, "rewards/accuracies": 0.925000011920929, "rewards/chosen": -1.21875, "rewards/margins": 2.1875, "rewards/rejected": -3.40625, "step": 210 }, { "epoch": 0.7051282051282052, "grad_norm": 41.23383631897795, "learning_rate": 4.251781472684085e-07, "logits/chosen": -1.6484375, "logits/rejected": -1.5546875, "logps/chosen": -148.0, "logps/rejected": -200.0, "loss": 0.2921, "rewards/accuracies": 0.8500000238418579, "rewards/chosen": -0.609375, "rewards/margins": 2.3125, "rewards/rejected": -2.921875, "step": 220 }, { "epoch": 0.7371794871794872, "grad_norm": 22.289875990424193, "learning_rate": 4.192399049881235e-07, "logits/chosen": -1.609375, "logits/rejected": -1.5859375, "logps/chosen": -127.0, "logps/rejected": -199.0, "loss": 0.2805, "rewards/accuracies": 0.925000011920929, "rewards/chosen": -0.318359375, "rewards/margins": 2.578125, "rewards/rejected": -2.90625, "step": 230 }, { "epoch": 0.7692307692307693, "grad_norm": 29.25683917677673, "learning_rate": 4.1330166270783846e-07, "logits/chosen": -1.59375, "logits/rejected": -1.5390625, "logps/chosen": -108.5, "logps/rejected": -180.0, "loss": 0.2541, "rewards/accuracies": 0.9125000238418579, "rewards/chosen": -0.267578125, "rewards/margins": 2.765625, "rewards/rejected": -3.03125, "step": 240 }, { "epoch": 0.8012820512820513, "grad_norm": 28.339641539576366, "learning_rate": 4.0736342042755347e-07, "logits/chosen": -1.6484375, "logits/rejected": -1.5625, "logps/chosen": -120.0, "logps/rejected": -215.0, "loss": 0.257, "rewards/accuracies": 0.887499988079071, "rewards/chosen": -0.50390625, "rewards/margins": 3.140625, "rewards/rejected": -3.640625, "step": 250 }, { "epoch": 0.8333333333333334, "grad_norm": 28.458003688730955, "learning_rate": 4.0142517814726837e-07, "logits/chosen": -1.6328125, "logits/rejected": -1.5703125, "logps/chosen": -124.0, "logps/rejected": -195.0, "loss": 0.2514, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.3203125, "rewards/margins": 2.6875, "rewards/rejected": -3.015625, "step": 260 }, { "epoch": 0.8653846153846154, "grad_norm": 19.712649681094366, "learning_rate": 3.9548693586698333e-07, "logits/chosen": -1.640625, "logits/rejected": -1.546875, "logps/chosen": -121.0, "logps/rejected": -177.0, "loss": 0.2635, "rewards/accuracies": 0.8500000238418579, "rewards/chosen": -0.11083984375, "rewards/margins": 2.453125, "rewards/rejected": -2.5625, "step": 270 }, { "epoch": 0.8974358974358975, "grad_norm": 46.09415755147797, "learning_rate": 3.8954869358669834e-07, "logits/chosen": -1.6640625, "logits/rejected": -1.546875, "logps/chosen": -118.5, "logps/rejected": -197.0, "loss": 0.2671, "rewards/accuracies": 0.875, "rewards/chosen": 0.0693359375, "rewards/margins": 2.78125, "rewards/rejected": -2.703125, "step": 280 }, { "epoch": 0.9294871794871795, "grad_norm": 25.186336345995475, "learning_rate": 3.836104513064133e-07, "logits/chosen": -1.5703125, "logits/rejected": -1.5390625, "logps/chosen": -129.0, "logps/rejected": -203.0, "loss": 0.2957, "rewards/accuracies": 0.824999988079071, "rewards/chosen": -1.03125, "rewards/margins": 2.21875, "rewards/rejected": -3.25, "step": 290 }, { "epoch": 0.9615384615384616, "grad_norm": 25.061643380626624, "learning_rate": 3.7767220902612825e-07, "logits/chosen": -1.609375, "logits/rejected": -1.578125, "logps/chosen": -134.0, "logps/rejected": -219.0, "loss": 0.2345, "rewards/accuracies": 0.887499988079071, "rewards/chosen": -0.90234375, "rewards/margins": 2.671875, "rewards/rejected": -3.578125, "step": 300 }, { "epoch": 0.9935897435897436, "grad_norm": 44.17174735441787, "learning_rate": 3.717339667458432e-07, "logits/chosen": -1.640625, "logits/rejected": -1.5703125, "logps/chosen": -119.0, "logps/rejected": -192.0, "loss": 0.2734, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.2392578125, "rewards/margins": 2.734375, "rewards/rejected": -2.96875, "step": 310 }, { "epoch": 1.0, "eval_logits/chosen": -1.5546875, "eval_logits/rejected": -1.5234375, "eval_logps/chosen": -170.0, "eval_logps/rejected": -184.0, "eval_loss": 0.6114453077316284, "eval_rewards/accuracies": 0.5892857313156128, "eval_rewards/chosen": -1.53125, "eval_rewards/margins": 0.8984375, "eval_rewards/rejected": -2.4375, "eval_runtime": 14.9255, "eval_samples_per_second": 13.4, "eval_steps_per_second": 0.469, "step": 312 }, { "epoch": 1.0256410256410255, "grad_norm": 16.449029965031666, "learning_rate": 3.6579572446555817e-07, "logits/chosen": -1.59375, "logits/rejected": -1.578125, "logps/chosen": -117.0, "logps/rejected": -195.0, "loss": 0.1531, "rewards/accuracies": 0.925000011920929, "rewards/chosen": -0.34765625, "rewards/margins": 3.1875, "rewards/rejected": -3.546875, "step": 320 }, { "epoch": 1.0576923076923077, "grad_norm": 6.562166239087373, "learning_rate": 3.598574821852731e-07, "logits/chosen": -1.59375, "logits/rejected": -1.5703125, "logps/chosen": -111.0, "logps/rejected": -187.0, "loss": 0.1452, "rewards/accuracies": 0.8999999761581421, "rewards/chosen": -0.1923828125, "rewards/margins": 3.40625, "rewards/rejected": -3.609375, "step": 330 }, { "epoch": 1.0897435897435896, "grad_norm": 25.028970432454248, "learning_rate": 3.5391923990498813e-07, "logits/chosen": -1.6484375, "logits/rejected": -1.609375, "logps/chosen": -134.0, "logps/rejected": -219.0, "loss": 0.1345, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -0.69921875, "rewards/margins": 3.609375, "rewards/rejected": -4.3125, "step": 340 }, { "epoch": 1.1217948717948718, "grad_norm": 15.596444813277682, "learning_rate": 3.479809976247031e-07, "logits/chosen": -1.65625, "logits/rejected": -1.6015625, "logps/chosen": -130.0, "logps/rejected": -207.0, "loss": 0.1092, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -0.00677490234375, "rewards/margins": 3.828125, "rewards/rejected": -3.828125, "step": 350 }, { "epoch": 1.1538461538461537, "grad_norm": 22.37742199079234, "learning_rate": 3.42042755344418e-07, "logits/chosen": -1.5859375, "logits/rejected": -1.5703125, "logps/chosen": -136.0, "logps/rejected": -242.0, "loss": 0.0918, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -0.640625, "rewards/margins": 4.4375, "rewards/rejected": -5.09375, "step": 360 }, { "epoch": 1.185897435897436, "grad_norm": 27.25128799333949, "learning_rate": 3.36104513064133e-07, "logits/chosen": -1.609375, "logits/rejected": -1.59375, "logps/chosen": -154.0, "logps/rejected": -221.0, "loss": 0.1215, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -0.333984375, "rewards/margins": 4.46875, "rewards/rejected": -4.78125, "step": 370 }, { "epoch": 1.217948717948718, "grad_norm": 16.802650370434502, "learning_rate": 3.3016627078384796e-07, "logits/chosen": -1.625, "logits/rejected": -1.515625, "logps/chosen": -121.0, "logps/rejected": -184.0, "loss": 0.1051, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -0.75390625, "rewards/margins": 3.75, "rewards/rejected": -4.5, "step": 380 }, { "epoch": 1.25, "grad_norm": 18.14599606798184, "learning_rate": 3.2422802850356297e-07, "logits/chosen": -1.65625, "logits/rejected": -1.5625, "logps/chosen": -118.0, "logps/rejected": -203.0, "loss": 0.1691, "rewards/accuracies": 0.925000011920929, "rewards/chosen": -0.4609375, "rewards/margins": 4.4375, "rewards/rejected": -4.875, "step": 390 }, { "epoch": 1.282051282051282, "grad_norm": 14.341477798166297, "learning_rate": 3.182897862232779e-07, "logits/chosen": -1.609375, "logits/rejected": -1.5546875, "logps/chosen": -112.0, "logps/rejected": -202.0, "loss": 0.1084, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -0.2578125, "rewards/margins": 4.65625, "rewards/rejected": -4.90625, "step": 400 }, { "epoch": 1.314102564102564, "grad_norm": 22.538352683910524, "learning_rate": 3.1235154394299283e-07, "logits/chosen": -1.671875, "logits/rejected": -1.625, "logps/chosen": -168.0, "logps/rejected": -270.0, "loss": 0.136, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.1796875, "rewards/margins": 4.71875, "rewards/rejected": -5.90625, "step": 410 }, { "epoch": 1.3461538461538463, "grad_norm": 7.69063972210861, "learning_rate": 3.0641330166270784e-07, "logits/chosen": -1.609375, "logits/rejected": -1.59375, "logps/chosen": -128.0, "logps/rejected": -232.0, "loss": 0.1123, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.0859375, "rewards/margins": 4.59375, "rewards/rejected": -5.6875, "step": 420 }, { "epoch": 1.3782051282051282, "grad_norm": 22.81215314331503, "learning_rate": 3.004750593824228e-07, "logits/chosen": -1.6328125, "logits/rejected": -1.5390625, "logps/chosen": -122.5, "logps/rejected": -209.0, "loss": 0.0971, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -0.33984375, "rewards/margins": 4.65625, "rewards/rejected": -5.0, "step": 430 }, { "epoch": 1.4102564102564101, "grad_norm": 16.88060293661355, "learning_rate": 2.9453681710213776e-07, "logits/chosen": -1.5859375, "logits/rejected": -1.4765625, "logps/chosen": -123.0, "logps/rejected": -205.0, "loss": 0.11, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -0.4296875, "rewards/margins": 4.5625, "rewards/rejected": -5.0, "step": 440 }, { "epoch": 1.4423076923076923, "grad_norm": 14.494261083829612, "learning_rate": 2.885985748218527e-07, "logits/chosen": -1.546875, "logits/rejected": -1.5546875, "logps/chosen": -118.5, "logps/rejected": -221.0, "loss": 0.0896, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.0859375, "rewards/margins": 4.59375, "rewards/rejected": -5.6875, "step": 450 }, { "epoch": 1.4743589743589745, "grad_norm": 16.655718879118293, "learning_rate": 2.8266033254156767e-07, "logits/chosen": -1.59375, "logits/rejected": -1.5859375, "logps/chosen": -142.0, "logps/rejected": -256.0, "loss": 0.1071, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.25, "rewards/margins": 4.84375, "rewards/rejected": -6.09375, "step": 460 }, { "epoch": 1.5064102564102564, "grad_norm": 20.51874312804957, "learning_rate": 2.7672209026128263e-07, "logits/chosen": -1.5625, "logits/rejected": -1.53125, "logps/chosen": -132.0, "logps/rejected": -231.0, "loss": 0.1012, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.015625, "rewards/margins": 4.65625, "rewards/rejected": -5.6875, "step": 470 }, { "epoch": 1.5384615384615383, "grad_norm": 28.03019085733964, "learning_rate": 2.7078384798099764e-07, "logits/chosen": -1.4453125, "logits/rejected": -1.4609375, "logps/chosen": -104.5, "logps/rejected": -203.0, "loss": 0.1074, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.6796875, "rewards/margins": 4.59375, "rewards/rejected": -6.3125, "step": 480 }, { "epoch": 1.5705128205128205, "grad_norm": 14.674518317046845, "learning_rate": 2.648456057007126e-07, "logits/chosen": -1.5625, "logits/rejected": -1.5390625, "logps/chosen": -139.0, "logps/rejected": -233.0, "loss": 0.1003, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.3125, "rewards/margins": 4.9375, "rewards/rejected": -6.28125, "step": 490 }, { "epoch": 1.6025641025641026, "grad_norm": 18.731133870495857, "learning_rate": 2.589073634204275e-07, "logits/chosen": -1.5703125, "logits/rejected": -1.5625, "logps/chosen": -144.0, "logps/rejected": -251.0, "loss": 0.1228, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -1.109375, "rewards/margins": 5.25, "rewards/rejected": -6.375, "step": 500 }, { "epoch": 1.6346153846153846, "grad_norm": 16.81631331394118, "learning_rate": 2.529691211401425e-07, "logits/chosen": -1.5625, "logits/rejected": -1.5, "logps/chosen": -131.0, "logps/rejected": -230.0, "loss": 0.0819, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -0.8359375, "rewards/margins": 5.375, "rewards/rejected": -6.21875, "step": 510 }, { "epoch": 1.6666666666666665, "grad_norm": 15.111001669085608, "learning_rate": 2.4703087885985747e-07, "logits/chosen": -1.546875, "logits/rejected": -1.4921875, "logps/chosen": -114.0, "logps/rejected": -202.0, "loss": 0.1091, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -0.87109375, "rewards/margins": 4.78125, "rewards/rejected": -5.625, "step": 520 }, { "epoch": 1.6987179487179487, "grad_norm": 26.75975769912583, "learning_rate": 2.410926365795724e-07, "logits/chosen": -1.5625, "logits/rejected": -1.515625, "logps/chosen": -140.0, "logps/rejected": -221.0, "loss": 0.1139, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.0625, "rewards/margins": 4.8125, "rewards/rejected": -5.875, "step": 530 }, { "epoch": 1.7307692307692308, "grad_norm": 13.759559493772942, "learning_rate": 2.351543942992874e-07, "logits/chosen": -1.578125, "logits/rejected": -1.5, "logps/chosen": -143.0, "logps/rejected": -229.0, "loss": 0.0978, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -0.9765625, "rewards/margins": 5.09375, "rewards/rejected": -6.0625, "step": 540 }, { "epoch": 1.7628205128205128, "grad_norm": 18.98900148314797, "learning_rate": 2.2921615201900234e-07, "logits/chosen": -1.4921875, "logits/rejected": -1.5, "logps/chosen": -114.0, "logps/rejected": -232.0, "loss": 0.1103, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -0.6953125, "rewards/margins": 5.03125, "rewards/rejected": -5.71875, "step": 550 }, { "epoch": 1.7948717948717947, "grad_norm": 7.478935072745169, "learning_rate": 2.2327790973871732e-07, "logits/chosen": -1.515625, "logits/rejected": -1.4375, "logps/chosen": -112.0, "logps/rejected": -177.0, "loss": 0.1139, "rewards/accuracies": 0.9375, "rewards/chosen": -0.6640625, "rewards/margins": 4.5625, "rewards/rejected": -5.21875, "step": 560 }, { "epoch": 1.8269230769230769, "grad_norm": 13.728806584543696, "learning_rate": 2.173396674584323e-07, "logits/chosen": -1.546875, "logits/rejected": -1.5546875, "logps/chosen": -130.0, "logps/rejected": -239.0, "loss": 0.1355, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -0.466796875, "rewards/margins": 4.875, "rewards/rejected": -5.34375, "step": 570 }, { "epoch": 1.858974358974359, "grad_norm": 9.926742756220793, "learning_rate": 2.1140142517814726e-07, "logits/chosen": -1.609375, "logits/rejected": -1.5078125, "logps/chosen": -158.0, "logps/rejected": -202.0, "loss": 0.1118, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -1.09375, "rewards/margins": 4.21875, "rewards/rejected": -5.3125, "step": 580 }, { "epoch": 1.891025641025641, "grad_norm": 12.431864869622972, "learning_rate": 2.0546318289786222e-07, "logits/chosen": -1.5546875, "logits/rejected": -1.53125, "logps/chosen": -122.0, "logps/rejected": -202.0, "loss": 0.1325, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -0.61328125, "rewards/margins": 4.65625, "rewards/rejected": -5.28125, "step": 590 }, { "epoch": 1.9230769230769231, "grad_norm": 12.291656293105886, "learning_rate": 1.9952494061757718e-07, "logits/chosen": -1.5703125, "logits/rejected": -1.5625, "logps/chosen": -123.0, "logps/rejected": -227.0, "loss": 0.1085, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -0.67578125, "rewards/margins": 4.8125, "rewards/rejected": -5.5, "step": 600 }, { "epoch": 1.9551282051282053, "grad_norm": 6.534649119157792, "learning_rate": 1.9358669833729216e-07, "logits/chosen": -1.5546875, "logits/rejected": -1.484375, "logps/chosen": -122.0, "logps/rejected": -202.0, "loss": 0.0935, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -0.56640625, "rewards/margins": 5.0, "rewards/rejected": -5.59375, "step": 610 }, { "epoch": 1.9871794871794872, "grad_norm": 10.91488724265466, "learning_rate": 1.876484560570071e-07, "logits/chosen": -1.5703125, "logits/rejected": -1.515625, "logps/chosen": -139.0, "logps/rejected": -236.0, "loss": 0.1071, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -1.1953125, "rewards/margins": 5.34375, "rewards/rejected": -6.53125, "step": 620 }, { "epoch": 2.0, "eval_logits/chosen": -1.5, "eval_logits/rejected": -1.46875, "eval_logps/chosen": -188.0, "eval_logps/rejected": -206.0, "eval_loss": 0.700976550579071, "eval_rewards/accuracies": 0.6428571343421936, "eval_rewards/chosen": -3.296875, "eval_rewards/margins": 1.3046875, "eval_rewards/rejected": -4.59375, "eval_runtime": 15.0135, "eval_samples_per_second": 13.321, "eval_steps_per_second": 0.466, "step": 624 }, { "epoch": 2.019230769230769, "grad_norm": 2.9708656343687063, "learning_rate": 1.8171021377672207e-07, "logits/chosen": -1.515625, "logits/rejected": -1.5, "logps/chosen": -150.0, "logps/rejected": -224.0, "loss": 0.071, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -0.98828125, "rewards/margins": 5.40625, "rewards/rejected": -6.40625, "step": 630 }, { "epoch": 2.051282051282051, "grad_norm": 3.1451139864714266, "learning_rate": 1.7577197149643706e-07, "logits/chosen": -1.5078125, "logits/rejected": -1.4609375, "logps/chosen": -121.5, "logps/rejected": -198.0, "loss": 0.0534, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -0.8203125, "rewards/margins": 4.9375, "rewards/rejected": -5.75, "step": 640 }, { "epoch": 2.0833333333333335, "grad_norm": 7.314670214652841, "learning_rate": 1.6983372921615202e-07, "logits/chosen": -1.5703125, "logits/rejected": -1.546875, "logps/chosen": -134.0, "logps/rejected": -247.0, "loss": 0.0546, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -0.73046875, "rewards/margins": 5.84375, "rewards/rejected": -6.5625, "step": 650 }, { "epoch": 2.1153846153846154, "grad_norm": 6.986967407965111, "learning_rate": 1.6389548693586697e-07, "logits/chosen": -1.5625, "logits/rejected": -1.53125, "logps/chosen": -123.5, "logps/rejected": -219.0, "loss": 0.0485, "rewards/accuracies": 1.0, "rewards/chosen": -0.9375, "rewards/margins": 5.5625, "rewards/rejected": -6.53125, "step": 660 }, { "epoch": 2.1474358974358974, "grad_norm": 3.152252590941172, "learning_rate": 1.5795724465558193e-07, "logits/chosen": -1.515625, "logits/rejected": -1.53125, "logps/chosen": -121.5, "logps/rejected": -264.0, "loss": 0.0391, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -0.97265625, "rewards/margins": 6.40625, "rewards/rejected": -7.375, "step": 670 }, { "epoch": 2.1794871794871793, "grad_norm": 5.616933414198294, "learning_rate": 1.520190023752969e-07, "logits/chosen": -1.5390625, "logits/rejected": -1.5078125, "logps/chosen": -136.0, "logps/rejected": -225.0, "loss": 0.0523, "rewards/accuracies": 0.949999988079071, "rewards/chosen": -1.265625, "rewards/margins": 5.4375, "rewards/rejected": -6.6875, "step": 680 }, { "epoch": 2.2115384615384617, "grad_norm": 4.462570602831943, "learning_rate": 1.4608076009501184e-07, "logits/chosen": -1.5390625, "logits/rejected": -1.53125, "logps/chosen": -141.0, "logps/rejected": -247.0, "loss": 0.0548, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -0.9921875, "rewards/margins": 5.90625, "rewards/rejected": -6.90625, "step": 690 }, { "epoch": 2.2435897435897436, "grad_norm": 6.786342065093859, "learning_rate": 1.4014251781472683e-07, "logits/chosen": -1.5546875, "logits/rejected": -1.546875, "logps/chosen": -115.0, "logps/rejected": -243.0, "loss": 0.0328, "rewards/accuracies": 1.0, "rewards/chosen": -0.7890625, "rewards/margins": 6.25, "rewards/rejected": -7.03125, "step": 700 }, { "epoch": 2.2756410256410255, "grad_norm": 10.783463457904992, "learning_rate": 1.342042755344418e-07, "logits/chosen": -1.5390625, "logits/rejected": -1.546875, "logps/chosen": -149.0, "logps/rejected": -266.0, "loss": 0.0556, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.203125, "rewards/margins": 6.125, "rewards/rejected": -7.3125, "step": 710 }, { "epoch": 2.3076923076923075, "grad_norm": 9.0679334215508, "learning_rate": 1.2826603325415677e-07, "logits/chosen": -1.53125, "logits/rejected": -1.4765625, "logps/chosen": -121.0, "logps/rejected": -217.0, "loss": 0.0438, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -0.80078125, "rewards/margins": 5.84375, "rewards/rejected": -6.65625, "step": 720 }, { "epoch": 2.33974358974359, "grad_norm": 2.164285473082669, "learning_rate": 1.2232779097387173e-07, "logits/chosen": -1.5078125, "logits/rejected": -1.5, "logps/chosen": -153.0, "logps/rejected": -237.0, "loss": 0.0414, "rewards/accuracies": 1.0, "rewards/chosen": -1.2421875, "rewards/margins": 6.15625, "rewards/rejected": -7.375, "step": 730 }, { "epoch": 2.371794871794872, "grad_norm": 8.601824238964523, "learning_rate": 1.163895486935867e-07, "logits/chosen": -1.4921875, "logits/rejected": -1.46875, "logps/chosen": -148.0, "logps/rejected": -238.0, "loss": 0.0358, "rewards/accuracies": 1.0, "rewards/chosen": -1.6953125, "rewards/margins": 5.9375, "rewards/rejected": -7.625, "step": 740 }, { "epoch": 2.4038461538461537, "grad_norm": 11.674320163023708, "learning_rate": 1.1045130641330165e-07, "logits/chosen": -1.5234375, "logits/rejected": -1.5390625, "logps/chosen": -133.0, "logps/rejected": -258.0, "loss": 0.0523, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.9453125, "rewards/margins": 6.4375, "rewards/rejected": -8.375, "step": 750 }, { "epoch": 2.435897435897436, "grad_norm": 4.6040374115250415, "learning_rate": 1.0451306413301662e-07, "logits/chosen": -1.515625, "logits/rejected": -1.5234375, "logps/chosen": -115.5, "logps/rejected": -258.0, "loss": 0.0425, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -0.984375, "rewards/margins": 7.40625, "rewards/rejected": -8.375, "step": 760 }, { "epoch": 2.467948717948718, "grad_norm": 6.059248925740205, "learning_rate": 9.857482185273158e-08, "logits/chosen": -1.53125, "logits/rejected": -1.515625, "logps/chosen": -137.0, "logps/rejected": -233.0, "loss": 0.037, "rewards/accuracies": 1.0, "rewards/chosen": -1.2265625, "rewards/margins": 5.90625, "rewards/rejected": -7.125, "step": 770 }, { "epoch": 2.5, "grad_norm": 5.569807408041334, "learning_rate": 9.263657957244655e-08, "logits/chosen": -1.546875, "logits/rejected": -1.53125, "logps/chosen": -136.0, "logps/rejected": -240.0, "loss": 0.0383, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.46875, "rewards/margins": 6.78125, "rewards/rejected": -8.25, "step": 780 }, { "epoch": 2.532051282051282, "grad_norm": 9.080566390829592, "learning_rate": 8.669833729216151e-08, "logits/chosen": -1.515625, "logits/rejected": -1.4375, "logps/chosen": -138.0, "logps/rejected": -233.0, "loss": 0.043, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -2.125, "rewards/margins": 6.1875, "rewards/rejected": -8.3125, "step": 790 }, { "epoch": 2.564102564102564, "grad_norm": 4.722752417362213, "learning_rate": 8.076009501187649e-08, "logits/chosen": -1.4609375, "logits/rejected": -1.515625, "logps/chosen": -129.0, "logps/rejected": -244.0, "loss": 0.043, "rewards/accuracies": 1.0, "rewards/chosen": -1.78125, "rewards/margins": 6.5625, "rewards/rejected": -8.375, "step": 800 }, { "epoch": 2.5961538461538463, "grad_norm": 8.320380078890436, "learning_rate": 7.482185273159145e-08, "logits/chosen": -1.5625, "logits/rejected": -1.484375, "logps/chosen": -152.0, "logps/rejected": -234.0, "loss": 0.0311, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.734375, "rewards/margins": 6.25, "rewards/rejected": -7.96875, "step": 810 }, { "epoch": 2.628205128205128, "grad_norm": 10.713886798327927, "learning_rate": 6.88836104513064e-08, "logits/chosen": -1.5078125, "logits/rejected": -1.4765625, "logps/chosen": -151.0, "logps/rejected": -252.0, "loss": 0.0368, "rewards/accuracies": 1.0, "rewards/chosen": -1.6875, "rewards/margins": 6.25, "rewards/rejected": -7.9375, "step": 820 }, { "epoch": 2.66025641025641, "grad_norm": 10.956365394123852, "learning_rate": 6.294536817102138e-08, "logits/chosen": -1.484375, "logits/rejected": -1.4375, "logps/chosen": -139.0, "logps/rejected": -224.0, "loss": 0.0434, "rewards/accuracies": 0.9750000238418579, "rewards/chosen": -1.7421875, "rewards/margins": 6.15625, "rewards/rejected": -7.875, "step": 830 }, { "epoch": 2.6923076923076925, "grad_norm": 5.167870965947855, "learning_rate": 5.700712589073634e-08, "logits/chosen": -1.5546875, "logits/rejected": -1.484375, "logps/chosen": -155.0, "logps/rejected": -286.0, "loss": 0.0367, "rewards/accuracies": 1.0, "rewards/chosen": -1.4140625, "rewards/margins": 7.15625, "rewards/rejected": -8.5625, "step": 840 }, { "epoch": 2.7243589743589745, "grad_norm": 10.357396384526298, "learning_rate": 5.10688836104513e-08, "logits/chosen": -1.4765625, "logits/rejected": -1.5, "logps/chosen": -144.0, "logps/rejected": -245.0, "loss": 0.0381, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.8125, "rewards/margins": 6.46875, "rewards/rejected": -8.3125, "step": 850 }, { "epoch": 2.7564102564102564, "grad_norm": 4.157699399133391, "learning_rate": 4.5130641330166267e-08, "logits/chosen": -1.46875, "logits/rejected": -1.421875, "logps/chosen": -126.0, "logps/rejected": -239.0, "loss": 0.0442, "rewards/accuracies": 0.9375, "rewards/chosen": -1.5859375, "rewards/margins": 6.25, "rewards/rejected": -7.8125, "step": 860 }, { "epoch": 2.7884615384615383, "grad_norm": 10.34195664309316, "learning_rate": 3.919239904988123e-08, "logits/chosen": -1.5234375, "logits/rejected": -1.5390625, "logps/chosen": -148.0, "logps/rejected": -268.0, "loss": 0.036, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -2.296875, "rewards/margins": 6.5, "rewards/rejected": -8.8125, "step": 870 }, { "epoch": 2.8205128205128203, "grad_norm": 13.885196758276543, "learning_rate": 3.32541567695962e-08, "logits/chosen": -1.53125, "logits/rejected": -1.4609375, "logps/chosen": -139.0, "logps/rejected": -237.0, "loss": 0.0386, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.7890625, "rewards/margins": 6.4375, "rewards/rejected": -8.25, "step": 880 }, { "epoch": 2.8525641025641026, "grad_norm": 3.298992157382736, "learning_rate": 2.7315914489311164e-08, "logits/chosen": -1.53125, "logits/rejected": -1.484375, "logps/chosen": -142.0, "logps/rejected": -264.0, "loss": 0.0326, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -2.0, "rewards/margins": 6.84375, "rewards/rejected": -8.875, "step": 890 }, { "epoch": 2.8846153846153846, "grad_norm": 9.55680354164709, "learning_rate": 2.1377672209026125e-08, "logits/chosen": -1.4609375, "logits/rejected": -1.4453125, "logps/chosen": -124.0, "logps/rejected": -247.0, "loss": 0.0386, "rewards/accuracies": 1.0, "rewards/chosen": -1.7734375, "rewards/margins": 7.0, "rewards/rejected": -8.8125, "step": 900 }, { "epoch": 2.9166666666666665, "grad_norm": 8.674386112287156, "learning_rate": 1.5439429928741092e-08, "logits/chosen": -1.5078125, "logits/rejected": -1.4765625, "logps/chosen": -118.0, "logps/rejected": -240.0, "loss": 0.0517, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.3984375, "rewards/margins": 7.0625, "rewards/rejected": -8.4375, "step": 910 }, { "epoch": 2.948717948717949, "grad_norm": 6.427610044394737, "learning_rate": 9.501187648456057e-09, "logits/chosen": -1.5, "logits/rejected": -1.4765625, "logps/chosen": -148.0, "logps/rejected": -256.0, "loss": 0.0481, "rewards/accuracies": 0.987500011920929, "rewards/chosen": -1.84375, "rewards/margins": 6.375, "rewards/rejected": -8.25, "step": 920 }, { "epoch": 2.980769230769231, "grad_norm": 20.775958318465445, "learning_rate": 3.562945368171021e-09, "logits/chosen": -1.53125, "logits/rejected": -1.5, "logps/chosen": -141.0, "logps/rejected": -232.0, "loss": 0.0498, "rewards/accuracies": 0.9624999761581421, "rewards/chosen": -1.640625, "rewards/margins": 6.0, "rewards/rejected": -7.625, "step": 930 }, { "epoch": 3.0, "eval_logits/chosen": -1.4921875, "eval_logits/rejected": -1.4609375, "eval_logps/chosen": -199.0, "eval_logps/rejected": -222.0, "eval_loss": 0.830273449420929, "eval_rewards/accuracies": 0.625, "eval_rewards/chosen": -4.4375, "eval_rewards/margins": 1.765625, "eval_rewards/rejected": -6.1875, "eval_runtime": 16.7003, "eval_samples_per_second": 11.976, "eval_steps_per_second": 0.419, "step": 936 }, { "epoch": 3.0, "step": 936, "total_flos": 0.0, "train_loss": 0.1719489719113733, "train_runtime": 6022.2267, "train_samples_per_second": 4.969, "train_steps_per_second": 0.155 } ], "logging_steps": 10, "max_steps": 936, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 8, "trial_name": null, "trial_params": null }