|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 3.0, |
|
"eval_steps": 500, |
|
"global_step": 936, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.003205128205128205, |
|
"grad_norm": 38.020048422719306, |
|
"learning_rate": 5.3191489361702125e-09, |
|
"logits/chosen": -1.484375, |
|
"logits/rejected": -1.4140625, |
|
"logps/chosen": -172.0, |
|
"logps/rejected": -160.0, |
|
"loss": 0.6914, |
|
"rewards/accuracies": 0.0, |
|
"rewards/chosen": 0.0, |
|
"rewards/margins": 0.0, |
|
"rewards/rejected": 0.0, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.03205128205128205, |
|
"grad_norm": 34.89707580881474, |
|
"learning_rate": 5.3191489361702123e-08, |
|
"logits/chosen": -1.484375, |
|
"logits/rejected": -1.4375, |
|
"logps/chosen": -140.0, |
|
"logps/rejected": -181.0, |
|
"loss": 0.691, |
|
"rewards/accuracies": 0.25, |
|
"rewards/chosen": -0.00592041015625, |
|
"rewards/margins": -0.003997802734375, |
|
"rewards/rejected": -0.0019073486328125, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0641025641025641, |
|
"grad_norm": 35.22039582592668, |
|
"learning_rate": 1.0638297872340425e-07, |
|
"logits/chosen": -1.4375, |
|
"logits/rejected": -1.40625, |
|
"logps/chosen": -116.0, |
|
"logps/rejected": -171.0, |
|
"loss": 0.6797, |
|
"rewards/accuracies": 0.48750001192092896, |
|
"rewards/chosen": 0.0260009765625, |
|
"rewards/margins": 0.0264892578125, |
|
"rewards/rejected": -0.0004749298095703125, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.09615384615384616, |
|
"grad_norm": 40.58215068588761, |
|
"learning_rate": 1.5957446808510638e-07, |
|
"logits/chosen": -1.46875, |
|
"logits/rejected": -1.4140625, |
|
"logps/chosen": -122.5, |
|
"logps/rejected": -180.0, |
|
"loss": 0.6562, |
|
"rewards/accuracies": 0.7250000238418579, |
|
"rewards/chosen": 0.11474609375, |
|
"rewards/margins": 0.08203125, |
|
"rewards/rejected": 0.03271484375, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.1282051282051282, |
|
"grad_norm": 35.55857804315714, |
|
"learning_rate": 2.127659574468085e-07, |
|
"logits/chosen": -1.53125, |
|
"logits/rejected": -1.4609375, |
|
"logps/chosen": -127.5, |
|
"logps/rejected": -180.0, |
|
"loss": 0.6092, |
|
"rewards/accuracies": 0.8125, |
|
"rewards/chosen": 0.259765625, |
|
"rewards/margins": 0.21484375, |
|
"rewards/rejected": 0.046142578125, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.16025641025641027, |
|
"grad_norm": 23.579694487139133, |
|
"learning_rate": 2.659574468085106e-07, |
|
"logits/chosen": -1.515625, |
|
"logits/rejected": -1.453125, |
|
"logps/chosen": -116.5, |
|
"logps/rejected": -174.0, |
|
"loss": 0.5579, |
|
"rewards/accuracies": 0.762499988079071, |
|
"rewards/chosen": 0.26953125, |
|
"rewards/margins": 0.365234375, |
|
"rewards/rejected": -0.095703125, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.19230769230769232, |
|
"grad_norm": 23.19132387624345, |
|
"learning_rate": 3.1914893617021275e-07, |
|
"logits/chosen": -1.453125, |
|
"logits/rejected": -1.40625, |
|
"logps/chosen": -99.5, |
|
"logps/rejected": -155.0, |
|
"loss": 0.5053, |
|
"rewards/accuracies": 0.8500000238418579, |
|
"rewards/chosen": 0.298828125, |
|
"rewards/margins": 0.6015625, |
|
"rewards/rejected": -0.30078125, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.22435897435897437, |
|
"grad_norm": 21.818865607569602, |
|
"learning_rate": 3.7234042553191484e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.484375, |
|
"logps/chosen": -100.5, |
|
"logps/rejected": -162.0, |
|
"loss": 0.3894, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": 0.3984375, |
|
"rewards/margins": 1.0703125, |
|
"rewards/rejected": -0.66796875, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.2564102564102564, |
|
"grad_norm": 26.315987904129013, |
|
"learning_rate": 4.25531914893617e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.4921875, |
|
"logps/chosen": -117.5, |
|
"logps/rejected": -170.0, |
|
"loss": 0.3909, |
|
"rewards/accuracies": 0.8500000238418579, |
|
"rewards/chosen": 0.026611328125, |
|
"rewards/margins": 1.1875, |
|
"rewards/rejected": -1.1640625, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.28846153846153844, |
|
"grad_norm": 28.51531224072289, |
|
"learning_rate": 4.787234042553192e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -114.5, |
|
"logps/rejected": -176.0, |
|
"loss": 0.387, |
|
"rewards/accuracies": 0.800000011920929, |
|
"rewards/chosen": -0.1240234375, |
|
"rewards/margins": 1.375, |
|
"rewards/rejected": -1.5, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.32051282051282054, |
|
"grad_norm": 17.39940529070616, |
|
"learning_rate": 4.96437054631829e-07, |
|
"logits/chosen": -1.59375, |
|
"logits/rejected": -1.5234375, |
|
"logps/chosen": -105.5, |
|
"logps/rejected": -169.0, |
|
"loss": 0.324, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": 0.390625, |
|
"rewards/margins": 1.6640625, |
|
"rewards/rejected": -1.2734375, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.3525641025641026, |
|
"grad_norm": 28.443023292937838, |
|
"learning_rate": 4.904988123515439e-07, |
|
"logits/chosen": -1.6015625, |
|
"logits/rejected": -1.5546875, |
|
"logps/chosen": -118.0, |
|
"logps/rejected": -199.0, |
|
"loss": 0.3077, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.052490234375, |
|
"rewards/margins": 1.75, |
|
"rewards/rejected": -1.703125, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.38461538461538464, |
|
"grad_norm": 19.86248937078093, |
|
"learning_rate": 4.845605700712589e-07, |
|
"logits/chosen": -1.640625, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -118.0, |
|
"logps/rejected": -182.0, |
|
"loss": 0.3212, |
|
"rewards/accuracies": 0.8374999761581421, |
|
"rewards/chosen": -0.41015625, |
|
"rewards/margins": 2.0625, |
|
"rewards/rejected": -2.46875, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.4166666666666667, |
|
"grad_norm": 15.34168051684897, |
|
"learning_rate": 4.786223277909738e-07, |
|
"logits/chosen": -1.625, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -127.0, |
|
"logps/rejected": -196.0, |
|
"loss": 0.3436, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -0.73046875, |
|
"rewards/margins": 2.25, |
|
"rewards/rejected": -2.984375, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.44871794871794873, |
|
"grad_norm": 31.34040320177804, |
|
"learning_rate": 4.7268408551068883e-07, |
|
"logits/chosen": -1.609375, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -138.0, |
|
"logps/rejected": -193.0, |
|
"loss": 0.335, |
|
"rewards/accuracies": 0.862500011920929, |
|
"rewards/chosen": -0.453125, |
|
"rewards/margins": 1.7578125, |
|
"rewards/rejected": -2.21875, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.4807692307692308, |
|
"grad_norm": 39.09253912133659, |
|
"learning_rate": 4.667458432304038e-07, |
|
"logits/chosen": -1.640625, |
|
"logits/rejected": -1.6328125, |
|
"logps/chosen": -128.0, |
|
"logps/rejected": -249.0, |
|
"loss": 0.2853, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -0.08056640625, |
|
"rewards/margins": 2.53125, |
|
"rewards/rejected": -2.609375, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.5128205128205128, |
|
"grad_norm": 17.194231471019577, |
|
"learning_rate": 4.6080760095011875e-07, |
|
"logits/chosen": -1.6015625, |
|
"logits/rejected": -1.5859375, |
|
"logps/chosen": -107.0, |
|
"logps/rejected": -191.0, |
|
"loss": 0.2666, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": -0.310546875, |
|
"rewards/margins": 2.71875, |
|
"rewards/rejected": -3.015625, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.5448717948717948, |
|
"grad_norm": 18.722209877304238, |
|
"learning_rate": 4.548693586698337e-07, |
|
"logits/chosen": -1.578125, |
|
"logits/rejected": -1.59375, |
|
"logps/chosen": -115.0, |
|
"logps/rejected": -175.0, |
|
"loss": 0.2696, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": -0.150390625, |
|
"rewards/margins": 2.234375, |
|
"rewards/rejected": -2.390625, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.5769230769230769, |
|
"grad_norm": 21.42273705026191, |
|
"learning_rate": 4.4893111638954866e-07, |
|
"logits/chosen": -1.5859375, |
|
"logits/rejected": -1.6171875, |
|
"logps/chosen": -121.5, |
|
"logps/rejected": -203.0, |
|
"loss": 0.2899, |
|
"rewards/accuracies": 0.9125000238418579, |
|
"rewards/chosen": -0.248046875, |
|
"rewards/margins": 2.390625, |
|
"rewards/rejected": -2.640625, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.6089743589743589, |
|
"grad_norm": 35.105792045016955, |
|
"learning_rate": 4.429928741092636e-07, |
|
"logits/chosen": -1.625, |
|
"logits/rejected": -1.546875, |
|
"logps/chosen": -126.5, |
|
"logps/rejected": -199.0, |
|
"loss": 0.2834, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -0.38671875, |
|
"rewards/margins": 2.609375, |
|
"rewards/rejected": -3.0, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.6410256410256411, |
|
"grad_norm": 25.037051509827627, |
|
"learning_rate": 4.3705463182897863e-07, |
|
"logits/chosen": -1.59375, |
|
"logits/rejected": -1.5078125, |
|
"logps/chosen": -116.5, |
|
"logps/rejected": -171.0, |
|
"loss": 0.3022, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -0.59375, |
|
"rewards/margins": 2.15625, |
|
"rewards/rejected": -2.75, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.6730769230769231, |
|
"grad_norm": 25.3426239738337, |
|
"learning_rate": 4.311163895486936e-07, |
|
"logits/chosen": -1.5859375, |
|
"logits/rejected": -1.5546875, |
|
"logps/chosen": -118.5, |
|
"logps/rejected": -185.0, |
|
"loss": 0.3209, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": -1.21875, |
|
"rewards/margins": 2.1875, |
|
"rewards/rejected": -3.40625, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.7051282051282052, |
|
"grad_norm": 41.23383631897795, |
|
"learning_rate": 4.251781472684085e-07, |
|
"logits/chosen": -1.6484375, |
|
"logits/rejected": -1.5546875, |
|
"logps/chosen": -148.0, |
|
"logps/rejected": -200.0, |
|
"loss": 0.2921, |
|
"rewards/accuracies": 0.8500000238418579, |
|
"rewards/chosen": -0.609375, |
|
"rewards/margins": 2.3125, |
|
"rewards/rejected": -2.921875, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.7371794871794872, |
|
"grad_norm": 22.289875990424193, |
|
"learning_rate": 4.192399049881235e-07, |
|
"logits/chosen": -1.609375, |
|
"logits/rejected": -1.5859375, |
|
"logps/chosen": -127.0, |
|
"logps/rejected": -199.0, |
|
"loss": 0.2805, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": -0.318359375, |
|
"rewards/margins": 2.578125, |
|
"rewards/rejected": -2.90625, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.7692307692307693, |
|
"grad_norm": 29.25683917677673, |
|
"learning_rate": 4.1330166270783846e-07, |
|
"logits/chosen": -1.59375, |
|
"logits/rejected": -1.5390625, |
|
"logps/chosen": -108.5, |
|
"logps/rejected": -180.0, |
|
"loss": 0.2541, |
|
"rewards/accuracies": 0.9125000238418579, |
|
"rewards/chosen": -0.267578125, |
|
"rewards/margins": 2.765625, |
|
"rewards/rejected": -3.03125, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.8012820512820513, |
|
"grad_norm": 28.339641539576366, |
|
"learning_rate": 4.0736342042755347e-07, |
|
"logits/chosen": -1.6484375, |
|
"logits/rejected": -1.5625, |
|
"logps/chosen": -120.0, |
|
"logps/rejected": -215.0, |
|
"loss": 0.257, |
|
"rewards/accuracies": 0.887499988079071, |
|
"rewards/chosen": -0.50390625, |
|
"rewards/margins": 3.140625, |
|
"rewards/rejected": -3.640625, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.8333333333333334, |
|
"grad_norm": 28.458003688730955, |
|
"learning_rate": 4.0142517814726837e-07, |
|
"logits/chosen": -1.6328125, |
|
"logits/rejected": -1.5703125, |
|
"logps/chosen": -124.0, |
|
"logps/rejected": -195.0, |
|
"loss": 0.2514, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -0.3203125, |
|
"rewards/margins": 2.6875, |
|
"rewards/rejected": -3.015625, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.8653846153846154, |
|
"grad_norm": 19.712649681094366, |
|
"learning_rate": 3.9548693586698333e-07, |
|
"logits/chosen": -1.640625, |
|
"logits/rejected": -1.546875, |
|
"logps/chosen": -121.0, |
|
"logps/rejected": -177.0, |
|
"loss": 0.2635, |
|
"rewards/accuracies": 0.8500000238418579, |
|
"rewards/chosen": -0.11083984375, |
|
"rewards/margins": 2.453125, |
|
"rewards/rejected": -2.5625, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.8974358974358975, |
|
"grad_norm": 46.09415755147797, |
|
"learning_rate": 3.8954869358669834e-07, |
|
"logits/chosen": -1.6640625, |
|
"logits/rejected": -1.546875, |
|
"logps/chosen": -118.5, |
|
"logps/rejected": -197.0, |
|
"loss": 0.2671, |
|
"rewards/accuracies": 0.875, |
|
"rewards/chosen": 0.0693359375, |
|
"rewards/margins": 2.78125, |
|
"rewards/rejected": -2.703125, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.9294871794871795, |
|
"grad_norm": 25.186336345995475, |
|
"learning_rate": 3.836104513064133e-07, |
|
"logits/chosen": -1.5703125, |
|
"logits/rejected": -1.5390625, |
|
"logps/chosen": -129.0, |
|
"logps/rejected": -203.0, |
|
"loss": 0.2957, |
|
"rewards/accuracies": 0.824999988079071, |
|
"rewards/chosen": -1.03125, |
|
"rewards/margins": 2.21875, |
|
"rewards/rejected": -3.25, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.9615384615384616, |
|
"grad_norm": 25.061643380626624, |
|
"learning_rate": 3.7767220902612825e-07, |
|
"logits/chosen": -1.609375, |
|
"logits/rejected": -1.578125, |
|
"logps/chosen": -134.0, |
|
"logps/rejected": -219.0, |
|
"loss": 0.2345, |
|
"rewards/accuracies": 0.887499988079071, |
|
"rewards/chosen": -0.90234375, |
|
"rewards/margins": 2.671875, |
|
"rewards/rejected": -3.578125, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.9935897435897436, |
|
"grad_norm": 44.17174735441787, |
|
"learning_rate": 3.717339667458432e-07, |
|
"logits/chosen": -1.640625, |
|
"logits/rejected": -1.5703125, |
|
"logps/chosen": -119.0, |
|
"logps/rejected": -192.0, |
|
"loss": 0.2734, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -0.2392578125, |
|
"rewards/margins": 2.734375, |
|
"rewards/rejected": -2.96875, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_logits/chosen": -1.5546875, |
|
"eval_logits/rejected": -1.5234375, |
|
"eval_logps/chosen": -170.0, |
|
"eval_logps/rejected": -184.0, |
|
"eval_loss": 0.6114453077316284, |
|
"eval_rewards/accuracies": 0.5892857313156128, |
|
"eval_rewards/chosen": -1.53125, |
|
"eval_rewards/margins": 0.8984375, |
|
"eval_rewards/rejected": -2.4375, |
|
"eval_runtime": 14.9255, |
|
"eval_samples_per_second": 13.4, |
|
"eval_steps_per_second": 0.469, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 1.0256410256410255, |
|
"grad_norm": 16.449029965031666, |
|
"learning_rate": 3.6579572446555817e-07, |
|
"logits/chosen": -1.59375, |
|
"logits/rejected": -1.578125, |
|
"logps/chosen": -117.0, |
|
"logps/rejected": -195.0, |
|
"loss": 0.1531, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": -0.34765625, |
|
"rewards/margins": 3.1875, |
|
"rewards/rejected": -3.546875, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 1.0576923076923077, |
|
"grad_norm": 6.562166239087373, |
|
"learning_rate": 3.598574821852731e-07, |
|
"logits/chosen": -1.59375, |
|
"logits/rejected": -1.5703125, |
|
"logps/chosen": -111.0, |
|
"logps/rejected": -187.0, |
|
"loss": 0.1452, |
|
"rewards/accuracies": 0.8999999761581421, |
|
"rewards/chosen": -0.1923828125, |
|
"rewards/margins": 3.40625, |
|
"rewards/rejected": -3.609375, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 1.0897435897435896, |
|
"grad_norm": 25.028970432454248, |
|
"learning_rate": 3.5391923990498813e-07, |
|
"logits/chosen": -1.6484375, |
|
"logits/rejected": -1.609375, |
|
"logps/chosen": -134.0, |
|
"logps/rejected": -219.0, |
|
"loss": 0.1345, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -0.69921875, |
|
"rewards/margins": 3.609375, |
|
"rewards/rejected": -4.3125, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 1.1217948717948718, |
|
"grad_norm": 15.596444813277682, |
|
"learning_rate": 3.479809976247031e-07, |
|
"logits/chosen": -1.65625, |
|
"logits/rejected": -1.6015625, |
|
"logps/chosen": -130.0, |
|
"logps/rejected": -207.0, |
|
"loss": 0.1092, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -0.00677490234375, |
|
"rewards/margins": 3.828125, |
|
"rewards/rejected": -3.828125, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 1.1538461538461537, |
|
"grad_norm": 22.37742199079234, |
|
"learning_rate": 3.42042755344418e-07, |
|
"logits/chosen": -1.5859375, |
|
"logits/rejected": -1.5703125, |
|
"logps/chosen": -136.0, |
|
"logps/rejected": -242.0, |
|
"loss": 0.0918, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.640625, |
|
"rewards/margins": 4.4375, |
|
"rewards/rejected": -5.09375, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 1.185897435897436, |
|
"grad_norm": 27.25128799333949, |
|
"learning_rate": 3.36104513064133e-07, |
|
"logits/chosen": -1.609375, |
|
"logits/rejected": -1.59375, |
|
"logps/chosen": -154.0, |
|
"logps/rejected": -221.0, |
|
"loss": 0.1215, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.333984375, |
|
"rewards/margins": 4.46875, |
|
"rewards/rejected": -4.78125, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 1.217948717948718, |
|
"grad_norm": 16.802650370434502, |
|
"learning_rate": 3.3016627078384796e-07, |
|
"logits/chosen": -1.625, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -121.0, |
|
"logps/rejected": -184.0, |
|
"loss": 0.1051, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -0.75390625, |
|
"rewards/margins": 3.75, |
|
"rewards/rejected": -4.5, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"grad_norm": 18.14599606798184, |
|
"learning_rate": 3.2422802850356297e-07, |
|
"logits/chosen": -1.65625, |
|
"logits/rejected": -1.5625, |
|
"logps/chosen": -118.0, |
|
"logps/rejected": -203.0, |
|
"loss": 0.1691, |
|
"rewards/accuracies": 0.925000011920929, |
|
"rewards/chosen": -0.4609375, |
|
"rewards/margins": 4.4375, |
|
"rewards/rejected": -4.875, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 1.282051282051282, |
|
"grad_norm": 14.341477798166297, |
|
"learning_rate": 3.182897862232779e-07, |
|
"logits/chosen": -1.609375, |
|
"logits/rejected": -1.5546875, |
|
"logps/chosen": -112.0, |
|
"logps/rejected": -202.0, |
|
"loss": 0.1084, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -0.2578125, |
|
"rewards/margins": 4.65625, |
|
"rewards/rejected": -4.90625, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 1.314102564102564, |
|
"grad_norm": 22.538352683910524, |
|
"learning_rate": 3.1235154394299283e-07, |
|
"logits/chosen": -1.671875, |
|
"logits/rejected": -1.625, |
|
"logps/chosen": -168.0, |
|
"logps/rejected": -270.0, |
|
"loss": 0.136, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.1796875, |
|
"rewards/margins": 4.71875, |
|
"rewards/rejected": -5.90625, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 1.3461538461538463, |
|
"grad_norm": 7.69063972210861, |
|
"learning_rate": 3.0641330166270784e-07, |
|
"logits/chosen": -1.609375, |
|
"logits/rejected": -1.59375, |
|
"logps/chosen": -128.0, |
|
"logps/rejected": -232.0, |
|
"loss": 0.1123, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.0859375, |
|
"rewards/margins": 4.59375, |
|
"rewards/rejected": -5.6875, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 1.3782051282051282, |
|
"grad_norm": 22.81215314331503, |
|
"learning_rate": 3.004750593824228e-07, |
|
"logits/chosen": -1.6328125, |
|
"logits/rejected": -1.5390625, |
|
"logps/chosen": -122.5, |
|
"logps/rejected": -209.0, |
|
"loss": 0.0971, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.33984375, |
|
"rewards/margins": 4.65625, |
|
"rewards/rejected": -5.0, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 1.4102564102564101, |
|
"grad_norm": 16.88060293661355, |
|
"learning_rate": 2.9453681710213776e-07, |
|
"logits/chosen": -1.5859375, |
|
"logits/rejected": -1.4765625, |
|
"logps/chosen": -123.0, |
|
"logps/rejected": -205.0, |
|
"loss": 0.11, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.4296875, |
|
"rewards/margins": 4.5625, |
|
"rewards/rejected": -5.0, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 1.4423076923076923, |
|
"grad_norm": 14.494261083829612, |
|
"learning_rate": 2.885985748218527e-07, |
|
"logits/chosen": -1.546875, |
|
"logits/rejected": -1.5546875, |
|
"logps/chosen": -118.5, |
|
"logps/rejected": -221.0, |
|
"loss": 0.0896, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.0859375, |
|
"rewards/margins": 4.59375, |
|
"rewards/rejected": -5.6875, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 1.4743589743589745, |
|
"grad_norm": 16.655718879118293, |
|
"learning_rate": 2.8266033254156767e-07, |
|
"logits/chosen": -1.59375, |
|
"logits/rejected": -1.5859375, |
|
"logps/chosen": -142.0, |
|
"logps/rejected": -256.0, |
|
"loss": 0.1071, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.25, |
|
"rewards/margins": 4.84375, |
|
"rewards/rejected": -6.09375, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 1.5064102564102564, |
|
"grad_norm": 20.51874312804957, |
|
"learning_rate": 2.7672209026128263e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.53125, |
|
"logps/chosen": -132.0, |
|
"logps/rejected": -231.0, |
|
"loss": 0.1012, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.015625, |
|
"rewards/margins": 4.65625, |
|
"rewards/rejected": -5.6875, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 1.5384615384615383, |
|
"grad_norm": 28.03019085733964, |
|
"learning_rate": 2.7078384798099764e-07, |
|
"logits/chosen": -1.4453125, |
|
"logits/rejected": -1.4609375, |
|
"logps/chosen": -104.5, |
|
"logps/rejected": -203.0, |
|
"loss": 0.1074, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.6796875, |
|
"rewards/margins": 4.59375, |
|
"rewards/rejected": -6.3125, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.5705128205128205, |
|
"grad_norm": 14.674518317046845, |
|
"learning_rate": 2.648456057007126e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.5390625, |
|
"logps/chosen": -139.0, |
|
"logps/rejected": -233.0, |
|
"loss": 0.1003, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.3125, |
|
"rewards/margins": 4.9375, |
|
"rewards/rejected": -6.28125, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.6025641025641026, |
|
"grad_norm": 18.731133870495857, |
|
"learning_rate": 2.589073634204275e-07, |
|
"logits/chosen": -1.5703125, |
|
"logits/rejected": -1.5625, |
|
"logps/chosen": -144.0, |
|
"logps/rejected": -251.0, |
|
"loss": 0.1228, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -1.109375, |
|
"rewards/margins": 5.25, |
|
"rewards/rejected": -6.375, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.6346153846153846, |
|
"grad_norm": 16.81631331394118, |
|
"learning_rate": 2.529691211401425e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.5, |
|
"logps/chosen": -131.0, |
|
"logps/rejected": -230.0, |
|
"loss": 0.0819, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.8359375, |
|
"rewards/margins": 5.375, |
|
"rewards/rejected": -6.21875, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.6666666666666665, |
|
"grad_norm": 15.111001669085608, |
|
"learning_rate": 2.4703087885985747e-07, |
|
"logits/chosen": -1.546875, |
|
"logits/rejected": -1.4921875, |
|
"logps/chosen": -114.0, |
|
"logps/rejected": -202.0, |
|
"loss": 0.1091, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.87109375, |
|
"rewards/margins": 4.78125, |
|
"rewards/rejected": -5.625, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.6987179487179487, |
|
"grad_norm": 26.75975769912583, |
|
"learning_rate": 2.410926365795724e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -140.0, |
|
"logps/rejected": -221.0, |
|
"loss": 0.1139, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.0625, |
|
"rewards/margins": 4.8125, |
|
"rewards/rejected": -5.875, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.7307692307692308, |
|
"grad_norm": 13.759559493772942, |
|
"learning_rate": 2.351543942992874e-07, |
|
"logits/chosen": -1.578125, |
|
"logits/rejected": -1.5, |
|
"logps/chosen": -143.0, |
|
"logps/rejected": -229.0, |
|
"loss": 0.0978, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -0.9765625, |
|
"rewards/margins": 5.09375, |
|
"rewards/rejected": -6.0625, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.7628205128205128, |
|
"grad_norm": 18.98900148314797, |
|
"learning_rate": 2.2921615201900234e-07, |
|
"logits/chosen": -1.4921875, |
|
"logits/rejected": -1.5, |
|
"logps/chosen": -114.0, |
|
"logps/rejected": -232.0, |
|
"loss": 0.1103, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -0.6953125, |
|
"rewards/margins": 5.03125, |
|
"rewards/rejected": -5.71875, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.7948717948717947, |
|
"grad_norm": 7.478935072745169, |
|
"learning_rate": 2.2327790973871732e-07, |
|
"logits/chosen": -1.515625, |
|
"logits/rejected": -1.4375, |
|
"logps/chosen": -112.0, |
|
"logps/rejected": -177.0, |
|
"loss": 0.1139, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -0.6640625, |
|
"rewards/margins": 4.5625, |
|
"rewards/rejected": -5.21875, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.8269230769230769, |
|
"grad_norm": 13.728806584543696, |
|
"learning_rate": 2.173396674584323e-07, |
|
"logits/chosen": -1.546875, |
|
"logits/rejected": -1.5546875, |
|
"logps/chosen": -130.0, |
|
"logps/rejected": -239.0, |
|
"loss": 0.1355, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -0.466796875, |
|
"rewards/margins": 4.875, |
|
"rewards/rejected": -5.34375, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.858974358974359, |
|
"grad_norm": 9.926742756220793, |
|
"learning_rate": 2.1140142517814726e-07, |
|
"logits/chosen": -1.609375, |
|
"logits/rejected": -1.5078125, |
|
"logps/chosen": -158.0, |
|
"logps/rejected": -202.0, |
|
"loss": 0.1118, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -1.09375, |
|
"rewards/margins": 4.21875, |
|
"rewards/rejected": -5.3125, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.891025641025641, |
|
"grad_norm": 12.431864869622972, |
|
"learning_rate": 2.0546318289786222e-07, |
|
"logits/chosen": -1.5546875, |
|
"logits/rejected": -1.53125, |
|
"logps/chosen": -122.0, |
|
"logps/rejected": -202.0, |
|
"loss": 0.1325, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.61328125, |
|
"rewards/margins": 4.65625, |
|
"rewards/rejected": -5.28125, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.9230769230769231, |
|
"grad_norm": 12.291656293105886, |
|
"learning_rate": 1.9952494061757718e-07, |
|
"logits/chosen": -1.5703125, |
|
"logits/rejected": -1.5625, |
|
"logps/chosen": -123.0, |
|
"logps/rejected": -227.0, |
|
"loss": 0.1085, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -0.67578125, |
|
"rewards/margins": 4.8125, |
|
"rewards/rejected": -5.5, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.9551282051282053, |
|
"grad_norm": 6.534649119157792, |
|
"learning_rate": 1.9358669833729216e-07, |
|
"logits/chosen": -1.5546875, |
|
"logits/rejected": -1.484375, |
|
"logps/chosen": -122.0, |
|
"logps/rejected": -202.0, |
|
"loss": 0.0935, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -0.56640625, |
|
"rewards/margins": 5.0, |
|
"rewards/rejected": -5.59375, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.9871794871794872, |
|
"grad_norm": 10.91488724265466, |
|
"learning_rate": 1.876484560570071e-07, |
|
"logits/chosen": -1.5703125, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -139.0, |
|
"logps/rejected": -236.0, |
|
"loss": 0.1071, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -1.1953125, |
|
"rewards/margins": 5.34375, |
|
"rewards/rejected": -6.53125, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_logits/chosen": -1.5, |
|
"eval_logits/rejected": -1.46875, |
|
"eval_logps/chosen": -188.0, |
|
"eval_logps/rejected": -206.0, |
|
"eval_loss": 0.700976550579071, |
|
"eval_rewards/accuracies": 0.6428571343421936, |
|
"eval_rewards/chosen": -3.296875, |
|
"eval_rewards/margins": 1.3046875, |
|
"eval_rewards/rejected": -4.59375, |
|
"eval_runtime": 15.0135, |
|
"eval_samples_per_second": 13.321, |
|
"eval_steps_per_second": 0.466, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 2.019230769230769, |
|
"grad_norm": 2.9708656343687063, |
|
"learning_rate": 1.8171021377672207e-07, |
|
"logits/chosen": -1.515625, |
|
"logits/rejected": -1.5, |
|
"logps/chosen": -150.0, |
|
"logps/rejected": -224.0, |
|
"loss": 0.071, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -0.98828125, |
|
"rewards/margins": 5.40625, |
|
"rewards/rejected": -6.40625, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 2.051282051282051, |
|
"grad_norm": 3.1451139864714266, |
|
"learning_rate": 1.7577197149643706e-07, |
|
"logits/chosen": -1.5078125, |
|
"logits/rejected": -1.4609375, |
|
"logps/chosen": -121.5, |
|
"logps/rejected": -198.0, |
|
"loss": 0.0534, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -0.8203125, |
|
"rewards/margins": 4.9375, |
|
"rewards/rejected": -5.75, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 2.0833333333333335, |
|
"grad_norm": 7.314670214652841, |
|
"learning_rate": 1.6983372921615202e-07, |
|
"logits/chosen": -1.5703125, |
|
"logits/rejected": -1.546875, |
|
"logps/chosen": -134.0, |
|
"logps/rejected": -247.0, |
|
"loss": 0.0546, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.73046875, |
|
"rewards/margins": 5.84375, |
|
"rewards/rejected": -6.5625, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 2.1153846153846154, |
|
"grad_norm": 6.986967407965111, |
|
"learning_rate": 1.6389548693586697e-07, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.53125, |
|
"logps/chosen": -123.5, |
|
"logps/rejected": -219.0, |
|
"loss": 0.0485, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.9375, |
|
"rewards/margins": 5.5625, |
|
"rewards/rejected": -6.53125, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 2.1474358974358974, |
|
"grad_norm": 3.152252590941172, |
|
"learning_rate": 1.5795724465558193e-07, |
|
"logits/chosen": -1.515625, |
|
"logits/rejected": -1.53125, |
|
"logps/chosen": -121.5, |
|
"logps/rejected": -264.0, |
|
"loss": 0.0391, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.97265625, |
|
"rewards/margins": 6.40625, |
|
"rewards/rejected": -7.375, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 2.1794871794871793, |
|
"grad_norm": 5.616933414198294, |
|
"learning_rate": 1.520190023752969e-07, |
|
"logits/chosen": -1.5390625, |
|
"logits/rejected": -1.5078125, |
|
"logps/chosen": -136.0, |
|
"logps/rejected": -225.0, |
|
"loss": 0.0523, |
|
"rewards/accuracies": 0.949999988079071, |
|
"rewards/chosen": -1.265625, |
|
"rewards/margins": 5.4375, |
|
"rewards/rejected": -6.6875, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 2.2115384615384617, |
|
"grad_norm": 4.462570602831943, |
|
"learning_rate": 1.4608076009501184e-07, |
|
"logits/chosen": -1.5390625, |
|
"logits/rejected": -1.53125, |
|
"logps/chosen": -141.0, |
|
"logps/rejected": -247.0, |
|
"loss": 0.0548, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -0.9921875, |
|
"rewards/margins": 5.90625, |
|
"rewards/rejected": -6.90625, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 2.2435897435897436, |
|
"grad_norm": 6.786342065093859, |
|
"learning_rate": 1.4014251781472683e-07, |
|
"logits/chosen": -1.5546875, |
|
"logits/rejected": -1.546875, |
|
"logps/chosen": -115.0, |
|
"logps/rejected": -243.0, |
|
"loss": 0.0328, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -0.7890625, |
|
"rewards/margins": 6.25, |
|
"rewards/rejected": -7.03125, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 2.2756410256410255, |
|
"grad_norm": 10.783463457904992, |
|
"learning_rate": 1.342042755344418e-07, |
|
"logits/chosen": -1.5390625, |
|
"logits/rejected": -1.546875, |
|
"logps/chosen": -149.0, |
|
"logps/rejected": -266.0, |
|
"loss": 0.0556, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.203125, |
|
"rewards/margins": 6.125, |
|
"rewards/rejected": -7.3125, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 2.3076923076923075, |
|
"grad_norm": 9.0679334215508, |
|
"learning_rate": 1.2826603325415677e-07, |
|
"logits/chosen": -1.53125, |
|
"logits/rejected": -1.4765625, |
|
"logps/chosen": -121.0, |
|
"logps/rejected": -217.0, |
|
"loss": 0.0438, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.80078125, |
|
"rewards/margins": 5.84375, |
|
"rewards/rejected": -6.65625, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 2.33974358974359, |
|
"grad_norm": 2.164285473082669, |
|
"learning_rate": 1.2232779097387173e-07, |
|
"logits/chosen": -1.5078125, |
|
"logits/rejected": -1.5, |
|
"logps/chosen": -153.0, |
|
"logps/rejected": -237.0, |
|
"loss": 0.0414, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.2421875, |
|
"rewards/margins": 6.15625, |
|
"rewards/rejected": -7.375, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 2.371794871794872, |
|
"grad_norm": 8.601824238964523, |
|
"learning_rate": 1.163895486935867e-07, |
|
"logits/chosen": -1.4921875, |
|
"logits/rejected": -1.46875, |
|
"logps/chosen": -148.0, |
|
"logps/rejected": -238.0, |
|
"loss": 0.0358, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.6953125, |
|
"rewards/margins": 5.9375, |
|
"rewards/rejected": -7.625, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 2.4038461538461537, |
|
"grad_norm": 11.674320163023708, |
|
"learning_rate": 1.1045130641330165e-07, |
|
"logits/chosen": -1.5234375, |
|
"logits/rejected": -1.5390625, |
|
"logps/chosen": -133.0, |
|
"logps/rejected": -258.0, |
|
"loss": 0.0523, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.9453125, |
|
"rewards/margins": 6.4375, |
|
"rewards/rejected": -8.375, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 2.435897435897436, |
|
"grad_norm": 4.6040374115250415, |
|
"learning_rate": 1.0451306413301662e-07, |
|
"logits/chosen": -1.515625, |
|
"logits/rejected": -1.5234375, |
|
"logps/chosen": -115.5, |
|
"logps/rejected": -258.0, |
|
"loss": 0.0425, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -0.984375, |
|
"rewards/margins": 7.40625, |
|
"rewards/rejected": -8.375, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 2.467948717948718, |
|
"grad_norm": 6.059248925740205, |
|
"learning_rate": 9.857482185273158e-08, |
|
"logits/chosen": -1.53125, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -137.0, |
|
"logps/rejected": -233.0, |
|
"loss": 0.037, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.2265625, |
|
"rewards/margins": 5.90625, |
|
"rewards/rejected": -7.125, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 2.5, |
|
"grad_norm": 5.569807408041334, |
|
"learning_rate": 9.263657957244655e-08, |
|
"logits/chosen": -1.546875, |
|
"logits/rejected": -1.53125, |
|
"logps/chosen": -136.0, |
|
"logps/rejected": -240.0, |
|
"loss": 0.0383, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.46875, |
|
"rewards/margins": 6.78125, |
|
"rewards/rejected": -8.25, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 2.532051282051282, |
|
"grad_norm": 9.080566390829592, |
|
"learning_rate": 8.669833729216151e-08, |
|
"logits/chosen": -1.515625, |
|
"logits/rejected": -1.4375, |
|
"logps/chosen": -138.0, |
|
"logps/rejected": -233.0, |
|
"loss": 0.043, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -2.125, |
|
"rewards/margins": 6.1875, |
|
"rewards/rejected": -8.3125, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 2.564102564102564, |
|
"grad_norm": 4.722752417362213, |
|
"learning_rate": 8.076009501187649e-08, |
|
"logits/chosen": -1.4609375, |
|
"logits/rejected": -1.515625, |
|
"logps/chosen": -129.0, |
|
"logps/rejected": -244.0, |
|
"loss": 0.043, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.78125, |
|
"rewards/margins": 6.5625, |
|
"rewards/rejected": -8.375, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 2.5961538461538463, |
|
"grad_norm": 8.320380078890436, |
|
"learning_rate": 7.482185273159145e-08, |
|
"logits/chosen": -1.5625, |
|
"logits/rejected": -1.484375, |
|
"logps/chosen": -152.0, |
|
"logps/rejected": -234.0, |
|
"loss": 0.0311, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.734375, |
|
"rewards/margins": 6.25, |
|
"rewards/rejected": -7.96875, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 2.628205128205128, |
|
"grad_norm": 10.713886798327927, |
|
"learning_rate": 6.88836104513064e-08, |
|
"logits/chosen": -1.5078125, |
|
"logits/rejected": -1.4765625, |
|
"logps/chosen": -151.0, |
|
"logps/rejected": -252.0, |
|
"loss": 0.0368, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.6875, |
|
"rewards/margins": 6.25, |
|
"rewards/rejected": -7.9375, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 2.66025641025641, |
|
"grad_norm": 10.956365394123852, |
|
"learning_rate": 6.294536817102138e-08, |
|
"logits/chosen": -1.484375, |
|
"logits/rejected": -1.4375, |
|
"logps/chosen": -139.0, |
|
"logps/rejected": -224.0, |
|
"loss": 0.0434, |
|
"rewards/accuracies": 0.9750000238418579, |
|
"rewards/chosen": -1.7421875, |
|
"rewards/margins": 6.15625, |
|
"rewards/rejected": -7.875, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 2.6923076923076925, |
|
"grad_norm": 5.167870965947855, |
|
"learning_rate": 5.700712589073634e-08, |
|
"logits/chosen": -1.5546875, |
|
"logits/rejected": -1.484375, |
|
"logps/chosen": -155.0, |
|
"logps/rejected": -286.0, |
|
"loss": 0.0367, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.4140625, |
|
"rewards/margins": 7.15625, |
|
"rewards/rejected": -8.5625, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 2.7243589743589745, |
|
"grad_norm": 10.357396384526298, |
|
"learning_rate": 5.10688836104513e-08, |
|
"logits/chosen": -1.4765625, |
|
"logits/rejected": -1.5, |
|
"logps/chosen": -144.0, |
|
"logps/rejected": -245.0, |
|
"loss": 0.0381, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.8125, |
|
"rewards/margins": 6.46875, |
|
"rewards/rejected": -8.3125, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 2.7564102564102564, |
|
"grad_norm": 4.157699399133391, |
|
"learning_rate": 4.5130641330166267e-08, |
|
"logits/chosen": -1.46875, |
|
"logits/rejected": -1.421875, |
|
"logps/chosen": -126.0, |
|
"logps/rejected": -239.0, |
|
"loss": 0.0442, |
|
"rewards/accuracies": 0.9375, |
|
"rewards/chosen": -1.5859375, |
|
"rewards/margins": 6.25, |
|
"rewards/rejected": -7.8125, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 2.7884615384615383, |
|
"grad_norm": 10.34195664309316, |
|
"learning_rate": 3.919239904988123e-08, |
|
"logits/chosen": -1.5234375, |
|
"logits/rejected": -1.5390625, |
|
"logps/chosen": -148.0, |
|
"logps/rejected": -268.0, |
|
"loss": 0.036, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -2.296875, |
|
"rewards/margins": 6.5, |
|
"rewards/rejected": -8.8125, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 2.8205128205128203, |
|
"grad_norm": 13.885196758276543, |
|
"learning_rate": 3.32541567695962e-08, |
|
"logits/chosen": -1.53125, |
|
"logits/rejected": -1.4609375, |
|
"logps/chosen": -139.0, |
|
"logps/rejected": -237.0, |
|
"loss": 0.0386, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.7890625, |
|
"rewards/margins": 6.4375, |
|
"rewards/rejected": -8.25, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 2.8525641025641026, |
|
"grad_norm": 3.298992157382736, |
|
"learning_rate": 2.7315914489311164e-08, |
|
"logits/chosen": -1.53125, |
|
"logits/rejected": -1.484375, |
|
"logps/chosen": -142.0, |
|
"logps/rejected": -264.0, |
|
"loss": 0.0326, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -2.0, |
|
"rewards/margins": 6.84375, |
|
"rewards/rejected": -8.875, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 2.8846153846153846, |
|
"grad_norm": 9.55680354164709, |
|
"learning_rate": 2.1377672209026125e-08, |
|
"logits/chosen": -1.4609375, |
|
"logits/rejected": -1.4453125, |
|
"logps/chosen": -124.0, |
|
"logps/rejected": -247.0, |
|
"loss": 0.0386, |
|
"rewards/accuracies": 1.0, |
|
"rewards/chosen": -1.7734375, |
|
"rewards/margins": 7.0, |
|
"rewards/rejected": -8.8125, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 2.9166666666666665, |
|
"grad_norm": 8.674386112287156, |
|
"learning_rate": 1.5439429928741092e-08, |
|
"logits/chosen": -1.5078125, |
|
"logits/rejected": -1.4765625, |
|
"logps/chosen": -118.0, |
|
"logps/rejected": -240.0, |
|
"loss": 0.0517, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.3984375, |
|
"rewards/margins": 7.0625, |
|
"rewards/rejected": -8.4375, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 2.948717948717949, |
|
"grad_norm": 6.427610044394737, |
|
"learning_rate": 9.501187648456057e-09, |
|
"logits/chosen": -1.5, |
|
"logits/rejected": -1.4765625, |
|
"logps/chosen": -148.0, |
|
"logps/rejected": -256.0, |
|
"loss": 0.0481, |
|
"rewards/accuracies": 0.987500011920929, |
|
"rewards/chosen": -1.84375, |
|
"rewards/margins": 6.375, |
|
"rewards/rejected": -8.25, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 2.980769230769231, |
|
"grad_norm": 20.775958318465445, |
|
"learning_rate": 3.562945368171021e-09, |
|
"logits/chosen": -1.53125, |
|
"logits/rejected": -1.5, |
|
"logps/chosen": -141.0, |
|
"logps/rejected": -232.0, |
|
"loss": 0.0498, |
|
"rewards/accuracies": 0.9624999761581421, |
|
"rewards/chosen": -1.640625, |
|
"rewards/margins": 6.0, |
|
"rewards/rejected": -7.625, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_logits/chosen": -1.4921875, |
|
"eval_logits/rejected": -1.4609375, |
|
"eval_logps/chosen": -199.0, |
|
"eval_logps/rejected": -222.0, |
|
"eval_loss": 0.830273449420929, |
|
"eval_rewards/accuracies": 0.625, |
|
"eval_rewards/chosen": -4.4375, |
|
"eval_rewards/margins": 1.765625, |
|
"eval_rewards/rejected": -6.1875, |
|
"eval_runtime": 16.7003, |
|
"eval_samples_per_second": 11.976, |
|
"eval_steps_per_second": 0.419, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 936, |
|
"total_flos": 0.0, |
|
"train_loss": 0.1719489719113733, |
|
"train_runtime": 6022.2267, |
|
"train_samples_per_second": 4.969, |
|
"train_steps_per_second": 0.155 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 936, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 0.0, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|