| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 55.87301587301587, | |
| "eval_steps": 40, | |
| "global_step": 880, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 7.928347587585449, | |
| "learning_rate": 7.5e-08, | |
| "logits/chosen": -2.7277705669403076, | |
| "logits/rejected": -2.7679762840270996, | |
| "logps/chosen": -128.4798583984375, | |
| "logps/rejected": -98.36178588867188, | |
| "loss": 0.6931, | |
| "rewards/accuracies": 0.109375, | |
| "rewards/chosen": -7.17043731128797e-05, | |
| "rewards/margins": 5.325676465872675e-05, | |
| "rewards/rejected": -0.0001249611668754369, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 9.242063522338867, | |
| "learning_rate": 1.75e-07, | |
| "logits/chosen": -2.6805009841918945, | |
| "logits/rejected": -2.712625026702881, | |
| "logps/chosen": -138.11959838867188, | |
| "logps/rejected": -96.05882263183594, | |
| "loss": 0.693, | |
| "rewards/accuracies": 0.546875, | |
| "rewards/chosen": 0.0008271098486147821, | |
| "rewards/margins": 0.00034989125560969114, | |
| "rewards/rejected": 0.00047721865121275187, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 8.156970977783203, | |
| "learning_rate": 2.75e-07, | |
| "logits/chosen": -2.7695400714874268, | |
| "logits/rejected": -2.7420883178710938, | |
| "logps/chosen": -133.04307556152344, | |
| "logps/rejected": -104.19400024414062, | |
| "loss": 0.6919, | |
| "rewards/accuracies": 0.609375, | |
| "rewards/chosen": 0.0034509659744799137, | |
| "rewards/margins": 0.0025641382671892643, | |
| "rewards/rejected": 0.0008868275908753276, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "grad_norm": 8.723258018493652, | |
| "learning_rate": 3.75e-07, | |
| "logits/chosen": -2.6869003772735596, | |
| "logits/rejected": -2.7081785202026367, | |
| "logps/chosen": -126.2979736328125, | |
| "logps/rejected": -98.4257583618164, | |
| "loss": 0.6889, | |
| "rewards/accuracies": 0.828125, | |
| "rewards/chosen": 0.011365305632352829, | |
| "rewards/margins": 0.008443659171462059, | |
| "rewards/rejected": 0.0029216469265520573, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "grad_norm": 10.34566593170166, | |
| "learning_rate": 4.7499999999999995e-07, | |
| "logits/chosen": -2.704402208328247, | |
| "logits/rejected": -2.764737367630005, | |
| "logps/chosen": -125.91941833496094, | |
| "logps/rejected": -99.02344512939453, | |
| "loss": 0.6832, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.02058127522468567, | |
| "rewards/margins": 0.020110297948122025, | |
| "rewards/rejected": 0.00047097797505557537, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "grad_norm": 9.291991233825684, | |
| "learning_rate": 4.999985633126757e-07, | |
| "logits/chosen": -2.716165781021118, | |
| "logits/rejected": -2.6845083236694336, | |
| "logps/chosen": -137.2718505859375, | |
| "logps/rejected": -97.50873565673828, | |
| "loss": 0.6785, | |
| "rewards/accuracies": 0.90625, | |
| "rewards/chosen": 0.03298673778772354, | |
| "rewards/margins": 0.029706846922636032, | |
| "rewards/rejected": 0.0032798887696117163, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "grad_norm": 9.161444664001465, | |
| "learning_rate": 4.999921780689761e-07, | |
| "logits/chosen": -2.735781192779541, | |
| "logits/rejected": -2.7485365867614746, | |
| "logps/chosen": -128.5098876953125, | |
| "logps/rejected": -90.1419906616211, | |
| "loss": 0.6693, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.04758748784661293, | |
| "rewards/margins": 0.04847026616334915, | |
| "rewards/rejected": -0.0008827749406918883, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "grad_norm": 8.495955467224121, | |
| "learning_rate": 4.99980684767309e-07, | |
| "logits/chosen": -2.7206833362579346, | |
| "logits/rejected": -2.750042676925659, | |
| "logps/chosen": -133.54827880859375, | |
| "logps/rejected": -108.23342895507812, | |
| "loss": 0.6669, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.0631466805934906, | |
| "rewards/margins": 0.05378420650959015, | |
| "rewards/rejected": 0.009362474083900452, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "grad_norm": 8.951277732849121, | |
| "learning_rate": 4.999640836425158e-07, | |
| "logits/chosen": -2.6706807613372803, | |
| "logits/rejected": -2.6151680946350098, | |
| "logps/chosen": -136.55226135253906, | |
| "logps/rejected": -97.76315307617188, | |
| "loss": 0.6578, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.08149851858615875, | |
| "rewards/margins": 0.07302143424749374, | |
| "rewards/rejected": 0.008477086201310158, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "grad_norm": 8.608317375183105, | |
| "learning_rate": 4.99942375033805e-07, | |
| "logits/chosen": -2.7676639556884766, | |
| "logits/rejected": -2.833590030670166, | |
| "logps/chosen": -133.7703857421875, | |
| "logps/rejected": -106.12245178222656, | |
| "loss": 0.6488, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.10222765803337097, | |
| "rewards/margins": 0.09170522540807724, | |
| "rewards/rejected": 0.010522443801164627, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "eval_logits/chosen": -2.806352138519287, | |
| "eval_logits/rejected": -2.8898861408233643, | |
| "eval_logps/chosen": -133.14093017578125, | |
| "eval_logps/rejected": -97.8843002319336, | |
| "eval_loss": 0.6592782139778137, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 0.09283973276615143, | |
| "eval_rewards/margins": 0.07038124650716782, | |
| "eval_rewards/rejected": 0.022458484396338463, | |
| "eval_runtime": 19.3683, | |
| "eval_samples_per_second": 1.446, | |
| "eval_steps_per_second": 1.446, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "grad_norm": 9.09485912322998, | |
| "learning_rate": 4.999155593847457e-07, | |
| "logits/chosen": -2.7024385929107666, | |
| "logits/rejected": -2.7564737796783447, | |
| "logps/chosen": -123.412353515625, | |
| "logps/rejected": -91.91966247558594, | |
| "loss": 0.6347, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.114817313849926, | |
| "rewards/margins": 0.12200291454792023, | |
| "rewards/rejected": -0.007185595575720072, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "grad_norm": 9.453149795532227, | |
| "learning_rate": 4.998836372432589e-07, | |
| "logits/chosen": -2.736665725708008, | |
| "logits/rejected": -2.757841110229492, | |
| "logps/chosen": -127.0240478515625, | |
| "logps/rejected": -103.66780853271484, | |
| "loss": 0.636, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.13144327700138092, | |
| "rewards/margins": 0.11972987651824951, | |
| "rewards/rejected": 0.011713397689163685, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "grad_norm": 8.674492835998535, | |
| "learning_rate": 4.998466092616052e-07, | |
| "logits/chosen": -2.710110664367676, | |
| "logits/rejected": -2.7205581665039062, | |
| "logps/chosen": -131.0464324951172, | |
| "logps/rejected": -99.71279907226562, | |
| "loss": 0.6166, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.15897569060325623, | |
| "rewards/margins": 0.16229701042175293, | |
| "rewards/rejected": -0.003321333322674036, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 3.56, | |
| "grad_norm": 9.434328079223633, | |
| "learning_rate": 4.99804476196373e-07, | |
| "logits/chosen": -2.685814380645752, | |
| "logits/rejected": -2.736269474029541, | |
| "logps/chosen": -125.750732421875, | |
| "logps/rejected": -93.69428253173828, | |
| "loss": 0.6131, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.16781488060951233, | |
| "rewards/margins": 0.16998659074306488, | |
| "rewards/rejected": -0.002171688713133335, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "grad_norm": 8.643455505371094, | |
| "learning_rate": 4.997572389084615e-07, | |
| "logits/chosen": -2.7429587841033936, | |
| "logits/rejected": -2.75490140914917, | |
| "logps/chosen": -131.4405517578125, | |
| "logps/rejected": -99.93754577636719, | |
| "loss": 0.614, | |
| "rewards/accuracies": 0.921875, | |
| "rewards/chosen": 0.18873561918735504, | |
| "rewards/margins": 0.17132429778575897, | |
| "rewards/rejected": 0.01741132140159607, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "grad_norm": 8.412778854370117, | |
| "learning_rate": 4.997048983630643e-07, | |
| "logits/chosen": -2.6886143684387207, | |
| "logits/rejected": -2.710732936859131, | |
| "logps/chosen": -123.77167510986328, | |
| "logps/rejected": -97.30846405029297, | |
| "loss": 0.6099, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.19313779473304749, | |
| "rewards/margins": 0.1793561577796936, | |
| "rewards/rejected": 0.013781649991869926, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "grad_norm": 7.851288795471191, | |
| "learning_rate": 4.99647455629649e-07, | |
| "logits/chosen": -2.744528293609619, | |
| "logits/rejected": -2.7117881774902344, | |
| "logps/chosen": -136.25863647460938, | |
| "logps/rejected": -95.11305236816406, | |
| "loss": 0.5968, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.21463464200496674, | |
| "rewards/margins": 0.20793242752552032, | |
| "rewards/rejected": 0.0067022331058979034, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 4.57, | |
| "grad_norm": 9.278959274291992, | |
| "learning_rate": 4.995849118819353e-07, | |
| "logits/chosen": -2.7611894607543945, | |
| "logits/rejected": -2.7447633743286133, | |
| "logps/chosen": -134.3462677001953, | |
| "logps/rejected": -103.75216674804688, | |
| "loss": 0.5828, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.25433310866355896, | |
| "rewards/margins": 0.24298511445522308, | |
| "rewards/rejected": 0.011347985826432705, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "grad_norm": 7.795536994934082, | |
| "learning_rate": 4.995172683978719e-07, | |
| "logits/chosen": -2.698429584503174, | |
| "logits/rejected": -2.7458348274230957, | |
| "logps/chosen": -127.65770721435547, | |
| "logps/rejected": -104.23495483398438, | |
| "loss": 0.5796, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.25434932112693787, | |
| "rewards/margins": 0.25114190578460693, | |
| "rewards/rejected": 0.0032073920592665672, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 5.08, | |
| "grad_norm": 7.788575649261475, | |
| "learning_rate": 4.994445265596091e-07, | |
| "logits/chosen": -2.6734955310821533, | |
| "logits/rejected": -2.7292165756225586, | |
| "logps/chosen": -123.53501892089844, | |
| "logps/rejected": -95.23420715332031, | |
| "loss": 0.5676, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.2950035631656647, | |
| "rewards/margins": 0.2805778682231903, | |
| "rewards/rejected": 0.014425676316022873, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 5.08, | |
| "eval_logits/chosen": -2.8074097633361816, | |
| "eval_logits/rejected": -2.890533685684204, | |
| "eval_logps/chosen": -131.55104064941406, | |
| "eval_logps/rejected": -97.56663513183594, | |
| "eval_loss": 0.6045485734939575, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 0.2518289387226105, | |
| "eval_rewards/margins": 0.1976027488708496, | |
| "eval_rewards/rejected": 0.05422618240118027, | |
| "eval_runtime": 19.3586, | |
| "eval_samples_per_second": 1.446, | |
| "eval_steps_per_second": 1.446, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 5.33, | |
| "grad_norm": 6.709670543670654, | |
| "learning_rate": 4.993666878534718e-07, | |
| "logits/chosen": -2.736870527267456, | |
| "logits/rejected": -2.778140068054199, | |
| "logps/chosen": -128.86972045898438, | |
| "logps/rejected": -104.70977020263672, | |
| "loss": 0.5637, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.31255263090133667, | |
| "rewards/margins": 0.2908375561237335, | |
| "rewards/rejected": 0.021715058013796806, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 5.59, | |
| "grad_norm": 8.178019523620605, | |
| "learning_rate": 4.99283753869928e-07, | |
| "logits/chosen": -2.686039447784424, | |
| "logits/rejected": -2.681716203689575, | |
| "logps/chosen": -129.52798461914062, | |
| "logps/rejected": -97.80433654785156, | |
| "loss": 0.5475, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 0.32084640860557556, | |
| "rewards/margins": 0.33009591698646545, | |
| "rewards/rejected": -0.00924946740269661, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 5.84, | |
| "grad_norm": 7.515178680419922, | |
| "learning_rate": 4.991957263035573e-07, | |
| "logits/chosen": -2.763765811920166, | |
| "logits/rejected": -2.7660324573516846, | |
| "logps/chosen": -128.16851806640625, | |
| "logps/rejected": -95.72579956054688, | |
| "loss": 0.5493, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.3080848455429077, | |
| "rewards/margins": 0.32350555062294006, | |
| "rewards/rejected": -0.015420722775161266, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 6.1, | |
| "grad_norm": 8.599517822265625, | |
| "learning_rate": 4.991026069530156e-07, | |
| "logits/chosen": -2.7107338905334473, | |
| "logits/rejected": -2.745981216430664, | |
| "logps/chosen": -131.5907745361328, | |
| "logps/rejected": -102.65729522705078, | |
| "loss": 0.5568, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.36698901653289795, | |
| "rewards/margins": 0.3141450881958008, | |
| "rewards/rejected": 0.05284389108419418, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 6.35, | |
| "grad_norm": 7.087887287139893, | |
| "learning_rate": 4.990043977209984e-07, | |
| "logits/chosen": -2.75876522064209, | |
| "logits/rejected": -2.800281047821045, | |
| "logps/chosen": -124.87574005126953, | |
| "logps/rejected": -99.60267639160156, | |
| "loss": 0.5332, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.3760470151901245, | |
| "rewards/margins": 0.37043848633766174, | |
| "rewards/rejected": 0.00560858054086566, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "grad_norm": 6.458583831787109, | |
| "learning_rate": 4.989011006142024e-07, | |
| "logits/chosen": -2.70744252204895, | |
| "logits/rejected": -2.7285239696502686, | |
| "logps/chosen": -129.41262817382812, | |
| "logps/rejected": -98.60986328125, | |
| "loss": 0.5286, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.3909401297569275, | |
| "rewards/margins": 0.3845981955528259, | |
| "rewards/rejected": 0.006341893225908279, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 6.86, | |
| "grad_norm": 8.084393501281738, | |
| "learning_rate": 4.987927177432842e-07, | |
| "logits/chosen": -2.731127977371216, | |
| "logits/rejected": -2.7352988719940186, | |
| "logps/chosen": -123.85149383544922, | |
| "logps/rejected": -97.99423217773438, | |
| "loss": 0.544, | |
| "rewards/accuracies": 0.890625, | |
| "rewards/chosen": 0.38825392723083496, | |
| "rewards/margins": 0.34452566504478455, | |
| "rewards/rejected": 0.043728262186050415, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "grad_norm": 7.936505317687988, | |
| "learning_rate": 4.987080943856886e-07, | |
| "logits/chosen": -2.6931982040405273, | |
| "logits/rejected": -2.683288812637329, | |
| "logps/chosen": -130.97227478027344, | |
| "logps/rejected": -93.71176147460938, | |
| "loss": 0.4903, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.45238468050956726, | |
| "rewards/margins": 0.48204928636550903, | |
| "rewards/rejected": -0.029664600268006325, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 7.37, | |
| "grad_norm": 6.827272415161133, | |
| "learning_rate": 4.985908168188602e-07, | |
| "logits/chosen": -2.697653293609619, | |
| "logits/rejected": -2.7467117309570312, | |
| "logps/chosen": -119.16637420654297, | |
| "logps/rejected": -95.0647201538086, | |
| "loss": 0.5088, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.4183903634548187, | |
| "rewards/margins": 0.43436184525489807, | |
| "rewards/rejected": -0.015971507877111435, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "grad_norm": 6.599144458770752, | |
| "learning_rate": 4.984684598278982e-07, | |
| "logits/chosen": -2.7136449813842773, | |
| "logits/rejected": -2.7430026531219482, | |
| "logps/chosen": -126.032958984375, | |
| "logps/rejected": -102.69947814941406, | |
| "loss": 0.5231, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.4575308859348297, | |
| "rewards/margins": 0.40544256567955017, | |
| "rewards/rejected": 0.052088312804698944, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 7.62, | |
| "eval_logits/chosen": -2.807668447494507, | |
| "eval_logits/rejected": -2.8901402950286865, | |
| "eval_logps/chosen": -129.97122192382812, | |
| "eval_logps/rejected": -97.28166961669922, | |
| "eval_loss": 0.5570675134658813, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 0.4098115563392639, | |
| "eval_rewards/margins": 0.3270883560180664, | |
| "eval_rewards/rejected": 0.0827232226729393, | |
| "eval_runtime": 19.4265, | |
| "eval_samples_per_second": 1.441, | |
| "eval_steps_per_second": 1.441, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 7.87, | |
| "grad_norm": 7.769857883453369, | |
| "learning_rate": 4.983410259129073e-07, | |
| "logits/chosen": -2.7467093467712402, | |
| "logits/rejected": -2.7454986572265625, | |
| "logps/chosen": -134.60739135742188, | |
| "logps/rejected": -103.20536804199219, | |
| "loss": 0.4995, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.4985056221485138, | |
| "rewards/margins": 0.4699307978153229, | |
| "rewards/rejected": 0.028574790805578232, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 8.13, | |
| "grad_norm": 7.224113941192627, | |
| "learning_rate": 4.982085176777285e-07, | |
| "logits/chosen": -2.725543975830078, | |
| "logits/rejected": -2.7536404132843018, | |
| "logps/chosen": -121.35929870605469, | |
| "logps/rejected": -97.49042510986328, | |
| "loss": 0.4988, | |
| "rewards/accuracies": 0.921875, | |
| "rewards/chosen": 0.4724448025226593, | |
| "rewards/margins": 0.4719679057598114, | |
| "rewards/rejected": 0.0004768962971866131, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "grad_norm": 5.916596412658691, | |
| "learning_rate": 4.980709378298851e-07, | |
| "logits/chosen": -2.7414817810058594, | |
| "logits/rejected": -2.7378463745117188, | |
| "logps/chosen": -119.76469421386719, | |
| "logps/rejected": -92.40216827392578, | |
| "loss": 0.4946, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.48038381338119507, | |
| "rewards/margins": 0.4809207320213318, | |
| "rewards/rejected": -0.0005369335412979126, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 8.63, | |
| "grad_norm": 6.414724826812744, | |
| "learning_rate": 4.979282891805286e-07, | |
| "logits/chosen": -2.7029452323913574, | |
| "logits/rejected": -2.737753391265869, | |
| "logps/chosen": -130.59829711914062, | |
| "logps/rejected": -99.5873794555664, | |
| "loss": 0.4564, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.5609064698219299, | |
| "rewards/margins": 0.5927521586418152, | |
| "rewards/rejected": -0.03184571862220764, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 8.89, | |
| "grad_norm": 7.006284236907959, | |
| "learning_rate": 4.977805746443806e-07, | |
| "logits/chosen": -2.719618320465088, | |
| "logits/rejected": -2.7250468730926514, | |
| "logps/chosen": -128.04678344726562, | |
| "logps/rejected": -101.81006622314453, | |
| "loss": 0.4761, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.5477940440177917, | |
| "rewards/margins": 0.5303957462310791, | |
| "rewards/rejected": 0.017398254945874214, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 9.14, | |
| "grad_norm": 7.040762901306152, | |
| "learning_rate": 4.97627797239673e-07, | |
| "logits/chosen": -2.6895217895507812, | |
| "logits/rejected": -2.7162435054779053, | |
| "logps/chosen": -133.3381805419922, | |
| "logps/rejected": -102.25975799560547, | |
| "loss": 0.4403, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.6185128688812256, | |
| "rewards/margins": 0.6498669981956482, | |
| "rewards/rejected": -0.03135409206151962, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "grad_norm": 6.93245267868042, | |
| "learning_rate": 4.974699600880869e-07, | |
| "logits/chosen": -2.6911263465881348, | |
| "logits/rejected": -2.7075552940368652, | |
| "logps/chosen": -126.95785522460938, | |
| "logps/rejected": -102.03314971923828, | |
| "loss": 0.4587, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.573100745677948, | |
| "rewards/margins": 0.5804403424263, | |
| "rewards/rejected": -0.0073395660147070885, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 9.65, | |
| "grad_norm": 7.084275245666504, | |
| "learning_rate": 4.973070664146885e-07, | |
| "logits/chosen": -2.7053210735321045, | |
| "logits/rejected": -2.7612156867980957, | |
| "logps/chosen": -119.92796325683594, | |
| "logps/rejected": -97.49836730957031, | |
| "loss": 0.483, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.5761281251907349, | |
| "rewards/margins": 0.5190825462341309, | |
| "rewards/rejected": 0.057045597583055496, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 9.9, | |
| "grad_norm": 5.3888258934021, | |
| "learning_rate": 4.971391195478631e-07, | |
| "logits/chosen": -2.750107765197754, | |
| "logits/rejected": -2.7442452907562256, | |
| "logps/chosen": -128.85630798339844, | |
| "logps/rejected": -96.89845275878906, | |
| "loss": 0.4517, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.6420415043830872, | |
| "rewards/margins": 0.6197397708892822, | |
| "rewards/rejected": 0.022301698103547096, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 10.16, | |
| "grad_norm": 6.74313497543335, | |
| "learning_rate": 4.969661229192477e-07, | |
| "logits/chosen": -2.7252540588378906, | |
| "logits/rejected": -2.709249258041382, | |
| "logps/chosen": -126.21170043945312, | |
| "logps/rejected": -105.25879669189453, | |
| "loss": 0.4607, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.6133719682693481, | |
| "rewards/margins": 0.6076129674911499, | |
| "rewards/rejected": 0.005758981220424175, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 10.16, | |
| "eval_logits/chosen": -2.807128429412842, | |
| "eval_logits/rejected": -2.8889076709747314, | |
| "eval_logps/chosen": -128.48606872558594, | |
| "eval_logps/rejected": -97.06517028808594, | |
| "eval_loss": 0.5175719857215881, | |
| "eval_rewards/accuracies": 0.8214285969734192, | |
| "eval_rewards/chosen": 0.5583271384239197, | |
| "eval_rewards/margins": 0.4539553225040436, | |
| "eval_rewards/rejected": 0.10437185317277908, | |
| "eval_runtime": 19.3848, | |
| "eval_samples_per_second": 1.444, | |
| "eval_steps_per_second": 1.444, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 10.41, | |
| "grad_norm": 5.443469524383545, | |
| "learning_rate": 4.967880800636598e-07, | |
| "logits/chosen": -2.7496747970581055, | |
| "logits/rejected": -2.768813133239746, | |
| "logps/chosen": -122.11544799804688, | |
| "logps/rejected": -94.54540252685547, | |
| "loss": 0.436, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.6616504788398743, | |
| "rewards/margins": 0.67127525806427, | |
| "rewards/rejected": -0.009624744765460491, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 10.67, | |
| "grad_norm": 5.551913738250732, | |
| "learning_rate": 4.966049946190263e-07, | |
| "logits/chosen": -2.7052814960479736, | |
| "logits/rejected": -2.7216036319732666, | |
| "logps/chosen": -131.8173065185547, | |
| "logps/rejected": -102.79641723632812, | |
| "loss": 0.4326, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.7013098001480103, | |
| "rewards/margins": 0.6718720197677612, | |
| "rewards/rejected": 0.02943783439695835, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 10.92, | |
| "grad_norm": 5.826549530029297, | |
| "learning_rate": 4.964168703263086e-07, | |
| "logits/chosen": -2.701007604598999, | |
| "logits/rejected": -2.741464614868164, | |
| "logps/chosen": -116.08711242675781, | |
| "logps/rejected": -94.73246765136719, | |
| "loss": 0.446, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.6390998363494873, | |
| "rewards/margins": 0.6307864785194397, | |
| "rewards/rejected": 0.008313396014273167, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 11.17, | |
| "grad_norm": 6.066051483154297, | |
| "learning_rate": 4.962237110294259e-07, | |
| "logits/chosen": -2.7041149139404297, | |
| "logits/rejected": -2.731684684753418, | |
| "logps/chosen": -124.84115600585938, | |
| "logps/rejected": -103.8166732788086, | |
| "loss": 0.4294, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.6835889220237732, | |
| "rewards/margins": 0.6921644806861877, | |
| "rewards/rejected": -0.00857558287680149, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 11.43, | |
| "grad_norm": 5.8145012855529785, | |
| "learning_rate": 4.960255206751773e-07, | |
| "logits/chosen": -2.7352046966552734, | |
| "logits/rejected": -2.7856600284576416, | |
| "logps/chosen": -123.47644805908203, | |
| "logps/rejected": -93.96688079833984, | |
| "loss": 0.4056, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.7594637274742126, | |
| "rewards/margins": 0.7790957093238831, | |
| "rewards/rejected": -0.019631966948509216, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 11.68, | |
| "grad_norm": 5.962335109710693, | |
| "learning_rate": 4.958223033131609e-07, | |
| "logits/chosen": -2.740914821624756, | |
| "logits/rejected": -2.7585391998291016, | |
| "logps/chosen": -124.35005950927734, | |
| "logps/rejected": -105.05398559570312, | |
| "loss": 0.4293, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.6812430620193481, | |
| "rewards/margins": 0.6962765455245972, | |
| "rewards/rejected": -0.015033497475087643, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 11.94, | |
| "grad_norm": 5.529379367828369, | |
| "learning_rate": 4.956140630956907e-07, | |
| "logits/chosen": -2.7024734020233154, | |
| "logits/rejected": -2.6746203899383545, | |
| "logps/chosen": -125.29593658447266, | |
| "logps/rejected": -94.42231750488281, | |
| "loss": 0.3963, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.7892162799835205, | |
| "rewards/margins": 0.8050926327705383, | |
| "rewards/rejected": -0.01587628945708275, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 12.19, | |
| "grad_norm": 5.774294376373291, | |
| "learning_rate": 4.954008042777125e-07, | |
| "logits/chosen": -2.7662813663482666, | |
| "logits/rejected": -2.8167452812194824, | |
| "logps/chosen": -124.10614776611328, | |
| "logps/rejected": -94.9889907836914, | |
| "loss": 0.3728, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.8182777166366577, | |
| "rewards/margins": 0.8937400579452515, | |
| "rewards/rejected": -0.07546230405569077, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 12.44, | |
| "grad_norm": 4.990906715393066, | |
| "learning_rate": 4.951825312167159e-07, | |
| "logits/chosen": -2.7352771759033203, | |
| "logits/rejected": -2.7206149101257324, | |
| "logps/chosen": -116.98539733886719, | |
| "logps/rejected": -96.68843078613281, | |
| "loss": 0.4435, | |
| "rewards/accuracies": 0.9375, | |
| "rewards/chosen": 0.7208267450332642, | |
| "rewards/margins": 0.6662319302558899, | |
| "rewards/rejected": 0.05459484085440636, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 12.7, | |
| "grad_norm": 7.139024257659912, | |
| "learning_rate": 4.949592483726464e-07, | |
| "logits/chosen": -2.728516101837158, | |
| "logits/rejected": -2.732177257537842, | |
| "logps/chosen": -125.51838684082031, | |
| "logps/rejected": -103.7790298461914, | |
| "loss": 0.391, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.7955451011657715, | |
| "rewards/margins": 0.8273122310638428, | |
| "rewards/rejected": -0.03176717832684517, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 12.7, | |
| "eval_logits/chosen": -2.8053383827209473, | |
| "eval_logits/rejected": -2.8861937522888184, | |
| "eval_logps/chosen": -127.08057403564453, | |
| "eval_logps/rejected": -96.8802261352539, | |
| "eval_loss": 0.4858916103839874, | |
| "eval_rewards/accuracies": 0.8214285969734192, | |
| "eval_rewards/chosen": 0.6988765597343445, | |
| "eval_rewards/margins": 0.5760089755058289, | |
| "eval_rewards/rejected": 0.12286762148141861, | |
| "eval_runtime": 19.4173, | |
| "eval_samples_per_second": 1.442, | |
| "eval_steps_per_second": 1.442, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 12.95, | |
| "grad_norm": 5.450782299041748, | |
| "learning_rate": 4.947309603078138e-07, | |
| "logits/chosen": -2.6719353199005127, | |
| "logits/rejected": -2.7134647369384766, | |
| "logps/chosen": -124.22660827636719, | |
| "logps/rejected": -99.73461151123047, | |
| "loss": 0.4062, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.7935754060745239, | |
| "rewards/margins": 0.783437192440033, | |
| "rewards/rejected": 0.010138224810361862, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 13.21, | |
| "grad_norm": 5.752992153167725, | |
| "learning_rate": 4.944976716867984e-07, | |
| "logits/chosen": -2.7026755809783936, | |
| "logits/rejected": -2.698920726776123, | |
| "logps/chosen": -123.63475036621094, | |
| "logps/rejected": -107.06332397460938, | |
| "loss": 0.3952, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.8116875886917114, | |
| "rewards/margins": 0.8335397839546204, | |
| "rewards/rejected": -0.02185220457613468, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 13.46, | |
| "grad_norm": 5.456747531890869, | |
| "learning_rate": 4.942593872763566e-07, | |
| "logits/chosen": -2.701968193054199, | |
| "logits/rejected": -2.786785840988159, | |
| "logps/chosen": -124.91243743896484, | |
| "logps/rejected": -101.95451354980469, | |
| "loss": 0.3673, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.8831655979156494, | |
| "rewards/margins": 0.9259040951728821, | |
| "rewards/rejected": -0.04273851588368416, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 13.71, | |
| "grad_norm": 4.753533363342285, | |
| "learning_rate": 4.940161119453231e-07, | |
| "logits/chosen": -2.7109193801879883, | |
| "logits/rejected": -2.6961448192596436, | |
| "logps/chosen": -123.21333312988281, | |
| "logps/rejected": -92.74653625488281, | |
| "loss": 0.3832, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.8407853245735168, | |
| "rewards/margins": 0.8698095083236694, | |
| "rewards/rejected": -0.02902432158589363, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 13.97, | |
| "grad_norm": 5.374657154083252, | |
| "learning_rate": 4.937678506645115e-07, | |
| "logits/chosen": -2.7288832664489746, | |
| "logits/rejected": -2.7349352836608887, | |
| "logps/chosen": -123.1498031616211, | |
| "logps/rejected": -95.77909851074219, | |
| "loss": 0.3644, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.8925550580024719, | |
| "rewards/margins": 0.9183064699172974, | |
| "rewards/rejected": -0.025751333683729172, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 14.22, | |
| "grad_norm": 5.842701435089111, | |
| "learning_rate": 4.935146085066125e-07, | |
| "logits/chosen": -2.7030577659606934, | |
| "logits/rejected": -2.7200875282287598, | |
| "logps/chosen": -120.81951904296875, | |
| "logps/rejected": -100.53494262695312, | |
| "loss": 0.3581, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.9156714677810669, | |
| "rewards/margins": 0.9704564809799194, | |
| "rewards/rejected": -0.05478499457240105, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 14.48, | |
| "grad_norm": 5.020945072174072, | |
| "learning_rate": 4.932563906460904e-07, | |
| "logits/chosen": -2.722069501876831, | |
| "logits/rejected": -2.731358051300049, | |
| "logps/chosen": -123.65387725830078, | |
| "logps/rejected": -103.83390808105469, | |
| "loss": 0.391, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.8787897229194641, | |
| "rewards/margins": 0.8743190765380859, | |
| "rewards/rejected": 0.004470685496926308, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 14.73, | |
| "grad_norm": 5.041106700897217, | |
| "learning_rate": 4.929932023590776e-07, | |
| "logits/chosen": -2.735006093978882, | |
| "logits/rejected": -2.7546191215515137, | |
| "logps/chosen": -124.65998840332031, | |
| "logps/rejected": -101.49640655517578, | |
| "loss": 0.3586, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.9206160306930542, | |
| "rewards/margins": 0.9452594518661499, | |
| "rewards/rejected": -0.024643439799547195, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 14.98, | |
| "grad_norm": 5.318466663360596, | |
| "learning_rate": 4.927250490232664e-07, | |
| "logits/chosen": -2.6883015632629395, | |
| "logits/rejected": -2.718212842941284, | |
| "logps/chosen": -120.88870239257812, | |
| "logps/rejected": -94.79769134521484, | |
| "loss": 0.3564, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.9339860677719116, | |
| "rewards/margins": 0.9901734590530396, | |
| "rewards/rejected": -0.05618742108345032, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 15.24, | |
| "grad_norm": 5.438621997833252, | |
| "learning_rate": 4.924519361177992e-07, | |
| "logits/chosen": -2.7316794395446777, | |
| "logits/rejected": -2.812800407409668, | |
| "logps/chosen": -120.06542205810547, | |
| "logps/rejected": -97.61270141601562, | |
| "loss": 0.3471, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 0.9851706624031067, | |
| "rewards/margins": 1.039989948272705, | |
| "rewards/rejected": -0.05481935292482376, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 15.24, | |
| "eval_logits/chosen": -2.8022568225860596, | |
| "eval_logits/rejected": -2.882326602935791, | |
| "eval_logps/chosen": -125.87542724609375, | |
| "eval_logps/rejected": -96.83832550048828, | |
| "eval_loss": 0.46029001474380493, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 0.8193913102149963, | |
| "eval_rewards/margins": 0.6923348307609558, | |
| "eval_rewards/rejected": 0.12705650925636292, | |
| "eval_runtime": 19.4065, | |
| "eval_samples_per_second": 1.443, | |
| "eval_steps_per_second": 1.443, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 15.49, | |
| "grad_norm": 5.301855087280273, | |
| "learning_rate": 4.921738692231572e-07, | |
| "logits/chosen": -2.661013126373291, | |
| "logits/rejected": -2.666978359222412, | |
| "logps/chosen": -121.69336700439453, | |
| "logps/rejected": -101.86541748046875, | |
| "loss": 0.3574, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 0.9682250022888184, | |
| "rewards/margins": 0.9734476208686829, | |
| "rewards/rejected": -0.0052225226536393166, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 15.75, | |
| "grad_norm": 4.478713512420654, | |
| "learning_rate": 4.918908540210451e-07, | |
| "logits/chosen": -2.702775239944458, | |
| "logits/rejected": -2.7098357677459717, | |
| "logps/chosen": -119.93941497802734, | |
| "logps/rejected": -94.76625061035156, | |
| "loss": 0.3517, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.8942417502403259, | |
| "rewards/margins": 0.9958410263061523, | |
| "rewards/rejected": -0.10159924626350403, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 4.398434162139893, | |
| "learning_rate": 4.916028962942762e-07, | |
| "logits/chosen": -2.7399990558624268, | |
| "logits/rejected": -2.7476425170898438, | |
| "logps/chosen": -124.2032470703125, | |
| "logps/rejected": -101.86943054199219, | |
| "loss": 0.3372, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.0369476079940796, | |
| "rewards/margins": 1.0664609670639038, | |
| "rewards/rejected": -0.029513416811823845, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 16.25, | |
| "grad_norm": 4.430797576904297, | |
| "learning_rate": 4.913100019266536e-07, | |
| "logits/chosen": -2.6950950622558594, | |
| "logits/rejected": -2.7181689739227295, | |
| "logps/chosen": -123.97085571289062, | |
| "logps/rejected": -97.85747528076172, | |
| "loss": 0.3226, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.0282835960388184, | |
| "rewards/margins": 1.1182292699813843, | |
| "rewards/rejected": -0.08994561433792114, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 16.51, | |
| "grad_norm": 6.383476734161377, | |
| "learning_rate": 4.910121769028503e-07, | |
| "logits/chosen": -2.7224864959716797, | |
| "logits/rejected": -2.727613925933838, | |
| "logps/chosen": -117.91517639160156, | |
| "logps/rejected": -99.52539825439453, | |
| "loss": 0.3658, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 0.9342272877693176, | |
| "rewards/margins": 0.9777376651763916, | |
| "rewards/rejected": -0.04351034015417099, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 16.76, | |
| "grad_norm": 4.453210353851318, | |
| "learning_rate": 4.907094273082864e-07, | |
| "logits/chosen": -2.7149100303649902, | |
| "logits/rejected": -2.710146188735962, | |
| "logps/chosen": -126.07495880126953, | |
| "logps/rejected": -100.1856689453125, | |
| "loss": 0.3162, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 1.1231064796447754, | |
| "rewards/margins": 1.175417423248291, | |
| "rewards/rejected": -0.05231098458170891, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 17.02, | |
| "grad_norm": 4.718712329864502, | |
| "learning_rate": 4.904017593290055e-07, | |
| "logits/chosen": -2.710063934326172, | |
| "logits/rejected": -2.7535319328308105, | |
| "logps/chosen": -117.01451873779297, | |
| "logps/rejected": -99.95758819580078, | |
| "loss": 0.322, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.033459186553955, | |
| "rewards/margins": 1.1137871742248535, | |
| "rewards/rejected": -0.08032786101102829, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 17.27, | |
| "grad_norm": 5.164140701293945, | |
| "learning_rate": 4.900891792515478e-07, | |
| "logits/chosen": -2.7342586517333984, | |
| "logits/rejected": -2.7561776638031006, | |
| "logps/chosen": -118.80335998535156, | |
| "logps/rejected": -104.33666229248047, | |
| "loss": 0.3314, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.078546404838562, | |
| "rewards/margins": 1.0902469158172607, | |
| "rewards/rejected": -0.011700518429279327, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 17.52, | |
| "grad_norm": 4.392764568328857, | |
| "learning_rate": 4.897716934628217e-07, | |
| "logits/chosen": -2.6796340942382812, | |
| "logits/rejected": -2.6666793823242188, | |
| "logps/chosen": -122.01671600341797, | |
| "logps/rejected": -97.74385070800781, | |
| "loss": 0.3047, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.0594760179519653, | |
| "rewards/margins": 1.1904582977294922, | |
| "rewards/rejected": -0.1309822052717209, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 17.78, | |
| "grad_norm": 4.647542953491211, | |
| "learning_rate": 4.894493084499736e-07, | |
| "logits/chosen": -2.7203075885772705, | |
| "logits/rejected": -2.7299890518188477, | |
| "logps/chosen": -120.08210754394531, | |
| "logps/rejected": -98.8072738647461, | |
| "loss": 0.3496, | |
| "rewards/accuracies": 0.921875, | |
| "rewards/chosen": 1.0658477544784546, | |
| "rewards/margins": 1.065575122833252, | |
| "rewards/rejected": 0.0002726661041378975, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 17.78, | |
| "eval_logits/chosen": -2.798058271408081, | |
| "eval_logits/rejected": -2.8771989345550537, | |
| "eval_logps/chosen": -124.81001281738281, | |
| "eval_logps/rejected": -96.83968353271484, | |
| "eval_loss": 0.44096532464027405, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 0.9259332418441772, | |
| "eval_rewards/margins": 0.799013078212738, | |
| "eval_rewards/rejected": 0.12692023813724518, | |
| "eval_runtime": 19.3925, | |
| "eval_samples_per_second": 1.444, | |
| "eval_steps_per_second": 1.444, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 18.03, | |
| "grad_norm": 4.158952236175537, | |
| "learning_rate": 4.891220308002547e-07, | |
| "logits/chosen": -2.6795365810394287, | |
| "logits/rejected": -2.7270519733428955, | |
| "logps/chosen": -120.39765167236328, | |
| "logps/rejected": -100.02141571044922, | |
| "loss": 0.319, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 1.0584629774093628, | |
| "rewards/margins": 1.19419264793396, | |
| "rewards/rejected": -0.13572952151298523, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 18.29, | |
| "grad_norm": 4.605132579803467, | |
| "learning_rate": 4.887898672008871e-07, | |
| "logits/chosen": -2.696953296661377, | |
| "logits/rejected": -2.718316078186035, | |
| "logps/chosen": -120.32958984375, | |
| "logps/rejected": -101.76454162597656, | |
| "loss": 0.331, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 1.0935906171798706, | |
| "rewards/margins": 1.1072776317596436, | |
| "rewards/rejected": -0.013687020167708397, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 18.54, | |
| "grad_norm": 5.824638843536377, | |
| "learning_rate": 4.884528244389268e-07, | |
| "logits/chosen": -2.6889617443084717, | |
| "logits/rejected": -2.690692663192749, | |
| "logps/chosen": -120.37040710449219, | |
| "logps/rejected": -98.27632141113281, | |
| "loss": 0.3204, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 1.1049582958221436, | |
| "rewards/margins": 1.175376057624817, | |
| "rewards/rejected": -0.07041767239570618, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 18.79, | |
| "grad_norm": 5.431331634521484, | |
| "learning_rate": 4.881109094011253e-07, | |
| "logits/chosen": -2.740906238555908, | |
| "logits/rejected": -2.725914478302002, | |
| "logps/chosen": -124.35327911376953, | |
| "logps/rejected": -100.71537780761719, | |
| "loss": 0.2884, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.2057325839996338, | |
| "rewards/margins": 1.3113868236541748, | |
| "rewards/rejected": -0.10565409809350967, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 19.05, | |
| "grad_norm": 4.33558988571167, | |
| "learning_rate": 4.877641290737883e-07, | |
| "logits/chosen": -2.700420379638672, | |
| "logits/rejected": -2.7527971267700195, | |
| "logps/chosen": -117.01264953613281, | |
| "logps/rejected": -101.50834655761719, | |
| "loss": 0.2757, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.150115728378296, | |
| "rewards/margins": 1.34402596950531, | |
| "rewards/rejected": -0.19391010701656342, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 19.3, | |
| "grad_norm": 4.987298488616943, | |
| "learning_rate": 4.874124905426338e-07, | |
| "logits/chosen": -2.715477705001831, | |
| "logits/rejected": -2.760146141052246, | |
| "logps/chosen": -117.20336151123047, | |
| "logps/rejected": -94.89369201660156, | |
| "loss": 0.3244, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 1.107708215713501, | |
| "rewards/margins": 1.1569105386734009, | |
| "rewards/rejected": -0.04920227825641632, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 19.56, | |
| "grad_norm": 4.090747833251953, | |
| "learning_rate": 4.870560009926464e-07, | |
| "logits/chosen": -2.6785225868225098, | |
| "logits/rejected": -2.6984167098999023, | |
| "logps/chosen": -112.27914428710938, | |
| "logps/rejected": -94.59768676757812, | |
| "loss": 0.2833, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.2028326988220215, | |
| "rewards/margins": 1.334948182106018, | |
| "rewards/rejected": -0.13211557269096375, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 19.81, | |
| "grad_norm": 5.0767083168029785, | |
| "learning_rate": 4.866946677079314e-07, | |
| "logits/chosen": -2.712416410446167, | |
| "logits/rejected": -2.7296035289764404, | |
| "logps/chosen": -121.65861511230469, | |
| "logps/rejected": -106.11611938476562, | |
| "loss": 0.2893, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.1842504739761353, | |
| "rewards/margins": 1.3191744089126587, | |
| "rewards/rejected": -0.13492396473884583, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 20.06, | |
| "grad_norm": 4.1403326988220215, | |
| "learning_rate": 4.863284980715648e-07, | |
| "logits/chosen": -2.658827781677246, | |
| "logits/rejected": -2.631199836730957, | |
| "logps/chosen": -129.3588409423828, | |
| "logps/rejected": -103.25465393066406, | |
| "loss": 0.2678, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.291969895362854, | |
| "rewards/margins": 1.4036282300949097, | |
| "rewards/rejected": -0.11165820807218552, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 20.32, | |
| "grad_norm": 4.350497245788574, | |
| "learning_rate": 4.859574995654441e-07, | |
| "logits/chosen": -2.761063814163208, | |
| "logits/rejected": -2.7751922607421875, | |
| "logps/chosen": -114.76087951660156, | |
| "logps/rejected": -100.04124450683594, | |
| "loss": 0.298, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.1642746925354004, | |
| "rewards/margins": 1.3020979166030884, | |
| "rewards/rejected": -0.13782335817813873, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 20.32, | |
| "eval_logits/chosen": -2.792503595352173, | |
| "eval_logits/rejected": -2.870570421218872, | |
| "eval_logps/chosen": -123.9360122680664, | |
| "eval_logps/rejected": -96.95860290527344, | |
| "eval_loss": 0.4256473481655121, | |
| "eval_rewards/accuracies": 0.8214285969734192, | |
| "eval_rewards/chosen": 1.0133329629898071, | |
| "eval_rewards/margins": 0.8983038663864136, | |
| "eval_rewards/rejected": 0.11502902954816818, | |
| "eval_runtime": 19.3701, | |
| "eval_samples_per_second": 1.446, | |
| "eval_steps_per_second": 1.446, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 20.57, | |
| "grad_norm": 4.334249973297119, | |
| "learning_rate": 4.855816797701336e-07, | |
| "logits/chosen": -2.690117359161377, | |
| "logits/rejected": -2.7377476692199707, | |
| "logps/chosen": -116.01565551757812, | |
| "logps/rejected": -102.1231689453125, | |
| "loss": 0.2872, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 1.2394335269927979, | |
| "rewards/margins": 1.3586478233337402, | |
| "rewards/rejected": -0.11921422928571701, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 20.83, | |
| "grad_norm": 4.597221851348877, | |
| "learning_rate": 4.85201046364711e-07, | |
| "logits/chosen": -2.741600751876831, | |
| "logits/rejected": -2.712954044342041, | |
| "logps/chosen": -124.32752990722656, | |
| "logps/rejected": -101.2073974609375, | |
| "loss": 0.2676, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.22231924533844, | |
| "rewards/margins": 1.4293614625930786, | |
| "rewards/rejected": -0.20704227685928345, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 21.08, | |
| "grad_norm": 3.682258367538452, | |
| "learning_rate": 4.848156071266095e-07, | |
| "logits/chosen": -2.6585566997528076, | |
| "logits/rejected": -2.704587459564209, | |
| "logps/chosen": -120.91087341308594, | |
| "logps/rejected": -98.83493041992188, | |
| "loss": 0.2705, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.2277172803878784, | |
| "rewards/margins": 1.3790152072906494, | |
| "rewards/rejected": -0.15129785239696503, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 21.33, | |
| "grad_norm": 3.7775039672851562, | |
| "learning_rate": 4.844253699314596e-07, | |
| "logits/chosen": -2.7104837894439697, | |
| "logits/rejected": -2.7378451824188232, | |
| "logps/chosen": -117.22598266601562, | |
| "logps/rejected": -97.59563446044922, | |
| "loss": 0.2755, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.2548604011535645, | |
| "rewards/margins": 1.3694878816604614, | |
| "rewards/rejected": -0.11462761461734772, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 21.59, | |
| "grad_norm": 4.84611701965332, | |
| "learning_rate": 4.840303427529273e-07, | |
| "logits/chosen": -2.6837968826293945, | |
| "logits/rejected": -2.7071595191955566, | |
| "logps/chosen": -121.88581085205078, | |
| "logps/rejected": -102.45454406738281, | |
| "loss": 0.286, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 1.2159500122070312, | |
| "rewards/margins": 1.3315218687057495, | |
| "rewards/rejected": -0.11557181179523468, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 21.84, | |
| "grad_norm": 3.760068655014038, | |
| "learning_rate": 4.836305336625523e-07, | |
| "logits/chosen": -2.709066390991211, | |
| "logits/rejected": -2.73175048828125, | |
| "logps/chosen": -117.1901626586914, | |
| "logps/rejected": -105.62377166748047, | |
| "loss": 0.2661, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.3225661516189575, | |
| "rewards/margins": 1.445369005203247, | |
| "rewards/rejected": -0.12280277907848358, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 22.1, | |
| "grad_norm": 3.8488616943359375, | |
| "learning_rate": 4.832259508295822e-07, | |
| "logits/chosen": -2.7598538398742676, | |
| "logits/rejected": -2.776313066482544, | |
| "logps/chosen": -117.15644836425781, | |
| "logps/rejected": -93.95286560058594, | |
| "loss": 0.2711, | |
| "rewards/accuracies": 0.953125, | |
| "rewards/chosen": 1.3003535270690918, | |
| "rewards/margins": 1.5021978616714478, | |
| "rewards/rejected": -0.20184442400932312, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 22.35, | |
| "grad_norm": 3.152104377746582, | |
| "learning_rate": 4.828166025208058e-07, | |
| "logits/chosen": -2.6503114700317383, | |
| "logits/rejected": -2.6512064933776855, | |
| "logps/chosen": -118.29179382324219, | |
| "logps/rejected": -97.83879852294922, | |
| "loss": 0.2394, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.2750773429870605, | |
| "rewards/margins": 1.585729718208313, | |
| "rewards/rejected": -0.31065240502357483, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 22.6, | |
| "grad_norm": 4.1737165451049805, | |
| "learning_rate": 4.824024971003845e-07, | |
| "logits/chosen": -2.708979845046997, | |
| "logits/rejected": -2.718264579772949, | |
| "logps/chosen": -113.75922393798828, | |
| "logps/rejected": -100.5313720703125, | |
| "loss": 0.2626, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.3227206468582153, | |
| "rewards/margins": 1.4637939929962158, | |
| "rewards/rejected": -0.1410733461380005, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 22.86, | |
| "grad_norm": 4.2186737060546875, | |
| "learning_rate": 4.819836430296808e-07, | |
| "logits/chosen": -2.688643217086792, | |
| "logits/rejected": -2.7126376628875732, | |
| "logps/chosen": -119.3924560546875, | |
| "logps/rejected": -102.88687896728516, | |
| "loss": 0.272, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.3319997787475586, | |
| "rewards/margins": 1.3775808811187744, | |
| "rewards/rejected": -0.04558102786540985, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 22.86, | |
| "eval_logits/chosen": -2.7863781452178955, | |
| "eval_logits/rejected": -2.8632652759552, | |
| "eval_logps/chosen": -123.22322082519531, | |
| "eval_logps/rejected": -97.17105865478516, | |
| "eval_loss": 0.4147447645664215, | |
| "eval_rewards/accuracies": 0.8214285969734192, | |
| "eval_rewards/chosen": 1.0846121311187744, | |
| "eval_rewards/margins": 0.9908290505409241, | |
| "eval_rewards/rejected": 0.09378316253423691, | |
| "eval_runtime": 19.3953, | |
| "eval_samples_per_second": 1.444, | |
| "eval_steps_per_second": 1.444, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 23.11, | |
| "grad_norm": 3.8801398277282715, | |
| "learning_rate": 4.815600488670862e-07, | |
| "logits/chosen": -2.661937713623047, | |
| "logits/rejected": -2.69406795501709, | |
| "logps/chosen": -120.42941284179688, | |
| "logps/rejected": -107.56249237060547, | |
| "loss": 0.2463, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.303568959236145, | |
| "rewards/margins": 1.5796161890029907, | |
| "rewards/rejected": -0.2760472595691681, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 23.37, | |
| "grad_norm": 3.6953530311584473, | |
| "learning_rate": 4.811317232678456e-07, | |
| "logits/chosen": -2.696389675140381, | |
| "logits/rejected": -2.7162017822265625, | |
| "logps/chosen": -115.95695495605469, | |
| "logps/rejected": -95.26593017578125, | |
| "loss": 0.2557, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.3928656578063965, | |
| "rewards/margins": 1.5144150257110596, | |
| "rewards/rejected": -0.12154923379421234, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 23.62, | |
| "grad_norm": 4.2311553955078125, | |
| "learning_rate": 4.806986749838806e-07, | |
| "logits/chosen": -2.735870599746704, | |
| "logits/rejected": -2.7359132766723633, | |
| "logps/chosen": -120.84663391113281, | |
| "logps/rejected": -107.17831420898438, | |
| "loss": 0.2922, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.2136850357055664, | |
| "rewards/margins": 1.3251327276229858, | |
| "rewards/rejected": -0.11144763976335526, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 23.87, | |
| "grad_norm": 3.250326633453369, | |
| "learning_rate": 4.802609128636112e-07, | |
| "logits/chosen": -2.6809797286987305, | |
| "logits/rejected": -2.700901508331299, | |
| "logps/chosen": -121.92070007324219, | |
| "logps/rejected": -99.91931915283203, | |
| "loss": 0.2062, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.4374094009399414, | |
| "rewards/margins": 1.8276853561401367, | |
| "rewards/rejected": -0.3902759850025177, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 24.13, | |
| "grad_norm": 4.59958553314209, | |
| "learning_rate": 4.798184458517744e-07, | |
| "logits/chosen": -2.7332396507263184, | |
| "logits/rejected": -2.768554210662842, | |
| "logps/chosen": -114.1657943725586, | |
| "logps/rejected": -100.7848892211914, | |
| "loss": 0.2353, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.4016469717025757, | |
| "rewards/margins": 1.5988764762878418, | |
| "rewards/rejected": -0.19722968339920044, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 24.38, | |
| "grad_norm": 3.6598093509674072, | |
| "learning_rate": 4.793712829892414e-07, | |
| "logits/chosen": -2.7024924755096436, | |
| "logits/rejected": -2.7085440158843994, | |
| "logps/chosen": -114.50930786132812, | |
| "logps/rejected": -104.57618713378906, | |
| "loss": 0.2538, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.4238942861557007, | |
| "rewards/margins": 1.5259473323822021, | |
| "rewards/rejected": -0.10205309092998505, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 24.63, | |
| "grad_norm": 3.3857998847961426, | |
| "learning_rate": 4.789194334128338e-07, | |
| "logits/chosen": -2.7139151096343994, | |
| "logits/rejected": -2.7087197303771973, | |
| "logps/chosen": -121.57815551757812, | |
| "logps/rejected": -102.66044616699219, | |
| "loss": 0.2324, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.3619202375411987, | |
| "rewards/margins": 1.667473316192627, | |
| "rewards/rejected": -0.305553138256073, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 24.89, | |
| "grad_norm": 4.038039207458496, | |
| "learning_rate": 4.784629063551354e-07, | |
| "logits/chosen": -2.651282548904419, | |
| "logits/rejected": -2.679689884185791, | |
| "logps/chosen": -118.9875717163086, | |
| "logps/rejected": -97.62678527832031, | |
| "loss": 0.2304, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 1.3316906690597534, | |
| "rewards/margins": 1.7062404155731201, | |
| "rewards/rejected": -0.3745496869087219, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 25.14, | |
| "grad_norm": 3.0911149978637695, | |
| "learning_rate": 4.780017111443047e-07, | |
| "logits/chosen": -2.718074083328247, | |
| "logits/rejected": -2.7447762489318848, | |
| "logps/chosen": -121.20558166503906, | |
| "logps/rejected": -105.63926696777344, | |
| "loss": 0.2411, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.4981815814971924, | |
| "rewards/margins": 1.644660472869873, | |
| "rewards/rejected": -0.14647886157035828, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 25.4, | |
| "grad_norm": 3.236288070678711, | |
| "learning_rate": 4.775358572038845e-07, | |
| "logits/chosen": -2.693341016769409, | |
| "logits/rejected": -2.681511878967285, | |
| "logps/chosen": -114.28794860839844, | |
| "logps/rejected": -91.69850158691406, | |
| "loss": 0.259, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.3284053802490234, | |
| "rewards/margins": 1.5037521123886108, | |
| "rewards/rejected": -0.17534679174423218, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 25.4, | |
| "eval_logits/chosen": -2.7796428203582764, | |
| "eval_logits/rejected": -2.8552372455596924, | |
| "eval_logps/chosen": -122.60997009277344, | |
| "eval_logps/rejected": -97.46685028076172, | |
| "eval_loss": 0.4054030478000641, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.1459368467330933, | |
| "eval_rewards/margins": 1.0817322731018066, | |
| "eval_rewards/rejected": 0.06420455127954483, | |
| "eval_runtime": 19.3983, | |
| "eval_samples_per_second": 1.443, | |
| "eval_steps_per_second": 1.443, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 25.65, | |
| "grad_norm": 3.2894155979156494, | |
| "learning_rate": 4.770653540526078e-07, | |
| "logits/chosen": -2.652012825012207, | |
| "logits/rejected": -2.690704345703125, | |
| "logps/chosen": -118.52864837646484, | |
| "logps/rejected": -102.0433578491211, | |
| "loss": 0.1918, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5184407234191895, | |
| "rewards/margins": 1.8730953931808472, | |
| "rewards/rejected": -0.3546547293663025, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 25.9, | |
| "grad_norm": 4.958095550537109, | |
| "learning_rate": 4.765902113042053e-07, | |
| "logits/chosen": -2.7191715240478516, | |
| "logits/rejected": -2.701702833175659, | |
| "logps/chosen": -115.93360900878906, | |
| "logps/rejected": -100.49573516845703, | |
| "loss": 0.2536, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.357507348060608, | |
| "rewards/margins": 1.542129635810852, | |
| "rewards/rejected": -0.18462230265140533, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 26.16, | |
| "grad_norm": 3.1129214763641357, | |
| "learning_rate": 4.7611043866720737e-07, | |
| "logits/chosen": -2.6934893131256104, | |
| "logits/rejected": -2.7433109283447266, | |
| "logps/chosen": -117.78741455078125, | |
| "logps/rejected": -102.46746063232422, | |
| "loss": 0.2137, | |
| "rewards/accuracies": 0.96875, | |
| "rewards/chosen": 1.484704613685608, | |
| "rewards/margins": 1.8401193618774414, | |
| "rewards/rejected": -0.3554147481918335, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 26.41, | |
| "grad_norm": 3.2823047637939453, | |
| "learning_rate": 4.7562604594474643e-07, | |
| "logits/chosen": -2.6928577423095703, | |
| "logits/rejected": -2.708937644958496, | |
| "logps/chosen": -124.94667053222656, | |
| "logps/rejected": -106.61329650878906, | |
| "loss": 0.2137, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.471519112586975, | |
| "rewards/margins": 1.7852158546447754, | |
| "rewards/rejected": -0.31369656324386597, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 26.67, | |
| "grad_norm": 3.7071120738983154, | |
| "learning_rate": 4.7513704303435674e-07, | |
| "logits/chosen": -2.670248031616211, | |
| "logits/rejected": -2.6764914989471436, | |
| "logps/chosen": -110.71532440185547, | |
| "logps/rejected": -101.84510803222656, | |
| "loss": 0.229, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.4525586366653442, | |
| "rewards/margins": 1.6690635681152344, | |
| "rewards/rejected": -0.21650481224060059, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 26.92, | |
| "grad_norm": 3.449319362640381, | |
| "learning_rate": 4.7464343992777165e-07, | |
| "logits/chosen": -2.688829183578491, | |
| "logits/rejected": -2.7286489009857178, | |
| "logps/chosen": -115.69140625, | |
| "logps/rejected": -103.97221374511719, | |
| "loss": 0.2231, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.4910831451416016, | |
| "rewards/margins": 1.7472798824310303, | |
| "rewards/rejected": -0.2561967670917511, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 27.17, | |
| "grad_norm": 2.4958574771881104, | |
| "learning_rate": 4.741452467107199e-07, | |
| "logits/chosen": -2.6915664672851562, | |
| "logits/rejected": -2.7021422386169434, | |
| "logps/chosen": -110.91499328613281, | |
| "logps/rejected": -98.70269775390625, | |
| "loss": 0.233, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.4354866743087769, | |
| "rewards/margins": 1.7168078422546387, | |
| "rewards/rejected": -0.28132107853889465, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 27.43, | |
| "grad_norm": 3.7154319286346436, | |
| "learning_rate": 4.7364247356271927e-07, | |
| "logits/chosen": -2.6695199012756348, | |
| "logits/rejected": -2.6698789596557617, | |
| "logps/chosen": -118.46913146972656, | |
| "logps/rejected": -102.86263275146484, | |
| "loss": 0.2337, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.4410640001296997, | |
| "rewards/margins": 1.6770535707473755, | |
| "rewards/rejected": -0.23598948121070862, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 27.68, | |
| "grad_norm": 3.101527690887451, | |
| "learning_rate": 4.731351307568687e-07, | |
| "logits/chosen": -2.6910083293914795, | |
| "logits/rejected": -2.6801414489746094, | |
| "logps/chosen": -115.44275665283203, | |
| "logps/rejected": -97.7877197265625, | |
| "loss": 0.2, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5193870067596436, | |
| "rewards/margins": 1.8750090599060059, | |
| "rewards/rejected": -0.35562214255332947, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 27.94, | |
| "grad_norm": 3.9017672538757324, | |
| "learning_rate": 4.7262322865963846e-07, | |
| "logits/chosen": -2.7122743129730225, | |
| "logits/rejected": -2.723294258117676, | |
| "logps/chosen": -116.90724182128906, | |
| "logps/rejected": -104.45236206054688, | |
| "loss": 0.1959, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5187357664108276, | |
| "rewards/margins": 1.918578863143921, | |
| "rewards/rejected": -0.39984312653541565, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 27.94, | |
| "eval_logits/chosen": -2.7724335193634033, | |
| "eval_logits/rejected": -2.8468399047851562, | |
| "eval_logps/chosen": -122.11509704589844, | |
| "eval_logps/rejected": -97.87715148925781, | |
| "eval_loss": 0.39711207151412964, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.1954243183135986, | |
| "eval_rewards/margins": 1.1722500324249268, | |
| "eval_rewards/rejected": 0.023174213245511055, | |
| "eval_runtime": 19.4446, | |
| "eval_samples_per_second": 1.44, | |
| "eval_steps_per_second": 1.44, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 28.19, | |
| "grad_norm": 4.0338592529296875, | |
| "learning_rate": 4.721067777306581e-07, | |
| "logits/chosen": -2.6722593307495117, | |
| "logits/rejected": -2.682727575302124, | |
| "logps/chosen": -119.6982650756836, | |
| "logps/rejected": -105.52352905273438, | |
| "loss": 0.2108, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.5048611164093018, | |
| "rewards/margins": 1.8753631114959717, | |
| "rewards/rejected": -0.3705020546913147, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 28.44, | |
| "grad_norm": 3.167048692703247, | |
| "learning_rate": 4.71585788522503e-07, | |
| "logits/chosen": -2.652644157409668, | |
| "logits/rejected": -2.6818065643310547, | |
| "logps/chosen": -113.42478942871094, | |
| "logps/rejected": -104.37007141113281, | |
| "loss": 0.2076, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.474198341369629, | |
| "rewards/margins": 1.7973999977111816, | |
| "rewards/rejected": -0.32320165634155273, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 28.7, | |
| "grad_norm": 3.4248368740081787, | |
| "learning_rate": 4.7106027168047833e-07, | |
| "logits/chosen": -2.7117884159088135, | |
| "logits/rejected": -2.77234148979187, | |
| "logps/chosen": -116.26769256591797, | |
| "logps/rejected": -101.22008514404297, | |
| "loss": 0.1858, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5492609739303589, | |
| "rewards/margins": 1.9960644245147705, | |
| "rewards/rejected": -0.4468035101890564, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 28.95, | |
| "grad_norm": 2.855328321456909, | |
| "learning_rate": 4.705302379424022e-07, | |
| "logits/chosen": -2.6717069149017334, | |
| "logits/rejected": -2.672755718231201, | |
| "logps/chosen": -116.75204467773438, | |
| "logps/rejected": -106.22608184814453, | |
| "loss": 0.2065, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5143406391143799, | |
| "rewards/margins": 1.8520642518997192, | |
| "rewards/rejected": -0.33772367238998413, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 29.21, | |
| "grad_norm": 2.8747594356536865, | |
| "learning_rate": 4.699956981383857e-07, | |
| "logits/chosen": -2.693591356277466, | |
| "logits/rejected": -2.662400960922241, | |
| "logps/chosen": -120.1160888671875, | |
| "logps/rejected": -91.72486877441406, | |
| "loss": 0.1874, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.6478428840637207, | |
| "rewards/margins": 2.0628483295440674, | |
| "rewards/rejected": -0.4150054454803467, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 29.46, | |
| "grad_norm": 2.8170723915100098, | |
| "learning_rate": 4.694566631906116e-07, | |
| "logits/chosen": -2.6773173809051514, | |
| "logits/rejected": -2.672825574874878, | |
| "logps/chosen": -113.59075927734375, | |
| "logps/rejected": -103.25709533691406, | |
| "loss": 0.1984, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.4863029718399048, | |
| "rewards/margins": 1.892244577407837, | |
| "rewards/rejected": -0.405941367149353, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 29.71, | |
| "grad_norm": 3.132080078125, | |
| "learning_rate": 4.689131441131118e-07, | |
| "logits/chosen": -2.6520447731018066, | |
| "logits/rejected": -2.6872832775115967, | |
| "logps/chosen": -118.07131958007812, | |
| "logps/rejected": -107.88346862792969, | |
| "loss": 0.2029, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.6008590459823608, | |
| "rewards/margins": 1.909461259841919, | |
| "rewards/rejected": -0.30860215425491333, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 29.97, | |
| "grad_norm": 3.000974416732788, | |
| "learning_rate": 4.6836515201154135e-07, | |
| "logits/chosen": -2.709559202194214, | |
| "logits/rejected": -2.723712205886841, | |
| "logps/chosen": -112.09906768798828, | |
| "logps/rejected": -106.09660339355469, | |
| "loss": 0.2037, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5191339254379272, | |
| "rewards/margins": 1.8524271249771118, | |
| "rewards/rejected": -0.3332933187484741, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 30.22, | |
| "grad_norm": 2.544147491455078, | |
| "learning_rate": 4.678126980829524e-07, | |
| "logits/chosen": -2.678325653076172, | |
| "logits/rejected": -2.699833631515503, | |
| "logps/chosen": -115.4670639038086, | |
| "logps/rejected": -101.5027847290039, | |
| "loss": 0.1813, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.4956222772598267, | |
| "rewards/margins": 1.9856685400009155, | |
| "rewards/rejected": -0.4900462329387665, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 30.48, | |
| "grad_norm": 2.9770796298980713, | |
| "learning_rate": 4.672557936155649e-07, | |
| "logits/chosen": -2.708721160888672, | |
| "logits/rejected": -2.728306770324707, | |
| "logps/chosen": -109.71602630615234, | |
| "logps/rejected": -98.1273422241211, | |
| "loss": 0.1946, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.5404164791107178, | |
| "rewards/margins": 1.9511140584945679, | |
| "rewards/rejected": -0.4106977581977844, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 30.48, | |
| "eval_logits/chosen": -2.764019012451172, | |
| "eval_logits/rejected": -2.8370797634124756, | |
| "eval_logps/chosen": -121.74720001220703, | |
| "eval_logps/rejected": -98.30638122558594, | |
| "eval_loss": 0.393496572971344, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.232213020324707, | |
| "eval_rewards/margins": 1.2519627809524536, | |
| "eval_rewards/rejected": -0.019749613478779793, | |
| "eval_runtime": 19.4183, | |
| "eval_samples_per_second": 1.442, | |
| "eval_steps_per_second": 1.442, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 30.73, | |
| "grad_norm": 3.826716661453247, | |
| "learning_rate": 4.6669444998853605e-07, | |
| "logits/chosen": -2.6732892990112305, | |
| "logits/rejected": -2.700220823287964, | |
| "logps/chosen": -117.65553283691406, | |
| "logps/rejected": -111.19263458251953, | |
| "loss": 0.1876, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.6948100328445435, | |
| "rewards/margins": 2.0582737922668457, | |
| "rewards/rejected": -0.3634639382362366, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 30.98, | |
| "grad_norm": 3.326704740524292, | |
| "learning_rate": 4.661286786717278e-07, | |
| "logits/chosen": -2.6515111923217773, | |
| "logits/rejected": -2.6504664421081543, | |
| "logps/chosen": -120.06623840332031, | |
| "logps/rejected": -101.47470092773438, | |
| "loss": 0.1828, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5970648527145386, | |
| "rewards/margins": 2.0543787479400635, | |
| "rewards/rejected": -0.45731377601623535, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 31.24, | |
| "grad_norm": 2.700209856033325, | |
| "learning_rate": 4.6555849122547263e-07, | |
| "logits/chosen": -2.6469173431396484, | |
| "logits/rejected": -2.666424036026001, | |
| "logps/chosen": -111.76505279541016, | |
| "logps/rejected": -103.40081024169922, | |
| "loss": 0.185, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.524863839149475, | |
| "rewards/margins": 1.9786373376846313, | |
| "rewards/rejected": -0.4537736773490906, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 31.49, | |
| "grad_norm": 2.592606544494629, | |
| "learning_rate": 4.649838993003372e-07, | |
| "logits/chosen": -2.7192447185516357, | |
| "logits/rejected": -2.726841926574707, | |
| "logps/chosen": -112.72746276855469, | |
| "logps/rejected": -100.51376342773438, | |
| "loss": 0.1678, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.5847443342208862, | |
| "rewards/margins": 2.0699594020843506, | |
| "rewards/rejected": -0.4852150082588196, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 31.75, | |
| "grad_norm": 2.768076181411743, | |
| "learning_rate": 4.644049146368843e-07, | |
| "logits/chosen": -2.6624693870544434, | |
| "logits/rejected": -2.677476406097412, | |
| "logps/chosen": -119.13687896728516, | |
| "logps/rejected": -103.49778747558594, | |
| "loss": 0.1873, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7677032947540283, | |
| "rewards/margins": 2.089017868041992, | |
| "rewards/rejected": -0.32131490111351013, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "grad_norm": 2.4647607803344727, | |
| "learning_rate": 4.638215490654329e-07, | |
| "logits/chosen": -2.6946330070495605, | |
| "logits/rejected": -2.6913187503814697, | |
| "logps/chosen": -118.64775085449219, | |
| "logps/rejected": -107.40900421142578, | |
| "loss": 0.1847, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.5691322088241577, | |
| "rewards/margins": 2.105942964553833, | |
| "rewards/rejected": -0.5368106365203857, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 32.25, | |
| "grad_norm": 3.723209857940674, | |
| "learning_rate": 4.6323381450581665e-07, | |
| "logits/chosen": -2.6984715461730957, | |
| "logits/rejected": -2.686072826385498, | |
| "logps/chosen": -118.19576263427734, | |
| "logps/rejected": -104.87760162353516, | |
| "loss": 0.1853, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.504220724105835, | |
| "rewards/margins": 2.023249864578247, | |
| "rewards/rejected": -0.519028902053833, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 32.51, | |
| "grad_norm": 2.6810081005096436, | |
| "learning_rate": 4.6264172296714e-07, | |
| "logits/chosen": -2.6358213424682617, | |
| "logits/rejected": -2.6452157497406006, | |
| "logps/chosen": -116.47501373291016, | |
| "logps/rejected": -104.9029769897461, | |
| "loss": 0.1948, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.6027023792266846, | |
| "rewards/margins": 2.105417251586914, | |
| "rewards/rejected": -0.5027150511741638, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 32.76, | |
| "grad_norm": 2.1455435752868652, | |
| "learning_rate": 4.6204528654753304e-07, | |
| "logits/chosen": -2.692194700241089, | |
| "logits/rejected": -2.705289125442505, | |
| "logps/chosen": -113.38731384277344, | |
| "logps/rejected": -103.49237823486328, | |
| "loss": 0.1617, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7136329412460327, | |
| "rewards/margins": 2.1900761127471924, | |
| "rewards/rejected": -0.47644317150115967, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 33.02, | |
| "grad_norm": 3.499128818511963, | |
| "learning_rate": 4.614445174339045e-07, | |
| "logits/chosen": -2.661409616470337, | |
| "logits/rejected": -2.700990915298462, | |
| "logps/chosen": -113.71514892578125, | |
| "logps/rejected": -103.91769409179688, | |
| "loss": 0.1507, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7427312135696411, | |
| "rewards/margins": 2.1958351135253906, | |
| "rewards/rejected": -0.45310381054878235, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 33.02, | |
| "eval_logits/chosen": -2.755352020263672, | |
| "eval_logits/rejected": -2.826873540878296, | |
| "eval_logps/chosen": -121.47989654541016, | |
| "eval_logps/rejected": -98.8564453125, | |
| "eval_loss": 0.3899853527545929, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2589442729949951, | |
| "eval_rewards/margins": 1.333700180053711, | |
| "eval_rewards/rejected": -0.07475597411394119, | |
| "eval_runtime": 19.3899, | |
| "eval_samples_per_second": 1.444, | |
| "eval_steps_per_second": 1.444, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 33.27, | |
| "grad_norm": 2.419506549835205, | |
| "learning_rate": 4.6083942790169206e-07, | |
| "logits/chosen": -2.6735386848449707, | |
| "logits/rejected": -2.696384906768799, | |
| "logps/chosen": -111.79097747802734, | |
| "logps/rejected": -98.7368392944336, | |
| "loss": 0.1581, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.5749679803848267, | |
| "rewards/margins": 2.265718460083008, | |
| "rewards/rejected": -0.690750777721405, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 33.52, | |
| "grad_norm": 2.1046078205108643, | |
| "learning_rate": 4.6023003031461227e-07, | |
| "logits/chosen": -2.6465351581573486, | |
| "logits/rejected": -2.6622884273529053, | |
| "logps/chosen": -114.73282623291016, | |
| "logps/rejected": -103.12973022460938, | |
| "loss": 0.1778, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.6710339784622192, | |
| "rewards/margins": 2.119300365447998, | |
| "rewards/rejected": -0.4482663571834564, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 33.78, | |
| "grad_norm": 3.0457003116607666, | |
| "learning_rate": 4.596163371244076e-07, | |
| "logits/chosen": -2.6422927379608154, | |
| "logits/rejected": -2.6462604999542236, | |
| "logps/chosen": -114.708740234375, | |
| "logps/rejected": -105.07784271240234, | |
| "loss": 0.16, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7936890125274658, | |
| "rewards/margins": 2.2034826278686523, | |
| "rewards/rejected": -0.40979355573654175, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 34.03, | |
| "grad_norm": 2.8257133960723877, | |
| "learning_rate": 4.5899836087059174e-07, | |
| "logits/chosen": -2.6918063163757324, | |
| "logits/rejected": -2.696005344390869, | |
| "logps/chosen": -117.54481506347656, | |
| "logps/rejected": -111.74267578125, | |
| "loss": 0.1753, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.625847339630127, | |
| "rewards/margins": 2.2450764179229736, | |
| "rewards/rejected": -0.6192291975021362, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 34.29, | |
| "grad_norm": 2.5381155014038086, | |
| "learning_rate": 4.58376114180194e-07, | |
| "logits/chosen": -2.684847116470337, | |
| "logits/rejected": -2.6875247955322266, | |
| "logps/chosen": -121.1004867553711, | |
| "logps/rejected": -101.69570922851562, | |
| "loss": 0.1562, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.6668047904968262, | |
| "rewards/margins": 2.3632285594940186, | |
| "rewards/rejected": -0.6964237689971924, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 34.54, | |
| "grad_norm": 2.519352912902832, | |
| "learning_rate": 4.577496097675009e-07, | |
| "logits/chosen": -2.6664843559265137, | |
| "logits/rejected": -2.681034803390503, | |
| "logps/chosen": -115.94261169433594, | |
| "logps/rejected": -108.1679458618164, | |
| "loss": 0.1516, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7702218294143677, | |
| "rewards/margins": 2.2215099334716797, | |
| "rewards/rejected": -0.4512881338596344, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 34.79, | |
| "grad_norm": 2.789398431777954, | |
| "learning_rate": 4.571188604337962e-07, | |
| "logits/chosen": -2.668377637863159, | |
| "logits/rejected": -2.688797950744629, | |
| "logps/chosen": -108.2844467163086, | |
| "logps/rejected": -104.94053649902344, | |
| "loss": 0.1838, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.559061050415039, | |
| "rewards/margins": 2.064011812210083, | |
| "rewards/rejected": -0.504950761795044, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 35.05, | |
| "grad_norm": 3.570829153060913, | |
| "learning_rate": 4.5648387906709995e-07, | |
| "logits/chosen": -2.6559903621673584, | |
| "logits/rejected": -2.650508165359497, | |
| "logps/chosen": -112.94985961914062, | |
| "logps/rejected": -101.36627197265625, | |
| "loss": 0.1484, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.709464192390442, | |
| "rewards/margins": 2.303464889526367, | |
| "rewards/rejected": -0.5940006375312805, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 35.3, | |
| "grad_norm": 2.487671136856079, | |
| "learning_rate": 4.558446786419045e-07, | |
| "logits/chosen": -2.6454193592071533, | |
| "logits/rejected": -2.667191982269287, | |
| "logps/chosen": -112.91888427734375, | |
| "logps/rejected": -100.34318542480469, | |
| "loss": 0.1571, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.6041696071624756, | |
| "rewards/margins": 2.3108744621276855, | |
| "rewards/rejected": -0.7067050337791443, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 35.56, | |
| "grad_norm": 2.396897315979004, | |
| "learning_rate": 4.552012722189099e-07, | |
| "logits/chosen": -2.6444239616394043, | |
| "logits/rejected": -2.6464180946350098, | |
| "logps/chosen": -115.7559814453125, | |
| "logps/rejected": -111.63291931152344, | |
| "loss": 0.1752, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7378344535827637, | |
| "rewards/margins": 2.1783785820007324, | |
| "rewards/rejected": -0.4405441880226135, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 35.56, | |
| "eval_logits/chosen": -2.746135711669922, | |
| "eval_logits/rejected": -2.8161184787750244, | |
| "eval_logps/chosen": -121.27652740478516, | |
| "eval_logps/rejected": -99.45245361328125, | |
| "eval_loss": 0.38749733567237854, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2792812585830688, | |
| "eval_rewards/margins": 1.4136368036270142, | |
| "eval_rewards/rejected": -0.1343555599451065, | |
| "eval_runtime": 19.3831, | |
| "eval_samples_per_second": 1.445, | |
| "eval_steps_per_second": 1.445, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 35.81, | |
| "grad_norm": 3.2688581943511963, | |
| "learning_rate": 4.5455367294475655e-07, | |
| "logits/chosen": -2.7136824131011963, | |
| "logits/rejected": -2.724392890930176, | |
| "logps/chosen": -114.86489868164062, | |
| "logps/rejected": -108.21212768554688, | |
| "loss": 0.1485, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7876478433609009, | |
| "rewards/margins": 2.3711612224578857, | |
| "rewards/rejected": -0.5835133790969849, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 36.06, | |
| "grad_norm": 2.39674711227417, | |
| "learning_rate": 4.5390189405175715e-07, | |
| "logits/chosen": -2.6487505435943604, | |
| "logits/rejected": -2.651470899581909, | |
| "logps/chosen": -111.85699462890625, | |
| "logps/rejected": -101.57424926757812, | |
| "loss": 0.1222, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7182648181915283, | |
| "rewards/margins": 2.550403356552124, | |
| "rewards/rejected": -0.8321384787559509, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 36.32, | |
| "grad_norm": 2.970404624938965, | |
| "learning_rate": 4.5324594885762576e-07, | |
| "logits/chosen": -2.670048952102661, | |
| "logits/rejected": -2.70831036567688, | |
| "logps/chosen": -113.16908264160156, | |
| "logps/rejected": -116.0752944946289, | |
| "loss": 0.1838, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.6036030054092407, | |
| "rewards/margins": 2.133262872695923, | |
| "rewards/rejected": -0.5296599268913269, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 36.57, | |
| "grad_norm": 3.4210455417633057, | |
| "learning_rate": 4.5258585076520595e-07, | |
| "logits/chosen": -2.647081136703491, | |
| "logits/rejected": -2.625157594680786, | |
| "logps/chosen": -119.30801391601562, | |
| "logps/rejected": -101.5343017578125, | |
| "loss": 0.1394, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7498310804367065, | |
| "rewards/margins": 2.5768487453460693, | |
| "rewards/rejected": -0.8270176649093628, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 36.83, | |
| "grad_norm": 2.8785665035247803, | |
| "learning_rate": 4.5192161326219716e-07, | |
| "logits/chosen": -2.678386688232422, | |
| "logits/rejected": -2.6918272972106934, | |
| "logps/chosen": -111.96650695800781, | |
| "logps/rejected": -100.6183853149414, | |
| "loss": 0.1539, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.703883409500122, | |
| "rewards/margins": 2.2555694580078125, | |
| "rewards/rejected": -0.5516858696937561, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 37.08, | |
| "grad_norm": 1.953771710395813, | |
| "learning_rate": 4.512532499208786e-07, | |
| "logits/chosen": -2.622621774673462, | |
| "logits/rejected": -2.6352713108062744, | |
| "logps/chosen": -117.12508392333984, | |
| "logps/rejected": -106.53787231445312, | |
| "loss": 0.1165, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.857337474822998, | |
| "rewards/margins": 2.5984973907470703, | |
| "rewards/rejected": -0.7411599159240723, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 37.33, | |
| "grad_norm": 1.7565056085586548, | |
| "learning_rate": 4.505807743978324e-07, | |
| "logits/chosen": -2.6742653846740723, | |
| "logits/rejected": -2.6801705360412598, | |
| "logps/chosen": -112.03057098388672, | |
| "logps/rejected": -106.16900634765625, | |
| "loss": 0.145, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8526411056518555, | |
| "rewards/margins": 2.4808454513549805, | |
| "rewards/rejected": -0.6282044649124146, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 37.59, | |
| "grad_norm": 2.013282299041748, | |
| "learning_rate": 4.4990420043366415e-07, | |
| "logits/chosen": -2.6183061599731445, | |
| "logits/rejected": -2.6616299152374268, | |
| "logps/chosen": -112.98906707763672, | |
| "logps/rejected": -105.28535461425781, | |
| "loss": 0.1353, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7617923021316528, | |
| "rewards/margins": 2.430490016937256, | |
| "rewards/rejected": -0.6686975359916687, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 37.84, | |
| "grad_norm": 2.2207536697387695, | |
| "learning_rate": 4.492235418527227e-07, | |
| "logits/chosen": -2.6781811714172363, | |
| "logits/rejected": -2.6666665077209473, | |
| "logps/chosen": -116.05792999267578, | |
| "logps/rejected": -107.16488647460938, | |
| "loss": 0.1468, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7365107536315918, | |
| "rewards/margins": 2.3593077659606934, | |
| "rewards/rejected": -0.6227970719337463, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 38.1, | |
| "grad_norm": 2.3460826873779297, | |
| "learning_rate": 4.48538812562817e-07, | |
| "logits/chosen": -2.63061785697937, | |
| "logits/rejected": -2.6308352947235107, | |
| "logps/chosen": -111.25851440429688, | |
| "logps/rejected": -103.23674011230469, | |
| "loss": 0.1414, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.6231989860534668, | |
| "rewards/margins": 2.478404998779297, | |
| "rewards/rejected": -0.8552060127258301, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 38.1, | |
| "eval_logits/chosen": -2.7357699871063232, | |
| "eval_logits/rejected": -2.804048538208008, | |
| "eval_logps/chosen": -121.1922607421875, | |
| "eval_logps/rejected": -100.1863021850586, | |
| "eval_loss": 0.38548800349235535, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2877070903778076, | |
| "eval_rewards/margins": 1.495449423789978, | |
| "eval_rewards/rejected": -0.207742378115654, | |
| "eval_runtime": 19.3741, | |
| "eval_samples_per_second": 1.445, | |
| "eval_steps_per_second": 1.445, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 38.35, | |
| "grad_norm": 2.2148759365081787, | |
| "learning_rate": 4.4785002655493244e-07, | |
| "logits/chosen": -2.68538236618042, | |
| "logits/rejected": -2.721736431121826, | |
| "logps/chosen": -115.57979583740234, | |
| "logps/rejected": -107.06050872802734, | |
| "loss": 0.1434, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.6038768291473389, | |
| "rewards/margins": 2.4257020950317383, | |
| "rewards/rejected": -0.8218252658843994, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 38.6, | |
| "grad_norm": 2.045954465866089, | |
| "learning_rate": 4.4715719790294473e-07, | |
| "logits/chosen": -2.665043592453003, | |
| "logits/rejected": -2.6772406101226807, | |
| "logps/chosen": -118.3508071899414, | |
| "logps/rejected": -105.97952270507812, | |
| "loss": 0.1027, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.952754259109497, | |
| "rewards/margins": 2.809091329574585, | |
| "rewards/rejected": -0.8563370108604431, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 38.86, | |
| "grad_norm": 3.2380034923553467, | |
| "learning_rate": 4.4646034076333254e-07, | |
| "logits/chosen": -2.6413214206695557, | |
| "logits/rejected": -2.614654064178467, | |
| "logps/chosen": -112.02613067626953, | |
| "logps/rejected": -107.12358856201172, | |
| "loss": 0.1423, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.760948657989502, | |
| "rewards/margins": 2.4380459785461426, | |
| "rewards/rejected": -0.6770972609519958, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 39.11, | |
| "grad_norm": 1.8163915872573853, | |
| "learning_rate": 4.4575946937488807e-07, | |
| "logits/chosen": -2.653783082962036, | |
| "logits/rejected": -2.697185516357422, | |
| "logps/chosen": -109.09898376464844, | |
| "logps/rejected": -105.73826599121094, | |
| "loss": 0.1531, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.7506372928619385, | |
| "rewards/margins": 2.376413345336914, | |
| "rewards/rejected": -0.6257758140563965, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 39.37, | |
| "grad_norm": 2.3035383224487305, | |
| "learning_rate": 4.4505459805842593e-07, | |
| "logits/chosen": -2.6597530841827393, | |
| "logits/rejected": -2.691429615020752, | |
| "logps/chosen": -108.01673889160156, | |
| "logps/rejected": -106.93608093261719, | |
| "loss": 0.1441, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7096694707870483, | |
| "rewards/margins": 2.528447389602661, | |
| "rewards/rejected": -0.8187777996063232, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 39.62, | |
| "grad_norm": 2.202437162399292, | |
| "learning_rate": 4.44345741216491e-07, | |
| "logits/chosen": -2.6261043548583984, | |
| "logits/rejected": -2.6052300930023193, | |
| "logps/chosen": -114.36175537109375, | |
| "logps/rejected": -107.27129364013672, | |
| "loss": 0.1432, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.74629807472229, | |
| "rewards/margins": 2.387127637863159, | |
| "rewards/rejected": -0.6408295035362244, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 39.87, | |
| "grad_norm": 2.515286684036255, | |
| "learning_rate": 4.43632913333064e-07, | |
| "logits/chosen": -2.6591076850891113, | |
| "logits/rejected": -2.661803722381592, | |
| "logps/chosen": -115.1274185180664, | |
| "logps/rejected": -105.72164154052734, | |
| "loss": 0.1232, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.884786605834961, | |
| "rewards/margins": 2.571641206741333, | |
| "rewards/rejected": -0.6868546009063721, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 40.13, | |
| "grad_norm": 2.1287059783935547, | |
| "learning_rate": 4.429161289732649e-07, | |
| "logits/chosen": -2.6356821060180664, | |
| "logits/rejected": -2.644843101501465, | |
| "logps/chosen": -117.84187316894531, | |
| "logps/rejected": -108.8236083984375, | |
| "loss": 0.1079, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9427728652954102, | |
| "rewards/margins": 2.9027700424194336, | |
| "rewards/rejected": -0.9599968194961548, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 40.38, | |
| "grad_norm": 2.836397886276245, | |
| "learning_rate": 4.4219540278305647e-07, | |
| "logits/chosen": -2.681767463684082, | |
| "logits/rejected": -2.682898998260498, | |
| "logps/chosen": -108.63336181640625, | |
| "logps/rejected": -104.24710845947266, | |
| "loss": 0.1326, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.57785165309906, | |
| "rewards/margins": 2.463282346725464, | |
| "rewards/rejected": -0.8854306936264038, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 40.63, | |
| "grad_norm": 2.018007516860962, | |
| "learning_rate": 4.414707494889439e-07, | |
| "logits/chosen": -2.619904041290283, | |
| "logits/rejected": -2.6312315464019775, | |
| "logps/chosen": -107.43942260742188, | |
| "logps/rejected": -104.7081069946289, | |
| "loss": 0.1121, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8468282222747803, | |
| "rewards/margins": 2.7092161178588867, | |
| "rewards/rejected": -0.8623881340026855, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 40.63, | |
| "eval_logits/chosen": -2.7261669635772705, | |
| "eval_logits/rejected": -2.792802572250366, | |
| "eval_logps/chosen": -121.07279205322266, | |
| "eval_logps/rejected": -100.8832778930664, | |
| "eval_loss": 0.3840921223163605, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2996548414230347, | |
| "eval_rewards/margins": 1.5770933628082275, | |
| "eval_rewards/rejected": -0.2774384915828705, | |
| "eval_runtime": 19.3804, | |
| "eval_samples_per_second": 1.445, | |
| "eval_steps_per_second": 1.445, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 40.89, | |
| "grad_norm": 2.168912410736084, | |
| "learning_rate": 4.4074218389767466e-07, | |
| "logits/chosen": -2.6166744232177734, | |
| "logits/rejected": -2.6106488704681396, | |
| "logps/chosen": -124.27295684814453, | |
| "logps/rejected": -111.69468688964844, | |
| "loss": 0.1185, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9064750671386719, | |
| "rewards/margins": 2.79134464263916, | |
| "rewards/rejected": -0.8848695755004883, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 41.14, | |
| "grad_norm": 1.7980531454086304, | |
| "learning_rate": 4.400097208959356e-07, | |
| "logits/chosen": -2.6521995067596436, | |
| "logits/rejected": -2.704921245574951, | |
| "logps/chosen": -114.37103271484375, | |
| "logps/rejected": -113.26276397705078, | |
| "loss": 0.1368, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.8770157098770142, | |
| "rewards/margins": 2.7274460792541504, | |
| "rewards/rejected": -0.8504302501678467, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 41.4, | |
| "grad_norm": 2.447611093521118, | |
| "learning_rate": 4.3927337545004894e-07, | |
| "logits/chosen": -2.6812400817871094, | |
| "logits/rejected": -2.725090503692627, | |
| "logps/chosen": -112.9039535522461, | |
| "logps/rejected": -103.30266571044922, | |
| "loss": 0.119, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7921515703201294, | |
| "rewards/margins": 2.6354293823242188, | |
| "rewards/rejected": -0.8432780504226685, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 41.65, | |
| "grad_norm": 2.0098862648010254, | |
| "learning_rate": 4.385331626056663e-07, | |
| "logits/chosen": -2.6464550495147705, | |
| "logits/rejected": -2.60159969329834, | |
| "logps/chosen": -118.65347290039062, | |
| "logps/rejected": -106.13660430908203, | |
| "loss": 0.1482, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.773016333580017, | |
| "rewards/margins": 2.4340577125549316, | |
| "rewards/rejected": -0.661041259765625, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 41.9, | |
| "grad_norm": 2.093533515930176, | |
| "learning_rate": 4.377890974874613e-07, | |
| "logits/chosen": -2.652855157852173, | |
| "logits/rejected": -2.6764514446258545, | |
| "logps/chosen": -110.1133804321289, | |
| "logps/rejected": -112.76753997802734, | |
| "loss": 0.1062, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.795628309249878, | |
| "rewards/margins": 2.800370693206787, | |
| "rewards/rejected": -1.0047426223754883, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 42.16, | |
| "grad_norm": 1.8786094188690186, | |
| "learning_rate": 4.370411952988206e-07, | |
| "logits/chosen": -2.5962324142456055, | |
| "logits/rejected": -2.580698251724243, | |
| "logps/chosen": -116.74523162841797, | |
| "logps/rejected": -108.6583251953125, | |
| "loss": 0.1241, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7898542881011963, | |
| "rewards/margins": 2.6399893760681152, | |
| "rewards/rejected": -0.8501349687576294, | |
| "step": 664 | |
| }, | |
| { | |
| "epoch": 42.41, | |
| "grad_norm": 1.1881071329116821, | |
| "learning_rate": 4.362894713215334e-07, | |
| "logits/chosen": -2.642057418823242, | |
| "logits/rejected": -2.6707074642181396, | |
| "logps/chosen": -114.59297180175781, | |
| "logps/rejected": -107.72032165527344, | |
| "loss": 0.1115, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9477055072784424, | |
| "rewards/margins": 2.8435428142547607, | |
| "rewards/rejected": -0.8958378434181213, | |
| "step": 668 | |
| }, | |
| { | |
| "epoch": 42.67, | |
| "grad_norm": 2.0914270877838135, | |
| "learning_rate": 4.355339409154788e-07, | |
| "logits/chosen": -2.650442123413086, | |
| "logits/rejected": -2.647216558456421, | |
| "logps/chosen": -108.08013916015625, | |
| "logps/rejected": -103.83348083496094, | |
| "loss": 0.093, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8842676877975464, | |
| "rewards/margins": 2.963129997253418, | |
| "rewards/rejected": -1.0788623094558716, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 42.92, | |
| "grad_norm": 2.257300615310669, | |
| "learning_rate": 4.347746195183122e-07, | |
| "logits/chosen": -2.6150929927825928, | |
| "logits/rejected": -2.6115312576293945, | |
| "logps/chosen": -113.52247619628906, | |
| "logps/rejected": -110.75169372558594, | |
| "loss": 0.1205, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.8281488418579102, | |
| "rewards/margins": 2.7573764324188232, | |
| "rewards/rejected": -0.9292274117469788, | |
| "step": 676 | |
| }, | |
| { | |
| "epoch": 43.17, | |
| "grad_norm": 1.82343327999115, | |
| "learning_rate": 4.340115226451501e-07, | |
| "logits/chosen": -2.6201157569885254, | |
| "logits/rejected": -2.6617937088012695, | |
| "logps/chosen": -108.35755157470703, | |
| "logps/rejected": -107.59652709960938, | |
| "loss": 0.0976, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8024065494537354, | |
| "rewards/margins": 2.9102468490600586, | |
| "rewards/rejected": -1.1078401803970337, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 43.17, | |
| "eval_logits/chosen": -2.7156684398651123, | |
| "eval_logits/rejected": -2.7805325984954834, | |
| "eval_logps/chosen": -121.03025817871094, | |
| "eval_logps/rejected": -101.69651794433594, | |
| "eval_loss": 0.38208508491516113, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.3039093017578125, | |
| "eval_rewards/margins": 1.6626710891723633, | |
| "eval_rewards/rejected": -0.358761727809906, | |
| "eval_runtime": 19.3666, | |
| "eval_samples_per_second": 1.446, | |
| "eval_steps_per_second": 1.446, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 43.43, | |
| "grad_norm": 1.9972604513168335, | |
| "learning_rate": 4.3324466588825223e-07, | |
| "logits/chosen": -2.6469597816467285, | |
| "logits/rejected": -2.6479005813598633, | |
| "logps/chosen": -117.90531921386719, | |
| "logps/rejected": -106.33170318603516, | |
| "loss": 0.1079, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8468865156173706, | |
| "rewards/margins": 2.780641555786133, | |
| "rewards/rejected": -0.9337549805641174, | |
| "step": 684 | |
| }, | |
| { | |
| "epoch": 43.68, | |
| "grad_norm": 1.7684298753738403, | |
| "learning_rate": 4.3247406491670435e-07, | |
| "logits/chosen": -2.653805732727051, | |
| "logits/rejected": -2.658804178237915, | |
| "logps/chosen": -108.92557525634766, | |
| "logps/rejected": -110.38461303710938, | |
| "loss": 0.1272, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9131863117218018, | |
| "rewards/margins": 2.63901948928833, | |
| "rewards/rejected": -0.7258330583572388, | |
| "step": 688 | |
| }, | |
| { | |
| "epoch": 43.94, | |
| "grad_norm": 1.980388879776001, | |
| "learning_rate": 4.3169973547609644e-07, | |
| "logits/chosen": -2.637269973754883, | |
| "logits/rejected": -2.6658825874328613, | |
| "logps/chosen": -110.72245788574219, | |
| "logps/rejected": -109.98155975341797, | |
| "loss": 0.1142, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.804459810256958, | |
| "rewards/margins": 2.8206331729888916, | |
| "rewards/rejected": -1.0161734819412231, | |
| "step": 692 | |
| }, | |
| { | |
| "epoch": 44.19, | |
| "grad_norm": 1.6069421768188477, | |
| "learning_rate": 4.3092169338820245e-07, | |
| "logits/chosen": -2.6507728099823, | |
| "logits/rejected": -2.5921573638916016, | |
| "logps/chosen": -115.70768737792969, | |
| "logps/rejected": -117.3571548461914, | |
| "loss": 0.1211, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.852295994758606, | |
| "rewards/margins": 2.9615478515625, | |
| "rewards/rejected": -1.1092519760131836, | |
| "step": 696 | |
| }, | |
| { | |
| "epoch": 44.44, | |
| "grad_norm": 2.384201765060425, | |
| "learning_rate": 4.30139954550656e-07, | |
| "logits/chosen": -2.6414434909820557, | |
| "logits/rejected": -2.638594627380371, | |
| "logps/chosen": -112.53083038330078, | |
| "logps/rejected": -103.69336700439453, | |
| "loss": 0.0863, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9861321449279785, | |
| "rewards/margins": 3.037813663482666, | |
| "rewards/rejected": -1.0516817569732666, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 44.7, | |
| "grad_norm": 2.6236279010772705, | |
| "learning_rate": 4.293545349366261e-07, | |
| "logits/chosen": -2.627302646636963, | |
| "logits/rejected": -2.648130416870117, | |
| "logps/chosen": -117.52839660644531, | |
| "logps/rejected": -113.75141143798828, | |
| "loss": 0.1266, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7409615516662598, | |
| "rewards/margins": 2.609776258468628, | |
| "rewards/rejected": -0.8688147068023682, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 44.95, | |
| "grad_norm": 1.3566991090774536, | |
| "learning_rate": 4.2856545059449054e-07, | |
| "logits/chosen": -2.609771251678467, | |
| "logits/rejected": -2.649744749069214, | |
| "logps/chosen": -110.12776947021484, | |
| "logps/rejected": -104.61019897460938, | |
| "loss": 0.0946, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9009833335876465, | |
| "rewards/margins": 3.0006396770477295, | |
| "rewards/rejected": -1.099656343460083, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 45.21, | |
| "grad_norm": 1.4097861051559448, | |
| "learning_rate": 4.27772717647508e-07, | |
| "logits/chosen": -2.639524459838867, | |
| "logits/rejected": -2.6454386711120605, | |
| "logps/chosen": -110.92047119140625, | |
| "logps/rejected": -104.04742431640625, | |
| "loss": 0.1045, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7894549369812012, | |
| "rewards/margins": 2.7535037994384766, | |
| "rewards/rejected": -0.9640489816665649, | |
| "step": 712 | |
| }, | |
| { | |
| "epoch": 45.46, | |
| "grad_norm": 2.119607448577881, | |
| "learning_rate": 4.2697635229348874e-07, | |
| "logits/chosen": -2.6511929035186768, | |
| "logits/rejected": -2.633582353591919, | |
| "logps/chosen": -109.69573974609375, | |
| "logps/rejected": -110.53610229492188, | |
| "loss": 0.1163, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7527060508728027, | |
| "rewards/margins": 2.6145551204681396, | |
| "rewards/rejected": -0.8618494868278503, | |
| "step": 716 | |
| }, | |
| { | |
| "epoch": 45.71, | |
| "grad_norm": 1.88676917552948, | |
| "learning_rate": 4.261763708044632e-07, | |
| "logits/chosen": -2.637509822845459, | |
| "logits/rejected": -2.6555514335632324, | |
| "logps/chosen": -114.49153137207031, | |
| "logps/rejected": -110.72467041015625, | |
| "loss": 0.1129, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.9959712028503418, | |
| "rewards/margins": 3.11696720123291, | |
| "rewards/rejected": -1.1209962368011475, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 45.71, | |
| "eval_logits/chosen": -2.704455614089966, | |
| "eval_logits/rejected": -2.7675421237945557, | |
| "eval_logps/chosen": -121.06656646728516, | |
| "eval_logps/rejected": -102.52320098876953, | |
| "eval_loss": 0.38231486082077026, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.3002774715423584, | |
| "eval_rewards/margins": 1.741708517074585, | |
| "eval_rewards/rejected": -0.44143104553222656, | |
| "eval_runtime": 19.3804, | |
| "eval_samples_per_second": 1.445, | |
| "eval_steps_per_second": 1.445, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 45.97, | |
| "grad_norm": 1.0807371139526367, | |
| "learning_rate": 4.2537278952635036e-07, | |
| "logits/chosen": -2.653682231903076, | |
| "logits/rejected": -2.6631548404693604, | |
| "logps/chosen": -112.33302307128906, | |
| "logps/rejected": -112.11505889892578, | |
| "loss": 0.0763, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9477115869522095, | |
| "rewards/margins": 3.1752381324768066, | |
| "rewards/rejected": -1.227526307106018, | |
| "step": 724 | |
| }, | |
| { | |
| "epoch": 46.22, | |
| "grad_norm": 1.5240466594696045, | |
| "learning_rate": 4.245656248786228e-07, | |
| "logits/chosen": -2.5998778343200684, | |
| "logits/rejected": -2.628650188446045, | |
| "logps/chosen": -116.5973892211914, | |
| "logps/rejected": -119.11909484863281, | |
| "loss": 0.0753, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.059324026107788, | |
| "rewards/margins": 3.2868502140045166, | |
| "rewards/rejected": -1.2275261878967285, | |
| "step": 728 | |
| }, | |
| { | |
| "epoch": 46.48, | |
| "grad_norm": 1.6910673379898071, | |
| "learning_rate": 4.237548933539718e-07, | |
| "logits/chosen": -2.609449863433838, | |
| "logits/rejected": -2.6589772701263428, | |
| "logps/chosen": -110.6764907836914, | |
| "logps/rejected": -105.00558471679688, | |
| "loss": 0.1179, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.9273223876953125, | |
| "rewards/margins": 2.8337433338165283, | |
| "rewards/rejected": -0.9064207077026367, | |
| "step": 732 | |
| }, | |
| { | |
| "epoch": 46.73, | |
| "grad_norm": 1.9393585920333862, | |
| "learning_rate": 4.229406115179702e-07, | |
| "logits/chosen": -2.674577474594116, | |
| "logits/rejected": -2.6538712978363037, | |
| "logps/chosen": -114.55770874023438, | |
| "logps/rejected": -111.56117248535156, | |
| "loss": 0.1062, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.6356512308120728, | |
| "rewards/margins": 2.7821154594421387, | |
| "rewards/rejected": -1.146464467048645, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 46.98, | |
| "grad_norm": 1.7621327638626099, | |
| "learning_rate": 4.221227960087338e-07, | |
| "logits/chosen": -2.6068484783172607, | |
| "logits/rejected": -2.5801186561584473, | |
| "logps/chosen": -109.57792663574219, | |
| "logps/rejected": -106.45953369140625, | |
| "loss": 0.0958, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8922866582870483, | |
| "rewards/margins": 3.114880323410034, | |
| "rewards/rejected": -1.2225935459136963, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 47.24, | |
| "grad_norm": 1.7458181381225586, | |
| "learning_rate": 4.213014635365815e-07, | |
| "logits/chosen": -2.6573941707611084, | |
| "logits/rejected": -2.653240919113159, | |
| "logps/chosen": -116.97773742675781, | |
| "logps/rejected": -108.12200164794922, | |
| "loss": 0.0816, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.0080180168151855, | |
| "rewards/margins": 3.296840190887451, | |
| "rewards/rejected": -1.2888221740722656, | |
| "step": 744 | |
| }, | |
| { | |
| "epoch": 47.49, | |
| "grad_norm": 1.9578934907913208, | |
| "learning_rate": 4.20476630883694e-07, | |
| "logits/chosen": -2.5878279209136963, | |
| "logits/rejected": -2.595855712890625, | |
| "logps/chosen": -115.75021362304688, | |
| "logps/rejected": -112.4002685546875, | |
| "loss": 0.075, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9529811143875122, | |
| "rewards/margins": 3.2216196060180664, | |
| "rewards/rejected": -1.2686389684677124, | |
| "step": 748 | |
| }, | |
| { | |
| "epoch": 47.75, | |
| "grad_norm": 1.6000980138778687, | |
| "learning_rate": 4.196483149037706e-07, | |
| "logits/chosen": -2.632962226867676, | |
| "logits/rejected": -2.6356112957000732, | |
| "logps/chosen": -110.47245788574219, | |
| "logps/rejected": -111.49674224853516, | |
| "loss": 0.1139, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.7682106494903564, | |
| "rewards/margins": 2.8430519104003906, | |
| "rewards/rejected": -1.074841022491455, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "grad_norm": 1.9791028499603271, | |
| "learning_rate": 4.188165325216849e-07, | |
| "logits/chosen": -2.6407368183135986, | |
| "logits/rejected": -2.6439175605773926, | |
| "logps/chosen": -107.70930480957031, | |
| "logps/rejected": -111.42974853515625, | |
| "loss": 0.1044, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9136321544647217, | |
| "rewards/margins": 3.008749485015869, | |
| "rewards/rejected": -1.095117211341858, | |
| "step": 756 | |
| }, | |
| { | |
| "epoch": 48.25, | |
| "grad_norm": 1.486619234085083, | |
| "learning_rate": 4.179813007331393e-07, | |
| "logits/chosen": -2.638111114501953, | |
| "logits/rejected": -2.6660499572753906, | |
| "logps/chosen": -112.31642150878906, | |
| "logps/rejected": -114.76564025878906, | |
| "loss": 0.1037, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.7412384748458862, | |
| "rewards/margins": 2.972817897796631, | |
| "rewards/rejected": -1.2315791845321655, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 48.25, | |
| "eval_logits/chosen": -2.69281268119812, | |
| "eval_logits/rejected": -2.754284143447876, | |
| "eval_logps/chosen": -121.21448516845703, | |
| "eval_logps/rejected": -103.48255920410156, | |
| "eval_loss": 0.3811449110507965, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2854851484298706, | |
| "eval_rewards/margins": 1.8228529691696167, | |
| "eval_rewards/rejected": -0.5373677015304565, | |
| "eval_runtime": 19.4147, | |
| "eval_samples_per_second": 1.442, | |
| "eval_steps_per_second": 1.442, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 48.51, | |
| "grad_norm": 1.6442840099334717, | |
| "learning_rate": 4.171426366043172e-07, | |
| "logits/chosen": -2.5812575817108154, | |
| "logits/rejected": -2.6027064323425293, | |
| "logps/chosen": -109.93325805664062, | |
| "logps/rejected": -106.68315887451172, | |
| "loss": 0.1014, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 2.025794506072998, | |
| "rewards/margins": 2.979997158050537, | |
| "rewards/rejected": -0.9542028903961182, | |
| "step": 764 | |
| }, | |
| { | |
| "epoch": 48.76, | |
| "grad_norm": 1.7033922672271729, | |
| "learning_rate": 4.163005572715348e-07, | |
| "logits/chosen": -2.6315484046936035, | |
| "logits/rejected": -2.652425765991211, | |
| "logps/chosen": -114.54792785644531, | |
| "logps/rejected": -113.0596923828125, | |
| "loss": 0.0923, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9352294206619263, | |
| "rewards/margins": 3.1911542415618896, | |
| "rewards/rejected": -1.2559247016906738, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 49.02, | |
| "grad_norm": 1.4940606355667114, | |
| "learning_rate": 4.154550799408906e-07, | |
| "logits/chosen": -2.6459126472473145, | |
| "logits/rejected": -2.59468936920166, | |
| "logps/chosen": -112.78709411621094, | |
| "logps/rejected": -111.69085693359375, | |
| "loss": 0.0632, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9187610149383545, | |
| "rewards/margins": 3.4088921546936035, | |
| "rewards/rejected": -1.4901307821273804, | |
| "step": 772 | |
| }, | |
| { | |
| "epoch": 49.27, | |
| "grad_norm": 1.6349976062774658, | |
| "learning_rate": 4.1460622188791386e-07, | |
| "logits/chosen": -2.626100778579712, | |
| "logits/rejected": -2.6048431396484375, | |
| "logps/chosen": -114.90809631347656, | |
| "logps/rejected": -110.02332305908203, | |
| "loss": 0.0862, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.9543546438217163, | |
| "rewards/margins": 3.2480387687683105, | |
| "rewards/rejected": -1.2936842441558838, | |
| "step": 776 | |
| }, | |
| { | |
| "epoch": 49.52, | |
| "grad_norm": 1.4356471300125122, | |
| "learning_rate": 4.13754000457212e-07, | |
| "logits/chosen": -2.6456077098846436, | |
| "logits/rejected": -2.676055908203125, | |
| "logps/chosen": -103.83885955810547, | |
| "logps/rejected": -109.37602233886719, | |
| "loss": 0.082, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8857074975967407, | |
| "rewards/margins": 3.213439702987671, | |
| "rewards/rejected": -1.3277320861816406, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 49.78, | |
| "grad_norm": 1.7946281433105469, | |
| "learning_rate": 4.128984330621156e-07, | |
| "logits/chosen": -2.6110281944274902, | |
| "logits/rejected": -2.605884075164795, | |
| "logps/chosen": -116.66584777832031, | |
| "logps/rejected": -114.77125549316406, | |
| "loss": 0.0825, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.0007410049438477, | |
| "rewards/margins": 3.1677379608154297, | |
| "rewards/rejected": -1.1669968366622925, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 50.03, | |
| "grad_norm": 2.0904555320739746, | |
| "learning_rate": 4.1203953718432304e-07, | |
| "logits/chosen": -2.5824971199035645, | |
| "logits/rejected": -2.5949392318725586, | |
| "logps/chosen": -113.75666809082031, | |
| "logps/rejected": -112.34939575195312, | |
| "loss": 0.0968, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8396023511886597, | |
| "rewards/margins": 3.167187213897705, | |
| "rewards/rejected": -1.3275845050811768, | |
| "step": 788 | |
| }, | |
| { | |
| "epoch": 50.29, | |
| "grad_norm": 1.8408178091049194, | |
| "learning_rate": 4.1117733037354313e-07, | |
| "logits/chosen": -2.5902395248413086, | |
| "logits/rejected": -2.590879440307617, | |
| "logps/chosen": -116.93998718261719, | |
| "logps/rejected": -112.17693328857422, | |
| "loss": 0.0719, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.067030429840088, | |
| "rewards/margins": 3.2486207485198975, | |
| "rewards/rejected": -1.1815906763076782, | |
| "step": 792 | |
| }, | |
| { | |
| "epoch": 50.54, | |
| "grad_norm": 1.2143826484680176, | |
| "learning_rate": 4.1031183024713657e-07, | |
| "logits/chosen": -2.6065564155578613, | |
| "logits/rejected": -2.5807137489318848, | |
| "logps/chosen": -116.70026397705078, | |
| "logps/rejected": -109.03843688964844, | |
| "loss": 0.0901, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 2.0011649131774902, | |
| "rewards/margins": 3.269892454147339, | |
| "rewards/rejected": -1.2687275409698486, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 50.79, | |
| "grad_norm": 1.2470941543579102, | |
| "learning_rate": 4.0944305448975594e-07, | |
| "logits/chosen": -2.6263880729675293, | |
| "logits/rejected": -2.6351208686828613, | |
| "logps/chosen": -109.2755126953125, | |
| "logps/rejected": -116.7694091796875, | |
| "loss": 0.085, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.79188871383667, | |
| "rewards/margins": 3.278684616088867, | |
| "rewards/rejected": -1.4867955446243286, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 50.79, | |
| "eval_logits/chosen": -2.681551694869995, | |
| "eval_logits/rejected": -2.741281270980835, | |
| "eval_logps/chosen": -121.30486297607422, | |
| "eval_logps/rejected": -104.38188171386719, | |
| "eval_loss": 0.38058552145957947, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2764482498168945, | |
| "eval_rewards/margins": 1.9037466049194336, | |
| "eval_rewards/rejected": -0.6272983551025391, | |
| "eval_runtime": 19.3679, | |
| "eval_samples_per_second": 1.446, | |
| "eval_steps_per_second": 1.446, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 51.05, | |
| "grad_norm": 2.3190038204193115, | |
| "learning_rate": 4.085710208529843e-07, | |
| "logits/chosen": -2.6023600101470947, | |
| "logits/rejected": -2.6153862476348877, | |
| "logps/chosen": -105.47126007080078, | |
| "logps/rejected": -109.3538818359375, | |
| "loss": 0.0898, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8685368299484253, | |
| "rewards/margins": 3.1377530097961426, | |
| "rewards/rejected": -1.2692162990570068, | |
| "step": 804 | |
| }, | |
| { | |
| "epoch": 51.3, | |
| "grad_norm": 1.2653313875198364, | |
| "learning_rate": 4.076957471549728e-07, | |
| "logits/chosen": -2.6036365032196045, | |
| "logits/rejected": -2.623936414718628, | |
| "logps/chosen": -112.60127258300781, | |
| "logps/rejected": -108.7287368774414, | |
| "loss": 0.0865, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9202613830566406, | |
| "rewards/margins": 3.164761543273926, | |
| "rewards/rejected": -1.2445003986358643, | |
| "step": 808 | |
| }, | |
| { | |
| "epoch": 51.56, | |
| "grad_norm": 1.1972923278808594, | |
| "learning_rate": 4.068172512800759e-07, | |
| "logits/chosen": -2.5859198570251465, | |
| "logits/rejected": -2.5816516876220703, | |
| "logps/chosen": -110.08045959472656, | |
| "logps/rejected": -117.87619018554688, | |
| "loss": 0.0605, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.0786402225494385, | |
| "rewards/margins": 3.6814520359039307, | |
| "rewards/rejected": -1.6028118133544922, | |
| "step": 812 | |
| }, | |
| { | |
| "epoch": 51.81, | |
| "grad_norm": 1.5057625770568848, | |
| "learning_rate": 4.059355511784868e-07, | |
| "logits/chosen": -2.619558572769165, | |
| "logits/rejected": -2.6305971145629883, | |
| "logps/chosen": -114.13056945800781, | |
| "logps/rejected": -113.92069244384766, | |
| "loss": 0.0933, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.7876722812652588, | |
| "rewards/margins": 3.144243001937866, | |
| "rewards/rejected": -1.3565707206726074, | |
| "step": 816 | |
| }, | |
| { | |
| "epoch": 52.06, | |
| "grad_norm": 1.5957279205322266, | |
| "learning_rate": 4.0505066486587e-07, | |
| "logits/chosen": -2.638758659362793, | |
| "logits/rejected": -2.615614175796509, | |
| "logps/chosen": -117.18038940429688, | |
| "logps/rejected": -118.73646545410156, | |
| "loss": 0.0666, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.000675678253174, | |
| "rewards/margins": 3.431554079055786, | |
| "rewards/rejected": -1.4308786392211914, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 52.32, | |
| "grad_norm": 1.7117719650268555, | |
| "learning_rate": 4.041626104229936e-07, | |
| "logits/chosen": -2.561399221420288, | |
| "logits/rejected": -2.5965375900268555, | |
| "logps/chosen": -108.37356567382812, | |
| "logps/rejected": -109.40271759033203, | |
| "loss": 0.0615, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.173064708709717, | |
| "rewards/margins": 3.6536784172058105, | |
| "rewards/rejected": -1.4806135892868042, | |
| "step": 824 | |
| }, | |
| { | |
| "epoch": 52.57, | |
| "grad_norm": 1.3572646379470825, | |
| "learning_rate": 4.032714059953595e-07, | |
| "logits/chosen": -2.6196985244750977, | |
| "logits/rejected": -2.614413022994995, | |
| "logps/chosen": -109.2529067993164, | |
| "logps/rejected": -104.56177520751953, | |
| "loss": 0.092, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.9012202024459839, | |
| "rewards/margins": 3.18809175491333, | |
| "rewards/rejected": -1.2868719100952148, | |
| "step": 828 | |
| }, | |
| { | |
| "epoch": 52.83, | |
| "grad_norm": 1.784955382347107, | |
| "learning_rate": 4.0237706979283305e-07, | |
| "logits/chosen": -2.6027987003326416, | |
| "logits/rejected": -2.6091623306274414, | |
| "logps/chosen": -116.41980743408203, | |
| "logps/rejected": -121.5534896850586, | |
| "loss": 0.0837, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.864450216293335, | |
| "rewards/margins": 3.336150646209717, | |
| "rewards/rejected": -1.4717004299163818, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 53.08, | |
| "grad_norm": 1.603405237197876, | |
| "learning_rate": 4.014796200892706e-07, | |
| "logits/chosen": -2.59523344039917, | |
| "logits/rejected": -2.59432053565979, | |
| "logps/chosen": -108.4609603881836, | |
| "logps/rejected": -112.68742370605469, | |
| "loss": 0.071, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.829506278038025, | |
| "rewards/margins": 3.3810737133026123, | |
| "rewards/rejected": -1.5515676736831665, | |
| "step": 836 | |
| }, | |
| { | |
| "epoch": 53.33, | |
| "grad_norm": 1.8830804824829102, | |
| "learning_rate": 4.005790752221464e-07, | |
| "logits/chosen": -2.6272573471069336, | |
| "logits/rejected": -2.627049446105957, | |
| "logps/chosen": -113.10144805908203, | |
| "logps/rejected": -108.62581634521484, | |
| "loss": 0.0714, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.913501262664795, | |
| "rewards/margins": 3.577221632003784, | |
| "rewards/rejected": -1.663720726966858, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 53.33, | |
| "eval_logits/chosen": -2.668611526489258, | |
| "eval_logits/rejected": -2.726428508758545, | |
| "eval_logps/chosen": -121.56246185302734, | |
| "eval_logps/rejected": -105.41184997558594, | |
| "eval_loss": 0.381181925535202, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2506877183914185, | |
| "eval_rewards/margins": 1.9809836149215698, | |
| "eval_rewards/rejected": -0.7302957773208618, | |
| "eval_runtime": 19.382, | |
| "eval_samples_per_second": 1.445, | |
| "eval_steps_per_second": 1.445, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 53.59, | |
| "grad_norm": 1.5219979286193848, | |
| "learning_rate": 3.996754535921777e-07, | |
| "logits/chosen": -2.594785451889038, | |
| "logits/rejected": -2.600487470626831, | |
| "logps/chosen": -106.41715240478516, | |
| "logps/rejected": -120.44597625732422, | |
| "loss": 0.0765, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9574780464172363, | |
| "rewards/margins": 3.235558032989502, | |
| "rewards/rejected": -1.2780797481536865, | |
| "step": 844 | |
| }, | |
| { | |
| "epoch": 53.84, | |
| "grad_norm": 0.8998934626579285, | |
| "learning_rate": 3.987687736629487e-07, | |
| "logits/chosen": -2.5739502906799316, | |
| "logits/rejected": -2.540557861328125, | |
| "logps/chosen": -116.07780456542969, | |
| "logps/rejected": -113.49636840820312, | |
| "loss": 0.0588, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 1.980076551437378, | |
| "rewards/margins": 3.6954421997070312, | |
| "rewards/rejected": -1.7153657674789429, | |
| "step": 848 | |
| }, | |
| { | |
| "epoch": 54.1, | |
| "grad_norm": 1.3017789125442505, | |
| "learning_rate": 3.9785905396053377e-07, | |
| "logits/chosen": -2.5990142822265625, | |
| "logits/rejected": -2.6168363094329834, | |
| "logps/chosen": -111.93960571289062, | |
| "logps/rejected": -116.76874542236328, | |
| "loss": 0.0872, | |
| "rewards/accuracies": 0.984375, | |
| "rewards/chosen": 2.0168068408966064, | |
| "rewards/margins": 3.3573083877563477, | |
| "rewards/rejected": -1.340501308441162, | |
| "step": 852 | |
| }, | |
| { | |
| "epoch": 54.35, | |
| "grad_norm": 1.2396210432052612, | |
| "learning_rate": 3.9694631307311825e-07, | |
| "logits/chosen": -2.5667662620544434, | |
| "logits/rejected": -2.5850954055786133, | |
| "logps/chosen": -111.22042846679688, | |
| "logps/rejected": -111.89250183105469, | |
| "loss": 0.0731, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9918500185012817, | |
| "rewards/margins": 3.535062313079834, | |
| "rewards/rejected": -1.5432121753692627, | |
| "step": 856 | |
| }, | |
| { | |
| "epoch": 54.6, | |
| "grad_norm": 1.335559606552124, | |
| "learning_rate": 3.960305696506192e-07, | |
| "logits/chosen": -2.5854077339172363, | |
| "logits/rejected": -2.5688228607177734, | |
| "logps/chosen": -116.70706939697266, | |
| "logps/rejected": -110.84555053710938, | |
| "loss": 0.0528, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.941176414489746, | |
| "rewards/margins": 3.603893518447876, | |
| "rewards/rejected": -1.6627171039581299, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 54.86, | |
| "grad_norm": 1.6855989694595337, | |
| "learning_rate": 3.95111842404304e-07, | |
| "logits/chosen": -2.5832574367523193, | |
| "logits/rejected": -2.5922787189483643, | |
| "logps/chosen": -115.02678680419922, | |
| "logps/rejected": -123.42115783691406, | |
| "loss": 0.0682, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.9162797927856445, | |
| "rewards/margins": 3.6613926887512207, | |
| "rewards/rejected": -1.7451128959655762, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 55.11, | |
| "grad_norm": 1.599083423614502, | |
| "learning_rate": 3.94190150106408e-07, | |
| "logits/chosen": -2.6082019805908203, | |
| "logits/rejected": -2.583937168121338, | |
| "logps/chosen": -106.68289947509766, | |
| "logps/rejected": -116.42233276367188, | |
| "loss": 0.0741, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8997607231140137, | |
| "rewards/margins": 3.2817296981811523, | |
| "rewards/rejected": -1.3819692134857178, | |
| "step": 868 | |
| }, | |
| { | |
| "epoch": 55.37, | |
| "grad_norm": 1.4026548862457275, | |
| "learning_rate": 3.9326551158975124e-07, | |
| "logits/chosen": -2.575761079788208, | |
| "logits/rejected": -2.6161446571350098, | |
| "logps/chosen": -109.90505981445312, | |
| "logps/rejected": -113.70225524902344, | |
| "loss": 0.0556, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 2.069807291030884, | |
| "rewards/margins": 3.6844887733459473, | |
| "rewards/rejected": -1.6146814823150635, | |
| "step": 872 | |
| }, | |
| { | |
| "epoch": 55.62, | |
| "grad_norm": 1.6232682466506958, | |
| "learning_rate": 3.923379457473534e-07, | |
| "logits/chosen": -2.569347858428955, | |
| "logits/rejected": -2.5830793380737305, | |
| "logps/chosen": -115.759033203125, | |
| "logps/rejected": -114.3780288696289, | |
| "loss": 0.0651, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.960666298866272, | |
| "rewards/margins": 3.5167884826660156, | |
| "rewards/rejected": -1.5561223030090332, | |
| "step": 876 | |
| }, | |
| { | |
| "epoch": 55.87, | |
| "grad_norm": 1.7684590816497803, | |
| "learning_rate": 3.914074715320479e-07, | |
| "logits/chosen": -2.5655999183654785, | |
| "logits/rejected": -2.5638012886047363, | |
| "logps/chosen": -109.46947479248047, | |
| "logps/rejected": -118.88379669189453, | |
| "loss": 0.0605, | |
| "rewards/accuracies": 1.0, | |
| "rewards/chosen": 1.8757377862930298, | |
| "rewards/margins": 3.77754807472229, | |
| "rewards/rejected": -1.9018104076385498, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 55.87, | |
| "eval_logits/chosen": -2.656400203704834, | |
| "eval_logits/rejected": -2.712291955947876, | |
| "eval_logps/chosen": -121.80836486816406, | |
| "eval_logps/rejected": -106.49579620361328, | |
| "eval_loss": 0.380399227142334, | |
| "eval_rewards/accuracies": 0.7857142686843872, | |
| "eval_rewards/chosen": 1.2260981798171997, | |
| "eval_rewards/margins": 2.0647895336151123, | |
| "eval_rewards/rejected": -0.8386915922164917, | |
| "eval_runtime": 19.382, | |
| "eval_samples_per_second": 1.445, | |
| "eval_steps_per_second": 1.445, | |
| "step": 880 | |
| } | |
| ], | |
| "logging_steps": 4, | |
| "max_steps": 2800, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 187, | |
| "save_steps": 40, | |
| "total_flos": 0.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |