| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 0.9972451790633609, | |
| "eval_steps": 100, | |
| "global_step": 181, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 4.4602671481986516, | |
| "learning_rate": 2.6315789473684208e-08, | |
| "logits/chosen": -0.7853389978408813, | |
| "logits/rejected": -0.651086151599884, | |
| "logps/chosen": -1158.1322021484375, | |
| "logps/rejected": -1143.569091796875, | |
| "loss": 0.6931, | |
| "rewards/accuracies": 0.0, | |
| "rewards/chosen": 0.0, | |
| "rewards/margins": 0.0, | |
| "rewards/rejected": 0.0, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 4.676508624329794, | |
| "learning_rate": 2.631578947368421e-07, | |
| "logits/chosen": -0.709392786026001, | |
| "logits/rejected": -0.8101767897605896, | |
| "logps/chosen": -995.121826171875, | |
| "logps/rejected": -1316.8297119140625, | |
| "loss": 0.693, | |
| "rewards/accuracies": 0.4305555522441864, | |
| "rewards/chosen": -0.0007655912195332348, | |
| "rewards/margins": -0.00027954205870628357, | |
| "rewards/rejected": -0.0004860491317231208, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 4.611380190345714, | |
| "learning_rate": 4.999529926121253e-07, | |
| "logits/chosen": -0.7076992988586426, | |
| "logits/rejected": -0.826177716255188, | |
| "logps/chosen": -942.4779052734375, | |
| "logps/rejected": -1303.2080078125, | |
| "loss": 0.6922, | |
| "rewards/accuracies": 0.581250011920929, | |
| "rewards/chosen": 0.0009432749939151108, | |
| "rewards/margins": 0.002406059531494975, | |
| "rewards/rejected": -0.0014627845957875252, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 4.545137388238839, | |
| "learning_rate": 4.943334645626589e-07, | |
| "logits/chosen": -0.7218513488769531, | |
| "logits/rejected": -0.8190325498580933, | |
| "logps/chosen": -986.5565185546875, | |
| "logps/rejected": -1347.8033447265625, | |
| "loss": 0.6883, | |
| "rewards/accuracies": 0.699999988079071, | |
| "rewards/chosen": 0.004364156164228916, | |
| "rewards/margins": 0.008785699494183064, | |
| "rewards/rejected": -0.004421542398631573, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 4.401192871872299, | |
| "learning_rate": 4.795540267200686e-07, | |
| "logits/chosen": -0.7325557470321655, | |
| "logits/rejected": -0.7795466184616089, | |
| "logps/chosen": -1022.5738525390625, | |
| "logps/rejected": -1272.192626953125, | |
| "loss": 0.6826, | |
| "rewards/accuracies": 0.7749999761581421, | |
| "rewards/chosen": 0.01184375025331974, | |
| "rewards/margins": 0.019737884402275085, | |
| "rewards/rejected": -0.007894134148955345, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 4.67072465163404, | |
| "learning_rate": 4.561687510272767e-07, | |
| "logits/chosen": -0.7071916460990906, | |
| "logits/rejected": -0.7952053546905518, | |
| "logps/chosen": -973.41259765625, | |
| "logps/rejected": -1254.411376953125, | |
| "loss": 0.6759, | |
| "rewards/accuracies": 0.75, | |
| "rewards/chosen": 0.010998593643307686, | |
| "rewards/margins": 0.031063910573720932, | |
| "rewards/rejected": -0.020065316930413246, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 4.340731797997725, | |
| "learning_rate": 4.2505433694179213e-07, | |
| "logits/chosen": -0.7227746248245239, | |
| "logits/rejected": -0.8140878677368164, | |
| "logps/chosen": -1035.677734375, | |
| "logps/rejected": -1306.7808837890625, | |
| "loss": 0.6642, | |
| "rewards/accuracies": 0.8687499761581421, | |
| "rewards/chosen": 0.017496155574917793, | |
| "rewards/margins": 0.06553898006677628, | |
| "rewards/rejected": -0.04804282635450363, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 4.496628183792233, | |
| "learning_rate": 3.873772445177015e-07, | |
| "logits/chosen": -0.7832368612289429, | |
| "logits/rejected": -0.7676926851272583, | |
| "logps/chosen": -1028.62890625, | |
| "logps/rejected": -1254.6610107421875, | |
| "loss": 0.6538, | |
| "rewards/accuracies": 0.762499988079071, | |
| "rewards/chosen": 0.01262822188436985, | |
| "rewards/margins": 0.07675839215517044, | |
| "rewards/rejected": -0.06413016468286514, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 4.939508062028092, | |
| "learning_rate": 3.445499645429106e-07, | |
| "logits/chosen": -0.7500189542770386, | |
| "logits/rejected": -0.8156210780143738, | |
| "logps/chosen": -1028.065185546875, | |
| "logps/rejected": -1316.736328125, | |
| "loss": 0.6433, | |
| "rewards/accuracies": 0.8062499761581421, | |
| "rewards/chosen": 0.0009801089763641357, | |
| "rewards/margins": 0.0968375876545906, | |
| "rewards/rejected": -0.09585747867822647, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 4.7399950008550205, | |
| "learning_rate": 2.981780651370224e-07, | |
| "logits/chosen": -0.7468028664588928, | |
| "logits/rejected": -0.7884877920150757, | |
| "logps/chosen": -996.8697509765625, | |
| "logps/rejected": -1192.45654296875, | |
| "loss": 0.6079, | |
| "rewards/accuracies": 0.8187500238418579, | |
| "rewards/chosen": 0.004719099495559931, | |
| "rewards/margins": 0.21119482815265656, | |
| "rewards/rejected": -0.20647573471069336, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 5.466151335512206, | |
| "learning_rate": 2.5e-07, | |
| "logits/chosen": -0.7582017183303833, | |
| "logits/rejected": -0.8466933369636536, | |
| "logps/chosen": -979.6242065429688, | |
| "logps/rejected": -1357.511962890625, | |
| "loss": 0.592, | |
| "rewards/accuracies": 0.8187500238418579, | |
| "rewards/chosen": -0.00921861082315445, | |
| "rewards/margins": 0.2683759033679962, | |
| "rewards/rejected": -0.27759450674057007, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "eval_logits/chosen": -0.7069126963615417, | |
| "eval_logits/rejected": -0.8721536993980408, | |
| "eval_logps/chosen": -955.1961059570312, | |
| "eval_logps/rejected": -1615.9989013671875, | |
| "eval_loss": 0.6292956471443176, | |
| "eval_rewards/accuracies": 0.7796609997749329, | |
| "eval_rewards/chosen": -0.0145129868760705, | |
| "eval_rewards/margins": 0.14818498492240906, | |
| "eval_rewards/rejected": -0.16269797086715698, | |
| "eval_runtime": 214.9002, | |
| "eval_samples_per_second": 8.674, | |
| "eval_steps_per_second": 0.275, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 4.779970412172738, | |
| "learning_rate": 2.018219348629775e-07, | |
| "logits/chosen": -0.779070258140564, | |
| "logits/rejected": -0.8311734199523926, | |
| "logps/chosen": -992.7100830078125, | |
| "logps/rejected": -1354.688720703125, | |
| "loss": 0.5575, | |
| "rewards/accuracies": 0.856249988079071, | |
| "rewards/chosen": -0.0526651069521904, | |
| "rewards/margins": 0.4234324097633362, | |
| "rewards/rejected": -0.4760975241661072, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 5.9180549707398695, | |
| "learning_rate": 1.554500354570894e-07, | |
| "logits/chosen": -0.7683907747268677, | |
| "logits/rejected": -0.8335941433906555, | |
| "logps/chosen": -1003.9620971679688, | |
| "logps/rejected": -1312.5740966796875, | |
| "loss": 0.5428, | |
| "rewards/accuracies": 0.8999999761581421, | |
| "rewards/chosen": -0.06494404375553131, | |
| "rewards/margins": 0.5937026739120483, | |
| "rewards/rejected": -0.6586467623710632, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 5.7132927229572354, | |
| "learning_rate": 1.126227554822985e-07, | |
| "logits/chosen": -0.798209011554718, | |
| "logits/rejected": -0.8723245859146118, | |
| "logps/chosen": -1054.9263916015625, | |
| "logps/rejected": -1349.2750244140625, | |
| "loss": 0.5312, | |
| "rewards/accuracies": 0.8125, | |
| "rewards/chosen": -0.15924784541130066, | |
| "rewards/margins": 0.5923423767089844, | |
| "rewards/rejected": -0.7515901923179626, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 10.497632384133095, | |
| "learning_rate": 7.494566305820788e-08, | |
| "logits/chosen": -0.8022674322128296, | |
| "logits/rejected": -0.882843017578125, | |
| "logps/chosen": -1055.48974609375, | |
| "logps/rejected": -1417.170654296875, | |
| "loss": 0.4988, | |
| "rewards/accuracies": 0.862500011920929, | |
| "rewards/chosen": -0.17785976827144623, | |
| "rewards/margins": 1.1309340000152588, | |
| "rewards/rejected": -1.3087936639785767, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 6.590111680094942, | |
| "learning_rate": 4.38312489727233e-08, | |
| "logits/chosen": -0.7941451668739319, | |
| "logits/rejected": -0.8360635042190552, | |
| "logps/chosen": -994.5494384765625, | |
| "logps/rejected": -1374.3560791015625, | |
| "loss": 0.4917, | |
| "rewards/accuracies": 0.8687499761581421, | |
| "rewards/chosen": -0.2062647044658661, | |
| "rewards/margins": 0.8967850804328918, | |
| "rewards/rejected": -1.1030497550964355, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 8.361359149599993, | |
| "learning_rate": 2.044597327993153e-08, | |
| "logits/chosen": -0.7812503576278687, | |
| "logits/rejected": -0.8671078681945801, | |
| "logps/chosen": -1042.375244140625, | |
| "logps/rejected": -1443.9189453125, | |
| "loss": 0.4813, | |
| "rewards/accuracies": 0.8374999761581421, | |
| "rewards/chosen": -0.2346261441707611, | |
| "rewards/margins": 0.8702405095100403, | |
| "rewards/rejected": -1.104866623878479, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 7.649602408833162, | |
| "learning_rate": 5.666535437341108e-09, | |
| "logits/chosen": -0.792350172996521, | |
| "logits/rejected": -0.8578357696533203, | |
| "logps/chosen": -1057.263427734375, | |
| "logps/rejected": -1363.985107421875, | |
| "loss": 0.4843, | |
| "rewards/accuracies": 0.8062499761581421, | |
| "rewards/chosen": -0.292695015668869, | |
| "rewards/margins": 1.1091253757476807, | |
| "rewards/rejected": -1.401820421218872, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 8.116244988839705, | |
| "learning_rate": 4.700738787466463e-11, | |
| "logits/chosen": -0.7855504751205444, | |
| "logits/rejected": -0.8505135774612427, | |
| "logps/chosen": -952.7326049804688, | |
| "logps/rejected": -1447.79541015625, | |
| "loss": 0.4771, | |
| "rewards/accuracies": 0.800000011920929, | |
| "rewards/chosen": -0.23469653725624084, | |
| "rewards/margins": 0.9054014086723328, | |
| "rewards/rejected": -1.140097975730896, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 181, | |
| "total_flos": 0.0, | |
| "train_loss": 0.5914445311983646, | |
| "train_runtime": 2922.1625, | |
| "train_samples_per_second": 3.971, | |
| "train_steps_per_second": 0.062 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 181, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "total_flos": 0.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |