| { | |
| "best_metric": 1.1799217587629869, | |
| "best_model_checkpoint": "results/checkpoints/mlm/v4_v6/512_64_1024/checkpoint-300000", | |
| "epoch": 68.000256, | |
| "eval_steps": 10000, | |
| "global_step": 300000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 2.001184, | |
| "grad_norm": 0.22232888638973236, | |
| "learning_rate": 0.001, | |
| "loss": 1.2046, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 2.001184, | |
| "eval_loss": 1.2169866000817773, | |
| "eval_runtime": 9.4105, | |
| "eval_samples_per_second": 233.569, | |
| "eval_steps_per_second": 0.956, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 4.002368, | |
| "grad_norm": 0.029476888477802277, | |
| "learning_rate": 0.001, | |
| "loss": 1.1796, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 4.002368, | |
| "eval_loss": 1.2071053582668596, | |
| "eval_runtime": 8.7108, | |
| "eval_samples_per_second": 252.33, | |
| "eval_steps_per_second": 1.033, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 6.003552, | |
| "grad_norm": 0.020134098827838898, | |
| "learning_rate": 0.001, | |
| "loss": 1.1695, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 6.003552, | |
| "eval_loss": 1.2035266431563263, | |
| "eval_runtime": 8.7636, | |
| "eval_samples_per_second": 250.809, | |
| "eval_steps_per_second": 1.027, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 9.000328, | |
| "grad_norm": 0.013419270515441895, | |
| "learning_rate": 0.001, | |
| "loss": 1.1623, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 9.000328, | |
| "eval_loss": 1.199994934919684, | |
| "eval_runtime": 8.7143, | |
| "eval_samples_per_second": 252.228, | |
| "eval_steps_per_second": 1.033, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 11.001512, | |
| "grad_norm": 0.012691290117800236, | |
| "learning_rate": 0.001, | |
| "loss": 1.1569, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 11.001512, | |
| "eval_loss": 1.196269909022808, | |
| "eval_runtime": 8.6947, | |
| "eval_samples_per_second": 252.798, | |
| "eval_steps_per_second": 1.035, | |
| "step": 50000 | |
| }, | |
| { | |
| "epoch": 13.002696, | |
| "grad_norm": 0.013640903867781162, | |
| "learning_rate": 0.001, | |
| "loss": 1.1525, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 13.002696, | |
| "eval_loss": 1.1957027129989803, | |
| "eval_runtime": 8.8343, | |
| "eval_samples_per_second": 248.802, | |
| "eval_steps_per_second": 1.019, | |
| "step": 60000 | |
| }, | |
| { | |
| "epoch": 15.00388, | |
| "grad_norm": 0.013336733914911747, | |
| "learning_rate": 0.001, | |
| "loss": 1.1487, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 15.00388, | |
| "eval_loss": 1.1923882669474761, | |
| "eval_runtime": 8.7699, | |
| "eval_samples_per_second": 250.63, | |
| "eval_steps_per_second": 1.026, | |
| "step": 70000 | |
| }, | |
| { | |
| "epoch": 18.000656, | |
| "grad_norm": 0.013540836051106453, | |
| "learning_rate": 0.001, | |
| "loss": 1.1454, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 18.000656, | |
| "eval_loss": 1.1914640151582543, | |
| "eval_runtime": 8.6988, | |
| "eval_samples_per_second": 252.679, | |
| "eval_steps_per_second": 1.035, | |
| "step": 80000 | |
| }, | |
| { | |
| "epoch": 20.00184, | |
| "grad_norm": 0.013072439469397068, | |
| "learning_rate": 0.001, | |
| "loss": 1.1425, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 20.00184, | |
| "eval_loss": 1.192123629600482, | |
| "eval_runtime": 8.6374, | |
| "eval_samples_per_second": 254.476, | |
| "eval_steps_per_second": 1.042, | |
| "step": 90000 | |
| }, | |
| { | |
| "epoch": 22.003024, | |
| "grad_norm": 0.01519271731376648, | |
| "learning_rate": 0.001, | |
| "loss": 1.1399, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 22.003024, | |
| "eval_loss": 1.1890448101005333, | |
| "eval_runtime": 8.7241, | |
| "eval_samples_per_second": 251.945, | |
| "eval_steps_per_second": 1.032, | |
| "step": 100000 | |
| }, | |
| { | |
| "epoch": 24.004208, | |
| "grad_norm": 0.013080044649541378, | |
| "learning_rate": 0.001, | |
| "loss": 1.1376, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 24.004208, | |
| "eval_loss": 1.1863106067113791, | |
| "eval_runtime": 8.7128, | |
| "eval_samples_per_second": 252.271, | |
| "eval_steps_per_second": 1.033, | |
| "step": 110000 | |
| }, | |
| { | |
| "epoch": 27.000984, | |
| "grad_norm": 0.012146814726293087, | |
| "learning_rate": 0.001, | |
| "loss": 1.1355, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 27.000984, | |
| "eval_loss": 1.188288665043113, | |
| "eval_runtime": 8.76, | |
| "eval_samples_per_second": 250.912, | |
| "eval_steps_per_second": 1.027, | |
| "step": 120000 | |
| }, | |
| { | |
| "epoch": 29.002168, | |
| "grad_norm": 0.012161113321781158, | |
| "learning_rate": 0.001, | |
| "loss": 1.1336, | |
| "step": 130000 | |
| }, | |
| { | |
| "epoch": 29.002168, | |
| "eval_loss": 1.187914588210265, | |
| "eval_runtime": 8.7346, | |
| "eval_samples_per_second": 251.644, | |
| "eval_steps_per_second": 1.03, | |
| "step": 130000 | |
| }, | |
| { | |
| "epoch": 31.003352, | |
| "grad_norm": 0.013448276557028294, | |
| "learning_rate": 0.001, | |
| "loss": 1.1319, | |
| "step": 140000 | |
| }, | |
| { | |
| "epoch": 31.003352, | |
| "eval_loss": 1.185675648549141, | |
| "eval_runtime": 8.9763, | |
| "eval_samples_per_second": 244.866, | |
| "eval_steps_per_second": 1.003, | |
| "step": 140000 | |
| }, | |
| { | |
| "epoch": 34.000128, | |
| "grad_norm": 0.014190836809575558, | |
| "learning_rate": 0.001, | |
| "loss": 1.1303, | |
| "step": 150000 | |
| }, | |
| { | |
| "epoch": 34.000128, | |
| "eval_loss": 1.1869828313599362, | |
| "eval_runtime": 9.017, | |
| "eval_samples_per_second": 243.761, | |
| "eval_steps_per_second": 0.998, | |
| "step": 150000 | |
| }, | |
| { | |
| "epoch": 36.001312, | |
| "grad_norm": 0.017434025183320045, | |
| "learning_rate": 0.001, | |
| "loss": 1.1288, | |
| "step": 160000 | |
| }, | |
| { | |
| "epoch": 36.001312, | |
| "eval_loss": 1.1848532211813299, | |
| "eval_runtime": 9.2402, | |
| "eval_samples_per_second": 237.873, | |
| "eval_steps_per_second": 0.974, | |
| "step": 160000 | |
| }, | |
| { | |
| "epoch": 38.002496, | |
| "grad_norm": 0.014645294286310673, | |
| "learning_rate": 0.001, | |
| "loss": 1.1274, | |
| "step": 170000 | |
| }, | |
| { | |
| "epoch": 38.002496, | |
| "eval_loss": 1.1855380815257557, | |
| "eval_runtime": 8.7057, | |
| "eval_samples_per_second": 252.478, | |
| "eval_steps_per_second": 1.034, | |
| "step": 170000 | |
| }, | |
| { | |
| "epoch": 40.00368, | |
| "grad_norm": 0.014773557893931866, | |
| "learning_rate": 0.001, | |
| "loss": 1.1261, | |
| "step": 180000 | |
| }, | |
| { | |
| "epoch": 40.00368, | |
| "eval_loss": 1.184328326940904, | |
| "eval_runtime": 8.717, | |
| "eval_samples_per_second": 252.152, | |
| "eval_steps_per_second": 1.032, | |
| "step": 180000 | |
| }, | |
| { | |
| "epoch": 43.000456, | |
| "grad_norm": 0.01418334897607565, | |
| "learning_rate": 0.001, | |
| "loss": 1.125, | |
| "step": 190000 | |
| }, | |
| { | |
| "epoch": 43.000456, | |
| "eval_loss": 1.1831796657343023, | |
| "eval_runtime": 8.8653, | |
| "eval_samples_per_second": 247.933, | |
| "eval_steps_per_second": 1.015, | |
| "step": 190000 | |
| }, | |
| { | |
| "epoch": 45.00164, | |
| "grad_norm": 0.015187916345894337, | |
| "learning_rate": 0.001, | |
| "loss": 1.1239, | |
| "step": 200000 | |
| }, | |
| { | |
| "epoch": 45.00164, | |
| "eval_loss": 1.183707485496527, | |
| "eval_runtime": 8.752, | |
| "eval_samples_per_second": 251.142, | |
| "eval_steps_per_second": 1.028, | |
| "step": 200000 | |
| }, | |
| { | |
| "epoch": 47.002824, | |
| "grad_norm": 0.01608165167272091, | |
| "learning_rate": 0.001, | |
| "loss": 1.1228, | |
| "step": 210000 | |
| }, | |
| { | |
| "epoch": 47.002824, | |
| "eval_loss": 1.1836327739651904, | |
| "eval_runtime": 8.8678, | |
| "eval_samples_per_second": 247.864, | |
| "eval_steps_per_second": 1.015, | |
| "step": 210000 | |
| }, | |
| { | |
| "epoch": 49.004008, | |
| "grad_norm": 0.014980562962591648, | |
| "learning_rate": 0.001, | |
| "loss": 1.1219, | |
| "step": 220000 | |
| }, | |
| { | |
| "epoch": 49.004008, | |
| "eval_loss": 1.182785976935125, | |
| "eval_runtime": 8.7558, | |
| "eval_samples_per_second": 251.035, | |
| "eval_steps_per_second": 1.028, | |
| "step": 220000 | |
| }, | |
| { | |
| "epoch": 52.000784, | |
| "grad_norm": 0.014229262247681618, | |
| "learning_rate": 0.001, | |
| "loss": 1.121, | |
| "step": 230000 | |
| }, | |
| { | |
| "epoch": 52.000784, | |
| "eval_loss": 1.1820512524995475, | |
| "eval_runtime": 8.7585, | |
| "eval_samples_per_second": 250.956, | |
| "eval_steps_per_second": 1.028, | |
| "step": 230000 | |
| }, | |
| { | |
| "epoch": 54.001968, | |
| "grad_norm": 0.015481743030250072, | |
| "learning_rate": 0.001, | |
| "loss": 1.1201, | |
| "step": 240000 | |
| }, | |
| { | |
| "epoch": 54.001968, | |
| "eval_loss": 1.184246257349048, | |
| "eval_runtime": 8.6149, | |
| "eval_samples_per_second": 255.14, | |
| "eval_steps_per_second": 1.045, | |
| "step": 240000 | |
| }, | |
| { | |
| "epoch": 56.003152, | |
| "grad_norm": 0.01397514808923006, | |
| "learning_rate": 0.001, | |
| "loss": 1.1193, | |
| "step": 250000 | |
| }, | |
| { | |
| "epoch": 56.003152, | |
| "eval_loss": 1.1817655188913954, | |
| "eval_runtime": 8.7701, | |
| "eval_samples_per_second": 250.626, | |
| "eval_steps_per_second": 1.026, | |
| "step": 250000 | |
| }, | |
| { | |
| "epoch": 58.004336, | |
| "grad_norm": 0.01819504424929619, | |
| "learning_rate": 0.001, | |
| "loss": 1.1186, | |
| "step": 260000 | |
| }, | |
| { | |
| "epoch": 58.004336, | |
| "eval_loss": 1.1822551848024445, | |
| "eval_runtime": 8.7888, | |
| "eval_samples_per_second": 250.091, | |
| "eval_steps_per_second": 1.024, | |
| "step": 260000 | |
| }, | |
| { | |
| "epoch": 61.001112, | |
| "grad_norm": 0.014879155904054642, | |
| "learning_rate": 0.001, | |
| "loss": 1.1178, | |
| "step": 270000 | |
| }, | |
| { | |
| "epoch": 61.001112, | |
| "eval_loss": 1.181130289503475, | |
| "eval_runtime": 8.7683, | |
| "eval_samples_per_second": 250.676, | |
| "eval_steps_per_second": 1.026, | |
| "step": 270000 | |
| }, | |
| { | |
| "epoch": 63.002296, | |
| "grad_norm": 0.015341316349804401, | |
| "learning_rate": 0.001, | |
| "loss": 1.1172, | |
| "step": 280000 | |
| }, | |
| { | |
| "epoch": 63.002296, | |
| "eval_loss": 1.180569705755674, | |
| "eval_runtime": 8.8978, | |
| "eval_samples_per_second": 247.027, | |
| "eval_steps_per_second": 1.011, | |
| "step": 280000 | |
| }, | |
| { | |
| "epoch": 65.00348, | |
| "grad_norm": 0.015073459595441818, | |
| "learning_rate": 0.001, | |
| "loss": 1.1165, | |
| "step": 290000 | |
| }, | |
| { | |
| "epoch": 65.00348, | |
| "eval_loss": 1.1800268809816898, | |
| "eval_runtime": 8.7943, | |
| "eval_samples_per_second": 249.933, | |
| "eval_steps_per_second": 1.023, | |
| "step": 290000 | |
| }, | |
| { | |
| "epoch": 68.000256, | |
| "grad_norm": 0.014499953016638756, | |
| "learning_rate": 0.001, | |
| "loss": 1.1159, | |
| "step": 300000 | |
| }, | |
| { | |
| "epoch": 68.000256, | |
| "eval_loss": 1.1799217587629869, | |
| "eval_runtime": 8.8138, | |
| "eval_samples_per_second": 249.381, | |
| "eval_steps_per_second": 1.021, | |
| "step": 300000 | |
| } | |
| ], | |
| "logging_steps": 10000, | |
| "max_steps": 1000000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 9223372036854775807, | |
| "save_steps": 10000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.872353950943884e+20, | |
| "train_batch_size": 64, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |