| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.0, | |
| "eval_steps": 500, | |
| "global_step": 784, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.002551020408163265, | |
| "grad_norm": 1.138124187948155, | |
| "learning_rate": 1.2658227848101267e-05, | |
| "loss": 1.7551, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.012755102040816327, | |
| "grad_norm": 0.7551830131389933, | |
| "learning_rate": 6.329113924050633e-05, | |
| "loss": 1.7707, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.025510204081632654, | |
| "grad_norm": 1.379764699722761, | |
| "learning_rate": 0.00012658227848101267, | |
| "loss": 1.7257, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.03826530612244898, | |
| "grad_norm": 0.4560275436076655, | |
| "learning_rate": 0.000189873417721519, | |
| "loss": 1.6272, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.05102040816326531, | |
| "grad_norm": 0.3675085939167979, | |
| "learning_rate": 0.00025316455696202533, | |
| "loss": 1.5808, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.06377551020408163, | |
| "grad_norm": 0.20864289230413074, | |
| "learning_rate": 0.00031645569620253165, | |
| "loss": 1.5291, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.07653061224489796, | |
| "grad_norm": 0.18719394627139643, | |
| "learning_rate": 0.000379746835443038, | |
| "loss": 1.5338, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.08928571428571429, | |
| "grad_norm": 0.1717343346810603, | |
| "learning_rate": 0.0004430379746835443, | |
| "loss": 1.5007, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.10204081632653061, | |
| "grad_norm": 0.15610479873752922, | |
| "learning_rate": 0.0005063291139240507, | |
| "loss": 1.4652, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.11479591836734694, | |
| "grad_norm": 0.13214894352385953, | |
| "learning_rate": 0.000569620253164557, | |
| "loss": 1.4388, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.12755102040816327, | |
| "grad_norm": 0.12168972456032588, | |
| "learning_rate": 0.0006329113924050633, | |
| "loss": 1.4225, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.14030612244897958, | |
| "grad_norm": 0.1323077190059568, | |
| "learning_rate": 0.0006962025316455697, | |
| "loss": 1.409, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.15306122448979592, | |
| "grad_norm": 0.37430096940846036, | |
| "learning_rate": 0.000759493670886076, | |
| "loss": 1.3951, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.16581632653061223, | |
| "grad_norm": 0.3814386987087306, | |
| "learning_rate": 0.0008227848101265824, | |
| "loss": 1.3972, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.17857142857142858, | |
| "grad_norm": 0.40216892416078964, | |
| "learning_rate": 0.0008860759493670886, | |
| "loss": 1.3864, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.1913265306122449, | |
| "grad_norm": 0.20332439264842958, | |
| "learning_rate": 0.0009493670886075949, | |
| "loss": 1.3861, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.20408163265306123, | |
| "grad_norm": 0.2924926145777638, | |
| "learning_rate": 0.0009999950356681913, | |
| "loss": 1.3891, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.21683673469387754, | |
| "grad_norm": 0.29723891288444315, | |
| "learning_rate": 0.000999821294405392, | |
| "loss": 1.3982, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.22959183673469388, | |
| "grad_norm": 0.23043075033731064, | |
| "learning_rate": 0.0009993994351217151, | |
| "loss": 1.3821, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.2423469387755102, | |
| "grad_norm": 0.15934631204828656, | |
| "learning_rate": 0.00099872966723379, | |
| "loss": 1.3937, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.25510204081632654, | |
| "grad_norm": 0.19636795516281652, | |
| "learning_rate": 0.0009978123232234147, | |
| "loss": 1.3393, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.26785714285714285, | |
| "grad_norm": 0.3357008076473798, | |
| "learning_rate": 0.0009966478584725086, | |
| "loss": 1.3873, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.28061224489795916, | |
| "grad_norm": 0.2979357511561018, | |
| "learning_rate": 0.0009952368510370538, | |
| "loss": 1.3575, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.29336734693877553, | |
| "grad_norm": 0.23158016835002, | |
| "learning_rate": 0.0009935800013601416, | |
| "loss": 1.3614, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.30612244897959184, | |
| "grad_norm": 0.19086278999416814, | |
| "learning_rate": 0.0009916781319242614, | |
| "loss": 1.3609, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.31887755102040816, | |
| "grad_norm": 0.2238120900111794, | |
| "learning_rate": 0.0009895321868430113, | |
| "loss": 1.3436, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.33163265306122447, | |
| "grad_norm": 0.22212987754769067, | |
| "learning_rate": 0.0009871432313924254, | |
| "loss": 1.3356, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.34438775510204084, | |
| "grad_norm": 0.18172190875737002, | |
| "learning_rate": 0.000984512451482158, | |
| "loss": 1.3428, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.35714285714285715, | |
| "grad_norm": 0.1965103083782127, | |
| "learning_rate": 0.0009816411530667814, | |
| "loss": 1.3053, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.36989795918367346, | |
| "grad_norm": 0.14562927658303307, | |
| "learning_rate": 0.000978530761497492, | |
| "loss": 1.3233, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.3826530612244898, | |
| "grad_norm": 0.19370329245107856, | |
| "learning_rate": 0.0009751828208145482, | |
| "loss": 1.3309, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.39540816326530615, | |
| "grad_norm": 0.25874954807518885, | |
| "learning_rate": 0.0009715989929807862, | |
| "loss": 1.3386, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.40816326530612246, | |
| "grad_norm": 0.32074900532811484, | |
| "learning_rate": 0.000967781057056601, | |
| "loss": 1.3197, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.42091836734693877, | |
| "grad_norm": 0.2345834124220609, | |
| "learning_rate": 0.0009637309083167956, | |
| "loss": 1.3353, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.4336734693877551, | |
| "grad_norm": 0.19716142253846314, | |
| "learning_rate": 0.0009594505573097414, | |
| "loss": 1.3148, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.44642857142857145, | |
| "grad_norm": 1.6838448936670876, | |
| "learning_rate": 0.0009549421288593157, | |
| "loss": 1.4963, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.45918367346938777, | |
| "grad_norm": 0.5400080339132108, | |
| "learning_rate": 0.0009502078610101092, | |
| "loss": 1.3763, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.4719387755102041, | |
| "grad_norm": 0.5636053179876761, | |
| "learning_rate": 0.0009452501039164315, | |
| "loss": 1.3562, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.4846938775510204, | |
| "grad_norm": 0.26363289666703416, | |
| "learning_rate": 0.0009400713186756625, | |
| "loss": 1.3671, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.49744897959183676, | |
| "grad_norm": 0.2063220498775075, | |
| "learning_rate": 0.0009346740761065305, | |
| "loss": 1.337, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.5102040816326531, | |
| "grad_norm": 0.851877134215801, | |
| "learning_rate": 0.0009290610554729234, | |
| "loss": 1.3404, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.5229591836734694, | |
| "grad_norm": 0.17433687232665127, | |
| "learning_rate": 0.0009232350431538657, | |
| "loss": 1.3278, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.5357142857142857, | |
| "grad_norm": 0.2552513397179755, | |
| "learning_rate": 0.0009171989312603226, | |
| "loss": 1.3621, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.548469387755102, | |
| "grad_norm": 0.18498941711074082, | |
| "learning_rate": 0.0009109557161995172, | |
| "loss": 1.3365, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.5612244897959183, | |
| "grad_norm": 0.19451053194971357, | |
| "learning_rate": 0.0009045084971874737, | |
| "loss": 1.3329, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.5739795918367347, | |
| "grad_norm": 0.14628015588879814, | |
| "learning_rate": 0.0008978604747105246, | |
| "loss": 1.3133, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.5867346938775511, | |
| "grad_norm": 0.2028020969079187, | |
| "learning_rate": 0.000891014948936546, | |
| "loss": 1.3337, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.5994897959183674, | |
| "grad_norm": 0.18332572064094557, | |
| "learning_rate": 0.0008839753180767108, | |
| "loss": 1.3132, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.6122448979591837, | |
| "grad_norm": 0.14205229763128208, | |
| "learning_rate": 0.0008767450766985694, | |
| "loss": 1.29, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.625, | |
| "grad_norm": 0.30410202409557807, | |
| "learning_rate": 0.000869327813991301, | |
| "loss": 1.3057, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.6377551020408163, | |
| "grad_norm": 0.19577309506455626, | |
| "learning_rate": 0.0008617272119839903, | |
| "loss": 1.332, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.6505102040816326, | |
| "grad_norm": 0.1272875527576836, | |
| "learning_rate": 0.0008539470437178196, | |
| "loss": 1.3206, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.6632653061224489, | |
| "grad_norm": 0.15194088157912253, | |
| "learning_rate": 0.0008459911713730799, | |
| "loss": 1.2914, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.6760204081632653, | |
| "grad_norm": 0.12663044164256262, | |
| "learning_rate": 0.0008378635443519327, | |
| "loss": 1.2917, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.6887755102040817, | |
| "grad_norm": 0.1344562065693353, | |
| "learning_rate": 0.0008295681973178737, | |
| "loss": 1.2994, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.701530612244898, | |
| "grad_norm": 0.13392241997034995, | |
| "learning_rate": 0.0008211092481928716, | |
| "loss": 1.297, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.14371429407675881, | |
| "learning_rate": 0.0008124908961131759, | |
| "loss": 1.3095, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.7270408163265306, | |
| "grad_norm": 0.15013152147820535, | |
| "learning_rate": 0.0008037174193448089, | |
| "loss": 1.2991, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.7397959183673469, | |
| "grad_norm": 0.13870289504861724, | |
| "learning_rate": 0.000794793173159778, | |
| "loss": 1.3066, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.7525510204081632, | |
| "grad_norm": 0.11463233811669157, | |
| "learning_rate": 0.0007857225876740584, | |
| "loss": 1.2914, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.7653061224489796, | |
| "grad_norm": 0.16118564397614718, | |
| "learning_rate": 0.000776510165648425, | |
| "loss": 1.286, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.7780612244897959, | |
| "grad_norm": 0.16328954399209802, | |
| "learning_rate": 0.000767160480253221, | |
| "loss": 1.3065, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.7908163265306123, | |
| "grad_norm": 0.3891017035371994, | |
| "learning_rate": 0.000757678172798175, | |
| "loss": 1.2978, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.8035714285714286, | |
| "grad_norm": 0.1374369729620582, | |
| "learning_rate": 0.0007480679504283911, | |
| "loss": 1.2737, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.8163265306122449, | |
| "grad_norm": 0.10483812728418156, | |
| "learning_rate": 0.00073833458378766, | |
| "loss": 1.2638, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.8290816326530612, | |
| "grad_norm": 0.20381163662422036, | |
| "learning_rate": 0.0007284829046502467, | |
| "loss": 1.2888, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.8418367346938775, | |
| "grad_norm": 0.11728927247718034, | |
| "learning_rate": 0.0007185178035223327, | |
| "loss": 1.2988, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.8545918367346939, | |
| "grad_norm": 0.10959677714791609, | |
| "learning_rate": 0.0007084442272143026, | |
| "loss": 1.2722, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.8673469387755102, | |
| "grad_norm": 0.1156080310279316, | |
| "learning_rate": 0.0006982671763850814, | |
| "loss": 1.2831, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.8801020408163265, | |
| "grad_norm": 0.1401568063137037, | |
| "learning_rate": 0.0006879917030597397, | |
| "loss": 1.2766, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.8928571428571429, | |
| "grad_norm": 0.1267536193584464, | |
| "learning_rate": 0.0006776229081216001, | |
| "loss": 1.2909, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.9056122448979592, | |
| "grad_norm": 0.1276445520009822, | |
| "learning_rate": 0.0006671659387800909, | |
| "loss": 1.271, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.9183673469387755, | |
| "grad_norm": 0.12028703336124712, | |
| "learning_rate": 0.0006566259860156014, | |
| "loss": 1.2681, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.9311224489795918, | |
| "grad_norm": 0.10748771516704643, | |
| "learning_rate": 0.0006460082820026094, | |
| "loss": 1.2634, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.9438775510204082, | |
| "grad_norm": 0.09058225138342332, | |
| "learning_rate": 0.0006353180975123595, | |
| "loss": 1.2785, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.9566326530612245, | |
| "grad_norm": 0.1471646323105399, | |
| "learning_rate": 0.000624560739296381, | |
| "loss": 1.2537, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.9693877551020408, | |
| "grad_norm": 0.12205622985562622, | |
| "learning_rate": 0.0006137415474521454, | |
| "loss": 1.2834, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.9821428571428571, | |
| "grad_norm": 0.1198227865731359, | |
| "learning_rate": 0.0006028658927721697, | |
| "loss": 1.268, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.9948979591836735, | |
| "grad_norm": 0.10689994479443513, | |
| "learning_rate": 0.0005919391740778833, | |
| "loss": 1.2705, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 1.864872694015503, | |
| "eval_runtime": 85.1881, | |
| "eval_samples_per_second": 152.568, | |
| "eval_steps_per_second": 4.778, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 1.0076530612244898, | |
| "grad_norm": 0.11143268558344442, | |
| "learning_rate": 0.0005809668155395793, | |
| "loss": 1.2488, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.0204081632653061, | |
| "grad_norm": 0.10688715702524573, | |
| "learning_rate": 0.0005699542639837844, | |
| "loss": 1.2129, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.0331632653061225, | |
| "grad_norm": 0.09930208201709802, | |
| "learning_rate": 0.0005589069861893798, | |
| "loss": 1.2152, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.0459183673469388, | |
| "grad_norm": 0.14907362895999168, | |
| "learning_rate": 0.0005478304661738199, | |
| "loss": 1.2204, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.058673469387755, | |
| "grad_norm": 0.12329968075769418, | |
| "learning_rate": 0.000536730202470791, | |
| "loss": 1.248, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.0714285714285714, | |
| "grad_norm": 0.10929541376542931, | |
| "learning_rate": 0.000525611705400666, | |
| "loss": 1.2511, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.0841836734693877, | |
| "grad_norm": 0.12233961076169748, | |
| "learning_rate": 0.000514480494335106, | |
| "loss": 1.2165, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.096938775510204, | |
| "grad_norm": 0.11362833380867698, | |
| "learning_rate": 0.0005033420949571712, | |
| "loss": 1.2353, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.1096938775510203, | |
| "grad_norm": 0.09720826313265056, | |
| "learning_rate": 0.0004922020365182968, | |
| "loss": 1.209, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.1224489795918366, | |
| "grad_norm": 0.1184275991715705, | |
| "learning_rate": 0.0004810658490934979, | |
| "loss": 1.2218, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.135204081632653, | |
| "grad_norm": 0.13198322755570885, | |
| "learning_rate": 0.0004699390608361665, | |
| "loss": 1.2035, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.1479591836734695, | |
| "grad_norm": 0.10874477518300978, | |
| "learning_rate": 0.0004588271952338212, | |
| "loss": 1.2172, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.1607142857142858, | |
| "grad_norm": 0.08852135264878797, | |
| "learning_rate": 0.00044773576836617336, | |
| "loss": 1.221, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.1734693877551021, | |
| "grad_norm": 0.10770572502518004, | |
| "learning_rate": 0.0004366702861668716, | |
| "loss": 1.2192, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.1862244897959184, | |
| "grad_norm": 0.15699433228433193, | |
| "learning_rate": 0.0004256362416902817, | |
| "loss": 1.2204, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.1989795918367347, | |
| "grad_norm": 0.09204241599916607, | |
| "learning_rate": 0.0004146391123846606, | |
| "loss": 1.2338, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.211734693877551, | |
| "grad_norm": 0.08836564479064991, | |
| "learning_rate": 0.00040368435737307733, | |
| "loss": 1.2248, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.2244897959183674, | |
| "grad_norm": 0.09509702477272594, | |
| "learning_rate": 0.00039277741474343054, | |
| "loss": 1.2168, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.2372448979591837, | |
| "grad_norm": 0.0894046781603559, | |
| "learning_rate": 0.00038192369884890886, | |
| "loss": 1.2232, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 0.09858037446272845, | |
| "learning_rate": 0.0003711285976202331, | |
| "loss": 1.2159, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.2627551020408163, | |
| "grad_norm": 0.10699315355393832, | |
| "learning_rate": 0.0003603974698910139, | |
| "loss": 1.2324, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.2755102040816326, | |
| "grad_norm": 0.09869982370664032, | |
| "learning_rate": 0.0003497356427375562, | |
| "loss": 1.2252, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.288265306122449, | |
| "grad_norm": 0.0845449914129218, | |
| "learning_rate": 0.0003391484088344257, | |
| "loss": 1.231, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.3010204081632653, | |
| "grad_norm": 0.09770167031364653, | |
| "learning_rate": 0.00032864102382709374, | |
| "loss": 1.2108, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.3137755102040816, | |
| "grad_norm": 0.08930615418315077, | |
| "learning_rate": 0.0003182187037229653, | |
| "loss": 1.2234, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.3265306122448979, | |
| "grad_norm": 0.11628427629387307, | |
| "learning_rate": 0.0003078866223020815, | |
| "loss": 1.2504, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.3392857142857144, | |
| "grad_norm": 0.08943286693904606, | |
| "learning_rate": 0.0002976499085487862, | |
| "loss": 1.2265, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.3520408163265305, | |
| "grad_norm": 0.10296423428005826, | |
| "learning_rate": 0.0002875136441056286, | |
| "loss": 1.2096, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.364795918367347, | |
| "grad_norm": 0.0881663874621136, | |
| "learning_rate": 0.00027748286075076836, | |
| "loss": 1.2155, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.3775510204081631, | |
| "grad_norm": 0.08112211915193616, | |
| "learning_rate": 0.00026756253790013193, | |
| "loss": 1.2279, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.3903061224489797, | |
| "grad_norm": 0.090452876583666, | |
| "learning_rate": 0.00025775760013556424, | |
| "loss": 1.2176, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.403061224489796, | |
| "grad_norm": 0.08241049413771702, | |
| "learning_rate": 0.00024807291476019994, | |
| "loss": 1.2235, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.4158163265306123, | |
| "grad_norm": 0.0930330073695628, | |
| "learning_rate": 0.00023851328938226808, | |
| "loss": 1.2039, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.4285714285714286, | |
| "grad_norm": 0.07990040558164058, | |
| "learning_rate": 0.0002290834695285316, | |
| "loss": 1.2134, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.441326530612245, | |
| "grad_norm": 0.08688446519906046, | |
| "learning_rate": 0.0002197881362885426, | |
| "loss": 1.1858, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.4540816326530612, | |
| "grad_norm": 0.08906821862405218, | |
| "learning_rate": 0.0002106319039908879, | |
| "loss": 1.2142, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.4668367346938775, | |
| "grad_norm": 0.09353827272017026, | |
| "learning_rate": 0.000201619317912573, | |
| "loss": 1.2176, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.4795918367346939, | |
| "grad_norm": 0.09869151116487236, | |
| "learning_rate": 0.00019275485202268573, | |
| "loss": 1.2111, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.4923469387755102, | |
| "grad_norm": 0.07871336680332862, | |
| "learning_rate": 0.00018404290676145857, | |
| "loss": 1.2009, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.5051020408163265, | |
| "grad_norm": 0.08210906329432778, | |
| "learning_rate": 0.00017548780685582949, | |
| "loss": 1.1945, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.5178571428571428, | |
| "grad_norm": 0.07607843649816663, | |
| "learning_rate": 0.00016709379917259027, | |
| "loss": 1.2013, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.5306122448979593, | |
| "grad_norm": 0.07966793194282133, | |
| "learning_rate": 0.00015886505061018413, | |
| "loss": 1.2372, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.5433673469387754, | |
| "grad_norm": 0.07879844228820579, | |
| "learning_rate": 0.00015080564603020142, | |
| "loss": 1.2051, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.556122448979592, | |
| "grad_norm": 0.07337572605792314, | |
| "learning_rate": 0.00014291958622959973, | |
| "loss": 1.2224, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.568877551020408, | |
| "grad_norm": 0.08717534192433574, | |
| "learning_rate": 0.0001352107859546533, | |
| "loss": 1.1978, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.5816326530612246, | |
| "grad_norm": 0.08338962851545824, | |
| "learning_rate": 0.00012768307195762168, | |
| "loss": 1.1933, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.5943877551020407, | |
| "grad_norm": 0.07500681809171242, | |
| "learning_rate": 0.00012034018109709716, | |
| "loss": 1.1924, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.6071428571428572, | |
| "grad_norm": 0.07959205675721448, | |
| "learning_rate": 0.0001131857584829783, | |
| "loss": 1.2101, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 1.6198979591836735, | |
| "grad_norm": 0.06971085806957752, | |
| "learning_rate": 0.00010622335566698877, | |
| "loss": 1.2142, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 1.6326530612244898, | |
| "grad_norm": 0.06828459321037969, | |
| "learning_rate": 9.94564288796384e-05, | |
| "loss": 1.2119, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 1.6454081632653061, | |
| "grad_norm": 0.07198545650406166, | |
| "learning_rate": 9.288833731450419e-05, | |
| "loss": 1.203, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 1.6581632653061225, | |
| "grad_norm": 0.0765334557855611, | |
| "learning_rate": 8.652234146068206e-05, | |
| "loss": 1.1987, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 1.6709183673469388, | |
| "grad_norm": 0.06996249721515567, | |
| "learning_rate": 8.036160148423449e-05, | |
| "loss": 1.205, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 1.683673469387755, | |
| "grad_norm": 0.07068726506681239, | |
| "learning_rate": 7.440917565944349e-05, | |
| "loss": 1.2168, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 1.6964285714285714, | |
| "grad_norm": 0.06352753939218142, | |
| "learning_rate": 6.866801885064056e-05, | |
| "loss": 1.1967, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 1.7091836734693877, | |
| "grad_norm": 0.06967499508626017, | |
| "learning_rate": 6.314098104537324e-05, | |
| "loss": 1.1747, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 1.7219387755102042, | |
| "grad_norm": 0.06527614945695559, | |
| "learning_rate": 5.783080593963219e-05, | |
| "loss": 1.1991, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 1.7346938775510203, | |
| "grad_norm": 0.06301111901417586, | |
| "learning_rate": 5.27401295758439e-05, | |
| "loss": 1.1996, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 1.7474489795918369, | |
| "grad_norm": 0.06944052501466691, | |
| "learning_rate": 4.787147903430383e-05, | |
| "loss": 1.1968, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 1.760204081632653, | |
| "grad_norm": 0.07413044994704315, | |
| "learning_rate": 4.322727117869951e-05, | |
| "loss": 1.1943, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 1.7729591836734695, | |
| "grad_norm": 0.06365480540319339, | |
| "learning_rate": 3.880981145634704e-05, | |
| "loss": 1.2191, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 1.7857142857142856, | |
| "grad_norm": 0.0713239761471288, | |
| "learning_rate": 3.462129275373577e-05, | |
| "loss": 1.2286, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 1.7984693877551021, | |
| "grad_norm": 0.08049708826438332, | |
| "learning_rate": 3.066379430795002e-05, | |
| "loss": 1.2161, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 1.8112244897959182, | |
| "grad_norm": 0.061716360746827195, | |
| "learning_rate": 2.6939280674508016e-05, | |
| "loss": 1.1999, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 1.8239795918367347, | |
| "grad_norm": 0.0631321070521143, | |
| "learning_rate": 2.3449600752129597e-05, | |
| "loss": 1.2079, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 1.836734693877551, | |
| "grad_norm": 0.0635532701163778, | |
| "learning_rate": 2.019648686491865e-05, | |
| "loss": 1.191, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 1.8494897959183674, | |
| "grad_norm": 0.06640321579684394, | |
| "learning_rate": 1.7181553902413438e-05, | |
| "loss": 1.2255, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 1.8622448979591837, | |
| "grad_norm": 0.06524200918052049, | |
| "learning_rate": 1.4406298517934068e-05, | |
| "loss": 1.2206, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 1.875, | |
| "grad_norm": 0.06477805270118185, | |
| "learning_rate": 1.1872098385623586e-05, | |
| "loss": 1.1887, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 1.8877551020408163, | |
| "grad_norm": 0.06497765016187285, | |
| "learning_rate": 9.580211516551862e-06, | |
| "loss": 1.2036, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 1.9005102040816326, | |
| "grad_norm": 0.06544507847669666, | |
| "learning_rate": 7.531775634222138e-06, | |
| "loss": 1.2224, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 1.913265306122449, | |
| "grad_norm": 0.06400465123455064, | |
| "learning_rate": 5.727807609789471e-06, | |
| "loss": 1.2099, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 1.9260204081632653, | |
| "grad_norm": 0.06135587896382674, | |
| "learning_rate": 4.169202957272522e-06, | |
| "loss": 1.1876, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 1.9387755102040818, | |
| "grad_norm": 0.06395336761154996, | |
| "learning_rate": 2.856735389008269e-06, | |
| "loss": 1.2315, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 1.9515306122448979, | |
| "grad_norm": 0.06406624485413773, | |
| "learning_rate": 1.7910564315704035e-06, | |
| "loss": 1.1941, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 1.9642857142857144, | |
| "grad_norm": 0.06823262080551822, | |
| "learning_rate": 9.726951023434348e-07, | |
| "loss": 1.2067, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 1.9770408163265305, | |
| "grad_norm": 0.06268227825632704, | |
| "learning_rate": 4.020576469108139e-07, | |
| "loss": 1.2145, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 1.989795918367347, | |
| "grad_norm": 0.06573014422017058, | |
| "learning_rate": 7.942733738924845e-08, | |
| "loss": 1.1867, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 1.8389793634414673, | |
| "eval_runtime": 82.5875, | |
| "eval_samples_per_second": 157.373, | |
| "eval_steps_per_second": 4.928, | |
| "step": 784 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "step": 784, | |
| "total_flos": 173253108695040.0, | |
| "train_loss": 1.2838877389321521, | |
| "train_runtime": 2377.8575, | |
| "train_samples_per_second": 42.195, | |
| "train_steps_per_second": 0.33 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 784, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": false, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 173253108695040.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |