| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 660, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.004545454545454545, | |
| "grad_norm": 3.4694509506225586, | |
| "learning_rate": 0.0004992424242424243, | |
| "loss": 1.6877, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.00909090909090909, | |
| "grad_norm": 2.844703435897827, | |
| "learning_rate": 0.0004984848484848485, | |
| "loss": 1.7528, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.013636363636363636, | |
| "grad_norm": 4.147863388061523, | |
| "learning_rate": 0.0004977272727272727, | |
| "loss": 2.6111, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.01818181818181818, | |
| "grad_norm": 2.755852699279785, | |
| "learning_rate": 0.000496969696969697, | |
| "loss": 1.9464, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.022727272727272728, | |
| "grad_norm": 4.124767780303955, | |
| "learning_rate": 0.0004962121212121212, | |
| "loss": 2.2121, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.02727272727272727, | |
| "grad_norm": 3.923773765563965, | |
| "learning_rate": 0.0004954545454545455, | |
| "loss": 2.6991, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.031818181818181815, | |
| "grad_norm": 4.66182279586792, | |
| "learning_rate": 0.0004946969696969697, | |
| "loss": 2.176, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.03636363636363636, | |
| "grad_norm": 3.7830166816711426, | |
| "learning_rate": 0.000493939393939394, | |
| "loss": 2.7265, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.04090909090909091, | |
| "grad_norm": 3.966615676879883, | |
| "learning_rate": 0.0004931818181818182, | |
| "loss": 3.1926, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.045454545454545456, | |
| "grad_norm": 3.281916618347168, | |
| "learning_rate": 0.0004924242424242425, | |
| "loss": 2.1706, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 4.288072109222412, | |
| "learning_rate": 0.0004916666666666666, | |
| "loss": 3.1348, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.05454545454545454, | |
| "grad_norm": 3.2815868854522705, | |
| "learning_rate": 0.0004909090909090909, | |
| "loss": 2.2711, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.05909090909090909, | |
| "grad_norm": 2.75382924079895, | |
| "learning_rate": 0.0004901515151515152, | |
| "loss": 1.762, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.06363636363636363, | |
| "grad_norm": 2.6767005920410156, | |
| "learning_rate": 0.0004893939393939393, | |
| "loss": 1.591, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.06818181818181818, | |
| "grad_norm": 3.613719940185547, | |
| "learning_rate": 0.0004886363636363636, | |
| "loss": 2.1644, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.07272727272727272, | |
| "grad_norm": 3.3433680534362793, | |
| "learning_rate": 0.00048787878787878784, | |
| "loss": 2.9727, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.07727272727272727, | |
| "grad_norm": 3.7183644771575928, | |
| "learning_rate": 0.0004871212121212121, | |
| "loss": 1.8928, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.08181818181818182, | |
| "grad_norm": 4.1484575271606445, | |
| "learning_rate": 0.0004863636363636364, | |
| "loss": 2.6002, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.08636363636363636, | |
| "grad_norm": 3.281487464904785, | |
| "learning_rate": 0.0004856060606060606, | |
| "loss": 1.9074, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.09090909090909091, | |
| "grad_norm": 3.0067665576934814, | |
| "learning_rate": 0.0004848484848484849, | |
| "loss": 2.0375, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.09545454545454546, | |
| "grad_norm": 2.8053739070892334, | |
| "learning_rate": 0.00048409090909090906, | |
| "loss": 1.7248, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 2.619422435760498, | |
| "learning_rate": 0.00048333333333333334, | |
| "loss": 2.6251, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.10454545454545454, | |
| "grad_norm": 3.1278717517852783, | |
| "learning_rate": 0.0004825757575757576, | |
| "loss": 2.7065, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.10909090909090909, | |
| "grad_norm": 2.723963975906372, | |
| "learning_rate": 0.00048181818181818184, | |
| "loss": 1.8838, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.11363636363636363, | |
| "grad_norm": 2.6069819927215576, | |
| "learning_rate": 0.0004810606060606061, | |
| "loss": 1.9516, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.11818181818181818, | |
| "grad_norm": 2.426720380783081, | |
| "learning_rate": 0.0004803030303030303, | |
| "loss": 2.1439, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.12272727272727273, | |
| "grad_norm": 2.3633666038513184, | |
| "learning_rate": 0.00047954545454545456, | |
| "loss": 2.0113, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.12727272727272726, | |
| "grad_norm": 2.988654136657715, | |
| "learning_rate": 0.0004787878787878788, | |
| "loss": 2.1691, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.1318181818181818, | |
| "grad_norm": 2.713346481323242, | |
| "learning_rate": 0.00047803030303030306, | |
| "loss": 2.1206, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.13636363636363635, | |
| "grad_norm": 2.9896864891052246, | |
| "learning_rate": 0.0004772727272727273, | |
| "loss": 2.139, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.1409090909090909, | |
| "grad_norm": 2.7176098823547363, | |
| "learning_rate": 0.0004765151515151515, | |
| "loss": 2.2194, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.14545454545454545, | |
| "grad_norm": 2.6730499267578125, | |
| "learning_rate": 0.0004757575757575758, | |
| "loss": 1.5875, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 3.921717643737793, | |
| "learning_rate": 0.000475, | |
| "loss": 1.9037, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.15454545454545454, | |
| "grad_norm": 2.800473690032959, | |
| "learning_rate": 0.0004742424242424243, | |
| "loss": 1.8628, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.1590909090909091, | |
| "grad_norm": 2.7188827991485596, | |
| "learning_rate": 0.0004734848484848485, | |
| "loss": 2.1262, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.16363636363636364, | |
| "grad_norm": 2.794339895248413, | |
| "learning_rate": 0.0004727272727272727, | |
| "loss": 2.3508, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.16818181818181818, | |
| "grad_norm": NaN, | |
| "learning_rate": 0.0004727272727272727, | |
| "loss": 2.4565, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.17272727272727273, | |
| "grad_norm": 3.1815218925476074, | |
| "learning_rate": 0.000471969696969697, | |
| "loss": 2.2242, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.17727272727272728, | |
| "grad_norm": 3.4017906188964844, | |
| "learning_rate": 0.0004712121212121212, | |
| "loss": 2.1975, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.18181818181818182, | |
| "grad_norm": 2.7533328533172607, | |
| "learning_rate": 0.00047045454545454544, | |
| "loss": 2.1, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.18636363636363637, | |
| "grad_norm": 2.8896608352661133, | |
| "learning_rate": 0.0004696969696969697, | |
| "loss": 1.8716, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.19090909090909092, | |
| "grad_norm": 2.501896858215332, | |
| "learning_rate": 0.00046893939393939394, | |
| "loss": 2.3533, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.19545454545454546, | |
| "grad_norm": 2.2779133319854736, | |
| "learning_rate": 0.0004681818181818182, | |
| "loss": 1.8062, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 2.78344988822937, | |
| "learning_rate": 0.00046742424242424244, | |
| "loss": 2.4259, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.20454545454545456, | |
| "grad_norm": 2.230311393737793, | |
| "learning_rate": 0.00046666666666666666, | |
| "loss": 1.8689, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.20909090909090908, | |
| "grad_norm": 3.4158501625061035, | |
| "learning_rate": 0.0004659090909090909, | |
| "loss": 2.3214, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.21363636363636362, | |
| "grad_norm": 2.355423927307129, | |
| "learning_rate": 0.00046515151515151516, | |
| "loss": 1.7059, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.21818181818181817, | |
| "grad_norm": 2.495224952697754, | |
| "learning_rate": 0.00046439393939393944, | |
| "loss": 1.3574, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.22272727272727272, | |
| "grad_norm": 2.829482078552246, | |
| "learning_rate": 0.00046363636363636366, | |
| "loss": 2.2364, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.22727272727272727, | |
| "grad_norm": 2.390627861022949, | |
| "learning_rate": 0.0004628787878787879, | |
| "loss": 1.5226, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.2318181818181818, | |
| "grad_norm": 2.2006781101226807, | |
| "learning_rate": 0.0004621212121212121, | |
| "loss": 1.6848, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.23636363636363636, | |
| "grad_norm": 2.737412452697754, | |
| "learning_rate": 0.0004613636363636364, | |
| "loss": 1.9071, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.2409090909090909, | |
| "grad_norm": 3.4992029666900635, | |
| "learning_rate": 0.00046060606060606066, | |
| "loss": 2.0172, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.24545454545454545, | |
| "grad_norm": 3.196709394454956, | |
| "learning_rate": 0.0004598484848484848, | |
| "loss": 2.6357, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 3.9549436569213867, | |
| "learning_rate": 0.0004590909090909091, | |
| "loss": 2.9822, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.2545454545454545, | |
| "grad_norm": 2.788527488708496, | |
| "learning_rate": 0.0004583333333333333, | |
| "loss": 2.3458, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.2590909090909091, | |
| "grad_norm": 3.2539544105529785, | |
| "learning_rate": 0.0004575757575757576, | |
| "loss": 1.9336, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.2636363636363636, | |
| "grad_norm": 2.859744071960449, | |
| "learning_rate": 0.0004568181818181819, | |
| "loss": 2.4852, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.2681818181818182, | |
| "grad_norm": 2.6832542419433594, | |
| "learning_rate": 0.00045606060606060605, | |
| "loss": 2.0347, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.2727272727272727, | |
| "grad_norm": 4.593046188354492, | |
| "learning_rate": 0.0004553030303030303, | |
| "loss": 2.2415, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.2772727272727273, | |
| "grad_norm": 3.3459599018096924, | |
| "learning_rate": 0.00045454545454545455, | |
| "loss": 2.2843, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.2818181818181818, | |
| "grad_norm": 2.1371498107910156, | |
| "learning_rate": 0.0004537878787878788, | |
| "loss": 1.9185, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.2863636363636364, | |
| "grad_norm": 2.3603177070617676, | |
| "learning_rate": 0.000453030303030303, | |
| "loss": 1.9426, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.2909090909090909, | |
| "grad_norm": 2.5435550212860107, | |
| "learning_rate": 0.00045227272727272727, | |
| "loss": 2.4551, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.29545454545454547, | |
| "grad_norm": 2.5501880645751953, | |
| "learning_rate": 0.00045151515151515154, | |
| "loss": 1.9113, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 2.8549928665161133, | |
| "learning_rate": 0.00045075757575757577, | |
| "loss": 2.2465, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.30454545454545456, | |
| "grad_norm": 2.396857976913452, | |
| "learning_rate": 0.00045000000000000004, | |
| "loss": 2.1836, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.3090909090909091, | |
| "grad_norm": 2.7043912410736084, | |
| "learning_rate": 0.0004492424242424242, | |
| "loss": 2.1715, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.31363636363636366, | |
| "grad_norm": 3.1579270362854004, | |
| "learning_rate": 0.0004484848484848485, | |
| "loss": 2.4971, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.3181818181818182, | |
| "grad_norm": 2.3673815727233887, | |
| "learning_rate": 0.00044772727272727276, | |
| "loss": 1.7927, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.32272727272727275, | |
| "grad_norm": 2.776143789291382, | |
| "learning_rate": 0.000446969696969697, | |
| "loss": 1.9065, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.32727272727272727, | |
| "grad_norm": 2.937574625015259, | |
| "learning_rate": 0.00044621212121212126, | |
| "loss": 1.6579, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.33181818181818185, | |
| "grad_norm": 2.937641143798828, | |
| "learning_rate": 0.00044545454545454543, | |
| "loss": 2.2259, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.33636363636363636, | |
| "grad_norm": 2.305844306945801, | |
| "learning_rate": 0.0004446969696969697, | |
| "loss": 2.0766, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 0.3409090909090909, | |
| "grad_norm": 3.1322784423828125, | |
| "learning_rate": 0.000443939393939394, | |
| "loss": 2.3208, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.34545454545454546, | |
| "grad_norm": 2.7713890075683594, | |
| "learning_rate": 0.0004431818181818182, | |
| "loss": 2.4553, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 2.275108575820923, | |
| "learning_rate": 0.00044242424242424243, | |
| "loss": 1.4249, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 0.35454545454545455, | |
| "grad_norm": 2.502997875213623, | |
| "learning_rate": 0.00044166666666666665, | |
| "loss": 2.4343, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 0.35909090909090907, | |
| "grad_norm": 2.1204617023468018, | |
| "learning_rate": 0.00044090909090909093, | |
| "loss": 1.9248, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 0.36363636363636365, | |
| "grad_norm": 2.9564898014068604, | |
| "learning_rate": 0.00044015151515151515, | |
| "loss": 1.6672, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.36818181818181817, | |
| "grad_norm": 3.0879478454589844, | |
| "learning_rate": 0.0004393939393939394, | |
| "loss": 2.0625, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 0.37272727272727274, | |
| "grad_norm": 3.1532368659973145, | |
| "learning_rate": 0.00043863636363636365, | |
| "loss": 2.3648, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 0.37727272727272726, | |
| "grad_norm": 2.1850852966308594, | |
| "learning_rate": 0.00043787878787878787, | |
| "loss": 2.4397, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 0.38181818181818183, | |
| "grad_norm": 2.241144895553589, | |
| "learning_rate": 0.00043712121212121215, | |
| "loss": 2.5193, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 0.38636363636363635, | |
| "grad_norm": 3.6500165462493896, | |
| "learning_rate": 0.00043636363636363637, | |
| "loss": 2.1096, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.39090909090909093, | |
| "grad_norm": 2.0548837184906006, | |
| "learning_rate": 0.0004356060606060606, | |
| "loss": 2.359, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 0.39545454545454545, | |
| "grad_norm": 2.4512407779693604, | |
| "learning_rate": 0.00043484848484848487, | |
| "loss": 1.5353, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 1.9420820474624634, | |
| "learning_rate": 0.0004340909090909091, | |
| "loss": 1.688, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 0.40454545454545454, | |
| "grad_norm": 1.845475673675537, | |
| "learning_rate": 0.00043333333333333337, | |
| "loss": 1.6642, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 0.4090909090909091, | |
| "grad_norm": 2.2735133171081543, | |
| "learning_rate": 0.0004325757575757576, | |
| "loss": 1.89, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.41363636363636364, | |
| "grad_norm": 2.203105926513672, | |
| "learning_rate": 0.0004318181818181818, | |
| "loss": 1.9556, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 0.41818181818181815, | |
| "grad_norm": 1.8748105764389038, | |
| "learning_rate": 0.00043106060606060603, | |
| "loss": 1.5452, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 0.42272727272727273, | |
| "grad_norm": 2.8958442211151123, | |
| "learning_rate": 0.0004303030303030303, | |
| "loss": 1.9343, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 0.42727272727272725, | |
| "grad_norm": 2.7512269020080566, | |
| "learning_rate": 0.0004295454545454546, | |
| "loss": 2.4008, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 0.4318181818181818, | |
| "grad_norm": 2.748307228088379, | |
| "learning_rate": 0.00042878787878787876, | |
| "loss": 2.3614, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.43636363636363634, | |
| "grad_norm": 3.7091145515441895, | |
| "learning_rate": 0.00042803030303030303, | |
| "loss": 1.5435, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.4409090909090909, | |
| "grad_norm": 2.0227293968200684, | |
| "learning_rate": 0.00042727272727272726, | |
| "loss": 1.5536, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 0.44545454545454544, | |
| "grad_norm": 1.868477702140808, | |
| "learning_rate": 0.00042651515151515153, | |
| "loss": 2.0019, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.2410340309143066, | |
| "learning_rate": 0.0004257575757575758, | |
| "loss": 2.0278, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 0.45454545454545453, | |
| "grad_norm": 2.4206206798553467, | |
| "learning_rate": 0.000425, | |
| "loss": 2.6757, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.4590909090909091, | |
| "grad_norm": 2.6481056213378906, | |
| "learning_rate": 0.00042424242424242425, | |
| "loss": 2.004, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 0.4636363636363636, | |
| "grad_norm": 2.493495225906372, | |
| "learning_rate": 0.0004234848484848485, | |
| "loss": 2.2102, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 0.4681818181818182, | |
| "grad_norm": 2.588595390319824, | |
| "learning_rate": 0.00042272727272727275, | |
| "loss": 2.3133, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 0.4727272727272727, | |
| "grad_norm": 2.185718536376953, | |
| "learning_rate": 0.00042196969696969703, | |
| "loss": 2.5506, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 0.4772727272727273, | |
| "grad_norm": 2.155470132827759, | |
| "learning_rate": 0.0004212121212121212, | |
| "loss": 2.2074, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.4818181818181818, | |
| "grad_norm": 2.518435001373291, | |
| "learning_rate": 0.0004204545454545455, | |
| "loss": 1.8589, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 0.4863636363636364, | |
| "grad_norm": 2.5512635707855225, | |
| "learning_rate": 0.0004196969696969697, | |
| "loss": 1.9953, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 0.4909090909090909, | |
| "grad_norm": 2.238809108734131, | |
| "learning_rate": 0.00041893939393939397, | |
| "loss": 2.2441, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 0.4954545454545455, | |
| "grad_norm": 1.8442784547805786, | |
| "learning_rate": 0.00041818181818181814, | |
| "loss": 1.8682, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 2.4844954013824463, | |
| "learning_rate": 0.0004174242424242424, | |
| "loss": 1.9522, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.5045454545454545, | |
| "grad_norm": 1.9704878330230713, | |
| "learning_rate": 0.0004166666666666667, | |
| "loss": 1.4167, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 0.509090909090909, | |
| "grad_norm": 2.2447972297668457, | |
| "learning_rate": 0.0004159090909090909, | |
| "loss": 1.7897, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 0.5136363636363637, | |
| "grad_norm": 2.530410051345825, | |
| "learning_rate": 0.0004151515151515152, | |
| "loss": 2.5473, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 0.5181818181818182, | |
| "grad_norm": 2.450526714324951, | |
| "learning_rate": 0.00041439393939393936, | |
| "loss": 2.0436, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 0.5227272727272727, | |
| "grad_norm": 2.4212632179260254, | |
| "learning_rate": 0.00041363636363636364, | |
| "loss": 2.1118, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.5272727272727272, | |
| "grad_norm": 1.9820351600646973, | |
| "learning_rate": 0.0004128787878787879, | |
| "loss": 1.9614, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 0.5318181818181819, | |
| "grad_norm": 2.467961549758911, | |
| "learning_rate": 0.00041212121212121214, | |
| "loss": 1.9572, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 0.5363636363636364, | |
| "grad_norm": 2.2693068981170654, | |
| "learning_rate": 0.00041136363636363636, | |
| "loss": 1.9033, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 0.5409090909090909, | |
| "grad_norm": 2.299119710922241, | |
| "learning_rate": 0.0004106060606060606, | |
| "loss": 1.9848, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 0.5454545454545454, | |
| "grad_norm": 2.0330560207366943, | |
| "learning_rate": 0.00040984848484848486, | |
| "loss": 1.3706, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 2.197603702545166, | |
| "learning_rate": 0.00040909090909090913, | |
| "loss": 2.1687, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 0.5545454545454546, | |
| "grad_norm": 2.7206549644470215, | |
| "learning_rate": 0.00040833333333333336, | |
| "loss": 2.3247, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 0.5590909090909091, | |
| "grad_norm": 2.882654905319214, | |
| "learning_rate": 0.0004075757575757576, | |
| "loss": 1.6946, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 0.5636363636363636, | |
| "grad_norm": 2.3815231323242188, | |
| "learning_rate": 0.0004068181818181818, | |
| "loss": 1.862, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 0.5681818181818182, | |
| "grad_norm": 2.4142932891845703, | |
| "learning_rate": 0.0004060606060606061, | |
| "loss": 2.0066, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.5727272727272728, | |
| "grad_norm": 2.6641104221343994, | |
| "learning_rate": 0.0004053030303030303, | |
| "loss": 1.9456, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 0.5772727272727273, | |
| "grad_norm": 2.983633518218994, | |
| "learning_rate": 0.0004045454545454546, | |
| "loss": 2.0049, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 0.5818181818181818, | |
| "grad_norm": 2.1993696689605713, | |
| "learning_rate": 0.0004037878787878788, | |
| "loss": 1.6259, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.5863636363636363, | |
| "grad_norm": 3.0398480892181396, | |
| "learning_rate": 0.000403030303030303, | |
| "loss": 1.6535, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 0.5909090909090909, | |
| "grad_norm": 2.298558235168457, | |
| "learning_rate": 0.0004022727272727273, | |
| "loss": 1.8693, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.5954545454545455, | |
| "grad_norm": 2.5126214027404785, | |
| "learning_rate": 0.0004015151515151515, | |
| "loss": 2.1854, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 2.0419557094573975, | |
| "learning_rate": 0.00040075757575757574, | |
| "loss": 1.5857, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 0.6045454545454545, | |
| "grad_norm": 2.4304699897766113, | |
| "learning_rate": 0.0004, | |
| "loss": 2.347, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 0.6090909090909091, | |
| "grad_norm": 3.098036050796509, | |
| "learning_rate": 0.00039924242424242424, | |
| "loss": 2.9748, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 0.6136363636363636, | |
| "grad_norm": 2.4403679370880127, | |
| "learning_rate": 0.0003984848484848485, | |
| "loss": 1.9342, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.6181818181818182, | |
| "grad_norm": 2.832394599914551, | |
| "learning_rate": 0.00039772727272727274, | |
| "loss": 2.132, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 0.6227272727272727, | |
| "grad_norm": 2.601243019104004, | |
| "learning_rate": 0.00039696969696969696, | |
| "loss": 2.3074, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 0.6272727272727273, | |
| "grad_norm": 2.2306132316589355, | |
| "learning_rate": 0.00039621212121212124, | |
| "loss": 1.6065, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 0.6318181818181818, | |
| "grad_norm": 2.393157720565796, | |
| "learning_rate": 0.00039545454545454546, | |
| "loss": 1.7411, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 0.6363636363636364, | |
| "grad_norm": 2.174208164215088, | |
| "learning_rate": 0.00039469696969696974, | |
| "loss": 1.3876, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.6409090909090909, | |
| "grad_norm": 2.3376457691192627, | |
| "learning_rate": 0.0003939393939393939, | |
| "loss": 2.3752, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 0.6454545454545455, | |
| "grad_norm": 2.141479969024658, | |
| "learning_rate": 0.0003931818181818182, | |
| "loss": 1.3948, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 2.4302890300750732, | |
| "learning_rate": 0.0003924242424242424, | |
| "loss": 2.5493, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 0.6545454545454545, | |
| "grad_norm": 1.9080986976623535, | |
| "learning_rate": 0.0003916666666666667, | |
| "loss": 1.4652, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 0.6590909090909091, | |
| "grad_norm": 2.438420057296753, | |
| "learning_rate": 0.00039090909090909096, | |
| "loss": 2.0606, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.6636363636363637, | |
| "grad_norm": 2.028294563293457, | |
| "learning_rate": 0.0003901515151515151, | |
| "loss": 1.8798, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 0.6681818181818182, | |
| "grad_norm": 2.4235315322875977, | |
| "learning_rate": 0.0003893939393939394, | |
| "loss": 1.8855, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 0.6727272727272727, | |
| "grad_norm": 2.430391788482666, | |
| "learning_rate": 0.0003886363636363636, | |
| "loss": 1.7753, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 0.6772727272727272, | |
| "grad_norm": 2.053199052810669, | |
| "learning_rate": 0.0003878787878787879, | |
| "loss": 2.1466, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 0.6818181818181818, | |
| "grad_norm": 2.067093849182129, | |
| "learning_rate": 0.0003871212121212122, | |
| "loss": 1.7715, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.6863636363636364, | |
| "grad_norm": 2.047165632247925, | |
| "learning_rate": 0.00038636363636363635, | |
| "loss": 1.8703, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 0.6909090909090909, | |
| "grad_norm": 2.567540168762207, | |
| "learning_rate": 0.0003856060606060606, | |
| "loss": 1.7973, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 0.6954545454545454, | |
| "grad_norm": 2.5282599925994873, | |
| "learning_rate": 0.00038484848484848485, | |
| "loss": 2.422, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 2.0428948402404785, | |
| "learning_rate": 0.0003840909090909091, | |
| "loss": 1.5709, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 0.7045454545454546, | |
| "grad_norm": 2.032672166824341, | |
| "learning_rate": 0.00038333333333333334, | |
| "loss": 1.8584, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.7090909090909091, | |
| "grad_norm": 2.4448535442352295, | |
| "learning_rate": 0.00038257575757575757, | |
| "loss": 2.069, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 0.7136363636363636, | |
| "grad_norm": 1.6503087282180786, | |
| "learning_rate": 0.00038181818181818184, | |
| "loss": 1.5194, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 0.7181818181818181, | |
| "grad_norm": 2.5853593349456787, | |
| "learning_rate": 0.00038106060606060607, | |
| "loss": 2.4603, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 0.7227272727272728, | |
| "grad_norm": 2.353992223739624, | |
| "learning_rate": 0.00038030303030303034, | |
| "loss": 1.4417, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 0.7272727272727273, | |
| "grad_norm": 2.382633686065674, | |
| "learning_rate": 0.0003795454545454545, | |
| "loss": 1.9239, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.7318181818181818, | |
| "grad_norm": 2.7274303436279297, | |
| "learning_rate": 0.0003787878787878788, | |
| "loss": 2.1116, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 0.7363636363636363, | |
| "grad_norm": 2.0137476921081543, | |
| "learning_rate": 0.00037803030303030306, | |
| "loss": 1.5707, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 0.740909090909091, | |
| "grad_norm": 1.977155089378357, | |
| "learning_rate": 0.0003772727272727273, | |
| "loss": 1.4972, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 0.7454545454545455, | |
| "grad_norm": 2.5506880283355713, | |
| "learning_rate": 0.0003765151515151515, | |
| "loss": 2.4583, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 2.2664661407470703, | |
| "learning_rate": 0.00037575757575757573, | |
| "loss": 2.1239, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.7545454545454545, | |
| "grad_norm": 2.393831968307495, | |
| "learning_rate": 0.000375, | |
| "loss": 2.2104, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 0.759090909090909, | |
| "grad_norm": 2.358670711517334, | |
| "learning_rate": 0.0003742424242424243, | |
| "loss": 1.7545, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 0.7636363636363637, | |
| "grad_norm": 2.0985164642333984, | |
| "learning_rate": 0.0003734848484848485, | |
| "loss": 1.7337, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 0.7681818181818182, | |
| "grad_norm": 1.711176872253418, | |
| "learning_rate": 0.00037272727272727273, | |
| "loss": 1.3195, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 0.7727272727272727, | |
| "grad_norm": 2.20684814453125, | |
| "learning_rate": 0.00037196969696969695, | |
| "loss": 1.7633, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.7772727272727272, | |
| "grad_norm": 2.0569570064544678, | |
| "learning_rate": 0.00037121212121212123, | |
| "loss": 1.8354, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 0.7818181818181819, | |
| "grad_norm": 2.4895520210266113, | |
| "learning_rate": 0.0003704545454545455, | |
| "loss": 2.6706, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 0.7863636363636364, | |
| "grad_norm": 2.3134992122650146, | |
| "learning_rate": 0.00036969696969696967, | |
| "loss": 2.077, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 0.7909090909090909, | |
| "grad_norm": 1.876047968864441, | |
| "learning_rate": 0.00036893939393939395, | |
| "loss": 1.5816, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 0.7954545454545454, | |
| "grad_norm": 2.301314353942871, | |
| "learning_rate": 0.00036818181818181817, | |
| "loss": 2.9433, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 2.4783785343170166, | |
| "learning_rate": 0.00036742424242424245, | |
| "loss": 2.3913, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 0.8045454545454546, | |
| "grad_norm": 2.3966879844665527, | |
| "learning_rate": 0.00036666666666666667, | |
| "loss": 2.6103, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 0.8090909090909091, | |
| "grad_norm": 2.1050190925598145, | |
| "learning_rate": 0.0003659090909090909, | |
| "loss": 1.6801, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 0.8136363636363636, | |
| "grad_norm": 2.3336639404296875, | |
| "learning_rate": 0.00036515151515151517, | |
| "loss": 1.936, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 0.8181818181818182, | |
| "grad_norm": 2.4867429733276367, | |
| "learning_rate": 0.0003643939393939394, | |
| "loss": 2.0285, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.8227272727272728, | |
| "grad_norm": 1.9529379606246948, | |
| "learning_rate": 0.00036363636363636367, | |
| "loss": 1.5503, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 0.8272727272727273, | |
| "grad_norm": 2.095381736755371, | |
| "learning_rate": 0.00036287878787878784, | |
| "loss": 2.5626, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 0.8318181818181818, | |
| "grad_norm": 2.156099319458008, | |
| "learning_rate": 0.0003621212121212121, | |
| "loss": 1.8788, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 0.8363636363636363, | |
| "grad_norm": 1.8726741075515747, | |
| "learning_rate": 0.0003613636363636364, | |
| "loss": 1.6386, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 0.8409090909090909, | |
| "grad_norm": 2.6056482791900635, | |
| "learning_rate": 0.0003606060606060606, | |
| "loss": 1.7965, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.8454545454545455, | |
| "grad_norm": 2.65775728225708, | |
| "learning_rate": 0.0003598484848484849, | |
| "loss": 1.775, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 1.9309563636779785, | |
| "learning_rate": 0.00035909090909090906, | |
| "loss": 1.7575, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 0.8545454545454545, | |
| "grad_norm": 1.874107003211975, | |
| "learning_rate": 0.00035833333333333333, | |
| "loss": 1.4686, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 0.8590909090909091, | |
| "grad_norm": 2.125084400177002, | |
| "learning_rate": 0.0003575757575757576, | |
| "loss": 1.3713, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 0.8636363636363636, | |
| "grad_norm": 2.016660690307617, | |
| "learning_rate": 0.00035681818181818183, | |
| "loss": 1.6914, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.8681818181818182, | |
| "grad_norm": 2.8699893951416016, | |
| "learning_rate": 0.0003560606060606061, | |
| "loss": 1.6524, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 0.8727272727272727, | |
| "grad_norm": 2.2474772930145264, | |
| "learning_rate": 0.0003553030303030303, | |
| "loss": 1.7333, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.8772727272727273, | |
| "grad_norm": 2.6996030807495117, | |
| "learning_rate": 0.00035454545454545455, | |
| "loss": 1.5828, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 0.8818181818181818, | |
| "grad_norm": 2.196274757385254, | |
| "learning_rate": 0.0003537878787878788, | |
| "loss": 1.6058, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 0.8863636363636364, | |
| "grad_norm": 2.4350290298461914, | |
| "learning_rate": 0.00035303030303030305, | |
| "loss": 2.0724, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.8909090909090909, | |
| "grad_norm": 2.047480821609497, | |
| "learning_rate": 0.0003522727272727273, | |
| "loss": 1.7112, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 0.8954545454545455, | |
| "grad_norm": 2.4547033309936523, | |
| "learning_rate": 0.0003515151515151515, | |
| "loss": 1.7747, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 2.9125847816467285, | |
| "learning_rate": 0.0003507575757575758, | |
| "loss": 2.0878, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 0.9045454545454545, | |
| "grad_norm": 2.168196678161621, | |
| "learning_rate": 0.00035, | |
| "loss": 1.5592, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 0.9090909090909091, | |
| "grad_norm": 2.364847183227539, | |
| "learning_rate": 0.0003492424242424243, | |
| "loss": 2.0301, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.9136363636363637, | |
| "grad_norm": 2.743267059326172, | |
| "learning_rate": 0.0003484848484848485, | |
| "loss": 1.8784, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 0.9181818181818182, | |
| "grad_norm": 2.2784361839294434, | |
| "learning_rate": 0.0003477272727272727, | |
| "loss": 1.5936, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 0.9227272727272727, | |
| "grad_norm": 2.875943422317505, | |
| "learning_rate": 0.000346969696969697, | |
| "loss": 1.9961, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 0.9272727272727272, | |
| "grad_norm": 2.0056071281433105, | |
| "learning_rate": 0.0003462121212121212, | |
| "loss": 1.8795, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 0.9318181818181818, | |
| "grad_norm": 2.3547298908233643, | |
| "learning_rate": 0.00034545454545454544, | |
| "loss": 2.1429, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.9363636363636364, | |
| "grad_norm": 2.7082138061523438, | |
| "learning_rate": 0.0003446969696969697, | |
| "loss": 1.7504, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 0.9409090909090909, | |
| "grad_norm": 2.335139751434326, | |
| "learning_rate": 0.00034393939393939394, | |
| "loss": 1.9774, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 0.9454545454545454, | |
| "grad_norm": 2.3550143241882324, | |
| "learning_rate": 0.0003431818181818182, | |
| "loss": 1.8602, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 2.622682809829712, | |
| "learning_rate": 0.00034242424242424244, | |
| "loss": 2.2283, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 0.9545454545454546, | |
| "grad_norm": 1.7282129526138306, | |
| "learning_rate": 0.00034166666666666666, | |
| "loss": 1.6025, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.9590909090909091, | |
| "grad_norm": 2.8252415657043457, | |
| "learning_rate": 0.0003409090909090909, | |
| "loss": 1.7649, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 0.9636363636363636, | |
| "grad_norm": 2.146219491958618, | |
| "learning_rate": 0.00034015151515151516, | |
| "loss": 1.9742, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 0.9681818181818181, | |
| "grad_norm": 2.124577045440674, | |
| "learning_rate": 0.00033939393939393943, | |
| "loss": 1.7412, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 0.9727272727272728, | |
| "grad_norm": 1.7649497985839844, | |
| "learning_rate": 0.00033863636363636366, | |
| "loss": 1.2667, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 0.9772727272727273, | |
| "grad_norm": 2.375659465789795, | |
| "learning_rate": 0.0003378787878787879, | |
| "loss": 1.7142, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.9818181818181818, | |
| "grad_norm": 1.9995368719100952, | |
| "learning_rate": 0.0003371212121212121, | |
| "loss": 1.4613, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 0.9863636363636363, | |
| "grad_norm": 2.2640981674194336, | |
| "learning_rate": 0.0003363636363636364, | |
| "loss": 1.9474, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 0.990909090909091, | |
| "grad_norm": 2.1680893898010254, | |
| "learning_rate": 0.00033560606060606066, | |
| "loss": 2.5352, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 0.9954545454545455, | |
| "grad_norm": 2.3969366550445557, | |
| "learning_rate": 0.0003348484848484848, | |
| "loss": 1.9012, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 3.3696913719177246, | |
| "learning_rate": 0.0003340909090909091, | |
| "loss": 2.2928, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_f1": 0.891, | |
| "eval_gen_len": 41.9182, | |
| "eval_loss": 1.8093845844268799, | |
| "eval_precision": 0.8891, | |
| "eval_recall": 0.8931, | |
| "eval_rouge1": 0.466, | |
| "eval_rouge2": 0.2146, | |
| "eval_rougeL": 0.3912, | |
| "eval_rougeLsum": 0.4301, | |
| "eval_runtime": 25.1921, | |
| "eval_samples_per_second": 4.366, | |
| "eval_steps_per_second": 0.556, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 1.0045454545454546, | |
| "grad_norm": 1.7403843402862549, | |
| "learning_rate": 0.0003333333333333333, | |
| "loss": 1.7294, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 1.009090909090909, | |
| "grad_norm": 1.5273452997207642, | |
| "learning_rate": 0.0003325757575757576, | |
| "loss": 1.3279, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 1.0136363636363637, | |
| "grad_norm": 1.427538275718689, | |
| "learning_rate": 0.0003318181818181819, | |
| "loss": 0.9647, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 1.018181818181818, | |
| "grad_norm": 1.5605067014694214, | |
| "learning_rate": 0.00033106060606060604, | |
| "loss": 1.3178, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 1.0227272727272727, | |
| "grad_norm": 1.6737922430038452, | |
| "learning_rate": 0.0003303030303030303, | |
| "loss": 1.403, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 1.0272727272727273, | |
| "grad_norm": 2.3249313831329346, | |
| "learning_rate": 0.00032954545454545454, | |
| "loss": 1.4907, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 1.0318181818181817, | |
| "grad_norm": 1.9939628839492798, | |
| "learning_rate": 0.0003287878787878788, | |
| "loss": 1.795, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 1.0363636363636364, | |
| "grad_norm": 2.482421398162842, | |
| "learning_rate": 0.000328030303030303, | |
| "loss": 1.7309, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 1.040909090909091, | |
| "grad_norm": 1.6090010404586792, | |
| "learning_rate": 0.00032727272727272726, | |
| "loss": 1.0976, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 1.0454545454545454, | |
| "grad_norm": 1.5481090545654297, | |
| "learning_rate": 0.00032651515151515154, | |
| "loss": 1.1785, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "grad_norm": 2.0420186519622803, | |
| "learning_rate": 0.00032575757575757576, | |
| "loss": 1.1853, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 1.0545454545454545, | |
| "grad_norm": 2.5020453929901123, | |
| "learning_rate": 0.00032500000000000004, | |
| "loss": 1.8213, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 1.059090909090909, | |
| "grad_norm": 1.5312837362289429, | |
| "learning_rate": 0.0003242424242424242, | |
| "loss": 1.016, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 1.0636363636363637, | |
| "grad_norm": 2.062110185623169, | |
| "learning_rate": 0.0003234848484848485, | |
| "loss": 1.2245, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 1.0681818181818181, | |
| "grad_norm": 1.7521977424621582, | |
| "learning_rate": 0.00032272727272727276, | |
| "loss": 1.4904, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 1.0727272727272728, | |
| "grad_norm": 1.431998372077942, | |
| "learning_rate": 0.000321969696969697, | |
| "loss": 1.2364, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 1.0772727272727272, | |
| "grad_norm": 1.8957371711730957, | |
| "learning_rate": 0.00032121212121212126, | |
| "loss": 1.9241, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 1.0818181818181818, | |
| "grad_norm": 1.6720540523529053, | |
| "learning_rate": 0.00032045454545454543, | |
| "loss": 1.0261, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 1.0863636363636364, | |
| "grad_norm": 1.8503271341323853, | |
| "learning_rate": 0.0003196969696969697, | |
| "loss": 1.6694, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 1.0909090909090908, | |
| "grad_norm": 2.055772066116333, | |
| "learning_rate": 0.000318939393939394, | |
| "loss": 1.6855, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 1.0954545454545455, | |
| "grad_norm": 4.034445285797119, | |
| "learning_rate": 0.0003181818181818182, | |
| "loss": 1.8592, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "grad_norm": 2.6941239833831787, | |
| "learning_rate": 0.0003174242424242424, | |
| "loss": 1.1528, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 1.1045454545454545, | |
| "grad_norm": 1.8258893489837646, | |
| "learning_rate": 0.00031666666666666665, | |
| "loss": 1.1762, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 1.1090909090909091, | |
| "grad_norm": 2.721888303756714, | |
| "learning_rate": 0.0003159090909090909, | |
| "loss": 1.5523, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 1.1136363636363635, | |
| "grad_norm": 2.35798978805542, | |
| "learning_rate": 0.00031515151515151515, | |
| "loss": 1.7533, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 1.1181818181818182, | |
| "grad_norm": 2.089695453643799, | |
| "learning_rate": 0.0003143939393939394, | |
| "loss": 1.4344, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 1.1227272727272728, | |
| "grad_norm": 2.0275492668151855, | |
| "learning_rate": 0.00031363636363636365, | |
| "loss": 1.5359, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 1.1272727272727272, | |
| "grad_norm": 2.290893077850342, | |
| "learning_rate": 0.00031287878787878787, | |
| "loss": 1.8292, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 1.1318181818181818, | |
| "grad_norm": 2.3136603832244873, | |
| "learning_rate": 0.00031212121212121214, | |
| "loss": 1.6828, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 1.1363636363636362, | |
| "grad_norm": 2.1181459426879883, | |
| "learning_rate": 0.00031136363636363637, | |
| "loss": 1.1531, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 1.1409090909090909, | |
| "grad_norm": 1.9240480661392212, | |
| "learning_rate": 0.0003106060606060606, | |
| "loss": 1.2515, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 1.1454545454545455, | |
| "grad_norm": 2.522502899169922, | |
| "learning_rate": 0.00030984848484848487, | |
| "loss": 1.4942, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "grad_norm": 1.5959993600845337, | |
| "learning_rate": 0.0003090909090909091, | |
| "loss": 1.1412, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 1.1545454545454545, | |
| "grad_norm": 1.740268588066101, | |
| "learning_rate": 0.00030833333333333337, | |
| "loss": 1.7009, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 1.1590909090909092, | |
| "grad_norm": 2.1791181564331055, | |
| "learning_rate": 0.0003075757575757576, | |
| "loss": 1.4727, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 1.1636363636363636, | |
| "grad_norm": 2.2325475215911865, | |
| "learning_rate": 0.0003068181818181818, | |
| "loss": 1.4379, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 1.1681818181818182, | |
| "grad_norm": 2.206281900405884, | |
| "learning_rate": 0.00030606060606060603, | |
| "loss": 1.5069, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 1.1727272727272728, | |
| "grad_norm": 2.6821632385253906, | |
| "learning_rate": 0.0003053030303030303, | |
| "loss": 1.7888, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 1.1772727272727272, | |
| "grad_norm": 2.567087173461914, | |
| "learning_rate": 0.0003045454545454546, | |
| "loss": 1.7501, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 1.1818181818181819, | |
| "grad_norm": 2.131887674331665, | |
| "learning_rate": 0.00030378787878787875, | |
| "loss": 1.3294, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 1.1863636363636363, | |
| "grad_norm": 1.5638073682785034, | |
| "learning_rate": 0.00030303030303030303, | |
| "loss": 1.3679, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 1.190909090909091, | |
| "grad_norm": 2.208799362182617, | |
| "learning_rate": 0.00030227272727272725, | |
| "loss": 1.2585, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 1.1954545454545455, | |
| "grad_norm": 2.379058599472046, | |
| "learning_rate": 0.00030151515151515153, | |
| "loss": 1.4692, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 2.2137858867645264, | |
| "learning_rate": 0.0003007575757575758, | |
| "loss": 1.5648, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 1.2045454545454546, | |
| "grad_norm": 1.94793701171875, | |
| "learning_rate": 0.0003, | |
| "loss": 1.2642, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 1.209090909090909, | |
| "grad_norm": 2.8321635723114014, | |
| "learning_rate": 0.00029924242424242425, | |
| "loss": 1.4013, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 1.2136363636363636, | |
| "grad_norm": 2.6073920726776123, | |
| "learning_rate": 0.00029848484848484847, | |
| "loss": 1.6666, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 1.2181818181818183, | |
| "grad_norm": 1.6753661632537842, | |
| "learning_rate": 0.00029772727272727275, | |
| "loss": 1.2472, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 1.2227272727272727, | |
| "grad_norm": 2.296635866165161, | |
| "learning_rate": 0.000296969696969697, | |
| "loss": 1.8099, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 1.2272727272727273, | |
| "grad_norm": 2.8359079360961914, | |
| "learning_rate": 0.0002962121212121212, | |
| "loss": 1.5522, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 1.231818181818182, | |
| "grad_norm": 2.3741962909698486, | |
| "learning_rate": 0.00029545454545454547, | |
| "loss": 1.5737, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 1.2363636363636363, | |
| "grad_norm": 1.9859591722488403, | |
| "learning_rate": 0.0002946969696969697, | |
| "loss": 1.5659, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 1.240909090909091, | |
| "grad_norm": 1.9343222379684448, | |
| "learning_rate": 0.00029393939393939397, | |
| "loss": 1.1204, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 1.2454545454545454, | |
| "grad_norm": 1.6376460790634155, | |
| "learning_rate": 0.00029318181818181814, | |
| "loss": 0.8886, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 1.9865474700927734, | |
| "learning_rate": 0.0002924242424242424, | |
| "loss": 1.5425, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 1.2545454545454544, | |
| "grad_norm": 2.1017825603485107, | |
| "learning_rate": 0.0002916666666666667, | |
| "loss": 1.1125, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 1.259090909090909, | |
| "grad_norm": 2.349350690841675, | |
| "learning_rate": 0.0002909090909090909, | |
| "loss": 1.6496, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 1.2636363636363637, | |
| "grad_norm": 1.8741109371185303, | |
| "learning_rate": 0.0002901515151515152, | |
| "loss": 1.3404, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 1.268181818181818, | |
| "grad_norm": 2.2605037689208984, | |
| "learning_rate": 0.00028939393939393936, | |
| "loss": 1.5495, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 1.2727272727272727, | |
| "grad_norm": 2.0851080417633057, | |
| "learning_rate": 0.00028863636363636363, | |
| "loss": 1.501, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 1.2772727272727273, | |
| "grad_norm": 2.2092325687408447, | |
| "learning_rate": 0.0002878787878787879, | |
| "loss": 1.5655, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 1.2818181818181817, | |
| "grad_norm": 2.343780755996704, | |
| "learning_rate": 0.00028712121212121213, | |
| "loss": 1.4229, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 1.2863636363636364, | |
| "grad_norm": 1.684411883354187, | |
| "learning_rate": 0.00028636363636363636, | |
| "loss": 1.4132, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 1.290909090909091, | |
| "grad_norm": 2.034984827041626, | |
| "learning_rate": 0.0002856060606060606, | |
| "loss": 1.1224, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 1.2954545454545454, | |
| "grad_norm": 1.9973840713500977, | |
| "learning_rate": 0.00028484848484848485, | |
| "loss": 1.4387, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "grad_norm": 2.2674074172973633, | |
| "learning_rate": 0.00028409090909090913, | |
| "loss": 1.6697, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 1.3045454545454547, | |
| "grad_norm": 1.596279501914978, | |
| "learning_rate": 0.00028333333333333335, | |
| "loss": 1.0433, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 1.309090909090909, | |
| "grad_norm": 1.874055027961731, | |
| "learning_rate": 0.0002825757575757576, | |
| "loss": 1.291, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 1.3136363636363637, | |
| "grad_norm": 2.2551302909851074, | |
| "learning_rate": 0.0002818181818181818, | |
| "loss": 1.3771, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 1.3181818181818181, | |
| "grad_norm": 2.502380847930908, | |
| "learning_rate": 0.0002810606060606061, | |
| "loss": 1.853, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 1.3227272727272728, | |
| "grad_norm": 1.9750282764434814, | |
| "learning_rate": 0.0002803030303030303, | |
| "loss": 1.4369, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 1.3272727272727272, | |
| "grad_norm": 2.4181363582611084, | |
| "learning_rate": 0.0002795454545454546, | |
| "loss": 1.3565, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 1.3318181818181818, | |
| "grad_norm": 2.0823867321014404, | |
| "learning_rate": 0.0002787878787878788, | |
| "loss": 1.4589, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 1.3363636363636364, | |
| "grad_norm": 2.147993326187134, | |
| "learning_rate": 0.000278030303030303, | |
| "loss": 1.3616, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 1.3409090909090908, | |
| "grad_norm": 3.184967517852783, | |
| "learning_rate": 0.0002772727272727273, | |
| "loss": 1.8248, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 1.3454545454545455, | |
| "grad_norm": 2.3890221118927, | |
| "learning_rate": 0.0002765151515151515, | |
| "loss": 1.4599, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "grad_norm": 1.9724668264389038, | |
| "learning_rate": 0.00027575757575757574, | |
| "loss": 1.2248, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 1.3545454545454545, | |
| "grad_norm": 2.1539180278778076, | |
| "learning_rate": 0.000275, | |
| "loss": 1.1712, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 1.3590909090909091, | |
| "grad_norm": 2.107490062713623, | |
| "learning_rate": 0.00027424242424242424, | |
| "loss": 1.3786, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 1.3636363636363638, | |
| "grad_norm": 2.052065372467041, | |
| "learning_rate": 0.0002734848484848485, | |
| "loss": 1.2121, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 1.3681818181818182, | |
| "grad_norm": 2.310871124267578, | |
| "learning_rate": 0.00027272727272727274, | |
| "loss": 1.4206, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 1.3727272727272728, | |
| "grad_norm": 2.1283962726593018, | |
| "learning_rate": 0.00027196969696969696, | |
| "loss": 1.8294, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 1.3772727272727272, | |
| "grad_norm": 1.676561951637268, | |
| "learning_rate": 0.00027121212121212124, | |
| "loss": 0.9432, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 1.3818181818181818, | |
| "grad_norm": 2.4148755073547363, | |
| "learning_rate": 0.00027045454545454546, | |
| "loss": 1.8412, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 1.3863636363636362, | |
| "grad_norm": 1.668143391609192, | |
| "learning_rate": 0.00026969696969696974, | |
| "loss": 0.9952, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 1.3909090909090909, | |
| "grad_norm": 2.411818504333496, | |
| "learning_rate": 0.0002689393939393939, | |
| "loss": 1.657, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 1.3954545454545455, | |
| "grad_norm": 2.2723898887634277, | |
| "learning_rate": 0.0002681818181818182, | |
| "loss": 1.6628, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "grad_norm": 1.8225884437561035, | |
| "learning_rate": 0.0002674242424242424, | |
| "loss": 1.3039, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 1.4045454545454545, | |
| "grad_norm": 1.860181450843811, | |
| "learning_rate": 0.0002666666666666667, | |
| "loss": 1.4974, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 1.4090909090909092, | |
| "grad_norm": 2.22611927986145, | |
| "learning_rate": 0.00026590909090909096, | |
| "loss": 1.3242, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 1.4136363636363636, | |
| "grad_norm": 2.4301326274871826, | |
| "learning_rate": 0.0002651515151515151, | |
| "loss": 1.4631, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 1.4181818181818182, | |
| "grad_norm": 2.2716891765594482, | |
| "learning_rate": 0.0002643939393939394, | |
| "loss": 1.4076, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 1.4227272727272728, | |
| "grad_norm": 1.8279646635055542, | |
| "learning_rate": 0.0002636363636363636, | |
| "loss": 1.1232, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 1.4272727272727272, | |
| "grad_norm": 1.9681382179260254, | |
| "learning_rate": 0.0002628787878787879, | |
| "loss": 1.0339, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 1.4318181818181819, | |
| "grad_norm": 2.138864278793335, | |
| "learning_rate": 0.0002621212121212122, | |
| "loss": 1.4739, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 1.4363636363636363, | |
| "grad_norm": 1.997995376586914, | |
| "learning_rate": 0.00026136363636363634, | |
| "loss": 1.6025, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 1.440909090909091, | |
| "grad_norm": 2.493382215499878, | |
| "learning_rate": 0.0002606060606060606, | |
| "loss": 1.9215, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 1.4454545454545453, | |
| "grad_norm": 2.0182077884674072, | |
| "learning_rate": 0.00025984848484848484, | |
| "loss": 0.8709, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "grad_norm": 2.0383524894714355, | |
| "learning_rate": 0.0002590909090909091, | |
| "loss": 1.1791, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 1.4545454545454546, | |
| "grad_norm": 2.159406900405884, | |
| "learning_rate": 0.00025833333333333334, | |
| "loss": 1.896, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.459090909090909, | |
| "grad_norm": 2.154700756072998, | |
| "learning_rate": 0.00025757575757575756, | |
| "loss": 1.4738, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 1.4636363636363636, | |
| "grad_norm": 2.071272134780884, | |
| "learning_rate": 0.00025681818181818184, | |
| "loss": 1.4454, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 1.4681818181818183, | |
| "grad_norm": 2.1091556549072266, | |
| "learning_rate": 0.00025606060606060606, | |
| "loss": 1.2386, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 1.4727272727272727, | |
| "grad_norm": 1.8080791234970093, | |
| "learning_rate": 0.00025530303030303034, | |
| "loss": 0.9288, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 1.4772727272727273, | |
| "grad_norm": 2.170426607131958, | |
| "learning_rate": 0.0002545454545454545, | |
| "loss": 1.6025, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 1.481818181818182, | |
| "grad_norm": 2.9302620887756348, | |
| "learning_rate": 0.0002537878787878788, | |
| "loss": 2.0105, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 1.4863636363636363, | |
| "grad_norm": 2.4640023708343506, | |
| "learning_rate": 0.00025303030303030306, | |
| "loss": 1.5101, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 1.490909090909091, | |
| "grad_norm": 1.6185390949249268, | |
| "learning_rate": 0.0002522727272727273, | |
| "loss": 0.9489, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 1.4954545454545456, | |
| "grad_norm": 1.4413659572601318, | |
| "learning_rate": 0.0002515151515151515, | |
| "loss": 0.8982, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 1.884208083152771, | |
| "learning_rate": 0.00025075757575757573, | |
| "loss": 1.2771, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.5045454545454544, | |
| "grad_norm": 1.9692877531051636, | |
| "learning_rate": 0.00025, | |
| "loss": 1.6345, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 1.509090909090909, | |
| "grad_norm": 2.3343496322631836, | |
| "learning_rate": 0.00024924242424242423, | |
| "loss": 1.6179, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 1.5136363636363637, | |
| "grad_norm": 2.2422614097595215, | |
| "learning_rate": 0.0002484848484848485, | |
| "loss": 1.3785, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 1.518181818181818, | |
| "grad_norm": 2.130425214767456, | |
| "learning_rate": 0.0002477272727272727, | |
| "loss": 1.6261, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 1.5227272727272727, | |
| "grad_norm": 2.13519287109375, | |
| "learning_rate": 0.000246969696969697, | |
| "loss": 1.4136, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 1.5272727272727273, | |
| "grad_norm": 1.268389344215393, | |
| "learning_rate": 0.0002462121212121212, | |
| "loss": 0.9329, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 1.5318181818181817, | |
| "grad_norm": 2.094594955444336, | |
| "learning_rate": 0.00024545454545454545, | |
| "loss": 1.3465, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 1.5363636363636364, | |
| "grad_norm": 2.156473159790039, | |
| "learning_rate": 0.00024469696969696967, | |
| "loss": 1.2741, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 1.540909090909091, | |
| "grad_norm": 1.898336410522461, | |
| "learning_rate": 0.00024393939393939392, | |
| "loss": 1.059, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 1.5454545454545454, | |
| "grad_norm": 1.8859503269195557, | |
| "learning_rate": 0.0002431818181818182, | |
| "loss": 1.4959, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "grad_norm": 2.060011863708496, | |
| "learning_rate": 0.00024242424242424245, | |
| "loss": 1.3644, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 1.5545454545454547, | |
| "grad_norm": 2.3816416263580322, | |
| "learning_rate": 0.00024166666666666667, | |
| "loss": 1.4375, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 1.559090909090909, | |
| "grad_norm": 3.5078084468841553, | |
| "learning_rate": 0.00024090909090909092, | |
| "loss": 1.422, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 1.5636363636363635, | |
| "grad_norm": 2.221052885055542, | |
| "learning_rate": 0.00024015151515151514, | |
| "loss": 1.3024, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 1.5681818181818183, | |
| "grad_norm": 2.4711906909942627, | |
| "learning_rate": 0.0002393939393939394, | |
| "loss": 1.4838, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 1.5727272727272728, | |
| "grad_norm": 1.9803431034088135, | |
| "learning_rate": 0.00023863636363636364, | |
| "loss": 1.1055, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 1.5772727272727272, | |
| "grad_norm": 2.3196969032287598, | |
| "learning_rate": 0.0002378787878787879, | |
| "loss": 1.8282, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 1.5818181818181818, | |
| "grad_norm": 2.094829797744751, | |
| "learning_rate": 0.00023712121212121214, | |
| "loss": 1.2518, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 1.5863636363636364, | |
| "grad_norm": 2.0144240856170654, | |
| "learning_rate": 0.00023636363636363636, | |
| "loss": 1.4561, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 1.5909090909090908, | |
| "grad_norm": 1.8540327548980713, | |
| "learning_rate": 0.0002356060606060606, | |
| "loss": 1.3567, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.5954545454545455, | |
| "grad_norm": 2.6601638793945312, | |
| "learning_rate": 0.00023484848484848486, | |
| "loss": 1.7279, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 2.5020227432250977, | |
| "learning_rate": 0.0002340909090909091, | |
| "loss": 1.7535, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 1.6045454545454545, | |
| "grad_norm": 2.0830986499786377, | |
| "learning_rate": 0.00023333333333333333, | |
| "loss": 1.2985, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 1.6090909090909091, | |
| "grad_norm": 1.8451659679412842, | |
| "learning_rate": 0.00023257575757575758, | |
| "loss": 1.0175, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 1.6136363636363638, | |
| "grad_norm": 2.271484375, | |
| "learning_rate": 0.00023181818181818183, | |
| "loss": 1.4269, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.6181818181818182, | |
| "grad_norm": 4.305004596710205, | |
| "learning_rate": 0.00023106060606060605, | |
| "loss": 1.3775, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 1.6227272727272726, | |
| "grad_norm": 1.8406462669372559, | |
| "learning_rate": 0.00023030303030303033, | |
| "loss": 1.2472, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 1.6272727272727274, | |
| "grad_norm": 2.6075923442840576, | |
| "learning_rate": 0.00022954545454545455, | |
| "loss": 1.3993, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 1.6318181818181818, | |
| "grad_norm": 2.324907064437866, | |
| "learning_rate": 0.0002287878787878788, | |
| "loss": 1.3076, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 1.6363636363636362, | |
| "grad_norm": 1.954463005065918, | |
| "learning_rate": 0.00022803030303030302, | |
| "loss": 1.4135, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.6409090909090909, | |
| "grad_norm": 1.8350000381469727, | |
| "learning_rate": 0.00022727272727272727, | |
| "loss": 1.2618, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 1.6454545454545455, | |
| "grad_norm": 2.2729427814483643, | |
| "learning_rate": 0.0002265151515151515, | |
| "loss": 1.2483, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "grad_norm": 1.7203510999679565, | |
| "learning_rate": 0.00022575757575757577, | |
| "loss": 0.9189, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 1.6545454545454545, | |
| "grad_norm": 2.2685649394989014, | |
| "learning_rate": 0.00022500000000000002, | |
| "loss": 1.4897, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 1.6590909090909092, | |
| "grad_norm": 2.502887010574341, | |
| "learning_rate": 0.00022424242424242424, | |
| "loss": 1.315, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.6636363636363636, | |
| "grad_norm": 1.8945139646530151, | |
| "learning_rate": 0.0002234848484848485, | |
| "loss": 1.1686, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 1.6681818181818182, | |
| "grad_norm": 2.207409620285034, | |
| "learning_rate": 0.00022272727272727272, | |
| "loss": 1.9896, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 1.6727272727272728, | |
| "grad_norm": 2.028022527694702, | |
| "learning_rate": 0.000221969696969697, | |
| "loss": 1.5135, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 1.6772727272727272, | |
| "grad_norm": 2.403134822845459, | |
| "learning_rate": 0.00022121212121212121, | |
| "loss": 1.4201, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 1.6818181818181817, | |
| "grad_norm": 2.022304058074951, | |
| "learning_rate": 0.00022045454545454546, | |
| "loss": 1.672, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.6863636363636365, | |
| "grad_norm": 2.3457093238830566, | |
| "learning_rate": 0.0002196969696969697, | |
| "loss": 1.4657, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 1.690909090909091, | |
| "grad_norm": 2.0904908180236816, | |
| "learning_rate": 0.00021893939393939394, | |
| "loss": 1.5864, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 1.6954545454545453, | |
| "grad_norm": 1.5914294719696045, | |
| "learning_rate": 0.00021818181818181818, | |
| "loss": 1.2828, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "grad_norm": 1.8519418239593506, | |
| "learning_rate": 0.00021742424242424243, | |
| "loss": 1.5195, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 1.7045454545454546, | |
| "grad_norm": 2.2076525688171387, | |
| "learning_rate": 0.00021666666666666668, | |
| "loss": 1.247, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.709090909090909, | |
| "grad_norm": 1.8584811687469482, | |
| "learning_rate": 0.0002159090909090909, | |
| "loss": 1.6614, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 1.7136363636363636, | |
| "grad_norm": 1.9943331480026245, | |
| "learning_rate": 0.00021515151515151516, | |
| "loss": 1.2409, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 1.7181818181818183, | |
| "grad_norm": 1.5293704271316528, | |
| "learning_rate": 0.00021439393939393938, | |
| "loss": 1.1563, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 1.7227272727272727, | |
| "grad_norm": 2.835527181625366, | |
| "learning_rate": 0.00021363636363636363, | |
| "loss": 1.3795, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 1.7272727272727273, | |
| "grad_norm": 2.491914987564087, | |
| "learning_rate": 0.0002128787878787879, | |
| "loss": 1.3017, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.731818181818182, | |
| "grad_norm": 3.2870216369628906, | |
| "learning_rate": 0.00021212121212121213, | |
| "loss": 1.1947, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 1.7363636363636363, | |
| "grad_norm": 2.1510424613952637, | |
| "learning_rate": 0.00021136363636363638, | |
| "loss": 1.7048, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 1.740909090909091, | |
| "grad_norm": 2.2372002601623535, | |
| "learning_rate": 0.0002106060606060606, | |
| "loss": 1.1627, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 1.7454545454545456, | |
| "grad_norm": 2.217090368270874, | |
| "learning_rate": 0.00020984848484848485, | |
| "loss": 1.8424, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "grad_norm": 1.8897899389266968, | |
| "learning_rate": 0.00020909090909090907, | |
| "loss": 1.3006, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.7545454545454544, | |
| "grad_norm": 1.7469165325164795, | |
| "learning_rate": 0.00020833333333333335, | |
| "loss": 1.496, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 1.759090909090909, | |
| "grad_norm": 1.8982353210449219, | |
| "learning_rate": 0.0002075757575757576, | |
| "loss": 1.2682, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 1.7636363636363637, | |
| "grad_norm": 2.0795273780822754, | |
| "learning_rate": 0.00020681818181818182, | |
| "loss": 2.2314, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 1.768181818181818, | |
| "grad_norm": 1.6244702339172363, | |
| "learning_rate": 0.00020606060606060607, | |
| "loss": 0.9585, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 1.7727272727272727, | |
| "grad_norm": 2.3025052547454834, | |
| "learning_rate": 0.0002053030303030303, | |
| "loss": 1.6639, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.7772727272727273, | |
| "grad_norm": 2.418119192123413, | |
| "learning_rate": 0.00020454545454545457, | |
| "loss": 1.8274, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 1.7818181818181817, | |
| "grad_norm": 1.70542573928833, | |
| "learning_rate": 0.0002037878787878788, | |
| "loss": 1.3115, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 1.7863636363636364, | |
| "grad_norm": 1.9681594371795654, | |
| "learning_rate": 0.00020303030303030304, | |
| "loss": 0.973, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 1.790909090909091, | |
| "grad_norm": 1.856879711151123, | |
| "learning_rate": 0.0002022727272727273, | |
| "loss": 1.2958, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 1.7954545454545454, | |
| "grad_norm": 2.1984550952911377, | |
| "learning_rate": 0.0002015151515151515, | |
| "loss": 1.5643, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 2.6989824771881104, | |
| "learning_rate": 0.00020075757575757576, | |
| "loss": 1.4334, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 1.8045454545454547, | |
| "grad_norm": 2.3298850059509277, | |
| "learning_rate": 0.0002, | |
| "loss": 1.5145, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 1.809090909090909, | |
| "grad_norm": 1.9490818977355957, | |
| "learning_rate": 0.00019924242424242426, | |
| "loss": 1.5346, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 1.8136363636363635, | |
| "grad_norm": 2.0060818195343018, | |
| "learning_rate": 0.00019848484848484848, | |
| "loss": 1.2493, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 1.8181818181818183, | |
| "grad_norm": 2.1615042686462402, | |
| "learning_rate": 0.00019772727272727273, | |
| "loss": 1.8834, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.8227272727272728, | |
| "grad_norm": 2.4424972534179688, | |
| "learning_rate": 0.00019696969696969695, | |
| "loss": 1.6012, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 1.8272727272727272, | |
| "grad_norm": 1.83026921749115, | |
| "learning_rate": 0.0001962121212121212, | |
| "loss": 1.4308, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 1.8318181818181818, | |
| "grad_norm": 2.6793453693389893, | |
| "learning_rate": 0.00019545454545454548, | |
| "loss": 1.6356, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 1.8363636363636364, | |
| "grad_norm": 2.1211740970611572, | |
| "learning_rate": 0.0001946969696969697, | |
| "loss": 1.0859, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 1.8409090909090908, | |
| "grad_norm": 1.6658772230148315, | |
| "learning_rate": 0.00019393939393939395, | |
| "loss": 1.3467, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.8454545454545455, | |
| "grad_norm": 1.7903032302856445, | |
| "learning_rate": 0.00019318181818181817, | |
| "loss": 1.0118, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "grad_norm": 1.7358275651931763, | |
| "learning_rate": 0.00019242424242424242, | |
| "loss": 1.4404, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 1.8545454545454545, | |
| "grad_norm": 2.0745861530303955, | |
| "learning_rate": 0.00019166666666666667, | |
| "loss": 1.3409, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 1.8590909090909091, | |
| "grad_norm": 2.1365160942077637, | |
| "learning_rate": 0.00019090909090909092, | |
| "loss": 1.4658, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 1.8636363636363638, | |
| "grad_norm": 2.040371894836426, | |
| "learning_rate": 0.00019015151515151517, | |
| "loss": 1.3165, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.8681818181818182, | |
| "grad_norm": 1.9335429668426514, | |
| "learning_rate": 0.0001893939393939394, | |
| "loss": 1.5063, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 1.8727272727272726, | |
| "grad_norm": 2.0025243759155273, | |
| "learning_rate": 0.00018863636363636364, | |
| "loss": 1.2584, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 1.8772727272727274, | |
| "grad_norm": 1.7558890581130981, | |
| "learning_rate": 0.00018787878787878787, | |
| "loss": 1.0937, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 1.8818181818181818, | |
| "grad_norm": 1.886003851890564, | |
| "learning_rate": 0.00018712121212121214, | |
| "loss": 1.3969, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 1.8863636363636362, | |
| "grad_norm": 2.6999497413635254, | |
| "learning_rate": 0.00018636363636363636, | |
| "loss": 1.1411, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.8909090909090909, | |
| "grad_norm": 1.923196792602539, | |
| "learning_rate": 0.00018560606060606061, | |
| "loss": 1.6597, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 1.8954545454545455, | |
| "grad_norm": 2.261504650115967, | |
| "learning_rate": 0.00018484848484848484, | |
| "loss": 1.738, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "grad_norm": 1.9908875226974487, | |
| "learning_rate": 0.00018409090909090909, | |
| "loss": 1.1275, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 1.9045454545454545, | |
| "grad_norm": 2.348240852355957, | |
| "learning_rate": 0.00018333333333333334, | |
| "loss": 1.5368, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 1.9090909090909092, | |
| "grad_norm": 3.1023001670837402, | |
| "learning_rate": 0.00018257575757575758, | |
| "loss": 1.1828, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.9136363636363636, | |
| "grad_norm": 2.2243757247924805, | |
| "learning_rate": 0.00018181818181818183, | |
| "loss": 1.4783, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 1.9181818181818182, | |
| "grad_norm": 1.824317216873169, | |
| "learning_rate": 0.00018106060606060606, | |
| "loss": 1.178, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 1.9227272727272728, | |
| "grad_norm": 2.1963822841644287, | |
| "learning_rate": 0.0001803030303030303, | |
| "loss": 1.5811, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 1.9272727272727272, | |
| "grad_norm": 1.8573740720748901, | |
| "learning_rate": 0.00017954545454545453, | |
| "loss": 1.2361, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 1.9318181818181817, | |
| "grad_norm": 2.223315715789795, | |
| "learning_rate": 0.0001787878787878788, | |
| "loss": 1.3236, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.9363636363636365, | |
| "grad_norm": 2.0890495777130127, | |
| "learning_rate": 0.00017803030303030305, | |
| "loss": 1.4358, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 1.940909090909091, | |
| "grad_norm": 1.8097957372665405, | |
| "learning_rate": 0.00017727272727272728, | |
| "loss": 1.0634, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 1.9454545454545453, | |
| "grad_norm": 2.378750801086426, | |
| "learning_rate": 0.00017651515151515153, | |
| "loss": 1.8565, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "grad_norm": 1.7777960300445557, | |
| "learning_rate": 0.00017575757575757575, | |
| "loss": 1.5755, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 1.9545454545454546, | |
| "grad_norm": 2.1310245990753174, | |
| "learning_rate": 0.000175, | |
| "loss": 1.4119, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.959090909090909, | |
| "grad_norm": 2.506479024887085, | |
| "learning_rate": 0.00017424242424242425, | |
| "loss": 1.5025, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 1.9636363636363636, | |
| "grad_norm": 2.091891288757324, | |
| "learning_rate": 0.0001734848484848485, | |
| "loss": 1.383, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 1.9681818181818183, | |
| "grad_norm": 1.7450625896453857, | |
| "learning_rate": 0.00017272727272727272, | |
| "loss": 1.4122, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 1.9727272727272727, | |
| "grad_norm": 1.7834117412567139, | |
| "learning_rate": 0.00017196969696969697, | |
| "loss": 1.0932, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 1.9772727272727273, | |
| "grad_norm": 1.6854950189590454, | |
| "learning_rate": 0.00017121212121212122, | |
| "loss": 0.9985, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.981818181818182, | |
| "grad_norm": 1.4569097757339478, | |
| "learning_rate": 0.00017045454545454544, | |
| "loss": 1.1354, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 1.9863636363636363, | |
| "grad_norm": 2.3083584308624268, | |
| "learning_rate": 0.00016969696969696972, | |
| "loss": 1.4856, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 1.990909090909091, | |
| "grad_norm": 2.1282572746276855, | |
| "learning_rate": 0.00016893939393939394, | |
| "loss": 1.942, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 1.9954545454545456, | |
| "grad_norm": 2.199822187423706, | |
| "learning_rate": 0.0001681818181818182, | |
| "loss": 1.4891, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 2.0030460357666016, | |
| "learning_rate": 0.0001674242424242424, | |
| "loss": 1.2939, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_f1": 0.8928, | |
| "eval_gen_len": 42.0, | |
| "eval_loss": 1.7928985357284546, | |
| "eval_precision": 0.8914, | |
| "eval_recall": 0.8944, | |
| "eval_rouge1": 0.4605, | |
| "eval_rouge2": 0.2125, | |
| "eval_rougeL": 0.3897, | |
| "eval_rougeLsum": 0.4259, | |
| "eval_runtime": 25.1108, | |
| "eval_samples_per_second": 4.381, | |
| "eval_steps_per_second": 0.558, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 2.0045454545454544, | |
| "grad_norm": 1.4425781965255737, | |
| "learning_rate": 0.00016666666666666666, | |
| "loss": 0.9057, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 2.0090909090909093, | |
| "grad_norm": 1.579765796661377, | |
| "learning_rate": 0.00016590909090909094, | |
| "loss": 0.7069, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 2.0136363636363637, | |
| "grad_norm": 1.8639825582504272, | |
| "learning_rate": 0.00016515151515151516, | |
| "loss": 1.1531, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 2.018181818181818, | |
| "grad_norm": 1.4890676736831665, | |
| "learning_rate": 0.0001643939393939394, | |
| "loss": 0.8112, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 2.022727272727273, | |
| "grad_norm": 1.7381901741027832, | |
| "learning_rate": 0.00016363636363636363, | |
| "loss": 1.2108, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 2.0272727272727273, | |
| "grad_norm": 1.6125924587249756, | |
| "learning_rate": 0.00016287878787878788, | |
| "loss": 1.0529, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 2.0318181818181817, | |
| "grad_norm": 1.8624428510665894, | |
| "learning_rate": 0.0001621212121212121, | |
| "loss": 1.006, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 2.036363636363636, | |
| "grad_norm": 1.719439148902893, | |
| "learning_rate": 0.00016136363636363638, | |
| "loss": 1.0881, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 2.040909090909091, | |
| "grad_norm": 2.446216106414795, | |
| "learning_rate": 0.00016060606060606063, | |
| "loss": 1.2399, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 2.0454545454545454, | |
| "grad_norm": 1.703517198562622, | |
| "learning_rate": 0.00015984848484848485, | |
| "loss": 0.931, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "grad_norm": 1.780228853225708, | |
| "learning_rate": 0.0001590909090909091, | |
| "loss": 0.9769, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 2.0545454545454547, | |
| "grad_norm": 2.015679121017456, | |
| "learning_rate": 0.00015833333333333332, | |
| "loss": 1.6044, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 2.059090909090909, | |
| "grad_norm": 2.084481716156006, | |
| "learning_rate": 0.00015757575757575757, | |
| "loss": 0.9933, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 2.0636363636363635, | |
| "grad_norm": 2.3098299503326416, | |
| "learning_rate": 0.00015681818181818182, | |
| "loss": 0.9405, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 2.0681818181818183, | |
| "grad_norm": 1.8041385412216187, | |
| "learning_rate": 0.00015606060606060607, | |
| "loss": 1.1748, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 2.0727272727272728, | |
| "grad_norm": 1.693158745765686, | |
| "learning_rate": 0.0001553030303030303, | |
| "loss": 0.9358, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 2.077272727272727, | |
| "grad_norm": 1.5484883785247803, | |
| "learning_rate": 0.00015454545454545454, | |
| "loss": 1.0664, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 2.081818181818182, | |
| "grad_norm": 1.4313092231750488, | |
| "learning_rate": 0.0001537878787878788, | |
| "loss": 0.6624, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 2.0863636363636364, | |
| "grad_norm": 2.218092679977417, | |
| "learning_rate": 0.00015303030303030302, | |
| "loss": 0.9856, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 2.090909090909091, | |
| "grad_norm": 2.030869960784912, | |
| "learning_rate": 0.0001522727272727273, | |
| "loss": 1.143, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 2.0954545454545457, | |
| "grad_norm": 2.190603017807007, | |
| "learning_rate": 0.00015151515151515152, | |
| "loss": 1.077, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "grad_norm": 1.3030821084976196, | |
| "learning_rate": 0.00015075757575757576, | |
| "loss": 0.6711, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 2.1045454545454545, | |
| "grad_norm": 1.8678494691848755, | |
| "learning_rate": 0.00015, | |
| "loss": 1.0674, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 2.109090909090909, | |
| "grad_norm": 1.407085657119751, | |
| "learning_rate": 0.00014924242424242424, | |
| "loss": 0.7024, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 2.1136363636363638, | |
| "grad_norm": 2.004911422729492, | |
| "learning_rate": 0.0001484848484848485, | |
| "loss": 0.795, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 2.118181818181818, | |
| "grad_norm": 2.25128173828125, | |
| "learning_rate": 0.00014772727272727274, | |
| "loss": 1.2232, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 2.1227272727272726, | |
| "grad_norm": 1.960771918296814, | |
| "learning_rate": 0.00014696969696969698, | |
| "loss": 1.0019, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 2.1272727272727274, | |
| "grad_norm": 1.9563887119293213, | |
| "learning_rate": 0.0001462121212121212, | |
| "loss": 0.9798, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 2.131818181818182, | |
| "grad_norm": 1.687361240386963, | |
| "learning_rate": 0.00014545454545454546, | |
| "loss": 0.755, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 2.1363636363636362, | |
| "grad_norm": 2.191286325454712, | |
| "learning_rate": 0.00014469696969696968, | |
| "loss": 1.0018, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 2.140909090909091, | |
| "grad_norm": 2.046880006790161, | |
| "learning_rate": 0.00014393939393939396, | |
| "loss": 1.2281, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 2.1454545454545455, | |
| "grad_norm": 2.4996211528778076, | |
| "learning_rate": 0.00014318181818181818, | |
| "loss": 1.0795, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "grad_norm": 1.8937994241714478, | |
| "learning_rate": 0.00014242424242424243, | |
| "loss": 1.0556, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 2.1545454545454543, | |
| "grad_norm": 2.250491142272949, | |
| "learning_rate": 0.00014166666666666668, | |
| "loss": 0.8816, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 2.159090909090909, | |
| "grad_norm": NaN, | |
| "learning_rate": 0.00014166666666666668, | |
| "loss": 0.0, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 2.1636363636363636, | |
| "grad_norm": 2.231706142425537, | |
| "learning_rate": 0.0001409090909090909, | |
| "loss": 1.2344, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 2.168181818181818, | |
| "grad_norm": 2.2170498371124268, | |
| "learning_rate": 0.00014015151515151515, | |
| "loss": 1.4409, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 2.172727272727273, | |
| "grad_norm": 2.3106095790863037, | |
| "learning_rate": 0.0001393939393939394, | |
| "loss": 0.9081, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 2.1772727272727272, | |
| "grad_norm": 1.9665738344192505, | |
| "learning_rate": 0.00013863636363636365, | |
| "loss": 1.3029, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 2.1818181818181817, | |
| "grad_norm": 2.321331739425659, | |
| "learning_rate": 0.00013787878787878787, | |
| "loss": 1.4714, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 2.1863636363636365, | |
| "grad_norm": 2.0038533210754395, | |
| "learning_rate": 0.00013712121212121212, | |
| "loss": 1.0879, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 2.190909090909091, | |
| "grad_norm": 1.6077767610549927, | |
| "learning_rate": 0.00013636363636363637, | |
| "loss": 0.6456, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 2.1954545454545453, | |
| "grad_norm": 1.5018125772476196, | |
| "learning_rate": 0.00013560606060606062, | |
| "loss": 0.6937, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "grad_norm": 1.5473967790603638, | |
| "learning_rate": 0.00013484848484848487, | |
| "loss": 0.6191, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 2.2045454545454546, | |
| "grad_norm": 2.5554354190826416, | |
| "learning_rate": 0.0001340909090909091, | |
| "loss": 1.4345, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 2.209090909090909, | |
| "grad_norm": 2.1666858196258545, | |
| "learning_rate": 0.00013333333333333334, | |
| "loss": 1.4182, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 2.213636363636364, | |
| "grad_norm": 2.2915759086608887, | |
| "learning_rate": 0.00013257575757575756, | |
| "loss": 0.8752, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 2.2181818181818183, | |
| "grad_norm": 2.24314546585083, | |
| "learning_rate": 0.0001318181818181818, | |
| "loss": 1.3214, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 2.2227272727272727, | |
| "grad_norm": 2.269216537475586, | |
| "learning_rate": 0.0001310606060606061, | |
| "loss": 0.9968, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 2.227272727272727, | |
| "grad_norm": 2.3108322620391846, | |
| "learning_rate": 0.0001303030303030303, | |
| "loss": 0.9695, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 2.231818181818182, | |
| "grad_norm": 2.3146250247955322, | |
| "learning_rate": 0.00012954545454545456, | |
| "loss": 1.4007, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 2.2363636363636363, | |
| "grad_norm": 1.9747002124786377, | |
| "learning_rate": 0.00012878787878787878, | |
| "loss": 0.8876, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 2.2409090909090907, | |
| "grad_norm": 2.0410826206207275, | |
| "learning_rate": 0.00012803030303030303, | |
| "loss": 0.9588, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 2.2454545454545456, | |
| "grad_norm": 2.2743778228759766, | |
| "learning_rate": 0.00012727272727272725, | |
| "loss": 1.2062, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "grad_norm": 2.272749662399292, | |
| "learning_rate": 0.00012651515151515153, | |
| "loss": 0.975, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 2.2545454545454544, | |
| "grad_norm": 2.297175884246826, | |
| "learning_rate": 0.00012575757575757575, | |
| "loss": 1.0806, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 2.2590909090909093, | |
| "grad_norm": 2.2274718284606934, | |
| "learning_rate": 0.000125, | |
| "loss": 0.9391, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 2.2636363636363637, | |
| "grad_norm": 2.4175453186035156, | |
| "learning_rate": 0.00012424242424242425, | |
| "loss": 1.2736, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 2.268181818181818, | |
| "grad_norm": 1.7530089616775513, | |
| "learning_rate": 0.0001234848484848485, | |
| "loss": 1.1917, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 2.2727272727272725, | |
| "grad_norm": 2.598747730255127, | |
| "learning_rate": 0.00012272727272727272, | |
| "loss": 1.5901, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 2.2772727272727273, | |
| "grad_norm": 2.0590171813964844, | |
| "learning_rate": 0.00012196969696969696, | |
| "loss": 1.0049, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 2.2818181818181817, | |
| "grad_norm": 1.6530457735061646, | |
| "learning_rate": 0.00012121212121212122, | |
| "loss": 0.6991, | |
| "step": 502 | |
| }, | |
| { | |
| "epoch": 2.286363636363636, | |
| "grad_norm": 1.4000625610351562, | |
| "learning_rate": 0.00012045454545454546, | |
| "loss": 0.7258, | |
| "step": 503 | |
| }, | |
| { | |
| "epoch": 2.290909090909091, | |
| "grad_norm": 3.4282798767089844, | |
| "learning_rate": 0.0001196969696969697, | |
| "loss": 0.7331, | |
| "step": 504 | |
| }, | |
| { | |
| "epoch": 2.2954545454545454, | |
| "grad_norm": 2.0328640937805176, | |
| "learning_rate": 0.00011893939393939394, | |
| "loss": 1.0245, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "grad_norm": 2.639125347137451, | |
| "learning_rate": 0.00011818181818181818, | |
| "loss": 0.939, | |
| "step": 506 | |
| }, | |
| { | |
| "epoch": 2.3045454545454547, | |
| "grad_norm": 2.069645643234253, | |
| "learning_rate": 0.00011742424242424243, | |
| "loss": 1.186, | |
| "step": 507 | |
| }, | |
| { | |
| "epoch": 2.309090909090909, | |
| "grad_norm": 2.103675603866577, | |
| "learning_rate": 0.00011666666666666667, | |
| "loss": 1.0986, | |
| "step": 508 | |
| }, | |
| { | |
| "epoch": 2.3136363636363635, | |
| "grad_norm": 2.022813320159912, | |
| "learning_rate": 0.00011590909090909091, | |
| "loss": 1.1106, | |
| "step": 509 | |
| }, | |
| { | |
| "epoch": 2.3181818181818183, | |
| "grad_norm": 2.1240341663360596, | |
| "learning_rate": 0.00011515151515151516, | |
| "loss": 0.9754, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 2.3227272727272728, | |
| "grad_norm": 2.11362361907959, | |
| "learning_rate": 0.0001143939393939394, | |
| "loss": 1.243, | |
| "step": 511 | |
| }, | |
| { | |
| "epoch": 2.327272727272727, | |
| "grad_norm": 1.9033676385879517, | |
| "learning_rate": 0.00011363636363636364, | |
| "loss": 0.7314, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 2.331818181818182, | |
| "grad_norm": 2.7902817726135254, | |
| "learning_rate": 0.00011287878787878789, | |
| "loss": 1.2161, | |
| "step": 513 | |
| }, | |
| { | |
| "epoch": 2.3363636363636364, | |
| "grad_norm": 2.1139214038848877, | |
| "learning_rate": 0.00011212121212121212, | |
| "loss": 1.4216, | |
| "step": 514 | |
| }, | |
| { | |
| "epoch": 2.340909090909091, | |
| "grad_norm": 2.2380800247192383, | |
| "learning_rate": 0.00011136363636363636, | |
| "loss": 1.0319, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 2.3454545454545457, | |
| "grad_norm": 1.9591755867004395, | |
| "learning_rate": 0.00011060606060606061, | |
| "loss": 0.7923, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "grad_norm": 2.6767358779907227, | |
| "learning_rate": 0.00010984848484848486, | |
| "loss": 0.9721, | |
| "step": 517 | |
| }, | |
| { | |
| "epoch": 2.3545454545454545, | |
| "grad_norm": 2.350008487701416, | |
| "learning_rate": 0.00010909090909090909, | |
| "loss": 1.1793, | |
| "step": 518 | |
| }, | |
| { | |
| "epoch": 2.359090909090909, | |
| "grad_norm": 2.0240652561187744, | |
| "learning_rate": 0.00010833333333333334, | |
| "loss": 1.1184, | |
| "step": 519 | |
| }, | |
| { | |
| "epoch": 2.3636363636363638, | |
| "grad_norm": 2.058748960494995, | |
| "learning_rate": 0.00010757575757575758, | |
| "loss": 1.1886, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 2.368181818181818, | |
| "grad_norm": 1.7921950817108154, | |
| "learning_rate": 0.00010681818181818181, | |
| "loss": 0.8511, | |
| "step": 521 | |
| }, | |
| { | |
| "epoch": 2.3727272727272726, | |
| "grad_norm": 2.027445077896118, | |
| "learning_rate": 0.00010606060606060606, | |
| "loss": 0.8641, | |
| "step": 522 | |
| }, | |
| { | |
| "epoch": 2.3772727272727274, | |
| "grad_norm": 1.8156445026397705, | |
| "learning_rate": 0.0001053030303030303, | |
| "loss": 0.8234, | |
| "step": 523 | |
| }, | |
| { | |
| "epoch": 2.381818181818182, | |
| "grad_norm": 2.3511455059051514, | |
| "learning_rate": 0.00010454545454545454, | |
| "loss": 1.048, | |
| "step": 524 | |
| }, | |
| { | |
| "epoch": 2.3863636363636362, | |
| "grad_norm": 1.489744782447815, | |
| "learning_rate": 0.0001037878787878788, | |
| "loss": 0.4886, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 2.390909090909091, | |
| "grad_norm": 2.0359721183776855, | |
| "learning_rate": 0.00010303030303030303, | |
| "loss": 1.0011, | |
| "step": 526 | |
| }, | |
| { | |
| "epoch": 2.3954545454545455, | |
| "grad_norm": 2.8290212154388428, | |
| "learning_rate": 0.00010227272727272728, | |
| "loss": 1.4443, | |
| "step": 527 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 1.991904616355896, | |
| "learning_rate": 0.00010151515151515152, | |
| "loss": 0.9877, | |
| "step": 528 | |
| }, | |
| { | |
| "epoch": 2.4045454545454543, | |
| "grad_norm": 1.8174313306808472, | |
| "learning_rate": 0.00010075757575757576, | |
| "loss": 0.9048, | |
| "step": 529 | |
| }, | |
| { | |
| "epoch": 2.409090909090909, | |
| "grad_norm": 1.66022527217865, | |
| "learning_rate": 0.0001, | |
| "loss": 0.9039, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 2.4136363636363636, | |
| "grad_norm": 1.6025142669677734, | |
| "learning_rate": 9.924242424242424e-05, | |
| "loss": 0.8169, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 2.418181818181818, | |
| "grad_norm": 1.871733546257019, | |
| "learning_rate": 9.848484848484848e-05, | |
| "loss": 1.039, | |
| "step": 532 | |
| }, | |
| { | |
| "epoch": 2.422727272727273, | |
| "grad_norm": 2.35320782661438, | |
| "learning_rate": 9.772727272727274e-05, | |
| "loss": 1.3449, | |
| "step": 533 | |
| }, | |
| { | |
| "epoch": 2.4272727272727272, | |
| "grad_norm": 1.9311728477478027, | |
| "learning_rate": 9.696969696969698e-05, | |
| "loss": 1.0332, | |
| "step": 534 | |
| }, | |
| { | |
| "epoch": 2.4318181818181817, | |
| "grad_norm": 1.6838319301605225, | |
| "learning_rate": 9.621212121212121e-05, | |
| "loss": 0.6631, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 2.4363636363636365, | |
| "grad_norm": 1.9957849979400635, | |
| "learning_rate": 9.545454545454546e-05, | |
| "loss": 1.0397, | |
| "step": 536 | |
| }, | |
| { | |
| "epoch": 2.440909090909091, | |
| "grad_norm": 2.338730573654175, | |
| "learning_rate": 9.46969696969697e-05, | |
| "loss": 1.1155, | |
| "step": 537 | |
| }, | |
| { | |
| "epoch": 2.4454545454545453, | |
| "grad_norm": 2.0578792095184326, | |
| "learning_rate": 9.393939393939393e-05, | |
| "loss": 1.0634, | |
| "step": 538 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "grad_norm": 2.0512609481811523, | |
| "learning_rate": 9.318181818181818e-05, | |
| "loss": 0.9052, | |
| "step": 539 | |
| }, | |
| { | |
| "epoch": 2.4545454545454546, | |
| "grad_norm": 2.2808845043182373, | |
| "learning_rate": 9.242424242424242e-05, | |
| "loss": 1.2479, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 2.459090909090909, | |
| "grad_norm": 1.7963327169418335, | |
| "learning_rate": 9.166666666666667e-05, | |
| "loss": 0.8655, | |
| "step": 541 | |
| }, | |
| { | |
| "epoch": 2.463636363636364, | |
| "grad_norm": 2.378777265548706, | |
| "learning_rate": 9.090909090909092e-05, | |
| "loss": 1.1019, | |
| "step": 542 | |
| }, | |
| { | |
| "epoch": 2.4681818181818183, | |
| "grad_norm": 1.7346596717834473, | |
| "learning_rate": 9.015151515151515e-05, | |
| "loss": 0.6478, | |
| "step": 543 | |
| }, | |
| { | |
| "epoch": 2.4727272727272727, | |
| "grad_norm": 1.8121107816696167, | |
| "learning_rate": 8.93939393939394e-05, | |
| "loss": 0.9549, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 2.4772727272727275, | |
| "grad_norm": 1.9102083444595337, | |
| "learning_rate": 8.863636363636364e-05, | |
| "loss": 0.9103, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 2.481818181818182, | |
| "grad_norm": 2.3148677349090576, | |
| "learning_rate": 8.787878787878787e-05, | |
| "loss": 1.1075, | |
| "step": 546 | |
| }, | |
| { | |
| "epoch": 2.4863636363636363, | |
| "grad_norm": 2.3098530769348145, | |
| "learning_rate": 8.712121212121212e-05, | |
| "loss": 1.0885, | |
| "step": 547 | |
| }, | |
| { | |
| "epoch": 2.4909090909090907, | |
| "grad_norm": 2.061582565307617, | |
| "learning_rate": 8.636363636363636e-05, | |
| "loss": 0.7894, | |
| "step": 548 | |
| }, | |
| { | |
| "epoch": 2.4954545454545456, | |
| "grad_norm": 2.3829803466796875, | |
| "learning_rate": 8.560606060606061e-05, | |
| "loss": 1.2397, | |
| "step": 549 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 2.111055612564087, | |
| "learning_rate": 8.484848484848486e-05, | |
| "loss": 1.0463, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 2.5045454545454544, | |
| "grad_norm": 1.883468508720398, | |
| "learning_rate": 8.40909090909091e-05, | |
| "loss": 0.9837, | |
| "step": 551 | |
| }, | |
| { | |
| "epoch": 2.509090909090909, | |
| "grad_norm": 1.8480087518692017, | |
| "learning_rate": 8.333333333333333e-05, | |
| "loss": 0.897, | |
| "step": 552 | |
| }, | |
| { | |
| "epoch": 2.5136363636363637, | |
| "grad_norm": 1.9513871669769287, | |
| "learning_rate": 8.257575757575758e-05, | |
| "loss": 0.9668, | |
| "step": 553 | |
| }, | |
| { | |
| "epoch": 2.518181818181818, | |
| "grad_norm": 1.5687415599822998, | |
| "learning_rate": 8.181818181818182e-05, | |
| "loss": 0.8729, | |
| "step": 554 | |
| }, | |
| { | |
| "epoch": 2.5227272727272725, | |
| "grad_norm": 1.959887981414795, | |
| "learning_rate": 8.106060606060605e-05, | |
| "loss": 0.9612, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 2.5272727272727273, | |
| "grad_norm": 2.1609091758728027, | |
| "learning_rate": 8.030303030303031e-05, | |
| "loss": 1.133, | |
| "step": 556 | |
| }, | |
| { | |
| "epoch": 2.5318181818181817, | |
| "grad_norm": 2.534611225128174, | |
| "learning_rate": 7.954545454545455e-05, | |
| "loss": 1.3566, | |
| "step": 557 | |
| }, | |
| { | |
| "epoch": 2.536363636363636, | |
| "grad_norm": 2.731877088546753, | |
| "learning_rate": 7.878787878787879e-05, | |
| "loss": 1.1991, | |
| "step": 558 | |
| }, | |
| { | |
| "epoch": 2.540909090909091, | |
| "grad_norm": 2.1953938007354736, | |
| "learning_rate": 7.803030303030304e-05, | |
| "loss": 0.932, | |
| "step": 559 | |
| }, | |
| { | |
| "epoch": 2.5454545454545454, | |
| "grad_norm": 2.260007381439209, | |
| "learning_rate": 7.727272727272727e-05, | |
| "loss": 1.0682, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "grad_norm": 2.9932045936584473, | |
| "learning_rate": 7.651515151515151e-05, | |
| "loss": 1.2489, | |
| "step": 561 | |
| }, | |
| { | |
| "epoch": 2.5545454545454547, | |
| "grad_norm": 2.4135005474090576, | |
| "learning_rate": 7.575757575757576e-05, | |
| "loss": 0.7289, | |
| "step": 562 | |
| }, | |
| { | |
| "epoch": 2.559090909090909, | |
| "grad_norm": 2.2235300540924072, | |
| "learning_rate": 7.5e-05, | |
| "loss": 0.7027, | |
| "step": 563 | |
| }, | |
| { | |
| "epoch": 2.5636363636363635, | |
| "grad_norm": 2.6621127128601074, | |
| "learning_rate": 7.424242424242426e-05, | |
| "loss": 1.2601, | |
| "step": 564 | |
| }, | |
| { | |
| "epoch": 2.5681818181818183, | |
| "grad_norm": 2.574686050415039, | |
| "learning_rate": 7.348484848484849e-05, | |
| "loss": 1.1076, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 2.5727272727272728, | |
| "grad_norm": 2.416339874267578, | |
| "learning_rate": 7.272727272727273e-05, | |
| "loss": 0.9473, | |
| "step": 566 | |
| }, | |
| { | |
| "epoch": 2.577272727272727, | |
| "grad_norm": 1.7082793712615967, | |
| "learning_rate": 7.196969696969698e-05, | |
| "loss": 0.7671, | |
| "step": 567 | |
| }, | |
| { | |
| "epoch": 2.581818181818182, | |
| "grad_norm": 2.220196008682251, | |
| "learning_rate": 7.121212121212121e-05, | |
| "loss": 1.1754, | |
| "step": 568 | |
| }, | |
| { | |
| "epoch": 2.5863636363636364, | |
| "grad_norm": 2.26267409324646, | |
| "learning_rate": 7.045454545454545e-05, | |
| "loss": 1.4229, | |
| "step": 569 | |
| }, | |
| { | |
| "epoch": 2.590909090909091, | |
| "grad_norm": 1.7881556749343872, | |
| "learning_rate": 6.96969696969697e-05, | |
| "loss": 0.8333, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 2.5954545454545457, | |
| "grad_norm": 2.156179904937744, | |
| "learning_rate": 6.893939393939393e-05, | |
| "loss": 1.1788, | |
| "step": 571 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "grad_norm": 1.6093627214431763, | |
| "learning_rate": 6.818181818181818e-05, | |
| "loss": 0.6442, | |
| "step": 572 | |
| }, | |
| { | |
| "epoch": 2.6045454545454545, | |
| "grad_norm": 1.984737753868103, | |
| "learning_rate": 6.742424242424243e-05, | |
| "loss": 0.9969, | |
| "step": 573 | |
| }, | |
| { | |
| "epoch": 2.6090909090909093, | |
| "grad_norm": 1.958917498588562, | |
| "learning_rate": 6.666666666666667e-05, | |
| "loss": 0.8534, | |
| "step": 574 | |
| }, | |
| { | |
| "epoch": 2.6136363636363638, | |
| "grad_norm": 3.060192346572876, | |
| "learning_rate": 6.59090909090909e-05, | |
| "loss": 1.4748, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 2.618181818181818, | |
| "grad_norm": 1.9157240390777588, | |
| "learning_rate": 6.515151515151516e-05, | |
| "loss": 0.8512, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 2.6227272727272726, | |
| "grad_norm": 1.6468448638916016, | |
| "learning_rate": 6.439393939393939e-05, | |
| "loss": 0.7536, | |
| "step": 577 | |
| }, | |
| { | |
| "epoch": 2.6272727272727274, | |
| "grad_norm": 1.9406344890594482, | |
| "learning_rate": 6.363636363636363e-05, | |
| "loss": 0.9798, | |
| "step": 578 | |
| }, | |
| { | |
| "epoch": 2.631818181818182, | |
| "grad_norm": 1.8992547988891602, | |
| "learning_rate": 6.287878787878788e-05, | |
| "loss": 1.1394, | |
| "step": 579 | |
| }, | |
| { | |
| "epoch": 2.6363636363636362, | |
| "grad_norm": 1.8168598413467407, | |
| "learning_rate": 6.212121212121213e-05, | |
| "loss": 0.8848, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 2.6409090909090907, | |
| "grad_norm": 2.8009986877441406, | |
| "learning_rate": 6.136363636363636e-05, | |
| "loss": 1.1817, | |
| "step": 581 | |
| }, | |
| { | |
| "epoch": 2.6454545454545455, | |
| "grad_norm": 1.8650470972061157, | |
| "learning_rate": 6.060606060606061e-05, | |
| "loss": 0.9148, | |
| "step": 582 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "grad_norm": 2.132161855697632, | |
| "learning_rate": 5.984848484848485e-05, | |
| "loss": 1.0103, | |
| "step": 583 | |
| }, | |
| { | |
| "epoch": 2.6545454545454543, | |
| "grad_norm": 2.488576650619507, | |
| "learning_rate": 5.909090909090909e-05, | |
| "loss": 1.5804, | |
| "step": 584 | |
| }, | |
| { | |
| "epoch": 2.659090909090909, | |
| "grad_norm": 1.7953377962112427, | |
| "learning_rate": 5.833333333333333e-05, | |
| "loss": 0.8419, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 2.6636363636363636, | |
| "grad_norm": 2.563900947570801, | |
| "learning_rate": 5.757575757575758e-05, | |
| "loss": 1.1122, | |
| "step": 586 | |
| }, | |
| { | |
| "epoch": 2.668181818181818, | |
| "grad_norm": 2.112504243850708, | |
| "learning_rate": 5.681818181818182e-05, | |
| "loss": 0.8345, | |
| "step": 587 | |
| }, | |
| { | |
| "epoch": 2.672727272727273, | |
| "grad_norm": 2.874629020690918, | |
| "learning_rate": 5.606060606060606e-05, | |
| "loss": 1.257, | |
| "step": 588 | |
| }, | |
| { | |
| "epoch": 2.6772727272727272, | |
| "grad_norm": 2.3965139389038086, | |
| "learning_rate": 5.5303030303030304e-05, | |
| "loss": 1.4174, | |
| "step": 589 | |
| }, | |
| { | |
| "epoch": 2.6818181818181817, | |
| "grad_norm": 2.149787425994873, | |
| "learning_rate": 5.4545454545454546e-05, | |
| "loss": 1.0162, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 2.6863636363636365, | |
| "grad_norm": 3.67689847946167, | |
| "learning_rate": 5.378787878787879e-05, | |
| "loss": 0.9925, | |
| "step": 591 | |
| }, | |
| { | |
| "epoch": 2.690909090909091, | |
| "grad_norm": 2.144545316696167, | |
| "learning_rate": 5.303030303030303e-05, | |
| "loss": 1.2257, | |
| "step": 592 | |
| }, | |
| { | |
| "epoch": 2.6954545454545453, | |
| "grad_norm": 1.9149094820022583, | |
| "learning_rate": 5.227272727272727e-05, | |
| "loss": 0.7236, | |
| "step": 593 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "grad_norm": 2.803966999053955, | |
| "learning_rate": 5.151515151515152e-05, | |
| "loss": 1.1317, | |
| "step": 594 | |
| }, | |
| { | |
| "epoch": 2.7045454545454546, | |
| "grad_norm": 2.1107089519500732, | |
| "learning_rate": 5.075757575757576e-05, | |
| "loss": 0.9265, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 2.709090909090909, | |
| "grad_norm": 2.037118911743164, | |
| "learning_rate": 5e-05, | |
| "loss": 0.6859, | |
| "step": 596 | |
| }, | |
| { | |
| "epoch": 2.713636363636364, | |
| "grad_norm": 2.310952663421631, | |
| "learning_rate": 4.924242424242424e-05, | |
| "loss": 0.98, | |
| "step": 597 | |
| }, | |
| { | |
| "epoch": 2.7181818181818183, | |
| "grad_norm": 1.9925788640975952, | |
| "learning_rate": 4.848484848484849e-05, | |
| "loss": 0.8919, | |
| "step": 598 | |
| }, | |
| { | |
| "epoch": 2.7227272727272727, | |
| "grad_norm": 2.466705083847046, | |
| "learning_rate": 4.772727272727273e-05, | |
| "loss": 1.3115, | |
| "step": 599 | |
| }, | |
| { | |
| "epoch": 2.7272727272727275, | |
| "grad_norm": 2.8948001861572266, | |
| "learning_rate": 4.6969696969696966e-05, | |
| "loss": 1.4843, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 2.731818181818182, | |
| "grad_norm": 1.8009178638458252, | |
| "learning_rate": 4.621212121212121e-05, | |
| "loss": 0.8387, | |
| "step": 601 | |
| }, | |
| { | |
| "epoch": 2.7363636363636363, | |
| "grad_norm": 1.7695908546447754, | |
| "learning_rate": 4.545454545454546e-05, | |
| "loss": 0.6376, | |
| "step": 602 | |
| }, | |
| { | |
| "epoch": 2.740909090909091, | |
| "grad_norm": 2.255938768386841, | |
| "learning_rate": 4.46969696969697e-05, | |
| "loss": 0.9277, | |
| "step": 603 | |
| }, | |
| { | |
| "epoch": 2.7454545454545456, | |
| "grad_norm": 2.6216013431549072, | |
| "learning_rate": 4.393939393939394e-05, | |
| "loss": 0.8539, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "grad_norm": 2.33111310005188, | |
| "learning_rate": 4.318181818181818e-05, | |
| "loss": 0.766, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 2.7545454545454544, | |
| "grad_norm": 2.256770610809326, | |
| "learning_rate": 4.242424242424243e-05, | |
| "loss": 0.9658, | |
| "step": 606 | |
| }, | |
| { | |
| "epoch": 2.759090909090909, | |
| "grad_norm": 2.4762847423553467, | |
| "learning_rate": 4.1666666666666665e-05, | |
| "loss": 0.8902, | |
| "step": 607 | |
| }, | |
| { | |
| "epoch": 2.7636363636363637, | |
| "grad_norm": 1.8913813829421997, | |
| "learning_rate": 4.090909090909091e-05, | |
| "loss": 0.6938, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 2.768181818181818, | |
| "grad_norm": 1.928743839263916, | |
| "learning_rate": 4.015151515151516e-05, | |
| "loss": 0.637, | |
| "step": 609 | |
| }, | |
| { | |
| "epoch": 2.7727272727272725, | |
| "grad_norm": 3.3095438480377197, | |
| "learning_rate": 3.939393939393939e-05, | |
| "loss": 0.9913, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 2.7772727272727273, | |
| "grad_norm": 2.611701011657715, | |
| "learning_rate": 3.8636363636363636e-05, | |
| "loss": 1.1573, | |
| "step": 611 | |
| }, | |
| { | |
| "epoch": 2.7818181818181817, | |
| "grad_norm": 2.022073745727539, | |
| "learning_rate": 3.787878787878788e-05, | |
| "loss": 0.9537, | |
| "step": 612 | |
| }, | |
| { | |
| "epoch": 2.786363636363636, | |
| "grad_norm": 2.9264447689056396, | |
| "learning_rate": 3.712121212121213e-05, | |
| "loss": 1.4692, | |
| "step": 613 | |
| }, | |
| { | |
| "epoch": 2.790909090909091, | |
| "grad_norm": 2.22469425201416, | |
| "learning_rate": 3.6363636363636364e-05, | |
| "loss": 1.0733, | |
| "step": 614 | |
| }, | |
| { | |
| "epoch": 2.7954545454545454, | |
| "grad_norm": 2.8329367637634277, | |
| "learning_rate": 3.560606060606061e-05, | |
| "loss": 1.5153, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "grad_norm": 1.8949047327041626, | |
| "learning_rate": 3.484848484848485e-05, | |
| "loss": 0.6316, | |
| "step": 616 | |
| }, | |
| { | |
| "epoch": 2.8045454545454547, | |
| "grad_norm": 2.597440242767334, | |
| "learning_rate": 3.409090909090909e-05, | |
| "loss": 1.6049, | |
| "step": 617 | |
| }, | |
| { | |
| "epoch": 2.809090909090909, | |
| "grad_norm": 2.0482330322265625, | |
| "learning_rate": 3.3333333333333335e-05, | |
| "loss": 1.0083, | |
| "step": 618 | |
| }, | |
| { | |
| "epoch": 2.8136363636363635, | |
| "grad_norm": 1.7359944581985474, | |
| "learning_rate": 3.257575757575758e-05, | |
| "loss": 0.9097, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 2.8181818181818183, | |
| "grad_norm": 1.9825539588928223, | |
| "learning_rate": 3.1818181818181814e-05, | |
| "loss": 0.7107, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 2.8227272727272728, | |
| "grad_norm": 1.7564197778701782, | |
| "learning_rate": 3.106060606060606e-05, | |
| "loss": 0.9089, | |
| "step": 621 | |
| }, | |
| { | |
| "epoch": 2.827272727272727, | |
| "grad_norm": 2.735137939453125, | |
| "learning_rate": 3.0303030303030306e-05, | |
| "loss": 1.4978, | |
| "step": 622 | |
| }, | |
| { | |
| "epoch": 2.831818181818182, | |
| "grad_norm": 2.702873706817627, | |
| "learning_rate": 2.9545454545454545e-05, | |
| "loss": 1.2556, | |
| "step": 623 | |
| }, | |
| { | |
| "epoch": 2.8363636363636364, | |
| "grad_norm": 1.9755101203918457, | |
| "learning_rate": 2.878787878787879e-05, | |
| "loss": 0.8022, | |
| "step": 624 | |
| }, | |
| { | |
| "epoch": 2.840909090909091, | |
| "grad_norm": 2.0104050636291504, | |
| "learning_rate": 2.803030303030303e-05, | |
| "loss": 0.8993, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 2.8454545454545457, | |
| "grad_norm": 2.915588855743408, | |
| "learning_rate": 2.7272727272727273e-05, | |
| "loss": 1.7374, | |
| "step": 626 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "grad_norm": 2.5364370346069336, | |
| "learning_rate": 2.6515151515151516e-05, | |
| "loss": 0.7967, | |
| "step": 627 | |
| }, | |
| { | |
| "epoch": 2.8545454545454545, | |
| "grad_norm": 2.731673002243042, | |
| "learning_rate": 2.575757575757576e-05, | |
| "loss": 1.5847, | |
| "step": 628 | |
| }, | |
| { | |
| "epoch": 2.8590909090909093, | |
| "grad_norm": 2.1468403339385986, | |
| "learning_rate": 2.5e-05, | |
| "loss": 1.0101, | |
| "step": 629 | |
| }, | |
| { | |
| "epoch": 2.8636363636363638, | |
| "grad_norm": 1.8162040710449219, | |
| "learning_rate": 2.4242424242424244e-05, | |
| "loss": 0.6688, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.868181818181818, | |
| "grad_norm": 2.285930871963501, | |
| "learning_rate": 2.3484848484848483e-05, | |
| "loss": 1.1906, | |
| "step": 631 | |
| }, | |
| { | |
| "epoch": 2.8727272727272726, | |
| "grad_norm": 2.0448861122131348, | |
| "learning_rate": 2.272727272727273e-05, | |
| "loss": 0.8868, | |
| "step": 632 | |
| }, | |
| { | |
| "epoch": 2.8772727272727274, | |
| "grad_norm": 2.98807692527771, | |
| "learning_rate": 2.196969696969697e-05, | |
| "loss": 1.1679, | |
| "step": 633 | |
| }, | |
| { | |
| "epoch": 2.881818181818182, | |
| "grad_norm": 1.9618700742721558, | |
| "learning_rate": 2.1212121212121215e-05, | |
| "loss": 1.0659, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 2.8863636363636362, | |
| "grad_norm": 2.200741767883301, | |
| "learning_rate": 2.0454545454545454e-05, | |
| "loss": 1.0582, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 2.8909090909090907, | |
| "grad_norm": 2.1953506469726562, | |
| "learning_rate": 1.9696969696969697e-05, | |
| "loss": 1.4356, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 2.8954545454545455, | |
| "grad_norm": 2.1912357807159424, | |
| "learning_rate": 1.893939393939394e-05, | |
| "loss": 1.0532, | |
| "step": 637 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "grad_norm": 2.125601053237915, | |
| "learning_rate": 1.8181818181818182e-05, | |
| "loss": 1.458, | |
| "step": 638 | |
| }, | |
| { | |
| "epoch": 2.9045454545454543, | |
| "grad_norm": 1.9430787563323975, | |
| "learning_rate": 1.7424242424242425e-05, | |
| "loss": 0.8313, | |
| "step": 639 | |
| }, | |
| { | |
| "epoch": 2.909090909090909, | |
| "grad_norm": 1.7255802154541016, | |
| "learning_rate": 1.6666666666666667e-05, | |
| "loss": 0.7894, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.9136363636363636, | |
| "grad_norm": 2.259798526763916, | |
| "learning_rate": 1.5909090909090907e-05, | |
| "loss": 1.0942, | |
| "step": 641 | |
| }, | |
| { | |
| "epoch": 2.918181818181818, | |
| "grad_norm": 2.4443533420562744, | |
| "learning_rate": 1.5151515151515153e-05, | |
| "loss": 1.5392, | |
| "step": 642 | |
| }, | |
| { | |
| "epoch": 2.922727272727273, | |
| "grad_norm": 2.437310218811035, | |
| "learning_rate": 1.4393939393939396e-05, | |
| "loss": 0.9475, | |
| "step": 643 | |
| }, | |
| { | |
| "epoch": 2.9272727272727272, | |
| "grad_norm": 2.1248443126678467, | |
| "learning_rate": 1.3636363636363637e-05, | |
| "loss": 0.6917, | |
| "step": 644 | |
| }, | |
| { | |
| "epoch": 2.9318181818181817, | |
| "grad_norm": 2.0161659717559814, | |
| "learning_rate": 1.287878787878788e-05, | |
| "loss": 1.0465, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 2.9363636363636365, | |
| "grad_norm": 1.825695514678955, | |
| "learning_rate": 1.2121212121212122e-05, | |
| "loss": 0.8001, | |
| "step": 646 | |
| }, | |
| { | |
| "epoch": 2.940909090909091, | |
| "grad_norm": 2.58219575881958, | |
| "learning_rate": 1.1363636363636365e-05, | |
| "loss": 0.9649, | |
| "step": 647 | |
| }, | |
| { | |
| "epoch": 2.9454545454545453, | |
| "grad_norm": 1.9554407596588135, | |
| "learning_rate": 1.0606060606060607e-05, | |
| "loss": 1.1447, | |
| "step": 648 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "grad_norm": 2.89900541305542, | |
| "learning_rate": 9.848484848484848e-06, | |
| "loss": 0.9461, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 2.9545454545454546, | |
| "grad_norm": 1.8475868701934814, | |
| "learning_rate": 9.090909090909091e-06, | |
| "loss": 1.1863, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 2.959090909090909, | |
| "grad_norm": 2.264302968978882, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 0.9459, | |
| "step": 651 | |
| }, | |
| { | |
| "epoch": 2.963636363636364, | |
| "grad_norm": 2.157198190689087, | |
| "learning_rate": 7.5757575757575764e-06, | |
| "loss": 1.4461, | |
| "step": 652 | |
| }, | |
| { | |
| "epoch": 2.9681818181818183, | |
| "grad_norm": 2.3027210235595703, | |
| "learning_rate": 6.818181818181818e-06, | |
| "loss": 0.8407, | |
| "step": 653 | |
| }, | |
| { | |
| "epoch": 2.9727272727272727, | |
| "grad_norm": 1.786800503730774, | |
| "learning_rate": 6.060606060606061e-06, | |
| "loss": 0.7051, | |
| "step": 654 | |
| }, | |
| { | |
| "epoch": 2.9772727272727275, | |
| "grad_norm": 2.4173872470855713, | |
| "learning_rate": 5.303030303030304e-06, | |
| "loss": 1.0276, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 2.981818181818182, | |
| "grad_norm": 3.755701780319214, | |
| "learning_rate": 4.5454545454545455e-06, | |
| "loss": 0.9572, | |
| "step": 656 | |
| }, | |
| { | |
| "epoch": 2.9863636363636363, | |
| "grad_norm": 2.0097804069519043, | |
| "learning_rate": 3.7878787878787882e-06, | |
| "loss": 0.9736, | |
| "step": 657 | |
| }, | |
| { | |
| "epoch": 2.990909090909091, | |
| "grad_norm": 1.773881196975708, | |
| "learning_rate": 3.0303030303030305e-06, | |
| "loss": 0.8404, | |
| "step": 658 | |
| }, | |
| { | |
| "epoch": 2.9954545454545456, | |
| "grad_norm": 2.139065980911255, | |
| "learning_rate": 2.2727272727272728e-06, | |
| "loss": 1.1301, | |
| "step": 659 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 1.8846311569213867, | |
| "learning_rate": 1.5151515151515152e-06, | |
| "loss": 0.7227, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_f1": 0.8952, | |
| "eval_gen_len": 41.8455, | |
| "eval_loss": 1.8697103261947632, | |
| "eval_precision": 0.8934, | |
| "eval_recall": 0.8971, | |
| "eval_rouge1": 0.4709, | |
| "eval_rouge2": 0.2223, | |
| "eval_rougeL": 0.3999, | |
| "eval_rougeLsum": 0.4391, | |
| "eval_runtime": 25.3276, | |
| "eval_samples_per_second": 4.343, | |
| "eval_steps_per_second": 0.553, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "step": 660, | |
| "total_flos": 2484005840363520.0, | |
| "train_loss": 1.4827006761774872, | |
| "train_runtime": 507.3071, | |
| "train_samples_per_second": 5.198, | |
| "train_steps_per_second": 1.301 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 660, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2484005840363520.0, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |