{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 5196, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00019245573518090838, "grad_norm": 12.822089195251465, "learning_rate": 1.282051282051282e-07, "loss": 1.7864, "step": 1 }, { "epoch": 0.00038491147036181676, "grad_norm": 15.502657890319824, "learning_rate": 2.564102564102564e-07, "loss": 1.809, "step": 2 }, { "epoch": 0.0005773672055427252, "grad_norm": 16.625526428222656, "learning_rate": 3.846153846153847e-07, "loss": 1.9262, "step": 3 }, { "epoch": 0.0007698229407236335, "grad_norm": 15.486868858337402, "learning_rate": 5.128205128205128e-07, "loss": 1.8607, "step": 4 }, { "epoch": 0.000962278675904542, "grad_norm": 14.205751419067383, "learning_rate": 6.41025641025641e-07, "loss": 1.832, "step": 5 }, { "epoch": 0.0011547344110854503, "grad_norm": 16.743728637695312, "learning_rate": 7.692307692307694e-07, "loss": 1.7619, "step": 6 }, { "epoch": 0.0013471901462663587, "grad_norm": 15.840396881103516, "learning_rate": 8.974358974358975e-07, "loss": 1.8003, "step": 7 }, { "epoch": 0.001539645881447267, "grad_norm": 17.014772415161133, "learning_rate": 1.0256410256410257e-06, "loss": 1.8521, "step": 8 }, { "epoch": 0.0017321016166281756, "grad_norm": 3.425260066986084, "learning_rate": 1.153846153846154e-06, "loss": 1.1337, "step": 9 }, { "epoch": 0.001924557351809084, "grad_norm": 12.898903846740723, "learning_rate": 1.282051282051282e-06, "loss": 1.7902, "step": 10 }, { "epoch": 0.002117013086989992, "grad_norm": 12.346709251403809, "learning_rate": 1.4102564102564104e-06, "loss": 1.8239, "step": 11 }, { "epoch": 0.0023094688221709007, "grad_norm": 11.96186637878418, "learning_rate": 1.5384615384615387e-06, "loss": 1.7284, "step": 12 }, { "epoch": 0.0025019245573518092, "grad_norm": 10.177848815917969, "learning_rate": 1.6666666666666667e-06, "loss": 1.7008, "step": 13 }, { "epoch": 0.0026943802925327174, "grad_norm": 10.066478729248047, "learning_rate": 1.794871794871795e-06, "loss": 1.7296, "step": 14 }, { "epoch": 0.002886836027713626, "grad_norm": 3.2294039726257324, "learning_rate": 1.9230769230769234e-06, "loss": 1.0516, "step": 15 }, { "epoch": 0.003079291762894534, "grad_norm": 9.396673202514648, "learning_rate": 2.0512820512820513e-06, "loss": 1.7109, "step": 16 }, { "epoch": 0.0032717474980754426, "grad_norm": 8.930295944213867, "learning_rate": 2.1794871794871797e-06, "loss": 1.6944, "step": 17 }, { "epoch": 0.003464203233256351, "grad_norm": 7.112697601318359, "learning_rate": 2.307692307692308e-06, "loss": 1.5955, "step": 18 }, { "epoch": 0.0036566589684372593, "grad_norm": 6.16896915435791, "learning_rate": 2.435897435897436e-06, "loss": 1.5548, "step": 19 }, { "epoch": 0.003849114703618168, "grad_norm": 5.964946746826172, "learning_rate": 2.564102564102564e-06, "loss": 1.5453, "step": 20 }, { "epoch": 0.0040415704387990765, "grad_norm": 5.10248327255249, "learning_rate": 2.6923076923076923e-06, "loss": 1.4643, "step": 21 }, { "epoch": 0.004234026173979984, "grad_norm": 4.836123466491699, "learning_rate": 2.8205128205128207e-06, "loss": 1.4925, "step": 22 }, { "epoch": 0.004426481909160893, "grad_norm": 4.450653076171875, "learning_rate": 2.948717948717949e-06, "loss": 1.4011, "step": 23 }, { "epoch": 0.004618937644341801, "grad_norm": 4.279849529266357, "learning_rate": 3.0769230769230774e-06, "loss": 1.4694, "step": 24 }, { "epoch": 0.00481139337952271, "grad_norm": 3.859330415725708, "learning_rate": 3.205128205128206e-06, "loss": 1.4068, "step": 25 }, { "epoch": 0.0050038491147036184, "grad_norm": 3.5212297439575195, "learning_rate": 3.3333333333333333e-06, "loss": 1.3321, "step": 26 }, { "epoch": 0.005196304849884526, "grad_norm": 3.2917630672454834, "learning_rate": 3.4615384615384617e-06, "loss": 1.3573, "step": 27 }, { "epoch": 0.005388760585065435, "grad_norm": 3.4271128177642822, "learning_rate": 3.58974358974359e-06, "loss": 1.1472, "step": 28 }, { "epoch": 0.005581216320246343, "grad_norm": 3.9574453830718994, "learning_rate": 3.7179487179487184e-06, "loss": 1.4178, "step": 29 }, { "epoch": 0.005773672055427252, "grad_norm": 2.950049877166748, "learning_rate": 3.846153846153847e-06, "loss": 1.2248, "step": 30 }, { "epoch": 0.00596612779060816, "grad_norm": 3.720195770263672, "learning_rate": 3.974358974358974e-06, "loss": 1.3861, "step": 31 }, { "epoch": 0.006158583525789068, "grad_norm": 3.2444863319396973, "learning_rate": 4.102564102564103e-06, "loss": 1.3506, "step": 32 }, { "epoch": 0.006351039260969977, "grad_norm": 3.5759549140930176, "learning_rate": 4.230769230769231e-06, "loss": 1.4098, "step": 33 }, { "epoch": 0.006543494996150885, "grad_norm": 3.14204478263855, "learning_rate": 4.358974358974359e-06, "loss": 1.258, "step": 34 }, { "epoch": 0.006735950731331794, "grad_norm": 2.818664789199829, "learning_rate": 4.487179487179488e-06, "loss": 1.0367, "step": 35 }, { "epoch": 0.006928406466512702, "grad_norm": 3.267183542251587, "learning_rate": 4.615384615384616e-06, "loss": 1.3263, "step": 36 }, { "epoch": 0.00712086220169361, "grad_norm": 2.8578402996063232, "learning_rate": 4.743589743589744e-06, "loss": 1.2931, "step": 37 }, { "epoch": 0.007313317936874519, "grad_norm": 2.805088520050049, "learning_rate": 4.871794871794872e-06, "loss": 1.271, "step": 38 }, { "epoch": 0.007505773672055427, "grad_norm": 3.186495065689087, "learning_rate": 5e-06, "loss": 1.3552, "step": 39 }, { "epoch": 0.007698229407236336, "grad_norm": 2.701211929321289, "learning_rate": 5.128205128205128e-06, "loss": 1.3347, "step": 40 }, { "epoch": 0.007890685142417244, "grad_norm": 2.6992037296295166, "learning_rate": 5.256410256410257e-06, "loss": 1.294, "step": 41 }, { "epoch": 0.008083140877598153, "grad_norm": 3.0867204666137695, "learning_rate": 5.384615384615385e-06, "loss": 1.2933, "step": 42 }, { "epoch": 0.008275596612779062, "grad_norm": 2.5067825317382812, "learning_rate": 5.512820512820514e-06, "loss": 1.2909, "step": 43 }, { "epoch": 0.008468052347959968, "grad_norm": 2.5910398960113525, "learning_rate": 5.641025641025641e-06, "loss": 1.2735, "step": 44 }, { "epoch": 0.008660508083140877, "grad_norm": 2.755016803741455, "learning_rate": 5.769230769230769e-06, "loss": 1.3035, "step": 45 }, { "epoch": 0.008852963818321785, "grad_norm": 2.2789146900177, "learning_rate": 5.897435897435898e-06, "loss": 1.1097, "step": 46 }, { "epoch": 0.009045419553502694, "grad_norm": 2.0922346115112305, "learning_rate": 6.025641025641026e-06, "loss": 1.0117, "step": 47 }, { "epoch": 0.009237875288683603, "grad_norm": 2.4330496788024902, "learning_rate": 6.153846153846155e-06, "loss": 1.2662, "step": 48 }, { "epoch": 0.009430331023864511, "grad_norm": 2.6654067039489746, "learning_rate": 6.282051282051282e-06, "loss": 1.3366, "step": 49 }, { "epoch": 0.00962278675904542, "grad_norm": 2.1606128215789795, "learning_rate": 6.410256410256412e-06, "loss": 1.2292, "step": 50 }, { "epoch": 0.009815242494226328, "grad_norm": 2.382218360900879, "learning_rate": 6.538461538461539e-06, "loss": 1.1865, "step": 51 }, { "epoch": 0.010007698229407237, "grad_norm": 2.0865025520324707, "learning_rate": 6.666666666666667e-06, "loss": 1.2384, "step": 52 }, { "epoch": 0.010200153964588145, "grad_norm": 2.2342097759246826, "learning_rate": 6.794871794871796e-06, "loss": 1.2919, "step": 53 }, { "epoch": 0.010392609699769052, "grad_norm": 2.0874555110931396, "learning_rate": 6.923076923076923e-06, "loss": 1.2766, "step": 54 }, { "epoch": 0.010585065434949961, "grad_norm": 2.268005847930908, "learning_rate": 7.051282051282053e-06, "loss": 1.2323, "step": 55 }, { "epoch": 0.01077752117013087, "grad_norm": 2.5772290229797363, "learning_rate": 7.17948717948718e-06, "loss": 1.1302, "step": 56 }, { "epoch": 0.010969976905311778, "grad_norm": 2.07559871673584, "learning_rate": 7.307692307692308e-06, "loss": 1.2162, "step": 57 }, { "epoch": 0.011162432640492687, "grad_norm": 2.216207265853882, "learning_rate": 7.435897435897437e-06, "loss": 1.1496, "step": 58 }, { "epoch": 0.011354888375673595, "grad_norm": 2.183626890182495, "learning_rate": 7.564102564102564e-06, "loss": 1.2705, "step": 59 }, { "epoch": 0.011547344110854504, "grad_norm": 2.187493085861206, "learning_rate": 7.692307692307694e-06, "loss": 1.22, "step": 60 }, { "epoch": 0.011739799846035412, "grad_norm": 2.0497353076934814, "learning_rate": 7.820512820512822e-06, "loss": 1.2344, "step": 61 }, { "epoch": 0.01193225558121632, "grad_norm": 2.324303388595581, "learning_rate": 7.948717948717949e-06, "loss": 1.1398, "step": 62 }, { "epoch": 0.01212471131639723, "grad_norm": 2.2123873233795166, "learning_rate": 8.076923076923077e-06, "loss": 1.1237, "step": 63 }, { "epoch": 0.012317167051578136, "grad_norm": 2.1354572772979736, "learning_rate": 8.205128205128205e-06, "loss": 1.1925, "step": 64 }, { "epoch": 0.012509622786759045, "grad_norm": 2.054776430130005, "learning_rate": 8.333333333333334e-06, "loss": 1.2531, "step": 65 }, { "epoch": 0.012702078521939953, "grad_norm": 2.3509910106658936, "learning_rate": 8.461538461538462e-06, "loss": 1.1556, "step": 66 }, { "epoch": 0.012894534257120862, "grad_norm": 2.21283221244812, "learning_rate": 8.58974358974359e-06, "loss": 1.242, "step": 67 }, { "epoch": 0.01308698999230177, "grad_norm": 2.0342981815338135, "learning_rate": 8.717948717948719e-06, "loss": 1.2512, "step": 68 }, { "epoch": 0.013279445727482679, "grad_norm": 2.1091084480285645, "learning_rate": 8.846153846153847e-06, "loss": 1.1677, "step": 69 }, { "epoch": 0.013471901462663588, "grad_norm": 2.142625570297241, "learning_rate": 8.974358974358976e-06, "loss": 1.2226, "step": 70 }, { "epoch": 0.013664357197844496, "grad_norm": 1.9051457643508911, "learning_rate": 9.102564102564104e-06, "loss": 1.115, "step": 71 }, { "epoch": 0.013856812933025405, "grad_norm": 2.168208360671997, "learning_rate": 9.230769230769232e-06, "loss": 1.1727, "step": 72 }, { "epoch": 0.014049268668206313, "grad_norm": 2.2827489376068115, "learning_rate": 9.358974358974359e-06, "loss": 1.1535, "step": 73 }, { "epoch": 0.01424172440338722, "grad_norm": 1.9166820049285889, "learning_rate": 9.487179487179487e-06, "loss": 1.2497, "step": 74 }, { "epoch": 0.014434180138568129, "grad_norm": 2.0673937797546387, "learning_rate": 9.615384615384616e-06, "loss": 1.1311, "step": 75 }, { "epoch": 0.014626635873749037, "grad_norm": 2.4237053394317627, "learning_rate": 9.743589743589744e-06, "loss": 1.1514, "step": 76 }, { "epoch": 0.014819091608929946, "grad_norm": 1.9658292531967163, "learning_rate": 9.871794871794872e-06, "loss": 1.2051, "step": 77 }, { "epoch": 0.015011547344110854, "grad_norm": 2.1604244709014893, "learning_rate": 1e-05, "loss": 1.1259, "step": 78 }, { "epoch": 0.015204003079291763, "grad_norm": 1.995413899421692, "learning_rate": 1.012820512820513e-05, "loss": 1.2145, "step": 79 }, { "epoch": 0.015396458814472672, "grad_norm": 2.4904634952545166, "learning_rate": 1.0256410256410256e-05, "loss": 1.1759, "step": 80 }, { "epoch": 0.01558891454965358, "grad_norm": 2.060880184173584, "learning_rate": 1.0384615384615386e-05, "loss": 1.243, "step": 81 }, { "epoch": 0.01578137028483449, "grad_norm": 2.1147937774658203, "learning_rate": 1.0512820512820514e-05, "loss": 1.1531, "step": 82 }, { "epoch": 0.015973826020015396, "grad_norm": 1.6522949934005737, "learning_rate": 1.0641025641025643e-05, "loss": 0.9506, "step": 83 }, { "epoch": 0.016166281755196306, "grad_norm": 2.2048704624176025, "learning_rate": 1.076923076923077e-05, "loss": 1.1701, "step": 84 }, { "epoch": 0.016358737490377213, "grad_norm": 1.9545150995254517, "learning_rate": 1.0897435897435898e-05, "loss": 1.2011, "step": 85 }, { "epoch": 0.016551193225558123, "grad_norm": 2.0530219078063965, "learning_rate": 1.1025641025641028e-05, "loss": 1.1103, "step": 86 }, { "epoch": 0.01674364896073903, "grad_norm": 1.9556024074554443, "learning_rate": 1.1153846153846154e-05, "loss": 1.0949, "step": 87 }, { "epoch": 0.016936104695919937, "grad_norm": 2.006627082824707, "learning_rate": 1.1282051282051283e-05, "loss": 1.2281, "step": 88 }, { "epoch": 0.017128560431100847, "grad_norm": 2.2797443866729736, "learning_rate": 1.1410256410256411e-05, "loss": 1.1694, "step": 89 }, { "epoch": 0.017321016166281754, "grad_norm": 1.9811910390853882, "learning_rate": 1.1538461538461538e-05, "loss": 1.1155, "step": 90 }, { "epoch": 0.017513471901462664, "grad_norm": 2.027745485305786, "learning_rate": 1.1666666666666668e-05, "loss": 1.1213, "step": 91 }, { "epoch": 0.01770592763664357, "grad_norm": 1.895713210105896, "learning_rate": 1.1794871794871796e-05, "loss": 1.1145, "step": 92 }, { "epoch": 0.01789838337182448, "grad_norm": 2.2578670978546143, "learning_rate": 1.1923076923076925e-05, "loss": 1.1143, "step": 93 }, { "epoch": 0.018090839107005388, "grad_norm": 1.4406471252441406, "learning_rate": 1.2051282051282051e-05, "loss": 0.9945, "step": 94 }, { "epoch": 0.0182832948421863, "grad_norm": 2.3857421875, "learning_rate": 1.217948717948718e-05, "loss": 1.1937, "step": 95 }, { "epoch": 0.018475750577367205, "grad_norm": 1.3613249063491821, "learning_rate": 1.230769230769231e-05, "loss": 0.9247, "step": 96 }, { "epoch": 0.018668206312548116, "grad_norm": 2.0658631324768066, "learning_rate": 1.2435897435897436e-05, "loss": 1.1109, "step": 97 }, { "epoch": 0.018860662047729022, "grad_norm": 2.048159122467041, "learning_rate": 1.2564102564102565e-05, "loss": 1.1868, "step": 98 }, { "epoch": 0.01905311778290993, "grad_norm": 1.9883164167404175, "learning_rate": 1.2692307692307693e-05, "loss": 1.2179, "step": 99 }, { "epoch": 0.01924557351809084, "grad_norm": 1.887096881866455, "learning_rate": 1.2820512820512823e-05, "loss": 1.1249, "step": 100 }, { "epoch": 0.019438029253271746, "grad_norm": 1.8913002014160156, "learning_rate": 1.294871794871795e-05, "loss": 1.1314, "step": 101 }, { "epoch": 0.019630484988452657, "grad_norm": 1.9547669887542725, "learning_rate": 1.3076923076923078e-05, "loss": 1.1349, "step": 102 }, { "epoch": 0.019822940723633563, "grad_norm": 1.970658540725708, "learning_rate": 1.3205128205128207e-05, "loss": 1.1626, "step": 103 }, { "epoch": 0.020015396458814474, "grad_norm": 2.0676755905151367, "learning_rate": 1.3333333333333333e-05, "loss": 1.1424, "step": 104 }, { "epoch": 0.02020785219399538, "grad_norm": 1.3007159233093262, "learning_rate": 1.3461538461538463e-05, "loss": 1.005, "step": 105 }, { "epoch": 0.02040030792917629, "grad_norm": 2.0430386066436768, "learning_rate": 1.3589743589743592e-05, "loss": 1.2203, "step": 106 }, { "epoch": 0.020592763664357198, "grad_norm": 2.1048197746276855, "learning_rate": 1.3717948717948718e-05, "loss": 1.1334, "step": 107 }, { "epoch": 0.020785219399538105, "grad_norm": 2.023083448410034, "learning_rate": 1.3846153846153847e-05, "loss": 1.1844, "step": 108 }, { "epoch": 0.020977675134719015, "grad_norm": 2.017812490463257, "learning_rate": 1.3974358974358975e-05, "loss": 1.1929, "step": 109 }, { "epoch": 0.021170130869899922, "grad_norm": 1.9722820520401, "learning_rate": 1.4102564102564105e-05, "loss": 1.146, "step": 110 }, { "epoch": 0.021362586605080832, "grad_norm": 2.0664279460906982, "learning_rate": 1.4230769230769232e-05, "loss": 1.0048, "step": 111 }, { "epoch": 0.02155504234026174, "grad_norm": 1.2306312322616577, "learning_rate": 1.435897435897436e-05, "loss": 0.9728, "step": 112 }, { "epoch": 0.02174749807544265, "grad_norm": 2.1424176692962646, "learning_rate": 1.4487179487179489e-05, "loss": 1.1453, "step": 113 }, { "epoch": 0.021939953810623556, "grad_norm": 2.203076124191284, "learning_rate": 1.4615384615384615e-05, "loss": 1.2156, "step": 114 }, { "epoch": 0.022132409545804466, "grad_norm": 2.018467903137207, "learning_rate": 1.4743589743589745e-05, "loss": 1.1461, "step": 115 }, { "epoch": 0.022324865280985373, "grad_norm": 1.85000741481781, "learning_rate": 1.4871794871794874e-05, "loss": 1.1281, "step": 116 }, { "epoch": 0.022517321016166283, "grad_norm": 2.1398367881774902, "learning_rate": 1.5000000000000002e-05, "loss": 1.1198, "step": 117 }, { "epoch": 0.02270977675134719, "grad_norm": 2.32318115234375, "learning_rate": 1.5128205128205129e-05, "loss": 1.114, "step": 118 }, { "epoch": 0.022902232486528097, "grad_norm": 2.074957847595215, "learning_rate": 1.5256410256410257e-05, "loss": 1.1538, "step": 119 }, { "epoch": 0.023094688221709007, "grad_norm": 2.0280978679656982, "learning_rate": 1.5384615384615387e-05, "loss": 1.1095, "step": 120 }, { "epoch": 0.023287143956889914, "grad_norm": 2.0076022148132324, "learning_rate": 1.5512820512820516e-05, "loss": 1.0825, "step": 121 }, { "epoch": 0.023479599692070825, "grad_norm": 2.159745454788208, "learning_rate": 1.5641025641025644e-05, "loss": 1.2233, "step": 122 }, { "epoch": 0.02367205542725173, "grad_norm": 1.1482303142547607, "learning_rate": 1.576923076923077e-05, "loss": 0.9763, "step": 123 }, { "epoch": 0.02386451116243264, "grad_norm": 2.2342629432678223, "learning_rate": 1.5897435897435897e-05, "loss": 1.1917, "step": 124 }, { "epoch": 0.02405696689761355, "grad_norm": 2.173105478286743, "learning_rate": 1.602564102564103e-05, "loss": 1.1897, "step": 125 }, { "epoch": 0.02424942263279446, "grad_norm": 1.921324372291565, "learning_rate": 1.6153846153846154e-05, "loss": 1.2244, "step": 126 }, { "epoch": 0.024441878367975366, "grad_norm": 1.0997803211212158, "learning_rate": 1.6282051282051282e-05, "loss": 0.9404, "step": 127 }, { "epoch": 0.024634334103156273, "grad_norm": 1.1733825206756592, "learning_rate": 1.641025641025641e-05, "loss": 0.9621, "step": 128 }, { "epoch": 0.024826789838337183, "grad_norm": 2.0403177738189697, "learning_rate": 1.653846153846154e-05, "loss": 1.1321, "step": 129 }, { "epoch": 0.02501924557351809, "grad_norm": 2.2034287452697754, "learning_rate": 1.6666666666666667e-05, "loss": 1.1288, "step": 130 }, { "epoch": 0.025211701308699, "grad_norm": 1.9954860210418701, "learning_rate": 1.6794871794871796e-05, "loss": 1.1416, "step": 131 }, { "epoch": 0.025404157043879907, "grad_norm": 2.0325286388397217, "learning_rate": 1.6923076923076924e-05, "loss": 1.1004, "step": 132 }, { "epoch": 0.025596612779060817, "grad_norm": 2.074540853500366, "learning_rate": 1.7051282051282053e-05, "loss": 1.136, "step": 133 }, { "epoch": 0.025789068514241724, "grad_norm": 2.2531938552856445, "learning_rate": 1.717948717948718e-05, "loss": 1.1463, "step": 134 }, { "epoch": 0.025981524249422634, "grad_norm": 2.1192843914031982, "learning_rate": 1.730769230769231e-05, "loss": 1.2, "step": 135 }, { "epoch": 0.02617397998460354, "grad_norm": 1.9313995838165283, "learning_rate": 1.7435897435897438e-05, "loss": 1.1567, "step": 136 }, { "epoch": 0.026366435719784448, "grad_norm": 2.152517080307007, "learning_rate": 1.7564102564102566e-05, "loss": 1.1183, "step": 137 }, { "epoch": 0.026558891454965358, "grad_norm": 2.020488977432251, "learning_rate": 1.7692307692307694e-05, "loss": 1.1914, "step": 138 }, { "epoch": 0.026751347190146265, "grad_norm": 2.1419143676757812, "learning_rate": 1.7820512820512823e-05, "loss": 1.1844, "step": 139 }, { "epoch": 0.026943802925327175, "grad_norm": 2.026850938796997, "learning_rate": 1.794871794871795e-05, "loss": 1.1305, "step": 140 }, { "epoch": 0.027136258660508082, "grad_norm": 1.9881304502487183, "learning_rate": 1.807692307692308e-05, "loss": 1.1148, "step": 141 }, { "epoch": 0.027328714395688992, "grad_norm": 1.998692274093628, "learning_rate": 1.8205128205128208e-05, "loss": 1.094, "step": 142 }, { "epoch": 0.0275211701308699, "grad_norm": 2.50514817237854, "learning_rate": 1.8333333333333333e-05, "loss": 1.1346, "step": 143 }, { "epoch": 0.02771362586605081, "grad_norm": 2.39231538772583, "learning_rate": 1.8461538461538465e-05, "loss": 1.1131, "step": 144 }, { "epoch": 0.027906081601231716, "grad_norm": 1.901489019393921, "learning_rate": 1.8589743589743593e-05, "loss": 1.1793, "step": 145 }, { "epoch": 0.028098537336412627, "grad_norm": 2.197298049926758, "learning_rate": 1.8717948717948718e-05, "loss": 1.1929, "step": 146 }, { "epoch": 0.028290993071593534, "grad_norm": 2.0326902866363525, "learning_rate": 1.8846153846153846e-05, "loss": 1.1342, "step": 147 }, { "epoch": 0.02848344880677444, "grad_norm": 2.216961622238159, "learning_rate": 1.8974358974358975e-05, "loss": 1.0769, "step": 148 }, { "epoch": 0.02867590454195535, "grad_norm": 1.140304684638977, "learning_rate": 1.9102564102564106e-05, "loss": 0.925, "step": 149 }, { "epoch": 0.028868360277136258, "grad_norm": 2.226663827896118, "learning_rate": 1.923076923076923e-05, "loss": 1.1478, "step": 150 }, { "epoch": 0.029060816012317168, "grad_norm": 2.1292662620544434, "learning_rate": 1.935897435897436e-05, "loss": 1.1396, "step": 151 }, { "epoch": 0.029253271747498075, "grad_norm": 2.017021656036377, "learning_rate": 1.9487179487179488e-05, "loss": 1.1398, "step": 152 }, { "epoch": 0.029445727482678985, "grad_norm": 2.2312324047088623, "learning_rate": 1.9615384615384617e-05, "loss": 1.1515, "step": 153 }, { "epoch": 0.029638183217859892, "grad_norm": 1.9998685121536255, "learning_rate": 1.9743589743589745e-05, "loss": 1.1344, "step": 154 }, { "epoch": 0.029830638953040802, "grad_norm": 1.1204949617385864, "learning_rate": 1.9871794871794873e-05, "loss": 0.9517, "step": 155 }, { "epoch": 0.03002309468822171, "grad_norm": 1.111180305480957, "learning_rate": 2e-05, "loss": 0.8972, "step": 156 }, { "epoch": 0.030215550423402616, "grad_norm": 2.190965414047241, "learning_rate": 1.9999998057286928e-05, "loss": 1.1266, "step": 157 }, { "epoch": 0.030408006158583526, "grad_norm": 2.132208824157715, "learning_rate": 1.9999992229148463e-05, "loss": 1.1165, "step": 158 }, { "epoch": 0.030600461893764433, "grad_norm": 2.4220693111419678, "learning_rate": 1.999998251558687e-05, "loss": 1.0923, "step": 159 }, { "epoch": 0.030792917628945343, "grad_norm": 2.010136127471924, "learning_rate": 1.999996891660592e-05, "loss": 1.1371, "step": 160 }, { "epoch": 0.03098537336412625, "grad_norm": 2.1428442001342773, "learning_rate": 1.9999951432210905e-05, "loss": 1.0209, "step": 161 }, { "epoch": 0.03117782909930716, "grad_norm": 2.006835460662842, "learning_rate": 1.9999930062408612e-05, "loss": 1.08, "step": 162 }, { "epoch": 0.03137028483448807, "grad_norm": 2.074542760848999, "learning_rate": 1.9999904807207348e-05, "loss": 1.072, "step": 163 }, { "epoch": 0.03156274056966898, "grad_norm": 1.9647022485733032, "learning_rate": 1.9999875666616918e-05, "loss": 1.0841, "step": 164 }, { "epoch": 0.03175519630484989, "grad_norm": 2.094486713409424, "learning_rate": 1.9999842640648656e-05, "loss": 1.1265, "step": 165 }, { "epoch": 0.03194765204003079, "grad_norm": 2.219172477722168, "learning_rate": 1.9999805729315383e-05, "loss": 1.1326, "step": 166 }, { "epoch": 0.0321401077752117, "grad_norm": 1.8253027200698853, "learning_rate": 1.999976493263145e-05, "loss": 1.0738, "step": 167 }, { "epoch": 0.03233256351039261, "grad_norm": 2.1188716888427734, "learning_rate": 1.99997202506127e-05, "loss": 1.0201, "step": 168 }, { "epoch": 0.032525019245573515, "grad_norm": 1.8973318338394165, "learning_rate": 1.99996716832765e-05, "loss": 1.1132, "step": 169 }, { "epoch": 0.032717474980754425, "grad_norm": 2.0979061126708984, "learning_rate": 1.9999619230641714e-05, "loss": 1.0555, "step": 170 }, { "epoch": 0.032909930715935336, "grad_norm": 2.499190330505371, "learning_rate": 1.999956289272873e-05, "loss": 1.0812, "step": 171 }, { "epoch": 0.033102386451116246, "grad_norm": 2.0691018104553223, "learning_rate": 1.9999502669559432e-05, "loss": 1.0962, "step": 172 }, { "epoch": 0.03329484218629715, "grad_norm": 2.364868640899658, "learning_rate": 1.999943856115722e-05, "loss": 1.1312, "step": 173 }, { "epoch": 0.03348729792147806, "grad_norm": 1.8841214179992676, "learning_rate": 1.9999370567547008e-05, "loss": 1.1411, "step": 174 }, { "epoch": 0.03367975365665897, "grad_norm": 2.040099859237671, "learning_rate": 1.999929868875521e-05, "loss": 1.1903, "step": 175 }, { "epoch": 0.03387220939183987, "grad_norm": 2.3591747283935547, "learning_rate": 1.999922292480975e-05, "loss": 1.0833, "step": 176 }, { "epoch": 0.034064665127020784, "grad_norm": 1.877581000328064, "learning_rate": 1.999914327574007e-05, "loss": 1.1074, "step": 177 }, { "epoch": 0.034257120862201694, "grad_norm": 1.9595013856887817, "learning_rate": 1.999905974157712e-05, "loss": 1.1314, "step": 178 }, { "epoch": 0.034449576597382604, "grad_norm": 2.207402229309082, "learning_rate": 1.9998972322353353e-05, "loss": 1.1579, "step": 179 }, { "epoch": 0.03464203233256351, "grad_norm": 2.2019686698913574, "learning_rate": 1.9998881018102735e-05, "loss": 0.9843, "step": 180 }, { "epoch": 0.03483448806774442, "grad_norm": 2.210948944091797, "learning_rate": 1.9998785828860744e-05, "loss": 1.1583, "step": 181 }, { "epoch": 0.03502694380292533, "grad_norm": 1.94996976852417, "learning_rate": 1.999868675466436e-05, "loss": 1.031, "step": 182 }, { "epoch": 0.03521939953810624, "grad_norm": 2.045884370803833, "learning_rate": 1.9998583795552086e-05, "loss": 1.1142, "step": 183 }, { "epoch": 0.03541185527328714, "grad_norm": 2.313939094543457, "learning_rate": 1.9998476951563914e-05, "loss": 1.1559, "step": 184 }, { "epoch": 0.03560431100846805, "grad_norm": 1.1143927574157715, "learning_rate": 1.999836622274137e-05, "loss": 0.9635, "step": 185 }, { "epoch": 0.03579676674364896, "grad_norm": 1.903009295463562, "learning_rate": 1.9998251609127465e-05, "loss": 1.2159, "step": 186 }, { "epoch": 0.035989222478829866, "grad_norm": 1.9762166738510132, "learning_rate": 1.9998133110766744e-05, "loss": 1.0963, "step": 187 }, { "epoch": 0.036181678214010776, "grad_norm": 1.9005874395370483, "learning_rate": 1.9998010727705237e-05, "loss": 1.1517, "step": 188 }, { "epoch": 0.036374133949191687, "grad_norm": 2.1902811527252197, "learning_rate": 1.9997884459990502e-05, "loss": 1.1811, "step": 189 }, { "epoch": 0.0365665896843726, "grad_norm": 2.223172903060913, "learning_rate": 1.9997754307671597e-05, "loss": 1.0531, "step": 190 }, { "epoch": 0.0367590454195535, "grad_norm": 1.78392493724823, "learning_rate": 1.999762027079909e-05, "loss": 1.0264, "step": 191 }, { "epoch": 0.03695150115473441, "grad_norm": 2.302363157272339, "learning_rate": 1.999748234942507e-05, "loss": 1.1008, "step": 192 }, { "epoch": 0.03714395688991532, "grad_norm": 2.0841126441955566, "learning_rate": 1.999734054360311e-05, "loss": 1.0804, "step": 193 }, { "epoch": 0.03733641262509623, "grad_norm": 2.021899938583374, "learning_rate": 1.9997194853388316e-05, "loss": 1.1621, "step": 194 }, { "epoch": 0.037528868360277134, "grad_norm": 1.9690407514572144, "learning_rate": 1.9997045278837298e-05, "loss": 1.1416, "step": 195 }, { "epoch": 0.037721324095458045, "grad_norm": 1.9700400829315186, "learning_rate": 1.9996891820008165e-05, "loss": 1.1493, "step": 196 }, { "epoch": 0.037913779830638955, "grad_norm": 2.160266637802124, "learning_rate": 1.9996734476960547e-05, "loss": 1.0752, "step": 197 }, { "epoch": 0.03810623556581986, "grad_norm": 2.356799840927124, "learning_rate": 1.9996573249755573e-05, "loss": 1.0964, "step": 198 }, { "epoch": 0.03829869130100077, "grad_norm": 2.0327095985412598, "learning_rate": 1.9996408138455896e-05, "loss": 1.1226, "step": 199 }, { "epoch": 0.03849114703618168, "grad_norm": 3.0659170150756836, "learning_rate": 1.999623914312566e-05, "loss": 1.0983, "step": 200 }, { "epoch": 0.03868360277136259, "grad_norm": 2.049588441848755, "learning_rate": 1.9996066263830533e-05, "loss": 1.0875, "step": 201 }, { "epoch": 0.03887605850654349, "grad_norm": 2.2685041427612305, "learning_rate": 1.999588950063768e-05, "loss": 1.1423, "step": 202 }, { "epoch": 0.0390685142417244, "grad_norm": 2.1024460792541504, "learning_rate": 1.9995708853615783e-05, "loss": 1.1237, "step": 203 }, { "epoch": 0.03926096997690531, "grad_norm": 1.8129985332489014, "learning_rate": 1.9995524322835035e-05, "loss": 1.1537, "step": 204 }, { "epoch": 0.039453425712086224, "grad_norm": 2.0320887565612793, "learning_rate": 1.9995335908367132e-05, "loss": 1.111, "step": 205 }, { "epoch": 0.03964588144726713, "grad_norm": 1.8262051343917847, "learning_rate": 1.9995143610285275e-05, "loss": 1.0067, "step": 206 }, { "epoch": 0.03983833718244804, "grad_norm": 2.10689115524292, "learning_rate": 1.9994947428664194e-05, "loss": 1.1224, "step": 207 }, { "epoch": 0.04003079291762895, "grad_norm": 2.1416141986846924, "learning_rate": 1.99947473635801e-05, "loss": 1.1291, "step": 208 }, { "epoch": 0.04022324865280985, "grad_norm": 2.124711275100708, "learning_rate": 1.9994543415110733e-05, "loss": 1.0204, "step": 209 }, { "epoch": 0.04041570438799076, "grad_norm": 2.1166152954101562, "learning_rate": 1.9994335583335336e-05, "loss": 1.1063, "step": 210 }, { "epoch": 0.04060816012317167, "grad_norm": 1.729125738143921, "learning_rate": 1.9994123868334655e-05, "loss": 1.061, "step": 211 }, { "epoch": 0.04080061585835258, "grad_norm": 1.9856681823730469, "learning_rate": 1.999390827019096e-05, "loss": 1.1302, "step": 212 }, { "epoch": 0.040993071593533485, "grad_norm": 2.037698268890381, "learning_rate": 1.999368878898801e-05, "loss": 1.1118, "step": 213 }, { "epoch": 0.041185527328714396, "grad_norm": 2.082167148590088, "learning_rate": 1.999346542481109e-05, "loss": 1.2069, "step": 214 }, { "epoch": 0.041377983063895306, "grad_norm": 2.0513463020324707, "learning_rate": 1.999323817774698e-05, "loss": 1.0668, "step": 215 }, { "epoch": 0.04157043879907621, "grad_norm": 1.8028476238250732, "learning_rate": 1.9993007047883988e-05, "loss": 1.0668, "step": 216 }, { "epoch": 0.04176289453425712, "grad_norm": 2.1010804176330566, "learning_rate": 1.9992772035311902e-05, "loss": 1.0504, "step": 217 }, { "epoch": 0.04195535026943803, "grad_norm": 1.9072201251983643, "learning_rate": 1.9992533140122044e-05, "loss": 1.0689, "step": 218 }, { "epoch": 0.04214780600461894, "grad_norm": 2.158442974090576, "learning_rate": 1.9992290362407232e-05, "loss": 1.074, "step": 219 }, { "epoch": 0.042340261739799843, "grad_norm": 1.8606246709823608, "learning_rate": 1.9992043702261795e-05, "loss": 1.1394, "step": 220 }, { "epoch": 0.042532717474980754, "grad_norm": 2.1072592735290527, "learning_rate": 1.999179315978157e-05, "loss": 1.0943, "step": 221 }, { "epoch": 0.042725173210161664, "grad_norm": 2.0352225303649902, "learning_rate": 1.9991538735063906e-05, "loss": 1.0793, "step": 222 }, { "epoch": 0.042917628945342574, "grad_norm": 2.1362555027008057, "learning_rate": 1.999128042820766e-05, "loss": 1.1638, "step": 223 }, { "epoch": 0.04311008468052348, "grad_norm": 2.0524795055389404, "learning_rate": 1.999101823931319e-05, "loss": 1.1725, "step": 224 }, { "epoch": 0.04330254041570439, "grad_norm": 1.973961591720581, "learning_rate": 1.999075216848237e-05, "loss": 1.0912, "step": 225 }, { "epoch": 0.0434949961508853, "grad_norm": 2.0757157802581787, "learning_rate": 1.999048221581858e-05, "loss": 1.2127, "step": 226 }, { "epoch": 0.0436874518860662, "grad_norm": 2.2070868015289307, "learning_rate": 1.9990208381426708e-05, "loss": 1.0485, "step": 227 }, { "epoch": 0.04387990762124711, "grad_norm": 2.063763380050659, "learning_rate": 1.9989930665413148e-05, "loss": 1.0384, "step": 228 }, { "epoch": 0.04407236335642802, "grad_norm": 2.2438125610351562, "learning_rate": 1.9989649067885808e-05, "loss": 1.1655, "step": 229 }, { "epoch": 0.04426481909160893, "grad_norm": 2.054109811782837, "learning_rate": 1.99893635889541e-05, "loss": 1.0873, "step": 230 }, { "epoch": 0.044457274826789836, "grad_norm": 2.274214029312134, "learning_rate": 1.9989074228728942e-05, "loss": 1.0926, "step": 231 }, { "epoch": 0.044649730561970746, "grad_norm": 2.40297532081604, "learning_rate": 1.998878098732277e-05, "loss": 1.1451, "step": 232 }, { "epoch": 0.04484218629715166, "grad_norm": 1.9875500202178955, "learning_rate": 1.998848386484951e-05, "loss": 1.1457, "step": 233 }, { "epoch": 0.04503464203233257, "grad_norm": 1.7313205003738403, "learning_rate": 1.9988182861424613e-05, "loss": 1.1048, "step": 234 }, { "epoch": 0.04522709776751347, "grad_norm": 1.8364102840423584, "learning_rate": 1.998787797716503e-05, "loss": 1.1879, "step": 235 }, { "epoch": 0.04541955350269438, "grad_norm": 2.029669761657715, "learning_rate": 1.9987569212189224e-05, "loss": 1.0332, "step": 236 }, { "epoch": 0.04561200923787529, "grad_norm": 2.001997470855713, "learning_rate": 1.9987256566617165e-05, "loss": 1.1231, "step": 237 }, { "epoch": 0.045804464973056194, "grad_norm": 2.0194694995880127, "learning_rate": 1.9986940040570323e-05, "loss": 1.1591, "step": 238 }, { "epoch": 0.045996920708237105, "grad_norm": 2.448617935180664, "learning_rate": 1.9986619634171682e-05, "loss": 1.0383, "step": 239 }, { "epoch": 0.046189376443418015, "grad_norm": 1.8933806419372559, "learning_rate": 1.9986295347545738e-05, "loss": 1.0934, "step": 240 }, { "epoch": 0.046381832178598925, "grad_norm": 1.1100465059280396, "learning_rate": 1.9985967180818493e-05, "loss": 0.9531, "step": 241 }, { "epoch": 0.04657428791377983, "grad_norm": 2.1387996673583984, "learning_rate": 1.9985635134117445e-05, "loss": 1.0524, "step": 242 }, { "epoch": 0.04676674364896074, "grad_norm": 2.0161595344543457, "learning_rate": 1.9985299207571614e-05, "loss": 1.0888, "step": 243 }, { "epoch": 0.04695919938414165, "grad_norm": 1.9262018203735352, "learning_rate": 1.998495940131152e-05, "loss": 1.1353, "step": 244 }, { "epoch": 0.04715165511932255, "grad_norm": 2.0804708003997803, "learning_rate": 1.998461571546919e-05, "loss": 1.0808, "step": 245 }, { "epoch": 0.04734411085450346, "grad_norm": 1.9080052375793457, "learning_rate": 1.998426815017817e-05, "loss": 1.1381, "step": 246 }, { "epoch": 0.04753656658968437, "grad_norm": 1.9068005084991455, "learning_rate": 1.998391670557349e-05, "loss": 1.1836, "step": 247 }, { "epoch": 0.04772902232486528, "grad_norm": 1.915165662765503, "learning_rate": 1.998356138179171e-05, "loss": 1.0609, "step": 248 }, { "epoch": 0.04792147806004619, "grad_norm": 2.2248706817626953, "learning_rate": 1.998320217897089e-05, "loss": 1.1005, "step": 249 }, { "epoch": 0.0481139337952271, "grad_norm": 1.8674229383468628, "learning_rate": 1.9982839097250588e-05, "loss": 1.0812, "step": 250 }, { "epoch": 0.04830638953040801, "grad_norm": 1.7297526597976685, "learning_rate": 1.998247213677188e-05, "loss": 0.9959, "step": 251 }, { "epoch": 0.04849884526558892, "grad_norm": 1.0943597555160522, "learning_rate": 1.998210129767735e-05, "loss": 0.9548, "step": 252 }, { "epoch": 0.04869130100076982, "grad_norm": 2.0170698165893555, "learning_rate": 1.9981726580111087e-05, "loss": 1.0855, "step": 253 }, { "epoch": 0.04888375673595073, "grad_norm": 2.2094550132751465, "learning_rate": 1.998134798421867e-05, "loss": 1.0936, "step": 254 }, { "epoch": 0.04907621247113164, "grad_norm": 1.9037797451019287, "learning_rate": 1.9980965510147214e-05, "loss": 1.0249, "step": 255 }, { "epoch": 0.049268668206312545, "grad_norm": 1.8956867456436157, "learning_rate": 1.9980579158045322e-05, "loss": 1.1117, "step": 256 }, { "epoch": 0.049461123941493455, "grad_norm": 1.036456823348999, "learning_rate": 1.9980188928063103e-05, "loss": 1.0112, "step": 257 }, { "epoch": 0.049653579676674366, "grad_norm": 1.9397242069244385, "learning_rate": 1.9979794820352187e-05, "loss": 1.1047, "step": 258 }, { "epoch": 0.049846035411855276, "grad_norm": 2.1860578060150146, "learning_rate": 1.9979396835065696e-05, "loss": 1.0315, "step": 259 }, { "epoch": 0.05003849114703618, "grad_norm": 1.8199071884155273, "learning_rate": 1.9978994972358265e-05, "loss": 1.1123, "step": 260 }, { "epoch": 0.05023094688221709, "grad_norm": 2.0878307819366455, "learning_rate": 1.9978589232386036e-05, "loss": 1.1737, "step": 261 }, { "epoch": 0.050423402617398, "grad_norm": 2.149130344390869, "learning_rate": 1.9978179615306658e-05, "loss": 1.1193, "step": 262 }, { "epoch": 0.05061585835257891, "grad_norm": 1.7451056241989136, "learning_rate": 1.997776612127928e-05, "loss": 1.1536, "step": 263 }, { "epoch": 0.050808314087759814, "grad_norm": 2.0781421661376953, "learning_rate": 1.997734875046456e-05, "loss": 1.106, "step": 264 }, { "epoch": 0.051000769822940724, "grad_norm": 1.9924591779708862, "learning_rate": 1.9976927503024677e-05, "loss": 1.0439, "step": 265 }, { "epoch": 0.051193225558121634, "grad_norm": 1.8670432567596436, "learning_rate": 1.997650237912329e-05, "loss": 1.0725, "step": 266 }, { "epoch": 0.05138568129330254, "grad_norm": 1.8259646892547607, "learning_rate": 1.9976073378925586e-05, "loss": 1.0405, "step": 267 }, { "epoch": 0.05157813702848345, "grad_norm": 1.9916901588439941, "learning_rate": 1.9975640502598243e-05, "loss": 1.0982, "step": 268 }, { "epoch": 0.05177059276366436, "grad_norm": 1.8680857419967651, "learning_rate": 1.997520375030946e-05, "loss": 1.1468, "step": 269 }, { "epoch": 0.05196304849884527, "grad_norm": 2.3491151332855225, "learning_rate": 1.9974763122228927e-05, "loss": 1.134, "step": 270 }, { "epoch": 0.05215550423402617, "grad_norm": 2.0595386028289795, "learning_rate": 1.997431861852785e-05, "loss": 1.1176, "step": 271 }, { "epoch": 0.05234795996920708, "grad_norm": 2.0007033348083496, "learning_rate": 1.9973870239378938e-05, "loss": 1.1129, "step": 272 }, { "epoch": 0.05254041570438799, "grad_norm": 1.9455167055130005, "learning_rate": 1.9973417984956404e-05, "loss": 1.1284, "step": 273 }, { "epoch": 0.052732871439568896, "grad_norm": 1.8305734395980835, "learning_rate": 1.997296185543597e-05, "loss": 1.1369, "step": 274 }, { "epoch": 0.052925327174749806, "grad_norm": 1.0071614980697632, "learning_rate": 1.9972501850994857e-05, "loss": 0.9114, "step": 275 }, { "epoch": 0.053117782909930716, "grad_norm": 1.9528825283050537, "learning_rate": 1.9972037971811802e-05, "loss": 1.1351, "step": 276 }, { "epoch": 0.05331023864511163, "grad_norm": 1.0603333711624146, "learning_rate": 1.997157021806704e-05, "loss": 0.9553, "step": 277 }, { "epoch": 0.05350269438029253, "grad_norm": 2.001891851425171, "learning_rate": 1.9971098589942314e-05, "loss": 1.0726, "step": 278 }, { "epoch": 0.05369515011547344, "grad_norm": 1.7532188892364502, "learning_rate": 1.9970623087620867e-05, "loss": 1.1241, "step": 279 }, { "epoch": 0.05388760585065435, "grad_norm": 1.9638525247573853, "learning_rate": 1.997014371128746e-05, "loss": 1.068, "step": 280 }, { "epoch": 0.05408006158583526, "grad_norm": 1.8798953294754028, "learning_rate": 1.996966046112834e-05, "loss": 1.0596, "step": 281 }, { "epoch": 0.054272517321016164, "grad_norm": 1.8201245069503784, "learning_rate": 1.9969173337331283e-05, "loss": 1.0751, "step": 282 }, { "epoch": 0.054464973056197075, "grad_norm": 1.845236897468567, "learning_rate": 1.9968682340085546e-05, "loss": 1.107, "step": 283 }, { "epoch": 0.054657428791377985, "grad_norm": 2.2483530044555664, "learning_rate": 1.996818746958191e-05, "loss": 1.0952, "step": 284 }, { "epoch": 0.05484988452655889, "grad_norm": 2.0356900691986084, "learning_rate": 1.996768872601265e-05, "loss": 1.066, "step": 285 }, { "epoch": 0.0550423402617398, "grad_norm": 1.838862657546997, "learning_rate": 1.996718610957155e-05, "loss": 1.1225, "step": 286 }, { "epoch": 0.05523479599692071, "grad_norm": 1.8150132894515991, "learning_rate": 1.9966679620453903e-05, "loss": 1.066, "step": 287 }, { "epoch": 0.05542725173210162, "grad_norm": 1.8981868028640747, "learning_rate": 1.9966169258856488e-05, "loss": 1.1009, "step": 288 }, { "epoch": 0.05561970746728252, "grad_norm": 1.955868124961853, "learning_rate": 1.9965655024977617e-05, "loss": 1.099, "step": 289 }, { "epoch": 0.05581216320246343, "grad_norm": 1.846192717552185, "learning_rate": 1.996513691901708e-05, "loss": 1.0287, "step": 290 }, { "epoch": 0.05600461893764434, "grad_norm": 1.7783055305480957, "learning_rate": 1.9964614941176194e-05, "loss": 0.9961, "step": 291 }, { "epoch": 0.056197074672825253, "grad_norm": 1.9272091388702393, "learning_rate": 1.996408909165776e-05, "loss": 1.0306, "step": 292 }, { "epoch": 0.05638953040800616, "grad_norm": 1.9080709218978882, "learning_rate": 1.99635593706661e-05, "loss": 1.0241, "step": 293 }, { "epoch": 0.05658198614318707, "grad_norm": 1.8557677268981934, "learning_rate": 1.996302577840703e-05, "loss": 1.0523, "step": 294 }, { "epoch": 0.05677444187836798, "grad_norm": 1.0520893335342407, "learning_rate": 1.9962488315087872e-05, "loss": 0.9096, "step": 295 }, { "epoch": 0.05696689761354888, "grad_norm": 1.9271793365478516, "learning_rate": 1.9961946980917457e-05, "loss": 1.058, "step": 296 }, { "epoch": 0.05715935334872979, "grad_norm": 2.0063111782073975, "learning_rate": 1.9961401776106113e-05, "loss": 0.9968, "step": 297 }, { "epoch": 0.0573518090839107, "grad_norm": 1.998648762702942, "learning_rate": 1.996085270086568e-05, "loss": 1.1263, "step": 298 }, { "epoch": 0.05754426481909161, "grad_norm": 1.8948326110839844, "learning_rate": 1.996029975540949e-05, "loss": 1.0912, "step": 299 }, { "epoch": 0.057736720554272515, "grad_norm": 2.044820547103882, "learning_rate": 1.9959742939952393e-05, "loss": 1.186, "step": 300 }, { "epoch": 0.057929176289453425, "grad_norm": 2.0060698986053467, "learning_rate": 1.995918225471073e-05, "loss": 1.0976, "step": 301 }, { "epoch": 0.058121632024634336, "grad_norm": 1.9796364307403564, "learning_rate": 1.9958617699902354e-05, "loss": 1.1375, "step": 302 }, { "epoch": 0.05831408775981524, "grad_norm": 2.122551679611206, "learning_rate": 1.9958049275746617e-05, "loss": 1.0859, "step": 303 }, { "epoch": 0.05850654349499615, "grad_norm": 1.786564826965332, "learning_rate": 1.9957476982464382e-05, "loss": 1.1212, "step": 304 }, { "epoch": 0.05869899923017706, "grad_norm": 1.9185203313827515, "learning_rate": 1.9956900820278e-05, "loss": 0.9609, "step": 305 }, { "epoch": 0.05889145496535797, "grad_norm": 2.0406312942504883, "learning_rate": 1.9956320789411338e-05, "loss": 1.1476, "step": 306 }, { "epoch": 0.05908391070053887, "grad_norm": 1.8319671154022217, "learning_rate": 1.9955736890089768e-05, "loss": 1.06, "step": 307 }, { "epoch": 0.059276366435719784, "grad_norm": 1.9506793022155762, "learning_rate": 1.995514912254015e-05, "loss": 1.048, "step": 308 }, { "epoch": 0.059468822170900694, "grad_norm": 1.7802228927612305, "learning_rate": 1.995455748699087e-05, "loss": 1.0628, "step": 309 }, { "epoch": 0.059661277906081604, "grad_norm": 1.8466390371322632, "learning_rate": 1.9953961983671792e-05, "loss": 1.1178, "step": 310 }, { "epoch": 0.05985373364126251, "grad_norm": 1.8039296865463257, "learning_rate": 1.9953362612814294e-05, "loss": 1.0751, "step": 311 }, { "epoch": 0.06004618937644342, "grad_norm": 2.070028066635132, "learning_rate": 1.9952759374651266e-05, "loss": 1.0708, "step": 312 }, { "epoch": 0.06023864511162433, "grad_norm": 1.0980297327041626, "learning_rate": 1.9952152269417088e-05, "loss": 0.9121, "step": 313 }, { "epoch": 0.06043110084680523, "grad_norm": 1.9291366338729858, "learning_rate": 1.9951541297347642e-05, "loss": 1.1029, "step": 314 }, { "epoch": 0.06062355658198614, "grad_norm": 1.8742951154708862, "learning_rate": 1.995092645868032e-05, "loss": 1.081, "step": 315 }, { "epoch": 0.06081601231716705, "grad_norm": 1.9387081861495972, "learning_rate": 1.9950307753654016e-05, "loss": 1.1246, "step": 316 }, { "epoch": 0.06100846805234796, "grad_norm": 2.033475160598755, "learning_rate": 1.9949685182509117e-05, "loss": 1.0838, "step": 317 }, { "epoch": 0.061200923787528866, "grad_norm": 2.0009639263153076, "learning_rate": 1.9949058745487524e-05, "loss": 1.044, "step": 318 }, { "epoch": 0.061393379522709776, "grad_norm": 2.098273515701294, "learning_rate": 1.9948428442832633e-05, "loss": 1.1293, "step": 319 }, { "epoch": 0.061585835257890686, "grad_norm": 1.8275796175003052, "learning_rate": 1.994779427478934e-05, "loss": 1.0483, "step": 320 }, { "epoch": 0.0617782909930716, "grad_norm": 1.8539766073226929, "learning_rate": 1.994715624160405e-05, "loss": 1.0825, "step": 321 }, { "epoch": 0.0619707467282525, "grad_norm": 2.0822293758392334, "learning_rate": 1.994651434352467e-05, "loss": 1.0593, "step": 322 }, { "epoch": 0.06216320246343341, "grad_norm": 2.0292410850524902, "learning_rate": 1.9945868580800593e-05, "loss": 1.0776, "step": 323 }, { "epoch": 0.06235565819861432, "grad_norm": 1.727033019065857, "learning_rate": 1.9945218953682736e-05, "loss": 1.0782, "step": 324 }, { "epoch": 0.06254811393379522, "grad_norm": 1.9884116649627686, "learning_rate": 1.9944565462423502e-05, "loss": 1.0875, "step": 325 }, { "epoch": 0.06274056966897613, "grad_norm": 1.898795247077942, "learning_rate": 1.99439081072768e-05, "loss": 0.9993, "step": 326 }, { "epoch": 0.06293302540415704, "grad_norm": 1.8701682090759277, "learning_rate": 1.994324688849804e-05, "loss": 1.0776, "step": 327 }, { "epoch": 0.06312548113933796, "grad_norm": 1.8718277215957642, "learning_rate": 1.994258180634414e-05, "loss": 1.1456, "step": 328 }, { "epoch": 0.06331793687451887, "grad_norm": 1.8626608848571777, "learning_rate": 1.9941912861073507e-05, "loss": 1.0685, "step": 329 }, { "epoch": 0.06351039260969978, "grad_norm": 2.0976054668426514, "learning_rate": 1.9941240052946055e-05, "loss": 1.1204, "step": 330 }, { "epoch": 0.06370284834488067, "grad_norm": 1.9197001457214355, "learning_rate": 1.9940563382223196e-05, "loss": 1.1056, "step": 331 }, { "epoch": 0.06389530408006158, "grad_norm": 1.9192863702774048, "learning_rate": 1.9939882849167853e-05, "loss": 1.1318, "step": 332 }, { "epoch": 0.06408775981524249, "grad_norm": 1.756412148475647, "learning_rate": 1.993919845404444e-05, "loss": 1.0828, "step": 333 }, { "epoch": 0.0642802155504234, "grad_norm": 1.804733395576477, "learning_rate": 1.9938510197118868e-05, "loss": 1.0581, "step": 334 }, { "epoch": 0.06447267128560431, "grad_norm": 1.925678014755249, "learning_rate": 1.9937818078658556e-05, "loss": 1.1149, "step": 335 }, { "epoch": 0.06466512702078522, "grad_norm": 1.9162620306015015, "learning_rate": 1.9937122098932428e-05, "loss": 1.1522, "step": 336 }, { "epoch": 0.06485758275596613, "grad_norm": 1.7437552213668823, "learning_rate": 1.9936422258210894e-05, "loss": 1.0731, "step": 337 }, { "epoch": 0.06505003849114703, "grad_norm": 1.802807092666626, "learning_rate": 1.9935718556765878e-05, "loss": 1.0235, "step": 338 }, { "epoch": 0.06524249422632794, "grad_norm": 1.8687171936035156, "learning_rate": 1.993501099487079e-05, "loss": 1.0161, "step": 339 }, { "epoch": 0.06543494996150885, "grad_norm": 1.0427062511444092, "learning_rate": 1.9934299572800556e-05, "loss": 0.9383, "step": 340 }, { "epoch": 0.06562740569668976, "grad_norm": 2.008777618408203, "learning_rate": 1.9933584290831593e-05, "loss": 1.0689, "step": 341 }, { "epoch": 0.06581986143187067, "grad_norm": 1.9812185764312744, "learning_rate": 1.9932865149241813e-05, "loss": 1.071, "step": 342 }, { "epoch": 0.06601231716705158, "grad_norm": 1.9773062467575073, "learning_rate": 1.9932142148310638e-05, "loss": 1.1289, "step": 343 }, { "epoch": 0.06620477290223249, "grad_norm": 1.7774780988693237, "learning_rate": 1.9931415288318985e-05, "loss": 1.091, "step": 344 }, { "epoch": 0.06639722863741339, "grad_norm": 2.137747287750244, "learning_rate": 1.9930684569549265e-05, "loss": 1.0723, "step": 345 }, { "epoch": 0.0665896843725943, "grad_norm": 2.0008246898651123, "learning_rate": 1.9929949992285397e-05, "loss": 1.0538, "step": 346 }, { "epoch": 0.06678214010777521, "grad_norm": 1.9377113580703735, "learning_rate": 1.9929211556812794e-05, "loss": 1.0339, "step": 347 }, { "epoch": 0.06697459584295612, "grad_norm": 1.8344753980636597, "learning_rate": 1.9928469263418376e-05, "loss": 1.1363, "step": 348 }, { "epoch": 0.06716705157813703, "grad_norm": 2.0816099643707275, "learning_rate": 1.9927723112390547e-05, "loss": 1.0178, "step": 349 }, { "epoch": 0.06735950731331794, "grad_norm": 0.8840500712394714, "learning_rate": 1.992697310401922e-05, "loss": 0.8425, "step": 350 }, { "epoch": 0.06755196304849885, "grad_norm": 2.039031982421875, "learning_rate": 1.992621923859581e-05, "loss": 1.0392, "step": 351 }, { "epoch": 0.06774441878367975, "grad_norm": 2.1741487979888916, "learning_rate": 1.9925461516413224e-05, "loss": 1.0477, "step": 352 }, { "epoch": 0.06793687451886066, "grad_norm": 1.8604975938796997, "learning_rate": 1.9924699937765863e-05, "loss": 1.0269, "step": 353 }, { "epoch": 0.06812933025404157, "grad_norm": 2.037583112716675, "learning_rate": 1.9923934502949645e-05, "loss": 1.0948, "step": 354 }, { "epoch": 0.06832178598922248, "grad_norm": 1.9853684902191162, "learning_rate": 1.992316521226196e-05, "loss": 1.038, "step": 355 }, { "epoch": 0.06851424172440339, "grad_norm": 1.8607470989227295, "learning_rate": 1.9922392066001724e-05, "loss": 0.9963, "step": 356 }, { "epoch": 0.0687066974595843, "grad_norm": 1.8960143327713013, "learning_rate": 1.9921615064469328e-05, "loss": 1.1849, "step": 357 }, { "epoch": 0.06889915319476521, "grad_norm": 1.7322030067443848, "learning_rate": 1.992083420796667e-05, "loss": 1.0686, "step": 358 }, { "epoch": 0.06909160892994612, "grad_norm": 2.041977882385254, "learning_rate": 1.9920049496797153e-05, "loss": 1.0509, "step": 359 }, { "epoch": 0.06928406466512702, "grad_norm": 1.9696863889694214, "learning_rate": 1.9919260931265666e-05, "loss": 1.1118, "step": 360 }, { "epoch": 0.06947652040030793, "grad_norm": 2.049968719482422, "learning_rate": 1.99184685116786e-05, "loss": 1.1194, "step": 361 }, { "epoch": 0.06966897613548884, "grad_norm": 1.967901587486267, "learning_rate": 1.991767223834384e-05, "loss": 1.0606, "step": 362 }, { "epoch": 0.06986143187066975, "grad_norm": 1.858583688735962, "learning_rate": 1.9916872111570782e-05, "loss": 1.0418, "step": 363 }, { "epoch": 0.07005388760585066, "grad_norm": 1.0986965894699097, "learning_rate": 1.9916068131670302e-05, "loss": 0.9747, "step": 364 }, { "epoch": 0.07024634334103157, "grad_norm": 2.4967737197875977, "learning_rate": 1.9915260298954786e-05, "loss": 1.0825, "step": 365 }, { "epoch": 0.07043879907621248, "grad_norm": 1.919541835784912, "learning_rate": 1.9914448613738107e-05, "loss": 1.1308, "step": 366 }, { "epoch": 0.07063125481139337, "grad_norm": 0.9795111417770386, "learning_rate": 1.9913633076335638e-05, "loss": 0.9576, "step": 367 }, { "epoch": 0.07082371054657428, "grad_norm": 2.0834782123565674, "learning_rate": 1.9912813687064255e-05, "loss": 1.0511, "step": 368 }, { "epoch": 0.0710161662817552, "grad_norm": 1.8799113035202026, "learning_rate": 1.991199044624232e-05, "loss": 1.003, "step": 369 }, { "epoch": 0.0712086220169361, "grad_norm": 1.8467507362365723, "learning_rate": 1.9911163354189702e-05, "loss": 1.0391, "step": 370 }, { "epoch": 0.07140107775211701, "grad_norm": 2.0189335346221924, "learning_rate": 1.991033241122776e-05, "loss": 1.071, "step": 371 }, { "epoch": 0.07159353348729793, "grad_norm": 2.0685229301452637, "learning_rate": 1.990949761767935e-05, "loss": 1.1198, "step": 372 }, { "epoch": 0.07178598922247884, "grad_norm": 1.9276286363601685, "learning_rate": 1.9908658973868823e-05, "loss": 1.0945, "step": 373 }, { "epoch": 0.07197844495765973, "grad_norm": 2.1261022090911865, "learning_rate": 1.9907816480122032e-05, "loss": 1.1694, "step": 374 }, { "epoch": 0.07217090069284064, "grad_norm": 2.046309471130371, "learning_rate": 1.9906970136766323e-05, "loss": 1.0982, "step": 375 }, { "epoch": 0.07236335642802155, "grad_norm": 2.1238198280334473, "learning_rate": 1.9906119944130527e-05, "loss": 1.1492, "step": 376 }, { "epoch": 0.07255581216320246, "grad_norm": 1.8909717798233032, "learning_rate": 1.990526590254499e-05, "loss": 1.0794, "step": 377 }, { "epoch": 0.07274826789838337, "grad_norm": 2.1433300971984863, "learning_rate": 1.9904408012341542e-05, "loss": 1.022, "step": 378 }, { "epoch": 0.07294072363356428, "grad_norm": 1.944788932800293, "learning_rate": 1.9903546273853503e-05, "loss": 1.0335, "step": 379 }, { "epoch": 0.0731331793687452, "grad_norm": 1.7993412017822266, "learning_rate": 1.9902680687415704e-05, "loss": 1.1729, "step": 380 }, { "epoch": 0.07332563510392609, "grad_norm": 2.169924736022949, "learning_rate": 1.9901811253364458e-05, "loss": 0.9873, "step": 381 }, { "epoch": 0.073518090839107, "grad_norm": 2.0918314456939697, "learning_rate": 1.9900937972037577e-05, "loss": 1.1523, "step": 382 }, { "epoch": 0.07371054657428791, "grad_norm": 1.8481202125549316, "learning_rate": 1.9900060843774364e-05, "loss": 1.1308, "step": 383 }, { "epoch": 0.07390300230946882, "grad_norm": 1.8293710947036743, "learning_rate": 1.989917986891563e-05, "loss": 1.1532, "step": 384 }, { "epoch": 0.07409545804464973, "grad_norm": 2.140244483947754, "learning_rate": 1.9898295047803665e-05, "loss": 1.0799, "step": 385 }, { "epoch": 0.07428791377983064, "grad_norm": 2.045227527618408, "learning_rate": 1.9897406380782262e-05, "loss": 1.1405, "step": 386 }, { "epoch": 0.07448036951501155, "grad_norm": 2.0215260982513428, "learning_rate": 1.9896513868196706e-05, "loss": 1.0717, "step": 387 }, { "epoch": 0.07467282525019246, "grad_norm": 2.1683692932128906, "learning_rate": 1.9895617510393773e-05, "loss": 1.0307, "step": 388 }, { "epoch": 0.07486528098537336, "grad_norm": 1.9758880138397217, "learning_rate": 1.9894717307721734e-05, "loss": 1.0851, "step": 389 }, { "epoch": 0.07505773672055427, "grad_norm": 1.7286189794540405, "learning_rate": 1.9893813260530368e-05, "loss": 1.1128, "step": 390 }, { "epoch": 0.07525019245573518, "grad_norm": 1.9432743787765503, "learning_rate": 1.989290536917093e-05, "loss": 0.9664, "step": 391 }, { "epoch": 0.07544264819091609, "grad_norm": 2.002781629562378, "learning_rate": 1.9891993633996164e-05, "loss": 1.0528, "step": 392 }, { "epoch": 0.075635103926097, "grad_norm": 1.921952486038208, "learning_rate": 1.9891078055360332e-05, "loss": 1.02, "step": 393 }, { "epoch": 0.07582755966127791, "grad_norm": 1.068670392036438, "learning_rate": 1.989015863361917e-05, "loss": 0.9773, "step": 394 }, { "epoch": 0.07602001539645882, "grad_norm": 1.9413923025131226, "learning_rate": 1.9889235369129913e-05, "loss": 1.1125, "step": 395 }, { "epoch": 0.07621247113163972, "grad_norm": 1.9656031131744385, "learning_rate": 1.9888308262251286e-05, "loss": 1.176, "step": 396 }, { "epoch": 0.07640492686682063, "grad_norm": 2.0729422569274902, "learning_rate": 1.9887377313343514e-05, "loss": 1.069, "step": 397 }, { "epoch": 0.07659738260200154, "grad_norm": 2.3098745346069336, "learning_rate": 1.988644252276831e-05, "loss": 1.0769, "step": 398 }, { "epoch": 0.07678983833718245, "grad_norm": 1.8306607007980347, "learning_rate": 1.9885503890888874e-05, "loss": 1.0606, "step": 399 }, { "epoch": 0.07698229407236336, "grad_norm": 2.4004573822021484, "learning_rate": 1.988456141806991e-05, "loss": 1.1003, "step": 400 }, { "epoch": 0.07717474980754427, "grad_norm": 2.3030784130096436, "learning_rate": 1.988361510467761e-05, "loss": 1.044, "step": 401 }, { "epoch": 0.07736720554272518, "grad_norm": 0.9968491792678833, "learning_rate": 1.9882664951079652e-05, "loss": 0.8923, "step": 402 }, { "epoch": 0.07755966127790608, "grad_norm": 2.032879590988159, "learning_rate": 1.988171095764521e-05, "loss": 1.0133, "step": 403 }, { "epoch": 0.07775211701308699, "grad_norm": 1.8498350381851196, "learning_rate": 1.9880753124744964e-05, "loss": 1.0867, "step": 404 }, { "epoch": 0.0779445727482679, "grad_norm": 2.009408950805664, "learning_rate": 1.9879791452751055e-05, "loss": 1.0984, "step": 405 }, { "epoch": 0.0781370284834488, "grad_norm": 2.1746535301208496, "learning_rate": 1.9878825942037147e-05, "loss": 1.0417, "step": 406 }, { "epoch": 0.07832948421862972, "grad_norm": 0.9810822606086731, "learning_rate": 1.987785659297838e-05, "loss": 0.904, "step": 407 }, { "epoch": 0.07852193995381063, "grad_norm": 2.107671022415161, "learning_rate": 1.9876883405951378e-05, "loss": 1.0549, "step": 408 }, { "epoch": 0.07871439568899154, "grad_norm": 2.4178225994110107, "learning_rate": 1.9875906381334277e-05, "loss": 1.0272, "step": 409 }, { "epoch": 0.07890685142417245, "grad_norm": 2.0550055503845215, "learning_rate": 1.9874925519506686e-05, "loss": 0.9923, "step": 410 }, { "epoch": 0.07909930715935334, "grad_norm": 2.1893374919891357, "learning_rate": 1.9873940820849714e-05, "loss": 1.0564, "step": 411 }, { "epoch": 0.07929176289453425, "grad_norm": 2.048220157623291, "learning_rate": 1.9872952285745958e-05, "loss": 1.139, "step": 412 }, { "epoch": 0.07948421862971516, "grad_norm": 1.8988646268844604, "learning_rate": 1.987195991457951e-05, "loss": 1.0641, "step": 413 }, { "epoch": 0.07967667436489607, "grad_norm": 1.9113562107086182, "learning_rate": 1.987096370773594e-05, "loss": 1.059, "step": 414 }, { "epoch": 0.07986913010007698, "grad_norm": 2.0818557739257812, "learning_rate": 1.9869963665602322e-05, "loss": 1.0276, "step": 415 }, { "epoch": 0.0800615858352579, "grad_norm": 1.956807017326355, "learning_rate": 1.9868959788567213e-05, "loss": 1.126, "step": 416 }, { "epoch": 0.0802540415704388, "grad_norm": 1.8188530206680298, "learning_rate": 1.9867952077020667e-05, "loss": 1.1835, "step": 417 }, { "epoch": 0.0804464973056197, "grad_norm": 1.8978670835494995, "learning_rate": 1.9866940531354214e-05, "loss": 1.0995, "step": 418 }, { "epoch": 0.08063895304080061, "grad_norm": 1.8703874349594116, "learning_rate": 1.9865925151960893e-05, "loss": 1.0056, "step": 419 }, { "epoch": 0.08083140877598152, "grad_norm": 2.131010055541992, "learning_rate": 1.9864905939235215e-05, "loss": 1.1252, "step": 420 }, { "epoch": 0.08102386451116243, "grad_norm": 1.9145139455795288, "learning_rate": 1.9863882893573188e-05, "loss": 1.0677, "step": 421 }, { "epoch": 0.08121632024634334, "grad_norm": 2.0749764442443848, "learning_rate": 1.9862856015372315e-05, "loss": 1.081, "step": 422 }, { "epoch": 0.08140877598152425, "grad_norm": 0.9918889403343201, "learning_rate": 1.9861825305031576e-05, "loss": 0.9055, "step": 423 }, { "epoch": 0.08160123171670516, "grad_norm": 2.1616575717926025, "learning_rate": 1.9860790762951447e-05, "loss": 1.0934, "step": 424 }, { "epoch": 0.08179368745188606, "grad_norm": 1.9510889053344727, "learning_rate": 1.985975238953389e-05, "loss": 1.1145, "step": 425 }, { "epoch": 0.08198614318706697, "grad_norm": 1.8759390115737915, "learning_rate": 1.985871018518236e-05, "loss": 1.0763, "step": 426 }, { "epoch": 0.08217859892224788, "grad_norm": 1.9842033386230469, "learning_rate": 1.9857664150301798e-05, "loss": 1.0472, "step": 427 }, { "epoch": 0.08237105465742879, "grad_norm": 1.724048137664795, "learning_rate": 1.985661428529863e-05, "loss": 1.1517, "step": 428 }, { "epoch": 0.0825635103926097, "grad_norm": 1.0283504724502563, "learning_rate": 1.985556059058078e-05, "loss": 0.8785, "step": 429 }, { "epoch": 0.08275596612779061, "grad_norm": 2.0816354751586914, "learning_rate": 1.9854503066557646e-05, "loss": 1.0595, "step": 430 }, { "epoch": 0.08294842186297152, "grad_norm": 1.015547752380371, "learning_rate": 1.9853441713640123e-05, "loss": 0.9407, "step": 431 }, { "epoch": 0.08314087759815242, "grad_norm": 1.887131929397583, "learning_rate": 1.9852376532240594e-05, "loss": 0.9853, "step": 432 }, { "epoch": 0.08333333333333333, "grad_norm": 1.8701651096343994, "learning_rate": 1.9851307522772926e-05, "loss": 1.0211, "step": 433 }, { "epoch": 0.08352578906851424, "grad_norm": 1.9214274883270264, "learning_rate": 1.9850234685652478e-05, "loss": 1.0719, "step": 434 }, { "epoch": 0.08371824480369515, "grad_norm": 1.7181062698364258, "learning_rate": 1.9849158021296083e-05, "loss": 1.0983, "step": 435 }, { "epoch": 0.08391070053887606, "grad_norm": 2.1615700721740723, "learning_rate": 1.9848077530122083e-05, "loss": 0.9326, "step": 436 }, { "epoch": 0.08410315627405697, "grad_norm": 1.783625602722168, "learning_rate": 1.9846993212550286e-05, "loss": 1.1539, "step": 437 }, { "epoch": 0.08429561200923788, "grad_norm": 1.9634573459625244, "learning_rate": 1.9845905069002004e-05, "loss": 1.0632, "step": 438 }, { "epoch": 0.08448806774441879, "grad_norm": 2.064742088317871, "learning_rate": 1.984481309990002e-05, "loss": 1.0204, "step": 439 }, { "epoch": 0.08468052347959969, "grad_norm": 1.9431153535842896, "learning_rate": 1.984371730566861e-05, "loss": 1.0388, "step": 440 }, { "epoch": 0.0848729792147806, "grad_norm": 1.9191973209381104, "learning_rate": 1.9842617686733546e-05, "loss": 1.0669, "step": 441 }, { "epoch": 0.08506543494996151, "grad_norm": 1.0412219762802124, "learning_rate": 1.9841514243522065e-05, "loss": 0.9307, "step": 442 }, { "epoch": 0.08525789068514242, "grad_norm": 0.9976733326911926, "learning_rate": 1.984040697646291e-05, "loss": 0.9571, "step": 443 }, { "epoch": 0.08545034642032333, "grad_norm": 1.971793293952942, "learning_rate": 1.98392958859863e-05, "loss": 1.0569, "step": 444 }, { "epoch": 0.08564280215550424, "grad_norm": 1.9992115497589111, "learning_rate": 1.9838180972523936e-05, "loss": 1.0777, "step": 445 }, { "epoch": 0.08583525789068515, "grad_norm": 1.8540982007980347, "learning_rate": 1.9837062236509013e-05, "loss": 1.1332, "step": 446 }, { "epoch": 0.08602771362586605, "grad_norm": 1.8273231983184814, "learning_rate": 1.9835939678376213e-05, "loss": 1.0257, "step": 447 }, { "epoch": 0.08622016936104696, "grad_norm": 1.8232215642929077, "learning_rate": 1.9834813298561686e-05, "loss": 1.074, "step": 448 }, { "epoch": 0.08641262509622787, "grad_norm": 2.028073310852051, "learning_rate": 1.9833683097503087e-05, "loss": 1.0785, "step": 449 }, { "epoch": 0.08660508083140878, "grad_norm": 2.0158872604370117, "learning_rate": 1.983254907563955e-05, "loss": 1.1016, "step": 450 }, { "epoch": 0.08679753656658969, "grad_norm": 1.934831976890564, "learning_rate": 1.983141123341168e-05, "loss": 1.1075, "step": 451 }, { "epoch": 0.0869899923017706, "grad_norm": 1.8577687740325928, "learning_rate": 1.9830269571261585e-05, "loss": 1.0142, "step": 452 }, { "epoch": 0.08718244803695151, "grad_norm": 1.7445321083068848, "learning_rate": 1.982912408963285e-05, "loss": 1.0305, "step": 453 }, { "epoch": 0.0873749037721324, "grad_norm": 1.948188304901123, "learning_rate": 1.9827974788970536e-05, "loss": 1.0868, "step": 454 }, { "epoch": 0.08756735950731331, "grad_norm": 2.034566879272461, "learning_rate": 1.9826821669721205e-05, "loss": 1.0884, "step": 455 }, { "epoch": 0.08775981524249422, "grad_norm": 2.072718381881714, "learning_rate": 1.9825664732332886e-05, "loss": 1.0592, "step": 456 }, { "epoch": 0.08795227097767513, "grad_norm": 1.9664205312728882, "learning_rate": 1.98245039772551e-05, "loss": 1.0747, "step": 457 }, { "epoch": 0.08814472671285604, "grad_norm": 2.13956356048584, "learning_rate": 1.982333940493885e-05, "loss": 1.0716, "step": 458 }, { "epoch": 0.08833718244803695, "grad_norm": 1.06049382686615, "learning_rate": 1.9822171015836625e-05, "loss": 0.9272, "step": 459 }, { "epoch": 0.08852963818321787, "grad_norm": 2.323936939239502, "learning_rate": 1.982099881040239e-05, "loss": 1.066, "step": 460 }, { "epoch": 0.08872209391839876, "grad_norm": 1.823792576789856, "learning_rate": 1.9819822789091597e-05, "loss": 1.0847, "step": 461 }, { "epoch": 0.08891454965357967, "grad_norm": 1.8866627216339111, "learning_rate": 1.9818642952361188e-05, "loss": 1.1211, "step": 462 }, { "epoch": 0.08910700538876058, "grad_norm": 1.824110507965088, "learning_rate": 1.9817459300669567e-05, "loss": 1.0453, "step": 463 }, { "epoch": 0.08929946112394149, "grad_norm": 2.1441428661346436, "learning_rate": 1.9816271834476642e-05, "loss": 1.1479, "step": 464 }, { "epoch": 0.0894919168591224, "grad_norm": 1.911728024482727, "learning_rate": 1.981508055424379e-05, "loss": 1.023, "step": 465 }, { "epoch": 0.08968437259430331, "grad_norm": 2.1533942222595215, "learning_rate": 1.981388546043388e-05, "loss": 1.107, "step": 466 }, { "epoch": 0.08987682832948422, "grad_norm": 1.9569392204284668, "learning_rate": 1.9812686553511246e-05, "loss": 1.0745, "step": 467 }, { "epoch": 0.09006928406466513, "grad_norm": 1.01942777633667, "learning_rate": 1.9811483833941726e-05, "loss": 0.9213, "step": 468 }, { "epoch": 0.09026173979984603, "grad_norm": 2.0068280696868896, "learning_rate": 1.9810277302192624e-05, "loss": 1.1049, "step": 469 }, { "epoch": 0.09045419553502694, "grad_norm": 2.019318103790283, "learning_rate": 1.9809066958732727e-05, "loss": 0.9636, "step": 470 }, { "epoch": 0.09064665127020785, "grad_norm": 1.8868746757507324, "learning_rate": 1.9807852804032306e-05, "loss": 0.9739, "step": 471 }, { "epoch": 0.09083910700538876, "grad_norm": 1.979896068572998, "learning_rate": 1.9806634838563113e-05, "loss": 1.0486, "step": 472 }, { "epoch": 0.09103156274056967, "grad_norm": 2.0071215629577637, "learning_rate": 1.9805413062798376e-05, "loss": 1.011, "step": 473 }, { "epoch": 0.09122401847575058, "grad_norm": 1.9919341802597046, "learning_rate": 1.980418747721281e-05, "loss": 1.0585, "step": 474 }, { "epoch": 0.09141647421093149, "grad_norm": 2.2145938873291016, "learning_rate": 1.980295808228261e-05, "loss": 0.9632, "step": 475 }, { "epoch": 0.09160892994611239, "grad_norm": 1.92658531665802, "learning_rate": 1.9801724878485438e-05, "loss": 1.0504, "step": 476 }, { "epoch": 0.0918013856812933, "grad_norm": 1.8957911729812622, "learning_rate": 1.980048786630046e-05, "loss": 0.993, "step": 477 }, { "epoch": 0.09199384141647421, "grad_norm": 1.8102710247039795, "learning_rate": 1.9799247046208297e-05, "loss": 1.0914, "step": 478 }, { "epoch": 0.09218629715165512, "grad_norm": 1.8429460525512695, "learning_rate": 1.9798002418691067e-05, "loss": 1.0703, "step": 479 }, { "epoch": 0.09237875288683603, "grad_norm": 1.7665605545043945, "learning_rate": 1.9796753984232357e-05, "loss": 1.1114, "step": 480 }, { "epoch": 0.09257120862201694, "grad_norm": 1.9230808019638062, "learning_rate": 1.979550174331724e-05, "loss": 1.0789, "step": 481 }, { "epoch": 0.09276366435719785, "grad_norm": 1.7684491872787476, "learning_rate": 1.9794245696432262e-05, "loss": 1.0645, "step": 482 }, { "epoch": 0.09295612009237875, "grad_norm": 1.9817627668380737, "learning_rate": 1.979298584406545e-05, "loss": 1.0845, "step": 483 }, { "epoch": 0.09314857582755966, "grad_norm": 2.1252405643463135, "learning_rate": 1.9791722186706318e-05, "loss": 1.0144, "step": 484 }, { "epoch": 0.09334103156274057, "grad_norm": 2.1359128952026367, "learning_rate": 1.979045472484584e-05, "loss": 1.0366, "step": 485 }, { "epoch": 0.09353348729792148, "grad_norm": 1.8751832246780396, "learning_rate": 1.9789183458976485e-05, "loss": 1.0818, "step": 486 }, { "epoch": 0.09372594303310239, "grad_norm": 2.0680389404296875, "learning_rate": 1.9787908389592197e-05, "loss": 1.0357, "step": 487 }, { "epoch": 0.0939183987682833, "grad_norm": 1.8467575311660767, "learning_rate": 1.9786629517188384e-05, "loss": 1.0464, "step": 488 }, { "epoch": 0.09411085450346421, "grad_norm": 1.9802943468093872, "learning_rate": 1.9785346842261957e-05, "loss": 1.0413, "step": 489 }, { "epoch": 0.0943033102386451, "grad_norm": 1.9998193979263306, "learning_rate": 1.9784060365311276e-05, "loss": 1.078, "step": 490 }, { "epoch": 0.09449576597382602, "grad_norm": 2.2239794731140137, "learning_rate": 1.97827700868362e-05, "loss": 1.1261, "step": 491 }, { "epoch": 0.09468822170900693, "grad_norm": 1.8362261056900024, "learning_rate": 1.9781476007338058e-05, "loss": 1.062, "step": 492 }, { "epoch": 0.09488067744418784, "grad_norm": 1.9770945310592651, "learning_rate": 1.978017812731965e-05, "loss": 1.053, "step": 493 }, { "epoch": 0.09507313317936875, "grad_norm": 1.838757038116455, "learning_rate": 1.9778876447285264e-05, "loss": 1.0547, "step": 494 }, { "epoch": 0.09526558891454966, "grad_norm": 1.8650875091552734, "learning_rate": 1.9777570967740648e-05, "loss": 1.0621, "step": 495 }, { "epoch": 0.09545804464973057, "grad_norm": 1.843515396118164, "learning_rate": 1.977626168919305e-05, "loss": 1.018, "step": 496 }, { "epoch": 0.09565050038491148, "grad_norm": 1.9621353149414062, "learning_rate": 1.977494861215117e-05, "loss": 1.1262, "step": 497 }, { "epoch": 0.09584295612009237, "grad_norm": 1.9379910230636597, "learning_rate": 1.9773631737125192e-05, "loss": 1.1076, "step": 498 }, { "epoch": 0.09603541185527328, "grad_norm": 2.0728769302368164, "learning_rate": 1.977231106462679e-05, "loss": 1.0357, "step": 499 }, { "epoch": 0.0962278675904542, "grad_norm": 2.0274953842163086, "learning_rate": 1.9770986595169096e-05, "loss": 1.1233, "step": 500 }, { "epoch": 0.0964203233256351, "grad_norm": 2.0266332626342773, "learning_rate": 1.9769658329266718e-05, "loss": 1.0848, "step": 501 }, { "epoch": 0.09661277906081601, "grad_norm": 1.8384894132614136, "learning_rate": 1.976832626743575e-05, "loss": 1.0968, "step": 502 }, { "epoch": 0.09680523479599693, "grad_norm": 1.831685185432434, "learning_rate": 1.9766990410193755e-05, "loss": 1.0704, "step": 503 }, { "epoch": 0.09699769053117784, "grad_norm": 1.9811996221542358, "learning_rate": 1.9765650758059766e-05, "loss": 1.1019, "step": 504 }, { "epoch": 0.09719014626635873, "grad_norm": 1.9394782781600952, "learning_rate": 1.97643073115543e-05, "loss": 1.0505, "step": 505 }, { "epoch": 0.09738260200153964, "grad_norm": 1.9711352586746216, "learning_rate": 1.9762960071199334e-05, "loss": 0.9832, "step": 506 }, { "epoch": 0.09757505773672055, "grad_norm": 1.8081459999084473, "learning_rate": 1.976160903751834e-05, "loss": 1.087, "step": 507 }, { "epoch": 0.09776751347190146, "grad_norm": 1.7603284120559692, "learning_rate": 1.9760254211036245e-05, "loss": 1.0364, "step": 508 }, { "epoch": 0.09795996920708237, "grad_norm": 1.915087103843689, "learning_rate": 1.9758895592279458e-05, "loss": 1.0721, "step": 509 }, { "epoch": 0.09815242494226328, "grad_norm": 1.7813336849212646, "learning_rate": 1.9757533181775865e-05, "loss": 1.1291, "step": 510 }, { "epoch": 0.0983448806774442, "grad_norm": 1.7126929759979248, "learning_rate": 1.9756166980054812e-05, "loss": 1.0705, "step": 511 }, { "epoch": 0.09853733641262509, "grad_norm": 1.811356782913208, "learning_rate": 1.9754796987647134e-05, "loss": 1.0951, "step": 512 }, { "epoch": 0.098729792147806, "grad_norm": 1.018143653869629, "learning_rate": 1.9753423205085127e-05, "loss": 0.9113, "step": 513 }, { "epoch": 0.09892224788298691, "grad_norm": 1.9739432334899902, "learning_rate": 1.9752045632902567e-05, "loss": 1.06, "step": 514 }, { "epoch": 0.09911470361816782, "grad_norm": 1.6870028972625732, "learning_rate": 1.97506642716347e-05, "loss": 1.1066, "step": 515 }, { "epoch": 0.09930715935334873, "grad_norm": 1.9246145486831665, "learning_rate": 1.9749279121818235e-05, "loss": 1.0121, "step": 516 }, { "epoch": 0.09949961508852964, "grad_norm": 1.8966962099075317, "learning_rate": 1.9747890183991376e-05, "loss": 1.1529, "step": 517 }, { "epoch": 0.09969207082371055, "grad_norm": 2.0046677589416504, "learning_rate": 1.9746497458693776e-05, "loss": 1.0772, "step": 518 }, { "epoch": 0.09988452655889145, "grad_norm": 1.9526045322418213, "learning_rate": 1.9745100946466563e-05, "loss": 0.9979, "step": 519 }, { "epoch": 0.10007698229407236, "grad_norm": 1.9230717420578003, "learning_rate": 1.9743700647852356e-05, "loss": 0.9828, "step": 520 }, { "epoch": 0.10026943802925327, "grad_norm": 1.8227673768997192, "learning_rate": 1.9742296563395218e-05, "loss": 1.0033, "step": 521 }, { "epoch": 0.10046189376443418, "grad_norm": 2.004889965057373, "learning_rate": 1.9740888693640698e-05, "loss": 1.078, "step": 522 }, { "epoch": 0.10065434949961509, "grad_norm": 1.7988289594650269, "learning_rate": 1.973947703913582e-05, "loss": 0.923, "step": 523 }, { "epoch": 0.100846805234796, "grad_norm": 1.7453898191452026, "learning_rate": 1.9738061600429062e-05, "loss": 1.0365, "step": 524 }, { "epoch": 0.10103926096997691, "grad_norm": 2.1042673587799072, "learning_rate": 1.9736642378070393e-05, "loss": 1.0579, "step": 525 }, { "epoch": 0.10123171670515782, "grad_norm": 1.8706365823745728, "learning_rate": 1.9735219372611232e-05, "loss": 1.0106, "step": 526 }, { "epoch": 0.10142417244033872, "grad_norm": 1.6723026037216187, "learning_rate": 1.9733792584604487e-05, "loss": 0.9984, "step": 527 }, { "epoch": 0.10161662817551963, "grad_norm": 1.8463131189346313, "learning_rate": 1.9732362014604515e-05, "loss": 1.1141, "step": 528 }, { "epoch": 0.10180908391070054, "grad_norm": 1.9170361757278442, "learning_rate": 1.973092766316716e-05, "loss": 1.0125, "step": 529 }, { "epoch": 0.10200153964588145, "grad_norm": 2.0003836154937744, "learning_rate": 1.972948953084973e-05, "loss": 0.9906, "step": 530 }, { "epoch": 0.10219399538106236, "grad_norm": 1.8767521381378174, "learning_rate": 1.9728047618210995e-05, "loss": 1.1314, "step": 531 }, { "epoch": 0.10238645111624327, "grad_norm": 1.8070870637893677, "learning_rate": 1.9726601925811204e-05, "loss": 1.1017, "step": 532 }, { "epoch": 0.10257890685142418, "grad_norm": 1.5967661142349243, "learning_rate": 1.9725152454212068e-05, "loss": 0.9788, "step": 533 }, { "epoch": 0.10277136258660508, "grad_norm": 1.8374614715576172, "learning_rate": 1.9723699203976768e-05, "loss": 1.1194, "step": 534 }, { "epoch": 0.10296381832178599, "grad_norm": 1.682363748550415, "learning_rate": 1.9722242175669955e-05, "loss": 1.0921, "step": 535 }, { "epoch": 0.1031562740569669, "grad_norm": 1.8782944679260254, "learning_rate": 1.9720781369857747e-05, "loss": 1.0255, "step": 536 }, { "epoch": 0.1033487297921478, "grad_norm": 1.7316479682922363, "learning_rate": 1.9719316787107728e-05, "loss": 1.1391, "step": 537 }, { "epoch": 0.10354118552732872, "grad_norm": 1.9749709367752075, "learning_rate": 1.9717848427988954e-05, "loss": 1.0727, "step": 538 }, { "epoch": 0.10373364126250963, "grad_norm": 1.855950117111206, "learning_rate": 1.971637629307194e-05, "loss": 0.9908, "step": 539 }, { "epoch": 0.10392609699769054, "grad_norm": 1.8045189380645752, "learning_rate": 1.9714900382928674e-05, "loss": 0.9219, "step": 540 }, { "epoch": 0.10411855273287143, "grad_norm": 1.9687528610229492, "learning_rate": 1.9713420698132614e-05, "loss": 1.0857, "step": 541 }, { "epoch": 0.10431100846805234, "grad_norm": 0.974544107913971, "learning_rate": 1.971193723925868e-05, "loss": 0.9757, "step": 542 }, { "epoch": 0.10450346420323325, "grad_norm": 1.768945336341858, "learning_rate": 1.9710450006883258e-05, "loss": 1.0488, "step": 543 }, { "epoch": 0.10469591993841416, "grad_norm": 0.9058737754821777, "learning_rate": 1.9708959001584197e-05, "loss": 0.8928, "step": 544 }, { "epoch": 0.10488837567359507, "grad_norm": 2.22385311126709, "learning_rate": 1.970746422394082e-05, "loss": 1.122, "step": 545 }, { "epoch": 0.10508083140877598, "grad_norm": 0.9252240061759949, "learning_rate": 1.970596567453391e-05, "loss": 0.9028, "step": 546 }, { "epoch": 0.1052732871439569, "grad_norm": 1.9228028059005737, "learning_rate": 1.9704463353945724e-05, "loss": 1.0755, "step": 547 }, { "epoch": 0.10546574287913779, "grad_norm": 2.069215774536133, "learning_rate": 1.9702957262759964e-05, "loss": 1.0918, "step": 548 }, { "epoch": 0.1056581986143187, "grad_norm": 2.1567542552948, "learning_rate": 1.9701447401561826e-05, "loss": 1.047, "step": 549 }, { "epoch": 0.10585065434949961, "grad_norm": 1.9342552423477173, "learning_rate": 1.9699933770937948e-05, "loss": 1.1015, "step": 550 }, { "epoch": 0.10604311008468052, "grad_norm": 1.9535177946090698, "learning_rate": 1.9698416371476434e-05, "loss": 0.9997, "step": 551 }, { "epoch": 0.10623556581986143, "grad_norm": 2.7614104747772217, "learning_rate": 1.969689520376687e-05, "loss": 1.0987, "step": 552 }, { "epoch": 0.10642802155504234, "grad_norm": 2.055088996887207, "learning_rate": 1.9695370268400288e-05, "loss": 1.1007, "step": 553 }, { "epoch": 0.10662047729022325, "grad_norm": 1.912510633468628, "learning_rate": 1.969384156596919e-05, "loss": 1.0789, "step": 554 }, { "epoch": 0.10681293302540416, "grad_norm": 1.903411626815796, "learning_rate": 1.9692309097067546e-05, "loss": 1.0581, "step": 555 }, { "epoch": 0.10700538876058506, "grad_norm": 1.8434637784957886, "learning_rate": 1.969077286229078e-05, "loss": 1.0157, "step": 556 }, { "epoch": 0.10719784449576597, "grad_norm": 2.0096569061279297, "learning_rate": 1.968923286223579e-05, "loss": 1.1297, "step": 557 }, { "epoch": 0.10739030023094688, "grad_norm": 2.015148878097534, "learning_rate": 1.9687689097500927e-05, "loss": 1.1228, "step": 558 }, { "epoch": 0.10758275596612779, "grad_norm": 2.0655734539031982, "learning_rate": 1.9686141568686015e-05, "loss": 1.0319, "step": 559 }, { "epoch": 0.1077752117013087, "grad_norm": 2.00614857673645, "learning_rate": 1.968459027639233e-05, "loss": 1.0363, "step": 560 }, { "epoch": 0.10796766743648961, "grad_norm": 1.9746959209442139, "learning_rate": 1.9683035221222617e-05, "loss": 1.117, "step": 561 }, { "epoch": 0.10816012317167052, "grad_norm": 1.7862404584884644, "learning_rate": 1.968147640378108e-05, "loss": 1.0235, "step": 562 }, { "epoch": 0.10835257890685142, "grad_norm": 2.0600969791412354, "learning_rate": 1.9679913824673388e-05, "loss": 1.0343, "step": 563 }, { "epoch": 0.10854503464203233, "grad_norm": 1.8007007837295532, "learning_rate": 1.9678347484506667e-05, "loss": 1.0694, "step": 564 }, { "epoch": 0.10873749037721324, "grad_norm": 1.9060628414154053, "learning_rate": 1.9676777383889506e-05, "loss": 1.0535, "step": 565 }, { "epoch": 0.10892994611239415, "grad_norm": 2.004653215408325, "learning_rate": 1.9675203523431964e-05, "loss": 1.1166, "step": 566 }, { "epoch": 0.10912240184757506, "grad_norm": 1.0712146759033203, "learning_rate": 1.9673625903745546e-05, "loss": 1.0181, "step": 567 }, { "epoch": 0.10931485758275597, "grad_norm": 2.315927505493164, "learning_rate": 1.9672044525443222e-05, "loss": 0.9983, "step": 568 }, { "epoch": 0.10950731331793688, "grad_norm": 1.891409158706665, "learning_rate": 1.9670459389139433e-05, "loss": 0.997, "step": 569 }, { "epoch": 0.10969976905311778, "grad_norm": 2.5097477436065674, "learning_rate": 1.9668870495450064e-05, "loss": 1.16, "step": 570 }, { "epoch": 0.10989222478829869, "grad_norm": 1.763094425201416, "learning_rate": 1.9667277844992476e-05, "loss": 1.0349, "step": 571 }, { "epoch": 0.1100846805234796, "grad_norm": 1.909881591796875, "learning_rate": 1.9665681438385475e-05, "loss": 0.9978, "step": 572 }, { "epoch": 0.11027713625866051, "grad_norm": 1.8126075267791748, "learning_rate": 1.9664081276249335e-05, "loss": 1.0984, "step": 573 }, { "epoch": 0.11046959199384142, "grad_norm": 2.0055270195007324, "learning_rate": 1.9662477359205786e-05, "loss": 1.0099, "step": 574 }, { "epoch": 0.11066204772902233, "grad_norm": 2.153329372406006, "learning_rate": 1.966086968787802e-05, "loss": 1.014, "step": 575 }, { "epoch": 0.11085450346420324, "grad_norm": 1.7271853685379028, "learning_rate": 1.9659258262890683e-05, "loss": 1.0863, "step": 576 }, { "epoch": 0.11104695919938413, "grad_norm": 1.9132412672042847, "learning_rate": 1.965764308486989e-05, "loss": 1.1253, "step": 577 }, { "epoch": 0.11123941493456505, "grad_norm": 1.6975780725479126, "learning_rate": 1.9656024154443195e-05, "loss": 1.13, "step": 578 }, { "epoch": 0.11143187066974596, "grad_norm": 1.930691123008728, "learning_rate": 1.965440147223963e-05, "loss": 1.0631, "step": 579 }, { "epoch": 0.11162432640492687, "grad_norm": 2.0817456245422363, "learning_rate": 1.9652775038889676e-05, "loss": 1.0652, "step": 580 }, { "epoch": 0.11181678214010778, "grad_norm": 1.8197139501571655, "learning_rate": 1.9651144855025265e-05, "loss": 1.103, "step": 581 }, { "epoch": 0.11200923787528869, "grad_norm": 1.7528233528137207, "learning_rate": 1.9649510921279803e-05, "loss": 1.0925, "step": 582 }, { "epoch": 0.1122016936104696, "grad_norm": 1.9419900178909302, "learning_rate": 1.964787323828813e-05, "loss": 1.0693, "step": 583 }, { "epoch": 0.11239414934565051, "grad_norm": 1.7717524766921997, "learning_rate": 1.9646231806686566e-05, "loss": 1.1371, "step": 584 }, { "epoch": 0.1125866050808314, "grad_norm": 1.75080406665802, "learning_rate": 1.9644586627112877e-05, "loss": 1.0661, "step": 585 }, { "epoch": 0.11277906081601231, "grad_norm": 1.903153896331787, "learning_rate": 1.964293770020628e-05, "loss": 1.1077, "step": 586 }, { "epoch": 0.11297151655119322, "grad_norm": 1.043135166168213, "learning_rate": 1.9641285026607453e-05, "loss": 0.9577, "step": 587 }, { "epoch": 0.11316397228637413, "grad_norm": 1.8222750425338745, "learning_rate": 1.9639628606958535e-05, "loss": 0.9664, "step": 588 }, { "epoch": 0.11335642802155504, "grad_norm": 2.1395421028137207, "learning_rate": 1.9637968441903112e-05, "loss": 0.9989, "step": 589 }, { "epoch": 0.11354888375673595, "grad_norm": 1.6749343872070312, "learning_rate": 1.963630453208623e-05, "loss": 1.0442, "step": 590 }, { "epoch": 0.11374133949191687, "grad_norm": 1.8441596031188965, "learning_rate": 1.9634636878154393e-05, "loss": 1.1011, "step": 591 }, { "epoch": 0.11393379522709776, "grad_norm": 1.7566505670547485, "learning_rate": 1.963296548075555e-05, "loss": 1.0822, "step": 592 }, { "epoch": 0.11412625096227867, "grad_norm": 1.7567025423049927, "learning_rate": 1.963129034053911e-05, "loss": 0.989, "step": 593 }, { "epoch": 0.11431870669745958, "grad_norm": 1.8125393390655518, "learning_rate": 1.9629611458155936e-05, "loss": 1.0389, "step": 594 }, { "epoch": 0.11451116243264049, "grad_norm": 1.9036537408828735, "learning_rate": 1.962792883425835e-05, "loss": 1.0655, "step": 595 }, { "epoch": 0.1147036181678214, "grad_norm": 2.057142734527588, "learning_rate": 1.962624246950012e-05, "loss": 1.0282, "step": 596 }, { "epoch": 0.11489607390300231, "grad_norm": 1.6065772771835327, "learning_rate": 1.9624552364536472e-05, "loss": 1.0344, "step": 597 }, { "epoch": 0.11508852963818322, "grad_norm": 0.9914276003837585, "learning_rate": 1.9622858520024087e-05, "loss": 0.9306, "step": 598 }, { "epoch": 0.11528098537336412, "grad_norm": 1.8407113552093506, "learning_rate": 1.9621160936621086e-05, "loss": 1.009, "step": 599 }, { "epoch": 0.11547344110854503, "grad_norm": 1.9886919260025024, "learning_rate": 1.961945961498706e-05, "loss": 1.0379, "step": 600 }, { "epoch": 0.11566589684372594, "grad_norm": 1.884535312652588, "learning_rate": 1.9617754555783045e-05, "loss": 1.0639, "step": 601 }, { "epoch": 0.11585835257890685, "grad_norm": 1.7920173406600952, "learning_rate": 1.9616045759671524e-05, "loss": 1.0109, "step": 602 }, { "epoch": 0.11605080831408776, "grad_norm": 1.819301962852478, "learning_rate": 1.9614333227316443e-05, "loss": 1.0452, "step": 603 }, { "epoch": 0.11624326404926867, "grad_norm": 1.9905906915664673, "learning_rate": 1.961261695938319e-05, "loss": 1.1223, "step": 604 }, { "epoch": 0.11643571978444958, "grad_norm": 1.6836289167404175, "learning_rate": 1.961089695653861e-05, "loss": 1.0391, "step": 605 }, { "epoch": 0.11662817551963048, "grad_norm": 1.6615184545516968, "learning_rate": 1.9609173219450998e-05, "loss": 1.0347, "step": 606 }, { "epoch": 0.11682063125481139, "grad_norm": 1.8804123401641846, "learning_rate": 1.9607445748790096e-05, "loss": 1.1475, "step": 607 }, { "epoch": 0.1170130869899923, "grad_norm": 0.9769880771636963, "learning_rate": 1.9605714545227105e-05, "loss": 0.9761, "step": 608 }, { "epoch": 0.11720554272517321, "grad_norm": 1.9053921699523926, "learning_rate": 1.9603979609434666e-05, "loss": 0.9815, "step": 609 }, { "epoch": 0.11739799846035412, "grad_norm": 1.7690304517745972, "learning_rate": 1.960224094208688e-05, "loss": 0.9917, "step": 610 }, { "epoch": 0.11759045419553503, "grad_norm": 1.8213820457458496, "learning_rate": 1.960049854385929e-05, "loss": 0.9546, "step": 611 }, { "epoch": 0.11778290993071594, "grad_norm": 1.846914529800415, "learning_rate": 1.9598752415428893e-05, "loss": 1.0954, "step": 612 }, { "epoch": 0.11797536566589685, "grad_norm": 1.7468945980072021, "learning_rate": 1.9597002557474136e-05, "loss": 1.036, "step": 613 }, { "epoch": 0.11816782140107775, "grad_norm": 1.8613026142120361, "learning_rate": 1.959524897067491e-05, "loss": 1.0865, "step": 614 }, { "epoch": 0.11836027713625866, "grad_norm": 1.9264206886291504, "learning_rate": 1.959349165571256e-05, "loss": 0.9994, "step": 615 }, { "epoch": 0.11855273287143957, "grad_norm": 1.689648985862732, "learning_rate": 1.9591730613269878e-05, "loss": 1.0364, "step": 616 }, { "epoch": 0.11874518860662048, "grad_norm": 2.080777645111084, "learning_rate": 1.9589965844031108e-05, "loss": 1.1069, "step": 617 }, { "epoch": 0.11893764434180139, "grad_norm": 1.9339449405670166, "learning_rate": 1.958819734868193e-05, "loss": 1.0624, "step": 618 }, { "epoch": 0.1191301000769823, "grad_norm": 1.9516017436981201, "learning_rate": 1.9586425127909488e-05, "loss": 0.9883, "step": 619 }, { "epoch": 0.11932255581216321, "grad_norm": 1.8677557706832886, "learning_rate": 1.9584649182402358e-05, "loss": 1.0618, "step": 620 }, { "epoch": 0.1195150115473441, "grad_norm": 1.9306186437606812, "learning_rate": 1.9582869512850576e-05, "loss": 1.0377, "step": 621 }, { "epoch": 0.11970746728252502, "grad_norm": 1.8844149112701416, "learning_rate": 1.9581086119945615e-05, "loss": 1.0273, "step": 622 }, { "epoch": 0.11989992301770593, "grad_norm": 1.7717770338058472, "learning_rate": 1.9579299004380402e-05, "loss": 1.0812, "step": 623 }, { "epoch": 0.12009237875288684, "grad_norm": 1.7238742113113403, "learning_rate": 1.9577508166849308e-05, "loss": 0.993, "step": 624 }, { "epoch": 0.12028483448806775, "grad_norm": 1.7734516859054565, "learning_rate": 1.9575713608048146e-05, "loss": 1.1625, "step": 625 }, { "epoch": 0.12047729022324866, "grad_norm": 1.9877711534500122, "learning_rate": 1.957391532867418e-05, "loss": 1.1511, "step": 626 }, { "epoch": 0.12066974595842957, "grad_norm": 1.9654442071914673, "learning_rate": 1.9572113329426123e-05, "loss": 1.0304, "step": 627 }, { "epoch": 0.12086220169361046, "grad_norm": 1.7001477479934692, "learning_rate": 1.9570307611004124e-05, "loss": 1.0758, "step": 628 }, { "epoch": 0.12105465742879137, "grad_norm": 1.8202993869781494, "learning_rate": 1.9568498174109783e-05, "loss": 1.0066, "step": 629 }, { "epoch": 0.12124711316397228, "grad_norm": 1.792492389678955, "learning_rate": 1.956668501944614e-05, "loss": 1.1291, "step": 630 }, { "epoch": 0.1214395688991532, "grad_norm": 2.0775163173675537, "learning_rate": 1.956486814771769e-05, "loss": 1.0128, "step": 631 }, { "epoch": 0.1216320246343341, "grad_norm": 1.7778408527374268, "learning_rate": 1.9563047559630356e-05, "loss": 1.0225, "step": 632 }, { "epoch": 0.12182448036951501, "grad_norm": 1.847807765007019, "learning_rate": 1.956122325589152e-05, "loss": 1.0558, "step": 633 }, { "epoch": 0.12201693610469592, "grad_norm": 1.7681615352630615, "learning_rate": 1.9559395237210002e-05, "loss": 1.0394, "step": 634 }, { "epoch": 0.12220939183987682, "grad_norm": 1.804386854171753, "learning_rate": 1.9557563504296064e-05, "loss": 1.1396, "step": 635 }, { "epoch": 0.12240184757505773, "grad_norm": 1.7099944353103638, "learning_rate": 1.955572805786141e-05, "loss": 1.0802, "step": 636 }, { "epoch": 0.12259430331023864, "grad_norm": 1.7447009086608887, "learning_rate": 1.955388889861919e-05, "loss": 1.0913, "step": 637 }, { "epoch": 0.12278675904541955, "grad_norm": 0.9787389636039734, "learning_rate": 1.9552046027284e-05, "loss": 0.9541, "step": 638 }, { "epoch": 0.12297921478060046, "grad_norm": 1.8490904569625854, "learning_rate": 1.9550199444571868e-05, "loss": 0.9432, "step": 639 }, { "epoch": 0.12317167051578137, "grad_norm": 2.0911710262298584, "learning_rate": 1.9548349151200275e-05, "loss": 1.0078, "step": 640 }, { "epoch": 0.12336412625096228, "grad_norm": 1.8651716709136963, "learning_rate": 1.9546495147888134e-05, "loss": 1.053, "step": 641 }, { "epoch": 0.1235565819861432, "grad_norm": 2.1007957458496094, "learning_rate": 1.954463743535581e-05, "loss": 1.0763, "step": 642 }, { "epoch": 0.12374903772132409, "grad_norm": 1.841106653213501, "learning_rate": 1.95427760143251e-05, "loss": 1.0845, "step": 643 }, { "epoch": 0.123941493456505, "grad_norm": 1.9699455499649048, "learning_rate": 1.9540910885519243e-05, "loss": 1.0461, "step": 644 }, { "epoch": 0.12413394919168591, "grad_norm": 1.774623155593872, "learning_rate": 1.9539042049662926e-05, "loss": 1.0162, "step": 645 }, { "epoch": 0.12432640492686682, "grad_norm": 1.8765196800231934, "learning_rate": 1.953716950748227e-05, "loss": 1.0493, "step": 646 }, { "epoch": 0.12451886066204773, "grad_norm": 1.9721708297729492, "learning_rate": 1.953529325970484e-05, "loss": 1.1348, "step": 647 }, { "epoch": 0.12471131639722864, "grad_norm": 1.871772050857544, "learning_rate": 1.953341330705963e-05, "loss": 1.0174, "step": 648 }, { "epoch": 0.12490377213240955, "grad_norm": 1.9706321954727173, "learning_rate": 1.9531529650277087e-05, "loss": 1.0637, "step": 649 }, { "epoch": 0.12509622786759045, "grad_norm": 2.1656877994537354, "learning_rate": 1.952964229008909e-05, "loss": 1.1391, "step": 650 }, { "epoch": 0.12528868360277137, "grad_norm": 1.806382656097412, "learning_rate": 1.9527751227228964e-05, "loss": 1.1237, "step": 651 }, { "epoch": 0.12548113933795227, "grad_norm": 1.9285237789154053, "learning_rate": 1.9525856462431463e-05, "loss": 1.0158, "step": 652 }, { "epoch": 0.1256735950731332, "grad_norm": 1.8296648263931274, "learning_rate": 1.9523957996432785e-05, "loss": 1.0418, "step": 653 }, { "epoch": 0.1258660508083141, "grad_norm": 1.9126784801483154, "learning_rate": 1.9522055829970564e-05, "loss": 1.1238, "step": 654 }, { "epoch": 0.12605850654349499, "grad_norm": 1.90383780002594, "learning_rate": 1.9520149963783873e-05, "loss": 1.023, "step": 655 }, { "epoch": 0.1262509622786759, "grad_norm": 1.9747416973114014, "learning_rate": 1.9518240398613226e-05, "loss": 1.0669, "step": 656 }, { "epoch": 0.1264434180138568, "grad_norm": 1.873603105545044, "learning_rate": 1.9516327135200565e-05, "loss": 0.9861, "step": 657 }, { "epoch": 0.12663587374903773, "grad_norm": 1.7834066152572632, "learning_rate": 1.951441017428928e-05, "loss": 1.1182, "step": 658 }, { "epoch": 0.12682832948421863, "grad_norm": 2.0852115154266357, "learning_rate": 1.951248951662418e-05, "loss": 0.8713, "step": 659 }, { "epoch": 0.12702078521939955, "grad_norm": 1.9413955211639404, "learning_rate": 1.9510565162951538e-05, "loss": 1.0676, "step": 660 }, { "epoch": 0.12721324095458045, "grad_norm": 1.8750298023223877, "learning_rate": 1.9508637114019037e-05, "loss": 1.082, "step": 661 }, { "epoch": 0.12740569668976134, "grad_norm": 1.7423945665359497, "learning_rate": 1.950670537057581e-05, "loss": 1.0576, "step": 662 }, { "epoch": 0.12759815242494227, "grad_norm": 1.818292260169983, "learning_rate": 1.9504769933372424e-05, "loss": 1.1287, "step": 663 }, { "epoch": 0.12779060816012316, "grad_norm": 1.8188986778259277, "learning_rate": 1.9502830803160873e-05, "loss": 0.9947, "step": 664 }, { "epoch": 0.1279830638953041, "grad_norm": 1.9943432807922363, "learning_rate": 1.9500887980694597e-05, "loss": 1.0222, "step": 665 }, { "epoch": 0.12817551963048499, "grad_norm": 1.9611740112304688, "learning_rate": 1.9498941466728462e-05, "loss": 1.0457, "step": 666 }, { "epoch": 0.1283679753656659, "grad_norm": 1.8990379571914673, "learning_rate": 1.949699126201877e-05, "loss": 0.9739, "step": 667 }, { "epoch": 0.1285604311008468, "grad_norm": 0.9880357980728149, "learning_rate": 1.9495037367323264e-05, "loss": 0.9853, "step": 668 }, { "epoch": 0.1287528868360277, "grad_norm": 1.750780463218689, "learning_rate": 1.9493079783401116e-05, "loss": 1.0149, "step": 669 }, { "epoch": 0.12894534257120863, "grad_norm": 1.894282341003418, "learning_rate": 1.949111851101292e-05, "loss": 0.976, "step": 670 }, { "epoch": 0.12913779830638952, "grad_norm": 1.6835464239120483, "learning_rate": 1.9489153550920726e-05, "loss": 1.0545, "step": 671 }, { "epoch": 0.12933025404157045, "grad_norm": 1.9264110326766968, "learning_rate": 1.9487184903888e-05, "loss": 1.0143, "step": 672 }, { "epoch": 0.12952270977675134, "grad_norm": 1.9531846046447754, "learning_rate": 1.9485212570679643e-05, "loss": 1.0893, "step": 673 }, { "epoch": 0.12971516551193227, "grad_norm": 1.9108635187149048, "learning_rate": 1.9483236552061996e-05, "loss": 0.9767, "step": 674 }, { "epoch": 0.12990762124711316, "grad_norm": 1.835564374923706, "learning_rate": 1.948125684880282e-05, "loss": 1.0638, "step": 675 }, { "epoch": 0.13010007698229406, "grad_norm": 1.7745845317840576, "learning_rate": 1.947927346167132e-05, "loss": 1.0006, "step": 676 }, { "epoch": 0.13029253271747498, "grad_norm": 1.862092137336731, "learning_rate": 1.9477286391438122e-05, "loss": 1.0371, "step": 677 }, { "epoch": 0.13048498845265588, "grad_norm": 2.0028343200683594, "learning_rate": 1.947529563887529e-05, "loss": 1.0369, "step": 678 }, { "epoch": 0.1306774441878368, "grad_norm": 1.8679611682891846, "learning_rate": 1.9473301204756315e-05, "loss": 1.0018, "step": 679 }, { "epoch": 0.1308698999230177, "grad_norm": 1.8468900918960571, "learning_rate": 1.947130308985612e-05, "loss": 1.0419, "step": 680 }, { "epoch": 0.13106235565819863, "grad_norm": 1.845431923866272, "learning_rate": 1.946930129495106e-05, "loss": 1.0513, "step": 681 }, { "epoch": 0.13125481139337952, "grad_norm": 1.8284912109375, "learning_rate": 1.9467295820818913e-05, "loss": 0.9758, "step": 682 }, { "epoch": 0.13144726712856042, "grad_norm": 2.0402565002441406, "learning_rate": 1.9465286668238895e-05, "loss": 1.1309, "step": 683 }, { "epoch": 0.13163972286374134, "grad_norm": 1.9071311950683594, "learning_rate": 1.9463273837991643e-05, "loss": 1.095, "step": 684 }, { "epoch": 0.13183217859892224, "grad_norm": 1.7187628746032715, "learning_rate": 1.9461257330859233e-05, "loss": 1.0461, "step": 685 }, { "epoch": 0.13202463433410316, "grad_norm": 1.9771403074264526, "learning_rate": 1.945923714762516e-05, "loss": 1.1084, "step": 686 }, { "epoch": 0.13221709006928406, "grad_norm": 1.8100978136062622, "learning_rate": 1.9457213289074353e-05, "loss": 0.9969, "step": 687 }, { "epoch": 0.13240954580446498, "grad_norm": 0.999468207359314, "learning_rate": 1.945518575599317e-05, "loss": 0.9531, "step": 688 }, { "epoch": 0.13260200153964588, "grad_norm": 1.7867693901062012, "learning_rate": 1.945315454916939e-05, "loss": 0.9845, "step": 689 }, { "epoch": 0.13279445727482678, "grad_norm": 2.030405044555664, "learning_rate": 1.945111966939222e-05, "loss": 1.0697, "step": 690 }, { "epoch": 0.1329869130100077, "grad_norm": 1.9287108182907104, "learning_rate": 1.9449081117452304e-05, "loss": 1.0837, "step": 691 }, { "epoch": 0.1331793687451886, "grad_norm": 1.7917439937591553, "learning_rate": 1.9447038894141707e-05, "loss": 1.079, "step": 692 }, { "epoch": 0.13337182448036952, "grad_norm": 1.6798985004425049, "learning_rate": 1.9444993000253912e-05, "loss": 1.0983, "step": 693 }, { "epoch": 0.13356428021555042, "grad_norm": 1.597659707069397, "learning_rate": 1.9442943436583845e-05, "loss": 1.085, "step": 694 }, { "epoch": 0.13375673595073134, "grad_norm": 1.7162896394729614, "learning_rate": 1.9440890203927846e-05, "loss": 1.0683, "step": 695 }, { "epoch": 0.13394919168591224, "grad_norm": 1.8342747688293457, "learning_rate": 1.9438833303083677e-05, "loss": 1.0327, "step": 696 }, { "epoch": 0.13414164742109314, "grad_norm": 1.6689069271087646, "learning_rate": 1.943677273485054e-05, "loss": 1.1206, "step": 697 }, { "epoch": 0.13433410315627406, "grad_norm": 1.8049540519714355, "learning_rate": 1.943470850002905e-05, "loss": 1.0248, "step": 698 }, { "epoch": 0.13452655889145496, "grad_norm": 1.0162373781204224, "learning_rate": 1.943264059942125e-05, "loss": 0.9161, "step": 699 }, { "epoch": 0.13471901462663588, "grad_norm": 1.958119511604309, "learning_rate": 1.9430569033830606e-05, "loss": 1.1096, "step": 700 }, { "epoch": 0.13491147036181678, "grad_norm": 0.9479613900184631, "learning_rate": 1.9428493804062013e-05, "loss": 0.8849, "step": 701 }, { "epoch": 0.1351039260969977, "grad_norm": 1.9936504364013672, "learning_rate": 1.9426414910921785e-05, "loss": 1.0403, "step": 702 }, { "epoch": 0.1352963818321786, "grad_norm": 1.726816177368164, "learning_rate": 1.942433235521766e-05, "loss": 1.0638, "step": 703 }, { "epoch": 0.1354888375673595, "grad_norm": 1.807843804359436, "learning_rate": 1.94222461377588e-05, "loss": 1.1033, "step": 704 }, { "epoch": 0.13568129330254042, "grad_norm": 1.8652249574661255, "learning_rate": 1.942015625935579e-05, "loss": 0.9907, "step": 705 }, { "epoch": 0.13587374903772131, "grad_norm": 1.7422491312026978, "learning_rate": 1.9418062720820636e-05, "loss": 1.0503, "step": 706 }, { "epoch": 0.13606620477290224, "grad_norm": 1.7758936882019043, "learning_rate": 1.9415965522966767e-05, "loss": 1.0422, "step": 707 }, { "epoch": 0.13625866050808313, "grad_norm": 2.0055336952209473, "learning_rate": 1.9413864666609036e-05, "loss": 1.0043, "step": 708 }, { "epoch": 0.13645111624326406, "grad_norm": 0.9495923519134521, "learning_rate": 1.941176015256371e-05, "loss": 0.9133, "step": 709 }, { "epoch": 0.13664357197844496, "grad_norm": 1.839900016784668, "learning_rate": 1.9409651981648486e-05, "loss": 0.9774, "step": 710 }, { "epoch": 0.13683602771362588, "grad_norm": 2.0253968238830566, "learning_rate": 1.9407540154682473e-05, "loss": 1.1266, "step": 711 }, { "epoch": 0.13702848344880678, "grad_norm": 1.7584038972854614, "learning_rate": 1.9405424672486218e-05, "loss": 0.9988, "step": 712 }, { "epoch": 0.13722093918398767, "grad_norm": 1.696969747543335, "learning_rate": 1.9403305535881664e-05, "loss": 1.0785, "step": 713 }, { "epoch": 0.1374133949191686, "grad_norm": 1.9572157859802246, "learning_rate": 1.940118274569219e-05, "loss": 1.0387, "step": 714 }, { "epoch": 0.1376058506543495, "grad_norm": 2.1646480560302734, "learning_rate": 1.9399056302742598e-05, "loss": 0.9452, "step": 715 }, { "epoch": 0.13779830638953042, "grad_norm": 2.0360054969787598, "learning_rate": 1.9396926207859085e-05, "loss": 1.0117, "step": 716 }, { "epoch": 0.1379907621247113, "grad_norm": 1.7679258584976196, "learning_rate": 1.9394792461869296e-05, "loss": 1.0578, "step": 717 }, { "epoch": 0.13818321785989224, "grad_norm": 1.8553813695907593, "learning_rate": 1.939265506560228e-05, "loss": 1.0105, "step": 718 }, { "epoch": 0.13837567359507313, "grad_norm": 1.863491177558899, "learning_rate": 1.9390514019888508e-05, "loss": 1.0122, "step": 719 }, { "epoch": 0.13856812933025403, "grad_norm": 2.064150094985962, "learning_rate": 1.9388369325559862e-05, "loss": 1.0633, "step": 720 }, { "epoch": 0.13876058506543495, "grad_norm": 2.0232086181640625, "learning_rate": 1.9386220983449652e-05, "loss": 1.133, "step": 721 }, { "epoch": 0.13895304080061585, "grad_norm": 1.7371606826782227, "learning_rate": 1.93840689943926e-05, "loss": 1.0152, "step": 722 }, { "epoch": 0.13914549653579678, "grad_norm": 1.7722785472869873, "learning_rate": 1.9381913359224844e-05, "loss": 1.0253, "step": 723 }, { "epoch": 0.13933795227097767, "grad_norm": 1.88107168674469, "learning_rate": 1.937975407878394e-05, "loss": 1.0857, "step": 724 }, { "epoch": 0.1395304080061586, "grad_norm": 1.9407353401184082, "learning_rate": 1.937759115390886e-05, "loss": 1.0026, "step": 725 }, { "epoch": 0.1397228637413395, "grad_norm": 1.8366038799285889, "learning_rate": 1.9375424585439994e-05, "loss": 0.995, "step": 726 }, { "epoch": 0.1399153194765204, "grad_norm": 1.9896798133850098, "learning_rate": 1.9373254374219146e-05, "loss": 1.0137, "step": 727 }, { "epoch": 0.1401077752117013, "grad_norm": 2.165299892425537, "learning_rate": 1.9371080521089536e-05, "loss": 0.9944, "step": 728 }, { "epoch": 0.1403002309468822, "grad_norm": 1.9276577234268188, "learning_rate": 1.9368903026895795e-05, "loss": 1.0369, "step": 729 }, { "epoch": 0.14049268668206313, "grad_norm": 2.167565107345581, "learning_rate": 1.9366721892483976e-05, "loss": 1.0448, "step": 730 }, { "epoch": 0.14068514241724403, "grad_norm": 1.7721621990203857, "learning_rate": 1.9364537118701542e-05, "loss": 1.0133, "step": 731 }, { "epoch": 0.14087759815242495, "grad_norm": 2.0589303970336914, "learning_rate": 1.9362348706397374e-05, "loss": 0.9684, "step": 732 }, { "epoch": 0.14107005388760585, "grad_norm": 1.9515771865844727, "learning_rate": 1.9360156656421756e-05, "loss": 1.1383, "step": 733 }, { "epoch": 0.14126250962278675, "grad_norm": 1.5488319396972656, "learning_rate": 1.9357960969626398e-05, "loss": 1.0224, "step": 734 }, { "epoch": 0.14145496535796767, "grad_norm": 1.8274539709091187, "learning_rate": 1.9355761646864417e-05, "loss": 1.0746, "step": 735 }, { "epoch": 0.14164742109314857, "grad_norm": 1.7872482538223267, "learning_rate": 1.935355868899034e-05, "loss": 1.0307, "step": 736 }, { "epoch": 0.1418398768283295, "grad_norm": 1.8236021995544434, "learning_rate": 1.935135209686012e-05, "loss": 0.9522, "step": 737 }, { "epoch": 0.1420323325635104, "grad_norm": 1.8238246440887451, "learning_rate": 1.93491418713311e-05, "loss": 1.1675, "step": 738 }, { "epoch": 0.1422247882986913, "grad_norm": 1.0203803777694702, "learning_rate": 1.9346928013262052e-05, "loss": 1.0319, "step": 739 }, { "epoch": 0.1424172440338722, "grad_norm": 1.9282431602478027, "learning_rate": 1.9344710523513157e-05, "loss": 0.9993, "step": 740 }, { "epoch": 0.1426096997690531, "grad_norm": 1.8842706680297852, "learning_rate": 1.9342489402945997e-05, "loss": 1.0621, "step": 741 }, { "epoch": 0.14280215550423403, "grad_norm": 1.772351861000061, "learning_rate": 1.9340264652423584e-05, "loss": 0.9429, "step": 742 }, { "epoch": 0.14299461123941493, "grad_norm": 1.8866921663284302, "learning_rate": 1.9338036272810314e-05, "loss": 1.1064, "step": 743 }, { "epoch": 0.14318706697459585, "grad_norm": 1.9042494297027588, "learning_rate": 1.9335804264972018e-05, "loss": 0.9967, "step": 744 }, { "epoch": 0.14337952270977675, "grad_norm": 1.964693546295166, "learning_rate": 1.933356862977592e-05, "loss": 1.0165, "step": 745 }, { "epoch": 0.14357197844495767, "grad_norm": 1.8669610023498535, "learning_rate": 1.9331329368090664e-05, "loss": 1.0561, "step": 746 }, { "epoch": 0.14376443418013857, "grad_norm": 1.8360422849655151, "learning_rate": 1.93290864807863e-05, "loss": 1.1563, "step": 747 }, { "epoch": 0.14395688991531946, "grad_norm": 1.8937022686004639, "learning_rate": 1.9326839968734278e-05, "loss": 1.0895, "step": 748 }, { "epoch": 0.1441493456505004, "grad_norm": 1.7559276819229126, "learning_rate": 1.9324589832807467e-05, "loss": 1.0141, "step": 749 }, { "epoch": 0.14434180138568128, "grad_norm": 1.801269769668579, "learning_rate": 1.9322336073880143e-05, "loss": 1.0842, "step": 750 }, { "epoch": 0.1445342571208622, "grad_norm": 1.855718970298767, "learning_rate": 1.932007869282799e-05, "loss": 1.0211, "step": 751 }, { "epoch": 0.1447267128560431, "grad_norm": 1.8507449626922607, "learning_rate": 1.9317817690528086e-05, "loss": 1.1162, "step": 752 }, { "epoch": 0.14491916859122403, "grad_norm": 1.0060285329818726, "learning_rate": 1.9315553067858933e-05, "loss": 0.9554, "step": 753 }, { "epoch": 0.14511162432640493, "grad_norm": 1.9339544773101807, "learning_rate": 1.931328482570044e-05, "loss": 1.163, "step": 754 }, { "epoch": 0.14530408006158582, "grad_norm": 1.7005418539047241, "learning_rate": 1.93110129649339e-05, "loss": 0.9483, "step": 755 }, { "epoch": 0.14549653579676675, "grad_norm": 1.756054401397705, "learning_rate": 1.9308737486442045e-05, "loss": 1.0619, "step": 756 }, { "epoch": 0.14568899153194764, "grad_norm": 1.9170414209365845, "learning_rate": 1.9306458391108984e-05, "loss": 1.1276, "step": 757 }, { "epoch": 0.14588144726712857, "grad_norm": 1.679626226425171, "learning_rate": 1.9304175679820247e-05, "loss": 0.9985, "step": 758 }, { "epoch": 0.14607390300230946, "grad_norm": 1.7036267518997192, "learning_rate": 1.9301889353462765e-05, "loss": 1.0142, "step": 759 }, { "epoch": 0.1462663587374904, "grad_norm": 1.7625948190689087, "learning_rate": 1.929959941292487e-05, "loss": 1.0215, "step": 760 }, { "epoch": 0.14645881447267128, "grad_norm": 0.9800795912742615, "learning_rate": 1.9297305859096305e-05, "loss": 0.9214, "step": 761 }, { "epoch": 0.14665127020785218, "grad_norm": 1.9728052616119385, "learning_rate": 1.929500869286821e-05, "loss": 1.0697, "step": 762 }, { "epoch": 0.1468437259430331, "grad_norm": 1.859421968460083, "learning_rate": 1.9292707915133134e-05, "loss": 1.1, "step": 763 }, { "epoch": 0.147036181678214, "grad_norm": 1.9221793413162231, "learning_rate": 1.9290403526785025e-05, "loss": 0.9823, "step": 764 }, { "epoch": 0.14722863741339492, "grad_norm": 1.963499903678894, "learning_rate": 1.9288095528719245e-05, "loss": 1.0665, "step": 765 }, { "epoch": 0.14742109314857582, "grad_norm": 2.0811595916748047, "learning_rate": 1.9285783921832537e-05, "loss": 1.1006, "step": 766 }, { "epoch": 0.14761354888375675, "grad_norm": 1.94386887550354, "learning_rate": 1.9283468707023068e-05, "loss": 1.0655, "step": 767 }, { "epoch": 0.14780600461893764, "grad_norm": 0.9830089807510376, "learning_rate": 1.928114988519039e-05, "loss": 0.9012, "step": 768 }, { "epoch": 0.14799846035411857, "grad_norm": 2.0522351264953613, "learning_rate": 1.9278827457235473e-05, "loss": 0.9684, "step": 769 }, { "epoch": 0.14819091608929946, "grad_norm": 1.7430166006088257, "learning_rate": 1.927650142406067e-05, "loss": 1.0186, "step": 770 }, { "epoch": 0.14838337182448036, "grad_norm": 1.981881856918335, "learning_rate": 1.927417178656975e-05, "loss": 1.0093, "step": 771 }, { "epoch": 0.14857582755966128, "grad_norm": 1.9995741844177246, "learning_rate": 1.9271838545667876e-05, "loss": 1.0438, "step": 772 }, { "epoch": 0.14876828329484218, "grad_norm": 1.7669248580932617, "learning_rate": 1.926950170226161e-05, "loss": 0.9754, "step": 773 }, { "epoch": 0.1489607390300231, "grad_norm": 1.790802240371704, "learning_rate": 1.9267161257258916e-05, "loss": 1.1056, "step": 774 }, { "epoch": 0.149153194765204, "grad_norm": 2.016782522201538, "learning_rate": 1.9264817211569154e-05, "loss": 1.0025, "step": 775 }, { "epoch": 0.14934565050038492, "grad_norm": 1.8168582916259766, "learning_rate": 1.926246956610309e-05, "loss": 0.9733, "step": 776 }, { "epoch": 0.14953810623556582, "grad_norm": 1.8395148515701294, "learning_rate": 1.926011832177288e-05, "loss": 1.0003, "step": 777 }, { "epoch": 0.14973056197074672, "grad_norm": 1.775355577468872, "learning_rate": 1.9257763479492086e-05, "loss": 1.0005, "step": 778 }, { "epoch": 0.14992301770592764, "grad_norm": 2.0468311309814453, "learning_rate": 1.9255405040175666e-05, "loss": 0.9266, "step": 779 }, { "epoch": 0.15011547344110854, "grad_norm": 1.788172960281372, "learning_rate": 1.9253043004739967e-05, "loss": 1.0404, "step": 780 }, { "epoch": 0.15030792917628946, "grad_norm": 1.8667372465133667, "learning_rate": 1.9250677374102752e-05, "loss": 1.0117, "step": 781 }, { "epoch": 0.15050038491147036, "grad_norm": 1.8807982206344604, "learning_rate": 1.9248308149183155e-05, "loss": 1.1426, "step": 782 }, { "epoch": 0.15069284064665128, "grad_norm": 1.8556149005889893, "learning_rate": 1.9245935330901733e-05, "loss": 1.0419, "step": 783 }, { "epoch": 0.15088529638183218, "grad_norm": 1.7991968393325806, "learning_rate": 1.9243558920180417e-05, "loss": 1.0587, "step": 784 }, { "epoch": 0.15107775211701308, "grad_norm": 1.0295864343643188, "learning_rate": 1.9241178917942552e-05, "loss": 0.9274, "step": 785 }, { "epoch": 0.151270207852194, "grad_norm": 1.7402684688568115, "learning_rate": 1.9238795325112867e-05, "loss": 0.9866, "step": 786 }, { "epoch": 0.1514626635873749, "grad_norm": 1.9365077018737793, "learning_rate": 1.923640814261749e-05, "loss": 1.0244, "step": 787 }, { "epoch": 0.15165511932255582, "grad_norm": 1.855178713798523, "learning_rate": 1.9234017371383946e-05, "loss": 1.1417, "step": 788 }, { "epoch": 0.15184757505773672, "grad_norm": 1.8731263875961304, "learning_rate": 1.9231623012341145e-05, "loss": 1.0281, "step": 789 }, { "epoch": 0.15204003079291764, "grad_norm": 1.8208426237106323, "learning_rate": 1.92292250664194e-05, "loss": 1.0159, "step": 790 }, { "epoch": 0.15223248652809854, "grad_norm": 1.7611134052276611, "learning_rate": 1.9226823534550418e-05, "loss": 1.0611, "step": 791 }, { "epoch": 0.15242494226327943, "grad_norm": 1.7592854499816895, "learning_rate": 1.9224418417667295e-05, "loss": 1.0512, "step": 792 }, { "epoch": 0.15261739799846036, "grad_norm": 1.9874253273010254, "learning_rate": 1.922200971670452e-05, "loss": 1.0744, "step": 793 }, { "epoch": 0.15280985373364125, "grad_norm": 1.8959243297576904, "learning_rate": 1.9219597432597975e-05, "loss": 1.0596, "step": 794 }, { "epoch": 0.15300230946882218, "grad_norm": 1.75994873046875, "learning_rate": 1.9217181566284942e-05, "loss": 1.0273, "step": 795 }, { "epoch": 0.15319476520400307, "grad_norm": 1.964096188545227, "learning_rate": 1.921476211870408e-05, "loss": 1.023, "step": 796 }, { "epoch": 0.153387220939184, "grad_norm": 1.9010738134384155, "learning_rate": 1.921233909079545e-05, "loss": 1.093, "step": 797 }, { "epoch": 0.1535796766743649, "grad_norm": 1.8385283946990967, "learning_rate": 1.92099124835005e-05, "loss": 1.0219, "step": 798 }, { "epoch": 0.1537721324095458, "grad_norm": 1.7191005945205688, "learning_rate": 1.9207482297762076e-05, "loss": 1.1217, "step": 799 }, { "epoch": 0.15396458814472672, "grad_norm": 1.8148480653762817, "learning_rate": 1.9205048534524405e-05, "loss": 1.0655, "step": 800 }, { "epoch": 0.1541570438799076, "grad_norm": 2.1281018257141113, "learning_rate": 1.9202611194733107e-05, "loss": 1.0402, "step": 801 }, { "epoch": 0.15434949961508854, "grad_norm": 1.8952546119689941, "learning_rate": 1.9200170279335194e-05, "loss": 1.0629, "step": 802 }, { "epoch": 0.15454195535026943, "grad_norm": 1.9646159410476685, "learning_rate": 1.9197725789279065e-05, "loss": 1.0021, "step": 803 }, { "epoch": 0.15473441108545036, "grad_norm": 1.8048688173294067, "learning_rate": 1.919527772551451e-05, "loss": 1.0693, "step": 804 }, { "epoch": 0.15492686682063125, "grad_norm": 1.8859809637069702, "learning_rate": 1.91928260889927e-05, "loss": 0.9513, "step": 805 }, { "epoch": 0.15511932255581215, "grad_norm": 1.968471646308899, "learning_rate": 1.9190370880666206e-05, "loss": 1.103, "step": 806 }, { "epoch": 0.15531177829099307, "grad_norm": 1.9892020225524902, "learning_rate": 1.9187912101488986e-05, "loss": 1.0569, "step": 807 }, { "epoch": 0.15550423402617397, "grad_norm": 1.8863886594772339, "learning_rate": 1.9185449752416367e-05, "loss": 1.137, "step": 808 }, { "epoch": 0.1556966897613549, "grad_norm": 1.9442975521087646, "learning_rate": 1.9182983834405088e-05, "loss": 1.0581, "step": 809 }, { "epoch": 0.1558891454965358, "grad_norm": 1.7951229810714722, "learning_rate": 1.918051434841326e-05, "loss": 0.9794, "step": 810 }, { "epoch": 0.15608160123171672, "grad_norm": 1.824581503868103, "learning_rate": 1.9178041295400383e-05, "loss": 0.9596, "step": 811 }, { "epoch": 0.1562740569668976, "grad_norm": 1.814587116241455, "learning_rate": 1.917556467632734e-05, "loss": 0.9584, "step": 812 }, { "epoch": 0.1564665127020785, "grad_norm": 1.9821945428848267, "learning_rate": 1.9173084492156406e-05, "loss": 0.9154, "step": 813 }, { "epoch": 0.15665896843725943, "grad_norm": 1.9843652248382568, "learning_rate": 1.917060074385124e-05, "loss": 1.1803, "step": 814 }, { "epoch": 0.15685142417244033, "grad_norm": 2.114295244216919, "learning_rate": 1.9168113432376885e-05, "loss": 1.0117, "step": 815 }, { "epoch": 0.15704387990762125, "grad_norm": 1.9193452596664429, "learning_rate": 1.9165622558699763e-05, "loss": 0.9599, "step": 816 }, { "epoch": 0.15723633564280215, "grad_norm": 2.21612286567688, "learning_rate": 1.9163128123787688e-05, "loss": 1.1139, "step": 817 }, { "epoch": 0.15742879137798307, "grad_norm": 1.6431894302368164, "learning_rate": 1.916063012860985e-05, "loss": 0.98, "step": 818 }, { "epoch": 0.15762124711316397, "grad_norm": 1.0864559412002563, "learning_rate": 1.9158128574136833e-05, "loss": 0.9225, "step": 819 }, { "epoch": 0.1578137028483449, "grad_norm": 2.1066713333129883, "learning_rate": 1.9155623461340594e-05, "loss": 1.0454, "step": 820 }, { "epoch": 0.1580061585835258, "grad_norm": 1.9968464374542236, "learning_rate": 1.9153114791194475e-05, "loss": 1.0646, "step": 821 }, { "epoch": 0.1581986143187067, "grad_norm": 1.7027695178985596, "learning_rate": 1.91506025646732e-05, "loss": 1.0267, "step": 822 }, { "epoch": 0.1583910700538876, "grad_norm": 1.8297191858291626, "learning_rate": 1.914808678275288e-05, "loss": 0.995, "step": 823 }, { "epoch": 0.1585835257890685, "grad_norm": 1.872512936592102, "learning_rate": 1.9145567446411003e-05, "loss": 0.9352, "step": 824 }, { "epoch": 0.15877598152424943, "grad_norm": 1.9383331537246704, "learning_rate": 1.9143044556626437e-05, "loss": 1.021, "step": 825 }, { "epoch": 0.15896843725943033, "grad_norm": 1.8862186670303345, "learning_rate": 1.9140518114379433e-05, "loss": 1.0426, "step": 826 }, { "epoch": 0.15916089299461125, "grad_norm": 1.9327762126922607, "learning_rate": 1.913798812065162e-05, "loss": 1.0066, "step": 827 }, { "epoch": 0.15935334872979215, "grad_norm": 1.7620855569839478, "learning_rate": 1.913545457642601e-05, "loss": 1.0, "step": 828 }, { "epoch": 0.15954580446497305, "grad_norm": 1.817885398864746, "learning_rate": 1.913291748268699e-05, "loss": 1.0615, "step": 829 }, { "epoch": 0.15973826020015397, "grad_norm": 1.7099207639694214, "learning_rate": 1.9130376840420337e-05, "loss": 1.0813, "step": 830 }, { "epoch": 0.15993071593533487, "grad_norm": 1.6124449968338013, "learning_rate": 1.912783265061319e-05, "loss": 1.0048, "step": 831 }, { "epoch": 0.1601231716705158, "grad_norm": 1.8943891525268555, "learning_rate": 1.912528491425408e-05, "loss": 1.0251, "step": 832 }, { "epoch": 0.1603156274056967, "grad_norm": 2.0991952419281006, "learning_rate": 1.9122733632332907e-05, "loss": 1.1226, "step": 833 }, { "epoch": 0.1605080831408776, "grad_norm": 1.8139768838882446, "learning_rate": 1.9120178805840956e-05, "loss": 1.0328, "step": 834 }, { "epoch": 0.1607005388760585, "grad_norm": 1.8370256423950195, "learning_rate": 1.911762043577089e-05, "loss": 0.9977, "step": 835 }, { "epoch": 0.1608929946112394, "grad_norm": 1.8669499158859253, "learning_rate": 1.9115058523116734e-05, "loss": 0.9854, "step": 836 }, { "epoch": 0.16108545034642033, "grad_norm": 1.6820703744888306, "learning_rate": 1.9112493068873907e-05, "loss": 1.0551, "step": 837 }, { "epoch": 0.16127790608160122, "grad_norm": 1.6766277551651, "learning_rate": 1.9109924074039198e-05, "loss": 1.103, "step": 838 }, { "epoch": 0.16147036181678215, "grad_norm": 1.718582034111023, "learning_rate": 1.910735153961077e-05, "loss": 1.0707, "step": 839 }, { "epoch": 0.16166281755196305, "grad_norm": 1.974519968032837, "learning_rate": 1.9104775466588162e-05, "loss": 1.0311, "step": 840 }, { "epoch": 0.16185527328714397, "grad_norm": 1.6999092102050781, "learning_rate": 1.9102195855972287e-05, "loss": 0.9859, "step": 841 }, { "epoch": 0.16204772902232487, "grad_norm": 1.9100570678710938, "learning_rate": 1.9099612708765432e-05, "loss": 1.1022, "step": 842 }, { "epoch": 0.16224018475750576, "grad_norm": 1.8535345792770386, "learning_rate": 1.9097026025971268e-05, "loss": 1.0511, "step": 843 }, { "epoch": 0.1624326404926867, "grad_norm": 1.8734996318817139, "learning_rate": 1.9094435808594823e-05, "loss": 1.0539, "step": 844 }, { "epoch": 0.16262509622786758, "grad_norm": 1.7165933847427368, "learning_rate": 1.909184205764251e-05, "loss": 1.0534, "step": 845 }, { "epoch": 0.1628175519630485, "grad_norm": 1.7317732572555542, "learning_rate": 1.908924477412211e-05, "loss": 0.9644, "step": 846 }, { "epoch": 0.1630100076982294, "grad_norm": 1.723581314086914, "learning_rate": 1.908664395904278e-05, "loss": 1.0719, "step": 847 }, { "epoch": 0.16320246343341033, "grad_norm": 1.9269397258758545, "learning_rate": 1.9084039613415052e-05, "loss": 1.0579, "step": 848 }, { "epoch": 0.16339491916859122, "grad_norm": 1.907837152481079, "learning_rate": 1.9081431738250815e-05, "loss": 1.108, "step": 849 }, { "epoch": 0.16358737490377212, "grad_norm": 1.8354381322860718, "learning_rate": 1.9078820334563346e-05, "loss": 1.0208, "step": 850 }, { "epoch": 0.16377983063895304, "grad_norm": 1.839447021484375, "learning_rate": 1.9076205403367287e-05, "loss": 1.0629, "step": 851 }, { "epoch": 0.16397228637413394, "grad_norm": 2.001133918762207, "learning_rate": 1.907358694567865e-05, "loss": 1.0667, "step": 852 }, { "epoch": 0.16416474210931487, "grad_norm": 1.8356646299362183, "learning_rate": 1.9070964962514816e-05, "loss": 1.0761, "step": 853 }, { "epoch": 0.16435719784449576, "grad_norm": 1.8363560438156128, "learning_rate": 1.9068339454894536e-05, "loss": 1.0059, "step": 854 }, { "epoch": 0.16454965357967669, "grad_norm": 1.9907485246658325, "learning_rate": 1.9065710423837936e-05, "loss": 1.0461, "step": 855 }, { "epoch": 0.16474210931485758, "grad_norm": 1.6925932168960571, "learning_rate": 1.9063077870366504e-05, "loss": 0.9761, "step": 856 }, { "epoch": 0.16493456505003848, "grad_norm": 1.841995120048523, "learning_rate": 1.9060441795503095e-05, "loss": 1.1256, "step": 857 }, { "epoch": 0.1651270207852194, "grad_norm": 1.8130391836166382, "learning_rate": 1.9057802200271943e-05, "loss": 1.0439, "step": 858 }, { "epoch": 0.1653194765204003, "grad_norm": 1.8468430042266846, "learning_rate": 1.9055159085698638e-05, "loss": 0.972, "step": 859 }, { "epoch": 0.16551193225558122, "grad_norm": 1.9957631826400757, "learning_rate": 1.905251245281015e-05, "loss": 1.0338, "step": 860 }, { "epoch": 0.16570438799076212, "grad_norm": 1.8274272680282593, "learning_rate": 1.90498623026348e-05, "loss": 1.0852, "step": 861 }, { "epoch": 0.16589684372594304, "grad_norm": 1.87775719165802, "learning_rate": 1.904720863620229e-05, "loss": 0.9864, "step": 862 }, { "epoch": 0.16608929946112394, "grad_norm": 1.8939504623413086, "learning_rate": 1.9044551454543683e-05, "loss": 0.9975, "step": 863 }, { "epoch": 0.16628175519630484, "grad_norm": 1.9896374940872192, "learning_rate": 1.9041890758691403e-05, "loss": 1.036, "step": 864 }, { "epoch": 0.16647421093148576, "grad_norm": 1.9141910076141357, "learning_rate": 1.9039226549679243e-05, "loss": 1.033, "step": 865 }, { "epoch": 0.16666666666666666, "grad_norm": 1.8497891426086426, "learning_rate": 1.903655882854237e-05, "loss": 1.0824, "step": 866 }, { "epoch": 0.16685912240184758, "grad_norm": 2.010185956954956, "learning_rate": 1.90338875963173e-05, "loss": 1.1061, "step": 867 }, { "epoch": 0.16705157813702848, "grad_norm": 1.9594199657440186, "learning_rate": 1.903121285404192e-05, "loss": 1.0955, "step": 868 }, { "epoch": 0.1672440338722094, "grad_norm": 1.7790595293045044, "learning_rate": 1.9028534602755485e-05, "loss": 1.0985, "step": 869 }, { "epoch": 0.1674364896073903, "grad_norm": 2.3293259143829346, "learning_rate": 1.902585284349861e-05, "loss": 0.9343, "step": 870 }, { "epoch": 0.1676289453425712, "grad_norm": 1.0938304662704468, "learning_rate": 1.9023167577313267e-05, "loss": 0.9271, "step": 871 }, { "epoch": 0.16782140107775212, "grad_norm": 1.8699318170547485, "learning_rate": 1.9020478805242803e-05, "loss": 0.9875, "step": 872 }, { "epoch": 0.16801385681293302, "grad_norm": 1.808601975440979, "learning_rate": 1.901778652833192e-05, "loss": 0.9839, "step": 873 }, { "epoch": 0.16820631254811394, "grad_norm": 1.60258150100708, "learning_rate": 1.901509074762668e-05, "loss": 0.9988, "step": 874 }, { "epoch": 0.16839876828329484, "grad_norm": 2.071719169616699, "learning_rate": 1.901239146417451e-05, "loss": 1.0161, "step": 875 }, { "epoch": 0.16859122401847576, "grad_norm": 1.7650020122528076, "learning_rate": 1.900968867902419e-05, "loss": 1.0098, "step": 876 }, { "epoch": 0.16878367975365666, "grad_norm": 1.8329802751541138, "learning_rate": 1.9006982393225878e-05, "loss": 1.0466, "step": 877 }, { "epoch": 0.16897613548883758, "grad_norm": 1.88319993019104, "learning_rate": 1.9004272607831077e-05, "loss": 1.0673, "step": 878 }, { "epoch": 0.16916859122401848, "grad_norm": 2.0712344646453857, "learning_rate": 1.9001559323892652e-05, "loss": 1.0881, "step": 879 }, { "epoch": 0.16936104695919937, "grad_norm": 1.769935131072998, "learning_rate": 1.8998842542464833e-05, "loss": 1.0519, "step": 880 }, { "epoch": 0.1695535026943803, "grad_norm": 1.8526180982589722, "learning_rate": 1.8996122264603202e-05, "loss": 1.031, "step": 881 }, { "epoch": 0.1697459584295612, "grad_norm": 1.8072012662887573, "learning_rate": 1.8993398491364702e-05, "loss": 1.1069, "step": 882 }, { "epoch": 0.16993841416474212, "grad_norm": 2.0149877071380615, "learning_rate": 1.899067122380764e-05, "loss": 1.0453, "step": 883 }, { "epoch": 0.17013086989992302, "grad_norm": 1.8691116571426392, "learning_rate": 1.8987940462991673e-05, "loss": 1.0435, "step": 884 }, { "epoch": 0.17032332563510394, "grad_norm": 2.2917675971984863, "learning_rate": 1.8985206209977813e-05, "loss": 0.9968, "step": 885 }, { "epoch": 0.17051578137028484, "grad_norm": 1.609276533126831, "learning_rate": 1.898246846582844e-05, "loss": 1.0338, "step": 886 }, { "epoch": 0.17070823710546573, "grad_norm": 1.957764744758606, "learning_rate": 1.897972723160729e-05, "loss": 1.0497, "step": 887 }, { "epoch": 0.17090069284064666, "grad_norm": 1.8542054891586304, "learning_rate": 1.8976982508379436e-05, "loss": 1.0472, "step": 888 }, { "epoch": 0.17109314857582755, "grad_norm": 2.0098819732666016, "learning_rate": 1.8974234297211327e-05, "loss": 1.0095, "step": 889 }, { "epoch": 0.17128560431100848, "grad_norm": 1.6726094484329224, "learning_rate": 1.897148259917076e-05, "loss": 1.0575, "step": 890 }, { "epoch": 0.17147806004618937, "grad_norm": 1.19387948513031, "learning_rate": 1.8968727415326885e-05, "loss": 0.9313, "step": 891 }, { "epoch": 0.1716705157813703, "grad_norm": 2.0022661685943604, "learning_rate": 1.896596874675021e-05, "loss": 1.0161, "step": 892 }, { "epoch": 0.1718629715165512, "grad_norm": 1.772822618484497, "learning_rate": 1.8963206594512593e-05, "loss": 1.0429, "step": 893 }, { "epoch": 0.1720554272517321, "grad_norm": 1.7540333271026611, "learning_rate": 1.8960440959687254e-05, "loss": 1.0505, "step": 894 }, { "epoch": 0.17224788298691301, "grad_norm": 1.7388813495635986, "learning_rate": 1.8957671843348754e-05, "loss": 1.1187, "step": 895 }, { "epoch": 0.1724403387220939, "grad_norm": 1.9682852029800415, "learning_rate": 1.895489924657301e-05, "loss": 1.0796, "step": 896 }, { "epoch": 0.17263279445727484, "grad_norm": 1.7046477794647217, "learning_rate": 1.8952123170437304e-05, "loss": 0.9628, "step": 897 }, { "epoch": 0.17282525019245573, "grad_norm": 1.7901122570037842, "learning_rate": 1.894934361602025e-05, "loss": 1.0176, "step": 898 }, { "epoch": 0.17301770592763666, "grad_norm": 1.7585915327072144, "learning_rate": 1.8946560584401833e-05, "loss": 1.0652, "step": 899 }, { "epoch": 0.17321016166281755, "grad_norm": 1.7957626581192017, "learning_rate": 1.8943774076663372e-05, "loss": 1.0247, "step": 900 }, { "epoch": 0.17340261739799845, "grad_norm": 1.6159627437591553, "learning_rate": 1.894098409388754e-05, "loss": 0.9572, "step": 901 }, { "epoch": 0.17359507313317937, "grad_norm": 1.7876954078674316, "learning_rate": 1.8938190637158377e-05, "loss": 1.0195, "step": 902 }, { "epoch": 0.17378752886836027, "grad_norm": 1.9027154445648193, "learning_rate": 1.893539370756125e-05, "loss": 1.0292, "step": 903 }, { "epoch": 0.1739799846035412, "grad_norm": 1.6486036777496338, "learning_rate": 1.893259330618289e-05, "loss": 1.144, "step": 904 }, { "epoch": 0.1741724403387221, "grad_norm": 2.005584716796875, "learning_rate": 1.892978943411137e-05, "loss": 0.9814, "step": 905 }, { "epoch": 0.17436489607390301, "grad_norm": 1.9639264345169067, "learning_rate": 1.8926982092436117e-05, "loss": 1.0134, "step": 906 }, { "epoch": 0.1745573518090839, "grad_norm": 1.887213110923767, "learning_rate": 1.89241712822479e-05, "loss": 1.0381, "step": 907 }, { "epoch": 0.1747498075442648, "grad_norm": 1.9802154302597046, "learning_rate": 1.8921357004638837e-05, "loss": 1.0286, "step": 908 }, { "epoch": 0.17494226327944573, "grad_norm": 1.821059226989746, "learning_rate": 1.8918539260702394e-05, "loss": 1.0876, "step": 909 }, { "epoch": 0.17513471901462663, "grad_norm": 1.7930537462234497, "learning_rate": 1.891571805153339e-05, "loss": 0.9766, "step": 910 }, { "epoch": 0.17532717474980755, "grad_norm": 1.7542009353637695, "learning_rate": 1.8912893378227984e-05, "loss": 0.9653, "step": 911 }, { "epoch": 0.17551963048498845, "grad_norm": 1.6874165534973145, "learning_rate": 1.891006524188368e-05, "loss": 1.0256, "step": 912 }, { "epoch": 0.17571208622016937, "grad_norm": 1.8666621446609497, "learning_rate": 1.890723364359933e-05, "loss": 1.0658, "step": 913 }, { "epoch": 0.17590454195535027, "grad_norm": 1.7374211549758911, "learning_rate": 1.890439858447513e-05, "loss": 1.0411, "step": 914 }, { "epoch": 0.17609699769053117, "grad_norm": 1.645519495010376, "learning_rate": 1.890156006561262e-05, "loss": 0.9538, "step": 915 }, { "epoch": 0.1762894534257121, "grad_norm": 1.8398423194885254, "learning_rate": 1.8898718088114688e-05, "loss": 1.0689, "step": 916 }, { "epoch": 0.17648190916089299, "grad_norm": 1.8389241695404053, "learning_rate": 1.8895872653085563e-05, "loss": 1.0464, "step": 917 }, { "epoch": 0.1766743648960739, "grad_norm": 2.086231231689453, "learning_rate": 1.8893023761630816e-05, "loss": 1.0614, "step": 918 }, { "epoch": 0.1768668206312548, "grad_norm": 1.8300734758377075, "learning_rate": 1.8890171414857366e-05, "loss": 1.0646, "step": 919 }, { "epoch": 0.17705927636643573, "grad_norm": 1.8184055089950562, "learning_rate": 1.888731561387347e-05, "loss": 1.1739, "step": 920 }, { "epoch": 0.17725173210161663, "grad_norm": 1.7660341262817383, "learning_rate": 1.8884456359788725e-05, "loss": 1.0628, "step": 921 }, { "epoch": 0.17744418783679752, "grad_norm": 1.0114213228225708, "learning_rate": 1.8881593653714075e-05, "loss": 0.9451, "step": 922 }, { "epoch": 0.17763664357197845, "grad_norm": 1.8874154090881348, "learning_rate": 1.8878727496761808e-05, "loss": 1.0482, "step": 923 }, { "epoch": 0.17782909930715934, "grad_norm": 1.9241492748260498, "learning_rate": 1.8875857890045544e-05, "loss": 1.0374, "step": 924 }, { "epoch": 0.17802155504234027, "grad_norm": 1.6878619194030762, "learning_rate": 1.8872984834680242e-05, "loss": 1.0077, "step": 925 }, { "epoch": 0.17821401077752116, "grad_norm": 1.9827361106872559, "learning_rate": 1.887010833178222e-05, "loss": 1.0752, "step": 926 }, { "epoch": 0.1784064665127021, "grad_norm": 1.8027729988098145, "learning_rate": 1.886722838246911e-05, "loss": 1.0553, "step": 927 }, { "epoch": 0.17859892224788299, "grad_norm": 1.835041880607605, "learning_rate": 1.8864344987859898e-05, "loss": 0.9973, "step": 928 }, { "epoch": 0.17879137798306388, "grad_norm": 1.9430558681488037, "learning_rate": 1.886145814907491e-05, "loss": 1.0732, "step": 929 }, { "epoch": 0.1789838337182448, "grad_norm": 1.938186764717102, "learning_rate": 1.88585678672358e-05, "loss": 1.0697, "step": 930 }, { "epoch": 0.1791762894534257, "grad_norm": 1.7199450731277466, "learning_rate": 1.8855674143465567e-05, "loss": 1.009, "step": 931 }, { "epoch": 0.17936874518860663, "grad_norm": 1.7696146965026855, "learning_rate": 1.885277697888855e-05, "loss": 1.0555, "step": 932 }, { "epoch": 0.17956120092378752, "grad_norm": 0.9714207649230957, "learning_rate": 1.884987637463042e-05, "loss": 0.9418, "step": 933 }, { "epoch": 0.17975365665896845, "grad_norm": 1.8125615119934082, "learning_rate": 1.8846972331818182e-05, "loss": 1.0621, "step": 934 }, { "epoch": 0.17994611239414934, "grad_norm": 1.8445812463760376, "learning_rate": 1.8844064851580182e-05, "loss": 1.0307, "step": 935 }, { "epoch": 0.18013856812933027, "grad_norm": 1.851665735244751, "learning_rate": 1.8841153935046098e-05, "loss": 0.9727, "step": 936 }, { "epoch": 0.18033102386451116, "grad_norm": 1.925013542175293, "learning_rate": 1.8838239583346948e-05, "loss": 1.1111, "step": 937 }, { "epoch": 0.18052347959969206, "grad_norm": 1.74274742603302, "learning_rate": 1.8835321797615083e-05, "loss": 0.9995, "step": 938 }, { "epoch": 0.18071593533487298, "grad_norm": 1.7983896732330322, "learning_rate": 1.8832400578984183e-05, "loss": 1.0525, "step": 939 }, { "epoch": 0.18090839107005388, "grad_norm": 1.9717377424240112, "learning_rate": 1.8829475928589272e-05, "loss": 0.9644, "step": 940 }, { "epoch": 0.1811008468052348, "grad_norm": 1.8048958778381348, "learning_rate": 1.8826547847566692e-05, "loss": 1.0454, "step": 941 }, { "epoch": 0.1812933025404157, "grad_norm": 1.7836039066314697, "learning_rate": 1.8823616337054136e-05, "loss": 1.0467, "step": 942 }, { "epoch": 0.18148575827559663, "grad_norm": 1.9517724514007568, "learning_rate": 1.8820681398190618e-05, "loss": 1.1058, "step": 943 }, { "epoch": 0.18167821401077752, "grad_norm": 1.874281644821167, "learning_rate": 1.8817743032116483e-05, "loss": 0.9662, "step": 944 }, { "epoch": 0.18187066974595842, "grad_norm": 1.9344921112060547, "learning_rate": 1.8814801239973413e-05, "loss": 0.9874, "step": 945 }, { "epoch": 0.18206312548113934, "grad_norm": 1.6398276090621948, "learning_rate": 1.8811856022904423e-05, "loss": 1.0832, "step": 946 }, { "epoch": 0.18225558121632024, "grad_norm": 1.0052623748779297, "learning_rate": 1.880890738205386e-05, "loss": 0.9568, "step": 947 }, { "epoch": 0.18244803695150116, "grad_norm": 1.9227954149246216, "learning_rate": 1.880595531856738e-05, "loss": 1.0276, "step": 948 }, { "epoch": 0.18264049268668206, "grad_norm": 1.767619252204895, "learning_rate": 1.8802999833592e-05, "loss": 0.998, "step": 949 }, { "epoch": 0.18283294842186298, "grad_norm": 0.9438467025756836, "learning_rate": 1.880004092827605e-05, "loss": 0.993, "step": 950 }, { "epoch": 0.18302540415704388, "grad_norm": 1.8609060049057007, "learning_rate": 1.8797078603769184e-05, "loss": 1.0433, "step": 951 }, { "epoch": 0.18321785989222478, "grad_norm": 1.75949227809906, "learning_rate": 1.87941128612224e-05, "loss": 1.0787, "step": 952 }, { "epoch": 0.1834103156274057, "grad_norm": 1.7850619554519653, "learning_rate": 1.8791143701788004e-05, "loss": 1.0621, "step": 953 }, { "epoch": 0.1836027713625866, "grad_norm": 1.5783555507659912, "learning_rate": 1.8788171126619653e-05, "loss": 1.1148, "step": 954 }, { "epoch": 0.18379522709776752, "grad_norm": 1.8430248498916626, "learning_rate": 1.8785195136872315e-05, "loss": 1.0273, "step": 955 }, { "epoch": 0.18398768283294842, "grad_norm": 1.79518723487854, "learning_rate": 1.8782215733702286e-05, "loss": 0.9435, "step": 956 }, { "epoch": 0.18418013856812934, "grad_norm": 1.7462091445922852, "learning_rate": 1.8779232918267194e-05, "loss": 0.9921, "step": 957 }, { "epoch": 0.18437259430331024, "grad_norm": 1.8421075344085693, "learning_rate": 1.8776246691725988e-05, "loss": 0.9917, "step": 958 }, { "epoch": 0.18456505003849114, "grad_norm": 1.698609709739685, "learning_rate": 1.8773257055238946e-05, "loss": 1.0652, "step": 959 }, { "epoch": 0.18475750577367206, "grad_norm": 1.9753444194793701, "learning_rate": 1.8770264009967667e-05, "loss": 1.0075, "step": 960 }, { "epoch": 0.18494996150885296, "grad_norm": 1.8233006000518799, "learning_rate": 1.876726755707508e-05, "loss": 0.989, "step": 961 }, { "epoch": 0.18514241724403388, "grad_norm": 1.813793420791626, "learning_rate": 1.8764267697725432e-05, "loss": 1.0366, "step": 962 }, { "epoch": 0.18533487297921478, "grad_norm": 1.7772347927093506, "learning_rate": 1.8761264433084294e-05, "loss": 1.0749, "step": 963 }, { "epoch": 0.1855273287143957, "grad_norm": 1.8173842430114746, "learning_rate": 1.8758257764318566e-05, "loss": 1.0448, "step": 964 }, { "epoch": 0.1857197844495766, "grad_norm": 1.8422237634658813, "learning_rate": 1.8755247692596467e-05, "loss": 0.9166, "step": 965 }, { "epoch": 0.1859122401847575, "grad_norm": 1.8423718214035034, "learning_rate": 1.8752234219087538e-05, "loss": 1.0763, "step": 966 }, { "epoch": 0.18610469591993842, "grad_norm": 2.0342793464660645, "learning_rate": 1.874921734496264e-05, "loss": 0.9247, "step": 967 }, { "epoch": 0.18629715165511931, "grad_norm": 1.8517146110534668, "learning_rate": 1.874619707139396e-05, "loss": 1.036, "step": 968 }, { "epoch": 0.18648960739030024, "grad_norm": 1.822220802307129, "learning_rate": 1.8743173399555e-05, "loss": 0.9724, "step": 969 }, { "epoch": 0.18668206312548113, "grad_norm": 1.901777982711792, "learning_rate": 1.8740146330620584e-05, "loss": 1.0395, "step": 970 }, { "epoch": 0.18687451886066206, "grad_norm": 1.7870514392852783, "learning_rate": 1.8737115865766865e-05, "loss": 1.04, "step": 971 }, { "epoch": 0.18706697459584296, "grad_norm": 1.8566724061965942, "learning_rate": 1.87340820061713e-05, "loss": 1.024, "step": 972 }, { "epoch": 0.18725943033102385, "grad_norm": 1.833913803100586, "learning_rate": 1.8731044753012674e-05, "loss": 1.0918, "step": 973 }, { "epoch": 0.18745188606620478, "grad_norm": 1.9312456846237183, "learning_rate": 1.872800410747109e-05, "loss": 0.9741, "step": 974 }, { "epoch": 0.18764434180138567, "grad_norm": 1.8630257844924927, "learning_rate": 1.8724960070727974e-05, "loss": 1.1241, "step": 975 }, { "epoch": 0.1878367975365666, "grad_norm": 0.979623556137085, "learning_rate": 1.8721912643966055e-05, "loss": 0.9371, "step": 976 }, { "epoch": 0.1880292532717475, "grad_norm": 1.0421589612960815, "learning_rate": 1.871886182836939e-05, "loss": 0.9364, "step": 977 }, { "epoch": 0.18822170900692842, "grad_norm": 1.923294186592102, "learning_rate": 1.871580762512336e-05, "loss": 1.0129, "step": 978 }, { "epoch": 0.1884141647421093, "grad_norm": 1.8897676467895508, "learning_rate": 1.871275003541464e-05, "loss": 1.0578, "step": 979 }, { "epoch": 0.1886066204772902, "grad_norm": 1.9939303398132324, "learning_rate": 1.8709689060431242e-05, "loss": 1.0375, "step": 980 }, { "epoch": 0.18879907621247113, "grad_norm": 1.75008225440979, "learning_rate": 1.8706624701362485e-05, "loss": 0.9813, "step": 981 }, { "epoch": 0.18899153194765203, "grad_norm": 1.8540035486221313, "learning_rate": 1.8703556959398998e-05, "loss": 1.0379, "step": 982 }, { "epoch": 0.18918398768283295, "grad_norm": 1.5888439416885376, "learning_rate": 1.8700485835732735e-05, "loss": 0.9546, "step": 983 }, { "epoch": 0.18937644341801385, "grad_norm": 1.9681564569473267, "learning_rate": 1.8697411331556958e-05, "loss": 1.1183, "step": 984 }, { "epoch": 0.18956889915319478, "grad_norm": 1.9214798212051392, "learning_rate": 1.8694333448066238e-05, "loss": 0.9255, "step": 985 }, { "epoch": 0.18976135488837567, "grad_norm": 1.9251784086227417, "learning_rate": 1.8691252186456465e-05, "loss": 0.9977, "step": 986 }, { "epoch": 0.18995381062355657, "grad_norm": 1.7150720357894897, "learning_rate": 1.8688167547924846e-05, "loss": 0.9714, "step": 987 }, { "epoch": 0.1901462663587375, "grad_norm": 1.9250214099884033, "learning_rate": 1.868507953366989e-05, "loss": 0.9933, "step": 988 }, { "epoch": 0.1903387220939184, "grad_norm": 1.750757098197937, "learning_rate": 1.8681988144891425e-05, "loss": 0.9743, "step": 989 }, { "epoch": 0.1905311778290993, "grad_norm": 1.8341478109359741, "learning_rate": 1.8678893382790583e-05, "loss": 1.0849, "step": 990 }, { "epoch": 0.1907236335642802, "grad_norm": 2.0515639781951904, "learning_rate": 1.8675795248569816e-05, "loss": 1.0745, "step": 991 }, { "epoch": 0.19091608929946113, "grad_norm": 1.0423712730407715, "learning_rate": 1.8672693743432875e-05, "loss": 0.8873, "step": 992 }, { "epoch": 0.19110854503464203, "grad_norm": 1.8037941455841064, "learning_rate": 1.8669588868584832e-05, "loss": 1.0347, "step": 993 }, { "epoch": 0.19130100076982295, "grad_norm": 2.019803285598755, "learning_rate": 1.866648062523206e-05, "loss": 1.1049, "step": 994 }, { "epoch": 0.19149345650500385, "grad_norm": 2.1451826095581055, "learning_rate": 1.866336901458225e-05, "loss": 1.0171, "step": 995 }, { "epoch": 0.19168591224018475, "grad_norm": 1.9621726274490356, "learning_rate": 1.866025403784439e-05, "loss": 1.03, "step": 996 }, { "epoch": 0.19187836797536567, "grad_norm": 1.7517309188842773, "learning_rate": 1.8657135696228777e-05, "loss": 0.981, "step": 997 }, { "epoch": 0.19207082371054657, "grad_norm": 1.7853367328643799, "learning_rate": 1.865401399094703e-05, "loss": 1.0487, "step": 998 }, { "epoch": 0.1922632794457275, "grad_norm": 2.146730422973633, "learning_rate": 1.8650888923212057e-05, "loss": 0.9861, "step": 999 }, { "epoch": 0.1924557351809084, "grad_norm": 0.9793761968612671, "learning_rate": 1.8647760494238082e-05, "loss": 0.9189, "step": 1000 }, { "epoch": 0.1926481909160893, "grad_norm": 1.868817687034607, "learning_rate": 1.8644628705240636e-05, "loss": 1.0055, "step": 1001 }, { "epoch": 0.1928406466512702, "grad_norm": 2.1687662601470947, "learning_rate": 1.864149355743655e-05, "loss": 1.0008, "step": 1002 }, { "epoch": 0.1930331023864511, "grad_norm": 1.9123890399932861, "learning_rate": 1.863835505204396e-05, "loss": 1.1333, "step": 1003 }, { "epoch": 0.19322555812163203, "grad_norm": 1.8504670858383179, "learning_rate": 1.8635213190282312e-05, "loss": 1.0287, "step": 1004 }, { "epoch": 0.19341801385681293, "grad_norm": 2.1333136558532715, "learning_rate": 1.863206797337235e-05, "loss": 1.009, "step": 1005 }, { "epoch": 0.19361046959199385, "grad_norm": 1.7836076021194458, "learning_rate": 1.862891940253613e-05, "loss": 0.9733, "step": 1006 }, { "epoch": 0.19380292532717475, "grad_norm": 0.9302992820739746, "learning_rate": 1.8625767478997003e-05, "loss": 0.9108, "step": 1007 }, { "epoch": 0.19399538106235567, "grad_norm": 1.951202154159546, "learning_rate": 1.8622612203979628e-05, "loss": 1.0476, "step": 1008 }, { "epoch": 0.19418783679753657, "grad_norm": 2.02200984954834, "learning_rate": 1.8619453578709957e-05, "loss": 0.9919, "step": 1009 }, { "epoch": 0.19438029253271746, "grad_norm": 1.8609297275543213, "learning_rate": 1.861629160441526e-05, "loss": 1.0626, "step": 1010 }, { "epoch": 0.1945727482678984, "grad_norm": 1.8526087999343872, "learning_rate": 1.8613126282324092e-05, "loss": 1.0691, "step": 1011 }, { "epoch": 0.19476520400307928, "grad_norm": 1.845171332359314, "learning_rate": 1.8609957613666316e-05, "loss": 1.021, "step": 1012 }, { "epoch": 0.1949576597382602, "grad_norm": 1.7719639539718628, "learning_rate": 1.8606785599673098e-05, "loss": 1.0055, "step": 1013 }, { "epoch": 0.1951501154734411, "grad_norm": 1.7800822257995605, "learning_rate": 1.86036102415769e-05, "loss": 1.0531, "step": 1014 }, { "epoch": 0.19534257120862203, "grad_norm": 2.0064759254455566, "learning_rate": 1.860043154061148e-05, "loss": 1.0525, "step": 1015 }, { "epoch": 0.19553502694380293, "grad_norm": 1.925990104675293, "learning_rate": 1.8597249498011906e-05, "loss": 1.0499, "step": 1016 }, { "epoch": 0.19572748267898382, "grad_norm": 1.6583619117736816, "learning_rate": 1.8594064115014528e-05, "loss": 1.0514, "step": 1017 }, { "epoch": 0.19591993841416475, "grad_norm": 2.0886902809143066, "learning_rate": 1.8590875392857008e-05, "loss": 1.0058, "step": 1018 }, { "epoch": 0.19611239414934564, "grad_norm": 1.8365750312805176, "learning_rate": 1.8587683332778305e-05, "loss": 1.0443, "step": 1019 }, { "epoch": 0.19630484988452657, "grad_norm": 1.7067655324935913, "learning_rate": 1.8584487936018663e-05, "loss": 1.0608, "step": 1020 }, { "epoch": 0.19649730561970746, "grad_norm": 3.241781711578369, "learning_rate": 1.858128920381963e-05, "loss": 1.0696, "step": 1021 }, { "epoch": 0.1966897613548884, "grad_norm": 2.1947684288024902, "learning_rate": 1.8578087137424057e-05, "loss": 0.9824, "step": 1022 }, { "epoch": 0.19688221709006928, "grad_norm": 1.9641575813293457, "learning_rate": 1.8574881738076076e-05, "loss": 1.0991, "step": 1023 }, { "epoch": 0.19707467282525018, "grad_norm": 1.8190679550170898, "learning_rate": 1.8571673007021124e-05, "loss": 1.0245, "step": 1024 }, { "epoch": 0.1972671285604311, "grad_norm": 1.632847785949707, "learning_rate": 1.856846094550593e-05, "loss": 1.0534, "step": 1025 }, { "epoch": 0.197459584295612, "grad_norm": 2.068267583847046, "learning_rate": 1.8565245554778516e-05, "loss": 0.9974, "step": 1026 }, { "epoch": 0.19765204003079292, "grad_norm": 1.9693275690078735, "learning_rate": 1.8562026836088198e-05, "loss": 1.0699, "step": 1027 }, { "epoch": 0.19784449576597382, "grad_norm": 2.0901858806610107, "learning_rate": 1.855880479068559e-05, "loss": 1.0121, "step": 1028 }, { "epoch": 0.19803695150115475, "grad_norm": 1.9260458946228027, "learning_rate": 1.855557941982258e-05, "loss": 1.0849, "step": 1029 }, { "epoch": 0.19822940723633564, "grad_norm": 1.7626837491989136, "learning_rate": 1.8552350724752383e-05, "loss": 1.0075, "step": 1030 }, { "epoch": 0.19842186297151654, "grad_norm": 1.8949304819107056, "learning_rate": 1.854911870672947e-05, "loss": 1.1387, "step": 1031 }, { "epoch": 0.19861431870669746, "grad_norm": 1.7402981519699097, "learning_rate": 1.854588336700962e-05, "loss": 1.0301, "step": 1032 }, { "epoch": 0.19880677444187836, "grad_norm": 1.6345536708831787, "learning_rate": 1.85426447068499e-05, "loss": 1.0361, "step": 1033 }, { "epoch": 0.19899923017705928, "grad_norm": 1.7676799297332764, "learning_rate": 1.853940272750867e-05, "loss": 0.9988, "step": 1034 }, { "epoch": 0.19919168591224018, "grad_norm": 1.7999532222747803, "learning_rate": 1.8536157430245577e-05, "loss": 1.0244, "step": 1035 }, { "epoch": 0.1993841416474211, "grad_norm": 1.8965489864349365, "learning_rate": 1.8532908816321557e-05, "loss": 1.0727, "step": 1036 }, { "epoch": 0.199576597382602, "grad_norm": 1.9054021835327148, "learning_rate": 1.8529656886998838e-05, "loss": 1.0685, "step": 1037 }, { "epoch": 0.1997690531177829, "grad_norm": 1.817887783050537, "learning_rate": 1.8526401643540924e-05, "loss": 1.1067, "step": 1038 }, { "epoch": 0.19996150885296382, "grad_norm": 1.9422781467437744, "learning_rate": 1.8523143087212623e-05, "loss": 0.975, "step": 1039 }, { "epoch": 0.20015396458814472, "grad_norm": 1.9313528537750244, "learning_rate": 1.851988121928002e-05, "loss": 1.0833, "step": 1040 }, { "epoch": 0.20034642032332564, "grad_norm": 1.6501028537750244, "learning_rate": 1.8516616041010495e-05, "loss": 1.0025, "step": 1041 }, { "epoch": 0.20053887605850654, "grad_norm": 1.608765959739685, "learning_rate": 1.85133475536727e-05, "loss": 1.0794, "step": 1042 }, { "epoch": 0.20073133179368746, "grad_norm": 1.7184313535690308, "learning_rate": 1.851007575853659e-05, "loss": 0.9518, "step": 1043 }, { "epoch": 0.20092378752886836, "grad_norm": 0.9287310838699341, "learning_rate": 1.8506800656873397e-05, "loss": 0.8879, "step": 1044 }, { "epoch": 0.20111624326404928, "grad_norm": 1.8343358039855957, "learning_rate": 1.850352224995563e-05, "loss": 1.0023, "step": 1045 }, { "epoch": 0.20130869899923018, "grad_norm": 1.8882182836532593, "learning_rate": 1.8500240539057093e-05, "loss": 1.0496, "step": 1046 }, { "epoch": 0.20150115473441108, "grad_norm": 1.8249081373214722, "learning_rate": 1.8496955525452877e-05, "loss": 0.9506, "step": 1047 }, { "epoch": 0.201693610469592, "grad_norm": 1.760257601737976, "learning_rate": 1.8493667210419337e-05, "loss": 1.0689, "step": 1048 }, { "epoch": 0.2018860662047729, "grad_norm": 2.1494243144989014, "learning_rate": 1.8490375595234135e-05, "loss": 0.9769, "step": 1049 }, { "epoch": 0.20207852193995382, "grad_norm": 0.9508609771728516, "learning_rate": 1.8487080681176195e-05, "loss": 0.9155, "step": 1050 }, { "epoch": 0.20227097767513472, "grad_norm": 1.8663274049758911, "learning_rate": 1.848378246952574e-05, "loss": 1.0206, "step": 1051 }, { "epoch": 0.20246343341031564, "grad_norm": 1.6670690774917603, "learning_rate": 1.848048096156426e-05, "loss": 1.0712, "step": 1052 }, { "epoch": 0.20265588914549654, "grad_norm": 1.9200055599212646, "learning_rate": 1.8477176158574533e-05, "loss": 1.023, "step": 1053 }, { "epoch": 0.20284834488067743, "grad_norm": 0.9380425810813904, "learning_rate": 1.8473868061840617e-05, "loss": 0.9584, "step": 1054 }, { "epoch": 0.20304080061585836, "grad_norm": 1.8937983512878418, "learning_rate": 1.847055667264785e-05, "loss": 0.994, "step": 1055 }, { "epoch": 0.20323325635103925, "grad_norm": 0.9107078909873962, "learning_rate": 1.8467241992282842e-05, "loss": 0.9207, "step": 1056 }, { "epoch": 0.20342571208622018, "grad_norm": 1.8114209175109863, "learning_rate": 1.8463924022033492e-05, "loss": 0.9767, "step": 1057 }, { "epoch": 0.20361816782140107, "grad_norm": 1.8540204763412476, "learning_rate": 1.8460602763188974e-05, "loss": 1.0149, "step": 1058 }, { "epoch": 0.203810623556582, "grad_norm": 1.8666528463363647, "learning_rate": 1.8457278217039735e-05, "loss": 1.0119, "step": 1059 }, { "epoch": 0.2040030792917629, "grad_norm": 1.8038444519042969, "learning_rate": 1.8453950384877504e-05, "loss": 1.0667, "step": 1060 }, { "epoch": 0.2041955350269438, "grad_norm": 1.9152212142944336, "learning_rate": 1.8450619267995283e-05, "loss": 1.0885, "step": 1061 }, { "epoch": 0.20438799076212472, "grad_norm": 1.750767707824707, "learning_rate": 1.844728486768736e-05, "loss": 1.0792, "step": 1062 }, { "epoch": 0.2045804464973056, "grad_norm": 1.5561922788619995, "learning_rate": 1.8443947185249287e-05, "loss": 0.9974, "step": 1063 }, { "epoch": 0.20477290223248654, "grad_norm": 1.6789472103118896, "learning_rate": 1.8440606221977893e-05, "loss": 0.9861, "step": 1064 }, { "epoch": 0.20496535796766743, "grad_norm": 1.8870909214019775, "learning_rate": 1.8437261979171294e-05, "loss": 0.9904, "step": 1065 }, { "epoch": 0.20515781370284836, "grad_norm": 1.9678127765655518, "learning_rate": 1.843391445812886e-05, "loss": 0.9707, "step": 1066 }, { "epoch": 0.20535026943802925, "grad_norm": 1.8673638105392456, "learning_rate": 1.843056366015125e-05, "loss": 1.0365, "step": 1067 }, { "epoch": 0.20554272517321015, "grad_norm": 1.6730300188064575, "learning_rate": 1.8427209586540392e-05, "loss": 1.115, "step": 1068 }, { "epoch": 0.20573518090839107, "grad_norm": 1.9282186031341553, "learning_rate": 1.8423852238599488e-05, "loss": 1.1257, "step": 1069 }, { "epoch": 0.20592763664357197, "grad_norm": 1.8388830423355103, "learning_rate": 1.842049161763301e-05, "loss": 1.0134, "step": 1070 }, { "epoch": 0.2061200923787529, "grad_norm": 1.644269347190857, "learning_rate": 1.84171277249467e-05, "loss": 1.0023, "step": 1071 }, { "epoch": 0.2063125481139338, "grad_norm": 1.7801090478897095, "learning_rate": 1.841376056184758e-05, "loss": 1.0929, "step": 1072 }, { "epoch": 0.20650500384911472, "grad_norm": 1.9444172382354736, "learning_rate": 1.8410390129643927e-05, "loss": 0.9899, "step": 1073 }, { "epoch": 0.2066974595842956, "grad_norm": 1.7458233833312988, "learning_rate": 1.8407016429645305e-05, "loss": 0.9575, "step": 1074 }, { "epoch": 0.2068899153194765, "grad_norm": 2.0225753784179688, "learning_rate": 1.8403639463162536e-05, "loss": 1.1052, "step": 1075 }, { "epoch": 0.20708237105465743, "grad_norm": 1.6574417352676392, "learning_rate": 1.8400259231507716e-05, "loss": 1.0638, "step": 1076 }, { "epoch": 0.20727482678983833, "grad_norm": 1.6813980340957642, "learning_rate": 1.839687573599421e-05, "loss": 1.0221, "step": 1077 }, { "epoch": 0.20746728252501925, "grad_norm": 1.9149795770645142, "learning_rate": 1.8393488977936654e-05, "loss": 1.0771, "step": 1078 }, { "epoch": 0.20765973826020015, "grad_norm": 1.6382049322128296, "learning_rate": 1.839009895865094e-05, "loss": 1.0251, "step": 1079 }, { "epoch": 0.20785219399538107, "grad_norm": 1.9716051816940308, "learning_rate": 1.8386705679454243e-05, "loss": 1.0837, "step": 1080 }, { "epoch": 0.20804464973056197, "grad_norm": 2.0258374214172363, "learning_rate": 1.8383309141664992e-05, "loss": 1.1278, "step": 1081 }, { "epoch": 0.20823710546574287, "grad_norm": 1.6770477294921875, "learning_rate": 1.837990934660288e-05, "loss": 1.063, "step": 1082 }, { "epoch": 0.2084295612009238, "grad_norm": 1.7865912914276123, "learning_rate": 1.8376506295588887e-05, "loss": 1.0724, "step": 1083 }, { "epoch": 0.2086220169361047, "grad_norm": 2.0237650871276855, "learning_rate": 1.8373099989945236e-05, "loss": 0.9722, "step": 1084 }, { "epoch": 0.2088144726712856, "grad_norm": 1.9553661346435547, "learning_rate": 1.836969043099542e-05, "loss": 1.026, "step": 1085 }, { "epoch": 0.2090069284064665, "grad_norm": 1.7831957340240479, "learning_rate": 1.83662776200642e-05, "loss": 1.0627, "step": 1086 }, { "epoch": 0.20919938414164743, "grad_norm": 1.9168123006820679, "learning_rate": 1.8362861558477597e-05, "loss": 1.0774, "step": 1087 }, { "epoch": 0.20939183987682833, "grad_norm": 1.8233869075775146, "learning_rate": 1.8359442247562896e-05, "loss": 1.0233, "step": 1088 }, { "epoch": 0.20958429561200922, "grad_norm": 2.010859727859497, "learning_rate": 1.835601968864865e-05, "loss": 0.8869, "step": 1089 }, { "epoch": 0.20977675134719015, "grad_norm": 1.8921481370925903, "learning_rate": 1.8352593883064664e-05, "loss": 1.0672, "step": 1090 }, { "epoch": 0.20996920708237105, "grad_norm": 1.8805242776870728, "learning_rate": 1.8349164832142015e-05, "loss": 1.0628, "step": 1091 }, { "epoch": 0.21016166281755197, "grad_norm": 1.7704228162765503, "learning_rate": 1.834573253721303e-05, "loss": 1.0324, "step": 1092 }, { "epoch": 0.21035411855273287, "grad_norm": 1.6177185773849487, "learning_rate": 1.83422969996113e-05, "loss": 1.0192, "step": 1093 }, { "epoch": 0.2105465742879138, "grad_norm": 1.7440119981765747, "learning_rate": 1.8338858220671683e-05, "loss": 1.0338, "step": 1094 }, { "epoch": 0.2107390300230947, "grad_norm": 1.7780840396881104, "learning_rate": 1.833541620173029e-05, "loss": 0.9533, "step": 1095 }, { "epoch": 0.21093148575827558, "grad_norm": 1.9631036520004272, "learning_rate": 1.833197094412449e-05, "loss": 1.1008, "step": 1096 }, { "epoch": 0.2111239414934565, "grad_norm": 1.7581381797790527, "learning_rate": 1.8328522449192915e-05, "loss": 1.0777, "step": 1097 }, { "epoch": 0.2113163972286374, "grad_norm": 1.6984455585479736, "learning_rate": 1.8325070718275452e-05, "loss": 1.0441, "step": 1098 }, { "epoch": 0.21150885296381833, "grad_norm": 1.8124308586120605, "learning_rate": 1.8321615752713242e-05, "loss": 0.9438, "step": 1099 }, { "epoch": 0.21170130869899922, "grad_norm": 1.8689213991165161, "learning_rate": 1.8318157553848694e-05, "loss": 0.9867, "step": 1100 }, { "epoch": 0.21189376443418015, "grad_norm": 1.8502029180526733, "learning_rate": 1.8314696123025456e-05, "loss": 1.0766, "step": 1101 }, { "epoch": 0.21208622016936104, "grad_norm": 1.9622750282287598, "learning_rate": 1.8311231461588446e-05, "loss": 1.0017, "step": 1102 }, { "epoch": 0.21227867590454197, "grad_norm": 1.793318748474121, "learning_rate": 1.830776357088383e-05, "loss": 0.9984, "step": 1103 }, { "epoch": 0.21247113163972287, "grad_norm": 1.8471704721450806, "learning_rate": 1.8304292452259037e-05, "loss": 1.0222, "step": 1104 }, { "epoch": 0.21266358737490376, "grad_norm": 1.885473608970642, "learning_rate": 1.8300818107062744e-05, "loss": 1.0161, "step": 1105 }, { "epoch": 0.2128560431100847, "grad_norm": 1.7097712755203247, "learning_rate": 1.8297340536644877e-05, "loss": 1.0857, "step": 1106 }, { "epoch": 0.21304849884526558, "grad_norm": 1.632116436958313, "learning_rate": 1.8293859742356625e-05, "loss": 1.0589, "step": 1107 }, { "epoch": 0.2132409545804465, "grad_norm": 1.7865877151489258, "learning_rate": 1.8290375725550417e-05, "loss": 0.9824, "step": 1108 }, { "epoch": 0.2134334103156274, "grad_norm": 1.8652921915054321, "learning_rate": 1.8286888487579954e-05, "loss": 0.961, "step": 1109 }, { "epoch": 0.21362586605080833, "grad_norm": 1.826355218887329, "learning_rate": 1.8283398029800167e-05, "loss": 1.0006, "step": 1110 }, { "epoch": 0.21381832178598922, "grad_norm": 1.707168698310852, "learning_rate": 1.827990435356725e-05, "loss": 0.9484, "step": 1111 }, { "epoch": 0.21401077752117012, "grad_norm": 1.7456754446029663, "learning_rate": 1.827640746023865e-05, "loss": 0.9068, "step": 1112 }, { "epoch": 0.21420323325635104, "grad_norm": 1.827331304550171, "learning_rate": 1.827290735117305e-05, "loss": 1.0244, "step": 1113 }, { "epoch": 0.21439568899153194, "grad_norm": 1.875224232673645, "learning_rate": 1.82694040277304e-05, "loss": 1.0135, "step": 1114 }, { "epoch": 0.21458814472671286, "grad_norm": 1.7828434705734253, "learning_rate": 1.8265897491271885e-05, "loss": 1.0461, "step": 1115 }, { "epoch": 0.21478060046189376, "grad_norm": 1.8433310985565186, "learning_rate": 1.826238774315995e-05, "loss": 0.8909, "step": 1116 }, { "epoch": 0.21497305619707469, "grad_norm": 1.804031252861023, "learning_rate": 1.8258874784758276e-05, "loss": 1.014, "step": 1117 }, { "epoch": 0.21516551193225558, "grad_norm": 1.7992842197418213, "learning_rate": 1.82553586174318e-05, "loss": 1.0486, "step": 1118 }, { "epoch": 0.21535796766743648, "grad_norm": 1.848079800605774, "learning_rate": 1.8251839242546696e-05, "loss": 1.0638, "step": 1119 }, { "epoch": 0.2155504234026174, "grad_norm": 1.8478999137878418, "learning_rate": 1.82483166614704e-05, "loss": 0.9732, "step": 1120 }, { "epoch": 0.2157428791377983, "grad_norm": 1.8616169691085815, "learning_rate": 1.8244790875571582e-05, "loss": 1.0612, "step": 1121 }, { "epoch": 0.21593533487297922, "grad_norm": 1.9023991823196411, "learning_rate": 1.8241261886220155e-05, "loss": 1.0503, "step": 1122 }, { "epoch": 0.21612779060816012, "grad_norm": 1.1413387060165405, "learning_rate": 1.8237729694787295e-05, "loss": 0.9433, "step": 1123 }, { "epoch": 0.21632024634334104, "grad_norm": 1.787695288658142, "learning_rate": 1.8234194302645393e-05, "loss": 1.027, "step": 1124 }, { "epoch": 0.21651270207852194, "grad_norm": 1.963045358657837, "learning_rate": 1.8230655711168112e-05, "loss": 0.9202, "step": 1125 }, { "epoch": 0.21670515781370284, "grad_norm": 1.711930274963379, "learning_rate": 1.8227113921730336e-05, "loss": 1.0207, "step": 1126 }, { "epoch": 0.21689761354888376, "grad_norm": 0.9423871636390686, "learning_rate": 1.822356893570821e-05, "loss": 0.9479, "step": 1127 }, { "epoch": 0.21709006928406466, "grad_norm": 1.8183504343032837, "learning_rate": 1.8220020754479104e-05, "loss": 0.9918, "step": 1128 }, { "epoch": 0.21728252501924558, "grad_norm": 1.918871521949768, "learning_rate": 1.821646937942164e-05, "loss": 1.1492, "step": 1129 }, { "epoch": 0.21747498075442648, "grad_norm": 2.0717124938964844, "learning_rate": 1.821291481191568e-05, "loss": 0.898, "step": 1130 }, { "epoch": 0.2176674364896074, "grad_norm": 1.7555091381072998, "learning_rate": 1.8209357053342325e-05, "loss": 1.0465, "step": 1131 }, { "epoch": 0.2178598922247883, "grad_norm": 1.7912899255752563, "learning_rate": 1.8205796105083917e-05, "loss": 1.0205, "step": 1132 }, { "epoch": 0.2180523479599692, "grad_norm": 1.930643081665039, "learning_rate": 1.8202231968524033e-05, "loss": 0.9695, "step": 1133 }, { "epoch": 0.21824480369515012, "grad_norm": 1.831152081489563, "learning_rate": 1.8198664645047496e-05, "loss": 0.9631, "step": 1134 }, { "epoch": 0.21843725943033102, "grad_norm": 1.735793113708496, "learning_rate": 1.8195094136040357e-05, "loss": 1.0294, "step": 1135 }, { "epoch": 0.21862971516551194, "grad_norm": 1.81747567653656, "learning_rate": 1.819152044288992e-05, "loss": 0.9889, "step": 1136 }, { "epoch": 0.21882217090069284, "grad_norm": 1.8590980768203735, "learning_rate": 1.818794356698471e-05, "loss": 1.0412, "step": 1137 }, { "epoch": 0.21901462663587376, "grad_norm": 2.285672903060913, "learning_rate": 1.8184363509714494e-05, "loss": 0.9631, "step": 1138 }, { "epoch": 0.21920708237105466, "grad_norm": 2.0010228157043457, "learning_rate": 1.818078027247028e-05, "loss": 1.0693, "step": 1139 }, { "epoch": 0.21939953810623555, "grad_norm": 1.8023267984390259, "learning_rate": 1.8177193856644315e-05, "loss": 0.9676, "step": 1140 }, { "epoch": 0.21959199384141648, "grad_norm": 1.6251740455627441, "learning_rate": 1.8173604263630066e-05, "loss": 0.9698, "step": 1141 }, { "epoch": 0.21978444957659737, "grad_norm": 1.848659634590149, "learning_rate": 1.8170011494822242e-05, "loss": 0.9922, "step": 1142 }, { "epoch": 0.2199769053117783, "grad_norm": 1.7886439561843872, "learning_rate": 1.816641555161679e-05, "loss": 0.9706, "step": 1143 }, { "epoch": 0.2201693610469592, "grad_norm": 1.7670153379440308, "learning_rate": 1.8162816435410892e-05, "loss": 1.1177, "step": 1144 }, { "epoch": 0.22036181678214012, "grad_norm": 1.720010757446289, "learning_rate": 1.8159214147602947e-05, "loss": 0.9645, "step": 1145 }, { "epoch": 0.22055427251732102, "grad_norm": 1.9693390130996704, "learning_rate": 1.8155608689592604e-05, "loss": 1.0013, "step": 1146 }, { "epoch": 0.2207467282525019, "grad_norm": 1.756457805633545, "learning_rate": 1.8152000062780735e-05, "loss": 0.9904, "step": 1147 }, { "epoch": 0.22093918398768284, "grad_norm": 1.805774211883545, "learning_rate": 1.8148388268569453e-05, "loss": 0.9492, "step": 1148 }, { "epoch": 0.22113163972286373, "grad_norm": 1.7399941682815552, "learning_rate": 1.814477330836208e-05, "loss": 1.0284, "step": 1149 }, { "epoch": 0.22132409545804466, "grad_norm": 1.838537335395813, "learning_rate": 1.8141155183563195e-05, "loss": 1.0929, "step": 1150 }, { "epoch": 0.22151655119322555, "grad_norm": 1.9893226623535156, "learning_rate": 1.8137533895578585e-05, "loss": 1.063, "step": 1151 }, { "epoch": 0.22170900692840648, "grad_norm": 1.577759027481079, "learning_rate": 1.8133909445815277e-05, "loss": 1.0447, "step": 1152 }, { "epoch": 0.22190146266358737, "grad_norm": 1.8358311653137207, "learning_rate": 1.813028183568153e-05, "loss": 1.0892, "step": 1153 }, { "epoch": 0.22209391839876827, "grad_norm": 1.0686964988708496, "learning_rate": 1.8126651066586818e-05, "loss": 0.9327, "step": 1154 }, { "epoch": 0.2222863741339492, "grad_norm": 1.7629766464233398, "learning_rate": 1.8123017139941853e-05, "loss": 0.9343, "step": 1155 }, { "epoch": 0.2224788298691301, "grad_norm": 1.798951506614685, "learning_rate": 1.811938005715857e-05, "loss": 1.1111, "step": 1156 }, { "epoch": 0.22267128560431101, "grad_norm": 1.740086317062378, "learning_rate": 1.8115739819650124e-05, "loss": 1.0597, "step": 1157 }, { "epoch": 0.2228637413394919, "grad_norm": 1.7950950860977173, "learning_rate": 1.8112096428830912e-05, "loss": 1.0726, "step": 1158 }, { "epoch": 0.22305619707467284, "grad_norm": 1.825238585472107, "learning_rate": 1.8108449886116543e-05, "loss": 1.1226, "step": 1159 }, { "epoch": 0.22324865280985373, "grad_norm": 1.8829210996627808, "learning_rate": 1.8104800192923856e-05, "loss": 1.021, "step": 1160 }, { "epoch": 0.22344110854503466, "grad_norm": 1.971191167831421, "learning_rate": 1.8101147350670905e-05, "loss": 0.953, "step": 1161 }, { "epoch": 0.22363356428021555, "grad_norm": 1.8453309535980225, "learning_rate": 1.8097491360776987e-05, "loss": 1.0679, "step": 1162 }, { "epoch": 0.22382602001539645, "grad_norm": 1.9674023389816284, "learning_rate": 1.8093832224662598e-05, "loss": 0.9755, "step": 1163 }, { "epoch": 0.22401847575057737, "grad_norm": 1.70065438747406, "learning_rate": 1.8090169943749477e-05, "loss": 0.9797, "step": 1164 }, { "epoch": 0.22421093148575827, "grad_norm": 1.9551289081573486, "learning_rate": 1.808650451946057e-05, "loss": 1.0253, "step": 1165 }, { "epoch": 0.2244033872209392, "grad_norm": 1.5888679027557373, "learning_rate": 1.8082835953220055e-05, "loss": 1.022, "step": 1166 }, { "epoch": 0.2245958429561201, "grad_norm": 1.6807249784469604, "learning_rate": 1.8079164246453322e-05, "loss": 1.0853, "step": 1167 }, { "epoch": 0.22478829869130101, "grad_norm": 1.8890981674194336, "learning_rate": 1.8075489400586993e-05, "loss": 1.0652, "step": 1168 }, { "epoch": 0.2249807544264819, "grad_norm": 1.772186279296875, "learning_rate": 1.807181141704889e-05, "loss": 1.0159, "step": 1169 }, { "epoch": 0.2251732101616628, "grad_norm": 0.9067140817642212, "learning_rate": 1.806813029726808e-05, "loss": 0.9559, "step": 1170 }, { "epoch": 0.22536566589684373, "grad_norm": 1.6426454782485962, "learning_rate": 1.806444604267483e-05, "loss": 0.9861, "step": 1171 }, { "epoch": 0.22555812163202463, "grad_norm": 1.745294451713562, "learning_rate": 1.8060758654700622e-05, "loss": 1.0108, "step": 1172 }, { "epoch": 0.22575057736720555, "grad_norm": 1.8601793050765991, "learning_rate": 1.8057068134778174e-05, "loss": 1.0605, "step": 1173 }, { "epoch": 0.22594303310238645, "grad_norm": 1.8416272401809692, "learning_rate": 1.8053374484341403e-05, "loss": 1.0164, "step": 1174 }, { "epoch": 0.22613548883756737, "grad_norm": 1.7557302713394165, "learning_rate": 1.8049677704825455e-05, "loss": 1.0863, "step": 1175 }, { "epoch": 0.22632794457274827, "grad_norm": 1.7593114376068115, "learning_rate": 1.8045977797666685e-05, "loss": 0.9973, "step": 1176 }, { "epoch": 0.22652040030792917, "grad_norm": 1.832168698310852, "learning_rate": 1.8042274764302664e-05, "loss": 0.8869, "step": 1177 }, { "epoch": 0.2267128560431101, "grad_norm": 1.7560023069381714, "learning_rate": 1.8038568606172172e-05, "loss": 1.0508, "step": 1178 }, { "epoch": 0.22690531177829099, "grad_norm": 1.8156278133392334, "learning_rate": 1.803485932471522e-05, "loss": 1.0583, "step": 1179 }, { "epoch": 0.2270977675134719, "grad_norm": 1.6816129684448242, "learning_rate": 1.803114692137302e-05, "loss": 1.0035, "step": 1180 }, { "epoch": 0.2272902232486528, "grad_norm": 1.7617651224136353, "learning_rate": 1.8027431397587993e-05, "loss": 1.0421, "step": 1181 }, { "epoch": 0.22748267898383373, "grad_norm": 1.6813771724700928, "learning_rate": 1.8023712754803783e-05, "loss": 1.038, "step": 1182 }, { "epoch": 0.22767513471901463, "grad_norm": 1.7724182605743408, "learning_rate": 1.8019990994465235e-05, "loss": 0.9613, "step": 1183 }, { "epoch": 0.22786759045419552, "grad_norm": 1.9362376928329468, "learning_rate": 1.801626611801842e-05, "loss": 1.0767, "step": 1184 }, { "epoch": 0.22806004618937645, "grad_norm": 2.093571901321411, "learning_rate": 1.801253812691061e-05, "loss": 1.0002, "step": 1185 }, { "epoch": 0.22825250192455734, "grad_norm": 1.7024599313735962, "learning_rate": 1.8008807022590283e-05, "loss": 0.9694, "step": 1186 }, { "epoch": 0.22844495765973827, "grad_norm": 1.7977104187011719, "learning_rate": 1.8005072806507132e-05, "loss": 1.0255, "step": 1187 }, { "epoch": 0.22863741339491916, "grad_norm": 2.2798855304718018, "learning_rate": 1.8001335480112067e-05, "loss": 1.0358, "step": 1188 }, { "epoch": 0.2288298691301001, "grad_norm": 1.7836755514144897, "learning_rate": 1.799759504485719e-05, "loss": 1.0198, "step": 1189 }, { "epoch": 0.22902232486528099, "grad_norm": 2.0673530101776123, "learning_rate": 1.7993851502195822e-05, "loss": 0.9821, "step": 1190 }, { "epoch": 0.22921478060046188, "grad_norm": 1.8233660459518433, "learning_rate": 1.7990104853582494e-05, "loss": 1.1074, "step": 1191 }, { "epoch": 0.2294072363356428, "grad_norm": 1.7256102561950684, "learning_rate": 1.798635510047293e-05, "loss": 0.9712, "step": 1192 }, { "epoch": 0.2295996920708237, "grad_norm": 1.5374864339828491, "learning_rate": 1.7982602244324074e-05, "loss": 1.0184, "step": 1193 }, { "epoch": 0.22979214780600463, "grad_norm": 1.7800910472869873, "learning_rate": 1.7978846286594073e-05, "loss": 1.0586, "step": 1194 }, { "epoch": 0.22998460354118552, "grad_norm": 1.6155908107757568, "learning_rate": 1.7975087228742272e-05, "loss": 1.0483, "step": 1195 }, { "epoch": 0.23017705927636645, "grad_norm": 1.812739610671997, "learning_rate": 1.7971325072229227e-05, "loss": 1.0085, "step": 1196 }, { "epoch": 0.23036951501154734, "grad_norm": 1.69160795211792, "learning_rate": 1.7967559818516697e-05, "loss": 0.9515, "step": 1197 }, { "epoch": 0.23056197074672824, "grad_norm": 1.69354248046875, "learning_rate": 1.7963791469067637e-05, "loss": 0.9843, "step": 1198 }, { "epoch": 0.23075442648190916, "grad_norm": 1.8023440837860107, "learning_rate": 1.796002002534622e-05, "loss": 1.037, "step": 1199 }, { "epoch": 0.23094688221709006, "grad_norm": 1.8462785482406616, "learning_rate": 1.795624548881781e-05, "loss": 1.0643, "step": 1200 }, { "epoch": 0.23113933795227098, "grad_norm": 1.996802806854248, "learning_rate": 1.7952467860948975e-05, "loss": 0.9829, "step": 1201 }, { "epoch": 0.23133179368745188, "grad_norm": 1.8798552751541138, "learning_rate": 1.7948687143207485e-05, "loss": 1.0581, "step": 1202 }, { "epoch": 0.2315242494226328, "grad_norm": 1.8175809383392334, "learning_rate": 1.7944903337062304e-05, "loss": 0.9505, "step": 1203 }, { "epoch": 0.2317167051578137, "grad_norm": 1.9423161745071411, "learning_rate": 1.7941116443983612e-05, "loss": 0.9593, "step": 1204 }, { "epoch": 0.2319091608929946, "grad_norm": 1.7487837076187134, "learning_rate": 1.793732646544277e-05, "loss": 1.0067, "step": 1205 }, { "epoch": 0.23210161662817552, "grad_norm": 1.8825668096542358, "learning_rate": 1.7933533402912354e-05, "loss": 1.0329, "step": 1206 }, { "epoch": 0.23229407236335642, "grad_norm": 1.6054389476776123, "learning_rate": 1.7929737257866123e-05, "loss": 0.9492, "step": 1207 }, { "epoch": 0.23248652809853734, "grad_norm": 1.8480180501937866, "learning_rate": 1.7925938031779044e-05, "loss": 1.0566, "step": 1208 }, { "epoch": 0.23267898383371824, "grad_norm": 1.8325356245040894, "learning_rate": 1.7922135726127276e-05, "loss": 1.0747, "step": 1209 }, { "epoch": 0.23287143956889916, "grad_norm": 1.872121810913086, "learning_rate": 1.7918330342388183e-05, "loss": 1.0237, "step": 1210 }, { "epoch": 0.23306389530408006, "grad_norm": 1.8002170324325562, "learning_rate": 1.791452188204031e-05, "loss": 1.0303, "step": 1211 }, { "epoch": 0.23325635103926096, "grad_norm": 1.7081952095031738, "learning_rate": 1.7910710346563417e-05, "loss": 1.0768, "step": 1212 }, { "epoch": 0.23344880677444188, "grad_norm": 1.76996648311615, "learning_rate": 1.7906895737438437e-05, "loss": 1.0116, "step": 1213 }, { "epoch": 0.23364126250962278, "grad_norm": 1.8718223571777344, "learning_rate": 1.7903078056147515e-05, "loss": 1.0709, "step": 1214 }, { "epoch": 0.2338337182448037, "grad_norm": 1.5310128927230835, "learning_rate": 1.7899257304173975e-05, "loss": 0.9604, "step": 1215 }, { "epoch": 0.2340261739799846, "grad_norm": 1.7675104141235352, "learning_rate": 1.7895433483002356e-05, "loss": 0.9809, "step": 1216 }, { "epoch": 0.23421862971516552, "grad_norm": 1.7944415807724, "learning_rate": 1.7891606594118362e-05, "loss": 1.0158, "step": 1217 }, { "epoch": 0.23441108545034642, "grad_norm": 2.0859057903289795, "learning_rate": 1.7887776639008912e-05, "loss": 1.1133, "step": 1218 }, { "epoch": 0.23460354118552734, "grad_norm": 1.7554936408996582, "learning_rate": 1.7883943619162102e-05, "loss": 1.0571, "step": 1219 }, { "epoch": 0.23479599692070824, "grad_norm": 1.7676604986190796, "learning_rate": 1.788010753606722e-05, "loss": 0.9529, "step": 1220 }, { "epoch": 0.23498845265588914, "grad_norm": 1.5903501510620117, "learning_rate": 1.7876268391214756e-05, "loss": 0.9976, "step": 1221 }, { "epoch": 0.23518090839107006, "grad_norm": 1.6037919521331787, "learning_rate": 1.7872426186096374e-05, "loss": 1.0401, "step": 1222 }, { "epoch": 0.23537336412625096, "grad_norm": 0.9709982872009277, "learning_rate": 1.786858092220494e-05, "loss": 0.8902, "step": 1223 }, { "epoch": 0.23556581986143188, "grad_norm": 1.7277920246124268, "learning_rate": 1.78647326010345e-05, "loss": 0.9767, "step": 1224 }, { "epoch": 0.23575827559661278, "grad_norm": 0.9047741293907166, "learning_rate": 1.786088122408029e-05, "loss": 0.9068, "step": 1225 }, { "epoch": 0.2359507313317937, "grad_norm": 1.9055099487304688, "learning_rate": 1.785702679283874e-05, "loss": 1.1505, "step": 1226 }, { "epoch": 0.2361431870669746, "grad_norm": 0.9288726449012756, "learning_rate": 1.785316930880745e-05, "loss": 0.9357, "step": 1227 }, { "epoch": 0.2363356428021555, "grad_norm": 1.7098684310913086, "learning_rate": 1.7849308773485226e-05, "loss": 0.969, "step": 1228 }, { "epoch": 0.23652809853733642, "grad_norm": 1.840373158454895, "learning_rate": 1.7845445188372047e-05, "loss": 0.9651, "step": 1229 }, { "epoch": 0.23672055427251731, "grad_norm": 1.9371370077133179, "learning_rate": 1.784157855496908e-05, "loss": 0.9595, "step": 1230 }, { "epoch": 0.23691301000769824, "grad_norm": 1.808653473854065, "learning_rate": 1.7837708874778683e-05, "loss": 0.8936, "step": 1231 }, { "epoch": 0.23710546574287913, "grad_norm": 1.819419026374817, "learning_rate": 1.783383614930438e-05, "loss": 0.8878, "step": 1232 }, { "epoch": 0.23729792147806006, "grad_norm": 1.772814154624939, "learning_rate": 1.78299603800509e-05, "loss": 1.0108, "step": 1233 }, { "epoch": 0.23749037721324096, "grad_norm": 1.695981740951538, "learning_rate": 1.782608156852414e-05, "loss": 0.9906, "step": 1234 }, { "epoch": 0.23768283294842185, "grad_norm": 1.7509560585021973, "learning_rate": 1.7822199716231186e-05, "loss": 0.9955, "step": 1235 }, { "epoch": 0.23787528868360278, "grad_norm": 1.8696647882461548, "learning_rate": 1.78183148246803e-05, "loss": 0.9274, "step": 1236 }, { "epoch": 0.23806774441878367, "grad_norm": 1.6070022583007812, "learning_rate": 1.781442689538093e-05, "loss": 0.9698, "step": 1237 }, { "epoch": 0.2382602001539646, "grad_norm": 1.897267460823059, "learning_rate": 1.7810535929843703e-05, "loss": 1.107, "step": 1238 }, { "epoch": 0.2384526558891455, "grad_norm": 1.9193824529647827, "learning_rate": 1.780664192958042e-05, "loss": 1.0508, "step": 1239 }, { "epoch": 0.23864511162432642, "grad_norm": 1.8030322790145874, "learning_rate": 1.780274489610407e-05, "loss": 1.0556, "step": 1240 }, { "epoch": 0.2388375673595073, "grad_norm": 1.7325868606567383, "learning_rate": 1.7798844830928818e-05, "loss": 1.0318, "step": 1241 }, { "epoch": 0.2390300230946882, "grad_norm": 1.8228734731674194, "learning_rate": 1.7794941735570005e-05, "loss": 1.0345, "step": 1242 }, { "epoch": 0.23922247882986913, "grad_norm": 1.743476152420044, "learning_rate": 1.7791035611544146e-05, "loss": 1.0488, "step": 1243 }, { "epoch": 0.23941493456505003, "grad_norm": 1.841093897819519, "learning_rate": 1.778712646036894e-05, "loss": 1.0806, "step": 1244 }, { "epoch": 0.23960739030023095, "grad_norm": 1.0069082975387573, "learning_rate": 1.778321428356326e-05, "loss": 0.9101, "step": 1245 }, { "epoch": 0.23979984603541185, "grad_norm": 1.8351349830627441, "learning_rate": 1.777929908264715e-05, "loss": 1.0576, "step": 1246 }, { "epoch": 0.23999230177059278, "grad_norm": 0.9957099556922913, "learning_rate": 1.7775380859141837e-05, "loss": 0.9264, "step": 1247 }, { "epoch": 0.24018475750577367, "grad_norm": 1.7579529285430908, "learning_rate": 1.777145961456971e-05, "loss": 1.0947, "step": 1248 }, { "epoch": 0.24037721324095457, "grad_norm": 1.727107286453247, "learning_rate": 1.7767535350454347e-05, "loss": 1.0202, "step": 1249 }, { "epoch": 0.2405696689761355, "grad_norm": 1.709868311882019, "learning_rate": 1.7763608068320487e-05, "loss": 1.1389, "step": 1250 }, { "epoch": 0.2407621247113164, "grad_norm": 1.717427372932434, "learning_rate": 1.775967776969405e-05, "loss": 0.9958, "step": 1251 }, { "epoch": 0.2409545804464973, "grad_norm": 1.6196602582931519, "learning_rate": 1.7755744456102123e-05, "loss": 0.9765, "step": 1252 }, { "epoch": 0.2411470361816782, "grad_norm": 1.794206142425537, "learning_rate": 1.7751808129072968e-05, "loss": 1.0123, "step": 1253 }, { "epoch": 0.24133949191685913, "grad_norm": 1.7889125347137451, "learning_rate": 1.7747868790136012e-05, "loss": 1.0263, "step": 1254 }, { "epoch": 0.24153194765204003, "grad_norm": 1.6667126417160034, "learning_rate": 1.7743926440821857e-05, "loss": 1.0379, "step": 1255 }, { "epoch": 0.24172440338722093, "grad_norm": 1.7728615999221802, "learning_rate": 1.7739981082662275e-05, "loss": 1.0555, "step": 1256 }, { "epoch": 0.24191685912240185, "grad_norm": 1.777949333190918, "learning_rate": 1.7736032717190208e-05, "loss": 1.0027, "step": 1257 }, { "epoch": 0.24210931485758275, "grad_norm": 1.6838446855545044, "learning_rate": 1.773208134593976e-05, "loss": 0.978, "step": 1258 }, { "epoch": 0.24230177059276367, "grad_norm": 1.857752799987793, "learning_rate": 1.772812697044621e-05, "loss": 1.0812, "step": 1259 }, { "epoch": 0.24249422632794457, "grad_norm": 1.7607934474945068, "learning_rate": 1.7724169592245996e-05, "loss": 0.9992, "step": 1260 }, { "epoch": 0.2426866820631255, "grad_norm": 1.7443361282348633, "learning_rate": 1.772020921287674e-05, "loss": 1.0517, "step": 1261 }, { "epoch": 0.2428791377983064, "grad_norm": 1.8854649066925049, "learning_rate": 1.7716245833877202e-05, "loss": 0.8986, "step": 1262 }, { "epoch": 0.24307159353348728, "grad_norm": 1.7788060903549194, "learning_rate": 1.7712279456787334e-05, "loss": 1.0414, "step": 1263 }, { "epoch": 0.2432640492686682, "grad_norm": 1.8839924335479736, "learning_rate": 1.7708310083148242e-05, "loss": 1.0034, "step": 1264 }, { "epoch": 0.2434565050038491, "grad_norm": 1.1061986684799194, "learning_rate": 1.7704337714502195e-05, "loss": 0.9332, "step": 1265 }, { "epoch": 0.24364896073903003, "grad_norm": 1.904776692390442, "learning_rate": 1.7700362352392632e-05, "loss": 1.0976, "step": 1266 }, { "epoch": 0.24384141647421093, "grad_norm": 1.9352667331695557, "learning_rate": 1.769638399836414e-05, "loss": 1.036, "step": 1267 }, { "epoch": 0.24403387220939185, "grad_norm": 1.694658875465393, "learning_rate": 1.769240265396249e-05, "loss": 0.9766, "step": 1268 }, { "epoch": 0.24422632794457275, "grad_norm": 1.8434773683547974, "learning_rate": 1.7688418320734596e-05, "loss": 1.0676, "step": 1269 }, { "epoch": 0.24441878367975364, "grad_norm": 1.8939404487609863, "learning_rate": 1.7684431000228548e-05, "loss": 1.0183, "step": 1270 }, { "epoch": 0.24461123941493457, "grad_norm": 1.8293036222457886, "learning_rate": 1.7680440693993586e-05, "loss": 0.9748, "step": 1271 }, { "epoch": 0.24480369515011546, "grad_norm": 1.837016224861145, "learning_rate": 1.7676447403580114e-05, "loss": 0.9581, "step": 1272 }, { "epoch": 0.2449961508852964, "grad_norm": 1.7812628746032715, "learning_rate": 1.7672451130539697e-05, "loss": 1.0667, "step": 1273 }, { "epoch": 0.24518860662047728, "grad_norm": 1.7685917615890503, "learning_rate": 1.7668451876425056e-05, "loss": 1.1035, "step": 1274 }, { "epoch": 0.2453810623556582, "grad_norm": 1.7301274538040161, "learning_rate": 1.766444964279007e-05, "loss": 1.0731, "step": 1275 }, { "epoch": 0.2455735180908391, "grad_norm": 1.724979043006897, "learning_rate": 1.766044443118978e-05, "loss": 1.0233, "step": 1276 }, { "epoch": 0.24576597382602003, "grad_norm": 1.7777516841888428, "learning_rate": 1.7656436243180385e-05, "loss": 1.0054, "step": 1277 }, { "epoch": 0.24595842956120093, "grad_norm": 1.7832542657852173, "learning_rate": 1.765242508031923e-05, "loss": 1.0269, "step": 1278 }, { "epoch": 0.24615088529638182, "grad_norm": 1.6509227752685547, "learning_rate": 1.7648410944164823e-05, "loss": 0.9269, "step": 1279 }, { "epoch": 0.24634334103156275, "grad_norm": 1.7656588554382324, "learning_rate": 1.7644393836276832e-05, "loss": 1.0168, "step": 1280 }, { "epoch": 0.24653579676674364, "grad_norm": 1.8130340576171875, "learning_rate": 1.7640373758216075e-05, "loss": 1.006, "step": 1281 }, { "epoch": 0.24672825250192457, "grad_norm": 1.8497333526611328, "learning_rate": 1.7636350711544518e-05, "loss": 1.0527, "step": 1282 }, { "epoch": 0.24692070823710546, "grad_norm": 1.7750145196914673, "learning_rate": 1.7632324697825288e-05, "loss": 0.9754, "step": 1283 }, { "epoch": 0.2471131639722864, "grad_norm": 1.7450518608093262, "learning_rate": 1.7628295718622666e-05, "loss": 1.0049, "step": 1284 }, { "epoch": 0.24730561970746728, "grad_norm": 1.7295411825180054, "learning_rate": 1.762426377550208e-05, "loss": 1.0718, "step": 1285 }, { "epoch": 0.24749807544264818, "grad_norm": 0.9832800626754761, "learning_rate": 1.762022887003011e-05, "loss": 0.905, "step": 1286 }, { "epoch": 0.2476905311778291, "grad_norm": 1.9119161367416382, "learning_rate": 1.761619100377449e-05, "loss": 0.9969, "step": 1287 }, { "epoch": 0.24788298691301, "grad_norm": 1.6869584321975708, "learning_rate": 1.7612150178304102e-05, "loss": 1.0315, "step": 1288 }, { "epoch": 0.24807544264819092, "grad_norm": 1.627511978149414, "learning_rate": 1.7608106395188983e-05, "loss": 0.9896, "step": 1289 }, { "epoch": 0.24826789838337182, "grad_norm": 1.7231568098068237, "learning_rate": 1.7604059656000313e-05, "loss": 1.0248, "step": 1290 }, { "epoch": 0.24846035411855275, "grad_norm": 1.6172611713409424, "learning_rate": 1.7600009962310417e-05, "loss": 1.0182, "step": 1291 }, { "epoch": 0.24865280985373364, "grad_norm": 2.021790027618408, "learning_rate": 1.7595957315692782e-05, "loss": 0.9912, "step": 1292 }, { "epoch": 0.24884526558891454, "grad_norm": 1.859525203704834, "learning_rate": 1.759190171772203e-05, "loss": 1.078, "step": 1293 }, { "epoch": 0.24903772132409546, "grad_norm": 1.68208909034729, "learning_rate": 1.7587843169973934e-05, "loss": 0.9745, "step": 1294 }, { "epoch": 0.24923017705927636, "grad_norm": 1.7649648189544678, "learning_rate": 1.7583781674025415e-05, "loss": 1.0394, "step": 1295 }, { "epoch": 0.24942263279445728, "grad_norm": 1.6666499376296997, "learning_rate": 1.757971723145453e-05, "loss": 1.0309, "step": 1296 }, { "epoch": 0.24961508852963818, "grad_norm": 1.6025043725967407, "learning_rate": 1.75756498438405e-05, "loss": 0.9921, "step": 1297 }, { "epoch": 0.2498075442648191, "grad_norm": 1.816798210144043, "learning_rate": 1.7571579512763665e-05, "loss": 1.06, "step": 1298 }, { "epoch": 0.25, "grad_norm": 1.614940881729126, "learning_rate": 1.756750623980553e-05, "loss": 1.0393, "step": 1299 }, { "epoch": 0.2501924557351809, "grad_norm": 1.6103315353393555, "learning_rate": 1.7563430026548737e-05, "loss": 0.9786, "step": 1300 }, { "epoch": 0.2503849114703618, "grad_norm": 1.058126449584961, "learning_rate": 1.7559350874577066e-05, "loss": 1.0101, "step": 1301 }, { "epoch": 0.25057736720554274, "grad_norm": 1.711808204650879, "learning_rate": 1.7555268785475438e-05, "loss": 1.0158, "step": 1302 }, { "epoch": 0.25076982294072364, "grad_norm": 1.819156527519226, "learning_rate": 1.755118376082992e-05, "loss": 1.0493, "step": 1303 }, { "epoch": 0.25096227867590454, "grad_norm": 1.6954679489135742, "learning_rate": 1.7547095802227723e-05, "loss": 0.9808, "step": 1304 }, { "epoch": 0.25115473441108543, "grad_norm": 1.9226264953613281, "learning_rate": 1.7543004911257186e-05, "loss": 1.0386, "step": 1305 }, { "epoch": 0.2513471901462664, "grad_norm": 1.7474966049194336, "learning_rate": 1.75389110895078e-05, "loss": 1.0898, "step": 1306 }, { "epoch": 0.2515396458814473, "grad_norm": 1.8988890647888184, "learning_rate": 1.7534814338570187e-05, "loss": 1.0258, "step": 1307 }, { "epoch": 0.2517321016166282, "grad_norm": 1.687604308128357, "learning_rate": 1.7530714660036112e-05, "loss": 1.1225, "step": 1308 }, { "epoch": 0.2519245573518091, "grad_norm": 0.8564708828926086, "learning_rate": 1.7526612055498468e-05, "loss": 0.8735, "step": 1309 }, { "epoch": 0.25211701308698997, "grad_norm": 1.9589653015136719, "learning_rate": 1.7522506526551303e-05, "loss": 0.9413, "step": 1310 }, { "epoch": 0.2523094688221709, "grad_norm": 1.763735055923462, "learning_rate": 1.7518398074789776e-05, "loss": 1.0326, "step": 1311 }, { "epoch": 0.2525019245573518, "grad_norm": 1.8903660774230957, "learning_rate": 1.7514286701810203e-05, "loss": 1.0755, "step": 1312 }, { "epoch": 0.2526943802925327, "grad_norm": 1.7349900007247925, "learning_rate": 1.751017240921003e-05, "loss": 1.0698, "step": 1313 }, { "epoch": 0.2528868360277136, "grad_norm": 1.749704122543335, "learning_rate": 1.7506055198587835e-05, "loss": 0.9767, "step": 1314 }, { "epoch": 0.2530792917628945, "grad_norm": 1.7060540914535522, "learning_rate": 1.7501935071543325e-05, "loss": 0.8928, "step": 1315 }, { "epoch": 0.25327174749807546, "grad_norm": 1.8056551218032837, "learning_rate": 1.7497812029677344e-05, "loss": 1.0696, "step": 1316 }, { "epoch": 0.25346420323325636, "grad_norm": 1.7558966875076294, "learning_rate": 1.7493686074591873e-05, "loss": 0.9184, "step": 1317 }, { "epoch": 0.25365665896843725, "grad_norm": 1.8495056629180908, "learning_rate": 1.7489557207890025e-05, "loss": 1.0018, "step": 1318 }, { "epoch": 0.25384911470361815, "grad_norm": 1.800793170928955, "learning_rate": 1.7485425431176032e-05, "loss": 1.0815, "step": 1319 }, { "epoch": 0.2540415704387991, "grad_norm": 1.952811598777771, "learning_rate": 1.748129074605527e-05, "loss": 1.0077, "step": 1320 }, { "epoch": 0.25423402617398, "grad_norm": 1.7768923044204712, "learning_rate": 1.7477153154134244e-05, "loss": 0.9185, "step": 1321 }, { "epoch": 0.2544264819091609, "grad_norm": 1.576505184173584, "learning_rate": 1.7473012657020577e-05, "loss": 0.9175, "step": 1322 }, { "epoch": 0.2546189376443418, "grad_norm": 1.6869131326675415, "learning_rate": 1.7468869256323036e-05, "loss": 0.9495, "step": 1323 }, { "epoch": 0.2548113933795227, "grad_norm": 1.8397648334503174, "learning_rate": 1.7464722953651504e-05, "loss": 1.0346, "step": 1324 }, { "epoch": 0.25500384911470364, "grad_norm": 1.7348251342773438, "learning_rate": 1.7460573750616996e-05, "loss": 1.0139, "step": 1325 }, { "epoch": 0.25519630484988454, "grad_norm": 1.7938367128372192, "learning_rate": 1.7456421648831658e-05, "loss": 1.0418, "step": 1326 }, { "epoch": 0.25538876058506543, "grad_norm": 1.7588385343551636, "learning_rate": 1.7452266649908756e-05, "loss": 1.0398, "step": 1327 }, { "epoch": 0.25558121632024633, "grad_norm": 1.6426222324371338, "learning_rate": 1.7448108755462684e-05, "loss": 0.9784, "step": 1328 }, { "epoch": 0.2557736720554272, "grad_norm": 1.8306379318237305, "learning_rate": 1.744394796710896e-05, "loss": 0.9632, "step": 1329 }, { "epoch": 0.2559661277906082, "grad_norm": 1.7624549865722656, "learning_rate": 1.7439784286464234e-05, "loss": 1.0239, "step": 1330 }, { "epoch": 0.2561585835257891, "grad_norm": 1.6674563884735107, "learning_rate": 1.743561771514626e-05, "loss": 1.0682, "step": 1331 }, { "epoch": 0.25635103926096997, "grad_norm": 1.8462331295013428, "learning_rate": 1.7431448254773943e-05, "loss": 1.0796, "step": 1332 }, { "epoch": 0.25654349499615087, "grad_norm": 1.751166582107544, "learning_rate": 1.7427275906967288e-05, "loss": 1.0622, "step": 1333 }, { "epoch": 0.2567359507313318, "grad_norm": 1.7317910194396973, "learning_rate": 1.7423100673347435e-05, "loss": 1.0154, "step": 1334 }, { "epoch": 0.2569284064665127, "grad_norm": 1.7881197929382324, "learning_rate": 1.741892255553664e-05, "loss": 0.9863, "step": 1335 }, { "epoch": 0.2571208622016936, "grad_norm": 1.92635178565979, "learning_rate": 1.741474155515827e-05, "loss": 1.0101, "step": 1336 }, { "epoch": 0.2573133179368745, "grad_norm": 1.8678805828094482, "learning_rate": 1.741055767383683e-05, "loss": 0.9825, "step": 1337 }, { "epoch": 0.2575057736720554, "grad_norm": 1.5679118633270264, "learning_rate": 1.7406370913197938e-05, "loss": 0.9571, "step": 1338 }, { "epoch": 0.25769822940723636, "grad_norm": 1.802016258239746, "learning_rate": 1.7402181274868323e-05, "loss": 0.9932, "step": 1339 }, { "epoch": 0.25789068514241725, "grad_norm": 1.0259318351745605, "learning_rate": 1.7397988760475842e-05, "loss": 0.9404, "step": 1340 }, { "epoch": 0.25808314087759815, "grad_norm": 1.7486754655838013, "learning_rate": 1.739379337164946e-05, "loss": 1.0702, "step": 1341 }, { "epoch": 0.25827559661277905, "grad_norm": 1.8874517679214478, "learning_rate": 1.738959511001928e-05, "loss": 1.081, "step": 1342 }, { "epoch": 0.25846805234795994, "grad_norm": 1.7347866296768188, "learning_rate": 1.7385393977216486e-05, "loss": 0.9043, "step": 1343 }, { "epoch": 0.2586605080831409, "grad_norm": 1.917789101600647, "learning_rate": 1.738118997487341e-05, "loss": 1.0167, "step": 1344 }, { "epoch": 0.2588529638183218, "grad_norm": 1.7709996700286865, "learning_rate": 1.7376983104623477e-05, "loss": 0.9877, "step": 1345 }, { "epoch": 0.2590454195535027, "grad_norm": 1.7030057907104492, "learning_rate": 1.737277336810124e-05, "loss": 0.9062, "step": 1346 }, { "epoch": 0.2592378752886836, "grad_norm": 1.7927793264389038, "learning_rate": 1.7368560766942362e-05, "loss": 0.9742, "step": 1347 }, { "epoch": 0.25943033102386454, "grad_norm": 1.974885106086731, "learning_rate": 1.736434530278362e-05, "loss": 1.0647, "step": 1348 }, { "epoch": 0.25962278675904543, "grad_norm": 1.8438800573349, "learning_rate": 1.73601269772629e-05, "loss": 0.9913, "step": 1349 }, { "epoch": 0.25981524249422633, "grad_norm": 1.8494850397109985, "learning_rate": 1.7355905792019195e-05, "loss": 1.0457, "step": 1350 }, { "epoch": 0.2600076982294072, "grad_norm": 1.992661714553833, "learning_rate": 1.7351681748692622e-05, "loss": 1.0228, "step": 1351 }, { "epoch": 0.2602001539645881, "grad_norm": 1.733710527420044, "learning_rate": 1.73474548489244e-05, "loss": 1.0338, "step": 1352 }, { "epoch": 0.2603926096997691, "grad_norm": 1.829520344734192, "learning_rate": 1.7343225094356857e-05, "loss": 0.9681, "step": 1353 }, { "epoch": 0.26058506543494997, "grad_norm": 1.976354956626892, "learning_rate": 1.7338992486633438e-05, "loss": 1.0252, "step": 1354 }, { "epoch": 0.26077752117013087, "grad_norm": 1.8615800142288208, "learning_rate": 1.733475702739869e-05, "loss": 0.9862, "step": 1355 }, { "epoch": 0.26096997690531176, "grad_norm": 1.6784212589263916, "learning_rate": 1.7330518718298263e-05, "loss": 0.9432, "step": 1356 }, { "epoch": 0.2611624326404927, "grad_norm": 1.9273643493652344, "learning_rate": 1.732627756097893e-05, "loss": 0.9529, "step": 1357 }, { "epoch": 0.2613548883756736, "grad_norm": 1.0276539325714111, "learning_rate": 1.732203355708856e-05, "loss": 0.9107, "step": 1358 }, { "epoch": 0.2615473441108545, "grad_norm": 1.8310490846633911, "learning_rate": 1.7317786708276124e-05, "loss": 1.0597, "step": 1359 }, { "epoch": 0.2617397998460354, "grad_norm": 1.8171107769012451, "learning_rate": 1.7313537016191706e-05, "loss": 0.9781, "step": 1360 }, { "epoch": 0.2619322555812163, "grad_norm": 1.8415398597717285, "learning_rate": 1.7309284482486494e-05, "loss": 0.9609, "step": 1361 }, { "epoch": 0.26212471131639725, "grad_norm": 1.808424472808838, "learning_rate": 1.7305029108812777e-05, "loss": 1.0246, "step": 1362 }, { "epoch": 0.26231716705157815, "grad_norm": 2.0431182384490967, "learning_rate": 1.730077089682395e-05, "loss": 1.0021, "step": 1363 }, { "epoch": 0.26250962278675904, "grad_norm": 1.7229528427124023, "learning_rate": 1.729650984817451e-05, "loss": 1.0615, "step": 1364 }, { "epoch": 0.26270207852193994, "grad_norm": 1.954764723777771, "learning_rate": 1.7292245964520056e-05, "loss": 1.0028, "step": 1365 }, { "epoch": 0.26289453425712084, "grad_norm": 1.7568436861038208, "learning_rate": 1.7287979247517285e-05, "loss": 0.9881, "step": 1366 }, { "epoch": 0.2630869899923018, "grad_norm": 1.8195327520370483, "learning_rate": 1.7283709698824004e-05, "loss": 0.8825, "step": 1367 }, { "epoch": 0.2632794457274827, "grad_norm": 1.807549238204956, "learning_rate": 1.727943732009911e-05, "loss": 0.9414, "step": 1368 }, { "epoch": 0.2634719014626636, "grad_norm": 1.665419101715088, "learning_rate": 1.7275162113002607e-05, "loss": 1.026, "step": 1369 }, { "epoch": 0.2636643571978445, "grad_norm": 1.8149460554122925, "learning_rate": 1.7270884079195596e-05, "loss": 1.0202, "step": 1370 }, { "epoch": 0.26385681293302543, "grad_norm": 1.875658631324768, "learning_rate": 1.7266603220340273e-05, "loss": 0.9394, "step": 1371 }, { "epoch": 0.2640492686682063, "grad_norm": 1.882163405418396, "learning_rate": 1.726231953809993e-05, "loss": 1.082, "step": 1372 }, { "epoch": 0.2642417244033872, "grad_norm": 1.8240703344345093, "learning_rate": 1.7258033034138972e-05, "loss": 0.9827, "step": 1373 }, { "epoch": 0.2644341801385681, "grad_norm": 1.9846265316009521, "learning_rate": 1.7253743710122877e-05, "loss": 0.8961, "step": 1374 }, { "epoch": 0.264626635873749, "grad_norm": 1.7519563436508179, "learning_rate": 1.7249451567718236e-05, "loss": 0.9887, "step": 1375 }, { "epoch": 0.26481909160892997, "grad_norm": 1.8336985111236572, "learning_rate": 1.7245156608592727e-05, "loss": 1.0265, "step": 1376 }, { "epoch": 0.26501154734411086, "grad_norm": 1.8857085704803467, "learning_rate": 1.7240858834415126e-05, "loss": 1.0037, "step": 1377 }, { "epoch": 0.26520400307929176, "grad_norm": 1.7547630071640015, "learning_rate": 1.7236558246855303e-05, "loss": 0.9228, "step": 1378 }, { "epoch": 0.26539645881447266, "grad_norm": 1.7152235507965088, "learning_rate": 1.7232254847584212e-05, "loss": 1.0608, "step": 1379 }, { "epoch": 0.26558891454965355, "grad_norm": 1.7370901107788086, "learning_rate": 1.7227948638273918e-05, "loss": 1.0041, "step": 1380 }, { "epoch": 0.2657813702848345, "grad_norm": 1.6386165618896484, "learning_rate": 1.7223639620597556e-05, "loss": 0.9868, "step": 1381 }, { "epoch": 0.2659738260200154, "grad_norm": 1.0060551166534424, "learning_rate": 1.721932779622937e-05, "loss": 0.96, "step": 1382 }, { "epoch": 0.2661662817551963, "grad_norm": 1.8396605253219604, "learning_rate": 1.721501316684469e-05, "loss": 1.1159, "step": 1383 }, { "epoch": 0.2663587374903772, "grad_norm": 1.8105412721633911, "learning_rate": 1.7210695734119926e-05, "loss": 0.9782, "step": 1384 }, { "epoch": 0.26655119322555815, "grad_norm": 1.8018509149551392, "learning_rate": 1.7206375499732586e-05, "loss": 1.1463, "step": 1385 }, { "epoch": 0.26674364896073904, "grad_norm": 1.8077846765518188, "learning_rate": 1.7202052465361268e-05, "loss": 0.9705, "step": 1386 }, { "epoch": 0.26693610469591994, "grad_norm": 1.6987974643707275, "learning_rate": 1.7197726632685653e-05, "loss": 1.0391, "step": 1387 }, { "epoch": 0.26712856043110084, "grad_norm": 1.6711596250534058, "learning_rate": 1.7193398003386514e-05, "loss": 0.9726, "step": 1388 }, { "epoch": 0.26732101616628173, "grad_norm": 1.7686970233917236, "learning_rate": 1.7189066579145704e-05, "loss": 1.0221, "step": 1389 }, { "epoch": 0.2675134719014627, "grad_norm": 1.8154048919677734, "learning_rate": 1.7184732361646168e-05, "loss": 0.9996, "step": 1390 }, { "epoch": 0.2677059276366436, "grad_norm": 1.779394507408142, "learning_rate": 1.718039535257194e-05, "loss": 1.0399, "step": 1391 }, { "epoch": 0.2678983833718245, "grad_norm": 1.8855408430099487, "learning_rate": 1.717605555360812e-05, "loss": 1.0439, "step": 1392 }, { "epoch": 0.2680908391070054, "grad_norm": 1.8445767164230347, "learning_rate": 1.7171712966440916e-05, "loss": 0.9833, "step": 1393 }, { "epoch": 0.26828329484218627, "grad_norm": 1.8350696563720703, "learning_rate": 1.71673675927576e-05, "loss": 1.01, "step": 1394 }, { "epoch": 0.2684757505773672, "grad_norm": 2.0257089138031006, "learning_rate": 1.7163019434246545e-05, "loss": 1.0609, "step": 1395 }, { "epoch": 0.2686682063125481, "grad_norm": 1.6797620058059692, "learning_rate": 1.7158668492597186e-05, "loss": 1.0513, "step": 1396 }, { "epoch": 0.268860662047729, "grad_norm": 1.6715037822723389, "learning_rate": 1.7154314769500052e-05, "loss": 1.0546, "step": 1397 }, { "epoch": 0.2690531177829099, "grad_norm": 1.692131519317627, "learning_rate": 1.7149958266646756e-05, "loss": 1.0242, "step": 1398 }, { "epoch": 0.26924557351809086, "grad_norm": 1.886602520942688, "learning_rate": 1.7145598985729976e-05, "loss": 0.9728, "step": 1399 }, { "epoch": 0.26943802925327176, "grad_norm": 1.510108232498169, "learning_rate": 1.7141236928443482e-05, "loss": 1.0239, "step": 1400 }, { "epoch": 0.26963048498845266, "grad_norm": 1.9030672311782837, "learning_rate": 1.7136872096482123e-05, "loss": 1.0234, "step": 1401 }, { "epoch": 0.26982294072363355, "grad_norm": 1.6661264896392822, "learning_rate": 1.713250449154182e-05, "loss": 0.9716, "step": 1402 }, { "epoch": 0.27001539645881445, "grad_norm": 1.672459363937378, "learning_rate": 1.7128134115319567e-05, "loss": 1.0058, "step": 1403 }, { "epoch": 0.2702078521939954, "grad_norm": 1.665493369102478, "learning_rate": 1.712376096951345e-05, "loss": 1.0422, "step": 1404 }, { "epoch": 0.2704003079291763, "grad_norm": 2.1093878746032715, "learning_rate": 1.711938505582262e-05, "loss": 1.0554, "step": 1405 }, { "epoch": 0.2705927636643572, "grad_norm": 1.7243448495864868, "learning_rate": 1.7115006375947304e-05, "loss": 1.0035, "step": 1406 }, { "epoch": 0.2707852193995381, "grad_norm": 1.8660968542099, "learning_rate": 1.7110624931588805e-05, "loss": 0.9071, "step": 1407 }, { "epoch": 0.270977675134719, "grad_norm": 1.950353741645813, "learning_rate": 1.7106240724449507e-05, "loss": 1.0344, "step": 1408 }, { "epoch": 0.27117013086989994, "grad_norm": 1.7867465019226074, "learning_rate": 1.7101853756232856e-05, "loss": 1.0574, "step": 1409 }, { "epoch": 0.27136258660508084, "grad_norm": 1.833276629447937, "learning_rate": 1.7097464028643376e-05, "loss": 0.8997, "step": 1410 }, { "epoch": 0.27155504234026173, "grad_norm": 1.7349493503570557, "learning_rate": 1.7093071543386667e-05, "loss": 1.0942, "step": 1411 }, { "epoch": 0.27174749807544263, "grad_norm": 1.99166738986969, "learning_rate": 1.7088676302169394e-05, "loss": 1.0613, "step": 1412 }, { "epoch": 0.2719399538106236, "grad_norm": 1.8484245538711548, "learning_rate": 1.7084278306699295e-05, "loss": 0.9649, "step": 1413 }, { "epoch": 0.2721324095458045, "grad_norm": 1.764248013496399, "learning_rate": 1.7079877558685178e-05, "loss": 1.0351, "step": 1414 }, { "epoch": 0.2723248652809854, "grad_norm": 1.7142534255981445, "learning_rate": 1.7075474059836923e-05, "loss": 1.0147, "step": 1415 }, { "epoch": 0.27251732101616627, "grad_norm": 1.048027753829956, "learning_rate": 1.7071067811865477e-05, "loss": 0.9572, "step": 1416 }, { "epoch": 0.27270977675134717, "grad_norm": 1.7482398748397827, "learning_rate": 1.7066658816482856e-05, "loss": 0.9821, "step": 1417 }, { "epoch": 0.2729022324865281, "grad_norm": 2.0123450756073, "learning_rate": 1.7062247075402137e-05, "loss": 1.0296, "step": 1418 }, { "epoch": 0.273094688221709, "grad_norm": 1.903601884841919, "learning_rate": 1.705783259033748e-05, "loss": 0.965, "step": 1419 }, { "epoch": 0.2732871439568899, "grad_norm": 1.7349658012390137, "learning_rate": 1.705341536300409e-05, "loss": 1.066, "step": 1420 }, { "epoch": 0.2734795996920708, "grad_norm": 0.969205915927887, "learning_rate": 1.7048995395118253e-05, "loss": 0.9331, "step": 1421 }, { "epoch": 0.27367205542725176, "grad_norm": 1.8710532188415527, "learning_rate": 1.7044572688397316e-05, "loss": 1.0358, "step": 1422 }, { "epoch": 0.27386451116243266, "grad_norm": 1.883230447769165, "learning_rate": 1.7040147244559688e-05, "loss": 1.0545, "step": 1423 }, { "epoch": 0.27405696689761355, "grad_norm": 1.9921497106552124, "learning_rate": 1.7035719065324837e-05, "loss": 0.9576, "step": 1424 }, { "epoch": 0.27424942263279445, "grad_norm": 1.786797285079956, "learning_rate": 1.7031288152413304e-05, "loss": 1.0137, "step": 1425 }, { "epoch": 0.27444187836797534, "grad_norm": 1.7009328603744507, "learning_rate": 1.7026854507546694e-05, "loss": 0.9465, "step": 1426 }, { "epoch": 0.2746343341031563, "grad_norm": 1.8068492412567139, "learning_rate": 1.7022418132447656e-05, "loss": 1.0471, "step": 1427 }, { "epoch": 0.2748267898383372, "grad_norm": 1.7573258876800537, "learning_rate": 1.7017979028839918e-05, "loss": 0.9715, "step": 1428 }, { "epoch": 0.2750192455735181, "grad_norm": 1.7368972301483154, "learning_rate": 1.7013537198448253e-05, "loss": 0.983, "step": 1429 }, { "epoch": 0.275211701308699, "grad_norm": 1.7669157981872559, "learning_rate": 1.700909264299851e-05, "loss": 0.935, "step": 1430 }, { "epoch": 0.2754041570438799, "grad_norm": 1.910591959953308, "learning_rate": 1.7004645364217584e-05, "loss": 1.0562, "step": 1431 }, { "epoch": 0.27559661277906083, "grad_norm": 0.9342668056488037, "learning_rate": 1.7000195363833434e-05, "loss": 0.9297, "step": 1432 }, { "epoch": 0.27578906851424173, "grad_norm": 1.724603533744812, "learning_rate": 1.6995742643575076e-05, "loss": 1.0257, "step": 1433 }, { "epoch": 0.2759815242494226, "grad_norm": 1.8259358406066895, "learning_rate": 1.6991287205172575e-05, "loss": 0.9676, "step": 1434 }, { "epoch": 0.2761739799846035, "grad_norm": 1.8452657461166382, "learning_rate": 1.6986829050357068e-05, "loss": 1.0805, "step": 1435 }, { "epoch": 0.2763664357197845, "grad_norm": 1.8732038736343384, "learning_rate": 1.698236818086073e-05, "loss": 1.118, "step": 1436 }, { "epoch": 0.27655889145496537, "grad_norm": 1.706579327583313, "learning_rate": 1.6977904598416803e-05, "loss": 1.0446, "step": 1437 }, { "epoch": 0.27675134719014627, "grad_norm": 1.7886749505996704, "learning_rate": 1.697343830475958e-05, "loss": 0.929, "step": 1438 }, { "epoch": 0.27694380292532716, "grad_norm": 0.9286351799964905, "learning_rate": 1.6968969301624402e-05, "loss": 0.872, "step": 1439 }, { "epoch": 0.27713625866050806, "grad_norm": 1.6689391136169434, "learning_rate": 1.696449759074767e-05, "loss": 0.995, "step": 1440 }, { "epoch": 0.277328714395689, "grad_norm": 1.6442140340805054, "learning_rate": 1.6960023173866834e-05, "loss": 1.0275, "step": 1441 }, { "epoch": 0.2775211701308699, "grad_norm": 0.9267822504043579, "learning_rate": 1.6955546052720398e-05, "loss": 0.8755, "step": 1442 }, { "epoch": 0.2777136258660508, "grad_norm": 1.802245020866394, "learning_rate": 1.695106622904791e-05, "loss": 0.9753, "step": 1443 }, { "epoch": 0.2779060816012317, "grad_norm": 1.7879210710525513, "learning_rate": 1.6946583704589973e-05, "loss": 1.0423, "step": 1444 }, { "epoch": 0.2780985373364126, "grad_norm": 1.692579984664917, "learning_rate": 1.6942098481088245e-05, "loss": 0.93, "step": 1445 }, { "epoch": 0.27829099307159355, "grad_norm": 1.565054178237915, "learning_rate": 1.693761056028542e-05, "loss": 0.9838, "step": 1446 }, { "epoch": 0.27848344880677445, "grad_norm": 1.7526954412460327, "learning_rate": 1.6933119943925243e-05, "loss": 0.994, "step": 1447 }, { "epoch": 0.27867590454195534, "grad_norm": 1.9390277862548828, "learning_rate": 1.692862663375252e-05, "loss": 1.0117, "step": 1448 }, { "epoch": 0.27886836027713624, "grad_norm": 1.6273677349090576, "learning_rate": 1.692413063151309e-05, "loss": 0.9235, "step": 1449 }, { "epoch": 0.2790608160123172, "grad_norm": 1.0046995878219604, "learning_rate": 1.6919631938953835e-05, "loss": 0.9116, "step": 1450 }, { "epoch": 0.2792532717474981, "grad_norm": 1.7830642461776733, "learning_rate": 1.6915130557822698e-05, "loss": 0.9815, "step": 1451 }, { "epoch": 0.279445727482679, "grad_norm": 1.6447077989578247, "learning_rate": 1.691062648986865e-05, "loss": 0.9248, "step": 1452 }, { "epoch": 0.2796381832178599, "grad_norm": 0.9249933362007141, "learning_rate": 1.690611973684172e-05, "loss": 0.8766, "step": 1453 }, { "epoch": 0.2798306389530408, "grad_norm": 1.8931480646133423, "learning_rate": 1.6901610300492963e-05, "loss": 0.9857, "step": 1454 }, { "epoch": 0.28002309468822173, "grad_norm": 1.5578685998916626, "learning_rate": 1.6897098182574502e-05, "loss": 0.9355, "step": 1455 }, { "epoch": 0.2802155504234026, "grad_norm": 1.6910210847854614, "learning_rate": 1.689258338483947e-05, "loss": 1.067, "step": 1456 }, { "epoch": 0.2804080061585835, "grad_norm": 1.639620065689087, "learning_rate": 1.6888065909042074e-05, "loss": 1.0014, "step": 1457 }, { "epoch": 0.2806004618937644, "grad_norm": 1.7743351459503174, "learning_rate": 1.688354575693754e-05, "loss": 0.9396, "step": 1458 }, { "epoch": 0.2807929176289453, "grad_norm": 1.8035134077072144, "learning_rate": 1.687902293028214e-05, "loss": 0.9773, "step": 1459 }, { "epoch": 0.28098537336412627, "grad_norm": 1.723802924156189, "learning_rate": 1.6874497430833182e-05, "loss": 0.993, "step": 1460 }, { "epoch": 0.28117782909930716, "grad_norm": 1.8689061403274536, "learning_rate": 1.686996926034902e-05, "loss": 1.0971, "step": 1461 }, { "epoch": 0.28137028483448806, "grad_norm": 1.9459902048110962, "learning_rate": 1.6865438420589036e-05, "loss": 0.9967, "step": 1462 }, { "epoch": 0.28156274056966896, "grad_norm": 1.9046951532363892, "learning_rate": 1.6860904913313656e-05, "loss": 1.0684, "step": 1463 }, { "epoch": 0.2817551963048499, "grad_norm": 1.7166478633880615, "learning_rate": 1.6856368740284342e-05, "loss": 0.9766, "step": 1464 }, { "epoch": 0.2819476520400308, "grad_norm": 1.767199993133545, "learning_rate": 1.685182990326359e-05, "loss": 0.9946, "step": 1465 }, { "epoch": 0.2821401077752117, "grad_norm": 1.681044101715088, "learning_rate": 1.6847288404014937e-05, "loss": 1.0311, "step": 1466 }, { "epoch": 0.2823325635103926, "grad_norm": 1.7135965824127197, "learning_rate": 1.684274424430294e-05, "loss": 1.0712, "step": 1467 }, { "epoch": 0.2825250192455735, "grad_norm": 1.7834993600845337, "learning_rate": 1.68381974258932e-05, "loss": 1.0661, "step": 1468 }, { "epoch": 0.28271747498075445, "grad_norm": 1.8556705713272095, "learning_rate": 1.6833647950552357e-05, "loss": 0.9496, "step": 1469 }, { "epoch": 0.28290993071593534, "grad_norm": 2.6503992080688477, "learning_rate": 1.682909582004807e-05, "loss": 0.9948, "step": 1470 }, { "epoch": 0.28310238645111624, "grad_norm": 1.7716891765594482, "learning_rate": 1.682454103614904e-05, "loss": 0.988, "step": 1471 }, { "epoch": 0.28329484218629714, "grad_norm": 1.7190643548965454, "learning_rate": 1.6819983600624986e-05, "loss": 0.9067, "step": 1472 }, { "epoch": 0.2834872979214781, "grad_norm": 1.8176895380020142, "learning_rate": 1.6815423515246674e-05, "loss": 0.9504, "step": 1473 }, { "epoch": 0.283679753656659, "grad_norm": 1.6780424118041992, "learning_rate": 1.681086078178589e-05, "loss": 1.029, "step": 1474 }, { "epoch": 0.2838722093918399, "grad_norm": 1.7558003664016724, "learning_rate": 1.680629540201545e-05, "loss": 0.8811, "step": 1475 }, { "epoch": 0.2840646651270208, "grad_norm": 1.7794471979141235, "learning_rate": 1.6801727377709195e-05, "loss": 0.9793, "step": 1476 }, { "epoch": 0.2842571208622017, "grad_norm": 1.855647087097168, "learning_rate": 1.6797156710642003e-05, "loss": 0.9426, "step": 1477 }, { "epoch": 0.2844495765973826, "grad_norm": 1.7893325090408325, "learning_rate": 1.679258340258977e-05, "loss": 0.886, "step": 1478 }, { "epoch": 0.2846420323325635, "grad_norm": 1.7456836700439453, "learning_rate": 1.678800745532942e-05, "loss": 1.0244, "step": 1479 }, { "epoch": 0.2848344880677444, "grad_norm": 1.8668369054794312, "learning_rate": 1.6783428870638904e-05, "loss": 1.014, "step": 1480 }, { "epoch": 0.2850269438029253, "grad_norm": 1.6180843114852905, "learning_rate": 1.67788476502972e-05, "loss": 1.022, "step": 1481 }, { "epoch": 0.2852193995381062, "grad_norm": 1.9680200815200806, "learning_rate": 1.6774263796084304e-05, "loss": 1.0654, "step": 1482 }, { "epoch": 0.28541185527328716, "grad_norm": 0.9683606028556824, "learning_rate": 1.676967730978124e-05, "loss": 0.8917, "step": 1483 }, { "epoch": 0.28560431100846806, "grad_norm": 1.856134295463562, "learning_rate": 1.6765088193170055e-05, "loss": 0.9863, "step": 1484 }, { "epoch": 0.28579676674364896, "grad_norm": 1.7471874952316284, "learning_rate": 1.676049644803381e-05, "loss": 1.0727, "step": 1485 }, { "epoch": 0.28598922247882985, "grad_norm": 1.6205285787582397, "learning_rate": 1.6755902076156606e-05, "loss": 0.9402, "step": 1486 }, { "epoch": 0.2861816782140108, "grad_norm": 1.7780214548110962, "learning_rate": 1.6751305079323538e-05, "loss": 0.9591, "step": 1487 }, { "epoch": 0.2863741339491917, "grad_norm": 1.5687700510025024, "learning_rate": 1.6746705459320746e-05, "loss": 0.9756, "step": 1488 }, { "epoch": 0.2865665896843726, "grad_norm": 1.8139265775680542, "learning_rate": 1.6742103217935372e-05, "loss": 0.9644, "step": 1489 }, { "epoch": 0.2867590454195535, "grad_norm": 1.7928632497787476, "learning_rate": 1.6737498356955586e-05, "loss": 0.9729, "step": 1490 }, { "epoch": 0.2869515011547344, "grad_norm": 2.1075594425201416, "learning_rate": 1.6732890878170573e-05, "loss": 0.9307, "step": 1491 }, { "epoch": 0.28714395688991534, "grad_norm": 1.886930227279663, "learning_rate": 1.672828078337053e-05, "loss": 1.0713, "step": 1492 }, { "epoch": 0.28733641262509624, "grad_norm": 1.808154821395874, "learning_rate": 1.672366807434668e-05, "loss": 1.0121, "step": 1493 }, { "epoch": 0.28752886836027713, "grad_norm": 1.666575312614441, "learning_rate": 1.6719052752891262e-05, "loss": 1.0397, "step": 1494 }, { "epoch": 0.28772132409545803, "grad_norm": 1.7579233646392822, "learning_rate": 1.6714434820797515e-05, "loss": 0.9979, "step": 1495 }, { "epoch": 0.2879137798306389, "grad_norm": 1.8627372980117798, "learning_rate": 1.67098142798597e-05, "loss": 1.0059, "step": 1496 }, { "epoch": 0.2881062355658199, "grad_norm": 2.0003764629364014, "learning_rate": 1.670519113187311e-05, "loss": 0.9474, "step": 1497 }, { "epoch": 0.2882986913010008, "grad_norm": 1.7871371507644653, "learning_rate": 1.6700565378634022e-05, "loss": 1.0392, "step": 1498 }, { "epoch": 0.28849114703618167, "grad_norm": 1.9526821374893188, "learning_rate": 1.669593702193974e-05, "loss": 1.0429, "step": 1499 }, { "epoch": 0.28868360277136257, "grad_norm": 1.7681679725646973, "learning_rate": 1.6691306063588583e-05, "loss": 1.018, "step": 1500 }, { "epoch": 0.2888760585065435, "grad_norm": 1.7056353092193604, "learning_rate": 1.668667250537987e-05, "loss": 0.9716, "step": 1501 }, { "epoch": 0.2890685142417244, "grad_norm": 1.9142543077468872, "learning_rate": 1.668203634911394e-05, "loss": 1.0261, "step": 1502 }, { "epoch": 0.2892609699769053, "grad_norm": 1.7905254364013672, "learning_rate": 1.6677397596592134e-05, "loss": 0.9771, "step": 1503 }, { "epoch": 0.2894534257120862, "grad_norm": 2.0082154273986816, "learning_rate": 1.667275624961681e-05, "loss": 1.0646, "step": 1504 }, { "epoch": 0.2896458814472671, "grad_norm": 1.7965528964996338, "learning_rate": 1.666811230999132e-05, "loss": 1.0835, "step": 1505 }, { "epoch": 0.28983833718244806, "grad_norm": 2.104262351989746, "learning_rate": 1.6663465779520042e-05, "loss": 0.9383, "step": 1506 }, { "epoch": 0.29003079291762895, "grad_norm": 1.6659682989120483, "learning_rate": 1.6658816660008344e-05, "loss": 0.976, "step": 1507 }, { "epoch": 0.29022324865280985, "grad_norm": 1.7418100833892822, "learning_rate": 1.6654164953262614e-05, "loss": 0.9839, "step": 1508 }, { "epoch": 0.29041570438799075, "grad_norm": 1.9407572746276855, "learning_rate": 1.6649510661090233e-05, "loss": 1.0608, "step": 1509 }, { "epoch": 0.29060816012317164, "grad_norm": 1.951568603515625, "learning_rate": 1.6644853785299592e-05, "loss": 1.0136, "step": 1510 }, { "epoch": 0.2908006158583526, "grad_norm": 1.8873028755187988, "learning_rate": 1.6640194327700087e-05, "loss": 1.0437, "step": 1511 }, { "epoch": 0.2909930715935335, "grad_norm": 1.8386814594268799, "learning_rate": 1.6635532290102114e-05, "loss": 0.9148, "step": 1512 }, { "epoch": 0.2911855273287144, "grad_norm": 1.7825829982757568, "learning_rate": 1.6630867674317078e-05, "loss": 0.9925, "step": 1513 }, { "epoch": 0.2913779830638953, "grad_norm": 1.0290942192077637, "learning_rate": 1.6626200482157378e-05, "loss": 0.9888, "step": 1514 }, { "epoch": 0.29157043879907624, "grad_norm": 1.7475014925003052, "learning_rate": 1.6621530715436413e-05, "loss": 0.9864, "step": 1515 }, { "epoch": 0.29176289453425713, "grad_norm": 1.7863366603851318, "learning_rate": 1.6616858375968596e-05, "loss": 0.9223, "step": 1516 }, { "epoch": 0.29195535026943803, "grad_norm": 1.9573452472686768, "learning_rate": 1.6612183465569322e-05, "loss": 0.9945, "step": 1517 }, { "epoch": 0.2921478060046189, "grad_norm": 1.6399321556091309, "learning_rate": 1.6607505986054998e-05, "loss": 0.9353, "step": 1518 }, { "epoch": 0.2923402617397998, "grad_norm": 1.86214280128479, "learning_rate": 1.660282593924302e-05, "loss": 1.0001, "step": 1519 }, { "epoch": 0.2925327174749808, "grad_norm": 1.8999090194702148, "learning_rate": 1.6598143326951784e-05, "loss": 0.9641, "step": 1520 }, { "epoch": 0.29272517321016167, "grad_norm": 1.8859182596206665, "learning_rate": 1.659345815100069e-05, "loss": 1.0563, "step": 1521 }, { "epoch": 0.29291762894534257, "grad_norm": 1.853476881980896, "learning_rate": 1.6588770413210126e-05, "loss": 1.0315, "step": 1522 }, { "epoch": 0.29311008468052346, "grad_norm": 1.6301045417785645, "learning_rate": 1.6584080115401478e-05, "loss": 1.0298, "step": 1523 }, { "epoch": 0.29330254041570436, "grad_norm": 1.723157286643982, "learning_rate": 1.657938725939713e-05, "loss": 0.9411, "step": 1524 }, { "epoch": 0.2934949961508853, "grad_norm": 1.530401349067688, "learning_rate": 1.6574691847020446e-05, "loss": 1.0005, "step": 1525 }, { "epoch": 0.2936874518860662, "grad_norm": 1.8671878576278687, "learning_rate": 1.6569993880095807e-05, "loss": 0.9743, "step": 1526 }, { "epoch": 0.2938799076212471, "grad_norm": 1.7657887935638428, "learning_rate": 1.6565293360448565e-05, "loss": 1.0485, "step": 1527 }, { "epoch": 0.294072363356428, "grad_norm": 1.7405316829681396, "learning_rate": 1.6560590289905074e-05, "loss": 1.0707, "step": 1528 }, { "epoch": 0.29426481909160895, "grad_norm": 1.7758798599243164, "learning_rate": 1.655588467029268e-05, "loss": 1.0043, "step": 1529 }, { "epoch": 0.29445727482678985, "grad_norm": 2.0639078617095947, "learning_rate": 1.6551176503439713e-05, "loss": 0.9658, "step": 1530 }, { "epoch": 0.29464973056197075, "grad_norm": 2.048430919647217, "learning_rate": 1.6546465791175498e-05, "loss": 0.966, "step": 1531 }, { "epoch": 0.29484218629715164, "grad_norm": 1.7092403173446655, "learning_rate": 1.6541752535330345e-05, "loss": 1.0177, "step": 1532 }, { "epoch": 0.29503464203233254, "grad_norm": 1.7112072706222534, "learning_rate": 1.6537036737735563e-05, "loss": 0.9577, "step": 1533 }, { "epoch": 0.2952270977675135, "grad_norm": 1.8516780138015747, "learning_rate": 1.6532318400223427e-05, "loss": 1.0174, "step": 1534 }, { "epoch": 0.2954195535026944, "grad_norm": 1.908063292503357, "learning_rate": 1.6527597524627226e-05, "loss": 1.1176, "step": 1535 }, { "epoch": 0.2956120092378753, "grad_norm": 1.554583191871643, "learning_rate": 1.6522874112781213e-05, "loss": 1.0162, "step": 1536 }, { "epoch": 0.2958044649730562, "grad_norm": 1.9764091968536377, "learning_rate": 1.6518148166520638e-05, "loss": 0.9962, "step": 1537 }, { "epoch": 0.29599692070823713, "grad_norm": 1.731709361076355, "learning_rate": 1.651341968768173e-05, "loss": 0.9522, "step": 1538 }, { "epoch": 0.29618937644341803, "grad_norm": 1.852522611618042, "learning_rate": 1.6508688678101706e-05, "loss": 0.9886, "step": 1539 }, { "epoch": 0.2963818321785989, "grad_norm": 1.6258785724639893, "learning_rate": 1.6503955139618765e-05, "loss": 1.013, "step": 1540 }, { "epoch": 0.2965742879137798, "grad_norm": 1.6885334253311157, "learning_rate": 1.6499219074072087e-05, "loss": 0.8929, "step": 1541 }, { "epoch": 0.2967667436489607, "grad_norm": 1.7770051956176758, "learning_rate": 1.6494480483301836e-05, "loss": 0.9704, "step": 1542 }, { "epoch": 0.29695919938414167, "grad_norm": 1.6170932054519653, "learning_rate": 1.648973936914916e-05, "loss": 0.9815, "step": 1543 }, { "epoch": 0.29715165511932257, "grad_norm": 1.8974127769470215, "learning_rate": 1.6484995733456178e-05, "loss": 1.0293, "step": 1544 }, { "epoch": 0.29734411085450346, "grad_norm": 1.881015658378601, "learning_rate": 1.6480249578066e-05, "loss": 0.9446, "step": 1545 }, { "epoch": 0.29753656658968436, "grad_norm": 1.6574127674102783, "learning_rate": 1.6475500904822707e-05, "loss": 1.0716, "step": 1546 }, { "epoch": 0.29772902232486526, "grad_norm": 1.6878477334976196, "learning_rate": 1.647074971557136e-05, "loss": 0.9982, "step": 1547 }, { "epoch": 0.2979214780600462, "grad_norm": 1.538177728652954, "learning_rate": 1.6465996012157996e-05, "loss": 1.0161, "step": 1548 }, { "epoch": 0.2981139337952271, "grad_norm": 1.5863555669784546, "learning_rate": 1.646123979642964e-05, "loss": 0.9734, "step": 1549 }, { "epoch": 0.298306389530408, "grad_norm": 1.6166596412658691, "learning_rate": 1.6456481070234277e-05, "loss": 0.9638, "step": 1550 }, { "epoch": 0.2984988452655889, "grad_norm": 2.014636993408203, "learning_rate": 1.645171983542088e-05, "loss": 0.9543, "step": 1551 }, { "epoch": 0.29869130100076985, "grad_norm": 1.8636096715927124, "learning_rate": 1.6446956093839385e-05, "loss": 1.0927, "step": 1552 }, { "epoch": 0.29888375673595075, "grad_norm": 1.8549911975860596, "learning_rate": 1.644218984734071e-05, "loss": 0.9317, "step": 1553 }, { "epoch": 0.29907621247113164, "grad_norm": 1.7594201564788818, "learning_rate": 1.6437421097776754e-05, "loss": 0.9097, "step": 1554 }, { "epoch": 0.29926866820631254, "grad_norm": 1.8618059158325195, "learning_rate": 1.643264984700037e-05, "loss": 0.9809, "step": 1555 }, { "epoch": 0.29946112394149343, "grad_norm": 1.7036622762680054, "learning_rate": 1.6427876096865394e-05, "loss": 1.0425, "step": 1556 }, { "epoch": 0.2996535796766744, "grad_norm": 1.8091943264007568, "learning_rate": 1.6423099849226633e-05, "loss": 0.9146, "step": 1557 }, { "epoch": 0.2998460354118553, "grad_norm": 1.9414712190628052, "learning_rate": 1.6418321105939863e-05, "loss": 1.1124, "step": 1558 }, { "epoch": 0.3000384911470362, "grad_norm": 1.801892638206482, "learning_rate": 1.6413539868861828e-05, "loss": 0.9908, "step": 1559 }, { "epoch": 0.3002309468822171, "grad_norm": 1.772099494934082, "learning_rate": 1.6408756139850243e-05, "loss": 1.0428, "step": 1560 }, { "epoch": 0.30042340261739797, "grad_norm": 1.719939947128296, "learning_rate": 1.640396992076379e-05, "loss": 0.9012, "step": 1561 }, { "epoch": 0.3006158583525789, "grad_norm": 1.922352910041809, "learning_rate": 1.639918121346212e-05, "loss": 1.0245, "step": 1562 }, { "epoch": 0.3008083140877598, "grad_norm": 1.545076847076416, "learning_rate": 1.639439001980585e-05, "loss": 1.073, "step": 1563 }, { "epoch": 0.3010007698229407, "grad_norm": 1.891070008277893, "learning_rate": 1.638959634165656e-05, "loss": 0.9447, "step": 1564 }, { "epoch": 0.3011932255581216, "grad_norm": 1.6073846817016602, "learning_rate": 1.6384800180876804e-05, "loss": 0.9192, "step": 1565 }, { "epoch": 0.30138568129330257, "grad_norm": 1.8814278841018677, "learning_rate": 1.6380001539330088e-05, "loss": 1.0125, "step": 1566 }, { "epoch": 0.30157813702848346, "grad_norm": 1.7187384366989136, "learning_rate": 1.6375200418880895e-05, "loss": 0.9738, "step": 1567 }, { "epoch": 0.30177059276366436, "grad_norm": 1.744933843612671, "learning_rate": 1.637039682139466e-05, "loss": 0.9743, "step": 1568 }, { "epoch": 0.30196304849884525, "grad_norm": 1.863205075263977, "learning_rate": 1.6365590748737786e-05, "loss": 1.0585, "step": 1569 }, { "epoch": 0.30215550423402615, "grad_norm": 2.0287206172943115, "learning_rate": 1.636078220277764e-05, "loss": 1.0263, "step": 1570 }, { "epoch": 0.3023479599692071, "grad_norm": 1.5740458965301514, "learning_rate": 1.6355971185382547e-05, "loss": 0.9275, "step": 1571 }, { "epoch": 0.302540415704388, "grad_norm": 1.830632209777832, "learning_rate": 1.635115769842179e-05, "loss": 0.9783, "step": 1572 }, { "epoch": 0.3027328714395689, "grad_norm": 1.854601263999939, "learning_rate": 1.6346341743765616e-05, "loss": 1.0767, "step": 1573 }, { "epoch": 0.3029253271747498, "grad_norm": 1.8612595796585083, "learning_rate": 1.634152332328522e-05, "loss": 1.0014, "step": 1574 }, { "epoch": 0.3031177829099307, "grad_norm": 1.7086471319198608, "learning_rate": 1.633670243885278e-05, "loss": 0.9435, "step": 1575 }, { "epoch": 0.30331023864511164, "grad_norm": 1.895244836807251, "learning_rate": 1.6331879092341402e-05, "loss": 1.044, "step": 1576 }, { "epoch": 0.30350269438029254, "grad_norm": 1.9149186611175537, "learning_rate": 1.6327053285625164e-05, "loss": 1.0127, "step": 1577 }, { "epoch": 0.30369515011547343, "grad_norm": 1.6396968364715576, "learning_rate": 1.63222250205791e-05, "loss": 1.0014, "step": 1578 }, { "epoch": 0.30388760585065433, "grad_norm": 1.7869336605072021, "learning_rate": 1.6317394299079193e-05, "loss": 1.0501, "step": 1579 }, { "epoch": 0.3040800615858353, "grad_norm": 1.813430666923523, "learning_rate": 1.631256112300239e-05, "loss": 0.9436, "step": 1580 }, { "epoch": 0.3042725173210162, "grad_norm": 2.158643960952759, "learning_rate": 1.6307725494226586e-05, "loss": 1.033, "step": 1581 }, { "epoch": 0.3044649730561971, "grad_norm": 0.9738618731498718, "learning_rate": 1.630288741463062e-05, "loss": 0.8895, "step": 1582 }, { "epoch": 0.30465742879137797, "grad_norm": 1.8254077434539795, "learning_rate": 1.62980468860943e-05, "loss": 1.0085, "step": 1583 }, { "epoch": 0.30484988452655887, "grad_norm": 1.8007192611694336, "learning_rate": 1.6293203910498375e-05, "loss": 1.0082, "step": 1584 }, { "epoch": 0.3050423402617398, "grad_norm": 1.8346091508865356, "learning_rate": 1.628835848972455e-05, "loss": 0.9483, "step": 1585 }, { "epoch": 0.3052347959969207, "grad_norm": 1.9840954542160034, "learning_rate": 1.6283510625655474e-05, "loss": 0.966, "step": 1586 }, { "epoch": 0.3054272517321016, "grad_norm": 1.8244198560714722, "learning_rate": 1.627866032017475e-05, "loss": 1.0509, "step": 1587 }, { "epoch": 0.3056197074672825, "grad_norm": 1.9669748544692993, "learning_rate": 1.6273807575166927e-05, "loss": 1.0518, "step": 1588 }, { "epoch": 0.30581216320246346, "grad_norm": 1.747692584991455, "learning_rate": 1.6268952392517505e-05, "loss": 0.9898, "step": 1589 }, { "epoch": 0.30600461893764436, "grad_norm": 2.00764536857605, "learning_rate": 1.626409477411293e-05, "loss": 0.9123, "step": 1590 }, { "epoch": 0.30619707467282525, "grad_norm": 1.7182563543319702, "learning_rate": 1.6259234721840595e-05, "loss": 1.0407, "step": 1591 }, { "epoch": 0.30638953040800615, "grad_norm": 1.8612865209579468, "learning_rate": 1.625437223758883e-05, "loss": 1.0217, "step": 1592 }, { "epoch": 0.30658198614318705, "grad_norm": 1.7401247024536133, "learning_rate": 1.6249507323246926e-05, "loss": 1.0128, "step": 1593 }, { "epoch": 0.306774441878368, "grad_norm": 1.6441352367401123, "learning_rate": 1.6244639980705104e-05, "loss": 1.0332, "step": 1594 }, { "epoch": 0.3069668976135489, "grad_norm": 1.6691603660583496, "learning_rate": 1.623977021185454e-05, "loss": 1.0467, "step": 1595 }, { "epoch": 0.3071593533487298, "grad_norm": 1.7422953844070435, "learning_rate": 1.6234898018587336e-05, "loss": 0.9704, "step": 1596 }, { "epoch": 0.3073518090839107, "grad_norm": 1.9007364511489868, "learning_rate": 1.623002340279656e-05, "loss": 0.9696, "step": 1597 }, { "epoch": 0.3075442648190916, "grad_norm": 1.841038465499878, "learning_rate": 1.6225146366376198e-05, "loss": 0.928, "step": 1598 }, { "epoch": 0.30773672055427254, "grad_norm": 1.6869525909423828, "learning_rate": 1.622026691122119e-05, "loss": 0.9035, "step": 1599 }, { "epoch": 0.30792917628945343, "grad_norm": 1.8111488819122314, "learning_rate": 1.621538503922741e-05, "loss": 0.9897, "step": 1600 }, { "epoch": 0.30812163202463433, "grad_norm": 1.7562575340270996, "learning_rate": 1.6210500752291682e-05, "loss": 1.0195, "step": 1601 }, { "epoch": 0.3083140877598152, "grad_norm": 1.8421173095703125, "learning_rate": 1.620561405231175e-05, "loss": 1.0301, "step": 1602 }, { "epoch": 0.3085065434949962, "grad_norm": 1.721318244934082, "learning_rate": 1.6200724941186305e-05, "loss": 1.0096, "step": 1603 }, { "epoch": 0.3086989992301771, "grad_norm": 1.7768428325653076, "learning_rate": 1.6195833420814983e-05, "loss": 1.0028, "step": 1604 }, { "epoch": 0.30889145496535797, "grad_norm": 2.0415077209472656, "learning_rate": 1.6190939493098344e-05, "loss": 0.9767, "step": 1605 }, { "epoch": 0.30908391070053887, "grad_norm": 1.6920299530029297, "learning_rate": 1.6186043159937884e-05, "loss": 0.9573, "step": 1606 }, { "epoch": 0.30927636643571976, "grad_norm": 1.8492498397827148, "learning_rate": 1.6181144423236038e-05, "loss": 1.0226, "step": 1607 }, { "epoch": 0.3094688221709007, "grad_norm": 0.9862644672393799, "learning_rate": 1.617624328489618e-05, "loss": 0.8903, "step": 1608 }, { "epoch": 0.3096612779060816, "grad_norm": 1.7790777683258057, "learning_rate": 1.6171339746822606e-05, "loss": 1.0072, "step": 1609 }, { "epoch": 0.3098537336412625, "grad_norm": 1.74509859085083, "learning_rate": 1.616643381092055e-05, "loss": 1.0318, "step": 1610 }, { "epoch": 0.3100461893764434, "grad_norm": 1.7997702360153198, "learning_rate": 1.616152547909618e-05, "loss": 0.9966, "step": 1611 }, { "epoch": 0.3102386451116243, "grad_norm": 1.852412462234497, "learning_rate": 1.6156614753256583e-05, "loss": 1.0166, "step": 1612 }, { "epoch": 0.31043110084680525, "grad_norm": 1.867826223373413, "learning_rate": 1.6151701635309797e-05, "loss": 0.986, "step": 1613 }, { "epoch": 0.31062355658198615, "grad_norm": 1.6695712804794312, "learning_rate": 1.6146786127164773e-05, "loss": 0.9579, "step": 1614 }, { "epoch": 0.31081601231716705, "grad_norm": 1.7458585500717163, "learning_rate": 1.6141868230731393e-05, "loss": 0.9957, "step": 1615 }, { "epoch": 0.31100846805234794, "grad_norm": 1.7383522987365723, "learning_rate": 1.6136947947920477e-05, "loss": 1.0399, "step": 1616 }, { "epoch": 0.3112009237875289, "grad_norm": 1.6937193870544434, "learning_rate": 1.6132025280643753e-05, "loss": 0.9394, "step": 1617 }, { "epoch": 0.3113933795227098, "grad_norm": 1.7102807760238647, "learning_rate": 1.6127100230813897e-05, "loss": 1.0342, "step": 1618 }, { "epoch": 0.3115858352578907, "grad_norm": 0.9480326771736145, "learning_rate": 1.6122172800344494e-05, "loss": 0.972, "step": 1619 }, { "epoch": 0.3117782909930716, "grad_norm": 1.663881540298462, "learning_rate": 1.6117242991150064e-05, "loss": 0.9103, "step": 1620 }, { "epoch": 0.3119707467282525, "grad_norm": 1.8087718486785889, "learning_rate": 1.611231080514605e-05, "loss": 0.9485, "step": 1621 }, { "epoch": 0.31216320246343343, "grad_norm": 1.5888209342956543, "learning_rate": 1.610737624424881e-05, "loss": 1.0429, "step": 1622 }, { "epoch": 0.3123556581986143, "grad_norm": 0.9585890769958496, "learning_rate": 1.6102439310375637e-05, "loss": 0.9587, "step": 1623 }, { "epoch": 0.3125481139337952, "grad_norm": 1.873613715171814, "learning_rate": 1.609750000544474e-05, "loss": 0.9504, "step": 1624 }, { "epoch": 0.3127405696689761, "grad_norm": 0.9027090072631836, "learning_rate": 1.6092558331375245e-05, "loss": 0.9498, "step": 1625 }, { "epoch": 0.312933025404157, "grad_norm": 1.7109851837158203, "learning_rate": 1.608761429008721e-05, "loss": 0.9589, "step": 1626 }, { "epoch": 0.31312548113933797, "grad_norm": 1.8576031923294067, "learning_rate": 1.6082667883501596e-05, "loss": 0.908, "step": 1627 }, { "epoch": 0.31331793687451887, "grad_norm": 1.7118754386901855, "learning_rate": 1.6077719113540303e-05, "loss": 1.0444, "step": 1628 }, { "epoch": 0.31351039260969976, "grad_norm": 1.8047579526901245, "learning_rate": 1.607276798212613e-05, "loss": 0.9955, "step": 1629 }, { "epoch": 0.31370284834488066, "grad_norm": 1.9328746795654297, "learning_rate": 1.6067814491182806e-05, "loss": 0.9247, "step": 1630 }, { "epoch": 0.3138953040800616, "grad_norm": 1.8870935440063477, "learning_rate": 1.606285864263498e-05, "loss": 0.9693, "step": 1631 }, { "epoch": 0.3140877598152425, "grad_norm": 1.7145658731460571, "learning_rate": 1.60579004384082e-05, "loss": 1.0732, "step": 1632 }, { "epoch": 0.3142802155504234, "grad_norm": 1.938612461090088, "learning_rate": 1.6052939880428942e-05, "loss": 0.9489, "step": 1633 }, { "epoch": 0.3144726712856043, "grad_norm": 1.7863969802856445, "learning_rate": 1.6047976970624603e-05, "loss": 0.9511, "step": 1634 }, { "epoch": 0.3146651270207852, "grad_norm": 1.7148983478546143, "learning_rate": 1.6043011710923476e-05, "loss": 0.9891, "step": 1635 }, { "epoch": 0.31485758275596615, "grad_norm": 1.7553948163986206, "learning_rate": 1.6038044103254775e-05, "loss": 1.013, "step": 1636 }, { "epoch": 0.31505003849114704, "grad_norm": 1.8945555686950684, "learning_rate": 1.603307414954863e-05, "loss": 0.8464, "step": 1637 }, { "epoch": 0.31524249422632794, "grad_norm": 1.742958903312683, "learning_rate": 1.6028101851736083e-05, "loss": 1.0019, "step": 1638 }, { "epoch": 0.31543494996150884, "grad_norm": 1.6972116231918335, "learning_rate": 1.6023127211749082e-05, "loss": 0.994, "step": 1639 }, { "epoch": 0.3156274056966898, "grad_norm": 1.7025740146636963, "learning_rate": 1.6018150231520486e-05, "loss": 1.0272, "step": 1640 }, { "epoch": 0.3158198614318707, "grad_norm": 1.8751802444458008, "learning_rate": 1.601317091298406e-05, "loss": 0.9727, "step": 1641 }, { "epoch": 0.3160123171670516, "grad_norm": 1.860803484916687, "learning_rate": 1.6008189258074483e-05, "loss": 0.9981, "step": 1642 }, { "epoch": 0.3162047729022325, "grad_norm": 1.6714084148406982, "learning_rate": 1.600320526872735e-05, "loss": 1.0339, "step": 1643 }, { "epoch": 0.3163972286374134, "grad_norm": 1.7508667707443237, "learning_rate": 1.599821894687914e-05, "loss": 1.0138, "step": 1644 }, { "epoch": 0.3165896843725943, "grad_norm": 1.7288819551467896, "learning_rate": 1.5993230294467257e-05, "loss": 0.9381, "step": 1645 }, { "epoch": 0.3167821401077752, "grad_norm": 1.8353605270385742, "learning_rate": 1.5988239313430004e-05, "loss": 1.0066, "step": 1646 }, { "epoch": 0.3169745958429561, "grad_norm": 1.7945586442947388, "learning_rate": 1.5983246005706592e-05, "loss": 0.9669, "step": 1647 }, { "epoch": 0.317167051578137, "grad_norm": 1.6595655679702759, "learning_rate": 1.5978250373237132e-05, "loss": 0.9667, "step": 1648 }, { "epoch": 0.3173595073133179, "grad_norm": 1.6445173025131226, "learning_rate": 1.597325241796264e-05, "loss": 0.9328, "step": 1649 }, { "epoch": 0.31755196304849886, "grad_norm": 0.9720272421836853, "learning_rate": 1.5968252141825038e-05, "loss": 0.8729, "step": 1650 }, { "epoch": 0.31774441878367976, "grad_norm": 1.6678732633590698, "learning_rate": 1.5963249546767144e-05, "loss": 1.0519, "step": 1651 }, { "epoch": 0.31793687451886066, "grad_norm": 1.6013301610946655, "learning_rate": 1.5958244634732673e-05, "loss": 1.0462, "step": 1652 }, { "epoch": 0.31812933025404155, "grad_norm": 1.7456437349319458, "learning_rate": 1.595323740766626e-05, "loss": 0.9508, "step": 1653 }, { "epoch": 0.3183217859892225, "grad_norm": 1.9124515056610107, "learning_rate": 1.5948227867513416e-05, "loss": 0.9757, "step": 1654 }, { "epoch": 0.3185142417244034, "grad_norm": 2.002638578414917, "learning_rate": 1.5943216016220563e-05, "loss": 1.0226, "step": 1655 }, { "epoch": 0.3187066974595843, "grad_norm": 1.6422607898712158, "learning_rate": 1.5938201855735017e-05, "loss": 0.9394, "step": 1656 }, { "epoch": 0.3188991531947652, "grad_norm": 1.7120252847671509, "learning_rate": 1.5933185388004998e-05, "loss": 0.9428, "step": 1657 }, { "epoch": 0.3190916089299461, "grad_norm": 1.951762318611145, "learning_rate": 1.5928166614979613e-05, "loss": 0.8596, "step": 1658 }, { "epoch": 0.31928406466512704, "grad_norm": 1.8206475973129272, "learning_rate": 1.5923145538608874e-05, "loss": 0.9487, "step": 1659 }, { "epoch": 0.31947652040030794, "grad_norm": 1.7178988456726074, "learning_rate": 1.591812216084368e-05, "loss": 0.9843, "step": 1660 }, { "epoch": 0.31966897613548884, "grad_norm": 1.9629712104797363, "learning_rate": 1.5913096483635827e-05, "loss": 1.0154, "step": 1661 }, { "epoch": 0.31986143187066973, "grad_norm": 1.7199310064315796, "learning_rate": 1.5908068508938e-05, "loss": 0.9709, "step": 1662 }, { "epoch": 0.32005388760585063, "grad_norm": 0.8834888339042664, "learning_rate": 1.5903038238703795e-05, "loss": 0.9322, "step": 1663 }, { "epoch": 0.3202463433410316, "grad_norm": 1.8330293893814087, "learning_rate": 1.5898005674887673e-05, "loss": 0.9693, "step": 1664 }, { "epoch": 0.3204387990762125, "grad_norm": 1.6615421772003174, "learning_rate": 1.589297081944501e-05, "loss": 1.0638, "step": 1665 }, { "epoch": 0.3206312548113934, "grad_norm": 1.9187180995941162, "learning_rate": 1.5887933674332048e-05, "loss": 1.0317, "step": 1666 }, { "epoch": 0.32082371054657427, "grad_norm": 1.7260987758636475, "learning_rate": 1.5882894241505946e-05, "loss": 0.9897, "step": 1667 }, { "epoch": 0.3210161662817552, "grad_norm": 1.9676867723464966, "learning_rate": 1.5877852522924733e-05, "loss": 1.033, "step": 1668 }, { "epoch": 0.3212086220169361, "grad_norm": 1.7443852424621582, "learning_rate": 1.5872808520547334e-05, "loss": 0.9672, "step": 1669 }, { "epoch": 0.321401077752117, "grad_norm": 1.781541109085083, "learning_rate": 1.5867762236333553e-05, "loss": 1.0343, "step": 1670 }, { "epoch": 0.3215935334872979, "grad_norm": 1.89210844039917, "learning_rate": 1.5862713672244092e-05, "loss": 1.0462, "step": 1671 }, { "epoch": 0.3217859892224788, "grad_norm": 1.545162558555603, "learning_rate": 1.585766283024053e-05, "loss": 0.9067, "step": 1672 }, { "epoch": 0.32197844495765976, "grad_norm": 1.6994256973266602, "learning_rate": 1.585260971228534e-05, "loss": 0.9786, "step": 1673 }, { "epoch": 0.32217090069284066, "grad_norm": 1.7282161712646484, "learning_rate": 1.5847554320341865e-05, "loss": 0.9445, "step": 1674 }, { "epoch": 0.32236335642802155, "grad_norm": 0.9777271151542664, "learning_rate": 1.5842496656374347e-05, "loss": 0.9919, "step": 1675 }, { "epoch": 0.32255581216320245, "grad_norm": 1.585505723953247, "learning_rate": 1.5837436722347902e-05, "loss": 0.9592, "step": 1676 }, { "epoch": 0.32274826789838335, "grad_norm": 2.0595457553863525, "learning_rate": 1.5832374520228527e-05, "loss": 0.9867, "step": 1677 }, { "epoch": 0.3229407236335643, "grad_norm": 1.7800649404525757, "learning_rate": 1.582731005198311e-05, "loss": 0.9797, "step": 1678 }, { "epoch": 0.3231331793687452, "grad_norm": 1.8043187856674194, "learning_rate": 1.5822243319579405e-05, "loss": 1.0029, "step": 1679 }, { "epoch": 0.3233256351039261, "grad_norm": 1.6553676128387451, "learning_rate": 1.581717432498606e-05, "loss": 0.9603, "step": 1680 }, { "epoch": 0.323518090839107, "grad_norm": 1.8246638774871826, "learning_rate": 1.5812103070172592e-05, "loss": 1.0288, "step": 1681 }, { "epoch": 0.32371054657428794, "grad_norm": 1.6993404626846313, "learning_rate": 1.5807029557109398e-05, "loss": 1.02, "step": 1682 }, { "epoch": 0.32390300230946883, "grad_norm": 1.899330973625183, "learning_rate": 1.580195378776776e-05, "loss": 0.9438, "step": 1683 }, { "epoch": 0.32409545804464973, "grad_norm": 0.9404033422470093, "learning_rate": 1.5796875764119826e-05, "loss": 0.917, "step": 1684 }, { "epoch": 0.3242879137798306, "grad_norm": 1.5981310606002808, "learning_rate": 1.5791795488138623e-05, "loss": 1.0151, "step": 1685 }, { "epoch": 0.3244803695150115, "grad_norm": 1.7941358089447021, "learning_rate": 1.578671296179806e-05, "loss": 0.9073, "step": 1686 }, { "epoch": 0.3246728252501925, "grad_norm": 1.7918518781661987, "learning_rate": 1.578162818707291e-05, "loss": 1.059, "step": 1687 }, { "epoch": 0.3248652809853734, "grad_norm": 1.680680274963379, "learning_rate": 1.577654116593883e-05, "loss": 1.035, "step": 1688 }, { "epoch": 0.32505773672055427, "grad_norm": 1.7011933326721191, "learning_rate": 1.577145190037234e-05, "loss": 1.0145, "step": 1689 }, { "epoch": 0.32525019245573517, "grad_norm": 1.8802440166473389, "learning_rate": 1.5766360392350837e-05, "loss": 1.0182, "step": 1690 }, { "epoch": 0.32544264819091606, "grad_norm": 1.7538574934005737, "learning_rate": 1.5761266643852587e-05, "loss": 0.8889, "step": 1691 }, { "epoch": 0.325635103926097, "grad_norm": 1.792249083518982, "learning_rate": 1.575617065685674e-05, "loss": 1.0816, "step": 1692 }, { "epoch": 0.3258275596612779, "grad_norm": 2.0103187561035156, "learning_rate": 1.5751072433343285e-05, "loss": 0.9355, "step": 1693 }, { "epoch": 0.3260200153964588, "grad_norm": 1.6156758069992065, "learning_rate": 1.5745971975293114e-05, "loss": 1.0246, "step": 1694 }, { "epoch": 0.3262124711316397, "grad_norm": 1.703883171081543, "learning_rate": 1.574086928468797e-05, "loss": 0.891, "step": 1695 }, { "epoch": 0.32640492686682065, "grad_norm": 1.8960217237472534, "learning_rate": 1.573576436351046e-05, "loss": 0.9745, "step": 1696 }, { "epoch": 0.32659738260200155, "grad_norm": 1.6109836101531982, "learning_rate": 1.5730657213744072e-05, "loss": 0.9262, "step": 1697 }, { "epoch": 0.32678983833718245, "grad_norm": 1.6797646284103394, "learning_rate": 1.5725547837373142e-05, "loss": 1.0308, "step": 1698 }, { "epoch": 0.32698229407236334, "grad_norm": 1.8447962999343872, "learning_rate": 1.5720436236382885e-05, "loss": 0.9636, "step": 1699 }, { "epoch": 0.32717474980754424, "grad_norm": 1.770391583442688, "learning_rate": 1.5715322412759374e-05, "loss": 0.9765, "step": 1700 }, { "epoch": 0.3273672055427252, "grad_norm": 1.7947641611099243, "learning_rate": 1.5710206368489555e-05, "loss": 0.9823, "step": 1701 }, { "epoch": 0.3275596612779061, "grad_norm": 1.906477928161621, "learning_rate": 1.5705088105561217e-05, "loss": 1.0531, "step": 1702 }, { "epoch": 0.327752117013087, "grad_norm": 1.8414548635482788, "learning_rate": 1.5699967625963032e-05, "loss": 1.0318, "step": 1703 }, { "epoch": 0.3279445727482679, "grad_norm": 1.9499144554138184, "learning_rate": 1.569484493168452e-05, "loss": 0.981, "step": 1704 }, { "epoch": 0.32813702848344883, "grad_norm": 1.6564552783966064, "learning_rate": 1.568972002471607e-05, "loss": 1.0188, "step": 1705 }, { "epoch": 0.32832948421862973, "grad_norm": 2.4806151390075684, "learning_rate": 1.5684592907048925e-05, "loss": 0.9342, "step": 1706 }, { "epoch": 0.3285219399538106, "grad_norm": 1.6865906715393066, "learning_rate": 1.5679463580675187e-05, "loss": 0.9732, "step": 1707 }, { "epoch": 0.3287143956889915, "grad_norm": 2.062652587890625, "learning_rate": 1.567433204758782e-05, "loss": 0.91, "step": 1708 }, { "epoch": 0.3289068514241724, "grad_norm": 1.7929702997207642, "learning_rate": 1.5669198309780644e-05, "loss": 1.0079, "step": 1709 }, { "epoch": 0.32909930715935337, "grad_norm": 1.7934985160827637, "learning_rate": 1.566406236924833e-05, "loss": 1.0278, "step": 1710 }, { "epoch": 0.32929176289453427, "grad_norm": 1.793951392173767, "learning_rate": 1.5658924227986415e-05, "loss": 1.0016, "step": 1711 }, { "epoch": 0.32948421862971516, "grad_norm": 1.7239779233932495, "learning_rate": 1.5653783887991282e-05, "loss": 0.9793, "step": 1712 }, { "epoch": 0.32967667436489606, "grad_norm": 1.7481112480163574, "learning_rate": 1.564864135126018e-05, "loss": 1.024, "step": 1713 }, { "epoch": 0.32986913010007696, "grad_norm": 1.8553564548492432, "learning_rate": 1.564349661979119e-05, "loss": 0.9513, "step": 1714 }, { "epoch": 0.3300615858352579, "grad_norm": 1.645546555519104, "learning_rate": 1.5638349695583276e-05, "loss": 1.0239, "step": 1715 }, { "epoch": 0.3302540415704388, "grad_norm": 1.8084485530853271, "learning_rate": 1.563320058063622e-05, "loss": 0.9394, "step": 1716 }, { "epoch": 0.3304464973056197, "grad_norm": 1.802618145942688, "learning_rate": 1.5628049276950687e-05, "loss": 0.9635, "step": 1717 }, { "epoch": 0.3306389530408006, "grad_norm": 1.8893779516220093, "learning_rate": 1.5622895786528173e-05, "loss": 0.9452, "step": 1718 }, { "epoch": 0.33083140877598155, "grad_norm": 1.7637345790863037, "learning_rate": 1.5617740111371023e-05, "loss": 1.0405, "step": 1719 }, { "epoch": 0.33102386451116245, "grad_norm": 1.957239031791687, "learning_rate": 1.5612582253482444e-05, "loss": 0.8714, "step": 1720 }, { "epoch": 0.33121632024634334, "grad_norm": 1.9569957256317139, "learning_rate": 1.560742221486648e-05, "loss": 0.9653, "step": 1721 }, { "epoch": 0.33140877598152424, "grad_norm": 0.9606925249099731, "learning_rate": 1.5602259997528028e-05, "loss": 0.9062, "step": 1722 }, { "epoch": 0.33160123171670514, "grad_norm": 1.869847297668457, "learning_rate": 1.5597095603472828e-05, "loss": 1.0369, "step": 1723 }, { "epoch": 0.3317936874518861, "grad_norm": 1.8176629543304443, "learning_rate": 1.5591929034707468e-05, "loss": 0.9297, "step": 1724 }, { "epoch": 0.331986143187067, "grad_norm": 1.8708559274673462, "learning_rate": 1.5586760293239383e-05, "loss": 0.9786, "step": 1725 }, { "epoch": 0.3321785989222479, "grad_norm": 1.7141287326812744, "learning_rate": 1.5581589381076843e-05, "loss": 1.027, "step": 1726 }, { "epoch": 0.3323710546574288, "grad_norm": 1.7462016344070435, "learning_rate": 1.557641630022897e-05, "loss": 0.9607, "step": 1727 }, { "epoch": 0.3325635103926097, "grad_norm": 1.7583905458450317, "learning_rate": 1.5571241052705724e-05, "loss": 1.0236, "step": 1728 }, { "epoch": 0.3327559661277906, "grad_norm": 1.8696091175079346, "learning_rate": 1.5566063640517918e-05, "loss": 0.9689, "step": 1729 }, { "epoch": 0.3329484218629715, "grad_norm": 1.852078914642334, "learning_rate": 1.5560884065677188e-05, "loss": 0.9935, "step": 1730 }, { "epoch": 0.3331408775981524, "grad_norm": 1.9050599336624146, "learning_rate": 1.5555702330196024e-05, "loss": 1.0304, "step": 1731 }, { "epoch": 0.3333333333333333, "grad_norm": 1.8081910610198975, "learning_rate": 1.5550518436087753e-05, "loss": 0.9328, "step": 1732 }, { "epoch": 0.33352578906851427, "grad_norm": 1.6070475578308105, "learning_rate": 1.554533238536653e-05, "loss": 1.0388, "step": 1733 }, { "epoch": 0.33371824480369516, "grad_norm": 1.7032333612442017, "learning_rate": 1.5540144180047367e-05, "loss": 1.0205, "step": 1734 }, { "epoch": 0.33391070053887606, "grad_norm": 1.6502132415771484, "learning_rate": 1.5534953822146097e-05, "loss": 0.9746, "step": 1735 }, { "epoch": 0.33410315627405696, "grad_norm": 1.8607031106948853, "learning_rate": 1.5529761313679396e-05, "loss": 1.05, "step": 1736 }, { "epoch": 0.33429561200923785, "grad_norm": 1.7787359952926636, "learning_rate": 1.5524566656664777e-05, "loss": 1.0195, "step": 1737 }, { "epoch": 0.3344880677444188, "grad_norm": 1.7438163757324219, "learning_rate": 1.5519369853120584e-05, "loss": 0.936, "step": 1738 }, { "epoch": 0.3346805234795997, "grad_norm": 1.5819883346557617, "learning_rate": 1.5514170905065996e-05, "loss": 1.0686, "step": 1739 }, { "epoch": 0.3348729792147806, "grad_norm": 1.6069012880325317, "learning_rate": 1.5508969814521026e-05, "loss": 1.0857, "step": 1740 }, { "epoch": 0.3350654349499615, "grad_norm": 1.8491158485412598, "learning_rate": 1.5503766583506522e-05, "loss": 0.9816, "step": 1741 }, { "epoch": 0.3352578906851424, "grad_norm": 1.7476420402526855, "learning_rate": 1.549856121404416e-05, "loss": 0.937, "step": 1742 }, { "epoch": 0.33545034642032334, "grad_norm": 1.7756730318069458, "learning_rate": 1.5493353708156442e-05, "loss": 0.9633, "step": 1743 }, { "epoch": 0.33564280215550424, "grad_norm": 1.619093894958496, "learning_rate": 1.548814406786671e-05, "loss": 1.0837, "step": 1744 }, { "epoch": 0.33583525789068513, "grad_norm": 1.8196501731872559, "learning_rate": 1.548293229519914e-05, "loss": 0.9647, "step": 1745 }, { "epoch": 0.33602771362586603, "grad_norm": 1.684403657913208, "learning_rate": 1.5477718392178716e-05, "loss": 0.8912, "step": 1746 }, { "epoch": 0.336220169361047, "grad_norm": 1.728331208229065, "learning_rate": 1.5472502360831265e-05, "loss": 0.9572, "step": 1747 }, { "epoch": 0.3364126250962279, "grad_norm": 1.669525384902954, "learning_rate": 1.5467284203183437e-05, "loss": 1.1018, "step": 1748 }, { "epoch": 0.3366050808314088, "grad_norm": 1.7439517974853516, "learning_rate": 1.5462063921262708e-05, "loss": 1.0457, "step": 1749 }, { "epoch": 0.3367975365665897, "grad_norm": 1.6491779088974, "learning_rate": 1.5456841517097385e-05, "loss": 1.0158, "step": 1750 }, { "epoch": 0.33698999230177057, "grad_norm": 1.7667906284332275, "learning_rate": 1.545161699271659e-05, "loss": 0.9586, "step": 1751 }, { "epoch": 0.3371824480369515, "grad_norm": 1.715625524520874, "learning_rate": 1.5446390350150272e-05, "loss": 0.99, "step": 1752 }, { "epoch": 0.3373749037721324, "grad_norm": 1.7603507041931152, "learning_rate": 1.544116159142921e-05, "loss": 0.9805, "step": 1753 }, { "epoch": 0.3375673595073133, "grad_norm": 0.9383662939071655, "learning_rate": 1.5435930718584993e-05, "loss": 0.9317, "step": 1754 }, { "epoch": 0.3377598152424942, "grad_norm": 1.713032841682434, "learning_rate": 1.5430697733650045e-05, "loss": 1.0013, "step": 1755 }, { "epoch": 0.33795227097767516, "grad_norm": 1.783747911453247, "learning_rate": 1.5425462638657597e-05, "loss": 0.9569, "step": 1756 }, { "epoch": 0.33814472671285606, "grad_norm": 1.689237356185913, "learning_rate": 1.542022543564171e-05, "loss": 1.0033, "step": 1757 }, { "epoch": 0.33833718244803695, "grad_norm": 1.646461009979248, "learning_rate": 1.541498612663726e-05, "loss": 1.0029, "step": 1758 }, { "epoch": 0.33852963818321785, "grad_norm": 1.6828275918960571, "learning_rate": 1.5409744713679942e-05, "loss": 0.9929, "step": 1759 }, { "epoch": 0.33872209391839875, "grad_norm": 1.8641035556793213, "learning_rate": 1.5404501198806267e-05, "loss": 0.966, "step": 1760 }, { "epoch": 0.3389145496535797, "grad_norm": 1.7769709825515747, "learning_rate": 1.5399255584053568e-05, "loss": 0.9975, "step": 1761 }, { "epoch": 0.3391070053887606, "grad_norm": 1.7866312265396118, "learning_rate": 1.5394007871459983e-05, "loss": 1.0045, "step": 1762 }, { "epoch": 0.3392994611239415, "grad_norm": 1.7538526058197021, "learning_rate": 1.538875806306448e-05, "loss": 1.0361, "step": 1763 }, { "epoch": 0.3394919168591224, "grad_norm": 2.205838441848755, "learning_rate": 1.5383506160906826e-05, "loss": 0.9743, "step": 1764 }, { "epoch": 0.3396843725943033, "grad_norm": 1.7115983963012695, "learning_rate": 1.5378252167027613e-05, "loss": 1.0269, "step": 1765 }, { "epoch": 0.33987682832948424, "grad_norm": 1.7475506067276, "learning_rate": 1.5372996083468242e-05, "loss": 0.941, "step": 1766 }, { "epoch": 0.34006928406466513, "grad_norm": 1.8076540231704712, "learning_rate": 1.536773791227092e-05, "loss": 0.9403, "step": 1767 }, { "epoch": 0.34026173979984603, "grad_norm": 1.8480545282363892, "learning_rate": 1.5362477655478677e-05, "loss": 1.0338, "step": 1768 }, { "epoch": 0.3404541955350269, "grad_norm": 1.6669305562973022, "learning_rate": 1.535721531513535e-05, "loss": 1.0443, "step": 1769 }, { "epoch": 0.3406466512702079, "grad_norm": 1.8368251323699951, "learning_rate": 1.535195089328557e-05, "loss": 0.9653, "step": 1770 }, { "epoch": 0.3408391070053888, "grad_norm": 0.9453438520431519, "learning_rate": 1.5346684391974792e-05, "loss": 0.9354, "step": 1771 }, { "epoch": 0.34103156274056967, "grad_norm": 1.7002607583999634, "learning_rate": 1.534141581324929e-05, "loss": 1.0769, "step": 1772 }, { "epoch": 0.34122401847575057, "grad_norm": 1.8152868747711182, "learning_rate": 1.533614515915612e-05, "loss": 1.0388, "step": 1773 }, { "epoch": 0.34141647421093146, "grad_norm": 1.757623553276062, "learning_rate": 1.533087243174315e-05, "loss": 1.04, "step": 1774 }, { "epoch": 0.3416089299461124, "grad_norm": 1.919758915901184, "learning_rate": 1.5325597633059075e-05, "loss": 0.9511, "step": 1775 }, { "epoch": 0.3418013856812933, "grad_norm": 1.7335333824157715, "learning_rate": 1.5320320765153367e-05, "loss": 1.0302, "step": 1776 }, { "epoch": 0.3419938414164742, "grad_norm": 1.8738481998443604, "learning_rate": 1.5315041830076317e-05, "loss": 0.9744, "step": 1777 }, { "epoch": 0.3421862971516551, "grad_norm": 1.7490358352661133, "learning_rate": 1.530976082987902e-05, "loss": 0.9796, "step": 1778 }, { "epoch": 0.342378752886836, "grad_norm": 1.8043572902679443, "learning_rate": 1.5304477766613365e-05, "loss": 1.0082, "step": 1779 }, { "epoch": 0.34257120862201695, "grad_norm": 1.5871425867080688, "learning_rate": 1.529919264233205e-05, "loss": 1.03, "step": 1780 }, { "epoch": 0.34276366435719785, "grad_norm": 1.6797513961791992, "learning_rate": 1.529390545908857e-05, "loss": 1.0928, "step": 1781 }, { "epoch": 0.34295612009237875, "grad_norm": 1.7837003469467163, "learning_rate": 1.528861621893722e-05, "loss": 1.0261, "step": 1782 }, { "epoch": 0.34314857582755964, "grad_norm": 1.6577980518341064, "learning_rate": 1.52833249239331e-05, "loss": 1.0051, "step": 1783 }, { "epoch": 0.3433410315627406, "grad_norm": 1.6752525568008423, "learning_rate": 1.527803157613209e-05, "loss": 0.9512, "step": 1784 }, { "epoch": 0.3435334872979215, "grad_norm": 1.9407802820205688, "learning_rate": 1.5272736177590902e-05, "loss": 0.9585, "step": 1785 }, { "epoch": 0.3437259430331024, "grad_norm": 1.8005989789962769, "learning_rate": 1.526743873036701e-05, "loss": 1.0382, "step": 1786 }, { "epoch": 0.3439183987682833, "grad_norm": 1.700411081314087, "learning_rate": 1.5262139236518695e-05, "loss": 0.9658, "step": 1787 }, { "epoch": 0.3441108545034642, "grad_norm": 1.7319666147232056, "learning_rate": 1.5256837698105047e-05, "loss": 1.0232, "step": 1788 }, { "epoch": 0.34430331023864513, "grad_norm": 1.9743072986602783, "learning_rate": 1.5251534117185935e-05, "loss": 1.0297, "step": 1789 }, { "epoch": 0.34449576597382603, "grad_norm": 1.812059998512268, "learning_rate": 1.5246228495822025e-05, "loss": 1.0263, "step": 1790 }, { "epoch": 0.3446882217090069, "grad_norm": 2.0064446926116943, "learning_rate": 1.5240920836074777e-05, "loss": 0.9941, "step": 1791 }, { "epoch": 0.3448806774441878, "grad_norm": 1.7347018718719482, "learning_rate": 1.5235611140006446e-05, "loss": 0.9202, "step": 1792 }, { "epoch": 0.3450731331793687, "grad_norm": 1.7446880340576172, "learning_rate": 1.5230299409680073e-05, "loss": 0.9867, "step": 1793 }, { "epoch": 0.34526558891454967, "grad_norm": 1.6364250183105469, "learning_rate": 1.5224985647159489e-05, "loss": 0.998, "step": 1794 }, { "epoch": 0.34545804464973057, "grad_norm": 1.6061142683029175, "learning_rate": 1.5219669854509322e-05, "loss": 0.9044, "step": 1795 }, { "epoch": 0.34565050038491146, "grad_norm": 1.7231050729751587, "learning_rate": 1.5214352033794981e-05, "loss": 0.9923, "step": 1796 }, { "epoch": 0.34584295612009236, "grad_norm": 1.6974034309387207, "learning_rate": 1.5209032187082668e-05, "loss": 0.9485, "step": 1797 }, { "epoch": 0.3460354118552733, "grad_norm": 1.663171648979187, "learning_rate": 1.5203710316439368e-05, "loss": 0.9791, "step": 1798 }, { "epoch": 0.3462278675904542, "grad_norm": 1.7451183795928955, "learning_rate": 1.5198386423932856e-05, "loss": 0.8731, "step": 1799 }, { "epoch": 0.3464203233256351, "grad_norm": 1.656074047088623, "learning_rate": 1.5193060511631692e-05, "loss": 0.9885, "step": 1800 }, { "epoch": 0.346612779060816, "grad_norm": 1.7435290813446045, "learning_rate": 1.5187732581605217e-05, "loss": 0.9412, "step": 1801 }, { "epoch": 0.3468052347959969, "grad_norm": 1.8893373012542725, "learning_rate": 1.518240263592356e-05, "loss": 0.9793, "step": 1802 }, { "epoch": 0.34699769053117785, "grad_norm": 0.968467652797699, "learning_rate": 1.5177070676657632e-05, "loss": 0.9738, "step": 1803 }, { "epoch": 0.34719014626635875, "grad_norm": 1.661312460899353, "learning_rate": 1.5171736705879127e-05, "loss": 1.0186, "step": 1804 }, { "epoch": 0.34738260200153964, "grad_norm": 1.8427565097808838, "learning_rate": 1.5166400725660522e-05, "loss": 1.0757, "step": 1805 }, { "epoch": 0.34757505773672054, "grad_norm": 1.962753176689148, "learning_rate": 1.5161062738075068e-05, "loss": 1.0142, "step": 1806 }, { "epoch": 0.34776751347190143, "grad_norm": 1.8477873802185059, "learning_rate": 1.5155722745196802e-05, "loss": 1.0406, "step": 1807 }, { "epoch": 0.3479599692070824, "grad_norm": 1.8629461526870728, "learning_rate": 1.5150380749100545e-05, "loss": 1.0145, "step": 1808 }, { "epoch": 0.3481524249422633, "grad_norm": 1.8187203407287598, "learning_rate": 1.514503675186188e-05, "loss": 1.0678, "step": 1809 }, { "epoch": 0.3483448806774442, "grad_norm": 1.8062211275100708, "learning_rate": 1.5139690755557186e-05, "loss": 1.1055, "step": 1810 }, { "epoch": 0.3485373364126251, "grad_norm": 1.7697410583496094, "learning_rate": 1.5134342762263606e-05, "loss": 0.9458, "step": 1811 }, { "epoch": 0.34872979214780603, "grad_norm": 1.891059160232544, "learning_rate": 1.5128992774059063e-05, "loss": 0.9142, "step": 1812 }, { "epoch": 0.3489222478829869, "grad_norm": 1.8962926864624023, "learning_rate": 1.5123640793022255e-05, "loss": 1.0509, "step": 1813 }, { "epoch": 0.3491147036181678, "grad_norm": 1.7912273406982422, "learning_rate": 1.5118286821232661e-05, "loss": 1.0551, "step": 1814 }, { "epoch": 0.3493071593533487, "grad_norm": 1.796258568763733, "learning_rate": 1.511293086077052e-05, "loss": 1.01, "step": 1815 }, { "epoch": 0.3494996150885296, "grad_norm": 0.9572339653968811, "learning_rate": 1.5107572913716859e-05, "loss": 0.9144, "step": 1816 }, { "epoch": 0.34969207082371057, "grad_norm": 0.9153892993927002, "learning_rate": 1.510221298215346e-05, "loss": 0.8672, "step": 1817 }, { "epoch": 0.34988452655889146, "grad_norm": 1.8201547861099243, "learning_rate": 1.5096851068162885e-05, "loss": 1.0036, "step": 1818 }, { "epoch": 0.35007698229407236, "grad_norm": 1.7657115459442139, "learning_rate": 1.5091487173828471e-05, "loss": 1.0024, "step": 1819 }, { "epoch": 0.35026943802925325, "grad_norm": 1.8045085668563843, "learning_rate": 1.5086121301234318e-05, "loss": 0.9945, "step": 1820 }, { "epoch": 0.3504618937644342, "grad_norm": 1.8675174713134766, "learning_rate": 1.5080753452465296e-05, "loss": 0.935, "step": 1821 }, { "epoch": 0.3506543494996151, "grad_norm": 1.0032260417938232, "learning_rate": 1.5075383629607043e-05, "loss": 0.9049, "step": 1822 }, { "epoch": 0.350846805234796, "grad_norm": 1.978115439414978, "learning_rate": 1.5070011834745963e-05, "loss": 0.9233, "step": 1823 }, { "epoch": 0.3510392609699769, "grad_norm": 1.5761922597885132, "learning_rate": 1.5064638069969228e-05, "loss": 0.9496, "step": 1824 }, { "epoch": 0.3512317167051578, "grad_norm": 1.76466703414917, "learning_rate": 1.5059262337364779e-05, "loss": 0.9874, "step": 1825 }, { "epoch": 0.35142417244033874, "grad_norm": 1.7005233764648438, "learning_rate": 1.505388463902131e-05, "loss": 1.0453, "step": 1826 }, { "epoch": 0.35161662817551964, "grad_norm": 1.7527347803115845, "learning_rate": 1.5048504977028289e-05, "loss": 0.9828, "step": 1827 }, { "epoch": 0.35180908391070054, "grad_norm": 1.7358801364898682, "learning_rate": 1.5043123353475944e-05, "loss": 0.9622, "step": 1828 }, { "epoch": 0.35200153964588143, "grad_norm": 0.9531983137130737, "learning_rate": 1.5037739770455263e-05, "loss": 0.9236, "step": 1829 }, { "epoch": 0.35219399538106233, "grad_norm": 1.897436499595642, "learning_rate": 1.5032354230058004e-05, "loss": 1.0767, "step": 1830 }, { "epoch": 0.3523864511162433, "grad_norm": 1.7305550575256348, "learning_rate": 1.502696673437667e-05, "loss": 0.9625, "step": 1831 }, { "epoch": 0.3525789068514242, "grad_norm": 1.7434419393539429, "learning_rate": 1.5021577285504538e-05, "loss": 1.0259, "step": 1832 }, { "epoch": 0.3527713625866051, "grad_norm": 1.7427966594696045, "learning_rate": 1.5016185885535639e-05, "loss": 1.0457, "step": 1833 }, { "epoch": 0.35296381832178597, "grad_norm": 1.6742581129074097, "learning_rate": 1.5010792536564762e-05, "loss": 0.9799, "step": 1834 }, { "epoch": 0.3531562740569669, "grad_norm": 1.6152372360229492, "learning_rate": 1.5005397240687447e-05, "loss": 1.017, "step": 1835 }, { "epoch": 0.3533487297921478, "grad_norm": 1.6158370971679688, "learning_rate": 1.5000000000000002e-05, "loss": 0.9691, "step": 1836 }, { "epoch": 0.3535411855273287, "grad_norm": 1.7204499244689941, "learning_rate": 1.4994600816599482e-05, "loss": 1.1004, "step": 1837 }, { "epoch": 0.3537336412625096, "grad_norm": 1.7515671253204346, "learning_rate": 1.4989199692583705e-05, "loss": 0.9887, "step": 1838 }, { "epoch": 0.3539260969976905, "grad_norm": 1.8158537149429321, "learning_rate": 1.498379663005123e-05, "loss": 1.0361, "step": 1839 }, { "epoch": 0.35411855273287146, "grad_norm": 1.7149584293365479, "learning_rate": 1.4978391631101383e-05, "loss": 0.8766, "step": 1840 }, { "epoch": 0.35431100846805236, "grad_norm": 1.627145767211914, "learning_rate": 1.4972984697834238e-05, "loss": 0.8509, "step": 1841 }, { "epoch": 0.35450346420323325, "grad_norm": 1.6957424879074097, "learning_rate": 1.4967575832350614e-05, "loss": 0.991, "step": 1842 }, { "epoch": 0.35469591993841415, "grad_norm": 1.001355767250061, "learning_rate": 1.4962165036752085e-05, "loss": 0.9685, "step": 1843 }, { "epoch": 0.35488837567359505, "grad_norm": 1.664493441581726, "learning_rate": 1.4956752313140978e-05, "loss": 0.9964, "step": 1844 }, { "epoch": 0.355080831408776, "grad_norm": 2.200862407684326, "learning_rate": 1.4951337663620368e-05, "loss": 0.9445, "step": 1845 }, { "epoch": 0.3552732871439569, "grad_norm": 1.8153678178787231, "learning_rate": 1.4945921090294076e-05, "loss": 1.0006, "step": 1846 }, { "epoch": 0.3554657428791378, "grad_norm": 1.8465709686279297, "learning_rate": 1.494050259526667e-05, "loss": 1.1014, "step": 1847 }, { "epoch": 0.3556581986143187, "grad_norm": 1.79034423828125, "learning_rate": 1.493508218064347e-05, "loss": 0.9557, "step": 1848 }, { "epoch": 0.35585065434949964, "grad_norm": 1.7255334854125977, "learning_rate": 1.4929659848530537e-05, "loss": 0.9405, "step": 1849 }, { "epoch": 0.35604311008468054, "grad_norm": 1.7785202264785767, "learning_rate": 1.4924235601034673e-05, "loss": 0.9204, "step": 1850 }, { "epoch": 0.35623556581986143, "grad_norm": 1.9693217277526855, "learning_rate": 1.4918809440263435e-05, "loss": 0.9728, "step": 1851 }, { "epoch": 0.35642802155504233, "grad_norm": 1.6653003692626953, "learning_rate": 1.4913381368325115e-05, "loss": 0.884, "step": 1852 }, { "epoch": 0.3566204772902232, "grad_norm": 1.7136613130569458, "learning_rate": 1.4907951387328752e-05, "loss": 1.0735, "step": 1853 }, { "epoch": 0.3568129330254042, "grad_norm": 1.8986228704452515, "learning_rate": 1.4902519499384122e-05, "loss": 1.0058, "step": 1854 }, { "epoch": 0.3570053887605851, "grad_norm": 1.693052887916565, "learning_rate": 1.489708570660175e-05, "loss": 0.951, "step": 1855 }, { "epoch": 0.35719784449576597, "grad_norm": 1.7758793830871582, "learning_rate": 1.4891650011092896e-05, "loss": 1.0251, "step": 1856 }, { "epoch": 0.35739030023094687, "grad_norm": 0.9435529708862305, "learning_rate": 1.4886212414969551e-05, "loss": 0.9106, "step": 1857 }, { "epoch": 0.35758275596612776, "grad_norm": 1.9558954238891602, "learning_rate": 1.488077292034446e-05, "loss": 0.9583, "step": 1858 }, { "epoch": 0.3577752117013087, "grad_norm": 1.7710357904434204, "learning_rate": 1.4875331529331097e-05, "loss": 1.0021, "step": 1859 }, { "epoch": 0.3579676674364896, "grad_norm": 1.700529932975769, "learning_rate": 1.4869888244043674e-05, "loss": 1.0239, "step": 1860 }, { "epoch": 0.3581601231716705, "grad_norm": 1.79655921459198, "learning_rate": 1.486444306659714e-05, "loss": 0.9699, "step": 1861 }, { "epoch": 0.3583525789068514, "grad_norm": 1.8918763399124146, "learning_rate": 1.4858995999107175e-05, "loss": 1.0352, "step": 1862 }, { "epoch": 0.35854503464203236, "grad_norm": 1.7336639165878296, "learning_rate": 1.4853547043690201e-05, "loss": 0.9867, "step": 1863 }, { "epoch": 0.35873749037721325, "grad_norm": 1.88038170337677, "learning_rate": 1.4848096202463373e-05, "loss": 0.9806, "step": 1864 }, { "epoch": 0.35892994611239415, "grad_norm": 1.765183687210083, "learning_rate": 1.4842643477544565e-05, "loss": 0.9848, "step": 1865 }, { "epoch": 0.35912240184757505, "grad_norm": 1.7156071662902832, "learning_rate": 1.4837188871052399e-05, "loss": 0.9358, "step": 1866 }, { "epoch": 0.35931485758275594, "grad_norm": 1.782707929611206, "learning_rate": 1.4831732385106221e-05, "loss": 1.0382, "step": 1867 }, { "epoch": 0.3595073133179369, "grad_norm": 1.8035192489624023, "learning_rate": 1.482627402182611e-05, "loss": 0.9793, "step": 1868 }, { "epoch": 0.3596997690531178, "grad_norm": 1.644707202911377, "learning_rate": 1.482081378333287e-05, "loss": 1.0671, "step": 1869 }, { "epoch": 0.3598922247882987, "grad_norm": 1.8341261148452759, "learning_rate": 1.4815351671748039e-05, "loss": 0.9814, "step": 1870 }, { "epoch": 0.3600846805234796, "grad_norm": 1.8055627346038818, "learning_rate": 1.4809887689193878e-05, "loss": 0.9508, "step": 1871 }, { "epoch": 0.36027713625866054, "grad_norm": 1.7969026565551758, "learning_rate": 1.4804421837793379e-05, "loss": 0.9645, "step": 1872 }, { "epoch": 0.36046959199384143, "grad_norm": 1.8895519971847534, "learning_rate": 1.4798954119670254e-05, "loss": 1.008, "step": 1873 }, { "epoch": 0.36066204772902233, "grad_norm": 1.850370168685913, "learning_rate": 1.479348453694895e-05, "loss": 0.934, "step": 1874 }, { "epoch": 0.3608545034642032, "grad_norm": 1.658559799194336, "learning_rate": 1.4788013091754629e-05, "loss": 0.9489, "step": 1875 }, { "epoch": 0.3610469591993841, "grad_norm": 1.8425266742706299, "learning_rate": 1.4782539786213184e-05, "loss": 1.0039, "step": 1876 }, { "epoch": 0.3612394149345651, "grad_norm": 1.7742093801498413, "learning_rate": 1.4777064622451222e-05, "loss": 1.0094, "step": 1877 }, { "epoch": 0.36143187066974597, "grad_norm": 1.7343881130218506, "learning_rate": 1.4771587602596085e-05, "loss": 1.0538, "step": 1878 }, { "epoch": 0.36162432640492687, "grad_norm": 1.6593002080917358, "learning_rate": 1.4766108728775822e-05, "loss": 0.9969, "step": 1879 }, { "epoch": 0.36181678214010776, "grad_norm": 1.8009185791015625, "learning_rate": 1.4760628003119213e-05, "loss": 1.0339, "step": 1880 }, { "epoch": 0.36200923787528866, "grad_norm": 1.823994517326355, "learning_rate": 1.4755145427755755e-05, "loss": 1.0146, "step": 1881 }, { "epoch": 0.3622016936104696, "grad_norm": 1.5586533546447754, "learning_rate": 1.4749661004815654e-05, "loss": 0.9343, "step": 1882 }, { "epoch": 0.3623941493456505, "grad_norm": 0.9568275809288025, "learning_rate": 1.474417473642985e-05, "loss": 0.9596, "step": 1883 }, { "epoch": 0.3625866050808314, "grad_norm": 1.8383195400238037, "learning_rate": 1.4738686624729987e-05, "loss": 0.9041, "step": 1884 }, { "epoch": 0.3627790608160123, "grad_norm": 1.7363063097000122, "learning_rate": 1.4733196671848435e-05, "loss": 1.061, "step": 1885 }, { "epoch": 0.36297151655119325, "grad_norm": 0.953652024269104, "learning_rate": 1.4727704879918272e-05, "loss": 0.968, "step": 1886 }, { "epoch": 0.36316397228637415, "grad_norm": 1.6911596059799194, "learning_rate": 1.4722211251073293e-05, "loss": 0.9449, "step": 1887 }, { "epoch": 0.36335642802155504, "grad_norm": 1.7536365985870361, "learning_rate": 1.4716715787448007e-05, "loss": 0.8973, "step": 1888 }, { "epoch": 0.36354888375673594, "grad_norm": 1.6344425678253174, "learning_rate": 1.4711218491177636e-05, "loss": 0.9418, "step": 1889 }, { "epoch": 0.36374133949191684, "grad_norm": 1.9258824586868286, "learning_rate": 1.4705719364398117e-05, "loss": 0.9604, "step": 1890 }, { "epoch": 0.3639337952270978, "grad_norm": 1.7928732633590698, "learning_rate": 1.4700218409246087e-05, "loss": 1.0507, "step": 1891 }, { "epoch": 0.3641262509622787, "grad_norm": 1.9834192991256714, "learning_rate": 1.469471562785891e-05, "loss": 1.0997, "step": 1892 }, { "epoch": 0.3643187066974596, "grad_norm": 1.7049096822738647, "learning_rate": 1.4689211022374645e-05, "loss": 1.0872, "step": 1893 }, { "epoch": 0.3645111624326405, "grad_norm": 1.9117679595947266, "learning_rate": 1.4683704594932072e-05, "loss": 0.9592, "step": 1894 }, { "epoch": 0.3647036181678214, "grad_norm": 1.756568431854248, "learning_rate": 1.4678196347670665e-05, "loss": 0.9899, "step": 1895 }, { "epoch": 0.3648960739030023, "grad_norm": 1.8233407735824585, "learning_rate": 1.4672686282730622e-05, "loss": 0.9564, "step": 1896 }, { "epoch": 0.3650885296381832, "grad_norm": 1.7915583848953247, "learning_rate": 1.4667174402252831e-05, "loss": 0.9611, "step": 1897 }, { "epoch": 0.3652809853733641, "grad_norm": 1.6582584381103516, "learning_rate": 1.4661660708378893e-05, "loss": 1.0075, "step": 1898 }, { "epoch": 0.365473441108545, "grad_norm": 1.6997991800308228, "learning_rate": 1.4656145203251116e-05, "loss": 0.9696, "step": 1899 }, { "epoch": 0.36566589684372597, "grad_norm": 1.4926577806472778, "learning_rate": 1.4650627889012507e-05, "loss": 0.9643, "step": 1900 }, { "epoch": 0.36585835257890686, "grad_norm": 1.787379264831543, "learning_rate": 1.4645108767806778e-05, "loss": 0.9739, "step": 1901 }, { "epoch": 0.36605080831408776, "grad_norm": 1.5647180080413818, "learning_rate": 1.4639587841778342e-05, "loss": 0.9411, "step": 1902 }, { "epoch": 0.36624326404926866, "grad_norm": 1.569069504737854, "learning_rate": 1.4634065113072316e-05, "loss": 0.9336, "step": 1903 }, { "epoch": 0.36643571978444955, "grad_norm": 1.7829434871673584, "learning_rate": 1.4628540583834511e-05, "loss": 1.0497, "step": 1904 }, { "epoch": 0.3666281755196305, "grad_norm": 1.7288581132888794, "learning_rate": 1.462301425621145e-05, "loss": 1.0149, "step": 1905 }, { "epoch": 0.3668206312548114, "grad_norm": 1.8584659099578857, "learning_rate": 1.4617486132350343e-05, "loss": 0.9957, "step": 1906 }, { "epoch": 0.3670130869899923, "grad_norm": 1.8501023054122925, "learning_rate": 1.4611956214399097e-05, "loss": 0.9084, "step": 1907 }, { "epoch": 0.3672055427251732, "grad_norm": 1.9918019771575928, "learning_rate": 1.4606424504506325e-05, "loss": 1.078, "step": 1908 }, { "epoch": 0.3673979984603541, "grad_norm": 1.648996353149414, "learning_rate": 1.4600891004821332e-05, "loss": 0.9808, "step": 1909 }, { "epoch": 0.36759045419553504, "grad_norm": 1.7351601123809814, "learning_rate": 1.4595355717494118e-05, "loss": 0.9646, "step": 1910 }, { "epoch": 0.36778290993071594, "grad_norm": 2.054368495941162, "learning_rate": 1.4589818644675378e-05, "loss": 1.0624, "step": 1911 }, { "epoch": 0.36797536566589684, "grad_norm": 1.7488799095153809, "learning_rate": 1.45842797885165e-05, "loss": 0.9769, "step": 1912 }, { "epoch": 0.36816782140107773, "grad_norm": 1.8232616186141968, "learning_rate": 1.4578739151169567e-05, "loss": 0.9007, "step": 1913 }, { "epoch": 0.3683602771362587, "grad_norm": 1.8748761415481567, "learning_rate": 1.4573196734787357e-05, "loss": 0.9725, "step": 1914 }, { "epoch": 0.3685527328714396, "grad_norm": 0.9844322204589844, "learning_rate": 1.4567652541523325e-05, "loss": 0.905, "step": 1915 }, { "epoch": 0.3687451886066205, "grad_norm": 1.824452519416809, "learning_rate": 1.4562106573531632e-05, "loss": 1.0371, "step": 1916 }, { "epoch": 0.3689376443418014, "grad_norm": 2.2875170707702637, "learning_rate": 1.4556558832967122e-05, "loss": 0.9793, "step": 1917 }, { "epoch": 0.36913010007698227, "grad_norm": 1.6063363552093506, "learning_rate": 1.455100932198533e-05, "loss": 0.9742, "step": 1918 }, { "epoch": 0.3693225558121632, "grad_norm": 1.7440885305404663, "learning_rate": 1.4545458042742477e-05, "loss": 0.9639, "step": 1919 }, { "epoch": 0.3695150115473441, "grad_norm": 2.064701557159424, "learning_rate": 1.4539904997395468e-05, "loss": 0.9666, "step": 1920 }, { "epoch": 0.369707467282525, "grad_norm": 1.7673022747039795, "learning_rate": 1.4534350188101905e-05, "loss": 0.9194, "step": 1921 }, { "epoch": 0.3698999230177059, "grad_norm": 1.6951756477355957, "learning_rate": 1.4528793617020063e-05, "loss": 0.951, "step": 1922 }, { "epoch": 0.37009237875288686, "grad_norm": 1.771060824394226, "learning_rate": 1.4523235286308907e-05, "loss": 0.9517, "step": 1923 }, { "epoch": 0.37028483448806776, "grad_norm": 1.7901089191436768, "learning_rate": 1.4517675198128086e-05, "loss": 0.9978, "step": 1924 }, { "epoch": 0.37047729022324866, "grad_norm": 1.6266430616378784, "learning_rate": 1.4512113354637931e-05, "loss": 1.0247, "step": 1925 }, { "epoch": 0.37066974595842955, "grad_norm": 1.6474885940551758, "learning_rate": 1.4506549757999456e-05, "loss": 1.0497, "step": 1926 }, { "epoch": 0.37086220169361045, "grad_norm": 1.8308584690093994, "learning_rate": 1.4500984410374353e-05, "loss": 0.989, "step": 1927 }, { "epoch": 0.3710546574287914, "grad_norm": 1.8437163829803467, "learning_rate": 1.4495417313924996e-05, "loss": 1.0045, "step": 1928 }, { "epoch": 0.3712471131639723, "grad_norm": 1.6356878280639648, "learning_rate": 1.4489848470814447e-05, "loss": 0.994, "step": 1929 }, { "epoch": 0.3714395688991532, "grad_norm": 1.624777913093567, "learning_rate": 1.4484277883206433e-05, "loss": 1.0028, "step": 1930 }, { "epoch": 0.3716320246343341, "grad_norm": 1.6456894874572754, "learning_rate": 1.4478705553265363e-05, "loss": 0.9101, "step": 1931 }, { "epoch": 0.371824480369515, "grad_norm": 1.755164384841919, "learning_rate": 1.4473131483156326e-05, "loss": 0.924, "step": 1932 }, { "epoch": 0.37201693610469594, "grad_norm": 1.7762017250061035, "learning_rate": 1.4467555675045088e-05, "loss": 0.9734, "step": 1933 }, { "epoch": 0.37220939183987684, "grad_norm": 1.772132158279419, "learning_rate": 1.4461978131098089e-05, "loss": 0.9483, "step": 1934 }, { "epoch": 0.37240184757505773, "grad_norm": 1.8451095819473267, "learning_rate": 1.4456398853482438e-05, "loss": 1.0793, "step": 1935 }, { "epoch": 0.37259430331023863, "grad_norm": 1.7906150817871094, "learning_rate": 1.4450817844365924e-05, "loss": 1.0455, "step": 1936 }, { "epoch": 0.3727867590454196, "grad_norm": 1.7000939846038818, "learning_rate": 1.444523510591701e-05, "loss": 1.0102, "step": 1937 }, { "epoch": 0.3729792147806005, "grad_norm": 1.7745611667633057, "learning_rate": 1.4439650640304822e-05, "loss": 0.9774, "step": 1938 }, { "epoch": 0.3731716705157814, "grad_norm": 1.8367114067077637, "learning_rate": 1.4434064449699168e-05, "loss": 1.0389, "step": 1939 }, { "epoch": 0.37336412625096227, "grad_norm": 1.8302029371261597, "learning_rate": 1.4428476536270517e-05, "loss": 0.9139, "step": 1940 }, { "epoch": 0.37355658198614317, "grad_norm": 1.825661540031433, "learning_rate": 1.4422886902190014e-05, "loss": 0.9879, "step": 1941 }, { "epoch": 0.3737490377213241, "grad_norm": 1.690522313117981, "learning_rate": 1.4417295549629469e-05, "loss": 0.9629, "step": 1942 }, { "epoch": 0.373941493456505, "grad_norm": 1.7939332723617554, "learning_rate": 1.4411702480761362e-05, "loss": 1.0092, "step": 1943 }, { "epoch": 0.3741339491916859, "grad_norm": 1.655850887298584, "learning_rate": 1.4406107697758838e-05, "loss": 1.0481, "step": 1944 }, { "epoch": 0.3743264049268668, "grad_norm": 1.7256571054458618, "learning_rate": 1.440051120279571e-05, "loss": 0.9911, "step": 1945 }, { "epoch": 0.3745188606620477, "grad_norm": 1.6558855772018433, "learning_rate": 1.4394912998046451e-05, "loss": 1.0133, "step": 1946 }, { "epoch": 0.37471131639722866, "grad_norm": 1.5809326171875, "learning_rate": 1.4389313085686205e-05, "loss": 1.0212, "step": 1947 }, { "epoch": 0.37490377213240955, "grad_norm": 1.5560415983200073, "learning_rate": 1.4383711467890776e-05, "loss": 1.0678, "step": 1948 }, { "epoch": 0.37509622786759045, "grad_norm": 1.7232911586761475, "learning_rate": 1.4378108146836632e-05, "loss": 0.8871, "step": 1949 }, { "epoch": 0.37528868360277134, "grad_norm": 1.86379075050354, "learning_rate": 1.4372503124700899e-05, "loss": 1.0044, "step": 1950 }, { "epoch": 0.3754811393379523, "grad_norm": 1.6726834774017334, "learning_rate": 1.436689640366137e-05, "loss": 0.9294, "step": 1951 }, { "epoch": 0.3756735950731332, "grad_norm": 1.7556164264678955, "learning_rate": 1.4361287985896495e-05, "loss": 0.9914, "step": 1952 }, { "epoch": 0.3758660508083141, "grad_norm": 1.6470776796340942, "learning_rate": 1.4355677873585383e-05, "loss": 0.9562, "step": 1953 }, { "epoch": 0.376058506543495, "grad_norm": 1.621160626411438, "learning_rate": 1.43500660689078e-05, "loss": 0.9376, "step": 1954 }, { "epoch": 0.3762509622786759, "grad_norm": 1.7888695001602173, "learning_rate": 1.4344452574044173e-05, "loss": 1.0268, "step": 1955 }, { "epoch": 0.37644341801385683, "grad_norm": 1.7422411441802979, "learning_rate": 1.4338837391175582e-05, "loss": 0.9306, "step": 1956 }, { "epoch": 0.37663587374903773, "grad_norm": 1.7942121028900146, "learning_rate": 1.4333220522483768e-05, "loss": 1.0233, "step": 1957 }, { "epoch": 0.3768283294842186, "grad_norm": 1.6585218906402588, "learning_rate": 1.4327601970151122e-05, "loss": 0.9579, "step": 1958 }, { "epoch": 0.3770207852193995, "grad_norm": 1.7867778539657593, "learning_rate": 1.432198173636069e-05, "loss": 1.0209, "step": 1959 }, { "epoch": 0.3772132409545804, "grad_norm": 1.7259944677352905, "learning_rate": 1.4316359823296174e-05, "loss": 1.0178, "step": 1960 }, { "epoch": 0.37740569668976137, "grad_norm": 4.83243465423584, "learning_rate": 1.4310736233141926e-05, "loss": 0.922, "step": 1961 }, { "epoch": 0.37759815242494227, "grad_norm": 1.8803330659866333, "learning_rate": 1.4305110968082953e-05, "loss": 0.9977, "step": 1962 }, { "epoch": 0.37779060816012316, "grad_norm": 1.8122565746307373, "learning_rate": 1.4299484030304906e-05, "loss": 0.9645, "step": 1963 }, { "epoch": 0.37798306389530406, "grad_norm": 1.755486011505127, "learning_rate": 1.4293855421994094e-05, "loss": 0.915, "step": 1964 }, { "epoch": 0.378175519630485, "grad_norm": 1.9173431396484375, "learning_rate": 1.4288225145337467e-05, "loss": 1.0457, "step": 1965 }, { "epoch": 0.3783679753656659, "grad_norm": 1.6177693605422974, "learning_rate": 1.4282593202522627e-05, "loss": 0.9553, "step": 1966 }, { "epoch": 0.3785604311008468, "grad_norm": 1.8555335998535156, "learning_rate": 1.4276959595737832e-05, "loss": 1.0855, "step": 1967 }, { "epoch": 0.3787528868360277, "grad_norm": 1.929799199104309, "learning_rate": 1.4271324327171969e-05, "loss": 0.9304, "step": 1968 }, { "epoch": 0.3789453425712086, "grad_norm": 1.8007659912109375, "learning_rate": 1.4265687399014584e-05, "loss": 1.0211, "step": 1969 }, { "epoch": 0.37913779830638955, "grad_norm": 1.7115871906280518, "learning_rate": 1.4260048813455867e-05, "loss": 1.0229, "step": 1970 }, { "epoch": 0.37933025404157045, "grad_norm": 1.7530657052993774, "learning_rate": 1.4254408572686642e-05, "loss": 1.0693, "step": 1971 }, { "epoch": 0.37952270977675134, "grad_norm": 1.8238295316696167, "learning_rate": 1.4248766678898386e-05, "loss": 0.9122, "step": 1972 }, { "epoch": 0.37971516551193224, "grad_norm": 1.6122338771820068, "learning_rate": 1.4243123134283217e-05, "loss": 0.9053, "step": 1973 }, { "epoch": 0.37990762124711314, "grad_norm": 1.673415184020996, "learning_rate": 1.4237477941033888e-05, "loss": 0.9975, "step": 1974 }, { "epoch": 0.3801000769822941, "grad_norm": 1.8511546850204468, "learning_rate": 1.4231831101343802e-05, "loss": 0.9626, "step": 1975 }, { "epoch": 0.380292532717475, "grad_norm": 1.4975576400756836, "learning_rate": 1.4226182617406996e-05, "loss": 0.986, "step": 1976 }, { "epoch": 0.3804849884526559, "grad_norm": 1.554913878440857, "learning_rate": 1.4220532491418144e-05, "loss": 0.9972, "step": 1977 }, { "epoch": 0.3806774441878368, "grad_norm": 1.775629997253418, "learning_rate": 1.4214880725572563e-05, "loss": 1.0087, "step": 1978 }, { "epoch": 0.38086989992301773, "grad_norm": 1.651060938835144, "learning_rate": 1.4209227322066206e-05, "loss": 0.8841, "step": 1979 }, { "epoch": 0.3810623556581986, "grad_norm": 1.8014286756515503, "learning_rate": 1.4203572283095657e-05, "loss": 0.9728, "step": 1980 }, { "epoch": 0.3812548113933795, "grad_norm": 1.762376308441162, "learning_rate": 1.4197915610858143e-05, "loss": 0.9638, "step": 1981 }, { "epoch": 0.3814472671285604, "grad_norm": 1.7320339679718018, "learning_rate": 1.419225730755152e-05, "loss": 1.0645, "step": 1982 }, { "epoch": 0.3816397228637413, "grad_norm": 1.7174746990203857, "learning_rate": 1.4186597375374283e-05, "loss": 0.9905, "step": 1983 }, { "epoch": 0.38183217859892227, "grad_norm": 1.8385252952575684, "learning_rate": 1.4180935816525554e-05, "loss": 1.0071, "step": 1984 }, { "epoch": 0.38202463433410316, "grad_norm": 0.9697132110595703, "learning_rate": 1.417527263320509e-05, "loss": 0.8468, "step": 1985 }, { "epoch": 0.38221709006928406, "grad_norm": 2.0253593921661377, "learning_rate": 1.4169607827613284e-05, "loss": 1.0395, "step": 1986 }, { "epoch": 0.38240954580446496, "grad_norm": 2.054898500442505, "learning_rate": 1.4163941401951147e-05, "loss": 1.0069, "step": 1987 }, { "epoch": 0.3826020015396459, "grad_norm": 1.7338184118270874, "learning_rate": 1.415827335842033e-05, "loss": 0.9927, "step": 1988 }, { "epoch": 0.3827944572748268, "grad_norm": 1.6496775150299072, "learning_rate": 1.4152603699223112e-05, "loss": 0.9868, "step": 1989 }, { "epoch": 0.3829869130100077, "grad_norm": 1.7666347026824951, "learning_rate": 1.4146932426562391e-05, "loss": 0.9564, "step": 1990 }, { "epoch": 0.3831793687451886, "grad_norm": 1.791212797164917, "learning_rate": 1.4141259542641706e-05, "loss": 1.0263, "step": 1991 }, { "epoch": 0.3833718244803695, "grad_norm": 1.7299693822860718, "learning_rate": 1.4135585049665207e-05, "loss": 1.0899, "step": 1992 }, { "epoch": 0.38356428021555045, "grad_norm": 1.8792473077774048, "learning_rate": 1.4129908949837684e-05, "loss": 0.9862, "step": 1993 }, { "epoch": 0.38375673595073134, "grad_norm": 1.9791984558105469, "learning_rate": 1.4124231245364534e-05, "loss": 0.8934, "step": 1994 }, { "epoch": 0.38394919168591224, "grad_norm": 1.8039600849151611, "learning_rate": 1.4118551938451797e-05, "loss": 1.0477, "step": 1995 }, { "epoch": 0.38414164742109314, "grad_norm": 1.6801077127456665, "learning_rate": 1.4112871031306118e-05, "loss": 1.0477, "step": 1996 }, { "epoch": 0.38433410315627403, "grad_norm": 1.8121223449707031, "learning_rate": 1.4107188526134774e-05, "loss": 1.0081, "step": 1997 }, { "epoch": 0.384526558891455, "grad_norm": 1.7899469137191772, "learning_rate": 1.4101504425145662e-05, "loss": 0.9686, "step": 1998 }, { "epoch": 0.3847190146266359, "grad_norm": 1.9349442720413208, "learning_rate": 1.4095818730547297e-05, "loss": 0.9467, "step": 1999 }, { "epoch": 0.3849114703618168, "grad_norm": 1.637374997138977, "learning_rate": 1.4090131444548814e-05, "loss": 0.9958, "step": 2000 }, { "epoch": 0.3851039260969977, "grad_norm": 1.6273151636123657, "learning_rate": 1.4084442569359964e-05, "loss": 0.9635, "step": 2001 }, { "epoch": 0.3852963818321786, "grad_norm": 1.8371778726577759, "learning_rate": 1.4078752107191118e-05, "loss": 1.0348, "step": 2002 }, { "epoch": 0.3854888375673595, "grad_norm": 1.560639500617981, "learning_rate": 1.4073060060253265e-05, "loss": 0.9934, "step": 2003 }, { "epoch": 0.3856812933025404, "grad_norm": 1.6575658321380615, "learning_rate": 1.4067366430758004e-05, "loss": 1.0171, "step": 2004 }, { "epoch": 0.3858737490377213, "grad_norm": 1.7119311094284058, "learning_rate": 1.4061671220917557e-05, "loss": 0.9354, "step": 2005 }, { "epoch": 0.3860662047729022, "grad_norm": 1.7783334255218506, "learning_rate": 1.4055974432944753e-05, "loss": 0.9984, "step": 2006 }, { "epoch": 0.38625866050808316, "grad_norm": 1.6485015153884888, "learning_rate": 1.4050276069053036e-05, "loss": 0.9115, "step": 2007 }, { "epoch": 0.38645111624326406, "grad_norm": 1.7605186700820923, "learning_rate": 1.4044576131456466e-05, "loss": 0.9653, "step": 2008 }, { "epoch": 0.38664357197844496, "grad_norm": 2.064528226852417, "learning_rate": 1.4038874622369714e-05, "loss": 0.9846, "step": 2009 }, { "epoch": 0.38683602771362585, "grad_norm": 1.8784908056259155, "learning_rate": 1.4033171544008053e-05, "loss": 0.9582, "step": 2010 }, { "epoch": 0.38702848344880675, "grad_norm": 1.869053602218628, "learning_rate": 1.4027466898587375e-05, "loss": 1.0082, "step": 2011 }, { "epoch": 0.3872209391839877, "grad_norm": 1.717316746711731, "learning_rate": 1.4021760688324175e-05, "loss": 0.9752, "step": 2012 }, { "epoch": 0.3874133949191686, "grad_norm": 0.9960609674453735, "learning_rate": 1.4016052915435565e-05, "loss": 1.019, "step": 2013 }, { "epoch": 0.3876058506543495, "grad_norm": 1.6758718490600586, "learning_rate": 1.4010343582139252e-05, "loss": 0.9974, "step": 2014 }, { "epoch": 0.3877983063895304, "grad_norm": 1.6149204969406128, "learning_rate": 1.4004632690653559e-05, "loss": 0.9908, "step": 2015 }, { "epoch": 0.38799076212471134, "grad_norm": 1.9212483167648315, "learning_rate": 1.3998920243197408e-05, "loss": 0.9454, "step": 2016 }, { "epoch": 0.38818321785989224, "grad_norm": 1.8701709508895874, "learning_rate": 1.3993206241990333e-05, "loss": 0.9856, "step": 2017 }, { "epoch": 0.38837567359507313, "grad_norm": 1.6525683403015137, "learning_rate": 1.3987490689252463e-05, "loss": 0.9247, "step": 2018 }, { "epoch": 0.38856812933025403, "grad_norm": 1.8931653499603271, "learning_rate": 1.3981773587204535e-05, "loss": 1.0949, "step": 2019 }, { "epoch": 0.3887605850654349, "grad_norm": 2.060068130493164, "learning_rate": 1.3976054938067885e-05, "loss": 0.9748, "step": 2020 }, { "epoch": 0.3889530408006159, "grad_norm": 1.5122934579849243, "learning_rate": 1.3970334744064451e-05, "loss": 0.9233, "step": 2021 }, { "epoch": 0.3891454965357968, "grad_norm": 2.8335907459259033, "learning_rate": 1.3964613007416777e-05, "loss": 0.9702, "step": 2022 }, { "epoch": 0.38933795227097767, "grad_norm": 1.7102752923965454, "learning_rate": 1.3958889730347997e-05, "loss": 0.9971, "step": 2023 }, { "epoch": 0.38953040800615857, "grad_norm": 1.8470264673233032, "learning_rate": 1.3953164915081852e-05, "loss": 1.0199, "step": 2024 }, { "epoch": 0.38972286374133946, "grad_norm": 1.741223692893982, "learning_rate": 1.3947438563842672e-05, "loss": 0.9761, "step": 2025 }, { "epoch": 0.3899153194765204, "grad_norm": 1.8601967096328735, "learning_rate": 1.3941710678855396e-05, "loss": 0.9552, "step": 2026 }, { "epoch": 0.3901077752117013, "grad_norm": 1.8208478689193726, "learning_rate": 1.3935981262345541e-05, "loss": 0.9976, "step": 2027 }, { "epoch": 0.3903002309468822, "grad_norm": 1.7925888299942017, "learning_rate": 1.3930250316539237e-05, "loss": 0.984, "step": 2028 }, { "epoch": 0.3904926866820631, "grad_norm": 1.8152565956115723, "learning_rate": 1.39245178436632e-05, "loss": 1.0463, "step": 2029 }, { "epoch": 0.39068514241724406, "grad_norm": 1.7336716651916504, "learning_rate": 1.3918783845944737e-05, "loss": 0.9954, "step": 2030 }, { "epoch": 0.39087759815242495, "grad_norm": 2.2292046546936035, "learning_rate": 1.391304832561175e-05, "loss": 0.9174, "step": 2031 }, { "epoch": 0.39107005388760585, "grad_norm": 1.7288459539413452, "learning_rate": 1.3907311284892737e-05, "loss": 1.0216, "step": 2032 }, { "epoch": 0.39126250962278675, "grad_norm": 1.637322187423706, "learning_rate": 1.3901572726016782e-05, "loss": 0.9253, "step": 2033 }, { "epoch": 0.39145496535796764, "grad_norm": 0.9683698415756226, "learning_rate": 1.389583265121356e-05, "loss": 0.9872, "step": 2034 }, { "epoch": 0.3916474210931486, "grad_norm": 1.0408151149749756, "learning_rate": 1.3890091062713332e-05, "loss": 0.9163, "step": 2035 }, { "epoch": 0.3918398768283295, "grad_norm": 1.9911779165267944, "learning_rate": 1.3884347962746949e-05, "loss": 0.9821, "step": 2036 }, { "epoch": 0.3920323325635104, "grad_norm": 2.588649034500122, "learning_rate": 1.3878603353545854e-05, "loss": 1.1045, "step": 2037 }, { "epoch": 0.3922247882986913, "grad_norm": 2.025618314743042, "learning_rate": 1.3872857237342067e-05, "loss": 1.0634, "step": 2038 }, { "epoch": 0.39241724403387224, "grad_norm": 0.9749510884284973, "learning_rate": 1.3867109616368208e-05, "loss": 0.8942, "step": 2039 }, { "epoch": 0.39260969976905313, "grad_norm": 1.807681918144226, "learning_rate": 1.3861360492857464e-05, "loss": 0.929, "step": 2040 }, { "epoch": 0.39280215550423403, "grad_norm": 1.8939077854156494, "learning_rate": 1.3855609869043618e-05, "loss": 0.9373, "step": 2041 }, { "epoch": 0.3929946112394149, "grad_norm": 1.6454027891159058, "learning_rate": 1.3849857747161036e-05, "loss": 0.9032, "step": 2042 }, { "epoch": 0.3931870669745958, "grad_norm": 1.7353684902191162, "learning_rate": 1.3844104129444655e-05, "loss": 0.9946, "step": 2043 }, { "epoch": 0.3933795227097768, "grad_norm": 1.8138177394866943, "learning_rate": 1.3838349018130007e-05, "loss": 0.9684, "step": 2044 }, { "epoch": 0.39357197844495767, "grad_norm": 1.9001697301864624, "learning_rate": 1.3832592415453193e-05, "loss": 0.9971, "step": 2045 }, { "epoch": 0.39376443418013857, "grad_norm": 1.7635122537612915, "learning_rate": 1.3826834323650899e-05, "loss": 1.0403, "step": 2046 }, { "epoch": 0.39395688991531946, "grad_norm": 1.9057657718658447, "learning_rate": 1.3821074744960394e-05, "loss": 0.954, "step": 2047 }, { "epoch": 0.39414934565050036, "grad_norm": 1.4782705307006836, "learning_rate": 1.3815313681619515e-05, "loss": 0.9798, "step": 2048 }, { "epoch": 0.3943418013856813, "grad_norm": 1.6358892917633057, "learning_rate": 1.3809551135866683e-05, "loss": 0.943, "step": 2049 }, { "epoch": 0.3945342571208622, "grad_norm": 1.8052687644958496, "learning_rate": 1.3803787109940891e-05, "loss": 0.9325, "step": 2050 }, { "epoch": 0.3947267128560431, "grad_norm": 1.6639820337295532, "learning_rate": 1.3798021606081713e-05, "loss": 0.9819, "step": 2051 }, { "epoch": 0.394919168591224, "grad_norm": 1.8043951988220215, "learning_rate": 1.3792254626529286e-05, "loss": 1.0134, "step": 2052 }, { "epoch": 0.39511162432640495, "grad_norm": 1.6742961406707764, "learning_rate": 1.3786486173524331e-05, "loss": 1.0323, "step": 2053 }, { "epoch": 0.39530408006158585, "grad_norm": 1.7478245496749878, "learning_rate": 1.378071624930814e-05, "loss": 0.9797, "step": 2054 }, { "epoch": 0.39549653579676675, "grad_norm": 1.9097031354904175, "learning_rate": 1.377494485612257e-05, "loss": 1.0842, "step": 2055 }, { "epoch": 0.39568899153194764, "grad_norm": 1.8818575143814087, "learning_rate": 1.3769171996210053e-05, "loss": 1.0266, "step": 2056 }, { "epoch": 0.39588144726712854, "grad_norm": 1.6143834590911865, "learning_rate": 1.3763397671813597e-05, "loss": 0.9203, "step": 2057 }, { "epoch": 0.3960739030023095, "grad_norm": 1.798588514328003, "learning_rate": 1.3757621885176768e-05, "loss": 0.9495, "step": 2058 }, { "epoch": 0.3962663587374904, "grad_norm": 1.7057496309280396, "learning_rate": 1.375184463854371e-05, "loss": 0.9649, "step": 2059 }, { "epoch": 0.3964588144726713, "grad_norm": 1.6676299571990967, "learning_rate": 1.3746065934159123e-05, "loss": 1.0196, "step": 2060 }, { "epoch": 0.3966512702078522, "grad_norm": 1.8378809690475464, "learning_rate": 1.3740285774268282e-05, "loss": 0.9631, "step": 2061 }, { "epoch": 0.3968437259430331, "grad_norm": 1.7750494480133057, "learning_rate": 1.3734504161117028e-05, "loss": 0.9016, "step": 2062 }, { "epoch": 0.39703618167821403, "grad_norm": 1.8523062467575073, "learning_rate": 1.3728721096951762e-05, "loss": 0.8813, "step": 2063 }, { "epoch": 0.3972286374133949, "grad_norm": 1.8497612476348877, "learning_rate": 1.3722936584019453e-05, "loss": 0.9487, "step": 2064 }, { "epoch": 0.3974210931485758, "grad_norm": 1.9014673233032227, "learning_rate": 1.3717150624567627e-05, "loss": 0.9133, "step": 2065 }, { "epoch": 0.3976135488837567, "grad_norm": 1.6362453699111938, "learning_rate": 1.371136322084438e-05, "loss": 1.0211, "step": 2066 }, { "epoch": 0.39780600461893767, "grad_norm": 1.84226655960083, "learning_rate": 1.3705574375098365e-05, "loss": 0.9428, "step": 2067 }, { "epoch": 0.39799846035411857, "grad_norm": 1.7137995958328247, "learning_rate": 1.3699784089578791e-05, "loss": 0.9609, "step": 2068 }, { "epoch": 0.39819091608929946, "grad_norm": 1.7027982473373413, "learning_rate": 1.3693992366535433e-05, "loss": 0.9381, "step": 2069 }, { "epoch": 0.39838337182448036, "grad_norm": 1.8677493333816528, "learning_rate": 1.3688199208218623e-05, "loss": 1.0187, "step": 2070 }, { "epoch": 0.39857582755966126, "grad_norm": 1.9954644441604614, "learning_rate": 1.3682404616879246e-05, "loss": 0.9758, "step": 2071 }, { "epoch": 0.3987682832948422, "grad_norm": 1.8320896625518799, "learning_rate": 1.3676608594768754e-05, "loss": 0.9589, "step": 2072 }, { "epoch": 0.3989607390300231, "grad_norm": 1.7886749505996704, "learning_rate": 1.3670811144139144e-05, "loss": 1.0457, "step": 2073 }, { "epoch": 0.399153194765204, "grad_norm": 1.6116259098052979, "learning_rate": 1.3665012267242974e-05, "loss": 0.951, "step": 2074 }, { "epoch": 0.3993456505003849, "grad_norm": 1.718163013458252, "learning_rate": 1.3659211966333357e-05, "loss": 1.0526, "step": 2075 }, { "epoch": 0.3995381062355658, "grad_norm": 1.8794950246810913, "learning_rate": 1.3653410243663953e-05, "loss": 1.0422, "step": 2076 }, { "epoch": 0.39973056197074675, "grad_norm": 1.728333592414856, "learning_rate": 1.364760710148898e-05, "loss": 0.9549, "step": 2077 }, { "epoch": 0.39992301770592764, "grad_norm": 1.887534499168396, "learning_rate": 1.3641802542063208e-05, "loss": 0.9647, "step": 2078 }, { "epoch": 0.40011547344110854, "grad_norm": 1.7697713375091553, "learning_rate": 1.3635996567641952e-05, "loss": 0.9717, "step": 2079 }, { "epoch": 0.40030792917628943, "grad_norm": 1.806296467781067, "learning_rate": 1.3630189180481083e-05, "loss": 1.0381, "step": 2080 }, { "epoch": 0.4005003849114704, "grad_norm": 1.862405776977539, "learning_rate": 1.3624380382837017e-05, "loss": 0.9947, "step": 2081 }, { "epoch": 0.4006928406466513, "grad_norm": 1.7284010648727417, "learning_rate": 1.3618570176966723e-05, "loss": 0.9294, "step": 2082 }, { "epoch": 0.4008852963818322, "grad_norm": 1.8863991498947144, "learning_rate": 1.361275856512771e-05, "loss": 1.0141, "step": 2083 }, { "epoch": 0.4010777521170131, "grad_norm": 1.7761805057525635, "learning_rate": 1.3606945549578039e-05, "loss": 0.9898, "step": 2084 }, { "epoch": 0.40127020785219397, "grad_norm": 1.9305976629257202, "learning_rate": 1.360113113257631e-05, "loss": 0.9908, "step": 2085 }, { "epoch": 0.4014626635873749, "grad_norm": 1.7332381010055542, "learning_rate": 1.3595315316381676e-05, "loss": 0.9411, "step": 2086 }, { "epoch": 0.4016551193225558, "grad_norm": 1.663541555404663, "learning_rate": 1.3589498103253828e-05, "loss": 0.9499, "step": 2087 }, { "epoch": 0.4018475750577367, "grad_norm": 1.8119417428970337, "learning_rate": 1.3583679495453e-05, "loss": 0.8928, "step": 2088 }, { "epoch": 0.4020400307929176, "grad_norm": 1.7064785957336426, "learning_rate": 1.3577859495239975e-05, "loss": 1.0262, "step": 2089 }, { "epoch": 0.40223248652809857, "grad_norm": 1.7013280391693115, "learning_rate": 1.3572038104876066e-05, "loss": 1.0146, "step": 2090 }, { "epoch": 0.40242494226327946, "grad_norm": 1.8446394205093384, "learning_rate": 1.3566215326623131e-05, "loss": 0.926, "step": 2091 }, { "epoch": 0.40261739799846036, "grad_norm": 1.7226033210754395, "learning_rate": 1.356039116274357e-05, "loss": 1.0443, "step": 2092 }, { "epoch": 0.40280985373364125, "grad_norm": 1.790978193283081, "learning_rate": 1.3554565615500313e-05, "loss": 0.9462, "step": 2093 }, { "epoch": 0.40300230946882215, "grad_norm": 1.7985692024230957, "learning_rate": 1.3548738687156839e-05, "loss": 1.0413, "step": 2094 }, { "epoch": 0.4031947652040031, "grad_norm": 0.8923540115356445, "learning_rate": 1.3542910379977158e-05, "loss": 0.9012, "step": 2095 }, { "epoch": 0.403387220939184, "grad_norm": 1.636090636253357, "learning_rate": 1.3537080696225815e-05, "loss": 0.9692, "step": 2096 }, { "epoch": 0.4035796766743649, "grad_norm": 1.6827718019485474, "learning_rate": 1.353124963816789e-05, "loss": 1.0121, "step": 2097 }, { "epoch": 0.4037721324095458, "grad_norm": 1.589532494544983, "learning_rate": 1.3525417208068996e-05, "loss": 1.0009, "step": 2098 }, { "epoch": 0.4039645881447267, "grad_norm": 1.7537076473236084, "learning_rate": 1.3519583408195283e-05, "loss": 0.987, "step": 2099 }, { "epoch": 0.40415704387990764, "grad_norm": 1.8689415454864502, "learning_rate": 1.3513748240813429e-05, "loss": 0.866, "step": 2100 }, { "epoch": 0.40434949961508854, "grad_norm": 1.8183119297027588, "learning_rate": 1.3507911708190646e-05, "loss": 0.8814, "step": 2101 }, { "epoch": 0.40454195535026943, "grad_norm": 1.6112996339797974, "learning_rate": 1.3502073812594677e-05, "loss": 1.0339, "step": 2102 }, { "epoch": 0.40473441108545033, "grad_norm": 1.8475773334503174, "learning_rate": 1.349623455629379e-05, "loss": 1.0126, "step": 2103 }, { "epoch": 0.4049268668206313, "grad_norm": 1.754859209060669, "learning_rate": 1.3490393941556787e-05, "loss": 1.0127, "step": 2104 }, { "epoch": 0.4051193225558122, "grad_norm": 2.0299603939056396, "learning_rate": 1.3484551970652997e-05, "loss": 1.0049, "step": 2105 }, { "epoch": 0.4053117782909931, "grad_norm": 1.6877714395523071, "learning_rate": 1.3478708645852272e-05, "loss": 0.9512, "step": 2106 }, { "epoch": 0.40550423402617397, "grad_norm": 2.039612293243408, "learning_rate": 1.3472863969424998e-05, "loss": 0.9732, "step": 2107 }, { "epoch": 0.40569668976135487, "grad_norm": 1.910059928894043, "learning_rate": 1.3467017943642074e-05, "loss": 0.9356, "step": 2108 }, { "epoch": 0.4058891454965358, "grad_norm": 1.7781462669372559, "learning_rate": 1.346117057077493e-05, "loss": 1.0055, "step": 2109 }, { "epoch": 0.4060816012317167, "grad_norm": 1.6408418416976929, "learning_rate": 1.3455321853095527e-05, "loss": 0.9338, "step": 2110 }, { "epoch": 0.4062740569668976, "grad_norm": 1.7563893795013428, "learning_rate": 1.3449471792876333e-05, "loss": 1.0378, "step": 2111 }, { "epoch": 0.4064665127020785, "grad_norm": 0.9516794085502625, "learning_rate": 1.3443620392390352e-05, "loss": 0.9528, "step": 2112 }, { "epoch": 0.4066589684372594, "grad_norm": 1.733107328414917, "learning_rate": 1.3437767653911097e-05, "loss": 0.9256, "step": 2113 }, { "epoch": 0.40685142417244036, "grad_norm": 1.6630865335464478, "learning_rate": 1.3431913579712612e-05, "loss": 1.0075, "step": 2114 }, { "epoch": 0.40704387990762125, "grad_norm": 1.8065743446350098, "learning_rate": 1.3426058172069451e-05, "loss": 0.9268, "step": 2115 }, { "epoch": 0.40723633564280215, "grad_norm": 1.8041304349899292, "learning_rate": 1.342020143325669e-05, "loss": 1.0093, "step": 2116 }, { "epoch": 0.40742879137798305, "grad_norm": 1.630800724029541, "learning_rate": 1.341434336554992e-05, "loss": 1.0045, "step": 2117 }, { "epoch": 0.407621247113164, "grad_norm": 1.8522273302078247, "learning_rate": 1.340848397122525e-05, "loss": 1.0148, "step": 2118 }, { "epoch": 0.4078137028483449, "grad_norm": 1.722927451133728, "learning_rate": 1.3402623252559308e-05, "loss": 0.979, "step": 2119 }, { "epoch": 0.4080061585835258, "grad_norm": 1.8997496366500854, "learning_rate": 1.3396761211829229e-05, "loss": 0.9994, "step": 2120 }, { "epoch": 0.4081986143187067, "grad_norm": 1.7310665845870972, "learning_rate": 1.3390897851312667e-05, "loss": 0.9959, "step": 2121 }, { "epoch": 0.4083910700538876, "grad_norm": 1.7372195720672607, "learning_rate": 1.338503317328779e-05, "loss": 0.9201, "step": 2122 }, { "epoch": 0.40858352578906854, "grad_norm": 1.961373209953308, "learning_rate": 1.3379167180033272e-05, "loss": 0.9361, "step": 2123 }, { "epoch": 0.40877598152424943, "grad_norm": 1.747282862663269, "learning_rate": 1.3373299873828303e-05, "loss": 0.9649, "step": 2124 }, { "epoch": 0.40896843725943033, "grad_norm": 1.8443785905838013, "learning_rate": 1.336743125695258e-05, "loss": 0.9508, "step": 2125 }, { "epoch": 0.4091608929946112, "grad_norm": 1.7303752899169922, "learning_rate": 1.336156133168631e-05, "loss": 0.9475, "step": 2126 }, { "epoch": 0.4093533487297921, "grad_norm": 1.8836239576339722, "learning_rate": 1.3355690100310213e-05, "loss": 0.9388, "step": 2127 }, { "epoch": 0.4095458044649731, "grad_norm": 1.5934606790542603, "learning_rate": 1.3349817565105507e-05, "loss": 0.9769, "step": 2128 }, { "epoch": 0.40973826020015397, "grad_norm": 1.583523154258728, "learning_rate": 1.334394372835393e-05, "loss": 0.8856, "step": 2129 }, { "epoch": 0.40993071593533487, "grad_norm": 1.882719874382019, "learning_rate": 1.333806859233771e-05, "loss": 0.9031, "step": 2130 }, { "epoch": 0.41012317167051576, "grad_norm": 1.912460207939148, "learning_rate": 1.3332192159339595e-05, "loss": 0.9626, "step": 2131 }, { "epoch": 0.4103156274056967, "grad_norm": 1.9093360900878906, "learning_rate": 1.3326314431642821e-05, "loss": 1.0122, "step": 2132 }, { "epoch": 0.4105080831408776, "grad_norm": 1.7700835466384888, "learning_rate": 1.3320435411531145e-05, "loss": 0.9835, "step": 2133 }, { "epoch": 0.4107005388760585, "grad_norm": 1.9040415287017822, "learning_rate": 1.3314555101288808e-05, "loss": 0.911, "step": 2134 }, { "epoch": 0.4108929946112394, "grad_norm": 1.5362709760665894, "learning_rate": 1.3308673503200566e-05, "loss": 1.0066, "step": 2135 }, { "epoch": 0.4110854503464203, "grad_norm": 1.8051828145980835, "learning_rate": 1.3302790619551673e-05, "loss": 0.9186, "step": 2136 }, { "epoch": 0.41127790608160125, "grad_norm": 1.9245274066925049, "learning_rate": 1.3296906452627874e-05, "loss": 0.9363, "step": 2137 }, { "epoch": 0.41147036181678215, "grad_norm": 1.8282278776168823, "learning_rate": 1.3291021004715423e-05, "loss": 1.0698, "step": 2138 }, { "epoch": 0.41166281755196305, "grad_norm": 1.7695746421813965, "learning_rate": 1.3285134278101065e-05, "loss": 1.013, "step": 2139 }, { "epoch": 0.41185527328714394, "grad_norm": 1.7928221225738525, "learning_rate": 1.3279246275072046e-05, "loss": 0.9707, "step": 2140 }, { "epoch": 0.41204772902232484, "grad_norm": 1.8057674169540405, "learning_rate": 1.3273356997916106e-05, "loss": 0.9629, "step": 2141 }, { "epoch": 0.4122401847575058, "grad_norm": 1.9168205261230469, "learning_rate": 1.3267466448921475e-05, "loss": 0.9445, "step": 2142 }, { "epoch": 0.4124326404926867, "grad_norm": 1.8468844890594482, "learning_rate": 1.326157463037689e-05, "loss": 0.9257, "step": 2143 }, { "epoch": 0.4126250962278676, "grad_norm": 1.6158725023269653, "learning_rate": 1.3255681544571568e-05, "loss": 0.9696, "step": 2144 }, { "epoch": 0.4128175519630485, "grad_norm": 1.6677511930465698, "learning_rate": 1.3249787193795227e-05, "loss": 1.0195, "step": 2145 }, { "epoch": 0.41301000769822943, "grad_norm": 1.7644315958023071, "learning_rate": 1.3243891580338074e-05, "loss": 0.896, "step": 2146 }, { "epoch": 0.4132024634334103, "grad_norm": 1.828275203704834, "learning_rate": 1.3237994706490802e-05, "loss": 0.975, "step": 2147 }, { "epoch": 0.4133949191685912, "grad_norm": 1.7779356241226196, "learning_rate": 1.3232096574544602e-05, "loss": 0.9942, "step": 2148 }, { "epoch": 0.4135873749037721, "grad_norm": 1.7545757293701172, "learning_rate": 1.3226197186791147e-05, "loss": 1.0465, "step": 2149 }, { "epoch": 0.413779830638953, "grad_norm": 2.0328774452209473, "learning_rate": 1.3220296545522601e-05, "loss": 0.9834, "step": 2150 }, { "epoch": 0.41397228637413397, "grad_norm": 2.4823338985443115, "learning_rate": 1.3214394653031616e-05, "loss": 1.0356, "step": 2151 }, { "epoch": 0.41416474210931487, "grad_norm": 1.7433737516403198, "learning_rate": 1.320849151161133e-05, "loss": 0.9086, "step": 2152 }, { "epoch": 0.41435719784449576, "grad_norm": 1.8171130418777466, "learning_rate": 1.320258712355536e-05, "loss": 0.9822, "step": 2153 }, { "epoch": 0.41454965357967666, "grad_norm": 1.736508846282959, "learning_rate": 1.3196681491157816e-05, "loss": 0.9474, "step": 2154 }, { "epoch": 0.4147421093148576, "grad_norm": 1.7251181602478027, "learning_rate": 1.3190774616713286e-05, "loss": 0.9887, "step": 2155 }, { "epoch": 0.4149345650500385, "grad_norm": 1.854845643043518, "learning_rate": 1.3184866502516846e-05, "loss": 0.9913, "step": 2156 }, { "epoch": 0.4151270207852194, "grad_norm": 1.637202501296997, "learning_rate": 1.3178957150864047e-05, "loss": 0.989, "step": 2157 }, { "epoch": 0.4153194765204003, "grad_norm": 2.003485918045044, "learning_rate": 1.3173046564050923e-05, "loss": 1.0016, "step": 2158 }, { "epoch": 0.4155119322555812, "grad_norm": 1.8424874544143677, "learning_rate": 1.3167134744373993e-05, "loss": 1.0094, "step": 2159 }, { "epoch": 0.41570438799076215, "grad_norm": 1.872542142868042, "learning_rate": 1.3161221694130247e-05, "loss": 1.0121, "step": 2160 }, { "epoch": 0.41589684372594304, "grad_norm": 1.0590440034866333, "learning_rate": 1.3155307415617156e-05, "loss": 0.9789, "step": 2161 }, { "epoch": 0.41608929946112394, "grad_norm": 1.6069647073745728, "learning_rate": 1.3149391911132674e-05, "loss": 0.928, "step": 2162 }, { "epoch": 0.41628175519630484, "grad_norm": 1.890120029449463, "learning_rate": 1.3143475182975222e-05, "loss": 0.9673, "step": 2163 }, { "epoch": 0.41647421093148573, "grad_norm": 1.5654081106185913, "learning_rate": 1.3137557233443707e-05, "loss": 0.973, "step": 2164 }, { "epoch": 0.4166666666666667, "grad_norm": 1.7641433477401733, "learning_rate": 1.3131638064837496e-05, "loss": 0.9598, "step": 2165 }, { "epoch": 0.4168591224018476, "grad_norm": 1.8171759843826294, "learning_rate": 1.3125717679456447e-05, "loss": 1.0279, "step": 2166 }, { "epoch": 0.4170515781370285, "grad_norm": 0.8760558366775513, "learning_rate": 1.3119796079600875e-05, "loss": 0.9293, "step": 2167 }, { "epoch": 0.4172440338722094, "grad_norm": 1.8492945432662964, "learning_rate": 1.3113873267571577e-05, "loss": 0.9772, "step": 2168 }, { "epoch": 0.4174364896073903, "grad_norm": 1.8009836673736572, "learning_rate": 1.310794924566982e-05, "loss": 0.9755, "step": 2169 }, { "epoch": 0.4176289453425712, "grad_norm": 1.6825505495071411, "learning_rate": 1.3102024016197337e-05, "loss": 1.0303, "step": 2170 }, { "epoch": 0.4178214010777521, "grad_norm": 1.7613478899002075, "learning_rate": 1.309609758145633e-05, "loss": 0.9832, "step": 2171 }, { "epoch": 0.418013856812933, "grad_norm": 1.727813720703125, "learning_rate": 1.3090169943749475e-05, "loss": 0.994, "step": 2172 }, { "epoch": 0.4182063125481139, "grad_norm": 1.7242426872253418, "learning_rate": 1.3084241105379909e-05, "loss": 0.9502, "step": 2173 }, { "epoch": 0.41839876828329486, "grad_norm": 1.5613230466842651, "learning_rate": 1.3078311068651241e-05, "loss": 1.0132, "step": 2174 }, { "epoch": 0.41859122401847576, "grad_norm": 1.6472156047821045, "learning_rate": 1.307237983586754e-05, "loss": 0.9257, "step": 2175 }, { "epoch": 0.41878367975365666, "grad_norm": 1.960904836654663, "learning_rate": 1.3066447409333345e-05, "loss": 1.0201, "step": 2176 }, { "epoch": 0.41897613548883755, "grad_norm": 1.7344560623168945, "learning_rate": 1.3060513791353655e-05, "loss": 1.0485, "step": 2177 }, { "epoch": 0.41916859122401845, "grad_norm": 1.684669852256775, "learning_rate": 1.3054578984233935e-05, "loss": 0.9865, "step": 2178 }, { "epoch": 0.4193610469591994, "grad_norm": 1.727281093597412, "learning_rate": 1.3048642990280109e-05, "loss": 1.0318, "step": 2179 }, { "epoch": 0.4195535026943803, "grad_norm": 1.5963596105575562, "learning_rate": 1.3042705811798565e-05, "loss": 0.9912, "step": 2180 }, { "epoch": 0.4197459584295612, "grad_norm": 1.729940414428711, "learning_rate": 1.3036767451096148e-05, "loss": 1.0519, "step": 2181 }, { "epoch": 0.4199384141647421, "grad_norm": 1.8296971321105957, "learning_rate": 1.3030827910480167e-05, "loss": 0.963, "step": 2182 }, { "epoch": 0.42013086989992304, "grad_norm": 1.6504627466201782, "learning_rate": 1.3024887192258384e-05, "loss": 0.9672, "step": 2183 }, { "epoch": 0.42032332563510394, "grad_norm": 1.7737363576889038, "learning_rate": 1.3018945298739022e-05, "loss": 0.9244, "step": 2184 }, { "epoch": 0.42051578137028484, "grad_norm": 1.8815377950668335, "learning_rate": 1.301300223223076e-05, "loss": 1.0183, "step": 2185 }, { "epoch": 0.42070823710546573, "grad_norm": 1.9204562902450562, "learning_rate": 1.300705799504273e-05, "loss": 0.888, "step": 2186 }, { "epoch": 0.42090069284064663, "grad_norm": 1.6020839214324951, "learning_rate": 1.300111258948453e-05, "loss": 0.9312, "step": 2187 }, { "epoch": 0.4210931485758276, "grad_norm": 1.7201136350631714, "learning_rate": 1.2995166017866194e-05, "loss": 0.9405, "step": 2188 }, { "epoch": 0.4212856043110085, "grad_norm": 1.6900391578674316, "learning_rate": 1.2989218282498222e-05, "loss": 0.9898, "step": 2189 }, { "epoch": 0.4214780600461894, "grad_norm": 1.765326976776123, "learning_rate": 1.2983269385691562e-05, "loss": 0.9828, "step": 2190 }, { "epoch": 0.42167051578137027, "grad_norm": 1.699736475944519, "learning_rate": 1.2977319329757616e-05, "loss": 0.8893, "step": 2191 }, { "epoch": 0.42186297151655117, "grad_norm": 1.5959320068359375, "learning_rate": 1.2971368117008232e-05, "loss": 0.9228, "step": 2192 }, { "epoch": 0.4220554272517321, "grad_norm": 1.7840313911437988, "learning_rate": 1.296541574975571e-05, "loss": 0.9655, "step": 2193 }, { "epoch": 0.422247882986913, "grad_norm": 1.6608049869537354, "learning_rate": 1.29594622303128e-05, "loss": 0.9042, "step": 2194 }, { "epoch": 0.4224403387220939, "grad_norm": 1.8267734050750732, "learning_rate": 1.2953507560992696e-05, "loss": 0.9715, "step": 2195 }, { "epoch": 0.4226327944572748, "grad_norm": 1.8622692823410034, "learning_rate": 1.2947551744109044e-05, "loss": 0.9266, "step": 2196 }, { "epoch": 0.42282525019245576, "grad_norm": 1.8044354915618896, "learning_rate": 1.2941594781975928e-05, "loss": 0.9829, "step": 2197 }, { "epoch": 0.42301770592763666, "grad_norm": 1.0172758102416992, "learning_rate": 1.2935636676907886e-05, "loss": 0.9644, "step": 2198 }, { "epoch": 0.42321016166281755, "grad_norm": 1.755821943283081, "learning_rate": 1.292967743121989e-05, "loss": 1.0467, "step": 2199 }, { "epoch": 0.42340261739799845, "grad_norm": 1.9822099208831787, "learning_rate": 1.2923717047227368e-05, "loss": 0.9019, "step": 2200 }, { "epoch": 0.42359507313317935, "grad_norm": 1.7691650390625, "learning_rate": 1.2917755527246179e-05, "loss": 0.986, "step": 2201 }, { "epoch": 0.4237875288683603, "grad_norm": 1.757727026939392, "learning_rate": 1.2911792873592628e-05, "loss": 0.9468, "step": 2202 }, { "epoch": 0.4239799846035412, "grad_norm": 1.78410005569458, "learning_rate": 1.2905829088583462e-05, "loss": 1.0062, "step": 2203 }, { "epoch": 0.4241724403387221, "grad_norm": 1.7611806392669678, "learning_rate": 1.2899864174535863e-05, "loss": 0.9386, "step": 2204 }, { "epoch": 0.424364896073903, "grad_norm": 1.7355962991714478, "learning_rate": 1.2893898133767454e-05, "loss": 0.8919, "step": 2205 }, { "epoch": 0.42455735180908394, "grad_norm": 1.7231559753417969, "learning_rate": 1.28879309685963e-05, "loss": 0.9126, "step": 2206 }, { "epoch": 0.42474980754426483, "grad_norm": 1.8716305494308472, "learning_rate": 1.2881962681340894e-05, "loss": 1.012, "step": 2207 }, { "epoch": 0.42494226327944573, "grad_norm": 1.8043121099472046, "learning_rate": 1.2875993274320173e-05, "loss": 1.1053, "step": 2208 }, { "epoch": 0.4251347190146266, "grad_norm": 1.5806753635406494, "learning_rate": 1.2870022749853508e-05, "loss": 0.9522, "step": 2209 }, { "epoch": 0.4253271747498075, "grad_norm": 1.84601891040802, "learning_rate": 1.2864051110260697e-05, "loss": 1.0343, "step": 2210 }, { "epoch": 0.4255196304849885, "grad_norm": 1.7185986042022705, "learning_rate": 1.2858078357861979e-05, "loss": 1.0011, "step": 2211 }, { "epoch": 0.4257120862201694, "grad_norm": 1.7174954414367676, "learning_rate": 1.2852104494978024e-05, "loss": 0.9939, "step": 2212 }, { "epoch": 0.42590454195535027, "grad_norm": 1.7660080194473267, "learning_rate": 1.2846129523929929e-05, "loss": 0.9446, "step": 2213 }, { "epoch": 0.42609699769053117, "grad_norm": 1.7342970371246338, "learning_rate": 1.284015344703923e-05, "loss": 1.0007, "step": 2214 }, { "epoch": 0.42628945342571206, "grad_norm": 1.8750807046890259, "learning_rate": 1.283417626662788e-05, "loss": 0.9907, "step": 2215 }, { "epoch": 0.426481909160893, "grad_norm": 1.7674480676651, "learning_rate": 1.2828197985018276e-05, "loss": 1.0495, "step": 2216 }, { "epoch": 0.4266743648960739, "grad_norm": 1.8894813060760498, "learning_rate": 1.2822218604533232e-05, "loss": 0.9409, "step": 2217 }, { "epoch": 0.4268668206312548, "grad_norm": 0.9724174737930298, "learning_rate": 1.281623812749599e-05, "loss": 0.9422, "step": 2218 }, { "epoch": 0.4270592763664357, "grad_norm": 1.728150725364685, "learning_rate": 1.2810256556230226e-05, "loss": 0.965, "step": 2219 }, { "epoch": 0.42725173210161665, "grad_norm": 1.554421305656433, "learning_rate": 1.2804273893060028e-05, "loss": 0.9872, "step": 2220 }, { "epoch": 0.42744418783679755, "grad_norm": 0.9607255458831787, "learning_rate": 1.2798290140309924e-05, "loss": 0.9438, "step": 2221 }, { "epoch": 0.42763664357197845, "grad_norm": 1.990488052368164, "learning_rate": 1.2792305300304848e-05, "loss": 0.9144, "step": 2222 }, { "epoch": 0.42782909930715934, "grad_norm": 1.7820043563842773, "learning_rate": 1.278631937537017e-05, "loss": 0.9763, "step": 2223 }, { "epoch": 0.42802155504234024, "grad_norm": 1.839257836341858, "learning_rate": 1.2780332367831678e-05, "loss": 0.981, "step": 2224 }, { "epoch": 0.4282140107775212, "grad_norm": 1.6478990316390991, "learning_rate": 1.2774344280015578e-05, "loss": 0.9708, "step": 2225 }, { "epoch": 0.4284064665127021, "grad_norm": 2.0120506286621094, "learning_rate": 1.2768355114248493e-05, "loss": 1.0008, "step": 2226 }, { "epoch": 0.428598922247883, "grad_norm": 1.7044005393981934, "learning_rate": 1.276236487285748e-05, "loss": 0.9082, "step": 2227 }, { "epoch": 0.4287913779830639, "grad_norm": 1.7084825038909912, "learning_rate": 1.2756373558169992e-05, "loss": 0.9702, "step": 2228 }, { "epoch": 0.4289838337182448, "grad_norm": 0.9720837473869324, "learning_rate": 1.2750381172513918e-05, "loss": 0.918, "step": 2229 }, { "epoch": 0.42917628945342573, "grad_norm": 2.050966739654541, "learning_rate": 1.274438771821755e-05, "loss": 0.9949, "step": 2230 }, { "epoch": 0.4293687451886066, "grad_norm": 1.8797640800476074, "learning_rate": 1.2738393197609602e-05, "loss": 1.0176, "step": 2231 }, { "epoch": 0.4295612009237875, "grad_norm": 1.8301764726638794, "learning_rate": 1.2732397613019203e-05, "loss": 0.9814, "step": 2232 }, { "epoch": 0.4297536566589684, "grad_norm": 1.748365879058838, "learning_rate": 1.272640096677589e-05, "loss": 0.9427, "step": 2233 }, { "epoch": 0.42994611239414937, "grad_norm": 1.956606388092041, "learning_rate": 1.2720403261209615e-05, "loss": 1.0727, "step": 2234 }, { "epoch": 0.43013856812933027, "grad_norm": 1.689342975616455, "learning_rate": 1.2714404498650743e-05, "loss": 0.9494, "step": 2235 }, { "epoch": 0.43033102386451116, "grad_norm": 1.7013987302780151, "learning_rate": 1.2708404681430054e-05, "loss": 0.9548, "step": 2236 }, { "epoch": 0.43052347959969206, "grad_norm": 1.7946473360061646, "learning_rate": 1.2702403811878727e-05, "loss": 0.9174, "step": 2237 }, { "epoch": 0.43071593533487296, "grad_norm": 0.9426779747009277, "learning_rate": 1.2696401892328356e-05, "loss": 0.9534, "step": 2238 }, { "epoch": 0.4309083910700539, "grad_norm": 1.6412510871887207, "learning_rate": 1.2690398925110943e-05, "loss": 0.9886, "step": 2239 }, { "epoch": 0.4311008468052348, "grad_norm": 1.8316644430160522, "learning_rate": 1.2684394912558898e-05, "loss": 0.9508, "step": 2240 }, { "epoch": 0.4312933025404157, "grad_norm": 1.6385741233825684, "learning_rate": 1.2678389857005033e-05, "loss": 1.0622, "step": 2241 }, { "epoch": 0.4314857582755966, "grad_norm": 1.7953205108642578, "learning_rate": 1.267238376078257e-05, "loss": 1.0016, "step": 2242 }, { "epoch": 0.4316782140107775, "grad_norm": 1.8662776947021484, "learning_rate": 1.2666376626225131e-05, "loss": 0.9349, "step": 2243 }, { "epoch": 0.43187066974595845, "grad_norm": 1.8582065105438232, "learning_rate": 1.2660368455666752e-05, "loss": 0.9972, "step": 2244 }, { "epoch": 0.43206312548113934, "grad_norm": 1.6141120195388794, "learning_rate": 1.2654359251441854e-05, "loss": 1.032, "step": 2245 }, { "epoch": 0.43225558121632024, "grad_norm": 1.690446138381958, "learning_rate": 1.2648349015885272e-05, "loss": 1.0019, "step": 2246 }, { "epoch": 0.43244803695150114, "grad_norm": 1.7220337390899658, "learning_rate": 1.2642337751332238e-05, "loss": 0.9405, "step": 2247 }, { "epoch": 0.4326404926866821, "grad_norm": 1.9969561100006104, "learning_rate": 1.2636325460118388e-05, "loss": 1.0287, "step": 2248 }, { "epoch": 0.432832948421863, "grad_norm": 1.6311569213867188, "learning_rate": 1.2630312144579748e-05, "loss": 1.0141, "step": 2249 }, { "epoch": 0.4330254041570439, "grad_norm": 1.6893813610076904, "learning_rate": 1.262429780705275e-05, "loss": 1.0071, "step": 2250 }, { "epoch": 0.4332178598922248, "grad_norm": 1.8757095336914062, "learning_rate": 1.2618282449874221e-05, "loss": 0.966, "step": 2251 }, { "epoch": 0.4334103156274057, "grad_norm": 1.7034698724746704, "learning_rate": 1.2612266075381385e-05, "loss": 1.0504, "step": 2252 }, { "epoch": 0.4336027713625866, "grad_norm": 1.6506664752960205, "learning_rate": 1.2606248685911856e-05, "loss": 0.921, "step": 2253 }, { "epoch": 0.4337952270977675, "grad_norm": 1.750646948814392, "learning_rate": 1.2600230283803648e-05, "loss": 0.9419, "step": 2254 }, { "epoch": 0.4339876828329484, "grad_norm": 1.60439133644104, "learning_rate": 1.2594210871395164e-05, "loss": 0.9988, "step": 2255 }, { "epoch": 0.4341801385681293, "grad_norm": 1.0411254167556763, "learning_rate": 1.2588190451025209e-05, "loss": 0.9515, "step": 2256 }, { "epoch": 0.4343725943033102, "grad_norm": 0.9826929569244385, "learning_rate": 1.2582169025032966e-05, "loss": 0.8949, "step": 2257 }, { "epoch": 0.43456505003849116, "grad_norm": 1.7708104848861694, "learning_rate": 1.2576146595758019e-05, "loss": 1.0034, "step": 2258 }, { "epoch": 0.43475750577367206, "grad_norm": 1.9281541109085083, "learning_rate": 1.2570123165540338e-05, "loss": 0.9937, "step": 2259 }, { "epoch": 0.43494996150885296, "grad_norm": 1.7999751567840576, "learning_rate": 1.2564098736720286e-05, "loss": 0.9909, "step": 2260 }, { "epoch": 0.43514241724403385, "grad_norm": 1.8252668380737305, "learning_rate": 1.2558073311638604e-05, "loss": 0.9811, "step": 2261 }, { "epoch": 0.4353348729792148, "grad_norm": 0.989126980304718, "learning_rate": 1.2552046892636427e-05, "loss": 0.9209, "step": 2262 }, { "epoch": 0.4355273287143957, "grad_norm": 1.039644718170166, "learning_rate": 1.2546019482055276e-05, "loss": 0.9931, "step": 2263 }, { "epoch": 0.4357197844495766, "grad_norm": 1.9120420217514038, "learning_rate": 1.2539991082237062e-05, "loss": 1.0181, "step": 2264 }, { "epoch": 0.4359122401847575, "grad_norm": 1.7508251667022705, "learning_rate": 1.2533961695524071e-05, "loss": 0.9732, "step": 2265 }, { "epoch": 0.4361046959199384, "grad_norm": 1.9827975034713745, "learning_rate": 1.2527931324258975e-05, "loss": 0.9873, "step": 2266 }, { "epoch": 0.43629715165511934, "grad_norm": 1.8357239961624146, "learning_rate": 1.2521899970784834e-05, "loss": 1.0083, "step": 2267 }, { "epoch": 0.43648960739030024, "grad_norm": 1.6103686094284058, "learning_rate": 1.2515867637445088e-05, "loss": 0.9782, "step": 2268 }, { "epoch": 0.43668206312548113, "grad_norm": 1.7222421169281006, "learning_rate": 1.2509834326583548e-05, "loss": 0.9431, "step": 2269 }, { "epoch": 0.43687451886066203, "grad_norm": 1.618837833404541, "learning_rate": 1.2503800040544417e-05, "loss": 0.9374, "step": 2270 }, { "epoch": 0.437066974595843, "grad_norm": 1.0346992015838623, "learning_rate": 1.249776478167227e-05, "loss": 0.9179, "step": 2271 }, { "epoch": 0.4372594303310239, "grad_norm": 1.80836820602417, "learning_rate": 1.2491728552312066e-05, "loss": 0.9635, "step": 2272 }, { "epoch": 0.4374518860662048, "grad_norm": 1.76581609249115, "learning_rate": 1.2485691354809134e-05, "loss": 0.9469, "step": 2273 }, { "epoch": 0.4376443418013857, "grad_norm": 1.781143069267273, "learning_rate": 1.2479653191509182e-05, "loss": 0.9359, "step": 2274 }, { "epoch": 0.43783679753656657, "grad_norm": 1.7076761722564697, "learning_rate": 1.2473614064758297e-05, "loss": 0.9341, "step": 2275 }, { "epoch": 0.4380292532717475, "grad_norm": 1.892514705657959, "learning_rate": 1.2467573976902936e-05, "loss": 0.9111, "step": 2276 }, { "epoch": 0.4382217090069284, "grad_norm": 1.7116341590881348, "learning_rate": 1.2461532930289932e-05, "loss": 1.0177, "step": 2277 }, { "epoch": 0.4384141647421093, "grad_norm": 1.6574907302856445, "learning_rate": 1.2455490927266485e-05, "loss": 1.0071, "step": 2278 }, { "epoch": 0.4386066204772902, "grad_norm": 1.8197227716445923, "learning_rate": 1.2449447970180172e-05, "loss": 0.996, "step": 2279 }, { "epoch": 0.4387990762124711, "grad_norm": 2.4165401458740234, "learning_rate": 1.2443404061378941e-05, "loss": 0.9799, "step": 2280 }, { "epoch": 0.43899153194765206, "grad_norm": 1.6268476247787476, "learning_rate": 1.2437359203211109e-05, "loss": 1.0153, "step": 2281 }, { "epoch": 0.43918398768283295, "grad_norm": 1.7256579399108887, "learning_rate": 1.2431313398025356e-05, "loss": 0.8734, "step": 2282 }, { "epoch": 0.43937644341801385, "grad_norm": 2.0695931911468506, "learning_rate": 1.242526664817074e-05, "loss": 0.983, "step": 2283 }, { "epoch": 0.43956889915319475, "grad_norm": 1.7578411102294922, "learning_rate": 1.2419218955996677e-05, "loss": 1.0043, "step": 2284 }, { "epoch": 0.4397613548883757, "grad_norm": 1.9387781620025635, "learning_rate": 1.2413170323852959e-05, "loss": 0.873, "step": 2285 }, { "epoch": 0.4399538106235566, "grad_norm": 1.722853422164917, "learning_rate": 1.2407120754089733e-05, "loss": 0.8412, "step": 2286 }, { "epoch": 0.4401462663587375, "grad_norm": 1.8679687976837158, "learning_rate": 1.2401070249057514e-05, "loss": 0.9789, "step": 2287 }, { "epoch": 0.4403387220939184, "grad_norm": 1.719572901725769, "learning_rate": 1.239501881110718e-05, "loss": 0.9808, "step": 2288 }, { "epoch": 0.4405311778290993, "grad_norm": 1.8278943300247192, "learning_rate": 1.2388966442589978e-05, "loss": 1.0696, "step": 2289 }, { "epoch": 0.44072363356428024, "grad_norm": 1.7576541900634766, "learning_rate": 1.2382913145857507e-05, "loss": 1.0014, "step": 2290 }, { "epoch": 0.44091608929946113, "grad_norm": 1.8827203512191772, "learning_rate": 1.2376858923261732e-05, "loss": 0.9696, "step": 2291 }, { "epoch": 0.44110854503464203, "grad_norm": 1.7177938222885132, "learning_rate": 1.2370803777154976e-05, "loss": 0.9806, "step": 2292 }, { "epoch": 0.4413010007698229, "grad_norm": 1.801792025566101, "learning_rate": 1.2364747709889923e-05, "loss": 0.9961, "step": 2293 }, { "epoch": 0.4414934565050038, "grad_norm": 1.676908016204834, "learning_rate": 1.2358690723819611e-05, "loss": 0.9522, "step": 2294 }, { "epoch": 0.4416859122401848, "grad_norm": 1.8079555034637451, "learning_rate": 1.2352632821297437e-05, "loss": 0.9756, "step": 2295 }, { "epoch": 0.44187836797536567, "grad_norm": 1.579372763633728, "learning_rate": 1.2346574004677154e-05, "loss": 0.9839, "step": 2296 }, { "epoch": 0.44207082371054657, "grad_norm": 1.6868959665298462, "learning_rate": 1.2340514276312874e-05, "loss": 1.046, "step": 2297 }, { "epoch": 0.44226327944572746, "grad_norm": 0.975253701210022, "learning_rate": 1.2334453638559057e-05, "loss": 0.9983, "step": 2298 }, { "epoch": 0.4424557351809084, "grad_norm": 1.6039180755615234, "learning_rate": 1.2328392093770517e-05, "loss": 0.9349, "step": 2299 }, { "epoch": 0.4426481909160893, "grad_norm": 1.719197154045105, "learning_rate": 1.2322329644302426e-05, "loss": 1.0026, "step": 2300 }, { "epoch": 0.4428406466512702, "grad_norm": 1.8221049308776855, "learning_rate": 1.2316266292510305e-05, "loss": 1.0555, "step": 2301 }, { "epoch": 0.4430331023864511, "grad_norm": 1.7465460300445557, "learning_rate": 1.2310202040750022e-05, "loss": 0.8497, "step": 2302 }, { "epoch": 0.443225558121632, "grad_norm": 1.6725069284439087, "learning_rate": 1.2304136891377795e-05, "loss": 0.9563, "step": 2303 }, { "epoch": 0.44341801385681295, "grad_norm": 1.6833906173706055, "learning_rate": 1.2298070846750197e-05, "loss": 1.0353, "step": 2304 }, { "epoch": 0.44361046959199385, "grad_norm": 1.8982234001159668, "learning_rate": 1.2292003909224144e-05, "loss": 0.9492, "step": 2305 }, { "epoch": 0.44380292532717475, "grad_norm": 1.7660242319107056, "learning_rate": 1.2285936081156897e-05, "loss": 1.0919, "step": 2306 }, { "epoch": 0.44399538106235564, "grad_norm": 1.8644734621047974, "learning_rate": 1.227986736490607e-05, "loss": 0.9978, "step": 2307 }, { "epoch": 0.44418783679753654, "grad_norm": 1.7476723194122314, "learning_rate": 1.2273797762829615e-05, "loss": 1.016, "step": 2308 }, { "epoch": 0.4443802925327175, "grad_norm": 1.7577075958251953, "learning_rate": 1.2267727277285832e-05, "loss": 1.0152, "step": 2309 }, { "epoch": 0.4445727482678984, "grad_norm": 1.7524163722991943, "learning_rate": 1.2261655910633365e-05, "loss": 0.9563, "step": 2310 }, { "epoch": 0.4447652040030793, "grad_norm": 1.5874872207641602, "learning_rate": 1.2255583665231196e-05, "loss": 0.9922, "step": 2311 }, { "epoch": 0.4449576597382602, "grad_norm": 1.7612804174423218, "learning_rate": 1.2249510543438652e-05, "loss": 0.9547, "step": 2312 }, { "epoch": 0.44515011547344113, "grad_norm": 1.5578655004501343, "learning_rate": 1.22434365476154e-05, "loss": 0.9072, "step": 2313 }, { "epoch": 0.44534257120862203, "grad_norm": 0.8939824104309082, "learning_rate": 1.2237361680121445e-05, "loss": 0.9113, "step": 2314 }, { "epoch": 0.4455350269438029, "grad_norm": 1.6066582202911377, "learning_rate": 1.2231285943317132e-05, "loss": 0.9424, "step": 2315 }, { "epoch": 0.4457274826789838, "grad_norm": 1.7614566087722778, "learning_rate": 1.2225209339563144e-05, "loss": 0.8776, "step": 2316 }, { "epoch": 0.4459199384141647, "grad_norm": 0.8848643898963928, "learning_rate": 1.2219131871220504e-05, "loss": 0.9564, "step": 2317 }, { "epoch": 0.44611239414934567, "grad_norm": 1.8702956438064575, "learning_rate": 1.2213053540650566e-05, "loss": 0.9566, "step": 2318 }, { "epoch": 0.44630484988452657, "grad_norm": 1.7226518392562866, "learning_rate": 1.2206974350215016e-05, "loss": 0.9922, "step": 2319 }, { "epoch": 0.44649730561970746, "grad_norm": 2.0815463066101074, "learning_rate": 1.2200894302275878e-05, "loss": 0.9619, "step": 2320 }, { "epoch": 0.44668976135488836, "grad_norm": 1.7683513164520264, "learning_rate": 1.2194813399195518e-05, "loss": 0.8838, "step": 2321 }, { "epoch": 0.4468822170900693, "grad_norm": 1.815604567527771, "learning_rate": 1.2188731643336615e-05, "loss": 1.0202, "step": 2322 }, { "epoch": 0.4470746728252502, "grad_norm": 1.6910731792449951, "learning_rate": 1.2182649037062203e-05, "loss": 0.9094, "step": 2323 }, { "epoch": 0.4472671285604311, "grad_norm": 2.0379199981689453, "learning_rate": 1.2176565582735624e-05, "loss": 0.9514, "step": 2324 }, { "epoch": 0.447459584295612, "grad_norm": 1.9798578023910522, "learning_rate": 1.2170481282720563e-05, "loss": 0.8981, "step": 2325 }, { "epoch": 0.4476520400307929, "grad_norm": 1.7403302192687988, "learning_rate": 1.2164396139381029e-05, "loss": 0.9123, "step": 2326 }, { "epoch": 0.44784449576597385, "grad_norm": 1.7368252277374268, "learning_rate": 1.215831015508136e-05, "loss": 0.9747, "step": 2327 }, { "epoch": 0.44803695150115475, "grad_norm": 1.7099689245224, "learning_rate": 1.2152223332186222e-05, "loss": 0.9494, "step": 2328 }, { "epoch": 0.44822940723633564, "grad_norm": 1.9472322463989258, "learning_rate": 1.2146135673060599e-05, "loss": 0.9899, "step": 2329 }, { "epoch": 0.44842186297151654, "grad_norm": 1.759556770324707, "learning_rate": 1.2140047180069814e-05, "loss": 0.9137, "step": 2330 }, { "epoch": 0.44861431870669743, "grad_norm": 1.8879690170288086, "learning_rate": 1.2133957855579501e-05, "loss": 0.9549, "step": 2331 }, { "epoch": 0.4488067744418784, "grad_norm": 1.6290510892868042, "learning_rate": 1.2127867701955622e-05, "loss": 0.8811, "step": 2332 }, { "epoch": 0.4489992301770593, "grad_norm": 1.7757058143615723, "learning_rate": 1.2121776721564465e-05, "loss": 1.0268, "step": 2333 }, { "epoch": 0.4491916859122402, "grad_norm": 1.7286909818649292, "learning_rate": 1.211568491677263e-05, "loss": 0.9837, "step": 2334 }, { "epoch": 0.4493841416474211, "grad_norm": 1.839980125427246, "learning_rate": 1.2109592289947045e-05, "loss": 0.9831, "step": 2335 }, { "epoch": 0.44957659738260203, "grad_norm": 1.8710060119628906, "learning_rate": 1.210349884345496e-05, "loss": 0.9616, "step": 2336 }, { "epoch": 0.4497690531177829, "grad_norm": 1.7398589849472046, "learning_rate": 1.209740457966393e-05, "loss": 0.9617, "step": 2337 }, { "epoch": 0.4499615088529638, "grad_norm": 1.7421640157699585, "learning_rate": 1.2091309500941841e-05, "loss": 1.0134, "step": 2338 }, { "epoch": 0.4501539645881447, "grad_norm": 1.6047028303146362, "learning_rate": 1.2085213609656891e-05, "loss": 0.9958, "step": 2339 }, { "epoch": 0.4503464203233256, "grad_norm": 1.0608611106872559, "learning_rate": 1.2079116908177592e-05, "loss": 0.9298, "step": 2340 }, { "epoch": 0.45053887605850657, "grad_norm": 1.7856225967407227, "learning_rate": 1.2073019398872778e-05, "loss": 0.9437, "step": 2341 }, { "epoch": 0.45073133179368746, "grad_norm": 1.5963258743286133, "learning_rate": 1.2066921084111582e-05, "loss": 1.0112, "step": 2342 }, { "epoch": 0.45092378752886836, "grad_norm": 1.9113974571228027, "learning_rate": 1.2060821966263465e-05, "loss": 1.0067, "step": 2343 }, { "epoch": 0.45111624326404925, "grad_norm": 1.6153010129928589, "learning_rate": 1.2054722047698192e-05, "loss": 0.9021, "step": 2344 }, { "epoch": 0.45130869899923015, "grad_norm": 1.9282292127609253, "learning_rate": 1.2048621330785842e-05, "loss": 1.0234, "step": 2345 }, { "epoch": 0.4515011547344111, "grad_norm": 1.8972386121749878, "learning_rate": 1.2042519817896805e-05, "loss": 1.0598, "step": 2346 }, { "epoch": 0.451693610469592, "grad_norm": 1.8205217123031616, "learning_rate": 1.2036417511401775e-05, "loss": 0.9896, "step": 2347 }, { "epoch": 0.4518860662047729, "grad_norm": 1.7705858945846558, "learning_rate": 1.2030314413671763e-05, "loss": 0.9758, "step": 2348 }, { "epoch": 0.4520785219399538, "grad_norm": 1.6301497220993042, "learning_rate": 1.202421052707808e-05, "loss": 1.0814, "step": 2349 }, { "epoch": 0.45227097767513474, "grad_norm": 1.6599459648132324, "learning_rate": 1.2018105853992346e-05, "loss": 0.9416, "step": 2350 }, { "epoch": 0.45246343341031564, "grad_norm": 1.70463228225708, "learning_rate": 1.2012000396786485e-05, "loss": 0.9901, "step": 2351 }, { "epoch": 0.45265588914549654, "grad_norm": 1.7657744884490967, "learning_rate": 1.200589415783273e-05, "loss": 0.9129, "step": 2352 }, { "epoch": 0.45284834488067743, "grad_norm": 1.8150169849395752, "learning_rate": 1.1999787139503614e-05, "loss": 1.0211, "step": 2353 }, { "epoch": 0.45304080061585833, "grad_norm": 1.4867533445358276, "learning_rate": 1.1993679344171973e-05, "loss": 0.8673, "step": 2354 }, { "epoch": 0.4532332563510393, "grad_norm": 1.7130755186080933, "learning_rate": 1.1987570774210948e-05, "loss": 0.926, "step": 2355 }, { "epoch": 0.4534257120862202, "grad_norm": 2.123491048812866, "learning_rate": 1.1981461431993978e-05, "loss": 0.9988, "step": 2356 }, { "epoch": 0.4536181678214011, "grad_norm": 1.710565209388733, "learning_rate": 1.1975351319894801e-05, "loss": 1.0184, "step": 2357 }, { "epoch": 0.45381062355658197, "grad_norm": 1.7636826038360596, "learning_rate": 1.1969240440287457e-05, "loss": 1.003, "step": 2358 }, { "epoch": 0.45400307929176287, "grad_norm": 1.7262758016586304, "learning_rate": 1.1963128795546286e-05, "loss": 0.9609, "step": 2359 }, { "epoch": 0.4541955350269438, "grad_norm": 1.792893409729004, "learning_rate": 1.1957016388045917e-05, "loss": 0.9874, "step": 2360 }, { "epoch": 0.4543879907621247, "grad_norm": 1.7145204544067383, "learning_rate": 1.1950903220161286e-05, "loss": 0.9055, "step": 2361 }, { "epoch": 0.4545804464973056, "grad_norm": 1.794786810874939, "learning_rate": 1.1944789294267614e-05, "loss": 0.9626, "step": 2362 }, { "epoch": 0.4547729022324865, "grad_norm": 2.554239273071289, "learning_rate": 1.1938674612740424e-05, "loss": 0.9202, "step": 2363 }, { "epoch": 0.45496535796766746, "grad_norm": 1.9347246885299683, "learning_rate": 1.1932559177955533e-05, "loss": 1.0477, "step": 2364 }, { "epoch": 0.45515781370284836, "grad_norm": 1.6353051662445068, "learning_rate": 1.1926442992289047e-05, "loss": 0.9779, "step": 2365 }, { "epoch": 0.45535026943802925, "grad_norm": 1.786901831626892, "learning_rate": 1.1920326058117364e-05, "loss": 0.9045, "step": 2366 }, { "epoch": 0.45554272517321015, "grad_norm": 1.7107051610946655, "learning_rate": 1.1914208377817171e-05, "loss": 0.9778, "step": 2367 }, { "epoch": 0.45573518090839105, "grad_norm": 2.0546085834503174, "learning_rate": 1.190808995376545e-05, "loss": 0.9931, "step": 2368 }, { "epoch": 0.455927636643572, "grad_norm": 2.1350810527801514, "learning_rate": 1.190197078833947e-05, "loss": 0.8733, "step": 2369 }, { "epoch": 0.4561200923787529, "grad_norm": 1.6949878931045532, "learning_rate": 1.1895850883916786e-05, "loss": 0.9573, "step": 2370 }, { "epoch": 0.4563125481139338, "grad_norm": 1.9167717695236206, "learning_rate": 1.1889730242875243e-05, "loss": 0.9499, "step": 2371 }, { "epoch": 0.4565050038491147, "grad_norm": 1.6809133291244507, "learning_rate": 1.188360886759297e-05, "loss": 1.0228, "step": 2372 }, { "epoch": 0.45669745958429564, "grad_norm": 2.066824436187744, "learning_rate": 1.1877486760448382e-05, "loss": 0.9733, "step": 2373 }, { "epoch": 0.45688991531947654, "grad_norm": 1.70389986038208, "learning_rate": 1.1871363923820182e-05, "loss": 0.9263, "step": 2374 }, { "epoch": 0.45708237105465743, "grad_norm": 1.8637346029281616, "learning_rate": 1.1865240360087349e-05, "loss": 0.9984, "step": 2375 }, { "epoch": 0.45727482678983833, "grad_norm": 1.7322235107421875, "learning_rate": 1.1859116071629148e-05, "loss": 0.991, "step": 2376 }, { "epoch": 0.4574672825250192, "grad_norm": 1.668251872062683, "learning_rate": 1.1852991060825128e-05, "loss": 0.9865, "step": 2377 }, { "epoch": 0.4576597382602002, "grad_norm": 2.0884711742401123, "learning_rate": 1.1846865330055115e-05, "loss": 1.0185, "step": 2378 }, { "epoch": 0.4578521939953811, "grad_norm": 0.9864965677261353, "learning_rate": 1.1840738881699218e-05, "loss": 0.9599, "step": 2379 }, { "epoch": 0.45804464973056197, "grad_norm": 1.5844217538833618, "learning_rate": 1.1834611718137825e-05, "loss": 1.0032, "step": 2380 }, { "epoch": 0.45823710546574287, "grad_norm": 1.6160128116607666, "learning_rate": 1.1828483841751597e-05, "loss": 0.9327, "step": 2381 }, { "epoch": 0.45842956120092376, "grad_norm": 1.84294593334198, "learning_rate": 1.1822355254921478e-05, "loss": 0.8838, "step": 2382 }, { "epoch": 0.4586220169361047, "grad_norm": 0.9895358681678772, "learning_rate": 1.181622596002868e-05, "loss": 1.0073, "step": 2383 }, { "epoch": 0.4588144726712856, "grad_norm": 1.940767765045166, "learning_rate": 1.18100959594547e-05, "loss": 0.9505, "step": 2384 }, { "epoch": 0.4590069284064665, "grad_norm": 0.9028760194778442, "learning_rate": 1.1803965255581304e-05, "loss": 0.9467, "step": 2385 }, { "epoch": 0.4591993841416474, "grad_norm": 1.7875146865844727, "learning_rate": 1.1797833850790527e-05, "loss": 0.9557, "step": 2386 }, { "epoch": 0.45939183987682836, "grad_norm": 1.866036295890808, "learning_rate": 1.179170174746469e-05, "loss": 1.0082, "step": 2387 }, { "epoch": 0.45958429561200925, "grad_norm": 1.9719218015670776, "learning_rate": 1.1785568947986368e-05, "loss": 0.9837, "step": 2388 }, { "epoch": 0.45977675134719015, "grad_norm": 1.6831638813018799, "learning_rate": 1.177943545473842e-05, "loss": 0.9385, "step": 2389 }, { "epoch": 0.45996920708237105, "grad_norm": 1.7502880096435547, "learning_rate": 1.1773301270103968e-05, "loss": 0.872, "step": 2390 }, { "epoch": 0.46016166281755194, "grad_norm": 1.8075807094573975, "learning_rate": 1.1767166396466404e-05, "loss": 0.9366, "step": 2391 }, { "epoch": 0.4603541185527329, "grad_norm": 1.7219760417938232, "learning_rate": 1.1761030836209384e-05, "loss": 0.9611, "step": 2392 }, { "epoch": 0.4605465742879138, "grad_norm": 1.8846087455749512, "learning_rate": 1.1754894591716841e-05, "loss": 0.9519, "step": 2393 }, { "epoch": 0.4607390300230947, "grad_norm": 2.3830111026763916, "learning_rate": 1.1748757665372965e-05, "loss": 0.9228, "step": 2394 }, { "epoch": 0.4609314857582756, "grad_norm": 1.6864556074142456, "learning_rate": 1.1742620059562211e-05, "loss": 1.0276, "step": 2395 }, { "epoch": 0.4611239414934565, "grad_norm": 1.7836157083511353, "learning_rate": 1.1736481776669307e-05, "loss": 1.0328, "step": 2396 }, { "epoch": 0.46131639722863743, "grad_norm": 1.642415165901184, "learning_rate": 1.1730342819079229e-05, "loss": 0.9401, "step": 2397 }, { "epoch": 0.46150885296381833, "grad_norm": 1.7203096151351929, "learning_rate": 1.1724203189177228e-05, "loss": 0.9886, "step": 2398 }, { "epoch": 0.4617013086989992, "grad_norm": 1.6408607959747314, "learning_rate": 1.1718062889348816e-05, "loss": 0.9464, "step": 2399 }, { "epoch": 0.4618937644341801, "grad_norm": 1.6255512237548828, "learning_rate": 1.1711921921979754e-05, "loss": 0.9607, "step": 2400 }, { "epoch": 0.4620862201693611, "grad_norm": 1.5947513580322266, "learning_rate": 1.1705780289456069e-05, "loss": 0.9421, "step": 2401 }, { "epoch": 0.46227867590454197, "grad_norm": 2.0536837577819824, "learning_rate": 1.1699637994164053e-05, "loss": 0.9297, "step": 2402 }, { "epoch": 0.46247113163972287, "grad_norm": 1.5925076007843018, "learning_rate": 1.1693495038490247e-05, "loss": 0.9009, "step": 2403 }, { "epoch": 0.46266358737490376, "grad_norm": 1.8813236951828003, "learning_rate": 1.1687351424821448e-05, "loss": 0.9999, "step": 2404 }, { "epoch": 0.46285604311008466, "grad_norm": 1.6630133390426636, "learning_rate": 1.168120715554472e-05, "loss": 0.9192, "step": 2405 }, { "epoch": 0.4630484988452656, "grad_norm": 1.5860397815704346, "learning_rate": 1.1675062233047365e-05, "loss": 0.9969, "step": 2406 }, { "epoch": 0.4632409545804465, "grad_norm": 1.8809325695037842, "learning_rate": 1.166891665971695e-05, "loss": 0.8889, "step": 2407 }, { "epoch": 0.4634334103156274, "grad_norm": 1.6948862075805664, "learning_rate": 1.1662770437941293e-05, "loss": 0.8981, "step": 2408 }, { "epoch": 0.4636258660508083, "grad_norm": 0.941698431968689, "learning_rate": 1.165662357010846e-05, "loss": 0.94, "step": 2409 }, { "epoch": 0.4638183217859892, "grad_norm": 1.7244641780853271, "learning_rate": 1.1650476058606776e-05, "loss": 0.9569, "step": 2410 }, { "epoch": 0.46401077752117015, "grad_norm": 1.629375696182251, "learning_rate": 1.1644327905824808e-05, "loss": 0.8802, "step": 2411 }, { "epoch": 0.46420323325635104, "grad_norm": 1.8318703174591064, "learning_rate": 1.1638179114151378e-05, "loss": 1.0359, "step": 2412 }, { "epoch": 0.46439568899153194, "grad_norm": 1.7172409296035767, "learning_rate": 1.163202968597555e-05, "loss": 0.9806, "step": 2413 }, { "epoch": 0.46458814472671284, "grad_norm": 1.8000929355621338, "learning_rate": 1.162587962368664e-05, "loss": 0.9598, "step": 2414 }, { "epoch": 0.4647806004618938, "grad_norm": 1.8672746419906616, "learning_rate": 1.161972892967421e-05, "loss": 0.9546, "step": 2415 }, { "epoch": 0.4649730561970747, "grad_norm": 1.6637725830078125, "learning_rate": 1.1613577606328068e-05, "loss": 0.8799, "step": 2416 }, { "epoch": 0.4651655119322556, "grad_norm": 1.6594940423965454, "learning_rate": 1.1607425656038263e-05, "loss": 0.9419, "step": 2417 }, { "epoch": 0.4653579676674365, "grad_norm": 1.7142961025238037, "learning_rate": 1.1601273081195087e-05, "loss": 0.9015, "step": 2418 }, { "epoch": 0.4655504234026174, "grad_norm": 1.6121093034744263, "learning_rate": 1.1595119884189085e-05, "loss": 0.8915, "step": 2419 }, { "epoch": 0.4657428791377983, "grad_norm": 1.8766065835952759, "learning_rate": 1.1588966067411033e-05, "loss": 1.0055, "step": 2420 }, { "epoch": 0.4659353348729792, "grad_norm": 1.8419183492660522, "learning_rate": 1.1582811633251949e-05, "loss": 0.9209, "step": 2421 }, { "epoch": 0.4661277906081601, "grad_norm": 1.7008318901062012, "learning_rate": 1.1576656584103097e-05, "loss": 0.9703, "step": 2422 }, { "epoch": 0.466320246343341, "grad_norm": 2.009002447128296, "learning_rate": 1.1570500922355971e-05, "loss": 0.9687, "step": 2423 }, { "epoch": 0.4665127020785219, "grad_norm": 1.8037961721420288, "learning_rate": 1.156434465040231e-05, "loss": 0.9622, "step": 2424 }, { "epoch": 0.46670515781370286, "grad_norm": 1.7638972997665405, "learning_rate": 1.1558187770634089e-05, "loss": 0.8816, "step": 2425 }, { "epoch": 0.46689761354888376, "grad_norm": 1.6208146810531616, "learning_rate": 1.1552030285443516e-05, "loss": 0.9517, "step": 2426 }, { "epoch": 0.46709006928406466, "grad_norm": 1.6956517696380615, "learning_rate": 1.1545872197223039e-05, "loss": 1.0613, "step": 2427 }, { "epoch": 0.46728252501924555, "grad_norm": 1.984990119934082, "learning_rate": 1.1539713508365336e-05, "loss": 0.9888, "step": 2428 }, { "epoch": 0.4674749807544265, "grad_norm": 1.6373353004455566, "learning_rate": 1.153355422126332e-05, "loss": 0.8758, "step": 2429 }, { "epoch": 0.4676674364896074, "grad_norm": 0.8930943608283997, "learning_rate": 1.152739433831014e-05, "loss": 0.9686, "step": 2430 }, { "epoch": 0.4678598922247883, "grad_norm": 1.7336249351501465, "learning_rate": 1.1521233861899168e-05, "loss": 1.0355, "step": 2431 }, { "epoch": 0.4680523479599692, "grad_norm": 1.6544311046600342, "learning_rate": 1.1515072794424013e-05, "loss": 0.9927, "step": 2432 }, { "epoch": 0.4682448036951501, "grad_norm": 1.8546185493469238, "learning_rate": 1.1508911138278516e-05, "loss": 0.9306, "step": 2433 }, { "epoch": 0.46843725943033104, "grad_norm": 1.6320329904556274, "learning_rate": 1.1502748895856737e-05, "loss": 0.9375, "step": 2434 }, { "epoch": 0.46862971516551194, "grad_norm": 1.862722396850586, "learning_rate": 1.1496586069552973e-05, "loss": 0.9496, "step": 2435 }, { "epoch": 0.46882217090069284, "grad_norm": 1.9692715406417847, "learning_rate": 1.1490422661761744e-05, "loss": 1.0032, "step": 2436 }, { "epoch": 0.46901462663587373, "grad_norm": 1.5954970121383667, "learning_rate": 1.14842586748778e-05, "loss": 0.8438, "step": 2437 }, { "epoch": 0.4692070823710547, "grad_norm": 1.867908000946045, "learning_rate": 1.1478094111296109e-05, "loss": 1.0152, "step": 2438 }, { "epoch": 0.4693995381062356, "grad_norm": 1.9273887872695923, "learning_rate": 1.1471928973411865e-05, "loss": 0.8954, "step": 2439 }, { "epoch": 0.4695919938414165, "grad_norm": 1.7977123260498047, "learning_rate": 1.146576326362049e-05, "loss": 1.0128, "step": 2440 }, { "epoch": 0.4697844495765974, "grad_norm": 1.7650774717330933, "learning_rate": 1.1459596984317622e-05, "loss": 0.9104, "step": 2441 }, { "epoch": 0.46997690531177827, "grad_norm": 1.6352444887161255, "learning_rate": 1.1453430137899129e-05, "loss": 0.9808, "step": 2442 }, { "epoch": 0.4701693610469592, "grad_norm": 1.879262924194336, "learning_rate": 1.1447262726761088e-05, "loss": 0.9046, "step": 2443 }, { "epoch": 0.4703618167821401, "grad_norm": 1.5640743970870972, "learning_rate": 1.1441094753299802e-05, "loss": 0.8888, "step": 2444 }, { "epoch": 0.470554272517321, "grad_norm": 1.7934092283248901, "learning_rate": 1.1434926219911792e-05, "loss": 0.9478, "step": 2445 }, { "epoch": 0.4707467282525019, "grad_norm": 1.5957752466201782, "learning_rate": 1.1428757128993801e-05, "loss": 0.9469, "step": 2446 }, { "epoch": 0.4709391839876828, "grad_norm": 1.8419419527053833, "learning_rate": 1.1422587482942778e-05, "loss": 1.0022, "step": 2447 }, { "epoch": 0.47113163972286376, "grad_norm": 1.831189751625061, "learning_rate": 1.1416417284155892e-05, "loss": 0.974, "step": 2448 }, { "epoch": 0.47132409545804466, "grad_norm": 1.8670446872711182, "learning_rate": 1.1410246535030533e-05, "loss": 0.875, "step": 2449 }, { "epoch": 0.47151655119322555, "grad_norm": 1.7855297327041626, "learning_rate": 1.1404075237964296e-05, "loss": 0.9597, "step": 2450 }, { "epoch": 0.47170900692840645, "grad_norm": 1.0037751197814941, "learning_rate": 1.1397903395354996e-05, "loss": 0.9494, "step": 2451 }, { "epoch": 0.4719014626635874, "grad_norm": 1.7400498390197754, "learning_rate": 1.1391731009600655e-05, "loss": 0.942, "step": 2452 }, { "epoch": 0.4720939183987683, "grad_norm": 1.901992678642273, "learning_rate": 1.138555808309951e-05, "loss": 1.0546, "step": 2453 }, { "epoch": 0.4722863741339492, "grad_norm": 1.8271876573562622, "learning_rate": 1.1379384618250002e-05, "loss": 0.9969, "step": 2454 }, { "epoch": 0.4724788298691301, "grad_norm": 1.7119807004928589, "learning_rate": 1.1373210617450793e-05, "loss": 0.9979, "step": 2455 }, { "epoch": 0.472671285604311, "grad_norm": 1.7102749347686768, "learning_rate": 1.1367036083100735e-05, "loss": 0.8864, "step": 2456 }, { "epoch": 0.47286374133949194, "grad_norm": 1.7240320444107056, "learning_rate": 1.1360861017598905e-05, "loss": 0.9969, "step": 2457 }, { "epoch": 0.47305619707467284, "grad_norm": 1.7498681545257568, "learning_rate": 1.1354685423344577e-05, "loss": 1.0206, "step": 2458 }, { "epoch": 0.47324865280985373, "grad_norm": 1.7679030895233154, "learning_rate": 1.1348509302737232e-05, "loss": 0.8698, "step": 2459 }, { "epoch": 0.47344110854503463, "grad_norm": 1.6931575536727905, "learning_rate": 1.1342332658176556e-05, "loss": 0.9016, "step": 2460 }, { "epoch": 0.4736335642802155, "grad_norm": 1.5806704759597778, "learning_rate": 1.1336155492062439e-05, "loss": 0.9062, "step": 2461 }, { "epoch": 0.4738260200153965, "grad_norm": 1.7470468282699585, "learning_rate": 1.1329977806794976e-05, "loss": 0.9844, "step": 2462 }, { "epoch": 0.4740184757505774, "grad_norm": 1.8262832164764404, "learning_rate": 1.1323799604774457e-05, "loss": 1.0072, "step": 2463 }, { "epoch": 0.47421093148575827, "grad_norm": 1.7990171909332275, "learning_rate": 1.1317620888401379e-05, "loss": 1.0206, "step": 2464 }, { "epoch": 0.47440338722093917, "grad_norm": 1.7142932415008545, "learning_rate": 1.1311441660076433e-05, "loss": 0.9872, "step": 2465 }, { "epoch": 0.4745958429561201, "grad_norm": 1.7087639570236206, "learning_rate": 1.130526192220052e-05, "loss": 0.9421, "step": 2466 }, { "epoch": 0.474788298691301, "grad_norm": 1.6382148265838623, "learning_rate": 1.1299081677174721e-05, "loss": 1.0363, "step": 2467 }, { "epoch": 0.4749807544264819, "grad_norm": 1.6526854038238525, "learning_rate": 1.1292900927400334e-05, "loss": 0.8932, "step": 2468 }, { "epoch": 0.4751732101616628, "grad_norm": 1.8506120443344116, "learning_rate": 1.1286719675278838e-05, "loss": 1.0086, "step": 2469 }, { "epoch": 0.4753656658968437, "grad_norm": 1.7848901748657227, "learning_rate": 1.1280537923211919e-05, "loss": 0.9524, "step": 2470 }, { "epoch": 0.47555812163202466, "grad_norm": 1.7623332738876343, "learning_rate": 1.1274355673601446e-05, "loss": 1.0374, "step": 2471 }, { "epoch": 0.47575057736720555, "grad_norm": 0.9106870889663696, "learning_rate": 1.1268172928849486e-05, "loss": 0.8747, "step": 2472 }, { "epoch": 0.47594303310238645, "grad_norm": 1.6627006530761719, "learning_rate": 1.12619896913583e-05, "loss": 0.9727, "step": 2473 }, { "epoch": 0.47613548883756734, "grad_norm": 1.6785444021224976, "learning_rate": 1.1255805963530339e-05, "loss": 1.0028, "step": 2474 }, { "epoch": 0.47632794457274824, "grad_norm": 1.7082171440124512, "learning_rate": 1.1249621747768247e-05, "loss": 0.9732, "step": 2475 }, { "epoch": 0.4765204003079292, "grad_norm": 1.7627836465835571, "learning_rate": 1.1243437046474854e-05, "loss": 0.8958, "step": 2476 }, { "epoch": 0.4767128560431101, "grad_norm": 1.796220302581787, "learning_rate": 1.1237251862053178e-05, "loss": 1.0092, "step": 2477 }, { "epoch": 0.476905311778291, "grad_norm": 1.8795274496078491, "learning_rate": 1.123106619690643e-05, "loss": 1.0085, "step": 2478 }, { "epoch": 0.4770977675134719, "grad_norm": 1.8828896284103394, "learning_rate": 1.1224880053438006e-05, "loss": 0.9293, "step": 2479 }, { "epoch": 0.47729022324865283, "grad_norm": 1.6245577335357666, "learning_rate": 1.1218693434051475e-05, "loss": 0.9643, "step": 2480 }, { "epoch": 0.47748267898383373, "grad_norm": 1.7668883800506592, "learning_rate": 1.1212506341150615e-05, "loss": 0.9524, "step": 2481 }, { "epoch": 0.4776751347190146, "grad_norm": 1.658494472503662, "learning_rate": 1.120631877713937e-05, "loss": 0.9959, "step": 2482 }, { "epoch": 0.4778675904541955, "grad_norm": 2.491473436355591, "learning_rate": 1.1200130744421877e-05, "loss": 0.9575, "step": 2483 }, { "epoch": 0.4780600461893764, "grad_norm": 1.6411579847335815, "learning_rate": 1.1193942245402443e-05, "loss": 0.9526, "step": 2484 }, { "epoch": 0.47825250192455737, "grad_norm": 2.039761781692505, "learning_rate": 1.1187753282485569e-05, "loss": 0.9607, "step": 2485 }, { "epoch": 0.47844495765973827, "grad_norm": 1.661043643951416, "learning_rate": 1.118156385807593e-05, "loss": 0.9796, "step": 2486 }, { "epoch": 0.47863741339491916, "grad_norm": 2.0359725952148438, "learning_rate": 1.1175373974578378e-05, "loss": 0.9389, "step": 2487 }, { "epoch": 0.47882986913010006, "grad_norm": 0.9338507056236267, "learning_rate": 1.1169183634397948e-05, "loss": 0.9227, "step": 2488 }, { "epoch": 0.479022324865281, "grad_norm": 1.6777851581573486, "learning_rate": 1.1162992839939852e-05, "loss": 1.0145, "step": 2489 }, { "epoch": 0.4792147806004619, "grad_norm": 1.664560317993164, "learning_rate": 1.1156801593609479e-05, "loss": 0.8947, "step": 2490 }, { "epoch": 0.4794072363356428, "grad_norm": 1.6090173721313477, "learning_rate": 1.1150609897812387e-05, "loss": 0.9772, "step": 2491 }, { "epoch": 0.4795996920708237, "grad_norm": 1.6112374067306519, "learning_rate": 1.114441775495432e-05, "loss": 1.0024, "step": 2492 }, { "epoch": 0.4797921478060046, "grad_norm": 1.928670048713684, "learning_rate": 1.1138225167441186e-05, "loss": 0.9418, "step": 2493 }, { "epoch": 0.47998460354118555, "grad_norm": 0.896676778793335, "learning_rate": 1.113203213767907e-05, "loss": 0.8839, "step": 2494 }, { "epoch": 0.48017705927636645, "grad_norm": 1.5596702098846436, "learning_rate": 1.1125838668074225e-05, "loss": 0.9194, "step": 2495 }, { "epoch": 0.48036951501154734, "grad_norm": 0.9425876140594482, "learning_rate": 1.1119644761033079e-05, "loss": 0.9225, "step": 2496 }, { "epoch": 0.48056197074672824, "grad_norm": 1.7130099534988403, "learning_rate": 1.1113450418962233e-05, "loss": 0.9941, "step": 2497 }, { "epoch": 0.48075442648190914, "grad_norm": 1.6306774616241455, "learning_rate": 1.1107255644268448e-05, "loss": 0.9542, "step": 2498 }, { "epoch": 0.4809468822170901, "grad_norm": 1.652103304862976, "learning_rate": 1.110106043935866e-05, "loss": 0.9879, "step": 2499 }, { "epoch": 0.481139337952271, "grad_norm": 1.6207376718521118, "learning_rate": 1.1094864806639971e-05, "loss": 0.991, "step": 2500 }, { "epoch": 0.4813317936874519, "grad_norm": 1.865201711654663, "learning_rate": 1.1088668748519646e-05, "loss": 0.9218, "step": 2501 }, { "epoch": 0.4815242494226328, "grad_norm": 1.7280155420303345, "learning_rate": 1.1082472267405123e-05, "loss": 0.9668, "step": 2502 }, { "epoch": 0.48171670515781373, "grad_norm": 1.7467422485351562, "learning_rate": 1.107627536570399e-05, "loss": 0.9701, "step": 2503 }, { "epoch": 0.4819091608929946, "grad_norm": 1.6940211057662964, "learning_rate": 1.1070078045824014e-05, "loss": 1.032, "step": 2504 }, { "epoch": 0.4821016166281755, "grad_norm": 1.708522081375122, "learning_rate": 1.1063880310173114e-05, "loss": 0.9458, "step": 2505 }, { "epoch": 0.4822940723633564, "grad_norm": 1.8541128635406494, "learning_rate": 1.105768216115938e-05, "loss": 0.9418, "step": 2506 }, { "epoch": 0.4824865280985373, "grad_norm": 1.5646147727966309, "learning_rate": 1.1051483601191051e-05, "loss": 0.9481, "step": 2507 }, { "epoch": 0.48267898383371827, "grad_norm": 1.6746430397033691, "learning_rate": 1.1045284632676535e-05, "loss": 0.8722, "step": 2508 }, { "epoch": 0.48287143956889916, "grad_norm": 1.6527259349822998, "learning_rate": 1.1039085258024396e-05, "loss": 0.9676, "step": 2509 }, { "epoch": 0.48306389530408006, "grad_norm": 1.8868647813796997, "learning_rate": 1.1032885479643351e-05, "loss": 0.9513, "step": 2510 }, { "epoch": 0.48325635103926096, "grad_norm": 2.008626937866211, "learning_rate": 1.1026685299942286e-05, "loss": 0.9582, "step": 2511 }, { "epoch": 0.48344880677444185, "grad_norm": 1.9393815994262695, "learning_rate": 1.1020484721330227e-05, "loss": 0.8996, "step": 2512 }, { "epoch": 0.4836412625096228, "grad_norm": 1.93147611618042, "learning_rate": 1.1014283746216366e-05, "loss": 1.0689, "step": 2513 }, { "epoch": 0.4838337182448037, "grad_norm": 1.9556524753570557, "learning_rate": 1.1008082377010045e-05, "loss": 0.9096, "step": 2514 }, { "epoch": 0.4840261739799846, "grad_norm": 1.766107439994812, "learning_rate": 1.1001880616120764e-05, "loss": 0.9915, "step": 2515 }, { "epoch": 0.4842186297151655, "grad_norm": 1.849605679512024, "learning_rate": 1.0995678465958168e-05, "loss": 0.88, "step": 2516 }, { "epoch": 0.48441108545034645, "grad_norm": 1.822379231452942, "learning_rate": 1.0989475928932057e-05, "loss": 0.8604, "step": 2517 }, { "epoch": 0.48460354118552734, "grad_norm": 1.7800196409225464, "learning_rate": 1.098327300745238e-05, "loss": 0.9783, "step": 2518 }, { "epoch": 0.48479599692070824, "grad_norm": 1.900404453277588, "learning_rate": 1.0977069703929242e-05, "loss": 0.9996, "step": 2519 }, { "epoch": 0.48498845265588914, "grad_norm": 1.8659120798110962, "learning_rate": 1.0970866020772884e-05, "loss": 0.9941, "step": 2520 }, { "epoch": 0.48518090839107003, "grad_norm": 1.7063900232315063, "learning_rate": 1.0964661960393703e-05, "loss": 1.017, "step": 2521 }, { "epoch": 0.485373364126251, "grad_norm": 1.687215805053711, "learning_rate": 1.0958457525202241e-05, "loss": 0.9649, "step": 2522 }, { "epoch": 0.4855658198614319, "grad_norm": 1.5848464965820312, "learning_rate": 1.0952252717609188e-05, "loss": 0.9759, "step": 2523 }, { "epoch": 0.4857582755966128, "grad_norm": 1.683504581451416, "learning_rate": 1.0946047540025373e-05, "loss": 0.9353, "step": 2524 }, { "epoch": 0.4859507313317937, "grad_norm": 1.7304487228393555, "learning_rate": 1.0939841994861773e-05, "loss": 0.927, "step": 2525 }, { "epoch": 0.48614318706697457, "grad_norm": 1.6588670015335083, "learning_rate": 1.0933636084529507e-05, "loss": 0.9761, "step": 2526 }, { "epoch": 0.4863356428021555, "grad_norm": 1.704005241394043, "learning_rate": 1.0927429811439838e-05, "loss": 0.9885, "step": 2527 }, { "epoch": 0.4865280985373364, "grad_norm": 1.631637454032898, "learning_rate": 1.0921223178004163e-05, "loss": 0.9777, "step": 2528 }, { "epoch": 0.4867205542725173, "grad_norm": 1.7077293395996094, "learning_rate": 1.0915016186634027e-05, "loss": 0.9251, "step": 2529 }, { "epoch": 0.4869130100076982, "grad_norm": 1.7995444536209106, "learning_rate": 1.0908808839741108e-05, "loss": 0.9527, "step": 2530 }, { "epoch": 0.48710546574287916, "grad_norm": 1.7117691040039062, "learning_rate": 1.0902601139737225e-05, "loss": 1.0516, "step": 2531 }, { "epoch": 0.48729792147806006, "grad_norm": 1.7033149003982544, "learning_rate": 1.0896393089034336e-05, "loss": 0.955, "step": 2532 }, { "epoch": 0.48749037721324096, "grad_norm": 0.9550507068634033, "learning_rate": 1.0890184690044534e-05, "loss": 0.937, "step": 2533 }, { "epoch": 0.48768283294842185, "grad_norm": 1.7512893676757812, "learning_rate": 1.0883975945180043e-05, "loss": 0.9903, "step": 2534 }, { "epoch": 0.48787528868360275, "grad_norm": 1.8685920238494873, "learning_rate": 1.0877766856853227e-05, "loss": 0.9622, "step": 2535 }, { "epoch": 0.4880677444187837, "grad_norm": 0.9323713779449463, "learning_rate": 1.0871557427476585e-05, "loss": 0.9471, "step": 2536 }, { "epoch": 0.4882602001539646, "grad_norm": 1.8061491250991821, "learning_rate": 1.0865347659462737e-05, "loss": 0.9147, "step": 2537 }, { "epoch": 0.4884526558891455, "grad_norm": 1.8753383159637451, "learning_rate": 1.0859137555224448e-05, "loss": 0.909, "step": 2538 }, { "epoch": 0.4886451116243264, "grad_norm": 0.9152477383613586, "learning_rate": 1.0852927117174607e-05, "loss": 0.9597, "step": 2539 }, { "epoch": 0.4888375673595073, "grad_norm": 1.8506150245666504, "learning_rate": 1.0846716347726233e-05, "loss": 0.9629, "step": 2540 }, { "epoch": 0.48903002309468824, "grad_norm": 1.6528825759887695, "learning_rate": 1.0840505249292477e-05, "loss": 0.9337, "step": 2541 }, { "epoch": 0.48922247882986913, "grad_norm": 1.8424479961395264, "learning_rate": 1.0834293824286612e-05, "loss": 0.9495, "step": 2542 }, { "epoch": 0.48941493456505003, "grad_norm": 1.6261227130889893, "learning_rate": 1.0828082075122044e-05, "loss": 0.9194, "step": 2543 }, { "epoch": 0.4896073903002309, "grad_norm": 1.6804587841033936, "learning_rate": 1.0821870004212305e-05, "loss": 0.986, "step": 2544 }, { "epoch": 0.4897998460354119, "grad_norm": 1.8019813299179077, "learning_rate": 1.081565761397104e-05, "loss": 0.9565, "step": 2545 }, { "epoch": 0.4899923017705928, "grad_norm": 1.7652204036712646, "learning_rate": 1.0809444906812034e-05, "loss": 0.9417, "step": 2546 }, { "epoch": 0.49018475750577367, "grad_norm": 1.6180081367492676, "learning_rate": 1.0803231885149191e-05, "loss": 0.9099, "step": 2547 }, { "epoch": 0.49037721324095457, "grad_norm": 1.8602015972137451, "learning_rate": 1.0797018551396527e-05, "loss": 1.0119, "step": 2548 }, { "epoch": 0.49056966897613546, "grad_norm": 0.9111103415489197, "learning_rate": 1.0790804907968194e-05, "loss": 0.8838, "step": 2549 }, { "epoch": 0.4907621247113164, "grad_norm": 1.8813616037368774, "learning_rate": 1.0784590957278452e-05, "loss": 1.0031, "step": 2550 }, { "epoch": 0.4909545804464973, "grad_norm": 1.841774344444275, "learning_rate": 1.0778376701741688e-05, "loss": 0.9657, "step": 2551 }, { "epoch": 0.4911470361816782, "grad_norm": 1.6066312789916992, "learning_rate": 1.0772162143772407e-05, "loss": 0.9944, "step": 2552 }, { "epoch": 0.4913394919168591, "grad_norm": 1.751482367515564, "learning_rate": 1.0765947285785228e-05, "loss": 0.8834, "step": 2553 }, { "epoch": 0.49153194765204006, "grad_norm": 1.7358108758926392, "learning_rate": 1.0759732130194887e-05, "loss": 0.9447, "step": 2554 }, { "epoch": 0.49172440338722095, "grad_norm": 1.8437845706939697, "learning_rate": 1.0753516679416237e-05, "loss": 0.8942, "step": 2555 }, { "epoch": 0.49191685912240185, "grad_norm": 1.959992527961731, "learning_rate": 1.0747300935864245e-05, "loss": 1.0067, "step": 2556 }, { "epoch": 0.49210931485758275, "grad_norm": 1.6747387647628784, "learning_rate": 1.0741084901953995e-05, "loss": 0.9883, "step": 2557 }, { "epoch": 0.49230177059276364, "grad_norm": 1.7554908990859985, "learning_rate": 1.0734868580100677e-05, "loss": 0.9509, "step": 2558 }, { "epoch": 0.4924942263279446, "grad_norm": 1.7112852334976196, "learning_rate": 1.0728651972719603e-05, "loss": 0.943, "step": 2559 }, { "epoch": 0.4926866820631255, "grad_norm": 1.9240753650665283, "learning_rate": 1.0722435082226186e-05, "loss": 1.0332, "step": 2560 }, { "epoch": 0.4928791377983064, "grad_norm": 1.7998155355453491, "learning_rate": 1.0716217911035952e-05, "loss": 1.0034, "step": 2561 }, { "epoch": 0.4930715935334873, "grad_norm": 1.7169820070266724, "learning_rate": 1.0710000461564535e-05, "loss": 0.9743, "step": 2562 }, { "epoch": 0.4932640492686682, "grad_norm": 1.509096384048462, "learning_rate": 1.0703782736227688e-05, "loss": 0.9281, "step": 2563 }, { "epoch": 0.49345650500384913, "grad_norm": 1.655415654182434, "learning_rate": 1.0697564737441254e-05, "loss": 1.0255, "step": 2564 }, { "epoch": 0.49364896073903003, "grad_norm": 1.703011155128479, "learning_rate": 1.0691346467621192e-05, "loss": 0.9842, "step": 2565 }, { "epoch": 0.4938414164742109, "grad_norm": 1.5411932468414307, "learning_rate": 1.0685127929183567e-05, "loss": 0.9208, "step": 2566 }, { "epoch": 0.4940338722093918, "grad_norm": 1.8359014987945557, "learning_rate": 1.0678909124544548e-05, "loss": 0.9513, "step": 2567 }, { "epoch": 0.4942263279445728, "grad_norm": 1.6655604839324951, "learning_rate": 1.0672690056120398e-05, "loss": 0.9182, "step": 2568 }, { "epoch": 0.49441878367975367, "grad_norm": 1.9323374032974243, "learning_rate": 1.0666470726327496e-05, "loss": 1.0974, "step": 2569 }, { "epoch": 0.49461123941493457, "grad_norm": 1.758662462234497, "learning_rate": 1.0660251137582316e-05, "loss": 0.9896, "step": 2570 }, { "epoch": 0.49480369515011546, "grad_norm": 1.606425404548645, "learning_rate": 1.0654031292301432e-05, "loss": 1.0199, "step": 2571 }, { "epoch": 0.49499615088529636, "grad_norm": 1.6762839555740356, "learning_rate": 1.0647811192901518e-05, "loss": 1.0265, "step": 2572 }, { "epoch": 0.4951886066204773, "grad_norm": 1.8858031034469604, "learning_rate": 1.0641590841799352e-05, "loss": 0.9672, "step": 2573 }, { "epoch": 0.4953810623556582, "grad_norm": 1.7840813398361206, "learning_rate": 1.0635370241411803e-05, "loss": 0.9871, "step": 2574 }, { "epoch": 0.4955735180908391, "grad_norm": 1.8249763250350952, "learning_rate": 1.0629149394155837e-05, "loss": 0.9258, "step": 2575 }, { "epoch": 0.49576597382602, "grad_norm": 1.645431637763977, "learning_rate": 1.0622928302448523e-05, "loss": 0.934, "step": 2576 }, { "epoch": 0.4959584295612009, "grad_norm": 1.598764181137085, "learning_rate": 1.0616706968707015e-05, "loss": 0.9166, "step": 2577 }, { "epoch": 0.49615088529638185, "grad_norm": 1.6708288192749023, "learning_rate": 1.0610485395348571e-05, "loss": 0.9516, "step": 2578 }, { "epoch": 0.49634334103156275, "grad_norm": 1.7206602096557617, "learning_rate": 1.0604263584790534e-05, "loss": 0.9886, "step": 2579 }, { "epoch": 0.49653579676674364, "grad_norm": 1.6740700006484985, "learning_rate": 1.0598041539450344e-05, "loss": 0.946, "step": 2580 }, { "epoch": 0.49672825250192454, "grad_norm": 1.7009068727493286, "learning_rate": 1.0591819261745528e-05, "loss": 0.9329, "step": 2581 }, { "epoch": 0.4969207082371055, "grad_norm": 1.7152082920074463, "learning_rate": 1.0585596754093712e-05, "loss": 0.988, "step": 2582 }, { "epoch": 0.4971131639722864, "grad_norm": 0.9035679697990417, "learning_rate": 1.0579374018912602e-05, "loss": 0.9249, "step": 2583 }, { "epoch": 0.4973056197074673, "grad_norm": 1.7057833671569824, "learning_rate": 1.0573151058619994e-05, "loss": 1.0094, "step": 2584 }, { "epoch": 0.4974980754426482, "grad_norm": 1.805561900138855, "learning_rate": 1.0566927875633776e-05, "loss": 0.9136, "step": 2585 }, { "epoch": 0.4976905311778291, "grad_norm": 1.8780978918075562, "learning_rate": 1.0560704472371919e-05, "loss": 0.9339, "step": 2586 }, { "epoch": 0.49788298691301003, "grad_norm": 0.8711561560630798, "learning_rate": 1.055448085125248e-05, "loss": 0.885, "step": 2587 }, { "epoch": 0.4980754426481909, "grad_norm": 1.7667337656021118, "learning_rate": 1.0548257014693602e-05, "loss": 0.9555, "step": 2588 }, { "epoch": 0.4982678983833718, "grad_norm": 1.763978362083435, "learning_rate": 1.0542032965113508e-05, "loss": 0.9531, "step": 2589 }, { "epoch": 0.4984603541185527, "grad_norm": 1.6618916988372803, "learning_rate": 1.053580870493051e-05, "loss": 0.9447, "step": 2590 }, { "epoch": 0.4986528098537336, "grad_norm": 1.6307361125946045, "learning_rate": 1.0529584236562995e-05, "loss": 0.9878, "step": 2591 }, { "epoch": 0.49884526558891457, "grad_norm": 1.8901039361953735, "learning_rate": 1.0523359562429441e-05, "loss": 0.9879, "step": 2592 }, { "epoch": 0.49903772132409546, "grad_norm": 1.743004322052002, "learning_rate": 1.051713468494839e-05, "loss": 0.9183, "step": 2593 }, { "epoch": 0.49923017705927636, "grad_norm": 1.8654699325561523, "learning_rate": 1.0510909606538478e-05, "loss": 0.9953, "step": 2594 }, { "epoch": 0.49942263279445726, "grad_norm": 1.67255437374115, "learning_rate": 1.0504684329618408e-05, "loss": 0.9342, "step": 2595 }, { "epoch": 0.4996150885296382, "grad_norm": 1.6896085739135742, "learning_rate": 1.0498458856606972e-05, "loss": 0.948, "step": 2596 }, { "epoch": 0.4998075442648191, "grad_norm": 1.8258219957351685, "learning_rate": 1.049223318992303e-05, "loss": 1.0103, "step": 2597 }, { "epoch": 0.5, "grad_norm": 1.6927515268325806, "learning_rate": 1.0486007331985516e-05, "loss": 1.0208, "step": 2598 }, { "epoch": 0.5001924557351809, "grad_norm": 1.664414882659912, "learning_rate": 1.047978128521344e-05, "loss": 1.0, "step": 2599 }, { "epoch": 0.5003849114703618, "grad_norm": 1.6496074199676514, "learning_rate": 1.0473555052025893e-05, "loss": 0.9407, "step": 2600 }, { "epoch": 0.5005773672055427, "grad_norm": 2.102339267730713, "learning_rate": 1.0467328634842024e-05, "loss": 0.9893, "step": 2601 }, { "epoch": 0.5007698229407236, "grad_norm": 1.9392521381378174, "learning_rate": 1.0461102036081069e-05, "loss": 0.9549, "step": 2602 }, { "epoch": 0.5009622786759046, "grad_norm": 1.6156729459762573, "learning_rate": 1.045487525816232e-05, "loss": 0.8917, "step": 2603 }, { "epoch": 0.5011547344110855, "grad_norm": 2.0820255279541016, "learning_rate": 1.044864830350515e-05, "loss": 0.9958, "step": 2604 }, { "epoch": 0.5013471901462664, "grad_norm": 1.6950922012329102, "learning_rate": 1.0442421174528997e-05, "loss": 1.0024, "step": 2605 }, { "epoch": 0.5015396458814473, "grad_norm": 1.6886099576950073, "learning_rate": 1.0436193873653362e-05, "loss": 0.9222, "step": 2606 }, { "epoch": 0.5017321016166282, "grad_norm": 0.8882932066917419, "learning_rate": 1.0429966403297818e-05, "loss": 0.9341, "step": 2607 }, { "epoch": 0.5019245573518091, "grad_norm": 1.794498324394226, "learning_rate": 1.0423738765882006e-05, "loss": 0.9314, "step": 2608 }, { "epoch": 0.50211701308699, "grad_norm": 1.8079392910003662, "learning_rate": 1.0417510963825625e-05, "loss": 0.8901, "step": 2609 }, { "epoch": 0.5023094688221709, "grad_norm": 1.793648362159729, "learning_rate": 1.0411282999548442e-05, "loss": 1.0387, "step": 2610 }, { "epoch": 0.5025019245573518, "grad_norm": 1.901085376739502, "learning_rate": 1.0405054875470287e-05, "loss": 0.9984, "step": 2611 }, { "epoch": 0.5026943802925328, "grad_norm": 1.7960476875305176, "learning_rate": 1.039882659401105e-05, "loss": 0.9759, "step": 2612 }, { "epoch": 0.5028868360277137, "grad_norm": 1.7626402378082275, "learning_rate": 1.0392598157590687e-05, "loss": 0.9723, "step": 2613 }, { "epoch": 0.5030792917628946, "grad_norm": 1.8104925155639648, "learning_rate": 1.0386369568629209e-05, "loss": 0.961, "step": 2614 }, { "epoch": 0.5032717474980755, "grad_norm": 1.946303367614746, "learning_rate": 1.038014082954669e-05, "loss": 0.9673, "step": 2615 }, { "epoch": 0.5034642032332564, "grad_norm": 1.8023290634155273, "learning_rate": 1.037391194276326e-05, "loss": 0.9577, "step": 2616 }, { "epoch": 0.5036566589684373, "grad_norm": 1.9239580631256104, "learning_rate": 1.0367682910699104e-05, "loss": 1.0539, "step": 2617 }, { "epoch": 0.5038491147036182, "grad_norm": 1.7491058111190796, "learning_rate": 1.0361453735774467e-05, "loss": 1.004, "step": 2618 }, { "epoch": 0.504041570438799, "grad_norm": 1.830714464187622, "learning_rate": 1.0355224420409653e-05, "loss": 0.952, "step": 2619 }, { "epoch": 0.5042340261739799, "grad_norm": 1.6921135187149048, "learning_rate": 1.0348994967025012e-05, "loss": 0.9931, "step": 2620 }, { "epoch": 0.5044264819091608, "grad_norm": 0.9464678764343262, "learning_rate": 1.0342765378040953e-05, "loss": 0.9309, "step": 2621 }, { "epoch": 0.5046189376443418, "grad_norm": 1.9240251779556274, "learning_rate": 1.0336535655877942e-05, "loss": 0.8995, "step": 2622 }, { "epoch": 0.5048113933795227, "grad_norm": 1.694456696510315, "learning_rate": 1.0330305802956484e-05, "loss": 1.0024, "step": 2623 }, { "epoch": 0.5050038491147036, "grad_norm": 1.7124693393707275, "learning_rate": 1.0324075821697146e-05, "loss": 0.9695, "step": 2624 }, { "epoch": 0.5051963048498845, "grad_norm": 1.732235074043274, "learning_rate": 1.0317845714520543e-05, "loss": 0.9637, "step": 2625 }, { "epoch": 0.5053887605850654, "grad_norm": 1.9292073249816895, "learning_rate": 1.0311615483847333e-05, "loss": 0.9623, "step": 2626 }, { "epoch": 0.5055812163202463, "grad_norm": 1.5642004013061523, "learning_rate": 1.0305385132098229e-05, "loss": 0.9776, "step": 2627 }, { "epoch": 0.5057736720554272, "grad_norm": 1.7604013681411743, "learning_rate": 1.0299154661693987e-05, "loss": 0.8965, "step": 2628 }, { "epoch": 0.5059661277906081, "grad_norm": 1.7320574522018433, "learning_rate": 1.029292407505541e-05, "loss": 0.9914, "step": 2629 }, { "epoch": 0.506158583525789, "grad_norm": 1.987414002418518, "learning_rate": 1.0286693374603347e-05, "loss": 0.9509, "step": 2630 }, { "epoch": 0.50635103926097, "grad_norm": 0.8902105689048767, "learning_rate": 1.028046256275869e-05, "loss": 0.9436, "step": 2631 }, { "epoch": 0.5065434949961509, "grad_norm": 1.7762867212295532, "learning_rate": 1.0274231641942378e-05, "loss": 0.9511, "step": 2632 }, { "epoch": 0.5067359507313318, "grad_norm": 1.5204285383224487, "learning_rate": 1.0268000614575386e-05, "loss": 0.9117, "step": 2633 }, { "epoch": 0.5069284064665127, "grad_norm": 1.7982538938522339, "learning_rate": 1.0261769483078734e-05, "loss": 0.9165, "step": 2634 }, { "epoch": 0.5071208622016936, "grad_norm": 1.6730221509933472, "learning_rate": 1.0255538249873482e-05, "loss": 1.0162, "step": 2635 }, { "epoch": 0.5073133179368745, "grad_norm": 1.7033743858337402, "learning_rate": 1.0249306917380731e-05, "loss": 0.9731, "step": 2636 }, { "epoch": 0.5075057736720554, "grad_norm": 1.8288662433624268, "learning_rate": 1.0243075488021617e-05, "loss": 0.9879, "step": 2637 }, { "epoch": 0.5076982294072363, "grad_norm": 1.7370675802230835, "learning_rate": 1.0236843964217317e-05, "loss": 0.9248, "step": 2638 }, { "epoch": 0.5078906851424172, "grad_norm": 1.8735297918319702, "learning_rate": 1.0230612348389048e-05, "loss": 0.9764, "step": 2639 }, { "epoch": 0.5080831408775982, "grad_norm": 1.6338409185409546, "learning_rate": 1.0224380642958052e-05, "loss": 0.9317, "step": 2640 }, { "epoch": 0.5082755966127791, "grad_norm": 1.579725980758667, "learning_rate": 1.0218148850345613e-05, "loss": 0.9465, "step": 2641 }, { "epoch": 0.50846805234796, "grad_norm": 1.740787386894226, "learning_rate": 1.0211916972973048e-05, "loss": 0.9425, "step": 2642 }, { "epoch": 0.5086605080831409, "grad_norm": 1.7199437618255615, "learning_rate": 1.0205685013261712e-05, "loss": 1.0563, "step": 2643 }, { "epoch": 0.5088529638183218, "grad_norm": 1.6943423748016357, "learning_rate": 1.0199452973632982e-05, "loss": 0.9705, "step": 2644 }, { "epoch": 0.5090454195535027, "grad_norm": 1.7635377645492554, "learning_rate": 1.019322085650827e-05, "loss": 0.9519, "step": 2645 }, { "epoch": 0.5092378752886836, "grad_norm": 1.700238823890686, "learning_rate": 1.0186988664309023e-05, "loss": 0.967, "step": 2646 }, { "epoch": 0.5094303310238645, "grad_norm": 1.5561318397521973, "learning_rate": 1.018075639945671e-05, "loss": 0.9499, "step": 2647 }, { "epoch": 0.5096227867590454, "grad_norm": 1.8046441078186035, "learning_rate": 1.0174524064372837e-05, "loss": 0.9743, "step": 2648 }, { "epoch": 0.5098152424942263, "grad_norm": 1.6736178398132324, "learning_rate": 1.0168291661478924e-05, "loss": 0.9129, "step": 2649 }, { "epoch": 0.5100076982294073, "grad_norm": 1.6682915687561035, "learning_rate": 1.016205919319653e-05, "loss": 1.0132, "step": 2650 }, { "epoch": 0.5102001539645882, "grad_norm": 1.7404675483703613, "learning_rate": 1.0155826661947232e-05, "loss": 0.9104, "step": 2651 }, { "epoch": 0.5103926096997691, "grad_norm": 2.396991014480591, "learning_rate": 1.0149594070152638e-05, "loss": 0.9425, "step": 2652 }, { "epoch": 0.51058506543495, "grad_norm": 1.7358639240264893, "learning_rate": 1.014336142023437e-05, "loss": 0.9326, "step": 2653 }, { "epoch": 0.5107775211701309, "grad_norm": 1.795857548713684, "learning_rate": 1.0137128714614082e-05, "loss": 0.947, "step": 2654 }, { "epoch": 0.5109699769053118, "grad_norm": 1.9969037771224976, "learning_rate": 1.0130895955713445e-05, "loss": 0.8753, "step": 2655 }, { "epoch": 0.5111624326404927, "grad_norm": 1.7478946447372437, "learning_rate": 1.0124663145954152e-05, "loss": 1.0194, "step": 2656 }, { "epoch": 0.5113548883756736, "grad_norm": 1.7508280277252197, "learning_rate": 1.0118430287757914e-05, "loss": 0.8897, "step": 2657 }, { "epoch": 0.5115473441108545, "grad_norm": 1.9347432851791382, "learning_rate": 1.011219738354646e-05, "loss": 1.0745, "step": 2658 }, { "epoch": 0.5117397998460355, "grad_norm": 1.668831706047058, "learning_rate": 1.0105964435741544e-05, "loss": 0.9758, "step": 2659 }, { "epoch": 0.5119322555812164, "grad_norm": 1.7541227340698242, "learning_rate": 1.0099731446764927e-05, "loss": 1.0125, "step": 2660 }, { "epoch": 0.5121247113163973, "grad_norm": 1.757276177406311, "learning_rate": 1.0093498419038394e-05, "loss": 0.8429, "step": 2661 }, { "epoch": 0.5123171670515781, "grad_norm": 1.865217685699463, "learning_rate": 1.008726535498374e-05, "loss": 0.9857, "step": 2662 }, { "epoch": 0.512509622786759, "grad_norm": 1.6601051092147827, "learning_rate": 1.0081032257022776e-05, "loss": 0.9632, "step": 2663 }, { "epoch": 0.5127020785219399, "grad_norm": 1.6460882425308228, "learning_rate": 1.007479912757733e-05, "loss": 0.9729, "step": 2664 }, { "epoch": 0.5128945342571208, "grad_norm": 1.9257735013961792, "learning_rate": 1.0068565969069236e-05, "loss": 0.9101, "step": 2665 }, { "epoch": 0.5130869899923017, "grad_norm": 1.5772889852523804, "learning_rate": 1.0062332783920337e-05, "loss": 0.921, "step": 2666 }, { "epoch": 0.5132794457274826, "grad_norm": 1.9605385065078735, "learning_rate": 1.0056099574552498e-05, "loss": 0.9101, "step": 2667 }, { "epoch": 0.5134719014626636, "grad_norm": 1.599408745765686, "learning_rate": 1.0049866343387582e-05, "loss": 1.0089, "step": 2668 }, { "epoch": 0.5136643571978445, "grad_norm": 1.7058508396148682, "learning_rate": 1.0043633092847468e-05, "loss": 0.9841, "step": 2669 }, { "epoch": 0.5138568129330254, "grad_norm": 1.7644081115722656, "learning_rate": 1.0037399825354036e-05, "loss": 0.9225, "step": 2670 }, { "epoch": 0.5140492686682063, "grad_norm": 1.9251432418823242, "learning_rate": 1.0031166543329179e-05, "loss": 1.0049, "step": 2671 }, { "epoch": 0.5142417244033872, "grad_norm": 1.8498098850250244, "learning_rate": 1.0024933249194792e-05, "loss": 0.9266, "step": 2672 }, { "epoch": 0.5144341801385681, "grad_norm": 1.6110475063323975, "learning_rate": 1.0018699945372776e-05, "loss": 0.8774, "step": 2673 }, { "epoch": 0.514626635873749, "grad_norm": 1.7248095273971558, "learning_rate": 1.0012466634285034e-05, "loss": 0.8597, "step": 2674 }, { "epoch": 0.5148190916089299, "grad_norm": 1.7393479347229004, "learning_rate": 1.0006233318353474e-05, "loss": 0.9337, "step": 2675 }, { "epoch": 0.5150115473441108, "grad_norm": 1.7079633474349976, "learning_rate": 1e-05, "loss": 1.0174, "step": 2676 }, { "epoch": 0.5152040030792918, "grad_norm": 1.852779746055603, "learning_rate": 9.993766681646531e-06, "loss": 0.9959, "step": 2677 }, { "epoch": 0.5153964588144727, "grad_norm": 0.9346885085105896, "learning_rate": 9.987533365714968e-06, "loss": 0.9839, "step": 2678 }, { "epoch": 0.5155889145496536, "grad_norm": 1.7506473064422607, "learning_rate": 9.981300054627227e-06, "loss": 0.8227, "step": 2679 }, { "epoch": 0.5157813702848345, "grad_norm": 1.8274807929992676, "learning_rate": 9.97506675080521e-06, "loss": 0.9407, "step": 2680 }, { "epoch": 0.5159738260200154, "grad_norm": 1.7469587326049805, "learning_rate": 9.968833456670824e-06, "loss": 0.9855, "step": 2681 }, { "epoch": 0.5161662817551963, "grad_norm": 1.5835057497024536, "learning_rate": 9.962600174645968e-06, "loss": 0.9003, "step": 2682 }, { "epoch": 0.5163587374903772, "grad_norm": 1.6277579069137573, "learning_rate": 9.956366907152536e-06, "loss": 0.9402, "step": 2683 }, { "epoch": 0.5165511932255581, "grad_norm": 0.8993666172027588, "learning_rate": 9.950133656612421e-06, "loss": 0.9376, "step": 2684 }, { "epoch": 0.516743648960739, "grad_norm": 1.8190206289291382, "learning_rate": 9.943900425447505e-06, "loss": 0.964, "step": 2685 }, { "epoch": 0.5169361046959199, "grad_norm": 1.7243624925613403, "learning_rate": 9.937667216079665e-06, "loss": 0.9561, "step": 2686 }, { "epoch": 0.5171285604311009, "grad_norm": 1.899775743484497, "learning_rate": 9.931434030930768e-06, "loss": 1.105, "step": 2687 }, { "epoch": 0.5173210161662818, "grad_norm": 0.8839601278305054, "learning_rate": 9.925200872422671e-06, "loss": 0.9132, "step": 2688 }, { "epoch": 0.5175134719014627, "grad_norm": 2.152015447616577, "learning_rate": 9.918967742977226e-06, "loss": 0.8954, "step": 2689 }, { "epoch": 0.5177059276366436, "grad_norm": 1.6577396392822266, "learning_rate": 9.912734645016262e-06, "loss": 0.927, "step": 2690 }, { "epoch": 0.5178983833718245, "grad_norm": 1.6232292652130127, "learning_rate": 9.90650158096161e-06, "loss": 1.0391, "step": 2691 }, { "epoch": 0.5180908391070054, "grad_norm": 1.7644054889678955, "learning_rate": 9.900268553235077e-06, "loss": 0.8917, "step": 2692 }, { "epoch": 0.5182832948421863, "grad_norm": 1.801323413848877, "learning_rate": 9.89403556425846e-06, "loss": 0.9176, "step": 2693 }, { "epoch": 0.5184757505773672, "grad_norm": 1.706662893295288, "learning_rate": 9.887802616453543e-06, "loss": 0.931, "step": 2694 }, { "epoch": 0.5186682063125481, "grad_norm": 1.8291906118392944, "learning_rate": 9.88156971224209e-06, "loss": 0.9976, "step": 2695 }, { "epoch": 0.5188606620477291, "grad_norm": 0.9232510924339294, "learning_rate": 9.87533685404585e-06, "loss": 0.8761, "step": 2696 }, { "epoch": 0.51905311778291, "grad_norm": 0.9049093127250671, "learning_rate": 9.869104044286558e-06, "loss": 0.9814, "step": 2697 }, { "epoch": 0.5192455735180909, "grad_norm": 1.9720518589019775, "learning_rate": 9.862871285385921e-06, "loss": 1.0042, "step": 2698 }, { "epoch": 0.5194380292532718, "grad_norm": 1.6203413009643555, "learning_rate": 9.856638579765633e-06, "loss": 0.9318, "step": 2699 }, { "epoch": 0.5196304849884527, "grad_norm": 1.765720009803772, "learning_rate": 9.850405929847367e-06, "loss": 1.0409, "step": 2700 }, { "epoch": 0.5198229407236336, "grad_norm": 1.7684149742126465, "learning_rate": 9.844173338052771e-06, "loss": 0.9908, "step": 2701 }, { "epoch": 0.5200153964588144, "grad_norm": 1.7169575691223145, "learning_rate": 9.837940806803474e-06, "loss": 0.9634, "step": 2702 }, { "epoch": 0.5202078521939953, "grad_norm": 1.7039082050323486, "learning_rate": 9.83170833852108e-06, "loss": 0.9297, "step": 2703 }, { "epoch": 0.5204003079291762, "grad_norm": 2.483220100402832, "learning_rate": 9.825475935627165e-06, "loss": 0.9691, "step": 2704 }, { "epoch": 0.5205927636643572, "grad_norm": 1.897674798965454, "learning_rate": 9.819243600543292e-06, "loss": 0.9788, "step": 2705 }, { "epoch": 0.5207852193995381, "grad_norm": 1.7572585344314575, "learning_rate": 9.81301133569098e-06, "loss": 0.963, "step": 2706 }, { "epoch": 0.520977675134719, "grad_norm": 1.5901063680648804, "learning_rate": 9.806779143491733e-06, "loss": 0.9203, "step": 2707 }, { "epoch": 0.5211701308698999, "grad_norm": 1.6806640625, "learning_rate": 9.800547026367022e-06, "loss": 0.8524, "step": 2708 }, { "epoch": 0.5213625866050808, "grad_norm": 1.7636572122573853, "learning_rate": 9.794314986738291e-06, "loss": 1.0527, "step": 2709 }, { "epoch": 0.5215550423402617, "grad_norm": 1.7680959701538086, "learning_rate": 9.788083027026953e-06, "loss": 0.9715, "step": 2710 }, { "epoch": 0.5217474980754426, "grad_norm": 1.8125712871551514, "learning_rate": 9.78185114965439e-06, "loss": 1.0034, "step": 2711 }, { "epoch": 0.5219399538106235, "grad_norm": 1.7768983840942383, "learning_rate": 9.775619357041952e-06, "loss": 0.8545, "step": 2712 }, { "epoch": 0.5221324095458044, "grad_norm": 1.876452088356018, "learning_rate": 9.769387651610955e-06, "loss": 0.9965, "step": 2713 }, { "epoch": 0.5223248652809854, "grad_norm": 1.6719523668289185, "learning_rate": 9.763156035782686e-06, "loss": 0.935, "step": 2714 }, { "epoch": 0.5225173210161663, "grad_norm": 1.656864881515503, "learning_rate": 9.756924511978388e-06, "loss": 0.9285, "step": 2715 }, { "epoch": 0.5227097767513472, "grad_norm": 1.8651610612869263, "learning_rate": 9.750693082619274e-06, "loss": 1.0017, "step": 2716 }, { "epoch": 0.5229022324865281, "grad_norm": 1.655044436454773, "learning_rate": 9.744461750126523e-06, "loss": 0.9664, "step": 2717 }, { "epoch": 0.523094688221709, "grad_norm": 1.782131552696228, "learning_rate": 9.738230516921272e-06, "loss": 0.9616, "step": 2718 }, { "epoch": 0.5232871439568899, "grad_norm": 1.9404141902923584, "learning_rate": 9.73199938542462e-06, "loss": 0.949, "step": 2719 }, { "epoch": 0.5234795996920708, "grad_norm": 1.8631519079208374, "learning_rate": 9.725768358057625e-06, "loss": 1.0062, "step": 2720 }, { "epoch": 0.5236720554272517, "grad_norm": 1.8416463136672974, "learning_rate": 9.719537437241311e-06, "loss": 0.8492, "step": 2721 }, { "epoch": 0.5238645111624326, "grad_norm": 1.6259980201721191, "learning_rate": 9.713306625396658e-06, "loss": 0.9458, "step": 2722 }, { "epoch": 0.5240569668976135, "grad_norm": 1.6901936531066895, "learning_rate": 9.707075924944595e-06, "loss": 0.9947, "step": 2723 }, { "epoch": 0.5242494226327945, "grad_norm": 1.7396188974380493, "learning_rate": 9.700845338306018e-06, "loss": 1.0049, "step": 2724 }, { "epoch": 0.5244418783679754, "grad_norm": 1.936281681060791, "learning_rate": 9.694614867901776e-06, "loss": 0.9772, "step": 2725 }, { "epoch": 0.5246343341031563, "grad_norm": 1.7554540634155273, "learning_rate": 9.68838451615267e-06, "loss": 0.912, "step": 2726 }, { "epoch": 0.5248267898383372, "grad_norm": 1.868207335472107, "learning_rate": 9.68215428547946e-06, "loss": 0.9839, "step": 2727 }, { "epoch": 0.5250192455735181, "grad_norm": 1.856107473373413, "learning_rate": 9.675924178302857e-06, "loss": 0.9815, "step": 2728 }, { "epoch": 0.525211701308699, "grad_norm": 1.7892992496490479, "learning_rate": 9.66969419704352e-06, "loss": 0.9383, "step": 2729 }, { "epoch": 0.5254041570438799, "grad_norm": 1.6498578786849976, "learning_rate": 9.663464344122064e-06, "loss": 0.8808, "step": 2730 }, { "epoch": 0.5255966127790608, "grad_norm": 2.1101608276367188, "learning_rate": 9.65723462195905e-06, "loss": 0.9568, "step": 2731 }, { "epoch": 0.5257890685142417, "grad_norm": 0.888077974319458, "learning_rate": 9.651005032974994e-06, "loss": 0.9031, "step": 2732 }, { "epoch": 0.5259815242494227, "grad_norm": 1.757644534111023, "learning_rate": 9.644775579590352e-06, "loss": 1.0133, "step": 2733 }, { "epoch": 0.5261739799846036, "grad_norm": 1.9336764812469482, "learning_rate": 9.638546264225538e-06, "loss": 0.9604, "step": 2734 }, { "epoch": 0.5263664357197845, "grad_norm": 1.6861035823822021, "learning_rate": 9.632317089300902e-06, "loss": 0.957, "step": 2735 }, { "epoch": 0.5265588914549654, "grad_norm": 1.7981489896774292, "learning_rate": 9.626088057236745e-06, "loss": 1.021, "step": 2736 }, { "epoch": 0.5267513471901463, "grad_norm": 1.54574453830719, "learning_rate": 9.619859170453314e-06, "loss": 0.9799, "step": 2737 }, { "epoch": 0.5269438029253272, "grad_norm": 1.7902891635894775, "learning_rate": 9.613630431370794e-06, "loss": 0.9842, "step": 2738 }, { "epoch": 0.5271362586605081, "grad_norm": 1.9360662698745728, "learning_rate": 9.607401842409318e-06, "loss": 0.944, "step": 2739 }, { "epoch": 0.527328714395689, "grad_norm": 1.896635890007019, "learning_rate": 9.601173405988955e-06, "loss": 0.8692, "step": 2740 }, { "epoch": 0.5275211701308699, "grad_norm": 1.6465919017791748, "learning_rate": 9.594945124529718e-06, "loss": 0.9148, "step": 2741 }, { "epoch": 0.5277136258660509, "grad_norm": 1.6439268589019775, "learning_rate": 9.588717000451563e-06, "loss": 0.8686, "step": 2742 }, { "epoch": 0.5279060816012318, "grad_norm": 1.6992038488388062, "learning_rate": 9.58248903617438e-06, "loss": 0.9656, "step": 2743 }, { "epoch": 0.5280985373364127, "grad_norm": 1.6367740631103516, "learning_rate": 9.576261234117998e-06, "loss": 0.953, "step": 2744 }, { "epoch": 0.5282909930715936, "grad_norm": 1.5830286741256714, "learning_rate": 9.570033596702185e-06, "loss": 0.9926, "step": 2745 }, { "epoch": 0.5284834488067744, "grad_norm": 1.8201066255569458, "learning_rate": 9.563806126346643e-06, "loss": 0.9275, "step": 2746 }, { "epoch": 0.5286759045419553, "grad_norm": 1.833636999130249, "learning_rate": 9.557578825471008e-06, "loss": 1.0015, "step": 2747 }, { "epoch": 0.5288683602771362, "grad_norm": 1.774916410446167, "learning_rate": 9.551351696494854e-06, "loss": 1.0241, "step": 2748 }, { "epoch": 0.5290608160123171, "grad_norm": 0.8611094355583191, "learning_rate": 9.545124741837684e-06, "loss": 0.8498, "step": 2749 }, { "epoch": 0.529253271747498, "grad_norm": 0.960064172744751, "learning_rate": 9.538897963918936e-06, "loss": 0.8913, "step": 2750 }, { "epoch": 0.5294457274826789, "grad_norm": 1.7007323503494263, "learning_rate": 9.532671365157979e-06, "loss": 0.9493, "step": 2751 }, { "epoch": 0.5296381832178599, "grad_norm": 1.7518436908721924, "learning_rate": 9.526444947974112e-06, "loss": 0.8636, "step": 2752 }, { "epoch": 0.5298306389530408, "grad_norm": 1.8486132621765137, "learning_rate": 9.520218714786564e-06, "loss": 0.9688, "step": 2753 }, { "epoch": 0.5300230946882217, "grad_norm": 1.531022548675537, "learning_rate": 9.51399266801449e-06, "loss": 0.9012, "step": 2754 }, { "epoch": 0.5302155504234026, "grad_norm": 1.753411889076233, "learning_rate": 9.507766810076972e-06, "loss": 0.9367, "step": 2755 }, { "epoch": 0.5304080061585835, "grad_norm": 0.8864457011222839, "learning_rate": 9.501541143393028e-06, "loss": 0.9405, "step": 2756 }, { "epoch": 0.5306004618937644, "grad_norm": 1.7424569129943848, "learning_rate": 9.495315670381592e-06, "loss": 0.9239, "step": 2757 }, { "epoch": 0.5307929176289453, "grad_norm": 2.003262758255005, "learning_rate": 9.489090393461523e-06, "loss": 0.9726, "step": 2758 }, { "epoch": 0.5309853733641262, "grad_norm": 1.5674717426300049, "learning_rate": 9.48286531505161e-06, "loss": 0.939, "step": 2759 }, { "epoch": 0.5311778290993071, "grad_norm": 0.9085109829902649, "learning_rate": 9.476640437570562e-06, "loss": 0.9246, "step": 2760 }, { "epoch": 0.5313702848344881, "grad_norm": 1.764864444732666, "learning_rate": 9.470415763437003e-06, "loss": 0.95, "step": 2761 }, { "epoch": 0.531562740569669, "grad_norm": 1.5910487174987793, "learning_rate": 9.46419129506949e-06, "loss": 0.961, "step": 2762 }, { "epoch": 0.5317551963048499, "grad_norm": 1.7415523529052734, "learning_rate": 9.457967034886493e-06, "loss": 0.996, "step": 2763 }, { "epoch": 0.5319476520400308, "grad_norm": 1.5188019275665283, "learning_rate": 9.4517429853064e-06, "loss": 1.0272, "step": 2764 }, { "epoch": 0.5321401077752117, "grad_norm": 1.6836899518966675, "learning_rate": 9.44551914874752e-06, "loss": 0.9227, "step": 2765 }, { "epoch": 0.5323325635103926, "grad_norm": 1.7348124980926514, "learning_rate": 9.439295527628083e-06, "loss": 0.9335, "step": 2766 }, { "epoch": 0.5325250192455735, "grad_norm": 1.8454375267028809, "learning_rate": 9.433072124366224e-06, "loss": 0.9518, "step": 2767 }, { "epoch": 0.5327174749807544, "grad_norm": 1.7137324810028076, "learning_rate": 9.426848941380007e-06, "loss": 0.9537, "step": 2768 }, { "epoch": 0.5329099307159353, "grad_norm": 1.7918413877487183, "learning_rate": 9.4206259810874e-06, "loss": 0.9893, "step": 2769 }, { "epoch": 0.5331023864511163, "grad_norm": 0.8947124481201172, "learning_rate": 9.41440324590629e-06, "loss": 0.9284, "step": 2770 }, { "epoch": 0.5332948421862972, "grad_norm": 1.8292086124420166, "learning_rate": 9.408180738254472e-06, "loss": 0.889, "step": 2771 }, { "epoch": 0.5334872979214781, "grad_norm": 1.7990657091140747, "learning_rate": 9.401958460549658e-06, "loss": 0.9077, "step": 2772 }, { "epoch": 0.533679753656659, "grad_norm": 1.765465259552002, "learning_rate": 9.395736415209468e-06, "loss": 1.0212, "step": 2773 }, { "epoch": 0.5338722093918399, "grad_norm": 0.9529819488525391, "learning_rate": 9.38951460465143e-06, "loss": 0.9122, "step": 2774 }, { "epoch": 0.5340646651270208, "grad_norm": 1.6302025318145752, "learning_rate": 9.383293031292985e-06, "loss": 0.9941, "step": 2775 }, { "epoch": 0.5342571208622017, "grad_norm": 1.7684575319290161, "learning_rate": 9.377071697551479e-06, "loss": 0.9697, "step": 2776 }, { "epoch": 0.5344495765973826, "grad_norm": 1.762911081314087, "learning_rate": 9.370850605844165e-06, "loss": 0.9392, "step": 2777 }, { "epoch": 0.5346420323325635, "grad_norm": 1.6029906272888184, "learning_rate": 9.364629758588199e-06, "loss": 0.8997, "step": 2778 }, { "epoch": 0.5348344880677445, "grad_norm": 1.8917847871780396, "learning_rate": 9.358409158200648e-06, "loss": 1.0318, "step": 2779 }, { "epoch": 0.5350269438029254, "grad_norm": 1.6695606708526611, "learning_rate": 9.352188807098482e-06, "loss": 0.9705, "step": 2780 }, { "epoch": 0.5352193995381063, "grad_norm": 1.8075644969940186, "learning_rate": 9.34596870769857e-06, "loss": 0.8862, "step": 2781 }, { "epoch": 0.5354118552732872, "grad_norm": 1.5545589923858643, "learning_rate": 9.339748862417686e-06, "loss": 0.9562, "step": 2782 }, { "epoch": 0.5356043110084681, "grad_norm": 1.6075382232666016, "learning_rate": 9.333529273672505e-06, "loss": 1.0061, "step": 2783 }, { "epoch": 0.535796766743649, "grad_norm": 1.8791543245315552, "learning_rate": 9.327309943879604e-06, "loss": 0.9562, "step": 2784 }, { "epoch": 0.5359892224788299, "grad_norm": 1.5333778858184814, "learning_rate": 9.321090875455457e-06, "loss": 0.8823, "step": 2785 }, { "epoch": 0.5361816782140107, "grad_norm": 1.6390790939331055, "learning_rate": 9.314872070816435e-06, "loss": 1.0015, "step": 2786 }, { "epoch": 0.5363741339491916, "grad_norm": 1.826635718345642, "learning_rate": 9.308653532378809e-06, "loss": 0.9091, "step": 2787 }, { "epoch": 0.5365665896843725, "grad_norm": 1.7442251443862915, "learning_rate": 9.302435262558748e-06, "loss": 0.9699, "step": 2788 }, { "epoch": 0.5367590454195535, "grad_norm": 1.7775946855545044, "learning_rate": 9.296217263772313e-06, "loss": 0.925, "step": 2789 }, { "epoch": 0.5369515011547344, "grad_norm": 1.6736632585525513, "learning_rate": 9.289999538435463e-06, "loss": 1.032, "step": 2790 }, { "epoch": 0.5371439568899153, "grad_norm": 1.6273735761642456, "learning_rate": 9.28378208896405e-06, "loss": 0.9389, "step": 2791 }, { "epoch": 0.5373364126250962, "grad_norm": 1.6790839433670044, "learning_rate": 9.277564917773816e-06, "loss": 1.0071, "step": 2792 }, { "epoch": 0.5375288683602771, "grad_norm": 1.6431087255477905, "learning_rate": 9.2713480272804e-06, "loss": 1.0477, "step": 2793 }, { "epoch": 0.537721324095458, "grad_norm": 1.8778955936431885, "learning_rate": 9.265131419899325e-06, "loss": 0.9144, "step": 2794 }, { "epoch": 0.5379137798306389, "grad_norm": 1.6644126176834106, "learning_rate": 9.258915098046008e-06, "loss": 0.936, "step": 2795 }, { "epoch": 0.5381062355658198, "grad_norm": 1.8462973833084106, "learning_rate": 9.252699064135759e-06, "loss": 0.9824, "step": 2796 }, { "epoch": 0.5382986913010007, "grad_norm": 0.9442023038864136, "learning_rate": 9.246483320583768e-06, "loss": 0.9229, "step": 2797 }, { "epoch": 0.5384911470361817, "grad_norm": 1.6928811073303223, "learning_rate": 9.240267869805118e-06, "loss": 0.9509, "step": 2798 }, { "epoch": 0.5386836027713626, "grad_norm": 1.748376488685608, "learning_rate": 9.234052714214777e-06, "loss": 1.001, "step": 2799 }, { "epoch": 0.5388760585065435, "grad_norm": 1.734100103378296, "learning_rate": 9.227837856227594e-06, "loss": 0.9428, "step": 2800 }, { "epoch": 0.5390685142417244, "grad_norm": 1.7140101194381714, "learning_rate": 9.221623298258315e-06, "loss": 0.9408, "step": 2801 }, { "epoch": 0.5392609699769053, "grad_norm": 1.8002983331680298, "learning_rate": 9.215409042721553e-06, "loss": 0.9279, "step": 2802 }, { "epoch": 0.5394534257120862, "grad_norm": 1.8716483116149902, "learning_rate": 9.209195092031811e-06, "loss": 0.9736, "step": 2803 }, { "epoch": 0.5396458814472671, "grad_norm": 1.7214860916137695, "learning_rate": 9.202981448603477e-06, "loss": 1.0205, "step": 2804 }, { "epoch": 0.539838337182448, "grad_norm": 1.67242431640625, "learning_rate": 9.196768114850814e-06, "loss": 1.0066, "step": 2805 }, { "epoch": 0.5400307929176289, "grad_norm": 1.716296911239624, "learning_rate": 9.190555093187968e-06, "loss": 0.9553, "step": 2806 }, { "epoch": 0.5402232486528099, "grad_norm": 1.8753520250320435, "learning_rate": 9.184342386028963e-06, "loss": 0.9631, "step": 2807 }, { "epoch": 0.5404157043879908, "grad_norm": 1.6706680059432983, "learning_rate": 9.178129995787698e-06, "loss": 0.9212, "step": 2808 }, { "epoch": 0.5406081601231717, "grad_norm": 1.7395124435424805, "learning_rate": 9.17191792487796e-06, "loss": 0.947, "step": 2809 }, { "epoch": 0.5408006158583526, "grad_norm": 1.7721867561340332, "learning_rate": 9.165706175713391e-06, "loss": 0.9504, "step": 2810 }, { "epoch": 0.5409930715935335, "grad_norm": 1.7275317907333374, "learning_rate": 9.159494750707527e-06, "loss": 0.9491, "step": 2811 }, { "epoch": 0.5411855273287144, "grad_norm": 1.677743911743164, "learning_rate": 9.153283652273768e-06, "loss": 0.9244, "step": 2812 }, { "epoch": 0.5413779830638953, "grad_norm": 1.598771333694458, "learning_rate": 9.147072882825394e-06, "loss": 0.9673, "step": 2813 }, { "epoch": 0.5415704387990762, "grad_norm": 1.6412098407745361, "learning_rate": 9.140862444775554e-06, "loss": 0.9523, "step": 2814 }, { "epoch": 0.5417628945342571, "grad_norm": 1.7778857946395874, "learning_rate": 9.134652340537265e-06, "loss": 0.97, "step": 2815 }, { "epoch": 0.541955350269438, "grad_norm": 1.7575709819793701, "learning_rate": 9.128442572523418e-06, "loss": 0.9536, "step": 2816 }, { "epoch": 0.542147806004619, "grad_norm": 1.801011085510254, "learning_rate": 9.122233143146777e-06, "loss": 1.0264, "step": 2817 }, { "epoch": 0.5423402617397999, "grad_norm": 1.9391707181930542, "learning_rate": 9.11602405481996e-06, "loss": 1.0187, "step": 2818 }, { "epoch": 0.5425327174749808, "grad_norm": 1.528388500213623, "learning_rate": 9.10981530995547e-06, "loss": 0.8896, "step": 2819 }, { "epoch": 0.5427251732101617, "grad_norm": 2.2085585594177246, "learning_rate": 9.103606910965666e-06, "loss": 1.0282, "step": 2820 }, { "epoch": 0.5429176289453426, "grad_norm": 1.728722095489502, "learning_rate": 9.097398860262777e-06, "loss": 0.9631, "step": 2821 }, { "epoch": 0.5431100846805235, "grad_norm": 1.669419288635254, "learning_rate": 9.091191160258895e-06, "loss": 0.9305, "step": 2822 }, { "epoch": 0.5433025404157044, "grad_norm": 1.6859018802642822, "learning_rate": 9.084983813365977e-06, "loss": 0.9921, "step": 2823 }, { "epoch": 0.5434949961508853, "grad_norm": 1.754764199256897, "learning_rate": 9.078776821995839e-06, "loss": 0.9055, "step": 2824 }, { "epoch": 0.5436874518860662, "grad_norm": 1.7501680850982666, "learning_rate": 9.072570188560164e-06, "loss": 0.9664, "step": 2825 }, { "epoch": 0.5438799076212472, "grad_norm": 0.9149741530418396, "learning_rate": 9.066363915470494e-06, "loss": 0.8986, "step": 2826 }, { "epoch": 0.5440723633564281, "grad_norm": 1.6467288732528687, "learning_rate": 9.060158005138229e-06, "loss": 0.9166, "step": 2827 }, { "epoch": 0.544264819091609, "grad_norm": 1.7493011951446533, "learning_rate": 9.05395245997463e-06, "loss": 0.9152, "step": 2828 }, { "epoch": 0.5444572748267898, "grad_norm": 1.6825838088989258, "learning_rate": 9.047747282390815e-06, "loss": 0.9339, "step": 2829 }, { "epoch": 0.5446497305619707, "grad_norm": 1.543800950050354, "learning_rate": 9.04154247479776e-06, "loss": 0.9503, "step": 2830 }, { "epoch": 0.5448421862971516, "grad_norm": 1.6420854330062866, "learning_rate": 9.0353380396063e-06, "loss": 1.0133, "step": 2831 }, { "epoch": 0.5450346420323325, "grad_norm": 1.7047592401504517, "learning_rate": 9.02913397922712e-06, "loss": 0.9512, "step": 2832 }, { "epoch": 0.5452270977675134, "grad_norm": 1.6143908500671387, "learning_rate": 9.022930296070761e-06, "loss": 0.944, "step": 2833 }, { "epoch": 0.5454195535026943, "grad_norm": 1.8498049974441528, "learning_rate": 9.016726992547621e-06, "loss": 0.9793, "step": 2834 }, { "epoch": 0.5456120092378753, "grad_norm": 1.649346947669983, "learning_rate": 9.010524071067947e-06, "loss": 0.9109, "step": 2835 }, { "epoch": 0.5458044649730562, "grad_norm": 1.6977261304855347, "learning_rate": 9.004321534041836e-06, "loss": 0.8795, "step": 2836 }, { "epoch": 0.5459969207082371, "grad_norm": 1.7800700664520264, "learning_rate": 8.99811938387924e-06, "loss": 0.8888, "step": 2837 }, { "epoch": 0.546189376443418, "grad_norm": 1.877732276916504, "learning_rate": 8.991917622989957e-06, "loss": 0.8706, "step": 2838 }, { "epoch": 0.5463818321785989, "grad_norm": 1.7536191940307617, "learning_rate": 8.985716253783636e-06, "loss": 0.9717, "step": 2839 }, { "epoch": 0.5465742879137798, "grad_norm": 1.9079275131225586, "learning_rate": 8.979515278669776e-06, "loss": 0.9489, "step": 2840 }, { "epoch": 0.5467667436489607, "grad_norm": 1.7857767343521118, "learning_rate": 8.973314700057717e-06, "loss": 0.9359, "step": 2841 }, { "epoch": 0.5469591993841416, "grad_norm": 1.6987247467041016, "learning_rate": 8.967114520356652e-06, "loss": 0.8808, "step": 2842 }, { "epoch": 0.5471516551193225, "grad_norm": 1.6191747188568115, "learning_rate": 8.960914741975608e-06, "loss": 0.9468, "step": 2843 }, { "epoch": 0.5473441108545035, "grad_norm": 1.631452202796936, "learning_rate": 8.954715367323468e-06, "loss": 0.9426, "step": 2844 }, { "epoch": 0.5475365665896844, "grad_norm": 1.713382363319397, "learning_rate": 8.948516398808952e-06, "loss": 0.9835, "step": 2845 }, { "epoch": 0.5477290223248653, "grad_norm": 1.7706388235092163, "learning_rate": 8.942317838840625e-06, "loss": 0.9719, "step": 2846 }, { "epoch": 0.5479214780600462, "grad_norm": 0.9782737493515015, "learning_rate": 8.936119689826888e-06, "loss": 0.9175, "step": 2847 }, { "epoch": 0.5481139337952271, "grad_norm": 1.8026446104049683, "learning_rate": 8.92992195417599e-06, "loss": 1.0025, "step": 2848 }, { "epoch": 0.548306389530408, "grad_norm": 1.7526710033416748, "learning_rate": 8.923724634296013e-06, "loss": 1.0375, "step": 2849 }, { "epoch": 0.5484988452655889, "grad_norm": 1.8215643167495728, "learning_rate": 8.917527732594882e-06, "loss": 0.9201, "step": 2850 }, { "epoch": 0.5486913010007698, "grad_norm": 1.5967341661453247, "learning_rate": 8.911331251480357e-06, "loss": 0.9046, "step": 2851 }, { "epoch": 0.5488837567359507, "grad_norm": 1.7772562503814697, "learning_rate": 8.905135193360032e-06, "loss": 0.9861, "step": 2852 }, { "epoch": 0.5490762124711316, "grad_norm": 1.629896879196167, "learning_rate": 8.898939560641342e-06, "loss": 0.929, "step": 2853 }, { "epoch": 0.5492686682063126, "grad_norm": 1.845942735671997, "learning_rate": 8.892744355731555e-06, "loss": 0.9288, "step": 2854 }, { "epoch": 0.5494611239414935, "grad_norm": 1.8825474977493286, "learning_rate": 8.88654958103777e-06, "loss": 0.9142, "step": 2855 }, { "epoch": 0.5496535796766744, "grad_norm": 1.5987112522125244, "learning_rate": 8.880355238966923e-06, "loss": 0.9431, "step": 2856 }, { "epoch": 0.5498460354118553, "grad_norm": 1.801778793334961, "learning_rate": 8.874161331925778e-06, "loss": 0.9046, "step": 2857 }, { "epoch": 0.5500384911470362, "grad_norm": 2.168508768081665, "learning_rate": 8.867967862320935e-06, "loss": 0.9293, "step": 2858 }, { "epoch": 0.5502309468822171, "grad_norm": 1.864163875579834, "learning_rate": 8.861774832558819e-06, "loss": 0.9592, "step": 2859 }, { "epoch": 0.550423402617398, "grad_norm": 1.7023168802261353, "learning_rate": 8.855582245045682e-06, "loss": 1.0487, "step": 2860 }, { "epoch": 0.5506158583525789, "grad_norm": 1.855760931968689, "learning_rate": 8.849390102187615e-06, "loss": 0.9347, "step": 2861 }, { "epoch": 0.5508083140877598, "grad_norm": 0.9254429340362549, "learning_rate": 8.843198406390524e-06, "loss": 0.9697, "step": 2862 }, { "epoch": 0.5510007698229408, "grad_norm": 2.01577091217041, "learning_rate": 8.83700716006015e-06, "loss": 0.9819, "step": 2863 }, { "epoch": 0.5511932255581217, "grad_norm": 1.780637502670288, "learning_rate": 8.830816365602053e-06, "loss": 0.9721, "step": 2864 }, { "epoch": 0.5513856812933026, "grad_norm": 1.836281180381775, "learning_rate": 8.824626025421625e-06, "loss": 0.986, "step": 2865 }, { "epoch": 0.5515781370284835, "grad_norm": 1.9009772539138794, "learning_rate": 8.818436141924072e-06, "loss": 1.0788, "step": 2866 }, { "epoch": 0.5517705927636644, "grad_norm": 1.6611988544464111, "learning_rate": 8.812246717514434e-06, "loss": 0.9686, "step": 2867 }, { "epoch": 0.5519630484988453, "grad_norm": 1.8770869970321655, "learning_rate": 8.806057754597559e-06, "loss": 1.0014, "step": 2868 }, { "epoch": 0.5521555042340262, "grad_norm": 1.7273067235946655, "learning_rate": 8.799869255578126e-06, "loss": 1.0017, "step": 2869 }, { "epoch": 0.552347959969207, "grad_norm": 1.6882126331329346, "learning_rate": 8.793681222860631e-06, "loss": 0.9142, "step": 2870 }, { "epoch": 0.5525404157043879, "grad_norm": 1.820639967918396, "learning_rate": 8.787493658849387e-06, "loss": 0.9597, "step": 2871 }, { "epoch": 0.552732871439569, "grad_norm": 1.7515637874603271, "learning_rate": 8.781306565948528e-06, "loss": 0.9842, "step": 2872 }, { "epoch": 0.5529253271747498, "grad_norm": 1.87770414352417, "learning_rate": 8.775119946562e-06, "loss": 0.8589, "step": 2873 }, { "epoch": 0.5531177829099307, "grad_norm": 0.9481867551803589, "learning_rate": 8.768933803093573e-06, "loss": 0.9177, "step": 2874 }, { "epoch": 0.5533102386451116, "grad_norm": 1.91827392578125, "learning_rate": 8.762748137946827e-06, "loss": 0.9837, "step": 2875 }, { "epoch": 0.5535026943802925, "grad_norm": 1.6901836395263672, "learning_rate": 8.756562953525151e-06, "loss": 1.0418, "step": 2876 }, { "epoch": 0.5536951501154734, "grad_norm": 1.6748099327087402, "learning_rate": 8.750378252231756e-06, "loss": 0.9919, "step": 2877 }, { "epoch": 0.5538876058506543, "grad_norm": 1.7069025039672852, "learning_rate": 8.744194036469665e-06, "loss": 0.9318, "step": 2878 }, { "epoch": 0.5540800615858352, "grad_norm": 1.5615936517715454, "learning_rate": 8.738010308641705e-06, "loss": 0.862, "step": 2879 }, { "epoch": 0.5542725173210161, "grad_norm": 1.6008946895599365, "learning_rate": 8.731827071150519e-06, "loss": 0.9796, "step": 2880 }, { "epoch": 0.5544649730561971, "grad_norm": 1.6812937259674072, "learning_rate": 8.72564432639856e-06, "loss": 0.8774, "step": 2881 }, { "epoch": 0.554657428791378, "grad_norm": 1.775210976600647, "learning_rate": 8.719462076788085e-06, "loss": 0.9394, "step": 2882 }, { "epoch": 0.5548498845265589, "grad_norm": 1.7067625522613525, "learning_rate": 8.713280324721163e-06, "loss": 1.0047, "step": 2883 }, { "epoch": 0.5550423402617398, "grad_norm": 1.646234154701233, "learning_rate": 8.70709907259967e-06, "loss": 1.0137, "step": 2884 }, { "epoch": 0.5552347959969207, "grad_norm": 1.5783158540725708, "learning_rate": 8.700918322825282e-06, "loss": 0.9036, "step": 2885 }, { "epoch": 0.5554272517321016, "grad_norm": 1.6035501956939697, "learning_rate": 8.694738077799487e-06, "loss": 0.9008, "step": 2886 }, { "epoch": 0.5556197074672825, "grad_norm": 1.6333136558532715, "learning_rate": 8.68855833992357e-06, "loss": 0.9333, "step": 2887 }, { "epoch": 0.5558121632024634, "grad_norm": 1.7241555452346802, "learning_rate": 8.682379111598626e-06, "loss": 0.9497, "step": 2888 }, { "epoch": 0.5560046189376443, "grad_norm": 1.8575297594070435, "learning_rate": 8.676200395225546e-06, "loss": 0.9127, "step": 2889 }, { "epoch": 0.5561970746728252, "grad_norm": 1.671816349029541, "learning_rate": 8.670022193205027e-06, "loss": 0.9904, "step": 2890 }, { "epoch": 0.5563895304080062, "grad_norm": 1.693889856338501, "learning_rate": 8.663844507937563e-06, "loss": 0.9288, "step": 2891 }, { "epoch": 0.5565819861431871, "grad_norm": 1.534420371055603, "learning_rate": 8.657667341823449e-06, "loss": 0.9832, "step": 2892 }, { "epoch": 0.556774441878368, "grad_norm": 1.737310528755188, "learning_rate": 8.651490697262773e-06, "loss": 1.0157, "step": 2893 }, { "epoch": 0.5569668976135489, "grad_norm": 1.7371097803115845, "learning_rate": 8.645314576655428e-06, "loss": 0.9968, "step": 2894 }, { "epoch": 0.5571593533487298, "grad_norm": 1.75102961063385, "learning_rate": 8.6391389824011e-06, "loss": 0.9741, "step": 2895 }, { "epoch": 0.5573518090839107, "grad_norm": 1.7553242444992065, "learning_rate": 8.632963916899268e-06, "loss": 0.9553, "step": 2896 }, { "epoch": 0.5575442648190916, "grad_norm": 1.7422131299972534, "learning_rate": 8.626789382549212e-06, "loss": 0.9854, "step": 2897 }, { "epoch": 0.5577367205542725, "grad_norm": 1.730526089668274, "learning_rate": 8.62061538175e-06, "loss": 0.9712, "step": 2898 }, { "epoch": 0.5579291762894534, "grad_norm": 1.630471110343933, "learning_rate": 8.614441916900493e-06, "loss": 1.0146, "step": 2899 }, { "epoch": 0.5581216320246344, "grad_norm": 2.00028657913208, "learning_rate": 8.60826899039935e-06, "loss": 1.0192, "step": 2900 }, { "epoch": 0.5583140877598153, "grad_norm": 1.8350640535354614, "learning_rate": 8.602096604645009e-06, "loss": 1.0603, "step": 2901 }, { "epoch": 0.5585065434949962, "grad_norm": 1.5071982145309448, "learning_rate": 8.59592476203571e-06, "loss": 0.9117, "step": 2902 }, { "epoch": 0.5586989992301771, "grad_norm": 1.596524953842163, "learning_rate": 8.589753464969473e-06, "loss": 0.9189, "step": 2903 }, { "epoch": 0.558891454965358, "grad_norm": 1.827192783355713, "learning_rate": 8.583582715844113e-06, "loss": 0.986, "step": 2904 }, { "epoch": 0.5590839107005389, "grad_norm": 1.6590081453323364, "learning_rate": 8.577412517057227e-06, "loss": 0.9524, "step": 2905 }, { "epoch": 0.5592763664357198, "grad_norm": 1.7904973030090332, "learning_rate": 8.571242871006202e-06, "loss": 0.9958, "step": 2906 }, { "epoch": 0.5594688221709007, "grad_norm": 1.8047839403152466, "learning_rate": 8.56507378008821e-06, "loss": 0.9985, "step": 2907 }, { "epoch": 0.5596612779060816, "grad_norm": 1.9825135469436646, "learning_rate": 8.558905246700202e-06, "loss": 0.8489, "step": 2908 }, { "epoch": 0.5598537336412626, "grad_norm": 1.623400092124939, "learning_rate": 8.552737273238919e-06, "loss": 0.8784, "step": 2909 }, { "epoch": 0.5600461893764435, "grad_norm": 1.6467536687850952, "learning_rate": 8.546569862100876e-06, "loss": 0.9128, "step": 2910 }, { "epoch": 0.5602386451116244, "grad_norm": 0.9603508114814758, "learning_rate": 8.540403015682382e-06, "loss": 0.892, "step": 2911 }, { "epoch": 0.5604311008468053, "grad_norm": 1.71060311794281, "learning_rate": 8.534236736379515e-06, "loss": 0.8771, "step": 2912 }, { "epoch": 0.5606235565819861, "grad_norm": 1.7929786443710327, "learning_rate": 8.528071026588137e-06, "loss": 0.9782, "step": 2913 }, { "epoch": 0.560816012317167, "grad_norm": 1.6755218505859375, "learning_rate": 8.521905888703894e-06, "loss": 1.0493, "step": 2914 }, { "epoch": 0.5610084680523479, "grad_norm": 1.8743388652801514, "learning_rate": 8.515741325122202e-06, "loss": 0.8865, "step": 2915 }, { "epoch": 0.5612009237875288, "grad_norm": 0.8616350889205933, "learning_rate": 8.509577338238255e-06, "loss": 0.8352, "step": 2916 }, { "epoch": 0.5613933795227097, "grad_norm": 1.7416044473648071, "learning_rate": 8.503413930447027e-06, "loss": 0.9708, "step": 2917 }, { "epoch": 0.5615858352578906, "grad_norm": 1.8031952381134033, "learning_rate": 8.497251104143263e-06, "loss": 1.0593, "step": 2918 }, { "epoch": 0.5617782909930716, "grad_norm": 1.6629844903945923, "learning_rate": 8.491088861721486e-06, "loss": 0.9643, "step": 2919 }, { "epoch": 0.5619707467282525, "grad_norm": 1.7399660348892212, "learning_rate": 8.484927205575985e-06, "loss": 0.8943, "step": 2920 }, { "epoch": 0.5621632024634334, "grad_norm": 1.714220643043518, "learning_rate": 8.478766138100834e-06, "loss": 0.9249, "step": 2921 }, { "epoch": 0.5623556581986143, "grad_norm": 1.6782300472259521, "learning_rate": 8.472605661689863e-06, "loss": 0.9667, "step": 2922 }, { "epoch": 0.5625481139337952, "grad_norm": 1.9140594005584717, "learning_rate": 8.46644577873668e-06, "loss": 0.9177, "step": 2923 }, { "epoch": 0.5627405696689761, "grad_norm": 1.5623764991760254, "learning_rate": 8.460286491634664e-06, "loss": 0.9155, "step": 2924 }, { "epoch": 0.562933025404157, "grad_norm": 1.5963772535324097, "learning_rate": 8.454127802776961e-06, "loss": 0.9294, "step": 2925 }, { "epoch": 0.5631254811393379, "grad_norm": 1.7991639375686646, "learning_rate": 8.447969714556484e-06, "loss": 0.9477, "step": 2926 }, { "epoch": 0.5633179368745188, "grad_norm": 1.9068244695663452, "learning_rate": 8.441812229365913e-06, "loss": 0.9209, "step": 2927 }, { "epoch": 0.5635103926096998, "grad_norm": 1.7739561796188354, "learning_rate": 8.43565534959769e-06, "loss": 0.8474, "step": 2928 }, { "epoch": 0.5637028483448807, "grad_norm": 1.7665737867355347, "learning_rate": 8.42949907764403e-06, "loss": 0.8932, "step": 2929 }, { "epoch": 0.5638953040800616, "grad_norm": 1.8083511590957642, "learning_rate": 8.423343415896904e-06, "loss": 0.9798, "step": 2930 }, { "epoch": 0.5640877598152425, "grad_norm": 1.6626704931259155, "learning_rate": 8.417188366748051e-06, "loss": 0.9924, "step": 2931 }, { "epoch": 0.5642802155504234, "grad_norm": 1.8067280054092407, "learning_rate": 8.411033932588969e-06, "loss": 0.9907, "step": 2932 }, { "epoch": 0.5644726712856043, "grad_norm": 1.6312150955200195, "learning_rate": 8.404880115810914e-06, "loss": 0.8916, "step": 2933 }, { "epoch": 0.5646651270207852, "grad_norm": 1.740869164466858, "learning_rate": 8.398726918804911e-06, "loss": 1.0291, "step": 2934 }, { "epoch": 0.5648575827559661, "grad_norm": 1.8676230907440186, "learning_rate": 8.392574343961739e-06, "loss": 0.9492, "step": 2935 }, { "epoch": 0.565050038491147, "grad_norm": 1.7246983051300049, "learning_rate": 8.386422393671934e-06, "loss": 0.9434, "step": 2936 }, { "epoch": 0.565242494226328, "grad_norm": 1.64383065700531, "learning_rate": 8.380271070325789e-06, "loss": 0.9716, "step": 2937 }, { "epoch": 0.5654349499615089, "grad_norm": 1.5677272081375122, "learning_rate": 8.374120376313362e-06, "loss": 1.0142, "step": 2938 }, { "epoch": 0.5656274056966898, "grad_norm": 1.6247469186782837, "learning_rate": 8.367970314024451e-06, "loss": 0.9866, "step": 2939 }, { "epoch": 0.5658198614318707, "grad_norm": 1.7564074993133545, "learning_rate": 8.361820885848623e-06, "loss": 0.9085, "step": 2940 }, { "epoch": 0.5660123171670516, "grad_norm": 1.8154188394546509, "learning_rate": 8.355672094175192e-06, "loss": 1.0384, "step": 2941 }, { "epoch": 0.5662047729022325, "grad_norm": 1.7659227848052979, "learning_rate": 8.349523941393224e-06, "loss": 0.9456, "step": 2942 }, { "epoch": 0.5663972286374134, "grad_norm": 1.6569443941116333, "learning_rate": 8.343376429891538e-06, "loss": 0.8884, "step": 2943 }, { "epoch": 0.5665896843725943, "grad_norm": 1.7725307941436768, "learning_rate": 8.337229562058707e-06, "loss": 1.0149, "step": 2944 }, { "epoch": 0.5667821401077752, "grad_norm": 1.6769144535064697, "learning_rate": 8.331083340283051e-06, "loss": 0.9744, "step": 2945 }, { "epoch": 0.5669745958429562, "grad_norm": 1.5744061470031738, "learning_rate": 8.324937766952638e-06, "loss": 0.919, "step": 2946 }, { "epoch": 0.5671670515781371, "grad_norm": 1.7301446199417114, "learning_rate": 8.318792844455282e-06, "loss": 0.9959, "step": 2947 }, { "epoch": 0.567359507313318, "grad_norm": 1.666870355606079, "learning_rate": 8.312648575178552e-06, "loss": 0.8693, "step": 2948 }, { "epoch": 0.5675519630484989, "grad_norm": 1.6792060136795044, "learning_rate": 8.306504961509755e-06, "loss": 1.0275, "step": 2949 }, { "epoch": 0.5677444187836798, "grad_norm": 1.847460389137268, "learning_rate": 8.300362005835947e-06, "loss": 0.9736, "step": 2950 }, { "epoch": 0.5679368745188607, "grad_norm": 0.9470747709274292, "learning_rate": 8.294219710543931e-06, "loss": 0.8863, "step": 2951 }, { "epoch": 0.5681293302540416, "grad_norm": 1.7566694021224976, "learning_rate": 8.28807807802025e-06, "loss": 0.9648, "step": 2952 }, { "epoch": 0.5683217859892224, "grad_norm": 1.8803002834320068, "learning_rate": 8.281937110651187e-06, "loss": 0.9391, "step": 2953 }, { "epoch": 0.5685142417244033, "grad_norm": 1.7637379169464111, "learning_rate": 8.275796810822774e-06, "loss": 1.0002, "step": 2954 }, { "epoch": 0.5687066974595842, "grad_norm": 1.7256768941879272, "learning_rate": 8.269657180920773e-06, "loss": 0.9718, "step": 2955 }, { "epoch": 0.5688991531947653, "grad_norm": 1.9248450994491577, "learning_rate": 8.263518223330698e-06, "loss": 0.926, "step": 2956 }, { "epoch": 0.5690916089299461, "grad_norm": 1.8923312425613403, "learning_rate": 8.25737994043779e-06, "loss": 1.019, "step": 2957 }, { "epoch": 0.569284064665127, "grad_norm": 1.6377313137054443, "learning_rate": 8.251242334627039e-06, "loss": 0.9763, "step": 2958 }, { "epoch": 0.5694765204003079, "grad_norm": 2.114448308944702, "learning_rate": 8.245105408283162e-06, "loss": 0.9249, "step": 2959 }, { "epoch": 0.5696689761354888, "grad_norm": 1.8890897035598755, "learning_rate": 8.238969163790617e-06, "loss": 0.9636, "step": 2960 }, { "epoch": 0.5698614318706697, "grad_norm": 1.7341420650482178, "learning_rate": 8.232833603533601e-06, "loss": 0.9279, "step": 2961 }, { "epoch": 0.5700538876058506, "grad_norm": 1.80060875415802, "learning_rate": 8.226698729896036e-06, "loss": 0.9096, "step": 2962 }, { "epoch": 0.5702463433410315, "grad_norm": 1.902990698814392, "learning_rate": 8.220564545261585e-06, "loss": 0.8863, "step": 2963 }, { "epoch": 0.5704387990762124, "grad_norm": 1.815102458000183, "learning_rate": 8.214431052013636e-06, "loss": 0.9308, "step": 2964 }, { "epoch": 0.5706312548113934, "grad_norm": 1.7460057735443115, "learning_rate": 8.208298252535314e-06, "loss": 1.0073, "step": 2965 }, { "epoch": 0.5708237105465743, "grad_norm": 1.9828581809997559, "learning_rate": 8.202166149209475e-06, "loss": 0.964, "step": 2966 }, { "epoch": 0.5710161662817552, "grad_norm": 1.8354816436767578, "learning_rate": 8.1960347444187e-06, "loss": 0.8836, "step": 2967 }, { "epoch": 0.5712086220169361, "grad_norm": 1.7165905237197876, "learning_rate": 8.189904040545302e-06, "loss": 1.0216, "step": 2968 }, { "epoch": 0.571401077752117, "grad_norm": 1.9106067419052124, "learning_rate": 8.183774039971322e-06, "loss": 0.9282, "step": 2969 }, { "epoch": 0.5715935334872979, "grad_norm": 1.6354033946990967, "learning_rate": 8.177644745078525e-06, "loss": 1.01, "step": 2970 }, { "epoch": 0.5717859892224788, "grad_norm": 1.909069299697876, "learning_rate": 8.171516158248406e-06, "loss": 1.0456, "step": 2971 }, { "epoch": 0.5719784449576597, "grad_norm": 1.6466954946517944, "learning_rate": 8.165388281862177e-06, "loss": 0.9929, "step": 2972 }, { "epoch": 0.5721709006928406, "grad_norm": 1.753554105758667, "learning_rate": 8.159261118300784e-06, "loss": 0.8903, "step": 2973 }, { "epoch": 0.5723633564280216, "grad_norm": 2.213350296020508, "learning_rate": 8.153134669944887e-06, "loss": 0.9208, "step": 2974 }, { "epoch": 0.5725558121632025, "grad_norm": 1.6472336053848267, "learning_rate": 8.147008939174875e-06, "loss": 0.9468, "step": 2975 }, { "epoch": 0.5727482678983834, "grad_norm": 1.952000617980957, "learning_rate": 8.140883928370855e-06, "loss": 0.9675, "step": 2976 }, { "epoch": 0.5729407236335643, "grad_norm": 1.5846456289291382, "learning_rate": 8.134759639912655e-06, "loss": 0.964, "step": 2977 }, { "epoch": 0.5731331793687452, "grad_norm": 1.758061170578003, "learning_rate": 8.12863607617982e-06, "loss": 0.8388, "step": 2978 }, { "epoch": 0.5733256351039261, "grad_norm": 1.8731545209884644, "learning_rate": 8.12251323955162e-06, "loss": 0.9636, "step": 2979 }, { "epoch": 0.573518090839107, "grad_norm": 1.6598634719848633, "learning_rate": 8.116391132407033e-06, "loss": 0.8807, "step": 2980 }, { "epoch": 0.5737105465742879, "grad_norm": 1.7117222547531128, "learning_rate": 8.11026975712476e-06, "loss": 0.9168, "step": 2981 }, { "epoch": 0.5739030023094688, "grad_norm": 1.8713873624801636, "learning_rate": 8.104149116083216e-06, "loss": 0.9744, "step": 2982 }, { "epoch": 0.5740954580446497, "grad_norm": 1.9112058877944946, "learning_rate": 8.098029211660534e-06, "loss": 0.9518, "step": 2983 }, { "epoch": 0.5742879137798307, "grad_norm": 1.8843671083450317, "learning_rate": 8.091910046234552e-06, "loss": 0.9087, "step": 2984 }, { "epoch": 0.5744803695150116, "grad_norm": 1.9149013757705688, "learning_rate": 8.085791622182832e-06, "loss": 1.0059, "step": 2985 }, { "epoch": 0.5746728252501925, "grad_norm": 1.6751363277435303, "learning_rate": 8.079673941882639e-06, "loss": 0.9408, "step": 2986 }, { "epoch": 0.5748652809853734, "grad_norm": 1.8452701568603516, "learning_rate": 8.073557007710957e-06, "loss": 0.9295, "step": 2987 }, { "epoch": 0.5750577367205543, "grad_norm": 1.6879347562789917, "learning_rate": 8.06744082204447e-06, "loss": 0.9044, "step": 2988 }, { "epoch": 0.5752501924557352, "grad_norm": 2.1773629188537598, "learning_rate": 8.061325387259578e-06, "loss": 0.9714, "step": 2989 }, { "epoch": 0.5754426481909161, "grad_norm": 1.7719420194625854, "learning_rate": 8.055210705732391e-06, "loss": 0.8865, "step": 2990 }, { "epoch": 0.575635103926097, "grad_norm": 1.685840368270874, "learning_rate": 8.04909677983872e-06, "loss": 0.9435, "step": 2991 }, { "epoch": 0.5758275596612779, "grad_norm": 1.6655336618423462, "learning_rate": 8.042983611954087e-06, "loss": 0.9102, "step": 2992 }, { "epoch": 0.5760200153964589, "grad_norm": 2.123296022415161, "learning_rate": 8.036871204453718e-06, "loss": 0.9706, "step": 2993 }, { "epoch": 0.5762124711316398, "grad_norm": 1.7848074436187744, "learning_rate": 8.030759559712545e-06, "loss": 0.9897, "step": 2994 }, { "epoch": 0.5764049268668207, "grad_norm": 1.7156323194503784, "learning_rate": 8.0246486801052e-06, "loss": 0.9861, "step": 2995 }, { "epoch": 0.5765973826020016, "grad_norm": 1.6514891386032104, "learning_rate": 8.018538568006027e-06, "loss": 0.9241, "step": 2996 }, { "epoch": 0.5767898383371824, "grad_norm": 1.7439475059509277, "learning_rate": 8.012429225789055e-06, "loss": 0.9733, "step": 2997 }, { "epoch": 0.5769822940723633, "grad_norm": 1.964775562286377, "learning_rate": 8.00632065582803e-06, "loss": 1.0217, "step": 2998 }, { "epoch": 0.5771747498075442, "grad_norm": 1.7098491191864014, "learning_rate": 8.00021286049639e-06, "loss": 0.9001, "step": 2999 }, { "epoch": 0.5773672055427251, "grad_norm": 1.857249140739441, "learning_rate": 7.994105842167274e-06, "loss": 0.8949, "step": 3000 }, { "epoch": 0.577559661277906, "grad_norm": 1.6955369710922241, "learning_rate": 7.987999603213518e-06, "loss": 0.9822, "step": 3001 }, { "epoch": 0.577752117013087, "grad_norm": 1.7742730379104614, "learning_rate": 7.981894146007658e-06, "loss": 0.9244, "step": 3002 }, { "epoch": 0.5779445727482679, "grad_norm": 1.824655294418335, "learning_rate": 7.975789472921922e-06, "loss": 0.9581, "step": 3003 }, { "epoch": 0.5781370284834488, "grad_norm": 1.8045625686645508, "learning_rate": 7.96968558632824e-06, "loss": 1.0254, "step": 3004 }, { "epoch": 0.5783294842186297, "grad_norm": 1.6778502464294434, "learning_rate": 7.963582488598227e-06, "loss": 0.9704, "step": 3005 }, { "epoch": 0.5785219399538106, "grad_norm": 1.6765117645263672, "learning_rate": 7.957480182103198e-06, "loss": 0.8823, "step": 3006 }, { "epoch": 0.5787143956889915, "grad_norm": 1.6389504671096802, "learning_rate": 7.95137866921416e-06, "loss": 0.8671, "step": 3007 }, { "epoch": 0.5789068514241724, "grad_norm": 1.5550858974456787, "learning_rate": 7.945277952301811e-06, "loss": 0.9456, "step": 3008 }, { "epoch": 0.5790993071593533, "grad_norm": 1.72085440158844, "learning_rate": 7.939178033736538e-06, "loss": 1.0094, "step": 3009 }, { "epoch": 0.5792917628945342, "grad_norm": 1.7494829893112183, "learning_rate": 7.93307891588842e-06, "loss": 1.0041, "step": 3010 }, { "epoch": 0.5794842186297152, "grad_norm": 1.7873876094818115, "learning_rate": 7.926980601127225e-06, "loss": 0.992, "step": 3011 }, { "epoch": 0.5796766743648961, "grad_norm": 1.9800448417663574, "learning_rate": 7.92088309182241e-06, "loss": 0.8435, "step": 3012 }, { "epoch": 0.579869130100077, "grad_norm": 1.7443515062332153, "learning_rate": 7.914786390343112e-06, "loss": 0.9594, "step": 3013 }, { "epoch": 0.5800615858352579, "grad_norm": 1.7304186820983887, "learning_rate": 7.908690499058162e-06, "loss": 0.979, "step": 3014 }, { "epoch": 0.5802540415704388, "grad_norm": 1.65768301486969, "learning_rate": 7.902595420336074e-06, "loss": 0.9446, "step": 3015 }, { "epoch": 0.5804464973056197, "grad_norm": 1.61051607131958, "learning_rate": 7.896501156545044e-06, "loss": 0.9181, "step": 3016 }, { "epoch": 0.5806389530408006, "grad_norm": 1.6247704029083252, "learning_rate": 7.890407710052957e-06, "loss": 0.966, "step": 3017 }, { "epoch": 0.5808314087759815, "grad_norm": 1.6123080253601074, "learning_rate": 7.884315083227373e-06, "loss": 0.9386, "step": 3018 }, { "epoch": 0.5810238645111624, "grad_norm": 0.9107101559638977, "learning_rate": 7.878223278435539e-06, "loss": 0.9371, "step": 3019 }, { "epoch": 0.5812163202463433, "grad_norm": 1.6647801399230957, "learning_rate": 7.872132298044382e-06, "loss": 0.8843, "step": 3020 }, { "epoch": 0.5814087759815243, "grad_norm": 1.5860421657562256, "learning_rate": 7.866042144420502e-06, "loss": 0.9742, "step": 3021 }, { "epoch": 0.5816012317167052, "grad_norm": 1.7504537105560303, "learning_rate": 7.859952819930188e-06, "loss": 0.9993, "step": 3022 }, { "epoch": 0.5817936874518861, "grad_norm": 1.7566826343536377, "learning_rate": 7.853864326939403e-06, "loss": 1.0004, "step": 3023 }, { "epoch": 0.581986143187067, "grad_norm": 1.6100963354110718, "learning_rate": 7.847776667813782e-06, "loss": 0.9635, "step": 3024 }, { "epoch": 0.5821785989222479, "grad_norm": 1.633142113685608, "learning_rate": 7.841689844918643e-06, "loss": 0.935, "step": 3025 }, { "epoch": 0.5823710546574288, "grad_norm": 1.7092989683151245, "learning_rate": 7.835603860618973e-06, "loss": 1.0052, "step": 3026 }, { "epoch": 0.5825635103926097, "grad_norm": 1.6931356191635132, "learning_rate": 7.829518717279439e-06, "loss": 0.9583, "step": 3027 }, { "epoch": 0.5827559661277906, "grad_norm": 1.8024014234542847, "learning_rate": 7.823434417264378e-06, "loss": 0.997, "step": 3028 }, { "epoch": 0.5829484218629715, "grad_norm": 1.786964774131775, "learning_rate": 7.817350962937802e-06, "loss": 0.9737, "step": 3029 }, { "epoch": 0.5831408775981525, "grad_norm": 1.7206529378890991, "learning_rate": 7.811268356663387e-06, "loss": 0.9519, "step": 3030 }, { "epoch": 0.5833333333333334, "grad_norm": 1.6637228727340698, "learning_rate": 7.805186600804489e-06, "loss": 0.9564, "step": 3031 }, { "epoch": 0.5835257890685143, "grad_norm": 1.7595679759979248, "learning_rate": 7.799105697724127e-06, "loss": 0.9108, "step": 3032 }, { "epoch": 0.5837182448036952, "grad_norm": 1.7303951978683472, "learning_rate": 7.793025649784991e-06, "loss": 0.9525, "step": 3033 }, { "epoch": 0.5839107005388761, "grad_norm": 1.8665261268615723, "learning_rate": 7.786946459349441e-06, "loss": 1.0046, "step": 3034 }, { "epoch": 0.584103156274057, "grad_norm": 1.6762267351150513, "learning_rate": 7.780868128779499e-06, "loss": 0.9512, "step": 3035 }, { "epoch": 0.5842956120092379, "grad_norm": 1.628430724143982, "learning_rate": 7.774790660436857e-06, "loss": 0.9701, "step": 3036 }, { "epoch": 0.5844880677444187, "grad_norm": 1.5883457660675049, "learning_rate": 7.768714056682873e-06, "loss": 0.9756, "step": 3037 }, { "epoch": 0.5846805234795996, "grad_norm": 0.9464545249938965, "learning_rate": 7.76263831987856e-06, "loss": 0.9011, "step": 3038 }, { "epoch": 0.5848729792147807, "grad_norm": 1.6757656335830688, "learning_rate": 7.756563452384605e-06, "loss": 0.9676, "step": 3039 }, { "epoch": 0.5850654349499615, "grad_norm": 1.796553611755371, "learning_rate": 7.750489456561351e-06, "loss": 0.9589, "step": 3040 }, { "epoch": 0.5852578906851424, "grad_norm": 1.5950746536254883, "learning_rate": 7.744416334768809e-06, "loss": 0.9715, "step": 3041 }, { "epoch": 0.5854503464203233, "grad_norm": 1.8375208377838135, "learning_rate": 7.738344089366638e-06, "loss": 0.9296, "step": 3042 }, { "epoch": 0.5856428021555042, "grad_norm": 1.8418134450912476, "learning_rate": 7.73227272271417e-06, "loss": 0.9209, "step": 3043 }, { "epoch": 0.5858352578906851, "grad_norm": 1.62941575050354, "learning_rate": 7.726202237170387e-06, "loss": 0.9131, "step": 3044 }, { "epoch": 0.586027713625866, "grad_norm": 1.942509651184082, "learning_rate": 7.720132635093935e-06, "loss": 0.8739, "step": 3045 }, { "epoch": 0.5862201693610469, "grad_norm": 0.8891316056251526, "learning_rate": 7.714063918843106e-06, "loss": 0.9305, "step": 3046 }, { "epoch": 0.5864126250962278, "grad_norm": 1.9507766962051392, "learning_rate": 7.70799609077586e-06, "loss": 0.9344, "step": 3047 }, { "epoch": 0.5866050808314087, "grad_norm": 1.7370929718017578, "learning_rate": 7.701929153249808e-06, "loss": 0.9978, "step": 3048 }, { "epoch": 0.5867975365665897, "grad_norm": 1.7096227407455444, "learning_rate": 7.695863108622208e-06, "loss": 0.9831, "step": 3049 }, { "epoch": 0.5869899923017706, "grad_norm": 1.721142053604126, "learning_rate": 7.689797959249984e-06, "loss": 0.9259, "step": 3050 }, { "epoch": 0.5871824480369515, "grad_norm": 1.7602863311767578, "learning_rate": 7.6837337074897e-06, "loss": 0.8795, "step": 3051 }, { "epoch": 0.5873749037721324, "grad_norm": 1.760566234588623, "learning_rate": 7.677670355697577e-06, "loss": 0.8368, "step": 3052 }, { "epoch": 0.5875673595073133, "grad_norm": 1.8040025234222412, "learning_rate": 7.671607906229488e-06, "loss": 0.8972, "step": 3053 }, { "epoch": 0.5877598152424942, "grad_norm": 1.6719255447387695, "learning_rate": 7.66554636144095e-06, "loss": 0.9486, "step": 3054 }, { "epoch": 0.5879522709776751, "grad_norm": 1.6952571868896484, "learning_rate": 7.659485723687131e-06, "loss": 0.9398, "step": 3055 }, { "epoch": 0.588144726712856, "grad_norm": 1.696886420249939, "learning_rate": 7.653425995322852e-06, "loss": 0.9093, "step": 3056 }, { "epoch": 0.5883371824480369, "grad_norm": 1.662434697151184, "learning_rate": 7.647367178702568e-06, "loss": 0.9904, "step": 3057 }, { "epoch": 0.5885296381832179, "grad_norm": 1.7197728157043457, "learning_rate": 7.641309276180394e-06, "loss": 0.984, "step": 3058 }, { "epoch": 0.5887220939183988, "grad_norm": 1.723658561706543, "learning_rate": 7.635252290110082e-06, "loss": 0.9129, "step": 3059 }, { "epoch": 0.5889145496535797, "grad_norm": 1.7255167961120605, "learning_rate": 7.629196222845027e-06, "loss": 0.9964, "step": 3060 }, { "epoch": 0.5891070053887606, "grad_norm": 1.5792759656906128, "learning_rate": 7.623141076738271e-06, "loss": 0.8868, "step": 3061 }, { "epoch": 0.5892994611239415, "grad_norm": 1.6306661367416382, "learning_rate": 7.617086854142498e-06, "loss": 0.9943, "step": 3062 }, { "epoch": 0.5894919168591224, "grad_norm": 1.8734837770462036, "learning_rate": 7.611033557410027e-06, "loss": 0.9348, "step": 3063 }, { "epoch": 0.5896843725943033, "grad_norm": 1.626111388206482, "learning_rate": 7.6049811888928235e-06, "loss": 0.8825, "step": 3064 }, { "epoch": 0.5898768283294842, "grad_norm": 1.838313102722168, "learning_rate": 7.598929750942492e-06, "loss": 0.9636, "step": 3065 }, { "epoch": 0.5900692840646651, "grad_norm": 2.4128808975219727, "learning_rate": 7.592879245910273e-06, "loss": 0.938, "step": 3066 }, { "epoch": 0.5902617397998461, "grad_norm": 1.6248093843460083, "learning_rate": 7.586829676147045e-06, "loss": 1.0066, "step": 3067 }, { "epoch": 0.590454195535027, "grad_norm": 1.8999649286270142, "learning_rate": 7.580781044003324e-06, "loss": 1.0448, "step": 3068 }, { "epoch": 0.5906466512702079, "grad_norm": 1.7558555603027344, "learning_rate": 7.574733351829263e-06, "loss": 0.9495, "step": 3069 }, { "epoch": 0.5908391070053888, "grad_norm": 1.840142846107483, "learning_rate": 7.5686866019746444e-06, "loss": 0.9618, "step": 3070 }, { "epoch": 0.5910315627405697, "grad_norm": 1.6699053049087524, "learning_rate": 7.562640796788893e-06, "loss": 0.9217, "step": 3071 }, { "epoch": 0.5912240184757506, "grad_norm": 1.6873536109924316, "learning_rate": 7.556595938621058e-06, "loss": 0.9419, "step": 3072 }, { "epoch": 0.5914164742109315, "grad_norm": 1.770682692527771, "learning_rate": 7.550552029819828e-06, "loss": 1.0167, "step": 3073 }, { "epoch": 0.5916089299461124, "grad_norm": 0.999251127243042, "learning_rate": 7.544509072733515e-06, "loss": 1.0401, "step": 3074 }, { "epoch": 0.5918013856812933, "grad_norm": 1.9162050485610962, "learning_rate": 7.53846706971007e-06, "loss": 0.9612, "step": 3075 }, { "epoch": 0.5919938414164743, "grad_norm": 1.8191437721252441, "learning_rate": 7.532426023097063e-06, "loss": 1.026, "step": 3076 }, { "epoch": 0.5921862971516552, "grad_norm": 1.6889057159423828, "learning_rate": 7.5263859352417025e-06, "loss": 1.0062, "step": 3077 }, { "epoch": 0.5923787528868361, "grad_norm": 0.8744052648544312, "learning_rate": 7.5203468084908185e-06, "loss": 0.8905, "step": 3078 }, { "epoch": 0.592571208622017, "grad_norm": 1.7647967338562012, "learning_rate": 7.514308645190867e-06, "loss": 0.8939, "step": 3079 }, { "epoch": 0.5927636643571979, "grad_norm": 1.8856765031814575, "learning_rate": 7.508271447687936e-06, "loss": 0.8917, "step": 3080 }, { "epoch": 0.5929561200923787, "grad_norm": 1.730196475982666, "learning_rate": 7.50223521832773e-06, "loss": 0.8386, "step": 3081 }, { "epoch": 0.5931485758275596, "grad_norm": 2.005169153213501, "learning_rate": 7.496199959455584e-06, "loss": 0.9225, "step": 3082 }, { "epoch": 0.5933410315627405, "grad_norm": 1.6016221046447754, "learning_rate": 7.490165673416455e-06, "loss": 0.9597, "step": 3083 }, { "epoch": 0.5935334872979214, "grad_norm": 0.8668866157531738, "learning_rate": 7.484132362554915e-06, "loss": 0.8897, "step": 3084 }, { "epoch": 0.5937259430331023, "grad_norm": 2.025111198425293, "learning_rate": 7.478100029215166e-06, "loss": 0.9786, "step": 3085 }, { "epoch": 0.5939183987682833, "grad_norm": 1.7242189645767212, "learning_rate": 7.472068675741024e-06, "loss": 0.9281, "step": 3086 }, { "epoch": 0.5941108545034642, "grad_norm": 1.7397632598876953, "learning_rate": 7.466038304475931e-06, "loss": 0.8382, "step": 3087 }, { "epoch": 0.5943033102386451, "grad_norm": 1.8191649913787842, "learning_rate": 7.4600089177629384e-06, "loss": 0.9869, "step": 3088 }, { "epoch": 0.594495765973826, "grad_norm": 1.7557562589645386, "learning_rate": 7.4539805179447234e-06, "loss": 0.9285, "step": 3089 }, { "epoch": 0.5946882217090069, "grad_norm": 1.7238134145736694, "learning_rate": 7.447953107363574e-06, "loss": 0.9649, "step": 3090 }, { "epoch": 0.5948806774441878, "grad_norm": 0.9079597592353821, "learning_rate": 7.4419266883614e-06, "loss": 0.9301, "step": 3091 }, { "epoch": 0.5950731331793687, "grad_norm": 0.8732458353042603, "learning_rate": 7.435901263279717e-06, "loss": 0.922, "step": 3092 }, { "epoch": 0.5952655889145496, "grad_norm": 1.6692142486572266, "learning_rate": 7.429876834459661e-06, "loss": 0.94, "step": 3093 }, { "epoch": 0.5954580446497305, "grad_norm": 1.7725183963775635, "learning_rate": 7.4238534042419805e-06, "loss": 0.9622, "step": 3094 }, { "epoch": 0.5956505003849115, "grad_norm": 1.8413903713226318, "learning_rate": 7.417830974967034e-06, "loss": 0.9298, "step": 3095 }, { "epoch": 0.5958429561200924, "grad_norm": 1.704624891281128, "learning_rate": 7.411809548974792e-06, "loss": 0.9267, "step": 3096 }, { "epoch": 0.5960354118552733, "grad_norm": 1.822956919670105, "learning_rate": 7.4057891286048356e-06, "loss": 0.9479, "step": 3097 }, { "epoch": 0.5962278675904542, "grad_norm": 1.774893879890442, "learning_rate": 7.399769716196353e-06, "loss": 1.0565, "step": 3098 }, { "epoch": 0.5964203233256351, "grad_norm": 1.6078121662139893, "learning_rate": 7.393751314088147e-06, "loss": 0.8112, "step": 3099 }, { "epoch": 0.596612779060816, "grad_norm": 1.6429210901260376, "learning_rate": 7.387733924618617e-06, "loss": 0.9288, "step": 3100 }, { "epoch": 0.5968052347959969, "grad_norm": 0.9268057942390442, "learning_rate": 7.38171755012578e-06, "loss": 0.9092, "step": 3101 }, { "epoch": 0.5969976905311778, "grad_norm": 1.7914083003997803, "learning_rate": 7.3757021929472495e-06, "loss": 0.9872, "step": 3102 }, { "epoch": 0.5971901462663587, "grad_norm": 1.7300490140914917, "learning_rate": 7.3696878554202525e-06, "loss": 0.958, "step": 3103 }, { "epoch": 0.5973826020015397, "grad_norm": 1.9645992517471313, "learning_rate": 7.3636745398816135e-06, "loss": 0.9569, "step": 3104 }, { "epoch": 0.5975750577367206, "grad_norm": 1.6007357835769653, "learning_rate": 7.35766224866776e-06, "loss": 0.9324, "step": 3105 }, { "epoch": 0.5977675134719015, "grad_norm": 1.7044273614883423, "learning_rate": 7.3516509841147276e-06, "loss": 1.0158, "step": 3106 }, { "epoch": 0.5979599692070824, "grad_norm": 1.6432082653045654, "learning_rate": 7.345640748558146e-06, "loss": 0.8864, "step": 3107 }, { "epoch": 0.5981524249422633, "grad_norm": 1.6368439197540283, "learning_rate": 7.33963154433325e-06, "loss": 0.8969, "step": 3108 }, { "epoch": 0.5983448806774442, "grad_norm": 1.609157919883728, "learning_rate": 7.333623373774868e-06, "loss": 0.934, "step": 3109 }, { "epoch": 0.5985373364126251, "grad_norm": 1.769007921218872, "learning_rate": 7.327616239217432e-06, "loss": 0.941, "step": 3110 }, { "epoch": 0.598729792147806, "grad_norm": 1.7063121795654297, "learning_rate": 7.321610142994971e-06, "loss": 0.9605, "step": 3111 }, { "epoch": 0.5989222478829869, "grad_norm": 1.651293396949768, "learning_rate": 7.315605087441107e-06, "loss": 0.9175, "step": 3112 }, { "epoch": 0.5991147036181679, "grad_norm": 1.733529806137085, "learning_rate": 7.309601074889061e-06, "loss": 0.9669, "step": 3113 }, { "epoch": 0.5993071593533488, "grad_norm": 1.7674496173858643, "learning_rate": 7.303598107671648e-06, "loss": 0.9292, "step": 3114 }, { "epoch": 0.5994996150885297, "grad_norm": 2.0784335136413574, "learning_rate": 7.297596188121277e-06, "loss": 0.8914, "step": 3115 }, { "epoch": 0.5996920708237106, "grad_norm": 1.8785899877548218, "learning_rate": 7.291595318569951e-06, "loss": 0.9318, "step": 3116 }, { "epoch": 0.5998845265588915, "grad_norm": 1.766610860824585, "learning_rate": 7.285595501349259e-06, "loss": 0.9053, "step": 3117 }, { "epoch": 0.6000769822940724, "grad_norm": 1.7902711629867554, "learning_rate": 7.279596738790389e-06, "loss": 0.8586, "step": 3118 }, { "epoch": 0.6002694380292533, "grad_norm": 1.8599613904953003, "learning_rate": 7.273599033224114e-06, "loss": 0.9097, "step": 3119 }, { "epoch": 0.6004618937644342, "grad_norm": 1.5772086381912231, "learning_rate": 7.267602386980801e-06, "loss": 0.9444, "step": 3120 }, { "epoch": 0.600654349499615, "grad_norm": 1.978500247001648, "learning_rate": 7.2616068023904e-06, "loss": 0.9465, "step": 3121 }, { "epoch": 0.6008468052347959, "grad_norm": 1.760067105293274, "learning_rate": 7.255612281782451e-06, "loss": 0.9141, "step": 3122 }, { "epoch": 0.601039260969977, "grad_norm": 0.8957152366638184, "learning_rate": 7.249618827486083e-06, "loss": 0.8587, "step": 3123 }, { "epoch": 0.6012317167051578, "grad_norm": 1.9353101253509521, "learning_rate": 7.243626441830009e-06, "loss": 0.9806, "step": 3124 }, { "epoch": 0.6014241724403387, "grad_norm": 1.6308897733688354, "learning_rate": 7.2376351271425235e-06, "loss": 0.8857, "step": 3125 }, { "epoch": 0.6016166281755196, "grad_norm": 1.7076072692871094, "learning_rate": 7.2316448857515076e-06, "loss": 0.9185, "step": 3126 }, { "epoch": 0.6018090839107005, "grad_norm": 1.5567580461502075, "learning_rate": 7.225655719984426e-06, "loss": 0.8813, "step": 3127 }, { "epoch": 0.6020015396458814, "grad_norm": 1.7650173902511597, "learning_rate": 7.219667632168326e-06, "loss": 0.8705, "step": 3128 }, { "epoch": 0.6021939953810623, "grad_norm": 1.7719755172729492, "learning_rate": 7.213680624629832e-06, "loss": 0.954, "step": 3129 }, { "epoch": 0.6023864511162432, "grad_norm": 1.8033884763717651, "learning_rate": 7.207694699695155e-06, "loss": 0.9326, "step": 3130 }, { "epoch": 0.6025789068514241, "grad_norm": 1.782778263092041, "learning_rate": 7.201709859690081e-06, "loss": 0.9477, "step": 3131 }, { "epoch": 0.6027713625866051, "grad_norm": 1.8047493696212769, "learning_rate": 7.1957261069399745e-06, "loss": 0.8708, "step": 3132 }, { "epoch": 0.602963818321786, "grad_norm": 1.7807023525238037, "learning_rate": 7.189743443769778e-06, "loss": 0.9409, "step": 3133 }, { "epoch": 0.6031562740569669, "grad_norm": 1.9129821062088013, "learning_rate": 7.183761872504011e-06, "loss": 0.9304, "step": 3134 }, { "epoch": 0.6033487297921478, "grad_norm": 1.626101016998291, "learning_rate": 7.1777813954667716e-06, "loss": 0.921, "step": 3135 }, { "epoch": 0.6035411855273287, "grad_norm": 1.9588425159454346, "learning_rate": 7.171802014981726e-06, "loss": 0.936, "step": 3136 }, { "epoch": 0.6037336412625096, "grad_norm": 2.0529778003692627, "learning_rate": 7.1658237333721216e-06, "loss": 0.9631, "step": 3137 }, { "epoch": 0.6039260969976905, "grad_norm": 0.8956499695777893, "learning_rate": 7.159846552960774e-06, "loss": 0.9019, "step": 3138 }, { "epoch": 0.6041185527328714, "grad_norm": 1.7404218912124634, "learning_rate": 7.153870476070073e-06, "loss": 0.9415, "step": 3139 }, { "epoch": 0.6043110084680523, "grad_norm": 1.6268503665924072, "learning_rate": 7.14789550502198e-06, "loss": 0.9084, "step": 3140 }, { "epoch": 0.6045034642032333, "grad_norm": 1.7708336114883423, "learning_rate": 7.141921642138025e-06, "loss": 1.0231, "step": 3141 }, { "epoch": 0.6046959199384142, "grad_norm": 1.707088828086853, "learning_rate": 7.135948889739307e-06, "loss": 0.9695, "step": 3142 }, { "epoch": 0.6048883756735951, "grad_norm": 1.6896693706512451, "learning_rate": 7.129977250146496e-06, "loss": 1.0124, "step": 3143 }, { "epoch": 0.605080831408776, "grad_norm": 1.6429632902145386, "learning_rate": 7.124006725679828e-06, "loss": 0.8634, "step": 3144 }, { "epoch": 0.6052732871439569, "grad_norm": 1.797355055809021, "learning_rate": 7.118037318659108e-06, "loss": 0.8911, "step": 3145 }, { "epoch": 0.6054657428791378, "grad_norm": 1.8120458126068115, "learning_rate": 7.112069031403704e-06, "loss": 0.8967, "step": 3146 }, { "epoch": 0.6056581986143187, "grad_norm": 1.693933129310608, "learning_rate": 7.106101866232546e-06, "loss": 0.9916, "step": 3147 }, { "epoch": 0.6058506543494996, "grad_norm": 1.604575276374817, "learning_rate": 7.100135825464138e-06, "loss": 0.9038, "step": 3148 }, { "epoch": 0.6060431100846805, "grad_norm": 1.89679753780365, "learning_rate": 7.094170911416543e-06, "loss": 0.9699, "step": 3149 }, { "epoch": 0.6062355658198614, "grad_norm": 1.655442476272583, "learning_rate": 7.088207126407374e-06, "loss": 0.9523, "step": 3150 }, { "epoch": 0.6064280215550424, "grad_norm": 1.8391051292419434, "learning_rate": 7.082244472753823e-06, "loss": 0.9908, "step": 3151 }, { "epoch": 0.6066204772902233, "grad_norm": 1.923933744430542, "learning_rate": 7.076282952772634e-06, "loss": 0.9494, "step": 3152 }, { "epoch": 0.6068129330254042, "grad_norm": 1.7745442390441895, "learning_rate": 7.070322568780111e-06, "loss": 0.9111, "step": 3153 }, { "epoch": 0.6070053887605851, "grad_norm": 1.6586029529571533, "learning_rate": 7.064363323092117e-06, "loss": 0.9303, "step": 3154 }, { "epoch": 0.607197844495766, "grad_norm": 1.7990206480026245, "learning_rate": 7.058405218024074e-06, "loss": 0.8972, "step": 3155 }, { "epoch": 0.6073903002309469, "grad_norm": 1.8297728300094604, "learning_rate": 7.052448255890958e-06, "loss": 0.9048, "step": 3156 }, { "epoch": 0.6075827559661278, "grad_norm": 1.7134735584259033, "learning_rate": 7.046492439007306e-06, "loss": 0.867, "step": 3157 }, { "epoch": 0.6077752117013087, "grad_norm": 1.734916090965271, "learning_rate": 7.040537769687204e-06, "loss": 0.9256, "step": 3158 }, { "epoch": 0.6079676674364896, "grad_norm": 1.8912674188613892, "learning_rate": 7.034584250244292e-06, "loss": 0.9796, "step": 3159 }, { "epoch": 0.6081601231716706, "grad_norm": 1.613027811050415, "learning_rate": 7.028631882991771e-06, "loss": 0.8988, "step": 3160 }, { "epoch": 0.6083525789068515, "grad_norm": 1.7163894176483154, "learning_rate": 7.022680670242387e-06, "loss": 0.9866, "step": 3161 }, { "epoch": 0.6085450346420324, "grad_norm": 0.9055702090263367, "learning_rate": 7.01673061430844e-06, "loss": 0.921, "step": 3162 }, { "epoch": 0.6087374903772133, "grad_norm": 1.6616817712783813, "learning_rate": 7.01078171750178e-06, "loss": 0.9702, "step": 3163 }, { "epoch": 0.6089299461123941, "grad_norm": 1.7192180156707764, "learning_rate": 7.004833982133808e-06, "loss": 1.0575, "step": 3164 }, { "epoch": 0.609122401847575, "grad_norm": 1.728763222694397, "learning_rate": 6.998887410515473e-06, "loss": 0.9602, "step": 3165 }, { "epoch": 0.6093148575827559, "grad_norm": 1.7535946369171143, "learning_rate": 6.992942004957271e-06, "loss": 0.9614, "step": 3166 }, { "epoch": 0.6095073133179368, "grad_norm": 2.1553993225097656, "learning_rate": 6.986997767769243e-06, "loss": 0.9441, "step": 3167 }, { "epoch": 0.6096997690531177, "grad_norm": 0.9410929083824158, "learning_rate": 6.981054701260981e-06, "loss": 0.9434, "step": 3168 }, { "epoch": 0.6098922247882987, "grad_norm": 1.6779158115386963, "learning_rate": 6.97511280774162e-06, "loss": 0.9381, "step": 3169 }, { "epoch": 0.6100846805234796, "grad_norm": 1.8352134227752686, "learning_rate": 6.969172089519836e-06, "loss": 0.852, "step": 3170 }, { "epoch": 0.6102771362586605, "grad_norm": 1.8777514696121216, "learning_rate": 6.963232548903853e-06, "loss": 0.9848, "step": 3171 }, { "epoch": 0.6104695919938414, "grad_norm": 1.7684664726257324, "learning_rate": 6.957294188201438e-06, "loss": 0.9384, "step": 3172 }, { "epoch": 0.6106620477290223, "grad_norm": 1.92123544216156, "learning_rate": 6.951357009719893e-06, "loss": 1.0142, "step": 3173 }, { "epoch": 0.6108545034642032, "grad_norm": 0.9332671761512756, "learning_rate": 6.9454210157660694e-06, "loss": 0.8899, "step": 3174 }, { "epoch": 0.6110469591993841, "grad_norm": 1.7205137014389038, "learning_rate": 6.939486208646349e-06, "loss": 0.9182, "step": 3175 }, { "epoch": 0.611239414934565, "grad_norm": 1.8418316841125488, "learning_rate": 6.933552590666659e-06, "loss": 0.9636, "step": 3176 }, { "epoch": 0.6114318706697459, "grad_norm": 1.724522590637207, "learning_rate": 6.927620164132463e-06, "loss": 0.8844, "step": 3177 }, { "epoch": 0.6116243264049269, "grad_norm": 1.9042211771011353, "learning_rate": 6.921688931348762e-06, "loss": 1.0362, "step": 3178 }, { "epoch": 0.6118167821401078, "grad_norm": 1.7500731945037842, "learning_rate": 6.915758894620093e-06, "loss": 0.9136, "step": 3179 }, { "epoch": 0.6120092378752887, "grad_norm": 1.6804414987564087, "learning_rate": 6.909830056250527e-06, "loss": 0.9258, "step": 3180 }, { "epoch": 0.6122016936104696, "grad_norm": 1.6197253465652466, "learning_rate": 6.903902418543671e-06, "loss": 0.9253, "step": 3181 }, { "epoch": 0.6123941493456505, "grad_norm": 1.7026760578155518, "learning_rate": 6.8979759838026676e-06, "loss": 0.9284, "step": 3182 }, { "epoch": 0.6125866050808314, "grad_norm": 2.2527170181274414, "learning_rate": 6.892050754330181e-06, "loss": 0.979, "step": 3183 }, { "epoch": 0.6127790608160123, "grad_norm": 1.6979202032089233, "learning_rate": 6.886126732428424e-06, "loss": 0.8549, "step": 3184 }, { "epoch": 0.6129715165511932, "grad_norm": 1.6735785007476807, "learning_rate": 6.8802039203991275e-06, "loss": 0.908, "step": 3185 }, { "epoch": 0.6131639722863741, "grad_norm": 1.6730470657348633, "learning_rate": 6.874282320543557e-06, "loss": 0.8493, "step": 3186 }, { "epoch": 0.613356428021555, "grad_norm": 1.8674790859222412, "learning_rate": 6.8683619351625065e-06, "loss": 0.8894, "step": 3187 }, { "epoch": 0.613548883756736, "grad_norm": 1.770477056503296, "learning_rate": 6.862442766556297e-06, "loss": 0.945, "step": 3188 }, { "epoch": 0.6137413394919169, "grad_norm": 1.7187254428863525, "learning_rate": 6.85652481702478e-06, "loss": 0.8622, "step": 3189 }, { "epoch": 0.6139337952270978, "grad_norm": 1.742129921913147, "learning_rate": 6.850608088867332e-06, "loss": 0.8889, "step": 3190 }, { "epoch": 0.6141262509622787, "grad_norm": 1.5799837112426758, "learning_rate": 6.844692584382848e-06, "loss": 0.9245, "step": 3191 }, { "epoch": 0.6143187066974596, "grad_norm": 1.6683019399642944, "learning_rate": 6.83877830586976e-06, "loss": 0.9512, "step": 3192 }, { "epoch": 0.6145111624326405, "grad_norm": 1.6474559307098389, "learning_rate": 6.832865255626013e-06, "loss": 0.9298, "step": 3193 }, { "epoch": 0.6147036181678214, "grad_norm": 1.6920115947723389, "learning_rate": 6.826953435949081e-06, "loss": 0.9155, "step": 3194 }, { "epoch": 0.6148960739030023, "grad_norm": 1.7854158878326416, "learning_rate": 6.821042849135957e-06, "loss": 0.9485, "step": 3195 }, { "epoch": 0.6150885296381832, "grad_norm": 1.7588282823562622, "learning_rate": 6.815133497483157e-06, "loss": 0.9636, "step": 3196 }, { "epoch": 0.6152809853733642, "grad_norm": 1.7492843866348267, "learning_rate": 6.809225383286715e-06, "loss": 0.9785, "step": 3197 }, { "epoch": 0.6154734411085451, "grad_norm": 1.8690872192382812, "learning_rate": 6.8033185088421874e-06, "loss": 0.8671, "step": 3198 }, { "epoch": 0.615665896843726, "grad_norm": 1.7724074125289917, "learning_rate": 6.797412876444645e-06, "loss": 0.9637, "step": 3199 }, { "epoch": 0.6158583525789069, "grad_norm": 1.7514842748641968, "learning_rate": 6.791508488388675e-06, "loss": 0.9561, "step": 3200 }, { "epoch": 0.6160508083140878, "grad_norm": 1.7365591526031494, "learning_rate": 6.785605346968387e-06, "loss": 1.0579, "step": 3201 }, { "epoch": 0.6162432640492687, "grad_norm": 1.7717989683151245, "learning_rate": 6.779703454477402e-06, "loss": 1.0581, "step": 3202 }, { "epoch": 0.6164357197844496, "grad_norm": 1.802849531173706, "learning_rate": 6.7738028132088565e-06, "loss": 0.9847, "step": 3203 }, { "epoch": 0.6166281755196305, "grad_norm": 1.9230589866638184, "learning_rate": 6.767903425455402e-06, "loss": 0.9406, "step": 3204 }, { "epoch": 0.6168206312548113, "grad_norm": 1.6573668718338013, "learning_rate": 6.762005293509202e-06, "loss": 0.921, "step": 3205 }, { "epoch": 0.6170130869899924, "grad_norm": 1.7774088382720947, "learning_rate": 6.7561084196619306e-06, "loss": 0.9151, "step": 3206 }, { "epoch": 0.6172055427251733, "grad_norm": 1.688793659210205, "learning_rate": 6.750212806204778e-06, "loss": 0.9419, "step": 3207 }, { "epoch": 0.6173979984603541, "grad_norm": 1.698472499847412, "learning_rate": 6.744318455428436e-06, "loss": 0.927, "step": 3208 }, { "epoch": 0.617590454195535, "grad_norm": 1.6036385297775269, "learning_rate": 6.738425369623115e-06, "loss": 0.9446, "step": 3209 }, { "epoch": 0.6177829099307159, "grad_norm": 1.7535037994384766, "learning_rate": 6.732533551078528e-06, "loss": 0.8567, "step": 3210 }, { "epoch": 0.6179753656658968, "grad_norm": 1.7296459674835205, "learning_rate": 6.7266430020839e-06, "loss": 0.8701, "step": 3211 }, { "epoch": 0.6181678214010777, "grad_norm": 1.8219283819198608, "learning_rate": 6.720753724927957e-06, "loss": 0.9682, "step": 3212 }, { "epoch": 0.6183602771362586, "grad_norm": 1.5691721439361572, "learning_rate": 6.714865721898938e-06, "loss": 0.942, "step": 3213 }, { "epoch": 0.6185527328714395, "grad_norm": 1.6300033330917358, "learning_rate": 6.708978995284581e-06, "loss": 0.9277, "step": 3214 }, { "epoch": 0.6187451886066204, "grad_norm": 1.5564969778060913, "learning_rate": 6.70309354737213e-06, "loss": 0.9816, "step": 3215 }, { "epoch": 0.6189376443418014, "grad_norm": 1.7144765853881836, "learning_rate": 6.697209380448333e-06, "loss": 0.9805, "step": 3216 }, { "epoch": 0.6191301000769823, "grad_norm": 1.7554031610488892, "learning_rate": 6.6913264967994375e-06, "loss": 1.0345, "step": 3217 }, { "epoch": 0.6193225558121632, "grad_norm": 1.7935832738876343, "learning_rate": 6.685444898711196e-06, "loss": 1.0069, "step": 3218 }, { "epoch": 0.6195150115473441, "grad_norm": 1.7465343475341797, "learning_rate": 6.679564588468861e-06, "loss": 0.9317, "step": 3219 }, { "epoch": 0.619707467282525, "grad_norm": 1.7861418724060059, "learning_rate": 6.673685568357182e-06, "loss": 0.8791, "step": 3220 }, { "epoch": 0.6198999230177059, "grad_norm": 1.6786564588546753, "learning_rate": 6.66780784066041e-06, "loss": 0.8567, "step": 3221 }, { "epoch": 0.6200923787528868, "grad_norm": 1.8464841842651367, "learning_rate": 6.661931407662292e-06, "loss": 0.9323, "step": 3222 }, { "epoch": 0.6202848344880677, "grad_norm": 1.9761865139007568, "learning_rate": 6.656056271646076e-06, "loss": 0.9152, "step": 3223 }, { "epoch": 0.6204772902232486, "grad_norm": 1.9605340957641602, "learning_rate": 6.650182434894496e-06, "loss": 0.9762, "step": 3224 }, { "epoch": 0.6206697459584296, "grad_norm": 1.7294394969940186, "learning_rate": 6.644309899689794e-06, "loss": 0.9904, "step": 3225 }, { "epoch": 0.6208622016936105, "grad_norm": 1.8550854921340942, "learning_rate": 6.638438668313695e-06, "loss": 0.9187, "step": 3226 }, { "epoch": 0.6210546574287914, "grad_norm": 1.7067784070968628, "learning_rate": 6.632568743047426e-06, "loss": 0.9901, "step": 3227 }, { "epoch": 0.6212471131639723, "grad_norm": 1.6767572164535522, "learning_rate": 6.6267001261717015e-06, "loss": 0.8918, "step": 3228 }, { "epoch": 0.6214395688991532, "grad_norm": 1.5912185907363892, "learning_rate": 6.6208328199667305e-06, "loss": 0.9238, "step": 3229 }, { "epoch": 0.6216320246343341, "grad_norm": 0.924150824546814, "learning_rate": 6.614966826712212e-06, "loss": 0.9186, "step": 3230 }, { "epoch": 0.621824480369515, "grad_norm": 1.8024675846099854, "learning_rate": 6.609102148687333e-06, "loss": 1.0614, "step": 3231 }, { "epoch": 0.6220169361046959, "grad_norm": 1.8041318655014038, "learning_rate": 6.603238788170771e-06, "loss": 0.9256, "step": 3232 }, { "epoch": 0.6222093918398768, "grad_norm": 1.6303967237472534, "learning_rate": 6.597376747440694e-06, "loss": 0.9962, "step": 3233 }, { "epoch": 0.6224018475750578, "grad_norm": 1.875462293624878, "learning_rate": 6.59151602877475e-06, "loss": 0.9883, "step": 3234 }, { "epoch": 0.6225943033102387, "grad_norm": 1.8030260801315308, "learning_rate": 6.5856566344500815e-06, "loss": 0.9089, "step": 3235 }, { "epoch": 0.6227867590454196, "grad_norm": 1.7845523357391357, "learning_rate": 6.579798566743314e-06, "loss": 0.9173, "step": 3236 }, { "epoch": 0.6229792147806005, "grad_norm": 1.8029476404190063, "learning_rate": 6.573941827930551e-06, "loss": 0.8976, "step": 3237 }, { "epoch": 0.6231716705157814, "grad_norm": 1.7241421937942505, "learning_rate": 6.568086420287389e-06, "loss": 0.9843, "step": 3238 }, { "epoch": 0.6233641262509623, "grad_norm": 1.724313497543335, "learning_rate": 6.562232346088903e-06, "loss": 0.9405, "step": 3239 }, { "epoch": 0.6235565819861432, "grad_norm": 1.8561499118804932, "learning_rate": 6.5563796076096484e-06, "loss": 0.9392, "step": 3240 }, { "epoch": 0.6237490377213241, "grad_norm": 1.733005404472351, "learning_rate": 6.550528207123667e-06, "loss": 0.9031, "step": 3241 }, { "epoch": 0.623941493456505, "grad_norm": 1.9871788024902344, "learning_rate": 6.544678146904475e-06, "loss": 0.955, "step": 3242 }, { "epoch": 0.624133949191686, "grad_norm": 1.6345582008361816, "learning_rate": 6.538829429225068e-06, "loss": 0.9295, "step": 3243 }, { "epoch": 0.6243264049268669, "grad_norm": 1.5734761953353882, "learning_rate": 6.532982056357928e-06, "loss": 0.8884, "step": 3244 }, { "epoch": 0.6245188606620478, "grad_norm": 1.8750501871109009, "learning_rate": 6.527136030575005e-06, "loss": 0.9633, "step": 3245 }, { "epoch": 0.6247113163972287, "grad_norm": 1.7924443483352661, "learning_rate": 6.521291354147727e-06, "loss": 0.9662, "step": 3246 }, { "epoch": 0.6249037721324096, "grad_norm": 0.9084703922271729, "learning_rate": 6.515448029347005e-06, "loss": 0.8778, "step": 3247 }, { "epoch": 0.6250962278675904, "grad_norm": 1.7733460664749146, "learning_rate": 6.5096060584432134e-06, "loss": 0.9595, "step": 3248 }, { "epoch": 0.6252886836027713, "grad_norm": 1.7360961437225342, "learning_rate": 6.503765443706212e-06, "loss": 0.8954, "step": 3249 }, { "epoch": 0.6254811393379522, "grad_norm": 1.8181066513061523, "learning_rate": 6.497926187405326e-06, "loss": 0.9448, "step": 3250 }, { "epoch": 0.6256735950731331, "grad_norm": 1.6166300773620605, "learning_rate": 6.492088291809355e-06, "loss": 0.8382, "step": 3251 }, { "epoch": 0.625866050808314, "grad_norm": 1.6044433116912842, "learning_rate": 6.486251759186573e-06, "loss": 0.9792, "step": 3252 }, { "epoch": 0.626058506543495, "grad_norm": 1.6275311708450317, "learning_rate": 6.48041659180472e-06, "loss": 0.9391, "step": 3253 }, { "epoch": 0.6262509622786759, "grad_norm": 1.8418091535568237, "learning_rate": 6.474582791931006e-06, "loss": 0.8718, "step": 3254 }, { "epoch": 0.6264434180138568, "grad_norm": 1.8523285388946533, "learning_rate": 6.468750361832112e-06, "loss": 0.9352, "step": 3255 }, { "epoch": 0.6266358737490377, "grad_norm": 1.8091304302215576, "learning_rate": 6.462919303774186e-06, "loss": 0.9476, "step": 3256 }, { "epoch": 0.6268283294842186, "grad_norm": 1.672059416770935, "learning_rate": 6.4570896200228415e-06, "loss": 0.9643, "step": 3257 }, { "epoch": 0.6270207852193995, "grad_norm": 1.7655494213104248, "learning_rate": 6.45126131284316e-06, "loss": 0.9651, "step": 3258 }, { "epoch": 0.6272132409545804, "grad_norm": 1.8237800598144531, "learning_rate": 6.445434384499687e-06, "loss": 0.9236, "step": 3259 }, { "epoch": 0.6274056966897613, "grad_norm": 1.8213757276535034, "learning_rate": 6.439608837256432e-06, "loss": 0.9937, "step": 3260 }, { "epoch": 0.6275981524249422, "grad_norm": 1.7549742460250854, "learning_rate": 6.43378467337687e-06, "loss": 0.9208, "step": 3261 }, { "epoch": 0.6277906081601232, "grad_norm": 1.5696029663085938, "learning_rate": 6.427961895123935e-06, "loss": 0.9533, "step": 3262 }, { "epoch": 0.6279830638953041, "grad_norm": 1.7207852602005005, "learning_rate": 6.422140504760025e-06, "loss": 0.9139, "step": 3263 }, { "epoch": 0.628175519630485, "grad_norm": 1.6829673051834106, "learning_rate": 6.4163205045469975e-06, "loss": 1.0153, "step": 3264 }, { "epoch": 0.6283679753656659, "grad_norm": 1.7586554288864136, "learning_rate": 6.410501896746172e-06, "loss": 0.9786, "step": 3265 }, { "epoch": 0.6285604311008468, "grad_norm": 2.7261741161346436, "learning_rate": 6.404684683618325e-06, "loss": 0.8109, "step": 3266 }, { "epoch": 0.6287528868360277, "grad_norm": 1.7530407905578613, "learning_rate": 6.398868867423691e-06, "loss": 0.9307, "step": 3267 }, { "epoch": 0.6289453425712086, "grad_norm": 1.6079076528549194, "learning_rate": 6.393054450421963e-06, "loss": 0.9091, "step": 3268 }, { "epoch": 0.6291377983063895, "grad_norm": 1.905810832977295, "learning_rate": 6.387241434872293e-06, "loss": 0.919, "step": 3269 }, { "epoch": 0.6293302540415704, "grad_norm": 1.6266740560531616, "learning_rate": 6.381429823033281e-06, "loss": 0.9685, "step": 3270 }, { "epoch": 0.6295227097767514, "grad_norm": 1.6186314821243286, "learning_rate": 6.375619617162985e-06, "loss": 0.9212, "step": 3271 }, { "epoch": 0.6297151655119323, "grad_norm": 1.7290632724761963, "learning_rate": 6.36981081951892e-06, "loss": 0.8993, "step": 3272 }, { "epoch": 0.6299076212471132, "grad_norm": 1.774246335029602, "learning_rate": 6.364003432358051e-06, "loss": 0.9613, "step": 3273 }, { "epoch": 0.6301000769822941, "grad_norm": 2.014669179916382, "learning_rate": 6.358197457936796e-06, "loss": 0.9688, "step": 3274 }, { "epoch": 0.630292532717475, "grad_norm": 1.4978059530258179, "learning_rate": 6.3523928985110216e-06, "loss": 0.9764, "step": 3275 }, { "epoch": 0.6304849884526559, "grad_norm": 1.8136916160583496, "learning_rate": 6.34658975633605e-06, "loss": 0.9232, "step": 3276 }, { "epoch": 0.6306774441878368, "grad_norm": 1.8452619314193726, "learning_rate": 6.340788033666645e-06, "loss": 1.005, "step": 3277 }, { "epoch": 0.6308698999230177, "grad_norm": 1.6856024265289307, "learning_rate": 6.334987732757028e-06, "loss": 0.965, "step": 3278 }, { "epoch": 0.6310623556581986, "grad_norm": 1.6409109830856323, "learning_rate": 6.329188855860859e-06, "loss": 0.9062, "step": 3279 }, { "epoch": 0.6312548113933796, "grad_norm": 1.7507151365280151, "learning_rate": 6.32339140523125e-06, "loss": 0.8677, "step": 3280 }, { "epoch": 0.6314472671285605, "grad_norm": 1.851258635520935, "learning_rate": 6.317595383120756e-06, "loss": 0.9892, "step": 3281 }, { "epoch": 0.6316397228637414, "grad_norm": 1.7153377532958984, "learning_rate": 6.311800791781381e-06, "loss": 0.8947, "step": 3282 }, { "epoch": 0.6318321785989223, "grad_norm": 1.9791998863220215, "learning_rate": 6.30600763346457e-06, "loss": 0.9181, "step": 3283 }, { "epoch": 0.6320246343341032, "grad_norm": 1.7142672538757324, "learning_rate": 6.300215910421212e-06, "loss": 0.8914, "step": 3284 }, { "epoch": 0.6322170900692841, "grad_norm": 1.868776559829712, "learning_rate": 6.294425624901638e-06, "loss": 1.012, "step": 3285 }, { "epoch": 0.632409545804465, "grad_norm": 1.4744089841842651, "learning_rate": 6.288636779155621e-06, "loss": 0.8557, "step": 3286 }, { "epoch": 0.6326020015396459, "grad_norm": 1.9101513624191284, "learning_rate": 6.282849375432374e-06, "loss": 0.891, "step": 3287 }, { "epoch": 0.6327944572748267, "grad_norm": 1.896141529083252, "learning_rate": 6.277063415980549e-06, "loss": 0.8874, "step": 3288 }, { "epoch": 0.6329869130100076, "grad_norm": 0.8723871111869812, "learning_rate": 6.271278903048239e-06, "loss": 0.8807, "step": 3289 }, { "epoch": 0.6331793687451887, "grad_norm": 1.8348798751831055, "learning_rate": 6.265495838882973e-06, "loss": 0.8804, "step": 3290 }, { "epoch": 0.6333718244803695, "grad_norm": 1.8826345205307007, "learning_rate": 6.2597142257317185e-06, "loss": 1.0014, "step": 3291 }, { "epoch": 0.6335642802155504, "grad_norm": 1.7536097764968872, "learning_rate": 6.25393406584088e-06, "loss": 0.8717, "step": 3292 }, { "epoch": 0.6337567359507313, "grad_norm": 1.7525378465652466, "learning_rate": 6.248155361456293e-06, "loss": 0.9102, "step": 3293 }, { "epoch": 0.6339491916859122, "grad_norm": 1.6756343841552734, "learning_rate": 6.2423781148232345e-06, "loss": 0.9505, "step": 3294 }, { "epoch": 0.6341416474210931, "grad_norm": 1.8366891145706177, "learning_rate": 6.236602328186405e-06, "loss": 0.9431, "step": 3295 }, { "epoch": 0.634334103156274, "grad_norm": 1.644197702407837, "learning_rate": 6.230828003789949e-06, "loss": 1.0018, "step": 3296 }, { "epoch": 0.6345265588914549, "grad_norm": 1.8687825202941895, "learning_rate": 6.225055143877434e-06, "loss": 0.9748, "step": 3297 }, { "epoch": 0.6347190146266358, "grad_norm": 1.742172122001648, "learning_rate": 6.219283750691864e-06, "loss": 0.8774, "step": 3298 }, { "epoch": 0.6349114703618168, "grad_norm": 1.6542288064956665, "learning_rate": 6.2135138264756715e-06, "loss": 0.9397, "step": 3299 }, { "epoch": 0.6351039260969977, "grad_norm": 1.6336995363235474, "learning_rate": 6.207745373470717e-06, "loss": 0.9293, "step": 3300 }, { "epoch": 0.6352963818321786, "grad_norm": 1.5556223392486572, "learning_rate": 6.201978393918291e-06, "loss": 0.9652, "step": 3301 }, { "epoch": 0.6354888375673595, "grad_norm": 1.7376117706298828, "learning_rate": 6.196212890059111e-06, "loss": 0.9365, "step": 3302 }, { "epoch": 0.6356812933025404, "grad_norm": 1.5776464939117432, "learning_rate": 6.190448864133321e-06, "loss": 0.9605, "step": 3303 }, { "epoch": 0.6358737490377213, "grad_norm": 1.6487114429473877, "learning_rate": 6.184686318380488e-06, "loss": 0.9696, "step": 3304 }, { "epoch": 0.6360662047729022, "grad_norm": 1.8035367727279663, "learning_rate": 6.17892525503961e-06, "loss": 0.9694, "step": 3305 }, { "epoch": 0.6362586605080831, "grad_norm": 1.9537098407745361, "learning_rate": 6.173165676349103e-06, "loss": 0.9965, "step": 3306 }, { "epoch": 0.636451116243264, "grad_norm": 1.9025707244873047, "learning_rate": 6.1674075845468114e-06, "loss": 0.9586, "step": 3307 }, { "epoch": 0.636643571978445, "grad_norm": 1.6091078519821167, "learning_rate": 6.1616509818699975e-06, "loss": 0.9575, "step": 3308 }, { "epoch": 0.6368360277136259, "grad_norm": 1.9612736701965332, "learning_rate": 6.155895870555348e-06, "loss": 0.8387, "step": 3309 }, { "epoch": 0.6370284834488068, "grad_norm": 1.8138504028320312, "learning_rate": 6.150142252838967e-06, "loss": 0.9071, "step": 3310 }, { "epoch": 0.6372209391839877, "grad_norm": 1.621180772781372, "learning_rate": 6.144390130956384e-06, "loss": 0.914, "step": 3311 }, { "epoch": 0.6374133949191686, "grad_norm": 1.6711097955703735, "learning_rate": 6.138639507142539e-06, "loss": 0.8889, "step": 3312 }, { "epoch": 0.6376058506543495, "grad_norm": 1.8578338623046875, "learning_rate": 6.132890383631796e-06, "loss": 0.9445, "step": 3313 }, { "epoch": 0.6377983063895304, "grad_norm": 1.7074776887893677, "learning_rate": 6.127142762657934e-06, "loss": 0.8762, "step": 3314 }, { "epoch": 0.6379907621247113, "grad_norm": 1.8341341018676758, "learning_rate": 6.12139664645415e-06, "loss": 1.0149, "step": 3315 }, { "epoch": 0.6381832178598922, "grad_norm": 1.7302956581115723, "learning_rate": 6.115652037253054e-06, "loss": 0.8575, "step": 3316 }, { "epoch": 0.6383756735950731, "grad_norm": 1.8176337480545044, "learning_rate": 6.109908937286673e-06, "loss": 0.9969, "step": 3317 }, { "epoch": 0.6385681293302541, "grad_norm": 1.781009316444397, "learning_rate": 6.104167348786443e-06, "loss": 0.9672, "step": 3318 }, { "epoch": 0.638760585065435, "grad_norm": 1.7821024656295776, "learning_rate": 6.098427273983221e-06, "loss": 0.9859, "step": 3319 }, { "epoch": 0.6389530408006159, "grad_norm": 1.656535267829895, "learning_rate": 6.092688715107265e-06, "loss": 0.9349, "step": 3320 }, { "epoch": 0.6391454965357968, "grad_norm": 1.8177406787872314, "learning_rate": 6.086951674388252e-06, "loss": 0.9496, "step": 3321 }, { "epoch": 0.6393379522709777, "grad_norm": 1.6509474515914917, "learning_rate": 6.0812161540552675e-06, "loss": 0.9101, "step": 3322 }, { "epoch": 0.6395304080061586, "grad_norm": 1.65707266330719, "learning_rate": 6.075482156336804e-06, "loss": 0.8917, "step": 3323 }, { "epoch": 0.6397228637413395, "grad_norm": 1.9060357809066772, "learning_rate": 6.069749683460765e-06, "loss": 0.9785, "step": 3324 }, { "epoch": 0.6399153194765204, "grad_norm": 1.6193355321884155, "learning_rate": 6.06401873765446e-06, "loss": 0.929, "step": 3325 }, { "epoch": 0.6401077752117013, "grad_norm": 1.6463682651519775, "learning_rate": 6.058289321144608e-06, "loss": 0.932, "step": 3326 }, { "epoch": 0.6403002309468823, "grad_norm": 1.9387314319610596, "learning_rate": 6.052561436157329e-06, "loss": 0.9641, "step": 3327 }, { "epoch": 0.6404926866820632, "grad_norm": 1.7617419958114624, "learning_rate": 6.046835084918152e-06, "loss": 0.9367, "step": 3328 }, { "epoch": 0.6406851424172441, "grad_norm": 1.7443351745605469, "learning_rate": 6.041110269652005e-06, "loss": 0.9265, "step": 3329 }, { "epoch": 0.640877598152425, "grad_norm": 1.7621960639953613, "learning_rate": 6.0353869925832255e-06, "loss": 0.9322, "step": 3330 }, { "epoch": 0.6410700538876059, "grad_norm": 1.7700227499008179, "learning_rate": 6.02966525593555e-06, "loss": 0.9347, "step": 3331 }, { "epoch": 0.6412625096227867, "grad_norm": 1.686730146408081, "learning_rate": 6.023945061932119e-06, "loss": 0.968, "step": 3332 }, { "epoch": 0.6414549653579676, "grad_norm": 0.9725883603096008, "learning_rate": 6.018226412795469e-06, "loss": 0.8927, "step": 3333 }, { "epoch": 0.6416474210931485, "grad_norm": 0.9579965472221375, "learning_rate": 6.0125093107475385e-06, "loss": 0.9297, "step": 3334 }, { "epoch": 0.6418398768283294, "grad_norm": 2.398865222930908, "learning_rate": 6.00679375800967e-06, "loss": 0.9463, "step": 3335 }, { "epoch": 0.6420323325635104, "grad_norm": 1.8209846019744873, "learning_rate": 6.001079756802592e-06, "loss": 0.899, "step": 3336 }, { "epoch": 0.6422247882986913, "grad_norm": 1.723073124885559, "learning_rate": 5.9953673093464425e-06, "loss": 0.9918, "step": 3337 }, { "epoch": 0.6424172440338722, "grad_norm": 1.6836910247802734, "learning_rate": 5.98965641786075e-06, "loss": 0.9614, "step": 3338 }, { "epoch": 0.6426096997690531, "grad_norm": 0.8900152444839478, "learning_rate": 5.983947084564439e-06, "loss": 0.9084, "step": 3339 }, { "epoch": 0.642802155504234, "grad_norm": 1.7993860244750977, "learning_rate": 5.978239311675826e-06, "loss": 0.9511, "step": 3340 }, { "epoch": 0.6429946112394149, "grad_norm": 1.782443881034851, "learning_rate": 5.97253310141263e-06, "loss": 1.0083, "step": 3341 }, { "epoch": 0.6431870669745958, "grad_norm": 1.8219255208969116, "learning_rate": 5.966828455991951e-06, "loss": 0.9579, "step": 3342 }, { "epoch": 0.6433795227097767, "grad_norm": 0.9101649522781372, "learning_rate": 5.96112537763029e-06, "loss": 0.9017, "step": 3343 }, { "epoch": 0.6435719784449576, "grad_norm": 1.7730344533920288, "learning_rate": 5.955423868543537e-06, "loss": 0.937, "step": 3344 }, { "epoch": 0.6437644341801386, "grad_norm": 1.6564887762069702, "learning_rate": 5.949723930946967e-06, "loss": 0.9192, "step": 3345 }, { "epoch": 0.6439568899153195, "grad_norm": 1.6564580202102661, "learning_rate": 5.944025567055251e-06, "loss": 0.9262, "step": 3346 }, { "epoch": 0.6441493456505004, "grad_norm": 1.8470607995986938, "learning_rate": 5.9383287790824475e-06, "loss": 0.8811, "step": 3347 }, { "epoch": 0.6443418013856813, "grad_norm": 1.7679998874664307, "learning_rate": 5.932633569242e-06, "loss": 0.912, "step": 3348 }, { "epoch": 0.6445342571208622, "grad_norm": 1.8303871154785156, "learning_rate": 5.92693993974674e-06, "loss": 0.9382, "step": 3349 }, { "epoch": 0.6447267128560431, "grad_norm": 1.7000925540924072, "learning_rate": 5.921247892808886e-06, "loss": 0.9311, "step": 3350 }, { "epoch": 0.644919168591224, "grad_norm": 1.7522213459014893, "learning_rate": 5.91555743064004e-06, "loss": 0.8935, "step": 3351 }, { "epoch": 0.6451116243264049, "grad_norm": 1.90994393825531, "learning_rate": 5.909868555451191e-06, "loss": 0.9493, "step": 3352 }, { "epoch": 0.6453040800615858, "grad_norm": 1.800954818725586, "learning_rate": 5.904181269452707e-06, "loss": 0.8049, "step": 3353 }, { "epoch": 0.6454965357967667, "grad_norm": 1.698570966720581, "learning_rate": 5.898495574854342e-06, "loss": 0.8456, "step": 3354 }, { "epoch": 0.6456889915319477, "grad_norm": 1.7127857208251953, "learning_rate": 5.89281147386523e-06, "loss": 0.9968, "step": 3355 }, { "epoch": 0.6458814472671286, "grad_norm": 1.8562018871307373, "learning_rate": 5.887128968693887e-06, "loss": 1.0011, "step": 3356 }, { "epoch": 0.6460739030023095, "grad_norm": 1.7820154428482056, "learning_rate": 5.881448061548208e-06, "loss": 0.9383, "step": 3357 }, { "epoch": 0.6462663587374904, "grad_norm": 1.6744425296783447, "learning_rate": 5.875768754635469e-06, "loss": 0.9287, "step": 3358 }, { "epoch": 0.6464588144726713, "grad_norm": 0.9217410087585449, "learning_rate": 5.870091050162321e-06, "loss": 0.9465, "step": 3359 }, { "epoch": 0.6466512702078522, "grad_norm": 1.6345635652542114, "learning_rate": 5.864414950334796e-06, "loss": 0.9189, "step": 3360 }, { "epoch": 0.6468437259430331, "grad_norm": 1.796751856803894, "learning_rate": 5.858740457358298e-06, "loss": 0.8817, "step": 3361 }, { "epoch": 0.647036181678214, "grad_norm": 1.690312385559082, "learning_rate": 5.853067573437612e-06, "loss": 0.9852, "step": 3362 }, { "epoch": 0.6472286374133949, "grad_norm": 1.905022382736206, "learning_rate": 5.847396300776893e-06, "loss": 0.9715, "step": 3363 }, { "epoch": 0.6474210931485759, "grad_norm": 1.7262775897979736, "learning_rate": 5.8417266415796745e-06, "loss": 0.909, "step": 3364 }, { "epoch": 0.6476135488837568, "grad_norm": 1.5987921953201294, "learning_rate": 5.836058598048859e-06, "loss": 0.9418, "step": 3365 }, { "epoch": 0.6478060046189377, "grad_norm": 1.8148651123046875, "learning_rate": 5.830392172386723e-06, "loss": 0.994, "step": 3366 }, { "epoch": 0.6479984603541186, "grad_norm": 1.8683109283447266, "learning_rate": 5.824727366794915e-06, "loss": 0.9567, "step": 3367 }, { "epoch": 0.6481909160892995, "grad_norm": 1.8322832584381104, "learning_rate": 5.819064183474451e-06, "loss": 0.952, "step": 3368 }, { "epoch": 0.6483833718244804, "grad_norm": 0.9256272315979004, "learning_rate": 5.813402624625722e-06, "loss": 0.9221, "step": 3369 }, { "epoch": 0.6485758275596613, "grad_norm": 0.9572224617004395, "learning_rate": 5.807742692448483e-06, "loss": 0.885, "step": 3370 }, { "epoch": 0.6487682832948422, "grad_norm": 1.7511030435562134, "learning_rate": 5.802084389141862e-06, "loss": 0.8635, "step": 3371 }, { "epoch": 0.648960739030023, "grad_norm": 1.8171534538269043, "learning_rate": 5.796427716904347e-06, "loss": 0.957, "step": 3372 }, { "epoch": 0.6491531947652041, "grad_norm": 2.2900426387786865, "learning_rate": 5.790772677933799e-06, "loss": 0.9543, "step": 3373 }, { "epoch": 0.649345650500385, "grad_norm": 1.926742672920227, "learning_rate": 5.78511927442744e-06, "loss": 0.9279, "step": 3374 }, { "epoch": 0.6495381062355658, "grad_norm": 1.7962862253189087, "learning_rate": 5.77946750858186e-06, "loss": 0.998, "step": 3375 }, { "epoch": 0.6497305619707467, "grad_norm": 1.8198928833007812, "learning_rate": 5.773817382593008e-06, "loss": 0.9234, "step": 3376 }, { "epoch": 0.6499230177059276, "grad_norm": 1.6461323499679565, "learning_rate": 5.7681688986562e-06, "loss": 0.8556, "step": 3377 }, { "epoch": 0.6501154734411085, "grad_norm": 1.643263816833496, "learning_rate": 5.7625220589661136e-06, "loss": 0.8976, "step": 3378 }, { "epoch": 0.6503079291762894, "grad_norm": 1.7362608909606934, "learning_rate": 5.756876865716787e-06, "loss": 0.8657, "step": 3379 }, { "epoch": 0.6505003849114703, "grad_norm": 1.6694568395614624, "learning_rate": 5.751233321101617e-06, "loss": 0.9473, "step": 3380 }, { "epoch": 0.6506928406466512, "grad_norm": 1.5742807388305664, "learning_rate": 5.745591427313365e-06, "loss": 0.9171, "step": 3381 }, { "epoch": 0.6508852963818321, "grad_norm": 1.7873681783676147, "learning_rate": 5.739951186544141e-06, "loss": 0.9024, "step": 3382 }, { "epoch": 0.6510777521170131, "grad_norm": 1.6715519428253174, "learning_rate": 5.7343126009854215e-06, "loss": 0.982, "step": 3383 }, { "epoch": 0.651270207852194, "grad_norm": 1.7497960329055786, "learning_rate": 5.728675672828037e-06, "loss": 0.8946, "step": 3384 }, { "epoch": 0.6514626635873749, "grad_norm": 1.5573481321334839, "learning_rate": 5.72304040426217e-06, "loss": 0.9297, "step": 3385 }, { "epoch": 0.6516551193225558, "grad_norm": 1.7341997623443604, "learning_rate": 5.717406797477371e-06, "loss": 0.8876, "step": 3386 }, { "epoch": 0.6518475750577367, "grad_norm": 0.9093543887138367, "learning_rate": 5.711774854662535e-06, "loss": 0.9212, "step": 3387 }, { "epoch": 0.6520400307929176, "grad_norm": 1.7846508026123047, "learning_rate": 5.706144578005908e-06, "loss": 0.8998, "step": 3388 }, { "epoch": 0.6522324865280985, "grad_norm": 1.7994836568832397, "learning_rate": 5.700515969695094e-06, "loss": 0.9671, "step": 3389 }, { "epoch": 0.6524249422632794, "grad_norm": 1.801116704940796, "learning_rate": 5.694889031917047e-06, "loss": 0.9491, "step": 3390 }, { "epoch": 0.6526173979984603, "grad_norm": 1.881959319114685, "learning_rate": 5.689263766858072e-06, "loss": 0.9842, "step": 3391 }, { "epoch": 0.6528098537336413, "grad_norm": 1.8578566312789917, "learning_rate": 5.683640176703824e-06, "loss": 0.9265, "step": 3392 }, { "epoch": 0.6530023094688222, "grad_norm": 1.7497116327285767, "learning_rate": 5.678018263639309e-06, "loss": 0.9032, "step": 3393 }, { "epoch": 0.6531947652040031, "grad_norm": 0.8865742087364197, "learning_rate": 5.672398029848881e-06, "loss": 0.9299, "step": 3394 }, { "epoch": 0.653387220939184, "grad_norm": 1.769739031791687, "learning_rate": 5.666779477516234e-06, "loss": 0.9556, "step": 3395 }, { "epoch": 0.6535796766743649, "grad_norm": 0.9176982641220093, "learning_rate": 5.66116260882442e-06, "loss": 0.9054, "step": 3396 }, { "epoch": 0.6537721324095458, "grad_norm": 1.6735881567001343, "learning_rate": 5.65554742595583e-06, "loss": 0.9256, "step": 3397 }, { "epoch": 0.6539645881447267, "grad_norm": 1.5571702718734741, "learning_rate": 5.6499339310922015e-06, "loss": 0.9388, "step": 3398 }, { "epoch": 0.6541570438799076, "grad_norm": 1.8811582326889038, "learning_rate": 5.6443221264146196e-06, "loss": 0.8605, "step": 3399 }, { "epoch": 0.6543494996150885, "grad_norm": 1.8953214883804321, "learning_rate": 5.638712014103507e-06, "loss": 0.886, "step": 3400 }, { "epoch": 0.6545419553502695, "grad_norm": 1.754567265510559, "learning_rate": 5.633103596338631e-06, "loss": 0.9831, "step": 3401 }, { "epoch": 0.6547344110854504, "grad_norm": 1.8926575183868408, "learning_rate": 5.627496875299102e-06, "loss": 0.9411, "step": 3402 }, { "epoch": 0.6549268668206313, "grad_norm": 1.6930826902389526, "learning_rate": 5.62189185316337e-06, "loss": 0.8933, "step": 3403 }, { "epoch": 0.6551193225558122, "grad_norm": 1.8461722135543823, "learning_rate": 5.616288532109225e-06, "loss": 0.9909, "step": 3404 }, { "epoch": 0.6553117782909931, "grad_norm": 1.6353554725646973, "learning_rate": 5.610686914313796e-06, "loss": 0.9731, "step": 3405 }, { "epoch": 0.655504234026174, "grad_norm": 1.698771357536316, "learning_rate": 5.6050870019535496e-06, "loss": 0.9709, "step": 3406 }, { "epoch": 0.6556966897613549, "grad_norm": 1.9085934162139893, "learning_rate": 5.599488797204291e-06, "loss": 0.9083, "step": 3407 }, { "epoch": 0.6558891454965358, "grad_norm": 0.9568070769309998, "learning_rate": 5.5938923022411615e-06, "loss": 0.9615, "step": 3408 }, { "epoch": 0.6560816012317167, "grad_norm": 1.7880637645721436, "learning_rate": 5.588297519238637e-06, "loss": 1.0197, "step": 3409 }, { "epoch": 0.6562740569668977, "grad_norm": 1.696130394935608, "learning_rate": 5.582704450370533e-06, "loss": 0.9767, "step": 3410 }, { "epoch": 0.6564665127020786, "grad_norm": 1.9392369985580444, "learning_rate": 5.5771130978099896e-06, "loss": 0.8901, "step": 3411 }, { "epoch": 0.6566589684372595, "grad_norm": 1.7083916664123535, "learning_rate": 5.571523463729487e-06, "loss": 0.8612, "step": 3412 }, { "epoch": 0.6568514241724404, "grad_norm": 1.766940712928772, "learning_rate": 5.565935550300836e-06, "loss": 1.0245, "step": 3413 }, { "epoch": 0.6570438799076213, "grad_norm": 1.708411693572998, "learning_rate": 5.560349359695181e-06, "loss": 0.9535, "step": 3414 }, { "epoch": 0.6572363356428021, "grad_norm": 1.670191764831543, "learning_rate": 5.554764894082994e-06, "loss": 0.8221, "step": 3415 }, { "epoch": 0.657428791377983, "grad_norm": 1.810355305671692, "learning_rate": 5.549182155634076e-06, "loss": 0.9612, "step": 3416 }, { "epoch": 0.6576212471131639, "grad_norm": 1.941634178161621, "learning_rate": 5.543601146517563e-06, "loss": 0.8933, "step": 3417 }, { "epoch": 0.6578137028483448, "grad_norm": 1.8155558109283447, "learning_rate": 5.5380218689019125e-06, "loss": 0.8816, "step": 3418 }, { "epoch": 0.6580061585835257, "grad_norm": 1.644957184791565, "learning_rate": 5.532444324954911e-06, "loss": 0.9753, "step": 3419 }, { "epoch": 0.6581986143187067, "grad_norm": 1.6677231788635254, "learning_rate": 5.526868516843673e-06, "loss": 1.0215, "step": 3420 }, { "epoch": 0.6583910700538876, "grad_norm": 1.793299674987793, "learning_rate": 5.521294446734637e-06, "loss": 0.8978, "step": 3421 }, { "epoch": 0.6585835257890685, "grad_norm": 1.7209322452545166, "learning_rate": 5.51572211679357e-06, "loss": 0.9936, "step": 3422 }, { "epoch": 0.6587759815242494, "grad_norm": 1.8610836267471313, "learning_rate": 5.510151529185554e-06, "loss": 0.8362, "step": 3423 }, { "epoch": 0.6589684372594303, "grad_norm": 1.8868011236190796, "learning_rate": 5.504582686075002e-06, "loss": 0.9202, "step": 3424 }, { "epoch": 0.6591608929946112, "grad_norm": 1.5614392757415771, "learning_rate": 5.499015589625649e-06, "loss": 0.8545, "step": 3425 }, { "epoch": 0.6593533487297921, "grad_norm": 1.7764856815338135, "learning_rate": 5.493450242000546e-06, "loss": 0.9722, "step": 3426 }, { "epoch": 0.659545804464973, "grad_norm": 1.8225010633468628, "learning_rate": 5.4878866453620746e-06, "loss": 0.9788, "step": 3427 }, { "epoch": 0.6597382602001539, "grad_norm": 1.8619908094406128, "learning_rate": 5.4823248018719184e-06, "loss": 0.9608, "step": 3428 }, { "epoch": 0.6599307159353349, "grad_norm": 1.6834542751312256, "learning_rate": 5.4767647136910984e-06, "loss": 0.8999, "step": 3429 }, { "epoch": 0.6601231716705158, "grad_norm": 1.8629670143127441, "learning_rate": 5.471206382979941e-06, "loss": 0.9221, "step": 3430 }, { "epoch": 0.6603156274056967, "grad_norm": 1.7257126569747925, "learning_rate": 5.465649811898098e-06, "loss": 0.8565, "step": 3431 }, { "epoch": 0.6605080831408776, "grad_norm": 1.7740085124969482, "learning_rate": 5.460095002604533e-06, "loss": 0.8931, "step": 3432 }, { "epoch": 0.6607005388760585, "grad_norm": 1.7363128662109375, "learning_rate": 5.4545419572575265e-06, "loss": 0.953, "step": 3433 }, { "epoch": 0.6608929946112394, "grad_norm": 1.8454254865646362, "learning_rate": 5.4489906780146716e-06, "loss": 0.9635, "step": 3434 }, { "epoch": 0.6610854503464203, "grad_norm": 1.7508928775787354, "learning_rate": 5.44344116703288e-06, "loss": 0.9736, "step": 3435 }, { "epoch": 0.6612779060816012, "grad_norm": 1.9192637205123901, "learning_rate": 5.43789342646837e-06, "loss": 0.9484, "step": 3436 }, { "epoch": 0.6614703618167821, "grad_norm": 0.9085773825645447, "learning_rate": 5.432347458476678e-06, "loss": 0.9001, "step": 3437 }, { "epoch": 0.6616628175519631, "grad_norm": 0.9292737245559692, "learning_rate": 5.4268032652126465e-06, "loss": 0.9142, "step": 3438 }, { "epoch": 0.661855273287144, "grad_norm": 1.6490066051483154, "learning_rate": 5.421260848830432e-06, "loss": 0.9434, "step": 3439 }, { "epoch": 0.6620477290223249, "grad_norm": 1.6521810293197632, "learning_rate": 5.415720211483499e-06, "loss": 0.9301, "step": 3440 }, { "epoch": 0.6622401847575058, "grad_norm": 2.8676137924194336, "learning_rate": 5.410181355324622e-06, "loss": 0.8729, "step": 3441 }, { "epoch": 0.6624326404926867, "grad_norm": 1.9626593589782715, "learning_rate": 5.404644282505882e-06, "loss": 0.9586, "step": 3442 }, { "epoch": 0.6626250962278676, "grad_norm": 1.8076010942459106, "learning_rate": 5.3991089951786725e-06, "loss": 0.9604, "step": 3443 }, { "epoch": 0.6628175519630485, "grad_norm": 1.7243555784225464, "learning_rate": 5.393575495493679e-06, "loss": 0.9466, "step": 3444 }, { "epoch": 0.6630100076982294, "grad_norm": 1.7615028619766235, "learning_rate": 5.388043785600907e-06, "loss": 0.8858, "step": 3445 }, { "epoch": 0.6632024634334103, "grad_norm": 1.7679067850112915, "learning_rate": 5.382513867649663e-06, "loss": 0.9664, "step": 3446 }, { "epoch": 0.6633949191685913, "grad_norm": 1.7351717948913574, "learning_rate": 5.376985743788552e-06, "loss": 0.863, "step": 3447 }, { "epoch": 0.6635873749037722, "grad_norm": 1.8214083909988403, "learning_rate": 5.37145941616549e-06, "loss": 0.9375, "step": 3448 }, { "epoch": 0.6637798306389531, "grad_norm": 1.7770386934280396, "learning_rate": 5.365934886927687e-06, "loss": 0.9515, "step": 3449 }, { "epoch": 0.663972286374134, "grad_norm": 1.793613314628601, "learning_rate": 5.360412158221661e-06, "loss": 0.9284, "step": 3450 }, { "epoch": 0.6641647421093149, "grad_norm": 1.8269356489181519, "learning_rate": 5.354891232193225e-06, "loss": 0.975, "step": 3451 }, { "epoch": 0.6643571978444958, "grad_norm": 1.562092661857605, "learning_rate": 5.349372110987496e-06, "loss": 0.9774, "step": 3452 }, { "epoch": 0.6645496535796767, "grad_norm": 1.8016637563705444, "learning_rate": 5.343854796748886e-06, "loss": 0.914, "step": 3453 }, { "epoch": 0.6647421093148576, "grad_norm": 1.7373336553573608, "learning_rate": 5.338339291621109e-06, "loss": 0.9026, "step": 3454 }, { "epoch": 0.6649345650500385, "grad_norm": 0.964300811290741, "learning_rate": 5.332825597747171e-06, "loss": 0.9106, "step": 3455 }, { "epoch": 0.6651270207852193, "grad_norm": 1.9491894245147705, "learning_rate": 5.32731371726938e-06, "loss": 0.992, "step": 3456 }, { "epoch": 0.6653194765204004, "grad_norm": 1.922573208808899, "learning_rate": 5.321803652329334e-06, "loss": 0.9094, "step": 3457 }, { "epoch": 0.6655119322555813, "grad_norm": 1.7878295183181763, "learning_rate": 5.316295405067929e-06, "loss": 0.9413, "step": 3458 }, { "epoch": 0.6657043879907621, "grad_norm": 1.5722486972808838, "learning_rate": 5.3107889776253544e-06, "loss": 0.9004, "step": 3459 }, { "epoch": 0.665896843725943, "grad_norm": 2.009685516357422, "learning_rate": 5.305284372141095e-06, "loss": 0.9912, "step": 3460 }, { "epoch": 0.6660892994611239, "grad_norm": 1.6209447383880615, "learning_rate": 5.299781590753916e-06, "loss": 1.0165, "step": 3461 }, { "epoch": 0.6662817551963048, "grad_norm": 1.9641523361206055, "learning_rate": 5.294280635601889e-06, "loss": 0.8787, "step": 3462 }, { "epoch": 0.6664742109314857, "grad_norm": 1.792001724243164, "learning_rate": 5.288781508822367e-06, "loss": 0.9928, "step": 3463 }, { "epoch": 0.6666666666666666, "grad_norm": 1.7955774068832397, "learning_rate": 5.283284212551997e-06, "loss": 0.9366, "step": 3464 }, { "epoch": 0.6668591224018475, "grad_norm": 1.6494568586349487, "learning_rate": 5.277788748926711e-06, "loss": 0.9377, "step": 3465 }, { "epoch": 0.6670515781370285, "grad_norm": 2.3010692596435547, "learning_rate": 5.2722951200817315e-06, "loss": 0.9651, "step": 3466 }, { "epoch": 0.6672440338722094, "grad_norm": 1.67125403881073, "learning_rate": 5.2668033281515676e-06, "loss": 0.8703, "step": 3467 }, { "epoch": 0.6674364896073903, "grad_norm": 1.6498452425003052, "learning_rate": 5.2613133752700145e-06, "loss": 0.9545, "step": 3468 }, { "epoch": 0.6676289453425712, "grad_norm": 1.6180171966552734, "learning_rate": 5.2558252635701535e-06, "loss": 0.8312, "step": 3469 }, { "epoch": 0.6678214010777521, "grad_norm": 1.7080947160720825, "learning_rate": 5.250338995184348e-06, "loss": 0.8749, "step": 3470 }, { "epoch": 0.668013856812933, "grad_norm": 1.50067138671875, "learning_rate": 5.244854572244249e-06, "loss": 0.9261, "step": 3471 }, { "epoch": 0.6682063125481139, "grad_norm": 1.7629857063293457, "learning_rate": 5.239371996880786e-06, "loss": 0.8962, "step": 3472 }, { "epoch": 0.6683987682832948, "grad_norm": 1.6186702251434326, "learning_rate": 5.2338912712241785e-06, "loss": 0.9602, "step": 3473 }, { "epoch": 0.6685912240184757, "grad_norm": 1.7863569259643555, "learning_rate": 5.228412397403916e-06, "loss": 0.9261, "step": 3474 }, { "epoch": 0.6687836797536567, "grad_norm": 1.749067783355713, "learning_rate": 5.222935377548777e-06, "loss": 1.0113, "step": 3475 }, { "epoch": 0.6689761354888376, "grad_norm": 1.7121068239212036, "learning_rate": 5.217460213786822e-06, "loss": 0.9403, "step": 3476 }, { "epoch": 0.6691685912240185, "grad_norm": 1.9137980937957764, "learning_rate": 5.211986908245374e-06, "loss": 0.9164, "step": 3477 }, { "epoch": 0.6693610469591994, "grad_norm": 1.6937050819396973, "learning_rate": 5.206515463051054e-06, "loss": 1.0324, "step": 3478 }, { "epoch": 0.6695535026943803, "grad_norm": 1.6444754600524902, "learning_rate": 5.201045880329749e-06, "loss": 1.0024, "step": 3479 }, { "epoch": 0.6697459584295612, "grad_norm": 1.6089844703674316, "learning_rate": 5.195578162206627e-06, "loss": 0.8336, "step": 3480 }, { "epoch": 0.6699384141647421, "grad_norm": 1.8064433336257935, "learning_rate": 5.190112310806126e-06, "loss": 1.0179, "step": 3481 }, { "epoch": 0.670130869899923, "grad_norm": 0.8969301581382751, "learning_rate": 5.184648328251965e-06, "loss": 0.8848, "step": 3482 }, { "epoch": 0.6703233256351039, "grad_norm": 1.7852959632873535, "learning_rate": 5.179186216667133e-06, "loss": 0.9681, "step": 3483 }, { "epoch": 0.6705157813702848, "grad_norm": 1.7706224918365479, "learning_rate": 5.1737259781738934e-06, "loss": 0.967, "step": 3484 }, { "epoch": 0.6707082371054658, "grad_norm": 1.704100489616394, "learning_rate": 5.1682676148937805e-06, "loss": 0.8792, "step": 3485 }, { "epoch": 0.6709006928406467, "grad_norm": 2.179177761077881, "learning_rate": 5.1628111289476025e-06, "loss": 0.9237, "step": 3486 }, { "epoch": 0.6710931485758276, "grad_norm": 2.4599878787994385, "learning_rate": 5.1573565224554365e-06, "loss": 0.9175, "step": 3487 }, { "epoch": 0.6712856043110085, "grad_norm": 1.5377827882766724, "learning_rate": 5.151903797536631e-06, "loss": 0.9209, "step": 3488 }, { "epoch": 0.6714780600461894, "grad_norm": 1.8712668418884277, "learning_rate": 5.146452956309798e-06, "loss": 0.9393, "step": 3489 }, { "epoch": 0.6716705157813703, "grad_norm": 1.6868737936019897, "learning_rate": 5.141004000892824e-06, "loss": 0.939, "step": 3490 }, { "epoch": 0.6718629715165512, "grad_norm": 1.6263906955718994, "learning_rate": 5.135556933402862e-06, "loss": 0.95, "step": 3491 }, { "epoch": 0.6720554272517321, "grad_norm": 0.9104430079460144, "learning_rate": 5.130111755956327e-06, "loss": 0.9395, "step": 3492 }, { "epoch": 0.672247882986913, "grad_norm": 1.6872116327285767, "learning_rate": 5.1246684706689074e-06, "loss": 0.9022, "step": 3493 }, { "epoch": 0.672440338722094, "grad_norm": 1.6843801736831665, "learning_rate": 5.119227079655544e-06, "loss": 0.8975, "step": 3494 }, { "epoch": 0.6726327944572749, "grad_norm": 1.589044213294983, "learning_rate": 5.1137875850304545e-06, "loss": 0.9914, "step": 3495 }, { "epoch": 0.6728252501924558, "grad_norm": 1.6750380992889404, "learning_rate": 5.108349988907111e-06, "loss": 0.9818, "step": 3496 }, { "epoch": 0.6730177059276367, "grad_norm": 1.7462462186813354, "learning_rate": 5.102914293398252e-06, "loss": 0.9279, "step": 3497 }, { "epoch": 0.6732101616628176, "grad_norm": 1.7899361848831177, "learning_rate": 5.097480500615879e-06, "loss": 0.9865, "step": 3498 }, { "epoch": 0.6734026173979984, "grad_norm": 1.9460341930389404, "learning_rate": 5.092048612671252e-06, "loss": 0.9656, "step": 3499 }, { "epoch": 0.6735950731331793, "grad_norm": 1.8290739059448242, "learning_rate": 5.086618631674888e-06, "loss": 0.9661, "step": 3500 }, { "epoch": 0.6737875288683602, "grad_norm": 1.741752028465271, "learning_rate": 5.081190559736569e-06, "loss": 0.9707, "step": 3501 }, { "epoch": 0.6739799846035411, "grad_norm": 1.849439263343811, "learning_rate": 5.075764398965331e-06, "loss": 0.9626, "step": 3502 }, { "epoch": 0.6741724403387221, "grad_norm": 1.67465341091156, "learning_rate": 5.070340151469468e-06, "loss": 0.8269, "step": 3503 }, { "epoch": 0.674364896073903, "grad_norm": 1.7456953525543213, "learning_rate": 5.064917819356532e-06, "loss": 0.9131, "step": 3504 }, { "epoch": 0.6745573518090839, "grad_norm": 1.6493909358978271, "learning_rate": 5.059497404733331e-06, "loss": 0.9644, "step": 3505 }, { "epoch": 0.6747498075442648, "grad_norm": 1.7014745473861694, "learning_rate": 5.054078909705926e-06, "loss": 0.9325, "step": 3506 }, { "epoch": 0.6749422632794457, "grad_norm": 0.8820096254348755, "learning_rate": 5.048662336379635e-06, "loss": 0.8765, "step": 3507 }, { "epoch": 0.6751347190146266, "grad_norm": 1.935365080833435, "learning_rate": 5.043247686859024e-06, "loss": 0.9149, "step": 3508 }, { "epoch": 0.6753271747498075, "grad_norm": 1.8217785358428955, "learning_rate": 5.037834963247922e-06, "loss": 1.0192, "step": 3509 }, { "epoch": 0.6755196304849884, "grad_norm": 1.6435409784317017, "learning_rate": 5.032424167649394e-06, "loss": 0.97, "step": 3510 }, { "epoch": 0.6757120862201693, "grad_norm": 1.8585370779037476, "learning_rate": 5.027015302165768e-06, "loss": 0.8559, "step": 3511 }, { "epoch": 0.6759045419553503, "grad_norm": 1.7686553001403809, "learning_rate": 5.021608368898621e-06, "loss": 0.9372, "step": 3512 }, { "epoch": 0.6760969976905312, "grad_norm": 1.9385324716567993, "learning_rate": 5.016203369948774e-06, "loss": 0.9675, "step": 3513 }, { "epoch": 0.6762894534257121, "grad_norm": 0.9401078224182129, "learning_rate": 5.0108003074163005e-06, "loss": 0.9109, "step": 3514 }, { "epoch": 0.676481909160893, "grad_norm": 1.8436037302017212, "learning_rate": 5.005399183400521e-06, "loss": 1.0093, "step": 3515 }, { "epoch": 0.6766743648960739, "grad_norm": 1.6753051280975342, "learning_rate": 5.000000000000003e-06, "loss": 0.8998, "step": 3516 }, { "epoch": 0.6768668206312548, "grad_norm": 1.661201000213623, "learning_rate": 4.994602759312557e-06, "loss": 0.8507, "step": 3517 }, { "epoch": 0.6770592763664357, "grad_norm": 1.5893633365631104, "learning_rate": 4.989207463435244e-06, "loss": 0.9459, "step": 3518 }, { "epoch": 0.6772517321016166, "grad_norm": 1.7016270160675049, "learning_rate": 4.9838141144643636e-06, "loss": 0.972, "step": 3519 }, { "epoch": 0.6774441878367975, "grad_norm": 0.8540983200073242, "learning_rate": 4.978422714495465e-06, "loss": 0.8667, "step": 3520 }, { "epoch": 0.6776366435719784, "grad_norm": 1.751617193222046, "learning_rate": 4.973033265623333e-06, "loss": 0.8754, "step": 3521 }, { "epoch": 0.6778290993071594, "grad_norm": 0.883199155330658, "learning_rate": 4.967645769942e-06, "loss": 0.9455, "step": 3522 }, { "epoch": 0.6780215550423403, "grad_norm": 0.8881901502609253, "learning_rate": 4.962260229544738e-06, "loss": 0.9152, "step": 3523 }, { "epoch": 0.6782140107775212, "grad_norm": 1.5736958980560303, "learning_rate": 4.956876646524059e-06, "loss": 0.9798, "step": 3524 }, { "epoch": 0.6784064665127021, "grad_norm": 1.7621546983718872, "learning_rate": 4.951495022971713e-06, "loss": 0.9015, "step": 3525 }, { "epoch": 0.678598922247883, "grad_norm": 1.5872132778167725, "learning_rate": 4.946115360978696e-06, "loss": 0.8964, "step": 3526 }, { "epoch": 0.6787913779830639, "grad_norm": 1.7109295129776, "learning_rate": 4.940737662635227e-06, "loss": 0.9937, "step": 3527 }, { "epoch": 0.6789838337182448, "grad_norm": 1.647711157798767, "learning_rate": 4.935361930030774e-06, "loss": 0.9399, "step": 3528 }, { "epoch": 0.6791762894534257, "grad_norm": 1.7020493745803833, "learning_rate": 4.929988165254041e-06, "loss": 0.9042, "step": 3529 }, { "epoch": 0.6793687451886066, "grad_norm": 1.5348321199417114, "learning_rate": 4.924616370392962e-06, "loss": 0.9086, "step": 3530 }, { "epoch": 0.6795612009237876, "grad_norm": 1.6763864755630493, "learning_rate": 4.919246547534709e-06, "loss": 0.8863, "step": 3531 }, { "epoch": 0.6797536566589685, "grad_norm": 1.560278296470642, "learning_rate": 4.9138786987656865e-06, "loss": 0.9043, "step": 3532 }, { "epoch": 0.6799461123941494, "grad_norm": 1.7472281455993652, "learning_rate": 4.908512826171533e-06, "loss": 0.8813, "step": 3533 }, { "epoch": 0.6801385681293303, "grad_norm": 1.7941151857376099, "learning_rate": 4.9031489318371195e-06, "loss": 0.9644, "step": 3534 }, { "epoch": 0.6803310238645112, "grad_norm": 1.738924503326416, "learning_rate": 4.897787017846547e-06, "loss": 0.9145, "step": 3535 }, { "epoch": 0.6805234795996921, "grad_norm": 0.9181484580039978, "learning_rate": 4.892427086283147e-06, "loss": 0.9093, "step": 3536 }, { "epoch": 0.680715935334873, "grad_norm": 1.7104320526123047, "learning_rate": 4.887069139229481e-06, "loss": 0.9074, "step": 3537 }, { "epoch": 0.6809083910700539, "grad_norm": 1.7616809606552124, "learning_rate": 4.881713178767341e-06, "loss": 0.9201, "step": 3538 }, { "epoch": 0.6811008468052347, "grad_norm": 1.7111355066299438, "learning_rate": 4.8763592069777454e-06, "loss": 0.8797, "step": 3539 }, { "epoch": 0.6812933025404158, "grad_norm": 1.5863111019134521, "learning_rate": 4.87100722594094e-06, "loss": 0.9595, "step": 3540 }, { "epoch": 0.6814857582755967, "grad_norm": 1.6166331768035889, "learning_rate": 4.865657237736397e-06, "loss": 0.9053, "step": 3541 }, { "epoch": 0.6816782140107776, "grad_norm": 1.7800498008728027, "learning_rate": 4.860309244442821e-06, "loss": 0.9915, "step": 3542 }, { "epoch": 0.6818706697459584, "grad_norm": 2.0758252143859863, "learning_rate": 4.854963248138121e-06, "loss": 0.9524, "step": 3543 }, { "epoch": 0.6820631254811393, "grad_norm": 1.8680355548858643, "learning_rate": 4.849619250899458e-06, "loss": 0.9392, "step": 3544 }, { "epoch": 0.6822555812163202, "grad_norm": 1.9671584367752075, "learning_rate": 4.844277254803198e-06, "loss": 0.9144, "step": 3545 }, { "epoch": 0.6824480369515011, "grad_norm": 1.6354202032089233, "learning_rate": 4.838937261924933e-06, "loss": 0.9043, "step": 3546 }, { "epoch": 0.682640492686682, "grad_norm": 1.802881121635437, "learning_rate": 4.83359927433948e-06, "loss": 0.9897, "step": 3547 }, { "epoch": 0.6828329484218629, "grad_norm": 1.6808555126190186, "learning_rate": 4.8282632941208725e-06, "loss": 0.9313, "step": 3548 }, { "epoch": 0.6830254041570438, "grad_norm": 1.6302028894424438, "learning_rate": 4.822929323342369e-06, "loss": 0.8836, "step": 3549 }, { "epoch": 0.6832178598922248, "grad_norm": 1.6222329139709473, "learning_rate": 4.817597364076441e-06, "loss": 1.0089, "step": 3550 }, { "epoch": 0.6834103156274057, "grad_norm": 1.8248558044433594, "learning_rate": 4.812267418394784e-06, "loss": 0.9048, "step": 3551 }, { "epoch": 0.6836027713625866, "grad_norm": 1.7333405017852783, "learning_rate": 4.806939488368308e-06, "loss": 0.9519, "step": 3552 }, { "epoch": 0.6837952270977675, "grad_norm": 1.7828094959259033, "learning_rate": 4.8016135760671435e-06, "loss": 0.9824, "step": 3553 }, { "epoch": 0.6839876828329484, "grad_norm": 1.839829921722412, "learning_rate": 4.796289683560631e-06, "loss": 0.9057, "step": 3554 }, { "epoch": 0.6841801385681293, "grad_norm": 1.8601003885269165, "learning_rate": 4.790967812917334e-06, "loss": 0.9751, "step": 3555 }, { "epoch": 0.6843725943033102, "grad_norm": 1.6036670207977295, "learning_rate": 4.78564796620502e-06, "loss": 0.9759, "step": 3556 }, { "epoch": 0.6845650500384911, "grad_norm": 1.7320500612258911, "learning_rate": 4.7803301454906804e-06, "loss": 0.891, "step": 3557 }, { "epoch": 0.684757505773672, "grad_norm": 1.685636043548584, "learning_rate": 4.775014352840512e-06, "loss": 0.8745, "step": 3558 }, { "epoch": 0.684949961508853, "grad_norm": 0.9304041862487793, "learning_rate": 4.76970059031993e-06, "loss": 0.9065, "step": 3559 }, { "epoch": 0.6851424172440339, "grad_norm": 1.785025954246521, "learning_rate": 4.764388859993556e-06, "loss": 0.9304, "step": 3560 }, { "epoch": 0.6853348729792148, "grad_norm": 1.6687031984329224, "learning_rate": 4.759079163925223e-06, "loss": 0.9175, "step": 3561 }, { "epoch": 0.6855273287143957, "grad_norm": 1.784484624862671, "learning_rate": 4.753771504177976e-06, "loss": 0.9368, "step": 3562 }, { "epoch": 0.6857197844495766, "grad_norm": 1.7726801633834839, "learning_rate": 4.7484658828140665e-06, "loss": 0.893, "step": 3563 }, { "epoch": 0.6859122401847575, "grad_norm": 1.666787028312683, "learning_rate": 4.743162301894952e-06, "loss": 0.962, "step": 3564 }, { "epoch": 0.6861046959199384, "grad_norm": 0.9010095000267029, "learning_rate": 4.7378607634813045e-06, "loss": 0.9351, "step": 3565 }, { "epoch": 0.6862971516551193, "grad_norm": 1.8619632720947266, "learning_rate": 4.732561269632992e-06, "loss": 0.966, "step": 3566 }, { "epoch": 0.6864896073903002, "grad_norm": 1.680849313735962, "learning_rate": 4.727263822409099e-06, "loss": 0.9024, "step": 3567 }, { "epoch": 0.6866820631254812, "grad_norm": 1.975286602973938, "learning_rate": 4.7219684238679066e-06, "loss": 0.956, "step": 3568 }, { "epoch": 0.6868745188606621, "grad_norm": 1.6593296527862549, "learning_rate": 4.716675076066902e-06, "loss": 0.9336, "step": 3569 }, { "epoch": 0.687066974595843, "grad_norm": 1.5858948230743408, "learning_rate": 4.711383781062779e-06, "loss": 0.9569, "step": 3570 }, { "epoch": 0.6872594303310239, "grad_norm": 0.8995071053504944, "learning_rate": 4.706094540911429e-06, "loss": 0.9448, "step": 3571 }, { "epoch": 0.6874518860662048, "grad_norm": 1.742321252822876, "learning_rate": 4.700807357667953e-06, "loss": 1.0261, "step": 3572 }, { "epoch": 0.6876443418013857, "grad_norm": 1.8734294176101685, "learning_rate": 4.6955222333866365e-06, "loss": 0.8509, "step": 3573 }, { "epoch": 0.6878367975365666, "grad_norm": 1.8507840633392334, "learning_rate": 4.6902391701209826e-06, "loss": 0.909, "step": 3574 }, { "epoch": 0.6880292532717475, "grad_norm": 1.7655754089355469, "learning_rate": 4.684958169923684e-06, "loss": 0.897, "step": 3575 }, { "epoch": 0.6882217090069284, "grad_norm": 1.6593842506408691, "learning_rate": 4.679679234846636e-06, "loss": 0.9472, "step": 3576 }, { "epoch": 0.6884141647421094, "grad_norm": 1.71574068069458, "learning_rate": 4.674402366940927e-06, "loss": 0.9566, "step": 3577 }, { "epoch": 0.6886066204772903, "grad_norm": 1.632213830947876, "learning_rate": 4.669127568256851e-06, "loss": 0.9267, "step": 3578 }, { "epoch": 0.6887990762124712, "grad_norm": 1.7090871334075928, "learning_rate": 4.663854840843885e-06, "loss": 0.9873, "step": 3579 }, { "epoch": 0.6889915319476521, "grad_norm": 1.8156142234802246, "learning_rate": 4.658584186750713e-06, "loss": 0.9158, "step": 3580 }, { "epoch": 0.689183987682833, "grad_norm": 1.6955251693725586, "learning_rate": 4.6533156080252076e-06, "loss": 0.8719, "step": 3581 }, { "epoch": 0.6893764434180139, "grad_norm": 1.5604935884475708, "learning_rate": 4.648049106714434e-06, "loss": 0.9099, "step": 3582 }, { "epoch": 0.6895688991531947, "grad_norm": 1.849888801574707, "learning_rate": 4.642784684864656e-06, "loss": 0.9191, "step": 3583 }, { "epoch": 0.6897613548883756, "grad_norm": 1.8071506023406982, "learning_rate": 4.637522344521323e-06, "loss": 0.9007, "step": 3584 }, { "epoch": 0.6899538106235565, "grad_norm": 2.4560303688049316, "learning_rate": 4.63226208772908e-06, "loss": 0.9467, "step": 3585 }, { "epoch": 0.6901462663587374, "grad_norm": 1.7255818843841553, "learning_rate": 4.627003916531761e-06, "loss": 0.977, "step": 3586 }, { "epoch": 0.6903387220939184, "grad_norm": 1.5953834056854248, "learning_rate": 4.621747832972389e-06, "loss": 0.9199, "step": 3587 }, { "epoch": 0.6905311778290993, "grad_norm": 1.7392430305480957, "learning_rate": 4.616493839093179e-06, "loss": 0.9379, "step": 3588 }, { "epoch": 0.6907236335642802, "grad_norm": 1.8367292881011963, "learning_rate": 4.6112419369355255e-06, "loss": 0.9296, "step": 3589 }, { "epoch": 0.6909160892994611, "grad_norm": 1.7075905799865723, "learning_rate": 4.60599212854002e-06, "loss": 0.9808, "step": 3590 }, { "epoch": 0.691108545034642, "grad_norm": 1.8431198596954346, "learning_rate": 4.600744415946438e-06, "loss": 0.895, "step": 3591 }, { "epoch": 0.6913010007698229, "grad_norm": 0.9042073488235474, "learning_rate": 4.595498801193736e-06, "loss": 0.949, "step": 3592 }, { "epoch": 0.6914934565050038, "grad_norm": 1.8543001413345337, "learning_rate": 4.590255286320062e-06, "loss": 0.9352, "step": 3593 }, { "epoch": 0.6916859122401847, "grad_norm": 1.6687097549438477, "learning_rate": 4.5850138733627435e-06, "loss": 0.9901, "step": 3594 }, { "epoch": 0.6918783679753656, "grad_norm": 1.6543142795562744, "learning_rate": 4.5797745643582935e-06, "loss": 0.935, "step": 3595 }, { "epoch": 0.6920708237105466, "grad_norm": 1.7264105081558228, "learning_rate": 4.5745373613424075e-06, "loss": 0.9461, "step": 3596 }, { "epoch": 0.6922632794457275, "grad_norm": 1.637157678604126, "learning_rate": 4.5693022663499584e-06, "loss": 1.0417, "step": 3597 }, { "epoch": 0.6924557351809084, "grad_norm": 1.773926854133606, "learning_rate": 4.564069281415009e-06, "loss": 0.957, "step": 3598 }, { "epoch": 0.6926481909160893, "grad_norm": 1.6433608531951904, "learning_rate": 4.5588384085707925e-06, "loss": 0.9037, "step": 3599 }, { "epoch": 0.6928406466512702, "grad_norm": 1.6686745882034302, "learning_rate": 4.5536096498497295e-06, "loss": 0.9686, "step": 3600 }, { "epoch": 0.6930331023864511, "grad_norm": 1.7095450162887573, "learning_rate": 4.548383007283412e-06, "loss": 0.9948, "step": 3601 }, { "epoch": 0.693225558121632, "grad_norm": 1.7358208894729614, "learning_rate": 4.543158482902617e-06, "loss": 0.9189, "step": 3602 }, { "epoch": 0.6934180138568129, "grad_norm": 1.6004054546356201, "learning_rate": 4.537936078737291e-06, "loss": 0.9229, "step": 3603 }, { "epoch": 0.6936104695919938, "grad_norm": 1.7015902996063232, "learning_rate": 4.532715796816565e-06, "loss": 0.9451, "step": 3604 }, { "epoch": 0.6938029253271748, "grad_norm": 1.6692826747894287, "learning_rate": 4.52749763916874e-06, "loss": 0.9518, "step": 3605 }, { "epoch": 0.6939953810623557, "grad_norm": 1.636660099029541, "learning_rate": 4.522281607821288e-06, "loss": 0.9578, "step": 3606 }, { "epoch": 0.6941878367975366, "grad_norm": 1.773856520652771, "learning_rate": 4.517067704800864e-06, "loss": 0.9519, "step": 3607 }, { "epoch": 0.6943802925327175, "grad_norm": 1.764016032218933, "learning_rate": 4.511855932133289e-06, "loss": 0.9707, "step": 3608 }, { "epoch": 0.6945727482678984, "grad_norm": 1.8073703050613403, "learning_rate": 4.506646291843561e-06, "loss": 0.9264, "step": 3609 }, { "epoch": 0.6947652040030793, "grad_norm": 1.658465027809143, "learning_rate": 4.501438785955845e-06, "loss": 0.9832, "step": 3610 }, { "epoch": 0.6949576597382602, "grad_norm": 1.8624742031097412, "learning_rate": 4.496233416493481e-06, "loss": 0.9859, "step": 3611 }, { "epoch": 0.6951501154734411, "grad_norm": 1.6144973039627075, "learning_rate": 4.491030185478976e-06, "loss": 0.9038, "step": 3612 }, { "epoch": 0.695342571208622, "grad_norm": 1.7121502161026, "learning_rate": 4.4858290949340065e-06, "loss": 0.9668, "step": 3613 }, { "epoch": 0.6955350269438029, "grad_norm": 1.5614014863967896, "learning_rate": 4.480630146879419e-06, "loss": 0.883, "step": 3614 }, { "epoch": 0.6957274826789839, "grad_norm": 1.6696916818618774, "learning_rate": 4.475433343335225e-06, "loss": 0.8828, "step": 3615 }, { "epoch": 0.6959199384141648, "grad_norm": 1.609911322593689, "learning_rate": 4.470238686320606e-06, "loss": 0.8706, "step": 3616 }, { "epoch": 0.6961123941493457, "grad_norm": 1.638105034828186, "learning_rate": 4.4650461778539055e-06, "loss": 0.8847, "step": 3617 }, { "epoch": 0.6963048498845266, "grad_norm": 1.9332939386367798, "learning_rate": 4.459855819952635e-06, "loss": 0.9502, "step": 3618 }, { "epoch": 0.6964973056197075, "grad_norm": 1.8208930492401123, "learning_rate": 4.454667614633471e-06, "loss": 0.9516, "step": 3619 }, { "epoch": 0.6966897613548884, "grad_norm": 0.9762057662010193, "learning_rate": 4.449481563912252e-06, "loss": 0.8994, "step": 3620 }, { "epoch": 0.6968822170900693, "grad_norm": 1.7704150676727295, "learning_rate": 4.444297669803981e-06, "loss": 0.9203, "step": 3621 }, { "epoch": 0.6970746728252502, "grad_norm": 1.9654228687286377, "learning_rate": 4.439115934322816e-06, "loss": 0.9927, "step": 3622 }, { "epoch": 0.697267128560431, "grad_norm": 0.9521400332450867, "learning_rate": 4.433936359482087e-06, "loss": 0.9362, "step": 3623 }, { "epoch": 0.6974595842956121, "grad_norm": 1.901347041130066, "learning_rate": 4.428758947294278e-06, "loss": 0.9587, "step": 3624 }, { "epoch": 0.697652040030793, "grad_norm": 1.8210316896438599, "learning_rate": 4.423583699771036e-06, "loss": 0.8535, "step": 3625 }, { "epoch": 0.6978444957659738, "grad_norm": 1.7348170280456543, "learning_rate": 4.418410618923163e-06, "loss": 0.9578, "step": 3626 }, { "epoch": 0.6980369515011547, "grad_norm": 1.6851931810379028, "learning_rate": 4.413239706760621e-06, "loss": 0.9771, "step": 3627 }, { "epoch": 0.6982294072363356, "grad_norm": 1.746250867843628, "learning_rate": 4.408070965292534e-06, "loss": 0.8785, "step": 3628 }, { "epoch": 0.6984218629715165, "grad_norm": 1.7753324508666992, "learning_rate": 4.402904396527172e-06, "loss": 0.9059, "step": 3629 }, { "epoch": 0.6986143187066974, "grad_norm": 0.9282997250556946, "learning_rate": 4.397740002471973e-06, "loss": 0.9033, "step": 3630 }, { "epoch": 0.6988067744418783, "grad_norm": 1.7341110706329346, "learning_rate": 4.392577785133521e-06, "loss": 0.9376, "step": 3631 }, { "epoch": 0.6989992301770592, "grad_norm": 1.8804572820663452, "learning_rate": 4.387417746517557e-06, "loss": 1.0277, "step": 3632 }, { "epoch": 0.6991916859122402, "grad_norm": 1.758665680885315, "learning_rate": 4.382259888628978e-06, "loss": 0.9101, "step": 3633 }, { "epoch": 0.6993841416474211, "grad_norm": 1.9642398357391357, "learning_rate": 4.377104213471831e-06, "loss": 0.9527, "step": 3634 }, { "epoch": 0.699576597382602, "grad_norm": 0.900040864944458, "learning_rate": 4.371950723049314e-06, "loss": 0.8437, "step": 3635 }, { "epoch": 0.6997690531177829, "grad_norm": 1.7725212574005127, "learning_rate": 4.3667994193637794e-06, "loss": 0.9206, "step": 3636 }, { "epoch": 0.6999615088529638, "grad_norm": 1.7199581861495972, "learning_rate": 4.361650304416728e-06, "loss": 0.8709, "step": 3637 }, { "epoch": 0.7001539645881447, "grad_norm": 2.0091516971588135, "learning_rate": 4.356503380208812e-06, "loss": 0.9527, "step": 3638 }, { "epoch": 0.7003464203233256, "grad_norm": 1.7010260820388794, "learning_rate": 4.351358648739826e-06, "loss": 0.8468, "step": 3639 }, { "epoch": 0.7005388760585065, "grad_norm": 1.919560194015503, "learning_rate": 4.34621611200872e-06, "loss": 0.9051, "step": 3640 }, { "epoch": 0.7007313317936874, "grad_norm": 1.8048046827316284, "learning_rate": 4.3410757720135886e-06, "loss": 0.9027, "step": 3641 }, { "epoch": 0.7009237875288684, "grad_norm": 1.7333379983901978, "learning_rate": 4.335937630751675e-06, "loss": 0.9475, "step": 3642 }, { "epoch": 0.7011162432640493, "grad_norm": 1.6370880603790283, "learning_rate": 4.330801690219361e-06, "loss": 0.9146, "step": 3643 }, { "epoch": 0.7013086989992302, "grad_norm": 1.6574703454971313, "learning_rate": 4.325667952412184e-06, "loss": 0.883, "step": 3644 }, { "epoch": 0.7015011547344111, "grad_norm": 1.7584916353225708, "learning_rate": 4.320536419324816e-06, "loss": 0.9785, "step": 3645 }, { "epoch": 0.701693610469592, "grad_norm": 1.7436972856521606, "learning_rate": 4.315407092951078e-06, "loss": 0.9266, "step": 3646 }, { "epoch": 0.7018860662047729, "grad_norm": 1.6227761507034302, "learning_rate": 4.310279975283932e-06, "loss": 0.9134, "step": 3647 }, { "epoch": 0.7020785219399538, "grad_norm": 1.8522777557373047, "learning_rate": 4.305155068315481e-06, "loss": 0.9518, "step": 3648 }, { "epoch": 0.7022709776751347, "grad_norm": 1.8011834621429443, "learning_rate": 4.30003237403697e-06, "loss": 0.9379, "step": 3649 }, { "epoch": 0.7024634334103156, "grad_norm": 1.920276403427124, "learning_rate": 4.2949118944387846e-06, "loss": 1.0526, "step": 3650 }, { "epoch": 0.7026558891454965, "grad_norm": 1.7253565788269043, "learning_rate": 4.289793631510449e-06, "loss": 0.9917, "step": 3651 }, { "epoch": 0.7028483448806775, "grad_norm": 2.140216827392578, "learning_rate": 4.284677587240625e-06, "loss": 0.9938, "step": 3652 }, { "epoch": 0.7030408006158584, "grad_norm": 1.6234476566314697, "learning_rate": 4.2795637636171165e-06, "loss": 0.9475, "step": 3653 }, { "epoch": 0.7032332563510393, "grad_norm": 1.6648706197738647, "learning_rate": 4.27445216262686e-06, "loss": 0.9402, "step": 3654 }, { "epoch": 0.7034257120862202, "grad_norm": 1.7492082118988037, "learning_rate": 4.2693427862559335e-06, "loss": 0.8974, "step": 3655 }, { "epoch": 0.7036181678214011, "grad_norm": 1.683445692062378, "learning_rate": 4.264235636489542e-06, "loss": 0.9874, "step": 3656 }, { "epoch": 0.703810623556582, "grad_norm": 1.6183136701583862, "learning_rate": 4.2591307153120335e-06, "loss": 0.9309, "step": 3657 }, { "epoch": 0.7040030792917629, "grad_norm": 1.6791754961013794, "learning_rate": 4.254028024706887e-06, "loss": 1.0304, "step": 3658 }, { "epoch": 0.7041955350269438, "grad_norm": 1.7862948179244995, "learning_rate": 4.248927566656717e-06, "loss": 0.8308, "step": 3659 }, { "epoch": 0.7043879907621247, "grad_norm": 1.8560540676116943, "learning_rate": 4.2438293431432665e-06, "loss": 0.8899, "step": 3660 }, { "epoch": 0.7045804464973057, "grad_norm": 1.8937824964523315, "learning_rate": 4.238733356147414e-06, "loss": 1.0277, "step": 3661 }, { "epoch": 0.7047729022324866, "grad_norm": 1.8942151069641113, "learning_rate": 4.233639607649168e-06, "loss": 0.9778, "step": 3662 }, { "epoch": 0.7049653579676675, "grad_norm": 1.9937940835952759, "learning_rate": 4.228548099627665e-06, "loss": 0.845, "step": 3663 }, { "epoch": 0.7051578137028484, "grad_norm": 1.7345616817474365, "learning_rate": 4.223458834061175e-06, "loss": 1.0314, "step": 3664 }, { "epoch": 0.7053502694380293, "grad_norm": 1.7550207376480103, "learning_rate": 4.218371812927092e-06, "loss": 0.874, "step": 3665 }, { "epoch": 0.7055427251732102, "grad_norm": 1.7918344736099243, "learning_rate": 4.213287038201943e-06, "loss": 0.8794, "step": 3666 }, { "epoch": 0.705735180908391, "grad_norm": 1.5447994470596313, "learning_rate": 4.208204511861379e-06, "loss": 0.936, "step": 3667 }, { "epoch": 0.7059276366435719, "grad_norm": 1.9018646478652954, "learning_rate": 4.203124235880179e-06, "loss": 0.9427, "step": 3668 }, { "epoch": 0.7061200923787528, "grad_norm": 1.8416558504104614, "learning_rate": 4.198046212232243e-06, "loss": 0.9258, "step": 3669 }, { "epoch": 0.7063125481139338, "grad_norm": 1.9400453567504883, "learning_rate": 4.192970442890602e-06, "loss": 0.9359, "step": 3670 }, { "epoch": 0.7065050038491147, "grad_norm": 1.727979302406311, "learning_rate": 4.187896929827414e-06, "loss": 0.8209, "step": 3671 }, { "epoch": 0.7066974595842956, "grad_norm": 0.8893999457359314, "learning_rate": 4.182825675013945e-06, "loss": 0.8929, "step": 3672 }, { "epoch": 0.7068899153194765, "grad_norm": 1.7074863910675049, "learning_rate": 4.1777566804206e-06, "loss": 0.8953, "step": 3673 }, { "epoch": 0.7070823710546574, "grad_norm": 1.8171278238296509, "learning_rate": 4.172689948016896e-06, "loss": 0.9017, "step": 3674 }, { "epoch": 0.7072748267898383, "grad_norm": 1.6707803010940552, "learning_rate": 4.167625479771476e-06, "loss": 0.9866, "step": 3675 }, { "epoch": 0.7074672825250192, "grad_norm": 1.6528441905975342, "learning_rate": 4.162563277652104e-06, "loss": 0.963, "step": 3676 }, { "epoch": 0.7076597382602001, "grad_norm": 1.8492140769958496, "learning_rate": 4.157503343625659e-06, "loss": 1.011, "step": 3677 }, { "epoch": 0.707852193995381, "grad_norm": 1.8094689846038818, "learning_rate": 4.152445679658139e-06, "loss": 0.8923, "step": 3678 }, { "epoch": 0.708044649730562, "grad_norm": 1.7770872116088867, "learning_rate": 4.147390287714667e-06, "loss": 1.012, "step": 3679 }, { "epoch": 0.7082371054657429, "grad_norm": 2.017263650894165, "learning_rate": 4.142337169759472e-06, "loss": 0.9438, "step": 3680 }, { "epoch": 0.7084295612009238, "grad_norm": 1.7695595026016235, "learning_rate": 4.137286327755913e-06, "loss": 0.9222, "step": 3681 }, { "epoch": 0.7086220169361047, "grad_norm": 1.8753374814987183, "learning_rate": 4.132237763666451e-06, "loss": 1.0347, "step": 3682 }, { "epoch": 0.7088144726712856, "grad_norm": 1.6318330764770508, "learning_rate": 4.1271914794526715e-06, "loss": 0.9587, "step": 3683 }, { "epoch": 0.7090069284064665, "grad_norm": 1.8075960874557495, "learning_rate": 4.12214747707527e-06, "loss": 1.0096, "step": 3684 }, { "epoch": 0.7091993841416474, "grad_norm": 1.816388487815857, "learning_rate": 4.117105758494056e-06, "loss": 0.9404, "step": 3685 }, { "epoch": 0.7093918398768283, "grad_norm": 1.9831154346466064, "learning_rate": 4.112066325667954e-06, "loss": 0.9439, "step": 3686 }, { "epoch": 0.7095842956120092, "grad_norm": 1.904625415802002, "learning_rate": 4.107029180554996e-06, "loss": 0.9102, "step": 3687 }, { "epoch": 0.7097767513471901, "grad_norm": 1.721761703491211, "learning_rate": 4.101994325112332e-06, "loss": 0.8735, "step": 3688 }, { "epoch": 0.7099692070823711, "grad_norm": 1.9413036108016968, "learning_rate": 4.096961761296211e-06, "loss": 0.8359, "step": 3689 }, { "epoch": 0.710161662817552, "grad_norm": 1.8073800802230835, "learning_rate": 4.0919314910620025e-06, "loss": 1.0395, "step": 3690 }, { "epoch": 0.7103541185527329, "grad_norm": 1.6554763317108154, "learning_rate": 4.086903516364179e-06, "loss": 0.9541, "step": 3691 }, { "epoch": 0.7105465742879138, "grad_norm": 1.6351912021636963, "learning_rate": 4.0818778391563255e-06, "loss": 0.906, "step": 3692 }, { "epoch": 0.7107390300230947, "grad_norm": 1.6374362707138062, "learning_rate": 4.07685446139113e-06, "loss": 0.9133, "step": 3693 }, { "epoch": 0.7109314857582756, "grad_norm": 1.7804003953933716, "learning_rate": 4.07183338502039e-06, "loss": 0.9571, "step": 3694 }, { "epoch": 0.7111239414934565, "grad_norm": 1.6902313232421875, "learning_rate": 4.066814611995006e-06, "loss": 0.9425, "step": 3695 }, { "epoch": 0.7113163972286374, "grad_norm": 1.5607993602752686, "learning_rate": 4.061798144264986e-06, "loss": 0.8847, "step": 3696 }, { "epoch": 0.7115088529638183, "grad_norm": 1.6833645105361938, "learning_rate": 4.056783983779442e-06, "loss": 0.9089, "step": 3697 }, { "epoch": 0.7117013086989993, "grad_norm": 1.716971755027771, "learning_rate": 4.051772132486589e-06, "loss": 0.8967, "step": 3698 }, { "epoch": 0.7118937644341802, "grad_norm": 2.139861822128296, "learning_rate": 4.046762592333744e-06, "loss": 0.8223, "step": 3699 }, { "epoch": 0.7120862201693611, "grad_norm": 1.748581886291504, "learning_rate": 4.041755365267323e-06, "loss": 0.964, "step": 3700 }, { "epoch": 0.712278675904542, "grad_norm": 1.8290610313415527, "learning_rate": 4.03675045323286e-06, "loss": 0.9766, "step": 3701 }, { "epoch": 0.7124711316397229, "grad_norm": 1.5929118394851685, "learning_rate": 4.0317478581749644e-06, "loss": 0.9271, "step": 3702 }, { "epoch": 0.7126635873749038, "grad_norm": 1.5934184789657593, "learning_rate": 4.02674758203736e-06, "loss": 0.9211, "step": 3703 }, { "epoch": 0.7128560431100847, "grad_norm": 1.7148069143295288, "learning_rate": 4.021749626762869e-06, "loss": 0.9533, "step": 3704 }, { "epoch": 0.7130484988452656, "grad_norm": 1.6647963523864746, "learning_rate": 4.01675399429341e-06, "loss": 0.9118, "step": 3705 }, { "epoch": 0.7132409545804465, "grad_norm": 1.5910618305206299, "learning_rate": 4.0117606865699975e-06, "loss": 0.9461, "step": 3706 }, { "epoch": 0.7134334103156275, "grad_norm": 1.747583270072937, "learning_rate": 4.006769705532746e-06, "loss": 0.9763, "step": 3707 }, { "epoch": 0.7136258660508084, "grad_norm": 1.812488079071045, "learning_rate": 4.001781053120863e-06, "loss": 0.8983, "step": 3708 }, { "epoch": 0.7138183217859893, "grad_norm": 1.724203109741211, "learning_rate": 3.996794731272653e-06, "loss": 0.8859, "step": 3709 }, { "epoch": 0.7140107775211701, "grad_norm": 1.9882450103759766, "learning_rate": 3.991810741925515e-06, "loss": 0.9857, "step": 3710 }, { "epoch": 0.714203233256351, "grad_norm": 1.897228717803955, "learning_rate": 3.986829087015941e-06, "loss": 0.94, "step": 3711 }, { "epoch": 0.7143956889915319, "grad_norm": 1.799200415611267, "learning_rate": 3.981849768479516e-06, "loss": 0.944, "step": 3712 }, { "epoch": 0.7145881447267128, "grad_norm": 1.6939918994903564, "learning_rate": 3.976872788250918e-06, "loss": 0.9655, "step": 3713 }, { "epoch": 0.7147806004618937, "grad_norm": 1.6385189294815063, "learning_rate": 3.971898148263915e-06, "loss": 0.901, "step": 3714 }, { "epoch": 0.7149730561970746, "grad_norm": 1.722093105316162, "learning_rate": 3.966925850451367e-06, "loss": 0.8844, "step": 3715 }, { "epoch": 0.7151655119322555, "grad_norm": 1.8588894605636597, "learning_rate": 3.961955896745224e-06, "loss": 0.8744, "step": 3716 }, { "epoch": 0.7153579676674365, "grad_norm": 1.6176855564117432, "learning_rate": 3.956988289076529e-06, "loss": 0.9144, "step": 3717 }, { "epoch": 0.7155504234026174, "grad_norm": 1.676416039466858, "learning_rate": 3.9520230293754e-06, "loss": 0.9076, "step": 3718 }, { "epoch": 0.7157428791377983, "grad_norm": 1.882345199584961, "learning_rate": 3.9470601195710575e-06, "loss": 0.8638, "step": 3719 }, { "epoch": 0.7159353348729792, "grad_norm": 1.7315891981124878, "learning_rate": 3.942099561591802e-06, "loss": 0.8926, "step": 3720 }, { "epoch": 0.7161277906081601, "grad_norm": 1.617348551750183, "learning_rate": 3.937141357365023e-06, "loss": 0.9027, "step": 3721 }, { "epoch": 0.716320246343341, "grad_norm": 1.6499409675598145, "learning_rate": 3.9321855088171934e-06, "loss": 0.8595, "step": 3722 }, { "epoch": 0.7165127020785219, "grad_norm": 1.598115086555481, "learning_rate": 3.927232017873872e-06, "loss": 0.9106, "step": 3723 }, { "epoch": 0.7167051578137028, "grad_norm": 2.043610095977783, "learning_rate": 3.922280886459701e-06, "loss": 0.978, "step": 3724 }, { "epoch": 0.7168976135488837, "grad_norm": 1.800264835357666, "learning_rate": 3.917332116498406e-06, "loss": 0.9518, "step": 3725 }, { "epoch": 0.7170900692840647, "grad_norm": 1.6401268243789673, "learning_rate": 3.912385709912794e-06, "loss": 0.9841, "step": 3726 }, { "epoch": 0.7172825250192456, "grad_norm": 1.624555230140686, "learning_rate": 3.9074416686247554e-06, "loss": 0.9558, "step": 3727 }, { "epoch": 0.7174749807544265, "grad_norm": 1.8923447132110596, "learning_rate": 3.902499994555261e-06, "loss": 0.9211, "step": 3728 }, { "epoch": 0.7176674364896074, "grad_norm": 1.8739162683486938, "learning_rate": 3.897560689624362e-06, "loss": 0.8901, "step": 3729 }, { "epoch": 0.7178598922247883, "grad_norm": 1.6195011138916016, "learning_rate": 3.892623755751189e-06, "loss": 0.9278, "step": 3730 }, { "epoch": 0.7180523479599692, "grad_norm": 1.9183586835861206, "learning_rate": 3.887689194853951e-06, "loss": 0.8873, "step": 3731 }, { "epoch": 0.7182448036951501, "grad_norm": 1.699321985244751, "learning_rate": 3.882757008849936e-06, "loss": 0.9206, "step": 3732 }, { "epoch": 0.718437259430331, "grad_norm": 1.6397508382797241, "learning_rate": 3.877827199655506e-06, "loss": 0.8955, "step": 3733 }, { "epoch": 0.7186297151655119, "grad_norm": 1.6325955390930176, "learning_rate": 3.872899769186107e-06, "loss": 0.9851, "step": 3734 }, { "epoch": 0.7188221709006929, "grad_norm": 0.9474324584007263, "learning_rate": 3.867974719356249e-06, "loss": 0.8754, "step": 3735 }, { "epoch": 0.7190146266358738, "grad_norm": 1.9148035049438477, "learning_rate": 3.8630520520795275e-06, "loss": 0.9366, "step": 3736 }, { "epoch": 0.7192070823710547, "grad_norm": 1.8077012300491333, "learning_rate": 3.858131769268607e-06, "loss": 0.898, "step": 3737 }, { "epoch": 0.7193995381062356, "grad_norm": 1.8866058588027954, "learning_rate": 3.853213872835229e-06, "loss": 0.8886, "step": 3738 }, { "epoch": 0.7195919938414165, "grad_norm": 1.6469823122024536, "learning_rate": 3.848298364690204e-06, "loss": 0.8924, "step": 3739 }, { "epoch": 0.7197844495765974, "grad_norm": 1.826834797859192, "learning_rate": 3.8433852467434175e-06, "loss": 0.8884, "step": 3740 }, { "epoch": 0.7199769053117783, "grad_norm": 1.6693713665008545, "learning_rate": 3.838474520903825e-06, "loss": 1.0084, "step": 3741 }, { "epoch": 0.7201693610469592, "grad_norm": 1.8152363300323486, "learning_rate": 3.833566189079453e-06, "loss": 0.921, "step": 3742 }, { "epoch": 0.7203618167821401, "grad_norm": 1.8509104251861572, "learning_rate": 3.8286602531773975e-06, "loss": 0.869, "step": 3743 }, { "epoch": 0.7205542725173211, "grad_norm": 1.786346673965454, "learning_rate": 3.823756715103822e-06, "loss": 0.8812, "step": 3744 }, { "epoch": 0.720746728252502, "grad_norm": 1.6753065586090088, "learning_rate": 3.8188555767639625e-06, "loss": 0.9816, "step": 3745 }, { "epoch": 0.7209391839876829, "grad_norm": 1.767048716545105, "learning_rate": 3.8139568400621184e-06, "loss": 0.8775, "step": 3746 }, { "epoch": 0.7211316397228638, "grad_norm": 1.6557140350341797, "learning_rate": 3.8090605069016596e-06, "loss": 0.9358, "step": 3747 }, { "epoch": 0.7213240954580447, "grad_norm": 1.8177536725997925, "learning_rate": 3.804166579185018e-06, "loss": 0.9841, "step": 3748 }, { "epoch": 0.7215165511932256, "grad_norm": 2.0585696697235107, "learning_rate": 3.799275058813694e-06, "loss": 0.9925, "step": 3749 }, { "epoch": 0.7217090069284064, "grad_norm": 1.5779823064804077, "learning_rate": 3.794385947688255e-06, "loss": 0.883, "step": 3750 }, { "epoch": 0.7219014626635873, "grad_norm": 2.012289524078369, "learning_rate": 3.7894992477083226e-06, "loss": 1.0255, "step": 3751 }, { "epoch": 0.7220939183987682, "grad_norm": 1.6775563955307007, "learning_rate": 3.78461496077259e-06, "loss": 0.91, "step": 3752 }, { "epoch": 0.7222863741339491, "grad_norm": 1.62815260887146, "learning_rate": 3.779733088778813e-06, "loss": 0.9628, "step": 3753 }, { "epoch": 0.7224788298691301, "grad_norm": 1.6503132581710815, "learning_rate": 3.774853633623806e-06, "loss": 0.9047, "step": 3754 }, { "epoch": 0.722671285604311, "grad_norm": 1.759012222290039, "learning_rate": 3.7699765972034442e-06, "loss": 0.941, "step": 3755 }, { "epoch": 0.7228637413394919, "grad_norm": 1.7295348644256592, "learning_rate": 3.7651019814126656e-06, "loss": 0.9034, "step": 3756 }, { "epoch": 0.7230561970746728, "grad_norm": 1.8059415817260742, "learning_rate": 3.7602297881454653e-06, "loss": 0.9262, "step": 3757 }, { "epoch": 0.7232486528098537, "grad_norm": 0.8651884198188782, "learning_rate": 3.755360019294898e-06, "loss": 0.8301, "step": 3758 }, { "epoch": 0.7234411085450346, "grad_norm": 1.821021556854248, "learning_rate": 3.750492676753077e-06, "loss": 0.8904, "step": 3759 }, { "epoch": 0.7236335642802155, "grad_norm": 1.7472025156021118, "learning_rate": 3.7456277624111725e-06, "loss": 0.9627, "step": 3760 }, { "epoch": 0.7238260200153964, "grad_norm": 1.7045828104019165, "learning_rate": 3.7407652781594094e-06, "loss": 0.9364, "step": 3761 }, { "epoch": 0.7240184757505773, "grad_norm": 1.9971729516983032, "learning_rate": 3.735905225887072e-06, "loss": 0.9446, "step": 3762 }, { "epoch": 0.7242109314857583, "grad_norm": 0.8987799882888794, "learning_rate": 3.7310476074824966e-06, "loss": 0.87, "step": 3763 }, { "epoch": 0.7244033872209392, "grad_norm": 1.666849970817566, "learning_rate": 3.726192424833075e-06, "loss": 0.9397, "step": 3764 }, { "epoch": 0.7245958429561201, "grad_norm": 1.865614891052246, "learning_rate": 3.7213396798252522e-06, "loss": 0.975, "step": 3765 }, { "epoch": 0.724788298691301, "grad_norm": 1.8458900451660156, "learning_rate": 3.7164893743445274e-06, "loss": 0.8818, "step": 3766 }, { "epoch": 0.7249807544264819, "grad_norm": 1.7316802740097046, "learning_rate": 3.7116415102754545e-06, "loss": 0.9247, "step": 3767 }, { "epoch": 0.7251732101616628, "grad_norm": 2.6089460849761963, "learning_rate": 3.7067960895016277e-06, "loss": 0.9083, "step": 3768 }, { "epoch": 0.7253656658968437, "grad_norm": 1.8129782676696777, "learning_rate": 3.701953113905703e-06, "loss": 0.8872, "step": 3769 }, { "epoch": 0.7255581216320246, "grad_norm": 1.6566029787063599, "learning_rate": 3.6971125853693834e-06, "loss": 0.9795, "step": 3770 }, { "epoch": 0.7257505773672055, "grad_norm": 1.6292799711227417, "learning_rate": 3.692274505773419e-06, "loss": 0.9571, "step": 3771 }, { "epoch": 0.7259430331023865, "grad_norm": 1.707647442817688, "learning_rate": 3.687438876997612e-06, "loss": 0.9549, "step": 3772 }, { "epoch": 0.7261354888375674, "grad_norm": 1.8775651454925537, "learning_rate": 3.6826057009208083e-06, "loss": 0.8891, "step": 3773 }, { "epoch": 0.7263279445727483, "grad_norm": 1.948901653289795, "learning_rate": 3.677774979420904e-06, "loss": 0.9716, "step": 3774 }, { "epoch": 0.7265204003079292, "grad_norm": 1.9079513549804688, "learning_rate": 3.67294671437484e-06, "loss": 0.9429, "step": 3775 }, { "epoch": 0.7267128560431101, "grad_norm": 1.9569562673568726, "learning_rate": 3.6681209076586035e-06, "loss": 0.9627, "step": 3776 }, { "epoch": 0.726905311778291, "grad_norm": 1.647851824760437, "learning_rate": 3.663297561147224e-06, "loss": 0.8567, "step": 3777 }, { "epoch": 0.7270977675134719, "grad_norm": 1.8918260335922241, "learning_rate": 3.6584766767147796e-06, "loss": 0.9834, "step": 3778 }, { "epoch": 0.7272902232486528, "grad_norm": 1.7165757417678833, "learning_rate": 3.6536582562343893e-06, "loss": 1.003, "step": 3779 }, { "epoch": 0.7274826789838337, "grad_norm": 1.7870110273361206, "learning_rate": 3.6488423015782128e-06, "loss": 0.9332, "step": 3780 }, { "epoch": 0.7276751347190146, "grad_norm": 1.7405707836151123, "learning_rate": 3.644028814617454e-06, "loss": 1.006, "step": 3781 }, { "epoch": 0.7278675904541956, "grad_norm": 1.7775375843048096, "learning_rate": 3.6392177972223596e-06, "loss": 0.889, "step": 3782 }, { "epoch": 0.7280600461893765, "grad_norm": 2.0872230529785156, "learning_rate": 3.6344092512622173e-06, "loss": 0.8771, "step": 3783 }, { "epoch": 0.7282525019245574, "grad_norm": 1.641983151435852, "learning_rate": 3.6296031786053455e-06, "loss": 0.8203, "step": 3784 }, { "epoch": 0.7284449576597383, "grad_norm": 1.9407005310058594, "learning_rate": 3.6247995811191107e-06, "loss": 0.8593, "step": 3785 }, { "epoch": 0.7286374133949192, "grad_norm": 1.8195254802703857, "learning_rate": 3.619998460669916e-06, "loss": 0.952, "step": 3786 }, { "epoch": 0.7288298691301001, "grad_norm": 1.7274514436721802, "learning_rate": 3.6151998191232005e-06, "loss": 0.8948, "step": 3787 }, { "epoch": 0.729022324865281, "grad_norm": 1.8413913249969482, "learning_rate": 3.610403658343443e-06, "loss": 0.9152, "step": 3788 }, { "epoch": 0.7292147806004619, "grad_norm": 1.648410677909851, "learning_rate": 3.6056099801941535e-06, "loss": 0.8771, "step": 3789 }, { "epoch": 0.7294072363356428, "grad_norm": 1.6907953023910522, "learning_rate": 3.600818786537883e-06, "loss": 0.922, "step": 3790 }, { "epoch": 0.7295996920708238, "grad_norm": 1.7866605520248413, "learning_rate": 3.5960300792362124e-06, "loss": 0.9481, "step": 3791 }, { "epoch": 0.7297921478060047, "grad_norm": 1.7889845371246338, "learning_rate": 3.591243860149759e-06, "loss": 0.9468, "step": 3792 }, { "epoch": 0.7299846035411856, "grad_norm": 1.9312835931777954, "learning_rate": 3.586460131138174e-06, "loss": 0.9935, "step": 3793 }, { "epoch": 0.7301770592763664, "grad_norm": 1.7008137702941895, "learning_rate": 3.5816788940601387e-06, "loss": 0.8846, "step": 3794 }, { "epoch": 0.7303695150115473, "grad_norm": 1.7842228412628174, "learning_rate": 3.576900150773368e-06, "loss": 0.9486, "step": 3795 }, { "epoch": 0.7305619707467282, "grad_norm": 1.7980836629867554, "learning_rate": 3.5721239031346067e-06, "loss": 0.911, "step": 3796 }, { "epoch": 0.7307544264819091, "grad_norm": 1.8791370391845703, "learning_rate": 3.5673501529996314e-06, "loss": 0.857, "step": 3797 }, { "epoch": 0.73094688221709, "grad_norm": 1.7578190565109253, "learning_rate": 3.5625789022232473e-06, "loss": 0.9594, "step": 3798 }, { "epoch": 0.7311393379522709, "grad_norm": 1.715129017829895, "learning_rate": 3.5578101526592877e-06, "loss": 0.9762, "step": 3799 }, { "epoch": 0.7313317936874519, "grad_norm": 1.7171292304992676, "learning_rate": 3.5530439061606202e-06, "loss": 0.932, "step": 3800 }, { "epoch": 0.7315242494226328, "grad_norm": 1.6815215349197388, "learning_rate": 3.5482801645791266e-06, "loss": 0.8634, "step": 3801 }, { "epoch": 0.7317167051578137, "grad_norm": 1.5651942491531372, "learning_rate": 3.5435189297657267e-06, "loss": 0.9067, "step": 3802 }, { "epoch": 0.7319091608929946, "grad_norm": 1.6557655334472656, "learning_rate": 3.5387602035703637e-06, "loss": 0.7843, "step": 3803 }, { "epoch": 0.7321016166281755, "grad_norm": 1.7977344989776611, "learning_rate": 3.534003987842005e-06, "loss": 0.8891, "step": 3804 }, { "epoch": 0.7322940723633564, "grad_norm": 0.8603512644767761, "learning_rate": 3.5292502844286447e-06, "loss": 0.8599, "step": 3805 }, { "epoch": 0.7324865280985373, "grad_norm": 1.7264420986175537, "learning_rate": 3.5244990951772972e-06, "loss": 0.9508, "step": 3806 }, { "epoch": 0.7326789838337182, "grad_norm": 1.6922509670257568, "learning_rate": 3.5197504219340027e-06, "loss": 0.9341, "step": 3807 }, { "epoch": 0.7328714395688991, "grad_norm": 1.9430891275405884, "learning_rate": 3.5150042665438233e-06, "loss": 0.9065, "step": 3808 }, { "epoch": 0.7330638953040801, "grad_norm": 1.6908938884735107, "learning_rate": 3.5102606308508423e-06, "loss": 0.9241, "step": 3809 }, { "epoch": 0.733256351039261, "grad_norm": 1.8099004030227661, "learning_rate": 3.505519516698165e-06, "loss": 0.8306, "step": 3810 }, { "epoch": 0.7334488067744419, "grad_norm": 1.9020333290100098, "learning_rate": 3.5007809259279146e-06, "loss": 0.9494, "step": 3811 }, { "epoch": 0.7336412625096228, "grad_norm": 2.0368494987487793, "learning_rate": 3.496044860381238e-06, "loss": 0.9005, "step": 3812 }, { "epoch": 0.7338337182448037, "grad_norm": 1.6994282007217407, "learning_rate": 3.4913113218982965e-06, "loss": 0.9525, "step": 3813 }, { "epoch": 0.7340261739799846, "grad_norm": 1.8050298690795898, "learning_rate": 3.486580312318273e-06, "loss": 0.8639, "step": 3814 }, { "epoch": 0.7342186297151655, "grad_norm": 1.8238089084625244, "learning_rate": 3.481851833479365e-06, "loss": 0.895, "step": 3815 }, { "epoch": 0.7344110854503464, "grad_norm": 1.8572880029678345, "learning_rate": 3.4771258872187917e-06, "loss": 0.873, "step": 3816 }, { "epoch": 0.7346035411855273, "grad_norm": 1.7538968324661255, "learning_rate": 3.472402475372778e-06, "loss": 1.0154, "step": 3817 }, { "epoch": 0.7347959969207082, "grad_norm": 1.7000271081924438, "learning_rate": 3.4676815997765756e-06, "loss": 0.9607, "step": 3818 }, { "epoch": 0.7349884526558892, "grad_norm": 1.8036025762557983, "learning_rate": 3.4629632622644437e-06, "loss": 0.9802, "step": 3819 }, { "epoch": 0.7351809083910701, "grad_norm": 2.039238929748535, "learning_rate": 3.4582474646696575e-06, "loss": 0.9385, "step": 3820 }, { "epoch": 0.735373364126251, "grad_norm": 1.794670581817627, "learning_rate": 3.453534208824507e-06, "loss": 0.9119, "step": 3821 }, { "epoch": 0.7355658198614319, "grad_norm": 1.695103406906128, "learning_rate": 3.448823496560292e-06, "loss": 0.929, "step": 3822 }, { "epoch": 0.7357582755966128, "grad_norm": 1.6967803239822388, "learning_rate": 3.444115329707325e-06, "loss": 0.802, "step": 3823 }, { "epoch": 0.7359507313317937, "grad_norm": 1.710740089416504, "learning_rate": 3.4394097100949286e-06, "loss": 0.8682, "step": 3824 }, { "epoch": 0.7361431870669746, "grad_norm": 1.7392624616622925, "learning_rate": 3.4347066395514383e-06, "loss": 0.9493, "step": 3825 }, { "epoch": 0.7363356428021555, "grad_norm": 1.8060885667800903, "learning_rate": 3.4300061199041967e-06, "loss": 0.9086, "step": 3826 }, { "epoch": 0.7365280985373364, "grad_norm": 1.7745487689971924, "learning_rate": 3.425308152979555e-06, "loss": 1.022, "step": 3827 }, { "epoch": 0.7367205542725174, "grad_norm": 1.6979912519454956, "learning_rate": 3.4206127406028744e-06, "loss": 0.9077, "step": 3828 }, { "epoch": 0.7369130100076983, "grad_norm": 1.6242289543151855, "learning_rate": 3.415919884598523e-06, "loss": 0.9208, "step": 3829 }, { "epoch": 0.7371054657428792, "grad_norm": 2.163660764694214, "learning_rate": 3.411229586789876e-06, "loss": 0.9002, "step": 3830 }, { "epoch": 0.7372979214780601, "grad_norm": 2.0429024696350098, "learning_rate": 3.4065418489993118e-06, "loss": 0.8961, "step": 3831 }, { "epoch": 0.737490377213241, "grad_norm": 2.788419723510742, "learning_rate": 3.401856673048217e-06, "loss": 0.8561, "step": 3832 }, { "epoch": 0.7376828329484219, "grad_norm": 1.6912580728530884, "learning_rate": 3.3971740607569868e-06, "loss": 0.9493, "step": 3833 }, { "epoch": 0.7378752886836027, "grad_norm": 1.8332105875015259, "learning_rate": 3.392494013945009e-06, "loss": 0.8532, "step": 3834 }, { "epoch": 0.7380677444187836, "grad_norm": 2.000765085220337, "learning_rate": 3.3878165344306824e-06, "loss": 0.8381, "step": 3835 }, { "epoch": 0.7382602001539645, "grad_norm": 1.7117446660995483, "learning_rate": 3.3831416240314085e-06, "loss": 0.9259, "step": 3836 }, { "epoch": 0.7384526558891455, "grad_norm": 1.6958849430084229, "learning_rate": 3.378469284563589e-06, "loss": 0.9746, "step": 3837 }, { "epoch": 0.7386451116243264, "grad_norm": 1.8388540744781494, "learning_rate": 3.3737995178426276e-06, "loss": 0.8761, "step": 3838 }, { "epoch": 0.7388375673595073, "grad_norm": 1.7091138362884521, "learning_rate": 3.3691323256829266e-06, "loss": 0.8853, "step": 3839 }, { "epoch": 0.7390300230946882, "grad_norm": 1.7589246034622192, "learning_rate": 3.3644677098978894e-06, "loss": 0.9005, "step": 3840 }, { "epoch": 0.7392224788298691, "grad_norm": 1.796491026878357, "learning_rate": 3.3598056722999185e-06, "loss": 0.8818, "step": 3841 }, { "epoch": 0.73941493456505, "grad_norm": 1.8350884914398193, "learning_rate": 3.3551462147004132e-06, "loss": 0.7787, "step": 3842 }, { "epoch": 0.7396073903002309, "grad_norm": 1.7753305435180664, "learning_rate": 3.3504893389097727e-06, "loss": 0.8643, "step": 3843 }, { "epoch": 0.7397998460354118, "grad_norm": 1.9164499044418335, "learning_rate": 3.3458350467373914e-06, "loss": 0.946, "step": 3844 }, { "epoch": 0.7399923017705927, "grad_norm": 1.8525071144104004, "learning_rate": 3.3411833399916584e-06, "loss": 0.8374, "step": 3845 }, { "epoch": 0.7401847575057737, "grad_norm": 1.6794849634170532, "learning_rate": 3.3365342204799613e-06, "loss": 0.8592, "step": 3846 }, { "epoch": 0.7403772132409546, "grad_norm": 1.6253608465194702, "learning_rate": 3.3318876900086815e-06, "loss": 0.9037, "step": 3847 }, { "epoch": 0.7405696689761355, "grad_norm": 1.7628328800201416, "learning_rate": 3.3272437503831945e-06, "loss": 0.9394, "step": 3848 }, { "epoch": 0.7407621247113164, "grad_norm": 0.912964403629303, "learning_rate": 3.3226024034078698e-06, "loss": 0.8613, "step": 3849 }, { "epoch": 0.7409545804464973, "grad_norm": 1.6622869968414307, "learning_rate": 3.3179636508860645e-06, "loss": 0.9551, "step": 3850 }, { "epoch": 0.7411470361816782, "grad_norm": 1.7354791164398193, "learning_rate": 3.3133274946201333e-06, "loss": 0.8849, "step": 3851 }, { "epoch": 0.7413394919168591, "grad_norm": 1.8365917205810547, "learning_rate": 3.308693936411421e-06, "loss": 0.8992, "step": 3852 }, { "epoch": 0.74153194765204, "grad_norm": 1.6966023445129395, "learning_rate": 3.304062978060263e-06, "loss": 0.9602, "step": 3853 }, { "epoch": 0.7417244033872209, "grad_norm": 1.7813916206359863, "learning_rate": 3.2994346213659824e-06, "loss": 0.9592, "step": 3854 }, { "epoch": 0.7419168591224018, "grad_norm": 1.9294930696487427, "learning_rate": 3.294808868126894e-06, "loss": 0.9342, "step": 3855 }, { "epoch": 0.7421093148575828, "grad_norm": 1.6844215393066406, "learning_rate": 3.290185720140301e-06, "loss": 0.977, "step": 3856 }, { "epoch": 0.7423017705927637, "grad_norm": 1.6281991004943848, "learning_rate": 3.285565179202491e-06, "loss": 0.9251, "step": 3857 }, { "epoch": 0.7424942263279446, "grad_norm": 2.1437790393829346, "learning_rate": 3.280947247108741e-06, "loss": 0.9856, "step": 3858 }, { "epoch": 0.7426866820631255, "grad_norm": 1.772961974143982, "learning_rate": 3.2763319256533177e-06, "loss": 0.8713, "step": 3859 }, { "epoch": 0.7428791377983064, "grad_norm": 1.7907110452651978, "learning_rate": 3.2717192166294685e-06, "loss": 0.9037, "step": 3860 }, { "epoch": 0.7430715935334873, "grad_norm": 1.7244598865509033, "learning_rate": 3.267109121829428e-06, "loss": 0.9412, "step": 3861 }, { "epoch": 0.7432640492686682, "grad_norm": 1.6671468019485474, "learning_rate": 3.262501643044417e-06, "loss": 0.9451, "step": 3862 }, { "epoch": 0.7434565050038491, "grad_norm": 1.7434862852096558, "learning_rate": 3.2578967820646313e-06, "loss": 0.8162, "step": 3863 }, { "epoch": 0.74364896073903, "grad_norm": 1.858473300933838, "learning_rate": 3.2532945406792573e-06, "loss": 0.8683, "step": 3864 }, { "epoch": 0.743841416474211, "grad_norm": 1.5833643674850464, "learning_rate": 3.248694920676464e-06, "loss": 0.8474, "step": 3865 }, { "epoch": 0.7440338722093919, "grad_norm": 1.7272731065750122, "learning_rate": 3.2440979238433977e-06, "loss": 0.9015, "step": 3866 }, { "epoch": 0.7442263279445728, "grad_norm": 1.8200198411941528, "learning_rate": 3.23950355196619e-06, "loss": 0.9288, "step": 3867 }, { "epoch": 0.7444187836797537, "grad_norm": 1.9386030435562134, "learning_rate": 3.234911806829948e-06, "loss": 0.9572, "step": 3868 }, { "epoch": 0.7446112394149346, "grad_norm": 1.8049795627593994, "learning_rate": 3.2303226902187625e-06, "loss": 0.89, "step": 3869 }, { "epoch": 0.7448036951501155, "grad_norm": 1.7297941446304321, "learning_rate": 3.225736203915698e-06, "loss": 0.917, "step": 3870 }, { "epoch": 0.7449961508852964, "grad_norm": 1.630277156829834, "learning_rate": 3.221152349702802e-06, "loss": 0.9119, "step": 3871 }, { "epoch": 0.7451886066204773, "grad_norm": 1.790122389793396, "learning_rate": 3.216571129361097e-06, "loss": 0.9335, "step": 3872 }, { "epoch": 0.7453810623556582, "grad_norm": 1.9039274454116821, "learning_rate": 3.2119925446705824e-06, "loss": 0.9947, "step": 3873 }, { "epoch": 0.7455735180908392, "grad_norm": 1.7523843050003052, "learning_rate": 3.2074165974102324e-06, "loss": 0.8743, "step": 3874 }, { "epoch": 0.7457659738260201, "grad_norm": 1.9392461776733398, "learning_rate": 3.2028432893579975e-06, "loss": 0.9609, "step": 3875 }, { "epoch": 0.745958429561201, "grad_norm": 1.714680552482605, "learning_rate": 3.1982726222908046e-06, "loss": 0.9292, "step": 3876 }, { "epoch": 0.7461508852963818, "grad_norm": 0.9127395749092102, "learning_rate": 3.1937045979845518e-06, "loss": 0.9263, "step": 3877 }, { "epoch": 0.7463433410315627, "grad_norm": 1.8035203218460083, "learning_rate": 3.1891392182141103e-06, "loss": 0.8232, "step": 3878 }, { "epoch": 0.7465357967667436, "grad_norm": 0.8980305790901184, "learning_rate": 3.1845764847533276e-06, "loss": 0.9004, "step": 3879 }, { "epoch": 0.7467282525019245, "grad_norm": 1.8180233240127563, "learning_rate": 3.1800163993750166e-06, "loss": 0.8533, "step": 3880 }, { "epoch": 0.7469207082371054, "grad_norm": 1.8277109861373901, "learning_rate": 3.1754589638509647e-06, "loss": 1.0293, "step": 3881 }, { "epoch": 0.7471131639722863, "grad_norm": 1.5936766862869263, "learning_rate": 3.1709041799519312e-06, "loss": 0.9378, "step": 3882 }, { "epoch": 0.7473056197074672, "grad_norm": 1.9058517217636108, "learning_rate": 3.166352049447644e-06, "loss": 0.9172, "step": 3883 }, { "epoch": 0.7474980754426482, "grad_norm": 1.6563794612884521, "learning_rate": 3.161802574106799e-06, "loss": 0.9639, "step": 3884 }, { "epoch": 0.7476905311778291, "grad_norm": 1.872847557067871, "learning_rate": 3.157255755697062e-06, "loss": 0.962, "step": 3885 }, { "epoch": 0.74788298691301, "grad_norm": 1.8390828371047974, "learning_rate": 3.152711595985065e-06, "loss": 0.9562, "step": 3886 }, { "epoch": 0.7480754426481909, "grad_norm": 1.8095601797103882, "learning_rate": 3.148170096736408e-06, "loss": 0.9127, "step": 3887 }, { "epoch": 0.7482678983833718, "grad_norm": 1.8260071277618408, "learning_rate": 3.143631259715658e-06, "loss": 0.9756, "step": 3888 }, { "epoch": 0.7484603541185527, "grad_norm": 1.7314852476119995, "learning_rate": 3.1390950866863456e-06, "loss": 0.9152, "step": 3889 }, { "epoch": 0.7486528098537336, "grad_norm": 1.79062819480896, "learning_rate": 3.134561579410966e-06, "loss": 0.9372, "step": 3890 }, { "epoch": 0.7488452655889145, "grad_norm": 1.6281756162643433, "learning_rate": 3.1300307396509833e-06, "loss": 0.8774, "step": 3891 }, { "epoch": 0.7490377213240954, "grad_norm": 1.8487001657485962, "learning_rate": 3.1255025691668184e-06, "loss": 1.0515, "step": 3892 }, { "epoch": 0.7492301770592764, "grad_norm": 1.832372784614563, "learning_rate": 3.120977069717861e-06, "loss": 0.9553, "step": 3893 }, { "epoch": 0.7494226327944573, "grad_norm": 1.708930253982544, "learning_rate": 3.116454243062459e-06, "loss": 0.7958, "step": 3894 }, { "epoch": 0.7496150885296382, "grad_norm": 1.6150094270706177, "learning_rate": 3.1119340909579265e-06, "loss": 0.9373, "step": 3895 }, { "epoch": 0.7498075442648191, "grad_norm": 1.8417260646820068, "learning_rate": 3.10741661516053e-06, "loss": 0.9557, "step": 3896 }, { "epoch": 0.75, "grad_norm": 1.649986982345581, "learning_rate": 3.102901817425504e-06, "loss": 0.9166, "step": 3897 }, { "epoch": 0.7501924557351809, "grad_norm": 2.2120673656463623, "learning_rate": 3.0983896995070397e-06, "loss": 0.9447, "step": 3898 }, { "epoch": 0.7503849114703618, "grad_norm": 1.6997156143188477, "learning_rate": 3.0938802631582865e-06, "loss": 0.9159, "step": 3899 }, { "epoch": 0.7505773672055427, "grad_norm": 1.703776240348816, "learning_rate": 3.089373510131354e-06, "loss": 0.8673, "step": 3900 }, { "epoch": 0.7507698229407236, "grad_norm": 1.7422881126403809, "learning_rate": 3.0848694421773075e-06, "loss": 0.8631, "step": 3901 }, { "epoch": 0.7509622786759046, "grad_norm": 1.8156569004058838, "learning_rate": 3.0803680610461674e-06, "loss": 0.9026, "step": 3902 }, { "epoch": 0.7511547344110855, "grad_norm": 1.7419483661651611, "learning_rate": 3.0758693684869147e-06, "loss": 0.8985, "step": 3903 }, { "epoch": 0.7513471901462664, "grad_norm": 1.647157073020935, "learning_rate": 3.071373366247482e-06, "loss": 0.9147, "step": 3904 }, { "epoch": 0.7515396458814473, "grad_norm": 1.7883245944976807, "learning_rate": 3.066880056074758e-06, "loss": 0.9871, "step": 3905 }, { "epoch": 0.7517321016166282, "grad_norm": 1.7706633806228638, "learning_rate": 3.0623894397145837e-06, "loss": 0.9311, "step": 3906 }, { "epoch": 0.7519245573518091, "grad_norm": 1.4799484014511108, "learning_rate": 3.057901518911758e-06, "loss": 0.8779, "step": 3907 }, { "epoch": 0.75211701308699, "grad_norm": 1.6384867429733276, "learning_rate": 3.0534162954100264e-06, "loss": 0.867, "step": 3908 }, { "epoch": 0.7523094688221709, "grad_norm": 1.7988359928131104, "learning_rate": 3.0489337709520918e-06, "loss": 0.9392, "step": 3909 }, { "epoch": 0.7525019245573518, "grad_norm": 1.5843770503997803, "learning_rate": 3.044453947279603e-06, "loss": 0.9676, "step": 3910 }, { "epoch": 0.7526943802925328, "grad_norm": 1.7816550731658936, "learning_rate": 3.0399768261331664e-06, "loss": 0.9022, "step": 3911 }, { "epoch": 0.7528868360277137, "grad_norm": 1.7060253620147705, "learning_rate": 3.0355024092523334e-06, "loss": 0.9055, "step": 3912 }, { "epoch": 0.7530792917628946, "grad_norm": 1.752164363861084, "learning_rate": 3.031030698375602e-06, "loss": 0.9138, "step": 3913 }, { "epoch": 0.7532717474980755, "grad_norm": 0.8799826502799988, "learning_rate": 3.0265616952404252e-06, "loss": 0.8513, "step": 3914 }, { "epoch": 0.7534642032332564, "grad_norm": 1.756686806678772, "learning_rate": 3.0220954015832004e-06, "loss": 0.7984, "step": 3915 }, { "epoch": 0.7536566589684373, "grad_norm": 1.8549926280975342, "learning_rate": 3.017631819139273e-06, "loss": 0.9236, "step": 3916 }, { "epoch": 0.7538491147036182, "grad_norm": 1.6138970851898193, "learning_rate": 3.013170949642935e-06, "loss": 0.9387, "step": 3917 }, { "epoch": 0.754041570438799, "grad_norm": 1.71934175491333, "learning_rate": 3.0087127948274264e-06, "loss": 0.9936, "step": 3918 }, { "epoch": 0.7542340261739799, "grad_norm": 1.874595046043396, "learning_rate": 3.004257356424927e-06, "loss": 0.9787, "step": 3919 }, { "epoch": 0.7544264819091608, "grad_norm": 1.5763262510299683, "learning_rate": 2.999804636166567e-06, "loss": 0.8968, "step": 3920 }, { "epoch": 0.7546189376443418, "grad_norm": 1.7726876735687256, "learning_rate": 2.995354635782417e-06, "loss": 0.9688, "step": 3921 }, { "epoch": 0.7548113933795227, "grad_norm": 1.554753065109253, "learning_rate": 2.990907357001491e-06, "loss": 0.8993, "step": 3922 }, { "epoch": 0.7550038491147036, "grad_norm": 1.9701539278030396, "learning_rate": 2.986462801551748e-06, "loss": 1.083, "step": 3923 }, { "epoch": 0.7551963048498845, "grad_norm": 1.8488991260528564, "learning_rate": 2.9820209711600858e-06, "loss": 0.9261, "step": 3924 }, { "epoch": 0.7553887605850654, "grad_norm": 1.914576768875122, "learning_rate": 2.977581867552346e-06, "loss": 0.9466, "step": 3925 }, { "epoch": 0.7555812163202463, "grad_norm": 1.726682186126709, "learning_rate": 2.9731454924533086e-06, "loss": 0.9345, "step": 3926 }, { "epoch": 0.7557736720554272, "grad_norm": 1.7761739492416382, "learning_rate": 2.9687118475866948e-06, "loss": 0.8944, "step": 3927 }, { "epoch": 0.7559661277906081, "grad_norm": 1.7333637475967407, "learning_rate": 2.9642809346751677e-06, "loss": 0.8857, "step": 3928 }, { "epoch": 0.756158583525789, "grad_norm": 1.7083461284637451, "learning_rate": 2.9598527554403187e-06, "loss": 0.8622, "step": 3929 }, { "epoch": 0.75635103926097, "grad_norm": 1.7392387390136719, "learning_rate": 2.9554273116026887e-06, "loss": 0.8689, "step": 3930 }, { "epoch": 0.7565434949961509, "grad_norm": 1.740929365158081, "learning_rate": 2.95100460488175e-06, "loss": 1.0077, "step": 3931 }, { "epoch": 0.7567359507313318, "grad_norm": 1.6594079732894897, "learning_rate": 2.9465846369959126e-06, "loss": 0.8918, "step": 3932 }, { "epoch": 0.7569284064665127, "grad_norm": 1.7169567346572876, "learning_rate": 2.942167409662524e-06, "loss": 0.9368, "step": 3933 }, { "epoch": 0.7571208622016936, "grad_norm": 1.7319214344024658, "learning_rate": 2.937752924597864e-06, "loss": 0.8963, "step": 3934 }, { "epoch": 0.7573133179368745, "grad_norm": 1.6905615329742432, "learning_rate": 2.933341183517148e-06, "loss": 1.0085, "step": 3935 }, { "epoch": 0.7575057736720554, "grad_norm": 1.7468857765197754, "learning_rate": 2.9289321881345257e-06, "loss": 0.869, "step": 3936 }, { "epoch": 0.7576982294072363, "grad_norm": 1.753749966621399, "learning_rate": 2.92452594016308e-06, "loss": 0.9221, "step": 3937 }, { "epoch": 0.7578906851424172, "grad_norm": 1.6252354383468628, "learning_rate": 2.9201224413148254e-06, "loss": 0.938, "step": 3938 }, { "epoch": 0.7580831408775982, "grad_norm": 1.675720453262329, "learning_rate": 2.9157216933007095e-06, "loss": 0.9196, "step": 3939 }, { "epoch": 0.7582755966127791, "grad_norm": 2.069101333618164, "learning_rate": 2.91132369783061e-06, "loss": 0.8679, "step": 3940 }, { "epoch": 0.75846805234796, "grad_norm": 1.7233712673187256, "learning_rate": 2.906928456613336e-06, "loss": 0.8973, "step": 3941 }, { "epoch": 0.7586605080831409, "grad_norm": 1.6812355518341064, "learning_rate": 2.902535971356625e-06, "loss": 0.8915, "step": 3942 }, { "epoch": 0.7588529638183218, "grad_norm": 1.6375269889831543, "learning_rate": 2.898146243767146e-06, "loss": 0.9764, "step": 3943 }, { "epoch": 0.7590454195535027, "grad_norm": 1.6960278749465942, "learning_rate": 2.893759275550494e-06, "loss": 0.9476, "step": 3944 }, { "epoch": 0.7592378752886836, "grad_norm": 1.8152192831039429, "learning_rate": 2.8893750684111977e-06, "loss": 0.9444, "step": 3945 }, { "epoch": 0.7594303310238645, "grad_norm": 1.741206407546997, "learning_rate": 2.884993624052701e-06, "loss": 0.9198, "step": 3946 }, { "epoch": 0.7596227867590454, "grad_norm": 0.881798505783081, "learning_rate": 2.8806149441773844e-06, "loss": 0.8835, "step": 3947 }, { "epoch": 0.7598152424942263, "grad_norm": 0.8986806273460388, "learning_rate": 2.876239030486554e-06, "loss": 0.9013, "step": 3948 }, { "epoch": 0.7600076982294073, "grad_norm": 0.8570490479469299, "learning_rate": 2.871865884680436e-06, "loss": 0.8727, "step": 3949 }, { "epoch": 0.7602001539645882, "grad_norm": 1.829601526260376, "learning_rate": 2.867495508458186e-06, "loss": 0.9147, "step": 3950 }, { "epoch": 0.7603926096997691, "grad_norm": 0.897922694683075, "learning_rate": 2.8631279035178796e-06, "loss": 0.9246, "step": 3951 }, { "epoch": 0.76058506543495, "grad_norm": 1.974184274673462, "learning_rate": 2.8587630715565185e-06, "loss": 0.8711, "step": 3952 }, { "epoch": 0.7607775211701309, "grad_norm": 1.671872854232788, "learning_rate": 2.8544010142700263e-06, "loss": 0.9809, "step": 3953 }, { "epoch": 0.7609699769053118, "grad_norm": 1.9260306358337402, "learning_rate": 2.850041733353247e-06, "loss": 0.8332, "step": 3954 }, { "epoch": 0.7611624326404927, "grad_norm": 1.8911657333374023, "learning_rate": 2.845685230499948e-06, "loss": 0.9511, "step": 3955 }, { "epoch": 0.7613548883756736, "grad_norm": 1.7182097434997559, "learning_rate": 2.8413315074028157e-06, "loss": 0.91, "step": 3956 }, { "epoch": 0.7615473441108545, "grad_norm": 1.6033059358596802, "learning_rate": 2.8369805657534576e-06, "loss": 0.9181, "step": 3957 }, { "epoch": 0.7617397998460355, "grad_norm": 1.63260817527771, "learning_rate": 2.8326324072423994e-06, "loss": 0.9893, "step": 3958 }, { "epoch": 0.7619322555812164, "grad_norm": 1.6555876731872559, "learning_rate": 2.828287033559086e-06, "loss": 0.9059, "step": 3959 }, { "epoch": 0.7621247113163973, "grad_norm": 1.7372273206710815, "learning_rate": 2.823944446391881e-06, "loss": 1.0009, "step": 3960 }, { "epoch": 0.7623171670515781, "grad_norm": 1.7335221767425537, "learning_rate": 2.819604647428067e-06, "loss": 0.9716, "step": 3961 }, { "epoch": 0.762509622786759, "grad_norm": 1.7508633136749268, "learning_rate": 2.8152676383538346e-06, "loss": 0.8777, "step": 3962 }, { "epoch": 0.7627020785219399, "grad_norm": 1.720937728881836, "learning_rate": 2.8109334208542994e-06, "loss": 0.9, "step": 3963 }, { "epoch": 0.7628945342571208, "grad_norm": 2.0624163150787354, "learning_rate": 2.8066019966134907e-06, "loss": 0.9414, "step": 3964 }, { "epoch": 0.7630869899923017, "grad_norm": 0.9367794990539551, "learning_rate": 2.8022733673143508e-06, "loss": 0.9215, "step": 3965 }, { "epoch": 0.7632794457274826, "grad_norm": 1.7649296522140503, "learning_rate": 2.7979475346387363e-06, "loss": 0.8565, "step": 3966 }, { "epoch": 0.7634719014626636, "grad_norm": 1.7157323360443115, "learning_rate": 2.793624500267418e-06, "loss": 0.9744, "step": 3967 }, { "epoch": 0.7636643571978445, "grad_norm": 1.694903016090393, "learning_rate": 2.7893042658800793e-06, "loss": 1.0297, "step": 3968 }, { "epoch": 0.7638568129330254, "grad_norm": 1.7097408771514893, "learning_rate": 2.784986833155314e-06, "loss": 0.9826, "step": 3969 }, { "epoch": 0.7640492686682063, "grad_norm": 0.8642791509628296, "learning_rate": 2.78067220377063e-06, "loss": 0.901, "step": 3970 }, { "epoch": 0.7642417244033872, "grad_norm": 2.1903867721557617, "learning_rate": 2.776360379402445e-06, "loss": 0.8755, "step": 3971 }, { "epoch": 0.7644341801385681, "grad_norm": 1.8496774435043335, "learning_rate": 2.7720513617260857e-06, "loss": 0.9548, "step": 3972 }, { "epoch": 0.764626635873749, "grad_norm": 1.7211551666259766, "learning_rate": 2.7677451524157882e-06, "loss": 0.9005, "step": 3973 }, { "epoch": 0.7648190916089299, "grad_norm": 1.739711880683899, "learning_rate": 2.763441753144701e-06, "loss": 0.8402, "step": 3974 }, { "epoch": 0.7650115473441108, "grad_norm": 1.7368019819259644, "learning_rate": 2.7591411655848755e-06, "loss": 0.9762, "step": 3975 }, { "epoch": 0.7652040030792918, "grad_norm": 1.6665337085723877, "learning_rate": 2.7548433914072736e-06, "loss": 0.932, "step": 3976 }, { "epoch": 0.7653964588144727, "grad_norm": 1.7674028873443604, "learning_rate": 2.7505484322817656e-06, "loss": 0.9717, "step": 3977 }, { "epoch": 0.7655889145496536, "grad_norm": 1.7991665601730347, "learning_rate": 2.746256289877126e-06, "loss": 0.8936, "step": 3978 }, { "epoch": 0.7657813702848345, "grad_norm": 1.9144983291625977, "learning_rate": 2.741966965861033e-06, "loss": 0.8645, "step": 3979 }, { "epoch": 0.7659738260200154, "grad_norm": 1.9745656251907349, "learning_rate": 2.7376804619000706e-06, "loss": 0.913, "step": 3980 }, { "epoch": 0.7661662817551963, "grad_norm": 1.7380291223526, "learning_rate": 2.7333967796597317e-06, "loss": 0.8787, "step": 3981 }, { "epoch": 0.7663587374903772, "grad_norm": 1.8387010097503662, "learning_rate": 2.729115920804408e-06, "loss": 0.8477, "step": 3982 }, { "epoch": 0.7665511932255581, "grad_norm": 1.901106834411621, "learning_rate": 2.7248378869973946e-06, "loss": 0.8607, "step": 3983 }, { "epoch": 0.766743648960739, "grad_norm": 1.633284568786621, "learning_rate": 2.720562679900892e-06, "loss": 0.8887, "step": 3984 }, { "epoch": 0.7669361046959199, "grad_norm": 1.9528007507324219, "learning_rate": 2.716290301175999e-06, "loss": 0.8147, "step": 3985 }, { "epoch": 0.7671285604311009, "grad_norm": 1.8058927059173584, "learning_rate": 2.712020752482717e-06, "loss": 0.9548, "step": 3986 }, { "epoch": 0.7673210161662818, "grad_norm": 1.6151124238967896, "learning_rate": 2.707754035479948e-06, "loss": 0.8966, "step": 3987 }, { "epoch": 0.7675134719014627, "grad_norm": 1.8441987037658691, "learning_rate": 2.703490151825492e-06, "loss": 0.9145, "step": 3988 }, { "epoch": 0.7677059276366436, "grad_norm": 1.7653534412384033, "learning_rate": 2.699229103176052e-06, "loss": 0.9683, "step": 3989 }, { "epoch": 0.7678983833718245, "grad_norm": 1.7361457347869873, "learning_rate": 2.694970891187225e-06, "loss": 0.9207, "step": 3990 }, { "epoch": 0.7680908391070054, "grad_norm": 1.625246286392212, "learning_rate": 2.690715517513508e-06, "loss": 0.9591, "step": 3991 }, { "epoch": 0.7682832948421863, "grad_norm": 1.8868553638458252, "learning_rate": 2.6864629838082957e-06, "loss": 0.872, "step": 3992 }, { "epoch": 0.7684757505773672, "grad_norm": 0.9860090017318726, "learning_rate": 2.6822132917238785e-06, "loss": 0.9603, "step": 3993 }, { "epoch": 0.7686682063125481, "grad_norm": 1.738326072692871, "learning_rate": 2.6779664429114448e-06, "loss": 0.8916, "step": 3994 }, { "epoch": 0.7688606620477291, "grad_norm": 1.8191266059875488, "learning_rate": 2.6737224390210726e-06, "loss": 0.9709, "step": 3995 }, { "epoch": 0.76905311778291, "grad_norm": 0.8872901797294617, "learning_rate": 2.669481281701739e-06, "loss": 0.8836, "step": 3996 }, { "epoch": 0.7692455735180909, "grad_norm": 1.7487057447433472, "learning_rate": 2.6652429726013162e-06, "loss": 0.895, "step": 3997 }, { "epoch": 0.7694380292532718, "grad_norm": 2.2640326023101807, "learning_rate": 2.661007513366567e-06, "loss": 0.938, "step": 3998 }, { "epoch": 0.7696304849884527, "grad_norm": 1.662664532661438, "learning_rate": 2.656774905643147e-06, "loss": 0.9058, "step": 3999 }, { "epoch": 0.7698229407236336, "grad_norm": 1.6709994077682495, "learning_rate": 2.652545151075606e-06, "loss": 0.8533, "step": 4000 }, { "epoch": 0.7700153964588144, "grad_norm": 1.726743221282959, "learning_rate": 2.6483182513073835e-06, "loss": 0.8763, "step": 4001 }, { "epoch": 0.7702078521939953, "grad_norm": 1.742574691772461, "learning_rate": 2.6440942079808096e-06, "loss": 0.9122, "step": 4002 }, { "epoch": 0.7704003079291762, "grad_norm": 1.8327423334121704, "learning_rate": 2.639873022737105e-06, "loss": 0.8637, "step": 4003 }, { "epoch": 0.7705927636643572, "grad_norm": 1.8438129425048828, "learning_rate": 2.635654697216382e-06, "loss": 1.0066, "step": 4004 }, { "epoch": 0.7707852193995381, "grad_norm": 1.7489243745803833, "learning_rate": 2.631439233057638e-06, "loss": 0.884, "step": 4005 }, { "epoch": 0.770977675134719, "grad_norm": 1.9445438385009766, "learning_rate": 2.6272266318987606e-06, "loss": 0.953, "step": 4006 }, { "epoch": 0.7711701308698999, "grad_norm": 1.552230715751648, "learning_rate": 2.6230168953765257e-06, "loss": 0.9031, "step": 4007 }, { "epoch": 0.7713625866050808, "grad_norm": 1.7662941217422485, "learning_rate": 2.6188100251265947e-06, "loss": 0.9269, "step": 4008 }, { "epoch": 0.7715550423402617, "grad_norm": 1.8908839225769043, "learning_rate": 2.6146060227835156e-06, "loss": 0.8954, "step": 4009 }, { "epoch": 0.7717474980754426, "grad_norm": 1.8092327117919922, "learning_rate": 2.6104048899807243e-06, "loss": 0.9381, "step": 4010 }, { "epoch": 0.7719399538106235, "grad_norm": 1.694856882095337, "learning_rate": 2.6062066283505404e-06, "loss": 0.7952, "step": 4011 }, { "epoch": 0.7721324095458044, "grad_norm": 1.79978346824646, "learning_rate": 2.6020112395241627e-06, "loss": 0.9608, "step": 4012 }, { "epoch": 0.7723248652809854, "grad_norm": 1.9115095138549805, "learning_rate": 2.5978187251316823e-06, "loss": 0.9268, "step": 4013 }, { "epoch": 0.7725173210161663, "grad_norm": 1.8108848333358765, "learning_rate": 2.5936290868020685e-06, "loss": 0.9623, "step": 4014 }, { "epoch": 0.7727097767513472, "grad_norm": 1.7389498949050903, "learning_rate": 2.5894423261631717e-06, "loss": 0.9196, "step": 4015 }, { "epoch": 0.7729022324865281, "grad_norm": 1.7370339632034302, "learning_rate": 2.5852584448417327e-06, "loss": 0.9459, "step": 4016 }, { "epoch": 0.773094688221709, "grad_norm": 2.0633273124694824, "learning_rate": 2.581077444463365e-06, "loss": 0.9254, "step": 4017 }, { "epoch": 0.7732871439568899, "grad_norm": 1.787894606590271, "learning_rate": 2.5768993266525653e-06, "loss": 0.8898, "step": 4018 }, { "epoch": 0.7734795996920708, "grad_norm": 1.684076189994812, "learning_rate": 2.5727240930327103e-06, "loss": 0.9531, "step": 4019 }, { "epoch": 0.7736720554272517, "grad_norm": 1.779449701309204, "learning_rate": 2.5685517452260566e-06, "loss": 0.9651, "step": 4020 }, { "epoch": 0.7738645111624326, "grad_norm": 0.8608153462409973, "learning_rate": 2.564382284853738e-06, "loss": 0.8894, "step": 4021 }, { "epoch": 0.7740569668976135, "grad_norm": 1.7776696681976318, "learning_rate": 2.5602157135357697e-06, "loss": 0.9892, "step": 4022 }, { "epoch": 0.7742494226327945, "grad_norm": 1.7945811748504639, "learning_rate": 2.5560520328910397e-06, "loss": 0.8318, "step": 4023 }, { "epoch": 0.7744418783679754, "grad_norm": 1.8053555488586426, "learning_rate": 2.55189124453732e-06, "loss": 0.9085, "step": 4024 }, { "epoch": 0.7746343341031563, "grad_norm": 1.656427264213562, "learning_rate": 2.5477333500912473e-06, "loss": 0.9454, "step": 4025 }, { "epoch": 0.7748267898383372, "grad_norm": 1.6920685768127441, "learning_rate": 2.5435783511683444e-06, "loss": 0.9682, "step": 4026 }, { "epoch": 0.7750192455735181, "grad_norm": 1.7086405754089355, "learning_rate": 2.539426249383006e-06, "loss": 0.9223, "step": 4027 }, { "epoch": 0.775211701308699, "grad_norm": 1.7059297561645508, "learning_rate": 2.5352770463484986e-06, "loss": 0.9125, "step": 4028 }, { "epoch": 0.7754041570438799, "grad_norm": 0.8374617695808411, "learning_rate": 2.531130743676966e-06, "loss": 0.8328, "step": 4029 }, { "epoch": 0.7755966127790608, "grad_norm": 1.7718596458435059, "learning_rate": 2.5269873429794235e-06, "loss": 0.9087, "step": 4030 }, { "epoch": 0.7757890685142417, "grad_norm": 1.9159125089645386, "learning_rate": 2.5228468458657585e-06, "loss": 1.0296, "step": 4031 }, { "epoch": 0.7759815242494227, "grad_norm": 1.6716376543045044, "learning_rate": 2.51870925394473e-06, "loss": 0.9394, "step": 4032 }, { "epoch": 0.7761739799846036, "grad_norm": 1.716349482536316, "learning_rate": 2.5145745688239696e-06, "loss": 0.9535, "step": 4033 }, { "epoch": 0.7763664357197845, "grad_norm": 1.7742855548858643, "learning_rate": 2.5104427921099783e-06, "loss": 0.7995, "step": 4034 }, { "epoch": 0.7765588914549654, "grad_norm": 1.9097943305969238, "learning_rate": 2.506313925408127e-06, "loss": 0.93, "step": 4035 }, { "epoch": 0.7767513471901463, "grad_norm": 1.6164729595184326, "learning_rate": 2.502187970322657e-06, "loss": 0.9632, "step": 4036 }, { "epoch": 0.7769438029253272, "grad_norm": 1.6353263854980469, "learning_rate": 2.4980649284566783e-06, "loss": 0.886, "step": 4037 }, { "epoch": 0.7771362586605081, "grad_norm": 1.6983253955841064, "learning_rate": 2.493944801412167e-06, "loss": 0.8731, "step": 4038 }, { "epoch": 0.777328714395689, "grad_norm": 1.812192440032959, "learning_rate": 2.489827590789968e-06, "loss": 0.8975, "step": 4039 }, { "epoch": 0.7775211701308699, "grad_norm": 1.7929126024246216, "learning_rate": 2.485713298189798e-06, "loss": 0.9198, "step": 4040 }, { "epoch": 0.7777136258660509, "grad_norm": 1.7959985733032227, "learning_rate": 2.4816019252102274e-06, "loss": 1.0131, "step": 4041 }, { "epoch": 0.7779060816012318, "grad_norm": 1.8222358226776123, "learning_rate": 2.4774934734487034e-06, "loss": 0.8862, "step": 4042 }, { "epoch": 0.7780985373364127, "grad_norm": 1.5747305154800415, "learning_rate": 2.473387944501533e-06, "loss": 0.9231, "step": 4043 }, { "epoch": 0.7782909930715936, "grad_norm": 1.7685227394104004, "learning_rate": 2.469285339963892e-06, "loss": 0.923, "step": 4044 }, { "epoch": 0.7784834488067744, "grad_norm": 1.7454628944396973, "learning_rate": 2.4651856614298143e-06, "loss": 0.9269, "step": 4045 }, { "epoch": 0.7786759045419553, "grad_norm": 1.7711107730865479, "learning_rate": 2.461088910492202e-06, "loss": 0.8806, "step": 4046 }, { "epoch": 0.7788683602771362, "grad_norm": 0.888532280921936, "learning_rate": 2.456995088742815e-06, "loss": 0.877, "step": 4047 }, { "epoch": 0.7790608160123171, "grad_norm": 1.8654848337173462, "learning_rate": 2.45290419777228e-06, "loss": 0.9692, "step": 4048 }, { "epoch": 0.779253271747498, "grad_norm": 1.7783645391464233, "learning_rate": 2.448816239170081e-06, "loss": 0.915, "step": 4049 }, { "epoch": 0.7794457274826789, "grad_norm": 1.8375135660171509, "learning_rate": 2.444731214524565e-06, "loss": 0.9099, "step": 4050 }, { "epoch": 0.7796381832178599, "grad_norm": 1.7210549116134644, "learning_rate": 2.440649125422937e-06, "loss": 0.9271, "step": 4051 }, { "epoch": 0.7798306389530408, "grad_norm": 1.6757333278656006, "learning_rate": 2.436569973451264e-06, "loss": 0.9127, "step": 4052 }, { "epoch": 0.7800230946882217, "grad_norm": 1.5941623449325562, "learning_rate": 2.432493760194469e-06, "loss": 0.9292, "step": 4053 }, { "epoch": 0.7802155504234026, "grad_norm": 1.8091968297958374, "learning_rate": 2.428420487236336e-06, "loss": 0.907, "step": 4054 }, { "epoch": 0.7804080061585835, "grad_norm": 1.781041145324707, "learning_rate": 2.4243501561595027e-06, "loss": 0.9136, "step": 4055 }, { "epoch": 0.7806004618937644, "grad_norm": 1.7897768020629883, "learning_rate": 2.420282768545469e-06, "loss": 0.9136, "step": 4056 }, { "epoch": 0.7807929176289453, "grad_norm": 1.8277232646942139, "learning_rate": 2.41621832597459e-06, "loss": 0.9285, "step": 4057 }, { "epoch": 0.7809853733641262, "grad_norm": 1.6590546369552612, "learning_rate": 2.412156830026069e-06, "loss": 0.8872, "step": 4058 }, { "epoch": 0.7811778290993071, "grad_norm": 1.8351997137069702, "learning_rate": 2.408098282277972e-06, "loss": 0.9976, "step": 4059 }, { "epoch": 0.7813702848344881, "grad_norm": 1.791701078414917, "learning_rate": 2.4040426843072206e-06, "loss": 0.8299, "step": 4060 }, { "epoch": 0.781562740569669, "grad_norm": 0.8749986886978149, "learning_rate": 2.3999900376895844e-06, "loss": 0.9054, "step": 4061 }, { "epoch": 0.7817551963048499, "grad_norm": 1.7632205486297607, "learning_rate": 2.395940343999691e-06, "loss": 0.9338, "step": 4062 }, { "epoch": 0.7819476520400308, "grad_norm": 1.9440624713897705, "learning_rate": 2.3918936048110197e-06, "loss": 0.9143, "step": 4063 }, { "epoch": 0.7821401077752117, "grad_norm": 1.6107124090194702, "learning_rate": 2.387849821695899e-06, "loss": 0.8103, "step": 4064 }, { "epoch": 0.7823325635103926, "grad_norm": 1.704505205154419, "learning_rate": 2.3838089962255127e-06, "loss": 0.9234, "step": 4065 }, { "epoch": 0.7825250192455735, "grad_norm": 1.6994004249572754, "learning_rate": 2.3797711299698924e-06, "loss": 0.9606, "step": 4066 }, { "epoch": 0.7827174749807544, "grad_norm": 1.7423814535140991, "learning_rate": 2.375736224497922e-06, "loss": 0.947, "step": 4067 }, { "epoch": 0.7829099307159353, "grad_norm": 1.6235370635986328, "learning_rate": 2.371704281377335e-06, "loss": 0.8677, "step": 4068 }, { "epoch": 0.7831023864511163, "grad_norm": 1.8227174282073975, "learning_rate": 2.3676753021747113e-06, "loss": 1.0437, "step": 4069 }, { "epoch": 0.7832948421862972, "grad_norm": 1.5885359048843384, "learning_rate": 2.363649288455482e-06, "loss": 0.9784, "step": 4070 }, { "epoch": 0.7834872979214781, "grad_norm": 1.7153465747833252, "learning_rate": 2.3596262417839256e-06, "loss": 0.8752, "step": 4071 }, { "epoch": 0.783679753656659, "grad_norm": 1.7843090295791626, "learning_rate": 2.3556061637231653e-06, "loss": 0.9542, "step": 4072 }, { "epoch": 0.7838722093918399, "grad_norm": 1.7992967367172241, "learning_rate": 2.3515890558351773e-06, "loss": 0.9761, "step": 4073 }, { "epoch": 0.7840646651270208, "grad_norm": 1.7027226686477661, "learning_rate": 2.347574919680773e-06, "loss": 0.8991, "step": 4074 }, { "epoch": 0.7842571208622017, "grad_norm": 1.7947041988372803, "learning_rate": 2.343563756819619e-06, "loss": 0.9599, "step": 4075 }, { "epoch": 0.7844495765973826, "grad_norm": 1.9267503023147583, "learning_rate": 2.339555568810221e-06, "loss": 1.0005, "step": 4076 }, { "epoch": 0.7846420323325635, "grad_norm": 1.8662915229797363, "learning_rate": 2.3355503572099324e-06, "loss": 0.8889, "step": 4077 }, { "epoch": 0.7848344880677445, "grad_norm": 1.644400715827942, "learning_rate": 2.3315481235749484e-06, "loss": 0.9568, "step": 4078 }, { "epoch": 0.7850269438029254, "grad_norm": 1.6382209062576294, "learning_rate": 2.3275488694603076e-06, "loss": 1.0112, "step": 4079 }, { "epoch": 0.7852193995381063, "grad_norm": 1.6913951635360718, "learning_rate": 2.323552596419889e-06, "loss": 0.9169, "step": 4080 }, { "epoch": 0.7854118552732872, "grad_norm": 1.802233338356018, "learning_rate": 2.319559306006417e-06, "loss": 0.9751, "step": 4081 }, { "epoch": 0.7856043110084681, "grad_norm": 1.6668671369552612, "learning_rate": 2.315568999771455e-06, "loss": 0.8335, "step": 4082 }, { "epoch": 0.785796766743649, "grad_norm": 1.6295757293701172, "learning_rate": 2.3115816792654057e-06, "loss": 0.8748, "step": 4083 }, { "epoch": 0.7859892224788299, "grad_norm": 1.8585103750228882, "learning_rate": 2.3075973460375134e-06, "loss": 0.9773, "step": 4084 }, { "epoch": 0.7861816782140107, "grad_norm": 0.8987467288970947, "learning_rate": 2.303616001635861e-06, "loss": 0.8894, "step": 4085 }, { "epoch": 0.7863741339491916, "grad_norm": 1.6927284002304077, "learning_rate": 2.2996376476073724e-06, "loss": 0.8485, "step": 4086 }, { "epoch": 0.7865665896843725, "grad_norm": 1.7210171222686768, "learning_rate": 2.295662285497805e-06, "loss": 0.9034, "step": 4087 }, { "epoch": 0.7867590454195535, "grad_norm": 1.8954977989196777, "learning_rate": 2.291689916851758e-06, "loss": 0.831, "step": 4088 }, { "epoch": 0.7869515011547344, "grad_norm": 1.8608659505844116, "learning_rate": 2.287720543212666e-06, "loss": 0.9985, "step": 4089 }, { "epoch": 0.7871439568899153, "grad_norm": 1.6163427829742432, "learning_rate": 2.2837541661228024e-06, "loss": 0.9033, "step": 4090 }, { "epoch": 0.7873364126250962, "grad_norm": 1.8723149299621582, "learning_rate": 2.2797907871232673e-06, "loss": 0.897, "step": 4091 }, { "epoch": 0.7875288683602771, "grad_norm": 2.0258073806762695, "learning_rate": 2.275830407754006e-06, "loss": 0.9096, "step": 4092 }, { "epoch": 0.787721324095458, "grad_norm": 1.7099132537841797, "learning_rate": 2.271873029553795e-06, "loss": 0.8397, "step": 4093 }, { "epoch": 0.7879137798306389, "grad_norm": 1.8233882188796997, "learning_rate": 2.267918654060244e-06, "loss": 0.9012, "step": 4094 }, { "epoch": 0.7881062355658198, "grad_norm": 1.7047005891799927, "learning_rate": 2.2639672828097956e-06, "loss": 0.8525, "step": 4095 }, { "epoch": 0.7882986913010007, "grad_norm": 1.6801788806915283, "learning_rate": 2.2600189173377263e-06, "loss": 0.8716, "step": 4096 }, { "epoch": 0.7884911470361817, "grad_norm": 0.8978151082992554, "learning_rate": 2.256073559178145e-06, "loss": 0.8674, "step": 4097 }, { "epoch": 0.7886836027713626, "grad_norm": 1.7653475999832153, "learning_rate": 2.2521312098639914e-06, "loss": 0.9572, "step": 4098 }, { "epoch": 0.7888760585065435, "grad_norm": 1.7373738288879395, "learning_rate": 2.2481918709270347e-06, "loss": 0.9904, "step": 4099 }, { "epoch": 0.7890685142417244, "grad_norm": 1.6675580739974976, "learning_rate": 2.2442555438978774e-06, "loss": 0.872, "step": 4100 }, { "epoch": 0.7892609699769053, "grad_norm": 1.8066377639770508, "learning_rate": 2.240322230305951e-06, "loss": 0.954, "step": 4101 }, { "epoch": 0.7894534257120862, "grad_norm": 0.8698446154594421, "learning_rate": 2.2363919316795135e-06, "loss": 0.8799, "step": 4102 }, { "epoch": 0.7896458814472671, "grad_norm": 1.6235077381134033, "learning_rate": 2.2324646495456546e-06, "loss": 0.902, "step": 4103 }, { "epoch": 0.789838337182448, "grad_norm": 1.745627999305725, "learning_rate": 2.2285403854302912e-06, "loss": 0.937, "step": 4104 }, { "epoch": 0.7900307929176289, "grad_norm": 1.8406028747558594, "learning_rate": 2.224619140858166e-06, "loss": 0.967, "step": 4105 }, { "epoch": 0.7902232486528099, "grad_norm": 1.881837248802185, "learning_rate": 2.2207009173528528e-06, "loss": 0.9025, "step": 4106 }, { "epoch": 0.7904157043879908, "grad_norm": 1.5573763847351074, "learning_rate": 2.2167857164367424e-06, "loss": 0.8921, "step": 4107 }, { "epoch": 0.7906081601231717, "grad_norm": 1.8537564277648926, "learning_rate": 2.2128735396310606e-06, "loss": 0.8747, "step": 4108 }, { "epoch": 0.7908006158583526, "grad_norm": 2.0814061164855957, "learning_rate": 2.2089643884558553e-06, "loss": 0.9632, "step": 4109 }, { "epoch": 0.7909930715935335, "grad_norm": 1.7369502782821655, "learning_rate": 2.2050582644299976e-06, "loss": 0.8816, "step": 4110 }, { "epoch": 0.7911855273287144, "grad_norm": 1.7999197244644165, "learning_rate": 2.201155169071184e-06, "loss": 1.0288, "step": 4111 }, { "epoch": 0.7913779830638953, "grad_norm": 1.627845048904419, "learning_rate": 2.1972551038959313e-06, "loss": 0.9286, "step": 4112 }, { "epoch": 0.7915704387990762, "grad_norm": 1.7556148767471313, "learning_rate": 2.1933580704195835e-06, "loss": 0.8168, "step": 4113 }, { "epoch": 0.7917628945342571, "grad_norm": 1.7163196802139282, "learning_rate": 2.189464070156302e-06, "loss": 1.0117, "step": 4114 }, { "epoch": 0.791955350269438, "grad_norm": 1.5705773830413818, "learning_rate": 2.1855731046190732e-06, "loss": 0.8678, "step": 4115 }, { "epoch": 0.792147806004619, "grad_norm": 1.532548189163208, "learning_rate": 2.1816851753197023e-06, "loss": 0.9434, "step": 4116 }, { "epoch": 0.7923402617397999, "grad_norm": 0.8407483696937561, "learning_rate": 2.177800283768817e-06, "loss": 0.8403, "step": 4117 }, { "epoch": 0.7925327174749808, "grad_norm": 3.66011643409729, "learning_rate": 2.173918431475861e-06, "loss": 0.9009, "step": 4118 }, { "epoch": 0.7927251732101617, "grad_norm": 1.9078726768493652, "learning_rate": 2.1700396199491015e-06, "loss": 1.0126, "step": 4119 }, { "epoch": 0.7929176289453426, "grad_norm": 1.6532087326049805, "learning_rate": 2.1661638506956208e-06, "loss": 0.8908, "step": 4120 }, { "epoch": 0.7931100846805235, "grad_norm": 1.682085633277893, "learning_rate": 2.1622911252213195e-06, "loss": 0.9527, "step": 4121 }, { "epoch": 0.7933025404157044, "grad_norm": 1.6870826482772827, "learning_rate": 2.1584214450309194e-06, "loss": 0.8926, "step": 4122 }, { "epoch": 0.7934949961508853, "grad_norm": 1.7931045293807983, "learning_rate": 2.1545548116279557e-06, "loss": 0.9725, "step": 4123 }, { "epoch": 0.7936874518860662, "grad_norm": 1.8187659978866577, "learning_rate": 2.1506912265147772e-06, "loss": 0.8105, "step": 4124 }, { "epoch": 0.7938799076212472, "grad_norm": 1.5757101774215698, "learning_rate": 2.146830691192553e-06, "loss": 0.945, "step": 4125 }, { "epoch": 0.7940723633564281, "grad_norm": 1.8198773860931396, "learning_rate": 2.1429732071612653e-06, "loss": 0.8876, "step": 4126 }, { "epoch": 0.794264819091609, "grad_norm": 1.789643406867981, "learning_rate": 2.1391187759197118e-06, "loss": 0.9436, "step": 4127 }, { "epoch": 0.7944572748267898, "grad_norm": 1.7685370445251465, "learning_rate": 2.1352673989655026e-06, "loss": 0.9415, "step": 4128 }, { "epoch": 0.7946497305619707, "grad_norm": 1.7043297290802002, "learning_rate": 2.131419077795063e-06, "loss": 0.9234, "step": 4129 }, { "epoch": 0.7948421862971516, "grad_norm": 1.789685845375061, "learning_rate": 2.1275738139036274e-06, "loss": 0.881, "step": 4130 }, { "epoch": 0.7950346420323325, "grad_norm": 1.8379076719284058, "learning_rate": 2.1237316087852465e-06, "loss": 0.8407, "step": 4131 }, { "epoch": 0.7952270977675134, "grad_norm": 1.603400707244873, "learning_rate": 2.119892463932781e-06, "loss": 0.9942, "step": 4132 }, { "epoch": 0.7954195535026943, "grad_norm": 1.579335331916809, "learning_rate": 2.116056380837902e-06, "loss": 0.88, "step": 4133 }, { "epoch": 0.7956120092378753, "grad_norm": 1.731886863708496, "learning_rate": 2.1122233609910903e-06, "loss": 0.9198, "step": 4134 }, { "epoch": 0.7958044649730562, "grad_norm": 0.8273346424102783, "learning_rate": 2.108393405881638e-06, "loss": 0.8905, "step": 4135 }, { "epoch": 0.7959969207082371, "grad_norm": 1.702808141708374, "learning_rate": 2.104566516997647e-06, "loss": 0.8653, "step": 4136 }, { "epoch": 0.796189376443418, "grad_norm": 1.9452027082443237, "learning_rate": 2.1007426958260245e-06, "loss": 0.9358, "step": 4137 }, { "epoch": 0.7963818321785989, "grad_norm": 1.7004059553146362, "learning_rate": 2.096921943852489e-06, "loss": 0.9803, "step": 4138 }, { "epoch": 0.7965742879137798, "grad_norm": 1.7824429273605347, "learning_rate": 2.093104262561569e-06, "loss": 0.8829, "step": 4139 }, { "epoch": 0.7967667436489607, "grad_norm": 1.800701379776001, "learning_rate": 2.08928965343659e-06, "loss": 0.8784, "step": 4140 }, { "epoch": 0.7969591993841416, "grad_norm": 1.615323781967163, "learning_rate": 2.0854781179596937e-06, "loss": 0.9204, "step": 4141 }, { "epoch": 0.7971516551193225, "grad_norm": 1.9221491813659668, "learning_rate": 2.0816696576118223e-06, "loss": 1.0165, "step": 4142 }, { "epoch": 0.7973441108545035, "grad_norm": 1.4252030849456787, "learning_rate": 2.077864273872727e-06, "loss": 0.9138, "step": 4143 }, { "epoch": 0.7975365665896844, "grad_norm": 1.5626823902130127, "learning_rate": 2.0740619682209607e-06, "loss": 0.9403, "step": 4144 }, { "epoch": 0.7977290223248653, "grad_norm": 1.7487009763717651, "learning_rate": 2.0702627421338817e-06, "loss": 0.9053, "step": 4145 }, { "epoch": 0.7979214780600462, "grad_norm": 1.7869491577148438, "learning_rate": 2.0664665970876496e-06, "loss": 1.0108, "step": 4146 }, { "epoch": 0.7981139337952271, "grad_norm": 1.7700741291046143, "learning_rate": 2.06267353455723e-06, "loss": 0.8933, "step": 4147 }, { "epoch": 0.798306389530408, "grad_norm": 0.913165807723999, "learning_rate": 2.05888355601639e-06, "loss": 0.8364, "step": 4148 }, { "epoch": 0.7984988452655889, "grad_norm": 1.8121180534362793, "learning_rate": 2.0550966629376955e-06, "loss": 0.9108, "step": 4149 }, { "epoch": 0.7986913010007698, "grad_norm": 2.0133259296417236, "learning_rate": 2.0513128567925177e-06, "loss": 0.8049, "step": 4150 }, { "epoch": 0.7988837567359507, "grad_norm": 1.747348666191101, "learning_rate": 2.0475321390510262e-06, "loss": 0.893, "step": 4151 }, { "epoch": 0.7990762124711316, "grad_norm": 0.8426925539970398, "learning_rate": 2.043754511182191e-06, "loss": 0.8638, "step": 4152 }, { "epoch": 0.7992686682063126, "grad_norm": 1.6918818950653076, "learning_rate": 2.0399799746537806e-06, "loss": 0.8923, "step": 4153 }, { "epoch": 0.7994611239414935, "grad_norm": 1.755738615989685, "learning_rate": 2.036208530932363e-06, "loss": 0.8642, "step": 4154 }, { "epoch": 0.7996535796766744, "grad_norm": 1.7846592664718628, "learning_rate": 2.0324401814833074e-06, "loss": 0.8771, "step": 4155 }, { "epoch": 0.7998460354118553, "grad_norm": 1.5918515920639038, "learning_rate": 2.0286749277707783e-06, "loss": 0.9704, "step": 4156 }, { "epoch": 0.8000384911470362, "grad_norm": 1.7065420150756836, "learning_rate": 2.024912771257732e-06, "loss": 0.8801, "step": 4157 }, { "epoch": 0.8002309468822171, "grad_norm": 1.7380180358886719, "learning_rate": 2.021153713405931e-06, "loss": 0.9264, "step": 4158 }, { "epoch": 0.800423402617398, "grad_norm": 2.009814977645874, "learning_rate": 2.0173977556759284e-06, "loss": 0.9261, "step": 4159 }, { "epoch": 0.8006158583525789, "grad_norm": 1.7124450206756592, "learning_rate": 2.013644899527074e-06, "loss": 0.9294, "step": 4160 }, { "epoch": 0.8008083140877598, "grad_norm": 1.768195629119873, "learning_rate": 2.009895146417512e-06, "loss": 0.8739, "step": 4161 }, { "epoch": 0.8010007698229408, "grad_norm": 1.8154486417770386, "learning_rate": 2.0061484978041813e-06, "loss": 0.972, "step": 4162 }, { "epoch": 0.8011932255581217, "grad_norm": 1.6953485012054443, "learning_rate": 2.002404955142815e-06, "loss": 0.9693, "step": 4163 }, { "epoch": 0.8013856812933026, "grad_norm": 1.7631884813308716, "learning_rate": 1.9986645198879385e-06, "loss": 0.8724, "step": 4164 }, { "epoch": 0.8015781370284835, "grad_norm": 0.9007226228713989, "learning_rate": 1.9949271934928703e-06, "loss": 0.86, "step": 4165 }, { "epoch": 0.8017705927636644, "grad_norm": 1.7080154418945312, "learning_rate": 1.9911929774097216e-06, "loss": 0.9088, "step": 4166 }, { "epoch": 0.8019630484988453, "grad_norm": 1.6333963871002197, "learning_rate": 1.9874618730893947e-06, "loss": 0.9125, "step": 4167 }, { "epoch": 0.8021555042340262, "grad_norm": 1.6962813138961792, "learning_rate": 1.983733881981581e-06, "loss": 0.9207, "step": 4168 }, { "epoch": 0.802347959969207, "grad_norm": 1.9219435453414917, "learning_rate": 1.9800090055347654e-06, "loss": 0.8877, "step": 4169 }, { "epoch": 0.8025404157043879, "grad_norm": 1.5853062868118286, "learning_rate": 1.9762872451962214e-06, "loss": 0.9354, "step": 4170 }, { "epoch": 0.802732871439569, "grad_norm": 1.7312614917755127, "learning_rate": 1.9725686024120093e-06, "loss": 0.8824, "step": 4171 }, { "epoch": 0.8029253271747498, "grad_norm": 1.6753257513046265, "learning_rate": 1.9688530786269854e-06, "loss": 0.9704, "step": 4172 }, { "epoch": 0.8031177829099307, "grad_norm": 1.8347809314727783, "learning_rate": 1.965140675284779e-06, "loss": 0.958, "step": 4173 }, { "epoch": 0.8033102386451116, "grad_norm": 1.6330296993255615, "learning_rate": 1.961431393827827e-06, "loss": 0.9175, "step": 4174 }, { "epoch": 0.8035026943802925, "grad_norm": 1.6899086236953735, "learning_rate": 1.95772523569734e-06, "loss": 0.9995, "step": 4175 }, { "epoch": 0.8036951501154734, "grad_norm": 1.725965976715088, "learning_rate": 1.9540222023333165e-06, "loss": 0.9076, "step": 4176 }, { "epoch": 0.8038876058506543, "grad_norm": 1.7115392684936523, "learning_rate": 1.950322295174545e-06, "loss": 0.8555, "step": 4177 }, { "epoch": 0.8040800615858352, "grad_norm": 2.325735092163086, "learning_rate": 1.9466255156585965e-06, "loss": 0.9329, "step": 4178 }, { "epoch": 0.8042725173210161, "grad_norm": 1.7065677642822266, "learning_rate": 1.9429318652218277e-06, "loss": 0.893, "step": 4179 }, { "epoch": 0.8044649730561971, "grad_norm": 1.7874064445495605, "learning_rate": 1.9392413452993787e-06, "loss": 0.904, "step": 4180 }, { "epoch": 0.804657428791378, "grad_norm": 1.701096534729004, "learning_rate": 1.9355539573251737e-06, "loss": 0.9194, "step": 4181 }, { "epoch": 0.8048498845265589, "grad_norm": 1.6897774934768677, "learning_rate": 1.93186970273192e-06, "loss": 0.9404, "step": 4182 }, { "epoch": 0.8050423402617398, "grad_norm": 1.9749809503555298, "learning_rate": 1.928188582951108e-06, "loss": 0.9543, "step": 4183 }, { "epoch": 0.8052347959969207, "grad_norm": 1.8024415969848633, "learning_rate": 1.9245105994130086e-06, "loss": 0.9133, "step": 4184 }, { "epoch": 0.8054272517321016, "grad_norm": 1.8628535270690918, "learning_rate": 1.920835753546679e-06, "loss": 0.8686, "step": 4185 }, { "epoch": 0.8056197074672825, "grad_norm": 1.520134449005127, "learning_rate": 1.9171640467799478e-06, "loss": 0.8965, "step": 4186 }, { "epoch": 0.8058121632024634, "grad_norm": 1.846358060836792, "learning_rate": 1.913495480539432e-06, "loss": 1.0196, "step": 4187 }, { "epoch": 0.8060046189376443, "grad_norm": 1.634227991104126, "learning_rate": 1.9098300562505266e-06, "loss": 0.9496, "step": 4188 }, { "epoch": 0.8061970746728252, "grad_norm": 1.7520923614501953, "learning_rate": 1.906167775337404e-06, "loss": 0.945, "step": 4189 }, { "epoch": 0.8063895304080062, "grad_norm": 1.7849434614181519, "learning_rate": 1.9025086392230163e-06, "loss": 0.9212, "step": 4190 }, { "epoch": 0.8065819861431871, "grad_norm": 1.973913550376892, "learning_rate": 1.8988526493290948e-06, "loss": 0.9248, "step": 4191 }, { "epoch": 0.806774441878368, "grad_norm": 1.7237037420272827, "learning_rate": 1.895199807076148e-06, "loss": 0.9191, "step": 4192 }, { "epoch": 0.8069668976135489, "grad_norm": 1.921294927597046, "learning_rate": 1.891550113883458e-06, "loss": 0.9146, "step": 4193 }, { "epoch": 0.8071593533487298, "grad_norm": 1.6559011936187744, "learning_rate": 1.8879035711690897e-06, "loss": 0.9469, "step": 4194 }, { "epoch": 0.8073518090839107, "grad_norm": 2.1886346340179443, "learning_rate": 1.8842601803498772e-06, "loss": 0.888, "step": 4195 }, { "epoch": 0.8075442648190916, "grad_norm": 1.6460084915161133, "learning_rate": 1.880619942841435e-06, "loss": 0.8677, "step": 4196 }, { "epoch": 0.8077367205542725, "grad_norm": 1.6948699951171875, "learning_rate": 1.8769828600581497e-06, "loss": 0.9895, "step": 4197 }, { "epoch": 0.8079291762894534, "grad_norm": 1.8468785285949707, "learning_rate": 1.8733489334131827e-06, "loss": 0.934, "step": 4198 }, { "epoch": 0.8081216320246344, "grad_norm": 1.7442662715911865, "learning_rate": 1.8697181643184703e-06, "loss": 0.8835, "step": 4199 }, { "epoch": 0.8083140877598153, "grad_norm": 1.787463903427124, "learning_rate": 1.8660905541847208e-06, "loss": 0.9641, "step": 4200 }, { "epoch": 0.8085065434949962, "grad_norm": 1.657860517501831, "learning_rate": 1.8624661044214154e-06, "loss": 0.877, "step": 4201 }, { "epoch": 0.8086989992301771, "grad_norm": 1.818753957748413, "learning_rate": 1.858844816436809e-06, "loss": 0.9232, "step": 4202 }, { "epoch": 0.808891454965358, "grad_norm": 0.8731621503829956, "learning_rate": 1.8552266916379225e-06, "loss": 0.9388, "step": 4203 }, { "epoch": 0.8090839107005389, "grad_norm": 1.8021739721298218, "learning_rate": 1.8516117314305526e-06, "loss": 0.9607, "step": 4204 }, { "epoch": 0.8092763664357198, "grad_norm": 1.5152666568756104, "learning_rate": 1.847999937219266e-06, "loss": 0.8147, "step": 4205 }, { "epoch": 0.8094688221709007, "grad_norm": 1.7942613363265991, "learning_rate": 1.8443913104073984e-06, "loss": 0.8727, "step": 4206 }, { "epoch": 0.8096612779060816, "grad_norm": 1.8418891429901123, "learning_rate": 1.8407858523970568e-06, "loss": 0.9542, "step": 4207 }, { "epoch": 0.8098537336412626, "grad_norm": 1.671880841255188, "learning_rate": 1.8371835645891134e-06, "loss": 0.9527, "step": 4208 }, { "epoch": 0.8100461893764435, "grad_norm": 1.663207769393921, "learning_rate": 1.833584448383211e-06, "loss": 0.9289, "step": 4209 }, { "epoch": 0.8102386451116244, "grad_norm": 1.6976065635681152, "learning_rate": 1.82998850517776e-06, "loss": 0.9382, "step": 4210 }, { "epoch": 0.8104311008468053, "grad_norm": 1.7451149225234985, "learning_rate": 1.826395736369937e-06, "loss": 0.8561, "step": 4211 }, { "epoch": 0.8106235565819861, "grad_norm": 1.8604519367218018, "learning_rate": 1.8228061433556866e-06, "loss": 0.977, "step": 4212 }, { "epoch": 0.810816012317167, "grad_norm": 1.8794054985046387, "learning_rate": 1.8192197275297174e-06, "loss": 0.8978, "step": 4213 }, { "epoch": 0.8110084680523479, "grad_norm": 1.7330206632614136, "learning_rate": 1.815636490285506e-06, "loss": 0.9227, "step": 4214 }, { "epoch": 0.8112009237875288, "grad_norm": 1.8505733013153076, "learning_rate": 1.8120564330152924e-06, "loss": 0.9015, "step": 4215 }, { "epoch": 0.8113933795227097, "grad_norm": 1.6796363592147827, "learning_rate": 1.808479557110081e-06, "loss": 0.9491, "step": 4216 }, { "epoch": 0.8115858352578906, "grad_norm": 1.878218412399292, "learning_rate": 1.8049058639596418e-06, "loss": 0.9308, "step": 4217 }, { "epoch": 0.8117782909930716, "grad_norm": 1.7754994630813599, "learning_rate": 1.801335354952507e-06, "loss": 0.9241, "step": 4218 }, { "epoch": 0.8119707467282525, "grad_norm": 1.771694302558899, "learning_rate": 1.7977680314759693e-06, "loss": 0.9203, "step": 4219 }, { "epoch": 0.8121632024634334, "grad_norm": 1.8847718238830566, "learning_rate": 1.7942038949160857e-06, "loss": 0.8729, "step": 4220 }, { "epoch": 0.8123556581986143, "grad_norm": 1.8215410709381104, "learning_rate": 1.7906429466576768e-06, "loss": 0.9399, "step": 4221 }, { "epoch": 0.8125481139337952, "grad_norm": 1.9568886756896973, "learning_rate": 1.7870851880843233e-06, "loss": 0.9825, "step": 4222 }, { "epoch": 0.8127405696689761, "grad_norm": 2.0534512996673584, "learning_rate": 1.7835306205783643e-06, "loss": 0.9358, "step": 4223 }, { "epoch": 0.812933025404157, "grad_norm": 1.7654199600219727, "learning_rate": 1.7799792455209019e-06, "loss": 0.8807, "step": 4224 }, { "epoch": 0.8131254811393379, "grad_norm": 1.6095807552337646, "learning_rate": 1.776431064291796e-06, "loss": 0.8955, "step": 4225 }, { "epoch": 0.8133179368745188, "grad_norm": 1.7856521606445312, "learning_rate": 1.7728860782696666e-06, "loss": 0.882, "step": 4226 }, { "epoch": 0.8135103926096998, "grad_norm": 1.734665870666504, "learning_rate": 1.7693442888318923e-06, "loss": 0.9345, "step": 4227 }, { "epoch": 0.8137028483448807, "grad_norm": 1.6986876726150513, "learning_rate": 1.765805697354608e-06, "loss": 0.9019, "step": 4228 }, { "epoch": 0.8138953040800616, "grad_norm": 0.9054495692253113, "learning_rate": 1.7622703052127078e-06, "loss": 0.9102, "step": 4229 }, { "epoch": 0.8140877598152425, "grad_norm": 1.755710244178772, "learning_rate": 1.7587381137798432e-06, "loss": 0.921, "step": 4230 }, { "epoch": 0.8142802155504234, "grad_norm": 1.6024792194366455, "learning_rate": 1.7552091244284197e-06, "loss": 0.9204, "step": 4231 }, { "epoch": 0.8144726712856043, "grad_norm": 1.6380693912506104, "learning_rate": 1.7516833385296016e-06, "loss": 0.9189, "step": 4232 }, { "epoch": 0.8146651270207852, "grad_norm": 2.064812660217285, "learning_rate": 1.7481607574533044e-06, "loss": 0.8332, "step": 4233 }, { "epoch": 0.8148575827559661, "grad_norm": 1.720259666442871, "learning_rate": 1.744641382568204e-06, "loss": 0.9808, "step": 4234 }, { "epoch": 0.815050038491147, "grad_norm": 1.7702584266662598, "learning_rate": 1.741125215241728e-06, "loss": 0.9313, "step": 4235 }, { "epoch": 0.815242494226328, "grad_norm": 1.6408809423446655, "learning_rate": 1.7376122568400533e-06, "loss": 1.0182, "step": 4236 }, { "epoch": 0.8154349499615089, "grad_norm": 1.726912021636963, "learning_rate": 1.7341025087281149e-06, "loss": 0.9077, "step": 4237 }, { "epoch": 0.8156274056966898, "grad_norm": 1.8319919109344482, "learning_rate": 1.730595972269602e-06, "loss": 0.9951, "step": 4238 }, { "epoch": 0.8158198614318707, "grad_norm": 1.8397530317306519, "learning_rate": 1.7270926488269502e-06, "loss": 0.9313, "step": 4239 }, { "epoch": 0.8160123171670516, "grad_norm": 1.856245517730713, "learning_rate": 1.7235925397613529e-06, "loss": 0.9689, "step": 4240 }, { "epoch": 0.8162047729022325, "grad_norm": 1.646837592124939, "learning_rate": 1.7200956464327512e-06, "loss": 0.9232, "step": 4241 }, { "epoch": 0.8163972286374134, "grad_norm": 1.9551154375076294, "learning_rate": 1.716601970199836e-06, "loss": 0.861, "step": 4242 }, { "epoch": 0.8165896843725943, "grad_norm": 1.7405091524124146, "learning_rate": 1.7131115124200494e-06, "loss": 0.9859, "step": 4243 }, { "epoch": 0.8167821401077752, "grad_norm": 1.6753772497177124, "learning_rate": 1.709624274449584e-06, "loss": 0.9308, "step": 4244 }, { "epoch": 0.8169745958429562, "grad_norm": 1.8177111148834229, "learning_rate": 1.7061402576433794e-06, "loss": 0.8794, "step": 4245 }, { "epoch": 0.8171670515781371, "grad_norm": 1.8208540678024292, "learning_rate": 1.7026594633551252e-06, "loss": 0.9008, "step": 4246 }, { "epoch": 0.817359507313318, "grad_norm": 1.803389549255371, "learning_rate": 1.6991818929372584e-06, "loss": 0.9693, "step": 4247 }, { "epoch": 0.8175519630484989, "grad_norm": 1.9115324020385742, "learning_rate": 1.6957075477409623e-06, "loss": 0.8886, "step": 4248 }, { "epoch": 0.8177444187836798, "grad_norm": 1.6749255657196045, "learning_rate": 1.6922364291161697e-06, "loss": 0.9003, "step": 4249 }, { "epoch": 0.8179368745188607, "grad_norm": 1.807468056678772, "learning_rate": 1.688768538411556e-06, "loss": 0.9866, "step": 4250 }, { "epoch": 0.8181293302540416, "grad_norm": 1.8065719604492188, "learning_rate": 1.6853038769745466e-06, "loss": 0.8801, "step": 4251 }, { "epoch": 0.8183217859892224, "grad_norm": 1.7839444875717163, "learning_rate": 1.6818424461513129e-06, "loss": 0.9524, "step": 4252 }, { "epoch": 0.8185142417244033, "grad_norm": 1.7329148054122925, "learning_rate": 1.6783842472867607e-06, "loss": 0.8754, "step": 4253 }, { "epoch": 0.8187066974595842, "grad_norm": 1.8929228782653809, "learning_rate": 1.674929281724551e-06, "loss": 0.9778, "step": 4254 }, { "epoch": 0.8188991531947653, "grad_norm": 1.772507667541504, "learning_rate": 1.6714775508070868e-06, "loss": 0.9281, "step": 4255 }, { "epoch": 0.8190916089299461, "grad_norm": 1.7080711126327515, "learning_rate": 1.6680290558755119e-06, "loss": 0.8903, "step": 4256 }, { "epoch": 0.819284064665127, "grad_norm": 1.702973484992981, "learning_rate": 1.6645837982697122e-06, "loss": 0.9571, "step": 4257 }, { "epoch": 0.8194765204003079, "grad_norm": 1.8499590158462524, "learning_rate": 1.6611417793283192e-06, "loss": 0.9306, "step": 4258 }, { "epoch": 0.8196689761354888, "grad_norm": 1.7541255950927734, "learning_rate": 1.6577030003887018e-06, "loss": 0.9135, "step": 4259 }, { "epoch": 0.8198614318706697, "grad_norm": 1.6664550304412842, "learning_rate": 1.6542674627869738e-06, "loss": 0.8819, "step": 4260 }, { "epoch": 0.8200538876058506, "grad_norm": 1.7327603101730347, "learning_rate": 1.6508351678579882e-06, "loss": 0.938, "step": 4261 }, { "epoch": 0.8202463433410315, "grad_norm": 1.9212325811386108, "learning_rate": 1.6474061169353362e-06, "loss": 0.9478, "step": 4262 }, { "epoch": 0.8204387990762124, "grad_norm": 1.8893914222717285, "learning_rate": 1.6439803113513509e-06, "loss": 0.94, "step": 4263 }, { "epoch": 0.8206312548113934, "grad_norm": 1.6505144834518433, "learning_rate": 1.640557752437103e-06, "loss": 0.8678, "step": 4264 }, { "epoch": 0.8208237105465743, "grad_norm": 0.8692054748535156, "learning_rate": 1.6371384415224046e-06, "loss": 0.8714, "step": 4265 }, { "epoch": 0.8210161662817552, "grad_norm": 1.7048276662826538, "learning_rate": 1.6337223799358025e-06, "loss": 0.8703, "step": 4266 }, { "epoch": 0.8212086220169361, "grad_norm": 1.8522233963012695, "learning_rate": 1.6303095690045822e-06, "loss": 1.0352, "step": 4267 }, { "epoch": 0.821401077752117, "grad_norm": 1.7438253164291382, "learning_rate": 1.6269000100547682e-06, "loss": 0.9445, "step": 4268 }, { "epoch": 0.8215935334872979, "grad_norm": 1.7960048913955688, "learning_rate": 1.6234937044111155e-06, "loss": 0.8408, "step": 4269 }, { "epoch": 0.8217859892224788, "grad_norm": 1.800661563873291, "learning_rate": 1.6200906533971206e-06, "loss": 0.8733, "step": 4270 }, { "epoch": 0.8219784449576597, "grad_norm": 1.7225450277328491, "learning_rate": 1.6166908583350138e-06, "loss": 0.9661, "step": 4271 }, { "epoch": 0.8221709006928406, "grad_norm": 1.6220556497573853, "learning_rate": 1.6132943205457607e-06, "loss": 1.0317, "step": 4272 }, { "epoch": 0.8223633564280216, "grad_norm": 1.7651034593582153, "learning_rate": 1.6099010413490624e-06, "loss": 0.8673, "step": 4273 }, { "epoch": 0.8225558121632025, "grad_norm": 1.7260187864303589, "learning_rate": 1.6065110220633495e-06, "loss": 0.8898, "step": 4274 }, { "epoch": 0.8227482678983834, "grad_norm": 1.8042688369750977, "learning_rate": 1.6031242640057908e-06, "loss": 0.9061, "step": 4275 }, { "epoch": 0.8229407236335643, "grad_norm": 1.661299467086792, "learning_rate": 1.599740768492286e-06, "loss": 0.9669, "step": 4276 }, { "epoch": 0.8231331793687452, "grad_norm": 2.0681376457214355, "learning_rate": 1.596360536837468e-06, "loss": 0.961, "step": 4277 }, { "epoch": 0.8233256351039261, "grad_norm": 1.9284907579421997, "learning_rate": 1.5929835703546992e-06, "loss": 0.8393, "step": 4278 }, { "epoch": 0.823518090839107, "grad_norm": 1.7041962146759033, "learning_rate": 1.589609870356076e-06, "loss": 0.8643, "step": 4279 }, { "epoch": 0.8237105465742879, "grad_norm": 1.7222951650619507, "learning_rate": 1.5862394381524239e-06, "loss": 0.9307, "step": 4280 }, { "epoch": 0.8239030023094688, "grad_norm": 1.7107068300247192, "learning_rate": 1.582872275053301e-06, "loss": 0.8792, "step": 4281 }, { "epoch": 0.8240954580446497, "grad_norm": 1.6639665365219116, "learning_rate": 1.579508382366992e-06, "loss": 0.8577, "step": 4282 }, { "epoch": 0.8242879137798307, "grad_norm": 1.8390802145004272, "learning_rate": 1.5761477614005127e-06, "loss": 0.9072, "step": 4283 }, { "epoch": 0.8244803695150116, "grad_norm": 1.781119704246521, "learning_rate": 1.5727904134596084e-06, "loss": 0.9687, "step": 4284 }, { "epoch": 0.8246728252501925, "grad_norm": 1.711091160774231, "learning_rate": 1.5694363398487545e-06, "loss": 0.908, "step": 4285 }, { "epoch": 0.8248652809853734, "grad_norm": 0.8537315726280212, "learning_rate": 1.566085541871145e-06, "loss": 0.8715, "step": 4286 }, { "epoch": 0.8250577367205543, "grad_norm": 1.7716619968414307, "learning_rate": 1.5627380208287114e-06, "loss": 0.9765, "step": 4287 }, { "epoch": 0.8252501924557352, "grad_norm": 1.9409608840942383, "learning_rate": 1.5593937780221092e-06, "loss": 0.9652, "step": 4288 }, { "epoch": 0.8254426481909161, "grad_norm": 1.7797962427139282, "learning_rate": 1.556052814750717e-06, "loss": 0.8295, "step": 4289 }, { "epoch": 0.825635103926097, "grad_norm": 1.8179157972335815, "learning_rate": 1.5527151323126422e-06, "loss": 0.8571, "step": 4290 }, { "epoch": 0.8258275596612779, "grad_norm": 1.9461023807525635, "learning_rate": 1.5493807320047183e-06, "loss": 0.9285, "step": 4291 }, { "epoch": 0.8260200153964589, "grad_norm": 1.6991040706634521, "learning_rate": 1.5460496151225002e-06, "loss": 0.8974, "step": 4292 }, { "epoch": 0.8262124711316398, "grad_norm": 1.8561325073242188, "learning_rate": 1.542721782960268e-06, "loss": 0.9413, "step": 4293 }, { "epoch": 0.8264049268668207, "grad_norm": 1.6219134330749512, "learning_rate": 1.5393972368110287e-06, "loss": 0.9175, "step": 4294 }, { "epoch": 0.8265973826020016, "grad_norm": 1.7772047519683838, "learning_rate": 1.536075977966509e-06, "loss": 0.9833, "step": 4295 }, { "epoch": 0.8267898383371824, "grad_norm": 1.6094797849655151, "learning_rate": 1.5327580077171589e-06, "loss": 0.9904, "step": 4296 }, { "epoch": 0.8269822940723633, "grad_norm": 1.7838640213012695, "learning_rate": 1.5294433273521525e-06, "loss": 0.7935, "step": 4297 }, { "epoch": 0.8271747498075442, "grad_norm": 1.7071502208709717, "learning_rate": 1.5261319381593831e-06, "loss": 0.9003, "step": 4298 }, { "epoch": 0.8273672055427251, "grad_norm": 1.82965886592865, "learning_rate": 1.522823841425467e-06, "loss": 0.8805, "step": 4299 }, { "epoch": 0.827559661277906, "grad_norm": 1.726851463317871, "learning_rate": 1.5195190384357405e-06, "loss": 0.9749, "step": 4300 }, { "epoch": 0.827752117013087, "grad_norm": 1.8934102058410645, "learning_rate": 1.5162175304742633e-06, "loss": 0.9168, "step": 4301 }, { "epoch": 0.8279445727482679, "grad_norm": 1.6925345659255981, "learning_rate": 1.5129193188238068e-06, "loss": 0.9727, "step": 4302 }, { "epoch": 0.8281370284834488, "grad_norm": 1.8061341047286987, "learning_rate": 1.5096244047658704e-06, "loss": 1.0359, "step": 4303 }, { "epoch": 0.8283294842186297, "grad_norm": 1.9418781995773315, "learning_rate": 1.5063327895806668e-06, "loss": 0.936, "step": 4304 }, { "epoch": 0.8285219399538106, "grad_norm": 1.560083031654358, "learning_rate": 1.5030444745471295e-06, "loss": 0.8828, "step": 4305 }, { "epoch": 0.8287143956889915, "grad_norm": 1.9550267457962036, "learning_rate": 1.499759460942909e-06, "loss": 0.8736, "step": 4306 }, { "epoch": 0.8289068514241724, "grad_norm": 1.8836151361465454, "learning_rate": 1.4964777500443728e-06, "loss": 0.8669, "step": 4307 }, { "epoch": 0.8290993071593533, "grad_norm": 1.7230879068374634, "learning_rate": 1.4931993431266056e-06, "loss": 0.8788, "step": 4308 }, { "epoch": 0.8292917628945342, "grad_norm": 1.8224726915359497, "learning_rate": 1.4899242414634097e-06, "loss": 0.9221, "step": 4309 }, { "epoch": 0.8294842186297152, "grad_norm": 1.5524909496307373, "learning_rate": 1.486652446327299e-06, "loss": 0.854, "step": 4310 }, { "epoch": 0.8296766743648961, "grad_norm": 1.9619694948196411, "learning_rate": 1.4833839589895072e-06, "loss": 0.9776, "step": 4311 }, { "epoch": 0.829869130100077, "grad_norm": 1.5696148872375488, "learning_rate": 1.48011878071998e-06, "loss": 0.9144, "step": 4312 }, { "epoch": 0.8300615858352579, "grad_norm": 2.028028726577759, "learning_rate": 1.476856912787379e-06, "loss": 0.9094, "step": 4313 }, { "epoch": 0.8302540415704388, "grad_norm": 1.6830319166183472, "learning_rate": 1.4735983564590784e-06, "loss": 0.9024, "step": 4314 }, { "epoch": 0.8304464973056197, "grad_norm": 1.9636989831924438, "learning_rate": 1.470343113001167e-06, "loss": 0.9765, "step": 4315 }, { "epoch": 0.8306389530408006, "grad_norm": 1.5231817960739136, "learning_rate": 1.467091183678444e-06, "loss": 0.875, "step": 4316 }, { "epoch": 0.8308314087759815, "grad_norm": 1.775160312652588, "learning_rate": 1.4638425697544235e-06, "loss": 0.9189, "step": 4317 }, { "epoch": 0.8310238645111624, "grad_norm": 1.6822912693023682, "learning_rate": 1.460597272491333e-06, "loss": 0.944, "step": 4318 }, { "epoch": 0.8312163202463433, "grad_norm": 1.823304295539856, "learning_rate": 1.4573552931501046e-06, "loss": 0.9519, "step": 4319 }, { "epoch": 0.8314087759815243, "grad_norm": 0.9431740045547485, "learning_rate": 1.4541166329903856e-06, "loss": 0.9068, "step": 4320 }, { "epoch": 0.8316012317167052, "grad_norm": 1.6673917770385742, "learning_rate": 1.4508812932705364e-06, "loss": 0.906, "step": 4321 }, { "epoch": 0.8317936874518861, "grad_norm": 1.9698683023452759, "learning_rate": 1.4476492752476223e-06, "loss": 0.976, "step": 4322 }, { "epoch": 0.831986143187067, "grad_norm": 1.6973397731781006, "learning_rate": 1.4444205801774204e-06, "loss": 0.9412, "step": 4323 }, { "epoch": 0.8321785989222479, "grad_norm": 1.7950079441070557, "learning_rate": 1.4411952093144167e-06, "loss": 0.9048, "step": 4324 }, { "epoch": 0.8323710546574288, "grad_norm": 1.9273159503936768, "learning_rate": 1.4379731639118045e-06, "loss": 0.8293, "step": 4325 }, { "epoch": 0.8325635103926097, "grad_norm": 1.7184327840805054, "learning_rate": 1.4347544452214869e-06, "loss": 0.886, "step": 4326 }, { "epoch": 0.8327559661277906, "grad_norm": 2.353425979614258, "learning_rate": 1.4315390544940732e-06, "loss": 0.9042, "step": 4327 }, { "epoch": 0.8329484218629715, "grad_norm": 1.924476981163025, "learning_rate": 1.4283269929788779e-06, "loss": 0.9154, "step": 4328 }, { "epoch": 0.8331408775981525, "grad_norm": 1.715088963508606, "learning_rate": 1.4251182619239257e-06, "loss": 1.0273, "step": 4329 }, { "epoch": 0.8333333333333334, "grad_norm": 1.9428443908691406, "learning_rate": 1.4219128625759426e-06, "loss": 0.8203, "step": 4330 }, { "epoch": 0.8335257890685143, "grad_norm": 1.9512563943862915, "learning_rate": 1.4187107961803704e-06, "loss": 0.8814, "step": 4331 }, { "epoch": 0.8337182448036952, "grad_norm": 0.9265878796577454, "learning_rate": 1.4155120639813392e-06, "loss": 0.9298, "step": 4332 }, { "epoch": 0.8339107005388761, "grad_norm": 1.6782152652740479, "learning_rate": 1.4123166672216971e-06, "loss": 0.9226, "step": 4333 }, { "epoch": 0.834103156274057, "grad_norm": 1.5828169584274292, "learning_rate": 1.409124607142992e-06, "loss": 0.9278, "step": 4334 }, { "epoch": 0.8342956120092379, "grad_norm": 1.710893988609314, "learning_rate": 1.4059358849854732e-06, "loss": 0.8731, "step": 4335 }, { "epoch": 0.8344880677444187, "grad_norm": 1.7301074266433716, "learning_rate": 1.4027505019880972e-06, "loss": 0.9195, "step": 4336 }, { "epoch": 0.8346805234795996, "grad_norm": 1.9170883893966675, "learning_rate": 1.3995684593885206e-06, "loss": 0.9295, "step": 4337 }, { "epoch": 0.8348729792147807, "grad_norm": 1.8658148050308228, "learning_rate": 1.3963897584231023e-06, "loss": 0.8623, "step": 4338 }, { "epoch": 0.8350654349499615, "grad_norm": 1.7214055061340332, "learning_rate": 1.3932144003269033e-06, "loss": 0.9246, "step": 4339 }, { "epoch": 0.8352578906851424, "grad_norm": 1.7628673315048218, "learning_rate": 1.3900423863336842e-06, "loss": 0.9108, "step": 4340 }, { "epoch": 0.8354503464203233, "grad_norm": 1.6332365274429321, "learning_rate": 1.3868737176759105e-06, "loss": 0.8799, "step": 4341 }, { "epoch": 0.8356428021555042, "grad_norm": 1.6727641820907593, "learning_rate": 1.3837083955847418e-06, "loss": 0.9271, "step": 4342 }, { "epoch": 0.8358352578906851, "grad_norm": 1.6236034631729126, "learning_rate": 1.3805464212900432e-06, "loss": 0.9192, "step": 4343 }, { "epoch": 0.836027713625866, "grad_norm": 1.6248812675476074, "learning_rate": 1.377387796020374e-06, "loss": 0.8858, "step": 4344 }, { "epoch": 0.8362201693610469, "grad_norm": 1.7886335849761963, "learning_rate": 1.3742325210029973e-06, "loss": 0.8103, "step": 4345 }, { "epoch": 0.8364126250962278, "grad_norm": 1.652933955192566, "learning_rate": 1.3710805974638697e-06, "loss": 1.0016, "step": 4346 }, { "epoch": 0.8366050808314087, "grad_norm": 1.7956198453903198, "learning_rate": 1.3679320266276518e-06, "loss": 0.8328, "step": 4347 }, { "epoch": 0.8367975365665897, "grad_norm": 1.6026407480239868, "learning_rate": 1.364786809717692e-06, "loss": 0.9534, "step": 4348 }, { "epoch": 0.8369899923017706, "grad_norm": 0.9663257598876953, "learning_rate": 1.3616449479560434e-06, "loss": 0.9229, "step": 4349 }, { "epoch": 0.8371824480369515, "grad_norm": 1.8057861328125, "learning_rate": 1.3585064425634542e-06, "loss": 0.7732, "step": 4350 }, { "epoch": 0.8373749037721324, "grad_norm": 1.7233569622039795, "learning_rate": 1.3553712947593655e-06, "loss": 0.9612, "step": 4351 }, { "epoch": 0.8375673595073133, "grad_norm": 2.1935484409332275, "learning_rate": 1.3522395057619186e-06, "loss": 0.8683, "step": 4352 }, { "epoch": 0.8377598152424942, "grad_norm": 1.7658956050872803, "learning_rate": 1.3491110767879444e-06, "loss": 0.9166, "step": 4353 }, { "epoch": 0.8379522709776751, "grad_norm": 1.7272753715515137, "learning_rate": 1.345986009052972e-06, "loss": 0.8775, "step": 4354 }, { "epoch": 0.838144726712856, "grad_norm": 1.5437290668487549, "learning_rate": 1.342864303771223e-06, "loss": 0.9321, "step": 4355 }, { "epoch": 0.8383371824480369, "grad_norm": 1.9196034669876099, "learning_rate": 1.339745962155613e-06, "loss": 0.9169, "step": 4356 }, { "epoch": 0.8385296381832179, "grad_norm": 1.6409907341003418, "learning_rate": 1.3366309854177517e-06, "loss": 0.9046, "step": 4357 }, { "epoch": 0.8387220939183988, "grad_norm": 1.719199538230896, "learning_rate": 1.3335193747679387e-06, "loss": 0.9425, "step": 4358 }, { "epoch": 0.8389145496535797, "grad_norm": 1.6527676582336426, "learning_rate": 1.330411131415169e-06, "loss": 0.9958, "step": 4359 }, { "epoch": 0.8391070053887606, "grad_norm": 1.6826039552688599, "learning_rate": 1.3273062565671258e-06, "loss": 0.9396, "step": 4360 }, { "epoch": 0.8392994611239415, "grad_norm": 1.758672833442688, "learning_rate": 1.324204751430186e-06, "loss": 0.8946, "step": 4361 }, { "epoch": 0.8394919168591224, "grad_norm": 1.6380798816680908, "learning_rate": 1.3211066172094178e-06, "loss": 0.8279, "step": 4362 }, { "epoch": 0.8396843725943033, "grad_norm": 1.7920273542404175, "learning_rate": 1.3180118551085763e-06, "loss": 0.9407, "step": 4363 }, { "epoch": 0.8398768283294842, "grad_norm": 1.9294027090072632, "learning_rate": 1.3149204663301118e-06, "loss": 0.8621, "step": 4364 }, { "epoch": 0.8400692840646651, "grad_norm": 1.7811110019683838, "learning_rate": 1.3118324520751558e-06, "loss": 0.8579, "step": 4365 }, { "epoch": 0.8402617397998461, "grad_norm": 1.8185826539993286, "learning_rate": 1.3087478135435361e-06, "loss": 0.9734, "step": 4366 }, { "epoch": 0.840454195535027, "grad_norm": 1.61343252658844, "learning_rate": 1.3056665519337663e-06, "loss": 1.0106, "step": 4367 }, { "epoch": 0.8406466512702079, "grad_norm": 1.859797477722168, "learning_rate": 1.3025886684430467e-06, "loss": 0.8914, "step": 4368 }, { "epoch": 0.8408391070053888, "grad_norm": 1.754708170890808, "learning_rate": 1.299514164267267e-06, "loss": 0.7904, "step": 4369 }, { "epoch": 0.8410315627405697, "grad_norm": 1.8280558586120605, "learning_rate": 1.2964430406010032e-06, "loss": 0.9085, "step": 4370 }, { "epoch": 0.8412240184757506, "grad_norm": 1.849546194076538, "learning_rate": 1.293375298637518e-06, "loss": 0.8961, "step": 4371 }, { "epoch": 0.8414164742109315, "grad_norm": 1.6942812204360962, "learning_rate": 1.2903109395687597e-06, "loss": 0.853, "step": 4372 }, { "epoch": 0.8416089299461124, "grad_norm": 1.7574023008346558, "learning_rate": 1.2872499645853609e-06, "loss": 0.8849, "step": 4373 }, { "epoch": 0.8418013856812933, "grad_norm": 1.7603715658187866, "learning_rate": 1.2841923748766427e-06, "loss": 0.9475, "step": 4374 }, { "epoch": 0.8419938414164743, "grad_norm": 1.867280125617981, "learning_rate": 1.2811381716306092e-06, "loss": 0.9657, "step": 4375 }, { "epoch": 0.8421862971516552, "grad_norm": 1.6992754936218262, "learning_rate": 1.278087356033947e-06, "loss": 0.8619, "step": 4376 }, { "epoch": 0.8423787528868361, "grad_norm": 1.8392752408981323, "learning_rate": 1.2750399292720284e-06, "loss": 0.9048, "step": 4377 }, { "epoch": 0.842571208622017, "grad_norm": 1.6466338634490967, "learning_rate": 1.271995892528909e-06, "loss": 0.8745, "step": 4378 }, { "epoch": 0.8427636643571979, "grad_norm": 1.744739294052124, "learning_rate": 1.2689552469873267e-06, "loss": 0.887, "step": 4379 }, { "epoch": 0.8429561200923787, "grad_norm": 2.1463263034820557, "learning_rate": 1.2659179938287035e-06, "loss": 0.8471, "step": 4380 }, { "epoch": 0.8431485758275596, "grad_norm": 1.6629698276519775, "learning_rate": 1.2628841342331389e-06, "loss": 0.9043, "step": 4381 }, { "epoch": 0.8433410315627405, "grad_norm": 1.757784128189087, "learning_rate": 1.2598536693794171e-06, "loss": 0.9238, "step": 4382 }, { "epoch": 0.8435334872979214, "grad_norm": 1.7084028720855713, "learning_rate": 1.2568266004450037e-06, "loss": 0.948, "step": 4383 }, { "epoch": 0.8437259430331023, "grad_norm": 1.974442720413208, "learning_rate": 1.2538029286060428e-06, "loss": 0.9606, "step": 4384 }, { "epoch": 0.8439183987682833, "grad_norm": 1.7231429815292358, "learning_rate": 1.2507826550373614e-06, "loss": 0.9209, "step": 4385 }, { "epoch": 0.8441108545034642, "grad_norm": 1.7420567274093628, "learning_rate": 1.2477657809124632e-06, "loss": 0.893, "step": 4386 }, { "epoch": 0.8443033102386451, "grad_norm": 1.8697997331619263, "learning_rate": 1.244752307403534e-06, "loss": 0.985, "step": 4387 }, { "epoch": 0.844495765973826, "grad_norm": 2.256493330001831, "learning_rate": 1.2417422356814345e-06, "loss": 0.897, "step": 4388 }, { "epoch": 0.8446882217090069, "grad_norm": 1.5790531635284424, "learning_rate": 1.2387355669157076e-06, "loss": 0.9565, "step": 4389 }, { "epoch": 0.8448806774441878, "grad_norm": 1.9034937620162964, "learning_rate": 1.2357323022745716e-06, "loss": 0.9379, "step": 4390 }, { "epoch": 0.8450731331793687, "grad_norm": 1.7414028644561768, "learning_rate": 1.2327324429249232e-06, "loss": 0.9392, "step": 4391 }, { "epoch": 0.8452655889145496, "grad_norm": 1.6371206045150757, "learning_rate": 1.2297359900323346e-06, "loss": 0.8588, "step": 4392 }, { "epoch": 0.8454580446497305, "grad_norm": 1.8576784133911133, "learning_rate": 1.2267429447610557e-06, "loss": 0.8482, "step": 4393 }, { "epoch": 0.8456505003849115, "grad_norm": 1.7783126831054688, "learning_rate": 1.2237533082740127e-06, "loss": 0.9558, "step": 4394 }, { "epoch": 0.8458429561200924, "grad_norm": 1.797315239906311, "learning_rate": 1.2207670817328065e-06, "loss": 0.8707, "step": 4395 }, { "epoch": 0.8460354118552733, "grad_norm": 1.7101528644561768, "learning_rate": 1.2177842662977136e-06, "loss": 0.9364, "step": 4396 }, { "epoch": 0.8462278675904542, "grad_norm": 1.6453410387039185, "learning_rate": 1.2148048631276876e-06, "loss": 0.9407, "step": 4397 }, { "epoch": 0.8464203233256351, "grad_norm": 2.000742197036743, "learning_rate": 1.2118288733803474e-06, "loss": 0.9503, "step": 4398 }, { "epoch": 0.846612779060816, "grad_norm": 1.6722896099090576, "learning_rate": 1.2088562982119957e-06, "loss": 0.9585, "step": 4399 }, { "epoch": 0.8468052347959969, "grad_norm": 0.8880183100700378, "learning_rate": 1.2058871387776039e-06, "loss": 0.8905, "step": 4400 }, { "epoch": 0.8469976905311778, "grad_norm": 1.8514553308486938, "learning_rate": 1.2029213962308172e-06, "loss": 0.9337, "step": 4401 }, { "epoch": 0.8471901462663587, "grad_norm": 1.6902151107788086, "learning_rate": 1.199959071723953e-06, "loss": 0.9123, "step": 4402 }, { "epoch": 0.8473826020015397, "grad_norm": 2.0223639011383057, "learning_rate": 1.197000166408001e-06, "loss": 1.0644, "step": 4403 }, { "epoch": 0.8475750577367206, "grad_norm": 1.7263312339782715, "learning_rate": 1.19404468143262e-06, "loss": 0.8748, "step": 4404 }, { "epoch": 0.8477675134719015, "grad_norm": 1.6143752336502075, "learning_rate": 1.1910926179461446e-06, "loss": 0.9377, "step": 4405 }, { "epoch": 0.8479599692070824, "grad_norm": 3.7136077880859375, "learning_rate": 1.188143977095576e-06, "loss": 0.8526, "step": 4406 }, { "epoch": 0.8481524249422633, "grad_norm": 1.66621732711792, "learning_rate": 1.185198760026587e-06, "loss": 0.9503, "step": 4407 }, { "epoch": 0.8483448806774442, "grad_norm": 1.791013240814209, "learning_rate": 1.1822569678835195e-06, "loss": 0.918, "step": 4408 }, { "epoch": 0.8485373364126251, "grad_norm": 1.7674356698989868, "learning_rate": 1.179318601809386e-06, "loss": 0.8827, "step": 4409 }, { "epoch": 0.848729792147806, "grad_norm": 1.5898926258087158, "learning_rate": 1.176383662945866e-06, "loss": 0.8522, "step": 4410 }, { "epoch": 0.8489222478829869, "grad_norm": 1.6146020889282227, "learning_rate": 1.1734521524333087e-06, "loss": 0.9583, "step": 4411 }, { "epoch": 0.8491147036181679, "grad_norm": 1.9493858814239502, "learning_rate": 1.1705240714107301e-06, "loss": 0.9823, "step": 4412 }, { "epoch": 0.8493071593533488, "grad_norm": 1.6889700889587402, "learning_rate": 1.167599421015818e-06, "loss": 0.8396, "step": 4413 }, { "epoch": 0.8494996150885297, "grad_norm": 1.672223448753357, "learning_rate": 1.164678202384919e-06, "loss": 0.8673, "step": 4414 }, { "epoch": 0.8496920708237106, "grad_norm": 1.8779722452163696, "learning_rate": 1.1617604166530527e-06, "loss": 0.9261, "step": 4415 }, { "epoch": 0.8498845265588915, "grad_norm": 1.8510607481002808, "learning_rate": 1.1588460649539036e-06, "loss": 0.8713, "step": 4416 }, { "epoch": 0.8500769822940724, "grad_norm": 1.801008701324463, "learning_rate": 1.1559351484198211e-06, "loss": 0.9058, "step": 4417 }, { "epoch": 0.8502694380292533, "grad_norm": 1.8866653442382812, "learning_rate": 1.153027668181821e-06, "loss": 1.021, "step": 4418 }, { "epoch": 0.8504618937644342, "grad_norm": 1.7555983066558838, "learning_rate": 1.1501236253695823e-06, "loss": 0.9344, "step": 4419 }, { "epoch": 0.850654349499615, "grad_norm": 1.8575620651245117, "learning_rate": 1.1472230211114498e-06, "loss": 0.8819, "step": 4420 }, { "epoch": 0.8508468052347959, "grad_norm": 1.605782389640808, "learning_rate": 1.1443258565344329e-06, "loss": 0.8483, "step": 4421 }, { "epoch": 0.851039260969977, "grad_norm": 1.8324559926986694, "learning_rate": 1.1414321327642019e-06, "loss": 0.8272, "step": 4422 }, { "epoch": 0.8512317167051578, "grad_norm": 0.8643845915794373, "learning_rate": 1.1385418509250933e-06, "loss": 0.8631, "step": 4423 }, { "epoch": 0.8514241724403387, "grad_norm": 1.796787977218628, "learning_rate": 1.1356550121401033e-06, "loss": 0.9059, "step": 4424 }, { "epoch": 0.8516166281755196, "grad_norm": 1.7553845643997192, "learning_rate": 1.1327716175308934e-06, "loss": 0.8934, "step": 4425 }, { "epoch": 0.8518090839107005, "grad_norm": 1.5919550657272339, "learning_rate": 1.129891668217783e-06, "loss": 0.9811, "step": 4426 }, { "epoch": 0.8520015396458814, "grad_norm": 0.8372297883033752, "learning_rate": 1.1270151653197569e-06, "loss": 0.8313, "step": 4427 }, { "epoch": 0.8521939953810623, "grad_norm": 1.6414622068405151, "learning_rate": 1.124142109954459e-06, "loss": 0.9459, "step": 4428 }, { "epoch": 0.8523864511162432, "grad_norm": 1.5787591934204102, "learning_rate": 1.1212725032381933e-06, "loss": 0.9529, "step": 4429 }, { "epoch": 0.8525789068514241, "grad_norm": 1.7654955387115479, "learning_rate": 1.1184063462859262e-06, "loss": 1.0215, "step": 4430 }, { "epoch": 0.8527713625866051, "grad_norm": 1.5774303674697876, "learning_rate": 1.1155436402112785e-06, "loss": 0.8637, "step": 4431 }, { "epoch": 0.852963818321786, "grad_norm": 1.9222971200942993, "learning_rate": 1.1126843861265347e-06, "loss": 0.9116, "step": 4432 }, { "epoch": 0.8531562740569669, "grad_norm": 1.5580732822418213, "learning_rate": 1.1098285851426372e-06, "loss": 0.9055, "step": 4433 }, { "epoch": 0.8533487297921478, "grad_norm": 1.5990151166915894, "learning_rate": 1.1069762383691862e-06, "loss": 0.8942, "step": 4434 }, { "epoch": 0.8535411855273287, "grad_norm": 2.0196633338928223, "learning_rate": 1.1041273469144399e-06, "loss": 0.8641, "step": 4435 }, { "epoch": 0.8537336412625096, "grad_norm": 1.6549125909805298, "learning_rate": 1.1012819118853147e-06, "loss": 0.9466, "step": 4436 }, { "epoch": 0.8539260969976905, "grad_norm": 2.0572917461395264, "learning_rate": 1.098439934387383e-06, "loss": 0.9179, "step": 4437 }, { "epoch": 0.8541185527328714, "grad_norm": 2.320844888687134, "learning_rate": 1.095601415524874e-06, "loss": 0.9259, "step": 4438 }, { "epoch": 0.8543110084680523, "grad_norm": 1.6620333194732666, "learning_rate": 1.092766356400673e-06, "loss": 0.8888, "step": 4439 }, { "epoch": 0.8545034642032333, "grad_norm": 1.808110237121582, "learning_rate": 1.0899347581163222e-06, "loss": 0.9659, "step": 4440 }, { "epoch": 0.8546959199384142, "grad_norm": 1.797581434249878, "learning_rate": 1.0871066217720173e-06, "loss": 0.9873, "step": 4441 }, { "epoch": 0.8548883756735951, "grad_norm": 1.5478016138076782, "learning_rate": 1.08428194846661e-06, "loss": 0.9049, "step": 4442 }, { "epoch": 0.855080831408776, "grad_norm": 1.7699617147445679, "learning_rate": 1.0814607392976062e-06, "loss": 0.9378, "step": 4443 }, { "epoch": 0.8552732871439569, "grad_norm": 1.8112612962722778, "learning_rate": 1.0786429953611665e-06, "loss": 0.8988, "step": 4444 }, { "epoch": 0.8554657428791378, "grad_norm": 1.6674292087554932, "learning_rate": 1.0758287177521032e-06, "loss": 0.996, "step": 4445 }, { "epoch": 0.8556581986143187, "grad_norm": 1.7372832298278809, "learning_rate": 1.073017907563887e-06, "loss": 0.963, "step": 4446 }, { "epoch": 0.8558506543494996, "grad_norm": 1.8815265893936157, "learning_rate": 1.0702105658886318e-06, "loss": 0.9793, "step": 4447 }, { "epoch": 0.8560431100846805, "grad_norm": 1.7649062871932983, "learning_rate": 1.0674066938171123e-06, "loss": 0.939, "step": 4448 }, { "epoch": 0.8562355658198614, "grad_norm": 1.5869271755218506, "learning_rate": 1.0646062924387513e-06, "loss": 0.9759, "step": 4449 }, { "epoch": 0.8564280215550424, "grad_norm": 1.8112019300460815, "learning_rate": 1.061809362841626e-06, "loss": 0.9141, "step": 4450 }, { "epoch": 0.8566204772902233, "grad_norm": 0.8680244088172913, "learning_rate": 1.0590159061124606e-06, "loss": 0.9309, "step": 4451 }, { "epoch": 0.8568129330254042, "grad_norm": 1.7235769033432007, "learning_rate": 1.0562259233366334e-06, "loss": 0.9951, "step": 4452 }, { "epoch": 0.8570053887605851, "grad_norm": 1.8003859519958496, "learning_rate": 1.053439415598171e-06, "loss": 0.9351, "step": 4453 }, { "epoch": 0.857197844495766, "grad_norm": 1.8437814712524414, "learning_rate": 1.0506563839797501e-06, "loss": 0.9457, "step": 4454 }, { "epoch": 0.8573903002309469, "grad_norm": 1.9507439136505127, "learning_rate": 1.0478768295626985e-06, "loss": 0.8406, "step": 4455 }, { "epoch": 0.8575827559661278, "grad_norm": 1.8357292413711548, "learning_rate": 1.0451007534269908e-06, "loss": 0.957, "step": 4456 }, { "epoch": 0.8577752117013087, "grad_norm": 1.682597041130066, "learning_rate": 1.04232815665125e-06, "loss": 0.9884, "step": 4457 }, { "epoch": 0.8579676674364896, "grad_norm": 1.6805672645568848, "learning_rate": 1.0395590403127487e-06, "loss": 0.9107, "step": 4458 }, { "epoch": 0.8581601231716706, "grad_norm": 1.8189362287521362, "learning_rate": 1.0367934054874074e-06, "loss": 0.9128, "step": 4459 }, { "epoch": 0.8583525789068515, "grad_norm": 1.6221890449523926, "learning_rate": 1.034031253249792e-06, "loss": 0.8995, "step": 4460 }, { "epoch": 0.8585450346420324, "grad_norm": 1.6082977056503296, "learning_rate": 1.0312725846731174e-06, "loss": 0.9388, "step": 4461 }, { "epoch": 0.8587374903772133, "grad_norm": 1.572615623474121, "learning_rate": 1.0285174008292431e-06, "loss": 0.944, "step": 4462 }, { "epoch": 0.8589299461123941, "grad_norm": 1.5556401014328003, "learning_rate": 1.025765702788677e-06, "loss": 0.9001, "step": 4463 }, { "epoch": 0.859122401847575, "grad_norm": 1.798750638961792, "learning_rate": 1.0230174916205681e-06, "loss": 0.8928, "step": 4464 }, { "epoch": 0.8593148575827559, "grad_norm": 1.809114694595337, "learning_rate": 1.0202727683927149e-06, "loss": 0.9273, "step": 4465 }, { "epoch": 0.8595073133179368, "grad_norm": 0.899472177028656, "learning_rate": 1.0175315341715598e-06, "loss": 0.9201, "step": 4466 }, { "epoch": 0.8596997690531177, "grad_norm": 1.824232578277588, "learning_rate": 1.0147937900221882e-06, "loss": 0.9495, "step": 4467 }, { "epoch": 0.8598922247882987, "grad_norm": 1.6713426113128662, "learning_rate": 1.012059537008332e-06, "loss": 0.9502, "step": 4468 }, { "epoch": 0.8600846805234796, "grad_norm": 1.7595478296279907, "learning_rate": 1.009328776192363e-06, "loss": 0.9036, "step": 4469 }, { "epoch": 0.8602771362586605, "grad_norm": 1.6076364517211914, "learning_rate": 1.0066015086353e-06, "loss": 0.8601, "step": 4470 }, { "epoch": 0.8604695919938414, "grad_norm": 1.7671749591827393, "learning_rate": 1.003877735396801e-06, "loss": 0.8929, "step": 4471 }, { "epoch": 0.8606620477290223, "grad_norm": 1.878847360610962, "learning_rate": 1.0011574575351702e-06, "loss": 0.8674, "step": 4472 }, { "epoch": 0.8608545034642032, "grad_norm": 0.8943829536437988, "learning_rate": 9.984406761073496e-07, "loss": 0.8963, "step": 4473 }, { "epoch": 0.8610469591993841, "grad_norm": 1.6800854206085205, "learning_rate": 9.957273921689237e-07, "loss": 0.8551, "step": 4474 }, { "epoch": 0.861239414934565, "grad_norm": 1.7575600147247314, "learning_rate": 9.930176067741216e-07, "loss": 0.9517, "step": 4475 }, { "epoch": 0.8614318706697459, "grad_norm": 1.6423978805541992, "learning_rate": 9.903113209758098e-07, "loss": 0.9339, "step": 4476 }, { "epoch": 0.8616243264049269, "grad_norm": 1.7209537029266357, "learning_rate": 9.87608535825494e-07, "loss": 0.8767, "step": 4477 }, { "epoch": 0.8618167821401078, "grad_norm": 0.9240423440933228, "learning_rate": 9.849092523733217e-07, "loss": 0.9224, "step": 4478 }, { "epoch": 0.8620092378752887, "grad_norm": 1.9497654438018799, "learning_rate": 9.822134716680832e-07, "loss": 0.8218, "step": 4479 }, { "epoch": 0.8622016936104696, "grad_norm": 1.6518983840942383, "learning_rate": 9.795211947571993e-07, "loss": 0.9407, "step": 4480 }, { "epoch": 0.8623941493456505, "grad_norm": 1.715362310409546, "learning_rate": 9.768324226867353e-07, "loss": 0.9336, "step": 4481 }, { "epoch": 0.8625866050808314, "grad_norm": 2.010856866836548, "learning_rate": 9.74147156501396e-07, "loss": 0.9364, "step": 4482 }, { "epoch": 0.8627790608160123, "grad_norm": 1.780233383178711, "learning_rate": 9.714653972445199e-07, "loss": 0.8251, "step": 4483 }, { "epoch": 0.8629715165511932, "grad_norm": 1.7593193054199219, "learning_rate": 9.687871459580845e-07, "loss": 0.9162, "step": 4484 }, { "epoch": 0.8631639722863741, "grad_norm": 1.6943724155426025, "learning_rate": 9.661124036827064e-07, "loss": 0.9424, "step": 4485 }, { "epoch": 0.863356428021555, "grad_norm": 1.59291410446167, "learning_rate": 9.634411714576353e-07, "loss": 0.8911, "step": 4486 }, { "epoch": 0.863548883756736, "grad_norm": 1.8310132026672363, "learning_rate": 9.60773450320759e-07, "loss": 0.9136, "step": 4487 }, { "epoch": 0.8637413394919169, "grad_norm": 1.770247459411621, "learning_rate": 9.58109241308599e-07, "loss": 0.8781, "step": 4488 }, { "epoch": 0.8639337952270978, "grad_norm": 1.717069149017334, "learning_rate": 9.55448545456319e-07, "loss": 0.8687, "step": 4489 }, { "epoch": 0.8641262509622787, "grad_norm": 1.7650303840637207, "learning_rate": 9.527913637977093e-07, "loss": 0.9767, "step": 4490 }, { "epoch": 0.8643187066974596, "grad_norm": 2.1053173542022705, "learning_rate": 9.501376973651999e-07, "loss": 0.9659, "step": 4491 }, { "epoch": 0.8645111624326405, "grad_norm": 1.7617216110229492, "learning_rate": 9.474875471898526e-07, "loss": 0.9003, "step": 4492 }, { "epoch": 0.8647036181678214, "grad_norm": 1.8802812099456787, "learning_rate": 9.448409143013626e-07, "loss": 1.0688, "step": 4493 }, { "epoch": 0.8648960739030023, "grad_norm": 1.665010929107666, "learning_rate": 9.421977997280596e-07, "loss": 0.9637, "step": 4494 }, { "epoch": 0.8650885296381832, "grad_norm": 1.965577483177185, "learning_rate": 9.395582044969076e-07, "loss": 0.871, "step": 4495 }, { "epoch": 0.8652809853733642, "grad_norm": 1.7655526399612427, "learning_rate": 9.369221296335007e-07, "loss": 0.809, "step": 4496 }, { "epoch": 0.8654734411085451, "grad_norm": 0.9457436203956604, "learning_rate": 9.342895761620663e-07, "loss": 0.9843, "step": 4497 }, { "epoch": 0.865665896843726, "grad_norm": 1.8821799755096436, "learning_rate": 9.31660545105465e-07, "loss": 0.9675, "step": 4498 }, { "epoch": 0.8658583525789069, "grad_norm": 1.6772260665893555, "learning_rate": 9.290350374851865e-07, "loss": 0.8678, "step": 4499 }, { "epoch": 0.8660508083140878, "grad_norm": 1.7377560138702393, "learning_rate": 9.264130543213512e-07, "loss": 0.947, "step": 4500 }, { "epoch": 0.8662432640492687, "grad_norm": 1.548792839050293, "learning_rate": 9.237945966327133e-07, "loss": 0.9132, "step": 4501 }, { "epoch": 0.8664357197844496, "grad_norm": 1.6864455938339233, "learning_rate": 9.211796654366545e-07, "loss": 0.9811, "step": 4502 }, { "epoch": 0.8666281755196305, "grad_norm": 1.7881170511245728, "learning_rate": 9.185682617491865e-07, "loss": 0.8339, "step": 4503 }, { "epoch": 0.8668206312548113, "grad_norm": 1.8653318881988525, "learning_rate": 9.15960386584952e-07, "loss": 0.9063, "step": 4504 }, { "epoch": 0.8670130869899924, "grad_norm": 0.9954100847244263, "learning_rate": 9.133560409572195e-07, "loss": 0.9171, "step": 4505 }, { "epoch": 0.8672055427251733, "grad_norm": 1.6978204250335693, "learning_rate": 9.107552258778907e-07, "loss": 0.897, "step": 4506 }, { "epoch": 0.8673979984603541, "grad_norm": 1.6766725778579712, "learning_rate": 9.081579423574916e-07, "loss": 0.9211, "step": 4507 }, { "epoch": 0.867590454195535, "grad_norm": 1.8919655084609985, "learning_rate": 9.055641914051783e-07, "loss": 0.908, "step": 4508 }, { "epoch": 0.8677829099307159, "grad_norm": 1.9327479600906372, "learning_rate": 9.029739740287358e-07, "loss": 0.9173, "step": 4509 }, { "epoch": 0.8679753656658968, "grad_norm": 1.6222783327102661, "learning_rate": 9.00387291234569e-07, "loss": 0.9597, "step": 4510 }, { "epoch": 0.8681678214010777, "grad_norm": 1.6995222568511963, "learning_rate": 8.978041440277163e-07, "loss": 1.0014, "step": 4511 }, { "epoch": 0.8683602771362586, "grad_norm": 2.417156934738159, "learning_rate": 8.952245334118415e-07, "loss": 0.8324, "step": 4512 }, { "epoch": 0.8685527328714395, "grad_norm": 1.7888003587722778, "learning_rate": 8.926484603892327e-07, "loss": 0.8735, "step": 4513 }, { "epoch": 0.8687451886066204, "grad_norm": 1.9681522846221924, "learning_rate": 8.900759259608038e-07, "loss": 0.9334, "step": 4514 }, { "epoch": 0.8689376443418014, "grad_norm": 0.9685776829719543, "learning_rate": 8.875069311260942e-07, "loss": 0.9593, "step": 4515 }, { "epoch": 0.8691301000769823, "grad_norm": 1.6864736080169678, "learning_rate": 8.849414768832687e-07, "loss": 0.8925, "step": 4516 }, { "epoch": 0.8693225558121632, "grad_norm": 1.6326510906219482, "learning_rate": 8.823795642291145e-07, "loss": 0.9152, "step": 4517 }, { "epoch": 0.8695150115473441, "grad_norm": 1.6503970623016357, "learning_rate": 8.79821194159044e-07, "loss": 0.93, "step": 4518 }, { "epoch": 0.869707467282525, "grad_norm": 1.6019023656845093, "learning_rate": 8.772663676670934e-07, "loss": 0.8682, "step": 4519 }, { "epoch": 0.8698999230177059, "grad_norm": 1.6742379665374756, "learning_rate": 8.747150857459219e-07, "loss": 0.9086, "step": 4520 }, { "epoch": 0.8700923787528868, "grad_norm": 1.7541180849075317, "learning_rate": 8.721673493868111e-07, "loss": 0.8774, "step": 4521 }, { "epoch": 0.8702848344880677, "grad_norm": 1.6651214361190796, "learning_rate": 8.696231595796633e-07, "loss": 0.946, "step": 4522 }, { "epoch": 0.8704772902232486, "grad_norm": 1.8287791013717651, "learning_rate": 8.670825173130081e-07, "loss": 0.907, "step": 4523 }, { "epoch": 0.8706697459584296, "grad_norm": 1.978616714477539, "learning_rate": 8.645454235739903e-07, "loss": 0.929, "step": 4524 }, { "epoch": 0.8708622016936105, "grad_norm": 1.7962009906768799, "learning_rate": 8.620118793483823e-07, "loss": 0.9137, "step": 4525 }, { "epoch": 0.8710546574287914, "grad_norm": 1.768997311592102, "learning_rate": 8.5948188562057e-07, "loss": 0.9698, "step": 4526 }, { "epoch": 0.8712471131639723, "grad_norm": 1.7460027933120728, "learning_rate": 8.56955443373565e-07, "loss": 0.8877, "step": 4527 }, { "epoch": 0.8714395688991532, "grad_norm": 1.7546968460083008, "learning_rate": 8.544325535889997e-07, "loss": 0.9021, "step": 4528 }, { "epoch": 0.8716320246343341, "grad_norm": 1.7033549547195435, "learning_rate": 8.519132172471212e-07, "loss": 0.8843, "step": 4529 }, { "epoch": 0.871824480369515, "grad_norm": 1.680238962173462, "learning_rate": 8.493974353268019e-07, "loss": 0.8902, "step": 4530 }, { "epoch": 0.8720169361046959, "grad_norm": 1.7112008333206177, "learning_rate": 8.468852088055291e-07, "loss": 0.9099, "step": 4531 }, { "epoch": 0.8722093918398768, "grad_norm": 1.6180204153060913, "learning_rate": 8.443765386594094e-07, "loss": 0.985, "step": 4532 }, { "epoch": 0.8724018475750578, "grad_norm": 1.7364057302474976, "learning_rate": 8.418714258631689e-07, "loss": 0.8279, "step": 4533 }, { "epoch": 0.8725943033102387, "grad_norm": 1.641308307647705, "learning_rate": 8.393698713901499e-07, "loss": 0.8845, "step": 4534 }, { "epoch": 0.8727867590454196, "grad_norm": 0.8223841786384583, "learning_rate": 8.36871876212314e-07, "loss": 0.924, "step": 4535 }, { "epoch": 0.8729792147806005, "grad_norm": 1.646851897239685, "learning_rate": 8.343774413002382e-07, "loss": 0.9319, "step": 4536 }, { "epoch": 0.8731716705157814, "grad_norm": 1.8828119039535522, "learning_rate": 8.318865676231158e-07, "loss": 0.8897, "step": 4537 }, { "epoch": 0.8733641262509623, "grad_norm": 1.880627989768982, "learning_rate": 8.293992561487596e-07, "loss": 0.9193, "step": 4538 }, { "epoch": 0.8735565819861432, "grad_norm": 1.7291349172592163, "learning_rate": 8.26915507843593e-07, "loss": 0.904, "step": 4539 }, { "epoch": 0.8737490377213241, "grad_norm": 2.4312212467193604, "learning_rate": 8.24435323672661e-07, "loss": 0.9305, "step": 4540 }, { "epoch": 0.873941493456505, "grad_norm": 1.7694976329803467, "learning_rate": 8.219587045996203e-07, "loss": 0.9386, "step": 4541 }, { "epoch": 0.874133949191686, "grad_norm": 1.9065922498703003, "learning_rate": 8.194856515867433e-07, "loss": 0.8363, "step": 4542 }, { "epoch": 0.8743264049268669, "grad_norm": 1.623293399810791, "learning_rate": 8.170161655949139e-07, "loss": 0.9275, "step": 4543 }, { "epoch": 0.8745188606620478, "grad_norm": 1.642863154411316, "learning_rate": 8.145502475836331e-07, "loss": 0.9967, "step": 4544 }, { "epoch": 0.8747113163972287, "grad_norm": 1.745168924331665, "learning_rate": 8.120878985110181e-07, "loss": 0.934, "step": 4545 }, { "epoch": 0.8749037721324096, "grad_norm": 1.8920714855194092, "learning_rate": 8.096291193337935e-07, "loss": 0.9796, "step": 4546 }, { "epoch": 0.8750962278675904, "grad_norm": 1.9377678632736206, "learning_rate": 8.071739110073018e-07, "loss": 0.9633, "step": 4547 }, { "epoch": 0.8752886836027713, "grad_norm": 1.8712385892868042, "learning_rate": 8.047222744854943e-07, "loss": 0.9616, "step": 4548 }, { "epoch": 0.8754811393379522, "grad_norm": 1.972001075744629, "learning_rate": 8.022742107209369e-07, "loss": 0.8705, "step": 4549 }, { "epoch": 0.8756735950731331, "grad_norm": 1.7095370292663574, "learning_rate": 7.998297206648075e-07, "loss": 0.9578, "step": 4550 }, { "epoch": 0.875866050808314, "grad_norm": 1.6083623170852661, "learning_rate": 7.973888052668943e-07, "loss": 0.8513, "step": 4551 }, { "epoch": 0.876058506543495, "grad_norm": 1.7789794206619263, "learning_rate": 7.949514654755963e-07, "loss": 0.8984, "step": 4552 }, { "epoch": 0.8762509622786759, "grad_norm": 1.7920817136764526, "learning_rate": 7.925177022379249e-07, "loss": 0.9401, "step": 4553 }, { "epoch": 0.8764434180138568, "grad_norm": 1.7534286975860596, "learning_rate": 7.900875164995003e-07, "loss": 0.8554, "step": 4554 }, { "epoch": 0.8766358737490377, "grad_norm": 1.97467041015625, "learning_rate": 7.876609092045528e-07, "loss": 0.9698, "step": 4555 }, { "epoch": 0.8768283294842186, "grad_norm": 3.238448143005371, "learning_rate": 7.852378812959227e-07, "loss": 0.8215, "step": 4556 }, { "epoch": 0.8770207852193995, "grad_norm": 0.946961522102356, "learning_rate": 7.828184337150613e-07, "loss": 0.8811, "step": 4557 }, { "epoch": 0.8772132409545804, "grad_norm": 1.5218231678009033, "learning_rate": 7.804025674020266e-07, "loss": 0.8475, "step": 4558 }, { "epoch": 0.8774056966897613, "grad_norm": 1.7492682933807373, "learning_rate": 7.779902832954833e-07, "loss": 0.921, "step": 4559 }, { "epoch": 0.8775981524249422, "grad_norm": 1.8103415966033936, "learning_rate": 7.755815823327084e-07, "loss": 0.967, "step": 4560 }, { "epoch": 0.8777906081601232, "grad_norm": 1.7757612466812134, "learning_rate": 7.731764654495832e-07, "loss": 0.9454, "step": 4561 }, { "epoch": 0.8779830638953041, "grad_norm": 1.7613611221313477, "learning_rate": 7.707749335806014e-07, "loss": 0.8882, "step": 4562 }, { "epoch": 0.878175519630485, "grad_norm": 1.8285560607910156, "learning_rate": 7.683769876588576e-07, "loss": 0.8332, "step": 4563 }, { "epoch": 0.8783679753656659, "grad_norm": 1.9224072694778442, "learning_rate": 7.659826286160565e-07, "loss": 0.8837, "step": 4564 }, { "epoch": 0.8785604311008468, "grad_norm": 1.6285111904144287, "learning_rate": 7.635918573825096e-07, "loss": 0.9368, "step": 4565 }, { "epoch": 0.8787528868360277, "grad_norm": 1.7547639608383179, "learning_rate": 7.612046748871327e-07, "loss": 0.9192, "step": 4566 }, { "epoch": 0.8789453425712086, "grad_norm": 1.7078752517700195, "learning_rate": 7.588210820574482e-07, "loss": 0.9591, "step": 4567 }, { "epoch": 0.8791377983063895, "grad_norm": 1.8804042339324951, "learning_rate": 7.564410798195832e-07, "loss": 0.9591, "step": 4568 }, { "epoch": 0.8793302540415704, "grad_norm": 1.9545427560806274, "learning_rate": 7.540646690982711e-07, "loss": 0.8947, "step": 4569 }, { "epoch": 0.8795227097767514, "grad_norm": 1.8150368928909302, "learning_rate": 7.516918508168469e-07, "loss": 0.9425, "step": 4570 }, { "epoch": 0.8797151655119323, "grad_norm": 1.649187684059143, "learning_rate": 7.493226258972519e-07, "loss": 0.9377, "step": 4571 }, { "epoch": 0.8799076212471132, "grad_norm": 1.760579228401184, "learning_rate": 7.46956995260033e-07, "loss": 0.9453, "step": 4572 }, { "epoch": 0.8801000769822941, "grad_norm": 1.9679381847381592, "learning_rate": 7.445949598243362e-07, "loss": 0.9226, "step": 4573 }, { "epoch": 0.880292532717475, "grad_norm": 1.7522602081298828, "learning_rate": 7.422365205079141e-07, "loss": 0.976, "step": 4574 }, { "epoch": 0.8804849884526559, "grad_norm": 1.6734532117843628, "learning_rate": 7.398816782271223e-07, "loss": 0.9291, "step": 4575 }, { "epoch": 0.8806774441878368, "grad_norm": 1.6494511365890503, "learning_rate": 7.375304338969135e-07, "loss": 0.9094, "step": 4576 }, { "epoch": 0.8808698999230177, "grad_norm": 1.8996403217315674, "learning_rate": 7.351827884308493e-07, "loss": 0.9279, "step": 4577 }, { "epoch": 0.8810623556581986, "grad_norm": 1.6665183305740356, "learning_rate": 7.328387427410877e-07, "loss": 0.9234, "step": 4578 }, { "epoch": 0.8812548113933796, "grad_norm": 1.6807657480239868, "learning_rate": 7.304982977383935e-07, "loss": 0.8557, "step": 4579 }, { "epoch": 0.8814472671285605, "grad_norm": 2.0911595821380615, "learning_rate": 7.281614543321269e-07, "loss": 0.8437, "step": 4580 }, { "epoch": 0.8816397228637414, "grad_norm": 1.8205580711364746, "learning_rate": 7.258282134302519e-07, "loss": 0.8678, "step": 4581 }, { "epoch": 0.8818321785989223, "grad_norm": 1.7864078283309937, "learning_rate": 7.234985759393321e-07, "loss": 0.9842, "step": 4582 }, { "epoch": 0.8820246343341032, "grad_norm": 1.7867627143859863, "learning_rate": 7.211725427645311e-07, "loss": 0.9134, "step": 4583 }, { "epoch": 0.8822170900692841, "grad_norm": 1.9057732820510864, "learning_rate": 7.188501148096117e-07, "loss": 0.8677, "step": 4584 }, { "epoch": 0.882409545804465, "grad_norm": 1.5514980554580688, "learning_rate": 7.165312929769352e-07, "loss": 0.9439, "step": 4585 }, { "epoch": 0.8826020015396459, "grad_norm": 1.787164330482483, "learning_rate": 7.142160781674645e-07, "loss": 0.9219, "step": 4586 }, { "epoch": 0.8827944572748267, "grad_norm": 1.799926519393921, "learning_rate": 7.119044712807577e-07, "loss": 0.892, "step": 4587 }, { "epoch": 0.8829869130100076, "grad_norm": 1.9440863132476807, "learning_rate": 7.095964732149741e-07, "loss": 0.9379, "step": 4588 }, { "epoch": 0.8831793687451887, "grad_norm": 2.01857328414917, "learning_rate": 7.072920848668674e-07, "loss": 0.9031, "step": 4589 }, { "epoch": 0.8833718244803695, "grad_norm": 1.8090003728866577, "learning_rate": 7.049913071317927e-07, "loss": 0.9475, "step": 4590 }, { "epoch": 0.8835642802155504, "grad_norm": 1.8560371398925781, "learning_rate": 7.026941409036991e-07, "loss": 0.8428, "step": 4591 }, { "epoch": 0.8837567359507313, "grad_norm": 1.846830129623413, "learning_rate": 7.004005870751341e-07, "loss": 0.9201, "step": 4592 }, { "epoch": 0.8839491916859122, "grad_norm": 1.742576241493225, "learning_rate": 6.981106465372389e-07, "loss": 0.8708, "step": 4593 }, { "epoch": 0.8841416474210931, "grad_norm": 1.601747751235962, "learning_rate": 6.958243201797554e-07, "loss": 1.0001, "step": 4594 }, { "epoch": 0.884334103156274, "grad_norm": 1.9314178228378296, "learning_rate": 6.935416088910185e-07, "loss": 0.9818, "step": 4595 }, { "epoch": 0.8845265588914549, "grad_norm": 0.9659202694892883, "learning_rate": 6.912625135579587e-07, "loss": 0.9803, "step": 4596 }, { "epoch": 0.8847190146266358, "grad_norm": 1.6690056324005127, "learning_rate": 6.889870350661009e-07, "loss": 0.8706, "step": 4597 }, { "epoch": 0.8849114703618168, "grad_norm": 1.8155843019485474, "learning_rate": 6.867151742995659e-07, "loss": 0.8716, "step": 4598 }, { "epoch": 0.8851039260969977, "grad_norm": 1.7352817058563232, "learning_rate": 6.844469321410685e-07, "loss": 0.9118, "step": 4599 }, { "epoch": 0.8852963818321786, "grad_norm": 1.7169686555862427, "learning_rate": 6.821823094719171e-07, "loss": 0.906, "step": 4600 }, { "epoch": 0.8854888375673595, "grad_norm": 1.771855354309082, "learning_rate": 6.799213071720156e-07, "loss": 0.9104, "step": 4601 }, { "epoch": 0.8856812933025404, "grad_norm": 1.7238482236862183, "learning_rate": 6.776639261198581e-07, "loss": 0.9184, "step": 4602 }, { "epoch": 0.8858737490377213, "grad_norm": 1.7848094701766968, "learning_rate": 6.754101671925339e-07, "loss": 0.9395, "step": 4603 }, { "epoch": 0.8860662047729022, "grad_norm": 0.9155448079109192, "learning_rate": 6.731600312657238e-07, "loss": 0.9273, "step": 4604 }, { "epoch": 0.8862586605080831, "grad_norm": 2.1516246795654297, "learning_rate": 6.709135192137018e-07, "loss": 0.9158, "step": 4605 }, { "epoch": 0.886451116243264, "grad_norm": 1.7109227180480957, "learning_rate": 6.68670631909335e-07, "loss": 0.7506, "step": 4606 }, { "epoch": 0.886643571978445, "grad_norm": 1.8024001121520996, "learning_rate": 6.664313702240788e-07, "loss": 0.9931, "step": 4607 }, { "epoch": 0.8868360277136259, "grad_norm": 1.559737205505371, "learning_rate": 6.641957350279838e-07, "loss": 0.8467, "step": 4608 }, { "epoch": 0.8870284834488068, "grad_norm": 1.9477784633636475, "learning_rate": 6.619637271896873e-07, "loss": 0.9116, "step": 4609 }, { "epoch": 0.8872209391839877, "grad_norm": 1.7992230653762817, "learning_rate": 6.597353475764201e-07, "loss": 0.9372, "step": 4610 }, { "epoch": 0.8874133949191686, "grad_norm": 2.300236940383911, "learning_rate": 6.57510597054003e-07, "loss": 0.8063, "step": 4611 }, { "epoch": 0.8876058506543495, "grad_norm": 0.9002736210823059, "learning_rate": 6.552894764868456e-07, "loss": 0.9448, "step": 4612 }, { "epoch": 0.8877983063895304, "grad_norm": 0.8136253356933594, "learning_rate": 6.530719867379498e-07, "loss": 0.8491, "step": 4613 }, { "epoch": 0.8879907621247113, "grad_norm": 1.9341810941696167, "learning_rate": 6.508581286689031e-07, "loss": 0.9963, "step": 4614 }, { "epoch": 0.8881832178598922, "grad_norm": 2.0188567638397217, "learning_rate": 6.48647903139884e-07, "loss": 0.9761, "step": 4615 }, { "epoch": 0.8883756735950731, "grad_norm": 1.5818471908569336, "learning_rate": 6.464413110096601e-07, "loss": 0.9531, "step": 4616 }, { "epoch": 0.8885681293302541, "grad_norm": 1.8132492303848267, "learning_rate": 6.442383531355856e-07, "loss": 0.8182, "step": 4617 }, { "epoch": 0.888760585065435, "grad_norm": 1.619665503501892, "learning_rate": 6.420390303736034e-07, "loss": 0.8305, "step": 4618 }, { "epoch": 0.8889530408006159, "grad_norm": 1.919082522392273, "learning_rate": 6.398433435782459e-07, "loss": 1.0031, "step": 4619 }, { "epoch": 0.8891454965357968, "grad_norm": 1.6701529026031494, "learning_rate": 6.37651293602628e-07, "loss": 0.8564, "step": 4620 }, { "epoch": 0.8893379522709777, "grad_norm": 1.4992094039916992, "learning_rate": 6.354628812984576e-07, "loss": 0.9739, "step": 4621 }, { "epoch": 0.8895304080061586, "grad_norm": 1.6888341903686523, "learning_rate": 6.332781075160244e-07, "loss": 0.8659, "step": 4622 }, { "epoch": 0.8897228637413395, "grad_norm": 1.8748725652694702, "learning_rate": 6.31096973104206e-07, "loss": 0.8955, "step": 4623 }, { "epoch": 0.8899153194765204, "grad_norm": 1.7143971920013428, "learning_rate": 6.28919478910468e-07, "loss": 0.8884, "step": 4624 }, { "epoch": 0.8901077752117013, "grad_norm": 1.6898267269134521, "learning_rate": 6.26745625780858e-07, "loss": 0.9435, "step": 4625 }, { "epoch": 0.8903002309468823, "grad_norm": 1.694114327430725, "learning_rate": 6.245754145600091e-07, "loss": 0.876, "step": 4626 }, { "epoch": 0.8904926866820632, "grad_norm": 1.8522770404815674, "learning_rate": 6.224088460911437e-07, "loss": 0.9518, "step": 4627 }, { "epoch": 0.8906851424172441, "grad_norm": 0.9747973084449768, "learning_rate": 6.202459212160638e-07, "loss": 0.9113, "step": 4628 }, { "epoch": 0.890877598152425, "grad_norm": 1.6128125190734863, "learning_rate": 6.180866407751595e-07, "loss": 0.8555, "step": 4629 }, { "epoch": 0.8910700538876059, "grad_norm": 0.8761985301971436, "learning_rate": 6.15931005607403e-07, "loss": 0.8427, "step": 4630 }, { "epoch": 0.8912625096227867, "grad_norm": 1.7442528009414673, "learning_rate": 6.137790165503499e-07, "loss": 0.9198, "step": 4631 }, { "epoch": 0.8914549653579676, "grad_norm": 1.7197226285934448, "learning_rate": 6.116306744401391e-07, "loss": 0.8894, "step": 4632 }, { "epoch": 0.8916474210931485, "grad_norm": 1.819201111793518, "learning_rate": 6.09485980111495e-07, "loss": 0.9948, "step": 4633 }, { "epoch": 0.8918398768283294, "grad_norm": 1.7525112628936768, "learning_rate": 6.073449343977211e-07, "loss": 0.8711, "step": 4634 }, { "epoch": 0.8920323325635104, "grad_norm": 1.7218610048294067, "learning_rate": 6.052075381307054e-07, "loss": 0.9227, "step": 4635 }, { "epoch": 0.8922247882986913, "grad_norm": 1.7466583251953125, "learning_rate": 6.030737921409169e-07, "loss": 0.8994, "step": 4636 }, { "epoch": 0.8924172440338722, "grad_norm": 1.8194990158081055, "learning_rate": 6.009436972574079e-07, "loss": 0.8556, "step": 4637 }, { "epoch": 0.8926096997690531, "grad_norm": 1.8914660215377808, "learning_rate": 5.988172543078097e-07, "loss": 0.8966, "step": 4638 }, { "epoch": 0.892802155504234, "grad_norm": 1.6951923370361328, "learning_rate": 5.966944641183369e-07, "loss": 0.9769, "step": 4639 }, { "epoch": 0.8929946112394149, "grad_norm": 1.659889578819275, "learning_rate": 5.945753275137844e-07, "loss": 0.8891, "step": 4640 }, { "epoch": 0.8931870669745958, "grad_norm": 1.6990644931793213, "learning_rate": 5.924598453175278e-07, "loss": 0.9166, "step": 4641 }, { "epoch": 0.8933795227097767, "grad_norm": 1.8403247594833374, "learning_rate": 5.903480183515197e-07, "loss": 1.0127, "step": 4642 }, { "epoch": 0.8935719784449576, "grad_norm": 0.8651690483093262, "learning_rate": 5.882398474362949e-07, "loss": 0.8539, "step": 4643 }, { "epoch": 0.8937644341801386, "grad_norm": 0.9350734949111938, "learning_rate": 5.861353333909692e-07, "loss": 0.9161, "step": 4644 }, { "epoch": 0.8939568899153195, "grad_norm": 1.7657092809677124, "learning_rate": 5.840344770332351e-07, "loss": 0.9505, "step": 4645 }, { "epoch": 0.8941493456505004, "grad_norm": 1.9987502098083496, "learning_rate": 5.819372791793654e-07, "loss": 0.8645, "step": 4646 }, { "epoch": 0.8943418013856813, "grad_norm": 1.664176344871521, "learning_rate": 5.798437406442104e-07, "loss": 0.9997, "step": 4647 }, { "epoch": 0.8945342571208622, "grad_norm": 1.6742953062057495, "learning_rate": 5.777538622412005e-07, "loss": 0.8582, "step": 4648 }, { "epoch": 0.8947267128560431, "grad_norm": 1.6213897466659546, "learning_rate": 5.756676447823406e-07, "loss": 0.9446, "step": 4649 }, { "epoch": 0.894919168591224, "grad_norm": 1.7622512578964233, "learning_rate": 5.735850890782158e-07, "loss": 0.8426, "step": 4650 }, { "epoch": 0.8951116243264049, "grad_norm": 1.7775484323501587, "learning_rate": 5.715061959379875e-07, "loss": 0.8734, "step": 4651 }, { "epoch": 0.8953040800615858, "grad_norm": 1.9662108421325684, "learning_rate": 5.694309661693942e-07, "loss": 0.8961, "step": 4652 }, { "epoch": 0.8954965357967667, "grad_norm": 1.971056342124939, "learning_rate": 5.67359400578752e-07, "loss": 0.8707, "step": 4653 }, { "epoch": 0.8956889915319477, "grad_norm": 1.9486395120620728, "learning_rate": 5.652914999709535e-07, "loss": 0.8978, "step": 4654 }, { "epoch": 0.8958814472671286, "grad_norm": 1.6971734762191772, "learning_rate": 5.632272651494631e-07, "loss": 0.943, "step": 4655 }, { "epoch": 0.8960739030023095, "grad_norm": 0.8335411548614502, "learning_rate": 5.611666969163243e-07, "loss": 0.8609, "step": 4656 }, { "epoch": 0.8962663587374904, "grad_norm": 1.8522565364837646, "learning_rate": 5.591097960721581e-07, "loss": 0.7508, "step": 4657 }, { "epoch": 0.8964588144726713, "grad_norm": 1.6077631711959839, "learning_rate": 5.570565634161562e-07, "loss": 0.8509, "step": 4658 }, { "epoch": 0.8966512702078522, "grad_norm": 1.6479111909866333, "learning_rate": 5.550069997460872e-07, "loss": 0.899, "step": 4659 }, { "epoch": 0.8968437259430331, "grad_norm": 1.6265314817428589, "learning_rate": 5.529611058582951e-07, "loss": 0.9333, "step": 4660 }, { "epoch": 0.897036181678214, "grad_norm": 1.986045479774475, "learning_rate": 5.509188825476964e-07, "loss": 0.9123, "step": 4661 }, { "epoch": 0.8972286374133949, "grad_norm": 1.809822678565979, "learning_rate": 5.488803306077806e-07, "loss": 0.9118, "step": 4662 }, { "epoch": 0.8974210931485759, "grad_norm": 1.7588878870010376, "learning_rate": 5.468454508306131e-07, "loss": 0.9169, "step": 4663 }, { "epoch": 0.8976135488837568, "grad_norm": 1.7753154039382935, "learning_rate": 5.448142440068316e-07, "loss": 0.9448, "step": 4664 }, { "epoch": 0.8978060046189377, "grad_norm": 0.8628824353218079, "learning_rate": 5.427867109256457e-07, "loss": 0.8604, "step": 4665 }, { "epoch": 0.8979984603541186, "grad_norm": 1.7549244165420532, "learning_rate": 5.407628523748398e-07, "loss": 0.8944, "step": 4666 }, { "epoch": 0.8981909160892995, "grad_norm": 1.7723881006240845, "learning_rate": 5.387426691407682e-07, "loss": 0.865, "step": 4667 }, { "epoch": 0.8983833718244804, "grad_norm": 1.8569858074188232, "learning_rate": 5.367261620083575e-07, "loss": 0.8609, "step": 4668 }, { "epoch": 0.8985758275596613, "grad_norm": 1.6620985269546509, "learning_rate": 5.347133317611075e-07, "loss": 0.8505, "step": 4669 }, { "epoch": 0.8987682832948422, "grad_norm": 1.8167173862457275, "learning_rate": 5.327041791810905e-07, "loss": 0.8625, "step": 4670 }, { "epoch": 0.898960739030023, "grad_norm": 1.7594622373580933, "learning_rate": 5.306987050489442e-07, "loss": 0.8987, "step": 4671 }, { "epoch": 0.8991531947652041, "grad_norm": 1.8659120798110962, "learning_rate": 5.286969101438821e-07, "loss": 0.9248, "step": 4672 }, { "epoch": 0.899345650500385, "grad_norm": 1.7423425912857056, "learning_rate": 5.266987952436875e-07, "loss": 0.8815, "step": 4673 }, { "epoch": 0.8995381062355658, "grad_norm": 1.7986290454864502, "learning_rate": 5.247043611247127e-07, "loss": 0.9493, "step": 4674 }, { "epoch": 0.8997305619707467, "grad_norm": 2.055258274078369, "learning_rate": 5.227136085618801e-07, "loss": 0.9573, "step": 4675 }, { "epoch": 0.8999230177059276, "grad_norm": 1.7386811971664429, "learning_rate": 5.207265383286831e-07, "loss": 0.8827, "step": 4676 }, { "epoch": 0.9001154734411085, "grad_norm": 1.7026543617248535, "learning_rate": 5.187431511971819e-07, "loss": 0.8693, "step": 4677 }, { "epoch": 0.9003079291762894, "grad_norm": 1.9607104063034058, "learning_rate": 5.167634479380068e-07, "loss": 0.8117, "step": 4678 }, { "epoch": 0.9005003849114703, "grad_norm": 1.7401719093322754, "learning_rate": 5.147874293203581e-07, "loss": 0.9175, "step": 4679 }, { "epoch": 0.9006928406466512, "grad_norm": 1.5865559577941895, "learning_rate": 5.128150961120026e-07, "loss": 0.8455, "step": 4680 }, { "epoch": 0.9008852963818321, "grad_norm": 1.6492749452590942, "learning_rate": 5.108464490792753e-07, "loss": 0.8708, "step": 4681 }, { "epoch": 0.9010777521170131, "grad_norm": 1.856960415840149, "learning_rate": 5.088814889870808e-07, "loss": 0.8487, "step": 4682 }, { "epoch": 0.901270207852194, "grad_norm": 1.7916319370269775, "learning_rate": 5.069202165988874e-07, "loss": 0.8879, "step": 4683 }, { "epoch": 0.9014626635873749, "grad_norm": 1.7260135412216187, "learning_rate": 5.049626326767366e-07, "loss": 0.9788, "step": 4684 }, { "epoch": 0.9016551193225558, "grad_norm": 1.660810947418213, "learning_rate": 5.030087379812299e-07, "loss": 0.9616, "step": 4685 }, { "epoch": 0.9018475750577367, "grad_norm": 1.7168588638305664, "learning_rate": 5.010585332715401e-07, "loss": 0.9084, "step": 4686 }, { "epoch": 0.9020400307929176, "grad_norm": 1.7697783708572388, "learning_rate": 4.991120193054056e-07, "loss": 0.9523, "step": 4687 }, { "epoch": 0.9022324865280985, "grad_norm": 1.956488847732544, "learning_rate": 4.97169196839129e-07, "loss": 0.904, "step": 4688 }, { "epoch": 0.9024249422632794, "grad_norm": 1.6730067729949951, "learning_rate": 4.952300666275789e-07, "loss": 0.9192, "step": 4689 }, { "epoch": 0.9026173979984603, "grad_norm": 1.7866907119750977, "learning_rate": 4.932946294241903e-07, "loss": 0.8938, "step": 4690 }, { "epoch": 0.9028098537336413, "grad_norm": 2.0164899826049805, "learning_rate": 4.913628859809638e-07, "loss": 0.9791, "step": 4691 }, { "epoch": 0.9030023094688222, "grad_norm": 3.0027291774749756, "learning_rate": 4.894348370484648e-07, "loss": 0.9902, "step": 4692 }, { "epoch": 0.9031947652040031, "grad_norm": 1.6987488269805908, "learning_rate": 4.875104833758204e-07, "loss": 0.8784, "step": 4693 }, { "epoch": 0.903387220939184, "grad_norm": 1.764320731163025, "learning_rate": 4.855898257107261e-07, "loss": 0.9944, "step": 4694 }, { "epoch": 0.9035796766743649, "grad_norm": 1.8861677646636963, "learning_rate": 4.836728647994371e-07, "loss": 1.0265, "step": 4695 }, { "epoch": 0.9037721324095458, "grad_norm": 0.8907985091209412, "learning_rate": 4.817596013867765e-07, "loss": 0.8962, "step": 4696 }, { "epoch": 0.9039645881447267, "grad_norm": 1.7811397314071655, "learning_rate": 4.798500362161273e-07, "loss": 0.9199, "step": 4697 }, { "epoch": 0.9041570438799076, "grad_norm": 1.6720751523971558, "learning_rate": 4.779441700294362e-07, "loss": 0.9478, "step": 4698 }, { "epoch": 0.9043494996150885, "grad_norm": 1.8561614751815796, "learning_rate": 4.7604200356721644e-07, "loss": 0.9882, "step": 4699 }, { "epoch": 0.9045419553502695, "grad_norm": 0.8970755934715271, "learning_rate": 4.7414353756853773e-07, "loss": 0.8724, "step": 4700 }, { "epoch": 0.9047344110854504, "grad_norm": 1.7764832973480225, "learning_rate": 4.7224877277103673e-07, "loss": 0.907, "step": 4701 }, { "epoch": 0.9049268668206313, "grad_norm": 1.8301312923431396, "learning_rate": 4.703577099109091e-07, "loss": 0.9104, "step": 4702 }, { "epoch": 0.9051193225558122, "grad_norm": 1.6448180675506592, "learning_rate": 4.6847034972291595e-07, "loss": 0.8552, "step": 4703 }, { "epoch": 0.9053117782909931, "grad_norm": 1.6822906732559204, "learning_rate": 4.6658669294037393e-07, "loss": 0.8198, "step": 4704 }, { "epoch": 0.905504234026174, "grad_norm": 1.585579752922058, "learning_rate": 4.647067402951655e-07, "loss": 0.9267, "step": 4705 }, { "epoch": 0.9056966897613549, "grad_norm": 1.765357255935669, "learning_rate": 4.628304925177318e-07, "loss": 0.9761, "step": 4706 }, { "epoch": 0.9058891454965358, "grad_norm": 1.7246525287628174, "learning_rate": 4.609579503370754e-07, "loss": 0.942, "step": 4707 }, { "epoch": 0.9060816012317167, "grad_norm": 0.9045140147209167, "learning_rate": 4.5908911448075746e-07, "loss": 0.858, "step": 4708 }, { "epoch": 0.9062740569668977, "grad_norm": 1.7995293140411377, "learning_rate": 4.572239856749028e-07, "loss": 0.8554, "step": 4709 }, { "epoch": 0.9064665127020786, "grad_norm": 1.5997288227081299, "learning_rate": 4.553625646441928e-07, "loss": 0.9198, "step": 4710 }, { "epoch": 0.9066589684372595, "grad_norm": 1.780782699584961, "learning_rate": 4.535048521118668e-07, "loss": 0.9032, "step": 4711 }, { "epoch": 0.9068514241724404, "grad_norm": 1.7443914413452148, "learning_rate": 4.5165084879972844e-07, "loss": 0.9124, "step": 4712 }, { "epoch": 0.9070438799076213, "grad_norm": 1.697629451751709, "learning_rate": 4.498005554281337e-07, "loss": 0.9287, "step": 4713 }, { "epoch": 0.9072363356428021, "grad_norm": 1.9009188413619995, "learning_rate": 4.4795397271600314e-07, "loss": 0.9062, "step": 4714 }, { "epoch": 0.907428791377983, "grad_norm": 1.6660504341125488, "learning_rate": 4.461111013808117e-07, "loss": 0.8864, "step": 4715 }, { "epoch": 0.9076212471131639, "grad_norm": 1.9300801753997803, "learning_rate": 4.4427194213859216e-07, "loss": 0.938, "step": 4716 }, { "epoch": 0.9078137028483448, "grad_norm": 1.6352487802505493, "learning_rate": 4.424364957039395e-07, "loss": 0.9074, "step": 4717 }, { "epoch": 0.9080061585835257, "grad_norm": 1.9206503629684448, "learning_rate": 4.4060476278999897e-07, "loss": 0.9059, "step": 4718 }, { "epoch": 0.9081986143187067, "grad_norm": 1.6867069005966187, "learning_rate": 4.3877674410848005e-07, "loss": 0.9151, "step": 4719 }, { "epoch": 0.9083910700538876, "grad_norm": 1.7568961381912231, "learning_rate": 4.3695244036964567e-07, "loss": 0.9424, "step": 4720 }, { "epoch": 0.9085835257890685, "grad_norm": 1.8440759181976318, "learning_rate": 4.351318522823134e-07, "loss": 0.9092, "step": 4721 }, { "epoch": 0.9087759815242494, "grad_norm": 1.7675895690917969, "learning_rate": 4.3331498055386057e-07, "loss": 0.8817, "step": 4722 }, { "epoch": 0.9089684372594303, "grad_norm": 1.999017596244812, "learning_rate": 4.3150182589021926e-07, "loss": 0.7584, "step": 4723 }, { "epoch": 0.9091608929946112, "grad_norm": 1.6449241638183594, "learning_rate": 4.296923889958771e-07, "loss": 0.944, "step": 4724 }, { "epoch": 0.9093533487297921, "grad_norm": 1.7319000959396362, "learning_rate": 4.278866705738782e-07, "loss": 0.8767, "step": 4725 }, { "epoch": 0.909545804464973, "grad_norm": 1.636411428451538, "learning_rate": 4.2608467132581934e-07, "loss": 0.9122, "step": 4726 }, { "epoch": 0.9097382602001539, "grad_norm": 1.721819281578064, "learning_rate": 4.2428639195185585e-07, "loss": 0.8436, "step": 4727 }, { "epoch": 0.9099307159353349, "grad_norm": 1.71918523311615, "learning_rate": 4.224918331506955e-07, "loss": 0.8742, "step": 4728 }, { "epoch": 0.9101231716705158, "grad_norm": 1.6895725727081299, "learning_rate": 4.2070099561960064e-07, "loss": 0.8101, "step": 4729 }, { "epoch": 0.9103156274056967, "grad_norm": 1.7954058647155762, "learning_rate": 4.189138800543868e-07, "loss": 0.9361, "step": 4730 }, { "epoch": 0.9105080831408776, "grad_norm": 1.7520251274108887, "learning_rate": 4.171304871494264e-07, "loss": 0.8568, "step": 4731 }, { "epoch": 0.9107005388760585, "grad_norm": 1.735666275024414, "learning_rate": 4.1535081759764286e-07, "loss": 0.8498, "step": 4732 }, { "epoch": 0.9108929946112394, "grad_norm": 1.7762409448623657, "learning_rate": 4.1357487209051416e-07, "loss": 0.9078, "step": 4733 }, { "epoch": 0.9110854503464203, "grad_norm": 1.682921290397644, "learning_rate": 4.118026513180695e-07, "loss": 0.8947, "step": 4734 }, { "epoch": 0.9112779060816012, "grad_norm": 0.9268428087234497, "learning_rate": 4.1003415596889363e-07, "loss": 0.9081, "step": 4735 }, { "epoch": 0.9114703618167821, "grad_norm": 2.248276472091675, "learning_rate": 4.082693867301224e-07, "loss": 0.9119, "step": 4736 }, { "epoch": 0.9116628175519631, "grad_norm": 1.8063682317733765, "learning_rate": 4.065083442874418e-07, "loss": 0.9767, "step": 4737 }, { "epoch": 0.911855273287144, "grad_norm": 1.954755425453186, "learning_rate": 4.0475102932509336e-07, "loss": 0.9265, "step": 4738 }, { "epoch": 0.9120477290223249, "grad_norm": 1.611634373664856, "learning_rate": 4.029974425258687e-07, "loss": 0.8541, "step": 4739 }, { "epoch": 0.9122401847575058, "grad_norm": 2.083188772201538, "learning_rate": 4.012475845711106e-07, "loss": 0.9195, "step": 4740 }, { "epoch": 0.9124326404926867, "grad_norm": 1.8747501373291016, "learning_rate": 3.99501456140714e-07, "loss": 0.9607, "step": 4741 }, { "epoch": 0.9126250962278676, "grad_norm": 1.7273318767547607, "learning_rate": 3.9775905791312407e-07, "loss": 0.9504, "step": 4742 }, { "epoch": 0.9128175519630485, "grad_norm": 1.7551336288452148, "learning_rate": 3.96020390565337e-07, "loss": 0.9061, "step": 4743 }, { "epoch": 0.9130100076982294, "grad_norm": 2.0117383003234863, "learning_rate": 3.9428545477289913e-07, "loss": 0.8519, "step": 4744 }, { "epoch": 0.9132024634334103, "grad_norm": 1.9607936143875122, "learning_rate": 3.9255425120990564e-07, "loss": 0.9258, "step": 4745 }, { "epoch": 0.9133949191685913, "grad_norm": 1.6973448991775513, "learning_rate": 3.908267805490051e-07, "loss": 0.8859, "step": 4746 }, { "epoch": 0.9135873749037722, "grad_norm": 1.661086916923523, "learning_rate": 3.8910304346139185e-07, "loss": 0.932, "step": 4747 }, { "epoch": 0.9137798306389531, "grad_norm": 1.803715467453003, "learning_rate": 3.8738304061681107e-07, "loss": 0.9402, "step": 4748 }, { "epoch": 0.913972286374134, "grad_norm": 1.7385083436965942, "learning_rate": 3.8566677268355943e-07, "loss": 0.9396, "step": 4749 }, { "epoch": 0.9141647421093149, "grad_norm": 1.8413794040679932, "learning_rate": 3.8395424032847686e-07, "loss": 0.8949, "step": 4750 }, { "epoch": 0.9143571978444958, "grad_norm": 1.7624759674072266, "learning_rate": 3.8224544421695766e-07, "loss": 0.8845, "step": 4751 }, { "epoch": 0.9145496535796767, "grad_norm": 1.773362398147583, "learning_rate": 3.8054038501294077e-07, "loss": 0.9811, "step": 4752 }, { "epoch": 0.9147421093148576, "grad_norm": 1.960214376449585, "learning_rate": 3.788390633789163e-07, "loss": 1.0414, "step": 4753 }, { "epoch": 0.9149345650500385, "grad_norm": 1.7714349031448364, "learning_rate": 3.7714147997591655e-07, "loss": 0.9155, "step": 4754 }, { "epoch": 0.9151270207852193, "grad_norm": 1.9158926010131836, "learning_rate": 3.7544763546352834e-07, "loss": 0.9664, "step": 4755 }, { "epoch": 0.9153194765204004, "grad_norm": 1.8404676914215088, "learning_rate": 3.7375753049987974e-07, "loss": 0.8579, "step": 4756 }, { "epoch": 0.9155119322555813, "grad_norm": 1.6016980409622192, "learning_rate": 3.7207116574165094e-07, "loss": 0.8735, "step": 4757 }, { "epoch": 0.9157043879907621, "grad_norm": 1.7089896202087402, "learning_rate": 3.703885418440656e-07, "loss": 0.8861, "step": 4758 }, { "epoch": 0.915896843725943, "grad_norm": 1.720229148864746, "learning_rate": 3.687096594608941e-07, "loss": 0.905, "step": 4759 }, { "epoch": 0.9160892994611239, "grad_norm": 1.8409990072250366, "learning_rate": 3.6703451924445467e-07, "loss": 0.9178, "step": 4760 }, { "epoch": 0.9162817551963048, "grad_norm": 1.7311615943908691, "learning_rate": 3.6536312184560996e-07, "loss": 0.8821, "step": 4761 }, { "epoch": 0.9164742109314857, "grad_norm": 1.8107564449310303, "learning_rate": 3.6369546791377054e-07, "loss": 0.8656, "step": 4762 }, { "epoch": 0.9166666666666666, "grad_norm": 1.8011131286621094, "learning_rate": 3.620315580968892e-07, "loss": 0.9416, "step": 4763 }, { "epoch": 0.9168591224018475, "grad_norm": 1.6764616966247559, "learning_rate": 3.603713930414676e-07, "loss": 0.8679, "step": 4764 }, { "epoch": 0.9170515781370285, "grad_norm": 0.978177547454834, "learning_rate": 3.587149733925488e-07, "loss": 0.9183, "step": 4765 }, { "epoch": 0.9172440338722094, "grad_norm": 1.8134093284606934, "learning_rate": 3.570622997937234e-07, "loss": 0.9553, "step": 4766 }, { "epoch": 0.9174364896073903, "grad_norm": 1.7197810411453247, "learning_rate": 3.5541337288712564e-07, "loss": 0.9439, "step": 4767 }, { "epoch": 0.9176289453425712, "grad_norm": 1.727913498878479, "learning_rate": 3.5376819331343404e-07, "loss": 0.8802, "step": 4768 }, { "epoch": 0.9178214010777521, "grad_norm": 2.5710127353668213, "learning_rate": 3.5212676171187065e-07, "loss": 0.8852, "step": 4769 }, { "epoch": 0.918013856812933, "grad_norm": 1.669385552406311, "learning_rate": 3.50489078720202e-07, "loss": 0.9426, "step": 4770 }, { "epoch": 0.9182063125481139, "grad_norm": 1.6277227401733398, "learning_rate": 3.4885514497473574e-07, "loss": 0.9061, "step": 4771 }, { "epoch": 0.9183987682832948, "grad_norm": 3.0289626121520996, "learning_rate": 3.472249611103273e-07, "loss": 0.9025, "step": 4772 }, { "epoch": 0.9185912240184757, "grad_norm": 1.766283392906189, "learning_rate": 3.455985277603713e-07, "loss": 0.8961, "step": 4773 }, { "epoch": 0.9187836797536567, "grad_norm": 1.7640736103057861, "learning_rate": 3.4397584555680653e-07, "loss": 0.8814, "step": 4774 }, { "epoch": 0.9189761354888376, "grad_norm": 1.674479365348816, "learning_rate": 3.4235691513011425e-07, "loss": 0.8357, "step": 4775 }, { "epoch": 0.9191685912240185, "grad_norm": 0.9121842384338379, "learning_rate": 3.4074173710931804e-07, "loss": 0.9338, "step": 4776 }, { "epoch": 0.9193610469591994, "grad_norm": 1.7225884199142456, "learning_rate": 3.3913031212198356e-07, "loss": 0.8454, "step": 4777 }, { "epoch": 0.9195535026943803, "grad_norm": 1.7789677381515503, "learning_rate": 3.375226407942178e-07, "loss": 0.9444, "step": 4778 }, { "epoch": 0.9197459584295612, "grad_norm": 1.629672646522522, "learning_rate": 3.359187237506689e-07, "loss": 0.9745, "step": 4779 }, { "epoch": 0.9199384141647421, "grad_norm": 1.8693183660507202, "learning_rate": 3.3431856161452835e-07, "loss": 0.8802, "step": 4780 }, { "epoch": 0.920130869899923, "grad_norm": 1.7442121505737305, "learning_rate": 3.327221550075266e-07, "loss": 0.88, "step": 4781 }, { "epoch": 0.9203233256351039, "grad_norm": 1.8631259202957153, "learning_rate": 3.3112950454993633e-07, "loss": 0.8031, "step": 4782 }, { "epoch": 0.9205157813702848, "grad_norm": 1.817678689956665, "learning_rate": 3.2954061086056924e-07, "loss": 0.9087, "step": 4783 }, { "epoch": 0.9207082371054658, "grad_norm": 4.617339611053467, "learning_rate": 3.2795547455677813e-07, "loss": 0.9518, "step": 4784 }, { "epoch": 0.9209006928406467, "grad_norm": 1.8480446338653564, "learning_rate": 3.26374096254457e-07, "loss": 0.8798, "step": 4785 }, { "epoch": 0.9210931485758276, "grad_norm": 1.780389428138733, "learning_rate": 3.247964765680389e-07, "loss": 0.8393, "step": 4786 }, { "epoch": 0.9212856043110085, "grad_norm": 1.7349390983581543, "learning_rate": 3.2322261611049457e-07, "loss": 0.9066, "step": 4787 }, { "epoch": 0.9214780600461894, "grad_norm": 1.7600276470184326, "learning_rate": 3.2165251549333585e-07, "loss": 0.9196, "step": 4788 }, { "epoch": 0.9216705157813703, "grad_norm": 2.061828851699829, "learning_rate": 3.20086175326616e-07, "loss": 0.8915, "step": 4789 }, { "epoch": 0.9218629715165512, "grad_norm": 1.6906039714813232, "learning_rate": 3.185235962189237e-07, "loss": 0.9053, "step": 4790 }, { "epoch": 0.9220554272517321, "grad_norm": 1.89994215965271, "learning_rate": 3.1696477877738664e-07, "loss": 0.9184, "step": 4791 }, { "epoch": 0.922247882986913, "grad_norm": 0.9526238441467285, "learning_rate": 3.1540972360767254e-07, "loss": 0.8956, "step": 4792 }, { "epoch": 0.922440338722094, "grad_norm": 1.7391424179077148, "learning_rate": 3.1385843131398697e-07, "loss": 0.932, "step": 4793 }, { "epoch": 0.9226327944572749, "grad_norm": 1.61171293258667, "learning_rate": 3.1231090249907336e-07, "loss": 0.8843, "step": 4794 }, { "epoch": 0.9228252501924558, "grad_norm": 0.896507740020752, "learning_rate": 3.107671377642119e-07, "loss": 0.9201, "step": 4795 }, { "epoch": 0.9230177059276367, "grad_norm": 1.8281115293502808, "learning_rate": 3.0922713770922155e-07, "loss": 0.9763, "step": 4796 }, { "epoch": 0.9232101616628176, "grad_norm": 1.64071524143219, "learning_rate": 3.076909029324571e-07, "loss": 0.9072, "step": 4797 }, { "epoch": 0.9234026173979984, "grad_norm": 1.7085450887680054, "learning_rate": 3.06158434030811e-07, "loss": 0.9643, "step": 4798 }, { "epoch": 0.9235950731331793, "grad_norm": 0.8372834920883179, "learning_rate": 3.0462973159971375e-07, "loss": 0.855, "step": 4799 }, { "epoch": 0.9237875288683602, "grad_norm": 1.7778486013412476, "learning_rate": 3.0310479623313125e-07, "loss": 0.8612, "step": 4800 }, { "epoch": 0.9239799846035411, "grad_norm": 1.6668615341186523, "learning_rate": 3.0158362852356627e-07, "loss": 0.8685, "step": 4801 }, { "epoch": 0.9241724403387221, "grad_norm": 1.7300249338150024, "learning_rate": 3.0006622906205707e-07, "loss": 0.9033, "step": 4802 }, { "epoch": 0.924364896073903, "grad_norm": 1.9477571249008179, "learning_rate": 2.985525984381754e-07, "loss": 0.8803, "step": 4803 }, { "epoch": 0.9245573518090839, "grad_norm": 1.9010552167892456, "learning_rate": 2.970427372400353e-07, "loss": 0.9389, "step": 4804 }, { "epoch": 0.9247498075442648, "grad_norm": 1.6804304122924805, "learning_rate": 2.955366460542797e-07, "loss": 0.8489, "step": 4805 }, { "epoch": 0.9249422632794457, "grad_norm": 1.746863603591919, "learning_rate": 2.940343254660905e-07, "loss": 0.9514, "step": 4806 }, { "epoch": 0.9251347190146266, "grad_norm": 1.7444335222244263, "learning_rate": 2.925357760591818e-07, "loss": 0.9569, "step": 4807 }, { "epoch": 0.9253271747498075, "grad_norm": 1.6852445602416992, "learning_rate": 2.910409984158058e-07, "loss": 0.918, "step": 4808 }, { "epoch": 0.9255196304849884, "grad_norm": 1.746703028678894, "learning_rate": 2.8954999311674557e-07, "loss": 0.9549, "step": 4809 }, { "epoch": 0.9257120862201693, "grad_norm": 1.695944905281067, "learning_rate": 2.880627607413211e-07, "loss": 0.8868, "step": 4810 }, { "epoch": 0.9259045419553503, "grad_norm": 1.7166118621826172, "learning_rate": 2.865793018673857e-07, "loss": 0.8989, "step": 4811 }, { "epoch": 0.9260969976905312, "grad_norm": 1.7035415172576904, "learning_rate": 2.8509961707132496e-07, "loss": 0.9602, "step": 4812 }, { "epoch": 0.9262894534257121, "grad_norm": 1.8007614612579346, "learning_rate": 2.8362370692806117e-07, "loss": 0.921, "step": 4813 }, { "epoch": 0.926481909160893, "grad_norm": 1.7436736822128296, "learning_rate": 2.821515720110479e-07, "loss": 1.0102, "step": 4814 }, { "epoch": 0.9266743648960739, "grad_norm": 1.7669144868850708, "learning_rate": 2.806832128922732e-07, "loss": 0.964, "step": 4815 }, { "epoch": 0.9268668206312548, "grad_norm": 1.6212629079818726, "learning_rate": 2.7921863014225504e-07, "loss": 0.9104, "step": 4816 }, { "epoch": 0.9270592763664357, "grad_norm": 1.840636968612671, "learning_rate": 2.7775782433004714e-07, "loss": 0.9321, "step": 4817 }, { "epoch": 0.9272517321016166, "grad_norm": 1.9070498943328857, "learning_rate": 2.7630079602323447e-07, "loss": 0.8954, "step": 4818 }, { "epoch": 0.9274441878367975, "grad_norm": 1.6181663274765015, "learning_rate": 2.7484754578793513e-07, "loss": 0.9066, "step": 4819 }, { "epoch": 0.9276366435719784, "grad_norm": 1.750315546989441, "learning_rate": 2.733980741887987e-07, "loss": 0.8834, "step": 4820 }, { "epoch": 0.9278290993071594, "grad_norm": 1.7466723918914795, "learning_rate": 2.7195238178900685e-07, "loss": 0.84, "step": 4821 }, { "epoch": 0.9280215550423403, "grad_norm": 1.867712140083313, "learning_rate": 2.7051046915027244e-07, "loss": 0.9528, "step": 4822 }, { "epoch": 0.9282140107775212, "grad_norm": 1.7667862176895142, "learning_rate": 2.690723368328407e-07, "loss": 0.9354, "step": 4823 }, { "epoch": 0.9284064665127021, "grad_norm": 1.8769786357879639, "learning_rate": 2.676379853954858e-07, "loss": 0.9211, "step": 4824 }, { "epoch": 0.928598922247883, "grad_norm": 2.0662314891815186, "learning_rate": 2.662074153955152e-07, "loss": 0.9026, "step": 4825 }, { "epoch": 0.9287913779830639, "grad_norm": 1.8869236707687378, "learning_rate": 2.6478062738876654e-07, "loss": 0.905, "step": 4826 }, { "epoch": 0.9289838337182448, "grad_norm": 1.676169514656067, "learning_rate": 2.6335762192960743e-07, "loss": 0.9462, "step": 4827 }, { "epoch": 0.9291762894534257, "grad_norm": 1.7572128772735596, "learning_rate": 2.6193839957093683e-07, "loss": 0.8911, "step": 4828 }, { "epoch": 0.9293687451886066, "grad_norm": 1.6695942878723145, "learning_rate": 2.605229608641824e-07, "loss": 0.9207, "step": 4829 }, { "epoch": 0.9295612009237876, "grad_norm": 1.7740347385406494, "learning_rate": 2.5911130635930203e-07, "loss": 0.9339, "step": 4830 }, { "epoch": 0.9297536566589685, "grad_norm": 1.9497209787368774, "learning_rate": 2.577034366047848e-07, "loss": 0.8779, "step": 4831 }, { "epoch": 0.9299461123941494, "grad_norm": 1.641097068786621, "learning_rate": 2.5629935214764866e-07, "loss": 0.8584, "step": 4832 }, { "epoch": 0.9301385681293303, "grad_norm": 1.663203477859497, "learning_rate": 2.5489905353343725e-07, "loss": 0.9112, "step": 4833 }, { "epoch": 0.9303310238645112, "grad_norm": 2.0471978187561035, "learning_rate": 2.5350254130622866e-07, "loss": 0.9307, "step": 4834 }, { "epoch": 0.9305234795996921, "grad_norm": 1.7074921131134033, "learning_rate": 2.5210981600862663e-07, "loss": 0.8738, "step": 4835 }, { "epoch": 0.930715935334873, "grad_norm": 1.8430519104003906, "learning_rate": 2.507208781817638e-07, "loss": 0.9313, "step": 4836 }, { "epoch": 0.9309083910700539, "grad_norm": 1.9072238206863403, "learning_rate": 2.4933572836530307e-07, "loss": 0.8517, "step": 4837 }, { "epoch": 0.9311008468052347, "grad_norm": 0.8866738080978394, "learning_rate": 2.4795436709743384e-07, "loss": 0.8882, "step": 4838 }, { "epoch": 0.9312933025404158, "grad_norm": 1.6284101009368896, "learning_rate": 2.465767949148734e-07, "loss": 0.8945, "step": 4839 }, { "epoch": 0.9314857582755967, "grad_norm": 1.7144591808319092, "learning_rate": 2.45203012352867e-07, "loss": 0.924, "step": 4840 }, { "epoch": 0.9316782140107776, "grad_norm": 1.811737060546875, "learning_rate": 2.4383301994518773e-07, "loss": 0.9275, "step": 4841 }, { "epoch": 0.9318706697459584, "grad_norm": 1.9508341550827026, "learning_rate": 2.424668182241363e-07, "loss": 0.9204, "step": 4842 }, { "epoch": 0.9320631254811393, "grad_norm": 1.7545024156570435, "learning_rate": 2.4110440772054155e-07, "loss": 0.8498, "step": 4843 }, { "epoch": 0.9322555812163202, "grad_norm": 1.7270668745040894, "learning_rate": 2.3974578896375555e-07, "loss": 0.9567, "step": 4844 }, { "epoch": 0.9324480369515011, "grad_norm": 1.9077086448669434, "learning_rate": 2.383909624816616e-07, "loss": 0.9216, "step": 4845 }, { "epoch": 0.932640492686682, "grad_norm": 1.998375654220581, "learning_rate": 2.370399288006664e-07, "loss": 0.824, "step": 4846 }, { "epoch": 0.9328329484218629, "grad_norm": 1.7505697011947632, "learning_rate": 2.3569268844570447e-07, "loss": 0.9499, "step": 4847 }, { "epoch": 0.9330254041570438, "grad_norm": 1.5878850221633911, "learning_rate": 2.3434924194023712e-07, "loss": 0.9487, "step": 4848 }, { "epoch": 0.9332178598922248, "grad_norm": 2.0753414630889893, "learning_rate": 2.3300958980624787e-07, "loss": 0.8266, "step": 4849 }, { "epoch": 0.9334103156274057, "grad_norm": 0.8450593948364258, "learning_rate": 2.3167373256425152e-07, "loss": 0.8478, "step": 4850 }, { "epoch": 0.9336027713625866, "grad_norm": 1.732397198677063, "learning_rate": 2.3034167073328283e-07, "loss": 0.982, "step": 4851 }, { "epoch": 0.9337952270977675, "grad_norm": 1.7365858554840088, "learning_rate": 2.2901340483090785e-07, "loss": 0.8597, "step": 4852 }, { "epoch": 0.9339876828329484, "grad_norm": 1.6468778848648071, "learning_rate": 2.2768893537321145e-07, "loss": 0.8229, "step": 4853 }, { "epoch": 0.9341801385681293, "grad_norm": 1.6674633026123047, "learning_rate": 2.2636826287480872e-07, "loss": 0.8662, "step": 4854 }, { "epoch": 0.9343725943033102, "grad_norm": 1.779270887374878, "learning_rate": 2.2505138784883585e-07, "loss": 0.8584, "step": 4855 }, { "epoch": 0.9345650500384911, "grad_norm": 1.791017770767212, "learning_rate": 2.2373831080695463e-07, "loss": 0.871, "step": 4856 }, { "epoch": 0.934757505773672, "grad_norm": 1.7366632223129272, "learning_rate": 2.2242903225935254e-07, "loss": 0.8347, "step": 4857 }, { "epoch": 0.934949961508853, "grad_norm": 0.9157231450080872, "learning_rate": 2.211235527147404e-07, "loss": 0.9054, "step": 4858 }, { "epoch": 0.9351424172440339, "grad_norm": 1.7156238555908203, "learning_rate": 2.1982187268035138e-07, "loss": 0.8027, "step": 4859 }, { "epoch": 0.9353348729792148, "grad_norm": 1.6320236921310425, "learning_rate": 2.1852399266194312e-07, "loss": 0.8525, "step": 4860 }, { "epoch": 0.9355273287143957, "grad_norm": 1.75459885597229, "learning_rate": 2.1722991316380005e-07, "loss": 0.9613, "step": 4861 }, { "epoch": 0.9357197844495766, "grad_norm": 1.5030946731567383, "learning_rate": 2.1593963468872435e-07, "loss": 0.8778, "step": 4862 }, { "epoch": 0.9359122401847575, "grad_norm": 1.933998465538025, "learning_rate": 2.1465315773804617e-07, "loss": 0.8966, "step": 4863 }, { "epoch": 0.9361046959199384, "grad_norm": 1.7168859243392944, "learning_rate": 2.1337048281161565e-07, "loss": 0.9089, "step": 4864 }, { "epoch": 0.9362971516551193, "grad_norm": 1.7307664155960083, "learning_rate": 2.1209161040780746e-07, "loss": 0.9192, "step": 4865 }, { "epoch": 0.9364896073903002, "grad_norm": 1.5924339294433594, "learning_rate": 2.1081654102351634e-07, "loss": 0.9066, "step": 4866 }, { "epoch": 0.9366820631254812, "grad_norm": 1.8014802932739258, "learning_rate": 2.0954527515416156e-07, "loss": 0.9248, "step": 4867 }, { "epoch": 0.9368745188606621, "grad_norm": 1.7934247255325317, "learning_rate": 2.082778132936858e-07, "loss": 0.9554, "step": 4868 }, { "epoch": 0.937066974595843, "grad_norm": 1.684612512588501, "learning_rate": 2.0701415593455065e-07, "loss": 0.9073, "step": 4869 }, { "epoch": 0.9372594303310239, "grad_norm": 1.6802936792373657, "learning_rate": 2.057543035677412e-07, "loss": 0.8814, "step": 4870 }, { "epoch": 0.9374518860662048, "grad_norm": 1.835364580154419, "learning_rate": 2.0449825668276246e-07, "loss": 0.894, "step": 4871 }, { "epoch": 0.9376443418013857, "grad_norm": 1.716550588607788, "learning_rate": 2.0324601576764525e-07, "loss": 0.9386, "step": 4872 }, { "epoch": 0.9378367975365666, "grad_norm": 1.8702348470687866, "learning_rate": 2.0199758130893478e-07, "loss": 0.9254, "step": 4873 }, { "epoch": 0.9380292532717475, "grad_norm": 1.9168022871017456, "learning_rate": 2.0075295379170413e-07, "loss": 0.9369, "step": 4874 }, { "epoch": 0.9382217090069284, "grad_norm": 1.8598748445510864, "learning_rate": 1.9951213369954204e-07, "loss": 0.9361, "step": 4875 }, { "epoch": 0.9384141647421094, "grad_norm": 1.6831399202346802, "learning_rate": 1.9827512151456175e-07, "loss": 0.9263, "step": 4876 }, { "epoch": 0.9386066204772903, "grad_norm": 1.615675687789917, "learning_rate": 1.9704191771739322e-07, "loss": 0.9464, "step": 4877 }, { "epoch": 0.9387990762124712, "grad_norm": 1.8768787384033203, "learning_rate": 1.95812522787191e-07, "loss": 0.9298, "step": 4878 }, { "epoch": 0.9389915319476521, "grad_norm": 1.781435489654541, "learning_rate": 1.9458693720162513e-07, "loss": 0.9496, "step": 4879 }, { "epoch": 0.939183987682833, "grad_norm": 1.5981781482696533, "learning_rate": 1.933651614368892e-07, "loss": 0.9068, "step": 4880 }, { "epoch": 0.9393764434180139, "grad_norm": 1.7023570537567139, "learning_rate": 1.921471959676957e-07, "loss": 0.8415, "step": 4881 }, { "epoch": 0.9395688991531947, "grad_norm": 1.7648065090179443, "learning_rate": 1.9093304126727497e-07, "loss": 1.0121, "step": 4882 }, { "epoch": 0.9397613548883756, "grad_norm": 1.8315552473068237, "learning_rate": 1.897226978073774e-07, "loss": 0.9405, "step": 4883 }, { "epoch": 0.9399538106235565, "grad_norm": 1.643708348274231, "learning_rate": 1.885161660582746e-07, "loss": 0.8857, "step": 4884 }, { "epoch": 0.9401462663587374, "grad_norm": 1.6086310148239136, "learning_rate": 1.8731344648875382e-07, "loss": 0.8516, "step": 4885 }, { "epoch": 0.9403387220939184, "grad_norm": 1.6000818014144897, "learning_rate": 1.8611453956612346e-07, "loss": 0.9711, "step": 4886 }, { "epoch": 0.9405311778290993, "grad_norm": 1.7069274187088013, "learning_rate": 1.8491944575621091e-07, "loss": 0.9249, "step": 4887 }, { "epoch": 0.9407236335642802, "grad_norm": 0.8661252856254578, "learning_rate": 1.8372816552336025e-07, "loss": 0.8818, "step": 4888 }, { "epoch": 0.9409160892994611, "grad_norm": 1.641238808631897, "learning_rate": 1.8254069933043572e-07, "loss": 0.993, "step": 4889 }, { "epoch": 0.941108545034642, "grad_norm": 1.6961842775344849, "learning_rate": 1.81357047638816e-07, "loss": 0.9762, "step": 4890 }, { "epoch": 0.9413010007698229, "grad_norm": 2.1548640727996826, "learning_rate": 1.8017721090840324e-07, "loss": 0.8638, "step": 4891 }, { "epoch": 0.9414934565050038, "grad_norm": 1.7392739057540894, "learning_rate": 1.7900118959761181e-07, "loss": 1.0076, "step": 4892 }, { "epoch": 0.9416859122401847, "grad_norm": 1.8730323314666748, "learning_rate": 1.7782898416337735e-07, "loss": 1.0342, "step": 4893 }, { "epoch": 0.9418783679753656, "grad_norm": 1.7638781070709229, "learning_rate": 1.7666059506115108e-07, "loss": 0.977, "step": 4894 }, { "epoch": 0.9420708237105466, "grad_norm": 1.772662878036499, "learning_rate": 1.754960227449032e-07, "loss": 0.9134, "step": 4895 }, { "epoch": 0.9422632794457275, "grad_norm": 1.7968339920043945, "learning_rate": 1.7433526766711727e-07, "loss": 0.8769, "step": 4896 }, { "epoch": 0.9424557351809084, "grad_norm": 1.8762168884277344, "learning_rate": 1.7317833027879928e-07, "loss": 0.8677, "step": 4897 }, { "epoch": 0.9426481909160893, "grad_norm": 1.6070098876953125, "learning_rate": 1.7202521102946624e-07, "loss": 0.8822, "step": 4898 }, { "epoch": 0.9428406466512702, "grad_norm": 1.8356244564056396, "learning_rate": 1.7087591036715534e-07, "loss": 0.9053, "step": 4899 }, { "epoch": 0.9430331023864511, "grad_norm": 1.9397823810577393, "learning_rate": 1.6973042873841827e-07, "loss": 0.9389, "step": 4900 }, { "epoch": 0.943225558121632, "grad_norm": 1.8684617280960083, "learning_rate": 1.6858876658832235e-07, "loss": 0.9242, "step": 4901 }, { "epoch": 0.9434180138568129, "grad_norm": 1.7800335884094238, "learning_rate": 1.6745092436045495e-07, "loss": 0.9593, "step": 4902 }, { "epoch": 0.9436104695919938, "grad_norm": 1.5409257411956787, "learning_rate": 1.6631690249691358e-07, "loss": 1.0059, "step": 4903 }, { "epoch": 0.9438029253271748, "grad_norm": 1.845908284187317, "learning_rate": 1.651867014383146e-07, "loss": 0.9788, "step": 4904 }, { "epoch": 0.9439953810623557, "grad_norm": 0.8025622963905334, "learning_rate": 1.6406032162379127e-07, "loss": 0.8094, "step": 4905 }, { "epoch": 0.9441878367975366, "grad_norm": 1.6414554119110107, "learning_rate": 1.629377634909868e-07, "loss": 1.0004, "step": 4906 }, { "epoch": 0.9443802925327175, "grad_norm": 1.8747586011886597, "learning_rate": 1.618190274760656e-07, "loss": 0.8644, "step": 4907 }, { "epoch": 0.9445727482678984, "grad_norm": 1.6366591453552246, "learning_rate": 1.6070411401370335e-07, "loss": 0.8733, "step": 4908 }, { "epoch": 0.9447652040030793, "grad_norm": 1.7805119752883911, "learning_rate": 1.5959302353709128e-07, "loss": 0.9041, "step": 4909 }, { "epoch": 0.9449576597382602, "grad_norm": 1.8747243881225586, "learning_rate": 1.584857564779363e-07, "loss": 0.9506, "step": 4910 }, { "epoch": 0.9451501154734411, "grad_norm": 1.6787868738174438, "learning_rate": 1.5738231326645758e-07, "loss": 0.9043, "step": 4911 }, { "epoch": 0.945342571208622, "grad_norm": 1.7745697498321533, "learning_rate": 1.5628269433139e-07, "loss": 0.8782, "step": 4912 }, { "epoch": 0.9455350269438029, "grad_norm": 1.7373838424682617, "learning_rate": 1.5518690009998282e-07, "loss": 0.8421, "step": 4913 }, { "epoch": 0.9457274826789839, "grad_norm": 1.6470859050750732, "learning_rate": 1.5409493099799888e-07, "loss": 0.9485, "step": 4914 }, { "epoch": 0.9459199384141648, "grad_norm": 1.6539729833602905, "learning_rate": 1.530067874497143e-07, "loss": 0.9438, "step": 4915 }, { "epoch": 0.9461123941493457, "grad_norm": 0.921311616897583, "learning_rate": 1.519224698779198e-07, "loss": 0.8901, "step": 4916 }, { "epoch": 0.9463048498845266, "grad_norm": 1.7064136266708374, "learning_rate": 1.5084197870391836e-07, "loss": 0.8797, "step": 4917 }, { "epoch": 0.9464973056197075, "grad_norm": 1.734123706817627, "learning_rate": 1.497653143475275e-07, "loss": 1.0208, "step": 4918 }, { "epoch": 0.9466897613548884, "grad_norm": 1.6997294425964355, "learning_rate": 1.486924772270759e-07, "loss": 0.9503, "step": 4919 }, { "epoch": 0.9468822170900693, "grad_norm": 2.161675453186035, "learning_rate": 1.4762346775940794e-07, "loss": 0.9684, "step": 4920 }, { "epoch": 0.9470746728252502, "grad_norm": 1.7576509714126587, "learning_rate": 1.465582863598791e-07, "loss": 0.894, "step": 4921 }, { "epoch": 0.947267128560431, "grad_norm": 1.9255119562149048, "learning_rate": 1.4549693344235726e-07, "loss": 0.8733, "step": 4922 }, { "epoch": 0.9474595842956121, "grad_norm": 1.8250627517700195, "learning_rate": 1.444394094192225e-07, "loss": 0.8689, "step": 4923 }, { "epoch": 0.947652040030793, "grad_norm": 1.773016095161438, "learning_rate": 1.4338571470137063e-07, "loss": 0.9645, "step": 4924 }, { "epoch": 0.9478444957659738, "grad_norm": 1.8170216083526611, "learning_rate": 1.4233584969820414e-07, "loss": 0.9009, "step": 4925 }, { "epoch": 0.9480369515011547, "grad_norm": 1.9065470695495605, "learning_rate": 1.4128981481764115e-07, "loss": 0.8958, "step": 4926 }, { "epoch": 0.9482294072363356, "grad_norm": 1.835839867591858, "learning_rate": 1.4024761046611213e-07, "loss": 1.0048, "step": 4927 }, { "epoch": 0.9484218629715165, "grad_norm": 1.8083370923995972, "learning_rate": 1.3920923704855648e-07, "loss": 0.9357, "step": 4928 }, { "epoch": 0.9486143187066974, "grad_norm": 1.7144523859024048, "learning_rate": 1.3817469496842596e-07, "loss": 0.9153, "step": 4929 }, { "epoch": 0.9488067744418783, "grad_norm": 1.6624090671539307, "learning_rate": 1.3714398462768563e-07, "loss": 0.8949, "step": 4930 }, { "epoch": 0.9489992301770592, "grad_norm": 1.5764845609664917, "learning_rate": 1.3611710642681076e-07, "loss": 0.8851, "step": 4931 }, { "epoch": 0.9491916859122402, "grad_norm": 1.6650302410125732, "learning_rate": 1.350940607647866e-07, "loss": 0.8913, "step": 4932 }, { "epoch": 0.9493841416474211, "grad_norm": 1.7398484945297241, "learning_rate": 1.3407484803910854e-07, "loss": 0.9492, "step": 4933 }, { "epoch": 0.949576597382602, "grad_norm": 1.8557682037353516, "learning_rate": 1.3305946864578645e-07, "loss": 0.8264, "step": 4934 }, { "epoch": 0.9497690531177829, "grad_norm": 1.7488837242126465, "learning_rate": 1.320479229793359e-07, "loss": 0.9428, "step": 4935 }, { "epoch": 0.9499615088529638, "grad_norm": 1.8158282041549683, "learning_rate": 1.3104021143278911e-07, "loss": 1.0401, "step": 4936 }, { "epoch": 0.9501539645881447, "grad_norm": 1.5517100095748901, "learning_rate": 1.3003633439768182e-07, "loss": 0.9299, "step": 4937 }, { "epoch": 0.9503464203233256, "grad_norm": 1.7988089323043823, "learning_rate": 1.2903629226406423e-07, "loss": 0.9427, "step": 4938 }, { "epoch": 0.9505388760585065, "grad_norm": 1.7607849836349487, "learning_rate": 1.280400854204944e-07, "loss": 0.9063, "step": 4939 }, { "epoch": 0.9507313317936874, "grad_norm": 1.8951563835144043, "learning_rate": 1.2704771425404382e-07, "loss": 0.8995, "step": 4940 }, { "epoch": 0.9509237875288684, "grad_norm": 2.0246710777282715, "learning_rate": 1.2605917915028743e-07, "loss": 0.9245, "step": 4941 }, { "epoch": 0.9511162432640493, "grad_norm": 1.6407110691070557, "learning_rate": 1.2507448049331571e-07, "loss": 0.9747, "step": 4942 }, { "epoch": 0.9513086989992302, "grad_norm": 0.9085986018180847, "learning_rate": 1.24093618665726e-07, "loss": 0.862, "step": 4943 }, { "epoch": 0.9515011547344111, "grad_norm": 1.6935354471206665, "learning_rate": 1.231165940486234e-07, "loss": 0.9051, "step": 4944 }, { "epoch": 0.951693610469592, "grad_norm": 1.7065576314926147, "learning_rate": 1.221434070216243e-07, "loss": 0.9055, "step": 4945 }, { "epoch": 0.9518860662047729, "grad_norm": 0.9075039625167847, "learning_rate": 1.2117405796285286e-07, "loss": 0.9092, "step": 4946 }, { "epoch": 0.9520785219399538, "grad_norm": 1.8683946132659912, "learning_rate": 1.2020854724894448e-07, "loss": 0.8796, "step": 4947 }, { "epoch": 0.9522709776751347, "grad_norm": 2.3253719806671143, "learning_rate": 1.192468752550402e-07, "loss": 0.9876, "step": 4948 }, { "epoch": 0.9524634334103156, "grad_norm": 1.7958213090896606, "learning_rate": 1.1828904235478888e-07, "loss": 0.9116, "step": 4949 }, { "epoch": 0.9526558891454965, "grad_norm": 1.6983802318572998, "learning_rate": 1.1733504892035064e-07, "loss": 0.8728, "step": 4950 }, { "epoch": 0.9528483448806775, "grad_norm": 1.7242679595947266, "learning_rate": 1.1638489532239339e-07, "loss": 0.9025, "step": 4951 }, { "epoch": 0.9530408006158584, "grad_norm": 1.5514590740203857, "learning_rate": 1.1543858193009183e-07, "loss": 0.9257, "step": 4952 }, { "epoch": 0.9532332563510393, "grad_norm": 1.8176358938217163, "learning_rate": 1.1449610911112741e-07, "loss": 0.9794, "step": 4953 }, { "epoch": 0.9534257120862202, "grad_norm": 1.9546122550964355, "learning_rate": 1.1355747723169386e-07, "loss": 0.9545, "step": 4954 }, { "epoch": 0.9536181678214011, "grad_norm": 1.679056167602539, "learning_rate": 1.1262268665648724e-07, "loss": 0.9105, "step": 4955 }, { "epoch": 0.953810623556582, "grad_norm": 1.7261285781860352, "learning_rate": 1.1169173774871478e-07, "loss": 0.9096, "step": 4956 }, { "epoch": 0.9540030792917629, "grad_norm": 1.77592933177948, "learning_rate": 1.1076463087008938e-07, "loss": 0.9135, "step": 4957 }, { "epoch": 0.9541955350269438, "grad_norm": 1.841078519821167, "learning_rate": 1.0984136638083176e-07, "loss": 0.8952, "step": 4958 }, { "epoch": 0.9543879907621247, "grad_norm": 1.586127758026123, "learning_rate": 1.0892194463966943e-07, "loss": 0.8959, "step": 4959 }, { "epoch": 0.9545804464973057, "grad_norm": 1.582200050354004, "learning_rate": 1.0800636600383662e-07, "loss": 0.9544, "step": 4960 }, { "epoch": 0.9547729022324866, "grad_norm": 0.9542142152786255, "learning_rate": 1.0709463082907545e-07, "loss": 0.9265, "step": 4961 }, { "epoch": 0.9549653579676675, "grad_norm": 1.7990862131118774, "learning_rate": 1.0618673946963365e-07, "loss": 0.9044, "step": 4962 }, { "epoch": 0.9551578137028484, "grad_norm": 1.6617053747177124, "learning_rate": 1.052826922782646e-07, "loss": 0.8602, "step": 4963 }, { "epoch": 0.9553502694380293, "grad_norm": 2.0394582748413086, "learning_rate": 1.0438248960623065e-07, "loss": 0.9381, "step": 4964 }, { "epoch": 0.9555427251732102, "grad_norm": 1.9113985300064087, "learning_rate": 1.0348613180329758e-07, "loss": 0.9223, "step": 4965 }, { "epoch": 0.955735180908391, "grad_norm": 1.6965852975845337, "learning_rate": 1.0259361921774014e-07, "loss": 0.9055, "step": 4966 }, { "epoch": 0.9559276366435719, "grad_norm": 1.6814028024673462, "learning_rate": 1.0170495219633647e-07, "loss": 0.8948, "step": 4967 }, { "epoch": 0.9561200923787528, "grad_norm": 1.6309480667114258, "learning_rate": 1.0082013108437038e-07, "loss": 0.8932, "step": 4968 }, { "epoch": 0.9563125481139338, "grad_norm": 1.8159620761871338, "learning_rate": 9.993915622563578e-08, "loss": 0.9231, "step": 4969 }, { "epoch": 0.9565050038491147, "grad_norm": 1.8377739191055298, "learning_rate": 9.906202796242547e-08, "loss": 1.0031, "step": 4970 }, { "epoch": 0.9566974595842956, "grad_norm": 1.5393259525299072, "learning_rate": 9.818874663554356e-08, "loss": 0.875, "step": 4971 }, { "epoch": 0.9568899153194765, "grad_norm": 1.7872649431228638, "learning_rate": 9.731931258429638e-08, "loss": 0.9425, "step": 4972 }, { "epoch": 0.9570823710546574, "grad_norm": 1.7148319482803345, "learning_rate": 9.645372614649595e-08, "loss": 0.758, "step": 4973 }, { "epoch": 0.9572748267898383, "grad_norm": 1.5794405937194824, "learning_rate": 9.55919876584599e-08, "loss": 0.9475, "step": 4974 }, { "epoch": 0.9574672825250192, "grad_norm": 1.7703711986541748, "learning_rate": 9.473409745500928e-08, "loss": 0.8753, "step": 4975 }, { "epoch": 0.9576597382602001, "grad_norm": 1.6674796342849731, "learning_rate": 9.388005586947191e-08, "loss": 0.8689, "step": 4976 }, { "epoch": 0.957852193995381, "grad_norm": 1.8466814756393433, "learning_rate": 9.302986323368012e-08, "loss": 0.8537, "step": 4977 }, { "epoch": 0.958044649730562, "grad_norm": 1.9694138765335083, "learning_rate": 9.218351987796859e-08, "loss": 0.8179, "step": 4978 }, { "epoch": 0.9582371054657429, "grad_norm": 0.923488438129425, "learning_rate": 9.134102613117757e-08, "loss": 0.8813, "step": 4979 }, { "epoch": 0.9584295612009238, "grad_norm": 1.8414655923843384, "learning_rate": 9.0502382320653e-08, "loss": 0.8971, "step": 4980 }, { "epoch": 0.9586220169361047, "grad_norm": 1.8651880025863647, "learning_rate": 8.966758877224202e-08, "loss": 0.9105, "step": 4981 }, { "epoch": 0.9588144726712856, "grad_norm": 0.8345015048980713, "learning_rate": 8.88366458102996e-08, "loss": 0.8624, "step": 4982 }, { "epoch": 0.9590069284064665, "grad_norm": 1.6514358520507812, "learning_rate": 8.800955375768194e-08, "loss": 0.9038, "step": 4983 }, { "epoch": 0.9591993841416474, "grad_norm": 1.947794795036316, "learning_rate": 8.718631293574753e-08, "loss": 0.9413, "step": 4984 }, { "epoch": 0.9593918398768283, "grad_norm": 1.7708135843276978, "learning_rate": 8.636692366436383e-08, "loss": 0.9269, "step": 4985 }, { "epoch": 0.9595842956120092, "grad_norm": 1.8726534843444824, "learning_rate": 8.555138626189619e-08, "loss": 0.9177, "step": 4986 }, { "epoch": 0.9597767513471901, "grad_norm": 1.6105023622512817, "learning_rate": 8.473970104521556e-08, "loss": 0.9702, "step": 4987 }, { "epoch": 0.9599692070823711, "grad_norm": 1.703658103942871, "learning_rate": 8.393186832969746e-08, "loss": 0.9987, "step": 4988 }, { "epoch": 0.960161662817552, "grad_norm": 1.826187014579773, "learning_rate": 8.31278884292186e-08, "loss": 0.9935, "step": 4989 }, { "epoch": 0.9603541185527329, "grad_norm": 1.819858193397522, "learning_rate": 8.232776165616019e-08, "loss": 0.9923, "step": 4990 }, { "epoch": 0.9605465742879138, "grad_norm": 1.6652424335479736, "learning_rate": 8.153148832140467e-08, "loss": 0.9219, "step": 4991 }, { "epoch": 0.9607390300230947, "grad_norm": 1.5833219289779663, "learning_rate": 8.07390687343379e-08, "loss": 0.9817, "step": 4992 }, { "epoch": 0.9609314857582756, "grad_norm": 1.7400418519973755, "learning_rate": 7.995050320285025e-08, "loss": 0.9481, "step": 4993 }, { "epoch": 0.9611239414934565, "grad_norm": 1.819797396659851, "learning_rate": 7.916579203333107e-08, "loss": 0.9097, "step": 4994 }, { "epoch": 0.9613163972286374, "grad_norm": 1.8352128267288208, "learning_rate": 7.838493553067539e-08, "loss": 0.9142, "step": 4995 }, { "epoch": 0.9615088529638183, "grad_norm": 2.016019105911255, "learning_rate": 7.760793399827937e-08, "loss": 0.7894, "step": 4996 }, { "epoch": 0.9617013086989993, "grad_norm": 1.7040338516235352, "learning_rate": 7.683478773804043e-08, "loss": 0.9654, "step": 4997 }, { "epoch": 0.9618937644341802, "grad_norm": 1.7433111667633057, "learning_rate": 7.606549705035937e-08, "loss": 0.8655, "step": 4998 }, { "epoch": 0.9620862201693611, "grad_norm": 2.044908285140991, "learning_rate": 7.53000622341371e-08, "loss": 0.8357, "step": 4999 }, { "epoch": 0.962278675904542, "grad_norm": 1.8219791650772095, "learning_rate": 7.453848358678018e-08, "loss": 0.9005, "step": 5000 }, { "epoch": 0.9624711316397229, "grad_norm": 1.7857296466827393, "learning_rate": 7.378076140419188e-08, "loss": 0.9849, "step": 5001 }, { "epoch": 0.9626635873749038, "grad_norm": 1.8978886604309082, "learning_rate": 7.302689598078117e-08, "loss": 0.8763, "step": 5002 }, { "epoch": 0.9628560431100847, "grad_norm": 1.6000418663024902, "learning_rate": 7.227688760945595e-08, "loss": 0.8665, "step": 5003 }, { "epoch": 0.9630484988452656, "grad_norm": 1.8286945819854736, "learning_rate": 7.153073658162646e-08, "loss": 0.8812, "step": 5004 }, { "epoch": 0.9632409545804465, "grad_norm": 1.717595100402832, "learning_rate": 7.078844318720523e-08, "loss": 0.9945, "step": 5005 }, { "epoch": 0.9634334103156275, "grad_norm": 1.8635538816452026, "learning_rate": 7.00500077146038e-08, "loss": 0.918, "step": 5006 }, { "epoch": 0.9636258660508084, "grad_norm": 1.6393096446990967, "learning_rate": 6.931543045073708e-08, "loss": 0.9759, "step": 5007 }, { "epoch": 0.9638183217859893, "grad_norm": 1.9613218307495117, "learning_rate": 6.858471168101788e-08, "loss": 0.8798, "step": 5008 }, { "epoch": 0.9640107775211701, "grad_norm": 1.8709243535995483, "learning_rate": 6.785785168936244e-08, "loss": 0.9082, "step": 5009 }, { "epoch": 0.964203233256351, "grad_norm": 1.729085087776184, "learning_rate": 6.713485075818815e-08, "loss": 0.8572, "step": 5010 }, { "epoch": 0.9643956889915319, "grad_norm": 0.9213813543319702, "learning_rate": 6.641570916840923e-08, "loss": 0.8501, "step": 5011 }, { "epoch": 0.9645881447267128, "grad_norm": 1.8068469762802124, "learning_rate": 6.570042719944436e-08, "loss": 0.873, "step": 5012 }, { "epoch": 0.9647806004618937, "grad_norm": 1.8859580755233765, "learning_rate": 6.498900512921013e-08, "loss": 0.8734, "step": 5013 }, { "epoch": 0.9649730561970746, "grad_norm": 1.8662978410720825, "learning_rate": 6.428144323412544e-08, "loss": 0.792, "step": 5014 }, { "epoch": 0.9651655119322555, "grad_norm": 1.7196509838104248, "learning_rate": 6.357774178910703e-08, "loss": 0.8476, "step": 5015 }, { "epoch": 0.9653579676674365, "grad_norm": 1.8772344589233398, "learning_rate": 6.287790106757396e-08, "loss": 0.9007, "step": 5016 }, { "epoch": 0.9655504234026174, "grad_norm": 1.8025811910629272, "learning_rate": 6.21819213414443e-08, "loss": 0.9023, "step": 5017 }, { "epoch": 0.9657428791377983, "grad_norm": 1.7576590776443481, "learning_rate": 6.148980288113504e-08, "loss": 0.9478, "step": 5018 }, { "epoch": 0.9659353348729792, "grad_norm": 2.060868263244629, "learning_rate": 6.080154595556331e-08, "loss": 1.0009, "step": 5019 }, { "epoch": 0.9661277906081601, "grad_norm": 1.8334498405456543, "learning_rate": 6.011715083214742e-08, "loss": 0.8271, "step": 5020 }, { "epoch": 0.966320246343341, "grad_norm": 1.694542407989502, "learning_rate": 5.943661777680354e-08, "loss": 0.9651, "step": 5021 }, { "epoch": 0.9665127020785219, "grad_norm": 1.645240068435669, "learning_rate": 5.875994705394794e-08, "loss": 1.0205, "step": 5022 }, { "epoch": 0.9667051578137028, "grad_norm": 1.8809397220611572, "learning_rate": 5.808713892649587e-08, "loss": 0.9657, "step": 5023 }, { "epoch": 0.9668976135488837, "grad_norm": 1.7664661407470703, "learning_rate": 5.7418193655861545e-08, "loss": 0.8707, "step": 5024 }, { "epoch": 0.9670900692840647, "grad_norm": 1.7392597198486328, "learning_rate": 5.675311150195928e-08, "loss": 0.9022, "step": 5025 }, { "epoch": 0.9672825250192456, "grad_norm": 1.5286766290664673, "learning_rate": 5.609189272320237e-08, "loss": 0.9164, "step": 5026 }, { "epoch": 0.9674749807544265, "grad_norm": 1.80027174949646, "learning_rate": 5.5434537576500856e-08, "loss": 0.8173, "step": 5027 }, { "epoch": 0.9676674364896074, "grad_norm": 1.6077117919921875, "learning_rate": 5.4781046317267103e-08, "loss": 0.9197, "step": 5028 }, { "epoch": 0.9678598922247883, "grad_norm": 1.7430278062820435, "learning_rate": 5.413141919940801e-08, "loss": 0.926, "step": 5029 }, { "epoch": 0.9680523479599692, "grad_norm": 2.166574239730835, "learning_rate": 5.348565647533388e-08, "loss": 0.9717, "step": 5030 }, { "epoch": 0.9682448036951501, "grad_norm": 1.7993981838226318, "learning_rate": 5.284375839594958e-08, "loss": 0.8211, "step": 5031 }, { "epoch": 0.968437259430331, "grad_norm": 1.594534993171692, "learning_rate": 5.220572521066003e-08, "loss": 0.943, "step": 5032 }, { "epoch": 0.9686297151655119, "grad_norm": 1.8174077272415161, "learning_rate": 5.1571557167369165e-08, "loss": 0.9998, "step": 5033 }, { "epoch": 0.9688221709006929, "grad_norm": 1.9759341478347778, "learning_rate": 5.094125451247656e-08, "loss": 0.9003, "step": 5034 }, { "epoch": 0.9690146266358738, "grad_norm": 1.7663795948028564, "learning_rate": 5.031481749088296e-08, "loss": 0.8997, "step": 5035 }, { "epoch": 0.9692070823710547, "grad_norm": 1.8109006881713867, "learning_rate": 4.9692246345985905e-08, "loss": 0.8914, "step": 5036 }, { "epoch": 0.9693995381062356, "grad_norm": 1.6059695482254028, "learning_rate": 4.9073541319680785e-08, "loss": 0.902, "step": 5037 }, { "epoch": 0.9695919938414165, "grad_norm": 1.6451672315597534, "learning_rate": 4.8458702652359744e-08, "loss": 0.9476, "step": 5038 }, { "epoch": 0.9697844495765974, "grad_norm": 1.8206524848937988, "learning_rate": 4.784773058291503e-08, "loss": 0.8941, "step": 5039 }, { "epoch": 0.9699769053117783, "grad_norm": 1.7027432918548584, "learning_rate": 4.7240625348735636e-08, "loss": 0.9673, "step": 5040 }, { "epoch": 0.9701693610469592, "grad_norm": 1.7841529846191406, "learning_rate": 4.663738718570621e-08, "loss": 0.9295, "step": 5041 }, { "epoch": 0.9703618167821401, "grad_norm": 1.5803463459014893, "learning_rate": 4.603801632821148e-08, "loss": 0.9286, "step": 5042 }, { "epoch": 0.9705542725173211, "grad_norm": 1.7207813262939453, "learning_rate": 4.544251300913405e-08, "loss": 0.9058, "step": 5043 }, { "epoch": 0.970746728252502, "grad_norm": 0.9200888276100159, "learning_rate": 4.485087745984884e-08, "loss": 0.9106, "step": 5044 }, { "epoch": 0.9709391839876829, "grad_norm": 1.556007742881775, "learning_rate": 4.4263109910234194e-08, "loss": 0.9459, "step": 5045 }, { "epoch": 0.9711316397228638, "grad_norm": 1.7532650232315063, "learning_rate": 4.367921058866187e-08, "loss": 0.8614, "step": 5046 }, { "epoch": 0.9713240954580447, "grad_norm": 1.857136607170105, "learning_rate": 4.3099179722002614e-08, "loss": 0.8935, "step": 5047 }, { "epoch": 0.9715165511932256, "grad_norm": 1.8851598501205444, "learning_rate": 4.252301753562171e-08, "loss": 0.8426, "step": 5048 }, { "epoch": 0.9717090069284064, "grad_norm": 1.6299688816070557, "learning_rate": 4.195072425338342e-08, "loss": 0.9705, "step": 5049 }, { "epoch": 0.9719014626635873, "grad_norm": 1.7968155145645142, "learning_rate": 4.1382300097647655e-08, "loss": 0.9204, "step": 5050 }, { "epoch": 0.9720939183987682, "grad_norm": 1.7433204650878906, "learning_rate": 4.0817745289272184e-08, "loss": 0.9112, "step": 5051 }, { "epoch": 0.9722863741339491, "grad_norm": 1.8956501483917236, "learning_rate": 4.025706004760932e-08, "loss": 0.8816, "step": 5052 }, { "epoch": 0.9724788298691301, "grad_norm": 1.6582077741622925, "learning_rate": 3.970024459051147e-08, "loss": 0.9273, "step": 5053 }, { "epoch": 0.972671285604311, "grad_norm": 1.8125413656234741, "learning_rate": 3.914729913432336e-08, "loss": 0.9392, "step": 5054 }, { "epoch": 0.9728637413394919, "grad_norm": 1.846343994140625, "learning_rate": 3.859822389388757e-08, "loss": 0.9367, "step": 5055 }, { "epoch": 0.9730561970746728, "grad_norm": 2.26226806640625, "learning_rate": 3.805301908254455e-08, "loss": 0.934, "step": 5056 }, { "epoch": 0.9732486528098537, "grad_norm": 1.7763423919677734, "learning_rate": 3.751168491212931e-08, "loss": 0.9554, "step": 5057 }, { "epoch": 0.9734411085450346, "grad_norm": 1.8089783191680908, "learning_rate": 3.697422159297248e-08, "loss": 0.9019, "step": 5058 }, { "epoch": 0.9736335642802155, "grad_norm": 1.8912146091461182, "learning_rate": 3.6440629333901466e-08, "loss": 0.8845, "step": 5059 }, { "epoch": 0.9738260200153964, "grad_norm": 1.8100903034210205, "learning_rate": 3.591090834224153e-08, "loss": 0.9222, "step": 5060 }, { "epoch": 0.9740184757505773, "grad_norm": 1.6927605867385864, "learning_rate": 3.538505882380916e-08, "loss": 0.9472, "step": 5061 }, { "epoch": 0.9742109314857583, "grad_norm": 1.8636614084243774, "learning_rate": 3.4863080982919794e-08, "loss": 0.8793, "step": 5062 }, { "epoch": 0.9744033872209392, "grad_norm": 1.6745274066925049, "learning_rate": 3.4344975022385654e-08, "loss": 0.9274, "step": 5063 }, { "epoch": 0.9745958429561201, "grad_norm": 0.8924670219421387, "learning_rate": 3.383074114351237e-08, "loss": 0.9494, "step": 5064 }, { "epoch": 0.974788298691301, "grad_norm": 1.8694889545440674, "learning_rate": 3.3320379546101234e-08, "loss": 0.8801, "step": 5065 }, { "epoch": 0.9749807544264819, "grad_norm": 2.0981969833374023, "learning_rate": 3.281389042844918e-08, "loss": 0.8548, "step": 5066 }, { "epoch": 0.9751732101616628, "grad_norm": 1.6610794067382812, "learning_rate": 3.23112739873499e-08, "loss": 0.9901, "step": 5067 }, { "epoch": 0.9753656658968437, "grad_norm": 1.59697425365448, "learning_rate": 3.181253041809052e-08, "loss": 0.8648, "step": 5068 }, { "epoch": 0.9755581216320246, "grad_norm": 1.9724822044372559, "learning_rate": 3.13176599144549e-08, "loss": 1.012, "step": 5069 }, { "epoch": 0.9757505773672055, "grad_norm": 1.643310308456421, "learning_rate": 3.082666266872036e-08, "loss": 0.9334, "step": 5070 }, { "epoch": 0.9759430331023865, "grad_norm": 0.897040069103241, "learning_rate": 3.033953887166097e-08, "loss": 0.8964, "step": 5071 }, { "epoch": 0.9761354888375674, "grad_norm": 0.8986654281616211, "learning_rate": 2.9856288712544204e-08, "loss": 0.8671, "step": 5072 }, { "epoch": 0.9763279445727483, "grad_norm": 1.9486429691314697, "learning_rate": 2.9376912379134314e-08, "loss": 0.8355, "step": 5073 }, { "epoch": 0.9765204003079292, "grad_norm": 1.5579249858856201, "learning_rate": 2.8901410057688982e-08, "loss": 0.8478, "step": 5074 }, { "epoch": 0.9767128560431101, "grad_norm": 1.6764822006225586, "learning_rate": 2.842978193296153e-08, "loss": 0.882, "step": 5075 }, { "epoch": 0.976905311778291, "grad_norm": 1.6065285205841064, "learning_rate": 2.796202818819871e-08, "loss": 0.7981, "step": 5076 }, { "epoch": 0.9770977675134719, "grad_norm": 1.5346144437789917, "learning_rate": 2.7498149005144025e-08, "loss": 0.9323, "step": 5077 }, { "epoch": 0.9772902232486528, "grad_norm": 1.7720375061035156, "learning_rate": 2.7038144564033307e-08, "loss": 0.9077, "step": 5078 }, { "epoch": 0.9774826789838337, "grad_norm": 0.939842939376831, "learning_rate": 2.658201504359803e-08, "loss": 0.9276, "step": 5079 }, { "epoch": 0.9776751347190146, "grad_norm": 2.007723331451416, "learning_rate": 2.6129760621063095e-08, "loss": 0.9324, "step": 5080 }, { "epoch": 0.9778675904541956, "grad_norm": 1.7180511951446533, "learning_rate": 2.5681381472151268e-08, "loss": 0.9325, "step": 5081 }, { "epoch": 0.9780600461893765, "grad_norm": 1.6762229204177856, "learning_rate": 2.5236877771074308e-08, "loss": 0.9496, "step": 5082 }, { "epoch": 0.9782525019245574, "grad_norm": 1.741676688194275, "learning_rate": 2.479624969054295e-08, "loss": 0.9748, "step": 5083 }, { "epoch": 0.9784449576597383, "grad_norm": 1.6695733070373535, "learning_rate": 2.4359497401758026e-08, "loss": 0.9211, "step": 5084 }, { "epoch": 0.9786374133949192, "grad_norm": 1.7717890739440918, "learning_rate": 2.3926621074417123e-08, "loss": 0.9191, "step": 5085 }, { "epoch": 0.9788298691301001, "grad_norm": 1.7004142999649048, "learning_rate": 2.349762087671126e-08, "loss": 0.9368, "step": 5086 }, { "epoch": 0.979022324865281, "grad_norm": 1.7372959852218628, "learning_rate": 2.3072496975325986e-08, "loss": 0.9327, "step": 5087 }, { "epoch": 0.9792147806004619, "grad_norm": 1.595108985900879, "learning_rate": 2.265124953543918e-08, "loss": 0.8907, "step": 5088 }, { "epoch": 0.9794072363356428, "grad_norm": 0.8623106479644775, "learning_rate": 2.2233878720723245e-08, "loss": 0.8629, "step": 5089 }, { "epoch": 0.9795996920708238, "grad_norm": 1.7444875240325928, "learning_rate": 2.182038469334513e-08, "loss": 0.938, "step": 5090 }, { "epoch": 0.9797921478060047, "grad_norm": 1.9056637287139893, "learning_rate": 2.1410767613965212e-08, "loss": 0.9541, "step": 5091 }, { "epoch": 0.9799846035411856, "grad_norm": 1.6589503288269043, "learning_rate": 2.1005027641736176e-08, "loss": 0.8599, "step": 5092 }, { "epoch": 0.9801770592763664, "grad_norm": 1.8093677759170532, "learning_rate": 2.0603164934306364e-08, "loss": 0.9083, "step": 5093 }, { "epoch": 0.9803695150115473, "grad_norm": 1.655653476715088, "learning_rate": 2.020517964781532e-08, "loss": 0.9028, "step": 5094 }, { "epoch": 0.9805619707467282, "grad_norm": 2.037905216217041, "learning_rate": 1.9811071936897132e-08, "loss": 0.9751, "step": 5095 }, { "epoch": 0.9807544264819091, "grad_norm": 1.7231525182724, "learning_rate": 1.9420841954681525e-08, "loss": 0.8841, "step": 5096 }, { "epoch": 0.98094688221709, "grad_norm": 1.6695948839187622, "learning_rate": 1.903448985278722e-08, "loss": 1.0628, "step": 5097 }, { "epoch": 0.9811393379522709, "grad_norm": 1.8612843751907349, "learning_rate": 1.86520157813308e-08, "loss": 0.9996, "step": 5098 }, { "epoch": 0.9813317936874519, "grad_norm": 1.8157663345336914, "learning_rate": 1.8273419888916733e-08, "loss": 0.9909, "step": 5099 }, { "epoch": 0.9815242494226328, "grad_norm": 1.798865795135498, "learning_rate": 1.7898702322648453e-08, "loss": 0.8884, "step": 5100 }, { "epoch": 0.9817167051578137, "grad_norm": 1.6262913942337036, "learning_rate": 1.7527863228118393e-08, "loss": 0.8249, "step": 5101 }, { "epoch": 0.9819091608929946, "grad_norm": 1.842873454093933, "learning_rate": 1.716090274941351e-08, "loss": 0.8963, "step": 5102 }, { "epoch": 0.9821016166281755, "grad_norm": 1.8302541971206665, "learning_rate": 1.6797821029113093e-08, "loss": 0.8272, "step": 5103 }, { "epoch": 0.9822940723633564, "grad_norm": 1.6816071271896362, "learning_rate": 1.6438618208290957e-08, "loss": 0.8802, "step": 5104 }, { "epoch": 0.9824865280985373, "grad_norm": 1.722398281097412, "learning_rate": 1.608329442651213e-08, "loss": 0.8256, "step": 5105 }, { "epoch": 0.9826789838337182, "grad_norm": 1.8138577938079834, "learning_rate": 1.5731849821833955e-08, "loss": 0.9196, "step": 5106 }, { "epoch": 0.9828714395688991, "grad_norm": 1.8999924659729004, "learning_rate": 1.5384284530810533e-08, "loss": 0.9164, "step": 5107 }, { "epoch": 0.9830638953040801, "grad_norm": 1.6980832815170288, "learning_rate": 1.5040598688482732e-08, "loss": 0.8507, "step": 5108 }, { "epoch": 0.983256351039261, "grad_norm": 1.6590511798858643, "learning_rate": 1.4700792428388177e-08, "loss": 0.9161, "step": 5109 }, { "epoch": 0.9834488067744419, "grad_norm": 1.815218210220337, "learning_rate": 1.436486588255681e-08, "loss": 0.9758, "step": 5110 }, { "epoch": 0.9836412625096228, "grad_norm": 1.7052456140518188, "learning_rate": 1.4032819181509783e-08, "loss": 0.9259, "step": 5111 }, { "epoch": 0.9838337182448037, "grad_norm": 1.9399404525756836, "learning_rate": 1.370465245426167e-08, "loss": 0.9567, "step": 5112 }, { "epoch": 0.9840261739799846, "grad_norm": 1.6699292659759521, "learning_rate": 1.3380365828318254e-08, "loss": 0.8789, "step": 5113 }, { "epoch": 0.9842186297151655, "grad_norm": 1.7068564891815186, "learning_rate": 1.3059959429679859e-08, "loss": 0.967, "step": 5114 }, { "epoch": 0.9844110854503464, "grad_norm": 1.55268132686615, "learning_rate": 1.274343338283801e-08, "loss": 0.8599, "step": 5115 }, { "epoch": 0.9846035411855273, "grad_norm": 0.7772056460380554, "learning_rate": 1.2430787810776556e-08, "loss": 0.8197, "step": 5116 }, { "epoch": 0.9847959969207082, "grad_norm": 1.6491152048110962, "learning_rate": 1.2122022834970548e-08, "loss": 0.8354, "step": 5117 }, { "epoch": 0.9849884526558892, "grad_norm": 1.7647578716278076, "learning_rate": 1.1817138575389576e-08, "loss": 0.9154, "step": 5118 }, { "epoch": 0.9851809083910701, "grad_norm": 1.7220553159713745, "learning_rate": 1.1516135150493323e-08, "loss": 0.9506, "step": 5119 }, { "epoch": 0.985373364126251, "grad_norm": 1.5396755933761597, "learning_rate": 1.1219012677234908e-08, "loss": 0.8491, "step": 5120 }, { "epoch": 0.9855658198614319, "grad_norm": 1.7235908508300781, "learning_rate": 1.0925771271058649e-08, "loss": 0.9324, "step": 5121 }, { "epoch": 0.9857582755966128, "grad_norm": 1.9007925987243652, "learning_rate": 1.0636411045902296e-08, "loss": 0.9982, "step": 5122 }, { "epoch": 0.9859507313317937, "grad_norm": 1.621556043624878, "learning_rate": 1.035093211419369e-08, "loss": 0.8682, "step": 5123 }, { "epoch": 0.9861431870669746, "grad_norm": 1.7614240646362305, "learning_rate": 1.0069334586854106e-08, "loss": 0.8207, "step": 5124 }, { "epoch": 0.9863356428021555, "grad_norm": 1.6785147190093994, "learning_rate": 9.791618573294914e-09, "loss": 0.9804, "step": 5125 }, { "epoch": 0.9865280985373364, "grad_norm": 1.948093056678772, "learning_rate": 9.517784181422018e-09, "loss": 0.9094, "step": 5126 }, { "epoch": 0.9867205542725174, "grad_norm": 1.8337852954864502, "learning_rate": 9.247831517632533e-09, "loss": 0.9756, "step": 5127 }, { "epoch": 0.9869130100076983, "grad_norm": 1.8649026155471802, "learning_rate": 8.981760686811448e-09, "loss": 0.9226, "step": 5128 }, { "epoch": 0.9871054657428792, "grad_norm": 1.7597006559371948, "learning_rate": 8.719571792341618e-09, "loss": 0.9377, "step": 5129 }, { "epoch": 0.9872979214780601, "grad_norm": 1.7643260955810547, "learning_rate": 8.461264936093783e-09, "loss": 0.9125, "step": 5130 }, { "epoch": 0.987490377213241, "grad_norm": 1.7881656885147095, "learning_rate": 8.20684021843099e-09, "loss": 0.8924, "step": 5131 }, { "epoch": 0.9876828329484219, "grad_norm": 1.6305475234985352, "learning_rate": 7.956297738207496e-09, "loss": 0.8397, "step": 5132 }, { "epoch": 0.9878752886836027, "grad_norm": 1.8011974096298218, "learning_rate": 7.70963759277099e-09, "loss": 0.8755, "step": 5133 }, { "epoch": 0.9880677444187836, "grad_norm": 1.776333212852478, "learning_rate": 7.466859877958143e-09, "loss": 0.9248, "step": 5134 }, { "epoch": 0.9882602001539645, "grad_norm": 1.6891127824783325, "learning_rate": 7.227964688099054e-09, "loss": 0.9339, "step": 5135 }, { "epoch": 0.9884526558891455, "grad_norm": 2.063249349594116, "learning_rate": 6.992952116013918e-09, "loss": 0.9193, "step": 5136 }, { "epoch": 0.9886451116243264, "grad_norm": 1.6560297012329102, "learning_rate": 6.76182225301747e-09, "loss": 0.9881, "step": 5137 }, { "epoch": 0.9888375673595073, "grad_norm": 1.8064961433410645, "learning_rate": 6.534575188911208e-09, "loss": 1.0011, "step": 5138 }, { "epoch": 0.9890300230946882, "grad_norm": 0.9157009124755859, "learning_rate": 6.3112110119900596e-09, "loss": 0.9147, "step": 5139 }, { "epoch": 0.9892224788298691, "grad_norm": 1.8693196773529053, "learning_rate": 6.091729809042379e-09, "loss": 0.9453, "step": 5140 }, { "epoch": 0.98941493456505, "grad_norm": 1.9743990898132324, "learning_rate": 5.876131665345508e-09, "loss": 0.7474, "step": 5141 }, { "epoch": 0.9896073903002309, "grad_norm": 1.8233686685562134, "learning_rate": 5.664416664666883e-09, "loss": 0.9497, "step": 5142 }, { "epoch": 0.9897998460354118, "grad_norm": 1.6983250379562378, "learning_rate": 5.4565848892695894e-09, "loss": 0.8928, "step": 5143 }, { "epoch": 0.9899923017705927, "grad_norm": 1.665627360343933, "learning_rate": 5.252636419902368e-09, "loss": 0.8895, "step": 5144 }, { "epoch": 0.9901847575057737, "grad_norm": 1.867610216140747, "learning_rate": 5.052571335808499e-09, "loss": 0.9384, "step": 5145 }, { "epoch": 0.9903772132409546, "grad_norm": 1.780145287513733, "learning_rate": 4.856389714723575e-09, "loss": 0.8893, "step": 5146 }, { "epoch": 0.9905696689761355, "grad_norm": 1.6347787380218506, "learning_rate": 4.6640916328710705e-09, "loss": 0.9878, "step": 5147 }, { "epoch": 0.9907621247113164, "grad_norm": 1.6712716817855835, "learning_rate": 4.475677164966774e-09, "loss": 0.9392, "step": 5148 }, { "epoch": 0.9909545804464973, "grad_norm": 1.660496473312378, "learning_rate": 4.2911463842176815e-09, "loss": 0.9233, "step": 5149 }, { "epoch": 0.9911470361816782, "grad_norm": 1.7979397773742676, "learning_rate": 4.110499362323106e-09, "loss": 1.0314, "step": 5150 }, { "epoch": 0.9913394919168591, "grad_norm": 1.743967056274414, "learning_rate": 3.933736169471347e-09, "loss": 0.8756, "step": 5151 }, { "epoch": 0.99153194765204, "grad_norm": 3.4190900325775146, "learning_rate": 3.760856874341912e-09, "loss": 0.8715, "step": 5152 }, { "epoch": 0.9917244033872209, "grad_norm": 1.6083959341049194, "learning_rate": 3.591861544106623e-09, "loss": 0.815, "step": 5153 }, { "epoch": 0.9919168591224018, "grad_norm": 1.6205967664718628, "learning_rate": 3.4267502444274013e-09, "loss": 0.9403, "step": 5154 }, { "epoch": 0.9921093148575828, "grad_norm": 1.5808223485946655, "learning_rate": 3.2655230394562644e-09, "loss": 0.8826, "step": 5155 }, { "epoch": 0.9923017705927637, "grad_norm": 1.7393923997879028, "learning_rate": 3.1081799918375454e-09, "loss": 0.8622, "step": 5156 }, { "epoch": 0.9924942263279446, "grad_norm": 1.721386194229126, "learning_rate": 2.954721162704566e-09, "loss": 0.8532, "step": 5157 }, { "epoch": 0.9926866820631255, "grad_norm": 1.7016346454620361, "learning_rate": 2.805146611684073e-09, "loss": 0.9182, "step": 5158 }, { "epoch": 0.9928791377983064, "grad_norm": 1.7341364622116089, "learning_rate": 2.6594563968918018e-09, "loss": 0.8765, "step": 5159 }, { "epoch": 0.9930715935334873, "grad_norm": 1.726991891860962, "learning_rate": 2.5176505749346937e-09, "loss": 0.9515, "step": 5160 }, { "epoch": 0.9932640492686682, "grad_norm": 1.6547901630401611, "learning_rate": 2.379729200908676e-09, "loss": 0.9262, "step": 5161 }, { "epoch": 0.9934565050038491, "grad_norm": 1.6144102811813354, "learning_rate": 2.245692328404214e-09, "loss": 0.8586, "step": 5162 }, { "epoch": 0.99364896073903, "grad_norm": 2.197000741958618, "learning_rate": 2.11554000949965e-09, "loss": 0.9884, "step": 5163 }, { "epoch": 0.993841416474211, "grad_norm": 1.6308985948562622, "learning_rate": 1.9892722947645328e-09, "loss": 0.9442, "step": 5164 }, { "epoch": 0.9940338722093919, "grad_norm": 1.7169991731643677, "learning_rate": 1.8668892332596165e-09, "loss": 0.8242, "step": 5165 }, { "epoch": 0.9942263279445728, "grad_norm": 1.7209423780441284, "learning_rate": 1.7483908725357546e-09, "loss": 0.8967, "step": 5166 }, { "epoch": 0.9944187836797537, "grad_norm": 1.963261365890503, "learning_rate": 1.6337772586338952e-09, "loss": 0.9457, "step": 5167 }, { "epoch": 0.9946112394149346, "grad_norm": 1.6533063650131226, "learning_rate": 1.5230484360873043e-09, "loss": 1.0188, "step": 5168 }, { "epoch": 0.9948036951501155, "grad_norm": 1.6715199947357178, "learning_rate": 1.4162044479182347e-09, "loss": 0.9114, "step": 5169 }, { "epoch": 0.9949961508852964, "grad_norm": 1.8019225597381592, "learning_rate": 1.3132453356412556e-09, "loss": 0.9162, "step": 5170 }, { "epoch": 0.9951886066204773, "grad_norm": 1.6687161922454834, "learning_rate": 1.214171139258813e-09, "loss": 0.9682, "step": 5171 }, { "epoch": 0.9953810623556582, "grad_norm": 1.6832813024520874, "learning_rate": 1.1189818972656697e-09, "loss": 0.9001, "step": 5172 }, { "epoch": 0.9955735180908392, "grad_norm": 1.723149061203003, "learning_rate": 1.0276776466477955e-09, "loss": 0.9117, "step": 5173 }, { "epoch": 0.9957659738260201, "grad_norm": 1.8294788599014282, "learning_rate": 9.40258422880147e-10, "loss": 0.8788, "step": 5174 }, { "epoch": 0.995958429561201, "grad_norm": 1.888719916343689, "learning_rate": 8.567242599299974e-10, "loss": 0.9134, "step": 5175 }, { "epoch": 0.9961508852963818, "grad_norm": 1.6505712270736694, "learning_rate": 7.770751902513862e-10, "loss": 0.8983, "step": 5176 }, { "epoch": 0.9963433410315627, "grad_norm": 0.8461357355117798, "learning_rate": 7.013112447940007e-10, "loss": 0.8625, "step": 5177 }, { "epoch": 0.9965357967667436, "grad_norm": 1.8349761962890625, "learning_rate": 6.294324529942942e-10, "loss": 0.8957, "step": 5178 }, { "epoch": 0.9967282525019245, "grad_norm": 1.849012017250061, "learning_rate": 5.614388427799266e-10, "loss": 0.8202, "step": 5179 }, { "epoch": 0.9969207082371054, "grad_norm": 1.773707628250122, "learning_rate": 4.973304405697654e-10, "loss": 0.9345, "step": 5180 }, { "epoch": 0.9971131639722863, "grad_norm": 1.8741662502288818, "learning_rate": 4.3710727127277417e-10, "loss": 0.8964, "step": 5181 }, { "epoch": 0.9973056197074672, "grad_norm": 1.6532068252563477, "learning_rate": 3.807693582869032e-10, "loss": 0.9381, "step": 5182 }, { "epoch": 0.9974980754426482, "grad_norm": 1.7500007152557373, "learning_rate": 3.283167235035301e-10, "loss": 0.8844, "step": 5183 }, { "epoch": 0.9976905311778291, "grad_norm": 1.7772070169448853, "learning_rate": 2.797493873019086e-10, "loss": 0.9269, "step": 5184 }, { "epoch": 0.99788298691301, "grad_norm": 1.9817670583724976, "learning_rate": 2.350673685536098e-10, "loss": 0.969, "step": 5185 }, { "epoch": 0.9980754426481909, "grad_norm": 1.5418437719345093, "learning_rate": 1.9427068461808086e-10, "loss": 0.9937, "step": 5186 }, { "epoch": 0.9982678983833718, "grad_norm": 1.738728404045105, "learning_rate": 1.5735935134708613e-10, "loss": 0.9953, "step": 5187 }, { "epoch": 0.9984603541185527, "grad_norm": 1.6960806846618652, "learning_rate": 1.2433338308137645e-10, "loss": 0.916, "step": 5188 }, { "epoch": 0.9986528098537336, "grad_norm": 1.7969945669174194, "learning_rate": 9.519279265512993e-11, "loss": 0.9096, "step": 5189 }, { "epoch": 0.9988452655889145, "grad_norm": 1.590344786643982, "learning_rate": 6.99375913892908e-11, "loss": 0.9867, "step": 5190 }, { "epoch": 0.9990377213240954, "grad_norm": 1.7562882900238037, "learning_rate": 4.856778909601012e-11, "loss": 0.9045, "step": 5191 }, { "epoch": 0.9992301770592764, "grad_norm": 1.7541764974594116, "learning_rate": 3.108339407975613e-11, "loss": 0.9464, "step": 5192 }, { "epoch": 0.9994226327944573, "grad_norm": 1.7172061204910278, "learning_rate": 1.74844131328733e-11, "loss": 0.9319, "step": 5193 }, { "epoch": 0.9996150885296382, "grad_norm": 1.7965987920761108, "learning_rate": 7.770851538912993e-12, "loss": 0.8512, "step": 5194 }, { "epoch": 0.9998075442648191, "grad_norm": 1.824587106704712, "learning_rate": 1.9427130737437183e-12, "loss": 0.8708, "step": 5195 }, { "epoch": 1.0, "grad_norm": 1.07023024559021, "learning_rate": 0.0, "loss": 0.9225, "step": 5196 }, { "epoch": 1.0, "step": 5196, "total_flos": 4.0873852180826685e+18, "train_loss": 0.9749902330562644, "train_runtime": 11353.9896, "train_samples_per_second": 58.575, "train_steps_per_second": 0.458 } ], "logging_steps": 1.0, "max_steps": 5196, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 2000, "total_flos": 4.0873852180826685e+18, "train_batch_size": 8, "trial_name": null, "trial_params": null }