ultra-lora-3000 / trainer_state.json
bziemba's picture
Upload folder using huggingface_hub
a72d112 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.7378939277478862,
"eval_steps": 500,
"global_step": 3000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004919292851652575,
"grad_norm": 2.650216579437256,
"learning_rate": 1.557377049180328e-06,
"loss": 1.9055,
"step": 20
},
{
"epoch": 0.00983858570330515,
"grad_norm": 2.133456230163574,
"learning_rate": 3.1967213114754097e-06,
"loss": 1.8043,
"step": 40
},
{
"epoch": 0.014757878554957724,
"grad_norm": 2.4590532779693604,
"learning_rate": 4.836065573770492e-06,
"loss": 1.7683,
"step": 60
},
{
"epoch": 0.0196771714066103,
"grad_norm": 1.522373080253601,
"learning_rate": 6.475409836065574e-06,
"loss": 1.6977,
"step": 80
},
{
"epoch": 0.024596464258262875,
"grad_norm": 1.6893309354782104,
"learning_rate": 8.114754098360657e-06,
"loss": 1.6298,
"step": 100
},
{
"epoch": 0.02951575710991545,
"grad_norm": 1.5092189311981201,
"learning_rate": 9.754098360655738e-06,
"loss": 1.4915,
"step": 120
},
{
"epoch": 0.03443504996156802,
"grad_norm": 1.3016860485076904,
"learning_rate": 1.139344262295082e-05,
"loss": 1.3592,
"step": 140
},
{
"epoch": 0.0393543428132206,
"grad_norm": 1.8532716035842896,
"learning_rate": 1.3032786885245902e-05,
"loss": 1.2786,
"step": 160
},
{
"epoch": 0.044273635664873176,
"grad_norm": 1.5091300010681152,
"learning_rate": 1.4672131147540986e-05,
"loss": 1.1882,
"step": 180
},
{
"epoch": 0.04919292851652575,
"grad_norm": 2.3855371475219727,
"learning_rate": 1.6311475409836068e-05,
"loss": 1.2969,
"step": 200
},
{
"epoch": 0.05411222136817832,
"grad_norm": 1.8920749425888062,
"learning_rate": 1.7950819672131146e-05,
"loss": 1.1571,
"step": 220
},
{
"epoch": 0.0590315142198309,
"grad_norm": 1.7724186182022095,
"learning_rate": 1.9590163934426232e-05,
"loss": 1.1782,
"step": 240
},
{
"epoch": 0.06395080707148347,
"grad_norm": 1.53757643699646,
"learning_rate": 2.122950819672131e-05,
"loss": 1.1917,
"step": 260
},
{
"epoch": 0.06887009992313604,
"grad_norm": 1.773545503616333,
"learning_rate": 2.2868852459016393e-05,
"loss": 1.1343,
"step": 280
},
{
"epoch": 0.07378939277478862,
"grad_norm": 2.211615800857544,
"learning_rate": 2.4508196721311478e-05,
"loss": 1.1242,
"step": 300
},
{
"epoch": 0.0787086856264412,
"grad_norm": 1.4129002094268799,
"learning_rate": 2.614754098360656e-05,
"loss": 1.0075,
"step": 320
},
{
"epoch": 0.08362797847809378,
"grad_norm": 1.8837412595748901,
"learning_rate": 2.778688524590164e-05,
"loss": 1.0791,
"step": 340
},
{
"epoch": 0.08854727132974635,
"grad_norm": 2.0091936588287354,
"learning_rate": 2.9426229508196725e-05,
"loss": 1.0921,
"step": 360
},
{
"epoch": 0.09346656418139893,
"grad_norm": 2.447704792022705,
"learning_rate": 3.106557377049181e-05,
"loss": 1.0776,
"step": 380
},
{
"epoch": 0.0983858570330515,
"grad_norm": 2.063652753829956,
"learning_rate": 3.2704918032786885e-05,
"loss": 1.0928,
"step": 400
},
{
"epoch": 0.10330514988470407,
"grad_norm": 1.985748529434204,
"learning_rate": 3.434426229508197e-05,
"loss": 1.1167,
"step": 420
},
{
"epoch": 0.10822444273635665,
"grad_norm": 2.069704532623291,
"learning_rate": 3.5983606557377056e-05,
"loss": 1.1432,
"step": 440
},
{
"epoch": 0.11314373558800922,
"grad_norm": 1.5993927717208862,
"learning_rate": 3.762295081967213e-05,
"loss": 1.0484,
"step": 460
},
{
"epoch": 0.1180630284396618,
"grad_norm": 1.8127434253692627,
"learning_rate": 3.9262295081967214e-05,
"loss": 1.0379,
"step": 480
},
{
"epoch": 0.12298232129131437,
"grad_norm": 1.3972879648208618,
"learning_rate": 4.09016393442623e-05,
"loss": 1.0084,
"step": 500
},
{
"epoch": 0.12790161414296694,
"grad_norm": 2.6233468055725098,
"learning_rate": 4.254098360655738e-05,
"loss": 0.9489,
"step": 520
},
{
"epoch": 0.13282090699461951,
"grad_norm": 1.9032576084136963,
"learning_rate": 4.4180327868852463e-05,
"loss": 0.9932,
"step": 540
},
{
"epoch": 0.1377401998462721,
"grad_norm": 2.0064117908477783,
"learning_rate": 4.581967213114754e-05,
"loss": 0.987,
"step": 560
},
{
"epoch": 0.14265949269792466,
"grad_norm": 1.6437580585479736,
"learning_rate": 4.745901639344262e-05,
"loss": 1.0212,
"step": 580
},
{
"epoch": 0.14757878554957723,
"grad_norm": 1.3102104663848877,
"learning_rate": 4.9098360655737706e-05,
"loss": 1.0538,
"step": 600
},
{
"epoch": 0.15249807840122984,
"grad_norm": 1.6655702590942383,
"learning_rate": 5.073770491803279e-05,
"loss": 0.9762,
"step": 620
},
{
"epoch": 0.1574173712528824,
"grad_norm": 1.488173007965088,
"learning_rate": 5.237704918032788e-05,
"loss": 1.0271,
"step": 640
},
{
"epoch": 0.16233666410453498,
"grad_norm": 2.1534035205841064,
"learning_rate": 5.401639344262295e-05,
"loss": 1.0124,
"step": 660
},
{
"epoch": 0.16725595695618756,
"grad_norm": 2.0473647117614746,
"learning_rate": 5.5655737704918035e-05,
"loss": 1.0228,
"step": 680
},
{
"epoch": 0.17217524980784013,
"grad_norm": 1.5801570415496826,
"learning_rate": 5.7295081967213114e-05,
"loss": 0.9668,
"step": 700
},
{
"epoch": 0.1770945426594927,
"grad_norm": 1.749709963798523,
"learning_rate": 5.89344262295082e-05,
"loss": 1.0632,
"step": 720
},
{
"epoch": 0.18201383551114528,
"grad_norm": 2.072209596633911,
"learning_rate": 6.0573770491803284e-05,
"loss": 0.9709,
"step": 740
},
{
"epoch": 0.18693312836279785,
"grad_norm": 1.6242225170135498,
"learning_rate": 6.221311475409836e-05,
"loss": 1.0601,
"step": 760
},
{
"epoch": 0.19185242121445042,
"grad_norm": 1.41958749294281,
"learning_rate": 6.385245901639346e-05,
"loss": 0.9846,
"step": 780
},
{
"epoch": 0.196771714066103,
"grad_norm": 1.4413362741470337,
"learning_rate": 6.549180327868852e-05,
"loss": 1.0331,
"step": 800
},
{
"epoch": 0.20169100691775557,
"grad_norm": 1.8630566596984863,
"learning_rate": 6.713114754098361e-05,
"loss": 1.0214,
"step": 820
},
{
"epoch": 0.20661029976940815,
"grad_norm": 1.689773678779602,
"learning_rate": 6.877049180327869e-05,
"loss": 0.9745,
"step": 840
},
{
"epoch": 0.21152959262106072,
"grad_norm": 1.9186772108078003,
"learning_rate": 7.040983606557377e-05,
"loss": 1.0628,
"step": 860
},
{
"epoch": 0.2164488854727133,
"grad_norm": 1.933374285697937,
"learning_rate": 7.204918032786886e-05,
"loss": 0.9864,
"step": 880
},
{
"epoch": 0.22136817832436587,
"grad_norm": 1.5824335813522339,
"learning_rate": 7.368852459016394e-05,
"loss": 1.0016,
"step": 900
},
{
"epoch": 0.22628747117601844,
"grad_norm": 1.591445803642273,
"learning_rate": 7.532786885245902e-05,
"loss": 1.005,
"step": 920
},
{
"epoch": 0.231206764027671,
"grad_norm": 1.7586395740509033,
"learning_rate": 7.69672131147541e-05,
"loss": 0.9775,
"step": 940
},
{
"epoch": 0.2361260568793236,
"grad_norm": 2.071493148803711,
"learning_rate": 7.860655737704918e-05,
"loss": 1.0228,
"step": 960
},
{
"epoch": 0.24104534973097616,
"grad_norm": 1.3450851440429688,
"learning_rate": 8.024590163934427e-05,
"loss": 1.0164,
"step": 980
},
{
"epoch": 0.24596464258262873,
"grad_norm": 1.8442620038986206,
"learning_rate": 8.188524590163935e-05,
"loss": 0.9911,
"step": 1000
},
{
"epoch": 0.25088393543428134,
"grad_norm": 1.098396897315979,
"learning_rate": 8.352459016393444e-05,
"loss": 0.9718,
"step": 1020
},
{
"epoch": 0.2558032282859339,
"grad_norm": 1.005915880203247,
"learning_rate": 8.51639344262295e-05,
"loss": 0.9436,
"step": 1040
},
{
"epoch": 0.2607225211375865,
"grad_norm": 1.113786220550537,
"learning_rate": 8.68032786885246e-05,
"loss": 1.0079,
"step": 1060
},
{
"epoch": 0.26564181398923903,
"grad_norm": 1.1104156970977783,
"learning_rate": 8.844262295081968e-05,
"loss": 0.9534,
"step": 1080
},
{
"epoch": 0.27056110684089163,
"grad_norm": 0.9686466455459595,
"learning_rate": 9.008196721311476e-05,
"loss": 0.9199,
"step": 1100
},
{
"epoch": 0.2754803996925442,
"grad_norm": 1.78913414478302,
"learning_rate": 9.172131147540985e-05,
"loss": 0.9132,
"step": 1120
},
{
"epoch": 0.2803996925441968,
"grad_norm": 1.5726219415664673,
"learning_rate": 9.336065573770493e-05,
"loss": 1.053,
"step": 1140
},
{
"epoch": 0.2853189853958493,
"grad_norm": 1.7471468448638916,
"learning_rate": 9.5e-05,
"loss": 0.9108,
"step": 1160
},
{
"epoch": 0.2902382782475019,
"grad_norm": 1.045501947402954,
"learning_rate": 9.663934426229508e-05,
"loss": 1.0033,
"step": 1180
},
{
"epoch": 0.29515757109915447,
"grad_norm": 1.3126252889633179,
"learning_rate": 9.827868852459016e-05,
"loss": 0.9574,
"step": 1200
},
{
"epoch": 0.30007686395080707,
"grad_norm": 1.1474781036376953,
"learning_rate": 9.991803278688525e-05,
"loss": 0.9781,
"step": 1220
},
{
"epoch": 0.3049961568024597,
"grad_norm": 1.435939073562622,
"learning_rate": 9.999926090641448e-05,
"loss": 0.9573,
"step": 1240
},
{
"epoch": 0.3099154496541122,
"grad_norm": 2.488921642303467,
"learning_rate": 9.999688600430955e-05,
"loss": 1.02,
"step": 1260
},
{
"epoch": 0.3148347425057648,
"grad_norm": 1.1515672206878662,
"learning_rate": 9.999287332412564e-05,
"loss": 0.9107,
"step": 1280
},
{
"epoch": 0.31975403535741737,
"grad_norm": 1.6522116661071777,
"learning_rate": 9.998722299730854e-05,
"loss": 0.8688,
"step": 1300
},
{
"epoch": 0.32467332820906997,
"grad_norm": 1.14382803440094,
"learning_rate": 9.997993520894937e-05,
"loss": 0.9889,
"step": 1320
},
{
"epoch": 0.3295926210607225,
"grad_norm": 1.205665111541748,
"learning_rate": 9.997101019777864e-05,
"loss": 1.014,
"step": 1340
},
{
"epoch": 0.3345119139123751,
"grad_norm": 1.1363279819488525,
"learning_rate": 9.996044825615826e-05,
"loss": 0.9774,
"step": 1360
},
{
"epoch": 0.33943120676402766,
"grad_norm": 1.490625023841858,
"learning_rate": 9.994824973007218e-05,
"loss": 1.0072,
"step": 1380
},
{
"epoch": 0.34435049961568026,
"grad_norm": 1.5652107000350952,
"learning_rate": 9.993441501911479e-05,
"loss": 1.0104,
"step": 1400
},
{
"epoch": 0.3492697924673328,
"grad_norm": 1.1277583837509155,
"learning_rate": 9.991894457647808e-05,
"loss": 0.9207,
"step": 1420
},
{
"epoch": 0.3541890853189854,
"grad_norm": 1.0140403509140015,
"learning_rate": 9.990183890893663e-05,
"loss": 0.9267,
"step": 1440
},
{
"epoch": 0.35910837817063795,
"grad_norm": 1.0925248861312866,
"learning_rate": 9.98830985768311e-05,
"loss": 0.989,
"step": 1460
},
{
"epoch": 0.36402767102229056,
"grad_norm": 1.3068804740905762,
"learning_rate": 9.986272419404982e-05,
"loss": 0.9893,
"step": 1480
},
{
"epoch": 0.3689469638739431,
"grad_norm": 1.398160457611084,
"learning_rate": 9.98407164280087e-05,
"loss": 0.9825,
"step": 1500
},
{
"epoch": 0.3738662567255957,
"grad_norm": 1.2048150300979614,
"learning_rate": 9.981707599962937e-05,
"loss": 0.916,
"step": 1520
},
{
"epoch": 0.37878554957724825,
"grad_norm": 0.9273918271064758,
"learning_rate": 9.979180368331558e-05,
"loss": 0.9421,
"step": 1540
},
{
"epoch": 0.38370484242890085,
"grad_norm": 1.247428297996521,
"learning_rate": 9.97649003069278e-05,
"loss": 0.9482,
"step": 1560
},
{
"epoch": 0.3886241352805534,
"grad_norm": 1.952083706855774,
"learning_rate": 9.973636675175613e-05,
"loss": 0.9189,
"step": 1580
},
{
"epoch": 0.393543428132206,
"grad_norm": 0.979943037033081,
"learning_rate": 9.970620395249137e-05,
"loss": 0.9506,
"step": 1600
},
{
"epoch": 0.39846272098385854,
"grad_norm": 0.9653918147087097,
"learning_rate": 9.967441289719452e-05,
"loss": 0.927,
"step": 1620
},
{
"epoch": 0.40338201383551114,
"grad_norm": 1.7574175596237183,
"learning_rate": 9.96409946272643e-05,
"loss": 0.9486,
"step": 1640
},
{
"epoch": 0.40830130668716375,
"grad_norm": 0.8972555994987488,
"learning_rate": 9.960595023740307e-05,
"loss": 0.9739,
"step": 1660
},
{
"epoch": 0.4132205995388163,
"grad_norm": 0.8190364837646484,
"learning_rate": 9.956928087558102e-05,
"loss": 0.943,
"step": 1680
},
{
"epoch": 0.4181398923904689,
"grad_norm": 0.9662739634513855,
"learning_rate": 9.953098774299847e-05,
"loss": 0.982,
"step": 1700
},
{
"epoch": 0.42305918524212144,
"grad_norm": 0.9831930994987488,
"learning_rate": 9.949107209404665e-05,
"loss": 1.0335,
"step": 1720
},
{
"epoch": 0.42797847809377404,
"grad_norm": 0.8506195545196533,
"learning_rate": 9.944953523626643e-05,
"loss": 0.9193,
"step": 1740
},
{
"epoch": 0.4328977709454266,
"grad_norm": 0.9357776641845703,
"learning_rate": 9.940637853030572e-05,
"loss": 0.9683,
"step": 1760
},
{
"epoch": 0.4378170637970792,
"grad_norm": 1.3345489501953125,
"learning_rate": 9.936160338987466e-05,
"loss": 0.956,
"step": 1780
},
{
"epoch": 0.44273635664873173,
"grad_norm": 0.8419784307479858,
"learning_rate": 9.931521128169947e-05,
"loss": 0.9407,
"step": 1800
},
{
"epoch": 0.44765564950038433,
"grad_norm": 0.9009922742843628,
"learning_rate": 9.926720372547438e-05,
"loss": 0.9174,
"step": 1820
},
{
"epoch": 0.4525749423520369,
"grad_norm": 1.1509584188461304,
"learning_rate": 9.921758229381177e-05,
"loss": 0.9644,
"step": 1840
},
{
"epoch": 0.4574942352036895,
"grad_norm": 1.1815032958984375,
"learning_rate": 9.916634861219076e-05,
"loss": 0.9572,
"step": 1860
},
{
"epoch": 0.462413528055342,
"grad_norm": 1.1411418914794922,
"learning_rate": 9.911350435890386e-05,
"loss": 0.9424,
"step": 1880
},
{
"epoch": 0.46733282090699463,
"grad_norm": 1.2081893682479858,
"learning_rate": 9.905905126500208e-05,
"loss": 0.9643,
"step": 1900
},
{
"epoch": 0.4722521137586472,
"grad_norm": 1.0384275913238525,
"learning_rate": 9.900299111423819e-05,
"loss": 0.9573,
"step": 1920
},
{
"epoch": 0.4771714066102998,
"grad_norm": 1.0825401544570923,
"learning_rate": 9.894532574300827e-05,
"loss": 0.9533,
"step": 1940
},
{
"epoch": 0.4820906994619523,
"grad_norm": 0.9118188619613647,
"learning_rate": 9.888605704029161e-05,
"loss": 0.952,
"step": 1960
},
{
"epoch": 0.4870099923136049,
"grad_norm": 1.1044224500656128,
"learning_rate": 9.882518694758875e-05,
"loss": 0.8827,
"step": 1980
},
{
"epoch": 0.49192928516525747,
"grad_norm": 0.9803332090377808,
"learning_rate": 9.876271745885792e-05,
"loss": 0.9254,
"step": 2000
},
{
"epoch": 0.49684857801691007,
"grad_norm": 1.0061184167861938,
"learning_rate": 9.869865062044979e-05,
"loss": 0.9428,
"step": 2020
},
{
"epoch": 0.5017678708685627,
"grad_norm": 0.9520462155342102,
"learning_rate": 9.863298853104032e-05,
"loss": 0.9292,
"step": 2040
},
{
"epoch": 0.5066871637202153,
"grad_norm": 1.5129801034927368,
"learning_rate": 9.856573334156209e-05,
"loss": 0.9351,
"step": 2060
},
{
"epoch": 0.5116064565718678,
"grad_norm": 1.4102259874343872,
"learning_rate": 9.84968872551338e-05,
"loss": 0.908,
"step": 2080
},
{
"epoch": 0.5165257494235204,
"grad_norm": 1.0974596738815308,
"learning_rate": 9.842645252698813e-05,
"loss": 0.9548,
"step": 2100
},
{
"epoch": 0.521445042275173,
"grad_norm": 0.9075796008110046,
"learning_rate": 9.835443146439786e-05,
"loss": 0.8862,
"step": 2120
},
{
"epoch": 0.5263643351268256,
"grad_norm": 1.2785767316818237,
"learning_rate": 9.828082642660026e-05,
"loss": 0.9333,
"step": 2140
},
{
"epoch": 0.5312836279784781,
"grad_norm": 1.128395915031433,
"learning_rate": 9.820563982471987e-05,
"loss": 0.9688,
"step": 2160
},
{
"epoch": 0.5362029208301307,
"grad_norm": 1.3578219413757324,
"learning_rate": 9.812887412168941e-05,
"loss": 0.9747,
"step": 2180
},
{
"epoch": 0.5411222136817833,
"grad_norm": 1.431168794631958,
"learning_rate": 9.805053183216923e-05,
"loss": 0.9358,
"step": 2200
},
{
"epoch": 0.5460415065334359,
"grad_norm": 0.8082225918769836,
"learning_rate": 9.797061552246486e-05,
"loss": 0.9133,
"step": 2220
},
{
"epoch": 0.5509607993850884,
"grad_norm": 1.1131757497787476,
"learning_rate": 9.788912781044292e-05,
"loss": 0.9679,
"step": 2240
},
{
"epoch": 0.555880092236741,
"grad_norm": 1.163212776184082,
"learning_rate": 9.780607136544543e-05,
"loss": 0.9928,
"step": 2260
},
{
"epoch": 0.5607993850883936,
"grad_norm": 1.0671954154968262,
"learning_rate": 9.772144890820234e-05,
"loss": 0.8596,
"step": 2280
},
{
"epoch": 0.5657186779400462,
"grad_norm": 1.1239769458770752,
"learning_rate": 9.763526321074241e-05,
"loss": 0.9382,
"step": 2300
},
{
"epoch": 0.5706379707916986,
"grad_norm": 1.2082741260528564,
"learning_rate": 9.754751709630237e-05,
"loss": 0.9354,
"step": 2320
},
{
"epoch": 0.5755572636433512,
"grad_norm": 1.0991088151931763,
"learning_rate": 9.745821343923451e-05,
"loss": 0.9695,
"step": 2340
},
{
"epoch": 0.5804765564950038,
"grad_norm": 1.2947261333465576,
"learning_rate": 9.736735516491247e-05,
"loss": 0.9162,
"step": 2360
},
{
"epoch": 0.5853958493466564,
"grad_norm": 1.3943086862564087,
"learning_rate": 9.727494524963536e-05,
"loss": 0.9511,
"step": 2380
},
{
"epoch": 0.5903151421983089,
"grad_norm": 1.0838426351547241,
"learning_rate": 9.718098672053044e-05,
"loss": 0.983,
"step": 2400
},
{
"epoch": 0.5952344350499615,
"grad_norm": 1.285778522491455,
"learning_rate": 9.708548265545375e-05,
"loss": 0.8961,
"step": 2420
},
{
"epoch": 0.6001537279016141,
"grad_norm": 0.9513759016990662,
"learning_rate": 9.698843618288942e-05,
"loss": 0.9459,
"step": 2440
},
{
"epoch": 0.6050730207532667,
"grad_norm": 0.8295150399208069,
"learning_rate": 9.688985048184717e-05,
"loss": 0.8766,
"step": 2460
},
{
"epoch": 0.6099923136049193,
"grad_norm": 0.8916402459144592,
"learning_rate": 9.67897287817581e-05,
"loss": 0.9322,
"step": 2480
},
{
"epoch": 0.6149116064565718,
"grad_norm": 1.0418959856033325,
"learning_rate": 9.6688074362369e-05,
"loss": 1.0029,
"step": 2500
},
{
"epoch": 0.6198308993082244,
"grad_norm": 1.5775642395019531,
"learning_rate": 9.658489055363486e-05,
"loss": 0.9241,
"step": 2520
},
{
"epoch": 0.624750192159877,
"grad_norm": 1.4654349088668823,
"learning_rate": 9.648018073560976e-05,
"loss": 0.9755,
"step": 2540
},
{
"epoch": 0.6296694850115296,
"grad_norm": 1.1109009981155396,
"learning_rate": 9.637394833833627e-05,
"loss": 0.9492,
"step": 2560
},
{
"epoch": 0.6345887778631821,
"grad_norm": 1.0579159259796143,
"learning_rate": 9.626619684173288e-05,
"loss": 0.9327,
"step": 2580
},
{
"epoch": 0.6395080707148347,
"grad_norm": 1.344231128692627,
"learning_rate": 9.615692977548026e-05,
"loss": 0.9245,
"step": 2600
},
{
"epoch": 0.6444273635664873,
"grad_norm": 1.1978901624679565,
"learning_rate": 9.604615071890541e-05,
"loss": 0.8989,
"step": 2620
},
{
"epoch": 0.6493466564181399,
"grad_norm": 1.2430976629257202,
"learning_rate": 9.593386330086458e-05,
"loss": 0.922,
"step": 2640
},
{
"epoch": 0.6542659492697924,
"grad_norm": 0.8552722334861755,
"learning_rate": 9.582007119962424e-05,
"loss": 0.9013,
"step": 2660
},
{
"epoch": 0.659185242121445,
"grad_norm": 0.8925425410270691,
"learning_rate": 9.570477814274077e-05,
"loss": 1.0079,
"step": 2680
},
{
"epoch": 0.6641045349730976,
"grad_norm": 1.0146712064743042,
"learning_rate": 9.558798790693817e-05,
"loss": 0.9355,
"step": 2700
},
{
"epoch": 0.6690238278247502,
"grad_norm": 1.0288127660751343,
"learning_rate": 9.54697043179845e-05,
"loss": 0.9069,
"step": 2720
},
{
"epoch": 0.6739431206764027,
"grad_norm": 1.180992603302002,
"learning_rate": 9.534993125056643e-05,
"loss": 0.9002,
"step": 2740
},
{
"epoch": 0.6788624135280553,
"grad_norm": 1.1395238637924194,
"learning_rate": 9.522867262816243e-05,
"loss": 0.9654,
"step": 2760
},
{
"epoch": 0.6837817063797079,
"grad_norm": 1.1949700117111206,
"learning_rate": 9.510593242291414e-05,
"loss": 0.8939,
"step": 2780
},
{
"epoch": 0.6887009992313605,
"grad_norm": 1.1609656810760498,
"learning_rate": 9.49817146554963e-05,
"loss": 0.9344,
"step": 2800
},
{
"epoch": 0.693620292083013,
"grad_norm": 1.0648009777069092,
"learning_rate": 9.485602339498509e-05,
"loss": 0.969,
"step": 2820
},
{
"epoch": 0.6985395849346656,
"grad_norm": 0.7220922112464905,
"learning_rate": 9.472886275872475e-05,
"loss": 0.8687,
"step": 2840
},
{
"epoch": 0.7034588777863182,
"grad_norm": 1.2389453649520874,
"learning_rate": 9.460023691219277e-05,
"loss": 0.8678,
"step": 2860
},
{
"epoch": 0.7083781706379708,
"grad_norm": 0.894183874130249,
"learning_rate": 9.447015006886338e-05,
"loss": 0.9072,
"step": 2880
},
{
"epoch": 0.7132974634896234,
"grad_norm": 1.036689281463623,
"learning_rate": 9.433860649006961e-05,
"loss": 0.9594,
"step": 2900
},
{
"epoch": 0.7182167563412759,
"grad_norm": 1.0629465579986572,
"learning_rate": 9.420561048486359e-05,
"loss": 0.9467,
"step": 2920
},
{
"epoch": 0.7231360491929285,
"grad_norm": 1.2931263446807861,
"learning_rate": 9.40711664098755e-05,
"loss": 0.9452,
"step": 2940
},
{
"epoch": 0.7280553420445811,
"grad_norm": 1.4071450233459473,
"learning_rate": 9.393527866917082e-05,
"loss": 0.8886,
"step": 2960
},
{
"epoch": 0.7329746348962337,
"grad_norm": 0.9684018492698669,
"learning_rate": 9.379795171410601e-05,
"loss": 0.8815,
"step": 2980
},
{
"epoch": 0.7378939277478862,
"grad_norm": 1.0788882970809937,
"learning_rate": 9.36591900431828e-05,
"loss": 0.9255,
"step": 3000
}
],
"logging_steps": 20,
"max_steps": 12198,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.19351181312e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}