LIMO / trainer_state.json
AnikiFan's picture
Upload trainer_state.json with huggingface_hub
4b0be9b verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 14.723716381418093,
"eval_steps": 500,
"global_step": 765,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.097799511002445,
"grad_norm": 2.395451784133911,
"learning_rate": 1.9997891995035914e-05,
"loss": 0.8739,
"num_input_tokens_seen": 163840,
"step": 5
},
{
"epoch": 0.19559902200489,
"grad_norm": 1.3395614624023438,
"learning_rate": 1.999156886888064e-05,
"loss": 0.7957,
"num_input_tokens_seen": 327680,
"step": 10
},
{
"epoch": 0.293398533007335,
"grad_norm": 1.125549554824829,
"learning_rate": 1.9981033287370443e-05,
"loss": 0.7822,
"num_input_tokens_seen": 491520,
"step": 15
},
{
"epoch": 0.39119804400978,
"grad_norm": 1.1936776638031006,
"learning_rate": 1.9966289692316944e-05,
"loss": 0.7815,
"num_input_tokens_seen": 655360,
"step": 20
},
{
"epoch": 0.4889975550122249,
"grad_norm": 1.230684757232666,
"learning_rate": 1.9947344299634464e-05,
"loss": 0.7432,
"num_input_tokens_seen": 819200,
"step": 25
},
{
"epoch": 0.58679706601467,
"grad_norm": 1.0630358457565308,
"learning_rate": 1.992420509671936e-05,
"loss": 0.7389,
"num_input_tokens_seen": 982368,
"step": 30
},
{
"epoch": 0.684596577017115,
"grad_norm": 1.1900779008865356,
"learning_rate": 1.9896881839082554e-05,
"loss": 0.7654,
"num_input_tokens_seen": 1146208,
"step": 35
},
{
"epoch": 0.78239608801956,
"grad_norm": 1.0942089557647705,
"learning_rate": 1.9865386046236597e-05,
"loss": 0.7753,
"num_input_tokens_seen": 1310048,
"step": 40
},
{
"epoch": 0.8801955990220048,
"grad_norm": 1.0910882949829102,
"learning_rate": 1.982973099683902e-05,
"loss": 0.7256,
"num_input_tokens_seen": 1473888,
"step": 45
},
{
"epoch": 0.9779951100244498,
"grad_norm": 1.0902478694915771,
"learning_rate": 1.9789931723094046e-05,
"loss": 0.727,
"num_input_tokens_seen": 1637728,
"step": 50
},
{
"epoch": 1.058679706601467,
"grad_norm": 1.3777414560317993,
"learning_rate": 1.9746005004415004e-05,
"loss": 0.582,
"num_input_tokens_seen": 1770848,
"step": 55
},
{
"epoch": 1.156479217603912,
"grad_norm": 1.3062232732772827,
"learning_rate": 1.9697969360350098e-05,
"loss": 0.5628,
"num_input_tokens_seen": 1934688,
"step": 60
},
{
"epoch": 1.254278728606357,
"grad_norm": 1.239915132522583,
"learning_rate": 1.9645845042774555e-05,
"loss": 0.5561,
"num_input_tokens_seen": 2098528,
"step": 65
},
{
"epoch": 1.352078239608802,
"grad_norm": 1.1612083911895752,
"learning_rate": 1.9589654027352412e-05,
"loss": 0.5778,
"num_input_tokens_seen": 2262368,
"step": 70
},
{
"epoch": 1.449877750611247,
"grad_norm": 1.2168060541152954,
"learning_rate": 1.9529420004271568e-05,
"loss": 0.5207,
"num_input_tokens_seen": 2426128,
"step": 75
},
{
"epoch": 1.5476772616136918,
"grad_norm": 1.2806190252304077,
"learning_rate": 1.9465168368255946e-05,
"loss": 0.5452,
"num_input_tokens_seen": 2589968,
"step": 80
},
{
"epoch": 1.6454767726161368,
"grad_norm": 1.17153000831604,
"learning_rate": 1.9396926207859085e-05,
"loss": 0.5379,
"num_input_tokens_seen": 2753808,
"step": 85
},
{
"epoch": 1.7432762836185818,
"grad_norm": 1.1548662185668945,
"learning_rate": 1.932472229404356e-05,
"loss": 0.5116,
"num_input_tokens_seen": 2917648,
"step": 90
},
{
"epoch": 1.8410757946210268,
"grad_norm": 1.3172167539596558,
"learning_rate": 1.924858706805112e-05,
"loss": 0.5532,
"num_input_tokens_seen": 3081488,
"step": 95
},
{
"epoch": 1.9388753056234718,
"grad_norm": 1.1930724382400513,
"learning_rate": 1.9168552628568632e-05,
"loss": 0.5659,
"num_input_tokens_seen": 3245328,
"step": 100
},
{
"epoch": 2.019559902200489,
"grad_norm": 1.7079554796218872,
"learning_rate": 1.9084652718195237e-05,
"loss": 0.6028,
"num_input_tokens_seen": 3378448,
"step": 105
},
{
"epoch": 2.117359413202934,
"grad_norm": 1.7393510341644287,
"learning_rate": 1.8996922709216456e-05,
"loss": 0.4228,
"num_input_tokens_seen": 3542288,
"step": 110
},
{
"epoch": 2.215158924205379,
"grad_norm": 1.2822431325912476,
"learning_rate": 1.8905399588691165e-05,
"loss": 0.3648,
"num_input_tokens_seen": 3706128,
"step": 115
},
{
"epoch": 2.312958435207824,
"grad_norm": 1.3198189735412598,
"learning_rate": 1.8810121942857848e-05,
"loss": 0.3822,
"num_input_tokens_seen": 3869968,
"step": 120
},
{
"epoch": 2.410757946210269,
"grad_norm": 1.2831072807312012,
"learning_rate": 1.8711129940866577e-05,
"loss": 0.3907,
"num_input_tokens_seen": 4033808,
"step": 125
},
{
"epoch": 2.508557457212714,
"grad_norm": 1.4272454977035522,
"learning_rate": 1.860846531784368e-05,
"loss": 0.3936,
"num_input_tokens_seen": 4197648,
"step": 130
},
{
"epoch": 2.606356968215159,
"grad_norm": 1.3846728801727295,
"learning_rate": 1.8502171357296144e-05,
"loss": 0.376,
"num_input_tokens_seen": 4361488,
"step": 135
},
{
"epoch": 2.704156479217604,
"grad_norm": 1.1824907064437866,
"learning_rate": 1.839229287286327e-05,
"loss": 0.3735,
"num_input_tokens_seen": 4525328,
"step": 140
},
{
"epoch": 2.801955990220049,
"grad_norm": 1.2617613077163696,
"learning_rate": 1.827887618942318e-05,
"loss": 0.4114,
"num_input_tokens_seen": 4689168,
"step": 145
},
{
"epoch": 2.899755501222494,
"grad_norm": 1.1593624353408813,
"learning_rate": 1.816196912356222e-05,
"loss": 0.3696,
"num_input_tokens_seen": 4853008,
"step": 150
},
{
"epoch": 2.997555012224939,
"grad_norm": 1.1450086832046509,
"learning_rate": 1.8041620963415418e-05,
"loss": 0.3888,
"num_input_tokens_seen": 5016848,
"step": 155
},
{
"epoch": 3.078239608801956,
"grad_norm": 1.4760794639587402,
"learning_rate": 1.7917882447886585e-05,
"loss": 0.2605,
"num_input_tokens_seen": 5149968,
"step": 160
},
{
"epoch": 3.176039119804401,
"grad_norm": 1.520880103111267,
"learning_rate": 1.7790805745256703e-05,
"loss": 0.2475,
"num_input_tokens_seen": 5313808,
"step": 165
},
{
"epoch": 3.273838630806846,
"grad_norm": 1.4768394231796265,
"learning_rate": 1.766044443118978e-05,
"loss": 0.2344,
"num_input_tokens_seen": 5477648,
"step": 170
},
{
"epoch": 3.371638141809291,
"grad_norm": 1.4958367347717285,
"learning_rate": 1.7526853466145248e-05,
"loss": 0.2665,
"num_input_tokens_seen": 5641488,
"step": 175
},
{
"epoch": 3.469437652811736,
"grad_norm": 1.4445388317108154,
"learning_rate": 1.7390089172206594e-05,
"loss": 0.2477,
"num_input_tokens_seen": 5805328,
"step": 180
},
{
"epoch": 3.567237163814181,
"grad_norm": 1.4173372983932495,
"learning_rate": 1.725020920933593e-05,
"loss": 0.2679,
"num_input_tokens_seen": 5969168,
"step": 185
},
{
"epoch": 3.665036674816626,
"grad_norm": 1.6111114025115967,
"learning_rate": 1.710727255106447e-05,
"loss": 0.234,
"num_input_tokens_seen": 6133008,
"step": 190
},
{
"epoch": 3.762836185819071,
"grad_norm": 1.4121896028518677,
"learning_rate": 1.696133945962927e-05,
"loss": 0.2587,
"num_input_tokens_seen": 6296848,
"step": 195
},
{
"epoch": 3.860635696821516,
"grad_norm": 1.3143867254257202,
"learning_rate": 1.681247146056654e-05,
"loss": 0.2606,
"num_input_tokens_seen": 6460688,
"step": 200
},
{
"epoch": 3.958435207823961,
"grad_norm": 1.2710639238357544,
"learning_rate": 1.6660731316772503e-05,
"loss": 0.2543,
"num_input_tokens_seen": 6624528,
"step": 205
},
{
"epoch": 4.039119804400978,
"grad_norm": 1.284611463546753,
"learning_rate": 1.650618300204242e-05,
"loss": 0.1856,
"num_input_tokens_seen": 6757128,
"step": 210
},
{
"epoch": 4.136919315403423,
"grad_norm": 1.6113184690475464,
"learning_rate": 1.634889167409923e-05,
"loss": 0.1462,
"num_input_tokens_seen": 6920968,
"step": 215
},
{
"epoch": 4.234718826405868,
"grad_norm": 1.230480432510376,
"learning_rate": 1.6188923647122946e-05,
"loss": 0.17,
"num_input_tokens_seen": 7084808,
"step": 220
},
{
"epoch": 4.332518337408313,
"grad_norm": 1.388247013092041,
"learning_rate": 1.6026346363792565e-05,
"loss": 0.1427,
"num_input_tokens_seen": 7248648,
"step": 225
},
{
"epoch": 4.430317848410758,
"grad_norm": 1.3794785737991333,
"learning_rate": 1.5861228366852148e-05,
"loss": 0.15,
"num_input_tokens_seen": 7412488,
"step": 230
},
{
"epoch": 4.528117359413203,
"grad_norm": 1.5547810792922974,
"learning_rate": 1.5693639270213138e-05,
"loss": 0.1603,
"num_input_tokens_seen": 7576264,
"step": 235
},
{
"epoch": 4.625916870415648,
"grad_norm": 1.49036705493927,
"learning_rate": 1.552364972960506e-05,
"loss": 0.1536,
"num_input_tokens_seen": 7740104,
"step": 240
},
{
"epoch": 4.723716381418093,
"grad_norm": 1.2922449111938477,
"learning_rate": 1.5351331412787004e-05,
"loss": 0.1415,
"num_input_tokens_seen": 7903944,
"step": 245
},
{
"epoch": 4.821515892420538,
"grad_norm": 1.5458725690841675,
"learning_rate": 1.5176756969332428e-05,
"loss": 0.159,
"num_input_tokens_seen": 8067784,
"step": 250
},
{
"epoch": 4.919315403422983,
"grad_norm": 1.3060288429260254,
"learning_rate": 1.5000000000000002e-05,
"loss": 0.1699,
"num_input_tokens_seen": 8231624,
"step": 255
},
{
"epoch": 5.0,
"grad_norm": 4.3814568519592285,
"learning_rate": 1.4821135025703491e-05,
"loss": 0.1295,
"num_input_tokens_seen": 8364744,
"step": 260
},
{
"epoch": 5.097799511002445,
"grad_norm": 1.496863842010498,
"learning_rate": 1.4640237456093636e-05,
"loss": 0.1019,
"num_input_tokens_seen": 8528584,
"step": 265
},
{
"epoch": 5.19559902200489,
"grad_norm": 1.3016966581344604,
"learning_rate": 1.4457383557765385e-05,
"loss": 0.0916,
"num_input_tokens_seen": 8692424,
"step": 270
},
{
"epoch": 5.293398533007335,
"grad_norm": 1.3272321224212646,
"learning_rate": 1.427265042210381e-05,
"loss": 0.0935,
"num_input_tokens_seen": 8856264,
"step": 275
},
{
"epoch": 5.39119804400978,
"grad_norm": 1.2260195016860962,
"learning_rate": 1.4086115932782316e-05,
"loss": 0.0679,
"num_input_tokens_seen": 9020104,
"step": 280
},
{
"epoch": 5.488997555012225,
"grad_norm": 1.1941609382629395,
"learning_rate": 1.3897858732926794e-05,
"loss": 0.101,
"num_input_tokens_seen": 9183944,
"step": 285
},
{
"epoch": 5.58679706601467,
"grad_norm": 1.3348464965820312,
"learning_rate": 1.3707958191959609e-05,
"loss": 0.0802,
"num_input_tokens_seen": 9347784,
"step": 290
},
{
"epoch": 5.684596577017115,
"grad_norm": 1.3610597848892212,
"learning_rate": 1.3516494372137368e-05,
"loss": 0.0812,
"num_input_tokens_seen": 9511624,
"step": 295
},
{
"epoch": 5.78239608801956,
"grad_norm": 1.1276705265045166,
"learning_rate": 1.3323547994796597e-05,
"loss": 0.0812,
"num_input_tokens_seen": 9675464,
"step": 300
},
{
"epoch": 5.880195599022005,
"grad_norm": 1.4052499532699585,
"learning_rate": 1.3129200406321545e-05,
"loss": 0.0884,
"num_input_tokens_seen": 9839304,
"step": 305
},
{
"epoch": 5.97799511002445,
"grad_norm": 1.2434982061386108,
"learning_rate": 1.2933533543848462e-05,
"loss": 0.0906,
"num_input_tokens_seen": 10003144,
"step": 310
},
{
"epoch": 6.058679706601467,
"grad_norm": 1.19767427444458,
"learning_rate": 1.2736629900720832e-05,
"loss": 0.0616,
"num_input_tokens_seen": 10136264,
"step": 315
},
{
"epoch": 6.156479217603912,
"grad_norm": 1.2783230543136597,
"learning_rate": 1.2538572491710079e-05,
"loss": 0.0447,
"num_input_tokens_seen": 10300104,
"step": 320
},
{
"epoch": 6.254278728606357,
"grad_norm": 1.1590359210968018,
"learning_rate": 1.2339444818016488e-05,
"loss": 0.0528,
"num_input_tokens_seen": 10463944,
"step": 325
},
{
"epoch": 6.352078239608802,
"grad_norm": 1.2891935110092163,
"learning_rate": 1.2139330832064975e-05,
"loss": 0.0429,
"num_input_tokens_seen": 10627784,
"step": 330
},
{
"epoch": 6.449877750611247,
"grad_norm": 1.0882583856582642,
"learning_rate": 1.1938314902110701e-05,
"loss": 0.0442,
"num_input_tokens_seen": 10791624,
"step": 335
},
{
"epoch": 6.547677261613692,
"grad_norm": 1.1069742441177368,
"learning_rate": 1.1736481776669307e-05,
"loss": 0.047,
"num_input_tokens_seen": 10955464,
"step": 340
},
{
"epoch": 6.645476772616137,
"grad_norm": 1.0986578464508057,
"learning_rate": 1.1533916548786856e-05,
"loss": 0.0437,
"num_input_tokens_seen": 11119304,
"step": 345
},
{
"epoch": 6.743276283618582,
"grad_norm": 1.1824274063110352,
"learning_rate": 1.133070462016454e-05,
"loss": 0.0466,
"num_input_tokens_seen": 11282568,
"step": 350
},
{
"epoch": 6.841075794621027,
"grad_norm": 1.197988510131836,
"learning_rate": 1.1126931665153213e-05,
"loss": 0.0498,
"num_input_tokens_seen": 11446408,
"step": 355
},
{
"epoch": 6.938875305623472,
"grad_norm": 1.1721147298812866,
"learning_rate": 1.092268359463302e-05,
"loss": 0.0409,
"num_input_tokens_seen": 11610248,
"step": 360
},
{
"epoch": 7.019559902200489,
"grad_norm": 0.9280807971954346,
"learning_rate": 1.0718046519793276e-05,
"loss": 0.0444,
"num_input_tokens_seen": 11743368,
"step": 365
},
{
"epoch": 7.117359413202934,
"grad_norm": 0.8121551871299744,
"learning_rate": 1.0513106715827897e-05,
"loss": 0.0212,
"num_input_tokens_seen": 11906632,
"step": 370
},
{
"epoch": 7.215158924205379,
"grad_norm": 0.974238932132721,
"learning_rate": 1.0307950585561705e-05,
"loss": 0.0275,
"num_input_tokens_seen": 12070472,
"step": 375
},
{
"epoch": 7.312958435207824,
"grad_norm": 0.8143863081932068,
"learning_rate": 1.01026646230229e-05,
"loss": 0.0234,
"num_input_tokens_seen": 12234312,
"step": 380
},
{
"epoch": 7.410757946210269,
"grad_norm": 0.9007371664047241,
"learning_rate": 9.897335376977104e-06,
"loss": 0.0227,
"num_input_tokens_seen": 12398152,
"step": 385
},
{
"epoch": 7.508557457212714,
"grad_norm": 0.8347233533859253,
"learning_rate": 9.692049414438298e-06,
"loss": 0.0264,
"num_input_tokens_seen": 12561992,
"step": 390
},
{
"epoch": 7.606356968215159,
"grad_norm": 1.0170607566833496,
"learning_rate": 9.486893284172103e-06,
"loss": 0.0251,
"num_input_tokens_seen": 12725832,
"step": 395
},
{
"epoch": 7.704156479217604,
"grad_norm": 0.8994267582893372,
"learning_rate": 9.281953480206725e-06,
"loss": 0.0237,
"num_input_tokens_seen": 12889672,
"step": 400
},
{
"epoch": 7.801955990220049,
"grad_norm": 1.0272514820098877,
"learning_rate": 9.07731640536698e-06,
"loss": 0.0228,
"num_input_tokens_seen": 13053512,
"step": 405
},
{
"epoch": 7.899755501222494,
"grad_norm": 0.8432409167289734,
"learning_rate": 8.87306833484679e-06,
"loss": 0.0238,
"num_input_tokens_seen": 13217352,
"step": 410
},
{
"epoch": 7.997555012224939,
"grad_norm": 0.9295158386230469,
"learning_rate": 8.669295379835467e-06,
"loss": 0.0222,
"num_input_tokens_seen": 13381192,
"step": 415
},
{
"epoch": 8.078239608801956,
"grad_norm": 0.6758769750595093,
"learning_rate": 8.466083451213145e-06,
"loss": 0.0109,
"num_input_tokens_seen": 13514312,
"step": 420
},
{
"epoch": 8.1760391198044,
"grad_norm": 0.5933005213737488,
"learning_rate": 8.263518223330698e-06,
"loss": 0.0118,
"num_input_tokens_seen": 13677448,
"step": 425
},
{
"epoch": 8.273838630806846,
"grad_norm": 0.7031515836715698,
"learning_rate": 8.0616850978893e-06,
"loss": 0.01,
"num_input_tokens_seen": 13841288,
"step": 430
},
{
"epoch": 8.37163814180929,
"grad_norm": 0.8007884621620178,
"learning_rate": 7.860669167935028e-06,
"loss": 0.0112,
"num_input_tokens_seen": 14005128,
"step": 435
},
{
"epoch": 8.469437652811736,
"grad_norm": 0.6427900195121765,
"learning_rate": 7.660555181983517e-06,
"loss": 0.0133,
"num_input_tokens_seen": 14168968,
"step": 440
},
{
"epoch": 8.56723716381418,
"grad_norm": 0.6638085246086121,
"learning_rate": 7.461427508289922e-06,
"loss": 0.0113,
"num_input_tokens_seen": 14332808,
"step": 445
},
{
"epoch": 8.665036674816626,
"grad_norm": 0.6730697751045227,
"learning_rate": 7.263370099279173e-06,
"loss": 0.0102,
"num_input_tokens_seen": 14496648,
"step": 450
},
{
"epoch": 8.76283618581907,
"grad_norm": 0.7171938419342041,
"learning_rate": 7.066466456151541e-06,
"loss": 0.01,
"num_input_tokens_seen": 14660488,
"step": 455
},
{
"epoch": 8.860635696821516,
"grad_norm": 0.7203698754310608,
"learning_rate": 6.870799593678459e-06,
"loss": 0.0116,
"num_input_tokens_seen": 14824328,
"step": 460
},
{
"epoch": 8.95843520782396,
"grad_norm": 0.7533681392669678,
"learning_rate": 6.6764520052034054e-06,
"loss": 0.0111,
"num_input_tokens_seen": 14988168,
"step": 465
},
{
"epoch": 9.039119804400977,
"grad_norm": 0.22926057875156403,
"learning_rate": 6.483505627862632e-06,
"loss": 0.0071,
"num_input_tokens_seen": 15121288,
"step": 470
},
{
"epoch": 9.136919315403423,
"grad_norm": 0.4467085301876068,
"learning_rate": 6.292041808040393e-06,
"loss": 0.0054,
"num_input_tokens_seen": 15285128,
"step": 475
},
{
"epoch": 9.234718826405867,
"grad_norm": 0.5730965733528137,
"learning_rate": 6.102141267073207e-06,
"loss": 0.0059,
"num_input_tokens_seen": 15448968,
"step": 480
},
{
"epoch": 9.332518337408313,
"grad_norm": 0.3801887333393097,
"learning_rate": 5.913884067217686e-06,
"loss": 0.0045,
"num_input_tokens_seen": 15612552,
"step": 485
},
{
"epoch": 9.430317848410757,
"grad_norm": 0.4900866448879242,
"learning_rate": 5.727349577896194e-06,
"loss": 0.004,
"num_input_tokens_seen": 15776392,
"step": 490
},
{
"epoch": 9.528117359413203,
"grad_norm": 0.3590303957462311,
"learning_rate": 5.542616442234618e-06,
"loss": 0.0042,
"num_input_tokens_seen": 15940232,
"step": 495
},
{
"epoch": 9.625916870415647,
"grad_norm": 0.36999812722206116,
"learning_rate": 5.3597625439063685e-06,
"loss": 0.0049,
"num_input_tokens_seen": 16104072,
"step": 500
},
{
"epoch": 9.723716381418093,
"grad_norm": 0.3589613735675812,
"learning_rate": 5.178864974296511e-06,
"loss": 0.0045,
"num_input_tokens_seen": 16267912,
"step": 505
},
{
"epoch": 9.821515892420537,
"grad_norm": 0.30231136083602905,
"learning_rate": 5.000000000000003e-06,
"loss": 0.004,
"num_input_tokens_seen": 16431752,
"step": 510
},
{
"epoch": 9.919315403422983,
"grad_norm": 0.34575000405311584,
"learning_rate": 4.823243030667576e-06,
"loss": 0.0042,
"num_input_tokens_seen": 16595592,
"step": 515
},
{
"epoch": 10.0,
"grad_norm": 0.6714850068092346,
"learning_rate": 4.648668587212998e-06,
"loss": 0.0037,
"num_input_tokens_seen": 16728712,
"step": 520
},
{
"epoch": 10.097799511002446,
"grad_norm": 0.19611337780952454,
"learning_rate": 4.476350270394942e-06,
"loss": 0.0021,
"num_input_tokens_seen": 16892552,
"step": 525
},
{
"epoch": 10.19559902200489,
"grad_norm": 0.13641950488090515,
"learning_rate": 4.306360729786867e-06,
"loss": 0.002,
"num_input_tokens_seen": 17056392,
"step": 530
},
{
"epoch": 10.293398533007334,
"grad_norm": 0.16702738404273987,
"learning_rate": 4.138771633147856e-06,
"loss": 0.0023,
"num_input_tokens_seen": 17220232,
"step": 535
},
{
"epoch": 10.39119804400978,
"grad_norm": 0.2278226763010025,
"learning_rate": 3.973653636207437e-06,
"loss": 0.002,
"num_input_tokens_seen": 17384072,
"step": 540
},
{
"epoch": 10.488997555012224,
"grad_norm": 0.1814439296722412,
"learning_rate": 3.8110763528770543e-06,
"loss": 0.0021,
"num_input_tokens_seen": 17547912,
"step": 545
},
{
"epoch": 10.58679706601467,
"grad_norm": 0.2528248429298401,
"learning_rate": 3.651108325900773e-06,
"loss": 0.0025,
"num_input_tokens_seen": 17711752,
"step": 550
},
{
"epoch": 10.684596577017114,
"grad_norm": 0.165152445435524,
"learning_rate": 3.493816997957582e-06,
"loss": 0.0021,
"num_input_tokens_seen": 17875592,
"step": 555
},
{
"epoch": 10.78239608801956,
"grad_norm": 0.08041153103113174,
"learning_rate": 3.339268683227499e-06,
"loss": 0.0017,
"num_input_tokens_seen": 18039432,
"step": 560
},
{
"epoch": 10.880195599022004,
"grad_norm": 0.24567244946956635,
"learning_rate": 3.1875285394334575e-06,
"loss": 0.0018,
"num_input_tokens_seen": 18203272,
"step": 565
},
{
"epoch": 10.97799511002445,
"grad_norm": 0.21529506146907806,
"learning_rate": 3.0386605403707347e-06,
"loss": 0.0018,
"num_input_tokens_seen": 18367112,
"step": 570
},
{
"epoch": 11.058679706601467,
"grad_norm": 0.05531800910830498,
"learning_rate": 2.8927274489355296e-06,
"loss": 0.0014,
"num_input_tokens_seen": 18500232,
"step": 575
},
{
"epoch": 11.156479217603913,
"grad_norm": 0.04145563766360283,
"learning_rate": 2.749790790664074e-06,
"loss": 0.0013,
"num_input_tokens_seen": 18662984,
"step": 580
},
{
"epoch": 11.254278728606357,
"grad_norm": 0.07673907279968262,
"learning_rate": 2.6099108277934105e-06,
"loss": 0.0014,
"num_input_tokens_seen": 18826824,
"step": 585
},
{
"epoch": 11.352078239608803,
"grad_norm": 0.054488956928253174,
"learning_rate": 2.4731465338547556e-06,
"loss": 0.0013,
"num_input_tokens_seen": 18990664,
"step": 590
},
{
"epoch": 11.449877750611247,
"grad_norm": 0.11207219213247299,
"learning_rate": 2.339555568810221e-06,
"loss": 0.0013,
"num_input_tokens_seen": 19154504,
"step": 595
},
{
"epoch": 11.547677261613693,
"grad_norm": 0.062526635825634,
"learning_rate": 2.209194254743295e-06,
"loss": 0.0012,
"num_input_tokens_seen": 19318344,
"step": 600
},
{
"epoch": 11.645476772616137,
"grad_norm": 0.05114143341779709,
"learning_rate": 2.0821175521134208e-06,
"loss": 0.0013,
"num_input_tokens_seen": 19482184,
"step": 605
},
{
"epoch": 11.743276283618583,
"grad_norm": 0.05199455842375755,
"learning_rate": 1.9583790365845823e-06,
"loss": 0.0013,
"num_input_tokens_seen": 19646024,
"step": 610
},
{
"epoch": 11.841075794621027,
"grad_norm": 0.057281751185655594,
"learning_rate": 1.8380308764377841e-06,
"loss": 0.0014,
"num_input_tokens_seen": 19809864,
"step": 615
},
{
"epoch": 11.938875305623473,
"grad_norm": 0.04751597344875336,
"learning_rate": 1.7211238105768213e-06,
"loss": 0.0012,
"num_input_tokens_seen": 19973704,
"step": 620
},
{
"epoch": 12.01955990220049,
"grad_norm": 0.04285968840122223,
"learning_rate": 1.607707127136734e-06,
"loss": 0.0014,
"num_input_tokens_seen": 20106824,
"step": 625
},
{
"epoch": 12.117359413202934,
"grad_norm": 0.044368330389261246,
"learning_rate": 1.4978286427038602e-06,
"loss": 0.0011,
"num_input_tokens_seen": 20270664,
"step": 630
},
{
"epoch": 12.21515892420538,
"grad_norm": 0.04820827767252922,
"learning_rate": 1.3915346821563235e-06,
"loss": 0.0012,
"num_input_tokens_seen": 20434504,
"step": 635
},
{
"epoch": 12.312958435207824,
"grad_norm": 0.034025732427835464,
"learning_rate": 1.2888700591334225e-06,
"loss": 0.0011,
"num_input_tokens_seen": 20598280,
"step": 640
},
{
"epoch": 12.41075794621027,
"grad_norm": 0.03853330388665199,
"learning_rate": 1.1898780571421554e-06,
"loss": 0.0011,
"num_input_tokens_seen": 20762120,
"step": 645
},
{
"epoch": 12.508557457212714,
"grad_norm": 0.039888255298137665,
"learning_rate": 1.0946004113088381e-06,
"loss": 0.0012,
"num_input_tokens_seen": 20925960,
"step": 650
},
{
"epoch": 12.60635696821516,
"grad_norm": 0.04366978630423546,
"learning_rate": 1.0030772907835484e-06,
"loss": 0.001,
"num_input_tokens_seen": 21089800,
"step": 655
},
{
"epoch": 12.704156479217604,
"grad_norm": 0.03416445106267929,
"learning_rate": 9.153472818047627e-07,
"loss": 0.001,
"num_input_tokens_seen": 21253640,
"step": 660
},
{
"epoch": 12.80195599022005,
"grad_norm": 0.04817335307598114,
"learning_rate": 8.31447371431372e-07,
"loss": 0.0011,
"num_input_tokens_seen": 21417480,
"step": 665
},
{
"epoch": 12.899755501222494,
"grad_norm": 0.03280309960246086,
"learning_rate": 7.514129319488839e-07,
"loss": 0.0011,
"num_input_tokens_seen": 21581320,
"step": 670
},
{
"epoch": 12.99755501222494,
"grad_norm": 0.04426475614309311,
"learning_rate": 6.752777059564431e-07,
"loss": 0.0011,
"num_input_tokens_seen": 21745160,
"step": 675
},
{
"epoch": 13.078239608801956,
"grad_norm": 0.03248964622616768,
"learning_rate": 6.030737921409169e-07,
"loss": 0.0011,
"num_input_tokens_seen": 21877864,
"step": 680
},
{
"epoch": 13.1760391198044,
"grad_norm": 0.0347786545753479,
"learning_rate": 5.348316317440549e-07,
"loss": 0.001,
"num_input_tokens_seen": 22041704,
"step": 685
},
{
"epoch": 13.273838630806846,
"grad_norm": 0.04258348420262337,
"learning_rate": 4.7057999572843516e-07,
"loss": 0.001,
"num_input_tokens_seen": 22205544,
"step": 690
},
{
"epoch": 13.37163814180929,
"grad_norm": 0.0364152230322361,
"learning_rate": 4.103459726475889e-07,
"loss": 0.0011,
"num_input_tokens_seen": 22369384,
"step": 695
},
{
"epoch": 13.469437652811736,
"grad_norm": 0.039297617971897125,
"learning_rate": 3.541549572254488e-07,
"loss": 0.0011,
"num_input_tokens_seen": 22533224,
"step": 700
},
{
"epoch": 13.56723716381418,
"grad_norm": 0.037363260984420776,
"learning_rate": 3.020306396499062e-07,
"loss": 0.0009,
"num_input_tokens_seen": 22697064,
"step": 705
},
{
"epoch": 13.665036674816626,
"grad_norm": 0.032989222556352615,
"learning_rate": 2.539949955849985e-07,
"loss": 0.0011,
"num_input_tokens_seen": 22860904,
"step": 710
},
{
"epoch": 13.76283618581907,
"grad_norm": 0.03354249894618988,
"learning_rate": 2.1006827690595478e-07,
"loss": 0.001,
"num_input_tokens_seen": 23024744,
"step": 715
},
{
"epoch": 13.860635696821516,
"grad_norm": 0.03322712704539299,
"learning_rate": 1.7026900316098217e-07,
"loss": 0.001,
"num_input_tokens_seen": 23188584,
"step": 720
},
{
"epoch": 13.95843520782396,
"grad_norm": 0.03775469958782196,
"learning_rate": 1.3461395376340502e-07,
"loss": 0.001,
"num_input_tokens_seen": 23352424,
"step": 725
},
{
"epoch": 14.039119804400977,
"grad_norm": 0.039018385112285614,
"learning_rate": 1.0311816091744698e-07,
"loss": 0.001,
"num_input_tokens_seen": 23485296,
"step": 730
},
{
"epoch": 14.136919315403423,
"grad_norm": 0.03440910577774048,
"learning_rate": 7.579490328064265e-08,
"loss": 0.001,
"num_input_tokens_seen": 23649136,
"step": 735
},
{
"epoch": 14.234718826405867,
"grad_norm": 0.03226885199546814,
"learning_rate": 5.265570036553813e-08,
"loss": 0.001,
"num_input_tokens_seen": 23812976,
"step": 740
},
{
"epoch": 14.332518337408313,
"grad_norm": 0.03353099152445793,
"learning_rate": 3.371030768305583e-08,
"loss": 0.0009,
"num_input_tokens_seen": 23976816,
"step": 745
},
{
"epoch": 14.430317848410757,
"grad_norm": 0.04043235257267952,
"learning_rate": 1.896671262955896e-08,
"loss": 0.001,
"num_input_tokens_seen": 24140656,
"step": 750
},
{
"epoch": 14.528117359413203,
"grad_norm": 0.0331251285970211,
"learning_rate": 8.431131119361891e-09,
"loss": 0.001,
"num_input_tokens_seen": 24304496,
"step": 755
},
{
"epoch": 14.625916870415647,
"grad_norm": 0.03793644160032272,
"learning_rate": 2.108004964086474e-09,
"loss": 0.001,
"num_input_tokens_seen": 24467760,
"step": 760
},
{
"epoch": 14.723716381418093,
"grad_norm": 0.03213175758719444,
"learning_rate": 0.0,
"loss": 0.001,
"num_input_tokens_seen": 24631600,
"step": 765
},
{
"epoch": 14.723716381418093,
"num_input_tokens_seen": 24631600,
"step": 765,
"total_flos": 1.936547143200768e+17,
"train_loss": 0.1555636612302776,
"train_runtime": 3426.6896,
"train_samples_per_second": 3.576,
"train_steps_per_second": 0.223
}
],
"logging_steps": 5,
"max_steps": 765,
"num_input_tokens_seen": 24631600,
"num_train_epochs": 15,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.936547143200768e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}