afull05's picture
Training in progress, step 1000, checkpoint
08f430e verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.031221096094631143,
"eval_steps": 334,
"global_step": 1000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 3.1221096094631145e-05,
"eval_loss": 2.5151805877685547,
"eval_runtime": 12858.7571,
"eval_samples_per_second": 2.214,
"eval_steps_per_second": 2.214,
"step": 1
},
{
"epoch": 0.0003122109609463114,
"grad_norm": 1.6301891803741455,
"learning_rate": 0.0002,
"loss": 1.2603,
"step": 10
},
{
"epoch": 0.0006244219218926228,
"grad_norm": 2.2189297676086426,
"learning_rate": 0.00019994965423831854,
"loss": 0.8201,
"step": 20
},
{
"epoch": 0.0009366328828389343,
"grad_norm": 1.6551645994186401,
"learning_rate": 0.00019979866764718843,
"loss": 0.7913,
"step": 30
},
{
"epoch": 0.0012488438437852456,
"grad_norm": 4.437395095825195,
"learning_rate": 0.00019954719225730847,
"loss": 1.2831,
"step": 40
},
{
"epoch": 0.001561054804731557,
"grad_norm": 7.653807163238525,
"learning_rate": 0.00019919548128307954,
"loss": 1.6268,
"step": 50
},
{
"epoch": 0.0018732657656778686,
"grad_norm": 1.2926839590072632,
"learning_rate": 0.00019874388886763944,
"loss": 1.1779,
"step": 60
},
{
"epoch": 0.00218547672662418,
"grad_norm": 1.5756969451904297,
"learning_rate": 0.00019819286972627066,
"loss": 0.6204,
"step": 70
},
{
"epoch": 0.002497687687570491,
"grad_norm": 1.8183478116989136,
"learning_rate": 0.00019754297868854073,
"loss": 0.8666,
"step": 80
},
{
"epoch": 0.0028098986485168027,
"grad_norm": 4.023663520812988,
"learning_rate": 0.00019679487013963564,
"loss": 1.1012,
"step": 90
},
{
"epoch": 0.003122109609463114,
"grad_norm": 3.675027847290039,
"learning_rate": 0.00019594929736144976,
"loss": 1.3794,
"step": 100
},
{
"epoch": 0.0034343205704094257,
"grad_norm": 1.1626209020614624,
"learning_rate": 0.00019500711177409454,
"loss": 1.1787,
"step": 110
},
{
"epoch": 0.003746531531355737,
"grad_norm": 1.4152737855911255,
"learning_rate": 0.00019396926207859084,
"loss": 0.471,
"step": 120
},
{
"epoch": 0.004058742492302049,
"grad_norm": 2.0098798274993896,
"learning_rate": 0.00019283679330160726,
"loss": 0.9037,
"step": 130
},
{
"epoch": 0.00437095345324836,
"grad_norm": 2.845428943634033,
"learning_rate": 0.00019161084574320696,
"loss": 1.2114,
"step": 140
},
{
"epoch": 0.004683164414194672,
"grad_norm": 5.3034749031066895,
"learning_rate": 0.00019029265382866214,
"loss": 1.397,
"step": 150
},
{
"epoch": 0.004995375375140982,
"grad_norm": 1.5945558547973633,
"learning_rate": 0.00018888354486549237,
"loss": 1.145,
"step": 160
},
{
"epoch": 0.005307586336087294,
"grad_norm": 1.1813420057296753,
"learning_rate": 0.00018738493770697852,
"loss": 0.5603,
"step": 170
},
{
"epoch": 0.005619797297033605,
"grad_norm": 2.35960054397583,
"learning_rate": 0.00018579834132349772,
"loss": 0.8248,
"step": 180
},
{
"epoch": 0.005932008257979917,
"grad_norm": 2.443915605545044,
"learning_rate": 0.00018412535328311814,
"loss": 1.1141,
"step": 190
},
{
"epoch": 0.006244219218926228,
"grad_norm": 5.689703941345215,
"learning_rate": 0.0001823676581429833,
"loss": 1.2009,
"step": 200
},
{
"epoch": 0.00655643017987254,
"grad_norm": 1.4314906597137451,
"learning_rate": 0.00018052702575310588,
"loss": 1.1061,
"step": 210
},
{
"epoch": 0.006868641140818851,
"grad_norm": 0.6448104977607727,
"learning_rate": 0.00017860530947427875,
"loss": 0.4016,
"step": 220
},
{
"epoch": 0.007180852101765163,
"grad_norm": 2.0396196842193604,
"learning_rate": 0.0001766044443118978,
"loss": 0.8709,
"step": 230
},
{
"epoch": 0.007493063062711474,
"grad_norm": 2.5875227451324463,
"learning_rate": 0.0001745264449675755,
"loss": 1.1121,
"step": 240
},
{
"epoch": 0.007805274023657786,
"grad_norm": 3.9609525203704834,
"learning_rate": 0.00017237340381050703,
"loss": 1.251,
"step": 250
},
{
"epoch": 0.008117484984604097,
"grad_norm": 1.2032607793807983,
"learning_rate": 0.00017014748877063214,
"loss": 1.1823,
"step": 260
},
{
"epoch": 0.008429695945550408,
"grad_norm": 1.186848521232605,
"learning_rate": 0.00016785094115571322,
"loss": 0.6219,
"step": 270
},
{
"epoch": 0.00874190690649672,
"grad_norm": 1.85453462600708,
"learning_rate": 0.00016548607339452853,
"loss": 0.5809,
"step": 280
},
{
"epoch": 0.009054117867443031,
"grad_norm": 2.0443332195281982,
"learning_rate": 0.00016305526670845226,
"loss": 1.2146,
"step": 290
},
{
"epoch": 0.009366328828389343,
"grad_norm": 7.1448516845703125,
"learning_rate": 0.00016056096871376667,
"loss": 1.2524,
"step": 300
},
{
"epoch": 0.009678539789335654,
"grad_norm": 1.334848165512085,
"learning_rate": 0.00015800569095711982,
"loss": 1.1966,
"step": 310
},
{
"epoch": 0.009990750750281965,
"grad_norm": 0.4558267295360565,
"learning_rate": 0.00015539200638661104,
"loss": 0.589,
"step": 320
},
{
"epoch": 0.010302961711228277,
"grad_norm": 1.8344190120697021,
"learning_rate": 0.00015272254676105025,
"loss": 0.5806,
"step": 330
},
{
"epoch": 0.010427846095606801,
"eval_loss": 0.9566133618354797,
"eval_runtime": 13592.0734,
"eval_samples_per_second": 2.095,
"eval_steps_per_second": 2.095,
"step": 334
},
{
"epoch": 0.010615172672174588,
"grad_norm": 1.9632649421691895,
"learning_rate": 0.00015000000000000001,
"loss": 1.0551,
"step": 340
},
{
"epoch": 0.0109273836331209,
"grad_norm": 4.136826992034912,
"learning_rate": 0.0001472271074772683,
"loss": 1.0784,
"step": 350
},
{
"epoch": 0.01123959459406721,
"grad_norm": 1.1779104471206665,
"learning_rate": 0.00014440666126057744,
"loss": 1.1613,
"step": 360
},
{
"epoch": 0.011551805555013523,
"grad_norm": 0.8325644731521606,
"learning_rate": 0.00014154150130018866,
"loss": 0.5119,
"step": 370
},
{
"epoch": 0.011864016515959834,
"grad_norm": 1.6711801290512085,
"learning_rate": 0.00013863451256931287,
"loss": 0.6156,
"step": 380
},
{
"epoch": 0.012176227476906146,
"grad_norm": 2.293975353240967,
"learning_rate": 0.00013568862215918717,
"loss": 1.0706,
"step": 390
},
{
"epoch": 0.012488438437852457,
"grad_norm": 2.2785656452178955,
"learning_rate": 0.00013270679633174218,
"loss": 1.2872,
"step": 400
},
{
"epoch": 0.012800649398798769,
"grad_norm": 1.2502048015594482,
"learning_rate": 0.0001296920375328275,
"loss": 1.0768,
"step": 410
},
{
"epoch": 0.01311286035974508,
"grad_norm": 0.7812928557395935,
"learning_rate": 0.00012664738136900348,
"loss": 0.5199,
"step": 420
},
{
"epoch": 0.01342507132069139,
"grad_norm": 2.0176918506622314,
"learning_rate": 0.00012357589355094275,
"loss": 0.8125,
"step": 430
},
{
"epoch": 0.013737282281637703,
"grad_norm": 2.014697313308716,
"learning_rate": 0.00012048066680651908,
"loss": 1.0261,
"step": 440
},
{
"epoch": 0.014049493242584013,
"grad_norm": 3.0161404609680176,
"learning_rate": 0.00011736481776669306,
"loss": 1.1352,
"step": 450
},
{
"epoch": 0.014361704203530326,
"grad_norm": 1.1186920404434204,
"learning_rate": 0.00011423148382732853,
"loss": 1.1374,
"step": 460
},
{
"epoch": 0.014673915164476636,
"grad_norm": 0.9820886850357056,
"learning_rate": 0.00011108381999010111,
"loss": 0.5135,
"step": 470
},
{
"epoch": 0.014986126125422949,
"grad_norm": 2.8473262786865234,
"learning_rate": 0.00010792499568567884,
"loss": 0.8812,
"step": 480
},
{
"epoch": 0.01529833708636926,
"grad_norm": 2.1481053829193115,
"learning_rate": 0.00010475819158237425,
"loss": 1.0178,
"step": 490
},
{
"epoch": 0.015610548047315572,
"grad_norm": 1.20015287399292,
"learning_rate": 0.00010158659638348081,
"loss": 1.0468,
"step": 500
},
{
"epoch": 0.015922759008261882,
"grad_norm": 1.3715260028839111,
"learning_rate": 9.84134036165192e-05,
"loss": 1.0121,
"step": 510
},
{
"epoch": 0.016234969969208195,
"grad_norm": 0.9032047390937805,
"learning_rate": 9.524180841762577e-05,
"loss": 0.5379,
"step": 520
},
{
"epoch": 0.016547180930154507,
"grad_norm": 1.6322988271713257,
"learning_rate": 9.207500431432115e-05,
"loss": 0.5695,
"step": 530
},
{
"epoch": 0.016859391891100816,
"grad_norm": 2.7707173824310303,
"learning_rate": 8.891618000989891e-05,
"loss": 0.9264,
"step": 540
},
{
"epoch": 0.01717160285204713,
"grad_norm": 2.905618667602539,
"learning_rate": 8.57685161726715e-05,
"loss": 1.0027,
"step": 550
},
{
"epoch": 0.01748381381299344,
"grad_norm": 1.0891023874282837,
"learning_rate": 8.263518223330697e-05,
"loss": 1.0168,
"step": 560
},
{
"epoch": 0.01779602477393975,
"grad_norm": 1.0451884269714355,
"learning_rate": 7.951933319348095e-05,
"loss": 0.3617,
"step": 570
},
{
"epoch": 0.018108235734886062,
"grad_norm": 1.5976777076721191,
"learning_rate": 7.642410644905726e-05,
"loss": 0.7168,
"step": 580
},
{
"epoch": 0.018420446695832374,
"grad_norm": 2.3960399627685547,
"learning_rate": 7.335261863099651e-05,
"loss": 0.9692,
"step": 590
},
{
"epoch": 0.018732657656778687,
"grad_norm": 2.764636516571045,
"learning_rate": 7.030796246717255e-05,
"loss": 1.0993,
"step": 600
},
{
"epoch": 0.019044868617724996,
"grad_norm": 1.0002310276031494,
"learning_rate": 6.729320366825784e-05,
"loss": 1.1104,
"step": 610
},
{
"epoch": 0.019357079578671308,
"grad_norm": 1.2536269426345825,
"learning_rate": 6.431137784081282e-05,
"loss": 0.5409,
"step": 620
},
{
"epoch": 0.01966929053961762,
"grad_norm": 1.9996048212051392,
"learning_rate": 6.136548743068713e-05,
"loss": 0.7138,
"step": 630
},
{
"epoch": 0.01998150150056393,
"grad_norm": 1.9154224395751953,
"learning_rate": 5.845849869981137e-05,
"loss": 0.9189,
"step": 640
},
{
"epoch": 0.02029371246151024,
"grad_norm": 0.8520050644874573,
"learning_rate": 5.559333873942259e-05,
"loss": 1.1057,
"step": 650
},
{
"epoch": 0.020605923422456554,
"grad_norm": 0.821565568447113,
"learning_rate": 5.277289252273174e-05,
"loss": 0.9571,
"step": 660
},
{
"epoch": 0.020855692191213603,
"eval_loss": 0.8139573931694031,
"eval_runtime": 11568.1075,
"eval_samples_per_second": 2.461,
"eval_steps_per_second": 2.461,
"step": 668
},
{
"epoch": 0.020918134383402866,
"grad_norm": 1.3406473398208618,
"learning_rate": 5.000000000000002e-05,
"loss": 0.4458,
"step": 670
},
{
"epoch": 0.021230345344349175,
"grad_norm": 1.6482555866241455,
"learning_rate": 4.727745323894976e-05,
"loss": 0.8605,
"step": 680
},
{
"epoch": 0.021542556305295488,
"grad_norm": 2.3034768104553223,
"learning_rate": 4.4607993613388976e-05,
"loss": 0.8842,
"step": 690
},
{
"epoch": 0.0218547672662418,
"grad_norm": 2.2205710411071777,
"learning_rate": 4.19943090428802e-05,
"loss": 1.0044,
"step": 700
},
{
"epoch": 0.022166978227188112,
"grad_norm": 1.0776264667510986,
"learning_rate": 3.943903128623335e-05,
"loss": 0.9987,
"step": 710
},
{
"epoch": 0.02247918918813442,
"grad_norm": 0.7974869012832642,
"learning_rate": 3.694473329154778e-05,
"loss": 0.4341,
"step": 720
},
{
"epoch": 0.022791400149080734,
"grad_norm": 1.410180926322937,
"learning_rate": 3.45139266054715e-05,
"loss": 0.7361,
"step": 730
},
{
"epoch": 0.023103611110027046,
"grad_norm": 1.5520089864730835,
"learning_rate": 3.21490588442868e-05,
"loss": 0.8486,
"step": 740
},
{
"epoch": 0.023415822070973355,
"grad_norm": 4.628422260284424,
"learning_rate": 2.9852511229367865e-05,
"loss": 1.1191,
"step": 750
},
{
"epoch": 0.023728033031919667,
"grad_norm": 1.1060361862182617,
"learning_rate": 2.7626596189492983e-05,
"loss": 1.018,
"step": 760
},
{
"epoch": 0.02404024399286598,
"grad_norm": 1.108621597290039,
"learning_rate": 2.5473555032424533e-05,
"loss": 0.4265,
"step": 770
},
{
"epoch": 0.024352454953812292,
"grad_norm": 1.4980049133300781,
"learning_rate": 2.339555568810221e-05,
"loss": 0.6406,
"step": 780
},
{
"epoch": 0.0246646659147586,
"grad_norm": 1.4313726425170898,
"learning_rate": 2.139469052572127e-05,
"loss": 0.9413,
"step": 790
},
{
"epoch": 0.024976876875704913,
"grad_norm": 1.4710307121276855,
"learning_rate": 1.947297424689414e-05,
"loss": 1.0135,
"step": 800
},
{
"epoch": 0.025289087836651226,
"grad_norm": 0.8172721266746521,
"learning_rate": 1.763234185701673e-05,
"loss": 0.9491,
"step": 810
},
{
"epoch": 0.025601298797597538,
"grad_norm": 1.1474344730377197,
"learning_rate": 1.587464671688187e-05,
"loss": 0.5334,
"step": 820
},
{
"epoch": 0.025913509758543847,
"grad_norm": 1.4124549627304077,
"learning_rate": 1.4201658676502294e-05,
"loss": 0.7033,
"step": 830
},
{
"epoch": 0.02622572071949016,
"grad_norm": 1.3929728269577026,
"learning_rate": 1.2615062293021507e-05,
"loss": 0.852,
"step": 840
},
{
"epoch": 0.026537931680436472,
"grad_norm": 0.782926619052887,
"learning_rate": 1.1116455134507664e-05,
"loss": 1.0648,
"step": 850
},
{
"epoch": 0.02685014264138278,
"grad_norm": 1.1986958980560303,
"learning_rate": 9.707346171337894e-06,
"loss": 0.9374,
"step": 860
},
{
"epoch": 0.027162353602329093,
"grad_norm": 1.0474522113800049,
"learning_rate": 8.38915425679304e-06,
"loss": 0.5091,
"step": 870
},
{
"epoch": 0.027474564563275405,
"grad_norm": 1.569217562675476,
"learning_rate": 7.163206698392744e-06,
"loss": 0.6196,
"step": 880
},
{
"epoch": 0.027786775524221718,
"grad_norm": 2.387129306793213,
"learning_rate": 6.030737921409169e-06,
"loss": 0.8911,
"step": 890
},
{
"epoch": 0.028098986485168027,
"grad_norm": 1.0116970539093018,
"learning_rate": 4.992888225905468e-06,
"loss": 0.893,
"step": 900
},
{
"epoch": 0.02841119744611434,
"grad_norm": 1.056663155555725,
"learning_rate": 4.050702638550275e-06,
"loss": 0.9925,
"step": 910
},
{
"epoch": 0.02872340840706065,
"grad_norm": 1.0909359455108643,
"learning_rate": 3.2051298603643753e-06,
"loss": 0.634,
"step": 920
},
{
"epoch": 0.029035619368006964,
"grad_norm": 1.5931735038757324,
"learning_rate": 2.4570213114592954e-06,
"loss": 0.6653,
"step": 930
},
{
"epoch": 0.029347830328953273,
"grad_norm": 1.73000967502594,
"learning_rate": 1.8071302737293295e-06,
"loss": 0.7891,
"step": 940
},
{
"epoch": 0.029660041289899585,
"grad_norm": 0.5561469197273254,
"learning_rate": 1.2561111323605712e-06,
"loss": 0.9679,
"step": 950
},
{
"epoch": 0.029972252250845897,
"grad_norm": 1.1528609991073608,
"learning_rate": 8.04518716920466e-07,
"loss": 0.9362,
"step": 960
},
{
"epoch": 0.030284463211792206,
"grad_norm": 0.6554343104362488,
"learning_rate": 4.5280774269154115e-07,
"loss": 0.3348,
"step": 970
},
{
"epoch": 0.03059667417273852,
"grad_norm": 1.3619601726531982,
"learning_rate": 2.0133235281156736e-07,
"loss": 0.5258,
"step": 980
},
{
"epoch": 0.03090888513368483,
"grad_norm": 1.2936415672302246,
"learning_rate": 5.0345761681491746e-08,
"loss": 0.9126,
"step": 990
},
{
"epoch": 0.031221096094631143,
"grad_norm": 1.708081841468811,
"learning_rate": 0.0,
"loss": 1.0155,
"step": 1000
}
],
"logging_steps": 10,
"max_steps": 1000,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 167,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7.641564097334477e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}