so100_test-x42chvlz4g / trainer_state.json
LegrandFrederic's picture
Upload trainer_state.json with huggingface_hub
7daf813 verified
raw
history blame
91 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 19.965384615384615,
"eval_steps": 500,
"global_step": 5191,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.038461538461538464,
"grad_norm": 7.0167975425720215,
"learning_rate": 7.692307692307694e-06,
"loss": 1.1358,
"step": 10
},
{
"epoch": 0.07692307692307693,
"grad_norm": 3.2799055576324463,
"learning_rate": 1.5384615384615387e-05,
"loss": 0.8043,
"step": 20
},
{
"epoch": 0.11538461538461539,
"grad_norm": 4.641620635986328,
"learning_rate": 2.307692307692308e-05,
"loss": 0.4223,
"step": 30
},
{
"epoch": 0.15384615384615385,
"grad_norm": 0.8846292495727539,
"learning_rate": 3.0769230769230774e-05,
"loss": 0.2722,
"step": 40
},
{
"epoch": 0.19230769230769232,
"grad_norm": 0.9276573657989502,
"learning_rate": 3.846153846153846e-05,
"loss": 0.1947,
"step": 50
},
{
"epoch": 0.23076923076923078,
"grad_norm": 1.1657756567001343,
"learning_rate": 4.615384615384616e-05,
"loss": 0.1486,
"step": 60
},
{
"epoch": 0.2692307692307692,
"grad_norm": 0.9154715538024902,
"learning_rate": 5.384615384615385e-05,
"loss": 0.1246,
"step": 70
},
{
"epoch": 0.3076923076923077,
"grad_norm": 1.1416549682617188,
"learning_rate": 6.153846153846155e-05,
"loss": 0.1014,
"step": 80
},
{
"epoch": 0.34615384615384615,
"grad_norm": 0.8195381760597229,
"learning_rate": 6.923076923076924e-05,
"loss": 0.0926,
"step": 90
},
{
"epoch": 0.38461538461538464,
"grad_norm": 0.45692142844200134,
"learning_rate": 7.692307692307693e-05,
"loss": 0.0815,
"step": 100
},
{
"epoch": 0.4230769230769231,
"grad_norm": 0.4946421980857849,
"learning_rate": 8.461538461538461e-05,
"loss": 0.0832,
"step": 110
},
{
"epoch": 0.46153846153846156,
"grad_norm": 0.6091125011444092,
"learning_rate": 9.230769230769232e-05,
"loss": 0.0721,
"step": 120
},
{
"epoch": 0.5,
"grad_norm": 0.610546886920929,
"learning_rate": 0.0001,
"loss": 0.0632,
"step": 130
},
{
"epoch": 0.5384615384615384,
"grad_norm": 0.5763479471206665,
"learning_rate": 0.0001076923076923077,
"loss": 0.0673,
"step": 140
},
{
"epoch": 0.5769230769230769,
"grad_norm": 0.821746826171875,
"learning_rate": 0.00011538461538461538,
"loss": 0.0641,
"step": 150
},
{
"epoch": 0.6153846153846154,
"grad_norm": 0.5488954186439514,
"learning_rate": 0.0001230769230769231,
"loss": 0.0625,
"step": 160
},
{
"epoch": 0.6538461538461539,
"grad_norm": 0.6605976819992065,
"learning_rate": 0.00013076923076923077,
"loss": 0.0579,
"step": 170
},
{
"epoch": 0.6923076923076923,
"grad_norm": 0.8009259104728699,
"learning_rate": 0.00013846153846153847,
"loss": 0.0523,
"step": 180
},
{
"epoch": 0.7307692307692307,
"grad_norm": 0.4878791272640228,
"learning_rate": 0.00014615384615384615,
"loss": 0.0538,
"step": 190
},
{
"epoch": 0.7692307692307693,
"grad_norm": 0.3475745618343353,
"learning_rate": 0.00015384615384615385,
"loss": 0.0519,
"step": 200
},
{
"epoch": 0.8076923076923077,
"grad_norm": 0.454436719417572,
"learning_rate": 0.00016153846153846155,
"loss": 0.0514,
"step": 210
},
{
"epoch": 0.8461538461538461,
"grad_norm": 0.7808321118354797,
"learning_rate": 0.00016923076923076923,
"loss": 0.0517,
"step": 220
},
{
"epoch": 0.8846153846153846,
"grad_norm": 0.6214762926101685,
"learning_rate": 0.00017692307692307693,
"loss": 0.0524,
"step": 230
},
{
"epoch": 0.9230769230769231,
"grad_norm": 0.47517019510269165,
"learning_rate": 0.00018461538461538463,
"loss": 0.0545,
"step": 240
},
{
"epoch": 0.9615384615384616,
"grad_norm": 0.7772664427757263,
"learning_rate": 0.00019230769230769233,
"loss": 0.0491,
"step": 250
},
{
"epoch": 1.0,
"grad_norm": 0.863855242729187,
"learning_rate": 0.0002,
"loss": 0.0482,
"step": 260
},
{
"epoch": 1.0384615384615385,
"grad_norm": 0.34070315957069397,
"learning_rate": 0.00019999797045691374,
"loss": 0.0467,
"step": 270
},
{
"epoch": 1.0769230769230769,
"grad_norm": 0.38879549503326416,
"learning_rate": 0.00019999188191003583,
"loss": 0.0392,
"step": 280
},
{
"epoch": 1.1153846153846154,
"grad_norm": 0.4438105821609497,
"learning_rate": 0.00019998173460650565,
"loss": 0.0424,
"step": 290
},
{
"epoch": 1.1538461538461537,
"grad_norm": 0.9575906991958618,
"learning_rate": 0.00019996752895821098,
"loss": 0.0503,
"step": 300
},
{
"epoch": 1.1923076923076923,
"grad_norm": 0.5975697636604309,
"learning_rate": 0.00019994926554177134,
"loss": 0.0487,
"step": 310
},
{
"epoch": 1.2307692307692308,
"grad_norm": 0.5060175061225891,
"learning_rate": 0.00019992694509851454,
"loss": 0.0376,
"step": 320
},
{
"epoch": 1.2692307692307692,
"grad_norm": 0.35833704471588135,
"learning_rate": 0.0001999005685344466,
"loss": 0.0378,
"step": 330
},
{
"epoch": 1.3076923076923077,
"grad_norm": 0.5617273449897766,
"learning_rate": 0.000199870136920215,
"loss": 0.0476,
"step": 340
},
{
"epoch": 1.3461538461538463,
"grad_norm": 0.611136794090271,
"learning_rate": 0.00019983565149106513,
"loss": 0.0442,
"step": 350
},
{
"epoch": 1.3846153846153846,
"grad_norm": 0.3568519949913025,
"learning_rate": 0.00019979711364679037,
"loss": 0.042,
"step": 360
},
{
"epoch": 1.4230769230769231,
"grad_norm": 0.3810388445854187,
"learning_rate": 0.00019975452495167494,
"loss": 0.0375,
"step": 370
},
{
"epoch": 1.4615384615384617,
"grad_norm": 0.5420308113098145,
"learning_rate": 0.00019970788713443073,
"loss": 0.0359,
"step": 380
},
{
"epoch": 1.5,
"grad_norm": 0.5682479739189148,
"learning_rate": 0.00019965720208812692,
"loss": 0.0371,
"step": 390
},
{
"epoch": 1.5384615384615383,
"grad_norm": 0.2949374318122864,
"learning_rate": 0.00019960247187011322,
"loss": 0.0368,
"step": 400
},
{
"epoch": 1.5769230769230769,
"grad_norm": 0.44852954149246216,
"learning_rate": 0.0001995436987019363,
"loss": 0.0375,
"step": 410
},
{
"epoch": 1.6153846153846154,
"grad_norm": 0.2495189756155014,
"learning_rate": 0.00019948088496924972,
"loss": 0.0363,
"step": 420
},
{
"epoch": 1.6538461538461537,
"grad_norm": 0.4514489769935608,
"learning_rate": 0.00019941403322171707,
"loss": 0.037,
"step": 430
},
{
"epoch": 1.6923076923076923,
"grad_norm": 0.26508140563964844,
"learning_rate": 0.00019934314617290832,
"loss": 0.0337,
"step": 440
},
{
"epoch": 1.7307692307692308,
"grad_norm": 0.23346170783042908,
"learning_rate": 0.00019926822670018993,
"loss": 0.0344,
"step": 450
},
{
"epoch": 1.7692307692307692,
"grad_norm": 0.3835722506046295,
"learning_rate": 0.00019918927784460781,
"loss": 0.0324,
"step": 460
},
{
"epoch": 1.8076923076923077,
"grad_norm": 0.3047066330909729,
"learning_rate": 0.00019910630281076407,
"loss": 0.0305,
"step": 470
},
{
"epoch": 1.8461538461538463,
"grad_norm": 0.22565406560897827,
"learning_rate": 0.0001990193049666868,
"loss": 0.033,
"step": 480
},
{
"epoch": 1.8846153846153846,
"grad_norm": 0.39445042610168457,
"learning_rate": 0.00019892828784369345,
"loss": 0.0301,
"step": 490
},
{
"epoch": 1.9230769230769231,
"grad_norm": 0.4797164797782898,
"learning_rate": 0.00019883325513624756,
"loss": 0.0343,
"step": 500
},
{
"epoch": 1.9615384615384617,
"grad_norm": 0.5288494825363159,
"learning_rate": 0.00019873421070180857,
"loss": 0.0318,
"step": 510
},
{
"epoch": 2.0,
"grad_norm": 0.3639772832393646,
"learning_rate": 0.00019863115856067536,
"loss": 0.0318,
"step": 520
},
{
"epoch": 2.0384615384615383,
"grad_norm": 0.4338575303554535,
"learning_rate": 0.0001985241028958232,
"loss": 0.029,
"step": 530
},
{
"epoch": 2.076923076923077,
"grad_norm": 0.25063231587409973,
"learning_rate": 0.00019841304805273382,
"loss": 0.0299,
"step": 540
},
{
"epoch": 2.1153846153846154,
"grad_norm": 0.17708472907543182,
"learning_rate": 0.0001982979985392189,
"loss": 0.024,
"step": 550
},
{
"epoch": 2.1538461538461537,
"grad_norm": 0.33794716000556946,
"learning_rate": 0.00019817895902523745,
"loss": 0.0285,
"step": 560
},
{
"epoch": 2.1923076923076925,
"grad_norm": 0.22588586807250977,
"learning_rate": 0.00019805593434270584,
"loss": 0.0309,
"step": 570
},
{
"epoch": 2.230769230769231,
"grad_norm": 0.27854955196380615,
"learning_rate": 0.00019792892948530195,
"loss": 0.0287,
"step": 580
},
{
"epoch": 2.269230769230769,
"grad_norm": 0.24123087525367737,
"learning_rate": 0.0001977979496082624,
"loss": 0.0259,
"step": 590
},
{
"epoch": 2.3076923076923075,
"grad_norm": 0.311997652053833,
"learning_rate": 0.0001976630000281733,
"loss": 0.0273,
"step": 600
},
{
"epoch": 2.3461538461538463,
"grad_norm": 0.2107621133327484,
"learning_rate": 0.00019752408622275436,
"loss": 0.0267,
"step": 610
},
{
"epoch": 2.3846153846153846,
"grad_norm": 0.49752968549728394,
"learning_rate": 0.0001973812138306366,
"loss": 0.0305,
"step": 620
},
{
"epoch": 2.423076923076923,
"grad_norm": 0.2595117688179016,
"learning_rate": 0.00019723438865113367,
"loss": 0.0272,
"step": 630
},
{
"epoch": 2.4615384615384617,
"grad_norm": 0.35266903042793274,
"learning_rate": 0.00019708361664400598,
"loss": 0.0289,
"step": 640
},
{
"epoch": 2.5,
"grad_norm": 0.3101799190044403,
"learning_rate": 0.00019692890392921934,
"loss": 0.0304,
"step": 650
},
{
"epoch": 2.5384615384615383,
"grad_norm": 0.3763001561164856,
"learning_rate": 0.00019677025678669607,
"loss": 0.0261,
"step": 660
},
{
"epoch": 2.5769230769230766,
"grad_norm": 0.17876401543617249,
"learning_rate": 0.00019660768165606046,
"loss": 0.0251,
"step": 670
},
{
"epoch": 2.6153846153846154,
"grad_norm": 0.2988975942134857,
"learning_rate": 0.00019644118513637715,
"loss": 0.0265,
"step": 680
},
{
"epoch": 2.6538461538461537,
"grad_norm": 0.32994213700294495,
"learning_rate": 0.00019627077398588335,
"loss": 0.0247,
"step": 690
},
{
"epoch": 2.6923076923076925,
"grad_norm": 0.25379645824432373,
"learning_rate": 0.00019609645512171442,
"loss": 0.0241,
"step": 700
},
{
"epoch": 2.730769230769231,
"grad_norm": 0.3905366361141205,
"learning_rate": 0.0001959182356196234,
"loss": 0.0249,
"step": 710
},
{
"epoch": 2.769230769230769,
"grad_norm": 0.22633816301822662,
"learning_rate": 0.00019573612271369336,
"loss": 0.0267,
"step": 720
},
{
"epoch": 2.8076923076923075,
"grad_norm": 0.2320065051317215,
"learning_rate": 0.00019555012379604417,
"loss": 0.0255,
"step": 730
},
{
"epoch": 2.8461538461538463,
"grad_norm": 0.21357214450836182,
"learning_rate": 0.0001953602464165321,
"loss": 0.0289,
"step": 740
},
{
"epoch": 2.8846153846153846,
"grad_norm": 0.23385067284107208,
"learning_rate": 0.00019516649828244363,
"loss": 0.023,
"step": 750
},
{
"epoch": 2.9230769230769234,
"grad_norm": 0.2579933702945709,
"learning_rate": 0.0001949688872581825,
"loss": 0.0258,
"step": 760
},
{
"epoch": 2.9615384615384617,
"grad_norm": 0.3598341643810272,
"learning_rate": 0.0001947674213649504,
"loss": 0.0282,
"step": 770
},
{
"epoch": 3.0,
"grad_norm": 0.2527090907096863,
"learning_rate": 0.00019456210878042164,
"loss": 0.0277,
"step": 780
},
{
"epoch": 3.0384615384615383,
"grad_norm": 0.3657386898994446,
"learning_rate": 0.00019435295783841092,
"loss": 0.0246,
"step": 790
},
{
"epoch": 3.076923076923077,
"grad_norm": 0.35666748881340027,
"learning_rate": 0.0001941399770285352,
"loss": 0.024,
"step": 800
},
{
"epoch": 3.1153846153846154,
"grad_norm": 0.19721713662147522,
"learning_rate": 0.00019392317499586907,
"loss": 0.0204,
"step": 810
},
{
"epoch": 3.1538461538461537,
"grad_norm": 0.23313289880752563,
"learning_rate": 0.00019370256054059386,
"loss": 0.0217,
"step": 820
},
{
"epoch": 3.1923076923076925,
"grad_norm": 0.2258790284395218,
"learning_rate": 0.00019347814261764046,
"loss": 0.0233,
"step": 830
},
{
"epoch": 3.230769230769231,
"grad_norm": 0.21273957192897797,
"learning_rate": 0.00019324993033632568,
"loss": 0.0215,
"step": 840
},
{
"epoch": 3.269230769230769,
"grad_norm": 0.2613265812397003,
"learning_rate": 0.00019301793295998273,
"loss": 0.0226,
"step": 850
},
{
"epoch": 3.3076923076923075,
"grad_norm": 0.35958781838417053,
"learning_rate": 0.00019278215990558504,
"loss": 0.0258,
"step": 860
},
{
"epoch": 3.3461538461538463,
"grad_norm": 0.27915194630622864,
"learning_rate": 0.000192542620743364,
"loss": 0.0255,
"step": 870
},
{
"epoch": 3.3846153846153846,
"grad_norm": 0.23008808493614197,
"learning_rate": 0.00019229932519642068,
"loss": 0.0259,
"step": 880
},
{
"epoch": 3.423076923076923,
"grad_norm": 0.3263101279735565,
"learning_rate": 0.00019205228314033093,
"loss": 0.0247,
"step": 890
},
{
"epoch": 3.4615384615384617,
"grad_norm": 0.27318865060806274,
"learning_rate": 0.00019180150460274478,
"loss": 0.0246,
"step": 900
},
{
"epoch": 3.5,
"grad_norm": 0.38367944955825806,
"learning_rate": 0.00019154699976297907,
"loss": 0.0218,
"step": 910
},
{
"epoch": 3.5384615384615383,
"grad_norm": 0.24464184045791626,
"learning_rate": 0.00019128877895160465,
"loss": 0.0205,
"step": 920
},
{
"epoch": 3.5769230769230766,
"grad_norm": 0.2480219155550003,
"learning_rate": 0.00019102685265002666,
"loss": 0.0248,
"step": 930
},
{
"epoch": 3.6153846153846154,
"grad_norm": 0.3331535756587982,
"learning_rate": 0.0001907612314900595,
"loss": 0.0232,
"step": 940
},
{
"epoch": 3.6538461538461537,
"grad_norm": 0.22589264810085297,
"learning_rate": 0.00019049192625349487,
"loss": 0.0208,
"step": 950
},
{
"epoch": 3.6923076923076925,
"grad_norm": 0.29576578736305237,
"learning_rate": 0.00019021894787166443,
"loss": 0.0239,
"step": 960
},
{
"epoch": 3.730769230769231,
"grad_norm": 0.249545156955719,
"learning_rate": 0.00018994230742499593,
"loss": 0.0225,
"step": 970
},
{
"epoch": 3.769230769230769,
"grad_norm": 0.2573404312133789,
"learning_rate": 0.00018966201614256347,
"loss": 0.0209,
"step": 980
},
{
"epoch": 3.8076923076923075,
"grad_norm": 0.32673683762550354,
"learning_rate": 0.00018937808540163173,
"loss": 0.023,
"step": 990
},
{
"epoch": 3.8461538461538463,
"grad_norm": 0.29287075996398926,
"learning_rate": 0.0001890905267271942,
"loss": 0.0231,
"step": 1000
},
{
"epoch": 3.8846153846153846,
"grad_norm": 0.2039879858493805,
"learning_rate": 0.0001887993517915052,
"loss": 0.0184,
"step": 1010
},
{
"epoch": 3.9230769230769234,
"grad_norm": 0.24603568017482758,
"learning_rate": 0.00018850457241360635,
"loss": 0.0262,
"step": 1020
},
{
"epoch": 3.9615384615384617,
"grad_norm": 0.2239179164171219,
"learning_rate": 0.00018820620055884658,
"loss": 0.0218,
"step": 1030
},
{
"epoch": 4.0,
"grad_norm": 0.17946171760559082,
"learning_rate": 0.0001879042483383966,
"loss": 0.0181,
"step": 1040
},
{
"epoch": 4.038461538461538,
"grad_norm": 0.11465075612068176,
"learning_rate": 0.00018759872800875726,
"loss": 0.0172,
"step": 1050
},
{
"epoch": 4.076923076923077,
"grad_norm": 0.17597812414169312,
"learning_rate": 0.000187289651971262,
"loss": 0.0169,
"step": 1060
},
{
"epoch": 4.115384615384615,
"grad_norm": 0.19563104212284088,
"learning_rate": 0.00018697703277157347,
"loss": 0.0198,
"step": 1070
},
{
"epoch": 4.153846153846154,
"grad_norm": 0.18617744743824005,
"learning_rate": 0.00018666088309917447,
"loss": 0.0218,
"step": 1080
},
{
"epoch": 4.1923076923076925,
"grad_norm": 0.23704974353313446,
"learning_rate": 0.00018634121578685255,
"loss": 0.0186,
"step": 1090
},
{
"epoch": 4.230769230769231,
"grad_norm": 0.32059594988822937,
"learning_rate": 0.00018601804381017945,
"loss": 0.0211,
"step": 1100
},
{
"epoch": 4.269230769230769,
"grad_norm": 0.28677231073379517,
"learning_rate": 0.00018569138028698415,
"loss": 0.0159,
"step": 1110
},
{
"epoch": 4.3076923076923075,
"grad_norm": 0.24357284605503082,
"learning_rate": 0.00018536123847682056,
"loss": 0.0189,
"step": 1120
},
{
"epoch": 4.346153846153846,
"grad_norm": 0.1930132806301117,
"learning_rate": 0.00018502763178042923,
"loss": 0.0202,
"step": 1130
},
{
"epoch": 4.384615384615385,
"grad_norm": 0.1683141440153122,
"learning_rate": 0.00018469057373919347,
"loss": 0.0202,
"step": 1140
},
{
"epoch": 4.423076923076923,
"grad_norm": 0.1439371407032013,
"learning_rate": 0.00018435007803458965,
"loss": 0.0186,
"step": 1150
},
{
"epoch": 4.461538461538462,
"grad_norm": 0.23720888793468475,
"learning_rate": 0.00018400615848763175,
"loss": 0.0209,
"step": 1160
},
{
"epoch": 4.5,
"grad_norm": 0.15303802490234375,
"learning_rate": 0.00018365882905831065,
"loss": 0.0232,
"step": 1170
},
{
"epoch": 4.538461538461538,
"grad_norm": 0.140608549118042,
"learning_rate": 0.0001833081038450271,
"loss": 0.0192,
"step": 1180
},
{
"epoch": 4.576923076923077,
"grad_norm": 0.31198394298553467,
"learning_rate": 0.00018295399708401975,
"loss": 0.0168,
"step": 1190
},
{
"epoch": 4.615384615384615,
"grad_norm": 0.1934264749288559,
"learning_rate": 0.00018259652314878721,
"loss": 0.0207,
"step": 1200
},
{
"epoch": 4.653846153846154,
"grad_norm": 0.33610162138938904,
"learning_rate": 0.00018223569654950454,
"loss": 0.0196,
"step": 1210
},
{
"epoch": 4.6923076923076925,
"grad_norm": 0.2503054738044739,
"learning_rate": 0.00018187153193243436,
"loss": 0.0201,
"step": 1220
},
{
"epoch": 4.730769230769231,
"grad_norm": 0.20254291594028473,
"learning_rate": 0.0001815040440793322,
"loss": 0.0199,
"step": 1230
},
{
"epoch": 4.769230769230769,
"grad_norm": 0.24005241692066193,
"learning_rate": 0.0001811332479068468,
"loss": 0.0214,
"step": 1240
},
{
"epoch": 4.8076923076923075,
"grad_norm": 0.21922606229782104,
"learning_rate": 0.00018075915846591428,
"loss": 0.0204,
"step": 1250
},
{
"epoch": 4.846153846153846,
"grad_norm": 0.2370215505361557,
"learning_rate": 0.00018038179094114742,
"loss": 0.022,
"step": 1260
},
{
"epoch": 4.884615384615385,
"grad_norm": 0.2668112814426422,
"learning_rate": 0.0001800011606502192,
"loss": 0.0188,
"step": 1270
},
{
"epoch": 4.923076923076923,
"grad_norm": 0.15615129470825195,
"learning_rate": 0.00017961728304324116,
"loss": 0.0182,
"step": 1280
},
{
"epoch": 4.961538461538462,
"grad_norm": 0.23041196167469025,
"learning_rate": 0.00017923017370213615,
"loss": 0.0188,
"step": 1290
},
{
"epoch": 5.0,
"grad_norm": 0.21733583509922028,
"learning_rate": 0.0001788398483400059,
"loss": 0.0198,
"step": 1300
},
{
"epoch": 5.038461538461538,
"grad_norm": 0.2194267213344574,
"learning_rate": 0.00017844632280049328,
"loss": 0.0181,
"step": 1310
},
{
"epoch": 5.076923076923077,
"grad_norm": 0.14587156474590302,
"learning_rate": 0.00017804961305713896,
"loss": 0.0151,
"step": 1320
},
{
"epoch": 5.115384615384615,
"grad_norm": 0.16885510087013245,
"learning_rate": 0.00017764973521273332,
"loss": 0.0215,
"step": 1330
},
{
"epoch": 5.153846153846154,
"grad_norm": 0.19568288326263428,
"learning_rate": 0.00017724670549866268,
"loss": 0.0191,
"step": 1340
},
{
"epoch": 5.1923076923076925,
"grad_norm": 0.15629301965236664,
"learning_rate": 0.00017684054027425035,
"loss": 0.0174,
"step": 1350
},
{
"epoch": 5.230769230769231,
"grad_norm": 0.17440061271190643,
"learning_rate": 0.00017643125602609288,
"loss": 0.0145,
"step": 1360
},
{
"epoch": 5.269230769230769,
"grad_norm": 0.21910502016544342,
"learning_rate": 0.0001760188693673905,
"loss": 0.0181,
"step": 1370
},
{
"epoch": 5.3076923076923075,
"grad_norm": 0.21728090941905975,
"learning_rate": 0.00017560339703727315,
"loss": 0.0137,
"step": 1380
},
{
"epoch": 5.346153846153846,
"grad_norm": 0.15279999375343323,
"learning_rate": 0.00017518485590012068,
"loss": 0.0152,
"step": 1390
},
{
"epoch": 5.384615384615385,
"grad_norm": 0.33734527230262756,
"learning_rate": 0.00017476326294487852,
"loss": 0.0169,
"step": 1400
},
{
"epoch": 5.423076923076923,
"grad_norm": 0.15465325117111206,
"learning_rate": 0.00017433863528436804,
"loss": 0.0159,
"step": 1410
},
{
"epoch": 5.461538461538462,
"grad_norm": 0.1761886179447174,
"learning_rate": 0.00017391099015459186,
"loss": 0.0179,
"step": 1420
},
{
"epoch": 5.5,
"grad_norm": 0.1088884100317955,
"learning_rate": 0.00017348034491403432,
"loss": 0.0163,
"step": 1430
},
{
"epoch": 5.538461538461538,
"grad_norm": 0.3224456012248993,
"learning_rate": 0.00017304671704295685,
"loss": 0.016,
"step": 1440
},
{
"epoch": 5.576923076923077,
"grad_norm": 0.2751074433326721,
"learning_rate": 0.0001726101241426884,
"loss": 0.0152,
"step": 1450
},
{
"epoch": 5.615384615384615,
"grad_norm": 0.20135682821273804,
"learning_rate": 0.00017217058393491098,
"loss": 0.0213,
"step": 1460
},
{
"epoch": 5.653846153846154,
"grad_norm": 0.13225609064102173,
"learning_rate": 0.00017172811426094045,
"loss": 0.0151,
"step": 1470
},
{
"epoch": 5.6923076923076925,
"grad_norm": 0.24074223637580872,
"learning_rate": 0.0001712827330810021,
"loss": 0.0181,
"step": 1480
},
{
"epoch": 5.730769230769231,
"grad_norm": 0.2650367319583893,
"learning_rate": 0.00017083445847350186,
"loss": 0.0181,
"step": 1490
},
{
"epoch": 5.769230769230769,
"grad_norm": 0.17957966029644012,
"learning_rate": 0.00017038330863429236,
"loss": 0.0153,
"step": 1500
},
{
"epoch": 5.8076923076923075,
"grad_norm": 0.1726105660200119,
"learning_rate": 0.00016992930187593425,
"loss": 0.0178,
"step": 1510
},
{
"epoch": 5.846153846153846,
"grad_norm": 0.29331403970718384,
"learning_rate": 0.00016947245662695317,
"loss": 0.0164,
"step": 1520
},
{
"epoch": 5.884615384615385,
"grad_norm": 0.1539272964000702,
"learning_rate": 0.0001690127914310914,
"loss": 0.0185,
"step": 1530
},
{
"epoch": 5.923076923076923,
"grad_norm": 0.2472430318593979,
"learning_rate": 0.00016855032494655537,
"loss": 0.0163,
"step": 1540
},
{
"epoch": 5.961538461538462,
"grad_norm": 0.23108868300914764,
"learning_rate": 0.00016808507594525822,
"loss": 0.0163,
"step": 1550
},
{
"epoch": 6.0,
"grad_norm": 0.1876610368490219,
"learning_rate": 0.00016761706331205783,
"loss": 0.0183,
"step": 1560
},
{
"epoch": 6.038461538461538,
"grad_norm": 0.2372998148202896,
"learning_rate": 0.00016714630604399021,
"loss": 0.0184,
"step": 1570
},
{
"epoch": 6.076923076923077,
"grad_norm": 0.23431254923343658,
"learning_rate": 0.00016667282324949863,
"loss": 0.0177,
"step": 1580
},
{
"epoch": 6.115384615384615,
"grad_norm": 0.18878187239170074,
"learning_rate": 0.00016619663414765764,
"loss": 0.0176,
"step": 1590
},
{
"epoch": 6.153846153846154,
"grad_norm": 0.17434252798557281,
"learning_rate": 0.00016571775806739335,
"loss": 0.0182,
"step": 1600
},
{
"epoch": 6.1923076923076925,
"grad_norm": 0.21172362565994263,
"learning_rate": 0.00016523621444669836,
"loss": 0.0186,
"step": 1610
},
{
"epoch": 6.230769230769231,
"grad_norm": 0.3032718300819397,
"learning_rate": 0.00016475202283184332,
"loss": 0.0158,
"step": 1620
},
{
"epoch": 6.269230769230769,
"grad_norm": 0.2207326591014862,
"learning_rate": 0.00016426520287658307,
"loss": 0.0148,
"step": 1630
},
{
"epoch": 6.3076923076923075,
"grad_norm": 0.1728544682264328,
"learning_rate": 0.00016377577434135908,
"loss": 0.018,
"step": 1640
},
{
"epoch": 6.346153846153846,
"grad_norm": 0.2233678251504898,
"learning_rate": 0.00016328375709249738,
"loss": 0.0143,
"step": 1650
},
{
"epoch": 6.384615384615385,
"grad_norm": 0.156491219997406,
"learning_rate": 0.00016278917110140205,
"loss": 0.015,
"step": 1660
},
{
"epoch": 6.423076923076923,
"grad_norm": 0.17857709527015686,
"learning_rate": 0.00016229203644374475,
"loss": 0.0164,
"step": 1670
},
{
"epoch": 6.461538461538462,
"grad_norm": 0.1415674239397049,
"learning_rate": 0.00016179237329864952,
"loss": 0.0157,
"step": 1680
},
{
"epoch": 6.5,
"grad_norm": 0.1668516844511032,
"learning_rate": 0.00016129020194787403,
"loss": 0.0168,
"step": 1690
},
{
"epoch": 6.538461538461538,
"grad_norm": 0.1995563507080078,
"learning_rate": 0.00016078554277498618,
"loss": 0.018,
"step": 1700
},
{
"epoch": 6.576923076923077,
"grad_norm": 0.18094874918460846,
"learning_rate": 0.00016027841626453665,
"loss": 0.0198,
"step": 1710
},
{
"epoch": 6.615384615384615,
"grad_norm": 0.17089825868606567,
"learning_rate": 0.00015976884300122746,
"loss": 0.0158,
"step": 1720
},
{
"epoch": 6.653846153846154,
"grad_norm": 0.17652678489685059,
"learning_rate": 0.00015925684366907653,
"loss": 0.0187,
"step": 1730
},
{
"epoch": 6.6923076923076925,
"grad_norm": 0.17994622886180878,
"learning_rate": 0.00015874243905057798,
"loss": 0.0156,
"step": 1740
},
{
"epoch": 6.730769230769231,
"grad_norm": 0.19167155027389526,
"learning_rate": 0.0001582256500258585,
"loss": 0.0198,
"step": 1750
},
{
"epoch": 6.769230769230769,
"grad_norm": 0.18259945511817932,
"learning_rate": 0.00015770649757182993,
"loss": 0.0164,
"step": 1760
},
{
"epoch": 6.8076923076923075,
"grad_norm": 0.22917133569717407,
"learning_rate": 0.0001571850027613378,
"loss": 0.0169,
"step": 1770
},
{
"epoch": 6.846153846153846,
"grad_norm": 0.20906826853752136,
"learning_rate": 0.00015666118676230576,
"loss": 0.0179,
"step": 1780
},
{
"epoch": 6.884615384615385,
"grad_norm": 0.3887175917625427,
"learning_rate": 0.00015613507083687668,
"loss": 0.0178,
"step": 1790
},
{
"epoch": 6.923076923076923,
"grad_norm": 0.1430666148662567,
"learning_rate": 0.00015560667634054932,
"loss": 0.0169,
"step": 1800
},
{
"epoch": 6.961538461538462,
"grad_norm": 0.34076541662216187,
"learning_rate": 0.00015507602472131162,
"loss": 0.0176,
"step": 1810
},
{
"epoch": 7.0,
"grad_norm": 0.248886838555336,
"learning_rate": 0.00015454313751877004,
"loss": 0.0178,
"step": 1820
},
{
"epoch": 7.038461538461538,
"grad_norm": 0.21486853063106537,
"learning_rate": 0.0001540080363632754,
"loss": 0.0162,
"step": 1830
},
{
"epoch": 7.076923076923077,
"grad_norm": 0.19121983647346497,
"learning_rate": 0.00015347074297504464,
"loss": 0.0147,
"step": 1840
},
{
"epoch": 7.115384615384615,
"grad_norm": 0.20204658806324005,
"learning_rate": 0.00015293127916327944,
"loss": 0.0167,
"step": 1850
},
{
"epoch": 7.153846153846154,
"grad_norm": 0.17696718871593475,
"learning_rate": 0.0001523896668252808,
"loss": 0.017,
"step": 1860
},
{
"epoch": 7.1923076923076925,
"grad_norm": 0.19161172211170197,
"learning_rate": 0.0001518459279455602,
"loss": 0.0143,
"step": 1870
},
{
"epoch": 7.230769230769231,
"grad_norm": 0.2208317518234253,
"learning_rate": 0.0001513000845949473,
"loss": 0.016,
"step": 1880
},
{
"epoch": 7.269230769230769,
"grad_norm": 0.16817378997802734,
"learning_rate": 0.00015075215892969418,
"loss": 0.0177,
"step": 1890
},
{
"epoch": 7.3076923076923075,
"grad_norm": 0.21887525916099548,
"learning_rate": 0.0001502021731905756,
"loss": 0.0142,
"step": 1900
},
{
"epoch": 7.346153846153846,
"grad_norm": 0.14425648748874664,
"learning_rate": 0.00014965014970198676,
"loss": 0.015,
"step": 1910
},
{
"epoch": 7.384615384615385,
"grad_norm": 0.2206214964389801,
"learning_rate": 0.0001490961108710367,
"loss": 0.0158,
"step": 1920
},
{
"epoch": 7.423076923076923,
"grad_norm": 0.13397306203842163,
"learning_rate": 0.00014854007918663903,
"loss": 0.0172,
"step": 1930
},
{
"epoch": 7.461538461538462,
"grad_norm": 0.2873019278049469,
"learning_rate": 0.00014798207721859898,
"loss": 0.0167,
"step": 1940
},
{
"epoch": 7.5,
"grad_norm": 0.20510359108448029,
"learning_rate": 0.00014742212761669724,
"loss": 0.0149,
"step": 1950
},
{
"epoch": 7.538461538461538,
"grad_norm": 0.2240099161863327,
"learning_rate": 0.00014686025310977065,
"loss": 0.0149,
"step": 1960
},
{
"epoch": 7.576923076923077,
"grad_norm": 0.2232583910226822,
"learning_rate": 0.0001462964765047897,
"loss": 0.0148,
"step": 1970
},
{
"epoch": 7.615384615384615,
"grad_norm": 0.16749520599842072,
"learning_rate": 0.00014573082068593254,
"loss": 0.0165,
"step": 1980
},
{
"epoch": 7.653846153846154,
"grad_norm": 0.18627752363681793,
"learning_rate": 0.00014516330861365638,
"loss": 0.0152,
"step": 1990
},
{
"epoch": 7.6923076923076925,
"grad_norm": 0.2765505313873291,
"learning_rate": 0.00014459396332376517,
"loss": 0.0154,
"step": 2000
},
{
"epoch": 7.730769230769231,
"grad_norm": 0.2526029050350189,
"learning_rate": 0.00014402280792647493,
"loss": 0.0149,
"step": 2010
},
{
"epoch": 7.769230769230769,
"grad_norm": 0.12884902954101562,
"learning_rate": 0.00014344986560547537,
"loss": 0.0164,
"step": 2020
},
{
"epoch": 7.8076923076923075,
"grad_norm": 0.10823468118906021,
"learning_rate": 0.00014287515961698902,
"loss": 0.0149,
"step": 2030
},
{
"epoch": 7.846153846153846,
"grad_norm": 0.15259294211864471,
"learning_rate": 0.0001422987132888272,
"loss": 0.0143,
"step": 2040
},
{
"epoch": 7.884615384615385,
"grad_norm": 0.13543514907360077,
"learning_rate": 0.00014172055001944312,
"loss": 0.0134,
"step": 2050
},
{
"epoch": 7.923076923076923,
"grad_norm": 0.2171320766210556,
"learning_rate": 0.0001411406932769821,
"loss": 0.0156,
"step": 2060
},
{
"epoch": 7.961538461538462,
"grad_norm": 0.09700020402669907,
"learning_rate": 0.000140559166598329,
"loss": 0.0141,
"step": 2070
},
{
"epoch": 8.0,
"grad_norm": 0.19601549208164215,
"learning_rate": 0.00013997599358815275,
"loss": 0.0147,
"step": 2080
},
{
"epoch": 8.038461538461538,
"grad_norm": 0.18070325255393982,
"learning_rate": 0.0001393911979179485,
"loss": 0.0159,
"step": 2090
},
{
"epoch": 8.076923076923077,
"grad_norm": 0.17102424800395966,
"learning_rate": 0.0001388048033250763,
"loss": 0.0171,
"step": 2100
},
{
"epoch": 8.115384615384615,
"grad_norm": 0.15269245207309723,
"learning_rate": 0.00013821683361179809,
"loss": 0.0136,
"step": 2110
},
{
"epoch": 8.153846153846153,
"grad_norm": 0.14295409619808197,
"learning_rate": 0.00013762731264431116,
"loss": 0.0116,
"step": 2120
},
{
"epoch": 8.192307692307692,
"grad_norm": 0.13391715288162231,
"learning_rate": 0.00013703626435177956,
"loss": 0.017,
"step": 2130
},
{
"epoch": 8.23076923076923,
"grad_norm": 0.18101289868354797,
"learning_rate": 0.00013644371272536284,
"loss": 0.0145,
"step": 2140
},
{
"epoch": 8.26923076923077,
"grad_norm": 0.13576363027095795,
"learning_rate": 0.00013584968181724213,
"loss": 0.0139,
"step": 2150
},
{
"epoch": 8.307692307692308,
"grad_norm": 0.14105717837810516,
"learning_rate": 0.00013525419573964386,
"loss": 0.0132,
"step": 2160
},
{
"epoch": 8.346153846153847,
"grad_norm": 0.09896686673164368,
"learning_rate": 0.00013465727866386108,
"loss": 0.0107,
"step": 2170
},
{
"epoch": 8.384615384615385,
"grad_norm": 0.27290794253349304,
"learning_rate": 0.0001340589548192723,
"loss": 0.0143,
"step": 2180
},
{
"epoch": 8.423076923076923,
"grad_norm": 0.1378784477710724,
"learning_rate": 0.00013345924849235793,
"loss": 0.0136,
"step": 2190
},
{
"epoch": 8.461538461538462,
"grad_norm": 0.18884290754795074,
"learning_rate": 0.00013285818402571458,
"loss": 0.0153,
"step": 2200
},
{
"epoch": 8.5,
"grad_norm": 0.13120310008525848,
"learning_rate": 0.00013225578581706687,
"loss": 0.0145,
"step": 2210
},
{
"epoch": 8.538461538461538,
"grad_norm": 0.1788390427827835,
"learning_rate": 0.00013165207831827724,
"loss": 0.0152,
"step": 2220
},
{
"epoch": 8.576923076923077,
"grad_norm": 0.16243167221546173,
"learning_rate": 0.00013104708603435325,
"loss": 0.0161,
"step": 2230
},
{
"epoch": 8.615384615384615,
"grad_norm": 0.24895057082176208,
"learning_rate": 0.00013044083352245308,
"loss": 0.0145,
"step": 2240
},
{
"epoch": 8.653846153846153,
"grad_norm": 0.15210068225860596,
"learning_rate": 0.0001298333453908886,
"loss": 0.0135,
"step": 2250
},
{
"epoch": 8.692307692307692,
"grad_norm": 0.19090086221694946,
"learning_rate": 0.00012922464629812653,
"loss": 0.0167,
"step": 2260
},
{
"epoch": 8.73076923076923,
"grad_norm": 0.2009580284357071,
"learning_rate": 0.0001286147609517876,
"loss": 0.0134,
"step": 2270
},
{
"epoch": 8.76923076923077,
"grad_norm": 0.20718801021575928,
"learning_rate": 0.00012800371410764356,
"loss": 0.0122,
"step": 2280
},
{
"epoch": 8.807692307692308,
"grad_norm": 0.18385674059391022,
"learning_rate": 0.00012739153056861243,
"loss": 0.0145,
"step": 2290
},
{
"epoch": 8.846153846153847,
"grad_norm": 0.19775976240634918,
"learning_rate": 0.00012677823518375146,
"loss": 0.0143,
"step": 2300
},
{
"epoch": 8.884615384615385,
"grad_norm": 0.14536786079406738,
"learning_rate": 0.00012616385284724898,
"loss": 0.0136,
"step": 2310
},
{
"epoch": 8.923076923076923,
"grad_norm": 0.1978665441274643,
"learning_rate": 0.00012554840849741332,
"loss": 0.0144,
"step": 2320
},
{
"epoch": 8.961538461538462,
"grad_norm": 0.16081364452838898,
"learning_rate": 0.00012493192711566104,
"loss": 0.0146,
"step": 2330
},
{
"epoch": 9.0,
"grad_norm": 0.18740719556808472,
"learning_rate": 0.00012431443372550266,
"loss": 0.017,
"step": 2340
},
{
"epoch": 9.038461538461538,
"grad_norm": 0.1770515739917755,
"learning_rate": 0.00012369595339152702,
"loss": 0.0154,
"step": 2350
},
{
"epoch": 9.076923076923077,
"grad_norm": 0.12580080330371857,
"learning_rate": 0.0001230765112183838,
"loss": 0.0149,
"step": 2360
},
{
"epoch": 9.115384615384615,
"grad_norm": 0.19087755680084229,
"learning_rate": 0.00012245613234976462,
"loss": 0.0145,
"step": 2370
},
{
"epoch": 9.153846153846153,
"grad_norm": 0.10227272659540176,
"learning_rate": 0.00012183484196738233,
"loss": 0.0126,
"step": 2380
},
{
"epoch": 9.192307692307692,
"grad_norm": 0.17450454831123352,
"learning_rate": 0.00012121266528994895,
"loss": 0.0145,
"step": 2390
},
{
"epoch": 9.23076923076923,
"grad_norm": 0.15834400057792664,
"learning_rate": 0.00012058962757215195,
"loss": 0.0122,
"step": 2400
},
{
"epoch": 9.26923076923077,
"grad_norm": 0.13027557730674744,
"learning_rate": 0.00011996575410362919,
"loss": 0.0137,
"step": 2410
},
{
"epoch": 9.307692307692308,
"grad_norm": 0.18278013169765472,
"learning_rate": 0.00011934107020794239,
"loss": 0.0133,
"step": 2420
},
{
"epoch": 9.346153846153847,
"grad_norm": 0.15627382695674896,
"learning_rate": 0.00011871560124154912,
"loss": 0.0158,
"step": 2430
},
{
"epoch": 9.384615384615385,
"grad_norm": 0.16056114435195923,
"learning_rate": 0.00011808937259277377,
"loss": 0.0131,
"step": 2440
},
{
"epoch": 9.423076923076923,
"grad_norm": 0.11140304803848267,
"learning_rate": 0.00011746240968077682,
"loss": 0.0142,
"step": 2450
},
{
"epoch": 9.461538461538462,
"grad_norm": 0.17418035864830017,
"learning_rate": 0.0001168347379545231,
"loss": 0.0123,
"step": 2460
},
{
"epoch": 9.5,
"grad_norm": 0.0985954999923706,
"learning_rate": 0.0001162063828917489,
"loss": 0.0123,
"step": 2470
},
{
"epoch": 9.538461538461538,
"grad_norm": 0.22951987385749817,
"learning_rate": 0.00011557736999792765,
"loss": 0.0127,
"step": 2480
},
{
"epoch": 9.576923076923077,
"grad_norm": 0.13590703904628754,
"learning_rate": 0.00011494772480523482,
"loss": 0.0135,
"step": 2490
},
{
"epoch": 9.615384615384615,
"grad_norm": 0.1775355190038681,
"learning_rate": 0.00011431747287151125,
"loss": 0.013,
"step": 2500
},
{
"epoch": 9.653846153846153,
"grad_norm": 0.22847086191177368,
"learning_rate": 0.00011368663977922613,
"loss": 0.0136,
"step": 2510
},
{
"epoch": 9.692307692307692,
"grad_norm": 0.1462956815958023,
"learning_rate": 0.00011305525113443822,
"loss": 0.0141,
"step": 2520
},
{
"epoch": 9.73076923076923,
"grad_norm": 0.1806713193655014,
"learning_rate": 0.00011242333256575677,
"loss": 0.0118,
"step": 2530
},
{
"epoch": 9.76923076923077,
"grad_norm": 0.14892129600048065,
"learning_rate": 0.00011179090972330093,
"loss": 0.012,
"step": 2540
},
{
"epoch": 9.807692307692308,
"grad_norm": 0.1719180941581726,
"learning_rate": 0.0001111580082776589,
"loss": 0.0134,
"step": 2550
},
{
"epoch": 9.846153846153847,
"grad_norm": 0.1946127712726593,
"learning_rate": 0.00011052465391884575,
"loss": 0.0146,
"step": 2560
},
{
"epoch": 9.884615384615385,
"grad_norm": 0.16166430711746216,
"learning_rate": 0.00010989087235526068,
"loss": 0.0145,
"step": 2570
},
{
"epoch": 9.923076923076923,
"grad_norm": 0.18934275209903717,
"learning_rate": 0.00010925668931264346,
"loss": 0.0109,
"step": 2580
},
{
"epoch": 9.961538461538462,
"grad_norm": 0.14358612895011902,
"learning_rate": 0.00010862213053303037,
"loss": 0.0099,
"step": 2590
},
{
"epoch": 10.0,
"grad_norm": 0.2569468319416046,
"learning_rate": 0.00010798722177370897,
"loss": 0.0147,
"step": 2600
},
{
"epoch": 10.038461538461538,
"grad_norm": 0.21339739859104156,
"learning_rate": 0.00010735198880617303,
"loss": 0.0137,
"step": 2610
},
{
"epoch": 10.076923076923077,
"grad_norm": 0.12013400346040726,
"learning_rate": 0.00010671645741507603,
"loss": 0.011,
"step": 2620
},
{
"epoch": 10.115384615384615,
"grad_norm": 0.18901623785495758,
"learning_rate": 0.00010608065339718483,
"loss": 0.0109,
"step": 2630
},
{
"epoch": 10.153846153846153,
"grad_norm": 0.10211600363254547,
"learning_rate": 0.00010544460256033237,
"loss": 0.0132,
"step": 2640
},
{
"epoch": 10.192307692307692,
"grad_norm": 0.17099827527999878,
"learning_rate": 0.00010480833072237022,
"loss": 0.0139,
"step": 2650
},
{
"epoch": 10.23076923076923,
"grad_norm": 0.15852560102939606,
"learning_rate": 0.00010417186371012063,
"loss": 0.0116,
"step": 2660
},
{
"epoch": 10.26923076923077,
"grad_norm": 0.17198602855205536,
"learning_rate": 0.000103535227358328,
"loss": 0.0114,
"step": 2670
},
{
"epoch": 10.307692307692308,
"grad_norm": 0.21135368943214417,
"learning_rate": 0.00010289844750861053,
"loss": 0.0126,
"step": 2680
},
{
"epoch": 10.346153846153847,
"grad_norm": 0.13126599788665771,
"learning_rate": 0.000102261550008411,
"loss": 0.013,
"step": 2690
},
{
"epoch": 10.384615384615385,
"grad_norm": 0.16771045327186584,
"learning_rate": 0.00010162456070994781,
"loss": 0.0111,
"step": 2700
},
{
"epoch": 10.423076923076923,
"grad_norm": 0.1977088302373886,
"learning_rate": 0.00010098750546916547,
"loss": 0.0123,
"step": 2710
},
{
"epoch": 10.461538461538462,
"grad_norm": 0.12443752586841583,
"learning_rate": 0.00010035041014468518,
"loss": 0.0114,
"step": 2720
},
{
"epoch": 10.5,
"grad_norm": 0.2196878045797348,
"learning_rate": 9.971330059675518e-05,
"loss": 0.0151,
"step": 2730
},
{
"epoch": 10.538461538461538,
"grad_norm": 0.1408672332763672,
"learning_rate": 9.907620268620098e-05,
"loss": 0.011,
"step": 2740
},
{
"epoch": 10.576923076923077,
"grad_norm": 0.14648661017417908,
"learning_rate": 9.84391422733758e-05,
"loss": 0.0101,
"step": 2750
},
{
"epoch": 10.615384615384615,
"grad_norm": 0.14232785999774933,
"learning_rate": 9.78021452171108e-05,
"loss": 0.0104,
"step": 2760
},
{
"epoch": 10.653846153846153,
"grad_norm": 0.16160699725151062,
"learning_rate": 9.716523737366537e-05,
"loss": 0.0127,
"step": 2770
},
{
"epoch": 10.692307692307692,
"grad_norm": 0.20870442688465118,
"learning_rate": 9.652844459567777e-05,
"loss": 0.0103,
"step": 2780
},
{
"epoch": 10.73076923076923,
"grad_norm": 0.15344356000423431,
"learning_rate": 9.589179273111551e-05,
"loss": 0.0101,
"step": 2790
},
{
"epoch": 10.76923076923077,
"grad_norm": 0.11750971525907516,
"learning_rate": 9.525530762222645e-05,
"loss": 0.0111,
"step": 2800
},
{
"epoch": 10.807692307692308,
"grad_norm": 0.117418073117733,
"learning_rate": 9.461901510448962e-05,
"loss": 0.0128,
"step": 2810
},
{
"epoch": 10.846153846153847,
"grad_norm": 0.317682683467865,
"learning_rate": 9.398294100556668e-05,
"loss": 0.0124,
"step": 2820
},
{
"epoch": 10.884615384615385,
"grad_norm": 0.14449702203273773,
"learning_rate": 9.334711114425335e-05,
"loss": 0.0117,
"step": 2830
},
{
"epoch": 10.923076923076923,
"grad_norm": 0.09939802438020706,
"learning_rate": 9.27115513294316e-05,
"loss": 0.012,
"step": 2840
},
{
"epoch": 10.961538461538462,
"grad_norm": 0.16178478300571442,
"learning_rate": 9.207628735902205e-05,
"loss": 0.0138,
"step": 2850
},
{
"epoch": 11.0,
"grad_norm": 0.21062442660331726,
"learning_rate": 9.144134501893671e-05,
"loss": 0.0122,
"step": 2860
},
{
"epoch": 11.038461538461538,
"grad_norm": 0.17146876454353333,
"learning_rate": 9.080675008203221e-05,
"loss": 0.0148,
"step": 2870
},
{
"epoch": 11.076923076923077,
"grad_norm": 0.13412779569625854,
"learning_rate": 9.017252830706394e-05,
"loss": 0.0099,
"step": 2880
},
{
"epoch": 11.115384615384615,
"grad_norm": 0.14295507967472076,
"learning_rate": 8.953870543764026e-05,
"loss": 0.0116,
"step": 2890
},
{
"epoch": 11.153846153846153,
"grad_norm": 0.13120034337043762,
"learning_rate": 8.890530720117767e-05,
"loss": 0.0127,
"step": 2900
},
{
"epoch": 11.192307692307692,
"grad_norm": 0.18492452800273895,
"learning_rate": 8.827235930785636e-05,
"loss": 0.0106,
"step": 2910
},
{
"epoch": 11.23076923076923,
"grad_norm": 0.15923990309238434,
"learning_rate": 8.763988744957672e-05,
"loss": 0.0134,
"step": 2920
},
{
"epoch": 11.26923076923077,
"grad_norm": 0.10775059461593628,
"learning_rate": 8.700791729891651e-05,
"loss": 0.0141,
"step": 2930
},
{
"epoch": 11.307692307692308,
"grad_norm": 0.1554538756608963,
"learning_rate": 8.637647450808879e-05,
"loss": 0.0103,
"step": 2940
},
{
"epoch": 11.346153846153847,
"grad_norm": 0.16043995320796967,
"learning_rate": 8.574558470790053e-05,
"loss": 0.0147,
"step": 2950
},
{
"epoch": 11.384615384615385,
"grad_norm": 0.1335715800523758,
"learning_rate": 8.511527350671233e-05,
"loss": 0.0115,
"step": 2960
},
{
"epoch": 11.423076923076923,
"grad_norm": 0.13701431453227997,
"learning_rate": 8.448556648939907e-05,
"loss": 0.0108,
"step": 2970
},
{
"epoch": 11.461538461538462,
"grad_norm": 0.09677750617265701,
"learning_rate": 8.385648921631116e-05,
"loss": 0.0095,
"step": 2980
},
{
"epoch": 11.5,
"grad_norm": 0.08720875531435013,
"learning_rate": 8.322806722223725e-05,
"loss": 0.0086,
"step": 2990
},
{
"epoch": 11.538461538461538,
"grad_norm": 0.10705883800983429,
"learning_rate": 8.260032601536758e-05,
"loss": 0.0103,
"step": 3000
},
{
"epoch": 11.576923076923077,
"grad_norm": 0.1649639904499054,
"learning_rate": 8.19732910762587e-05,
"loss": 0.0093,
"step": 3010
},
{
"epoch": 11.615384615384615,
"grad_norm": 0.0938439667224884,
"learning_rate": 8.134698785679909e-05,
"loss": 0.0087,
"step": 3020
},
{
"epoch": 11.653846153846153,
"grad_norm": 0.15994666516780853,
"learning_rate": 8.072144177917615e-05,
"loss": 0.0097,
"step": 3030
},
{
"epoch": 11.692307692307692,
"grad_norm": 0.16346988081932068,
"learning_rate": 8.009667823484424e-05,
"loss": 0.0087,
"step": 3040
},
{
"epoch": 11.73076923076923,
"grad_norm": 0.12906785309314728,
"learning_rate": 7.947272258349396e-05,
"loss": 0.0113,
"step": 3050
},
{
"epoch": 11.76923076923077,
"grad_norm": 0.1393740475177765,
"learning_rate": 7.884960015202289e-05,
"loss": 0.0078,
"step": 3060
},
{
"epoch": 11.807692307692308,
"grad_norm": 0.11777480691671371,
"learning_rate": 7.822733623350749e-05,
"loss": 0.0108,
"step": 3070
},
{
"epoch": 11.846153846153847,
"grad_norm": 0.11602358520030975,
"learning_rate": 7.760595608617646e-05,
"loss": 0.0078,
"step": 3080
},
{
"epoch": 11.884615384615385,
"grad_norm": 0.09884156286716461,
"learning_rate": 7.698548493238537e-05,
"loss": 0.0118,
"step": 3090
},
{
"epoch": 11.923076923076923,
"grad_norm": 0.16078071296215057,
"learning_rate": 7.636594795759307e-05,
"loss": 0.0108,
"step": 3100
},
{
"epoch": 11.961538461538462,
"grad_norm": 0.07723239809274673,
"learning_rate": 7.574737030933921e-05,
"loss": 0.0095,
"step": 3110
},
{
"epoch": 12.0,
"grad_norm": 0.1335863620042801,
"learning_rate": 7.512977709622362e-05,
"loss": 0.0108,
"step": 3120
},
{
"epoch": 12.038461538461538,
"grad_norm": 0.20832855999469757,
"learning_rate": 7.451319338688705e-05,
"loss": 0.0107,
"step": 3130
},
{
"epoch": 12.076923076923077,
"grad_norm": 0.0902862548828125,
"learning_rate": 7.389764420899348e-05,
"loss": 0.0116,
"step": 3140
},
{
"epoch": 12.115384615384615,
"grad_norm": 0.0983407124876976,
"learning_rate": 7.328315454821455e-05,
"loss": 0.0117,
"step": 3150
},
{
"epoch": 12.153846153846153,
"grad_norm": 0.10898351669311523,
"learning_rate": 7.266974934721506e-05,
"loss": 0.0087,
"step": 3160
},
{
"epoch": 12.192307692307692,
"grad_norm": 0.12344610691070557,
"learning_rate": 7.205745350464081e-05,
"loss": 0.0076,
"step": 3170
},
{
"epoch": 12.23076923076923,
"grad_norm": 0.1027444526553154,
"learning_rate": 7.144629187410756e-05,
"loss": 0.0089,
"step": 3180
},
{
"epoch": 12.26923076923077,
"grad_norm": 0.13259388506412506,
"learning_rate": 7.083628926319259e-05,
"loss": 0.0093,
"step": 3190
},
{
"epoch": 12.307692307692308,
"grad_norm": 0.15261848270893097,
"learning_rate": 7.022747043242753e-05,
"loss": 0.0088,
"step": 3200
},
{
"epoch": 12.346153846153847,
"grad_norm": 0.18796484172344208,
"learning_rate": 6.961986009429342e-05,
"loss": 0.0094,
"step": 3210
},
{
"epoch": 12.384615384615385,
"grad_norm": 0.12565913796424866,
"learning_rate": 6.901348291221737e-05,
"loss": 0.0092,
"step": 3220
},
{
"epoch": 12.423076923076923,
"grad_norm": 0.11127506196498871,
"learning_rate": 6.840836349957179e-05,
"loss": 0.0078,
"step": 3230
},
{
"epoch": 12.461538461538462,
"grad_norm": 0.17799825966358185,
"learning_rate": 6.780452641867508e-05,
"loss": 0.0098,
"step": 3240
},
{
"epoch": 12.5,
"grad_norm": 0.18342117965221405,
"learning_rate": 6.720199617979468e-05,
"loss": 0.0083,
"step": 3250
},
{
"epoch": 12.538461538461538,
"grad_norm": 0.14820097386837006,
"learning_rate": 6.660079724015226e-05,
"loss": 0.0087,
"step": 3260
},
{
"epoch": 12.576923076923077,
"grad_norm": 0.12141295522451401,
"learning_rate": 6.600095400293078e-05,
"loss": 0.0108,
"step": 3270
},
{
"epoch": 12.615384615384615,
"grad_norm": 0.10008122026920319,
"learning_rate": 6.540249081628416e-05,
"loss": 0.0097,
"step": 3280
},
{
"epoch": 12.653846153846153,
"grad_norm": 0.14060907065868378,
"learning_rate": 6.480543197234886e-05,
"loss": 0.0111,
"step": 3290
},
{
"epoch": 12.692307692307692,
"grad_norm": 0.16219834983348846,
"learning_rate": 6.420980170625788e-05,
"loss": 0.0113,
"step": 3300
},
{
"epoch": 12.73076923076923,
"grad_norm": 0.1932661533355713,
"learning_rate": 6.361562419515693e-05,
"loss": 0.0093,
"step": 3310
},
{
"epoch": 12.76923076923077,
"grad_norm": 0.14726077020168304,
"learning_rate": 6.302292355722325e-05,
"loss": 0.0099,
"step": 3320
},
{
"epoch": 12.807692307692308,
"grad_norm": 0.1557319462299347,
"learning_rate": 6.243172385068646e-05,
"loss": 0.0087,
"step": 3330
},
{
"epoch": 12.846153846153847,
"grad_norm": 0.21162110567092896,
"learning_rate": 6.184204907285215e-05,
"loss": 0.0098,
"step": 3340
},
{
"epoch": 12.884615384615385,
"grad_norm": 0.11167220026254654,
"learning_rate": 6.125392315912761e-05,
"loss": 0.0089,
"step": 3350
},
{
"epoch": 12.923076923076923,
"grad_norm": 0.10831273347139359,
"learning_rate": 6.066736998205054e-05,
"loss": 0.0088,
"step": 3360
},
{
"epoch": 12.961538461538462,
"grad_norm": 0.07814571261405945,
"learning_rate": 6.00824133503198e-05,
"loss": 0.0092,
"step": 3370
},
{
"epoch": 13.0,
"grad_norm": 0.11000088602304459,
"learning_rate": 5.94990770078292e-05,
"loss": 0.0096,
"step": 3380
},
{
"epoch": 13.038461538461538,
"grad_norm": 0.1632835417985916,
"learning_rate": 5.8917384632703534e-05,
"loss": 0.0068,
"step": 3390
},
{
"epoch": 13.076923076923077,
"grad_norm": 0.12230722606182098,
"learning_rate": 5.8337359836337526e-05,
"loss": 0.0092,
"step": 3400
},
{
"epoch": 13.115384615384615,
"grad_norm": 0.12406644970178604,
"learning_rate": 5.775902616243751e-05,
"loss": 0.0095,
"step": 3410
},
{
"epoch": 13.153846153846153,
"grad_norm": 0.10183268785476685,
"learning_rate": 5.7182407086065706e-05,
"loss": 0.0074,
"step": 3420
},
{
"epoch": 13.192307692307692,
"grad_norm": 0.11906790733337402,
"learning_rate": 5.660752601268731e-05,
"loss": 0.0064,
"step": 3430
},
{
"epoch": 13.23076923076923,
"grad_norm": 0.13318245112895966,
"learning_rate": 5.603440627722043e-05,
"loss": 0.0097,
"step": 3440
},
{
"epoch": 13.26923076923077,
"grad_norm": 0.10035967826843262,
"learning_rate": 5.546307114308902e-05,
"loss": 0.0105,
"step": 3450
},
{
"epoch": 13.307692307692308,
"grad_norm": 0.11527876555919647,
"learning_rate": 5.4893543801278514e-05,
"loss": 0.0077,
"step": 3460
},
{
"epoch": 13.346153846153847,
"grad_norm": 0.18029938638210297,
"learning_rate": 5.4325847369394486e-05,
"loss": 0.0099,
"step": 3470
},
{
"epoch": 13.384615384615385,
"grad_norm": 0.13629525899887085,
"learning_rate": 5.376000489072437e-05,
"loss": 0.0082,
"step": 3480
},
{
"epoch": 13.423076923076923,
"grad_norm": 0.0793733075261116,
"learning_rate": 5.3196039333301864e-05,
"loss": 0.0072,
"step": 3490
},
{
"epoch": 13.461538461538462,
"grad_norm": 0.14016038179397583,
"learning_rate": 5.263397358897499e-05,
"loss": 0.0101,
"step": 3500
},
{
"epoch": 13.5,
"grad_norm": 0.13849851489067078,
"learning_rate": 5.207383047247665e-05,
"loss": 0.0086,
"step": 3510
},
{
"epoch": 13.538461538461538,
"grad_norm": 0.10788647085428238,
"learning_rate": 5.1515632720498666e-05,
"loss": 0.0075,
"step": 3520
},
{
"epoch": 13.576923076923077,
"grad_norm": 0.14721888303756714,
"learning_rate": 5.095940299076874e-05,
"loss": 0.0111,
"step": 3530
},
{
"epoch": 13.615384615384615,
"grad_norm": 0.11311782151460648,
"learning_rate": 5.0405163861130925e-05,
"loss": 0.0083,
"step": 3540
},
{
"epoch": 13.653846153846153,
"grad_norm": 0.18341895937919617,
"learning_rate": 4.985293782862913e-05,
"loss": 0.0074,
"step": 3550
},
{
"epoch": 13.692307692307692,
"grad_norm": 0.10047091543674469,
"learning_rate": 4.930274730859391e-05,
"loss": 0.0076,
"step": 3560
},
{
"epoch": 13.73076923076923,
"grad_norm": 0.10870840400457382,
"learning_rate": 4.875461463373251e-05,
"loss": 0.0078,
"step": 3570
},
{
"epoch": 13.76923076923077,
"grad_norm": 0.12459917366504669,
"learning_rate": 4.820856205322257e-05,
"loss": 0.0093,
"step": 3580
},
{
"epoch": 13.807692307692308,
"grad_norm": 0.07429447025060654,
"learning_rate": 4.766461173180888e-05,
"loss": 0.0086,
"step": 3590
},
{
"epoch": 13.846153846153847,
"grad_norm": 0.08076323568820953,
"learning_rate": 4.7122785748903755e-05,
"loss": 0.0079,
"step": 3600
},
{
"epoch": 13.884615384615385,
"grad_norm": 0.09467586874961853,
"learning_rate": 4.6583106097690744e-05,
"loss": 0.0069,
"step": 3610
},
{
"epoch": 13.923076923076923,
"grad_norm": 0.10673276335000992,
"learning_rate": 4.604559468423188e-05,
"loss": 0.0084,
"step": 3620
},
{
"epoch": 13.961538461538462,
"grad_norm": 0.12460212409496307,
"learning_rate": 4.551027332657868e-05,
"loss": 0.0067,
"step": 3630
},
{
"epoch": 14.0,
"grad_norm": 0.13634030520915985,
"learning_rate": 4.4977163753886366e-05,
"loss": 0.008,
"step": 3640
},
{
"epoch": 14.038461538461538,
"grad_norm": 0.10553108155727386,
"learning_rate": 4.4446287605531826e-05,
"loss": 0.0064,
"step": 3650
},
{
"epoch": 14.076923076923077,
"grad_norm": 0.11521629989147186,
"learning_rate": 4.391766643023542e-05,
"loss": 0.0084,
"step": 3660
},
{
"epoch": 14.115384615384615,
"grad_norm": 0.10882071405649185,
"learning_rate": 4.3391321685186184e-05,
"loss": 0.0078,
"step": 3670
},
{
"epoch": 14.153846153846153,
"grad_norm": 0.061356205493211746,
"learning_rate": 4.2867274735170914e-05,
"loss": 0.0066,
"step": 3680
},
{
"epoch": 14.192307692307692,
"grad_norm": 0.12024400383234024,
"learning_rate": 4.234554685170683e-05,
"loss": 0.0062,
"step": 3690
},
{
"epoch": 14.23076923076923,
"grad_norm": 0.1351236253976822,
"learning_rate": 4.182615921217835e-05,
"loss": 0.0084,
"step": 3700
},
{
"epoch": 14.26923076923077,
"grad_norm": 0.06800480931997299,
"learning_rate": 4.130913289897732e-05,
"loss": 0.0056,
"step": 3710
},
{
"epoch": 14.307692307692308,
"grad_norm": 0.11439716070890427,
"learning_rate": 4.0794488898647365e-05,
"loss": 0.0083,
"step": 3720
},
{
"epoch": 14.346153846153847,
"grad_norm": 0.07181525975465775,
"learning_rate": 4.0282248101031884e-05,
"loss": 0.007,
"step": 3730
},
{
"epoch": 14.384615384615385,
"grad_norm": 0.08257249742746353,
"learning_rate": 3.977243129842627e-05,
"loss": 0.0067,
"step": 3740
},
{
"epoch": 14.423076923076923,
"grad_norm": 0.06434876471757889,
"learning_rate": 3.9265059184733934e-05,
"loss": 0.0074,
"step": 3750
},
{
"epoch": 14.461538461538462,
"grad_norm": 0.11237592250108719,
"learning_rate": 3.8760152354626075e-05,
"loss": 0.0065,
"step": 3760
},
{
"epoch": 14.5,
"grad_norm": 0.1452643871307373,
"learning_rate": 3.825773130270609e-05,
"loss": 0.0081,
"step": 3770
},
{
"epoch": 14.538461538461538,
"grad_norm": 0.08552118390798569,
"learning_rate": 3.7757816422677406e-05,
"loss": 0.0069,
"step": 3780
},
{
"epoch": 14.576923076923077,
"grad_norm": 0.12589380145072937,
"learning_rate": 3.7260428006515835e-05,
"loss": 0.0056,
"step": 3790
},
{
"epoch": 14.615384615384615,
"grad_norm": 0.0928504690527916,
"learning_rate": 3.6765586243645735e-05,
"loss": 0.0082,
"step": 3800
},
{
"epoch": 14.653846153846153,
"grad_norm": 0.11787436157464981,
"learning_rate": 3.6273311220120696e-05,
"loss": 0.0082,
"step": 3810
},
{
"epoch": 14.692307692307692,
"grad_norm": 0.1571059226989746,
"learning_rate": 3.578362291780818e-05,
"loss": 0.0065,
"step": 3820
},
{
"epoch": 14.73076923076923,
"grad_norm": 0.09746450930833817,
"learning_rate": 3.529654121357827e-05,
"loss": 0.0073,
"step": 3830
},
{
"epoch": 14.76923076923077,
"grad_norm": 0.10719246417284012,
"learning_rate": 3.481208587849712e-05,
"loss": 0.0075,
"step": 3840
},
{
"epoch": 14.807692307692308,
"grad_norm": 0.09190470725297928,
"learning_rate": 3.4330276577024255e-05,
"loss": 0.0062,
"step": 3850
},
{
"epoch": 14.846153846153847,
"grad_norm": 0.1003689393401146,
"learning_rate": 3.3851132866214394e-05,
"loss": 0.0063,
"step": 3860
},
{
"epoch": 14.884615384615385,
"grad_norm": 0.10899902880191803,
"learning_rate": 3.337467419492368e-05,
"loss": 0.0069,
"step": 3870
},
{
"epoch": 14.923076923076923,
"grad_norm": 0.10312393307685852,
"learning_rate": 3.290091990302011e-05,
"loss": 0.0065,
"step": 3880
},
{
"epoch": 14.961538461538462,
"grad_norm": 0.10359740257263184,
"learning_rate": 3.2429889220598655e-05,
"loss": 0.0062,
"step": 3890
},
{
"epoch": 15.0,
"grad_norm": 0.07049980014562607,
"learning_rate": 3.19616012672006e-05,
"loss": 0.0064,
"step": 3900
},
{
"epoch": 15.038461538461538,
"grad_norm": 0.06978161633014679,
"learning_rate": 3.1496075051037574e-05,
"loss": 0.0065,
"step": 3910
},
{
"epoch": 15.076923076923077,
"grad_norm": 0.16202770173549652,
"learning_rate": 3.1033329468219775e-05,
"loss": 0.01,
"step": 3920
},
{
"epoch": 15.115384615384615,
"grad_norm": 0.09229037910699844,
"learning_rate": 3.057338330198919e-05,
"loss": 0.0068,
"step": 3930
},
{
"epoch": 15.153846153846153,
"grad_norm": 0.12501412630081177,
"learning_rate": 3.0116255221957047e-05,
"loss": 0.0061,
"step": 3940
},
{
"epoch": 15.192307692307692,
"grad_norm": 0.07364345341920853,
"learning_rate": 2.9661963783346058e-05,
"loss": 0.0063,
"step": 3950
},
{
"epoch": 15.23076923076923,
"grad_norm": 0.129119873046875,
"learning_rate": 2.92105274262372e-05,
"loss": 0.007,
"step": 3960
},
{
"epoch": 15.26923076923077,
"grad_norm": 0.09468358010053635,
"learning_rate": 2.876196447482117e-05,
"loss": 0.0065,
"step": 3970
},
{
"epoch": 15.307692307692308,
"grad_norm": 0.11235329508781433,
"learning_rate": 2.8316293136654725e-05,
"loss": 0.0087,
"step": 3980
},
{
"epoch": 15.346153846153847,
"grad_norm": 0.1262669563293457,
"learning_rate": 2.7873531501921545e-05,
"loss": 0.0106,
"step": 3990
},
{
"epoch": 15.384615384615385,
"grad_norm": 0.10660561174154282,
"learning_rate": 2.7433697542697946e-05,
"loss": 0.0099,
"step": 4000
},
{
"epoch": 15.423076923076923,
"grad_norm": 0.07230859249830246,
"learning_rate": 2.6996809112223277e-05,
"loss": 0.0075,
"step": 4010
},
{
"epoch": 15.461538461538462,
"grad_norm": 0.15842293202877045,
"learning_rate": 2.6562883944175456e-05,
"loss": 0.008,
"step": 4020
},
{
"epoch": 15.5,
"grad_norm": 0.0934629887342453,
"learning_rate": 2.6131939651950966e-05,
"loss": 0.0056,
"step": 4030
},
{
"epoch": 15.538461538461538,
"grad_norm": 0.131788432598114,
"learning_rate": 2.5703993727950016e-05,
"loss": 0.0071,
"step": 4040
},
{
"epoch": 15.576923076923077,
"grad_norm": 0.09212549030780792,
"learning_rate": 2.5279063542866376e-05,
"loss": 0.0055,
"step": 4050
},
{
"epoch": 15.615384615384615,
"grad_norm": 0.1430416852235794,
"learning_rate": 2.4857166344982462e-05,
"loss": 0.0065,
"step": 4060
},
{
"epoch": 15.653846153846153,
"grad_norm": 0.08664490282535553,
"learning_rate": 2.44383192594691e-05,
"loss": 0.0059,
"step": 4070
},
{
"epoch": 15.692307692307692,
"grad_norm": 0.08127816766500473,
"learning_rate": 2.4022539287690416e-05,
"loss": 0.0066,
"step": 4080
},
{
"epoch": 15.73076923076923,
"grad_norm": 0.10413156449794769,
"learning_rate": 2.360984330651378e-05,
"loss": 0.0069,
"step": 4090
},
{
"epoch": 15.76923076923077,
"grad_norm": 0.08507948368787766,
"learning_rate": 2.3200248067624654e-05,
"loss": 0.0071,
"step": 4100
},
{
"epoch": 15.807692307692308,
"grad_norm": 0.0667034238576889,
"learning_rate": 2.2793770196846743e-05,
"loss": 0.0061,
"step": 4110
},
{
"epoch": 15.846153846153847,
"grad_norm": 0.09667922556400299,
"learning_rate": 2.239042619346713e-05,
"loss": 0.0075,
"step": 4120
},
{
"epoch": 15.884615384615385,
"grad_norm": 0.0950574055314064,
"learning_rate": 2.1990232429566484e-05,
"loss": 0.005,
"step": 4130
},
{
"epoch": 15.923076923076923,
"grad_norm": 0.09804387390613556,
"learning_rate": 2.159320514935449e-05,
"loss": 0.0073,
"step": 4140
},
{
"epoch": 15.961538461538462,
"grad_norm": 0.06155986711382866,
"learning_rate": 2.119936046851062e-05,
"loss": 0.0084,
"step": 4150
},
{
"epoch": 16.0,
"grad_norm": 0.12021949142217636,
"learning_rate": 2.080871437352985e-05,
"loss": 0.0056,
"step": 4160
},
{
"epoch": 16.03846153846154,
"grad_norm": 0.11103606224060059,
"learning_rate": 2.042128272107382e-05,
"loss": 0.0074,
"step": 4170
},
{
"epoch": 16.076923076923077,
"grad_norm": 0.07725587487220764,
"learning_rate": 2.0037081237327125e-05,
"loss": 0.0073,
"step": 4180
},
{
"epoch": 16.115384615384617,
"grad_norm": 0.081053227186203,
"learning_rate": 1.965612551735908e-05,
"loss": 0.0053,
"step": 4190
},
{
"epoch": 16.153846153846153,
"grad_norm": 0.09564225375652313,
"learning_rate": 1.927843102449065e-05,
"loss": 0.0059,
"step": 4200
},
{
"epoch": 16.192307692307693,
"grad_norm": 0.10796211659908295,
"learning_rate": 1.890401308966676e-05,
"loss": 0.0052,
"step": 4210
},
{
"epoch": 16.23076923076923,
"grad_norm": 0.05812790244817734,
"learning_rate": 1.853288691083407e-05,
"loss": 0.0054,
"step": 4220
},
{
"epoch": 16.26923076923077,
"grad_norm": 0.07313115149736404,
"learning_rate": 1.816506755232391e-05,
"loss": 0.006,
"step": 4230
},
{
"epoch": 16.307692307692307,
"grad_norm": 0.07596628367900848,
"learning_rate": 1.7800569944241053e-05,
"loss": 0.0054,
"step": 4240
},
{
"epoch": 16.346153846153847,
"grad_norm": 0.10533738136291504,
"learning_rate": 1.74394088818575e-05,
"loss": 0.007,
"step": 4250
},
{
"epoch": 16.384615384615383,
"grad_norm": 0.0967307761311531,
"learning_rate": 1.708159902501201e-05,
"loss": 0.0056,
"step": 4260
},
{
"epoch": 16.423076923076923,
"grad_norm": 0.06359889358282089,
"learning_rate": 1.672715489751496e-05,
"loss": 0.006,
"step": 4270
},
{
"epoch": 16.46153846153846,
"grad_norm": 0.08679856359958649,
"learning_rate": 1.637609088655896e-05,
"loss": 0.0053,
"step": 4280
},
{
"epoch": 16.5,
"grad_norm": 0.07200346887111664,
"learning_rate": 1.6028421242134707e-05,
"loss": 0.0051,
"step": 4290
},
{
"epoch": 16.53846153846154,
"grad_norm": 0.06287889182567596,
"learning_rate": 1.568416007645269e-05,
"loss": 0.0055,
"step": 4300
},
{
"epoch": 16.576923076923077,
"grad_norm": 0.08229623734951019,
"learning_rate": 1.53433213633703e-05,
"loss": 0.0062,
"step": 4310
},
{
"epoch": 16.615384615384617,
"grad_norm": 0.060530465096235275,
"learning_rate": 1.5005918937824548e-05,
"loss": 0.0067,
"step": 4320
},
{
"epoch": 16.653846153846153,
"grad_norm": 0.094852514564991,
"learning_rate": 1.4671966495270672e-05,
"loss": 0.0057,
"step": 4330
},
{
"epoch": 16.692307692307693,
"grad_norm": 0.09454119205474854,
"learning_rate": 1.4341477591126096e-05,
"loss": 0.0055,
"step": 4340
},
{
"epoch": 16.73076923076923,
"grad_norm": 0.06650886684656143,
"learning_rate": 1.4014465640220243e-05,
"loss": 0.0067,
"step": 4350
},
{
"epoch": 16.76923076923077,
"grad_norm": 0.0640147477388382,
"learning_rate": 1.3690943916249943e-05,
"loss": 0.0056,
"step": 4360
},
{
"epoch": 16.807692307692307,
"grad_norm": 0.0917372778058052,
"learning_rate": 1.3370925551240798e-05,
"loss": 0.006,
"step": 4370
},
{
"epoch": 16.846153846153847,
"grad_norm": 0.09550613164901733,
"learning_rate": 1.3054423535014005e-05,
"loss": 0.0063,
"step": 4380
},
{
"epoch": 16.884615384615383,
"grad_norm": 0.0998285710811615,
"learning_rate": 1.2741450714659153e-05,
"loss": 0.0055,
"step": 4390
},
{
"epoch": 16.923076923076923,
"grad_norm": 0.16476678848266602,
"learning_rate": 1.2432019794012684e-05,
"loss": 0.0069,
"step": 4400
},
{
"epoch": 16.96153846153846,
"grad_norm": 0.18762223422527313,
"learning_rate": 1.2126143333142337e-05,
"loss": 0.007,
"step": 4410
},
{
"epoch": 17.0,
"grad_norm": 0.07807065546512604,
"learning_rate": 1.1823833747837233e-05,
"loss": 0.006,
"step": 4420
},
{
"epoch": 17.03846153846154,
"grad_norm": 0.086389921605587,
"learning_rate": 1.1525103309103934e-05,
"loss": 0.0046,
"step": 4430
},
{
"epoch": 17.076923076923077,
"grad_norm": 0.09868927299976349,
"learning_rate": 1.1229964142668425e-05,
"loss": 0.0061,
"step": 4440
},
{
"epoch": 17.115384615384617,
"grad_norm": 0.06974314898252487,
"learning_rate": 1.0938428228483733e-05,
"loss": 0.0049,
"step": 4450
},
{
"epoch": 17.153846153846153,
"grad_norm": 0.06945253908634186,
"learning_rate": 1.0650507400243859e-05,
"loss": 0.0065,
"step": 4460
},
{
"epoch": 17.192307692307693,
"grad_norm": 0.08085139840841293,
"learning_rate": 1.0366213344903342e-05,
"loss": 0.005,
"step": 4470
},
{
"epoch": 17.23076923076923,
"grad_norm": 0.08081810176372528,
"learning_rate": 1.00855576022029e-05,
"loss": 0.0051,
"step": 4480
},
{
"epoch": 17.26923076923077,
"grad_norm": 0.04182474687695503,
"learning_rate": 9.808551564200908e-06,
"loss": 0.0056,
"step": 4490
},
{
"epoch": 17.307692307692307,
"grad_norm": 0.1365455687046051,
"learning_rate": 9.535206474811198e-06,
"loss": 0.0068,
"step": 4500
},
{
"epoch": 17.346153846153847,
"grad_norm": 0.10072525590658188,
"learning_rate": 9.265533429346484e-06,
"loss": 0.0058,
"step": 4510
},
{
"epoch": 17.384615384615383,
"grad_norm": 0.0997309759259224,
"learning_rate": 8.999543374068098e-06,
"loss": 0.0055,
"step": 4520
},
{
"epoch": 17.423076923076923,
"grad_norm": 0.05286664515733719,
"learning_rate": 8.737247105741542e-06,
"loss": 0.0036,
"step": 4530
},
{
"epoch": 17.46153846153846,
"grad_norm": 0.0948430672287941,
"learning_rate": 8.478655271198389e-06,
"loss": 0.0057,
"step": 4540
},
{
"epoch": 17.5,
"grad_norm": 0.043210502713918686,
"learning_rate": 8.223778366904056e-06,
"loss": 0.0058,
"step": 4550
},
{
"epoch": 17.53846153846154,
"grad_norm": 0.08293162286281586,
"learning_rate": 7.972626738531708e-06,
"loss": 0.0057,
"step": 4560
},
{
"epoch": 17.576923076923077,
"grad_norm": 0.049041640013456345,
"learning_rate": 7.725210580542387e-06,
"loss": 0.0049,
"step": 4570
},
{
"epoch": 17.615384615384617,
"grad_norm": 0.08410376310348511,
"learning_rate": 7.48153993577112e-06,
"loss": 0.005,
"step": 4580
},
{
"epoch": 17.653846153846153,
"grad_norm": 0.07442956417798996,
"learning_rate": 7.241624695019366e-06,
"loss": 0.0046,
"step": 4590
},
{
"epoch": 17.692307692307693,
"grad_norm": 0.04315830394625664,
"learning_rate": 7.005474596653494e-06,
"loss": 0.0063,
"step": 4600
},
{
"epoch": 17.73076923076923,
"grad_norm": 0.10330154001712799,
"learning_rate": 6.773099226209501e-06,
"loss": 0.0056,
"step": 4610
},
{
"epoch": 17.76923076923077,
"grad_norm": 0.0930248498916626,
"learning_rate": 6.544508016003903e-06,
"loss": 0.0058,
"step": 4620
},
{
"epoch": 17.807692307692307,
"grad_norm": 0.22107692062854767,
"learning_rate": 6.319710244750898e-06,
"loss": 0.0058,
"step": 4630
},
{
"epoch": 17.846153846153847,
"grad_norm": 0.07021880894899368,
"learning_rate": 6.0987150371857534e-06,
"loss": 0.005,
"step": 4640
},
{
"epoch": 17.884615384615383,
"grad_norm": 0.12697994709014893,
"learning_rate": 5.88153136369437e-06,
"loss": 0.0053,
"step": 4650
},
{
"epoch": 17.923076923076923,
"grad_norm": 0.0982799381017685,
"learning_rate": 5.668168039949206e-06,
"loss": 0.006,
"step": 4660
},
{
"epoch": 17.96153846153846,
"grad_norm": 0.07556256651878357,
"learning_rate": 5.458633726551443e-06,
"loss": 0.0052,
"step": 4670
},
{
"epoch": 18.0,
"grad_norm": 0.0795573741197586,
"learning_rate": 5.252936928679431e-06,
"loss": 0.0057,
"step": 4680
},
{
"epoch": 18.03846153846154,
"grad_norm": 0.0670771524310112,
"learning_rate": 5.051085995743421e-06,
"loss": 0.0076,
"step": 4690
},
{
"epoch": 18.076923076923077,
"grad_norm": 0.11251858621835709,
"learning_rate": 4.853089121046739e-06,
"loss": 0.0065,
"step": 4700
},
{
"epoch": 18.115384615384617,
"grad_norm": 0.10217051953077316,
"learning_rate": 4.6589543414531364e-06,
"loss": 0.0043,
"step": 4710
},
{
"epoch": 18.153846153846153,
"grad_norm": 0.050674889236688614,
"learning_rate": 4.468689537060633e-06,
"loss": 0.0049,
"step": 4720
},
{
"epoch": 18.192307692307693,
"grad_norm": 0.053426776081323624,
"learning_rate": 4.282302430881568e-06,
"loss": 0.0059,
"step": 4730
},
{
"epoch": 18.23076923076923,
"grad_norm": 0.09936069697141647,
"learning_rate": 4.099800588529212e-06,
"loss": 0.0057,
"step": 4740
},
{
"epoch": 18.26923076923077,
"grad_norm": 0.1067737266421318,
"learning_rate": 3.921191417910608e-06,
"loss": 0.0048,
"step": 4750
},
{
"epoch": 18.307692307692307,
"grad_norm": 0.058368340134620667,
"learning_rate": 3.746482168925902e-06,
"loss": 0.0068,
"step": 4760
},
{
"epoch": 18.346153846153847,
"grad_norm": 0.10528615117073059,
"learning_rate": 3.5756799331740588e-06,
"loss": 0.0064,
"step": 4770
},
{
"epoch": 18.384615384615383,
"grad_norm": 0.0884661078453064,
"learning_rate": 3.4087916436650104e-06,
"loss": 0.0066,
"step": 4780
},
{
"epoch": 18.423076923076923,
"grad_norm": 0.03825301676988602,
"learning_rate": 3.2458240745382617e-06,
"loss": 0.005,
"step": 4790
},
{
"epoch": 18.46153846153846,
"grad_norm": 0.08059193193912506,
"learning_rate": 3.0867838407878436e-06,
"loss": 0.0064,
"step": 4800
},
{
"epoch": 18.5,
"grad_norm": 0.0991053432226181,
"learning_rate": 2.931677397993915e-06,
"loss": 0.0048,
"step": 4810
},
{
"epoch": 18.53846153846154,
"grad_norm": 0.05811507999897003,
"learning_rate": 2.7805110420606407e-06,
"loss": 0.0073,
"step": 4820
},
{
"epoch": 18.576923076923077,
"grad_norm": 0.10655184090137482,
"learning_rate": 2.633290908960684e-06,
"loss": 0.004,
"step": 4830
},
{
"epoch": 18.615384615384617,
"grad_norm": 0.07305694371461868,
"learning_rate": 2.490022974486095e-06,
"loss": 0.0053,
"step": 4840
},
{
"epoch": 18.653846153846153,
"grad_norm": 0.06774792075157166,
"learning_rate": 2.350713054005793e-06,
"loss": 0.0049,
"step": 4850
},
{
"epoch": 18.692307692307693,
"grad_norm": 0.049805670976638794,
"learning_rate": 2.2153668022295015e-06,
"loss": 0.0044,
"step": 4860
},
{
"epoch": 18.73076923076923,
"grad_norm": 0.048138655722141266,
"learning_rate": 2.0839897129782183e-06,
"loss": 0.0055,
"step": 4870
},
{
"epoch": 18.76923076923077,
"grad_norm": 0.090663842856884,
"learning_rate": 1.956587118961195e-06,
"loss": 0.0047,
"step": 4880
},
{
"epoch": 18.807692307692307,
"grad_norm": 0.11241704970598221,
"learning_rate": 1.833164191559511e-06,
"loss": 0.0044,
"step": 4890
},
{
"epoch": 18.846153846153847,
"grad_norm": 0.04880325496196747,
"learning_rate": 1.713725940616151e-06,
"loss": 0.0038,
"step": 4900
},
{
"epoch": 18.884615384615383,
"grad_norm": 0.04214165359735489,
"learning_rate": 1.5982772142326462e-06,
"loss": 0.0043,
"step": 4910
},
{
"epoch": 18.923076923076923,
"grad_norm": 0.07064809650182724,
"learning_rate": 1.4868226985722877e-06,
"loss": 0.0056,
"step": 4920
},
{
"epoch": 18.96153846153846,
"grad_norm": 0.06411836296319962,
"learning_rate": 1.3793669176698887e-06,
"loss": 0.0056,
"step": 4930
},
{
"epoch": 19.0,
"grad_norm": 0.10901802033185959,
"learning_rate": 1.2759142332482099e-06,
"loss": 0.0057,
"step": 4940
},
{
"epoch": 19.03846153846154,
"grad_norm": 0.05795470252633095,
"learning_rate": 1.1764688445408677e-06,
"loss": 0.0049,
"step": 4950
},
{
"epoch": 19.076923076923077,
"grad_norm": 0.03703946992754936,
"learning_rate": 1.0810347881218708e-06,
"loss": 0.0059,
"step": 4960
},
{
"epoch": 19.115384615384617,
"grad_norm": 0.05625122785568237,
"learning_rate": 9.896159377418057e-07,
"loss": 0.0065,
"step": 4970
},
{
"epoch": 19.153846153846153,
"grad_norm": 0.08023211359977722,
"learning_rate": 9.022160041705974e-07,
"loss": 0.0053,
"step": 4980
},
{
"epoch": 19.192307692307693,
"grad_norm": 0.06640934944152832,
"learning_rate": 8.188385350468508e-07,
"loss": 0.0046,
"step": 4990
},
{
"epoch": 19.23076923076923,
"grad_norm": 0.10493505746126175,
"learning_rate": 7.394869147339112e-07,
"loss": 0.0061,
"step": 5000
},
{
"epoch": 19.26923076923077,
"grad_norm": 0.09153269976377487,
"learning_rate": 6.641643641823958e-07,
"loss": 0.0067,
"step": 5010
},
{
"epoch": 19.307692307692307,
"grad_norm": 0.06915758550167084,
"learning_rate": 5.928739407995543e-07,
"loss": 0.0052,
"step": 5020
},
{
"epoch": 19.346153846153847,
"grad_norm": 0.08392194658517838,
"learning_rate": 5.256185383251011e-07,
"loss": 0.0061,
"step": 5030
},
{
"epoch": 19.384615384615383,
"grad_norm": 0.04674766585230827,
"learning_rate": 4.6240088671378746e-07,
"loss": 0.0046,
"step": 5040
},
{
"epoch": 19.423076923076923,
"grad_norm": 0.052531130611896515,
"learning_rate": 4.0322355202455644e-07,
"loss": 0.006,
"step": 5050
},
{
"epoch": 19.46153846153846,
"grad_norm": 0.062260791659355164,
"learning_rate": 3.4808893631642636e-07,
"loss": 0.0043,
"step": 5060
},
{
"epoch": 19.5,
"grad_norm": 0.06215400993824005,
"learning_rate": 2.969992775509578e-07,
"loss": 0.0045,
"step": 5070
},
{
"epoch": 19.53846153846154,
"grad_norm": 0.07574478536844254,
"learning_rate": 2.4995664950141493e-07,
"loss": 0.0038,
"step": 5080
},
{
"epoch": 19.576923076923077,
"grad_norm": 0.05041692778468132,
"learning_rate": 2.0696296166862194e-07,
"loss": 0.0038,
"step": 5090
},
{
"epoch": 19.615384615384617,
"grad_norm": 0.10670372098684311,
"learning_rate": 1.6801995920341375e-07,
"loss": 0.0049,
"step": 5100
},
{
"epoch": 19.653846153846153,
"grad_norm": 0.05979029834270477,
"learning_rate": 1.3312922283581496e-07,
"loss": 0.0048,
"step": 5110
},
{
"epoch": 19.692307692307693,
"grad_norm": 0.05246563255786896,
"learning_rate": 1.0229216881088022e-07,
"loss": 0.0047,
"step": 5120
},
{
"epoch": 19.73076923076923,
"grad_norm": 0.08008202165365219,
"learning_rate": 7.551004883120661e-08,
"loss": 0.0045,
"step": 5130
},
{
"epoch": 19.76923076923077,
"grad_norm": 0.04513508453965187,
"learning_rate": 5.278395000611891e-08,
"loss": 0.0047,
"step": 5140
},
{
"epoch": 19.807692307692307,
"grad_norm": 0.07471077144145966,
"learning_rate": 3.4114794807571514e-08,
"loss": 0.0056,
"step": 5150
},
{
"epoch": 19.846153846153847,
"grad_norm": 0.07386625558137894,
"learning_rate": 1.9503341032633958e-08,
"loss": 0.0052,
"step": 5160
},
{
"epoch": 19.884615384615383,
"grad_norm": 0.05373087525367737,
"learning_rate": 8.950181772826582e-09,
"loss": 0.0062,
"step": 5170
},
{
"epoch": 19.923076923076923,
"grad_norm": 0.0932120531797409,
"learning_rate": 2.455745389962072e-09,
"loss": 0.0047,
"step": 5180
},
{
"epoch": 19.96153846153846,
"grad_norm": 0.08532081544399261,
"learning_rate": 2.0295498825984027e-11,
"loss": 0.0037,
"step": 5190
},
{
"epoch": 19.965384615384615,
"step": 5191,
"total_flos": 7.394345918457876e+17,
"train_loss": 0.02140558246369566,
"train_runtime": 6034.0474,
"train_samples_per_second": 55.058,
"train_steps_per_second": 0.86
}
],
"logging_steps": 10,
"max_steps": 5191,
"num_input_tokens_seen": 0,
"num_train_epochs": 20,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7.394345918457876e+17,
"train_batch_size": 64,
"trial_name": null,
"trial_params": null
}