{ "best_metric": null, "best_model_checkpoint": null, "epoch": 19.89010989010989, "eval_steps": 500, "global_step": 1810, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.10989010989010989, "grad_norm": 6.35939884185791, "learning_rate": 2.1978021978021977e-05, "loss": 1.3933, "step": 10 }, { "epoch": 0.21978021978021978, "grad_norm": 3.3019938468933105, "learning_rate": 4.3956043956043955e-05, "loss": 0.6013, "step": 20 }, { "epoch": 0.32967032967032966, "grad_norm": 1.7010605335235596, "learning_rate": 6.593406593406594e-05, "loss": 0.3661, "step": 30 }, { "epoch": 0.43956043956043955, "grad_norm": 1.9770668745040894, "learning_rate": 8.791208791208791e-05, "loss": 0.2746, "step": 40 }, { "epoch": 0.5494505494505495, "grad_norm": 1.2615512609481812, "learning_rate": 0.0001098901098901099, "loss": 0.2108, "step": 50 }, { "epoch": 0.6593406593406593, "grad_norm": 0.9937906265258789, "learning_rate": 0.00013186813186813188, "loss": 0.1888, "step": 60 }, { "epoch": 0.7692307692307693, "grad_norm": 0.7049073576927185, "learning_rate": 0.00015384615384615385, "loss": 0.166, "step": 70 }, { "epoch": 0.8791208791208791, "grad_norm": 0.6901200413703918, "learning_rate": 0.00017582417582417582, "loss": 0.1526, "step": 80 }, { "epoch": 0.989010989010989, "grad_norm": 1.3578896522521973, "learning_rate": 0.0001978021978021978, "loss": 0.1373, "step": 90 }, { "epoch": 1.098901098901099, "grad_norm": 0.9918624758720398, "learning_rate": 0.00019998647325745995, "loss": 0.1303, "step": 100 }, { "epoch": 1.2087912087912087, "grad_norm": 0.9448220133781433, "learning_rate": 0.00019993971884561357, "loss": 0.1236, "step": 110 }, { "epoch": 1.3186813186813187, "grad_norm": 0.5491310358047485, "learning_rate": 0.00019985958537951462, "loss": 0.1164, "step": 120 }, { "epoch": 1.4285714285714286, "grad_norm": 0.6634742617607117, "learning_rate": 0.00019974609962308986, "loss": 0.1024, "step": 130 }, { "epoch": 1.5384615384615383, "grad_norm": 0.652184784412384, "learning_rate": 0.00019959929947966, "loss": 0.108, "step": 140 }, { "epoch": 1.6483516483516483, "grad_norm": 0.42199188470840454, "learning_rate": 0.00019941923397928047, "loss": 0.0963, "step": 150 }, { "epoch": 1.7582417582417582, "grad_norm": 0.7065568566322327, "learning_rate": 0.0001992059632623657, "loss": 0.099, "step": 160 }, { "epoch": 1.8681318681318682, "grad_norm": 0.6244102120399475, "learning_rate": 0.0001989595585596026, "loss": 0.0897, "step": 170 }, { "epoch": 1.978021978021978, "grad_norm": 0.6415707468986511, "learning_rate": 0.00019868010216816034, "loss": 0.0781, "step": 180 }, { "epoch": 2.087912087912088, "grad_norm": 0.3453947603702545, "learning_rate": 0.00019836768742420352, "loss": 0.079, "step": 190 }, { "epoch": 2.197802197802198, "grad_norm": 0.37823784351348877, "learning_rate": 0.00019802241867171878, "loss": 0.075, "step": 200 }, { "epoch": 2.3076923076923075, "grad_norm": 0.8650999069213867, "learning_rate": 0.00019764441122766476, "loss": 0.0796, "step": 210 }, { "epoch": 2.4175824175824174, "grad_norm": 0.4968554973602295, "learning_rate": 0.000197233791343457, "loss": 0.0754, "step": 220 }, { "epoch": 2.5274725274725274, "grad_norm": 0.5782051682472229, "learning_rate": 0.00019679069616280133, "loss": 0.076, "step": 230 }, { "epoch": 2.6373626373626373, "grad_norm": 0.4181021749973297, "learning_rate": 0.00019631527367588864, "loss": 0.0769, "step": 240 }, { "epoch": 2.7472527472527473, "grad_norm": 0.41538622975349426, "learning_rate": 0.00019580768266996756, "loss": 0.0653, "step": 250 }, { "epoch": 2.857142857142857, "grad_norm": 0.3886262774467468, "learning_rate": 0.00019526809267631073, "loss": 0.0723, "step": 260 }, { "epoch": 2.967032967032967, "grad_norm": 0.466183602809906, "learning_rate": 0.0001946966839135928, "loss": 0.0702, "step": 270 }, { "epoch": 3.076923076923077, "grad_norm": 0.34900644421577454, "learning_rate": 0.00019409364722769882, "loss": 0.0638, "step": 280 }, { "epoch": 3.186813186813187, "grad_norm": 0.5098071694374084, "learning_rate": 0.00019345918402798314, "loss": 0.0645, "step": 290 }, { "epoch": 3.2967032967032965, "grad_norm": 0.45593976974487305, "learning_rate": 0.00019279350622000054, "loss": 0.0691, "step": 300 }, { "epoch": 3.4065934065934065, "grad_norm": 0.3932492733001709, "learning_rate": 0.0001920968361347314, "loss": 0.0706, "step": 310 }, { "epoch": 3.5164835164835164, "grad_norm": 0.3770773708820343, "learning_rate": 0.00019136940645432494, "loss": 0.0646, "step": 320 }, { "epoch": 3.6263736263736264, "grad_norm": 0.432841420173645, "learning_rate": 0.00019061146013438527, "loss": 0.0557, "step": 330 }, { "epoch": 3.7362637362637363, "grad_norm": 0.5559947490692139, "learning_rate": 0.00018982325032282615, "loss": 0.0729, "step": 340 }, { "epoch": 3.8461538461538463, "grad_norm": 0.5822933912277222, "learning_rate": 0.00018900504027532185, "loss": 0.0632, "step": 350 }, { "epoch": 3.956043956043956, "grad_norm": 0.5372587442398071, "learning_rate": 0.0001881571032673816, "loss": 0.0666, "step": 360 }, { "epoch": 4.065934065934066, "grad_norm": 0.4164665937423706, "learning_rate": 0.000187279722503078, "loss": 0.0566, "step": 370 }, { "epoch": 4.175824175824176, "grad_norm": 0.570026695728302, "learning_rate": 0.00018637319102045912, "loss": 0.0621, "step": 380 }, { "epoch": 4.285714285714286, "grad_norm": 0.3998757302761078, "learning_rate": 0.00018543781159367607, "loss": 0.0587, "step": 390 }, { "epoch": 4.395604395604396, "grad_norm": 0.3779265880584717, "learning_rate": 0.00018447389663185906, "loss": 0.0568, "step": 400 }, { "epoch": 4.5054945054945055, "grad_norm": 0.4073057174682617, "learning_rate": 0.0001834817680747751, "loss": 0.0615, "step": 410 }, { "epoch": 4.615384615384615, "grad_norm": 0.3638319969177246, "learning_rate": 0.00018246175728530296, "loss": 0.0548, "step": 420 }, { "epoch": 4.725274725274725, "grad_norm": 0.2937782108783722, "learning_rate": 0.00018141420493876035, "loss": 0.0601, "step": 430 }, { "epoch": 4.835164835164835, "grad_norm": 0.4016660749912262, "learning_rate": 0.00018033946090912148, "loss": 0.0563, "step": 440 }, { "epoch": 4.945054945054945, "grad_norm": 0.39952585101127625, "learning_rate": 0.00017923788415216175, "loss": 0.0554, "step": 450 }, { "epoch": 5.054945054945055, "grad_norm": 0.3408401906490326, "learning_rate": 0.00017810984258556957, "loss": 0.0515, "step": 460 }, { "epoch": 5.164835164835165, "grad_norm": 0.3611335754394531, "learning_rate": 0.00017695571296606465, "loss": 0.0511, "step": 470 }, { "epoch": 5.274725274725275, "grad_norm": 0.24698808789253235, "learning_rate": 0.00017577588076356465, "loss": 0.0531, "step": 480 }, { "epoch": 5.384615384615385, "grad_norm": 0.40450191497802734, "learning_rate": 0.000174570740032441, "loss": 0.048, "step": 490 }, { "epoch": 5.4945054945054945, "grad_norm": 0.44147250056266785, "learning_rate": 0.00017334069327990816, "loss": 0.0545, "step": 500 }, { "epoch": 5.604395604395604, "grad_norm": 0.45513996481895447, "learning_rate": 0.00017208615133158946, "loss": 0.0505, "step": 510 }, { "epoch": 5.714285714285714, "grad_norm": 0.3807731568813324, "learning_rate": 0.00017080753319430452, "loss": 0.0509, "step": 520 }, { "epoch": 5.824175824175824, "grad_norm": 0.288381427526474, "learning_rate": 0.00016950526591612462, "loss": 0.0493, "step": 530 }, { "epoch": 5.934065934065934, "grad_norm": 0.234198197722435, "learning_rate": 0.00016817978444374194, "loss": 0.0552, "step": 540 }, { "epoch": 6.043956043956044, "grad_norm": 0.33030226826667786, "learning_rate": 0.00016683153147720097, "loss": 0.0453, "step": 550 }, { "epoch": 6.153846153846154, "grad_norm": 0.30306366086006165, "learning_rate": 0.00016546095732204022, "loss": 0.0521, "step": 560 }, { "epoch": 6.263736263736264, "grad_norm": 0.32483500242233276, "learning_rate": 0.00016406851973889392, "loss": 0.0473, "step": 570 }, { "epoch": 6.373626373626374, "grad_norm": 0.3211219310760498, "learning_rate": 0.00016265468379060365, "loss": 0.0527, "step": 580 }, { "epoch": 6.483516483516484, "grad_norm": 0.2360507994890213, "learning_rate": 0.00016121992168689108, "loss": 0.0449, "step": 590 }, { "epoch": 6.593406593406593, "grad_norm": 0.3557351231575012, "learning_rate": 0.00015976471262664382, "loss": 0.0471, "step": 600 }, { "epoch": 6.7032967032967035, "grad_norm": 0.3993636667728424, "learning_rate": 0.00015828954263786687, "loss": 0.0461, "step": 610 }, { "epoch": 6.813186813186813, "grad_norm": 0.4391331076622009, "learning_rate": 0.00015679490441535324, "loss": 0.0431, "step": 620 }, { "epoch": 6.923076923076923, "grad_norm": 0.311146080493927, "learning_rate": 0.0001552812971561278, "loss": 0.0497, "step": 630 }, { "epoch": 7.032967032967033, "grad_norm": 0.29637405276298523, "learning_rate": 0.00015374922639271963, "loss": 0.0466, "step": 640 }, { "epoch": 7.142857142857143, "grad_norm": 0.24265693128108978, "learning_rate": 0.0001521992038243182, "loss": 0.048, "step": 650 }, { "epoch": 7.252747252747253, "grad_norm": 0.3331180214881897, "learning_rate": 0.0001506317471458701, "loss": 0.0444, "step": 660 }, { "epoch": 7.362637362637362, "grad_norm": 0.29099053144454956, "learning_rate": 0.00014904737987517293, "loss": 0.0446, "step": 670 }, { "epoch": 7.472527472527473, "grad_norm": 0.26169735193252563, "learning_rate": 0.00014744663117802475, "loss": 0.0442, "step": 680 }, { "epoch": 7.582417582417582, "grad_norm": 0.27879634499549866, "learning_rate": 0.00014583003569148712, "loss": 0.0424, "step": 690 }, { "epoch": 7.6923076923076925, "grad_norm": 0.3208893835544586, "learning_rate": 0.00014419813334532036, "loss": 0.0421, "step": 700 }, { "epoch": 7.802197802197802, "grad_norm": 0.2651744484901428, "learning_rate": 0.0001425514691816516, "loss": 0.046, "step": 710 }, { "epoch": 7.912087912087912, "grad_norm": 0.30851149559020996, "learning_rate": 0.00014089059317293524, "loss": 0.046, "step": 720 }, { "epoch": 8.021978021978022, "grad_norm": 0.32210397720336914, "learning_rate": 0.00013921606003826627, "loss": 0.0448, "step": 730 }, { "epoch": 8.131868131868131, "grad_norm": 0.49959078431129456, "learning_rate": 0.00013752842905810896, "loss": 0.0447, "step": 740 }, { "epoch": 8.241758241758241, "grad_norm": 0.5785067677497864, "learning_rate": 0.00013582826388750153, "loss": 0.0397, "step": 750 }, { "epoch": 8.351648351648352, "grad_norm": 0.28836700320243835, "learning_rate": 0.00013411613236779995, "loss": 0.0493, "step": 760 }, { "epoch": 8.461538461538462, "grad_norm": 0.26861655712127686, "learning_rate": 0.00013239260633702364, "loss": 0.0442, "step": 770 }, { "epoch": 8.571428571428571, "grad_norm": 0.3270936608314514, "learning_rate": 0.00013065826143886616, "loss": 0.0489, "step": 780 }, { "epoch": 8.68131868131868, "grad_norm": 0.24692483246326447, "learning_rate": 0.00012891367693043478, "loss": 0.0441, "step": 790 }, { "epoch": 8.791208791208792, "grad_norm": 0.32261839509010315, "learning_rate": 0.00012715943548878355, "loss": 0.042, "step": 800 }, { "epoch": 8.901098901098901, "grad_norm": 0.22817495465278625, "learning_rate": 0.00012539612301630377, "loss": 0.0415, "step": 810 }, { "epoch": 9.010989010989011, "grad_norm": 0.16383874416351318, "learning_rate": 0.00012362432844503725, "loss": 0.0366, "step": 820 }, { "epoch": 9.12087912087912, "grad_norm": 0.35975322127342224, "learning_rate": 0.00012184464353997796, "loss": 0.0424, "step": 830 }, { "epoch": 9.23076923076923, "grad_norm": 0.22110863029956818, "learning_rate": 0.00012005766270142724, "loss": 0.0334, "step": 840 }, { "epoch": 9.340659340659341, "grad_norm": 0.26277679204940796, "learning_rate": 0.00011826398276646897, "loss": 0.041, "step": 850 }, { "epoch": 9.45054945054945, "grad_norm": 0.29648274183273315, "learning_rate": 0.00011646420280963081, "loss": 0.0358, "step": 860 }, { "epoch": 9.56043956043956, "grad_norm": 0.27250102162361145, "learning_rate": 0.00011465892394279814, "loss": 0.0402, "step": 870 }, { "epoch": 9.67032967032967, "grad_norm": 0.24232593178749084, "learning_rate": 0.00011284874911444763, "loss": 0.0372, "step": 880 }, { "epoch": 9.780219780219781, "grad_norm": 0.25964832305908203, "learning_rate": 0.00011103428290826736, "loss": 0.0427, "step": 890 }, { "epoch": 9.89010989010989, "grad_norm": 0.31330835819244385, "learning_rate": 0.00010921613134123068, "loss": 0.0397, "step": 900 }, { "epoch": 10.0, "grad_norm": 0.3337639570236206, "learning_rate": 0.00010739490166119155, "loss": 0.0363, "step": 910 }, { "epoch": 10.10989010989011, "grad_norm": 0.1793065071105957, "learning_rate": 0.00010557120214406875, "loss": 0.0341, "step": 920 }, { "epoch": 10.219780219780219, "grad_norm": 0.32348737120628357, "learning_rate": 0.00010374564189068641, "loss": 0.0381, "step": 930 }, { "epoch": 10.32967032967033, "grad_norm": 0.3930850625038147, "learning_rate": 0.00010191883062333965, "loss": 0.0398, "step": 940 }, { "epoch": 10.43956043956044, "grad_norm": 0.2554318308830261, "learning_rate": 0.00010009137848215195, "loss": 0.034, "step": 950 }, { "epoch": 10.54945054945055, "grad_norm": 0.2729535698890686, "learning_rate": 9.826389582129352e-05, "loss": 0.0408, "step": 960 }, { "epoch": 10.659340659340659, "grad_norm": 0.2841566503047943, "learning_rate": 9.64369930051278e-05, "loss": 0.0373, "step": 970 }, { "epoch": 10.76923076923077, "grad_norm": 0.18831486999988556, "learning_rate": 9.4611280204355e-05, "loss": 0.0354, "step": 980 }, { "epoch": 10.87912087912088, "grad_norm": 0.25787901878356934, "learning_rate": 9.278736719221964e-05, "loss": 0.0357, "step": 990 }, { "epoch": 10.989010989010989, "grad_norm": 0.268319696187973, "learning_rate": 9.096586314085162e-05, "loss": 0.0357, "step": 1000 }, { "epoch": 11.098901098901099, "grad_norm": 0.29258257150650024, "learning_rate": 8.914737641780738e-05, "loss": 0.0298, "step": 1010 }, { "epoch": 11.208791208791208, "grad_norm": 0.26910606026649475, "learning_rate": 8.733251438288032e-05, "loss": 0.0385, "step": 1020 }, { "epoch": 11.31868131868132, "grad_norm": 0.275666207075119, "learning_rate": 8.552188318524737e-05, "loss": 0.0371, "step": 1030 }, { "epoch": 11.428571428571429, "grad_norm": 0.3032730221748352, "learning_rate": 8.371608756102028e-05, "loss": 0.0298, "step": 1040 }, { "epoch": 11.538461538461538, "grad_norm": 0.19398105144500732, "learning_rate": 8.191573063126842e-05, "loss": 0.032, "step": 1050 }, { "epoch": 11.648351648351648, "grad_norm": 0.23560473322868347, "learning_rate": 8.012141370058151e-05, "loss": 0.0294, "step": 1060 }, { "epoch": 11.758241758241759, "grad_norm": 0.24311786890029907, "learning_rate": 7.833373605623855e-05, "loss": 0.034, "step": 1070 }, { "epoch": 11.868131868131869, "grad_norm": 0.1695002168416977, "learning_rate": 7.655329476805095e-05, "loss": 0.0345, "step": 1080 }, { "epoch": 11.978021978021978, "grad_norm": 0.1906553953886032, "learning_rate": 7.478068448894577e-05, "loss": 0.03, "step": 1090 }, { "epoch": 12.087912087912088, "grad_norm": 0.18722419440746307, "learning_rate": 7.301649725635689e-05, "loss": 0.0327, "step": 1100 }, { "epoch": 12.197802197802197, "grad_norm": 0.2109653502702713, "learning_rate": 7.126132229448883e-05, "loss": 0.03, "step": 1110 }, { "epoch": 12.307692307692308, "grad_norm": 0.3317616879940033, "learning_rate": 6.95157458175211e-05, "loss": 0.0289, "step": 1120 }, { "epoch": 12.417582417582418, "grad_norm": 0.18462032079696655, "learning_rate": 6.778035083381708e-05, "loss": 0.0336, "step": 1130 }, { "epoch": 12.527472527472527, "grad_norm": 0.17815662920475006, "learning_rate": 6.60557169512042e-05, "loss": 0.0317, "step": 1140 }, { "epoch": 12.637362637362637, "grad_norm": 0.19909383356571198, "learning_rate": 6.434242018338947e-05, "loss": 0.0317, "step": 1150 }, { "epoch": 12.747252747252748, "grad_norm": 0.21948197484016418, "learning_rate": 6.264103275757578e-05, "loss": 0.0288, "step": 1160 }, { "epoch": 12.857142857142858, "grad_norm": 0.30106085538864136, "learning_rate": 6.095212292334232e-05, "loss": 0.0298, "step": 1170 }, { "epoch": 12.967032967032967, "grad_norm": 0.3014363944530487, "learning_rate": 5.927625476285426e-05, "loss": 0.0318, "step": 1180 }, { "epoch": 13.076923076923077, "grad_norm": 0.5095160007476807, "learning_rate": 5.761398800246354e-05, "loss": 0.0272, "step": 1190 }, { "epoch": 13.186813186813186, "grad_norm": 0.19524632394313812, "learning_rate": 5.596587782576509e-05, "loss": 0.0294, "step": 1200 }, { "epoch": 13.296703296703297, "grad_norm": 0.20751602947711945, "learning_rate": 5.433247468816977e-05, "loss": 0.0295, "step": 1210 }, { "epoch": 13.406593406593407, "grad_norm": 0.12134165316820145, "learning_rate": 5.271432413305687e-05, "loss": 0.0258, "step": 1220 }, { "epoch": 13.516483516483516, "grad_norm": 0.2958245277404785, "learning_rate": 5.111196660956703e-05, "loss": 0.0328, "step": 1230 }, { "epoch": 13.626373626373626, "grad_norm": 0.1860036551952362, "learning_rate": 4.952593729209671e-05, "loss": 0.028, "step": 1240 }, { "epoch": 13.736263736263737, "grad_norm": 0.16853293776512146, "learning_rate": 4.7956765901554065e-05, "loss": 0.0259, "step": 1250 }, { "epoch": 13.846153846153847, "grad_norm": 0.22942283749580383, "learning_rate": 4.640497652843672e-05, "loss": 0.0303, "step": 1260 }, { "epoch": 13.956043956043956, "grad_norm": 0.1646496206521988, "learning_rate": 4.4871087457789584e-05, "loss": 0.0263, "step": 1270 }, { "epoch": 14.065934065934066, "grad_norm": 0.22191093862056732, "learning_rate": 4.3355610996102e-05, "loss": 0.033, "step": 1280 }, { "epoch": 14.175824175824175, "grad_norm": 0.19444125890731812, "learning_rate": 4.185905330020143e-05, "loss": 0.0268, "step": 1290 }, { "epoch": 14.285714285714286, "grad_norm": 0.16687805950641632, "learning_rate": 4.038191420820139e-05, "loss": 0.0293, "step": 1300 }, { "epoch": 14.395604395604396, "grad_norm": 0.21555913984775543, "learning_rate": 3.892468707255923e-05, "loss": 0.0269, "step": 1310 }, { "epoch": 14.505494505494505, "grad_norm": 0.1728578358888626, "learning_rate": 3.7487858595300884e-05, "loss": 0.0222, "step": 1320 }, { "epoch": 14.615384615384615, "grad_norm": 0.25217944383621216, "learning_rate": 3.607190866546578e-05, "loss": 0.023, "step": 1330 }, { "epoch": 14.725274725274724, "grad_norm": 0.45261967182159424, "learning_rate": 3.467731019882838e-05, "loss": 0.0366, "step": 1340 }, { "epoch": 14.835164835164836, "grad_norm": 0.19587543606758118, "learning_rate": 3.330452897994773e-05, "loss": 0.0249, "step": 1350 }, { "epoch": 14.945054945054945, "grad_norm": 0.16712340712547302, "learning_rate": 3.195402350659945e-05, "loss": 0.0291, "step": 1360 }, { "epoch": 15.054945054945055, "grad_norm": 0.3442732095718384, "learning_rate": 3.0626244836640993e-05, "loss": 0.0267, "step": 1370 }, { "epoch": 15.164835164835164, "grad_norm": 0.20348604023456573, "learning_rate": 2.9321636437362253e-05, "loss": 0.0255, "step": 1380 }, { "epoch": 15.274725274725276, "grad_norm": 0.18801550567150116, "learning_rate": 2.8040634037370728e-05, "loss": 0.0247, "step": 1390 }, { "epoch": 15.384615384615385, "grad_norm": 0.2544891834259033, "learning_rate": 2.6783665481062113e-05, "loss": 0.0275, "step": 1400 }, { "epoch": 15.494505494505495, "grad_norm": 0.3431093394756317, "learning_rate": 2.5551150585723415e-05, "loss": 0.0238, "step": 1410 }, { "epoch": 15.604395604395604, "grad_norm": 0.17770768702030182, "learning_rate": 2.4343501001317605e-05, "loss": 0.0251, "step": 1420 }, { "epoch": 15.714285714285714, "grad_norm": 0.170986145734787, "learning_rate": 2.3161120072995757e-05, "loss": 0.0243, "step": 1430 }, { "epoch": 15.824175824175825, "grad_norm": 0.19961877167224884, "learning_rate": 2.2004402706383043e-05, "loss": 0.0237, "step": 1440 }, { "epoch": 15.934065934065934, "grad_norm": 0.38295048475265503, "learning_rate": 2.0873735235683535e-05, "loss": 0.0239, "step": 1450 }, { "epoch": 16.043956043956044, "grad_norm": 0.15348948538303375, "learning_rate": 1.976949529464771e-05, "loss": 0.0282, "step": 1460 }, { "epoch": 16.153846153846153, "grad_norm": 0.3392385244369507, "learning_rate": 1.8692051690445743e-05, "loss": 0.0283, "step": 1470 }, { "epoch": 16.263736263736263, "grad_norm": 0.15572497248649597, "learning_rate": 1.764176428048908e-05, "loss": 0.0245, "step": 1480 }, { "epoch": 16.373626373626372, "grad_norm": 0.2224312424659729, "learning_rate": 1.661898385224079e-05, "loss": 0.0258, "step": 1490 }, { "epoch": 16.483516483516482, "grad_norm": 0.17235592007637024, "learning_rate": 1.5624052006055544e-05, "loss": 0.0248, "step": 1500 }, { "epoch": 16.593406593406595, "grad_norm": 0.17149613797664642, "learning_rate": 1.4657301041087812e-05, "loss": 0.0243, "step": 1510 }, { "epoch": 16.703296703296704, "grad_norm": 0.18543320894241333, "learning_rate": 1.3719053844306706e-05, "loss": 0.0272, "step": 1520 }, { "epoch": 16.813186813186814, "grad_norm": 0.1908874362707138, "learning_rate": 1.2809623782654223e-05, "loss": 0.0235, "step": 1530 }, { "epoch": 16.923076923076923, "grad_norm": 0.18022581934928894, "learning_rate": 1.1929314598383423e-05, "loss": 0.0263, "step": 1540 }, { "epoch": 17.032967032967033, "grad_norm": 0.1982187032699585, "learning_rate": 1.1078420307610783e-05, "loss": 0.0252, "step": 1550 }, { "epoch": 17.142857142857142, "grad_norm": 0.1451161950826645, "learning_rate": 1.0257225102117363e-05, "loss": 0.025, "step": 1560 }, { "epoch": 17.252747252747252, "grad_norm": 0.1547311544418335, "learning_rate": 9.466003254430934e-06, "loss": 0.024, "step": 1570 }, { "epoch": 17.36263736263736, "grad_norm": 0.20238442718982697, "learning_rate": 8.705019026221317e-06, "loss": 0.0249, "step": 1580 }, { "epoch": 17.47252747252747, "grad_norm": 0.18659718334674835, "learning_rate": 7.974526580038988e-06, "loss": 0.0257, "step": 1590 }, { "epoch": 17.582417582417584, "grad_norm": 0.2173370122909546, "learning_rate": 7.2747698944269915e-06, "loss": 0.0248, "step": 1600 }, { "epoch": 17.692307692307693, "grad_norm": 0.12479448318481445, "learning_rate": 6.605982682433953e-06, "loss": 0.0263, "step": 1610 }, { "epoch": 17.802197802197803, "grad_norm": 0.17071430385112762, "learning_rate": 5.968388313555895e-06, "loss": 0.0225, "step": 1620 }, { "epoch": 17.912087912087912, "grad_norm": 0.2871854901313782, "learning_rate": 5.3621997391326565e-06, "loss": 0.0191, "step": 1630 }, { "epoch": 18.021978021978022, "grad_norm": 0.39435771107673645, "learning_rate": 4.7876194212238434e-06, "loss": 0.0291, "step": 1640 }, { "epoch": 18.13186813186813, "grad_norm": 0.18047714233398438, "learning_rate": 4.24483926498812e-06, "loss": 0.0207, "step": 1650 }, { "epoch": 18.24175824175824, "grad_norm": 0.19172823429107666, "learning_rate": 3.734040554588514e-06, "loss": 0.0214, "step": 1660 }, { "epoch": 18.35164835164835, "grad_norm": 0.2549908757209778, "learning_rate": 3.255393892644909e-06, "loss": 0.0205, "step": 1670 }, { "epoch": 18.46153846153846, "grad_norm": 0.21838967502117157, "learning_rate": 2.8090591432542e-06, "loss": 0.0214, "step": 1680 }, { "epoch": 18.571428571428573, "grad_norm": 0.2188936024904251, "learning_rate": 2.3951853785969537e-06, "loss": 0.0227, "step": 1690 }, { "epoch": 18.681318681318682, "grad_norm": 0.15396256744861603, "learning_rate": 2.0139108291485574e-06, "loss": 0.0286, "step": 1700 }, { "epoch": 18.791208791208792, "grad_norm": 0.22776196897029877, "learning_rate": 1.6653628375112972e-06, "loss": 0.0248, "step": 1710 }, { "epoch": 18.9010989010989, "grad_norm": 0.17969872057437897, "learning_rate": 1.349657815883032e-06, "loss": 0.0227, "step": 1720 }, { "epoch": 19.01098901098901, "grad_norm": 0.22254760563373566, "learning_rate": 1.0669012071764073e-06, "loss": 0.0197, "step": 1730 }, { "epoch": 19.12087912087912, "grad_norm": 0.1676974892616272, "learning_rate": 8.171874498018039e-07, "loss": 0.0264, "step": 1740 }, { "epoch": 19.23076923076923, "grad_norm": 0.21863488852977753, "learning_rate": 6.005999461256684e-07, "loss": 0.0224, "step": 1750 }, { "epoch": 19.34065934065934, "grad_norm": 0.13759975135326385, "learning_rate": 4.172110346148506e-07, "loss": 0.0265, "step": 1760 }, { "epoch": 19.45054945054945, "grad_norm": 0.16070395708084106, "learning_rate": 2.670819656760526e-07, "loss": 0.0233, "step": 1770 }, { "epoch": 19.560439560439562, "grad_norm": 0.2525140643119812, "learning_rate": 1.502628811987483e-07, "loss": 0.0241, "step": 1780 }, { "epoch": 19.67032967032967, "grad_norm": 0.16685707867145538, "learning_rate": 6.679279780821279e-08, "loss": 0.0248, "step": 1790 }, { "epoch": 19.78021978021978, "grad_norm": 0.17456871271133423, "learning_rate": 1.6699593834224036e-08, "loss": 0.0259, "step": 1800 }, { "epoch": 19.89010989010989, "grad_norm": 0.18164990842342377, "learning_rate": 0.0, "loss": 0.0233, "step": 1810 }, { "epoch": 19.89010989010989, "step": 1810, "total_flos": 2.654273592491328e+17, "train_loss": 0.05974875727246479, "train_runtime": 2083.4502, "train_samples_per_second": 55.6, "train_steps_per_second": 0.869 } ], "logging_steps": 10, "max_steps": 1810, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 10000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.654273592491328e+17, "train_batch_size": 64, "trial_name": null, "trial_params": null }