{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 20, "global_step": 11621, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0017211333663217229, "grad_norm": 0.512730062007904, "learning_rate": 0.0002, "loss": 1.7072, "step": 20 }, { "epoch": 0.0034422667326434457, "grad_norm": 0.3541152775287628, "learning_rate": 0.0002, "loss": 1.3871, "step": 40 }, { "epoch": 0.005163400098965169, "grad_norm": 0.40835729241371155, "learning_rate": 0.0002, "loss": 1.2637, "step": 60 }, { "epoch": 0.0068845334652868915, "grad_norm": 0.3723342716693878, "learning_rate": 0.0002, "loss": 1.2378, "step": 80 }, { "epoch": 0.008605666831608615, "grad_norm": 0.3391265571117401, "learning_rate": 0.0002, "loss": 1.1991, "step": 100 }, { "epoch": 0.010326800197930338, "grad_norm": 0.3430226445198059, "learning_rate": 0.0002, "loss": 1.1191, "step": 120 }, { "epoch": 0.01204793356425206, "grad_norm": 0.39266398549079895, "learning_rate": 0.0002, "loss": 1.1358, "step": 140 }, { "epoch": 0.013769066930573783, "grad_norm": 0.38184505701065063, "learning_rate": 0.0002, "loss": 1.0851, "step": 160 }, { "epoch": 0.015490200296895506, "grad_norm": 0.3323766887187958, "learning_rate": 0.0002, "loss": 1.0519, "step": 180 }, { "epoch": 0.01721133366321723, "grad_norm": 0.3520768880844116, "learning_rate": 0.0002, "loss": 0.9788, "step": 200 }, { "epoch": 0.018932467029538953, "grad_norm": 0.4066319465637207, "learning_rate": 0.0002, "loss": 0.9647, "step": 220 }, { "epoch": 0.020653600395860675, "grad_norm": 0.3871042728424072, "learning_rate": 0.0002, "loss": 0.9981, "step": 240 }, { "epoch": 0.022374733762182398, "grad_norm": 0.3759310841560364, "learning_rate": 0.0002, "loss": 0.9527, "step": 260 }, { "epoch": 0.02409586712850412, "grad_norm": 0.411364883184433, "learning_rate": 0.0002, "loss": 1.0079, "step": 280 }, { "epoch": 0.025817000494825843, "grad_norm": 0.38831627368927, "learning_rate": 0.0002, "loss": 1.0129, "step": 300 }, { "epoch": 0.027538133861147566, "grad_norm": 0.36452218890190125, "learning_rate": 0.0002, "loss": 0.9021, "step": 320 }, { "epoch": 0.02925926722746929, "grad_norm": 0.40188145637512207, "learning_rate": 0.0002, "loss": 0.9138, "step": 340 }, { "epoch": 0.03098040059379101, "grad_norm": 0.3994237184524536, "learning_rate": 0.0002, "loss": 0.9164, "step": 360 }, { "epoch": 0.03270153396011274, "grad_norm": 0.4226379692554474, "learning_rate": 0.0002, "loss": 0.8986, "step": 380 }, { "epoch": 0.03442266732643446, "grad_norm": 0.4331601560115814, "learning_rate": 0.0002, "loss": 0.8443, "step": 400 }, { "epoch": 0.03614380069275618, "grad_norm": 0.373415470123291, "learning_rate": 0.0002, "loss": 0.8437, "step": 420 }, { "epoch": 0.037864934059077905, "grad_norm": 0.49005845189094543, "learning_rate": 0.0002, "loss": 0.8387, "step": 440 }, { "epoch": 0.03958606742539963, "grad_norm": 0.5034841895103455, "learning_rate": 0.0002, "loss": 0.8108, "step": 460 }, { "epoch": 0.04130720079172135, "grad_norm": 0.3401569426059723, "learning_rate": 0.0002, "loss": 0.7975, "step": 480 }, { "epoch": 0.04302833415804307, "grad_norm": 0.37207746505737305, "learning_rate": 0.0002, "loss": 0.8135, "step": 500 }, { "epoch": 0.044749467524364796, "grad_norm": 0.39017221331596375, "learning_rate": 0.0002, "loss": 0.7053, "step": 520 }, { "epoch": 0.04647060089068652, "grad_norm": 0.42643848061561584, "learning_rate": 0.0002, "loss": 0.8119, "step": 540 }, { "epoch": 0.04819173425700824, "grad_norm": 0.4266549050807953, "learning_rate": 0.0002, "loss": 0.7676, "step": 560 }, { "epoch": 0.049912867623329964, "grad_norm": 0.37006089091300964, "learning_rate": 0.0002, "loss": 0.7675, "step": 580 }, { "epoch": 0.051634000989651686, "grad_norm": 0.3692554235458374, "learning_rate": 0.0002, "loss": 0.7843, "step": 600 }, { "epoch": 0.05335513435597341, "grad_norm": 0.41451218724250793, "learning_rate": 0.0002, "loss": 0.7111, "step": 620 }, { "epoch": 0.05507626772229513, "grad_norm": 0.46577128767967224, "learning_rate": 0.0002, "loss": 0.7248, "step": 640 }, { "epoch": 0.056797401088616854, "grad_norm": 0.4749889373779297, "learning_rate": 0.0002, "loss": 0.7454, "step": 660 }, { "epoch": 0.05851853445493858, "grad_norm": 0.44663748145103455, "learning_rate": 0.0002, "loss": 0.6829, "step": 680 }, { "epoch": 0.0602396678212603, "grad_norm": 0.3434062600135803, "learning_rate": 0.0002, "loss": 0.6625, "step": 700 }, { "epoch": 0.06196080118758202, "grad_norm": 0.45600661635398865, "learning_rate": 0.0002, "loss": 0.6623, "step": 720 }, { "epoch": 0.06368193455390375, "grad_norm": 0.41260451078414917, "learning_rate": 0.0002, "loss": 0.6497, "step": 740 }, { "epoch": 0.06540306792022547, "grad_norm": 0.3827366232872009, "learning_rate": 0.0002, "loss": 0.6133, "step": 760 }, { "epoch": 0.0671242012865472, "grad_norm": 0.44087541103363037, "learning_rate": 0.0002, "loss": 0.6782, "step": 780 }, { "epoch": 0.06884533465286892, "grad_norm": 0.3596974313259125, "learning_rate": 0.0002, "loss": 0.6294, "step": 800 }, { "epoch": 0.07056646801919064, "grad_norm": 0.42401421070098877, "learning_rate": 0.0002, "loss": 0.6761, "step": 820 }, { "epoch": 0.07228760138551237, "grad_norm": 0.34142622351646423, "learning_rate": 0.0002, "loss": 0.6274, "step": 840 }, { "epoch": 0.07400873475183409, "grad_norm": 0.3514570891857147, "learning_rate": 0.0002, "loss": 0.6053, "step": 860 }, { "epoch": 0.07572986811815581, "grad_norm": 0.3995071053504944, "learning_rate": 0.0002, "loss": 0.5953, "step": 880 }, { "epoch": 0.07745100148447753, "grad_norm": 0.3967471420764923, "learning_rate": 0.0002, "loss": 0.6375, "step": 900 }, { "epoch": 0.07917213485079926, "grad_norm": 0.434865266084671, "learning_rate": 0.0002, "loss": 0.6053, "step": 920 }, { "epoch": 0.08089326821712098, "grad_norm": 0.42927464842796326, "learning_rate": 0.0002, "loss": 0.5947, "step": 940 }, { "epoch": 0.0826144015834427, "grad_norm": 0.6075530052185059, "learning_rate": 0.0002, "loss": 0.6331, "step": 960 }, { "epoch": 0.08433553494976442, "grad_norm": 0.46515706181526184, "learning_rate": 0.0002, "loss": 0.569, "step": 980 }, { "epoch": 0.08605666831608615, "grad_norm": 0.2592894732952118, "learning_rate": 0.0002, "loss": 0.5572, "step": 1000 }, { "epoch": 0.08777780168240787, "grad_norm": 0.4780791699886322, "learning_rate": 0.0002, "loss": 0.6014, "step": 1020 }, { "epoch": 0.08949893504872959, "grad_norm": 0.37773773074150085, "learning_rate": 0.0002, "loss": 0.5783, "step": 1040 }, { "epoch": 0.09122006841505131, "grad_norm": 0.5149890780448914, "learning_rate": 0.0002, "loss": 0.5485, "step": 1060 }, { "epoch": 0.09294120178137304, "grad_norm": 0.522530734539032, "learning_rate": 0.0002, "loss": 0.5707, "step": 1080 }, { "epoch": 0.09466233514769476, "grad_norm": 0.42418375611305237, "learning_rate": 0.0002, "loss": 0.5817, "step": 1100 }, { "epoch": 0.09638346851401648, "grad_norm": 0.3230820596218109, "learning_rate": 0.0002, "loss": 0.5426, "step": 1120 }, { "epoch": 0.0981046018803382, "grad_norm": 0.46259307861328125, "learning_rate": 0.0002, "loss": 0.5304, "step": 1140 }, { "epoch": 0.09982573524665993, "grad_norm": 0.41511914134025574, "learning_rate": 0.0002, "loss": 0.5343, "step": 1160 }, { "epoch": 0.10154686861298165, "grad_norm": 0.5695448517799377, "learning_rate": 0.0002, "loss": 0.5773, "step": 1180 }, { "epoch": 0.10326800197930337, "grad_norm": 0.3354031443595886, "learning_rate": 0.0002, "loss": 0.5377, "step": 1200 }, { "epoch": 0.1049891353456251, "grad_norm": 0.5198836922645569, "learning_rate": 0.0002, "loss": 0.5454, "step": 1220 }, { "epoch": 0.10671026871194682, "grad_norm": 0.3364623188972473, "learning_rate": 0.0002, "loss": 0.5057, "step": 1240 }, { "epoch": 0.10843140207826854, "grad_norm": 0.3844848871231079, "learning_rate": 0.0002, "loss": 0.571, "step": 1260 }, { "epoch": 0.11015253544459026, "grad_norm": 0.45747458934783936, "learning_rate": 0.0002, "loss": 0.5425, "step": 1280 }, { "epoch": 0.11187366881091199, "grad_norm": 0.4278349280357361, "learning_rate": 0.0002, "loss": 0.4814, "step": 1300 }, { "epoch": 0.11359480217723371, "grad_norm": 0.4377232491970062, "learning_rate": 0.0002, "loss": 0.5083, "step": 1320 }, { "epoch": 0.11531593554355543, "grad_norm": 0.48636892437934875, "learning_rate": 0.0002, "loss": 0.5317, "step": 1340 }, { "epoch": 0.11703706890987715, "grad_norm": 0.39063599705696106, "learning_rate": 0.0002, "loss": 0.482, "step": 1360 }, { "epoch": 0.11875820227619888, "grad_norm": 0.45347273349761963, "learning_rate": 0.0002, "loss": 0.4754, "step": 1380 }, { "epoch": 0.1204793356425206, "grad_norm": 0.44759300351142883, "learning_rate": 0.0002, "loss": 0.4773, "step": 1400 }, { "epoch": 0.12220046900884232, "grad_norm": 0.44881579279899597, "learning_rate": 0.0002, "loss": 0.5236, "step": 1420 }, { "epoch": 0.12392160237516404, "grad_norm": 0.5386276245117188, "learning_rate": 0.0002, "loss": 0.5021, "step": 1440 }, { "epoch": 0.12564273574148577, "grad_norm": 0.48680126667022705, "learning_rate": 0.0002, "loss": 0.4784, "step": 1460 }, { "epoch": 0.1273638691078075, "grad_norm": 0.44588467478752136, "learning_rate": 0.0002, "loss": 0.4905, "step": 1480 }, { "epoch": 0.1290850024741292, "grad_norm": 0.4805804491043091, "learning_rate": 0.0002, "loss": 0.4866, "step": 1500 }, { "epoch": 0.13080613584045095, "grad_norm": 0.44360053539276123, "learning_rate": 0.0002, "loss": 0.4637, "step": 1520 }, { "epoch": 0.13252726920677266, "grad_norm": 0.3818538784980774, "learning_rate": 0.0002, "loss": 0.4512, "step": 1540 }, { "epoch": 0.1342484025730944, "grad_norm": 0.44495511054992676, "learning_rate": 0.0002, "loss": 0.4691, "step": 1560 }, { "epoch": 0.1359695359394161, "grad_norm": 0.4303475618362427, "learning_rate": 0.0002, "loss": 0.4551, "step": 1580 }, { "epoch": 0.13769066930573784, "grad_norm": 0.4990542531013489, "learning_rate": 0.0002, "loss": 0.4426, "step": 1600 }, { "epoch": 0.13941180267205955, "grad_norm": 0.4822593331336975, "learning_rate": 0.0002, "loss": 0.4746, "step": 1620 }, { "epoch": 0.14113293603838128, "grad_norm": 0.39103543758392334, "learning_rate": 0.0002, "loss": 0.4503, "step": 1640 }, { "epoch": 0.142854069404703, "grad_norm": 0.3490641117095947, "learning_rate": 0.0002, "loss": 0.4869, "step": 1660 }, { "epoch": 0.14457520277102473, "grad_norm": 0.3039151728153229, "learning_rate": 0.0002, "loss": 0.4885, "step": 1680 }, { "epoch": 0.14629633613734644, "grad_norm": 0.45525145530700684, "learning_rate": 0.0002, "loss": 0.4465, "step": 1700 }, { "epoch": 0.14801746950366818, "grad_norm": 0.3779124915599823, "learning_rate": 0.0002, "loss": 0.4077, "step": 1720 }, { "epoch": 0.14973860286998988, "grad_norm": 0.3984091579914093, "learning_rate": 0.0002, "loss": 0.4212, "step": 1740 }, { "epoch": 0.15145973623631162, "grad_norm": 0.37952274084091187, "learning_rate": 0.0002, "loss": 0.4158, "step": 1760 }, { "epoch": 0.15318086960263333, "grad_norm": 0.34842512011528015, "learning_rate": 0.0002, "loss": 0.4188, "step": 1780 }, { "epoch": 0.15490200296895507, "grad_norm": 0.36399173736572266, "learning_rate": 0.0002, "loss": 0.4347, "step": 1800 }, { "epoch": 0.15662313633527677, "grad_norm": 0.38256826996803284, "learning_rate": 0.0002, "loss": 0.4179, "step": 1820 }, { "epoch": 0.1583442697015985, "grad_norm": 0.3715302050113678, "learning_rate": 0.0002, "loss": 0.4222, "step": 1840 }, { "epoch": 0.16006540306792022, "grad_norm": 0.3852146565914154, "learning_rate": 0.0002, "loss": 0.4005, "step": 1860 }, { "epoch": 0.16178653643424196, "grad_norm": 0.3589220643043518, "learning_rate": 0.0002, "loss": 0.4099, "step": 1880 }, { "epoch": 0.16350766980056367, "grad_norm": 0.5060593485832214, "learning_rate": 0.0002, "loss": 0.4158, "step": 1900 }, { "epoch": 0.1652288031668854, "grad_norm": 0.3826269507408142, "learning_rate": 0.0002, "loss": 0.3874, "step": 1920 }, { "epoch": 0.1669499365332071, "grad_norm": 0.3576675057411194, "learning_rate": 0.0002, "loss": 0.4226, "step": 1940 }, { "epoch": 0.16867106989952885, "grad_norm": 0.37530502676963806, "learning_rate": 0.0002, "loss": 0.4014, "step": 1960 }, { "epoch": 0.17039220326585056, "grad_norm": 0.3857693374156952, "learning_rate": 0.0002, "loss": 0.413, "step": 1980 }, { "epoch": 0.1721133366321723, "grad_norm": 0.42137113213539124, "learning_rate": 0.0002, "loss": 0.4212, "step": 2000 }, { "epoch": 0.173834469998494, "grad_norm": 0.37976202368736267, "learning_rate": 0.0002, "loss": 0.3925, "step": 2020 }, { "epoch": 0.17555560336481574, "grad_norm": 0.3828701078891754, "learning_rate": 0.0002, "loss": 0.4004, "step": 2040 }, { "epoch": 0.17727673673113745, "grad_norm": 0.5018408894538879, "learning_rate": 0.0002, "loss": 0.384, "step": 2060 }, { "epoch": 0.17899787009745918, "grad_norm": 0.4595503509044647, "learning_rate": 0.0002, "loss": 0.4179, "step": 2080 }, { "epoch": 0.1807190034637809, "grad_norm": 0.5221086144447327, "learning_rate": 0.0002, "loss": 0.3855, "step": 2100 }, { "epoch": 0.18244013683010263, "grad_norm": 0.5442349314689636, "learning_rate": 0.0002, "loss": 0.3579, "step": 2120 }, { "epoch": 0.18416127019642434, "grad_norm": 0.5430803298950195, "learning_rate": 0.0002, "loss": 0.3871, "step": 2140 }, { "epoch": 0.18588240356274607, "grad_norm": 0.2919999063014984, "learning_rate": 0.0002, "loss": 0.3947, "step": 2160 }, { "epoch": 0.18760353692906778, "grad_norm": 0.3997620940208435, "learning_rate": 0.0002, "loss": 0.3989, "step": 2180 }, { "epoch": 0.18932467029538952, "grad_norm": 0.38948822021484375, "learning_rate": 0.0002, "loss": 0.4292, "step": 2200 }, { "epoch": 0.19104580366171123, "grad_norm": 0.4080924689769745, "learning_rate": 0.0002, "loss": 0.398, "step": 2220 }, { "epoch": 0.19276693702803296, "grad_norm": 0.3982754647731781, "learning_rate": 0.0002, "loss": 0.4022, "step": 2240 }, { "epoch": 0.19448807039435467, "grad_norm": 0.41448819637298584, "learning_rate": 0.0002, "loss": 0.3968, "step": 2260 }, { "epoch": 0.1962092037606764, "grad_norm": 0.42457613348960876, "learning_rate": 0.0002, "loss": 0.4024, "step": 2280 }, { "epoch": 0.19793033712699812, "grad_norm": 0.2920919358730316, "learning_rate": 0.0002, "loss": 0.387, "step": 2300 }, { "epoch": 0.19965147049331985, "grad_norm": 0.4027709662914276, "learning_rate": 0.0002, "loss": 0.3835, "step": 2320 }, { "epoch": 0.20137260385964156, "grad_norm": 0.41888585686683655, "learning_rate": 0.0002, "loss": 0.3985, "step": 2340 }, { "epoch": 0.2030937372259633, "grad_norm": 0.4292152225971222, "learning_rate": 0.0002, "loss": 0.3823, "step": 2360 }, { "epoch": 0.204814870592285, "grad_norm": 0.4724443554878235, "learning_rate": 0.0002, "loss": 0.3589, "step": 2380 }, { "epoch": 0.20653600395860675, "grad_norm": 0.32000842690467834, "learning_rate": 0.0002, "loss": 0.383, "step": 2400 }, { "epoch": 0.20825713732492845, "grad_norm": 0.3765513300895691, "learning_rate": 0.0002, "loss": 0.371, "step": 2420 }, { "epoch": 0.2099782706912502, "grad_norm": 0.39066755771636963, "learning_rate": 0.0002, "loss": 0.3684, "step": 2440 }, { "epoch": 0.2116994040575719, "grad_norm": 0.42424216866493225, "learning_rate": 0.0002, "loss": 0.3653, "step": 2460 }, { "epoch": 0.21342053742389364, "grad_norm": 0.5066993832588196, "learning_rate": 0.0002, "loss": 0.3566, "step": 2480 }, { "epoch": 0.21514167079021534, "grad_norm": 0.44107547402381897, "learning_rate": 0.0002, "loss": 0.3748, "step": 2500 }, { "epoch": 0.21686280415653708, "grad_norm": 0.3890049457550049, "learning_rate": 0.0002, "loss": 0.3378, "step": 2520 }, { "epoch": 0.2185839375228588, "grad_norm": 0.3157922327518463, "learning_rate": 0.0002, "loss": 0.3462, "step": 2540 }, { "epoch": 0.22030507088918053, "grad_norm": 0.34473904967308044, "learning_rate": 0.0002, "loss": 0.3523, "step": 2560 }, { "epoch": 0.22202620425550224, "grad_norm": 0.38636985421180725, "learning_rate": 0.0002, "loss": 0.3694, "step": 2580 }, { "epoch": 0.22374733762182397, "grad_norm": 0.4448566436767578, "learning_rate": 0.0002, "loss": 0.3752, "step": 2600 }, { "epoch": 0.22546847098814568, "grad_norm": 0.3706173002719879, "learning_rate": 0.0002, "loss": 0.3706, "step": 2620 }, { "epoch": 0.22718960435446742, "grad_norm": 0.47221776843070984, "learning_rate": 0.0002, "loss": 0.3865, "step": 2640 }, { "epoch": 0.22891073772078913, "grad_norm": 0.39190933108329773, "learning_rate": 0.0002, "loss": 0.362, "step": 2660 }, { "epoch": 0.23063187108711086, "grad_norm": 0.36031046509742737, "learning_rate": 0.0002, "loss": 0.352, "step": 2680 }, { "epoch": 0.2323530044534326, "grad_norm": 0.4583648145198822, "learning_rate": 0.0002, "loss": 0.3343, "step": 2700 }, { "epoch": 0.2340741378197543, "grad_norm": 0.35927554965019226, "learning_rate": 0.0002, "loss": 0.3748, "step": 2720 }, { "epoch": 0.23579527118607604, "grad_norm": 0.45593491196632385, "learning_rate": 0.0002, "loss": 0.356, "step": 2740 }, { "epoch": 0.23751640455239775, "grad_norm": 0.5094373226165771, "learning_rate": 0.0002, "loss": 0.3636, "step": 2760 }, { "epoch": 0.2392375379187195, "grad_norm": 0.47478821873664856, "learning_rate": 0.0002, "loss": 0.3537, "step": 2780 }, { "epoch": 0.2409586712850412, "grad_norm": 0.37840044498443604, "learning_rate": 0.0002, "loss": 0.3507, "step": 2800 }, { "epoch": 0.24267980465136293, "grad_norm": 0.41650140285491943, "learning_rate": 0.0002, "loss": 0.3405, "step": 2820 }, { "epoch": 0.24440093801768464, "grad_norm": 0.47170814871788025, "learning_rate": 0.0002, "loss": 0.3564, "step": 2840 }, { "epoch": 0.24612207138400638, "grad_norm": 0.3727279305458069, "learning_rate": 0.0002, "loss": 0.3621, "step": 2860 }, { "epoch": 0.2478432047503281, "grad_norm": 0.2921255826950073, "learning_rate": 0.0002, "loss": 0.3504, "step": 2880 }, { "epoch": 0.24956433811664983, "grad_norm": 0.30992376804351807, "learning_rate": 0.0002, "loss": 0.3403, "step": 2900 }, { "epoch": 0.25128547148297153, "grad_norm": 0.3726598620414734, "learning_rate": 0.0002, "loss": 0.337, "step": 2920 }, { "epoch": 0.25300660484929327, "grad_norm": 0.46260905265808105, "learning_rate": 0.0002, "loss": 0.3428, "step": 2940 }, { "epoch": 0.254727738215615, "grad_norm": 0.4512200653553009, "learning_rate": 0.0002, "loss": 0.3456, "step": 2960 }, { "epoch": 0.2564488715819367, "grad_norm": 0.3799881041049957, "learning_rate": 0.0002, "loss": 0.3479, "step": 2980 }, { "epoch": 0.2581700049482584, "grad_norm": 0.36525946855545044, "learning_rate": 0.0002, "loss": 0.3324, "step": 3000 }, { "epoch": 0.25989113831458016, "grad_norm": 0.43525341153144836, "learning_rate": 0.0002, "loss": 0.3499, "step": 3020 }, { "epoch": 0.2616122716809019, "grad_norm": 0.39706552028656006, "learning_rate": 0.0002, "loss": 0.3319, "step": 3040 }, { "epoch": 0.2633334050472236, "grad_norm": 0.36725783348083496, "learning_rate": 0.0002, "loss": 0.3581, "step": 3060 }, { "epoch": 0.2650545384135453, "grad_norm": 0.4533792734146118, "learning_rate": 0.0002, "loss": 0.3377, "step": 3080 }, { "epoch": 0.26677567177986705, "grad_norm": 0.3752067983150482, "learning_rate": 0.0002, "loss": 0.3331, "step": 3100 }, { "epoch": 0.2684968051461888, "grad_norm": 0.31743401288986206, "learning_rate": 0.0002, "loss": 0.3118, "step": 3120 }, { "epoch": 0.27021793851251047, "grad_norm": 0.45233970880508423, "learning_rate": 0.0002, "loss": 0.334, "step": 3140 }, { "epoch": 0.2719390718788322, "grad_norm": 0.32593509554862976, "learning_rate": 0.0002, "loss": 0.3208, "step": 3160 }, { "epoch": 0.27366020524515394, "grad_norm": 0.3870694935321808, "learning_rate": 0.0002, "loss": 0.3398, "step": 3180 }, { "epoch": 0.2753813386114757, "grad_norm": 0.39160028100013733, "learning_rate": 0.0002, "loss": 0.3209, "step": 3200 }, { "epoch": 0.27710247197779736, "grad_norm": 0.3718438148498535, "learning_rate": 0.0002, "loss": 0.2982, "step": 3220 }, { "epoch": 0.2788236053441191, "grad_norm": 0.4262580871582031, "learning_rate": 0.0002, "loss": 0.2977, "step": 3240 }, { "epoch": 0.28054473871044083, "grad_norm": 0.33147528767585754, "learning_rate": 0.0002, "loss": 0.3345, "step": 3260 }, { "epoch": 0.28226587207676257, "grad_norm": 0.3187743127346039, "learning_rate": 0.0002, "loss": 0.3325, "step": 3280 }, { "epoch": 0.28398700544308425, "grad_norm": 0.4593104422092438, "learning_rate": 0.0002, "loss": 0.3101, "step": 3300 }, { "epoch": 0.285708138809406, "grad_norm": 0.431805819272995, "learning_rate": 0.0002, "loss": 0.2995, "step": 3320 }, { "epoch": 0.2874292721757277, "grad_norm": 0.3650576174259186, "learning_rate": 0.0002, "loss": 0.3176, "step": 3340 }, { "epoch": 0.28915040554204946, "grad_norm": 0.3000945746898651, "learning_rate": 0.0002, "loss": 0.3364, "step": 3360 }, { "epoch": 0.29087153890837114, "grad_norm": 0.3499468266963959, "learning_rate": 0.0002, "loss": 0.2956, "step": 3380 }, { "epoch": 0.2925926722746929, "grad_norm": 0.41105055809020996, "learning_rate": 0.0002, "loss": 0.3516, "step": 3400 }, { "epoch": 0.2943138056410146, "grad_norm": 0.4257635772228241, "learning_rate": 0.0002, "loss": 0.3098, "step": 3420 }, { "epoch": 0.29603493900733635, "grad_norm": 0.3407900035381317, "learning_rate": 0.0002, "loss": 0.302, "step": 3440 }, { "epoch": 0.29775607237365803, "grad_norm": 0.3562329113483429, "learning_rate": 0.0002, "loss": 0.3066, "step": 3460 }, { "epoch": 0.29947720573997977, "grad_norm": 0.4639800786972046, "learning_rate": 0.0002, "loss": 0.3108, "step": 3480 }, { "epoch": 0.3011983391063015, "grad_norm": 0.47634968161582947, "learning_rate": 0.0002, "loss": 0.3185, "step": 3500 }, { "epoch": 0.30291947247262324, "grad_norm": 0.42725369334220886, "learning_rate": 0.0002, "loss": 0.3081, "step": 3520 }, { "epoch": 0.3046406058389449, "grad_norm": 0.38221365213394165, "learning_rate": 0.0002, "loss": 0.3282, "step": 3540 }, { "epoch": 0.30636173920526666, "grad_norm": 0.3222472369670868, "learning_rate": 0.0002, "loss": 0.2975, "step": 3560 }, { "epoch": 0.3080828725715884, "grad_norm": 0.3551533818244934, "learning_rate": 0.0002, "loss": 0.308, "step": 3580 }, { "epoch": 0.30980400593791013, "grad_norm": 0.35602936148643494, "learning_rate": 0.0002, "loss": 0.3225, "step": 3600 }, { "epoch": 0.3115251393042318, "grad_norm": 0.36850786209106445, "learning_rate": 0.0002, "loss": 0.293, "step": 3620 }, { "epoch": 0.31324627267055355, "grad_norm": 0.4266447126865387, "learning_rate": 0.0002, "loss": 0.3055, "step": 3640 }, { "epoch": 0.3149674060368753, "grad_norm": 0.3386904001235962, "learning_rate": 0.0002, "loss": 0.3046, "step": 3660 }, { "epoch": 0.316688539403197, "grad_norm": 0.39958760142326355, "learning_rate": 0.0002, "loss": 0.3177, "step": 3680 }, { "epoch": 0.3184096727695187, "grad_norm": 0.3587161600589752, "learning_rate": 0.0002, "loss": 0.3183, "step": 3700 }, { "epoch": 0.32013080613584044, "grad_norm": 0.37195342779159546, "learning_rate": 0.0002, "loss": 0.2984, "step": 3720 }, { "epoch": 0.3218519395021622, "grad_norm": 0.42094019055366516, "learning_rate": 0.0002, "loss": 0.3019, "step": 3740 }, { "epoch": 0.3235730728684839, "grad_norm": 0.2889210879802704, "learning_rate": 0.0002, "loss": 0.3229, "step": 3760 }, { "epoch": 0.3252942062348056, "grad_norm": 0.4362521767616272, "learning_rate": 0.0002, "loss": 0.2891, "step": 3780 }, { "epoch": 0.32701533960112733, "grad_norm": 0.3560026288032532, "learning_rate": 0.0002, "loss": 0.3042, "step": 3800 }, { "epoch": 0.32873647296744907, "grad_norm": 0.5475091338157654, "learning_rate": 0.0002, "loss": 0.2949, "step": 3820 }, { "epoch": 0.3304576063337708, "grad_norm": 0.3368379473686218, "learning_rate": 0.0002, "loss": 0.3203, "step": 3840 }, { "epoch": 0.3321787397000925, "grad_norm": 0.45856744050979614, "learning_rate": 0.0002, "loss": 0.3107, "step": 3860 }, { "epoch": 0.3338998730664142, "grad_norm": 0.47050952911376953, "learning_rate": 0.0002, "loss": 0.2882, "step": 3880 }, { "epoch": 0.33562100643273596, "grad_norm": 0.5037484169006348, "learning_rate": 0.0002, "loss": 0.3033, "step": 3900 }, { "epoch": 0.3373421397990577, "grad_norm": 0.34310051798820496, "learning_rate": 0.0002, "loss": 0.3099, "step": 3920 }, { "epoch": 0.3390632731653794, "grad_norm": 0.47114118933677673, "learning_rate": 0.0002, "loss": 0.2892, "step": 3940 }, { "epoch": 0.3407844065317011, "grad_norm": 0.4255194365978241, "learning_rate": 0.0002, "loss": 0.3022, "step": 3960 }, { "epoch": 0.34250553989802285, "grad_norm": 0.5251383185386658, "learning_rate": 0.0002, "loss": 0.3114, "step": 3980 }, { "epoch": 0.3442266732643446, "grad_norm": 0.29961732029914856, "learning_rate": 0.0002, "loss": 0.3093, "step": 4000 }, { "epoch": 0.3459478066306663, "grad_norm": 0.4707966446876526, "learning_rate": 0.0002, "loss": 0.3139, "step": 4020 }, { "epoch": 0.347668939996988, "grad_norm": 0.33052656054496765, "learning_rate": 0.0002, "loss": 0.277, "step": 4040 }, { "epoch": 0.34939007336330974, "grad_norm": 0.3629648685455322, "learning_rate": 0.0002, "loss": 0.2934, "step": 4060 }, { "epoch": 0.3511112067296315, "grad_norm": 0.3852795362472534, "learning_rate": 0.0002, "loss": 0.2885, "step": 4080 }, { "epoch": 0.3528323400959532, "grad_norm": 0.3913141191005707, "learning_rate": 0.0002, "loss": 0.3009, "step": 4100 }, { "epoch": 0.3545534734622749, "grad_norm": 0.4364122152328491, "learning_rate": 0.0002, "loss": 0.2825, "step": 4120 }, { "epoch": 0.35627460682859663, "grad_norm": 0.3607100248336792, "learning_rate": 0.0002, "loss": 0.2969, "step": 4140 }, { "epoch": 0.35799574019491837, "grad_norm": 0.41620948910713196, "learning_rate": 0.0002, "loss": 0.2735, "step": 4160 }, { "epoch": 0.3597168735612401, "grad_norm": 0.36478644609451294, "learning_rate": 0.0002, "loss": 0.2826, "step": 4180 }, { "epoch": 0.3614380069275618, "grad_norm": 0.39005762338638306, "learning_rate": 0.0002, "loss": 0.2945, "step": 4200 }, { "epoch": 0.3631591402938835, "grad_norm": 0.3086032271385193, "learning_rate": 0.0002, "loss": 0.3057, "step": 4220 }, { "epoch": 0.36488027366020526, "grad_norm": 0.4346630275249481, "learning_rate": 0.0002, "loss": 0.2667, "step": 4240 }, { "epoch": 0.366601407026527, "grad_norm": 0.3394578993320465, "learning_rate": 0.0002, "loss": 0.3107, "step": 4260 }, { "epoch": 0.3683225403928487, "grad_norm": 0.44261375069618225, "learning_rate": 0.0002, "loss": 0.2968, "step": 4280 }, { "epoch": 0.3700436737591704, "grad_norm": 0.36678528785705566, "learning_rate": 0.0002, "loss": 0.3023, "step": 4300 }, { "epoch": 0.37176480712549215, "grad_norm": 0.43799522519111633, "learning_rate": 0.0002, "loss": 0.2781, "step": 4320 }, { "epoch": 0.3734859404918139, "grad_norm": 0.3344813883304596, "learning_rate": 0.0002, "loss": 0.2818, "step": 4340 }, { "epoch": 0.37520707385813556, "grad_norm": 0.519148588180542, "learning_rate": 0.0002, "loss": 0.2897, "step": 4360 }, { "epoch": 0.3769282072244573, "grad_norm": 0.36793655157089233, "learning_rate": 0.0002, "loss": 0.3068, "step": 4380 }, { "epoch": 0.37864934059077904, "grad_norm": 0.37348249554634094, "learning_rate": 0.0002, "loss": 0.2849, "step": 4400 }, { "epoch": 0.3803704739571008, "grad_norm": 0.44873690605163574, "learning_rate": 0.0002, "loss": 0.2902, "step": 4420 }, { "epoch": 0.38209160732342246, "grad_norm": 0.32298997044563293, "learning_rate": 0.0002, "loss": 0.2892, "step": 4440 }, { "epoch": 0.3838127406897442, "grad_norm": 0.4323394298553467, "learning_rate": 0.0002, "loss": 0.288, "step": 4460 }, { "epoch": 0.38553387405606593, "grad_norm": 0.3986116945743561, "learning_rate": 0.0002, "loss": 0.2962, "step": 4480 }, { "epoch": 0.38725500742238766, "grad_norm": 0.4980430603027344, "learning_rate": 0.0002, "loss": 0.2846, "step": 4500 }, { "epoch": 0.38897614078870935, "grad_norm": 0.4195205569267273, "learning_rate": 0.0002, "loss": 0.2806, "step": 4520 }, { "epoch": 0.3906972741550311, "grad_norm": 0.2984451949596405, "learning_rate": 0.0002, "loss": 0.2796, "step": 4540 }, { "epoch": 0.3924184075213528, "grad_norm": 0.3897689878940582, "learning_rate": 0.0002, "loss": 0.2842, "step": 4560 }, { "epoch": 0.39413954088767456, "grad_norm": 0.40885892510414124, "learning_rate": 0.0002, "loss": 0.2801, "step": 4580 }, { "epoch": 0.39586067425399624, "grad_norm": 0.41099223494529724, "learning_rate": 0.0002, "loss": 0.276, "step": 4600 }, { "epoch": 0.397581807620318, "grad_norm": 0.4116320312023163, "learning_rate": 0.0002, "loss": 0.2898, "step": 4620 }, { "epoch": 0.3993029409866397, "grad_norm": 0.37863150238990784, "learning_rate": 0.0002, "loss": 0.268, "step": 4640 }, { "epoch": 0.40102407435296145, "grad_norm": 0.43875718116760254, "learning_rate": 0.0002, "loss": 0.277, "step": 4660 }, { "epoch": 0.4027452077192831, "grad_norm": 0.3862336277961731, "learning_rate": 0.0002, "loss": 0.269, "step": 4680 }, { "epoch": 0.40446634108560486, "grad_norm": 0.330501914024353, "learning_rate": 0.0002, "loss": 0.2754, "step": 4700 }, { "epoch": 0.4061874744519266, "grad_norm": 0.3258209526538849, "learning_rate": 0.0002, "loss": 0.2933, "step": 4720 }, { "epoch": 0.40790860781824834, "grad_norm": 0.4297473430633545, "learning_rate": 0.0002, "loss": 0.2726, "step": 4740 }, { "epoch": 0.40962974118457, "grad_norm": 0.36448150873184204, "learning_rate": 0.0002, "loss": 0.2645, "step": 4760 }, { "epoch": 0.41135087455089175, "grad_norm": 0.4198833107948303, "learning_rate": 0.0002, "loss": 0.3029, "step": 4780 }, { "epoch": 0.4130720079172135, "grad_norm": 0.35666197538375854, "learning_rate": 0.0002, "loss": 0.2443, "step": 4800 }, { "epoch": 0.4147931412835352, "grad_norm": 0.38097846508026123, "learning_rate": 0.0002, "loss": 0.2706, "step": 4820 }, { "epoch": 0.4165142746498569, "grad_norm": 0.3574432134628296, "learning_rate": 0.0002, "loss": 0.2791, "step": 4840 }, { "epoch": 0.41823540801617864, "grad_norm": 0.46495670080184937, "learning_rate": 0.0002, "loss": 0.2753, "step": 4860 }, { "epoch": 0.4199565413825004, "grad_norm": 0.4777359664440155, "learning_rate": 0.0002, "loss": 0.2717, "step": 4880 }, { "epoch": 0.4216776747488221, "grad_norm": 0.4285059869289398, "learning_rate": 0.0002, "loss": 0.271, "step": 4900 }, { "epoch": 0.4233988081151438, "grad_norm": 0.45529553294181824, "learning_rate": 0.0002, "loss": 0.268, "step": 4920 }, { "epoch": 0.42511994148146554, "grad_norm": 0.3446490466594696, "learning_rate": 0.0002, "loss": 0.2659, "step": 4940 }, { "epoch": 0.42684107484778727, "grad_norm": 0.3138258457183838, "learning_rate": 0.0002, "loss": 0.2845, "step": 4960 }, { "epoch": 0.428562208214109, "grad_norm": 0.3556279242038727, "learning_rate": 0.0002, "loss": 0.2688, "step": 4980 }, { "epoch": 0.4302833415804307, "grad_norm": 0.4338746964931488, "learning_rate": 0.0002, "loss": 0.272, "step": 5000 }, { "epoch": 0.4320044749467524, "grad_norm": 0.3764958083629608, "learning_rate": 0.0002, "loss": 0.2618, "step": 5020 }, { "epoch": 0.43372560831307416, "grad_norm": 0.3759072721004486, "learning_rate": 0.0002, "loss": 0.2657, "step": 5040 }, { "epoch": 0.4354467416793959, "grad_norm": 0.3582591414451599, "learning_rate": 0.0002, "loss": 0.3033, "step": 5060 }, { "epoch": 0.4371678750457176, "grad_norm": 0.3076396584510803, "learning_rate": 0.0002, "loss": 0.2719, "step": 5080 }, { "epoch": 0.4388890084120393, "grad_norm": 0.4947834610939026, "learning_rate": 0.0002, "loss": 0.2661, "step": 5100 }, { "epoch": 0.44061014177836105, "grad_norm": 0.4582163095474243, "learning_rate": 0.0002, "loss": 0.2687, "step": 5120 }, { "epoch": 0.4423312751446828, "grad_norm": 0.3605941832065582, "learning_rate": 0.0002, "loss": 0.2692, "step": 5140 }, { "epoch": 0.44405240851100447, "grad_norm": 0.37437519431114197, "learning_rate": 0.0002, "loss": 0.264, "step": 5160 }, { "epoch": 0.4457735418773262, "grad_norm": 0.3707126975059509, "learning_rate": 0.0002, "loss": 0.2535, "step": 5180 }, { "epoch": 0.44749467524364794, "grad_norm": 0.3294564485549927, "learning_rate": 0.0002, "loss": 0.2781, "step": 5200 }, { "epoch": 0.4492158086099697, "grad_norm": 0.3913413882255554, "learning_rate": 0.0002, "loss": 0.2532, "step": 5220 }, { "epoch": 0.45093694197629136, "grad_norm": 0.31665486097335815, "learning_rate": 0.0002, "loss": 0.2731, "step": 5240 }, { "epoch": 0.4526580753426131, "grad_norm": 0.405577689409256, "learning_rate": 0.0002, "loss": 0.2559, "step": 5260 }, { "epoch": 0.45437920870893483, "grad_norm": 0.41928738355636597, "learning_rate": 0.0002, "loss": 0.2874, "step": 5280 }, { "epoch": 0.45610034207525657, "grad_norm": 0.4400196671485901, "learning_rate": 0.0002, "loss": 0.2797, "step": 5300 }, { "epoch": 0.45782147544157825, "grad_norm": 0.29811346530914307, "learning_rate": 0.0002, "loss": 0.2619, "step": 5320 }, { "epoch": 0.4595426088079, "grad_norm": 0.3616848289966583, "learning_rate": 0.0002, "loss": 0.2658, "step": 5340 }, { "epoch": 0.4612637421742217, "grad_norm": 0.2878625690937042, "learning_rate": 0.0002, "loss": 0.2606, "step": 5360 }, { "epoch": 0.46298487554054346, "grad_norm": 0.40930724143981934, "learning_rate": 0.0002, "loss": 0.2653, "step": 5380 }, { "epoch": 0.4647060089068652, "grad_norm": 0.29266178607940674, "learning_rate": 0.0002, "loss": 0.2698, "step": 5400 }, { "epoch": 0.4664271422731869, "grad_norm": 0.4216326177120209, "learning_rate": 0.0002, "loss": 0.2622, "step": 5420 }, { "epoch": 0.4681482756395086, "grad_norm": 0.4683050215244293, "learning_rate": 0.0002, "loss": 0.2618, "step": 5440 }, { "epoch": 0.46986940900583035, "grad_norm": 0.41073721647262573, "learning_rate": 0.0002, "loss": 0.2627, "step": 5460 }, { "epoch": 0.4715905423721521, "grad_norm": 0.3913174867630005, "learning_rate": 0.0002, "loss": 0.257, "step": 5480 }, { "epoch": 0.47331167573847377, "grad_norm": 0.39269042015075684, "learning_rate": 0.0002, "loss": 0.2659, "step": 5500 }, { "epoch": 0.4750328091047955, "grad_norm": 0.4052613079547882, "learning_rate": 0.0002, "loss": 0.2523, "step": 5520 }, { "epoch": 0.47675394247111724, "grad_norm": 0.36656561493873596, "learning_rate": 0.0002, "loss": 0.2627, "step": 5540 }, { "epoch": 0.478475075837439, "grad_norm": 0.3679077625274658, "learning_rate": 0.0002, "loss": 0.264, "step": 5560 }, { "epoch": 0.48019620920376066, "grad_norm": 0.49448278546333313, "learning_rate": 0.0002, "loss": 0.2688, "step": 5580 }, { "epoch": 0.4819173425700824, "grad_norm": 0.39515119791030884, "learning_rate": 0.0002, "loss": 0.2498, "step": 5600 }, { "epoch": 0.48363847593640413, "grad_norm": 0.3903290033340454, "learning_rate": 0.0002, "loss": 0.2744, "step": 5620 }, { "epoch": 0.48535960930272587, "grad_norm": 0.41950464248657227, "learning_rate": 0.0002, "loss": 0.2645, "step": 5640 }, { "epoch": 0.48708074266904755, "grad_norm": 0.3693583607673645, "learning_rate": 0.0002, "loss": 0.261, "step": 5660 }, { "epoch": 0.4888018760353693, "grad_norm": 0.3864137530326843, "learning_rate": 0.0002, "loss": 0.2685, "step": 5680 }, { "epoch": 0.490523009401691, "grad_norm": 0.5029241442680359, "learning_rate": 0.0002, "loss": 0.2531, "step": 5700 }, { "epoch": 0.49224414276801276, "grad_norm": 0.37595513463020325, "learning_rate": 0.0002, "loss": 0.2775, "step": 5720 }, { "epoch": 0.49396527613433444, "grad_norm": 0.4152761399745941, "learning_rate": 0.0002, "loss": 0.2448, "step": 5740 }, { "epoch": 0.4956864095006562, "grad_norm": 0.36774635314941406, "learning_rate": 0.0002, "loss": 0.2731, "step": 5760 }, { "epoch": 0.4974075428669779, "grad_norm": 0.347969114780426, "learning_rate": 0.0002, "loss": 0.2689, "step": 5780 }, { "epoch": 0.49912867623329965, "grad_norm": 0.37085452675819397, "learning_rate": 0.0002, "loss": 0.2545, "step": 5800 }, { "epoch": 0.5008498095996213, "grad_norm": 0.3683365285396576, "learning_rate": 0.0002, "loss": 0.2755, "step": 5820 }, { "epoch": 0.5025709429659431, "grad_norm": 0.2994697690010071, "learning_rate": 0.0002, "loss": 0.2565, "step": 5840 }, { "epoch": 0.5042920763322648, "grad_norm": 0.3561314046382904, "learning_rate": 0.0002, "loss": 0.2697, "step": 5860 }, { "epoch": 0.5060132096985865, "grad_norm": 0.47199949622154236, "learning_rate": 0.0002, "loss": 0.2457, "step": 5880 }, { "epoch": 0.5077343430649083, "grad_norm": 0.3466067612171173, "learning_rate": 0.0002, "loss": 0.2571, "step": 5900 }, { "epoch": 0.50945547643123, "grad_norm": 0.26111066341400146, "learning_rate": 0.0002, "loss": 0.2509, "step": 5920 }, { "epoch": 0.5111766097975516, "grad_norm": 0.32691895961761475, "learning_rate": 0.0002, "loss": 0.2575, "step": 5940 }, { "epoch": 0.5128977431638734, "grad_norm": 0.31905749440193176, "learning_rate": 0.0002, "loss": 0.2549, "step": 5960 }, { "epoch": 0.5146188765301951, "grad_norm": 0.3610023558139801, "learning_rate": 0.0002, "loss": 0.263, "step": 5980 }, { "epoch": 0.5163400098965168, "grad_norm": 0.2800861895084381, "learning_rate": 0.0002, "loss": 0.2517, "step": 6000 }, { "epoch": 0.5180611432628386, "grad_norm": 0.33658984303474426, "learning_rate": 0.0002, "loss": 0.2633, "step": 6020 }, { "epoch": 0.5197822766291603, "grad_norm": 0.4304841160774231, "learning_rate": 0.0002, "loss": 0.2759, "step": 6040 }, { "epoch": 0.5215034099954821, "grad_norm": 0.41126030683517456, "learning_rate": 0.0002, "loss": 0.2719, "step": 6060 }, { "epoch": 0.5232245433618038, "grad_norm": 0.5020288825035095, "learning_rate": 0.0002, "loss": 0.2532, "step": 6080 }, { "epoch": 0.5249456767281254, "grad_norm": 0.32016468048095703, "learning_rate": 0.0002, "loss": 0.2601, "step": 6100 }, { "epoch": 0.5266668100944472, "grad_norm": 0.33081743121147156, "learning_rate": 0.0002, "loss": 0.2544, "step": 6120 }, { "epoch": 0.5283879434607689, "grad_norm": 0.3016170561313629, "learning_rate": 0.0002, "loss": 0.2454, "step": 6140 }, { "epoch": 0.5301090768270906, "grad_norm": 0.3886088728904724, "learning_rate": 0.0002, "loss": 0.2693, "step": 6160 }, { "epoch": 0.5318302101934124, "grad_norm": 0.37182730436325073, "learning_rate": 0.0002, "loss": 0.2599, "step": 6180 }, { "epoch": 0.5335513435597341, "grad_norm": 0.2848903238773346, "learning_rate": 0.0002, "loss": 0.26, "step": 6200 }, { "epoch": 0.5352724769260558, "grad_norm": 0.45262956619262695, "learning_rate": 0.0002, "loss": 0.2587, "step": 6220 }, { "epoch": 0.5369936102923776, "grad_norm": 0.3710488975048065, "learning_rate": 0.0002, "loss": 0.2608, "step": 6240 }, { "epoch": 0.5387147436586992, "grad_norm": 0.39627817273139954, "learning_rate": 0.0002, "loss": 0.2505, "step": 6260 }, { "epoch": 0.5404358770250209, "grad_norm": 0.35057738423347473, "learning_rate": 0.0002, "loss": 0.2619, "step": 6280 }, { "epoch": 0.5421570103913427, "grad_norm": 0.3486057221889496, "learning_rate": 0.0002, "loss": 0.2472, "step": 6300 }, { "epoch": 0.5438781437576644, "grad_norm": 0.330469012260437, "learning_rate": 0.0002, "loss": 0.2458, "step": 6320 }, { "epoch": 0.5455992771239861, "grad_norm": 0.34958067536354065, "learning_rate": 0.0002, "loss": 0.2462, "step": 6340 }, { "epoch": 0.5473204104903079, "grad_norm": 0.3149743676185608, "learning_rate": 0.0002, "loss": 0.2465, "step": 6360 }, { "epoch": 0.5490415438566296, "grad_norm": 0.33875513076782227, "learning_rate": 0.0002, "loss": 0.2593, "step": 6380 }, { "epoch": 0.5507626772229514, "grad_norm": 0.8838099241256714, "learning_rate": 0.0002, "loss": 0.2712, "step": 6400 }, { "epoch": 0.552483810589273, "grad_norm": 0.3521359860897064, "learning_rate": 0.0002, "loss": 0.2441, "step": 6420 }, { "epoch": 0.5542049439555947, "grad_norm": 0.4659444987773895, "learning_rate": 0.0002, "loss": 0.2526, "step": 6440 }, { "epoch": 0.5559260773219165, "grad_norm": 0.4129142165184021, "learning_rate": 0.0002, "loss": 0.2586, "step": 6460 }, { "epoch": 0.5576472106882382, "grad_norm": 0.36697232723236084, "learning_rate": 0.0002, "loss": 0.2367, "step": 6480 }, { "epoch": 0.5593683440545599, "grad_norm": 0.37690311670303345, "learning_rate": 0.0002, "loss": 0.253, "step": 6500 }, { "epoch": 0.5610894774208817, "grad_norm": 0.2961209714412689, "learning_rate": 0.0002, "loss": 0.2441, "step": 6520 }, { "epoch": 0.5628106107872034, "grad_norm": 0.32476598024368286, "learning_rate": 0.0002, "loss": 0.2552, "step": 6540 }, { "epoch": 0.5645317441535251, "grad_norm": 0.36377888917922974, "learning_rate": 0.0002, "loss": 0.259, "step": 6560 }, { "epoch": 0.5662528775198468, "grad_norm": 0.3714245557785034, "learning_rate": 0.0002, "loss": 0.2499, "step": 6580 }, { "epoch": 0.5679740108861685, "grad_norm": 0.3322594463825226, "learning_rate": 0.0002, "loss": 0.2437, "step": 6600 }, { "epoch": 0.5696951442524902, "grad_norm": 0.38657745718955994, "learning_rate": 0.0002, "loss": 0.2452, "step": 6620 }, { "epoch": 0.571416277618812, "grad_norm": 0.34600910544395447, "learning_rate": 0.0002, "loss": 0.2454, "step": 6640 }, { "epoch": 0.5731374109851337, "grad_norm": 0.3906444311141968, "learning_rate": 0.0002, "loss": 0.254, "step": 6660 }, { "epoch": 0.5748585443514554, "grad_norm": 0.37352487444877625, "learning_rate": 0.0002, "loss": 0.2489, "step": 6680 }, { "epoch": 0.5765796777177772, "grad_norm": 0.3105640411376953, "learning_rate": 0.0002, "loss": 0.2523, "step": 6700 }, { "epoch": 0.5783008110840989, "grad_norm": 0.3537706732749939, "learning_rate": 0.0002, "loss": 0.2505, "step": 6720 }, { "epoch": 0.5800219444504207, "grad_norm": 0.45570480823516846, "learning_rate": 0.0002, "loss": 0.2386, "step": 6740 }, { "epoch": 0.5817430778167423, "grad_norm": 0.36528217792510986, "learning_rate": 0.0002, "loss": 0.2382, "step": 6760 }, { "epoch": 0.583464211183064, "grad_norm": 0.3420511782169342, "learning_rate": 0.0002, "loss": 0.2456, "step": 6780 }, { "epoch": 0.5851853445493858, "grad_norm": 0.35168027877807617, "learning_rate": 0.0002, "loss": 0.2578, "step": 6800 }, { "epoch": 0.5869064779157075, "grad_norm": 0.3705766201019287, "learning_rate": 0.0002, "loss": 0.2481, "step": 6820 }, { "epoch": 0.5886276112820292, "grad_norm": 0.4245602786540985, "learning_rate": 0.0002, "loss": 0.2603, "step": 6840 }, { "epoch": 0.590348744648351, "grad_norm": 0.42988812923431396, "learning_rate": 0.0002, "loss": 0.2308, "step": 6860 }, { "epoch": 0.5920698780146727, "grad_norm": 0.3902256190776825, "learning_rate": 0.0002, "loss": 0.2382, "step": 6880 }, { "epoch": 0.5937910113809944, "grad_norm": 0.3394165337085724, "learning_rate": 0.0002, "loss": 0.2264, "step": 6900 }, { "epoch": 0.5955121447473161, "grad_norm": 0.40591520071029663, "learning_rate": 0.0002, "loss": 0.2487, "step": 6920 }, { "epoch": 0.5972332781136378, "grad_norm": 0.37964189052581787, "learning_rate": 0.0002, "loss": 0.2466, "step": 6940 }, { "epoch": 0.5989544114799595, "grad_norm": 0.4443446099758148, "learning_rate": 0.0002, "loss": 0.2384, "step": 6960 }, { "epoch": 0.6006755448462813, "grad_norm": 0.40669143199920654, "learning_rate": 0.0002, "loss": 0.244, "step": 6980 }, { "epoch": 0.602396678212603, "grad_norm": 0.40179166197776794, "learning_rate": 0.0002, "loss": 0.2368, "step": 7000 }, { "epoch": 0.6041178115789247, "grad_norm": 0.4410141408443451, "learning_rate": 0.0002, "loss": 0.2388, "step": 7020 }, { "epoch": 0.6058389449452465, "grad_norm": 0.43160775303840637, "learning_rate": 0.0002, "loss": 0.2377, "step": 7040 }, { "epoch": 0.6075600783115682, "grad_norm": 0.40169423818588257, "learning_rate": 0.0002, "loss": 0.2566, "step": 7060 }, { "epoch": 0.6092812116778898, "grad_norm": 0.43439412117004395, "learning_rate": 0.0002, "loss": 0.2559, "step": 7080 }, { "epoch": 0.6110023450442116, "grad_norm": 0.40532273054122925, "learning_rate": 0.0002, "loss": 0.242, "step": 7100 }, { "epoch": 0.6127234784105333, "grad_norm": 0.3553959131240845, "learning_rate": 0.0002, "loss": 0.2497, "step": 7120 }, { "epoch": 0.614444611776855, "grad_norm": 0.33834314346313477, "learning_rate": 0.0002, "loss": 0.2441, "step": 7140 }, { "epoch": 0.6161657451431768, "grad_norm": 0.432019978761673, "learning_rate": 0.0002, "loss": 0.2701, "step": 7160 }, { "epoch": 0.6178868785094985, "grad_norm": 0.35808584094047546, "learning_rate": 0.0002, "loss": 0.2436, "step": 7180 }, { "epoch": 0.6196080118758203, "grad_norm": 0.5244463682174683, "learning_rate": 0.0002, "loss": 0.2356, "step": 7200 }, { "epoch": 0.621329145242142, "grad_norm": 0.3416566848754883, "learning_rate": 0.0002, "loss": 0.2498, "step": 7220 }, { "epoch": 0.6230502786084636, "grad_norm": 0.377354234457016, "learning_rate": 0.0002, "loss": 0.2361, "step": 7240 }, { "epoch": 0.6247714119747854, "grad_norm": 0.3780601918697357, "learning_rate": 0.0002, "loss": 0.2414, "step": 7260 }, { "epoch": 0.6264925453411071, "grad_norm": 0.4537530541419983, "learning_rate": 0.0002, "loss": 0.2553, "step": 7280 }, { "epoch": 0.6282136787074288, "grad_norm": 0.34764549136161804, "learning_rate": 0.0002, "loss": 0.2294, "step": 7300 }, { "epoch": 0.6299348120737506, "grad_norm": 0.4038061499595642, "learning_rate": 0.0002, "loss": 0.2267, "step": 7320 }, { "epoch": 0.6316559454400723, "grad_norm": 0.3201640546321869, "learning_rate": 0.0002, "loss": 0.2407, "step": 7340 }, { "epoch": 0.633377078806394, "grad_norm": 0.437897652387619, "learning_rate": 0.0002, "loss": 0.238, "step": 7360 }, { "epoch": 0.6350982121727158, "grad_norm": 0.32471516728401184, "learning_rate": 0.0002, "loss": 0.2403, "step": 7380 }, { "epoch": 0.6368193455390374, "grad_norm": 0.2739992141723633, "learning_rate": 0.0002, "loss": 0.2507, "step": 7400 }, { "epoch": 0.6385404789053591, "grad_norm": 0.33521589636802673, "learning_rate": 0.0002, "loss": 0.2462, "step": 7420 }, { "epoch": 0.6402616122716809, "grad_norm": 0.3575187027454376, "learning_rate": 0.0002, "loss": 0.2445, "step": 7440 }, { "epoch": 0.6419827456380026, "grad_norm": 0.3016495704650879, "learning_rate": 0.0002, "loss": 0.2324, "step": 7460 }, { "epoch": 0.6437038790043244, "grad_norm": 0.4337838888168335, "learning_rate": 0.0002, "loss": 0.2508, "step": 7480 }, { "epoch": 0.6454250123706461, "grad_norm": 0.35893139243125916, "learning_rate": 0.0002, "loss": 0.2318, "step": 7500 }, { "epoch": 0.6471461457369678, "grad_norm": 0.3144209682941437, "learning_rate": 0.0002, "loss": 0.2384, "step": 7520 }, { "epoch": 0.6488672791032896, "grad_norm": 0.30142101645469666, "learning_rate": 0.0002, "loss": 0.2439, "step": 7540 }, { "epoch": 0.6505884124696112, "grad_norm": 0.3650546669960022, "learning_rate": 0.0002, "loss": 0.2495, "step": 7560 }, { "epoch": 0.6523095458359329, "grad_norm": 0.41195932030677795, "learning_rate": 0.0002, "loss": 0.2497, "step": 7580 }, { "epoch": 0.6540306792022547, "grad_norm": 0.34587377309799194, "learning_rate": 0.0002, "loss": 0.2372, "step": 7600 }, { "epoch": 0.6557518125685764, "grad_norm": 0.35166653990745544, "learning_rate": 0.0002, "loss": 0.2332, "step": 7620 }, { "epoch": 0.6574729459348981, "grad_norm": 0.46442800760269165, "learning_rate": 0.0002, "loss": 0.232, "step": 7640 }, { "epoch": 0.6591940793012199, "grad_norm": 0.32510513067245483, "learning_rate": 0.0002, "loss": 0.2286, "step": 7660 }, { "epoch": 0.6609152126675416, "grad_norm": 0.3486369848251343, "learning_rate": 0.0002, "loss": 0.2311, "step": 7680 }, { "epoch": 0.6626363460338633, "grad_norm": 0.4640563130378723, "learning_rate": 0.0002, "loss": 0.2386, "step": 7700 }, { "epoch": 0.664357479400185, "grad_norm": 0.4044354557991028, "learning_rate": 0.0002, "loss": 0.2408, "step": 7720 }, { "epoch": 0.6660786127665067, "grad_norm": 0.38926753401756287, "learning_rate": 0.0002, "loss": 0.2537, "step": 7740 }, { "epoch": 0.6677997461328284, "grad_norm": 0.428072988986969, "learning_rate": 0.0002, "loss": 0.231, "step": 7760 }, { "epoch": 0.6695208794991502, "grad_norm": 0.2939753830432892, "learning_rate": 0.0002, "loss": 0.2424, "step": 7780 }, { "epoch": 0.6712420128654719, "grad_norm": 0.41152870655059814, "learning_rate": 0.0002, "loss": 0.2422, "step": 7800 }, { "epoch": 0.6729631462317937, "grad_norm": 0.35313352942466736, "learning_rate": 0.0002, "loss": 0.2401, "step": 7820 }, { "epoch": 0.6746842795981154, "grad_norm": 0.4280535876750946, "learning_rate": 0.0002, "loss": 0.2309, "step": 7840 }, { "epoch": 0.6764054129644371, "grad_norm": 0.36120399832725525, "learning_rate": 0.0002, "loss": 0.2339, "step": 7860 }, { "epoch": 0.6781265463307588, "grad_norm": 0.39646589756011963, "learning_rate": 0.0002, "loss": 0.2464, "step": 7880 }, { "epoch": 0.6798476796970805, "grad_norm": 0.4060419797897339, "learning_rate": 0.0002, "loss": 0.2331, "step": 7900 }, { "epoch": 0.6815688130634022, "grad_norm": 0.4130091369152069, "learning_rate": 0.0002, "loss": 0.2331, "step": 7920 }, { "epoch": 0.683289946429724, "grad_norm": 0.31750190258026123, "learning_rate": 0.0002, "loss": 0.2371, "step": 7940 }, { "epoch": 0.6850110797960457, "grad_norm": 0.4835965037345886, "learning_rate": 0.0002, "loss": 0.2292, "step": 7960 }, { "epoch": 0.6867322131623674, "grad_norm": 0.3932303488254547, "learning_rate": 0.0002, "loss": 0.2243, "step": 7980 }, { "epoch": 0.6884533465286892, "grad_norm": 0.3820064067840576, "learning_rate": 0.0002, "loss": 0.2288, "step": 8000 }, { "epoch": 0.6901744798950109, "grad_norm": 0.40602800250053406, "learning_rate": 0.0002, "loss": 0.2259, "step": 8020 }, { "epoch": 0.6918956132613326, "grad_norm": 0.3742893636226654, "learning_rate": 0.0002, "loss": 0.2395, "step": 8040 }, { "epoch": 0.6936167466276543, "grad_norm": 0.39692652225494385, "learning_rate": 0.0002, "loss": 0.2296, "step": 8060 }, { "epoch": 0.695337879993976, "grad_norm": 0.31871652603149414, "learning_rate": 0.0002, "loss": 0.2444, "step": 8080 }, { "epoch": 0.6970590133602977, "grad_norm": 0.3136846125125885, "learning_rate": 0.0002, "loss": 0.2344, "step": 8100 }, { "epoch": 0.6987801467266195, "grad_norm": 0.3090713918209076, "learning_rate": 0.0002, "loss": 0.2353, "step": 8120 }, { "epoch": 0.7005012800929412, "grad_norm": 0.33116286993026733, "learning_rate": 0.0002, "loss": 0.2205, "step": 8140 }, { "epoch": 0.702222413459263, "grad_norm": 0.2979792058467865, "learning_rate": 0.0002, "loss": 0.2371, "step": 8160 }, { "epoch": 0.7039435468255847, "grad_norm": 0.3515206277370453, "learning_rate": 0.0002, "loss": 0.2293, "step": 8180 }, { "epoch": 0.7056646801919064, "grad_norm": 0.2888669967651367, "learning_rate": 0.0002, "loss": 0.234, "step": 8200 }, { "epoch": 0.707385813558228, "grad_norm": 0.3582359552383423, "learning_rate": 0.0002, "loss": 0.2291, "step": 8220 }, { "epoch": 0.7091069469245498, "grad_norm": 0.4168799817562103, "learning_rate": 0.0002, "loss": 0.2241, "step": 8240 }, { "epoch": 0.7108280802908715, "grad_norm": 0.3614865243434906, "learning_rate": 0.0002, "loss": 0.2323, "step": 8260 }, { "epoch": 0.7125492136571933, "grad_norm": 0.35036012530326843, "learning_rate": 0.0002, "loss": 0.2285, "step": 8280 }, { "epoch": 0.714270347023515, "grad_norm": 0.460549533367157, "learning_rate": 0.0002, "loss": 0.2303, "step": 8300 }, { "epoch": 0.7159914803898367, "grad_norm": 0.44449788331985474, "learning_rate": 0.0002, "loss": 0.2279, "step": 8320 }, { "epoch": 0.7177126137561585, "grad_norm": 0.29927289485931396, "learning_rate": 0.0002, "loss": 0.2235, "step": 8340 }, { "epoch": 0.7194337471224802, "grad_norm": 0.3563356101512909, "learning_rate": 0.0002, "loss": 0.2307, "step": 8360 }, { "epoch": 0.7211548804888018, "grad_norm": 0.42358216643333435, "learning_rate": 0.0002, "loss": 0.227, "step": 8380 }, { "epoch": 0.7228760138551236, "grad_norm": 0.42043471336364746, "learning_rate": 0.0002, "loss": 0.2199, "step": 8400 }, { "epoch": 0.7245971472214453, "grad_norm": 0.3534243702888489, "learning_rate": 0.0002, "loss": 0.2384, "step": 8420 }, { "epoch": 0.726318280587767, "grad_norm": 0.35375455021858215, "learning_rate": 0.0002, "loss": 0.2337, "step": 8440 }, { "epoch": 0.7280394139540888, "grad_norm": 0.31629830598831177, "learning_rate": 0.0002, "loss": 0.2326, "step": 8460 }, { "epoch": 0.7297605473204105, "grad_norm": 0.3449802100658417, "learning_rate": 0.0002, "loss": 0.221, "step": 8480 }, { "epoch": 0.7314816806867322, "grad_norm": 0.44133901596069336, "learning_rate": 0.0002, "loss": 0.2304, "step": 8500 }, { "epoch": 0.733202814053054, "grad_norm": 0.3913390040397644, "learning_rate": 0.0002, "loss": 0.2277, "step": 8520 }, { "epoch": 0.7349239474193756, "grad_norm": 0.3438524901866913, "learning_rate": 0.0002, "loss": 0.2484, "step": 8540 }, { "epoch": 0.7366450807856973, "grad_norm": 0.34400486946105957, "learning_rate": 0.0002, "loss": 0.2257, "step": 8560 }, { "epoch": 0.7383662141520191, "grad_norm": 0.40113121271133423, "learning_rate": 0.0002, "loss": 0.2246, "step": 8580 }, { "epoch": 0.7400873475183408, "grad_norm": 0.3908839821815491, "learning_rate": 0.0002, "loss": 0.2178, "step": 8600 }, { "epoch": 0.7418084808846626, "grad_norm": 0.3656678795814514, "learning_rate": 0.0002, "loss": 0.2345, "step": 8620 }, { "epoch": 0.7435296142509843, "grad_norm": 0.3811242878437042, "learning_rate": 0.0002, "loss": 0.2387, "step": 8640 }, { "epoch": 0.745250747617306, "grad_norm": 0.3093118965625763, "learning_rate": 0.0002, "loss": 0.2179, "step": 8660 }, { "epoch": 0.7469718809836278, "grad_norm": 0.42958253622055054, "learning_rate": 0.0002, "loss": 0.2212, "step": 8680 }, { "epoch": 0.7486930143499494, "grad_norm": 0.3202790319919586, "learning_rate": 0.0002, "loss": 0.2394, "step": 8700 }, { "epoch": 0.7504141477162711, "grad_norm": 0.42061832547187805, "learning_rate": 0.0002, "loss": 0.2282, "step": 8720 }, { "epoch": 0.7521352810825929, "grad_norm": 0.2962038516998291, "learning_rate": 0.0002, "loss": 0.2322, "step": 8740 }, { "epoch": 0.7538564144489146, "grad_norm": 0.3707858920097351, "learning_rate": 0.0002, "loss": 0.2251, "step": 8760 }, { "epoch": 0.7555775478152363, "grad_norm": 0.424730509519577, "learning_rate": 0.0002, "loss": 0.2345, "step": 8780 }, { "epoch": 0.7572986811815581, "grad_norm": 0.34581223130226135, "learning_rate": 0.0002, "loss": 0.2196, "step": 8800 }, { "epoch": 0.7590198145478798, "grad_norm": 0.436617374420166, "learning_rate": 0.0002, "loss": 0.2332, "step": 8820 }, { "epoch": 0.7607409479142015, "grad_norm": 0.3432689905166626, "learning_rate": 0.0002, "loss": 0.231, "step": 8840 }, { "epoch": 0.7624620812805232, "grad_norm": 0.36506009101867676, "learning_rate": 0.0002, "loss": 0.2299, "step": 8860 }, { "epoch": 0.7641832146468449, "grad_norm": 0.29211345314979553, "learning_rate": 0.0002, "loss": 0.2169, "step": 8880 }, { "epoch": 0.7659043480131666, "grad_norm": 0.3884661793708801, "learning_rate": 0.0002, "loss": 0.2205, "step": 8900 }, { "epoch": 0.7676254813794884, "grad_norm": 0.3466391861438751, "learning_rate": 0.0002, "loss": 0.2249, "step": 8920 }, { "epoch": 0.7693466147458101, "grad_norm": 0.3165854811668396, "learning_rate": 0.0002, "loss": 0.2436, "step": 8940 }, { "epoch": 0.7710677481121319, "grad_norm": 0.42315399646759033, "learning_rate": 0.0002, "loss": 0.2207, "step": 8960 }, { "epoch": 0.7727888814784536, "grad_norm": 0.3629299998283386, "learning_rate": 0.0002, "loss": 0.2356, "step": 8980 }, { "epoch": 0.7745100148447753, "grad_norm": 0.4586748480796814, "learning_rate": 0.0002, "loss": 0.2152, "step": 9000 }, { "epoch": 0.776231148211097, "grad_norm": 0.27728915214538574, "learning_rate": 0.0002, "loss": 0.2284, "step": 9020 }, { "epoch": 0.7779522815774187, "grad_norm": 0.3490050435066223, "learning_rate": 0.0002, "loss": 0.2166, "step": 9040 }, { "epoch": 0.7796734149437404, "grad_norm": 0.2839438021183014, "learning_rate": 0.0002, "loss": 0.2283, "step": 9060 }, { "epoch": 0.7813945483100622, "grad_norm": 0.35332831740379333, "learning_rate": 0.0002, "loss": 0.2336, "step": 9080 }, { "epoch": 0.7831156816763839, "grad_norm": 0.4003342390060425, "learning_rate": 0.0002, "loss": 0.2279, "step": 9100 }, { "epoch": 0.7848368150427056, "grad_norm": 0.32392072677612305, "learning_rate": 0.0002, "loss": 0.235, "step": 9120 }, { "epoch": 0.7865579484090274, "grad_norm": 0.3040711581707001, "learning_rate": 0.0002, "loss": 0.2234, "step": 9140 }, { "epoch": 0.7882790817753491, "grad_norm": 0.35234126448631287, "learning_rate": 0.0002, "loss": 0.2215, "step": 9160 }, { "epoch": 0.7900002151416707, "grad_norm": 0.3502811789512634, "learning_rate": 0.0002, "loss": 0.225, "step": 9180 }, { "epoch": 0.7917213485079925, "grad_norm": 0.3539245128631592, "learning_rate": 0.0002, "loss": 0.2289, "step": 9200 }, { "epoch": 0.7934424818743142, "grad_norm": 0.3154076635837555, "learning_rate": 0.0002, "loss": 0.2282, "step": 9220 }, { "epoch": 0.795163615240636, "grad_norm": 0.3497600853443146, "learning_rate": 0.0002, "loss": 0.2159, "step": 9240 }, { "epoch": 0.7968847486069577, "grad_norm": 0.3394204080104828, "learning_rate": 0.0002, "loss": 0.2175, "step": 9260 }, { "epoch": 0.7986058819732794, "grad_norm": 0.42728209495544434, "learning_rate": 0.0002, "loss": 0.2316, "step": 9280 }, { "epoch": 0.8003270153396012, "grad_norm": 0.30218786001205444, "learning_rate": 0.0002, "loss": 0.2268, "step": 9300 }, { "epoch": 0.8020481487059229, "grad_norm": 0.3418307602405548, "learning_rate": 0.0002, "loss": 0.2315, "step": 9320 }, { "epoch": 0.8037692820722446, "grad_norm": 0.38792407512664795, "learning_rate": 0.0002, "loss": 0.2302, "step": 9340 }, { "epoch": 0.8054904154385663, "grad_norm": 0.2837861180305481, "learning_rate": 0.0002, "loss": 0.2193, "step": 9360 }, { "epoch": 0.807211548804888, "grad_norm": 0.4480486214160919, "learning_rate": 0.0002, "loss": 0.2264, "step": 9380 }, { "epoch": 0.8089326821712097, "grad_norm": 0.25541505217552185, "learning_rate": 0.0002, "loss": 0.2302, "step": 9400 }, { "epoch": 0.8106538155375315, "grad_norm": 0.28958380222320557, "learning_rate": 0.0002, "loss": 0.2269, "step": 9420 }, { "epoch": 0.8123749489038532, "grad_norm": 0.2757406532764435, "learning_rate": 0.0002, "loss": 0.2155, "step": 9440 }, { "epoch": 0.8140960822701749, "grad_norm": 0.3484242558479309, "learning_rate": 0.0002, "loss": 0.2346, "step": 9460 }, { "epoch": 0.8158172156364967, "grad_norm": 0.23964551091194153, "learning_rate": 0.0002, "loss": 0.2208, "step": 9480 }, { "epoch": 0.8175383490028184, "grad_norm": 0.3568115234375, "learning_rate": 0.0002, "loss": 0.2187, "step": 9500 }, { "epoch": 0.81925948236914, "grad_norm": 0.4668004512786865, "learning_rate": 0.0002, "loss": 0.2306, "step": 9520 }, { "epoch": 0.8209806157354618, "grad_norm": 0.33082979917526245, "learning_rate": 0.0002, "loss": 0.2135, "step": 9540 }, { "epoch": 0.8227017491017835, "grad_norm": 0.3717847764492035, "learning_rate": 0.0002, "loss": 0.2203, "step": 9560 }, { "epoch": 0.8244228824681052, "grad_norm": 0.37038078904151917, "learning_rate": 0.0002, "loss": 0.2271, "step": 9580 }, { "epoch": 0.826144015834427, "grad_norm": 0.33178070187568665, "learning_rate": 0.0002, "loss": 0.2222, "step": 9600 }, { "epoch": 0.8278651492007487, "grad_norm": 0.27431583404541016, "learning_rate": 0.0002, "loss": 0.2304, "step": 9620 }, { "epoch": 0.8295862825670705, "grad_norm": 0.33868762850761414, "learning_rate": 0.0002, "loss": 0.2197, "step": 9640 }, { "epoch": 0.8313074159333922, "grad_norm": 0.34042665362358093, "learning_rate": 0.0002, "loss": 0.2194, "step": 9660 }, { "epoch": 0.8330285492997138, "grad_norm": 0.431411474943161, "learning_rate": 0.0002, "loss": 0.2228, "step": 9680 }, { "epoch": 0.8347496826660356, "grad_norm": 0.2958236634731293, "learning_rate": 0.0002, "loss": 0.2268, "step": 9700 }, { "epoch": 0.8364708160323573, "grad_norm": 0.31208258867263794, "learning_rate": 0.0002, "loss": 0.2371, "step": 9720 }, { "epoch": 0.838191949398679, "grad_norm": 0.3337579667568207, "learning_rate": 0.0002, "loss": 0.2312, "step": 9740 }, { "epoch": 0.8399130827650008, "grad_norm": 0.5392299294471741, "learning_rate": 0.0002, "loss": 0.2324, "step": 9760 }, { "epoch": 0.8416342161313225, "grad_norm": 0.3481907248497009, "learning_rate": 0.0002, "loss": 0.2241, "step": 9780 }, { "epoch": 0.8433553494976442, "grad_norm": 0.29136040806770325, "learning_rate": 0.0002, "loss": 0.2147, "step": 9800 }, { "epoch": 0.845076482863966, "grad_norm": 0.2877753973007202, "learning_rate": 0.0002, "loss": 0.1988, "step": 9820 }, { "epoch": 0.8467976162302876, "grad_norm": 0.4024602174758911, "learning_rate": 0.0002, "loss": 0.2161, "step": 9840 }, { "epoch": 0.8485187495966093, "grad_norm": 0.28684961795806885, "learning_rate": 0.0002, "loss": 0.2199, "step": 9860 }, { "epoch": 0.8502398829629311, "grad_norm": 0.31743690371513367, "learning_rate": 0.0002, "loss": 0.2196, "step": 9880 }, { "epoch": 0.8519610163292528, "grad_norm": 0.3548375368118286, "learning_rate": 0.0002, "loss": 0.2179, "step": 9900 }, { "epoch": 0.8536821496955745, "grad_norm": 0.31174436211586, "learning_rate": 0.0002, "loss": 0.2236, "step": 9920 }, { "epoch": 0.8554032830618963, "grad_norm": 0.2791374623775482, "learning_rate": 0.0002, "loss": 0.2257, "step": 9940 }, { "epoch": 0.857124416428218, "grad_norm": 0.3355172574520111, "learning_rate": 0.0002, "loss": 0.2244, "step": 9960 }, { "epoch": 0.8588455497945398, "grad_norm": 0.3883482813835144, "learning_rate": 0.0002, "loss": 0.2178, "step": 9980 }, { "epoch": 0.8605666831608614, "grad_norm": 0.3601789176464081, "learning_rate": 0.0002, "loss": 0.227, "step": 10000 }, { "epoch": 0.8622878165271831, "grad_norm": 0.24994397163391113, "learning_rate": 0.0002, "loss": 0.2165, "step": 10020 }, { "epoch": 0.8640089498935049, "grad_norm": 0.40082284808158875, "learning_rate": 0.0002, "loss": 0.228, "step": 10040 }, { "epoch": 0.8657300832598266, "grad_norm": 0.47781533002853394, "learning_rate": 0.0002, "loss": 0.2283, "step": 10060 }, { "epoch": 0.8674512166261483, "grad_norm": 0.3652552366256714, "learning_rate": 0.0002, "loss": 0.2207, "step": 10080 }, { "epoch": 0.8691723499924701, "grad_norm": 0.38674041628837585, "learning_rate": 0.0002, "loss": 0.2162, "step": 10100 }, { "epoch": 0.8708934833587918, "grad_norm": 0.32635965943336487, "learning_rate": 0.0002, "loss": 0.2173, "step": 10120 }, { "epoch": 0.8726146167251135, "grad_norm": 0.28276535868644714, "learning_rate": 0.0002, "loss": 0.2203, "step": 10140 }, { "epoch": 0.8743357500914352, "grad_norm": 0.363016813993454, "learning_rate": 0.0002, "loss": 0.2355, "step": 10160 }, { "epoch": 0.8760568834577569, "grad_norm": 0.302642822265625, "learning_rate": 0.0002, "loss": 0.2099, "step": 10180 }, { "epoch": 0.8777780168240786, "grad_norm": 0.25143080949783325, "learning_rate": 0.0002, "loss": 0.2233, "step": 10200 }, { "epoch": 0.8794991501904004, "grad_norm": 0.37836262583732605, "learning_rate": 0.0002, "loss": 0.2218, "step": 10220 }, { "epoch": 0.8812202835567221, "grad_norm": 0.3723512887954712, "learning_rate": 0.0002, "loss": 0.219, "step": 10240 }, { "epoch": 0.8829414169230438, "grad_norm": 0.33276501297950745, "learning_rate": 0.0002, "loss": 0.2153, "step": 10260 }, { "epoch": 0.8846625502893656, "grad_norm": 0.3877188563346863, "learning_rate": 0.0002, "loss": 0.2213, "step": 10280 }, { "epoch": 0.8863836836556873, "grad_norm": 0.36920130252838135, "learning_rate": 0.0002, "loss": 0.2099, "step": 10300 }, { "epoch": 0.8881048170220089, "grad_norm": 0.2991655170917511, "learning_rate": 0.0002, "loss": 0.218, "step": 10320 }, { "epoch": 0.8898259503883307, "grad_norm": 0.4785827100276947, "learning_rate": 0.0002, "loss": 0.2226, "step": 10340 }, { "epoch": 0.8915470837546524, "grad_norm": 0.30903199315071106, "learning_rate": 0.0002, "loss": 0.2059, "step": 10360 }, { "epoch": 0.8932682171209742, "grad_norm": 0.3614109754562378, "learning_rate": 0.0002, "loss": 0.2271, "step": 10380 }, { "epoch": 0.8949893504872959, "grad_norm": 0.3223751485347748, "learning_rate": 0.0002, "loss": 0.2223, "step": 10400 }, { "epoch": 0.8967104838536176, "grad_norm": 0.3175138533115387, "learning_rate": 0.0002, "loss": 0.2369, "step": 10420 }, { "epoch": 0.8984316172199394, "grad_norm": 0.37820857763290405, "learning_rate": 0.0002, "loss": 0.2206, "step": 10440 }, { "epoch": 0.9001527505862611, "grad_norm": 0.4233035445213318, "learning_rate": 0.0002, "loss": 0.2169, "step": 10460 }, { "epoch": 0.9018738839525827, "grad_norm": 0.4142135977745056, "learning_rate": 0.0002, "loss": 0.21, "step": 10480 }, { "epoch": 0.9035950173189045, "grad_norm": 0.43564194440841675, "learning_rate": 0.0002, "loss": 0.2231, "step": 10500 }, { "epoch": 0.9053161506852262, "grad_norm": 0.3781276345252991, "learning_rate": 0.0002, "loss": 0.2258, "step": 10520 }, { "epoch": 0.9070372840515479, "grad_norm": 0.4279311001300812, "learning_rate": 0.0002, "loss": 0.2287, "step": 10540 }, { "epoch": 0.9087584174178697, "grad_norm": 0.3078250586986542, "learning_rate": 0.0002, "loss": 0.212, "step": 10560 }, { "epoch": 0.9104795507841914, "grad_norm": 0.3039957880973816, "learning_rate": 0.0002, "loss": 0.2137, "step": 10580 }, { "epoch": 0.9122006841505131, "grad_norm": 0.3359488546848297, "learning_rate": 0.0002, "loss": 0.2206, "step": 10600 }, { "epoch": 0.9139218175168349, "grad_norm": 0.30003640055656433, "learning_rate": 0.0002, "loss": 0.213, "step": 10620 }, { "epoch": 0.9156429508831565, "grad_norm": 0.3175880014896393, "learning_rate": 0.0002, "loss": 0.2099, "step": 10640 }, { "epoch": 0.9173640842494782, "grad_norm": 0.3824600577354431, "learning_rate": 0.0002, "loss": 0.2126, "step": 10660 }, { "epoch": 0.9190852176158, "grad_norm": 0.3589145839214325, "learning_rate": 0.0002, "loss": 0.2234, "step": 10680 }, { "epoch": 0.9208063509821217, "grad_norm": 0.3821096420288086, "learning_rate": 0.0002, "loss": 0.2163, "step": 10700 }, { "epoch": 0.9225274843484434, "grad_norm": 0.41358160972595215, "learning_rate": 0.0002, "loss": 0.2207, "step": 10720 }, { "epoch": 0.9242486177147652, "grad_norm": 0.2841893434524536, "learning_rate": 0.0002, "loss": 0.2418, "step": 10740 }, { "epoch": 0.9259697510810869, "grad_norm": 0.29627254605293274, "learning_rate": 0.0002, "loss": 0.2259, "step": 10760 }, { "epoch": 0.9276908844474087, "grad_norm": 0.3516784906387329, "learning_rate": 0.0002, "loss": 0.2117, "step": 10780 }, { "epoch": 0.9294120178137304, "grad_norm": 0.37627148628234863, "learning_rate": 0.0002, "loss": 0.2236, "step": 10800 }, { "epoch": 0.931133151180052, "grad_norm": 0.3956553041934967, "learning_rate": 0.0002, "loss": 0.2091, "step": 10820 }, { "epoch": 0.9328542845463738, "grad_norm": 0.3255549967288971, "learning_rate": 0.0002, "loss": 0.2247, "step": 10840 }, { "epoch": 0.9345754179126955, "grad_norm": 0.3365786075592041, "learning_rate": 0.0002, "loss": 0.1955, "step": 10860 }, { "epoch": 0.9362965512790172, "grad_norm": 0.3629949688911438, "learning_rate": 0.0002, "loss": 0.2137, "step": 10880 }, { "epoch": 0.938017684645339, "grad_norm": 0.3127789795398712, "learning_rate": 0.0002, "loss": 0.2077, "step": 10900 }, { "epoch": 0.9397388180116607, "grad_norm": 0.3616786003112793, "learning_rate": 0.0002, "loss": 0.2125, "step": 10920 }, { "epoch": 0.9414599513779824, "grad_norm": 0.33411502838134766, "learning_rate": 0.0002, "loss": 0.2137, "step": 10940 }, { "epoch": 0.9431810847443042, "grad_norm": 0.3387938439846039, "learning_rate": 0.0002, "loss": 0.2089, "step": 10960 }, { "epoch": 0.9449022181106258, "grad_norm": 0.42915311455726624, "learning_rate": 0.0002, "loss": 0.2144, "step": 10980 }, { "epoch": 0.9466233514769475, "grad_norm": 0.4593462646007538, "learning_rate": 0.0002, "loss": 0.2136, "step": 11000 }, { "epoch": 0.9483444848432693, "grad_norm": 0.3778610825538635, "learning_rate": 0.0002, "loss": 0.2198, "step": 11020 }, { "epoch": 0.950065618209591, "grad_norm": 0.34387239813804626, "learning_rate": 0.0002, "loss": 0.2215, "step": 11040 }, { "epoch": 0.9517867515759127, "grad_norm": 0.38071706891059875, "learning_rate": 0.0002, "loss": 0.2159, "step": 11060 }, { "epoch": 0.9535078849422345, "grad_norm": 0.3698328137397766, "learning_rate": 0.0002, "loss": 0.214, "step": 11080 }, { "epoch": 0.9552290183085562, "grad_norm": 0.43745896220207214, "learning_rate": 0.0002, "loss": 0.2252, "step": 11100 }, { "epoch": 0.956950151674878, "grad_norm": 0.2951034903526306, "learning_rate": 0.0002, "loss": 0.2099, "step": 11120 }, { "epoch": 0.9586712850411996, "grad_norm": 0.35736939311027527, "learning_rate": 0.0002, "loss": 0.222, "step": 11140 }, { "epoch": 0.9603924184075213, "grad_norm": 0.3355426490306854, "learning_rate": 0.0002, "loss": 0.2142, "step": 11160 }, { "epoch": 0.9621135517738431, "grad_norm": 0.2925025522708893, "learning_rate": 0.0002, "loss": 0.2085, "step": 11180 }, { "epoch": 0.9638346851401648, "grad_norm": 0.36077678203582764, "learning_rate": 0.0002, "loss": 0.2097, "step": 11200 }, { "epoch": 0.9655558185064865, "grad_norm": 0.3234957158565521, "learning_rate": 0.0002, "loss": 0.2209, "step": 11220 }, { "epoch": 0.9672769518728083, "grad_norm": 0.32046082615852356, "learning_rate": 0.0002, "loss": 0.2051, "step": 11240 }, { "epoch": 0.96899808523913, "grad_norm": 0.3421325087547302, "learning_rate": 0.0002, "loss": 0.2288, "step": 11260 }, { "epoch": 0.9707192186054517, "grad_norm": 0.35989081859588623, "learning_rate": 0.0002, "loss": 0.2061, "step": 11280 }, { "epoch": 0.9724403519717734, "grad_norm": 0.3135656714439392, "learning_rate": 0.0002, "loss": 0.2238, "step": 11300 }, { "epoch": 0.9741614853380951, "grad_norm": 0.29025885462760925, "learning_rate": 0.0002, "loss": 0.2111, "step": 11320 }, { "epoch": 0.9758826187044168, "grad_norm": 0.4417757987976074, "learning_rate": 0.0002, "loss": 0.2, "step": 11340 }, { "epoch": 0.9776037520707386, "grad_norm": 0.3659359812736511, "learning_rate": 0.0002, "loss": 0.216, "step": 11360 }, { "epoch": 0.9793248854370603, "grad_norm": 0.3737911283969879, "learning_rate": 0.0002, "loss": 0.2139, "step": 11380 }, { "epoch": 0.981046018803382, "grad_norm": 0.38346412777900696, "learning_rate": 0.0002, "loss": 0.2072, "step": 11400 }, { "epoch": 0.9827671521697038, "grad_norm": 0.32871776819229126, "learning_rate": 0.0002, "loss": 0.2197, "step": 11420 }, { "epoch": 0.9844882855360255, "grad_norm": 0.29058969020843506, "learning_rate": 0.0002, "loss": 0.211, "step": 11440 }, { "epoch": 0.9862094189023471, "grad_norm": 0.39237943291664124, "learning_rate": 0.0002, "loss": 0.2053, "step": 11460 }, { "epoch": 0.9879305522686689, "grad_norm": 0.3916817307472229, "learning_rate": 0.0002, "loss": 0.2177, "step": 11480 }, { "epoch": 0.9896516856349906, "grad_norm": 0.2479276806116104, "learning_rate": 0.0002, "loss": 0.2128, "step": 11500 }, { "epoch": 0.9913728190013124, "grad_norm": 0.33634355664253235, "learning_rate": 0.0002, "loss": 0.2097, "step": 11520 }, { "epoch": 0.9930939523676341, "grad_norm": 0.336535781621933, "learning_rate": 0.0002, "loss": 0.208, "step": 11540 }, { "epoch": 0.9948150857339558, "grad_norm": 0.4115926921367645, "learning_rate": 0.0002, "loss": 0.2057, "step": 11560 }, { "epoch": 0.9965362191002776, "grad_norm": 0.3465426564216614, "learning_rate": 0.0002, "loss": 0.204, "step": 11580 }, { "epoch": 0.9982573524665993, "grad_norm": 0.32643455266952515, "learning_rate": 0.0002, "loss": 0.2068, "step": 11600 }, { "epoch": 0.9999784858329209, "grad_norm": 0.3390646278858185, "learning_rate": 0.0002, "loss": 0.2095, "step": 11620 }, { "epoch": 1.0, "eval_loss": 0.2866213619709015, "eval_runtime": 903.225, "eval_samples_per_second": 4.146, "eval_steps_per_second": 0.519, "step": 11621 } ], "logging_steps": 20, "max_steps": 13000, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 77, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 7.484935677507797e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }