{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.498407643312102, "eval_steps": 500, "global_step": 1569, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01592356687898089, "grad_norm": 4.3297224044799805, "learning_rate": 2.5316455696202533e-05, "loss": 0.86, "step": 10 }, { "epoch": 0.03184713375796178, "grad_norm": 6.637640476226807, "learning_rate": 5.0632911392405066e-05, "loss": 0.4332, "step": 20 }, { "epoch": 0.04777070063694268, "grad_norm": 1.9116368293762207, "learning_rate": 7.59493670886076e-05, "loss": 0.3014, "step": 30 }, { "epoch": 0.06369426751592357, "grad_norm": 1.2928426265716553, "learning_rate": 0.00010126582278481013, "loss": 0.2232, "step": 40 }, { "epoch": 0.07961783439490445, "grad_norm": 1.4354435205459595, "learning_rate": 0.00012658227848101267, "loss": 0.1771, "step": 50 }, { "epoch": 0.09554140127388536, "grad_norm": 1.710893154144287, "learning_rate": 0.0001518987341772152, "loss": 0.2136, "step": 60 }, { "epoch": 0.11146496815286625, "grad_norm": 1.6920921802520752, "learning_rate": 0.00017721518987341773, "loss": 0.1644, "step": 70 }, { "epoch": 0.12738853503184713, "grad_norm": 1.2916326522827148, "learning_rate": 0.00019999977772170748, "loss": 0.1593, "step": 80 }, { "epoch": 0.14331210191082802, "grad_norm": 1.599898099899292, "learning_rate": 0.00019997310552224846, "loss": 0.1578, "step": 90 }, { "epoch": 0.1592356687898089, "grad_norm": 1.4364933967590332, "learning_rate": 0.00019990199125039174, "loss": 0.155, "step": 100 }, { "epoch": 0.1751592356687898, "grad_norm": 1.7148686647415161, "learning_rate": 0.00019978646651929572, "loss": 0.1463, "step": 110 }, { "epoch": 0.1910828025477707, "grad_norm": 1.2754517793655396, "learning_rate": 0.0001996265826843568, "loss": 0.1471, "step": 120 }, { "epoch": 0.2070063694267516, "grad_norm": 1.0604585409164429, "learning_rate": 0.00019942241082037982, "loss": 0.124, "step": 130 }, { "epoch": 0.2229299363057325, "grad_norm": 0.6741822361946106, "learning_rate": 0.00019917404168998256, "loss": 0.0979, "step": 140 }, { "epoch": 0.23885350318471338, "grad_norm": 1.2134864330291748, "learning_rate": 0.00019888158570324795, "loss": 0.1105, "step": 150 }, { "epoch": 0.25477707006369427, "grad_norm": 0.6342090368270874, "learning_rate": 0.00019854517286864245, "loss": 0.099, "step": 160 }, { "epoch": 0.27070063694267515, "grad_norm": 1.0112156867980957, "learning_rate": 0.000198164952735222, "loss": 0.0928, "step": 170 }, { "epoch": 0.28662420382165604, "grad_norm": 1.0317986011505127, "learning_rate": 0.00019774109432615147, "loss": 0.0963, "step": 180 }, { "epoch": 0.30254777070063693, "grad_norm": 0.7419966459274292, "learning_rate": 0.00019727378606356703, "loss": 0.1069, "step": 190 }, { "epoch": 0.3184713375796178, "grad_norm": 1.373571753501892, "learning_rate": 0.00019676323568481498, "loss": 0.1103, "step": 200 }, { "epoch": 0.3343949044585987, "grad_norm": 0.6964902877807617, "learning_rate": 0.00019620967015010395, "loss": 0.0827, "step": 210 }, { "epoch": 0.3503184713375796, "grad_norm": 0.9365194439888, "learning_rate": 0.00019561333554161224, "loss": 0.1046, "step": 220 }, { "epoch": 0.3662420382165605, "grad_norm": 1.899908423423767, "learning_rate": 0.00019497449695409408, "loss": 0.1031, "step": 230 }, { "epoch": 0.3821656050955414, "grad_norm": 0.8126282095909119, "learning_rate": 0.00019429343837703455, "loss": 0.1069, "step": 240 }, { "epoch": 0.3980891719745223, "grad_norm": 1.244973063468933, "learning_rate": 0.00019357046256840473, "loss": 0.082, "step": 250 }, { "epoch": 0.4140127388535032, "grad_norm": 0.8132479786872864, "learning_rate": 0.00019280589092007352, "loss": 0.0867, "step": 260 }, { "epoch": 0.4299363057324841, "grad_norm": 0.8742226958274841, "learning_rate": 0.0001920000633149362, "loss": 0.0761, "step": 270 }, { "epoch": 0.445859872611465, "grad_norm": 0.6832584738731384, "learning_rate": 0.00019115333797582254, "loss": 0.0845, "step": 280 }, { "epoch": 0.46178343949044587, "grad_norm": 0.6030533909797668, "learning_rate": 0.00019026609130625257, "loss": 0.0722, "step": 290 }, { "epoch": 0.47770700636942676, "grad_norm": 0.7257299423217773, "learning_rate": 0.0001893387177231099, "loss": 0.0701, "step": 300 }, { "epoch": 0.49363057324840764, "grad_norm": 0.6390131711959839, "learning_rate": 0.00018837162948130752, "loss": 0.0696, "step": 310 }, { "epoch": 0.5095541401273885, "grad_norm": 1.8211143016815186, "learning_rate": 0.00018736525649052394, "loss": 0.0757, "step": 320 }, { "epoch": 0.5254777070063694, "grad_norm": 1.0733026266098022, "learning_rate": 0.00018632004612409103, "loss": 0.0835, "step": 330 }, { "epoch": 0.5414012738853503, "grad_norm": 0.7683069705963135, "learning_rate": 0.00018523646302011867, "loss": 0.0765, "step": 340 }, { "epoch": 0.5573248407643312, "grad_norm": 0.5842871069908142, "learning_rate": 0.00018411498887494396, "loss": 0.073, "step": 350 }, { "epoch": 0.5732484076433121, "grad_norm": 0.6933338046073914, "learning_rate": 0.0001829561222289984, "loss": 0.0648, "step": 360 }, { "epoch": 0.589171974522293, "grad_norm": 0.42543935775756836, "learning_rate": 0.000181760378245186, "loss": 0.072, "step": 370 }, { "epoch": 0.6050955414012739, "grad_norm": 0.72137451171875, "learning_rate": 0.0001805282884798732, "loss": 0.0661, "step": 380 }, { "epoch": 0.6210191082802548, "grad_norm": 0.6393111944198608, "learning_rate": 0.00017926040064659014, "loss": 0.0841, "step": 390 }, { "epoch": 0.6369426751592356, "grad_norm": 0.46522364020347595, "learning_rate": 0.00017795727837255015, "loss": 0.0668, "step": 400 }, { "epoch": 0.6528662420382165, "grad_norm": 0.790262758731842, "learning_rate": 0.0001766195009480949, "loss": 0.0702, "step": 410 }, { "epoch": 0.6687898089171974, "grad_norm": 0.7343723177909851, "learning_rate": 0.00017524766306917618, "loss": 0.0751, "step": 420 }, { "epoch": 0.6847133757961783, "grad_norm": 0.497380793094635, "learning_rate": 0.00017384237457298987, "loss": 0.0663, "step": 430 }, { "epoch": 0.7006369426751592, "grad_norm": 0.5845641493797302, "learning_rate": 0.00017240426016687863, "loss": 0.0656, "step": 440 }, { "epoch": 0.7165605095541401, "grad_norm": 1.0296058654785156, "learning_rate": 0.00017093395915062428, "loss": 0.0664, "step": 450 }, { "epoch": 0.732484076433121, "grad_norm": 0.7704136371612549, "learning_rate": 0.00016943212513225345, "loss": 0.0597, "step": 460 }, { "epoch": 0.7484076433121019, "grad_norm": 0.7734206318855286, "learning_rate": 0.00016789942573748232, "loss": 0.0638, "step": 470 }, { "epoch": 0.7643312101910829, "grad_norm": 0.6289462447166443, "learning_rate": 0.00016633654231293013, "loss": 0.0573, "step": 480 }, { "epoch": 0.7802547770700637, "grad_norm": 0.6616397500038147, "learning_rate": 0.00016474416962323325, "loss": 0.0506, "step": 490 }, { "epoch": 0.7961783439490446, "grad_norm": 0.3910081386566162, "learning_rate": 0.00016312301554219426, "loss": 0.0457, "step": 500 }, { "epoch": 0.8121019108280255, "grad_norm": 0.6165742874145508, "learning_rate": 0.00016147380073810346, "loss": 0.0503, "step": 510 }, { "epoch": 0.8280254777070064, "grad_norm": 0.6931295990943909, "learning_rate": 0.00015979725835337294, "loss": 0.0481, "step": 520 }, { "epoch": 0.8439490445859873, "grad_norm": 0.7499469518661499, "learning_rate": 0.00015809413367862512, "loss": 0.0725, "step": 530 }, { "epoch": 0.8598726114649682, "grad_norm": 0.4135420620441437, "learning_rate": 0.00015636518382138107, "loss": 0.0646, "step": 540 }, { "epoch": 0.8757961783439491, "grad_norm": 0.4710856080055237, "learning_rate": 0.00015461117736949577, "loss": 0.0538, "step": 550 }, { "epoch": 0.89171974522293, "grad_norm": 0.428799033164978, "learning_rate": 0.00015283289404948976, "loss": 0.0593, "step": 560 }, { "epoch": 0.9076433121019108, "grad_norm": 0.4142407774925232, "learning_rate": 0.0001510311243799295, "loss": 0.0605, "step": 570 }, { "epoch": 0.9235668789808917, "grad_norm": 0.46757104992866516, "learning_rate": 0.0001492066693200096, "loss": 0.0634, "step": 580 }, { "epoch": 0.9394904458598726, "grad_norm": 0.4138467609882355, "learning_rate": 0.0001473603399134948, "loss": 0.0586, "step": 590 }, { "epoch": 0.9554140127388535, "grad_norm": 0.43409624695777893, "learning_rate": 0.00014549295692817778, "loss": 0.0579, "step": 600 }, { "epoch": 0.9713375796178344, "grad_norm": 0.44268661737442017, "learning_rate": 0.0001436053504910151, "loss": 0.0637, "step": 610 }, { "epoch": 0.9872611464968153, "grad_norm": 0.3840467929840088, "learning_rate": 0.00014169835971910238, "loss": 0.0511, "step": 620 }, { "epoch": 1.0031847133757963, "grad_norm": 0.4400401711463928, "learning_rate": 0.00013977283234665273, "loss": 0.0477, "step": 630 }, { "epoch": 1.019108280254777, "grad_norm": 0.3332836925983429, "learning_rate": 0.00013782962434814492, "loss": 0.0478, "step": 640 }, { "epoch": 1.035031847133758, "grad_norm": 0.3282802700996399, "learning_rate": 0.00013586959955780824, "loss": 0.0508, "step": 650 }, { "epoch": 1.0509554140127388, "grad_norm": 0.6200388073921204, "learning_rate": 0.00013389362928561317, "loss": 0.0479, "step": 660 }, { "epoch": 1.0668789808917198, "grad_norm": 0.5026119947433472, "learning_rate": 0.0001319025919299394, "loss": 0.0402, "step": 670 }, { "epoch": 1.0828025477707006, "grad_norm": 0.440782755613327, "learning_rate": 0.00012989737258709203, "loss": 0.0508, "step": 680 }, { "epoch": 1.0987261146496816, "grad_norm": 0.34029659628868103, "learning_rate": 0.0001278788626578407, "loss": 0.044, "step": 690 }, { "epoch": 1.1146496815286624, "grad_norm": 0.46119582653045654, "learning_rate": 0.00012584795945115603, "loss": 0.0426, "step": 700 }, { "epoch": 1.1305732484076434, "grad_norm": 0.47696712613105774, "learning_rate": 0.0001238055657853198, "loss": 0.0515, "step": 710 }, { "epoch": 1.1464968152866242, "grad_norm": 0.49719110131263733, "learning_rate": 0.00012175258958658564, "loss": 0.0453, "step": 720 }, { "epoch": 1.1624203821656052, "grad_norm": 0.36977869272232056, "learning_rate": 0.0001196899434855693, "loss": 0.0529, "step": 730 }, { "epoch": 1.178343949044586, "grad_norm": 0.3333928883075714, "learning_rate": 0.00011761854441154767, "loss": 0.0429, "step": 740 }, { "epoch": 1.194267515923567, "grad_norm": 0.3594333231449127, "learning_rate": 0.0001155393131848467, "loss": 0.044, "step": 750 }, { "epoch": 1.2101910828025477, "grad_norm": 0.33035799860954285, "learning_rate": 0.00011345317410749964, "loss": 0.0425, "step": 760 }, { "epoch": 1.2261146496815287, "grad_norm": 0.431429386138916, "learning_rate": 0.00011136105455235766, "loss": 0.0395, "step": 770 }, { "epoch": 1.2420382165605095, "grad_norm": 0.34141987562179565, "learning_rate": 0.00010926388455083522, "loss": 0.0455, "step": 780 }, { "epoch": 1.2579617834394905, "grad_norm": 0.48105695843696594, "learning_rate": 0.00010716259637947357, "loss": 0.033, "step": 790 }, { "epoch": 1.2738853503184713, "grad_norm": 0.40072181820869446, "learning_rate": 0.0001050581241455064, "loss": 0.0404, "step": 800 }, { "epoch": 1.2898089171974523, "grad_norm": 0.5137969851493835, "learning_rate": 0.00010295140337161146, "loss": 0.0388, "step": 810 }, { "epoch": 1.305732484076433, "grad_norm": 0.6727393865585327, "learning_rate": 0.00010084337058003303, "loss": 0.0455, "step": 820 }, { "epoch": 1.321656050955414, "grad_norm": 0.36769551038742065, "learning_rate": 9.873496287626019e-05, "loss": 0.0416, "step": 830 }, { "epoch": 1.3375796178343948, "grad_norm": 0.40487590432167053, "learning_rate": 9.662711753244551e-05, "loss": 0.0394, "step": 840 }, { "epoch": 1.3535031847133758, "grad_norm": 0.3262194097042084, "learning_rate": 9.452077157074994e-05, "loss": 0.0371, "step": 850 }, { "epoch": 1.3694267515923566, "grad_norm": 0.49285051226615906, "learning_rate": 9.241686134679867e-05, "loss": 0.0434, "step": 860 }, { "epoch": 1.3853503184713376, "grad_norm": 0.4286457896232605, "learning_rate": 9.031632213343339e-05, "loss": 0.0388, "step": 870 }, { "epoch": 1.4012738853503186, "grad_norm": 0.4375673234462738, "learning_rate": 8.822008770494572e-05, "loss": 0.0419, "step": 880 }, { "epoch": 1.4171974522292994, "grad_norm": 0.34244051575660706, "learning_rate": 8.612908992197705e-05, "loss": 0.0478, "step": 890 }, { "epoch": 1.4331210191082802, "grad_norm": 0.325065016746521, "learning_rate": 8.404425831726894e-05, "loss": 0.04, "step": 900 }, { "epoch": 1.4490445859872612, "grad_norm": 0.4267027676105499, "learning_rate": 8.196651968244826e-05, "loss": 0.0377, "step": 910 }, { "epoch": 1.4649681528662422, "grad_norm": 0.5654813051223755, "learning_rate": 7.989679765603108e-05, "loss": 0.0445, "step": 920 }, { "epoch": 1.480891719745223, "grad_norm": 0.39815276861190796, "learning_rate": 7.783601231282812e-05, "loss": 0.0553, "step": 930 }, { "epoch": 1.4968152866242037, "grad_norm": 0.46311256289482117, "learning_rate": 7.578507975493448e-05, "loss": 0.0373, "step": 940 }, { "epoch": 1.5127388535031847, "grad_norm": 0.37769854068756104, "learning_rate": 7.374491170448525e-05, "loss": 0.0368, "step": 950 }, { "epoch": 1.5286624203821657, "grad_norm": 0.2921994924545288, "learning_rate": 7.17164150983584e-05, "loss": 0.0428, "step": 960 }, { "epoch": 1.5445859872611465, "grad_norm": 0.48072826862335205, "learning_rate": 6.970049168500474e-05, "loss": 0.0329, "step": 970 }, { "epoch": 1.5605095541401273, "grad_norm": 0.37165966629981995, "learning_rate": 6.769803762358443e-05, "loss": 0.0376, "step": 980 }, { "epoch": 1.5764331210191083, "grad_norm": 0.309915155172348, "learning_rate": 6.570994308558812e-05, "loss": 0.032, "step": 990 }, { "epoch": 1.5923566878980893, "grad_norm": 0.41205787658691406, "learning_rate": 6.373709185911998e-05, "loss": 0.036, "step": 1000 }, { "epoch": 1.60828025477707, "grad_norm": 0.45491883158683777, "learning_rate": 6.17803609560181e-05, "loss": 0.0453, "step": 1010 }, { "epoch": 1.6242038216560508, "grad_norm": 0.4330281615257263, "learning_rate": 5.98406202219875e-05, "loss": 0.0319, "step": 1020 }, { "epoch": 1.6401273885350318, "grad_norm": 0.2400377243757248, "learning_rate": 5.791873194991872e-05, "loss": 0.0342, "step": 1030 }, { "epoch": 1.6560509554140128, "grad_norm": 0.26502475142478943, "learning_rate": 5.601555049656382e-05, "loss": 0.0318, "step": 1040 }, { "epoch": 1.6719745222929936, "grad_norm": 0.2682415246963501, "learning_rate": 5.41319219027404e-05, "loss": 0.0301, "step": 1050 }, { "epoch": 1.6878980891719744, "grad_norm": 0.38560014963150024, "learning_rate": 5.226868351723244e-05, "loss": 0.0432, "step": 1060 }, { "epoch": 1.7038216560509554, "grad_norm": 0.1889456957578659, "learning_rate": 5.042666362455498e-05, "loss": 0.0255, "step": 1070 }, { "epoch": 1.7197452229299364, "grad_norm": 0.3188437223434448, "learning_rate": 4.860668107674823e-05, "loss": 0.037, "step": 1080 }, { "epoch": 1.7356687898089171, "grad_norm": 0.7025167942047119, "learning_rate": 4.6809544929365004e-05, "loss": 0.0357, "step": 1090 }, { "epoch": 1.7515923566878981, "grad_norm": 0.39571383595466614, "learning_rate": 4.503605408181286e-05, "loss": 0.0273, "step": 1100 }, { "epoch": 1.767515923566879, "grad_norm": 0.2839336693286896, "learning_rate": 4.3286996922211034e-05, "loss": 0.0302, "step": 1110 }, { "epoch": 1.78343949044586, "grad_norm": 0.3293486535549164, "learning_rate": 4.156315097692037e-05, "loss": 0.0308, "step": 1120 }, { "epoch": 1.799363057324841, "grad_norm": 0.3942393958568573, "learning_rate": 3.986528256490141e-05, "loss": 0.0293, "step": 1130 }, { "epoch": 1.8152866242038217, "grad_norm": 0.31392380595207214, "learning_rate": 3.8194146457054655e-05, "loss": 0.0249, "step": 1140 }, { "epoch": 1.8312101910828025, "grad_norm": 0.3941175639629364, "learning_rate": 3.655048554069478e-05, "loss": 0.032, "step": 1150 }, { "epoch": 1.8471337579617835, "grad_norm": 0.2796083092689514, "learning_rate": 3.4935030489306883e-05, "loss": 0.034, "step": 1160 }, { "epoch": 1.8630573248407645, "grad_norm": 0.30333149433135986, "learning_rate": 3.334849943773323e-05, "loss": 0.0351, "step": 1170 }, { "epoch": 1.8789808917197452, "grad_norm": 0.3254244029521942, "learning_rate": 3.179159766293282e-05, "loss": 0.0298, "step": 1180 }, { "epoch": 1.894904458598726, "grad_norm": 0.20831990242004395, "learning_rate": 3.0265017270457775e-05, "loss": 0.0288, "step": 1190 }, { "epoch": 1.910828025477707, "grad_norm": 0.31671157479286194, "learning_rate": 2.8769436886784408e-05, "loss": 0.0334, "step": 1200 }, { "epoch": 1.926751592356688, "grad_norm": 0.5833495259284973, "learning_rate": 2.730552135763632e-05, "loss": 0.0425, "step": 1210 }, { "epoch": 1.9426751592356688, "grad_norm": 0.32988494634628296, "learning_rate": 2.5873921452433915e-05, "loss": 0.0328, "step": 1220 }, { "epoch": 1.9585987261146496, "grad_norm": 0.20500314235687256, "learning_rate": 2.4475273575000936e-05, "loss": 0.0273, "step": 1230 }, { "epoch": 1.9745222929936306, "grad_norm": 0.33419910073280334, "learning_rate": 2.3110199480657525e-05, "loss": 0.0264, "step": 1240 }, { "epoch": 1.9904458598726116, "grad_norm": 0.20512598752975464, "learning_rate": 2.1779305999824884e-05, "loss": 0.0287, "step": 1250 }, { "epoch": 2.0063694267515926, "grad_norm": 0.4389275312423706, "learning_rate": 2.0483184768264596e-05, "loss": 0.0306, "step": 1260 }, { "epoch": 2.022292993630573, "grad_norm": 0.18635383248329163, "learning_rate": 1.9222411964072884e-05, "loss": 0.028, "step": 1270 }, { "epoch": 2.038216560509554, "grad_norm": 0.32153043150901794, "learning_rate": 1.799754805154603e-05, "loss": 0.0274, "step": 1280 }, { "epoch": 2.054140127388535, "grad_norm": 0.448691189289093, "learning_rate": 1.6809137532031704e-05, "loss": 0.0306, "step": 1290 }, { "epoch": 2.070063694267516, "grad_norm": 0.38386714458465576, "learning_rate": 1.565770870187585e-05, "loss": 0.0247, "step": 1300 }, { "epoch": 2.0859872611464967, "grad_norm": 0.18705415725708008, "learning_rate": 1.4543773417573925e-05, "loss": 0.025, "step": 1310 }, { "epoch": 2.1019108280254777, "grad_norm": 0.23585732281208038, "learning_rate": 1.3467826868229994e-05, "loss": 0.0265, "step": 1320 }, { "epoch": 2.1178343949044587, "grad_norm": 0.36259782314300537, "learning_rate": 1.243034735542512e-05, "loss": 0.0305, "step": 1330 }, { "epoch": 2.1337579617834397, "grad_norm": 0.34997355937957764, "learning_rate": 1.1431796080593283e-05, "loss": 0.032, "step": 1340 }, { "epoch": 2.1496815286624202, "grad_norm": 0.3579126298427582, "learning_rate": 1.0472616939998492e-05, "loss": 0.024, "step": 1350 }, { "epoch": 2.1656050955414012, "grad_norm": 0.41504374146461487, "learning_rate": 9.553236327405246e-06, "loss": 0.0259, "step": 1360 }, { "epoch": 2.1815286624203822, "grad_norm": 0.2632668912410736, "learning_rate": 8.674062944529216e-06, "loss": 0.0301, "step": 1370 }, { "epoch": 2.1974522292993632, "grad_norm": 0.3617275059223175, "learning_rate": 7.835487619352811e-06, "loss": 0.0232, "step": 1380 }, { "epoch": 2.213375796178344, "grad_norm": 0.3129204213619232, "learning_rate": 7.037883132386547e-06, "loss": 0.0355, "step": 1390 }, { "epoch": 2.229299363057325, "grad_norm": 0.2824781835079193, "learning_rate": 6.2816040509530165e-06, "loss": 0.0269, "step": 1400 }, { "epoch": 2.245222929936306, "grad_norm": 0.20426948368549347, "learning_rate": 5.566986571567401e-06, "loss": 0.0355, "step": 1410 }, { "epoch": 2.261146496815287, "grad_norm": 0.22026273608207703, "learning_rate": 4.8943483704846475e-06, "loss": 0.0267, "step": 1420 }, { "epoch": 2.2770700636942673, "grad_norm": 0.27001041173934937, "learning_rate": 4.263988462479484e-06, "loss": 0.0286, "step": 1430 }, { "epoch": 2.2929936305732483, "grad_norm": 0.22612400352954865, "learning_rate": 3.676187067922421e-06, "loss": 0.0263, "step": 1440 }, { "epoch": 2.3089171974522293, "grad_norm": 0.3205125033855438, "learning_rate": 3.131205488210409e-06, "loss": 0.0311, "step": 1450 }, { "epoch": 2.3248407643312103, "grad_norm": 0.2743224501609802, "learning_rate": 2.6292859896079213e-06, "loss": 0.0233, "step": 1460 }, { "epoch": 2.340764331210191, "grad_norm": 0.23394609987735748, "learning_rate": 2.170651695549786e-06, "loss": 0.0161, "step": 1470 }, { "epoch": 2.356687898089172, "grad_norm": 0.23162613809108734, "learning_rate": 1.7555064874538397e-06, "loss": 0.0243, "step": 1480 }, { "epoch": 2.372611464968153, "grad_norm": 0.3181140720844269, "learning_rate": 1.3840349140874619e-06, "loss": 0.0224, "step": 1490 }, { "epoch": 2.388535031847134, "grad_norm": 0.29145529866218567, "learning_rate": 1.0564021095281652e-06, "loss": 0.0345, "step": 1500 }, { "epoch": 2.404458598726115, "grad_norm": 0.321485698223114, "learning_rate": 7.727537197548707e-07, "loss": 0.0205, "step": 1510 }, { "epoch": 2.4203821656050954, "grad_norm": 0.2429160177707672, "learning_rate": 5.332158379024122e-07, "loss": 0.0217, "step": 1520 }, { "epoch": 2.4363057324840764, "grad_norm": 0.22818560898303986, "learning_rate": 3.3789494820803957e-07, "loss": 0.0219, "step": 1530 }, { "epoch": 2.4522292993630574, "grad_norm": 0.2582818865776062, "learning_rate": 1.8687787867489592e-07, "loss": 0.0186, "step": 1540 }, { "epoch": 2.468152866242038, "grad_norm": 0.2430548369884491, "learning_rate": 8.023176247348163e-08, "loss": 0.0344, "step": 1550 }, { "epoch": 2.484076433121019, "grad_norm": 0.38517120480537415, "learning_rate": 1.8004008098226887e-08, "loss": 0.0225, "step": 1560 }, { "epoch": 2.498407643312102, "step": 1569, "total_flos": 5.687859742190496e+16, "train_loss": 0.06483107545027846, "train_runtime": 706.8123, "train_samples_per_second": 35.517, "train_steps_per_second": 2.22 } ], "logging_steps": 10, "max_steps": 1569, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.687859742190496e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }