{ "best_metric": null, "best_model_checkpoint": null, "epoch": 9.321127579192096, "eval_steps": 1000000, "global_step": 64148, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07265329845975008, "grad_norm": 0.39395979046821594, "learning_rate": 9.992734670154026e-06, "loss": 8.1621, "step": 500 }, { "epoch": 0.14530659691950015, "grad_norm": 0.5780870318412781, "learning_rate": 9.985469340308051e-06, "loss": 7.0196, "step": 1000 }, { "epoch": 0.21795989537925023, "grad_norm": 0.5540204048156738, "learning_rate": 9.978204010462076e-06, "loss": 6.6142, "step": 1500 }, { "epoch": 0.2906131938390003, "grad_norm": 0.5443109273910522, "learning_rate": 9.970938680616102e-06, "loss": 6.3813, "step": 2000 }, { "epoch": 0.36326649229875035, "grad_norm": 0.6196579337120056, "learning_rate": 9.963673350770125e-06, "loss": 6.2207, "step": 2500 }, { "epoch": 0.43591979075850046, "grad_norm": 0.5789604187011719, "learning_rate": 9.95640802092415e-06, "loss": 6.0957, "step": 3000 }, { "epoch": 0.5085730892182505, "grad_norm": 0.63157719373703, "learning_rate": 9.949142691078175e-06, "loss": 5.9962, "step": 3500 }, { "epoch": 0.5812263876780006, "grad_norm": 0.5619252920150757, "learning_rate": 9.9418773612322e-06, "loss": 5.9116, "step": 4000 }, { "epoch": 0.6538796861377506, "grad_norm": 0.6025997996330261, "learning_rate": 9.934612031386226e-06, "loss": 5.8332, "step": 4500 }, { "epoch": 0.7265329845975007, "grad_norm": 0.6860383152961731, "learning_rate": 9.927346701540251e-06, "loss": 5.7695, "step": 5000 }, { "epoch": 0.7991862830572508, "grad_norm": 0.7735297679901123, "learning_rate": 9.920081371694276e-06, "loss": 5.7068, "step": 5500 }, { "epoch": 0.8718395815170009, "grad_norm": 0.6047775745391846, "learning_rate": 9.912816041848301e-06, "loss": 5.65, "step": 6000 }, { "epoch": 0.9444928799767509, "grad_norm": 0.8214231133460999, "learning_rate": 9.905550712002325e-06, "loss": 5.6052, "step": 6500 }, { "epoch": 1.017146178436501, "grad_norm": 0.8017473816871643, "learning_rate": 9.89828538215635e-06, "loss": 5.5515, "step": 7000 }, { "epoch": 1.0897994768962511, "grad_norm": 0.7803598642349243, "learning_rate": 9.891020052310377e-06, "loss": 5.5058, "step": 7500 }, { "epoch": 1.1624527753560012, "grad_norm": 0.7102193236351013, "learning_rate": 9.8837547224644e-06, "loss": 5.4684, "step": 8000 }, { "epoch": 1.2351060738157513, "grad_norm": 0.7121894359588623, "learning_rate": 9.876489392618425e-06, "loss": 5.4271, "step": 8500 }, { "epoch": 1.3077593722755014, "grad_norm": 0.7547861337661743, "learning_rate": 9.86922406277245e-06, "loss": 5.397, "step": 9000 }, { "epoch": 1.3804126707352513, "grad_norm": 0.7697557210922241, "learning_rate": 9.861958732926476e-06, "loss": 5.3563, "step": 9500 }, { "epoch": 1.4530659691950014, "grad_norm": 0.8195068836212158, "learning_rate": 9.854693403080501e-06, "loss": 5.3228, "step": 10000 }, { "epoch": 1.5257192676547515, "grad_norm": 0.7637478113174438, "learning_rate": 9.847428073234524e-06, "loss": 5.2946, "step": 10500 }, { "epoch": 1.5983725661145016, "grad_norm": 0.8972837328910828, "learning_rate": 9.840162743388551e-06, "loss": 5.2588, "step": 11000 }, { "epoch": 1.6710258645742515, "grad_norm": 0.9061026573181152, "learning_rate": 9.832897413542576e-06, "loss": 5.2321, "step": 11500 }, { "epoch": 1.7436791630340016, "grad_norm": 0.7589295506477356, "learning_rate": 9.8256320836966e-06, "loss": 5.2065, "step": 12000 }, { "epoch": 1.8163324614937517, "grad_norm": 0.8363276720046997, "learning_rate": 9.818366753850625e-06, "loss": 5.1819, "step": 12500 }, { "epoch": 1.8889857599535018, "grad_norm": 0.8336675763130188, "learning_rate": 9.81110142400465e-06, "loss": 5.1575, "step": 13000 }, { "epoch": 1.961639058413252, "grad_norm": 0.8027577996253967, "learning_rate": 9.803836094158675e-06, "loss": 5.1314, "step": 13500 }, { "epoch": 2.034292356873002, "grad_norm": 0.7707503437995911, "learning_rate": 9.7965707643127e-06, "loss": 5.1088, "step": 14000 }, { "epoch": 2.106945655332752, "grad_norm": 0.8998398184776306, "learning_rate": 9.789305434466726e-06, "loss": 5.0808, "step": 14500 }, { "epoch": 2.1795989537925022, "grad_norm": 0.8320822715759277, "learning_rate": 9.782040104620751e-06, "loss": 5.0634, "step": 15000 }, { "epoch": 2.2522522522522523, "grad_norm": 0.8281899094581604, "learning_rate": 9.774774774774776e-06, "loss": 5.0444, "step": 15500 }, { "epoch": 2.3249055507120024, "grad_norm": 0.8026254177093506, "learning_rate": 9.767509444928801e-06, "loss": 5.0179, "step": 16000 }, { "epoch": 2.3975588491717525, "grad_norm": 0.776054322719574, "learning_rate": 9.760244115082825e-06, "loss": 5.0009, "step": 16500 }, { "epoch": 2.4702121476315027, "grad_norm": 0.8994656801223755, "learning_rate": 9.75297878523685e-06, "loss": 4.9812, "step": 17000 }, { "epoch": 2.5428654460912528, "grad_norm": 0.8583637475967407, "learning_rate": 9.745713455390875e-06, "loss": 4.9628, "step": 17500 }, { "epoch": 2.615518744551003, "grad_norm": 0.893622636795044, "learning_rate": 9.7384481255449e-06, "loss": 4.9464, "step": 18000 }, { "epoch": 2.688172043010753, "grad_norm": 0.8372634053230286, "learning_rate": 9.731182795698925e-06, "loss": 4.9304, "step": 18500 }, { "epoch": 2.7608253414705026, "grad_norm": 0.9035200476646423, "learning_rate": 9.72391746585295e-06, "loss": 4.9127, "step": 19000 }, { "epoch": 2.8334786399302527, "grad_norm": 0.8660902380943298, "learning_rate": 9.716652136006976e-06, "loss": 4.8998, "step": 19500 }, { "epoch": 2.906131938390003, "grad_norm": 0.9463688731193542, "learning_rate": 9.709386806161001e-06, "loss": 4.8778, "step": 20000 }, { "epoch": 2.978785236849753, "grad_norm": 0.8671707510948181, "learning_rate": 9.702121476315024e-06, "loss": 4.8658, "step": 20500 }, { "epoch": 3.051438535309503, "grad_norm": 0.9653003215789795, "learning_rate": 9.69485614646905e-06, "loss": 4.853, "step": 21000 }, { "epoch": 3.124091833769253, "grad_norm": 0.8984787464141846, "learning_rate": 9.687590816623077e-06, "loss": 4.8364, "step": 21500 }, { "epoch": 3.1967451322290033, "grad_norm": 0.8803947567939758, "learning_rate": 9.6803254867771e-06, "loss": 4.8177, "step": 22000 }, { "epoch": 3.2693984306887534, "grad_norm": 0.8390781283378601, "learning_rate": 9.673060156931125e-06, "loss": 4.8077, "step": 22500 }, { "epoch": 3.3420517291485035, "grad_norm": 0.9399817585945129, "learning_rate": 9.66579482708515e-06, "loss": 4.7942, "step": 23000 }, { "epoch": 3.4147050276082536, "grad_norm": 0.864647626876831, "learning_rate": 9.658529497239176e-06, "loss": 4.7819, "step": 23500 }, { "epoch": 3.4873583260680037, "grad_norm": 0.863230288028717, "learning_rate": 9.6512641673932e-06, "loss": 4.7675, "step": 24000 }, { "epoch": 3.5600116245277533, "grad_norm": 0.9053711891174316, "learning_rate": 9.643998837547224e-06, "loss": 4.7562, "step": 24500 }, { "epoch": 3.6326649229875034, "grad_norm": 0.9143590331077576, "learning_rate": 9.636733507701251e-06, "loss": 4.7408, "step": 25000 }, { "epoch": 3.7053182214472535, "grad_norm": 0.8819567561149597, "learning_rate": 9.629468177855276e-06, "loss": 4.7347, "step": 25500 }, { "epoch": 3.7779715199070036, "grad_norm": 0.9433513283729553, "learning_rate": 9.6222028480093e-06, "loss": 4.7227, "step": 26000 }, { "epoch": 3.8506248183667537, "grad_norm": 0.9881005883216858, "learning_rate": 9.614937518163325e-06, "loss": 4.7096, "step": 26500 }, { "epoch": 3.923278116826504, "grad_norm": 0.890139639377594, "learning_rate": 9.60767218831735e-06, "loss": 4.7, "step": 27000 }, { "epoch": 3.995931415286254, "grad_norm": 0.9619746804237366, "learning_rate": 9.600406858471375e-06, "loss": 4.6879, "step": 27500 }, { "epoch": 4.068584713746004, "grad_norm": 0.8994712233543396, "learning_rate": 9.5931415286254e-06, "loss": 4.6762, "step": 28000 }, { "epoch": 4.141238012205754, "grad_norm": 0.9636611938476562, "learning_rate": 9.585876198779426e-06, "loss": 4.6643, "step": 28500 }, { "epoch": 4.213891310665504, "grad_norm": 0.8346700072288513, "learning_rate": 9.57861086893345e-06, "loss": 4.6541, "step": 29000 }, { "epoch": 4.286544609125254, "grad_norm": 0.8618379831314087, "learning_rate": 9.571345539087476e-06, "loss": 4.6473, "step": 29500 }, { "epoch": 4.3591979075850045, "grad_norm": 0.9965615272521973, "learning_rate": 9.564080209241501e-06, "loss": 4.636, "step": 30000 }, { "epoch": 4.431851206044755, "grad_norm": 0.9595836400985718, "learning_rate": 9.556814879395525e-06, "loss": 4.6262, "step": 30500 }, { "epoch": 4.504504504504505, "grad_norm": 0.918498694896698, "learning_rate": 9.54954954954955e-06, "loss": 4.6139, "step": 31000 }, { "epoch": 4.577157802964255, "grad_norm": 0.9323708415031433, "learning_rate": 9.542284219703575e-06, "loss": 4.6138, "step": 31500 }, { "epoch": 4.649811101424005, "grad_norm": 0.9415541291236877, "learning_rate": 9.5350188898576e-06, "loss": 4.6008, "step": 32000 }, { "epoch": 4.722464399883755, "grad_norm": 0.9424421787261963, "learning_rate": 9.527753560011625e-06, "loss": 4.5893, "step": 32500 }, { "epoch": 4.795117698343505, "grad_norm": 0.942669689655304, "learning_rate": 9.52048823016565e-06, "loss": 4.5866, "step": 33000 }, { "epoch": 4.867770996803255, "grad_norm": 0.9807618856430054, "learning_rate": 9.513222900319676e-06, "loss": 4.5736, "step": 33500 }, { "epoch": 4.940424295263005, "grad_norm": 0.960895299911499, "learning_rate": 9.5059575704737e-06, "loss": 4.5619, "step": 34000 }, { "epoch": 5.013077593722755, "grad_norm": 0.9801125526428223, "learning_rate": 9.498692240627724e-06, "loss": 4.5551, "step": 34500 }, { "epoch": 5.0857308921825055, "grad_norm": 0.9244375228881836, "learning_rate": 9.49142691078175e-06, "loss": 4.5408, "step": 35000 }, { "epoch": 5.158384190642255, "grad_norm": 0.9349271059036255, "learning_rate": 9.484161580935776e-06, "loss": 4.5358, "step": 35500 }, { "epoch": 5.231037489102005, "grad_norm": 0.9496700763702393, "learning_rate": 9.4768962510898e-06, "loss": 4.5284, "step": 36000 }, { "epoch": 5.303690787561755, "grad_norm": 0.8824469447135925, "learning_rate": 9.469630921243825e-06, "loss": 4.5225, "step": 36500 }, { "epoch": 5.376344086021505, "grad_norm": 0.9746178984642029, "learning_rate": 9.46236559139785e-06, "loss": 4.5084, "step": 37000 }, { "epoch": 5.448997384481255, "grad_norm": 0.9742151498794556, "learning_rate": 9.455100261551875e-06, "loss": 4.5036, "step": 37500 }, { "epoch": 5.521650682941005, "grad_norm": 1.0122525691986084, "learning_rate": 9.4478349317059e-06, "loss": 4.4951, "step": 38000 }, { "epoch": 5.594303981400755, "grad_norm": 0.9454242587089539, "learning_rate": 9.440569601859924e-06, "loss": 4.4844, "step": 38500 }, { "epoch": 5.6669572798605055, "grad_norm": 0.9594370722770691, "learning_rate": 9.43330427201395e-06, "loss": 4.4776, "step": 39000 }, { "epoch": 5.739610578320256, "grad_norm": 0.9644564986228943, "learning_rate": 9.426038942167976e-06, "loss": 4.4748, "step": 39500 }, { "epoch": 5.812263876780006, "grad_norm": 0.9564418792724609, "learning_rate": 9.418773612322e-06, "loss": 4.4592, "step": 40000 }, { "epoch": 5.884917175239756, "grad_norm": 0.9269986152648926, "learning_rate": 9.411508282476025e-06, "loss": 4.455, "step": 40500 }, { "epoch": 5.957570473699506, "grad_norm": 0.9712278842926025, "learning_rate": 9.40424295263005e-06, "loss": 4.4455, "step": 41000 }, { "epoch": 6.030223772159256, "grad_norm": 0.9372689723968506, "learning_rate": 9.396977622784075e-06, "loss": 4.4318, "step": 41500 }, { "epoch": 6.102877070619006, "grad_norm": 0.942115843296051, "learning_rate": 9.3897122929381e-06, "loss": 4.4264, "step": 42000 }, { "epoch": 6.175530369078756, "grad_norm": 1.0541362762451172, "learning_rate": 9.382446963092125e-06, "loss": 4.4163, "step": 42500 }, { "epoch": 6.248183667538506, "grad_norm": 0.9615015387535095, "learning_rate": 9.37518163324615e-06, "loss": 4.4096, "step": 43000 }, { "epoch": 6.320836965998256, "grad_norm": 1.0179786682128906, "learning_rate": 9.367916303400176e-06, "loss": 4.4034, "step": 43500 }, { "epoch": 6.3934902644580065, "grad_norm": 0.9816853404045105, "learning_rate": 9.360650973554201e-06, "loss": 4.3956, "step": 44000 }, { "epoch": 6.466143562917757, "grad_norm": 1.0108842849731445, "learning_rate": 9.353385643708224e-06, "loss": 4.3811, "step": 44500 }, { "epoch": 6.538796861377507, "grad_norm": 1.0273536443710327, "learning_rate": 9.34612031386225e-06, "loss": 4.3748, "step": 45000 }, { "epoch": 6.611450159837257, "grad_norm": 0.9984000325202942, "learning_rate": 9.338854984016275e-06, "loss": 4.3711, "step": 45500 }, { "epoch": 6.684103458297007, "grad_norm": 0.9320149421691895, "learning_rate": 9.3315896541703e-06, "loss": 4.3621, "step": 46000 }, { "epoch": 6.756756756756757, "grad_norm": 1.1042180061340332, "learning_rate": 9.324324324324325e-06, "loss": 4.3516, "step": 46500 }, { "epoch": 6.829410055216507, "grad_norm": 0.9890114665031433, "learning_rate": 9.31705899447835e-06, "loss": 4.3464, "step": 47000 }, { "epoch": 6.902063353676257, "grad_norm": 1.034157156944275, "learning_rate": 9.309793664632375e-06, "loss": 4.3374, "step": 47500 }, { "epoch": 6.974716652136007, "grad_norm": 0.9817051887512207, "learning_rate": 9.3025283347864e-06, "loss": 4.3284, "step": 48000 }, { "epoch": 7.047369950595757, "grad_norm": 1.0558934211730957, "learning_rate": 9.295263004940424e-06, "loss": 4.3154, "step": 48500 }, { "epoch": 7.1200232490555075, "grad_norm": 0.9428159594535828, "learning_rate": 9.28799767509445e-06, "loss": 4.3088, "step": 49000 }, { "epoch": 7.192676547515257, "grad_norm": 1.0073195695877075, "learning_rate": 9.280732345248476e-06, "loss": 4.2983, "step": 49500 }, { "epoch": 7.265329845975007, "grad_norm": 0.9252042174339294, "learning_rate": 9.2734670154025e-06, "loss": 4.2912, "step": 50000 }, { "epoch": 7.337983144434757, "grad_norm": 1.0058454275131226, "learning_rate": 9.266201685556525e-06, "loss": 4.2814, "step": 50500 }, { "epoch": 7.410636442894507, "grad_norm": 1.0393654108047485, "learning_rate": 9.25893635571055e-06, "loss": 4.2693, "step": 51000 }, { "epoch": 7.483289741354257, "grad_norm": 0.9429093599319458, "learning_rate": 9.251671025864575e-06, "loss": 4.2604, "step": 51500 }, { "epoch": 7.555943039814007, "grad_norm": 0.9522086381912231, "learning_rate": 9.2444056960186e-06, "loss": 4.2499, "step": 52000 }, { "epoch": 7.628596338273757, "grad_norm": 0.9854516386985779, "learning_rate": 9.237140366172624e-06, "loss": 4.2411, "step": 52500 }, { "epoch": 7.7012496367335075, "grad_norm": 0.9321109056472778, "learning_rate": 9.22987503632665e-06, "loss": 4.2257, "step": 53000 }, { "epoch": 7.773902935193258, "grad_norm": 0.9426267147064209, "learning_rate": 9.222609706480676e-06, "loss": 4.2225, "step": 53500 }, { "epoch": 7.846556233653008, "grad_norm": 0.9784730672836304, "learning_rate": 9.2153443766347e-06, "loss": 4.2077, "step": 54000 }, { "epoch": 7.919209532112758, "grad_norm": 0.9224970936775208, "learning_rate": 9.208079046788724e-06, "loss": 4.2029, "step": 54500 }, { "epoch": 7.991862830572508, "grad_norm": 0.9461036920547485, "learning_rate": 9.20081371694275e-06, "loss": 4.1932, "step": 55000 }, { "epoch": 8.064516129032258, "grad_norm": 0.9398289322853088, "learning_rate": 9.193548387096775e-06, "loss": 4.182, "step": 55500 }, { "epoch": 8.137169427492008, "grad_norm": 0.9409332275390625, "learning_rate": 9.1862830572508e-06, "loss": 4.1696, "step": 56000 }, { "epoch": 8.209822725951758, "grad_norm": 0.9343535304069519, "learning_rate": 9.179017727404825e-06, "loss": 4.1627, "step": 56500 }, { "epoch": 8.282476024411508, "grad_norm": 0.96622234582901, "learning_rate": 9.17175239755885e-06, "loss": 4.1569, "step": 57000 }, { "epoch": 8.355129322871258, "grad_norm": 0.9742193818092346, "learning_rate": 9.164487067712876e-06, "loss": 4.1494, "step": 57500 }, { "epoch": 8.427782621331009, "grad_norm": 0.9126195907592773, "learning_rate": 9.1572217378669e-06, "loss": 4.1413, "step": 58000 }, { "epoch": 8.500435919790759, "grad_norm": 0.978380024433136, "learning_rate": 9.149956408020924e-06, "loss": 4.1355, "step": 58500 }, { "epoch": 8.573089218250509, "grad_norm": 0.943013608455658, "learning_rate": 9.14269107817495e-06, "loss": 4.1277, "step": 59000 }, { "epoch": 8.645742516710259, "grad_norm": 0.9417549967765808, "learning_rate": 9.135425748328975e-06, "loss": 4.1254, "step": 59500 }, { "epoch": 8.718395815170009, "grad_norm": 0.9427275061607361, "learning_rate": 9.128160418483e-06, "loss": 4.1183, "step": 60000 }, { "epoch": 8.791049113629759, "grad_norm": 0.972618818283081, "learning_rate": 9.120895088637025e-06, "loss": 4.1094, "step": 60500 }, { "epoch": 8.86370241208951, "grad_norm": 0.9143289923667908, "learning_rate": 9.11362975879105e-06, "loss": 4.1035, "step": 61000 }, { "epoch": 8.93635571054926, "grad_norm": 0.961862325668335, "learning_rate": 9.106364428945075e-06, "loss": 4.1035, "step": 61500 }, { "epoch": 9.00900900900901, "grad_norm": 0.9252375960350037, "learning_rate": 9.0990990990991e-06, "loss": 4.0923, "step": 62000 }, { "epoch": 9.08166230746876, "grad_norm": 0.9321468472480774, "learning_rate": 9.091833769253124e-06, "loss": 4.0855, "step": 62500 }, { "epoch": 9.15431560592851, "grad_norm": 0.9290264844894409, "learning_rate": 9.084568439407149e-06, "loss": 4.0847, "step": 63000 }, { "epoch": 9.22696890438826, "grad_norm": 1.0008461475372314, "learning_rate": 9.077303109561176e-06, "loss": 4.0791, "step": 63500 }, { "epoch": 9.29962220284801, "grad_norm": 1.0144729614257812, "learning_rate": 9.0700377797152e-06, "loss": 4.0724, "step": 64000 }, { "epoch": 9.321127579192096, "step": 64148, "total_flos": 5.363892569191219e+17, "train_loss": 4.763968430426513, "train_runtime": 93599.1484, "train_samples_per_second": 705.822, "train_steps_per_second": 7.353 } ], "logging_steps": 500, "max_steps": 688200, "num_input_tokens_seen": 0, "num_train_epochs": 100, "save_steps": 1000000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": false, "should_training_stop": false }, "attributes": {} } }, "total_flos": 5.363892569191219e+17, "train_batch_size": 96, "trial_name": null, "trial_params": null }