{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 740, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.006756756756756757, "grad_norm": 3.1352566768118555, "learning_rate": 1.0810810810810812e-06, "loss": 1.0885, "step": 1 }, { "epoch": 0.013513513513513514, "grad_norm": 3.0707759310558114, "learning_rate": 2.1621621621621623e-06, "loss": 1.081, "step": 2 }, { "epoch": 0.02027027027027027, "grad_norm": 3.1041877911160487, "learning_rate": 3.2432432432432437e-06, "loss": 1.0779, "step": 3 }, { "epoch": 0.02702702702702703, "grad_norm": 2.894341433491889, "learning_rate": 4.324324324324325e-06, "loss": 1.066, "step": 4 }, { "epoch": 0.033783783783783786, "grad_norm": 2.4474574648074654, "learning_rate": 5.405405405405406e-06, "loss": 1.0464, "step": 5 }, { "epoch": 0.04054054054054054, "grad_norm": 1.7416847938696791, "learning_rate": 6.486486486486487e-06, "loss": 1.0062, "step": 6 }, { "epoch": 0.0472972972972973, "grad_norm": 1.737012288070095, "learning_rate": 7.567567567567569e-06, "loss": 0.9661, "step": 7 }, { "epoch": 0.05405405405405406, "grad_norm": 1.54166088046008, "learning_rate": 8.64864864864865e-06, "loss": 0.9625, "step": 8 }, { "epoch": 0.060810810810810814, "grad_norm": 1.1305140237848132, "learning_rate": 9.729729729729732e-06, "loss": 0.9355, "step": 9 }, { "epoch": 0.06756756756756757, "grad_norm": 1.6829884643585495, "learning_rate": 1.0810810810810812e-05, "loss": 0.8814, "step": 10 }, { "epoch": 0.07432432432432433, "grad_norm": 1.2528889517867892, "learning_rate": 1.1891891891891894e-05, "loss": 0.8733, "step": 11 }, { "epoch": 0.08108108108108109, "grad_norm": 0.9520381902650208, "learning_rate": 1.2972972972972975e-05, "loss": 0.8579, "step": 12 }, { "epoch": 0.08783783783783784, "grad_norm": 1.4164792552755203, "learning_rate": 1.4054054054054055e-05, "loss": 0.82, "step": 13 }, { "epoch": 0.0945945945945946, "grad_norm": 1.1661768167320599, "learning_rate": 1.5135135135135138e-05, "loss": 0.823, "step": 14 }, { "epoch": 0.10135135135135136, "grad_norm": 0.966545525495581, "learning_rate": 1.6216216216216218e-05, "loss": 0.8177, "step": 15 }, { "epoch": 0.10810810810810811, "grad_norm": 0.9970915856247566, "learning_rate": 1.72972972972973e-05, "loss": 0.7828, "step": 16 }, { "epoch": 0.11486486486486487, "grad_norm": 0.7510807795940354, "learning_rate": 1.8378378378378383e-05, "loss": 0.7752, "step": 17 }, { "epoch": 0.12162162162162163, "grad_norm": 0.6801364883047033, "learning_rate": 1.9459459459459463e-05, "loss": 0.7802, "step": 18 }, { "epoch": 0.12837837837837837, "grad_norm": 0.8052966224978529, "learning_rate": 2.054054054054054e-05, "loss": 0.7803, "step": 19 }, { "epoch": 0.13513513513513514, "grad_norm": 0.6731724951990521, "learning_rate": 2.1621621621621624e-05, "loss": 0.774, "step": 20 }, { "epoch": 0.14189189189189189, "grad_norm": 0.5240064900372621, "learning_rate": 2.2702702702702705e-05, "loss": 0.7503, "step": 21 }, { "epoch": 0.14864864864864866, "grad_norm": 0.5238422824665931, "learning_rate": 2.378378378378379e-05, "loss": 0.7452, "step": 22 }, { "epoch": 0.1554054054054054, "grad_norm": 0.5685661395814289, "learning_rate": 2.4864864864864866e-05, "loss": 0.7357, "step": 23 }, { "epoch": 0.16216216216216217, "grad_norm": 0.5559414871229672, "learning_rate": 2.594594594594595e-05, "loss": 0.7529, "step": 24 }, { "epoch": 0.16891891891891891, "grad_norm": 0.44354381932505604, "learning_rate": 2.702702702702703e-05, "loss": 0.7531, "step": 25 }, { "epoch": 0.17567567567567569, "grad_norm": 0.5353496542990213, "learning_rate": 2.810810810810811e-05, "loss": 0.7443, "step": 26 }, { "epoch": 0.18243243243243243, "grad_norm": 0.5011572381597165, "learning_rate": 2.918918918918919e-05, "loss": 0.7309, "step": 27 }, { "epoch": 0.1891891891891892, "grad_norm": 0.46554848262029747, "learning_rate": 3.0270270270270275e-05, "loss": 0.73, "step": 28 }, { "epoch": 0.19594594594594594, "grad_norm": 0.5645942632228904, "learning_rate": 3.135135135135135e-05, "loss": 0.7218, "step": 29 }, { "epoch": 0.20270270270270271, "grad_norm": 0.48122141248903777, "learning_rate": 3.2432432432432436e-05, "loss": 0.7272, "step": 30 }, { "epoch": 0.20945945945945946, "grad_norm": 0.4302278552572059, "learning_rate": 3.351351351351351e-05, "loss": 0.7138, "step": 31 }, { "epoch": 0.21621621621621623, "grad_norm": 0.38359657484495446, "learning_rate": 3.45945945945946e-05, "loss": 0.716, "step": 32 }, { "epoch": 0.22297297297297297, "grad_norm": 0.37734924853300067, "learning_rate": 3.567567567567568e-05, "loss": 0.7258, "step": 33 }, { "epoch": 0.22972972972972974, "grad_norm": 0.5860873173612458, "learning_rate": 3.6756756756756765e-05, "loss": 0.7146, "step": 34 }, { "epoch": 0.23648648648648649, "grad_norm": 0.7521759711776034, "learning_rate": 3.783783783783784e-05, "loss": 0.727, "step": 35 }, { "epoch": 0.24324324324324326, "grad_norm": 0.46287517636393716, "learning_rate": 3.8918918918918926e-05, "loss": 0.6978, "step": 36 }, { "epoch": 0.25, "grad_norm": 0.3949073220604405, "learning_rate": 4e-05, "loss": 0.7143, "step": 37 }, { "epoch": 0.25675675675675674, "grad_norm": 0.5390567372094507, "learning_rate": 4.108108108108108e-05, "loss": 0.7147, "step": 38 }, { "epoch": 0.2635135135135135, "grad_norm": 0.5260020316144047, "learning_rate": 4.2162162162162164e-05, "loss": 0.7042, "step": 39 }, { "epoch": 0.2702702702702703, "grad_norm": 0.6245421084330798, "learning_rate": 4.324324324324325e-05, "loss": 0.7066, "step": 40 }, { "epoch": 0.27702702702702703, "grad_norm": 0.4274410936399283, "learning_rate": 4.4324324324324325e-05, "loss": 0.7002, "step": 41 }, { "epoch": 0.28378378378378377, "grad_norm": 0.5645116869141522, "learning_rate": 4.540540540540541e-05, "loss": 0.7017, "step": 42 }, { "epoch": 0.2905405405405405, "grad_norm": 0.8531985478315304, "learning_rate": 4.6486486486486486e-05, "loss": 0.7105, "step": 43 }, { "epoch": 0.2972972972972973, "grad_norm": 0.8463718591267281, "learning_rate": 4.756756756756758e-05, "loss": 0.6926, "step": 44 }, { "epoch": 0.30405405405405406, "grad_norm": 0.527363152011757, "learning_rate": 4.8648648648648654e-05, "loss": 0.698, "step": 45 }, { "epoch": 0.3108108108108108, "grad_norm": 0.8921529578247511, "learning_rate": 4.972972972972973e-05, "loss": 0.7113, "step": 46 }, { "epoch": 0.31756756756756754, "grad_norm": 0.844585341198514, "learning_rate": 5.081081081081081e-05, "loss": 0.6918, "step": 47 }, { "epoch": 0.32432432432432434, "grad_norm": 0.6610922858373763, "learning_rate": 5.18918918918919e-05, "loss": 0.6874, "step": 48 }, { "epoch": 0.3310810810810811, "grad_norm": 0.9289696025390548, "learning_rate": 5.2972972972972976e-05, "loss": 0.6965, "step": 49 }, { "epoch": 0.33783783783783783, "grad_norm": 0.7104021291109865, "learning_rate": 5.405405405405406e-05, "loss": 0.7025, "step": 50 }, { "epoch": 0.34459459459459457, "grad_norm": 0.7179610269369078, "learning_rate": 5.513513513513514e-05, "loss": 0.6906, "step": 51 }, { "epoch": 0.35135135135135137, "grad_norm": 0.6612006329248038, "learning_rate": 5.621621621621622e-05, "loss": 0.7026, "step": 52 }, { "epoch": 0.3581081081081081, "grad_norm": 0.6920205226216654, "learning_rate": 5.7297297297297305e-05, "loss": 0.6849, "step": 53 }, { "epoch": 0.36486486486486486, "grad_norm": 0.5619919459693192, "learning_rate": 5.837837837837838e-05, "loss": 0.6827, "step": 54 }, { "epoch": 0.3716216216216216, "grad_norm": 0.789835519140916, "learning_rate": 5.945945945945946e-05, "loss": 0.6753, "step": 55 }, { "epoch": 0.3783783783783784, "grad_norm": 0.8468946375056802, "learning_rate": 6.054054054054055e-05, "loss": 0.6803, "step": 56 }, { "epoch": 0.38513513513513514, "grad_norm": 1.1480634289873959, "learning_rate": 6.162162162162163e-05, "loss": 0.6849, "step": 57 }, { "epoch": 0.3918918918918919, "grad_norm": 0.8338876354018716, "learning_rate": 6.27027027027027e-05, "loss": 0.6822, "step": 58 }, { "epoch": 0.39864864864864863, "grad_norm": 0.8619635833656949, "learning_rate": 6.378378378378379e-05, "loss": 0.6885, "step": 59 }, { "epoch": 0.40540540540540543, "grad_norm": 0.9841339840879071, "learning_rate": 6.486486486486487e-05, "loss": 0.6941, "step": 60 }, { "epoch": 0.41216216216216217, "grad_norm": 0.7846924650076373, "learning_rate": 6.594594594594596e-05, "loss": 0.6734, "step": 61 }, { "epoch": 0.4189189189189189, "grad_norm": 0.6084845178399872, "learning_rate": 6.702702702702703e-05, "loss": 0.6766, "step": 62 }, { "epoch": 0.42567567567567566, "grad_norm": 0.8121605538110969, "learning_rate": 6.810810810810811e-05, "loss": 0.6882, "step": 63 }, { "epoch": 0.43243243243243246, "grad_norm": 0.9962451486752875, "learning_rate": 6.91891891891892e-05, "loss": 0.6837, "step": 64 }, { "epoch": 0.4391891891891892, "grad_norm": 0.6516040658377318, "learning_rate": 7.027027027027028e-05, "loss": 0.6841, "step": 65 }, { "epoch": 0.44594594594594594, "grad_norm": 0.683694787120212, "learning_rate": 7.135135135135136e-05, "loss": 0.6952, "step": 66 }, { "epoch": 0.4527027027027027, "grad_norm": 0.6582713522549495, "learning_rate": 7.243243243243243e-05, "loss": 0.6727, "step": 67 }, { "epoch": 0.4594594594594595, "grad_norm": 0.673908455931949, "learning_rate": 7.351351351351353e-05, "loss": 0.6841, "step": 68 }, { "epoch": 0.46621621621621623, "grad_norm": 0.8017390116132226, "learning_rate": 7.45945945945946e-05, "loss": 0.67, "step": 69 }, { "epoch": 0.47297297297297297, "grad_norm": 0.9799335757769192, "learning_rate": 7.567567567567568e-05, "loss": 0.6914, "step": 70 }, { "epoch": 0.4797297297297297, "grad_norm": 1.5088822309237815, "learning_rate": 7.675675675675675e-05, "loss": 0.6925, "step": 71 }, { "epoch": 0.4864864864864865, "grad_norm": 0.7235799329993131, "learning_rate": 7.783783783783785e-05, "loss": 0.689, "step": 72 }, { "epoch": 0.49324324324324326, "grad_norm": 1.467550439828714, "learning_rate": 7.891891891891892e-05, "loss": 0.6927, "step": 73 }, { "epoch": 0.5, "grad_norm": 0.9680632895596434, "learning_rate": 8e-05, "loss": 0.6729, "step": 74 }, { "epoch": 0.5067567567567568, "grad_norm": 1.4048772257060618, "learning_rate": 7.999955497902857e-05, "loss": 0.6949, "step": 75 }, { "epoch": 0.5135135135135135, "grad_norm": 1.187179427424338, "learning_rate": 7.999821992601645e-05, "loss": 0.693, "step": 76 }, { "epoch": 0.5202702702702703, "grad_norm": 0.9529005283405405, "learning_rate": 7.999599487066996e-05, "loss": 0.6839, "step": 77 }, { "epoch": 0.527027027027027, "grad_norm": 0.9173343088851188, "learning_rate": 7.999287986249894e-05, "loss": 0.6766, "step": 78 }, { "epoch": 0.5337837837837838, "grad_norm": 0.7318824071301274, "learning_rate": 7.998887497081555e-05, "loss": 0.6847, "step": 79 }, { "epoch": 0.5405405405405406, "grad_norm": 0.6529765700425837, "learning_rate": 7.998398028473287e-05, "loss": 0.678, "step": 80 }, { "epoch": 0.5472972972972973, "grad_norm": 0.6928101408737816, "learning_rate": 7.997819591316278e-05, "loss": 0.6954, "step": 81 }, { "epoch": 0.5540540540540541, "grad_norm": 0.5879827240121825, "learning_rate": 7.99715219848136e-05, "loss": 0.6642, "step": 82 }, { "epoch": 0.5608108108108109, "grad_norm": 0.680605729950147, "learning_rate": 7.996395864818727e-05, "loss": 0.6804, "step": 83 }, { "epoch": 0.5675675675675675, "grad_norm": 0.5423663061022331, "learning_rate": 7.995550607157592e-05, "loss": 0.6688, "step": 84 }, { "epoch": 0.5743243243243243, "grad_norm": 0.606163015240631, "learning_rate": 7.994616444305826e-05, "loss": 0.6667, "step": 85 }, { "epoch": 0.581081081081081, "grad_norm": 0.5950010094955055, "learning_rate": 7.993593397049533e-05, "loss": 0.679, "step": 86 }, { "epoch": 0.5878378378378378, "grad_norm": 0.6429149705782599, "learning_rate": 7.992481488152585e-05, "loss": 0.6685, "step": 87 }, { "epoch": 0.5945945945945946, "grad_norm": 0.46140817883542623, "learning_rate": 7.991280742356124e-05, "loss": 0.6603, "step": 88 }, { "epoch": 0.6013513513513513, "grad_norm": 0.6795669467936494, "learning_rate": 7.989991186378e-05, "loss": 0.6699, "step": 89 }, { "epoch": 0.6081081081081081, "grad_norm": 0.5738680266825441, "learning_rate": 7.988612848912186e-05, "loss": 0.6819, "step": 90 }, { "epoch": 0.6148648648648649, "grad_norm": 0.5789217573047567, "learning_rate": 7.987145760628138e-05, "loss": 0.6708, "step": 91 }, { "epoch": 0.6216216216216216, "grad_norm": 0.49449419241667797, "learning_rate": 7.985589954170107e-05, "loss": 0.6665, "step": 92 }, { "epoch": 0.6283783783783784, "grad_norm": 0.4987045947630836, "learning_rate": 7.983945464156419e-05, "loss": 0.6597, "step": 93 }, { "epoch": 0.6351351351351351, "grad_norm": 0.40082592041703324, "learning_rate": 7.982212327178699e-05, "loss": 0.6675, "step": 94 }, { "epoch": 0.6418918918918919, "grad_norm": 0.3431463481882859, "learning_rate": 7.980390581801064e-05, "loss": 0.6616, "step": 95 }, { "epoch": 0.6486486486486487, "grad_norm": 0.44148520818681297, "learning_rate": 7.97848026855926e-05, "loss": 0.6761, "step": 96 }, { "epoch": 0.6554054054054054, "grad_norm": 0.4038247824809848, "learning_rate": 7.976481429959758e-05, "loss": 0.6596, "step": 97 }, { "epoch": 0.6621621621621622, "grad_norm": 0.3448527603506924, "learning_rate": 7.974394110478813e-05, "loss": 0.6536, "step": 98 }, { "epoch": 0.668918918918919, "grad_norm": 0.30698753802117784, "learning_rate": 7.972218356561471e-05, "loss": 0.6772, "step": 99 }, { "epoch": 0.6756756756756757, "grad_norm": 0.30195068085071514, "learning_rate": 7.96995421662054e-05, "loss": 0.6527, "step": 100 }, { "epoch": 0.6824324324324325, "grad_norm": 0.3166797371131855, "learning_rate": 7.967601741035507e-05, "loss": 0.6555, "step": 101 }, { "epoch": 0.6891891891891891, "grad_norm": 0.32820963864069536, "learning_rate": 7.965160982151422e-05, "loss": 0.6696, "step": 102 }, { "epoch": 0.6959459459459459, "grad_norm": 0.3483528954236439, "learning_rate": 7.962631994277728e-05, "loss": 0.663, "step": 103 }, { "epoch": 0.7027027027027027, "grad_norm": 0.4983197508582807, "learning_rate": 7.960014833687055e-05, "loss": 0.6633, "step": 104 }, { "epoch": 0.7094594594594594, "grad_norm": 0.7805957154226273, "learning_rate": 7.957309558613974e-05, "loss": 0.6587, "step": 105 }, { "epoch": 0.7162162162162162, "grad_norm": 1.106633618810969, "learning_rate": 7.954516229253691e-05, "loss": 0.662, "step": 106 }, { "epoch": 0.722972972972973, "grad_norm": 0.757015003567712, "learning_rate": 7.951634907760713e-05, "loss": 0.6593, "step": 107 }, { "epoch": 0.7297297297297297, "grad_norm": 0.8427739970040843, "learning_rate": 7.948665658247463e-05, "loss": 0.6601, "step": 108 }, { "epoch": 0.7364864864864865, "grad_norm": 0.7431541528418906, "learning_rate": 7.945608546782858e-05, "loss": 0.667, "step": 109 }, { "epoch": 0.7432432432432432, "grad_norm": 0.6845263081020915, "learning_rate": 7.942463641390834e-05, "loss": 0.6569, "step": 110 }, { "epoch": 0.75, "grad_norm": 0.8712514657793697, "learning_rate": 7.939231012048833e-05, "loss": 0.6632, "step": 111 }, { "epoch": 0.7567567567567568, "grad_norm": 0.6774356142156878, "learning_rate": 7.935910730686246e-05, "loss": 0.6706, "step": 112 }, { "epoch": 0.7635135135135135, "grad_norm": 0.7702128806997793, "learning_rate": 7.932502871182818e-05, "loss": 0.6679, "step": 113 }, { "epoch": 0.7702702702702703, "grad_norm": 0.6032796273856512, "learning_rate": 7.929007509366994e-05, "loss": 0.6612, "step": 114 }, { "epoch": 0.777027027027027, "grad_norm": 0.7262342972595842, "learning_rate": 7.925424723014239e-05, "loss": 0.6639, "step": 115 }, { "epoch": 0.7837837837837838, "grad_norm": 0.6319433116490037, "learning_rate": 7.921754591845307e-05, "loss": 0.662, "step": 116 }, { "epoch": 0.7905405405405406, "grad_norm": 0.6175612756087657, "learning_rate": 7.917997197524467e-05, "loss": 0.6566, "step": 117 }, { "epoch": 0.7972972972972973, "grad_norm": 0.7923618524838222, "learning_rate": 7.914152623657678e-05, "loss": 0.6469, "step": 118 }, { "epoch": 0.8040540540540541, "grad_norm": 0.7707195088131242, "learning_rate": 7.910220955790746e-05, "loss": 0.6563, "step": 119 }, { "epoch": 0.8108108108108109, "grad_norm": 0.4152288654221399, "learning_rate": 7.906202281407398e-05, "loss": 0.6519, "step": 120 }, { "epoch": 0.8175675675675675, "grad_norm": 0.5439040195272984, "learning_rate": 7.902096689927355e-05, "loss": 0.6727, "step": 121 }, { "epoch": 0.8243243243243243, "grad_norm": 0.5159119091886112, "learning_rate": 7.897904272704333e-05, "loss": 0.6563, "step": 122 }, { "epoch": 0.831081081081081, "grad_norm": 0.49623067194442205, "learning_rate": 7.893625123024011e-05, "loss": 0.6523, "step": 123 }, { "epoch": 0.8378378378378378, "grad_norm": 0.4176229579696035, "learning_rate": 7.889259336101957e-05, "loss": 0.6461, "step": 124 }, { "epoch": 0.8445945945945946, "grad_norm": 0.5424146541267013, "learning_rate": 7.884807009081506e-05, "loss": 0.6535, "step": 125 }, { "epoch": 0.8513513513513513, "grad_norm": 0.53245673115827, "learning_rate": 7.880268241031604e-05, "loss": 0.6516, "step": 126 }, { "epoch": 0.8581081081081081, "grad_norm": 0.4218647663890849, "learning_rate": 7.875643132944599e-05, "loss": 0.6557, "step": 127 }, { "epoch": 0.8648648648648649, "grad_norm": 0.31388453196671723, "learning_rate": 7.870931787733996e-05, "loss": 0.6504, "step": 128 }, { "epoch": 0.8716216216216216, "grad_norm": 0.42120451899053063, "learning_rate": 7.866134310232167e-05, "loss": 0.6637, "step": 129 }, { "epoch": 0.8783783783783784, "grad_norm": 0.3569049182508329, "learning_rate": 7.861250807188014e-05, "loss": 0.6565, "step": 130 }, { "epoch": 0.8851351351351351, "grad_norm": 0.46980061573845006, "learning_rate": 7.856281387264603e-05, "loss": 0.6643, "step": 131 }, { "epoch": 0.8918918918918919, "grad_norm": 0.5602330500767254, "learning_rate": 7.851226161036739e-05, "loss": 0.6541, "step": 132 }, { "epoch": 0.8986486486486487, "grad_norm": 0.555307896703473, "learning_rate": 7.846085240988503e-05, "loss": 0.6498, "step": 133 }, { "epoch": 0.9054054054054054, "grad_norm": 0.4906170522953401, "learning_rate": 7.840858741510758e-05, "loss": 0.6482, "step": 134 }, { "epoch": 0.9121621621621622, "grad_norm": 0.4486181714024936, "learning_rate": 7.835546778898599e-05, "loss": 0.6457, "step": 135 }, { "epoch": 0.918918918918919, "grad_norm": 0.5704154766929971, "learning_rate": 7.830149471348763e-05, "loss": 0.6508, "step": 136 }, { "epoch": 0.9256756756756757, "grad_norm": 0.738054877738817, "learning_rate": 7.824666938957004e-05, "loss": 0.6531, "step": 137 }, { "epoch": 0.9324324324324325, "grad_norm": 0.7131460924145132, "learning_rate": 7.819099303715414e-05, "loss": 0.6577, "step": 138 }, { "epoch": 0.9391891891891891, "grad_norm": 0.5314442328007468, "learning_rate": 7.813446689509718e-05, "loss": 0.6472, "step": 139 }, { "epoch": 0.9459459459459459, "grad_norm": 0.4354387291682998, "learning_rate": 7.807709222116506e-05, "loss": 0.6566, "step": 140 }, { "epoch": 0.9527027027027027, "grad_norm": 0.4367558414891408, "learning_rate": 7.801887029200448e-05, "loss": 0.6548, "step": 141 }, { "epoch": 0.9594594594594594, "grad_norm": 0.49114789228090255, "learning_rate": 7.795980240311436e-05, "loss": 0.6651, "step": 142 }, { "epoch": 0.9662162162162162, "grad_norm": 0.5563118895714915, "learning_rate": 7.789988986881719e-05, "loss": 0.6534, "step": 143 }, { "epoch": 0.972972972972973, "grad_norm": 0.5068498475757595, "learning_rate": 7.78391340222297e-05, "loss": 0.6635, "step": 144 }, { "epoch": 0.9797297297297297, "grad_norm": 0.40377422580342515, "learning_rate": 7.777753621523316e-05, "loss": 0.662, "step": 145 }, { "epoch": 0.9864864864864865, "grad_norm": 0.48943576691657636, "learning_rate": 7.771509781844338e-05, "loss": 0.6465, "step": 146 }, { "epoch": 0.9932432432432432, "grad_norm": 0.45222666736741574, "learning_rate": 7.765182022118014e-05, "loss": 0.6576, "step": 147 }, { "epoch": 1.0, "grad_norm": 0.3524115634430567, "learning_rate": 7.758770483143634e-05, "loss": 0.66, "step": 148 }, { "epoch": 1.0067567567567568, "grad_norm": 0.3465470718278503, "learning_rate": 7.752275307584664e-05, "loss": 0.6528, "step": 149 }, { "epoch": 1.0135135135135136, "grad_norm": 0.4289730330622286, "learning_rate": 7.745696639965569e-05, "loss": 0.6284, "step": 150 }, { "epoch": 1.0202702702702702, "grad_norm": 0.4078417015334056, "learning_rate": 7.739034626668605e-05, "loss": 0.6409, "step": 151 }, { "epoch": 1.027027027027027, "grad_norm": 0.4224099719890233, "learning_rate": 7.732289415930549e-05, "loss": 0.63, "step": 152 }, { "epoch": 1.0337837837837838, "grad_norm": 0.43135712270900994, "learning_rate": 7.725461157839417e-05, "loss": 0.6606, "step": 153 }, { "epoch": 1.0405405405405406, "grad_norm": 0.4832560652031643, "learning_rate": 7.71855000433111e-05, "loss": 0.6439, "step": 154 }, { "epoch": 1.0472972972972974, "grad_norm": 0.5148651119907238, "learning_rate": 7.711556109186039e-05, "loss": 0.6405, "step": 155 }, { "epoch": 1.054054054054054, "grad_norm": 0.49381906164540984, "learning_rate": 7.704479628025704e-05, "loss": 0.635, "step": 156 }, { "epoch": 1.0608108108108107, "grad_norm": 0.4238088291952583, "learning_rate": 7.697320718309235e-05, "loss": 0.6363, "step": 157 }, { "epoch": 1.0675675675675675, "grad_norm": 0.34639148621347143, "learning_rate": 7.690079539329875e-05, "loss": 0.6434, "step": 158 }, { "epoch": 1.0743243243243243, "grad_norm": 0.3670601156648982, "learning_rate": 7.682756252211453e-05, "loss": 0.6321, "step": 159 }, { "epoch": 1.0810810810810811, "grad_norm": 0.48931637412654205, "learning_rate": 7.675351019904785e-05, "loss": 0.6371, "step": 160 }, { "epoch": 1.087837837837838, "grad_norm": 0.5514281223158661, "learning_rate": 7.667864007184054e-05, "loss": 0.6313, "step": 161 }, { "epoch": 1.0945945945945945, "grad_norm": 0.48466826872522856, "learning_rate": 7.660295380643144e-05, "loss": 0.6358, "step": 162 }, { "epoch": 1.1013513513513513, "grad_norm": 0.46142889218279065, "learning_rate": 7.652645308691933e-05, "loss": 0.6418, "step": 163 }, { "epoch": 1.1081081081081081, "grad_norm": 0.4371908363892829, "learning_rate": 7.644913961552544e-05, "loss": 0.6327, "step": 164 }, { "epoch": 1.114864864864865, "grad_norm": 0.4159116225859721, "learning_rate": 7.637101511255554e-05, "loss": 0.6332, "step": 165 }, { "epoch": 1.1216216216216217, "grad_norm": 0.48524137415448176, "learning_rate": 7.629208131636179e-05, "loss": 0.6385, "step": 166 }, { "epoch": 1.1283783783783783, "grad_norm": 0.502011807585537, "learning_rate": 7.621233998330387e-05, "loss": 0.6534, "step": 167 }, { "epoch": 1.135135135135135, "grad_norm": 0.4849012769427631, "learning_rate": 7.61317928877101e-05, "loss": 0.6382, "step": 168 }, { "epoch": 1.1418918918918919, "grad_norm": 0.40497697606883803, "learning_rate": 7.605044182183779e-05, "loss": 0.6335, "step": 169 }, { "epoch": 1.1486486486486487, "grad_norm": 0.26814241993209303, "learning_rate": 7.596828859583347e-05, "loss": 0.6226, "step": 170 }, { "epoch": 1.1554054054054055, "grad_norm": 0.36135092127292284, "learning_rate": 7.588533503769257e-05, "loss": 0.6286, "step": 171 }, { "epoch": 1.1621621621621623, "grad_norm": 0.4608397427011372, "learning_rate": 7.580158299321872e-05, "loss": 0.6438, "step": 172 }, { "epoch": 1.1689189189189189, "grad_norm": 0.4171913645594413, "learning_rate": 7.571703432598275e-05, "loss": 0.6356, "step": 173 }, { "epoch": 1.1756756756756757, "grad_norm": 0.4370902558666279, "learning_rate": 7.563169091728115e-05, "loss": 0.6422, "step": 174 }, { "epoch": 1.1824324324324325, "grad_norm": 0.4283980244016275, "learning_rate": 7.554555466609425e-05, "loss": 0.6407, "step": 175 }, { "epoch": 1.1891891891891893, "grad_norm": 0.4381556639196284, "learning_rate": 7.545862748904394e-05, "loss": 0.6434, "step": 176 }, { "epoch": 1.195945945945946, "grad_norm": 0.44668924620805633, "learning_rate": 7.537091132035111e-05, "loss": 0.6219, "step": 177 }, { "epoch": 1.2027027027027026, "grad_norm": 0.37602920076180757, "learning_rate": 7.528240811179245e-05, "loss": 0.6419, "step": 178 }, { "epoch": 1.2094594594594594, "grad_norm": 0.2672376131285308, "learning_rate": 7.519311983265718e-05, "loss": 0.6366, "step": 179 }, { "epoch": 1.2162162162162162, "grad_norm": 0.29369317421661273, "learning_rate": 7.510304846970311e-05, "loss": 0.6341, "step": 180 }, { "epoch": 1.222972972972973, "grad_norm": 0.3602685800886801, "learning_rate": 7.501219602711253e-05, "loss": 0.6432, "step": 181 }, { "epoch": 1.2297297297297298, "grad_norm": 0.3237536586353385, "learning_rate": 7.492056452644753e-05, "loss": 0.6415, "step": 182 }, { "epoch": 1.2364864864864864, "grad_norm": 0.28545557770169117, "learning_rate": 7.48281560066051e-05, "loss": 0.6364, "step": 183 }, { "epoch": 1.2432432432432432, "grad_norm": 0.2623342902523158, "learning_rate": 7.473497252377171e-05, "loss": 0.6333, "step": 184 }, { "epoch": 1.25, "grad_norm": 0.3160766346456171, "learning_rate": 7.464101615137756e-05, "loss": 0.6256, "step": 185 }, { "epoch": 1.2567567567567568, "grad_norm": 0.4173934303596852, "learning_rate": 7.454628898005043e-05, "loss": 0.6521, "step": 186 }, { "epoch": 1.2635135135135136, "grad_norm": 0.4424346331081593, "learning_rate": 7.445079311756924e-05, "loss": 0.6541, "step": 187 }, { "epoch": 1.2702702702702702, "grad_norm": 0.38764273759912543, "learning_rate": 7.435453068881706e-05, "loss": 0.6385, "step": 188 }, { "epoch": 1.277027027027027, "grad_norm": 0.3811463582022792, "learning_rate": 7.425750383573384e-05, "loss": 0.6295, "step": 189 }, { "epoch": 1.2837837837837838, "grad_norm": 0.5272316642769805, "learning_rate": 7.415971471726884e-05, "loss": 0.6307, "step": 190 }, { "epoch": 1.2905405405405406, "grad_norm": 0.7635509141776905, "learning_rate": 7.406116550933246e-05, "loss": 0.6313, "step": 191 }, { "epoch": 1.2972972972972974, "grad_norm": 0.9673085818387788, "learning_rate": 7.396185840474792e-05, "loss": 0.6495, "step": 192 }, { "epoch": 1.304054054054054, "grad_norm": 0.8691278543652353, "learning_rate": 7.386179561320243e-05, "loss": 0.6285, "step": 193 }, { "epoch": 1.3108108108108107, "grad_norm": 0.7233647856746338, "learning_rate": 7.376097936119803e-05, "loss": 0.6435, "step": 194 }, { "epoch": 1.3175675675675675, "grad_norm": 0.4888365541378487, "learning_rate": 7.365941189200201e-05, "loss": 0.6345, "step": 195 }, { "epoch": 1.3243243243243243, "grad_norm": 0.41440516733521193, "learning_rate": 7.35570954655971e-05, "loss": 0.6522, "step": 196 }, { "epoch": 1.3310810810810811, "grad_norm": 0.39733170188465466, "learning_rate": 7.345403235863105e-05, "loss": 0.636, "step": 197 }, { "epoch": 1.3378378378378377, "grad_norm": 0.37538788641282456, "learning_rate": 7.335022486436608e-05, "loss": 0.6405, "step": 198 }, { "epoch": 1.3445945945945945, "grad_norm": 0.4912257786205075, "learning_rate": 7.324567529262775e-05, "loss": 0.6341, "step": 199 }, { "epoch": 1.3513513513513513, "grad_norm": 0.3964231351313485, "learning_rate": 7.31403859697537e-05, "loss": 0.6329, "step": 200 }, { "epoch": 1.3581081081081081, "grad_norm": 0.23258287711574407, "learning_rate": 7.303435923854172e-05, "loss": 0.6346, "step": 201 }, { "epoch": 1.364864864864865, "grad_norm": 0.36875878813699686, "learning_rate": 7.292759745819781e-05, "loss": 0.6423, "step": 202 }, { "epoch": 1.3716216216216215, "grad_norm": 0.3816088768921796, "learning_rate": 7.282010300428351e-05, "loss": 0.6247, "step": 203 }, { "epoch": 1.3783783783783785, "grad_norm": 0.32935135005176236, "learning_rate": 7.271187826866312e-05, "loss": 0.6306, "step": 204 }, { "epoch": 1.385135135135135, "grad_norm": 0.24849408204445086, "learning_rate": 7.260292565945049e-05, "loss": 0.6425, "step": 205 }, { "epoch": 1.3918918918918919, "grad_norm": 0.22496521412221002, "learning_rate": 7.249324760095544e-05, "loss": 0.64, "step": 206 }, { "epoch": 1.3986486486486487, "grad_norm": 0.30450316180118336, "learning_rate": 7.238284653362977e-05, "loss": 0.6497, "step": 207 }, { "epoch": 1.4054054054054055, "grad_norm": 0.2922608158162527, "learning_rate": 7.227172491401299e-05, "loss": 0.6339, "step": 208 }, { "epoch": 1.4121621621621623, "grad_norm": 0.23745259799881574, "learning_rate": 7.215988521467763e-05, "loss": 0.629, "step": 209 }, { "epoch": 1.4189189189189189, "grad_norm": 0.29205738332283676, "learning_rate": 7.204732992417431e-05, "loss": 0.6276, "step": 210 }, { "epoch": 1.4256756756756757, "grad_norm": 0.2764709836681376, "learning_rate": 7.193406154697625e-05, "loss": 0.6325, "step": 211 }, { "epoch": 1.4324324324324325, "grad_norm": 0.30413188245977635, "learning_rate": 7.18200826034236e-05, "loss": 0.6347, "step": 212 }, { "epoch": 1.4391891891891893, "grad_norm": 0.32848243797717097, "learning_rate": 7.17053956296674e-05, "loss": 0.6173, "step": 213 }, { "epoch": 1.445945945945946, "grad_norm": 0.3056033484058255, "learning_rate": 7.159000317761305e-05, "loss": 0.6339, "step": 214 }, { "epoch": 1.4527027027027026, "grad_norm": 0.3085891817121419, "learning_rate": 7.14739078148636e-05, "loss": 0.6364, "step": 215 }, { "epoch": 1.4594594594594594, "grad_norm": 0.3326838165486728, "learning_rate": 7.135711212466264e-05, "loss": 0.6422, "step": 216 }, { "epoch": 1.4662162162162162, "grad_norm": 0.3216850978230415, "learning_rate": 7.123961870583671e-05, "loss": 0.6342, "step": 217 }, { "epoch": 1.472972972972973, "grad_norm": 0.35372809908349234, "learning_rate": 7.112143017273759e-05, "loss": 0.637, "step": 218 }, { "epoch": 1.4797297297297298, "grad_norm": 0.3689688937687713, "learning_rate": 7.100254915518408e-05, "loss": 0.6342, "step": 219 }, { "epoch": 1.4864864864864864, "grad_norm": 0.4017045063635094, "learning_rate": 7.088297829840346e-05, "loss": 0.6286, "step": 220 }, { "epoch": 1.4932432432432432, "grad_norm": 0.4446976891407491, "learning_rate": 7.076272026297268e-05, "loss": 0.6342, "step": 221 }, { "epoch": 1.5, "grad_norm": 0.432808144889846, "learning_rate": 7.064177772475912e-05, "loss": 0.6452, "step": 222 }, { "epoch": 1.5067567567567568, "grad_norm": 0.3443564755955631, "learning_rate": 7.052015337486109e-05, "loss": 0.6378, "step": 223 }, { "epoch": 1.5135135135135136, "grad_norm": 0.30789638902879063, "learning_rate": 7.03978499195479e-05, "loss": 0.6284, "step": 224 }, { "epoch": 1.5202702702702702, "grad_norm": 0.3444085722973672, "learning_rate": 7.027487008019969e-05, "loss": 0.6439, "step": 225 }, { "epoch": 1.527027027027027, "grad_norm": 0.3698979548144048, "learning_rate": 7.015121659324678e-05, "loss": 0.6328, "step": 226 }, { "epoch": 1.5337837837837838, "grad_norm": 0.4166719230003736, "learning_rate": 7.002689221010897e-05, "loss": 0.6295, "step": 227 }, { "epoch": 1.5405405405405406, "grad_norm": 0.4418395627095002, "learning_rate": 6.990189969713416e-05, "loss": 0.6303, "step": 228 }, { "epoch": 1.5472972972972974, "grad_norm": 0.4102988370573012, "learning_rate": 6.977624183553676e-05, "loss": 0.6431, "step": 229 }, { "epoch": 1.554054054054054, "grad_norm": 0.39270575425283927, "learning_rate": 6.964992142133602e-05, "loss": 0.6333, "step": 230 }, { "epoch": 1.560810810810811, "grad_norm": 0.3396456921939933, "learning_rate": 6.952294126529356e-05, "loss": 0.6274, "step": 231 }, { "epoch": 1.5675675675675675, "grad_norm": 0.29161162699565907, "learning_rate": 6.939530419285104e-05, "loss": 0.6346, "step": 232 }, { "epoch": 1.5743243243243243, "grad_norm": 0.3270645478162509, "learning_rate": 6.926701304406713e-05, "loss": 0.6307, "step": 233 }, { "epoch": 1.5810810810810811, "grad_norm": 0.33343232181767685, "learning_rate": 6.913807067355445e-05, "loss": 0.6338, "step": 234 }, { "epoch": 1.5878378378378377, "grad_norm": 0.34421561990305105, "learning_rate": 6.90084799504159e-05, "loss": 0.6417, "step": 235 }, { "epoch": 1.5945945945945947, "grad_norm": 0.31139357648881777, "learning_rate": 6.887824375818099e-05, "loss": 0.6399, "step": 236 }, { "epoch": 1.6013513513513513, "grad_norm": 0.2657664210184094, "learning_rate": 6.874736499474154e-05, "loss": 0.6411, "step": 237 }, { "epoch": 1.6081081081081081, "grad_norm": 0.25481549866865094, "learning_rate": 6.861584657228728e-05, "loss": 0.6418, "step": 238 }, { "epoch": 1.614864864864865, "grad_norm": 0.27638932514157166, "learning_rate": 6.848369141724104e-05, "loss": 0.6435, "step": 239 }, { "epoch": 1.6216216216216215, "grad_norm": 0.32918191799417873, "learning_rate": 6.835090247019354e-05, "loss": 0.6269, "step": 240 }, { "epoch": 1.6283783783783785, "grad_norm": 0.334162128338808, "learning_rate": 6.821748268583813e-05, "loss": 0.6382, "step": 241 }, { "epoch": 1.635135135135135, "grad_norm": 0.3172895628393372, "learning_rate": 6.808343503290491e-05, "loss": 0.627, "step": 242 }, { "epoch": 1.6418918918918919, "grad_norm": 0.3353087016871893, "learning_rate": 6.79487624940947e-05, "loss": 0.6228, "step": 243 }, { "epoch": 1.6486486486486487, "grad_norm": 0.3718369962163942, "learning_rate": 6.781346806601273e-05, "loss": 0.6305, "step": 244 }, { "epoch": 1.6554054054054053, "grad_norm": 0.39658980608696004, "learning_rate": 6.767755475910185e-05, "loss": 0.6227, "step": 245 }, { "epoch": 1.6621621621621623, "grad_norm": 0.4269305577093925, "learning_rate": 6.754102559757569e-05, "loss": 0.6338, "step": 246 }, { "epoch": 1.6689189189189189, "grad_norm": 0.48834528749062367, "learning_rate": 6.740388361935125e-05, "loss": 0.6285, "step": 247 }, { "epoch": 1.6756756756756757, "grad_norm": 0.5054797184183533, "learning_rate": 6.726613187598132e-05, "loss": 0.6363, "step": 248 }, { "epoch": 1.6824324324324325, "grad_norm": 0.46558382405228205, "learning_rate": 6.712777343258666e-05, "loss": 0.6342, "step": 249 }, { "epoch": 1.689189189189189, "grad_norm": 0.3349216901360824, "learning_rate": 6.698881136778771e-05, "loss": 0.638, "step": 250 }, { "epoch": 1.695945945945946, "grad_norm": 0.296031271653257, "learning_rate": 6.684924877363613e-05, "loss": 0.6449, "step": 251 }, { "epoch": 1.7027027027027026, "grad_norm": 0.34651680386096584, "learning_rate": 6.670908875554594e-05, "loss": 0.6373, "step": 252 }, { "epoch": 1.7094594594594594, "grad_norm": 0.34431135065378676, "learning_rate": 6.656833443222458e-05, "loss": 0.6222, "step": 253 }, { "epoch": 1.7162162162162162, "grad_norm": 0.3856861975419152, "learning_rate": 6.642698893560327e-05, "loss": 0.6389, "step": 254 }, { "epoch": 1.722972972972973, "grad_norm": 0.3419900662624624, "learning_rate": 6.628505541076755e-05, "loss": 0.6345, "step": 255 }, { "epoch": 1.7297297297297298, "grad_norm": 0.25513443874276925, "learning_rate": 6.614253701588718e-05, "loss": 0.6337, "step": 256 }, { "epoch": 1.7364864864864864, "grad_norm": 0.40205274571625865, "learning_rate": 6.599943692214587e-05, "loss": 0.6329, "step": 257 }, { "epoch": 1.7432432432432432, "grad_norm": 0.38134230469861735, "learning_rate": 6.585575831367078e-05, "loss": 0.6315, "step": 258 }, { "epoch": 1.75, "grad_norm": 0.2108025719904083, "learning_rate": 6.571150438746157e-05, "loss": 0.6332, "step": 259 }, { "epoch": 1.7567567567567568, "grad_norm": 0.2896853669566718, "learning_rate": 6.55666783533194e-05, "loss": 0.6254, "step": 260 }, { "epoch": 1.7635135135135136, "grad_norm": 0.3811309603864002, "learning_rate": 6.542128343377536e-05, "loss": 0.6355, "step": 261 }, { "epoch": 1.7702702702702702, "grad_norm": 0.30773816113154867, "learning_rate": 6.527532286401889e-05, "loss": 0.6337, "step": 262 }, { "epoch": 1.777027027027027, "grad_norm": 0.22231601239252788, "learning_rate": 6.51287998918257e-05, "loss": 0.6204, "step": 263 }, { "epoch": 1.7837837837837838, "grad_norm": 0.2581820667256423, "learning_rate": 6.498171777748557e-05, "loss": 0.6335, "step": 264 }, { "epoch": 1.7905405405405406, "grad_norm": 0.3087437556000368, "learning_rate": 6.483407979372975e-05, "loss": 0.6221, "step": 265 }, { "epoch": 1.7972972972972974, "grad_norm": 0.229624863667569, "learning_rate": 6.468588922565822e-05, "loss": 0.6305, "step": 266 }, { "epoch": 1.804054054054054, "grad_norm": 0.15965170109349433, "learning_rate": 6.453714937066648e-05, "loss": 0.6277, "step": 267 }, { "epoch": 1.810810810810811, "grad_norm": 0.16500327907249202, "learning_rate": 6.438786353837228e-05, "loss": 0.6279, "step": 268 }, { "epoch": 1.8175675675675675, "grad_norm": 0.19228879276238184, "learning_rate": 6.423803505054193e-05, "loss": 0.637, "step": 269 }, { "epoch": 1.8243243243243243, "grad_norm": 0.1960334960484057, "learning_rate": 6.408766724101638e-05, "loss": 0.6344, "step": 270 }, { "epoch": 1.8310810810810811, "grad_norm": 0.22246057335109667, "learning_rate": 6.393676345563708e-05, "loss": 0.6315, "step": 271 }, { "epoch": 1.8378378378378377, "grad_norm": 0.25500558074002594, "learning_rate": 6.378532705217148e-05, "loss": 0.6267, "step": 272 }, { "epoch": 1.8445945945945947, "grad_norm": 0.2797665813089635, "learning_rate": 6.363336140023833e-05, "loss": 0.6198, "step": 273 }, { "epoch": 1.8513513513513513, "grad_norm": 0.32237396509938987, "learning_rate": 6.348086988123274e-05, "loss": 0.6302, "step": 274 }, { "epoch": 1.8581081081081081, "grad_norm": 0.3837879616068762, "learning_rate": 6.332785588825094e-05, "loss": 0.6366, "step": 275 }, { "epoch": 1.864864864864865, "grad_norm": 0.3847808366177522, "learning_rate": 6.317432282601469e-05, "loss": 0.6405, "step": 276 }, { "epoch": 1.8716216216216215, "grad_norm": 0.3485054619073992, "learning_rate": 6.302027411079562e-05, "loss": 0.6273, "step": 277 }, { "epoch": 1.8783783783783785, "grad_norm": 0.3171200624449193, "learning_rate": 6.286571317033915e-05, "loss": 0.6337, "step": 278 }, { "epoch": 1.885135135135135, "grad_norm": 0.27373328598715324, "learning_rate": 6.271064344378832e-05, "loss": 0.6403, "step": 279 }, { "epoch": 1.8918918918918919, "grad_norm": 0.3045549423645127, "learning_rate": 6.255506838160711e-05, "loss": 0.6317, "step": 280 }, { "epoch": 1.8986486486486487, "grad_norm": 0.23258625831405336, "learning_rate": 6.239899144550383e-05, "loss": 0.6195, "step": 281 }, { "epoch": 1.9054054054054053, "grad_norm": 0.30422094244714903, "learning_rate": 6.224241610835391e-05, "loss": 0.6422, "step": 282 }, { "epoch": 1.9121621621621623, "grad_norm": 0.3974220280836707, "learning_rate": 6.208534585412282e-05, "loss": 0.6341, "step": 283 }, { "epoch": 1.9189189189189189, "grad_norm": 0.3588942161699914, "learning_rate": 6.19277841777884e-05, "loss": 0.6277, "step": 284 }, { "epoch": 1.9256756756756757, "grad_norm": 0.23348715119029995, "learning_rate": 6.176973458526317e-05, "loss": 0.6196, "step": 285 }, { "epoch": 1.9324324324324325, "grad_norm": 0.18377971391888995, "learning_rate": 6.161120059331628e-05, "loss": 0.6332, "step": 286 }, { "epoch": 1.939189189189189, "grad_norm": 0.2385215605440726, "learning_rate": 6.14521857294953e-05, "loss": 0.6479, "step": 287 }, { "epoch": 1.945945945945946, "grad_norm": 0.2797923846654065, "learning_rate": 6.129269353204769e-05, "loss": 0.6354, "step": 288 }, { "epoch": 1.9527027027027026, "grad_norm": 0.3046327128914545, "learning_rate": 6.113272754984206e-05, "loss": 0.6394, "step": 289 }, { "epoch": 1.9594594594594594, "grad_norm": 0.2685120513408917, "learning_rate": 6.0972291342289274e-05, "loss": 0.6285, "step": 290 }, { "epoch": 1.9662162162162162, "grad_norm": 0.19923838754652695, "learning_rate": 6.081138847926317e-05, "loss": 0.6239, "step": 291 }, { "epoch": 1.972972972972973, "grad_norm": 0.23688012061160577, "learning_rate": 6.065002254102116e-05, "loss": 0.6391, "step": 292 }, { "epoch": 1.9797297297297298, "grad_norm": 0.2444120597492392, "learning_rate": 6.048819711812457e-05, "loss": 0.6315, "step": 293 }, { "epoch": 1.9864864864864864, "grad_norm": 0.2374660611243683, "learning_rate": 6.032591581135878e-05, "loss": 0.635, "step": 294 }, { "epoch": 1.9932432432432432, "grad_norm": 0.19031195961334121, "learning_rate": 6.0163182231652985e-05, "loss": 0.6266, "step": 295 }, { "epoch": 2.0, "grad_norm": 0.2396442769161295, "learning_rate": 6.000000000000001e-05, "loss": 0.6384, "step": 296 }, { "epoch": 2.0067567567567566, "grad_norm": 0.2791243307653228, "learning_rate": 5.983637274737558e-05, "loss": 0.6251, "step": 297 }, { "epoch": 2.0135135135135136, "grad_norm": 0.3160836645945337, "learning_rate": 5.967230411465768e-05, "loss": 0.6261, "step": 298 }, { "epoch": 2.02027027027027, "grad_norm": 0.2816061126247031, "learning_rate": 5.950779775254539e-05, "loss": 0.6146, "step": 299 }, { "epoch": 2.027027027027027, "grad_norm": 0.3070948777190884, "learning_rate": 5.934285732147778e-05, "loss": 0.6234, "step": 300 }, { "epoch": 2.0337837837837838, "grad_norm": 0.31200072769112625, "learning_rate": 5.91774864915524e-05, "loss": 0.6129, "step": 301 }, { "epoch": 2.0405405405405403, "grad_norm": 0.28687016611033417, "learning_rate": 5.90116889424436e-05, "loss": 0.6144, "step": 302 }, { "epoch": 2.0472972972972974, "grad_norm": 0.3063930815670496, "learning_rate": 5.884546836332072e-05, "loss": 0.6176, "step": 303 }, { "epoch": 2.054054054054054, "grad_norm": 0.24478115745724477, "learning_rate": 5.867882845276593e-05, "loss": 0.6164, "step": 304 }, { "epoch": 2.060810810810811, "grad_norm": 0.17845425335065362, "learning_rate": 5.851177291869197e-05, "loss": 0.6118, "step": 305 }, { "epoch": 2.0675675675675675, "grad_norm": 0.20058855750068053, "learning_rate": 5.834430547825964e-05, "loss": 0.6146, "step": 306 }, { "epoch": 2.074324324324324, "grad_norm": 0.21828538712863907, "learning_rate": 5.8176429857795104e-05, "loss": 0.6176, "step": 307 }, { "epoch": 2.081081081081081, "grad_norm": 0.2218763340787443, "learning_rate": 5.8008149792706936e-05, "loss": 0.6059, "step": 308 }, { "epoch": 2.0878378378378377, "grad_norm": 0.26010217837913424, "learning_rate": 5.783946902740304e-05, "loss": 0.6172, "step": 309 }, { "epoch": 2.0945945945945947, "grad_norm": 0.2968234389164342, "learning_rate": 5.767039131520733e-05, "loss": 0.6176, "step": 310 }, { "epoch": 2.1013513513513513, "grad_norm": 0.29905738657768327, "learning_rate": 5.750092041827618e-05, "loss": 0.5998, "step": 311 }, { "epoch": 2.108108108108108, "grad_norm": 0.287828129320029, "learning_rate": 5.7331060107514754e-05, "loss": 0.597, "step": 312 }, { "epoch": 2.114864864864865, "grad_norm": 0.26831745019142034, "learning_rate": 5.716081416249307e-05, "loss": 0.6131, "step": 313 }, { "epoch": 2.1216216216216215, "grad_norm": 0.28531963857984527, "learning_rate": 5.699018637136192e-05, "loss": 0.6056, "step": 314 }, { "epoch": 2.1283783783783785, "grad_norm": 0.18651675715743835, "learning_rate": 5.681918053076858e-05, "loss": 0.6167, "step": 315 }, { "epoch": 2.135135135135135, "grad_norm": 0.21775207949583675, "learning_rate": 5.664780044577231e-05, "loss": 0.6219, "step": 316 }, { "epoch": 2.141891891891892, "grad_norm": 0.25323829124729375, "learning_rate": 5.6476049929759714e-05, "loss": 0.6196, "step": 317 }, { "epoch": 2.1486486486486487, "grad_norm": 0.19732115665472313, "learning_rate": 5.6303932804359857e-05, "loss": 0.6156, "step": 318 }, { "epoch": 2.1554054054054053, "grad_norm": 0.16849505056239242, "learning_rate": 5.613145289935926e-05, "loss": 0.6042, "step": 319 }, { "epoch": 2.1621621621621623, "grad_norm": 0.19969762470742003, "learning_rate": 5.595861405261666e-05, "loss": 0.6031, "step": 320 }, { "epoch": 2.168918918918919, "grad_norm": 0.2036043502047779, "learning_rate": 5.578542010997764e-05, "loss": 0.633, "step": 321 }, { "epoch": 2.175675675675676, "grad_norm": 0.16807494045865312, "learning_rate": 5.561187492518903e-05, "loss": 0.6118, "step": 322 }, { "epoch": 2.1824324324324325, "grad_norm": 0.1717048805694881, "learning_rate": 5.5437982359813156e-05, "loss": 0.6116, "step": 323 }, { "epoch": 2.189189189189189, "grad_norm": 0.17594564977890167, "learning_rate": 5.526374628314195e-05, "loss": 0.6162, "step": 324 }, { "epoch": 2.195945945945946, "grad_norm": 0.21489547073195236, "learning_rate": 5.50891705721108e-05, "loss": 0.5984, "step": 325 }, { "epoch": 2.2027027027027026, "grad_norm": 0.21447613988889852, "learning_rate": 5.4914259111212355e-05, "loss": 0.6198, "step": 326 }, { "epoch": 2.2094594594594597, "grad_norm": 0.2207232176733451, "learning_rate": 5.473901579241e-05, "loss": 0.6092, "step": 327 }, { "epoch": 2.2162162162162162, "grad_norm": 0.19937702101442842, "learning_rate": 5.4563444515051354e-05, "loss": 0.6065, "step": 328 }, { "epoch": 2.222972972972973, "grad_norm": 0.19293343910472205, "learning_rate": 5.438754918578144e-05, "loss": 0.6063, "step": 329 }, { "epoch": 2.22972972972973, "grad_norm": 0.17732267437010826, "learning_rate": 5.4211333718455756e-05, "loss": 0.6076, "step": 330 }, { "epoch": 2.2364864864864864, "grad_norm": 0.18101481051912677, "learning_rate": 5.4034802034053223e-05, "loss": 0.6154, "step": 331 }, { "epoch": 2.2432432432432434, "grad_norm": 0.19133561762748252, "learning_rate": 5.3857958060588955e-05, "loss": 0.5988, "step": 332 }, { "epoch": 2.25, "grad_norm": 0.18341857255842758, "learning_rate": 5.368080573302676e-05, "loss": 0.6093, "step": 333 }, { "epoch": 2.2567567567567566, "grad_norm": 0.2486112524831487, "learning_rate": 5.3503348993191706e-05, "loss": 0.6159, "step": 334 }, { "epoch": 2.2635135135135136, "grad_norm": 0.21869922307823292, "learning_rate": 5.332559178968231e-05, "loss": 0.6128, "step": 335 }, { "epoch": 2.27027027027027, "grad_norm": 0.1808704522354482, "learning_rate": 5.314753807778276e-05, "loss": 0.6157, "step": 336 }, { "epoch": 2.277027027027027, "grad_norm": 0.19637386821805067, "learning_rate": 5.296919181937485e-05, "loss": 0.6206, "step": 337 }, { "epoch": 2.2837837837837838, "grad_norm": 0.1867094463718795, "learning_rate": 5.279055698284982e-05, "loss": 0.6209, "step": 338 }, { "epoch": 2.2905405405405403, "grad_norm": 0.17359407157166826, "learning_rate": 5.261163754302011e-05, "loss": 0.6012, "step": 339 }, { "epoch": 2.2972972972972974, "grad_norm": 0.18569999759822897, "learning_rate": 5.2432437481030855e-05, "loss": 0.6048, "step": 340 }, { "epoch": 2.304054054054054, "grad_norm": 0.1773148171307398, "learning_rate": 5.225296078427135e-05, "loss": 0.5963, "step": 341 }, { "epoch": 2.310810810810811, "grad_norm": 0.16243539870707618, "learning_rate": 5.207321144628628e-05, "loss": 0.6178, "step": 342 }, { "epoch": 2.3175675675675675, "grad_norm": 0.18693063507419128, "learning_rate": 5.18931934666869e-05, "loss": 0.6136, "step": 343 }, { "epoch": 2.3243243243243246, "grad_norm": 0.1661326777898068, "learning_rate": 5.171291085106202e-05, "loss": 0.6235, "step": 344 }, { "epoch": 2.331081081081081, "grad_norm": 0.15331403054989767, "learning_rate": 5.153236761088888e-05, "loss": 0.6098, "step": 345 }, { "epoch": 2.3378378378378377, "grad_norm": 0.1570395157158325, "learning_rate": 5.135156776344389e-05, "loss": 0.6137, "step": 346 }, { "epoch": 2.3445945945945947, "grad_norm": 0.17430865513764107, "learning_rate": 5.117051533171321e-05, "loss": 0.6177, "step": 347 }, { "epoch": 2.3513513513513513, "grad_norm": 0.21717070798995733, "learning_rate": 5.098921434430333e-05, "loss": 0.6203, "step": 348 }, { "epoch": 2.358108108108108, "grad_norm": 0.20392614118792024, "learning_rate": 5.080766883535129e-05, "loss": 0.6136, "step": 349 }, { "epoch": 2.364864864864865, "grad_norm": 0.21588477709369264, "learning_rate": 5.062588284443505e-05, "loss": 0.6145, "step": 350 }, { "epoch": 2.3716216216216215, "grad_norm": 0.21049543340334342, "learning_rate": 5.0443860416483536e-05, "loss": 0.6019, "step": 351 }, { "epoch": 2.3783783783783785, "grad_norm": 0.1730662628064374, "learning_rate": 5.026160560168661e-05, "loss": 0.6058, "step": 352 }, { "epoch": 2.385135135135135, "grad_norm": 0.17831683554874725, "learning_rate": 5.0079122455405014e-05, "loss": 0.6208, "step": 353 }, { "epoch": 2.391891891891892, "grad_norm": 0.17744701669503546, "learning_rate": 4.989641503808011e-05, "loss": 0.609, "step": 354 }, { "epoch": 2.3986486486486487, "grad_norm": 0.17994097783132235, "learning_rate": 4.971348741514349e-05, "loss": 0.6066, "step": 355 }, { "epoch": 2.4054054054054053, "grad_norm": 0.17459787928555573, "learning_rate": 4.95303436569266e-05, "loss": 0.6146, "step": 356 }, { "epoch": 2.4121621621621623, "grad_norm": 0.16284962487352878, "learning_rate": 4.934698783857011e-05, "loss": 0.6044, "step": 357 }, { "epoch": 2.418918918918919, "grad_norm": 0.17767021224773352, "learning_rate": 4.91634240399332e-05, "loss": 0.6211, "step": 358 }, { "epoch": 2.4256756756756754, "grad_norm": 0.1496311992502912, "learning_rate": 4.8979656345502904e-05, "loss": 0.602, "step": 359 }, { "epoch": 2.4324324324324325, "grad_norm": 0.16981969578628817, "learning_rate": 4.8795688844303114e-05, "loss": 0.6148, "step": 360 }, { "epoch": 2.439189189189189, "grad_norm": 0.2089769520148497, "learning_rate": 4.861152562980362e-05, "loss": 0.6123, "step": 361 }, { "epoch": 2.445945945945946, "grad_norm": 0.20528877980388496, "learning_rate": 4.8427170799829055e-05, "loss": 0.6095, "step": 362 }, { "epoch": 2.4527027027027026, "grad_norm": 0.1827881621296737, "learning_rate": 4.824262845646771e-05, "loss": 0.6112, "step": 363 }, { "epoch": 2.4594594594594597, "grad_norm": 0.19580070665324928, "learning_rate": 4.805790270598021e-05, "loss": 0.6209, "step": 364 }, { "epoch": 2.4662162162162162, "grad_norm": 0.23276386582983333, "learning_rate": 4.787299765870822e-05, "loss": 0.6083, "step": 365 }, { "epoch": 2.472972972972973, "grad_norm": 0.20338848706607754, "learning_rate": 4.768791742898292e-05, "loss": 0.615, "step": 366 }, { "epoch": 2.47972972972973, "grad_norm": 0.1521765603212478, "learning_rate": 4.7502666135033486e-05, "loss": 0.6154, "step": 367 }, { "epoch": 2.4864864864864864, "grad_norm": 0.1559952589700639, "learning_rate": 4.731724789889547e-05, "loss": 0.6026, "step": 368 }, { "epoch": 2.4932432432432434, "grad_norm": 0.14084577612463858, "learning_rate": 4.7131666846319036e-05, "loss": 0.607, "step": 369 }, { "epoch": 2.5, "grad_norm": 0.1450823104838612, "learning_rate": 4.694592710667723e-05, "loss": 0.6173, "step": 370 }, { "epoch": 2.506756756756757, "grad_norm": 0.14280439260452932, "learning_rate": 4.676003281287397e-05, "loss": 0.6035, "step": 371 }, { "epoch": 2.5135135135135136, "grad_norm": 0.1731629109803712, "learning_rate": 4.657398810125225e-05, "loss": 0.6044, "step": 372 }, { "epoch": 2.52027027027027, "grad_norm": 0.14551348661799535, "learning_rate": 4.638779711150198e-05, "loss": 0.6223, "step": 373 }, { "epoch": 2.527027027027027, "grad_norm": 0.15724571828258985, "learning_rate": 4.620146398656792e-05, "loss": 0.6174, "step": 374 }, { "epoch": 2.5337837837837838, "grad_norm": 0.1622156186595054, "learning_rate": 4.601499287255748e-05, "loss": 0.6159, "step": 375 }, { "epoch": 2.5405405405405403, "grad_norm": 0.16326776251404446, "learning_rate": 4.582838791864846e-05, "loss": 0.6178, "step": 376 }, { "epoch": 2.5472972972972974, "grad_norm": 0.16705938089441114, "learning_rate": 4.5641653276996774e-05, "loss": 0.6069, "step": 377 }, { "epoch": 2.554054054054054, "grad_norm": 0.1240389952109112, "learning_rate": 4.5454793102644006e-05, "loss": 0.6129, "step": 378 }, { "epoch": 2.560810810810811, "grad_norm": 0.16500872440232617, "learning_rate": 4.5267811553424945e-05, "loss": 0.6213, "step": 379 }, { "epoch": 2.5675675675675675, "grad_norm": 0.17252309025914744, "learning_rate": 4.5080712789875154e-05, "loss": 0.6172, "step": 380 }, { "epoch": 2.5743243243243246, "grad_norm": 0.163801171559839, "learning_rate": 4.489350097513829e-05, "loss": 0.6205, "step": 381 }, { "epoch": 2.581081081081081, "grad_norm": 0.14139376100405637, "learning_rate": 4.470618027487354e-05, "loss": 0.6101, "step": 382 }, { "epoch": 2.5878378378378377, "grad_norm": 0.14839912258194626, "learning_rate": 4.451875485716292e-05, "loss": 0.6214, "step": 383 }, { "epoch": 2.5945945945945947, "grad_norm": 0.13849384656147448, "learning_rate": 4.4331228892418473e-05, "loss": 0.6111, "step": 384 }, { "epoch": 2.6013513513513513, "grad_norm": 0.16314353286789263, "learning_rate": 4.414360655328957e-05, "loss": 0.6201, "step": 385 }, { "epoch": 2.608108108108108, "grad_norm": 0.13166529990196998, "learning_rate": 4.395589201457e-05, "loss": 0.6167, "step": 386 }, { "epoch": 2.614864864864865, "grad_norm": 0.14086485553732253, "learning_rate": 4.376808945310505e-05, "loss": 0.6219, "step": 387 }, { "epoch": 2.6216216216216215, "grad_norm": 0.15193755776745674, "learning_rate": 4.358020304769867e-05, "loss": 0.5994, "step": 388 }, { "epoch": 2.6283783783783785, "grad_norm": 0.14946480137653986, "learning_rate": 4.339223697902037e-05, "loss": 0.61, "step": 389 }, { "epoch": 2.635135135135135, "grad_norm": 0.15531265023719382, "learning_rate": 4.320419542951228e-05, "loss": 0.6179, "step": 390 }, { "epoch": 2.641891891891892, "grad_norm": 0.14264508419628688, "learning_rate": 4.3016082583296067e-05, "loss": 0.6167, "step": 391 }, { "epoch": 2.6486486486486487, "grad_norm": 0.12288922241404748, "learning_rate": 4.2827902626079784e-05, "loss": 0.6195, "step": 392 }, { "epoch": 2.6554054054054053, "grad_norm": 0.19330457974395565, "learning_rate": 4.263965974506483e-05, "loss": 0.6207, "step": 393 }, { "epoch": 2.6621621621621623, "grad_norm": 0.16883088734435447, "learning_rate": 4.2451358128852654e-05, "loss": 0.5989, "step": 394 }, { "epoch": 2.668918918918919, "grad_norm": 0.1599604515688246, "learning_rate": 4.22630019673517e-05, "loss": 0.6142, "step": 395 }, { "epoch": 2.6756756756756754, "grad_norm": 0.15518611864455714, "learning_rate": 4.207459545168405e-05, "loss": 0.6102, "step": 396 }, { "epoch": 2.6824324324324325, "grad_norm": 0.16742006705568305, "learning_rate": 4.188614277409224e-05, "loss": 0.6171, "step": 397 }, { "epoch": 2.689189189189189, "grad_norm": 0.15553602245497833, "learning_rate": 4.169764812784594e-05, "loss": 0.6143, "step": 398 }, { "epoch": 2.695945945945946, "grad_norm": 0.1595818544532621, "learning_rate": 4.1509115707148695e-05, "loss": 0.6047, "step": 399 }, { "epoch": 2.7027027027027026, "grad_norm": 0.13690560623925258, "learning_rate": 4.132054970704454e-05, "loss": 0.6092, "step": 400 }, { "epoch": 2.7094594594594597, "grad_norm": 0.1341753352699138, "learning_rate": 4.1131954323324734e-05, "loss": 0.6023, "step": 401 }, { "epoch": 2.7162162162162162, "grad_norm": 0.14452976105100054, "learning_rate": 4.094333375243428e-05, "loss": 0.6121, "step": 402 }, { "epoch": 2.722972972972973, "grad_norm": 0.17855418886582433, "learning_rate": 4.075469219137868e-05, "loss": 0.6192, "step": 403 }, { "epoch": 2.72972972972973, "grad_norm": 0.15337873200421298, "learning_rate": 4.056603383763049e-05, "loss": 0.6137, "step": 404 }, { "epoch": 2.7364864864864864, "grad_norm": 0.14193840867721655, "learning_rate": 4.0377362889035875e-05, "loss": 0.5971, "step": 405 }, { "epoch": 2.743243243243243, "grad_norm": 0.14049198222822518, "learning_rate": 4.0188683543721295e-05, "loss": 0.61, "step": 406 }, { "epoch": 2.75, "grad_norm": 0.1643045132978014, "learning_rate": 4e-05, "loss": 0.6224, "step": 407 }, { "epoch": 2.756756756756757, "grad_norm": 0.14527602919509278, "learning_rate": 3.981131645627872e-05, "loss": 0.6164, "step": 408 }, { "epoch": 2.7635135135135136, "grad_norm": 0.1384596570378767, "learning_rate": 3.9622637110964125e-05, "loss": 0.6158, "step": 409 }, { "epoch": 2.77027027027027, "grad_norm": 0.18210804820042478, "learning_rate": 3.943396616236953e-05, "loss": 0.6149, "step": 410 }, { "epoch": 2.777027027027027, "grad_norm": 0.16943522003893646, "learning_rate": 3.9245307808621325e-05, "loss": 0.6159, "step": 411 }, { "epoch": 2.7837837837837838, "grad_norm": 0.16603181052517613, "learning_rate": 3.905666624756573e-05, "loss": 0.6186, "step": 412 }, { "epoch": 2.7905405405405403, "grad_norm": 0.15368628411505494, "learning_rate": 3.886804567667528e-05, "loss": 0.6155, "step": 413 }, { "epoch": 2.7972972972972974, "grad_norm": 0.1721774875064551, "learning_rate": 3.867945029295546e-05, "loss": 0.6099, "step": 414 }, { "epoch": 2.804054054054054, "grad_norm": 0.1893984092920587, "learning_rate": 3.8490884292851325e-05, "loss": 0.6146, "step": 415 }, { "epoch": 2.810810810810811, "grad_norm": 0.1792007727966493, "learning_rate": 3.830235187215408e-05, "loss": 0.6195, "step": 416 }, { "epoch": 2.8175675675675675, "grad_norm": 0.1569338853010942, "learning_rate": 3.8113857225907783e-05, "loss": 0.6173, "step": 417 }, { "epoch": 2.8243243243243246, "grad_norm": 0.17031740028829662, "learning_rate": 3.792540454831596e-05, "loss": 0.6163, "step": 418 }, { "epoch": 2.831081081081081, "grad_norm": 0.15004651120970428, "learning_rate": 3.7736998032648305e-05, "loss": 0.6047, "step": 419 }, { "epoch": 2.8378378378378377, "grad_norm": 0.142075802466718, "learning_rate": 3.754864187114736e-05, "loss": 0.6253, "step": 420 }, { "epoch": 2.8445945945945947, "grad_norm": 0.15599780395707422, "learning_rate": 3.736034025493519e-05, "loss": 0.593, "step": 421 }, { "epoch": 2.8513513513513513, "grad_norm": 0.1563201754706643, "learning_rate": 3.717209737392022e-05, "loss": 0.6128, "step": 422 }, { "epoch": 2.858108108108108, "grad_norm": 0.15805265922848155, "learning_rate": 3.698391741670394e-05, "loss": 0.6059, "step": 423 }, { "epoch": 2.864864864864865, "grad_norm": 0.17933457522155424, "learning_rate": 3.679580457048772e-05, "loss": 0.6086, "step": 424 }, { "epoch": 2.8716216216216215, "grad_norm": 0.17371124474695815, "learning_rate": 3.660776302097965e-05, "loss": 0.6079, "step": 425 }, { "epoch": 2.8783783783783785, "grad_norm": 0.1315584411453832, "learning_rate": 3.641979695230135e-05, "loss": 0.6039, "step": 426 }, { "epoch": 2.885135135135135, "grad_norm": 0.148272614562961, "learning_rate": 3.6231910546894956e-05, "loss": 0.6161, "step": 427 }, { "epoch": 2.891891891891892, "grad_norm": 0.12013666775227018, "learning_rate": 3.6044107985430015e-05, "loss": 0.6113, "step": 428 }, { "epoch": 2.8986486486486487, "grad_norm": 0.15454206210883253, "learning_rate": 3.585639344671043e-05, "loss": 0.6113, "step": 429 }, { "epoch": 2.9054054054054053, "grad_norm": 0.130251935264837, "learning_rate": 3.5668771107581526e-05, "loss": 0.613, "step": 430 }, { "epoch": 2.9121621621621623, "grad_norm": 0.1370783380801396, "learning_rate": 3.5481245142837095e-05, "loss": 0.6168, "step": 431 }, { "epoch": 2.918918918918919, "grad_norm": 0.13116479132749334, "learning_rate": 3.5293819725126464e-05, "loss": 0.6136, "step": 432 }, { "epoch": 2.9256756756756754, "grad_norm": 0.14031976128589324, "learning_rate": 3.5106499024861715e-05, "loss": 0.6175, "step": 433 }, { "epoch": 2.9324324324324325, "grad_norm": 0.12054376627925889, "learning_rate": 3.491928721012485e-05, "loss": 0.6134, "step": 434 }, { "epoch": 2.939189189189189, "grad_norm": 0.14272909957039542, "learning_rate": 3.4732188446575055e-05, "loss": 0.6096, "step": 435 }, { "epoch": 2.945945945945946, "grad_norm": 0.1343117824779506, "learning_rate": 3.454520689735602e-05, "loss": 0.6203, "step": 436 }, { "epoch": 2.9527027027027026, "grad_norm": 0.11376422264610185, "learning_rate": 3.435834672300324e-05, "loss": 0.6057, "step": 437 }, { "epoch": 2.9594594594594597, "grad_norm": 0.14208820005659473, "learning_rate": 3.417161208135155e-05, "loss": 0.6193, "step": 438 }, { "epoch": 2.9662162162162162, "grad_norm": 0.11247340517088436, "learning_rate": 3.398500712744254e-05, "loss": 0.6218, "step": 439 }, { "epoch": 2.972972972972973, "grad_norm": 0.12141011100880744, "learning_rate": 3.379853601343209e-05, "loss": 0.6124, "step": 440 }, { "epoch": 2.97972972972973, "grad_norm": 0.11441162073496848, "learning_rate": 3.361220288849804e-05, "loss": 0.6146, "step": 441 }, { "epoch": 2.9864864864864864, "grad_norm": 0.11128267132483809, "learning_rate": 3.342601189874777e-05, "loss": 0.6141, "step": 442 }, { "epoch": 2.993243243243243, "grad_norm": 0.13308500191665273, "learning_rate": 3.323996718712605e-05, "loss": 0.6154, "step": 443 }, { "epoch": 3.0, "grad_norm": 0.13634885719680506, "learning_rate": 3.305407289332279e-05, "loss": 0.6063, "step": 444 }, { "epoch": 3.0067567567567566, "grad_norm": 0.14117643344569666, "learning_rate": 3.2868333153680964e-05, "loss": 0.5952, "step": 445 }, { "epoch": 3.0135135135135136, "grad_norm": 0.1441355574419299, "learning_rate": 3.2682752101104536e-05, "loss": 0.5807, "step": 446 }, { "epoch": 3.02027027027027, "grad_norm": 0.13317791462913253, "learning_rate": 3.249733386496653e-05, "loss": 0.6011, "step": 447 }, { "epoch": 3.027027027027027, "grad_norm": 0.1488227563399244, "learning_rate": 3.231208257101709e-05, "loss": 0.6055, "step": 448 }, { "epoch": 3.0337837837837838, "grad_norm": 0.13698170944523078, "learning_rate": 3.212700234129179e-05, "loss": 0.6012, "step": 449 }, { "epoch": 3.0405405405405403, "grad_norm": 0.13426258482203746, "learning_rate": 3.194209729401979e-05, "loss": 0.5955, "step": 450 }, { "epoch": 3.0472972972972974, "grad_norm": 0.12763528667729904, "learning_rate": 3.175737154353231e-05, "loss": 0.5829, "step": 451 }, { "epoch": 3.054054054054054, "grad_norm": 0.12302649236306719, "learning_rate": 3.157282920017096e-05, "loss": 0.6031, "step": 452 }, { "epoch": 3.060810810810811, "grad_norm": 0.12836088172431226, "learning_rate": 3.1388474370196395e-05, "loss": 0.6029, "step": 453 }, { "epoch": 3.0675675675675675, "grad_norm": 0.12713682511389784, "learning_rate": 3.12043111556969e-05, "loss": 0.5898, "step": 454 }, { "epoch": 3.074324324324324, "grad_norm": 0.13954610740683865, "learning_rate": 3.1020343654497096e-05, "loss": 0.5882, "step": 455 }, { "epoch": 3.081081081081081, "grad_norm": 0.14394153328876694, "learning_rate": 3.083657596006681e-05, "loss": 0.5878, "step": 456 }, { "epoch": 3.0878378378378377, "grad_norm": 0.11738094672449853, "learning_rate": 3.065301216142991e-05, "loss": 0.5916, "step": 457 }, { "epoch": 3.0945945945945947, "grad_norm": 0.1473396296692266, "learning_rate": 3.046965634307341e-05, "loss": 0.6003, "step": 458 }, { "epoch": 3.1013513513513513, "grad_norm": 0.12128542181194722, "learning_rate": 3.028651258485652e-05, "loss": 0.5807, "step": 459 }, { "epoch": 3.108108108108108, "grad_norm": 0.12076772408917641, "learning_rate": 3.010358496191991e-05, "loss": 0.5968, "step": 460 }, { "epoch": 3.114864864864865, "grad_norm": 0.12207943508497456, "learning_rate": 2.9920877544595002e-05, "loss": 0.5911, "step": 461 }, { "epoch": 3.1216216216216215, "grad_norm": 0.1193967138970814, "learning_rate": 2.9738394398313405e-05, "loss": 0.5867, "step": 462 }, { "epoch": 3.1283783783783785, "grad_norm": 0.11942196780407818, "learning_rate": 2.955613958351647e-05, "loss": 0.5927, "step": 463 }, { "epoch": 3.135135135135135, "grad_norm": 0.12137977118056532, "learning_rate": 2.9374117155564957e-05, "loss": 0.5923, "step": 464 }, { "epoch": 3.141891891891892, "grad_norm": 0.1204663577707489, "learning_rate": 2.919233116464872e-05, "loss": 0.5807, "step": 465 }, { "epoch": 3.1486486486486487, "grad_norm": 0.1277645131662919, "learning_rate": 2.9010785655696698e-05, "loss": 0.602, "step": 466 }, { "epoch": 3.1554054054054053, "grad_norm": 0.14373007650684674, "learning_rate": 2.88294846682868e-05, "loss": 0.5952, "step": 467 }, { "epoch": 3.1621621621621623, "grad_norm": 0.12531967044629988, "learning_rate": 2.864843223655613e-05, "loss": 0.6065, "step": 468 }, { "epoch": 3.168918918918919, "grad_norm": 0.1344060345828822, "learning_rate": 2.8467632389111126e-05, "loss": 0.5879, "step": 469 }, { "epoch": 3.175675675675676, "grad_norm": 0.11307986878514827, "learning_rate": 2.828708914893799e-05, "loss": 0.5923, "step": 470 }, { "epoch": 3.1824324324324325, "grad_norm": 0.1205066746149822, "learning_rate": 2.8106806533313106e-05, "loss": 0.6042, "step": 471 }, { "epoch": 3.189189189189189, "grad_norm": 0.11358248639127554, "learning_rate": 2.7926788553713734e-05, "loss": 0.6017, "step": 472 }, { "epoch": 3.195945945945946, "grad_norm": 0.11273620191303046, "learning_rate": 2.7747039215728667e-05, "loss": 0.5909, "step": 473 }, { "epoch": 3.2027027027027026, "grad_norm": 0.1527832963288818, "learning_rate": 2.7567562518969155e-05, "loss": 0.5884, "step": 474 }, { "epoch": 3.2094594594594597, "grad_norm": 0.1251866201941187, "learning_rate": 2.7388362456979906e-05, "loss": 0.5942, "step": 475 }, { "epoch": 3.2162162162162162, "grad_norm": 0.131897788334417, "learning_rate": 2.7209443017150193e-05, "loss": 0.6028, "step": 476 }, { "epoch": 3.222972972972973, "grad_norm": 0.10949010701632424, "learning_rate": 2.703080818062517e-05, "loss": 0.5899, "step": 477 }, { "epoch": 3.22972972972973, "grad_norm": 0.1085593433907961, "learning_rate": 2.6852461922217253e-05, "loss": 0.5915, "step": 478 }, { "epoch": 3.2364864864864864, "grad_norm": 0.10851474030600673, "learning_rate": 2.66744082103177e-05, "loss": 0.5956, "step": 479 }, { "epoch": 3.2432432432432434, "grad_norm": 0.10556652124147405, "learning_rate": 2.6496651006808308e-05, "loss": 0.5923, "step": 480 }, { "epoch": 3.25, "grad_norm": 0.11511281568946971, "learning_rate": 2.6319194266973256e-05, "loss": 0.5982, "step": 481 }, { "epoch": 3.2567567567567566, "grad_norm": 0.10661745620059551, "learning_rate": 2.614204193941107e-05, "loss": 0.6029, "step": 482 }, { "epoch": 3.2635135135135136, "grad_norm": 0.12822931878201987, "learning_rate": 2.5965197965946783e-05, "loss": 0.5935, "step": 483 }, { "epoch": 3.27027027027027, "grad_norm": 0.1156142465379123, "learning_rate": 2.5788666281544258e-05, "loss": 0.6142, "step": 484 }, { "epoch": 3.277027027027027, "grad_norm": 0.13148336347712752, "learning_rate": 2.561245081421857e-05, "loss": 0.5942, "step": 485 }, { "epoch": 3.2837837837837838, "grad_norm": 0.10123146621891457, "learning_rate": 2.5436555484948643e-05, "loss": 0.5989, "step": 486 }, { "epoch": 3.2905405405405403, "grad_norm": 0.1375355982170109, "learning_rate": 2.5260984207590015e-05, "loss": 0.5913, "step": 487 }, { "epoch": 3.2972972972972974, "grad_norm": 0.1071253679793487, "learning_rate": 2.5085740888787662e-05, "loss": 0.6034, "step": 488 }, { "epoch": 3.304054054054054, "grad_norm": 0.13183672834367724, "learning_rate": 2.4910829427889205e-05, "loss": 0.6016, "step": 489 }, { "epoch": 3.310810810810811, "grad_norm": 0.11417300702406326, "learning_rate": 2.473625371685806e-05, "loss": 0.5964, "step": 490 }, { "epoch": 3.3175675675675675, "grad_norm": 0.11783207513787236, "learning_rate": 2.4562017640186847e-05, "loss": 0.5983, "step": 491 }, { "epoch": 3.3243243243243246, "grad_norm": 0.12341779828696402, "learning_rate": 2.4388125074810986e-05, "loss": 0.5925, "step": 492 }, { "epoch": 3.331081081081081, "grad_norm": 0.11637168120668157, "learning_rate": 2.4214579890022373e-05, "loss": 0.5771, "step": 493 }, { "epoch": 3.3378378378378377, "grad_norm": 0.10923183772815541, "learning_rate": 2.404138594738335e-05, "loss": 0.5919, "step": 494 }, { "epoch": 3.3445945945945947, "grad_norm": 0.10526844464204445, "learning_rate": 2.386854710064075e-05, "loss": 0.607, "step": 495 }, { "epoch": 3.3513513513513513, "grad_norm": 0.10474734642926019, "learning_rate": 2.369606719564015e-05, "loss": 0.5907, "step": 496 }, { "epoch": 3.358108108108108, "grad_norm": 0.11107800655708087, "learning_rate": 2.35239500702403e-05, "loss": 0.5981, "step": 497 }, { "epoch": 3.364864864864865, "grad_norm": 0.10187850179282668, "learning_rate": 2.3352199554227698e-05, "loss": 0.5938, "step": 498 }, { "epoch": 3.3716216216216215, "grad_norm": 0.12632250115529653, "learning_rate": 2.318081946923144e-05, "loss": 0.6117, "step": 499 }, { "epoch": 3.3783783783783785, "grad_norm": 0.10055327704116511, "learning_rate": 2.3009813628638085e-05, "loss": 0.5964, "step": 500 }, { "epoch": 3.385135135135135, "grad_norm": 0.11916682604562856, "learning_rate": 2.283918583750695e-05, "loss": 0.5871, "step": 501 }, { "epoch": 3.391891891891892, "grad_norm": 0.1010245925435501, "learning_rate": 2.266893989248527e-05, "loss": 0.6004, "step": 502 }, { "epoch": 3.3986486486486487, "grad_norm": 0.10446456567129823, "learning_rate": 2.2499079581723846e-05, "loss": 0.5942, "step": 503 }, { "epoch": 3.4054054054054053, "grad_norm": 0.10099772014538502, "learning_rate": 2.2329608684792676e-05, "loss": 0.5956, "step": 504 }, { "epoch": 3.4121621621621623, "grad_norm": 0.1005267977448251, "learning_rate": 2.216053097259697e-05, "loss": 0.5898, "step": 505 }, { "epoch": 3.418918918918919, "grad_norm": 0.10513240417556334, "learning_rate": 2.1991850207293064e-05, "loss": 0.5997, "step": 506 }, { "epoch": 3.4256756756756754, "grad_norm": 0.09929924802833451, "learning_rate": 2.1823570142204902e-05, "loss": 0.5932, "step": 507 }, { "epoch": 3.4324324324324325, "grad_norm": 0.10494976279274904, "learning_rate": 2.1655694521740376e-05, "loss": 0.5935, "step": 508 }, { "epoch": 3.439189189189189, "grad_norm": 0.10732590635113842, "learning_rate": 2.1488227081308054e-05, "loss": 0.589, "step": 509 }, { "epoch": 3.445945945945946, "grad_norm": 0.1154273309217794, "learning_rate": 2.132117154723408e-05, "loss": 0.5927, "step": 510 }, { "epoch": 3.4527027027027026, "grad_norm": 0.10250291685305624, "learning_rate": 2.115453163667929e-05, "loss": 0.6047, "step": 511 }, { "epoch": 3.4594594594594597, "grad_norm": 0.1053185431025155, "learning_rate": 2.0988311057556397e-05, "loss": 0.5841, "step": 512 }, { "epoch": 3.4662162162162162, "grad_norm": 0.09414135194355852, "learning_rate": 2.0822513508447608e-05, "loss": 0.5863, "step": 513 }, { "epoch": 3.472972972972973, "grad_norm": 0.10071155506936441, "learning_rate": 2.065714267852223e-05, "loss": 0.5923, "step": 514 }, { "epoch": 3.47972972972973, "grad_norm": 0.10121850558468308, "learning_rate": 2.049220224745463e-05, "loss": 0.5962, "step": 515 }, { "epoch": 3.4864864864864864, "grad_norm": 0.10161225481650848, "learning_rate": 2.032769588534233e-05, "loss": 0.5925, "step": 516 }, { "epoch": 3.4932432432432434, "grad_norm": 0.09239639095614875, "learning_rate": 2.0163627252624427e-05, "loss": 0.5944, "step": 517 }, { "epoch": 3.5, "grad_norm": 0.11360448222663692, "learning_rate": 2.0000000000000012e-05, "loss": 0.6116, "step": 518 }, { "epoch": 3.506756756756757, "grad_norm": 0.09631002455489597, "learning_rate": 1.9836817768347015e-05, "loss": 0.5947, "step": 519 }, { "epoch": 3.5135135135135136, "grad_norm": 0.1167172974336559, "learning_rate": 1.9674084188641235e-05, "loss": 0.5843, "step": 520 }, { "epoch": 3.52027027027027, "grad_norm": 0.10349743444704453, "learning_rate": 1.9511802881875438e-05, "loss": 0.6026, "step": 521 }, { "epoch": 3.527027027027027, "grad_norm": 0.11753598782061575, "learning_rate": 1.9349977458978846e-05, "loss": 0.5892, "step": 522 }, { "epoch": 3.5337837837837838, "grad_norm": 0.09146602359405204, "learning_rate": 1.9188611520736846e-05, "loss": 0.5916, "step": 523 }, { "epoch": 3.5405405405405403, "grad_norm": 0.10892529356211986, "learning_rate": 1.902770865771074e-05, "loss": 0.5854, "step": 524 }, { "epoch": 3.5472972972972974, "grad_norm": 0.09418036132529382, "learning_rate": 1.886727245015794e-05, "loss": 0.5993, "step": 525 }, { "epoch": 3.554054054054054, "grad_norm": 0.1025611805390442, "learning_rate": 1.8707306467952323e-05, "loss": 0.5916, "step": 526 }, { "epoch": 3.560810810810811, "grad_norm": 0.08852437916251449, "learning_rate": 1.8547814270504705e-05, "loss": 0.5946, "step": 527 }, { "epoch": 3.5675675675675675, "grad_norm": 0.0999923376365888, "learning_rate": 1.838879940668373e-05, "loss": 0.6, "step": 528 }, { "epoch": 3.5743243243243246, "grad_norm": 0.10873141849605454, "learning_rate": 1.823026541473684e-05, "loss": 0.6017, "step": 529 }, { "epoch": 3.581081081081081, "grad_norm": 0.10599260470783008, "learning_rate": 1.8072215822211613e-05, "loss": 0.5968, "step": 530 }, { "epoch": 3.5878378378378377, "grad_norm": 0.11188333764585037, "learning_rate": 1.7914654145877187e-05, "loss": 0.6064, "step": 531 }, { "epoch": 3.5945945945945947, "grad_norm": 0.09583819123841314, "learning_rate": 1.77575838916461e-05, "loss": 0.5876, "step": 532 }, { "epoch": 3.6013513513513513, "grad_norm": 0.10206469144841168, "learning_rate": 1.760100855449619e-05, "loss": 0.5948, "step": 533 }, { "epoch": 3.608108108108108, "grad_norm": 0.10036289538603702, "learning_rate": 1.7444931618392894e-05, "loss": 0.6107, "step": 534 }, { "epoch": 3.614864864864865, "grad_norm": 0.08314263844698841, "learning_rate": 1.7289356556211687e-05, "loss": 0.6059, "step": 535 }, { "epoch": 3.6216216216216215, "grad_norm": 0.10484951196819213, "learning_rate": 1.7134286829660855e-05, "loss": 0.5961, "step": 536 }, { "epoch": 3.6283783783783785, "grad_norm": 0.07757732001833932, "learning_rate": 1.697972588920439e-05, "loss": 0.603, "step": 537 }, { "epoch": 3.635135135135135, "grad_norm": 0.09636592692846442, "learning_rate": 1.6825677173985332e-05, "loss": 0.597, "step": 538 }, { "epoch": 3.641891891891892, "grad_norm": 0.0934961698224681, "learning_rate": 1.6672144111749066e-05, "loss": 0.6065, "step": 539 }, { "epoch": 3.6486486486486487, "grad_norm": 0.0956796061229755, "learning_rate": 1.6519130118767258e-05, "loss": 0.5978, "step": 540 }, { "epoch": 3.6554054054054053, "grad_norm": 0.1161980558323059, "learning_rate": 1.6366638599761676e-05, "loss": 0.5914, "step": 541 }, { "epoch": 3.6621621621621623, "grad_norm": 0.08312306306871219, "learning_rate": 1.621467294782854e-05, "loss": 0.5855, "step": 542 }, { "epoch": 3.668918918918919, "grad_norm": 0.11701156024970683, "learning_rate": 1.606323654436293e-05, "loss": 0.5904, "step": 543 }, { "epoch": 3.6756756756756754, "grad_norm": 0.09466465341207426, "learning_rate": 1.591233275898363e-05, "loss": 0.5915, "step": 544 }, { "epoch": 3.6824324324324325, "grad_norm": 0.10284506462525071, "learning_rate": 1.5761964949458076e-05, "loss": 0.5847, "step": 545 }, { "epoch": 3.689189189189189, "grad_norm": 0.10382209320021844, "learning_rate": 1.5612136461627726e-05, "loss": 0.596, "step": 546 }, { "epoch": 3.695945945945946, "grad_norm": 0.09709377064685741, "learning_rate": 1.546285062933352e-05, "loss": 0.5909, "step": 547 }, { "epoch": 3.7027027027027026, "grad_norm": 0.11272439075208067, "learning_rate": 1.5314110774341803e-05, "loss": 0.5953, "step": 548 }, { "epoch": 3.7094594594594597, "grad_norm": 0.11904326038877579, "learning_rate": 1.5165920206270257e-05, "loss": 0.5929, "step": 549 }, { "epoch": 3.7162162162162162, "grad_norm": 0.1039606133991536, "learning_rate": 1.5018282222514451e-05, "loss": 0.5877, "step": 550 }, { "epoch": 3.722972972972973, "grad_norm": 0.11190632151336002, "learning_rate": 1.4871200108174306e-05, "loss": 0.5917, "step": 551 }, { "epoch": 3.72972972972973, "grad_norm": 0.09272507731002902, "learning_rate": 1.4724677135981118e-05, "loss": 0.5959, "step": 552 }, { "epoch": 3.7364864864864864, "grad_norm": 0.09431866458492948, "learning_rate": 1.457871656622463e-05, "loss": 0.5887, "step": 553 }, { "epoch": 3.743243243243243, "grad_norm": 0.08680813466014699, "learning_rate": 1.4433321646680614e-05, "loss": 0.6032, "step": 554 }, { "epoch": 3.75, "grad_norm": 0.0844092691423774, "learning_rate": 1.4288495612538427e-05, "loss": 0.6062, "step": 555 }, { "epoch": 3.756756756756757, "grad_norm": 0.0874893569172525, "learning_rate": 1.4144241686329236e-05, "loss": 0.5833, "step": 556 }, { "epoch": 3.7635135135135136, "grad_norm": 0.10957613038759607, "learning_rate": 1.400056307785413e-05, "loss": 0.6056, "step": 557 }, { "epoch": 3.77027027027027, "grad_norm": 0.07882052024046089, "learning_rate": 1.3857462984112831e-05, "loss": 0.5912, "step": 558 }, { "epoch": 3.777027027027027, "grad_norm": 0.08565900595306047, "learning_rate": 1.371494458923246e-05, "loss": 0.5919, "step": 559 }, { "epoch": 3.7837837837837838, "grad_norm": 0.09674275812299338, "learning_rate": 1.3573011064396751e-05, "loss": 0.5985, "step": 560 }, { "epoch": 3.7905405405405403, "grad_norm": 0.08500752883113108, "learning_rate": 1.3431665567775439e-05, "loss": 0.5983, "step": 561 }, { "epoch": 3.7972972972972974, "grad_norm": 0.11583654081493171, "learning_rate": 1.3290911244454066e-05, "loss": 0.5903, "step": 562 }, { "epoch": 3.804054054054054, "grad_norm": 0.07940239999432974, "learning_rate": 1.3150751226363886e-05, "loss": 0.5931, "step": 563 }, { "epoch": 3.810810810810811, "grad_norm": 0.10093820387387048, "learning_rate": 1.3011188632212307e-05, "loss": 0.6033, "step": 564 }, { "epoch": 3.8175675675675675, "grad_norm": 0.0890166406955722, "learning_rate": 1.2872226567413346e-05, "loss": 0.5924, "step": 565 }, { "epoch": 3.8243243243243246, "grad_norm": 0.07708649898799794, "learning_rate": 1.2733868124018694e-05, "loss": 0.5891, "step": 566 }, { "epoch": 3.831081081081081, "grad_norm": 0.11011562257155211, "learning_rate": 1.2596116380648761e-05, "loss": 0.5898, "step": 567 }, { "epoch": 3.8378378378378377, "grad_norm": 0.07996964493884025, "learning_rate": 1.2458974402424312e-05, "loss": 0.5937, "step": 568 }, { "epoch": 3.8445945945945947, "grad_norm": 0.09078629614798739, "learning_rate": 1.2322445240898158e-05, "loss": 0.5908, "step": 569 }, { "epoch": 3.8513513513513513, "grad_norm": 0.08453528616636899, "learning_rate": 1.2186531933987294e-05, "loss": 0.5895, "step": 570 }, { "epoch": 3.858108108108108, "grad_norm": 0.08207256114306587, "learning_rate": 1.2051237505905302e-05, "loss": 0.6, "step": 571 }, { "epoch": 3.864864864864865, "grad_norm": 0.08328736066936966, "learning_rate": 1.19165649670951e-05, "loss": 0.5901, "step": 572 }, { "epoch": 3.8716216216216215, "grad_norm": 0.0796730356081974, "learning_rate": 1.1782517314161872e-05, "loss": 0.5825, "step": 573 }, { "epoch": 3.8783783783783785, "grad_norm": 0.09614075098109777, "learning_rate": 1.164909752980648e-05, "loss": 0.5949, "step": 574 }, { "epoch": 3.885135135135135, "grad_norm": 0.08310475221153353, "learning_rate": 1.1516308582758983e-05, "loss": 0.604, "step": 575 }, { "epoch": 3.891891891891892, "grad_norm": 0.08282668808548053, "learning_rate": 1.1384153427712729e-05, "loss": 0.6002, "step": 576 }, { "epoch": 3.8986486486486487, "grad_norm": 0.0986705467432139, "learning_rate": 1.1252635005258466e-05, "loss": 0.5853, "step": 577 }, { "epoch": 3.9054054054054053, "grad_norm": 0.08062037787963591, "learning_rate": 1.1121756241819023e-05, "loss": 0.6047, "step": 578 }, { "epoch": 3.9121621621621623, "grad_norm": 0.07523397720678549, "learning_rate": 1.0991520049584112e-05, "loss": 0.5842, "step": 579 }, { "epoch": 3.918918918918919, "grad_norm": 0.08533444706267539, "learning_rate": 1.0861929326445572e-05, "loss": 0.5952, "step": 580 }, { "epoch": 3.9256756756756754, "grad_norm": 0.0867340533978853, "learning_rate": 1.0732986955932869e-05, "loss": 0.5978, "step": 581 }, { "epoch": 3.9324324324324325, "grad_norm": 0.06754168833386048, "learning_rate": 1.0604695807148971e-05, "loss": 0.5987, "step": 582 }, { "epoch": 3.939189189189189, "grad_norm": 0.0796662865647217, "learning_rate": 1.0477058734706436e-05, "loss": 0.588, "step": 583 }, { "epoch": 3.945945945945946, "grad_norm": 0.0813837559182795, "learning_rate": 1.0350078578664005e-05, "loss": 0.6009, "step": 584 }, { "epoch": 3.9527027027027026, "grad_norm": 0.06912060191449508, "learning_rate": 1.0223758164463246e-05, "loss": 0.6055, "step": 585 }, { "epoch": 3.9594594594594597, "grad_norm": 0.06784642823506538, "learning_rate": 1.0098100302865865e-05, "loss": 0.5882, "step": 586 }, { "epoch": 3.9662162162162162, "grad_norm": 0.0760135339338161, "learning_rate": 9.973107789891024e-06, "loss": 0.596, "step": 587 }, { "epoch": 3.972972972972973, "grad_norm": 0.07657252850751217, "learning_rate": 9.848783406753224e-06, "loss": 0.5962, "step": 588 }, { "epoch": 3.97972972972973, "grad_norm": 0.08050666474861323, "learning_rate": 9.725129919800339e-06, "loss": 0.5957, "step": 589 }, { "epoch": 3.9864864864864864, "grad_norm": 0.0778632825822388, "learning_rate": 9.60215008045211e-06, "loss": 0.5952, "step": 590 }, { "epoch": 3.993243243243243, "grad_norm": 0.07984922923288139, "learning_rate": 9.479846625138909e-06, "loss": 0.6023, "step": 591 }, { "epoch": 4.0, "grad_norm": 0.0916488655332951, "learning_rate": 9.358222275240884e-06, "loss": 0.5842, "step": 592 }, { "epoch": 4.006756756756757, "grad_norm": 0.08766439343096606, "learning_rate": 9.237279737027326e-06, "loss": 0.5851, "step": 593 }, { "epoch": 4.013513513513513, "grad_norm": 0.07964740730111751, "learning_rate": 9.117021701596567e-06, "loss": 0.5776, "step": 594 }, { "epoch": 4.02027027027027, "grad_norm": 0.08933138991359177, "learning_rate": 8.99745084481594e-06, "loss": 0.5953, "step": 595 }, { "epoch": 4.027027027027027, "grad_norm": 0.07914528740585995, "learning_rate": 8.87856982726243e-06, "loss": 0.5892, "step": 596 }, { "epoch": 4.033783783783784, "grad_norm": 0.07297723048835614, "learning_rate": 8.7603812941633e-06, "loss": 0.5823, "step": 597 }, { "epoch": 4.04054054054054, "grad_norm": 0.10229836912930287, "learning_rate": 8.642887875337376e-06, "loss": 0.5767, "step": 598 }, { "epoch": 4.047297297297297, "grad_norm": 0.0781799178126612, "learning_rate": 8.526092185136394e-06, "loss": 0.5784, "step": 599 }, { "epoch": 4.054054054054054, "grad_norm": 0.09402345888004014, "learning_rate": 8.409996822386972e-06, "loss": 0.5828, "step": 600 }, { "epoch": 4.0608108108108105, "grad_norm": 0.07679009840525959, "learning_rate": 8.294604370332613e-06, "loss": 0.5754, "step": 601 }, { "epoch": 4.0675675675675675, "grad_norm": 0.07923822298946714, "learning_rate": 8.17991739657641e-06, "loss": 0.573, "step": 602 }, { "epoch": 4.074324324324325, "grad_norm": 0.08082893460744046, "learning_rate": 8.06593845302376e-06, "loss": 0.5916, "step": 603 }, { "epoch": 4.081081081081081, "grad_norm": 0.07578230123409459, "learning_rate": 7.952670075825702e-06, "loss": 0.5851, "step": 604 }, { "epoch": 4.087837837837838, "grad_norm": 0.07539180401042113, "learning_rate": 7.840114785322384e-06, "loss": 0.5939, "step": 605 }, { "epoch": 4.094594594594595, "grad_norm": 0.07932706141071251, "learning_rate": 7.728275085987041e-06, "loss": 0.5908, "step": 606 }, { "epoch": 4.101351351351352, "grad_norm": 0.07344646624670495, "learning_rate": 7.6171534663702416e-06, "loss": 0.5915, "step": 607 }, { "epoch": 4.108108108108108, "grad_norm": 0.07204192586718106, "learning_rate": 7.50675239904457e-06, "loss": 0.5855, "step": 608 }, { "epoch": 4.114864864864865, "grad_norm": 0.06717242761020562, "learning_rate": 7.397074340549508e-06, "loss": 0.5919, "step": 609 }, { "epoch": 4.121621621621622, "grad_norm": 0.07615198090948691, "learning_rate": 7.288121731336901e-06, "loss": 0.5883, "step": 610 }, { "epoch": 4.128378378378378, "grad_norm": 0.07186386521547546, "learning_rate": 7.1798969957165025e-06, "loss": 0.5666, "step": 611 }, { "epoch": 4.135135135135135, "grad_norm": 0.07771245688676058, "learning_rate": 7.072402541802197e-06, "loss": 0.572, "step": 612 }, { "epoch": 4.141891891891892, "grad_norm": 0.06560942291485741, "learning_rate": 6.965640761458274e-06, "loss": 0.5853, "step": 613 }, { "epoch": 4.148648648648648, "grad_norm": 0.07554536141277478, "learning_rate": 6.859614030246318e-06, "loss": 0.5746, "step": 614 }, { "epoch": 4.155405405405405, "grad_norm": 0.0766866599574075, "learning_rate": 6.754324707372264e-06, "loss": 0.5827, "step": 615 }, { "epoch": 4.162162162162162, "grad_norm": 0.0685589990578116, "learning_rate": 6.649775135633944e-06, "loss": 0.5708, "step": 616 }, { "epoch": 4.168918918918919, "grad_norm": 0.06830496975914782, "learning_rate": 6.545967641368958e-06, "loss": 0.5772, "step": 617 }, { "epoch": 4.175675675675675, "grad_norm": 0.06792930742316021, "learning_rate": 6.4429045344029136e-06, "loss": 0.5801, "step": 618 }, { "epoch": 4.1824324324324325, "grad_norm": 0.0728029236733438, "learning_rate": 6.340588107997994e-06, "loss": 0.5666, "step": 619 }, { "epoch": 4.1891891891891895, "grad_norm": 0.07156462363762543, "learning_rate": 6.239020638801987e-06, "loss": 0.5848, "step": 620 }, { "epoch": 4.195945945945946, "grad_norm": 0.06757664374848818, "learning_rate": 6.1382043867975836e-06, "loss": 0.5679, "step": 621 }, { "epoch": 4.202702702702703, "grad_norm": 0.07016074613879056, "learning_rate": 6.038141595252094e-06, "loss": 0.5839, "step": 622 }, { "epoch": 4.20945945945946, "grad_norm": 0.06809385798805276, "learning_rate": 5.9388344906675485e-06, "loss": 0.575, "step": 623 }, { "epoch": 4.216216216216216, "grad_norm": 0.061129888925191694, "learning_rate": 5.840285282731173e-06, "loss": 0.5914, "step": 624 }, { "epoch": 4.222972972972973, "grad_norm": 0.0728841561672064, "learning_rate": 5.742496164266174e-06, "loss": 0.578, "step": 625 }, { "epoch": 4.22972972972973, "grad_norm": 0.07404139765077528, "learning_rate": 5.645469311182958e-06, "loss": 0.5788, "step": 626 }, { "epoch": 4.236486486486487, "grad_norm": 0.06599738371244251, "learning_rate": 5.549206882430773e-06, "loss": 0.5938, "step": 627 }, { "epoch": 4.243243243243243, "grad_norm": 0.07571258452222435, "learning_rate": 5.453711019949581e-06, "loss": 0.5761, "step": 628 }, { "epoch": 4.25, "grad_norm": 0.06873039122654918, "learning_rate": 5.358983848622452e-06, "loss": 0.5865, "step": 629 }, { "epoch": 4.256756756756757, "grad_norm": 0.05980111880094078, "learning_rate": 5.265027476228297e-06, "loss": 0.58, "step": 630 }, { "epoch": 4.263513513513513, "grad_norm": 0.07210626853825096, "learning_rate": 5.171843993394903e-06, "loss": 0.5756, "step": 631 }, { "epoch": 4.27027027027027, "grad_norm": 0.07396976700448737, "learning_rate": 5.079435473552474e-06, "loss": 0.5811, "step": 632 }, { "epoch": 4.277027027027027, "grad_norm": 0.09171081358728068, "learning_rate": 4.987803972887482e-06, "loss": 0.5871, "step": 633 }, { "epoch": 4.283783783783784, "grad_norm": 0.06701824447243687, "learning_rate": 4.896951530296896e-06, "loss": 0.579, "step": 634 }, { "epoch": 4.29054054054054, "grad_norm": 0.07518459749761504, "learning_rate": 4.806880167342831e-06, "loss": 0.5822, "step": 635 }, { "epoch": 4.297297297297297, "grad_norm": 0.07098354334790964, "learning_rate": 4.7175918882075465e-06, "loss": 0.5861, "step": 636 }, { "epoch": 4.304054054054054, "grad_norm": 0.07479138842797883, "learning_rate": 4.6290886796488946e-06, "loss": 0.5863, "step": 637 }, { "epoch": 4.3108108108108105, "grad_norm": 0.06686921193950773, "learning_rate": 4.54137251095605e-06, "loss": 0.5916, "step": 638 }, { "epoch": 4.3175675675675675, "grad_norm": 0.07897960962457094, "learning_rate": 4.454445333905768e-06, "loss": 0.5948, "step": 639 }, { "epoch": 4.324324324324325, "grad_norm": 0.07496951321971122, "learning_rate": 4.3683090827188666e-06, "loss": 0.5813, "step": 640 }, { "epoch": 4.331081081081081, "grad_norm": 0.07536718259174548, "learning_rate": 4.282965674017265e-06, "loss": 0.5872, "step": 641 }, { "epoch": 4.337837837837838, "grad_norm": 0.06854859712850378, "learning_rate": 4.198417006781283e-06, "loss": 0.5841, "step": 642 }, { "epoch": 4.344594594594595, "grad_norm": 0.07618804161631497, "learning_rate": 4.114664962307439e-06, "loss": 0.5817, "step": 643 }, { "epoch": 4.351351351351352, "grad_norm": 0.07601070844404975, "learning_rate": 4.031711404166525e-06, "loss": 0.584, "step": 644 }, { "epoch": 4.358108108108108, "grad_norm": 0.06914624176338743, "learning_rate": 3.949558178162209e-06, "loss": 0.5888, "step": 645 }, { "epoch": 4.364864864864865, "grad_norm": 0.07027655538494314, "learning_rate": 3.86820711228991e-06, "loss": 0.5867, "step": 646 }, { "epoch": 4.371621621621622, "grad_norm": 0.0760270326599887, "learning_rate": 3.7876600166961353e-06, "loss": 0.5961, "step": 647 }, { "epoch": 4.378378378378378, "grad_norm": 0.06782313812781374, "learning_rate": 3.707918683638223e-06, "loss": 0.5944, "step": 648 }, { "epoch": 4.385135135135135, "grad_norm": 0.06192386606423058, "learning_rate": 3.628984887444462e-06, "loss": 0.5969, "step": 649 }, { "epoch": 4.391891891891892, "grad_norm": 0.062017594322198465, "learning_rate": 3.550860384474568e-06, "loss": 0.5863, "step": 650 }, { "epoch": 4.398648648648648, "grad_norm": 0.058946604304473224, "learning_rate": 3.473546913080674e-06, "loss": 0.5847, "step": 651 }, { "epoch": 4.405405405405405, "grad_norm": 0.08803265906901768, "learning_rate": 3.397046193568558e-06, "loss": 0.5707, "step": 652 }, { "epoch": 4.412162162162162, "grad_norm": 0.06583037908170926, "learning_rate": 3.3213599281594688e-06, "loss": 0.582, "step": 653 }, { "epoch": 4.418918918918919, "grad_norm": 0.06086938296974132, "learning_rate": 3.246489800952155e-06, "loss": 0.5953, "step": 654 }, { "epoch": 4.425675675675675, "grad_norm": 0.05932401857518893, "learning_rate": 3.172437477885475e-06, "loss": 0.5819, "step": 655 }, { "epoch": 4.4324324324324325, "grad_norm": 0.057752799006265246, "learning_rate": 3.099204606701256e-06, "loss": 0.5766, "step": 656 }, { "epoch": 4.4391891891891895, "grad_norm": 0.06397066535508446, "learning_rate": 3.026792816907671e-06, "loss": 0.588, "step": 657 }, { "epoch": 4.445945945945946, "grad_norm": 0.06029045959058417, "learning_rate": 2.955203719742965e-06, "loss": 0.591, "step": 658 }, { "epoch": 4.452702702702703, "grad_norm": 0.05838249627908107, "learning_rate": 2.884438908139626e-06, "loss": 0.5929, "step": 659 }, { "epoch": 4.45945945945946, "grad_norm": 0.06075718886251464, "learning_rate": 2.814499956688912e-06, "loss": 0.5813, "step": 660 }, { "epoch": 4.466216216216216, "grad_norm": 0.0719277704750579, "learning_rate": 2.7453884216058368e-06, "loss": 0.5795, "step": 661 }, { "epoch": 4.472972972972973, "grad_norm": 0.06075719822797747, "learning_rate": 2.677105840694507e-06, "loss": 0.5674, "step": 662 }, { "epoch": 4.47972972972973, "grad_norm": 0.05975207139447743, "learning_rate": 2.6096537333139616e-06, "loss": 0.5805, "step": 663 }, { "epoch": 4.486486486486487, "grad_norm": 0.06213586333263577, "learning_rate": 2.5430336003443045e-06, "loss": 0.5906, "step": 664 }, { "epoch": 4.493243243243243, "grad_norm": 0.05895116376323496, "learning_rate": 2.4772469241533648e-06, "loss": 0.5844, "step": 665 }, { "epoch": 4.5, "grad_norm": 0.06765046020471, "learning_rate": 2.4122951685636674e-06, "loss": 0.5835, "step": 666 }, { "epoch": 4.506756756756757, "grad_norm": 0.06650006977809148, "learning_rate": 2.3481797788198745e-06, "loss": 0.5789, "step": 667 }, { "epoch": 4.513513513513513, "grad_norm": 0.05846137750044436, "learning_rate": 2.284902181556632e-06, "loss": 0.5745, "step": 668 }, { "epoch": 4.52027027027027, "grad_norm": 0.05989337419115531, "learning_rate": 2.2224637847668484e-06, "loss": 0.5733, "step": 669 }, { "epoch": 4.527027027027027, "grad_norm": 0.06939319651650724, "learning_rate": 2.1608659777703033e-06, "loss": 0.581, "step": 670 }, { "epoch": 4.533783783783784, "grad_norm": 0.054679826375991857, "learning_rate": 2.100110131182813e-06, "loss": 0.5931, "step": 671 }, { "epoch": 4.54054054054054, "grad_norm": 0.05548818045052255, "learning_rate": 2.0401975968856514e-06, "loss": 0.5786, "step": 672 }, { "epoch": 4.547297297297297, "grad_norm": 0.059389572689147435, "learning_rate": 1.981129707995542e-06, "loss": 0.5923, "step": 673 }, { "epoch": 4.554054054054054, "grad_norm": 0.06975826873383643, "learning_rate": 1.9229077788349393e-06, "loss": 0.577, "step": 674 }, { "epoch": 4.5608108108108105, "grad_norm": 0.07220930223960455, "learning_rate": 1.865533104902828e-06, "loss": 0.5916, "step": 675 }, { "epoch": 4.5675675675675675, "grad_norm": 0.053811020975530956, "learning_rate": 1.8090069628458583e-06, "loss": 0.5704, "step": 676 }, { "epoch": 4.574324324324325, "grad_norm": 0.05538660298263909, "learning_rate": 1.7533306104299663e-06, "loss": 0.5856, "step": 677 }, { "epoch": 4.581081081081081, "grad_norm": 0.05829509686058199, "learning_rate": 1.6985052865123641e-06, "loss": 0.5754, "step": 678 }, { "epoch": 4.587837837837838, "grad_norm": 0.06323909786498662, "learning_rate": 1.6445322110140116e-06, "loss": 0.5814, "step": 679 }, { "epoch": 4.594594594594595, "grad_norm": 0.056972721072935675, "learning_rate": 1.59141258489242e-06, "loss": 0.5756, "step": 680 }, { "epoch": 4.601351351351351, "grad_norm": 0.05816167859998497, "learning_rate": 1.53914759011498e-06, "loss": 0.5889, "step": 681 }, { "epoch": 4.608108108108108, "grad_norm": 0.059044794864098894, "learning_rate": 1.4877383896326269e-06, "loss": 0.5781, "step": 682 }, { "epoch": 4.614864864864865, "grad_norm": 0.06237941099681927, "learning_rate": 1.4371861273539778e-06, "loss": 0.576, "step": 683 }, { "epoch": 4.621621621621622, "grad_norm": 0.06531112971211704, "learning_rate": 1.3874919281198662e-06, "loss": 0.5799, "step": 684 }, { "epoch": 4.628378378378378, "grad_norm": 0.05265860386748638, "learning_rate": 1.3386568976783453e-06, "loss": 0.5891, "step": 685 }, { "epoch": 4.635135135135135, "grad_norm": 0.05785779557411594, "learning_rate": 1.2906821226600453e-06, "loss": 0.5768, "step": 686 }, { "epoch": 4.641891891891892, "grad_norm": 0.05639941023635892, "learning_rate": 1.2435686705540228e-06, "loss": 0.5842, "step": 687 }, { "epoch": 4.648648648648649, "grad_norm": 0.06338285927327728, "learning_rate": 1.1973175896839684e-06, "loss": 0.573, "step": 688 }, { "epoch": 4.655405405405405, "grad_norm": 0.054671739680034734, "learning_rate": 1.1519299091849523e-06, "loss": 0.5774, "step": 689 }, { "epoch": 4.662162162162162, "grad_norm": 0.057698821845826635, "learning_rate": 1.1074066389804395e-06, "loss": 0.5932, "step": 690 }, { "epoch": 4.668918918918919, "grad_norm": 0.05697783615575523, "learning_rate": 1.0637487697598937e-06, "loss": 0.5823, "step": 691 }, { "epoch": 4.675675675675675, "grad_norm": 0.054051563368746565, "learning_rate": 1.0209572729566708e-06, "loss": 0.5787, "step": 692 }, { "epoch": 4.6824324324324325, "grad_norm": 0.05344229805841218, "learning_rate": 9.790331007264543e-07, "loss": 0.5844, "step": 693 }, { "epoch": 4.6891891891891895, "grad_norm": 0.06959576323504683, "learning_rate": 9.379771859260267e-07, "loss": 0.5817, "step": 694 }, { "epoch": 4.695945945945946, "grad_norm": 0.06318026825084985, "learning_rate": 8.977904420925543e-07, "loss": 0.5892, "step": 695 }, { "epoch": 4.702702702702703, "grad_norm": 0.056039993579189235, "learning_rate": 8.584737634232154e-07, "loss": 0.5835, "step": 696 }, { "epoch": 4.70945945945946, "grad_norm": 0.05269309541455996, "learning_rate": 8.200280247553461e-07, "loss": 0.5803, "step": 697 }, { "epoch": 4.716216216216216, "grad_norm": 0.05510521164357771, "learning_rate": 7.824540815469306e-07, "loss": 0.5709, "step": 698 }, { "epoch": 4.722972972972973, "grad_norm": 0.05770652073739158, "learning_rate": 7.457527698576217e-07, "loss": 0.5855, "step": 699 }, { "epoch": 4.72972972972973, "grad_norm": 0.055593016439349656, "learning_rate": 7.099249063300751e-07, "loss": 0.5847, "step": 700 }, { "epoch": 4.736486486486487, "grad_norm": 0.05730248759901012, "learning_rate": 6.749712881718306e-07, "loss": 0.5639, "step": 701 }, { "epoch": 4.743243243243243, "grad_norm": 0.05973974788088902, "learning_rate": 6.408926931375403e-07, "loss": 0.5808, "step": 702 }, { "epoch": 4.75, "grad_norm": 0.05557100230846073, "learning_rate": 6.076898795116792e-07, "loss": 0.5848, "step": 703 }, { "epoch": 4.756756756756757, "grad_norm": 0.0563081652270169, "learning_rate": 5.753635860916617e-07, "loss": 0.5888, "step": 704 }, { "epoch": 4.763513513513513, "grad_norm": 0.0536729552491486, "learning_rate": 5.43914532171419e-07, "loss": 0.5873, "step": 705 }, { "epoch": 4.77027027027027, "grad_norm": 0.049916139412812977, "learning_rate": 5.133434175253715e-07, "loss": 0.5792, "step": 706 }, { "epoch": 4.777027027027027, "grad_norm": 0.05386272187182479, "learning_rate": 4.83650922392882e-07, "loss": 0.5829, "step": 707 }, { "epoch": 4.783783783783784, "grad_norm": 0.05503179443312982, "learning_rate": 4.5483770746309383e-07, "loss": 0.5873, "step": 708 }, { "epoch": 4.79054054054054, "grad_norm": 0.05614293253442416, "learning_rate": 4.269044138602585e-07, "loss": 0.5879, "step": 709 }, { "epoch": 4.797297297297297, "grad_norm": 0.053484276503556256, "learning_rate": 3.998516631294491e-07, "loss": 0.5794, "step": 710 }, { "epoch": 4.804054054054054, "grad_norm": 0.055818505320336, "learning_rate": 3.736800572227317e-07, "loss": 0.5905, "step": 711 }, { "epoch": 4.8108108108108105, "grad_norm": 0.05644903653041531, "learning_rate": 3.483901784857846e-07, "loss": 0.5721, "step": 712 }, { "epoch": 4.8175675675675675, "grad_norm": 0.04844049723060667, "learning_rate": 3.239825896449267e-07, "loss": 0.5826, "step": 713 }, { "epoch": 4.824324324324325, "grad_norm": 0.053088243917807275, "learning_rate": 3.004578337945985e-07, "loss": 0.5811, "step": 714 }, { "epoch": 4.831081081081081, "grad_norm": 0.05140464370774195, "learning_rate": 2.778164343852918e-07, "loss": 0.5791, "step": 715 }, { "epoch": 4.837837837837838, "grad_norm": 0.05679451911188569, "learning_rate": 2.5605889521188364e-07, "loss": 0.5787, "step": 716 }, { "epoch": 4.844594594594595, "grad_norm": 0.06082278783885325, "learning_rate": 2.351857004024316e-07, "loss": 0.5868, "step": 717 }, { "epoch": 4.851351351351351, "grad_norm": 0.05433042972998646, "learning_rate": 2.1519731440740487e-07, "loss": 0.5812, "step": 718 }, { "epoch": 4.858108108108108, "grad_norm": 0.055224715989184006, "learning_rate": 1.9609418198935916e-07, "loss": 0.5881, "step": 719 }, { "epoch": 4.864864864864865, "grad_norm": 0.05348313369078817, "learning_rate": 1.778767282130156e-07, "loss": 0.5945, "step": 720 }, { "epoch": 4.871621621621622, "grad_norm": 0.05163946442780112, "learning_rate": 1.6054535843582854e-07, "loss": 0.5891, "step": 721 }, { "epoch": 4.878378378378378, "grad_norm": 0.055008800596469655, "learning_rate": 1.4410045829893915e-07, "loss": 0.5817, "step": 722 }, { "epoch": 4.885135135135135, "grad_norm": 0.05385755694710143, "learning_rate": 1.2854239371863142e-07, "loss": 0.5803, "step": 723 }, { "epoch": 4.891891891891892, "grad_norm": 0.050256418798137074, "learning_rate": 1.1387151087814297e-07, "loss": 0.5776, "step": 724 }, { "epoch": 4.898648648648649, "grad_norm": 0.06422788348062423, "learning_rate": 1.0008813622001345e-07, "loss": 0.5865, "step": 725 }, { "epoch": 4.905405405405405, "grad_norm": 0.058912661052812065, "learning_rate": 8.719257643877044e-08, "loss": 0.5904, "step": 726 }, { "epoch": 4.912162162162162, "grad_norm": 0.05385241384320565, "learning_rate": 7.51851184741481e-08, "loss": 0.5944, "step": 727 }, { "epoch": 4.918918918918919, "grad_norm": 0.05443985642845785, "learning_rate": 6.40660295046791e-08, "loss": 0.5857, "step": 728 }, { "epoch": 4.925675675675675, "grad_norm": 0.05281227387100857, "learning_rate": 5.3835556941743695e-08, "loss": 0.5742, "step": 729 }, { "epoch": 4.9324324324324325, "grad_norm": 0.05531460046068238, "learning_rate": 4.449392842408529e-08, "loss": 0.5743, "step": 730 }, { "epoch": 4.9391891891891895, "grad_norm": 0.05538814823209324, "learning_rate": 3.6041351812743374e-08, "loss": 0.5877, "step": 731 }, { "epoch": 4.945945945945946, "grad_norm": 0.055186199974581227, "learning_rate": 2.8478015186399477e-08, "loss": 0.5842, "step": 732 }, { "epoch": 4.952702702702703, "grad_norm": 0.055151710910165075, "learning_rate": 2.1804086837229344e-08, "loss": 0.5846, "step": 733 }, { "epoch": 4.95945945945946, "grad_norm": 0.05558334617444456, "learning_rate": 1.601971526713708e-08, "loss": 0.5766, "step": 734 }, { "epoch": 4.966216216216216, "grad_norm": 0.049437407136724947, "learning_rate": 1.112502918445113e-08, "loss": 0.586, "step": 735 }, { "epoch": 4.972972972972973, "grad_norm": 0.05838258812809817, "learning_rate": 7.12013750107321e-09, "loss": 0.5918, "step": 736 }, { "epoch": 4.97972972972973, "grad_norm": 0.05252990242194222, "learning_rate": 4.00512933004471e-09, "loss": 0.5853, "step": 737 }, { "epoch": 4.986486486486487, "grad_norm": 0.05539911735758133, "learning_rate": 1.7800739835616143e-09, "loss": 0.5731, "step": 738 }, { "epoch": 4.993243243243243, "grad_norm": 0.059325115613032664, "learning_rate": 4.450209714379483e-10, "loss": 0.5814, "step": 739 }, { "epoch": 5.0, "grad_norm": 0.059385070432276595, "learning_rate": 0.0, "loss": 0.577, "step": 740 }, { "epoch": 5.0, "step": 740, "total_flos": 4.168626538020864e+16, "train_loss": 0.6272868978010642, "train_runtime": 42334.4318, "train_samples_per_second": 8.913, "train_steps_per_second": 0.017 } ], "logging_steps": 1, "max_steps": 740, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.168626538020864e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }