kodcode_32B / trainer_state.json
marianna13's picture
Upload folder using huggingface_hub
772c863 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 5.0,
"eval_steps": 500,
"global_step": 740,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006756756756756757,
"grad_norm": 3.1352566768118555,
"learning_rate": 1.0810810810810812e-06,
"loss": 1.0885,
"step": 1
},
{
"epoch": 0.013513513513513514,
"grad_norm": 3.0707759310558114,
"learning_rate": 2.1621621621621623e-06,
"loss": 1.081,
"step": 2
},
{
"epoch": 0.02027027027027027,
"grad_norm": 3.1041877911160487,
"learning_rate": 3.2432432432432437e-06,
"loss": 1.0779,
"step": 3
},
{
"epoch": 0.02702702702702703,
"grad_norm": 2.894341433491889,
"learning_rate": 4.324324324324325e-06,
"loss": 1.066,
"step": 4
},
{
"epoch": 0.033783783783783786,
"grad_norm": 2.4474574648074654,
"learning_rate": 5.405405405405406e-06,
"loss": 1.0464,
"step": 5
},
{
"epoch": 0.04054054054054054,
"grad_norm": 1.7416847938696791,
"learning_rate": 6.486486486486487e-06,
"loss": 1.0062,
"step": 6
},
{
"epoch": 0.0472972972972973,
"grad_norm": 1.737012288070095,
"learning_rate": 7.567567567567569e-06,
"loss": 0.9661,
"step": 7
},
{
"epoch": 0.05405405405405406,
"grad_norm": 1.54166088046008,
"learning_rate": 8.64864864864865e-06,
"loss": 0.9625,
"step": 8
},
{
"epoch": 0.060810810810810814,
"grad_norm": 1.1305140237848132,
"learning_rate": 9.729729729729732e-06,
"loss": 0.9355,
"step": 9
},
{
"epoch": 0.06756756756756757,
"grad_norm": 1.6829884643585495,
"learning_rate": 1.0810810810810812e-05,
"loss": 0.8814,
"step": 10
},
{
"epoch": 0.07432432432432433,
"grad_norm": 1.2528889517867892,
"learning_rate": 1.1891891891891894e-05,
"loss": 0.8733,
"step": 11
},
{
"epoch": 0.08108108108108109,
"grad_norm": 0.9520381902650208,
"learning_rate": 1.2972972972972975e-05,
"loss": 0.8579,
"step": 12
},
{
"epoch": 0.08783783783783784,
"grad_norm": 1.4164792552755203,
"learning_rate": 1.4054054054054055e-05,
"loss": 0.82,
"step": 13
},
{
"epoch": 0.0945945945945946,
"grad_norm": 1.1661768167320599,
"learning_rate": 1.5135135135135138e-05,
"loss": 0.823,
"step": 14
},
{
"epoch": 0.10135135135135136,
"grad_norm": 0.966545525495581,
"learning_rate": 1.6216216216216218e-05,
"loss": 0.8177,
"step": 15
},
{
"epoch": 0.10810810810810811,
"grad_norm": 0.9970915856247566,
"learning_rate": 1.72972972972973e-05,
"loss": 0.7828,
"step": 16
},
{
"epoch": 0.11486486486486487,
"grad_norm": 0.7510807795940354,
"learning_rate": 1.8378378378378383e-05,
"loss": 0.7752,
"step": 17
},
{
"epoch": 0.12162162162162163,
"grad_norm": 0.6801364883047033,
"learning_rate": 1.9459459459459463e-05,
"loss": 0.7802,
"step": 18
},
{
"epoch": 0.12837837837837837,
"grad_norm": 0.8052966224978529,
"learning_rate": 2.054054054054054e-05,
"loss": 0.7803,
"step": 19
},
{
"epoch": 0.13513513513513514,
"grad_norm": 0.6731724951990521,
"learning_rate": 2.1621621621621624e-05,
"loss": 0.774,
"step": 20
},
{
"epoch": 0.14189189189189189,
"grad_norm": 0.5240064900372621,
"learning_rate": 2.2702702702702705e-05,
"loss": 0.7503,
"step": 21
},
{
"epoch": 0.14864864864864866,
"grad_norm": 0.5238422824665931,
"learning_rate": 2.378378378378379e-05,
"loss": 0.7452,
"step": 22
},
{
"epoch": 0.1554054054054054,
"grad_norm": 0.5685661395814289,
"learning_rate": 2.4864864864864866e-05,
"loss": 0.7357,
"step": 23
},
{
"epoch": 0.16216216216216217,
"grad_norm": 0.5559414871229672,
"learning_rate": 2.594594594594595e-05,
"loss": 0.7529,
"step": 24
},
{
"epoch": 0.16891891891891891,
"grad_norm": 0.44354381932505604,
"learning_rate": 2.702702702702703e-05,
"loss": 0.7531,
"step": 25
},
{
"epoch": 0.17567567567567569,
"grad_norm": 0.5353496542990213,
"learning_rate": 2.810810810810811e-05,
"loss": 0.7443,
"step": 26
},
{
"epoch": 0.18243243243243243,
"grad_norm": 0.5011572381597165,
"learning_rate": 2.918918918918919e-05,
"loss": 0.7309,
"step": 27
},
{
"epoch": 0.1891891891891892,
"grad_norm": 0.46554848262029747,
"learning_rate": 3.0270270270270275e-05,
"loss": 0.73,
"step": 28
},
{
"epoch": 0.19594594594594594,
"grad_norm": 0.5645942632228904,
"learning_rate": 3.135135135135135e-05,
"loss": 0.7218,
"step": 29
},
{
"epoch": 0.20270270270270271,
"grad_norm": 0.48122141248903777,
"learning_rate": 3.2432432432432436e-05,
"loss": 0.7272,
"step": 30
},
{
"epoch": 0.20945945945945946,
"grad_norm": 0.4302278552572059,
"learning_rate": 3.351351351351351e-05,
"loss": 0.7138,
"step": 31
},
{
"epoch": 0.21621621621621623,
"grad_norm": 0.38359657484495446,
"learning_rate": 3.45945945945946e-05,
"loss": 0.716,
"step": 32
},
{
"epoch": 0.22297297297297297,
"grad_norm": 0.37734924853300067,
"learning_rate": 3.567567567567568e-05,
"loss": 0.7258,
"step": 33
},
{
"epoch": 0.22972972972972974,
"grad_norm": 0.5860873173612458,
"learning_rate": 3.6756756756756765e-05,
"loss": 0.7146,
"step": 34
},
{
"epoch": 0.23648648648648649,
"grad_norm": 0.7521759711776034,
"learning_rate": 3.783783783783784e-05,
"loss": 0.727,
"step": 35
},
{
"epoch": 0.24324324324324326,
"grad_norm": 0.46287517636393716,
"learning_rate": 3.8918918918918926e-05,
"loss": 0.6978,
"step": 36
},
{
"epoch": 0.25,
"grad_norm": 0.3949073220604405,
"learning_rate": 4e-05,
"loss": 0.7143,
"step": 37
},
{
"epoch": 0.25675675675675674,
"grad_norm": 0.5390567372094507,
"learning_rate": 4.108108108108108e-05,
"loss": 0.7147,
"step": 38
},
{
"epoch": 0.2635135135135135,
"grad_norm": 0.5260020316144047,
"learning_rate": 4.2162162162162164e-05,
"loss": 0.7042,
"step": 39
},
{
"epoch": 0.2702702702702703,
"grad_norm": 0.6245421084330798,
"learning_rate": 4.324324324324325e-05,
"loss": 0.7066,
"step": 40
},
{
"epoch": 0.27702702702702703,
"grad_norm": 0.4274410936399283,
"learning_rate": 4.4324324324324325e-05,
"loss": 0.7002,
"step": 41
},
{
"epoch": 0.28378378378378377,
"grad_norm": 0.5645116869141522,
"learning_rate": 4.540540540540541e-05,
"loss": 0.7017,
"step": 42
},
{
"epoch": 0.2905405405405405,
"grad_norm": 0.8531985478315304,
"learning_rate": 4.6486486486486486e-05,
"loss": 0.7105,
"step": 43
},
{
"epoch": 0.2972972972972973,
"grad_norm": 0.8463718591267281,
"learning_rate": 4.756756756756758e-05,
"loss": 0.6926,
"step": 44
},
{
"epoch": 0.30405405405405406,
"grad_norm": 0.527363152011757,
"learning_rate": 4.8648648648648654e-05,
"loss": 0.698,
"step": 45
},
{
"epoch": 0.3108108108108108,
"grad_norm": 0.8921529578247511,
"learning_rate": 4.972972972972973e-05,
"loss": 0.7113,
"step": 46
},
{
"epoch": 0.31756756756756754,
"grad_norm": 0.844585341198514,
"learning_rate": 5.081081081081081e-05,
"loss": 0.6918,
"step": 47
},
{
"epoch": 0.32432432432432434,
"grad_norm": 0.6610922858373763,
"learning_rate": 5.18918918918919e-05,
"loss": 0.6874,
"step": 48
},
{
"epoch": 0.3310810810810811,
"grad_norm": 0.9289696025390548,
"learning_rate": 5.2972972972972976e-05,
"loss": 0.6965,
"step": 49
},
{
"epoch": 0.33783783783783783,
"grad_norm": 0.7104021291109865,
"learning_rate": 5.405405405405406e-05,
"loss": 0.7025,
"step": 50
},
{
"epoch": 0.34459459459459457,
"grad_norm": 0.7179610269369078,
"learning_rate": 5.513513513513514e-05,
"loss": 0.6906,
"step": 51
},
{
"epoch": 0.35135135135135137,
"grad_norm": 0.6612006329248038,
"learning_rate": 5.621621621621622e-05,
"loss": 0.7026,
"step": 52
},
{
"epoch": 0.3581081081081081,
"grad_norm": 0.6920205226216654,
"learning_rate": 5.7297297297297305e-05,
"loss": 0.6849,
"step": 53
},
{
"epoch": 0.36486486486486486,
"grad_norm": 0.5619919459693192,
"learning_rate": 5.837837837837838e-05,
"loss": 0.6827,
"step": 54
},
{
"epoch": 0.3716216216216216,
"grad_norm": 0.789835519140916,
"learning_rate": 5.945945945945946e-05,
"loss": 0.6753,
"step": 55
},
{
"epoch": 0.3783783783783784,
"grad_norm": 0.8468946375056802,
"learning_rate": 6.054054054054055e-05,
"loss": 0.6803,
"step": 56
},
{
"epoch": 0.38513513513513514,
"grad_norm": 1.1480634289873959,
"learning_rate": 6.162162162162163e-05,
"loss": 0.6849,
"step": 57
},
{
"epoch": 0.3918918918918919,
"grad_norm": 0.8338876354018716,
"learning_rate": 6.27027027027027e-05,
"loss": 0.6822,
"step": 58
},
{
"epoch": 0.39864864864864863,
"grad_norm": 0.8619635833656949,
"learning_rate": 6.378378378378379e-05,
"loss": 0.6885,
"step": 59
},
{
"epoch": 0.40540540540540543,
"grad_norm": 0.9841339840879071,
"learning_rate": 6.486486486486487e-05,
"loss": 0.6941,
"step": 60
},
{
"epoch": 0.41216216216216217,
"grad_norm": 0.7846924650076373,
"learning_rate": 6.594594594594596e-05,
"loss": 0.6734,
"step": 61
},
{
"epoch": 0.4189189189189189,
"grad_norm": 0.6084845178399872,
"learning_rate": 6.702702702702703e-05,
"loss": 0.6766,
"step": 62
},
{
"epoch": 0.42567567567567566,
"grad_norm": 0.8121605538110969,
"learning_rate": 6.810810810810811e-05,
"loss": 0.6882,
"step": 63
},
{
"epoch": 0.43243243243243246,
"grad_norm": 0.9962451486752875,
"learning_rate": 6.91891891891892e-05,
"loss": 0.6837,
"step": 64
},
{
"epoch": 0.4391891891891892,
"grad_norm": 0.6516040658377318,
"learning_rate": 7.027027027027028e-05,
"loss": 0.6841,
"step": 65
},
{
"epoch": 0.44594594594594594,
"grad_norm": 0.683694787120212,
"learning_rate": 7.135135135135136e-05,
"loss": 0.6952,
"step": 66
},
{
"epoch": 0.4527027027027027,
"grad_norm": 0.6582713522549495,
"learning_rate": 7.243243243243243e-05,
"loss": 0.6727,
"step": 67
},
{
"epoch": 0.4594594594594595,
"grad_norm": 0.673908455931949,
"learning_rate": 7.351351351351353e-05,
"loss": 0.6841,
"step": 68
},
{
"epoch": 0.46621621621621623,
"grad_norm": 0.8017390116132226,
"learning_rate": 7.45945945945946e-05,
"loss": 0.67,
"step": 69
},
{
"epoch": 0.47297297297297297,
"grad_norm": 0.9799335757769192,
"learning_rate": 7.567567567567568e-05,
"loss": 0.6914,
"step": 70
},
{
"epoch": 0.4797297297297297,
"grad_norm": 1.5088822309237815,
"learning_rate": 7.675675675675675e-05,
"loss": 0.6925,
"step": 71
},
{
"epoch": 0.4864864864864865,
"grad_norm": 0.7235799329993131,
"learning_rate": 7.783783783783785e-05,
"loss": 0.689,
"step": 72
},
{
"epoch": 0.49324324324324326,
"grad_norm": 1.467550439828714,
"learning_rate": 7.891891891891892e-05,
"loss": 0.6927,
"step": 73
},
{
"epoch": 0.5,
"grad_norm": 0.9680632895596434,
"learning_rate": 8e-05,
"loss": 0.6729,
"step": 74
},
{
"epoch": 0.5067567567567568,
"grad_norm": 1.4048772257060618,
"learning_rate": 7.999955497902857e-05,
"loss": 0.6949,
"step": 75
},
{
"epoch": 0.5135135135135135,
"grad_norm": 1.187179427424338,
"learning_rate": 7.999821992601645e-05,
"loss": 0.693,
"step": 76
},
{
"epoch": 0.5202702702702703,
"grad_norm": 0.9529005283405405,
"learning_rate": 7.999599487066996e-05,
"loss": 0.6839,
"step": 77
},
{
"epoch": 0.527027027027027,
"grad_norm": 0.9173343088851188,
"learning_rate": 7.999287986249894e-05,
"loss": 0.6766,
"step": 78
},
{
"epoch": 0.5337837837837838,
"grad_norm": 0.7318824071301274,
"learning_rate": 7.998887497081555e-05,
"loss": 0.6847,
"step": 79
},
{
"epoch": 0.5405405405405406,
"grad_norm": 0.6529765700425837,
"learning_rate": 7.998398028473287e-05,
"loss": 0.678,
"step": 80
},
{
"epoch": 0.5472972972972973,
"grad_norm": 0.6928101408737816,
"learning_rate": 7.997819591316278e-05,
"loss": 0.6954,
"step": 81
},
{
"epoch": 0.5540540540540541,
"grad_norm": 0.5879827240121825,
"learning_rate": 7.99715219848136e-05,
"loss": 0.6642,
"step": 82
},
{
"epoch": 0.5608108108108109,
"grad_norm": 0.680605729950147,
"learning_rate": 7.996395864818727e-05,
"loss": 0.6804,
"step": 83
},
{
"epoch": 0.5675675675675675,
"grad_norm": 0.5423663061022331,
"learning_rate": 7.995550607157592e-05,
"loss": 0.6688,
"step": 84
},
{
"epoch": 0.5743243243243243,
"grad_norm": 0.606163015240631,
"learning_rate": 7.994616444305826e-05,
"loss": 0.6667,
"step": 85
},
{
"epoch": 0.581081081081081,
"grad_norm": 0.5950010094955055,
"learning_rate": 7.993593397049533e-05,
"loss": 0.679,
"step": 86
},
{
"epoch": 0.5878378378378378,
"grad_norm": 0.6429149705782599,
"learning_rate": 7.992481488152585e-05,
"loss": 0.6685,
"step": 87
},
{
"epoch": 0.5945945945945946,
"grad_norm": 0.46140817883542623,
"learning_rate": 7.991280742356124e-05,
"loss": 0.6603,
"step": 88
},
{
"epoch": 0.6013513513513513,
"grad_norm": 0.6795669467936494,
"learning_rate": 7.989991186378e-05,
"loss": 0.6699,
"step": 89
},
{
"epoch": 0.6081081081081081,
"grad_norm": 0.5738680266825441,
"learning_rate": 7.988612848912186e-05,
"loss": 0.6819,
"step": 90
},
{
"epoch": 0.6148648648648649,
"grad_norm": 0.5789217573047567,
"learning_rate": 7.987145760628138e-05,
"loss": 0.6708,
"step": 91
},
{
"epoch": 0.6216216216216216,
"grad_norm": 0.49449419241667797,
"learning_rate": 7.985589954170107e-05,
"loss": 0.6665,
"step": 92
},
{
"epoch": 0.6283783783783784,
"grad_norm": 0.4987045947630836,
"learning_rate": 7.983945464156419e-05,
"loss": 0.6597,
"step": 93
},
{
"epoch": 0.6351351351351351,
"grad_norm": 0.40082592041703324,
"learning_rate": 7.982212327178699e-05,
"loss": 0.6675,
"step": 94
},
{
"epoch": 0.6418918918918919,
"grad_norm": 0.3431463481882859,
"learning_rate": 7.980390581801064e-05,
"loss": 0.6616,
"step": 95
},
{
"epoch": 0.6486486486486487,
"grad_norm": 0.44148520818681297,
"learning_rate": 7.97848026855926e-05,
"loss": 0.6761,
"step": 96
},
{
"epoch": 0.6554054054054054,
"grad_norm": 0.4038247824809848,
"learning_rate": 7.976481429959758e-05,
"loss": 0.6596,
"step": 97
},
{
"epoch": 0.6621621621621622,
"grad_norm": 0.3448527603506924,
"learning_rate": 7.974394110478813e-05,
"loss": 0.6536,
"step": 98
},
{
"epoch": 0.668918918918919,
"grad_norm": 0.30698753802117784,
"learning_rate": 7.972218356561471e-05,
"loss": 0.6772,
"step": 99
},
{
"epoch": 0.6756756756756757,
"grad_norm": 0.30195068085071514,
"learning_rate": 7.96995421662054e-05,
"loss": 0.6527,
"step": 100
},
{
"epoch": 0.6824324324324325,
"grad_norm": 0.3166797371131855,
"learning_rate": 7.967601741035507e-05,
"loss": 0.6555,
"step": 101
},
{
"epoch": 0.6891891891891891,
"grad_norm": 0.32820963864069536,
"learning_rate": 7.965160982151422e-05,
"loss": 0.6696,
"step": 102
},
{
"epoch": 0.6959459459459459,
"grad_norm": 0.3483528954236439,
"learning_rate": 7.962631994277728e-05,
"loss": 0.663,
"step": 103
},
{
"epoch": 0.7027027027027027,
"grad_norm": 0.4983197508582807,
"learning_rate": 7.960014833687055e-05,
"loss": 0.6633,
"step": 104
},
{
"epoch": 0.7094594594594594,
"grad_norm": 0.7805957154226273,
"learning_rate": 7.957309558613974e-05,
"loss": 0.6587,
"step": 105
},
{
"epoch": 0.7162162162162162,
"grad_norm": 1.106633618810969,
"learning_rate": 7.954516229253691e-05,
"loss": 0.662,
"step": 106
},
{
"epoch": 0.722972972972973,
"grad_norm": 0.757015003567712,
"learning_rate": 7.951634907760713e-05,
"loss": 0.6593,
"step": 107
},
{
"epoch": 0.7297297297297297,
"grad_norm": 0.8427739970040843,
"learning_rate": 7.948665658247463e-05,
"loss": 0.6601,
"step": 108
},
{
"epoch": 0.7364864864864865,
"grad_norm": 0.7431541528418906,
"learning_rate": 7.945608546782858e-05,
"loss": 0.667,
"step": 109
},
{
"epoch": 0.7432432432432432,
"grad_norm": 0.6845263081020915,
"learning_rate": 7.942463641390834e-05,
"loss": 0.6569,
"step": 110
},
{
"epoch": 0.75,
"grad_norm": 0.8712514657793697,
"learning_rate": 7.939231012048833e-05,
"loss": 0.6632,
"step": 111
},
{
"epoch": 0.7567567567567568,
"grad_norm": 0.6774356142156878,
"learning_rate": 7.935910730686246e-05,
"loss": 0.6706,
"step": 112
},
{
"epoch": 0.7635135135135135,
"grad_norm": 0.7702128806997793,
"learning_rate": 7.932502871182818e-05,
"loss": 0.6679,
"step": 113
},
{
"epoch": 0.7702702702702703,
"grad_norm": 0.6032796273856512,
"learning_rate": 7.929007509366994e-05,
"loss": 0.6612,
"step": 114
},
{
"epoch": 0.777027027027027,
"grad_norm": 0.7262342972595842,
"learning_rate": 7.925424723014239e-05,
"loss": 0.6639,
"step": 115
},
{
"epoch": 0.7837837837837838,
"grad_norm": 0.6319433116490037,
"learning_rate": 7.921754591845307e-05,
"loss": 0.662,
"step": 116
},
{
"epoch": 0.7905405405405406,
"grad_norm": 0.6175612756087657,
"learning_rate": 7.917997197524467e-05,
"loss": 0.6566,
"step": 117
},
{
"epoch": 0.7972972972972973,
"grad_norm": 0.7923618524838222,
"learning_rate": 7.914152623657678e-05,
"loss": 0.6469,
"step": 118
},
{
"epoch": 0.8040540540540541,
"grad_norm": 0.7707195088131242,
"learning_rate": 7.910220955790746e-05,
"loss": 0.6563,
"step": 119
},
{
"epoch": 0.8108108108108109,
"grad_norm": 0.4152288654221399,
"learning_rate": 7.906202281407398e-05,
"loss": 0.6519,
"step": 120
},
{
"epoch": 0.8175675675675675,
"grad_norm": 0.5439040195272984,
"learning_rate": 7.902096689927355e-05,
"loss": 0.6727,
"step": 121
},
{
"epoch": 0.8243243243243243,
"grad_norm": 0.5159119091886112,
"learning_rate": 7.897904272704333e-05,
"loss": 0.6563,
"step": 122
},
{
"epoch": 0.831081081081081,
"grad_norm": 0.49623067194442205,
"learning_rate": 7.893625123024011e-05,
"loss": 0.6523,
"step": 123
},
{
"epoch": 0.8378378378378378,
"grad_norm": 0.4176229579696035,
"learning_rate": 7.889259336101957e-05,
"loss": 0.6461,
"step": 124
},
{
"epoch": 0.8445945945945946,
"grad_norm": 0.5424146541267013,
"learning_rate": 7.884807009081506e-05,
"loss": 0.6535,
"step": 125
},
{
"epoch": 0.8513513513513513,
"grad_norm": 0.53245673115827,
"learning_rate": 7.880268241031604e-05,
"loss": 0.6516,
"step": 126
},
{
"epoch": 0.8581081081081081,
"grad_norm": 0.4218647663890849,
"learning_rate": 7.875643132944599e-05,
"loss": 0.6557,
"step": 127
},
{
"epoch": 0.8648648648648649,
"grad_norm": 0.31388453196671723,
"learning_rate": 7.870931787733996e-05,
"loss": 0.6504,
"step": 128
},
{
"epoch": 0.8716216216216216,
"grad_norm": 0.42120451899053063,
"learning_rate": 7.866134310232167e-05,
"loss": 0.6637,
"step": 129
},
{
"epoch": 0.8783783783783784,
"grad_norm": 0.3569049182508329,
"learning_rate": 7.861250807188014e-05,
"loss": 0.6565,
"step": 130
},
{
"epoch": 0.8851351351351351,
"grad_norm": 0.46980061573845006,
"learning_rate": 7.856281387264603e-05,
"loss": 0.6643,
"step": 131
},
{
"epoch": 0.8918918918918919,
"grad_norm": 0.5602330500767254,
"learning_rate": 7.851226161036739e-05,
"loss": 0.6541,
"step": 132
},
{
"epoch": 0.8986486486486487,
"grad_norm": 0.555307896703473,
"learning_rate": 7.846085240988503e-05,
"loss": 0.6498,
"step": 133
},
{
"epoch": 0.9054054054054054,
"grad_norm": 0.4906170522953401,
"learning_rate": 7.840858741510758e-05,
"loss": 0.6482,
"step": 134
},
{
"epoch": 0.9121621621621622,
"grad_norm": 0.4486181714024936,
"learning_rate": 7.835546778898599e-05,
"loss": 0.6457,
"step": 135
},
{
"epoch": 0.918918918918919,
"grad_norm": 0.5704154766929971,
"learning_rate": 7.830149471348763e-05,
"loss": 0.6508,
"step": 136
},
{
"epoch": 0.9256756756756757,
"grad_norm": 0.738054877738817,
"learning_rate": 7.824666938957004e-05,
"loss": 0.6531,
"step": 137
},
{
"epoch": 0.9324324324324325,
"grad_norm": 0.7131460924145132,
"learning_rate": 7.819099303715414e-05,
"loss": 0.6577,
"step": 138
},
{
"epoch": 0.9391891891891891,
"grad_norm": 0.5314442328007468,
"learning_rate": 7.813446689509718e-05,
"loss": 0.6472,
"step": 139
},
{
"epoch": 0.9459459459459459,
"grad_norm": 0.4354387291682998,
"learning_rate": 7.807709222116506e-05,
"loss": 0.6566,
"step": 140
},
{
"epoch": 0.9527027027027027,
"grad_norm": 0.4367558414891408,
"learning_rate": 7.801887029200448e-05,
"loss": 0.6548,
"step": 141
},
{
"epoch": 0.9594594594594594,
"grad_norm": 0.49114789228090255,
"learning_rate": 7.795980240311436e-05,
"loss": 0.6651,
"step": 142
},
{
"epoch": 0.9662162162162162,
"grad_norm": 0.5563118895714915,
"learning_rate": 7.789988986881719e-05,
"loss": 0.6534,
"step": 143
},
{
"epoch": 0.972972972972973,
"grad_norm": 0.5068498475757595,
"learning_rate": 7.78391340222297e-05,
"loss": 0.6635,
"step": 144
},
{
"epoch": 0.9797297297297297,
"grad_norm": 0.40377422580342515,
"learning_rate": 7.777753621523316e-05,
"loss": 0.662,
"step": 145
},
{
"epoch": 0.9864864864864865,
"grad_norm": 0.48943576691657636,
"learning_rate": 7.771509781844338e-05,
"loss": 0.6465,
"step": 146
},
{
"epoch": 0.9932432432432432,
"grad_norm": 0.45222666736741574,
"learning_rate": 7.765182022118014e-05,
"loss": 0.6576,
"step": 147
},
{
"epoch": 1.0,
"grad_norm": 0.3524115634430567,
"learning_rate": 7.758770483143634e-05,
"loss": 0.66,
"step": 148
},
{
"epoch": 1.0067567567567568,
"grad_norm": 0.3465470718278503,
"learning_rate": 7.752275307584664e-05,
"loss": 0.6528,
"step": 149
},
{
"epoch": 1.0135135135135136,
"grad_norm": 0.4289730330622286,
"learning_rate": 7.745696639965569e-05,
"loss": 0.6284,
"step": 150
},
{
"epoch": 1.0202702702702702,
"grad_norm": 0.4078417015334056,
"learning_rate": 7.739034626668605e-05,
"loss": 0.6409,
"step": 151
},
{
"epoch": 1.027027027027027,
"grad_norm": 0.4224099719890233,
"learning_rate": 7.732289415930549e-05,
"loss": 0.63,
"step": 152
},
{
"epoch": 1.0337837837837838,
"grad_norm": 0.43135712270900994,
"learning_rate": 7.725461157839417e-05,
"loss": 0.6606,
"step": 153
},
{
"epoch": 1.0405405405405406,
"grad_norm": 0.4832560652031643,
"learning_rate": 7.71855000433111e-05,
"loss": 0.6439,
"step": 154
},
{
"epoch": 1.0472972972972974,
"grad_norm": 0.5148651119907238,
"learning_rate": 7.711556109186039e-05,
"loss": 0.6405,
"step": 155
},
{
"epoch": 1.054054054054054,
"grad_norm": 0.49381906164540984,
"learning_rate": 7.704479628025704e-05,
"loss": 0.635,
"step": 156
},
{
"epoch": 1.0608108108108107,
"grad_norm": 0.4238088291952583,
"learning_rate": 7.697320718309235e-05,
"loss": 0.6363,
"step": 157
},
{
"epoch": 1.0675675675675675,
"grad_norm": 0.34639148621347143,
"learning_rate": 7.690079539329875e-05,
"loss": 0.6434,
"step": 158
},
{
"epoch": 1.0743243243243243,
"grad_norm": 0.3670601156648982,
"learning_rate": 7.682756252211453e-05,
"loss": 0.6321,
"step": 159
},
{
"epoch": 1.0810810810810811,
"grad_norm": 0.48931637412654205,
"learning_rate": 7.675351019904785e-05,
"loss": 0.6371,
"step": 160
},
{
"epoch": 1.087837837837838,
"grad_norm": 0.5514281223158661,
"learning_rate": 7.667864007184054e-05,
"loss": 0.6313,
"step": 161
},
{
"epoch": 1.0945945945945945,
"grad_norm": 0.48466826872522856,
"learning_rate": 7.660295380643144e-05,
"loss": 0.6358,
"step": 162
},
{
"epoch": 1.1013513513513513,
"grad_norm": 0.46142889218279065,
"learning_rate": 7.652645308691933e-05,
"loss": 0.6418,
"step": 163
},
{
"epoch": 1.1081081081081081,
"grad_norm": 0.4371908363892829,
"learning_rate": 7.644913961552544e-05,
"loss": 0.6327,
"step": 164
},
{
"epoch": 1.114864864864865,
"grad_norm": 0.4159116225859721,
"learning_rate": 7.637101511255554e-05,
"loss": 0.6332,
"step": 165
},
{
"epoch": 1.1216216216216217,
"grad_norm": 0.48524137415448176,
"learning_rate": 7.629208131636179e-05,
"loss": 0.6385,
"step": 166
},
{
"epoch": 1.1283783783783783,
"grad_norm": 0.502011807585537,
"learning_rate": 7.621233998330387e-05,
"loss": 0.6534,
"step": 167
},
{
"epoch": 1.135135135135135,
"grad_norm": 0.4849012769427631,
"learning_rate": 7.61317928877101e-05,
"loss": 0.6382,
"step": 168
},
{
"epoch": 1.1418918918918919,
"grad_norm": 0.40497697606883803,
"learning_rate": 7.605044182183779e-05,
"loss": 0.6335,
"step": 169
},
{
"epoch": 1.1486486486486487,
"grad_norm": 0.26814241993209303,
"learning_rate": 7.596828859583347e-05,
"loss": 0.6226,
"step": 170
},
{
"epoch": 1.1554054054054055,
"grad_norm": 0.36135092127292284,
"learning_rate": 7.588533503769257e-05,
"loss": 0.6286,
"step": 171
},
{
"epoch": 1.1621621621621623,
"grad_norm": 0.4608397427011372,
"learning_rate": 7.580158299321872e-05,
"loss": 0.6438,
"step": 172
},
{
"epoch": 1.1689189189189189,
"grad_norm": 0.4171913645594413,
"learning_rate": 7.571703432598275e-05,
"loss": 0.6356,
"step": 173
},
{
"epoch": 1.1756756756756757,
"grad_norm": 0.4370902558666279,
"learning_rate": 7.563169091728115e-05,
"loss": 0.6422,
"step": 174
},
{
"epoch": 1.1824324324324325,
"grad_norm": 0.4283980244016275,
"learning_rate": 7.554555466609425e-05,
"loss": 0.6407,
"step": 175
},
{
"epoch": 1.1891891891891893,
"grad_norm": 0.4381556639196284,
"learning_rate": 7.545862748904394e-05,
"loss": 0.6434,
"step": 176
},
{
"epoch": 1.195945945945946,
"grad_norm": 0.44668924620805633,
"learning_rate": 7.537091132035111e-05,
"loss": 0.6219,
"step": 177
},
{
"epoch": 1.2027027027027026,
"grad_norm": 0.37602920076180757,
"learning_rate": 7.528240811179245e-05,
"loss": 0.6419,
"step": 178
},
{
"epoch": 1.2094594594594594,
"grad_norm": 0.2672376131285308,
"learning_rate": 7.519311983265718e-05,
"loss": 0.6366,
"step": 179
},
{
"epoch": 1.2162162162162162,
"grad_norm": 0.29369317421661273,
"learning_rate": 7.510304846970311e-05,
"loss": 0.6341,
"step": 180
},
{
"epoch": 1.222972972972973,
"grad_norm": 0.3602685800886801,
"learning_rate": 7.501219602711253e-05,
"loss": 0.6432,
"step": 181
},
{
"epoch": 1.2297297297297298,
"grad_norm": 0.3237536586353385,
"learning_rate": 7.492056452644753e-05,
"loss": 0.6415,
"step": 182
},
{
"epoch": 1.2364864864864864,
"grad_norm": 0.28545557770169117,
"learning_rate": 7.48281560066051e-05,
"loss": 0.6364,
"step": 183
},
{
"epoch": 1.2432432432432432,
"grad_norm": 0.2623342902523158,
"learning_rate": 7.473497252377171e-05,
"loss": 0.6333,
"step": 184
},
{
"epoch": 1.25,
"grad_norm": 0.3160766346456171,
"learning_rate": 7.464101615137756e-05,
"loss": 0.6256,
"step": 185
},
{
"epoch": 1.2567567567567568,
"grad_norm": 0.4173934303596852,
"learning_rate": 7.454628898005043e-05,
"loss": 0.6521,
"step": 186
},
{
"epoch": 1.2635135135135136,
"grad_norm": 0.4424346331081593,
"learning_rate": 7.445079311756924e-05,
"loss": 0.6541,
"step": 187
},
{
"epoch": 1.2702702702702702,
"grad_norm": 0.38764273759912543,
"learning_rate": 7.435453068881706e-05,
"loss": 0.6385,
"step": 188
},
{
"epoch": 1.277027027027027,
"grad_norm": 0.3811463582022792,
"learning_rate": 7.425750383573384e-05,
"loss": 0.6295,
"step": 189
},
{
"epoch": 1.2837837837837838,
"grad_norm": 0.5272316642769805,
"learning_rate": 7.415971471726884e-05,
"loss": 0.6307,
"step": 190
},
{
"epoch": 1.2905405405405406,
"grad_norm": 0.7635509141776905,
"learning_rate": 7.406116550933246e-05,
"loss": 0.6313,
"step": 191
},
{
"epoch": 1.2972972972972974,
"grad_norm": 0.9673085818387788,
"learning_rate": 7.396185840474792e-05,
"loss": 0.6495,
"step": 192
},
{
"epoch": 1.304054054054054,
"grad_norm": 0.8691278543652353,
"learning_rate": 7.386179561320243e-05,
"loss": 0.6285,
"step": 193
},
{
"epoch": 1.3108108108108107,
"grad_norm": 0.7233647856746338,
"learning_rate": 7.376097936119803e-05,
"loss": 0.6435,
"step": 194
},
{
"epoch": 1.3175675675675675,
"grad_norm": 0.4888365541378487,
"learning_rate": 7.365941189200201e-05,
"loss": 0.6345,
"step": 195
},
{
"epoch": 1.3243243243243243,
"grad_norm": 0.41440516733521193,
"learning_rate": 7.35570954655971e-05,
"loss": 0.6522,
"step": 196
},
{
"epoch": 1.3310810810810811,
"grad_norm": 0.39733170188465466,
"learning_rate": 7.345403235863105e-05,
"loss": 0.636,
"step": 197
},
{
"epoch": 1.3378378378378377,
"grad_norm": 0.37538788641282456,
"learning_rate": 7.335022486436608e-05,
"loss": 0.6405,
"step": 198
},
{
"epoch": 1.3445945945945945,
"grad_norm": 0.4912257786205075,
"learning_rate": 7.324567529262775e-05,
"loss": 0.6341,
"step": 199
},
{
"epoch": 1.3513513513513513,
"grad_norm": 0.3964231351313485,
"learning_rate": 7.31403859697537e-05,
"loss": 0.6329,
"step": 200
},
{
"epoch": 1.3581081081081081,
"grad_norm": 0.23258287711574407,
"learning_rate": 7.303435923854172e-05,
"loss": 0.6346,
"step": 201
},
{
"epoch": 1.364864864864865,
"grad_norm": 0.36875878813699686,
"learning_rate": 7.292759745819781e-05,
"loss": 0.6423,
"step": 202
},
{
"epoch": 1.3716216216216215,
"grad_norm": 0.3816088768921796,
"learning_rate": 7.282010300428351e-05,
"loss": 0.6247,
"step": 203
},
{
"epoch": 1.3783783783783785,
"grad_norm": 0.32935135005176236,
"learning_rate": 7.271187826866312e-05,
"loss": 0.6306,
"step": 204
},
{
"epoch": 1.385135135135135,
"grad_norm": 0.24849408204445086,
"learning_rate": 7.260292565945049e-05,
"loss": 0.6425,
"step": 205
},
{
"epoch": 1.3918918918918919,
"grad_norm": 0.22496521412221002,
"learning_rate": 7.249324760095544e-05,
"loss": 0.64,
"step": 206
},
{
"epoch": 1.3986486486486487,
"grad_norm": 0.30450316180118336,
"learning_rate": 7.238284653362977e-05,
"loss": 0.6497,
"step": 207
},
{
"epoch": 1.4054054054054055,
"grad_norm": 0.2922608158162527,
"learning_rate": 7.227172491401299e-05,
"loss": 0.6339,
"step": 208
},
{
"epoch": 1.4121621621621623,
"grad_norm": 0.23745259799881574,
"learning_rate": 7.215988521467763e-05,
"loss": 0.629,
"step": 209
},
{
"epoch": 1.4189189189189189,
"grad_norm": 0.29205738332283676,
"learning_rate": 7.204732992417431e-05,
"loss": 0.6276,
"step": 210
},
{
"epoch": 1.4256756756756757,
"grad_norm": 0.2764709836681376,
"learning_rate": 7.193406154697625e-05,
"loss": 0.6325,
"step": 211
},
{
"epoch": 1.4324324324324325,
"grad_norm": 0.30413188245977635,
"learning_rate": 7.18200826034236e-05,
"loss": 0.6347,
"step": 212
},
{
"epoch": 1.4391891891891893,
"grad_norm": 0.32848243797717097,
"learning_rate": 7.17053956296674e-05,
"loss": 0.6173,
"step": 213
},
{
"epoch": 1.445945945945946,
"grad_norm": 0.3056033484058255,
"learning_rate": 7.159000317761305e-05,
"loss": 0.6339,
"step": 214
},
{
"epoch": 1.4527027027027026,
"grad_norm": 0.3085891817121419,
"learning_rate": 7.14739078148636e-05,
"loss": 0.6364,
"step": 215
},
{
"epoch": 1.4594594594594594,
"grad_norm": 0.3326838165486728,
"learning_rate": 7.135711212466264e-05,
"loss": 0.6422,
"step": 216
},
{
"epoch": 1.4662162162162162,
"grad_norm": 0.3216850978230415,
"learning_rate": 7.123961870583671e-05,
"loss": 0.6342,
"step": 217
},
{
"epoch": 1.472972972972973,
"grad_norm": 0.35372809908349234,
"learning_rate": 7.112143017273759e-05,
"loss": 0.637,
"step": 218
},
{
"epoch": 1.4797297297297298,
"grad_norm": 0.3689688937687713,
"learning_rate": 7.100254915518408e-05,
"loss": 0.6342,
"step": 219
},
{
"epoch": 1.4864864864864864,
"grad_norm": 0.4017045063635094,
"learning_rate": 7.088297829840346e-05,
"loss": 0.6286,
"step": 220
},
{
"epoch": 1.4932432432432432,
"grad_norm": 0.4446976891407491,
"learning_rate": 7.076272026297268e-05,
"loss": 0.6342,
"step": 221
},
{
"epoch": 1.5,
"grad_norm": 0.432808144889846,
"learning_rate": 7.064177772475912e-05,
"loss": 0.6452,
"step": 222
},
{
"epoch": 1.5067567567567568,
"grad_norm": 0.3443564755955631,
"learning_rate": 7.052015337486109e-05,
"loss": 0.6378,
"step": 223
},
{
"epoch": 1.5135135135135136,
"grad_norm": 0.30789638902879063,
"learning_rate": 7.03978499195479e-05,
"loss": 0.6284,
"step": 224
},
{
"epoch": 1.5202702702702702,
"grad_norm": 0.3444085722973672,
"learning_rate": 7.027487008019969e-05,
"loss": 0.6439,
"step": 225
},
{
"epoch": 1.527027027027027,
"grad_norm": 0.3698979548144048,
"learning_rate": 7.015121659324678e-05,
"loss": 0.6328,
"step": 226
},
{
"epoch": 1.5337837837837838,
"grad_norm": 0.4166719230003736,
"learning_rate": 7.002689221010897e-05,
"loss": 0.6295,
"step": 227
},
{
"epoch": 1.5405405405405406,
"grad_norm": 0.4418395627095002,
"learning_rate": 6.990189969713416e-05,
"loss": 0.6303,
"step": 228
},
{
"epoch": 1.5472972972972974,
"grad_norm": 0.4102988370573012,
"learning_rate": 6.977624183553676e-05,
"loss": 0.6431,
"step": 229
},
{
"epoch": 1.554054054054054,
"grad_norm": 0.39270575425283927,
"learning_rate": 6.964992142133602e-05,
"loss": 0.6333,
"step": 230
},
{
"epoch": 1.560810810810811,
"grad_norm": 0.3396456921939933,
"learning_rate": 6.952294126529356e-05,
"loss": 0.6274,
"step": 231
},
{
"epoch": 1.5675675675675675,
"grad_norm": 0.29161162699565907,
"learning_rate": 6.939530419285104e-05,
"loss": 0.6346,
"step": 232
},
{
"epoch": 1.5743243243243243,
"grad_norm": 0.3270645478162509,
"learning_rate": 6.926701304406713e-05,
"loss": 0.6307,
"step": 233
},
{
"epoch": 1.5810810810810811,
"grad_norm": 0.33343232181767685,
"learning_rate": 6.913807067355445e-05,
"loss": 0.6338,
"step": 234
},
{
"epoch": 1.5878378378378377,
"grad_norm": 0.34421561990305105,
"learning_rate": 6.90084799504159e-05,
"loss": 0.6417,
"step": 235
},
{
"epoch": 1.5945945945945947,
"grad_norm": 0.31139357648881777,
"learning_rate": 6.887824375818099e-05,
"loss": 0.6399,
"step": 236
},
{
"epoch": 1.6013513513513513,
"grad_norm": 0.2657664210184094,
"learning_rate": 6.874736499474154e-05,
"loss": 0.6411,
"step": 237
},
{
"epoch": 1.6081081081081081,
"grad_norm": 0.25481549866865094,
"learning_rate": 6.861584657228728e-05,
"loss": 0.6418,
"step": 238
},
{
"epoch": 1.614864864864865,
"grad_norm": 0.27638932514157166,
"learning_rate": 6.848369141724104e-05,
"loss": 0.6435,
"step": 239
},
{
"epoch": 1.6216216216216215,
"grad_norm": 0.32918191799417873,
"learning_rate": 6.835090247019354e-05,
"loss": 0.6269,
"step": 240
},
{
"epoch": 1.6283783783783785,
"grad_norm": 0.334162128338808,
"learning_rate": 6.821748268583813e-05,
"loss": 0.6382,
"step": 241
},
{
"epoch": 1.635135135135135,
"grad_norm": 0.3172895628393372,
"learning_rate": 6.808343503290491e-05,
"loss": 0.627,
"step": 242
},
{
"epoch": 1.6418918918918919,
"grad_norm": 0.3353087016871893,
"learning_rate": 6.79487624940947e-05,
"loss": 0.6228,
"step": 243
},
{
"epoch": 1.6486486486486487,
"grad_norm": 0.3718369962163942,
"learning_rate": 6.781346806601273e-05,
"loss": 0.6305,
"step": 244
},
{
"epoch": 1.6554054054054053,
"grad_norm": 0.39658980608696004,
"learning_rate": 6.767755475910185e-05,
"loss": 0.6227,
"step": 245
},
{
"epoch": 1.6621621621621623,
"grad_norm": 0.4269305577093925,
"learning_rate": 6.754102559757569e-05,
"loss": 0.6338,
"step": 246
},
{
"epoch": 1.6689189189189189,
"grad_norm": 0.48834528749062367,
"learning_rate": 6.740388361935125e-05,
"loss": 0.6285,
"step": 247
},
{
"epoch": 1.6756756756756757,
"grad_norm": 0.5054797184183533,
"learning_rate": 6.726613187598132e-05,
"loss": 0.6363,
"step": 248
},
{
"epoch": 1.6824324324324325,
"grad_norm": 0.46558382405228205,
"learning_rate": 6.712777343258666e-05,
"loss": 0.6342,
"step": 249
},
{
"epoch": 1.689189189189189,
"grad_norm": 0.3349216901360824,
"learning_rate": 6.698881136778771e-05,
"loss": 0.638,
"step": 250
},
{
"epoch": 1.695945945945946,
"grad_norm": 0.296031271653257,
"learning_rate": 6.684924877363613e-05,
"loss": 0.6449,
"step": 251
},
{
"epoch": 1.7027027027027026,
"grad_norm": 0.34651680386096584,
"learning_rate": 6.670908875554594e-05,
"loss": 0.6373,
"step": 252
},
{
"epoch": 1.7094594594594594,
"grad_norm": 0.34431135065378676,
"learning_rate": 6.656833443222458e-05,
"loss": 0.6222,
"step": 253
},
{
"epoch": 1.7162162162162162,
"grad_norm": 0.3856861975419152,
"learning_rate": 6.642698893560327e-05,
"loss": 0.6389,
"step": 254
},
{
"epoch": 1.722972972972973,
"grad_norm": 0.3419900662624624,
"learning_rate": 6.628505541076755e-05,
"loss": 0.6345,
"step": 255
},
{
"epoch": 1.7297297297297298,
"grad_norm": 0.25513443874276925,
"learning_rate": 6.614253701588718e-05,
"loss": 0.6337,
"step": 256
},
{
"epoch": 1.7364864864864864,
"grad_norm": 0.40205274571625865,
"learning_rate": 6.599943692214587e-05,
"loss": 0.6329,
"step": 257
},
{
"epoch": 1.7432432432432432,
"grad_norm": 0.38134230469861735,
"learning_rate": 6.585575831367078e-05,
"loss": 0.6315,
"step": 258
},
{
"epoch": 1.75,
"grad_norm": 0.2108025719904083,
"learning_rate": 6.571150438746157e-05,
"loss": 0.6332,
"step": 259
},
{
"epoch": 1.7567567567567568,
"grad_norm": 0.2896853669566718,
"learning_rate": 6.55666783533194e-05,
"loss": 0.6254,
"step": 260
},
{
"epoch": 1.7635135135135136,
"grad_norm": 0.3811309603864002,
"learning_rate": 6.542128343377536e-05,
"loss": 0.6355,
"step": 261
},
{
"epoch": 1.7702702702702702,
"grad_norm": 0.30773816113154867,
"learning_rate": 6.527532286401889e-05,
"loss": 0.6337,
"step": 262
},
{
"epoch": 1.777027027027027,
"grad_norm": 0.22231601239252788,
"learning_rate": 6.51287998918257e-05,
"loss": 0.6204,
"step": 263
},
{
"epoch": 1.7837837837837838,
"grad_norm": 0.2581820667256423,
"learning_rate": 6.498171777748557e-05,
"loss": 0.6335,
"step": 264
},
{
"epoch": 1.7905405405405406,
"grad_norm": 0.3087437556000368,
"learning_rate": 6.483407979372975e-05,
"loss": 0.6221,
"step": 265
},
{
"epoch": 1.7972972972972974,
"grad_norm": 0.229624863667569,
"learning_rate": 6.468588922565822e-05,
"loss": 0.6305,
"step": 266
},
{
"epoch": 1.804054054054054,
"grad_norm": 0.15965170109349433,
"learning_rate": 6.453714937066648e-05,
"loss": 0.6277,
"step": 267
},
{
"epoch": 1.810810810810811,
"grad_norm": 0.16500327907249202,
"learning_rate": 6.438786353837228e-05,
"loss": 0.6279,
"step": 268
},
{
"epoch": 1.8175675675675675,
"grad_norm": 0.19228879276238184,
"learning_rate": 6.423803505054193e-05,
"loss": 0.637,
"step": 269
},
{
"epoch": 1.8243243243243243,
"grad_norm": 0.1960334960484057,
"learning_rate": 6.408766724101638e-05,
"loss": 0.6344,
"step": 270
},
{
"epoch": 1.8310810810810811,
"grad_norm": 0.22246057335109667,
"learning_rate": 6.393676345563708e-05,
"loss": 0.6315,
"step": 271
},
{
"epoch": 1.8378378378378377,
"grad_norm": 0.25500558074002594,
"learning_rate": 6.378532705217148e-05,
"loss": 0.6267,
"step": 272
},
{
"epoch": 1.8445945945945947,
"grad_norm": 0.2797665813089635,
"learning_rate": 6.363336140023833e-05,
"loss": 0.6198,
"step": 273
},
{
"epoch": 1.8513513513513513,
"grad_norm": 0.32237396509938987,
"learning_rate": 6.348086988123274e-05,
"loss": 0.6302,
"step": 274
},
{
"epoch": 1.8581081081081081,
"grad_norm": 0.3837879616068762,
"learning_rate": 6.332785588825094e-05,
"loss": 0.6366,
"step": 275
},
{
"epoch": 1.864864864864865,
"grad_norm": 0.3847808366177522,
"learning_rate": 6.317432282601469e-05,
"loss": 0.6405,
"step": 276
},
{
"epoch": 1.8716216216216215,
"grad_norm": 0.3485054619073992,
"learning_rate": 6.302027411079562e-05,
"loss": 0.6273,
"step": 277
},
{
"epoch": 1.8783783783783785,
"grad_norm": 0.3171200624449193,
"learning_rate": 6.286571317033915e-05,
"loss": 0.6337,
"step": 278
},
{
"epoch": 1.885135135135135,
"grad_norm": 0.27373328598715324,
"learning_rate": 6.271064344378832e-05,
"loss": 0.6403,
"step": 279
},
{
"epoch": 1.8918918918918919,
"grad_norm": 0.3045549423645127,
"learning_rate": 6.255506838160711e-05,
"loss": 0.6317,
"step": 280
},
{
"epoch": 1.8986486486486487,
"grad_norm": 0.23258625831405336,
"learning_rate": 6.239899144550383e-05,
"loss": 0.6195,
"step": 281
},
{
"epoch": 1.9054054054054053,
"grad_norm": 0.30422094244714903,
"learning_rate": 6.224241610835391e-05,
"loss": 0.6422,
"step": 282
},
{
"epoch": 1.9121621621621623,
"grad_norm": 0.3974220280836707,
"learning_rate": 6.208534585412282e-05,
"loss": 0.6341,
"step": 283
},
{
"epoch": 1.9189189189189189,
"grad_norm": 0.3588942161699914,
"learning_rate": 6.19277841777884e-05,
"loss": 0.6277,
"step": 284
},
{
"epoch": 1.9256756756756757,
"grad_norm": 0.23348715119029995,
"learning_rate": 6.176973458526317e-05,
"loss": 0.6196,
"step": 285
},
{
"epoch": 1.9324324324324325,
"grad_norm": 0.18377971391888995,
"learning_rate": 6.161120059331628e-05,
"loss": 0.6332,
"step": 286
},
{
"epoch": 1.939189189189189,
"grad_norm": 0.2385215605440726,
"learning_rate": 6.14521857294953e-05,
"loss": 0.6479,
"step": 287
},
{
"epoch": 1.945945945945946,
"grad_norm": 0.2797923846654065,
"learning_rate": 6.129269353204769e-05,
"loss": 0.6354,
"step": 288
},
{
"epoch": 1.9527027027027026,
"grad_norm": 0.3046327128914545,
"learning_rate": 6.113272754984206e-05,
"loss": 0.6394,
"step": 289
},
{
"epoch": 1.9594594594594594,
"grad_norm": 0.2685120513408917,
"learning_rate": 6.0972291342289274e-05,
"loss": 0.6285,
"step": 290
},
{
"epoch": 1.9662162162162162,
"grad_norm": 0.19923838754652695,
"learning_rate": 6.081138847926317e-05,
"loss": 0.6239,
"step": 291
},
{
"epoch": 1.972972972972973,
"grad_norm": 0.23688012061160577,
"learning_rate": 6.065002254102116e-05,
"loss": 0.6391,
"step": 292
},
{
"epoch": 1.9797297297297298,
"grad_norm": 0.2444120597492392,
"learning_rate": 6.048819711812457e-05,
"loss": 0.6315,
"step": 293
},
{
"epoch": 1.9864864864864864,
"grad_norm": 0.2374660611243683,
"learning_rate": 6.032591581135878e-05,
"loss": 0.635,
"step": 294
},
{
"epoch": 1.9932432432432432,
"grad_norm": 0.19031195961334121,
"learning_rate": 6.0163182231652985e-05,
"loss": 0.6266,
"step": 295
},
{
"epoch": 2.0,
"grad_norm": 0.2396442769161295,
"learning_rate": 6.000000000000001e-05,
"loss": 0.6384,
"step": 296
},
{
"epoch": 2.0067567567567566,
"grad_norm": 0.2791243307653228,
"learning_rate": 5.983637274737558e-05,
"loss": 0.6251,
"step": 297
},
{
"epoch": 2.0135135135135136,
"grad_norm": 0.3160836645945337,
"learning_rate": 5.967230411465768e-05,
"loss": 0.6261,
"step": 298
},
{
"epoch": 2.02027027027027,
"grad_norm": 0.2816061126247031,
"learning_rate": 5.950779775254539e-05,
"loss": 0.6146,
"step": 299
},
{
"epoch": 2.027027027027027,
"grad_norm": 0.3070948777190884,
"learning_rate": 5.934285732147778e-05,
"loss": 0.6234,
"step": 300
},
{
"epoch": 2.0337837837837838,
"grad_norm": 0.31200072769112625,
"learning_rate": 5.91774864915524e-05,
"loss": 0.6129,
"step": 301
},
{
"epoch": 2.0405405405405403,
"grad_norm": 0.28687016611033417,
"learning_rate": 5.90116889424436e-05,
"loss": 0.6144,
"step": 302
},
{
"epoch": 2.0472972972972974,
"grad_norm": 0.3063930815670496,
"learning_rate": 5.884546836332072e-05,
"loss": 0.6176,
"step": 303
},
{
"epoch": 2.054054054054054,
"grad_norm": 0.24478115745724477,
"learning_rate": 5.867882845276593e-05,
"loss": 0.6164,
"step": 304
},
{
"epoch": 2.060810810810811,
"grad_norm": 0.17845425335065362,
"learning_rate": 5.851177291869197e-05,
"loss": 0.6118,
"step": 305
},
{
"epoch": 2.0675675675675675,
"grad_norm": 0.20058855750068053,
"learning_rate": 5.834430547825964e-05,
"loss": 0.6146,
"step": 306
},
{
"epoch": 2.074324324324324,
"grad_norm": 0.21828538712863907,
"learning_rate": 5.8176429857795104e-05,
"loss": 0.6176,
"step": 307
},
{
"epoch": 2.081081081081081,
"grad_norm": 0.2218763340787443,
"learning_rate": 5.8008149792706936e-05,
"loss": 0.6059,
"step": 308
},
{
"epoch": 2.0878378378378377,
"grad_norm": 0.26010217837913424,
"learning_rate": 5.783946902740304e-05,
"loss": 0.6172,
"step": 309
},
{
"epoch": 2.0945945945945947,
"grad_norm": 0.2968234389164342,
"learning_rate": 5.767039131520733e-05,
"loss": 0.6176,
"step": 310
},
{
"epoch": 2.1013513513513513,
"grad_norm": 0.29905738657768327,
"learning_rate": 5.750092041827618e-05,
"loss": 0.5998,
"step": 311
},
{
"epoch": 2.108108108108108,
"grad_norm": 0.287828129320029,
"learning_rate": 5.7331060107514754e-05,
"loss": 0.597,
"step": 312
},
{
"epoch": 2.114864864864865,
"grad_norm": 0.26831745019142034,
"learning_rate": 5.716081416249307e-05,
"loss": 0.6131,
"step": 313
},
{
"epoch": 2.1216216216216215,
"grad_norm": 0.28531963857984527,
"learning_rate": 5.699018637136192e-05,
"loss": 0.6056,
"step": 314
},
{
"epoch": 2.1283783783783785,
"grad_norm": 0.18651675715743835,
"learning_rate": 5.681918053076858e-05,
"loss": 0.6167,
"step": 315
},
{
"epoch": 2.135135135135135,
"grad_norm": 0.21775207949583675,
"learning_rate": 5.664780044577231e-05,
"loss": 0.6219,
"step": 316
},
{
"epoch": 2.141891891891892,
"grad_norm": 0.25323829124729375,
"learning_rate": 5.6476049929759714e-05,
"loss": 0.6196,
"step": 317
},
{
"epoch": 2.1486486486486487,
"grad_norm": 0.19732115665472313,
"learning_rate": 5.6303932804359857e-05,
"loss": 0.6156,
"step": 318
},
{
"epoch": 2.1554054054054053,
"grad_norm": 0.16849505056239242,
"learning_rate": 5.613145289935926e-05,
"loss": 0.6042,
"step": 319
},
{
"epoch": 2.1621621621621623,
"grad_norm": 0.19969762470742003,
"learning_rate": 5.595861405261666e-05,
"loss": 0.6031,
"step": 320
},
{
"epoch": 2.168918918918919,
"grad_norm": 0.2036043502047779,
"learning_rate": 5.578542010997764e-05,
"loss": 0.633,
"step": 321
},
{
"epoch": 2.175675675675676,
"grad_norm": 0.16807494045865312,
"learning_rate": 5.561187492518903e-05,
"loss": 0.6118,
"step": 322
},
{
"epoch": 2.1824324324324325,
"grad_norm": 0.1717048805694881,
"learning_rate": 5.5437982359813156e-05,
"loss": 0.6116,
"step": 323
},
{
"epoch": 2.189189189189189,
"grad_norm": 0.17594564977890167,
"learning_rate": 5.526374628314195e-05,
"loss": 0.6162,
"step": 324
},
{
"epoch": 2.195945945945946,
"grad_norm": 0.21489547073195236,
"learning_rate": 5.50891705721108e-05,
"loss": 0.5984,
"step": 325
},
{
"epoch": 2.2027027027027026,
"grad_norm": 0.21447613988889852,
"learning_rate": 5.4914259111212355e-05,
"loss": 0.6198,
"step": 326
},
{
"epoch": 2.2094594594594597,
"grad_norm": 0.2207232176733451,
"learning_rate": 5.473901579241e-05,
"loss": 0.6092,
"step": 327
},
{
"epoch": 2.2162162162162162,
"grad_norm": 0.19937702101442842,
"learning_rate": 5.4563444515051354e-05,
"loss": 0.6065,
"step": 328
},
{
"epoch": 2.222972972972973,
"grad_norm": 0.19293343910472205,
"learning_rate": 5.438754918578144e-05,
"loss": 0.6063,
"step": 329
},
{
"epoch": 2.22972972972973,
"grad_norm": 0.17732267437010826,
"learning_rate": 5.4211333718455756e-05,
"loss": 0.6076,
"step": 330
},
{
"epoch": 2.2364864864864864,
"grad_norm": 0.18101481051912677,
"learning_rate": 5.4034802034053223e-05,
"loss": 0.6154,
"step": 331
},
{
"epoch": 2.2432432432432434,
"grad_norm": 0.19133561762748252,
"learning_rate": 5.3857958060588955e-05,
"loss": 0.5988,
"step": 332
},
{
"epoch": 2.25,
"grad_norm": 0.18341857255842758,
"learning_rate": 5.368080573302676e-05,
"loss": 0.6093,
"step": 333
},
{
"epoch": 2.2567567567567566,
"grad_norm": 0.2486112524831487,
"learning_rate": 5.3503348993191706e-05,
"loss": 0.6159,
"step": 334
},
{
"epoch": 2.2635135135135136,
"grad_norm": 0.21869922307823292,
"learning_rate": 5.332559178968231e-05,
"loss": 0.6128,
"step": 335
},
{
"epoch": 2.27027027027027,
"grad_norm": 0.1808704522354482,
"learning_rate": 5.314753807778276e-05,
"loss": 0.6157,
"step": 336
},
{
"epoch": 2.277027027027027,
"grad_norm": 0.19637386821805067,
"learning_rate": 5.296919181937485e-05,
"loss": 0.6206,
"step": 337
},
{
"epoch": 2.2837837837837838,
"grad_norm": 0.1867094463718795,
"learning_rate": 5.279055698284982e-05,
"loss": 0.6209,
"step": 338
},
{
"epoch": 2.2905405405405403,
"grad_norm": 0.17359407157166826,
"learning_rate": 5.261163754302011e-05,
"loss": 0.6012,
"step": 339
},
{
"epoch": 2.2972972972972974,
"grad_norm": 0.18569999759822897,
"learning_rate": 5.2432437481030855e-05,
"loss": 0.6048,
"step": 340
},
{
"epoch": 2.304054054054054,
"grad_norm": 0.1773148171307398,
"learning_rate": 5.225296078427135e-05,
"loss": 0.5963,
"step": 341
},
{
"epoch": 2.310810810810811,
"grad_norm": 0.16243539870707618,
"learning_rate": 5.207321144628628e-05,
"loss": 0.6178,
"step": 342
},
{
"epoch": 2.3175675675675675,
"grad_norm": 0.18693063507419128,
"learning_rate": 5.18931934666869e-05,
"loss": 0.6136,
"step": 343
},
{
"epoch": 2.3243243243243246,
"grad_norm": 0.1661326777898068,
"learning_rate": 5.171291085106202e-05,
"loss": 0.6235,
"step": 344
},
{
"epoch": 2.331081081081081,
"grad_norm": 0.15331403054989767,
"learning_rate": 5.153236761088888e-05,
"loss": 0.6098,
"step": 345
},
{
"epoch": 2.3378378378378377,
"grad_norm": 0.1570395157158325,
"learning_rate": 5.135156776344389e-05,
"loss": 0.6137,
"step": 346
},
{
"epoch": 2.3445945945945947,
"grad_norm": 0.17430865513764107,
"learning_rate": 5.117051533171321e-05,
"loss": 0.6177,
"step": 347
},
{
"epoch": 2.3513513513513513,
"grad_norm": 0.21717070798995733,
"learning_rate": 5.098921434430333e-05,
"loss": 0.6203,
"step": 348
},
{
"epoch": 2.358108108108108,
"grad_norm": 0.20392614118792024,
"learning_rate": 5.080766883535129e-05,
"loss": 0.6136,
"step": 349
},
{
"epoch": 2.364864864864865,
"grad_norm": 0.21588477709369264,
"learning_rate": 5.062588284443505e-05,
"loss": 0.6145,
"step": 350
},
{
"epoch": 2.3716216216216215,
"grad_norm": 0.21049543340334342,
"learning_rate": 5.0443860416483536e-05,
"loss": 0.6019,
"step": 351
},
{
"epoch": 2.3783783783783785,
"grad_norm": 0.1730662628064374,
"learning_rate": 5.026160560168661e-05,
"loss": 0.6058,
"step": 352
},
{
"epoch": 2.385135135135135,
"grad_norm": 0.17831683554874725,
"learning_rate": 5.0079122455405014e-05,
"loss": 0.6208,
"step": 353
},
{
"epoch": 2.391891891891892,
"grad_norm": 0.17744701669503546,
"learning_rate": 4.989641503808011e-05,
"loss": 0.609,
"step": 354
},
{
"epoch": 2.3986486486486487,
"grad_norm": 0.17994097783132235,
"learning_rate": 4.971348741514349e-05,
"loss": 0.6066,
"step": 355
},
{
"epoch": 2.4054054054054053,
"grad_norm": 0.17459787928555573,
"learning_rate": 4.95303436569266e-05,
"loss": 0.6146,
"step": 356
},
{
"epoch": 2.4121621621621623,
"grad_norm": 0.16284962487352878,
"learning_rate": 4.934698783857011e-05,
"loss": 0.6044,
"step": 357
},
{
"epoch": 2.418918918918919,
"grad_norm": 0.17767021224773352,
"learning_rate": 4.91634240399332e-05,
"loss": 0.6211,
"step": 358
},
{
"epoch": 2.4256756756756754,
"grad_norm": 0.1496311992502912,
"learning_rate": 4.8979656345502904e-05,
"loss": 0.602,
"step": 359
},
{
"epoch": 2.4324324324324325,
"grad_norm": 0.16981969578628817,
"learning_rate": 4.8795688844303114e-05,
"loss": 0.6148,
"step": 360
},
{
"epoch": 2.439189189189189,
"grad_norm": 0.2089769520148497,
"learning_rate": 4.861152562980362e-05,
"loss": 0.6123,
"step": 361
},
{
"epoch": 2.445945945945946,
"grad_norm": 0.20528877980388496,
"learning_rate": 4.8427170799829055e-05,
"loss": 0.6095,
"step": 362
},
{
"epoch": 2.4527027027027026,
"grad_norm": 0.1827881621296737,
"learning_rate": 4.824262845646771e-05,
"loss": 0.6112,
"step": 363
},
{
"epoch": 2.4594594594594597,
"grad_norm": 0.19580070665324928,
"learning_rate": 4.805790270598021e-05,
"loss": 0.6209,
"step": 364
},
{
"epoch": 2.4662162162162162,
"grad_norm": 0.23276386582983333,
"learning_rate": 4.787299765870822e-05,
"loss": 0.6083,
"step": 365
},
{
"epoch": 2.472972972972973,
"grad_norm": 0.20338848706607754,
"learning_rate": 4.768791742898292e-05,
"loss": 0.615,
"step": 366
},
{
"epoch": 2.47972972972973,
"grad_norm": 0.1521765603212478,
"learning_rate": 4.7502666135033486e-05,
"loss": 0.6154,
"step": 367
},
{
"epoch": 2.4864864864864864,
"grad_norm": 0.1559952589700639,
"learning_rate": 4.731724789889547e-05,
"loss": 0.6026,
"step": 368
},
{
"epoch": 2.4932432432432434,
"grad_norm": 0.14084577612463858,
"learning_rate": 4.7131666846319036e-05,
"loss": 0.607,
"step": 369
},
{
"epoch": 2.5,
"grad_norm": 0.1450823104838612,
"learning_rate": 4.694592710667723e-05,
"loss": 0.6173,
"step": 370
},
{
"epoch": 2.506756756756757,
"grad_norm": 0.14280439260452932,
"learning_rate": 4.676003281287397e-05,
"loss": 0.6035,
"step": 371
},
{
"epoch": 2.5135135135135136,
"grad_norm": 0.1731629109803712,
"learning_rate": 4.657398810125225e-05,
"loss": 0.6044,
"step": 372
},
{
"epoch": 2.52027027027027,
"grad_norm": 0.14551348661799535,
"learning_rate": 4.638779711150198e-05,
"loss": 0.6223,
"step": 373
},
{
"epoch": 2.527027027027027,
"grad_norm": 0.15724571828258985,
"learning_rate": 4.620146398656792e-05,
"loss": 0.6174,
"step": 374
},
{
"epoch": 2.5337837837837838,
"grad_norm": 0.1622156186595054,
"learning_rate": 4.601499287255748e-05,
"loss": 0.6159,
"step": 375
},
{
"epoch": 2.5405405405405403,
"grad_norm": 0.16326776251404446,
"learning_rate": 4.582838791864846e-05,
"loss": 0.6178,
"step": 376
},
{
"epoch": 2.5472972972972974,
"grad_norm": 0.16705938089441114,
"learning_rate": 4.5641653276996774e-05,
"loss": 0.6069,
"step": 377
},
{
"epoch": 2.554054054054054,
"grad_norm": 0.1240389952109112,
"learning_rate": 4.5454793102644006e-05,
"loss": 0.6129,
"step": 378
},
{
"epoch": 2.560810810810811,
"grad_norm": 0.16500872440232617,
"learning_rate": 4.5267811553424945e-05,
"loss": 0.6213,
"step": 379
},
{
"epoch": 2.5675675675675675,
"grad_norm": 0.17252309025914744,
"learning_rate": 4.5080712789875154e-05,
"loss": 0.6172,
"step": 380
},
{
"epoch": 2.5743243243243246,
"grad_norm": 0.163801171559839,
"learning_rate": 4.489350097513829e-05,
"loss": 0.6205,
"step": 381
},
{
"epoch": 2.581081081081081,
"grad_norm": 0.14139376100405637,
"learning_rate": 4.470618027487354e-05,
"loss": 0.6101,
"step": 382
},
{
"epoch": 2.5878378378378377,
"grad_norm": 0.14839912258194626,
"learning_rate": 4.451875485716292e-05,
"loss": 0.6214,
"step": 383
},
{
"epoch": 2.5945945945945947,
"grad_norm": 0.13849384656147448,
"learning_rate": 4.4331228892418473e-05,
"loss": 0.6111,
"step": 384
},
{
"epoch": 2.6013513513513513,
"grad_norm": 0.16314353286789263,
"learning_rate": 4.414360655328957e-05,
"loss": 0.6201,
"step": 385
},
{
"epoch": 2.608108108108108,
"grad_norm": 0.13166529990196998,
"learning_rate": 4.395589201457e-05,
"loss": 0.6167,
"step": 386
},
{
"epoch": 2.614864864864865,
"grad_norm": 0.14086485553732253,
"learning_rate": 4.376808945310505e-05,
"loss": 0.6219,
"step": 387
},
{
"epoch": 2.6216216216216215,
"grad_norm": 0.15193755776745674,
"learning_rate": 4.358020304769867e-05,
"loss": 0.5994,
"step": 388
},
{
"epoch": 2.6283783783783785,
"grad_norm": 0.14946480137653986,
"learning_rate": 4.339223697902037e-05,
"loss": 0.61,
"step": 389
},
{
"epoch": 2.635135135135135,
"grad_norm": 0.15531265023719382,
"learning_rate": 4.320419542951228e-05,
"loss": 0.6179,
"step": 390
},
{
"epoch": 2.641891891891892,
"grad_norm": 0.14264508419628688,
"learning_rate": 4.3016082583296067e-05,
"loss": 0.6167,
"step": 391
},
{
"epoch": 2.6486486486486487,
"grad_norm": 0.12288922241404748,
"learning_rate": 4.2827902626079784e-05,
"loss": 0.6195,
"step": 392
},
{
"epoch": 2.6554054054054053,
"grad_norm": 0.19330457974395565,
"learning_rate": 4.263965974506483e-05,
"loss": 0.6207,
"step": 393
},
{
"epoch": 2.6621621621621623,
"grad_norm": 0.16883088734435447,
"learning_rate": 4.2451358128852654e-05,
"loss": 0.5989,
"step": 394
},
{
"epoch": 2.668918918918919,
"grad_norm": 0.1599604515688246,
"learning_rate": 4.22630019673517e-05,
"loss": 0.6142,
"step": 395
},
{
"epoch": 2.6756756756756754,
"grad_norm": 0.15518611864455714,
"learning_rate": 4.207459545168405e-05,
"loss": 0.6102,
"step": 396
},
{
"epoch": 2.6824324324324325,
"grad_norm": 0.16742006705568305,
"learning_rate": 4.188614277409224e-05,
"loss": 0.6171,
"step": 397
},
{
"epoch": 2.689189189189189,
"grad_norm": 0.15553602245497833,
"learning_rate": 4.169764812784594e-05,
"loss": 0.6143,
"step": 398
},
{
"epoch": 2.695945945945946,
"grad_norm": 0.1595818544532621,
"learning_rate": 4.1509115707148695e-05,
"loss": 0.6047,
"step": 399
},
{
"epoch": 2.7027027027027026,
"grad_norm": 0.13690560623925258,
"learning_rate": 4.132054970704454e-05,
"loss": 0.6092,
"step": 400
},
{
"epoch": 2.7094594594594597,
"grad_norm": 0.1341753352699138,
"learning_rate": 4.1131954323324734e-05,
"loss": 0.6023,
"step": 401
},
{
"epoch": 2.7162162162162162,
"grad_norm": 0.14452976105100054,
"learning_rate": 4.094333375243428e-05,
"loss": 0.6121,
"step": 402
},
{
"epoch": 2.722972972972973,
"grad_norm": 0.17855418886582433,
"learning_rate": 4.075469219137868e-05,
"loss": 0.6192,
"step": 403
},
{
"epoch": 2.72972972972973,
"grad_norm": 0.15337873200421298,
"learning_rate": 4.056603383763049e-05,
"loss": 0.6137,
"step": 404
},
{
"epoch": 2.7364864864864864,
"grad_norm": 0.14193840867721655,
"learning_rate": 4.0377362889035875e-05,
"loss": 0.5971,
"step": 405
},
{
"epoch": 2.743243243243243,
"grad_norm": 0.14049198222822518,
"learning_rate": 4.0188683543721295e-05,
"loss": 0.61,
"step": 406
},
{
"epoch": 2.75,
"grad_norm": 0.1643045132978014,
"learning_rate": 4e-05,
"loss": 0.6224,
"step": 407
},
{
"epoch": 2.756756756756757,
"grad_norm": 0.14527602919509278,
"learning_rate": 3.981131645627872e-05,
"loss": 0.6164,
"step": 408
},
{
"epoch": 2.7635135135135136,
"grad_norm": 0.1384596570378767,
"learning_rate": 3.9622637110964125e-05,
"loss": 0.6158,
"step": 409
},
{
"epoch": 2.77027027027027,
"grad_norm": 0.18210804820042478,
"learning_rate": 3.943396616236953e-05,
"loss": 0.6149,
"step": 410
},
{
"epoch": 2.777027027027027,
"grad_norm": 0.16943522003893646,
"learning_rate": 3.9245307808621325e-05,
"loss": 0.6159,
"step": 411
},
{
"epoch": 2.7837837837837838,
"grad_norm": 0.16603181052517613,
"learning_rate": 3.905666624756573e-05,
"loss": 0.6186,
"step": 412
},
{
"epoch": 2.7905405405405403,
"grad_norm": 0.15368628411505494,
"learning_rate": 3.886804567667528e-05,
"loss": 0.6155,
"step": 413
},
{
"epoch": 2.7972972972972974,
"grad_norm": 0.1721774875064551,
"learning_rate": 3.867945029295546e-05,
"loss": 0.6099,
"step": 414
},
{
"epoch": 2.804054054054054,
"grad_norm": 0.1893984092920587,
"learning_rate": 3.8490884292851325e-05,
"loss": 0.6146,
"step": 415
},
{
"epoch": 2.810810810810811,
"grad_norm": 0.1792007727966493,
"learning_rate": 3.830235187215408e-05,
"loss": 0.6195,
"step": 416
},
{
"epoch": 2.8175675675675675,
"grad_norm": 0.1569338853010942,
"learning_rate": 3.8113857225907783e-05,
"loss": 0.6173,
"step": 417
},
{
"epoch": 2.8243243243243246,
"grad_norm": 0.17031740028829662,
"learning_rate": 3.792540454831596e-05,
"loss": 0.6163,
"step": 418
},
{
"epoch": 2.831081081081081,
"grad_norm": 0.15004651120970428,
"learning_rate": 3.7736998032648305e-05,
"loss": 0.6047,
"step": 419
},
{
"epoch": 2.8378378378378377,
"grad_norm": 0.142075802466718,
"learning_rate": 3.754864187114736e-05,
"loss": 0.6253,
"step": 420
},
{
"epoch": 2.8445945945945947,
"grad_norm": 0.15599780395707422,
"learning_rate": 3.736034025493519e-05,
"loss": 0.593,
"step": 421
},
{
"epoch": 2.8513513513513513,
"grad_norm": 0.1563201754706643,
"learning_rate": 3.717209737392022e-05,
"loss": 0.6128,
"step": 422
},
{
"epoch": 2.858108108108108,
"grad_norm": 0.15805265922848155,
"learning_rate": 3.698391741670394e-05,
"loss": 0.6059,
"step": 423
},
{
"epoch": 2.864864864864865,
"grad_norm": 0.17933457522155424,
"learning_rate": 3.679580457048772e-05,
"loss": 0.6086,
"step": 424
},
{
"epoch": 2.8716216216216215,
"grad_norm": 0.17371124474695815,
"learning_rate": 3.660776302097965e-05,
"loss": 0.6079,
"step": 425
},
{
"epoch": 2.8783783783783785,
"grad_norm": 0.1315584411453832,
"learning_rate": 3.641979695230135e-05,
"loss": 0.6039,
"step": 426
},
{
"epoch": 2.885135135135135,
"grad_norm": 0.148272614562961,
"learning_rate": 3.6231910546894956e-05,
"loss": 0.6161,
"step": 427
},
{
"epoch": 2.891891891891892,
"grad_norm": 0.12013666775227018,
"learning_rate": 3.6044107985430015e-05,
"loss": 0.6113,
"step": 428
},
{
"epoch": 2.8986486486486487,
"grad_norm": 0.15454206210883253,
"learning_rate": 3.585639344671043e-05,
"loss": 0.6113,
"step": 429
},
{
"epoch": 2.9054054054054053,
"grad_norm": 0.130251935264837,
"learning_rate": 3.5668771107581526e-05,
"loss": 0.613,
"step": 430
},
{
"epoch": 2.9121621621621623,
"grad_norm": 0.1370783380801396,
"learning_rate": 3.5481245142837095e-05,
"loss": 0.6168,
"step": 431
},
{
"epoch": 2.918918918918919,
"grad_norm": 0.13116479132749334,
"learning_rate": 3.5293819725126464e-05,
"loss": 0.6136,
"step": 432
},
{
"epoch": 2.9256756756756754,
"grad_norm": 0.14031976128589324,
"learning_rate": 3.5106499024861715e-05,
"loss": 0.6175,
"step": 433
},
{
"epoch": 2.9324324324324325,
"grad_norm": 0.12054376627925889,
"learning_rate": 3.491928721012485e-05,
"loss": 0.6134,
"step": 434
},
{
"epoch": 2.939189189189189,
"grad_norm": 0.14272909957039542,
"learning_rate": 3.4732188446575055e-05,
"loss": 0.6096,
"step": 435
},
{
"epoch": 2.945945945945946,
"grad_norm": 0.1343117824779506,
"learning_rate": 3.454520689735602e-05,
"loss": 0.6203,
"step": 436
},
{
"epoch": 2.9527027027027026,
"grad_norm": 0.11376422264610185,
"learning_rate": 3.435834672300324e-05,
"loss": 0.6057,
"step": 437
},
{
"epoch": 2.9594594594594597,
"grad_norm": 0.14208820005659473,
"learning_rate": 3.417161208135155e-05,
"loss": 0.6193,
"step": 438
},
{
"epoch": 2.9662162162162162,
"grad_norm": 0.11247340517088436,
"learning_rate": 3.398500712744254e-05,
"loss": 0.6218,
"step": 439
},
{
"epoch": 2.972972972972973,
"grad_norm": 0.12141011100880744,
"learning_rate": 3.379853601343209e-05,
"loss": 0.6124,
"step": 440
},
{
"epoch": 2.97972972972973,
"grad_norm": 0.11441162073496848,
"learning_rate": 3.361220288849804e-05,
"loss": 0.6146,
"step": 441
},
{
"epoch": 2.9864864864864864,
"grad_norm": 0.11128267132483809,
"learning_rate": 3.342601189874777e-05,
"loss": 0.6141,
"step": 442
},
{
"epoch": 2.993243243243243,
"grad_norm": 0.13308500191665273,
"learning_rate": 3.323996718712605e-05,
"loss": 0.6154,
"step": 443
},
{
"epoch": 3.0,
"grad_norm": 0.13634885719680506,
"learning_rate": 3.305407289332279e-05,
"loss": 0.6063,
"step": 444
},
{
"epoch": 3.0067567567567566,
"grad_norm": 0.14117643344569666,
"learning_rate": 3.2868333153680964e-05,
"loss": 0.5952,
"step": 445
},
{
"epoch": 3.0135135135135136,
"grad_norm": 0.1441355574419299,
"learning_rate": 3.2682752101104536e-05,
"loss": 0.5807,
"step": 446
},
{
"epoch": 3.02027027027027,
"grad_norm": 0.13317791462913253,
"learning_rate": 3.249733386496653e-05,
"loss": 0.6011,
"step": 447
},
{
"epoch": 3.027027027027027,
"grad_norm": 0.1488227563399244,
"learning_rate": 3.231208257101709e-05,
"loss": 0.6055,
"step": 448
},
{
"epoch": 3.0337837837837838,
"grad_norm": 0.13698170944523078,
"learning_rate": 3.212700234129179e-05,
"loss": 0.6012,
"step": 449
},
{
"epoch": 3.0405405405405403,
"grad_norm": 0.13426258482203746,
"learning_rate": 3.194209729401979e-05,
"loss": 0.5955,
"step": 450
},
{
"epoch": 3.0472972972972974,
"grad_norm": 0.12763528667729904,
"learning_rate": 3.175737154353231e-05,
"loss": 0.5829,
"step": 451
},
{
"epoch": 3.054054054054054,
"grad_norm": 0.12302649236306719,
"learning_rate": 3.157282920017096e-05,
"loss": 0.6031,
"step": 452
},
{
"epoch": 3.060810810810811,
"grad_norm": 0.12836088172431226,
"learning_rate": 3.1388474370196395e-05,
"loss": 0.6029,
"step": 453
},
{
"epoch": 3.0675675675675675,
"grad_norm": 0.12713682511389784,
"learning_rate": 3.12043111556969e-05,
"loss": 0.5898,
"step": 454
},
{
"epoch": 3.074324324324324,
"grad_norm": 0.13954610740683865,
"learning_rate": 3.1020343654497096e-05,
"loss": 0.5882,
"step": 455
},
{
"epoch": 3.081081081081081,
"grad_norm": 0.14394153328876694,
"learning_rate": 3.083657596006681e-05,
"loss": 0.5878,
"step": 456
},
{
"epoch": 3.0878378378378377,
"grad_norm": 0.11738094672449853,
"learning_rate": 3.065301216142991e-05,
"loss": 0.5916,
"step": 457
},
{
"epoch": 3.0945945945945947,
"grad_norm": 0.1473396296692266,
"learning_rate": 3.046965634307341e-05,
"loss": 0.6003,
"step": 458
},
{
"epoch": 3.1013513513513513,
"grad_norm": 0.12128542181194722,
"learning_rate": 3.028651258485652e-05,
"loss": 0.5807,
"step": 459
},
{
"epoch": 3.108108108108108,
"grad_norm": 0.12076772408917641,
"learning_rate": 3.010358496191991e-05,
"loss": 0.5968,
"step": 460
},
{
"epoch": 3.114864864864865,
"grad_norm": 0.12207943508497456,
"learning_rate": 2.9920877544595002e-05,
"loss": 0.5911,
"step": 461
},
{
"epoch": 3.1216216216216215,
"grad_norm": 0.1193967138970814,
"learning_rate": 2.9738394398313405e-05,
"loss": 0.5867,
"step": 462
},
{
"epoch": 3.1283783783783785,
"grad_norm": 0.11942196780407818,
"learning_rate": 2.955613958351647e-05,
"loss": 0.5927,
"step": 463
},
{
"epoch": 3.135135135135135,
"grad_norm": 0.12137977118056532,
"learning_rate": 2.9374117155564957e-05,
"loss": 0.5923,
"step": 464
},
{
"epoch": 3.141891891891892,
"grad_norm": 0.1204663577707489,
"learning_rate": 2.919233116464872e-05,
"loss": 0.5807,
"step": 465
},
{
"epoch": 3.1486486486486487,
"grad_norm": 0.1277645131662919,
"learning_rate": 2.9010785655696698e-05,
"loss": 0.602,
"step": 466
},
{
"epoch": 3.1554054054054053,
"grad_norm": 0.14373007650684674,
"learning_rate": 2.88294846682868e-05,
"loss": 0.5952,
"step": 467
},
{
"epoch": 3.1621621621621623,
"grad_norm": 0.12531967044629988,
"learning_rate": 2.864843223655613e-05,
"loss": 0.6065,
"step": 468
},
{
"epoch": 3.168918918918919,
"grad_norm": 0.1344060345828822,
"learning_rate": 2.8467632389111126e-05,
"loss": 0.5879,
"step": 469
},
{
"epoch": 3.175675675675676,
"grad_norm": 0.11307986878514827,
"learning_rate": 2.828708914893799e-05,
"loss": 0.5923,
"step": 470
},
{
"epoch": 3.1824324324324325,
"grad_norm": 0.1205066746149822,
"learning_rate": 2.8106806533313106e-05,
"loss": 0.6042,
"step": 471
},
{
"epoch": 3.189189189189189,
"grad_norm": 0.11358248639127554,
"learning_rate": 2.7926788553713734e-05,
"loss": 0.6017,
"step": 472
},
{
"epoch": 3.195945945945946,
"grad_norm": 0.11273620191303046,
"learning_rate": 2.7747039215728667e-05,
"loss": 0.5909,
"step": 473
},
{
"epoch": 3.2027027027027026,
"grad_norm": 0.1527832963288818,
"learning_rate": 2.7567562518969155e-05,
"loss": 0.5884,
"step": 474
},
{
"epoch": 3.2094594594594597,
"grad_norm": 0.1251866201941187,
"learning_rate": 2.7388362456979906e-05,
"loss": 0.5942,
"step": 475
},
{
"epoch": 3.2162162162162162,
"grad_norm": 0.131897788334417,
"learning_rate": 2.7209443017150193e-05,
"loss": 0.6028,
"step": 476
},
{
"epoch": 3.222972972972973,
"grad_norm": 0.10949010701632424,
"learning_rate": 2.703080818062517e-05,
"loss": 0.5899,
"step": 477
},
{
"epoch": 3.22972972972973,
"grad_norm": 0.1085593433907961,
"learning_rate": 2.6852461922217253e-05,
"loss": 0.5915,
"step": 478
},
{
"epoch": 3.2364864864864864,
"grad_norm": 0.10851474030600673,
"learning_rate": 2.66744082103177e-05,
"loss": 0.5956,
"step": 479
},
{
"epoch": 3.2432432432432434,
"grad_norm": 0.10556652124147405,
"learning_rate": 2.6496651006808308e-05,
"loss": 0.5923,
"step": 480
},
{
"epoch": 3.25,
"grad_norm": 0.11511281568946971,
"learning_rate": 2.6319194266973256e-05,
"loss": 0.5982,
"step": 481
},
{
"epoch": 3.2567567567567566,
"grad_norm": 0.10661745620059551,
"learning_rate": 2.614204193941107e-05,
"loss": 0.6029,
"step": 482
},
{
"epoch": 3.2635135135135136,
"grad_norm": 0.12822931878201987,
"learning_rate": 2.5965197965946783e-05,
"loss": 0.5935,
"step": 483
},
{
"epoch": 3.27027027027027,
"grad_norm": 0.1156142465379123,
"learning_rate": 2.5788666281544258e-05,
"loss": 0.6142,
"step": 484
},
{
"epoch": 3.277027027027027,
"grad_norm": 0.13148336347712752,
"learning_rate": 2.561245081421857e-05,
"loss": 0.5942,
"step": 485
},
{
"epoch": 3.2837837837837838,
"grad_norm": 0.10123146621891457,
"learning_rate": 2.5436555484948643e-05,
"loss": 0.5989,
"step": 486
},
{
"epoch": 3.2905405405405403,
"grad_norm": 0.1375355982170109,
"learning_rate": 2.5260984207590015e-05,
"loss": 0.5913,
"step": 487
},
{
"epoch": 3.2972972972972974,
"grad_norm": 0.1071253679793487,
"learning_rate": 2.5085740888787662e-05,
"loss": 0.6034,
"step": 488
},
{
"epoch": 3.304054054054054,
"grad_norm": 0.13183672834367724,
"learning_rate": 2.4910829427889205e-05,
"loss": 0.6016,
"step": 489
},
{
"epoch": 3.310810810810811,
"grad_norm": 0.11417300702406326,
"learning_rate": 2.473625371685806e-05,
"loss": 0.5964,
"step": 490
},
{
"epoch": 3.3175675675675675,
"grad_norm": 0.11783207513787236,
"learning_rate": 2.4562017640186847e-05,
"loss": 0.5983,
"step": 491
},
{
"epoch": 3.3243243243243246,
"grad_norm": 0.12341779828696402,
"learning_rate": 2.4388125074810986e-05,
"loss": 0.5925,
"step": 492
},
{
"epoch": 3.331081081081081,
"grad_norm": 0.11637168120668157,
"learning_rate": 2.4214579890022373e-05,
"loss": 0.5771,
"step": 493
},
{
"epoch": 3.3378378378378377,
"grad_norm": 0.10923183772815541,
"learning_rate": 2.404138594738335e-05,
"loss": 0.5919,
"step": 494
},
{
"epoch": 3.3445945945945947,
"grad_norm": 0.10526844464204445,
"learning_rate": 2.386854710064075e-05,
"loss": 0.607,
"step": 495
},
{
"epoch": 3.3513513513513513,
"grad_norm": 0.10474734642926019,
"learning_rate": 2.369606719564015e-05,
"loss": 0.5907,
"step": 496
},
{
"epoch": 3.358108108108108,
"grad_norm": 0.11107800655708087,
"learning_rate": 2.35239500702403e-05,
"loss": 0.5981,
"step": 497
},
{
"epoch": 3.364864864864865,
"grad_norm": 0.10187850179282668,
"learning_rate": 2.3352199554227698e-05,
"loss": 0.5938,
"step": 498
},
{
"epoch": 3.3716216216216215,
"grad_norm": 0.12632250115529653,
"learning_rate": 2.318081946923144e-05,
"loss": 0.6117,
"step": 499
},
{
"epoch": 3.3783783783783785,
"grad_norm": 0.10055327704116511,
"learning_rate": 2.3009813628638085e-05,
"loss": 0.5964,
"step": 500
},
{
"epoch": 3.385135135135135,
"grad_norm": 0.11916682604562856,
"learning_rate": 2.283918583750695e-05,
"loss": 0.5871,
"step": 501
},
{
"epoch": 3.391891891891892,
"grad_norm": 0.1010245925435501,
"learning_rate": 2.266893989248527e-05,
"loss": 0.6004,
"step": 502
},
{
"epoch": 3.3986486486486487,
"grad_norm": 0.10446456567129823,
"learning_rate": 2.2499079581723846e-05,
"loss": 0.5942,
"step": 503
},
{
"epoch": 3.4054054054054053,
"grad_norm": 0.10099772014538502,
"learning_rate": 2.2329608684792676e-05,
"loss": 0.5956,
"step": 504
},
{
"epoch": 3.4121621621621623,
"grad_norm": 0.1005267977448251,
"learning_rate": 2.216053097259697e-05,
"loss": 0.5898,
"step": 505
},
{
"epoch": 3.418918918918919,
"grad_norm": 0.10513240417556334,
"learning_rate": 2.1991850207293064e-05,
"loss": 0.5997,
"step": 506
},
{
"epoch": 3.4256756756756754,
"grad_norm": 0.09929924802833451,
"learning_rate": 2.1823570142204902e-05,
"loss": 0.5932,
"step": 507
},
{
"epoch": 3.4324324324324325,
"grad_norm": 0.10494976279274904,
"learning_rate": 2.1655694521740376e-05,
"loss": 0.5935,
"step": 508
},
{
"epoch": 3.439189189189189,
"grad_norm": 0.10732590635113842,
"learning_rate": 2.1488227081308054e-05,
"loss": 0.589,
"step": 509
},
{
"epoch": 3.445945945945946,
"grad_norm": 0.1154273309217794,
"learning_rate": 2.132117154723408e-05,
"loss": 0.5927,
"step": 510
},
{
"epoch": 3.4527027027027026,
"grad_norm": 0.10250291685305624,
"learning_rate": 2.115453163667929e-05,
"loss": 0.6047,
"step": 511
},
{
"epoch": 3.4594594594594597,
"grad_norm": 0.1053185431025155,
"learning_rate": 2.0988311057556397e-05,
"loss": 0.5841,
"step": 512
},
{
"epoch": 3.4662162162162162,
"grad_norm": 0.09414135194355852,
"learning_rate": 2.0822513508447608e-05,
"loss": 0.5863,
"step": 513
},
{
"epoch": 3.472972972972973,
"grad_norm": 0.10071155506936441,
"learning_rate": 2.065714267852223e-05,
"loss": 0.5923,
"step": 514
},
{
"epoch": 3.47972972972973,
"grad_norm": 0.10121850558468308,
"learning_rate": 2.049220224745463e-05,
"loss": 0.5962,
"step": 515
},
{
"epoch": 3.4864864864864864,
"grad_norm": 0.10161225481650848,
"learning_rate": 2.032769588534233e-05,
"loss": 0.5925,
"step": 516
},
{
"epoch": 3.4932432432432434,
"grad_norm": 0.09239639095614875,
"learning_rate": 2.0163627252624427e-05,
"loss": 0.5944,
"step": 517
},
{
"epoch": 3.5,
"grad_norm": 0.11360448222663692,
"learning_rate": 2.0000000000000012e-05,
"loss": 0.6116,
"step": 518
},
{
"epoch": 3.506756756756757,
"grad_norm": 0.09631002455489597,
"learning_rate": 1.9836817768347015e-05,
"loss": 0.5947,
"step": 519
},
{
"epoch": 3.5135135135135136,
"grad_norm": 0.1167172974336559,
"learning_rate": 1.9674084188641235e-05,
"loss": 0.5843,
"step": 520
},
{
"epoch": 3.52027027027027,
"grad_norm": 0.10349743444704453,
"learning_rate": 1.9511802881875438e-05,
"loss": 0.6026,
"step": 521
},
{
"epoch": 3.527027027027027,
"grad_norm": 0.11753598782061575,
"learning_rate": 1.9349977458978846e-05,
"loss": 0.5892,
"step": 522
},
{
"epoch": 3.5337837837837838,
"grad_norm": 0.09146602359405204,
"learning_rate": 1.9188611520736846e-05,
"loss": 0.5916,
"step": 523
},
{
"epoch": 3.5405405405405403,
"grad_norm": 0.10892529356211986,
"learning_rate": 1.902770865771074e-05,
"loss": 0.5854,
"step": 524
},
{
"epoch": 3.5472972972972974,
"grad_norm": 0.09418036132529382,
"learning_rate": 1.886727245015794e-05,
"loss": 0.5993,
"step": 525
},
{
"epoch": 3.554054054054054,
"grad_norm": 0.1025611805390442,
"learning_rate": 1.8707306467952323e-05,
"loss": 0.5916,
"step": 526
},
{
"epoch": 3.560810810810811,
"grad_norm": 0.08852437916251449,
"learning_rate": 1.8547814270504705e-05,
"loss": 0.5946,
"step": 527
},
{
"epoch": 3.5675675675675675,
"grad_norm": 0.0999923376365888,
"learning_rate": 1.838879940668373e-05,
"loss": 0.6,
"step": 528
},
{
"epoch": 3.5743243243243246,
"grad_norm": 0.10873141849605454,
"learning_rate": 1.823026541473684e-05,
"loss": 0.6017,
"step": 529
},
{
"epoch": 3.581081081081081,
"grad_norm": 0.10599260470783008,
"learning_rate": 1.8072215822211613e-05,
"loss": 0.5968,
"step": 530
},
{
"epoch": 3.5878378378378377,
"grad_norm": 0.11188333764585037,
"learning_rate": 1.7914654145877187e-05,
"loss": 0.6064,
"step": 531
},
{
"epoch": 3.5945945945945947,
"grad_norm": 0.09583819123841314,
"learning_rate": 1.77575838916461e-05,
"loss": 0.5876,
"step": 532
},
{
"epoch": 3.6013513513513513,
"grad_norm": 0.10206469144841168,
"learning_rate": 1.760100855449619e-05,
"loss": 0.5948,
"step": 533
},
{
"epoch": 3.608108108108108,
"grad_norm": 0.10036289538603702,
"learning_rate": 1.7444931618392894e-05,
"loss": 0.6107,
"step": 534
},
{
"epoch": 3.614864864864865,
"grad_norm": 0.08314263844698841,
"learning_rate": 1.7289356556211687e-05,
"loss": 0.6059,
"step": 535
},
{
"epoch": 3.6216216216216215,
"grad_norm": 0.10484951196819213,
"learning_rate": 1.7134286829660855e-05,
"loss": 0.5961,
"step": 536
},
{
"epoch": 3.6283783783783785,
"grad_norm": 0.07757732001833932,
"learning_rate": 1.697972588920439e-05,
"loss": 0.603,
"step": 537
},
{
"epoch": 3.635135135135135,
"grad_norm": 0.09636592692846442,
"learning_rate": 1.6825677173985332e-05,
"loss": 0.597,
"step": 538
},
{
"epoch": 3.641891891891892,
"grad_norm": 0.0934961698224681,
"learning_rate": 1.6672144111749066e-05,
"loss": 0.6065,
"step": 539
},
{
"epoch": 3.6486486486486487,
"grad_norm": 0.0956796061229755,
"learning_rate": 1.6519130118767258e-05,
"loss": 0.5978,
"step": 540
},
{
"epoch": 3.6554054054054053,
"grad_norm": 0.1161980558323059,
"learning_rate": 1.6366638599761676e-05,
"loss": 0.5914,
"step": 541
},
{
"epoch": 3.6621621621621623,
"grad_norm": 0.08312306306871219,
"learning_rate": 1.621467294782854e-05,
"loss": 0.5855,
"step": 542
},
{
"epoch": 3.668918918918919,
"grad_norm": 0.11701156024970683,
"learning_rate": 1.606323654436293e-05,
"loss": 0.5904,
"step": 543
},
{
"epoch": 3.6756756756756754,
"grad_norm": 0.09466465341207426,
"learning_rate": 1.591233275898363e-05,
"loss": 0.5915,
"step": 544
},
{
"epoch": 3.6824324324324325,
"grad_norm": 0.10284506462525071,
"learning_rate": 1.5761964949458076e-05,
"loss": 0.5847,
"step": 545
},
{
"epoch": 3.689189189189189,
"grad_norm": 0.10382209320021844,
"learning_rate": 1.5612136461627726e-05,
"loss": 0.596,
"step": 546
},
{
"epoch": 3.695945945945946,
"grad_norm": 0.09709377064685741,
"learning_rate": 1.546285062933352e-05,
"loss": 0.5909,
"step": 547
},
{
"epoch": 3.7027027027027026,
"grad_norm": 0.11272439075208067,
"learning_rate": 1.5314110774341803e-05,
"loss": 0.5953,
"step": 548
},
{
"epoch": 3.7094594594594597,
"grad_norm": 0.11904326038877579,
"learning_rate": 1.5165920206270257e-05,
"loss": 0.5929,
"step": 549
},
{
"epoch": 3.7162162162162162,
"grad_norm": 0.1039606133991536,
"learning_rate": 1.5018282222514451e-05,
"loss": 0.5877,
"step": 550
},
{
"epoch": 3.722972972972973,
"grad_norm": 0.11190632151336002,
"learning_rate": 1.4871200108174306e-05,
"loss": 0.5917,
"step": 551
},
{
"epoch": 3.72972972972973,
"grad_norm": 0.09272507731002902,
"learning_rate": 1.4724677135981118e-05,
"loss": 0.5959,
"step": 552
},
{
"epoch": 3.7364864864864864,
"grad_norm": 0.09431866458492948,
"learning_rate": 1.457871656622463e-05,
"loss": 0.5887,
"step": 553
},
{
"epoch": 3.743243243243243,
"grad_norm": 0.08680813466014699,
"learning_rate": 1.4433321646680614e-05,
"loss": 0.6032,
"step": 554
},
{
"epoch": 3.75,
"grad_norm": 0.0844092691423774,
"learning_rate": 1.4288495612538427e-05,
"loss": 0.6062,
"step": 555
},
{
"epoch": 3.756756756756757,
"grad_norm": 0.0874893569172525,
"learning_rate": 1.4144241686329236e-05,
"loss": 0.5833,
"step": 556
},
{
"epoch": 3.7635135135135136,
"grad_norm": 0.10957613038759607,
"learning_rate": 1.400056307785413e-05,
"loss": 0.6056,
"step": 557
},
{
"epoch": 3.77027027027027,
"grad_norm": 0.07882052024046089,
"learning_rate": 1.3857462984112831e-05,
"loss": 0.5912,
"step": 558
},
{
"epoch": 3.777027027027027,
"grad_norm": 0.08565900595306047,
"learning_rate": 1.371494458923246e-05,
"loss": 0.5919,
"step": 559
},
{
"epoch": 3.7837837837837838,
"grad_norm": 0.09674275812299338,
"learning_rate": 1.3573011064396751e-05,
"loss": 0.5985,
"step": 560
},
{
"epoch": 3.7905405405405403,
"grad_norm": 0.08500752883113108,
"learning_rate": 1.3431665567775439e-05,
"loss": 0.5983,
"step": 561
},
{
"epoch": 3.7972972972972974,
"grad_norm": 0.11583654081493171,
"learning_rate": 1.3290911244454066e-05,
"loss": 0.5903,
"step": 562
},
{
"epoch": 3.804054054054054,
"grad_norm": 0.07940239999432974,
"learning_rate": 1.3150751226363886e-05,
"loss": 0.5931,
"step": 563
},
{
"epoch": 3.810810810810811,
"grad_norm": 0.10093820387387048,
"learning_rate": 1.3011188632212307e-05,
"loss": 0.6033,
"step": 564
},
{
"epoch": 3.8175675675675675,
"grad_norm": 0.0890166406955722,
"learning_rate": 1.2872226567413346e-05,
"loss": 0.5924,
"step": 565
},
{
"epoch": 3.8243243243243246,
"grad_norm": 0.07708649898799794,
"learning_rate": 1.2733868124018694e-05,
"loss": 0.5891,
"step": 566
},
{
"epoch": 3.831081081081081,
"grad_norm": 0.11011562257155211,
"learning_rate": 1.2596116380648761e-05,
"loss": 0.5898,
"step": 567
},
{
"epoch": 3.8378378378378377,
"grad_norm": 0.07996964493884025,
"learning_rate": 1.2458974402424312e-05,
"loss": 0.5937,
"step": 568
},
{
"epoch": 3.8445945945945947,
"grad_norm": 0.09078629614798739,
"learning_rate": 1.2322445240898158e-05,
"loss": 0.5908,
"step": 569
},
{
"epoch": 3.8513513513513513,
"grad_norm": 0.08453528616636899,
"learning_rate": 1.2186531933987294e-05,
"loss": 0.5895,
"step": 570
},
{
"epoch": 3.858108108108108,
"grad_norm": 0.08207256114306587,
"learning_rate": 1.2051237505905302e-05,
"loss": 0.6,
"step": 571
},
{
"epoch": 3.864864864864865,
"grad_norm": 0.08328736066936966,
"learning_rate": 1.19165649670951e-05,
"loss": 0.5901,
"step": 572
},
{
"epoch": 3.8716216216216215,
"grad_norm": 0.0796730356081974,
"learning_rate": 1.1782517314161872e-05,
"loss": 0.5825,
"step": 573
},
{
"epoch": 3.8783783783783785,
"grad_norm": 0.09614075098109777,
"learning_rate": 1.164909752980648e-05,
"loss": 0.5949,
"step": 574
},
{
"epoch": 3.885135135135135,
"grad_norm": 0.08310475221153353,
"learning_rate": 1.1516308582758983e-05,
"loss": 0.604,
"step": 575
},
{
"epoch": 3.891891891891892,
"grad_norm": 0.08282668808548053,
"learning_rate": 1.1384153427712729e-05,
"loss": 0.6002,
"step": 576
},
{
"epoch": 3.8986486486486487,
"grad_norm": 0.0986705467432139,
"learning_rate": 1.1252635005258466e-05,
"loss": 0.5853,
"step": 577
},
{
"epoch": 3.9054054054054053,
"grad_norm": 0.08062037787963591,
"learning_rate": 1.1121756241819023e-05,
"loss": 0.6047,
"step": 578
},
{
"epoch": 3.9121621621621623,
"grad_norm": 0.07523397720678549,
"learning_rate": 1.0991520049584112e-05,
"loss": 0.5842,
"step": 579
},
{
"epoch": 3.918918918918919,
"grad_norm": 0.08533444706267539,
"learning_rate": 1.0861929326445572e-05,
"loss": 0.5952,
"step": 580
},
{
"epoch": 3.9256756756756754,
"grad_norm": 0.0867340533978853,
"learning_rate": 1.0732986955932869e-05,
"loss": 0.5978,
"step": 581
},
{
"epoch": 3.9324324324324325,
"grad_norm": 0.06754168833386048,
"learning_rate": 1.0604695807148971e-05,
"loss": 0.5987,
"step": 582
},
{
"epoch": 3.939189189189189,
"grad_norm": 0.0796662865647217,
"learning_rate": 1.0477058734706436e-05,
"loss": 0.588,
"step": 583
},
{
"epoch": 3.945945945945946,
"grad_norm": 0.0813837559182795,
"learning_rate": 1.0350078578664005e-05,
"loss": 0.6009,
"step": 584
},
{
"epoch": 3.9527027027027026,
"grad_norm": 0.06912060191449508,
"learning_rate": 1.0223758164463246e-05,
"loss": 0.6055,
"step": 585
},
{
"epoch": 3.9594594594594597,
"grad_norm": 0.06784642823506538,
"learning_rate": 1.0098100302865865e-05,
"loss": 0.5882,
"step": 586
},
{
"epoch": 3.9662162162162162,
"grad_norm": 0.0760135339338161,
"learning_rate": 9.973107789891024e-06,
"loss": 0.596,
"step": 587
},
{
"epoch": 3.972972972972973,
"grad_norm": 0.07657252850751217,
"learning_rate": 9.848783406753224e-06,
"loss": 0.5962,
"step": 588
},
{
"epoch": 3.97972972972973,
"grad_norm": 0.08050666474861323,
"learning_rate": 9.725129919800339e-06,
"loss": 0.5957,
"step": 589
},
{
"epoch": 3.9864864864864864,
"grad_norm": 0.0778632825822388,
"learning_rate": 9.60215008045211e-06,
"loss": 0.5952,
"step": 590
},
{
"epoch": 3.993243243243243,
"grad_norm": 0.07984922923288139,
"learning_rate": 9.479846625138909e-06,
"loss": 0.6023,
"step": 591
},
{
"epoch": 4.0,
"grad_norm": 0.0916488655332951,
"learning_rate": 9.358222275240884e-06,
"loss": 0.5842,
"step": 592
},
{
"epoch": 4.006756756756757,
"grad_norm": 0.08766439343096606,
"learning_rate": 9.237279737027326e-06,
"loss": 0.5851,
"step": 593
},
{
"epoch": 4.013513513513513,
"grad_norm": 0.07964740730111751,
"learning_rate": 9.117021701596567e-06,
"loss": 0.5776,
"step": 594
},
{
"epoch": 4.02027027027027,
"grad_norm": 0.08933138991359177,
"learning_rate": 8.99745084481594e-06,
"loss": 0.5953,
"step": 595
},
{
"epoch": 4.027027027027027,
"grad_norm": 0.07914528740585995,
"learning_rate": 8.87856982726243e-06,
"loss": 0.5892,
"step": 596
},
{
"epoch": 4.033783783783784,
"grad_norm": 0.07297723048835614,
"learning_rate": 8.7603812941633e-06,
"loss": 0.5823,
"step": 597
},
{
"epoch": 4.04054054054054,
"grad_norm": 0.10229836912930287,
"learning_rate": 8.642887875337376e-06,
"loss": 0.5767,
"step": 598
},
{
"epoch": 4.047297297297297,
"grad_norm": 0.0781799178126612,
"learning_rate": 8.526092185136394e-06,
"loss": 0.5784,
"step": 599
},
{
"epoch": 4.054054054054054,
"grad_norm": 0.09402345888004014,
"learning_rate": 8.409996822386972e-06,
"loss": 0.5828,
"step": 600
},
{
"epoch": 4.0608108108108105,
"grad_norm": 0.07679009840525959,
"learning_rate": 8.294604370332613e-06,
"loss": 0.5754,
"step": 601
},
{
"epoch": 4.0675675675675675,
"grad_norm": 0.07923822298946714,
"learning_rate": 8.17991739657641e-06,
"loss": 0.573,
"step": 602
},
{
"epoch": 4.074324324324325,
"grad_norm": 0.08082893460744046,
"learning_rate": 8.06593845302376e-06,
"loss": 0.5916,
"step": 603
},
{
"epoch": 4.081081081081081,
"grad_norm": 0.07578230123409459,
"learning_rate": 7.952670075825702e-06,
"loss": 0.5851,
"step": 604
},
{
"epoch": 4.087837837837838,
"grad_norm": 0.07539180401042113,
"learning_rate": 7.840114785322384e-06,
"loss": 0.5939,
"step": 605
},
{
"epoch": 4.094594594594595,
"grad_norm": 0.07932706141071251,
"learning_rate": 7.728275085987041e-06,
"loss": 0.5908,
"step": 606
},
{
"epoch": 4.101351351351352,
"grad_norm": 0.07344646624670495,
"learning_rate": 7.6171534663702416e-06,
"loss": 0.5915,
"step": 607
},
{
"epoch": 4.108108108108108,
"grad_norm": 0.07204192586718106,
"learning_rate": 7.50675239904457e-06,
"loss": 0.5855,
"step": 608
},
{
"epoch": 4.114864864864865,
"grad_norm": 0.06717242761020562,
"learning_rate": 7.397074340549508e-06,
"loss": 0.5919,
"step": 609
},
{
"epoch": 4.121621621621622,
"grad_norm": 0.07615198090948691,
"learning_rate": 7.288121731336901e-06,
"loss": 0.5883,
"step": 610
},
{
"epoch": 4.128378378378378,
"grad_norm": 0.07186386521547546,
"learning_rate": 7.1798969957165025e-06,
"loss": 0.5666,
"step": 611
},
{
"epoch": 4.135135135135135,
"grad_norm": 0.07771245688676058,
"learning_rate": 7.072402541802197e-06,
"loss": 0.572,
"step": 612
},
{
"epoch": 4.141891891891892,
"grad_norm": 0.06560942291485741,
"learning_rate": 6.965640761458274e-06,
"loss": 0.5853,
"step": 613
},
{
"epoch": 4.148648648648648,
"grad_norm": 0.07554536141277478,
"learning_rate": 6.859614030246318e-06,
"loss": 0.5746,
"step": 614
},
{
"epoch": 4.155405405405405,
"grad_norm": 0.0766866599574075,
"learning_rate": 6.754324707372264e-06,
"loss": 0.5827,
"step": 615
},
{
"epoch": 4.162162162162162,
"grad_norm": 0.0685589990578116,
"learning_rate": 6.649775135633944e-06,
"loss": 0.5708,
"step": 616
},
{
"epoch": 4.168918918918919,
"grad_norm": 0.06830496975914782,
"learning_rate": 6.545967641368958e-06,
"loss": 0.5772,
"step": 617
},
{
"epoch": 4.175675675675675,
"grad_norm": 0.06792930742316021,
"learning_rate": 6.4429045344029136e-06,
"loss": 0.5801,
"step": 618
},
{
"epoch": 4.1824324324324325,
"grad_norm": 0.0728029236733438,
"learning_rate": 6.340588107997994e-06,
"loss": 0.5666,
"step": 619
},
{
"epoch": 4.1891891891891895,
"grad_norm": 0.07156462363762543,
"learning_rate": 6.239020638801987e-06,
"loss": 0.5848,
"step": 620
},
{
"epoch": 4.195945945945946,
"grad_norm": 0.06757664374848818,
"learning_rate": 6.1382043867975836e-06,
"loss": 0.5679,
"step": 621
},
{
"epoch": 4.202702702702703,
"grad_norm": 0.07016074613879056,
"learning_rate": 6.038141595252094e-06,
"loss": 0.5839,
"step": 622
},
{
"epoch": 4.20945945945946,
"grad_norm": 0.06809385798805276,
"learning_rate": 5.9388344906675485e-06,
"loss": 0.575,
"step": 623
},
{
"epoch": 4.216216216216216,
"grad_norm": 0.061129888925191694,
"learning_rate": 5.840285282731173e-06,
"loss": 0.5914,
"step": 624
},
{
"epoch": 4.222972972972973,
"grad_norm": 0.0728841561672064,
"learning_rate": 5.742496164266174e-06,
"loss": 0.578,
"step": 625
},
{
"epoch": 4.22972972972973,
"grad_norm": 0.07404139765077528,
"learning_rate": 5.645469311182958e-06,
"loss": 0.5788,
"step": 626
},
{
"epoch": 4.236486486486487,
"grad_norm": 0.06599738371244251,
"learning_rate": 5.549206882430773e-06,
"loss": 0.5938,
"step": 627
},
{
"epoch": 4.243243243243243,
"grad_norm": 0.07571258452222435,
"learning_rate": 5.453711019949581e-06,
"loss": 0.5761,
"step": 628
},
{
"epoch": 4.25,
"grad_norm": 0.06873039122654918,
"learning_rate": 5.358983848622452e-06,
"loss": 0.5865,
"step": 629
},
{
"epoch": 4.256756756756757,
"grad_norm": 0.05980111880094078,
"learning_rate": 5.265027476228297e-06,
"loss": 0.58,
"step": 630
},
{
"epoch": 4.263513513513513,
"grad_norm": 0.07210626853825096,
"learning_rate": 5.171843993394903e-06,
"loss": 0.5756,
"step": 631
},
{
"epoch": 4.27027027027027,
"grad_norm": 0.07396976700448737,
"learning_rate": 5.079435473552474e-06,
"loss": 0.5811,
"step": 632
},
{
"epoch": 4.277027027027027,
"grad_norm": 0.09171081358728068,
"learning_rate": 4.987803972887482e-06,
"loss": 0.5871,
"step": 633
},
{
"epoch": 4.283783783783784,
"grad_norm": 0.06701824447243687,
"learning_rate": 4.896951530296896e-06,
"loss": 0.579,
"step": 634
},
{
"epoch": 4.29054054054054,
"grad_norm": 0.07518459749761504,
"learning_rate": 4.806880167342831e-06,
"loss": 0.5822,
"step": 635
},
{
"epoch": 4.297297297297297,
"grad_norm": 0.07098354334790964,
"learning_rate": 4.7175918882075465e-06,
"loss": 0.5861,
"step": 636
},
{
"epoch": 4.304054054054054,
"grad_norm": 0.07479138842797883,
"learning_rate": 4.6290886796488946e-06,
"loss": 0.5863,
"step": 637
},
{
"epoch": 4.3108108108108105,
"grad_norm": 0.06686921193950773,
"learning_rate": 4.54137251095605e-06,
"loss": 0.5916,
"step": 638
},
{
"epoch": 4.3175675675675675,
"grad_norm": 0.07897960962457094,
"learning_rate": 4.454445333905768e-06,
"loss": 0.5948,
"step": 639
},
{
"epoch": 4.324324324324325,
"grad_norm": 0.07496951321971122,
"learning_rate": 4.3683090827188666e-06,
"loss": 0.5813,
"step": 640
},
{
"epoch": 4.331081081081081,
"grad_norm": 0.07536718259174548,
"learning_rate": 4.282965674017265e-06,
"loss": 0.5872,
"step": 641
},
{
"epoch": 4.337837837837838,
"grad_norm": 0.06854859712850378,
"learning_rate": 4.198417006781283e-06,
"loss": 0.5841,
"step": 642
},
{
"epoch": 4.344594594594595,
"grad_norm": 0.07618804161631497,
"learning_rate": 4.114664962307439e-06,
"loss": 0.5817,
"step": 643
},
{
"epoch": 4.351351351351352,
"grad_norm": 0.07601070844404975,
"learning_rate": 4.031711404166525e-06,
"loss": 0.584,
"step": 644
},
{
"epoch": 4.358108108108108,
"grad_norm": 0.06914624176338743,
"learning_rate": 3.949558178162209e-06,
"loss": 0.5888,
"step": 645
},
{
"epoch": 4.364864864864865,
"grad_norm": 0.07027655538494314,
"learning_rate": 3.86820711228991e-06,
"loss": 0.5867,
"step": 646
},
{
"epoch": 4.371621621621622,
"grad_norm": 0.0760270326599887,
"learning_rate": 3.7876600166961353e-06,
"loss": 0.5961,
"step": 647
},
{
"epoch": 4.378378378378378,
"grad_norm": 0.06782313812781374,
"learning_rate": 3.707918683638223e-06,
"loss": 0.5944,
"step": 648
},
{
"epoch": 4.385135135135135,
"grad_norm": 0.06192386606423058,
"learning_rate": 3.628984887444462e-06,
"loss": 0.5969,
"step": 649
},
{
"epoch": 4.391891891891892,
"grad_norm": 0.062017594322198465,
"learning_rate": 3.550860384474568e-06,
"loss": 0.5863,
"step": 650
},
{
"epoch": 4.398648648648648,
"grad_norm": 0.058946604304473224,
"learning_rate": 3.473546913080674e-06,
"loss": 0.5847,
"step": 651
},
{
"epoch": 4.405405405405405,
"grad_norm": 0.08803265906901768,
"learning_rate": 3.397046193568558e-06,
"loss": 0.5707,
"step": 652
},
{
"epoch": 4.412162162162162,
"grad_norm": 0.06583037908170926,
"learning_rate": 3.3213599281594688e-06,
"loss": 0.582,
"step": 653
},
{
"epoch": 4.418918918918919,
"grad_norm": 0.06086938296974132,
"learning_rate": 3.246489800952155e-06,
"loss": 0.5953,
"step": 654
},
{
"epoch": 4.425675675675675,
"grad_norm": 0.05932401857518893,
"learning_rate": 3.172437477885475e-06,
"loss": 0.5819,
"step": 655
},
{
"epoch": 4.4324324324324325,
"grad_norm": 0.057752799006265246,
"learning_rate": 3.099204606701256e-06,
"loss": 0.5766,
"step": 656
},
{
"epoch": 4.4391891891891895,
"grad_norm": 0.06397066535508446,
"learning_rate": 3.026792816907671e-06,
"loss": 0.588,
"step": 657
},
{
"epoch": 4.445945945945946,
"grad_norm": 0.06029045959058417,
"learning_rate": 2.955203719742965e-06,
"loss": 0.591,
"step": 658
},
{
"epoch": 4.452702702702703,
"grad_norm": 0.05838249627908107,
"learning_rate": 2.884438908139626e-06,
"loss": 0.5929,
"step": 659
},
{
"epoch": 4.45945945945946,
"grad_norm": 0.06075718886251464,
"learning_rate": 2.814499956688912e-06,
"loss": 0.5813,
"step": 660
},
{
"epoch": 4.466216216216216,
"grad_norm": 0.0719277704750579,
"learning_rate": 2.7453884216058368e-06,
"loss": 0.5795,
"step": 661
},
{
"epoch": 4.472972972972973,
"grad_norm": 0.06075719822797747,
"learning_rate": 2.677105840694507e-06,
"loss": 0.5674,
"step": 662
},
{
"epoch": 4.47972972972973,
"grad_norm": 0.05975207139447743,
"learning_rate": 2.6096537333139616e-06,
"loss": 0.5805,
"step": 663
},
{
"epoch": 4.486486486486487,
"grad_norm": 0.06213586333263577,
"learning_rate": 2.5430336003443045e-06,
"loss": 0.5906,
"step": 664
},
{
"epoch": 4.493243243243243,
"grad_norm": 0.05895116376323496,
"learning_rate": 2.4772469241533648e-06,
"loss": 0.5844,
"step": 665
},
{
"epoch": 4.5,
"grad_norm": 0.06765046020471,
"learning_rate": 2.4122951685636674e-06,
"loss": 0.5835,
"step": 666
},
{
"epoch": 4.506756756756757,
"grad_norm": 0.06650006977809148,
"learning_rate": 2.3481797788198745e-06,
"loss": 0.5789,
"step": 667
},
{
"epoch": 4.513513513513513,
"grad_norm": 0.05846137750044436,
"learning_rate": 2.284902181556632e-06,
"loss": 0.5745,
"step": 668
},
{
"epoch": 4.52027027027027,
"grad_norm": 0.05989337419115531,
"learning_rate": 2.2224637847668484e-06,
"loss": 0.5733,
"step": 669
},
{
"epoch": 4.527027027027027,
"grad_norm": 0.06939319651650724,
"learning_rate": 2.1608659777703033e-06,
"loss": 0.581,
"step": 670
},
{
"epoch": 4.533783783783784,
"grad_norm": 0.054679826375991857,
"learning_rate": 2.100110131182813e-06,
"loss": 0.5931,
"step": 671
},
{
"epoch": 4.54054054054054,
"grad_norm": 0.05548818045052255,
"learning_rate": 2.0401975968856514e-06,
"loss": 0.5786,
"step": 672
},
{
"epoch": 4.547297297297297,
"grad_norm": 0.059389572689147435,
"learning_rate": 1.981129707995542e-06,
"loss": 0.5923,
"step": 673
},
{
"epoch": 4.554054054054054,
"grad_norm": 0.06975826873383643,
"learning_rate": 1.9229077788349393e-06,
"loss": 0.577,
"step": 674
},
{
"epoch": 4.5608108108108105,
"grad_norm": 0.07220930223960455,
"learning_rate": 1.865533104902828e-06,
"loss": 0.5916,
"step": 675
},
{
"epoch": 4.5675675675675675,
"grad_norm": 0.053811020975530956,
"learning_rate": 1.8090069628458583e-06,
"loss": 0.5704,
"step": 676
},
{
"epoch": 4.574324324324325,
"grad_norm": 0.05538660298263909,
"learning_rate": 1.7533306104299663e-06,
"loss": 0.5856,
"step": 677
},
{
"epoch": 4.581081081081081,
"grad_norm": 0.05829509686058199,
"learning_rate": 1.6985052865123641e-06,
"loss": 0.5754,
"step": 678
},
{
"epoch": 4.587837837837838,
"grad_norm": 0.06323909786498662,
"learning_rate": 1.6445322110140116e-06,
"loss": 0.5814,
"step": 679
},
{
"epoch": 4.594594594594595,
"grad_norm": 0.056972721072935675,
"learning_rate": 1.59141258489242e-06,
"loss": 0.5756,
"step": 680
},
{
"epoch": 4.601351351351351,
"grad_norm": 0.05816167859998497,
"learning_rate": 1.53914759011498e-06,
"loss": 0.5889,
"step": 681
},
{
"epoch": 4.608108108108108,
"grad_norm": 0.059044794864098894,
"learning_rate": 1.4877383896326269e-06,
"loss": 0.5781,
"step": 682
},
{
"epoch": 4.614864864864865,
"grad_norm": 0.06237941099681927,
"learning_rate": 1.4371861273539778e-06,
"loss": 0.576,
"step": 683
},
{
"epoch": 4.621621621621622,
"grad_norm": 0.06531112971211704,
"learning_rate": 1.3874919281198662e-06,
"loss": 0.5799,
"step": 684
},
{
"epoch": 4.628378378378378,
"grad_norm": 0.05265860386748638,
"learning_rate": 1.3386568976783453e-06,
"loss": 0.5891,
"step": 685
},
{
"epoch": 4.635135135135135,
"grad_norm": 0.05785779557411594,
"learning_rate": 1.2906821226600453e-06,
"loss": 0.5768,
"step": 686
},
{
"epoch": 4.641891891891892,
"grad_norm": 0.05639941023635892,
"learning_rate": 1.2435686705540228e-06,
"loss": 0.5842,
"step": 687
},
{
"epoch": 4.648648648648649,
"grad_norm": 0.06338285927327728,
"learning_rate": 1.1973175896839684e-06,
"loss": 0.573,
"step": 688
},
{
"epoch": 4.655405405405405,
"grad_norm": 0.054671739680034734,
"learning_rate": 1.1519299091849523e-06,
"loss": 0.5774,
"step": 689
},
{
"epoch": 4.662162162162162,
"grad_norm": 0.057698821845826635,
"learning_rate": 1.1074066389804395e-06,
"loss": 0.5932,
"step": 690
},
{
"epoch": 4.668918918918919,
"grad_norm": 0.05697783615575523,
"learning_rate": 1.0637487697598937e-06,
"loss": 0.5823,
"step": 691
},
{
"epoch": 4.675675675675675,
"grad_norm": 0.054051563368746565,
"learning_rate": 1.0209572729566708e-06,
"loss": 0.5787,
"step": 692
},
{
"epoch": 4.6824324324324325,
"grad_norm": 0.05344229805841218,
"learning_rate": 9.790331007264543e-07,
"loss": 0.5844,
"step": 693
},
{
"epoch": 4.6891891891891895,
"grad_norm": 0.06959576323504683,
"learning_rate": 9.379771859260267e-07,
"loss": 0.5817,
"step": 694
},
{
"epoch": 4.695945945945946,
"grad_norm": 0.06318026825084985,
"learning_rate": 8.977904420925543e-07,
"loss": 0.5892,
"step": 695
},
{
"epoch": 4.702702702702703,
"grad_norm": 0.056039993579189235,
"learning_rate": 8.584737634232154e-07,
"loss": 0.5835,
"step": 696
},
{
"epoch": 4.70945945945946,
"grad_norm": 0.05269309541455996,
"learning_rate": 8.200280247553461e-07,
"loss": 0.5803,
"step": 697
},
{
"epoch": 4.716216216216216,
"grad_norm": 0.05510521164357771,
"learning_rate": 7.824540815469306e-07,
"loss": 0.5709,
"step": 698
},
{
"epoch": 4.722972972972973,
"grad_norm": 0.05770652073739158,
"learning_rate": 7.457527698576217e-07,
"loss": 0.5855,
"step": 699
},
{
"epoch": 4.72972972972973,
"grad_norm": 0.055593016439349656,
"learning_rate": 7.099249063300751e-07,
"loss": 0.5847,
"step": 700
},
{
"epoch": 4.736486486486487,
"grad_norm": 0.05730248759901012,
"learning_rate": 6.749712881718306e-07,
"loss": 0.5639,
"step": 701
},
{
"epoch": 4.743243243243243,
"grad_norm": 0.05973974788088902,
"learning_rate": 6.408926931375403e-07,
"loss": 0.5808,
"step": 702
},
{
"epoch": 4.75,
"grad_norm": 0.05557100230846073,
"learning_rate": 6.076898795116792e-07,
"loss": 0.5848,
"step": 703
},
{
"epoch": 4.756756756756757,
"grad_norm": 0.0563081652270169,
"learning_rate": 5.753635860916617e-07,
"loss": 0.5888,
"step": 704
},
{
"epoch": 4.763513513513513,
"grad_norm": 0.0536729552491486,
"learning_rate": 5.43914532171419e-07,
"loss": 0.5873,
"step": 705
},
{
"epoch": 4.77027027027027,
"grad_norm": 0.049916139412812977,
"learning_rate": 5.133434175253715e-07,
"loss": 0.5792,
"step": 706
},
{
"epoch": 4.777027027027027,
"grad_norm": 0.05386272187182479,
"learning_rate": 4.83650922392882e-07,
"loss": 0.5829,
"step": 707
},
{
"epoch": 4.783783783783784,
"grad_norm": 0.05503179443312982,
"learning_rate": 4.5483770746309383e-07,
"loss": 0.5873,
"step": 708
},
{
"epoch": 4.79054054054054,
"grad_norm": 0.05614293253442416,
"learning_rate": 4.269044138602585e-07,
"loss": 0.5879,
"step": 709
},
{
"epoch": 4.797297297297297,
"grad_norm": 0.053484276503556256,
"learning_rate": 3.998516631294491e-07,
"loss": 0.5794,
"step": 710
},
{
"epoch": 4.804054054054054,
"grad_norm": 0.055818505320336,
"learning_rate": 3.736800572227317e-07,
"loss": 0.5905,
"step": 711
},
{
"epoch": 4.8108108108108105,
"grad_norm": 0.05644903653041531,
"learning_rate": 3.483901784857846e-07,
"loss": 0.5721,
"step": 712
},
{
"epoch": 4.8175675675675675,
"grad_norm": 0.04844049723060667,
"learning_rate": 3.239825896449267e-07,
"loss": 0.5826,
"step": 713
},
{
"epoch": 4.824324324324325,
"grad_norm": 0.053088243917807275,
"learning_rate": 3.004578337945985e-07,
"loss": 0.5811,
"step": 714
},
{
"epoch": 4.831081081081081,
"grad_norm": 0.05140464370774195,
"learning_rate": 2.778164343852918e-07,
"loss": 0.5791,
"step": 715
},
{
"epoch": 4.837837837837838,
"grad_norm": 0.05679451911188569,
"learning_rate": 2.5605889521188364e-07,
"loss": 0.5787,
"step": 716
},
{
"epoch": 4.844594594594595,
"grad_norm": 0.06082278783885325,
"learning_rate": 2.351857004024316e-07,
"loss": 0.5868,
"step": 717
},
{
"epoch": 4.851351351351351,
"grad_norm": 0.05433042972998646,
"learning_rate": 2.1519731440740487e-07,
"loss": 0.5812,
"step": 718
},
{
"epoch": 4.858108108108108,
"grad_norm": 0.055224715989184006,
"learning_rate": 1.9609418198935916e-07,
"loss": 0.5881,
"step": 719
},
{
"epoch": 4.864864864864865,
"grad_norm": 0.05348313369078817,
"learning_rate": 1.778767282130156e-07,
"loss": 0.5945,
"step": 720
},
{
"epoch": 4.871621621621622,
"grad_norm": 0.05163946442780112,
"learning_rate": 1.6054535843582854e-07,
"loss": 0.5891,
"step": 721
},
{
"epoch": 4.878378378378378,
"grad_norm": 0.055008800596469655,
"learning_rate": 1.4410045829893915e-07,
"loss": 0.5817,
"step": 722
},
{
"epoch": 4.885135135135135,
"grad_norm": 0.05385755694710143,
"learning_rate": 1.2854239371863142e-07,
"loss": 0.5803,
"step": 723
},
{
"epoch": 4.891891891891892,
"grad_norm": 0.050256418798137074,
"learning_rate": 1.1387151087814297e-07,
"loss": 0.5776,
"step": 724
},
{
"epoch": 4.898648648648649,
"grad_norm": 0.06422788348062423,
"learning_rate": 1.0008813622001345e-07,
"loss": 0.5865,
"step": 725
},
{
"epoch": 4.905405405405405,
"grad_norm": 0.058912661052812065,
"learning_rate": 8.719257643877044e-08,
"loss": 0.5904,
"step": 726
},
{
"epoch": 4.912162162162162,
"grad_norm": 0.05385241384320565,
"learning_rate": 7.51851184741481e-08,
"loss": 0.5944,
"step": 727
},
{
"epoch": 4.918918918918919,
"grad_norm": 0.05443985642845785,
"learning_rate": 6.40660295046791e-08,
"loss": 0.5857,
"step": 728
},
{
"epoch": 4.925675675675675,
"grad_norm": 0.05281227387100857,
"learning_rate": 5.3835556941743695e-08,
"loss": 0.5742,
"step": 729
},
{
"epoch": 4.9324324324324325,
"grad_norm": 0.05531460046068238,
"learning_rate": 4.449392842408529e-08,
"loss": 0.5743,
"step": 730
},
{
"epoch": 4.9391891891891895,
"grad_norm": 0.05538814823209324,
"learning_rate": 3.6041351812743374e-08,
"loss": 0.5877,
"step": 731
},
{
"epoch": 4.945945945945946,
"grad_norm": 0.055186199974581227,
"learning_rate": 2.8478015186399477e-08,
"loss": 0.5842,
"step": 732
},
{
"epoch": 4.952702702702703,
"grad_norm": 0.055151710910165075,
"learning_rate": 2.1804086837229344e-08,
"loss": 0.5846,
"step": 733
},
{
"epoch": 4.95945945945946,
"grad_norm": 0.05558334617444456,
"learning_rate": 1.601971526713708e-08,
"loss": 0.5766,
"step": 734
},
{
"epoch": 4.966216216216216,
"grad_norm": 0.049437407136724947,
"learning_rate": 1.112502918445113e-08,
"loss": 0.586,
"step": 735
},
{
"epoch": 4.972972972972973,
"grad_norm": 0.05838258812809817,
"learning_rate": 7.12013750107321e-09,
"loss": 0.5918,
"step": 736
},
{
"epoch": 4.97972972972973,
"grad_norm": 0.05252990242194222,
"learning_rate": 4.00512933004471e-09,
"loss": 0.5853,
"step": 737
},
{
"epoch": 4.986486486486487,
"grad_norm": 0.05539911735758133,
"learning_rate": 1.7800739835616143e-09,
"loss": 0.5731,
"step": 738
},
{
"epoch": 4.993243243243243,
"grad_norm": 0.059325115613032664,
"learning_rate": 4.450209714379483e-10,
"loss": 0.5814,
"step": 739
},
{
"epoch": 5.0,
"grad_norm": 0.059385070432276595,
"learning_rate": 0.0,
"loss": 0.577,
"step": 740
},
{
"epoch": 5.0,
"step": 740,
"total_flos": 4.168626538020864e+16,
"train_loss": 0.6272868978010642,
"train_runtime": 42334.4318,
"train_samples_per_second": 8.913,
"train_steps_per_second": 0.017
}
],
"logging_steps": 1,
"max_steps": 740,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.168626538020864e+16,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}