diff --git "a/trainer_state.json" "b/trainer_state.json" --- "a/trainer_state.json" +++ "b/trainer_state.json" @@ -1,31212 +1,18236 @@ { + "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, - "epoch": 0.9998557067962099, + "epoch": 1.0, "eval_steps": 500, - "global_step": 5197, + "global_step": 2599, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { - "epoch": 0.0, - "learning_rate": 1.282051282051282e-07, - "loss": 2.0435, + "epoch": 0.00038476337052712584, + "grad_norm": 254.13031445424159, + "learning_rate": 2.564102564102564e-07, + "loss": 3.5757, "step": 1 }, { - "epoch": 0.0, - "learning_rate": 2.564102564102564e-07, - "loss": 2.1579, + "epoch": 0.0007695267410542517, + "grad_norm": 257.2445854171958, + "learning_rate": 5.128205128205128e-07, + "loss": 3.5918, "step": 2 }, { - "epoch": 0.0, - "learning_rate": 3.846153846153847e-07, - "loss": 2.0936, + "epoch": 0.0011542901115813775, + "grad_norm": 267.3563361262743, + "learning_rate": 7.692307692307694e-07, + "loss": 3.7205, "step": 3 }, { - "epoch": 0.0, - "learning_rate": 5.128205128205128e-07, - "loss": 2.0988, + "epoch": 0.0015390534821085034, + "grad_norm": 277.93793291694885, + "learning_rate": 1.0256410256410257e-06, + "loss": 3.5852, "step": 4 }, { - "epoch": 0.0, - "learning_rate": 6.41025641025641e-07, - "loss": 2.0727, + "epoch": 0.001923816852635629, + "grad_norm": 238.89050589475377, + "learning_rate": 1.282051282051282e-06, + "loss": 3.4711, "step": 5 }, { - "epoch": 0.0, - "learning_rate": 7.692307692307694e-07, - "loss": 2.1138, + "epoch": 0.002308580223162755, + "grad_norm": 208.69241863782162, + "learning_rate": 1.5384615384615387e-06, + "loss": 3.3499, "step": 6 }, { - "epoch": 0.0, - "learning_rate": 8.974358974358975e-07, - "loss": 2.094, + "epoch": 0.0026933435936898806, + "grad_norm": 179.36354907798489, + "learning_rate": 1.794871794871795e-06, + "loss": 3.226, "step": 7 }, { - "epoch": 0.0, - "learning_rate": 1.0256410256410257e-06, - "loss": 1.9901, + "epoch": 0.0030781069642170067, + "grad_norm": 32.31632679981777, + "learning_rate": 2.0512820512820513e-06, + "loss": 2.866, "step": 8 }, { - "epoch": 0.0, - "learning_rate": 1.153846153846154e-06, - "loss": 2.0892, + "epoch": 0.0034628703347441324, + "grad_norm": 34.36410317645356, + "learning_rate": 2.307692307692308e-06, + "loss": 2.9734, "step": 9 }, { - "epoch": 0.0, - "learning_rate": 1.282051282051282e-06, - "loss": 1.8482, + "epoch": 0.003847633705271258, + "grad_norm": 36.35297461054895, + "learning_rate": 2.564102564102564e-06, + "loss": 2.5415, "step": 10 }, { - "epoch": 0.0, - "learning_rate": 1.4102564102564104e-06, - "loss": 1.8128, + "epoch": 0.004232397075798384, + "grad_norm": 45.30552498372646, + "learning_rate": 2.8205128205128207e-06, + "loss": 2.6819, "step": 11 }, { - "epoch": 0.0, - "learning_rate": 1.5384615384615387e-06, - "loss": 1.7343, + "epoch": 0.00461716044632551, + "grad_norm": 37.33218120390648, + "learning_rate": 3.0769230769230774e-06, + "loss": 2.5895, "step": 12 }, { - "epoch": 0.0, - "learning_rate": 1.6666666666666667e-06, - "loss": 1.7195, + "epoch": 0.005001923816852636, + "grad_norm": 31.29133891137258, + "learning_rate": 3.3333333333333333e-06, + "loss": 2.4569, "step": 13 }, { - "epoch": 0.0, - "learning_rate": 1.794871794871795e-06, - "loss": 1.6839, + "epoch": 0.005386687187379761, + "grad_norm": 30.350351589310346, + "learning_rate": 3.58974358974359e-06, + "loss": 2.3896, "step": 14 }, { - "epoch": 0.0, - "learning_rate": 1.9230769230769234e-06, - "loss": 1.6196, + "epoch": 0.005771450557906887, + "grad_norm": 31.919355167578146, + "learning_rate": 3.846153846153847e-06, + "loss": 2.3929, "step": 15 }, { - "epoch": 0.0, - "learning_rate": 2.0512820512820513e-06, - "loss": 1.5573, + "epoch": 0.0061562139284340135, + "grad_norm": 32.832009014984976, + "learning_rate": 4.102564102564103e-06, + "loss": 2.296, "step": 16 }, { - "epoch": 0.0, - "learning_rate": 2.1794871794871797e-06, - "loss": 1.5891, + "epoch": 0.006540977298961139, + "grad_norm": 32.49182168295507, + "learning_rate": 4.358974358974359e-06, + "loss": 2.3011, "step": 17 }, { - "epoch": 0.0, - "learning_rate": 2.307692307692308e-06, - "loss": 1.455, + "epoch": 0.006925740669488265, + "grad_norm": 9.843599017874173, + "learning_rate": 4.615384615384616e-06, + "loss": 0.9788, "step": 18 }, { - "epoch": 0.0, - "learning_rate": 2.435897435897436e-06, - "loss": 1.4842, + "epoch": 0.007310504040015391, + "grad_norm": 31.56813308794278, + "learning_rate": 4.871794871794872e-06, + "loss": 2.2194, "step": 19 }, { - "epoch": 0.0, - "learning_rate": 2.564102564102564e-06, - "loss": 1.4643, + "epoch": 0.007695267410542516, + "grad_norm": 32.53550983067877, + "learning_rate": 5.128205128205128e-06, + "loss": 2.2128, "step": 20 }, { - "epoch": 0.0, - "learning_rate": 2.6923076923076923e-06, - "loss": 1.466, + "epoch": 0.008080030781069641, + "grad_norm": 9.148059887262804, + "learning_rate": 5.384615384615385e-06, + "loss": 0.9732, "step": 21 }, { - "epoch": 0.0, - "learning_rate": 2.8205128205128207e-06, - "loss": 1.4158, + "epoch": 0.008464794151596768, + "grad_norm": 31.672291858317898, + "learning_rate": 5.641025641025641e-06, + "loss": 2.1545, "step": 22 }, { - "epoch": 0.0, - "learning_rate": 2.948717948717949e-06, - "loss": 1.3401, + "epoch": 0.008849557522123894, + "grad_norm": 26.529352600584385, + "learning_rate": 5.897435897435898e-06, + "loss": 2.0046, "step": 23 }, { - "epoch": 0.0, - "learning_rate": 3.0769230769230774e-06, - "loss": 1.416, + "epoch": 0.00923432089265102, + "grad_norm": 29.30554260862775, + "learning_rate": 6.153846153846155e-06, + "loss": 2.0447, "step": 24 }, { - "epoch": 0.0, - "learning_rate": 3.205128205128206e-06, - "loss": 1.3375, + "epoch": 0.009619084263178146, + "grad_norm": 27.770757938491702, + "learning_rate": 6.410256410256412e-06, + "loss": 1.9348, "step": 25 }, { - "epoch": 0.01, - "learning_rate": 3.3333333333333333e-06, - "loss": 1.3556, + "epoch": 0.010003847633705272, + "grad_norm": 28.702033049904912, + "learning_rate": 6.666666666666667e-06, + "loss": 1.9243, "step": 26 }, { - "epoch": 0.01, - "learning_rate": 3.4615384615384617e-06, - "loss": 1.3364, + "epoch": 0.010388611004232396, + "grad_norm": 5.200533143929214, + "learning_rate": 6.923076923076923e-06, + "loss": 0.8978, "step": 27 }, { - "epoch": 0.01, - "learning_rate": 3.58974358974359e-06, - "loss": 1.2624, + "epoch": 0.010773374374759523, + "grad_norm": 26.876589183258098, + "learning_rate": 7.17948717948718e-06, + "loss": 1.874, "step": 28 }, { - "epoch": 0.01, - "learning_rate": 3.7179487179487184e-06, - "loss": 1.286, + "epoch": 0.011158137745286649, + "grad_norm": 3.767429880190435, + "learning_rate": 7.435897435897437e-06, + "loss": 0.8539, "step": 29 }, { - "epoch": 0.01, - "learning_rate": 3.846153846153847e-06, - "loss": 1.2331, + "epoch": 0.011542901115813775, + "grad_norm": 27.387433673435865, + "learning_rate": 7.692307692307694e-06, + "loss": 1.8029, "step": 30 }, { - "epoch": 0.01, - "learning_rate": 3.974358974358974e-06, - "loss": 1.2614, + "epoch": 0.0119276644863409, + "grad_norm": 27.540558092447068, + "learning_rate": 7.948717948717949e-06, + "loss": 1.7961, "step": 31 }, { - "epoch": 0.01, - "learning_rate": 4.102564102564103e-06, - "loss": 1.2398, + "epoch": 0.012312427856868027, + "grad_norm": 29.8647581604738, + "learning_rate": 8.205128205128205e-06, + "loss": 1.787, "step": 32 }, { - "epoch": 0.01, - "learning_rate": 4.230769230769231e-06, - "loss": 1.2722, + "epoch": 0.012697191227395151, + "grad_norm": 29.88974003644155, + "learning_rate": 8.461538461538462e-06, + "loss": 1.7628, "step": 33 }, { - "epoch": 0.01, - "learning_rate": 4.358974358974359e-06, - "loss": 1.2289, + "epoch": 0.013081954597922277, + "grad_norm": 29.644502192656933, + "learning_rate": 8.717948717948719e-06, + "loss": 1.7221, "step": 34 }, { - "epoch": 0.01, - "learning_rate": 4.487179487179488e-06, - "loss": 1.2745, + "epoch": 0.013466717968449404, + "grad_norm": 27.408957414385196, + "learning_rate": 8.974358974358976e-06, + "loss": 1.6144, "step": 35 }, { - "epoch": 0.01, - "learning_rate": 4.615384615384616e-06, - "loss": 1.2522, + "epoch": 0.01385148133897653, + "grad_norm": 30.110366205643277, + "learning_rate": 9.230769230769232e-06, + "loss": 1.6487, "step": 36 }, { - "epoch": 0.01, - "learning_rate": 4.743589743589744e-06, - "loss": 0.9359, + "epoch": 0.014236244709503656, + "grad_norm": 30.117120450350164, + "learning_rate": 9.487179487179487e-06, + "loss": 1.5961, "step": 37 }, { - "epoch": 0.01, - "learning_rate": 4.871794871794872e-06, - "loss": 1.2977, + "epoch": 0.014621008080030782, + "grad_norm": 30.820280105757266, + "learning_rate": 9.743589743589744e-06, + "loss": 1.5426, "step": 38 }, { - "epoch": 0.01, - "learning_rate": 5e-06, - "loss": 1.2712, + "epoch": 0.015005771450557906, + "grad_norm": 27.066492655502547, + "learning_rate": 1e-05, + "loss": 1.4483, "step": 39 }, { - "epoch": 0.01, - "learning_rate": 5.128205128205128e-06, - "loss": 1.2233, + "epoch": 0.015390534821085032, + "grad_norm": 28.860853185373603, + "learning_rate": 1.0256410256410256e-05, + "loss": 1.4626, "step": 40 }, { - "epoch": 0.01, - "learning_rate": 5.256410256410257e-06, - "loss": 1.2418, + "epoch": 0.015775298191612157, + "grad_norm": 29.240007065853185, + "learning_rate": 1.0512820512820514e-05, + "loss": 1.4247, "step": 41 }, { - "epoch": 0.01, - "learning_rate": 5.384615384615385e-06, - "loss": 1.2694, + "epoch": 0.016160061562139283, + "grad_norm": 27.475140231075894, + "learning_rate": 1.076923076923077e-05, + "loss": 1.3474, "step": 42 }, { - "epoch": 0.01, - "learning_rate": 5.512820512820514e-06, - "loss": 1.1514, + "epoch": 0.01654482493266641, + "grad_norm": 26.2135141381118, + "learning_rate": 1.1025641025641028e-05, + "loss": 1.2871, "step": 43 }, { - "epoch": 0.01, - "learning_rate": 5.641025641025641e-06, - "loss": 1.2713, + "epoch": 0.016929588303193535, + "grad_norm": 26.698452815899614, + "learning_rate": 1.1282051282051283e-05, + "loss": 1.2287, "step": 44 }, { - "epoch": 0.01, - "learning_rate": 5.769230769230769e-06, - "loss": 1.2247, + "epoch": 0.01731435167372066, + "grad_norm": 24.106345542034916, + "learning_rate": 1.1538461538461538e-05, + "loss": 1.2215, "step": 45 }, { - "epoch": 0.01, - "learning_rate": 5.897435897435898e-06, - "loss": 1.1625, + "epoch": 0.017699115044247787, + "grad_norm": 7.038745110721678, + "learning_rate": 1.1794871794871796e-05, + "loss": 0.8096, "step": 46 }, { - "epoch": 0.01, - "learning_rate": 6.025641025641026e-06, - "loss": 1.1565, + "epoch": 0.018083878414774913, + "grad_norm": 23.605195445723812, + "learning_rate": 1.2051282051282051e-05, + "loss": 1.0822, "step": 47 }, { - "epoch": 0.01, - "learning_rate": 6.153846153846155e-06, - "loss": 1.1202, + "epoch": 0.01846864178530204, + "grad_norm": 23.63091371891358, + "learning_rate": 1.230769230769231e-05, + "loss": 1.0464, "step": 48 }, { - "epoch": 0.01, - "learning_rate": 6.282051282051282e-06, - "loss": 1.1412, + "epoch": 0.018853405155829166, + "grad_norm": 5.5664432307623075, + "learning_rate": 1.2564102564102565e-05, + "loss": 0.7375, "step": 49 }, { - "epoch": 0.01, - "learning_rate": 6.410256410256412e-06, - "loss": 1.1675, + "epoch": 0.019238168526356292, + "grad_norm": 19.974451474664452, + "learning_rate": 1.2820512820512823e-05, + "loss": 0.9526, "step": 50 }, { - "epoch": 0.01, - "learning_rate": 6.538461538461539e-06, - "loss": 1.2088, + "epoch": 0.019622931896883418, + "grad_norm": 3.3760320416005656, + "learning_rate": 1.3076923076923078e-05, + "loss": 0.7532, "step": 51 }, { - "epoch": 0.01, - "learning_rate": 6.666666666666667e-06, - "loss": 1.1778, + "epoch": 0.020007695267410544, + "grad_norm": 14.39249779220866, + "learning_rate": 1.3333333333333333e-05, + "loss": 0.9341, "step": 52 }, { - "epoch": 0.01, - "learning_rate": 6.794871794871796e-06, - "loss": 1.1548, + "epoch": 0.020392458637937667, + "grad_norm": 12.687699664808068, + "learning_rate": 1.3589743589743592e-05, + "loss": 0.8814, "step": 53 }, { - "epoch": 0.01, - "learning_rate": 6.923076923076923e-06, - "loss": 1.1275, + "epoch": 0.020777222008464793, + "grad_norm": 10.842834958475416, + "learning_rate": 1.3846153846153847e-05, + "loss": 0.8616, "step": 54 }, { - "epoch": 0.01, - "learning_rate": 7.051282051282053e-06, - "loss": 1.1325, + "epoch": 0.02116198537899192, + "grad_norm": 8.624617725412142, + "learning_rate": 1.4102564102564105e-05, + "loss": 0.8838, "step": 55 }, { - "epoch": 0.01, - "learning_rate": 7.17948717948718e-06, - "loss": 0.8788, + "epoch": 0.021546748749519045, + "grad_norm": 7.908834956833458, + "learning_rate": 1.435897435897436e-05, + "loss": 0.8112, "step": 56 }, { - "epoch": 0.01, - "learning_rate": 7.307692307692308e-06, - "loss": 1.1943, + "epoch": 0.02193151212004617, + "grad_norm": 5.932806290956269, + "learning_rate": 1.4615384615384615e-05, + "loss": 0.8717, "step": 57 }, { - "epoch": 0.01, - "learning_rate": 7.435897435897437e-06, - "loss": 1.1885, + "epoch": 0.022316275490573297, + "grad_norm": 5.431041160682564, + "learning_rate": 1.4871794871794874e-05, + "loss": 0.82, "step": 58 }, { - "epoch": 0.01, - "learning_rate": 7.564102564102564e-06, - "loss": 1.186, + "epoch": 0.022701038861100423, + "grad_norm": 4.396732578817546, + "learning_rate": 1.5128205128205129e-05, + "loss": 0.8758, "step": 59 }, { - "epoch": 0.01, - "learning_rate": 7.692307692307694e-06, - "loss": 1.1695, + "epoch": 0.02308580223162755, + "grad_norm": 4.171630720674021, + "learning_rate": 1.5384615384615387e-05, + "loss": 0.7962, "step": 60 }, { - "epoch": 0.01, - "learning_rate": 7.820512820512822e-06, - "loss": 1.1342, + "epoch": 0.023470565602154676, + "grad_norm": 3.7386243152162493, + "learning_rate": 1.5641025641025644e-05, + "loss": 0.8091, "step": 61 }, { - "epoch": 0.01, - "learning_rate": 7.948717948717949e-06, - "loss": 1.0974, + "epoch": 0.0238553289726818, + "grad_norm": 3.370246964882419, + "learning_rate": 1.5897435897435897e-05, + "loss": 0.7966, "step": 62 }, { - "epoch": 0.01, - "learning_rate": 8.076923076923077e-06, - "loss": 1.0832, + "epoch": 0.024240092343208928, + "grad_norm": 3.4358022596354996, + "learning_rate": 1.6153846153846154e-05, + "loss": 0.791, "step": 63 }, { - "epoch": 0.01, - "learning_rate": 8.205128205128205e-06, - "loss": 1.1105, + "epoch": 0.024624855713736054, + "grad_norm": 3.1436683433105936, + "learning_rate": 1.641025641025641e-05, + "loss": 0.7691, "step": 64 }, { - "epoch": 0.01, - "learning_rate": 8.333333333333334e-06, - "loss": 1.1017, + "epoch": 0.025009619084263177, + "grad_norm": 3.020165763044837, + "learning_rate": 1.6666666666666667e-05, + "loss": 0.8055, "step": 65 }, { - "epoch": 0.01, - "learning_rate": 8.461538461538462e-06, - "loss": 1.061, + "epoch": 0.025394382454790303, + "grad_norm": 3.1252991574318516, + "learning_rate": 1.6923076923076924e-05, + "loss": 0.8292, "step": 66 }, { - "epoch": 0.01, - "learning_rate": 8.58974358974359e-06, - "loss": 0.9965, + "epoch": 0.02577914582531743, + "grad_norm": 3.514375202196702, + "learning_rate": 1.717948717948718e-05, + "loss": 0.7686, "step": 67 }, { - "epoch": 0.01, - "learning_rate": 8.717948717948719e-06, - "loss": 1.1237, + "epoch": 0.026163909195844555, + "grad_norm": 3.461641859263959, + "learning_rate": 1.7435897435897438e-05, + "loss": 0.7998, "step": 68 }, { - "epoch": 0.01, - "learning_rate": 8.846153846153847e-06, - "loss": 1.1001, + "epoch": 0.02654867256637168, + "grad_norm": 3.9509509844288297, + "learning_rate": 1.7692307692307694e-05, + "loss": 0.8028, "step": 69 }, { - "epoch": 0.01, - "learning_rate": 8.974358974358976e-06, - "loss": 1.1055, + "epoch": 0.026933435936898807, + "grad_norm": 3.8138092429022916, + "learning_rate": 1.794871794871795e-05, + "loss": 0.8056, "step": 70 }, { - "epoch": 0.01, - "learning_rate": 9.102564102564104e-06, - "loss": 1.0615, + "epoch": 0.027318199307425933, + "grad_norm": 4.442396197495926, + "learning_rate": 1.8205128205128208e-05, + "loss": 0.7645, "step": 71 }, { - "epoch": 0.01, - "learning_rate": 9.230769230769232e-06, - "loss": 1.0063, + "epoch": 0.02770296267795306, + "grad_norm": 4.809972610493762, + "learning_rate": 1.8461538461538465e-05, + "loss": 0.7595, "step": 72 }, { - "epoch": 0.01, - "learning_rate": 9.358974358974359e-06, - "loss": 1.1002, + "epoch": 0.028087726048480186, + "grad_norm": 5.250477834596211, + "learning_rate": 1.8717948717948718e-05, + "loss": 0.755, "step": 73 }, { - "epoch": 0.01, - "learning_rate": 9.487179487179487e-06, - "loss": 1.0157, + "epoch": 0.02847248941900731, + "grad_norm": 5.3762574614609795, + "learning_rate": 1.8974358974358975e-05, + "loss": 0.7563, "step": 74 }, { - "epoch": 0.01, - "learning_rate": 9.615384615384616e-06, - "loss": 1.0458, + "epoch": 0.028857252789534438, + "grad_norm": 5.586483902415329, + "learning_rate": 1.923076923076923e-05, + "loss": 0.7506, "step": 75 }, { - "epoch": 0.01, - "learning_rate": 9.743589743589744e-06, - "loss": 1.0382, + "epoch": 0.029242016160061564, + "grad_norm": 5.501125579868852, + "learning_rate": 1.9487179487179488e-05, + "loss": 0.8167, "step": 76 }, { - "epoch": 0.01, - "learning_rate": 9.871794871794872e-06, - "loss": 1.0266, + "epoch": 0.029626779530588686, + "grad_norm": 6.247960459622417, + "learning_rate": 1.9743589743589745e-05, + "loss": 0.7516, "step": 77 }, { - "epoch": 0.02, - "learning_rate": 1e-05, - "loss": 1.075, + "epoch": 0.030011542901115813, + "grad_norm": 6.994317979635179, + "learning_rate": 2e-05, + "loss": 0.7356, "step": 78 }, { - "epoch": 0.02, - "learning_rate": 1.012820512820513e-05, - "loss": 1.1223, + "epoch": 0.03039630627164294, + "grad_norm": 6.673345062828693, + "learning_rate": 1.9999992235312136e-05, + "loss": 0.7932, "step": 79 }, { - "epoch": 0.02, - "learning_rate": 1.0256410256410256e-05, - "loss": 1.1071, + "epoch": 0.030781069642170065, + "grad_norm": 6.968509254023697, + "learning_rate": 1.9999968941260596e-05, + "loss": 0.7405, "step": 80 }, { - "epoch": 0.02, - "learning_rate": 1.0384615384615386e-05, - "loss": 1.1471, + "epoch": 0.03116583301269719, + "grad_norm": 7.095747234304572, + "learning_rate": 1.9999930117881548e-05, + "loss": 0.7415, "step": 81 }, { - "epoch": 0.02, - "learning_rate": 1.0512820512820514e-05, - "loss": 1.1882, + "epoch": 0.031550596383224314, + "grad_norm": 7.241986280109527, + "learning_rate": 1.99998757652353e-05, + "loss": 0.6663, "step": 82 }, { - "epoch": 0.02, - "learning_rate": 1.0641025641025643e-05, - "loss": 1.1812, + "epoch": 0.03193535975375144, + "grad_norm": 5.8828563511663905, + "learning_rate": 1.999980588340624e-05, + "loss": 0.7303, "step": 83 }, { - "epoch": 0.02, - "learning_rate": 1.076923076923077e-05, - "loss": 1.0865, + "epoch": 0.032320123124278566, + "grad_norm": 5.6897276351311765, + "learning_rate": 1.9999720472502902e-05, + "loss": 0.7241, "step": 84 }, { - "epoch": 0.02, - "learning_rate": 1.0897435897435898e-05, - "loss": 1.0104, + "epoch": 0.03270488649480569, + "grad_norm": 5.224159530934288, + "learning_rate": 1.9999619532657915e-05, + "loss": 0.6774, "step": 85 }, { - "epoch": 0.02, - "learning_rate": 1.1025641025641028e-05, - "loss": 1.0445, + "epoch": 0.03308964986533282, + "grad_norm": 4.677186466229448, + "learning_rate": 1.9999503064028043e-05, + "loss": 0.6925, "step": 86 }, { - "epoch": 0.02, - "learning_rate": 1.1153846153846154e-05, - "loss": 1.1112, + "epoch": 0.033474413235859944, + "grad_norm": 4.12600648346499, + "learning_rate": 1.9999371066794146e-05, + "loss": 0.663, "step": 87 }, { - "epoch": 0.02, - "learning_rate": 1.1282051282051283e-05, - "loss": 1.0677, + "epoch": 0.03385917660638707, + "grad_norm": 3.839369991867343, + "learning_rate": 1.999922354116121e-05, + "loss": 0.6839, "step": 88 }, { - "epoch": 0.02, - "learning_rate": 1.1410256410256411e-05, - "loss": 0.8353, + "epoch": 0.034243939976914196, + "grad_norm": 3.076387634827698, + "learning_rate": 1.9999060487358333e-05, + "loss": 0.7178, "step": 89 }, { - "epoch": 0.02, - "learning_rate": 1.1538461538461538e-05, - "loss": 0.974, + "epoch": 0.03462870334744132, + "grad_norm": 2.9578937316286713, + "learning_rate": 1.9998881905638727e-05, + "loss": 0.698, "step": 90 }, { - "epoch": 0.02, - "learning_rate": 1.1666666666666668e-05, - "loss": 1.0766, + "epoch": 0.03501346671796845, + "grad_norm": 2.6837233020137834, + "learning_rate": 1.999868779627972e-05, + "loss": 0.7042, "step": 91 }, { - "epoch": 0.02, - "learning_rate": 1.1794871794871796e-05, - "loss": 1.1667, + "epoch": 0.035398230088495575, + "grad_norm": 2.6437466466902317, + "learning_rate": 1.9998478159582747e-05, + "loss": 0.6448, "step": 92 }, { - "epoch": 0.02, - "learning_rate": 1.1923076923076925e-05, - "loss": 1.1263, + "epoch": 0.0357829934590227, + "grad_norm": 2.491994547670168, + "learning_rate": 1.9998252995873367e-05, + "loss": 0.6566, "step": 93 }, { - "epoch": 0.02, - "learning_rate": 1.2051282051282051e-05, - "loss": 0.9795, + "epoch": 0.03616775682954983, + "grad_norm": 2.224843142204464, + "learning_rate": 1.9998012305501243e-05, + "loss": 0.6734, "step": 94 }, { - "epoch": 0.02, - "learning_rate": 1.217948717948718e-05, - "loss": 0.8291, + "epoch": 0.03655252020007695, + "grad_norm": 2.002424886497015, + "learning_rate": 1.999775608884015e-05, + "loss": 0.644, "step": 95 }, { - "epoch": 0.02, - "learning_rate": 1.230769230769231e-05, - "loss": 1.03, + "epoch": 0.03693728357060408, + "grad_norm": 1.8827428543698392, + "learning_rate": 1.9997484346287973e-05, + "loss": 0.6575, "step": 96 }, { - "epoch": 0.02, - "learning_rate": 1.2435897435897436e-05, - "loss": 1.0783, + "epoch": 0.037322046941131205, + "grad_norm": 1.758486971958877, + "learning_rate": 1.9997197078266723e-05, + "loss": 0.6884, "step": 97 }, { - "epoch": 0.02, - "learning_rate": 1.2564102564102565e-05, - "loss": 1.0617, + "epoch": 0.03770681031165833, + "grad_norm": 1.7076111992527954, + "learning_rate": 1.99968942852225e-05, + "loss": 0.6579, "step": 98 }, { - "epoch": 0.02, - "learning_rate": 1.2692307692307693e-05, - "loss": 0.9693, + "epoch": 0.03809157368218546, + "grad_norm": 1.6159220839510806, + "learning_rate": 1.9996575967625525e-05, + "loss": 0.7062, "step": 99 }, { - "epoch": 0.02, - "learning_rate": 1.2820512820512823e-05, - "loss": 0.7842, + "epoch": 0.038476337052712584, + "grad_norm": 1.5655083625503003, + "learning_rate": 1.999624212597013e-05, + "loss": 0.6465, "step": 100 }, { - "epoch": 0.02, - "learning_rate": 1.294871794871795e-05, - "loss": 1.0361, + "epoch": 0.03886110042323971, + "grad_norm": 1.4738280504319872, + "learning_rate": 1.9995892760774738e-05, + "loss": 0.681, "step": 101 }, { - "epoch": 0.02, - "learning_rate": 1.3076923076923078e-05, - "loss": 1.1101, + "epoch": 0.039245863793766836, + "grad_norm": 1.4517968174779372, + "learning_rate": 1.9995527872581903e-05, + "loss": 0.6379, "step": 102 }, { - "epoch": 0.02, - "learning_rate": 1.3205128205128207e-05, - "loss": 1.0835, + "epoch": 0.03963062716429396, + "grad_norm": 1.3777795281664882, + "learning_rate": 1.9995147461958267e-05, + "loss": 0.6149, "step": 103 }, { - "epoch": 0.02, - "learning_rate": 1.3333333333333333e-05, - "loss": 1.063, + "epoch": 0.04001539053482109, + "grad_norm": 1.342152136177547, + "learning_rate": 1.999475152949459e-05, + "loss": 0.6416, "step": 104 }, { - "epoch": 0.02, - "learning_rate": 1.3461538461538463e-05, - "loss": 0.7536, + "epoch": 0.040400153905348214, + "grad_norm": 1.326935205991647, + "learning_rate": 1.9994340075805724e-05, + "loss": 0.6754, "step": 105 }, { - "epoch": 0.02, - "learning_rate": 1.3589743589743592e-05, - "loss": 1.0256, + "epoch": 0.04078491727587533, + "grad_norm": 1.3292407957163603, + "learning_rate": 1.9993913101530635e-05, + "loss": 0.6295, "step": 106 }, { - "epoch": 0.02, - "learning_rate": 1.3717948717948718e-05, - "loss": 1.0802, + "epoch": 0.04116968064640246, + "grad_norm": 1.4379231077150227, + "learning_rate": 1.9993470607332387e-05, + "loss": 0.6989, "step": 107 }, { - "epoch": 0.02, - "learning_rate": 1.3846153846153847e-05, - "loss": 1.0175, + "epoch": 0.041554444016929586, + "grad_norm": 1.2595832637925053, + "learning_rate": 1.9993012593898146e-05, + "loss": 0.6304, "step": 108 }, { - "epoch": 0.02, - "learning_rate": 1.3974358974358975e-05, - "loss": 1.0289, + "epoch": 0.04193920738745671, + "grad_norm": 1.2979375230866474, + "learning_rate": 1.9992539061939175e-05, + "loss": 0.6562, "step": 109 }, { - "epoch": 0.02, - "learning_rate": 1.4102564102564105e-05, - "loss": 1.0013, + "epoch": 0.04232397075798384, + "grad_norm": 1.2999323380976717, + "learning_rate": 1.9992050012190845e-05, + "loss": 0.6901, "step": 110 }, { - "epoch": 0.02, - "learning_rate": 1.4230769230769232e-05, - "loss": 1.0626, + "epoch": 0.042708734128510964, + "grad_norm": 1.316108382596062, + "learning_rate": 1.9991545445412614e-05, + "loss": 0.7188, "step": 111 }, { - "epoch": 0.02, - "learning_rate": 1.435897435897436e-05, - "loss": 1.1833, + "epoch": 0.04309349749903809, + "grad_norm": 1.3060309695535761, + "learning_rate": 1.9991025362388044e-05, + "loss": 0.6616, "step": 112 }, { - "epoch": 0.02, - "learning_rate": 1.4487179487179489e-05, - "loss": 1.0086, + "epoch": 0.043478260869565216, + "grad_norm": 1.2443838752128427, + "learning_rate": 1.9990489763924796e-05, + "loss": 0.6349, "step": 113 }, { - "epoch": 0.02, - "learning_rate": 1.4615384615384615e-05, - "loss": 1.0494, + "epoch": 0.04386302424009234, + "grad_norm": 1.3130530580889228, + "learning_rate": 1.9989938650854618e-05, + "loss": 0.6613, "step": 114 }, { - "epoch": 0.02, - "learning_rate": 1.4743589743589745e-05, - "loss": 1.0293, + "epoch": 0.04424778761061947, + "grad_norm": 1.2261283738623239, + "learning_rate": 1.9989372024033352e-05, + "loss": 0.6708, "step": 115 }, { - "epoch": 0.02, - "learning_rate": 1.4871794871794874e-05, - "loss": 0.8056, + "epoch": 0.044632550981146595, + "grad_norm": 1.3038843273958582, + "learning_rate": 1.9988789884340938e-05, + "loss": 0.6192, "step": 116 }, { - "epoch": 0.02, - "learning_rate": 1.5000000000000002e-05, - "loss": 1.1147, + "epoch": 0.04501731435167372, + "grad_norm": 1.4120208742499956, + "learning_rate": 1.9988192232681398e-05, + "loss": 0.6336, "step": 117 }, { - "epoch": 0.02, - "learning_rate": 1.5128205128205129e-05, - "loss": 1.037, + "epoch": 0.04540207772220085, + "grad_norm": 1.2195993648497905, + "learning_rate": 1.9987579069982856e-05, + "loss": 0.6394, "step": 118 }, { - "epoch": 0.02, - "learning_rate": 1.5256410256410257e-05, - "loss": 1.1532, + "epoch": 0.04578684109272797, + "grad_norm": 1.125537935201837, + "learning_rate": 1.9986950397197503e-05, + "loss": 0.5942, "step": 119 }, { - "epoch": 0.02, - "learning_rate": 1.5384615384615387e-05, - "loss": 1.0542, + "epoch": 0.0461716044632551, + "grad_norm": 1.2153397010654559, + "learning_rate": 1.998630621530164e-05, + "loss": 0.6446, "step": 120 }, { - "epoch": 0.02, - "learning_rate": 1.5512820512820516e-05, - "loss": 1.1333, + "epoch": 0.046556367833782225, + "grad_norm": 1.2953707503536012, + "learning_rate": 1.9985646525295634e-05, + "loss": 0.7142, "step": 121 }, { - "epoch": 0.02, - "learning_rate": 1.5641025641025644e-05, - "loss": 1.1772, + "epoch": 0.04694113120430935, + "grad_norm": 1.2523530409252028, + "learning_rate": 1.9984971328203945e-05, + "loss": 0.6264, "step": 122 }, { - "epoch": 0.02, - "learning_rate": 1.576923076923077e-05, - "loss": 0.9767, + "epoch": 0.04732589457483648, + "grad_norm": 1.199712399258201, + "learning_rate": 1.9984280625075115e-05, + "loss": 0.6448, "step": 123 }, { - "epoch": 0.02, - "learning_rate": 1.5897435897435897e-05, - "loss": 1.0509, + "epoch": 0.0477106579453636, + "grad_norm": 1.300287757337604, + "learning_rate": 1.998357441698176e-05, + "loss": 0.665, "step": 124 }, { - "epoch": 0.02, - "learning_rate": 1.602564102564103e-05, - "loss": 1.1079, + "epoch": 0.04809542131589073, + "grad_norm": 1.1336330482099135, + "learning_rate": 1.9982852705020572e-05, + "loss": 0.6774, "step": 125 }, { - "epoch": 0.02, - "learning_rate": 1.6153846153846154e-05, - "loss": 1.0725, + "epoch": 0.048480184686417856, + "grad_norm": 1.1747213826291132, + "learning_rate": 1.9982115490312334e-05, + "loss": 0.6421, "step": 126 }, { - "epoch": 0.02, - "learning_rate": 1.6282051282051282e-05, - "loss": 1.0932, + "epoch": 0.04886494805694498, + "grad_norm": 1.0676330334698418, + "learning_rate": 1.9981362774001886e-05, + "loss": 0.5974, "step": 127 }, { - "epoch": 0.02, - "learning_rate": 1.641025641025641e-05, - "loss": 0.8877, + "epoch": 0.04924971142747211, + "grad_norm": 1.1233697150234012, + "learning_rate": 1.9980594557258158e-05, + "loss": 0.6234, "step": 128 }, { - "epoch": 0.02, - "learning_rate": 1.653846153846154e-05, - "loss": 1.0874, + "epoch": 0.04963447479799923, + "grad_norm": 1.1615140644814288, + "learning_rate": 1.9979810841274135e-05, + "loss": 0.6302, "step": 129 }, { - "epoch": 0.03, - "learning_rate": 1.6666666666666667e-05, - "loss": 1.0427, + "epoch": 0.05001923816852635, + "grad_norm": 6.7712469071523484, + "learning_rate": 1.9979011627266884e-05, + "loss": 0.6549, "step": 130 }, { - "epoch": 0.03, - "learning_rate": 1.6794871794871796e-05, - "loss": 1.1156, + "epoch": 0.05040400153905348, + "grad_norm": 1.2914742932851113, + "learning_rate": 1.997819691647753e-05, + "loss": 0.6789, "step": 131 }, { - "epoch": 0.03, - "learning_rate": 1.6923076923076924e-05, - "loss": 1.0239, + "epoch": 0.050788764909580605, + "grad_norm": 1.2634656232320374, + "learning_rate": 1.9977366710171274e-05, + "loss": 0.6587, "step": 132 }, { - "epoch": 0.03, - "learning_rate": 1.7051282051282053e-05, - "loss": 1.0242, + "epoch": 0.05117352828010773, + "grad_norm": 1.1448977841599888, + "learning_rate": 1.9976521009637366e-05, + "loss": 0.6961, "step": 133 }, { - "epoch": 0.03, - "learning_rate": 1.717948717948718e-05, - "loss": 1.043, + "epoch": 0.05155829165063486, + "grad_norm": 1.037074348327444, + "learning_rate": 1.9975659816189137e-05, + "loss": 0.5646, "step": 134 }, { - "epoch": 0.03, - "learning_rate": 1.730769230769231e-05, - "loss": 0.9751, + "epoch": 0.051943055021161984, + "grad_norm": 1.116088415767267, + "learning_rate": 1.9974783131163957e-05, + "loss": 0.6625, "step": 135 }, { - "epoch": 0.03, - "learning_rate": 1.7435897435897438e-05, - "loss": 1.0387, + "epoch": 0.05232781839168911, + "grad_norm": 1.2104716807398774, + "learning_rate": 1.997389095592327e-05, + "loss": 0.6646, "step": 136 }, { - "epoch": 0.03, - "learning_rate": 1.7564102564102566e-05, - "loss": 1.0659, + "epoch": 0.052712581762216236, + "grad_norm": 1.3217444586361569, + "learning_rate": 1.9972983291852565e-05, + "loss": 0.67, "step": 137 }, { - "epoch": 0.03, - "learning_rate": 1.7692307692307694e-05, - "loss": 1.0436, + "epoch": 0.05309734513274336, + "grad_norm": 1.1006101156696524, + "learning_rate": 1.9972060140361384e-05, + "loss": 0.6677, "step": 138 }, { - "epoch": 0.03, - "learning_rate": 1.7820512820512823e-05, - "loss": 1.0762, + "epoch": 0.05348210850327049, + "grad_norm": 1.0784897490006204, + "learning_rate": 1.9971121502883332e-05, + "loss": 0.6334, "step": 139 }, { - "epoch": 0.03, - "learning_rate": 1.794871794871795e-05, - "loss": 1.1233, + "epoch": 0.053866871873797614, + "grad_norm": 1.1661655288947685, + "learning_rate": 1.997016738087605e-05, + "loss": 0.6301, "step": 140 }, { - "epoch": 0.03, - "learning_rate": 1.807692307692308e-05, - "loss": 1.0594, + "epoch": 0.05425163524432474, + "grad_norm": 1.1668903475042323, + "learning_rate": 1.9969197775821227e-05, + "loss": 0.641, "step": 141 }, { - "epoch": 0.03, - "learning_rate": 1.8205128205128208e-05, - "loss": 1.0631, + "epoch": 0.05463639861485187, + "grad_norm": 1.0996564057168, + "learning_rate": 1.9968212689224603e-05, + "loss": 0.627, "step": 142 }, { - "epoch": 0.03, - "learning_rate": 1.8333333333333333e-05, - "loss": 1.1144, + "epoch": 0.05502116198537899, + "grad_norm": 1.1947559959717744, + "learning_rate": 1.9967212122615958e-05, + "loss": 0.6789, "step": 143 }, { - "epoch": 0.03, - "learning_rate": 1.8461538461538465e-05, - "loss": 1.1106, + "epoch": 0.05540592535590612, + "grad_norm": 1.0372546268565326, + "learning_rate": 1.9966196077549106e-05, + "loss": 0.5918, "step": 144 }, { - "epoch": 0.03, - "learning_rate": 1.8589743589743593e-05, - "loss": 1.0429, + "epoch": 0.055790688726433245, + "grad_norm": 1.0705655962518679, + "learning_rate": 1.99651645556019e-05, + "loss": 0.6492, "step": 145 }, { - "epoch": 0.03, - "learning_rate": 1.8717948717948718e-05, - "loss": 1.0777, + "epoch": 0.05617545209696037, + "grad_norm": 1.1285747538240676, + "learning_rate": 1.996411755837623e-05, + "loss": 0.6353, "step": 146 }, { - "epoch": 0.03, - "learning_rate": 1.8846153846153846e-05, - "loss": 1.0221, + "epoch": 0.0565602154674875, + "grad_norm": 1.0339759507662283, + "learning_rate": 1.996305508749802e-05, + "loss": 0.56, "step": 147 }, { - "epoch": 0.03, - "learning_rate": 1.8974358974358975e-05, - "loss": 1.018, + "epoch": 0.05694497883801462, + "grad_norm": 1.1598427021338287, + "learning_rate": 1.9961977144617225e-05, + "loss": 0.656, "step": 148 }, { - "epoch": 0.03, - "learning_rate": 1.9102564102564106e-05, - "loss": 1.0424, + "epoch": 0.05732974220854175, + "grad_norm": 0.9932149026259058, + "learning_rate": 1.996088373140781e-05, + "loss": 0.6792, "step": 149 }, { - "epoch": 0.03, - "learning_rate": 1.923076923076923e-05, - "loss": 1.0226, + "epoch": 0.057714505579068875, + "grad_norm": 1.0974932677516867, + "learning_rate": 1.995977484956779e-05, + "loss": 0.6175, "step": 150 }, { - "epoch": 0.03, - "learning_rate": 1.935897435897436e-05, - "loss": 1.0694, + "epoch": 0.058099268949596, + "grad_norm": 1.1029243100245116, + "learning_rate": 1.9958650500819183e-05, + "loss": 0.6801, "step": 151 }, { - "epoch": 0.03, - "learning_rate": 1.9487179487179488e-05, - "loss": 0.8062, + "epoch": 0.05848403232012313, + "grad_norm": 1.2406324862091482, + "learning_rate": 1.9957510686908034e-05, + "loss": 0.6773, "step": 152 }, { - "epoch": 0.03, - "learning_rate": 1.9615384615384617e-05, - "loss": 1.0228, + "epoch": 0.05886879569065025, + "grad_norm": 1.109718071919507, + "learning_rate": 1.9956355409604402e-05, + "loss": 0.6268, "step": 153 }, { - "epoch": 0.03, - "learning_rate": 1.9743589743589745e-05, - "loss": 0.9552, + "epoch": 0.05925355906117737, + "grad_norm": 1.1013935363089709, + "learning_rate": 1.9955184670702363e-05, + "loss": 0.6008, "step": 154 }, { - "epoch": 0.03, - "learning_rate": 1.9871794871794873e-05, - "loss": 0.9583, + "epoch": 0.0596383224317045, + "grad_norm": 1.1282394625054482, + "learning_rate": 1.9953998472019996e-05, + "loss": 0.6168, "step": 155 }, { - "epoch": 0.03, - "learning_rate": 2e-05, - "loss": 1.1494, + "epoch": 0.060023085802231625, + "grad_norm": 1.100110118679096, + "learning_rate": 1.9952796815399403e-05, + "loss": 0.6588, "step": 156 }, { - "epoch": 0.03, - "learning_rate": 1.9999998058057616e-05, - "loss": 0.9939, + "epoch": 0.06040784917275875, + "grad_norm": 1.0598573935390456, + "learning_rate": 1.9951579702706668e-05, + "loss": 0.6683, "step": 157 }, { - "epoch": 0.03, - "learning_rate": 1.9999992232231216e-05, - "loss": 1.075, + "epoch": 0.06079261254328588, + "grad_norm": 1.0763908496901589, + "learning_rate": 1.9950347135831907e-05, + "loss": 0.6424, "step": 158 }, { - "epoch": 0.03, - "learning_rate": 1.999998252252306e-05, - "loss": 1.0494, + "epoch": 0.061177375913813004, + "grad_norm": 1.1300517092037785, + "learning_rate": 1.994909911668921e-05, + "loss": 0.6564, "step": 159 }, { - "epoch": 0.03, - "learning_rate": 1.9999968928936924e-05, - "loss": 1.1033, + "epoch": 0.06156213928434013, + "grad_norm": 1.1843132314582407, + "learning_rate": 1.994783564721667e-05, + "loss": 0.6832, "step": 160 }, { - "epoch": 0.03, - "learning_rate": 1.999995145147809e-05, - "loss": 1.0241, + "epoch": 0.061946902654867256, + "grad_norm": 1.1616836324557385, + "learning_rate": 1.994655672937638e-05, + "loss": 0.6034, "step": 161 }, { - "epoch": 0.03, - "learning_rate": 1.9999930090153335e-05, - "loss": 1.0491, + "epoch": 0.06233166602539438, + "grad_norm": 1.0510187186161004, + "learning_rate": 1.994526236515442e-05, + "loss": 0.6368, "step": 162 }, { - "epoch": 0.03, - "learning_rate": 1.9999904844970963e-05, - "loss": 1.0413, + "epoch": 0.06271642939592151, + "grad_norm": 1.1474656837532322, + "learning_rate": 1.9943952556560863e-05, + "loss": 0.6537, "step": 163 }, { - "epoch": 0.03, - "learning_rate": 1.999987571594078e-05, - "loss": 0.8145, + "epoch": 0.06310119276644863, + "grad_norm": 1.1275115147039416, + "learning_rate": 1.9942627305629747e-05, + "loss": 0.6285, "step": 164 }, { - "epoch": 0.03, - "learning_rate": 1.99998427030741e-05, - "loss": 1.0603, + "epoch": 0.06348595613697576, + "grad_norm": 1.05766439835012, + "learning_rate": 1.9941286614419113e-05, + "loss": 0.6411, "step": 165 }, { - "epoch": 0.03, - "learning_rate": 1.999980580638374e-05, - "loss": 1.0931, + "epoch": 0.06387071950750288, + "grad_norm": 1.2223024111751775, + "learning_rate": 1.9939930485010968e-05, + "loss": 0.6358, "step": 166 }, { - "epoch": 0.03, - "learning_rate": 1.999976502588403e-05, - "loss": 1.0459, + "epoch": 0.06425548287803001, + "grad_norm": 1.167180448325711, + "learning_rate": 1.99385589195113e-05, + "loss": 0.6534, "step": 167 }, { - "epoch": 0.03, - "learning_rate": 1.9999720361590812e-05, - "loss": 1.0239, + "epoch": 0.06464024624855713, + "grad_norm": 1.1872690819504401, + "learning_rate": 1.9937171920050057e-05, + "loss": 0.7056, "step": 168 }, { - "epoch": 0.03, - "learning_rate": 1.9999671813521435e-05, - "loss": 1.0575, + "epoch": 0.06502500961908426, + "grad_norm": 1.055421253952468, + "learning_rate": 1.9935769488781167e-05, + "loss": 0.6556, "step": 169 }, { - "epoch": 0.03, - "learning_rate": 1.999961938169475e-05, - "loss": 1.0214, + "epoch": 0.06540977298961138, + "grad_norm": 1.219639157389629, + "learning_rate": 1.993435162788252e-05, + "loss": 0.6108, "step": 170 }, { - "epoch": 0.03, - "learning_rate": 1.9999563066131124e-05, - "loss": 0.9953, + "epoch": 0.06579453636013852, + "grad_norm": 1.1002952979253235, + "learning_rate": 1.9932918339555965e-05, + "loss": 0.6655, "step": 171 }, { - "epoch": 0.03, - "learning_rate": 1.9999502866852427e-05, - "loss": 1.0834, + "epoch": 0.06617929973066564, + "grad_norm": 1.1180793801605866, + "learning_rate": 1.9931469626027305e-05, + "loss": 0.6293, "step": 172 }, { - "epoch": 0.03, - "learning_rate": 1.999943878388204e-05, - "loss": 0.9889, + "epoch": 0.06656406310119277, + "grad_norm": 1.204505706577808, + "learning_rate": 1.9930005489546308e-05, + "loss": 0.6013, "step": 173 }, { - "epoch": 0.03, - "learning_rate": 1.9999370817244853e-05, - "loss": 1.1028, + "epoch": 0.06694882647171989, + "grad_norm": 1.1619204193994284, + "learning_rate": 1.9928525932386678e-05, + "loss": 0.6204, "step": 174 }, { - "epoch": 0.03, - "learning_rate": 1.9999298966967264e-05, - "loss": 1.1047, + "epoch": 0.06733358984224702, + "grad_norm": 1.060267423327307, + "learning_rate": 1.9927030956846083e-05, + "loss": 0.6454, "step": 175 }, { - "epoch": 0.03, - "learning_rate": 1.9999223233077178e-05, - "loss": 1.0105, + "epoch": 0.06771835321277414, + "grad_norm": 1.0578506852888985, + "learning_rate": 1.9925520565246125e-05, + "loss": 0.6457, "step": 176 }, { - "epoch": 0.03, - "learning_rate": 1.999914361560401e-05, - "loss": 1.0366, + "epoch": 0.06810311658330127, + "grad_norm": 1.1033529969083768, + "learning_rate": 1.9923994759932344e-05, + "loss": 0.6384, "step": 177 }, { - "epoch": 0.03, - "learning_rate": 1.9999060114578682e-05, - "loss": 1.0049, + "epoch": 0.06848787995382839, + "grad_norm": 1.1511149389566608, + "learning_rate": 1.9922453543274223e-05, + "loss": 0.6399, "step": 178 }, { - "epoch": 0.03, - "learning_rate": 1.9998972730033624e-05, - "loss": 1.1073, + "epoch": 0.06887264332435553, + "grad_norm": 1.1113673807067372, + "learning_rate": 1.9920896917665178e-05, + "loss": 0.6543, "step": 179 }, { - "epoch": 0.03, - "learning_rate": 1.9998881462002778e-05, - "loss": 0.7187, + "epoch": 0.06925740669488265, + "grad_norm": 1.164517240929667, + "learning_rate": 1.9919324885522548e-05, + "loss": 0.6208, "step": 180 }, { - "epoch": 0.03, - "learning_rate": 1.9998786310521585e-05, - "loss": 1.0329, + "epoch": 0.06964217006540978, + "grad_norm": 1.200720762152209, + "learning_rate": 1.99177374492876e-05, + "loss": 0.683, "step": 181 }, { - "epoch": 0.04, - "learning_rate": 1.9998687275627008e-05, - "loss": 1.1577, + "epoch": 0.0700269334359369, + "grad_norm": 1.169347142479691, + "learning_rate": 1.9916134611425522e-05, + "loss": 0.6661, "step": 182 }, { - "epoch": 0.04, - "learning_rate": 1.9998584357357503e-05, - "loss": 1.0279, + "epoch": 0.07041169680646403, + "grad_norm": 1.0787968837227557, + "learning_rate": 1.991451637442543e-05, + "loss": 0.6465, "step": 183 }, { - "epoch": 0.04, - "learning_rate": 1.9998477555753054e-05, - "loss": 1.0151, + "epoch": 0.07079646017699115, + "grad_norm": 1.0428951280332095, + "learning_rate": 1.9912882740800336e-05, + "loss": 0.5622, "step": 184 }, { - "epoch": 0.04, - "learning_rate": 1.9998366870855134e-05, - "loss": 1.0677, + "epoch": 0.07118122354751828, + "grad_norm": 1.0960853558441503, + "learning_rate": 1.9911233713087172e-05, + "loss": 0.645, "step": 185 }, { - "epoch": 0.04, - "learning_rate": 1.999825230270673e-05, - "loss": 1.0772, + "epoch": 0.0715659869180454, + "grad_norm": 1.1510127893105204, + "learning_rate": 1.990956929384678e-05, + "loss": 0.6407, "step": 186 }, { - "epoch": 0.04, - "learning_rate": 1.9998133851352342e-05, - "loss": 0.7761, + "epoch": 0.07195075028857253, + "grad_norm": 1.1583501691874019, + "learning_rate": 1.9907889485663897e-05, + "loss": 0.6419, "step": 187 }, { - "epoch": 0.04, - "learning_rate": 1.9998011516837974e-05, - "loss": 0.9573, + "epoch": 0.07233551365909965, + "grad_norm": 0.9148005220041145, + "learning_rate": 1.9906194291147155e-05, + "loss": 0.677, "step": 188 }, { - "epoch": 0.04, - "learning_rate": 1.999788529921114e-05, - "loss": 0.9497, + "epoch": 0.07272027702962677, + "grad_norm": 1.0356109844798198, + "learning_rate": 1.9904483712929094e-05, + "loss": 0.6288, "step": 189 }, { - "epoch": 0.04, - "learning_rate": 1.999775519852086e-05, - "loss": 0.7959, + "epoch": 0.0731050404001539, + "grad_norm": 1.209218311886088, + "learning_rate": 1.990275775366613e-05, + "loss": 0.6758, "step": 190 }, { - "epoch": 0.04, - "learning_rate": 1.999762121481767e-05, - "loss": 1.072, + "epoch": 0.07348980377068103, + "grad_norm": 1.0872223468299071, + "learning_rate": 1.990101641603857e-05, + "loss": 0.6486, "step": 191 }, { - "epoch": 0.04, - "learning_rate": 1.99974833481536e-05, - "loss": 1.022, + "epoch": 0.07387456714120816, + "grad_norm": 1.0908358595075844, + "learning_rate": 1.9899259702750604e-05, + "loss": 0.5689, "step": 192 }, { - "epoch": 0.04, - "learning_rate": 1.9997341598582197e-05, - "loss": 1.1381, + "epoch": 0.07425933051173528, + "grad_norm": 1.1285445479605456, + "learning_rate": 1.9897487616530296e-05, + "loss": 0.6217, "step": 193 }, { - "epoch": 0.04, - "learning_rate": 1.9997195966158518e-05, - "loss": 1.064, + "epoch": 0.07464409388226241, + "grad_norm": 1.0369453248413463, + "learning_rate": 1.9895700160129593e-05, + "loss": 0.6856, "step": 194 }, { - "epoch": 0.04, - "learning_rate": 1.9997046450939122e-05, - "loss": 1.001, + "epoch": 0.07502885725278953, + "grad_norm": 1.1615318548831766, + "learning_rate": 1.9893897336324292e-05, + "loss": 0.628, "step": 195 }, { - "epoch": 0.04, - "learning_rate": 1.9996893052982083e-05, - "loss": 1.0791, + "epoch": 0.07541362062331666, + "grad_norm": 1.258347650709502, + "learning_rate": 1.9892079147914072e-05, + "loss": 0.6301, "step": 196 }, { - "epoch": 0.04, - "learning_rate": 1.9996735772346973e-05, - "loss": 1.1284, + "epoch": 0.07579838399384378, + "grad_norm": 1.0497984638900062, + "learning_rate": 1.9890245597722465e-05, + "loss": 0.6116, "step": 197 }, { - "epoch": 0.04, - "learning_rate": 1.9996574609094887e-05, - "loss": 1.0721, + "epoch": 0.07618314736437092, + "grad_norm": 1.129761979883271, + "learning_rate": 1.988839668859686e-05, + "loss": 0.6336, "step": 198 }, { - "epoch": 0.04, - "learning_rate": 1.9996409563288404e-05, - "loss": 1.002, + "epoch": 0.07656791073489803, + "grad_norm": 1.2378021881347194, + "learning_rate": 1.9886532423408495e-05, + "loss": 0.6298, "step": 199 }, { - "epoch": 0.04, - "learning_rate": 1.9996240634991645e-05, - "loss": 1.0455, + "epoch": 0.07695267410542517, + "grad_norm": 0.8359964521465918, + "learning_rate": 1.9884652805052465e-05, + "loss": 0.7224, "step": 200 }, { - "epoch": 0.04, - "learning_rate": 1.9996067824270204e-05, - "loss": 1.0623, + "epoch": 0.07733743747595229, + "grad_norm": 1.1729266641239005, + "learning_rate": 1.988275783644769e-05, + "loss": 0.661, "step": 201 }, { - "epoch": 0.04, - "learning_rate": 1.999589113119121e-05, - "loss": 1.0972, + "epoch": 0.07772220084647942, + "grad_norm": 1.1887145416267784, + "learning_rate": 1.988084752053695e-05, + "loss": 0.6421, "step": 202 }, { - "epoch": 0.04, - "learning_rate": 1.9995710555823277e-05, - "loss": 1.0513, + "epoch": 0.07810696421700654, + "grad_norm": 1.1761526162291613, + "learning_rate": 1.9878921860286832e-05, + "loss": 0.6609, "step": 203 }, { - "epoch": 0.04, - "learning_rate": 1.999552609823655e-05, - "loss": 0.978, + "epoch": 0.07849172758753367, + "grad_norm": 1.1264395130473355, + "learning_rate": 1.9876980858687777e-05, + "loss": 0.6251, "step": 204 }, { - "epoch": 0.04, - "learning_rate": 1.999533775850266e-05, - "loss": 0.7688, + "epoch": 0.07887649095806079, + "grad_norm": 1.053652805945505, + "learning_rate": 1.987502451875403e-05, + "loss": 0.6617, "step": 205 }, { - "epoch": 0.04, - "learning_rate": 1.9995145536694764e-05, - "loss": 0.9796, + "epoch": 0.07926125432858792, + "grad_norm": 1.0917282405225264, + "learning_rate": 1.9873052843523676e-05, + "loss": 0.613, "step": 206 }, { - "epoch": 0.04, - "learning_rate": 1.9994949432887512e-05, - "loss": 1.0161, + "epoch": 0.07964601769911504, + "grad_norm": 1.1768310410265819, + "learning_rate": 1.98710658360586e-05, + "loss": 0.6195, "step": 207 }, { - "epoch": 0.04, - "learning_rate": 1.999474944715708e-05, - "loss": 1.052, + "epoch": 0.08003078106964218, + "grad_norm": 1.0052383243236918, + "learning_rate": 1.9869063499444495e-05, + "loss": 0.559, "step": 208 }, { - "epoch": 0.04, - "learning_rate": 1.9994545579581125e-05, - "loss": 1.0784, + "epoch": 0.0804155444401693, + "grad_norm": 1.0339362339859088, + "learning_rate": 1.9867045836790867e-05, + "loss": 0.5802, "step": 209 }, { - "epoch": 0.04, - "learning_rate": 1.9994337830238836e-05, - "loss": 1.1168, + "epoch": 0.08080030781069643, + "grad_norm": 1.1799582749441122, + "learning_rate": 1.9865012851231022e-05, + "loss": 0.6247, "step": 210 }, { - "epoch": 0.04, - "learning_rate": 1.9994126199210897e-05, - "loss": 1.0636, + "epoch": 0.08118507118122355, + "grad_norm": 0.7705707100517312, + "learning_rate": 1.986296454592206e-05, + "loss": 0.6198, "step": 211 }, { - "epoch": 0.04, - "learning_rate": 1.999391068657951e-05, - "loss": 0.9491, + "epoch": 0.08156983455175067, + "grad_norm": 1.0815551467504874, + "learning_rate": 1.9860900924044873e-05, + "loss": 0.6539, "step": 212 }, { - "epoch": 0.04, - "learning_rate": 1.9993691292428364e-05, - "loss": 1.0042, + "epoch": 0.0819545979222778, + "grad_norm": 1.0258855036945687, + "learning_rate": 1.9858821988804132e-05, + "loss": 0.6695, "step": 213 }, { - "epoch": 0.04, - "learning_rate": 1.9993468016842684e-05, - "loss": 0.9829, + "epoch": 0.08233936129280492, + "grad_norm": 1.2284010782200772, + "learning_rate": 1.98567277434283e-05, + "loss": 0.6201, "step": 214 }, { - "epoch": 0.04, - "learning_rate": 1.999324085990918e-05, - "loss": 1.0609, + "epoch": 0.08272412466333205, + "grad_norm": 0.7494702781559469, + "learning_rate": 1.98546181911696e-05, + "loss": 0.6648, "step": 215 }, { - "epoch": 0.04, - "learning_rate": 1.9993009821716076e-05, - "loss": 1.0603, + "epoch": 0.08310888803385917, + "grad_norm": 1.1996863441863357, + "learning_rate": 1.985249333530404e-05, + "loss": 0.6428, "step": 216 }, { - "epoch": 0.04, - "learning_rate": 1.9992774902353104e-05, - "loss": 0.7666, + "epoch": 0.0834936514043863, + "grad_norm": 1.217248115851413, + "learning_rate": 1.9850353179131392e-05, + "loss": 0.6651, "step": 217 }, { - "epoch": 0.04, - "learning_rate": 1.999253610191151e-05, - "loss": 0.7652, + "epoch": 0.08387841477491342, + "grad_norm": 1.2120481138932582, + "learning_rate": 1.984819772597518e-05, + "loss": 0.6489, "step": 218 }, { - "epoch": 0.04, - "learning_rate": 1.999229342048404e-05, - "loss": 1.0581, + "epoch": 0.08426317814544056, + "grad_norm": 1.059103911653318, + "learning_rate": 1.984602697918269e-05, + "loss": 0.6264, "step": 219 }, { - "epoch": 0.04, - "learning_rate": 1.9992046858164942e-05, - "loss": 0.7, + "epoch": 0.08464794151596768, + "grad_norm": 1.0456030106682446, + "learning_rate": 1.9843840942124956e-05, + "loss": 0.581, "step": 220 }, { - "epoch": 0.04, - "learning_rate": 1.999179641504999e-05, - "loss": 1.0207, + "epoch": 0.08503270488649481, + "grad_norm": 1.2216361352828575, + "learning_rate": 1.984163961819676e-05, + "loss": 0.6573, "step": 221 }, { - "epoch": 0.04, - "learning_rate": 1.9991542091236438e-05, - "loss": 1.0079, + "epoch": 0.08541746825702193, + "grad_norm": 1.1916367686728033, + "learning_rate": 1.9839423010816616e-05, + "loss": 0.6579, "step": 222 }, { - "epoch": 0.04, - "learning_rate": 1.9991283886823075e-05, - "loss": 1.0847, + "epoch": 0.08580223162754906, + "grad_norm": 1.0352386000662999, + "learning_rate": 1.9837191123426777e-05, + "loss": 0.6266, "step": 223 }, { - "epoch": 0.04, - "learning_rate": 1.9991021801910177e-05, - "loss": 1.0495, + "epoch": 0.08618699499807618, + "grad_norm": 1.1910403288053584, + "learning_rate": 1.983494395949323e-05, + "loss": 0.6931, "step": 224 }, { - "epoch": 0.04, - "learning_rate": 1.999075583659954e-05, - "loss": 1.0109, + "epoch": 0.08657175836860331, + "grad_norm": 1.1409720269395809, + "learning_rate": 1.9832681522505676e-05, + "loss": 0.6767, "step": 225 }, { - "epoch": 0.04, - "learning_rate": 1.999048599099446e-05, - "loss": 0.7419, + "epoch": 0.08695652173913043, + "grad_norm": 1.0063345436239648, + "learning_rate": 1.983040381597754e-05, + "loss": 0.5888, "step": 226 }, { - "epoch": 0.04, - "learning_rate": 1.9990212265199738e-05, - "loss": 0.9118, + "epoch": 0.08734128510965757, + "grad_norm": 1.1435755328889867, + "learning_rate": 1.9828110843445954e-05, + "loss": 0.5822, "step": 227 }, { - "epoch": 0.04, - "learning_rate": 1.998993465932169e-05, - "loss": 1.047, + "epoch": 0.08772604848018468, + "grad_norm": 1.0952301566939102, + "learning_rate": 1.9825802608471767e-05, + "loss": 0.6656, "step": 228 }, { - "epoch": 0.04, - "learning_rate": 1.9989653173468137e-05, - "loss": 0.9947, + "epoch": 0.08811081185071182, + "grad_norm": 1.1518328905312722, + "learning_rate": 1.982347911463952e-05, + "loss": 0.6555, "step": 229 }, { - "epoch": 0.04, - "learning_rate": 1.99893678077484e-05, - "loss": 0.987, + "epoch": 0.08849557522123894, + "grad_norm": 1.2390625100117387, + "learning_rate": 1.982114036555746e-05, + "loss": 0.6482, "step": 230 }, { - "epoch": 0.04, - "learning_rate": 1.9989078562273313e-05, - "loss": 0.9699, + "epoch": 0.08888033859176607, + "grad_norm": 1.057561184150887, + "learning_rate": 1.9818786364857506e-05, + "loss": 0.627, "step": 231 }, { - "epoch": 0.04, - "learning_rate": 1.9988785437155222e-05, - "loss": 1.0319, + "epoch": 0.08926510196229319, + "grad_norm": 1.128740399710003, + "learning_rate": 1.9816417116195287e-05, + "loss": 0.6897, "step": 232 }, { - "epoch": 0.04, - "learning_rate": 1.9988488432507963e-05, - "loss": 1.0254, + "epoch": 0.08964986533282032, + "grad_norm": 1.0980773508753734, + "learning_rate": 1.9814032623250093e-05, + "loss": 0.6605, "step": 233 }, { - "epoch": 0.05, - "learning_rate": 1.9988187548446895e-05, - "loss": 0.982, + "epoch": 0.09003462870334744, + "grad_norm": 1.1646685009374829, + "learning_rate": 1.9811632889724888e-05, + "loss": 0.6257, "step": 234 }, { - "epoch": 0.05, - "learning_rate": 1.998788278508888e-05, - "loss": 1.0027, + "epoch": 0.09041939207387456, + "grad_norm": 1.1271704780975378, + "learning_rate": 1.9809217919346318e-05, + "loss": 0.6305, "step": 235 }, { - "epoch": 0.05, - "learning_rate": 1.9987574142552274e-05, - "loss": 1.0828, + "epoch": 0.0908041554444017, + "grad_norm": 1.2223933035753873, + "learning_rate": 1.9806787715864674e-05, + "loss": 0.6574, "step": 236 }, { - "epoch": 0.05, - "learning_rate": 1.9987261620956964e-05, - "loss": 1.046, + "epoch": 0.09118891881492881, + "grad_norm": 1.0925020063547646, + "learning_rate": 1.9804342283053916e-05, + "loss": 0.603, "step": 237 }, { - "epoch": 0.05, - "learning_rate": 1.9986945220424326e-05, - "loss": 1.0721, + "epoch": 0.09157368218545595, + "grad_norm": 1.0978191955442718, + "learning_rate": 1.980188162471164e-05, + "loss": 0.6434, "step": 238 }, { - "epoch": 0.05, - "learning_rate": 1.998662494107724e-05, - "loss": 1.0075, + "epoch": 0.09195844555598306, + "grad_norm": 1.101107320040046, + "learning_rate": 1.97994057446591e-05, + "loss": 0.6515, "step": 239 }, { - "epoch": 0.05, - "learning_rate": 1.99863007830401e-05, - "loss": 1.0368, + "epoch": 0.0923432089265102, + "grad_norm": 1.1033264795034836, + "learning_rate": 1.9796914646741187e-05, + "loss": 0.6649, "step": 240 }, { - "epoch": 0.05, - "learning_rate": 1.9985972746438815e-05, - "loss": 1.0265, + "epoch": 0.09272797229703732, + "grad_norm": 1.2293219321630533, + "learning_rate": 1.9794408334826415e-05, + "loss": 0.6432, "step": 241 }, { - "epoch": 0.05, - "learning_rate": 1.9985640831400778e-05, - "loss": 0.9143, + "epoch": 0.09311273566756445, + "grad_norm": 1.0016849688947362, + "learning_rate": 1.9791886812806932e-05, + "loss": 0.6124, "step": 242 }, { - "epoch": 0.05, - "learning_rate": 1.998530503805491e-05, - "loss": 1.0358, + "epoch": 0.09349749903809157, + "grad_norm": 0.9953416111573515, + "learning_rate": 1.9789350084598504e-05, + "loss": 0.5692, "step": 243 }, { - "epoch": 0.05, - "learning_rate": 1.9984965366531624e-05, - "loss": 1.0282, + "epoch": 0.0938822624086187, + "grad_norm": 1.0593409450207172, + "learning_rate": 1.9786798154140507e-05, + "loss": 0.6162, "step": 244 }, { - "epoch": 0.05, - "learning_rate": 1.9984621816962843e-05, - "loss": 1.0295, + "epoch": 0.09426702577914582, + "grad_norm": 1.1466660525556431, + "learning_rate": 1.9784231025395936e-05, + "loss": 0.6784, "step": 245 }, { - "epoch": 0.05, - "learning_rate": 1.9984274389482005e-05, - "loss": 1.1162, + "epoch": 0.09465178914967295, + "grad_norm": 1.0890828127311494, + "learning_rate": 1.9781648702351383e-05, + "loss": 0.6137, "step": 246 }, { - "epoch": 0.05, - "learning_rate": 1.9983923084224047e-05, - "loss": 0.9639, + "epoch": 0.09503655252020007, + "grad_norm": 1.0293014307639217, + "learning_rate": 1.977905118901703e-05, + "loss": 0.5661, "step": 247 }, { - "epoch": 0.05, - "learning_rate": 1.9983567901325404e-05, - "loss": 1.0348, + "epoch": 0.0954213158907272, + "grad_norm": 1.1746537217582962, + "learning_rate": 1.977643848942665e-05, + "loss": 0.6527, "step": 248 }, { - "epoch": 0.05, - "learning_rate": 1.9983208840924028e-05, - "loss": 1.0227, + "epoch": 0.09580607926125433, + "grad_norm": 1.0911917990586926, + "learning_rate": 1.9773810607637612e-05, + "loss": 0.5943, "step": 249 }, { - "epoch": 0.05, - "learning_rate": 1.998284590315937e-05, - "loss": 0.9554, + "epoch": 0.09619084263178146, + "grad_norm": 1.183725808532673, + "learning_rate": 1.9771167547730844e-05, + "loss": 0.6495, "step": 250 }, { - "epoch": 0.05, - "learning_rate": 1.9982479088172403e-05, - "loss": 0.7349, + "epoch": 0.09657560600230858, + "grad_norm": 1.02868988292841, + "learning_rate": 1.976850931381086e-05, + "loss": 0.6016, "step": 251 }, { - "epoch": 0.05, - "learning_rate": 1.9982108396105584e-05, - "loss": 0.9799, + "epoch": 0.09696036937283571, + "grad_norm": 1.0887609865250065, + "learning_rate": 1.9765835910005726e-05, + "loss": 0.6229, "step": 252 }, { - "epoch": 0.05, - "learning_rate": 1.9981733827102884e-05, - "loss": 0.9873, + "epoch": 0.09734513274336283, + "grad_norm": 1.0885545997746096, + "learning_rate": 1.9763147340467067e-05, + "loss": 0.6013, "step": 253 }, { - "epoch": 0.05, - "learning_rate": 1.998135538130979e-05, - "loss": 1.0044, + "epoch": 0.09772989611388996, + "grad_norm": 0.9304772229910628, + "learning_rate": 1.9760443609370074e-05, + "loss": 0.705, "step": 254 }, { - "epoch": 0.05, - "learning_rate": 1.998097305887328e-05, - "loss": 1.0546, + "epoch": 0.09811465948441708, + "grad_norm": 1.154776109527096, + "learning_rate": 1.9757724720913466e-05, + "loss": 0.6676, "step": 255 }, { - "epoch": 0.05, - "learning_rate": 1.9980586859941846e-05, - "loss": 0.9427, + "epoch": 0.09849942285494422, + "grad_norm": 1.1131330808934279, + "learning_rate": 1.975499067931951e-05, + "loss": 0.6133, "step": 256 }, { - "epoch": 0.05, - "learning_rate": 1.998019678466548e-05, - "loss": 0.9668, + "epoch": 0.09888418622547133, + "grad_norm": 1.0305366476479474, + "learning_rate": 1.9752241488834002e-05, + "loss": 0.6132, "step": 257 }, { - "epoch": 0.05, - "learning_rate": 1.997980283319568e-05, - "loss": 1.0277, + "epoch": 0.09926894959599845, + "grad_norm": 1.120726103115804, + "learning_rate": 1.974947715372626e-05, + "loss": 0.659, "step": 258 }, { - "epoch": 0.05, - "learning_rate": 1.9979405005685466e-05, - "loss": 1.086, + "epoch": 0.09965371296652559, + "grad_norm": 1.0656929800438308, + "learning_rate": 1.9746697678289128e-05, + "loss": 0.6077, "step": 259 }, { - "epoch": 0.05, - "learning_rate": 1.9979003302289336e-05, - "loss": 1.0453, + "epoch": 0.1000384763370527, + "grad_norm": 1.0322567374725347, + "learning_rate": 1.9743903066838954e-05, + "loss": 0.6342, "step": 260 }, { - "epoch": 0.05, - "learning_rate": 1.997859772316331e-05, - "loss": 1.0419, + "epoch": 0.10042323970757984, + "grad_norm": 0.989531442161535, + "learning_rate": 1.9741093323715597e-05, + "loss": 0.5577, "step": 261 }, { - "epoch": 0.05, - "learning_rate": 1.9978188268464912e-05, - "loss": 0.919, + "epoch": 0.10080800307810696, + "grad_norm": 0.9844125597109435, + "learning_rate": 1.9738268453282414e-05, + "loss": 0.5961, "step": 262 }, { - "epoch": 0.05, - "learning_rate": 1.997777493835317e-05, - "loss": 1.0195, + "epoch": 0.10119276644863409, + "grad_norm": 1.0323940763409276, + "learning_rate": 1.973542845992625e-05, + "loss": 0.6229, "step": 263 }, { - "epoch": 0.05, - "learning_rate": 1.9977357732988616e-05, - "loss": 1.0306, + "epoch": 0.10157752981916121, + "grad_norm": 1.1989945566571036, + "learning_rate": 1.9732573348057437e-05, + "loss": 0.6695, "step": 264 }, { - "epoch": 0.05, - "learning_rate": 1.9976936652533288e-05, - "loss": 1.0256, + "epoch": 0.10196229318968834, + "grad_norm": 1.1604985389161269, + "learning_rate": 1.9729703122109788e-05, + "loss": 0.6562, "step": 265 }, { - "epoch": 0.05, - "learning_rate": 1.997651169715073e-05, - "loss": 1.0129, + "epoch": 0.10234705656021546, + "grad_norm": 1.0047992028948332, + "learning_rate": 1.9726817786540584e-05, + "loss": 0.6026, "step": 266 }, { - "epoch": 0.05, - "learning_rate": 1.9976082867005985e-05, - "loss": 1.0456, + "epoch": 0.1027318199307426, + "grad_norm": 1.1286551617517824, + "learning_rate": 1.9723917345830568e-05, + "loss": 0.6604, "step": 267 }, { - "epoch": 0.05, - "learning_rate": 1.997565016226561e-05, - "loss": 1.0965, + "epoch": 0.10311658330126972, + "grad_norm": 1.0055043272676698, + "learning_rate": 1.9721001804483947e-05, + "loss": 0.6046, "step": 268 }, { - "epoch": 0.05, - "learning_rate": 1.997521358309766e-05, - "loss": 0.935, + "epoch": 0.10350134667179685, + "grad_norm": 0.9775689738131411, + "learning_rate": 1.9718071167028376e-05, + "loss": 0.6388, "step": 269 }, { - "epoch": 0.05, - "learning_rate": 1.99747731296717e-05, - "loss": 1.0057, + "epoch": 0.10388611004232397, + "grad_norm": 1.0318428114000868, + "learning_rate": 1.971512543801495e-05, + "loss": 0.604, "step": 270 }, { - "epoch": 0.05, - "learning_rate": 1.9974328802158798e-05, - "loss": 1.0005, + "epoch": 0.1042708734128511, + "grad_norm": 1.1408672157901112, + "learning_rate": 1.9712164622018197e-05, + "loss": 0.5693, "step": 271 }, { - "epoch": 0.05, - "learning_rate": 1.997388060073152e-05, - "loss": 1.0504, + "epoch": 0.10465563678337822, + "grad_norm": 1.2658781951032352, + "learning_rate": 1.9709188723636088e-05, + "loss": 0.6683, "step": 272 }, { - "epoch": 0.05, - "learning_rate": 1.9973428525563948e-05, - "loss": 1.0689, + "epoch": 0.10504040015390535, + "grad_norm": 1.1028278123104212, + "learning_rate": 1.9706197747490004e-05, + "loss": 0.605, "step": 273 }, { - "epoch": 0.05, - "learning_rate": 1.9972972576831656e-05, - "loss": 1.0149, + "epoch": 0.10542516352443247, + "grad_norm": 1.0762724315919445, + "learning_rate": 1.9703191698224742e-05, + "loss": 0.5848, "step": 274 }, { - "epoch": 0.05, - "learning_rate": 1.9972512754711738e-05, - "loss": 1.0045, + "epoch": 0.1058099268949596, + "grad_norm": 1.0808013763084716, + "learning_rate": 1.9700170580508514e-05, + "loss": 0.6183, "step": 275 }, { - "epoch": 0.05, - "learning_rate": 1.997204905938278e-05, - "loss": 1.0141, + "epoch": 0.10619469026548672, + "grad_norm": 1.2037069261515014, + "learning_rate": 1.969713439903292e-05, + "loss": 0.6955, "step": 276 }, { - "epoch": 0.05, - "learning_rate": 1.9971581491024873e-05, - "loss": 1.0781, + "epoch": 0.10657945363601386, + "grad_norm": 1.061017223837028, + "learning_rate": 1.9694083158512965e-05, + "loss": 0.656, "step": 277 }, { - "epoch": 0.05, - "learning_rate": 1.997111004981962e-05, - "loss": 1.0708, + "epoch": 0.10696421700654098, + "grad_norm": 1.1258832159160497, + "learning_rate": 1.9691016863687037e-05, + "loss": 0.6261, "step": 278 }, { - "epoch": 0.05, - "learning_rate": 1.9970634735950117e-05, - "loss": 1.0586, + "epoch": 0.10734898037706811, + "grad_norm": 1.1240232443303002, + "learning_rate": 1.9687935519316897e-05, + "loss": 0.6169, "step": 279 }, { - "epoch": 0.05, - "learning_rate": 1.9970155549600978e-05, - "loss": 1.0439, + "epoch": 0.10773374374759523, + "grad_norm": 1.0409490227935438, + "learning_rate": 1.9684839130187678e-05, + "loss": 0.675, "step": 280 }, { - "epoch": 0.05, - "learning_rate": 1.9969672490958304e-05, - "loss": 1.0546, + "epoch": 0.10811850711812235, + "grad_norm": 1.0194005983440315, + "learning_rate": 1.9681727701107885e-05, + "loss": 0.6535, "step": 281 }, { - "epoch": 0.05, - "learning_rate": 1.996918556020972e-05, - "loss": 1.036, + "epoch": 0.10850327048864948, + "grad_norm": 1.086161289236328, + "learning_rate": 1.967860123690937e-05, + "loss": 0.6312, "step": 282 }, { - "epoch": 0.05, - "learning_rate": 1.996869475754434e-05, - "loss": 1.0048, + "epoch": 0.1088880338591766, + "grad_norm": 0.9861333750306404, + "learning_rate": 1.967545974244734e-05, + "loss": 0.6512, "step": 283 }, { - "epoch": 0.05, - "learning_rate": 1.9968200083152784e-05, - "loss": 1.0732, + "epoch": 0.10927279722970373, + "grad_norm": 1.0066886431566908, + "learning_rate": 1.9672303222600333e-05, + "loss": 0.6418, "step": 284 }, { - "epoch": 0.05, - "learning_rate": 1.9967701537227175e-05, - "loss": 0.9336, + "epoch": 0.10965756060023085, + "grad_norm": 1.1474331802351048, + "learning_rate": 1.9669131682270232e-05, + "loss": 0.6865, "step": 285 }, { - "epoch": 0.06, - "learning_rate": 1.996719911996115e-05, - "loss": 1.0025, + "epoch": 0.11004232397075799, + "grad_norm": 0.9601309908617793, + "learning_rate": 1.966594512638224e-05, + "loss": 0.5927, "step": 286 }, { - "epoch": 0.06, - "learning_rate": 1.996669283154984e-05, - "loss": 0.8779, + "epoch": 0.1104270873412851, + "grad_norm": 1.015263788241438, + "learning_rate": 1.966274355988488e-05, + "loss": 0.6471, "step": 287 }, { - "epoch": 0.06, - "learning_rate": 1.996618267218988e-05, - "loss": 0.9774, + "epoch": 0.11081185071181224, + "grad_norm": 1.1246927708945536, + "learning_rate": 1.9659526987749987e-05, + "loss": 0.607, "step": 288 }, { - "epoch": 0.06, - "learning_rate": 1.996566864207941e-05, - "loss": 1.0096, + "epoch": 0.11119661408233936, + "grad_norm": 2.6453484421561817, + "learning_rate": 1.965629541497269e-05, + "loss": 0.7486, "step": 289 }, { - "epoch": 0.06, - "learning_rate": 1.9965150741418072e-05, - "loss": 1.0476, + "epoch": 0.11158137745286649, + "grad_norm": 1.0687433484027908, + "learning_rate": 1.9653048846571427e-05, + "loss": 0.6224, "step": 290 }, { - "epoch": 0.06, - "learning_rate": 1.9964628970407018e-05, - "loss": 1.0273, + "epoch": 0.11196614082339361, + "grad_norm": 1.1730634777920015, + "learning_rate": 1.964978728758791e-05, + "loss": 0.6665, "step": 291 }, { - "epoch": 0.06, - "learning_rate": 1.9964103329248892e-05, - "loss": 0.964, + "epoch": 0.11235090419392074, + "grad_norm": 1.0217247422196731, + "learning_rate": 1.9646510743087144e-05, + "loss": 0.6384, "step": 292 }, { - "epoch": 0.06, - "learning_rate": 1.996357381814785e-05, - "loss": 1.0109, + "epoch": 0.11273566756444786, + "grad_norm": 0.994668479022376, + "learning_rate": 1.9643219218157395e-05, + "loss": 0.603, "step": 293 }, { - "epoch": 0.06, - "learning_rate": 1.996304043730955e-05, - "loss": 1.0887, + "epoch": 0.113120430934975, + "grad_norm": 1.0143097523994742, + "learning_rate": 1.963991271791019e-05, + "loss": 0.62, "step": 294 }, { - "epoch": 0.06, - "learning_rate": 1.9962503186941143e-05, - "loss": 1.0046, + "epoch": 0.11350519430550211, + "grad_norm": 1.0326430178755834, + "learning_rate": 1.9636591247480323e-05, + "loss": 0.6127, "step": 295 }, { - "epoch": 0.06, - "learning_rate": 1.9961962067251298e-05, - "loss": 1.047, + "epoch": 0.11388995767602925, + "grad_norm": 1.0724698144718223, + "learning_rate": 1.963325481202583e-05, + "loss": 0.6482, "step": 296 }, { - "epoch": 0.06, - "learning_rate": 1.9961417078450177e-05, - "loss": 0.7934, + "epoch": 0.11427472104655637, + "grad_norm": 1.0746600644411541, + "learning_rate": 1.9629903416727987e-05, + "loss": 0.6321, "step": 297 }, { - "epoch": 0.06, - "learning_rate": 1.996086822074945e-05, - "loss": 1.0101, + "epoch": 0.1146594844170835, + "grad_norm": 1.1220829021290435, + "learning_rate": 1.96265370667913e-05, + "loss": 0.6111, "step": 298 }, { - "epoch": 0.06, - "learning_rate": 1.9960315494362286e-05, - "loss": 1.0211, + "epoch": 0.11504424778761062, + "grad_norm": 1.101414551326402, + "learning_rate": 1.9623155767443498e-05, + "loss": 0.5903, "step": 299 }, { - "epoch": 0.06, - "learning_rate": 1.9959758899503355e-05, - "loss": 0.9895, + "epoch": 0.11542901115813775, + "grad_norm": 1.0629968314044134, + "learning_rate": 1.9619759523935532e-05, + "loss": 0.6069, "step": 300 }, { - "epoch": 0.06, - "learning_rate": 1.995919843638883e-05, - "loss": 1.1163, + "epoch": 0.11581377452866487, + "grad_norm": 1.0938575071329402, + "learning_rate": 1.961634834154156e-05, + "loss": 0.6464, "step": 301 }, { - "epoch": 0.06, - "learning_rate": 1.9958634105236395e-05, - "loss": 1.1154, + "epoch": 0.116198537899192, + "grad_norm": 1.0819311123887316, + "learning_rate": 1.9612922225558924e-05, + "loss": 0.6529, "step": 302 }, { - "epoch": 0.06, - "learning_rate": 1.9958065906265228e-05, - "loss": 0.9873, + "epoch": 0.11658330126971912, + "grad_norm": 1.1158872036780034, + "learning_rate": 1.960948118130818e-05, + "loss": 0.6456, "step": 303 }, { - "epoch": 0.06, - "learning_rate": 1.9957493839696013e-05, - "loss": 0.9926, + "epoch": 0.11696806464024626, + "grad_norm": 1.0501610105602521, + "learning_rate": 1.9606025214133046e-05, + "loss": 0.6542, "step": 304 }, { - "epoch": 0.06, - "learning_rate": 1.9956917905750926e-05, - "loss": 1.0755, + "epoch": 0.11735282801077337, + "grad_norm": 1.0893528243798984, + "learning_rate": 1.960255432940043e-05, + "loss": 0.6518, "step": 305 }, { - "epoch": 0.06, - "learning_rate": 1.995633810465366e-05, - "loss": 0.9618, + "epoch": 0.1177375913813005, + "grad_norm": 1.0424268387570856, + "learning_rate": 1.9599068532500394e-05, + "loss": 0.5985, "step": 306 }, { - "epoch": 0.06, - "learning_rate": 1.99557544366294e-05, - "loss": 1.0251, + "epoch": 0.11812235475182763, + "grad_norm": 1.086663352558066, + "learning_rate": 1.9595567828846166e-05, + "loss": 0.6233, "step": 307 }, { - "epoch": 0.06, - "learning_rate": 1.9955166901904838e-05, - "loss": 0.9409, + "epoch": 0.11850711812235475, + "grad_norm": 1.0545055600245634, + "learning_rate": 1.9592052223874115e-05, + "loss": 0.587, "step": 308 }, { - "epoch": 0.06, - "learning_rate": 1.9954575500708164e-05, - "loss": 1.0244, + "epoch": 0.11889188149288188, + "grad_norm": 1.1141641876238628, + "learning_rate": 1.9588521723043764e-05, + "loss": 0.6719, "step": 309 }, { - "epoch": 0.06, - "learning_rate": 1.995398023326907e-05, - "loss": 0.9579, + "epoch": 0.119276644863409, + "grad_norm": 1.029853017492757, + "learning_rate": 1.9584976331837758e-05, + "loss": 0.5939, "step": 310 }, { - "epoch": 0.06, - "learning_rate": 1.9953381099818756e-05, - "loss": 1.0114, + "epoch": 0.11966140823393613, + "grad_norm": 1.0009200013305726, + "learning_rate": 1.9581416055761865e-05, + "loss": 0.5911, "step": 311 }, { - "epoch": 0.06, - "learning_rate": 1.9952778100589912e-05, - "loss": 0.9119, + "epoch": 0.12004617160446325, + "grad_norm": 1.0528645162721275, + "learning_rate": 1.9577840900344974e-05, + "loss": 0.6415, "step": 312 }, { - "epoch": 0.06, - "learning_rate": 1.9952171235816747e-05, - "loss": 1.0409, + "epoch": 0.12043093497499038, + "grad_norm": 1.0069708545703138, + "learning_rate": 1.957425087113908e-05, + "loss": 0.6251, "step": 313 }, { - "epoch": 0.06, - "learning_rate": 1.9951560505734948e-05, - "loss": 1.1015, + "epoch": 0.1208156983455175, + "grad_norm": 1.1165323677684353, + "learning_rate": 1.9570645973719273e-05, + "loss": 0.6311, "step": 314 }, { - "epoch": 0.06, - "learning_rate": 1.9950945910581718e-05, - "loss": 0.9308, + "epoch": 0.12120046171604464, + "grad_norm": 1.1337336698853253, + "learning_rate": 1.9567026213683728e-05, + "loss": 0.6651, "step": 315 }, { - "epoch": 0.06, - "learning_rate": 1.9950327450595766e-05, - "loss": 1.0666, + "epoch": 0.12158522508657175, + "grad_norm": 1.07315145372861, + "learning_rate": 1.956339159665371e-05, + "loss": 0.6303, "step": 316 }, { - "epoch": 0.06, - "learning_rate": 1.9949705126017286e-05, - "loss": 1.0662, + "epoch": 0.12196998845709889, + "grad_norm": 1.0119176213417673, + "learning_rate": 1.9559742128273558e-05, + "loss": 0.5944, "step": 317 }, { - "epoch": 0.06, - "learning_rate": 1.9949078937087988e-05, - "loss": 1.0236, + "epoch": 0.12235475182762601, + "grad_norm": 1.0758861980125103, + "learning_rate": 1.9556077814210662e-05, + "loss": 0.6179, "step": 318 }, { - "epoch": 0.06, - "learning_rate": 1.994844888405107e-05, - "loss": 0.9933, + "epoch": 0.12273951519815314, + "grad_norm": 1.222591817040764, + "learning_rate": 1.955239866015547e-05, + "loss": 0.619, "step": 319 }, { - "epoch": 0.06, - "learning_rate": 1.9947814967151246e-05, - "loss": 0.956, + "epoch": 0.12312427856868026, + "grad_norm": 1.1363089074935568, + "learning_rate": 1.954870467182149e-05, + "loss": 0.6041, "step": 320 }, { - "epoch": 0.06, - "learning_rate": 1.9947177186634716e-05, - "loss": 0.9638, + "epoch": 0.12350904193920739, + "grad_norm": 1.0621992042738941, + "learning_rate": 1.9544995854945248e-05, + "loss": 0.6231, "step": 321 }, { - "epoch": 0.06, - "learning_rate": 1.9946535542749187e-05, - "loss": 0.6978, + "epoch": 0.12389380530973451, + "grad_norm": 1.0821456117748103, + "learning_rate": 1.9541272215286304e-05, + "loss": 0.65, "step": 322 }, { - "epoch": 0.06, - "learning_rate": 1.9945890035743866e-05, - "loss": 1.1306, + "epoch": 0.12427856868026164, + "grad_norm": 1.1883402330002892, + "learning_rate": 1.9537533758627242e-05, + "loss": 0.6468, "step": 323 }, { - "epoch": 0.06, - "learning_rate": 1.9945240665869465e-05, - "loss": 1.0411, + "epoch": 0.12466333205078876, + "grad_norm": 1.154278539731404, + "learning_rate": 1.9533780490773645e-05, + "loss": 0.6944, "step": 324 }, { - "epoch": 0.06, - "learning_rate": 1.9944587433378187e-05, - "loss": 1.08, + "epoch": 0.1250480954213159, + "grad_norm": 1.0766934543994864, + "learning_rate": 1.953001241755411e-05, + "loss": 0.5877, "step": 325 }, { - "epoch": 0.06, - "learning_rate": 1.994393033852374e-05, - "loss": 0.8562, + "epoch": 0.12543285879184302, + "grad_norm": 1.1291619751964748, + "learning_rate": 1.952622954482022e-05, + "loss": 0.6338, "step": 326 }, { - "epoch": 0.06, - "learning_rate": 1.9943269381561334e-05, - "loss": 1.107, + "epoch": 0.12581762216237014, + "grad_norm": 1.941232125386861, + "learning_rate": 1.9522431878446536e-05, + "loss": 0.6943, "step": 327 }, { - "epoch": 0.06, - "learning_rate": 1.994260456274768e-05, - "loss": 1.0317, + "epoch": 0.12620238553289725, + "grad_norm": 1.1656551295037048, + "learning_rate": 1.95186194243306e-05, + "loss": 0.6739, "step": 328 }, { - "epoch": 0.06, - "learning_rate": 1.9941935882340976e-05, - "loss": 0.9818, + "epoch": 0.1265871489034244, + "grad_norm": 1.1042034842645334, + "learning_rate": 1.9514792188392914e-05, + "loss": 0.5946, "step": 329 }, { - "epoch": 0.06, - "learning_rate": 1.994126334060094e-05, - "loss": 0.959, + "epoch": 0.12697191227395152, + "grad_norm": 1.2340340118075774, + "learning_rate": 1.9510950176576933e-05, + "loss": 0.6319, "step": 330 }, { - "epoch": 0.06, - "learning_rate": 1.994058693778878e-05, - "loss": 1.0226, + "epoch": 0.12735667564447864, + "grad_norm": 1.0970385374287575, + "learning_rate": 1.950709339484907e-05, + "loss": 0.6528, "step": 331 }, { - "epoch": 0.06, - "learning_rate": 1.9939906674167192e-05, - "loss": 0.9951, + "epoch": 0.12774143901500576, + "grad_norm": 0.990143224823563, + "learning_rate": 1.9503221849198655e-05, + "loss": 0.6348, "step": 332 }, { - "epoch": 0.06, - "learning_rate": 1.993922255000039e-05, - "loss": 1.0242, + "epoch": 0.1281262023855329, + "grad_norm": 1.0633573498195679, + "learning_rate": 1.9499335545637968e-05, + "loss": 0.6436, "step": 333 }, { - "epoch": 0.06, - "learning_rate": 1.993853456555408e-05, - "loss": 0.9392, + "epoch": 0.12851096575606002, + "grad_norm": 1.1347977991244929, + "learning_rate": 1.9495434490202188e-05, + "loss": 0.6473, "step": 334 }, { - "epoch": 0.06, - "learning_rate": 1.9937842721095468e-05, - "loss": 1.0317, + "epoch": 0.12889572912658714, + "grad_norm": 1.1158650614515702, + "learning_rate": 1.9491518688949417e-05, + "loss": 0.6937, "step": 335 }, { - "epoch": 0.06, - "learning_rate": 1.9937147016893257e-05, - "loss": 1.0457, + "epoch": 0.12928049249711426, + "grad_norm": 1.2415714681628507, + "learning_rate": 1.948758814796064e-05, + "loss": 0.686, "step": 336 }, { - "epoch": 0.06, - "learning_rate": 1.9936447453217646e-05, - "loss": 0.9401, + "epoch": 0.1296652558676414, + "grad_norm": 1.0464639505858888, + "learning_rate": 1.9483642873339753e-05, + "loss": 0.6142, "step": 337 }, { - "epoch": 0.07, - "learning_rate": 1.9935744030340347e-05, - "loss": 0.9969, + "epoch": 0.13005001923816853, + "grad_norm": 1.2507711111692448, + "learning_rate": 1.9479682871213515e-05, + "loss": 0.6248, "step": 338 }, { - "epoch": 0.07, - "learning_rate": 1.9935036748534555e-05, - "loss": 1.0251, + "epoch": 0.13043478260869565, + "grad_norm": 1.0928688652395708, + "learning_rate": 1.947570814773156e-05, + "loss": 0.6408, "step": 339 }, { - "epoch": 0.07, - "learning_rate": 1.993432560807497e-05, - "loss": 0.9947, + "epoch": 0.13081954597922277, + "grad_norm": 1.178222345967826, + "learning_rate": 1.9471718709066392e-05, + "loss": 0.6478, "step": 340 }, { - "epoch": 0.07, - "learning_rate": 1.993361060923779e-05, - "loss": 1.0605, + "epoch": 0.13120430934974991, + "grad_norm": 1.2583356574306597, + "learning_rate": 1.9467714561413358e-05, + "loss": 0.634, "step": 341 }, { - "epoch": 0.07, - "learning_rate": 1.9932891752300717e-05, - "loss": 0.9203, + "epoch": 0.13158907272027703, + "grad_norm": 1.1560490101406473, + "learning_rate": 1.9463695710990648e-05, + "loss": 0.6275, "step": 342 }, { - "epoch": 0.07, - "learning_rate": 1.9932169037542947e-05, - "loss": 1.0823, + "epoch": 0.13197383609080415, + "grad_norm": 1.2609691127778428, + "learning_rate": 1.9459662164039283e-05, + "loss": 0.6354, "step": 343 }, { - "epoch": 0.07, - "learning_rate": 1.9931442465245164e-05, - "loss": 0.9826, + "epoch": 0.13235859946133127, + "grad_norm": 1.1393322673360609, + "learning_rate": 1.9455613926823115e-05, + "loss": 0.612, "step": 344 }, { - "epoch": 0.07, - "learning_rate": 1.9930712035689576e-05, - "loss": 0.9899, + "epoch": 0.13274336283185842, + "grad_norm": 1.1608156032118326, + "learning_rate": 1.9451551005628803e-05, + "loss": 0.5705, "step": 345 }, { - "epoch": 0.07, - "learning_rate": 1.992997774915986e-05, - "loss": 1.0071, + "epoch": 0.13312812620238554, + "grad_norm": 1.1109495952182638, + "learning_rate": 1.9447473406765803e-05, + "loss": 0.6368, "step": 346 }, { - "epoch": 0.07, - "learning_rate": 1.992923960594121e-05, - "loss": 1.0676, + "epoch": 0.13351288957291266, + "grad_norm": 1.11284703532172, + "learning_rate": 1.9443381136566382e-05, + "loss": 0.6367, "step": 347 }, { - "epoch": 0.07, - "learning_rate": 1.9928497606320308e-05, - "loss": 0.9934, + "epoch": 0.13389765294343978, + "grad_norm": 0.9831224948542597, + "learning_rate": 1.943927420138557e-05, + "loss": 0.5998, "step": 348 }, { - "epoch": 0.07, - "learning_rate": 1.992775175058535e-05, - "loss": 1.0784, + "epoch": 0.13428241631396692, + "grad_norm": 1.0639721298133291, + "learning_rate": 1.9435152607601187e-05, + "loss": 0.5893, "step": 349 }, { - "epoch": 0.07, - "learning_rate": 1.9927002039026002e-05, - "loss": 0.8996, + "epoch": 0.13466717968449404, + "grad_norm": 0.9959457949268063, + "learning_rate": 1.9431016361613816e-05, + "loss": 0.5955, "step": 350 }, { - "epoch": 0.07, - "learning_rate": 1.9926248471933453e-05, - "loss": 1.0347, + "epoch": 0.13505194305502116, + "grad_norm": 1.1683559966452763, + "learning_rate": 1.9426865469846773e-05, + "loss": 0.6813, "step": 351 }, { - "epoch": 0.07, - "learning_rate": 1.9925491049600382e-05, - "loss": 0.7679, + "epoch": 0.13543670642554828, + "grad_norm": 1.163303206856684, + "learning_rate": 1.942269993874615e-05, + "loss": 0.6512, "step": 352 }, { - "epoch": 0.07, - "learning_rate": 1.9924729772320953e-05, - "loss": 0.955, + "epoch": 0.1358214697960754, + "grad_norm": 1.0684541440847537, + "learning_rate": 1.9418519774780748e-05, + "loss": 0.6384, "step": 353 }, { - "epoch": 0.07, - "learning_rate": 1.9923964640390846e-05, - "loss": 1.0915, + "epoch": 0.13620623316660255, + "grad_norm": 0.9991361068639223, + "learning_rate": 1.9414324984442102e-05, + "loss": 0.6062, "step": 354 }, { - "epoch": 0.07, - "learning_rate": 1.9923195654107227e-05, - "loss": 0.944, + "epoch": 0.13659099653712967, + "grad_norm": 1.0235362404371509, + "learning_rate": 1.9410115574244462e-05, + "loss": 0.6598, "step": 355 }, { - "epoch": 0.07, - "learning_rate": 1.992242281376876e-05, - "loss": 0.9906, + "epoch": 0.13697575990765679, + "grad_norm": 1.0579306087008729, + "learning_rate": 1.9405891550724778e-05, + "loss": 0.6483, "step": 356 }, { - "epoch": 0.07, - "learning_rate": 1.9921646119675606e-05, - "loss": 1.0243, + "epoch": 0.1373605232781839, + "grad_norm": 1.021569142075857, + "learning_rate": 1.9401652920442694e-05, + "loss": 0.5813, "step": 357 }, { - "epoch": 0.07, - "learning_rate": 1.9920865572129426e-05, - "loss": 1.1007, + "epoch": 0.13774528664871105, + "grad_norm": 1.1108703115764964, + "learning_rate": 1.939739968998054e-05, + "loss": 0.6437, "step": 358 }, { - "epoch": 0.07, - "learning_rate": 1.9920081171433377e-05, - "loss": 1.0777, + "epoch": 0.13813005001923817, + "grad_norm": 1.0158557363053349, + "learning_rate": 1.939313186594331e-05, + "loss": 0.6356, "step": 359 }, { - "epoch": 0.07, - "learning_rate": 1.991929291789211e-05, - "loss": 1.0413, + "epoch": 0.1385148133897653, + "grad_norm": 1.0619012371195269, + "learning_rate": 1.938884945495868e-05, + "loss": 0.6102, "step": 360 }, { - "epoch": 0.07, - "learning_rate": 1.9918500811811778e-05, - "loss": 0.9728, + "epoch": 0.1388995767602924, + "grad_norm": 1.186765590767799, + "learning_rate": 1.938455246367696e-05, + "loss": 0.6439, "step": 361 }, { - "epoch": 0.07, - "learning_rate": 1.991770485350002e-05, - "loss": 0.9625, + "epoch": 0.13928434013081956, + "grad_norm": 1.1049512447919285, + "learning_rate": 1.938024089877111e-05, + "loss": 0.6561, "step": 362 }, { - "epoch": 0.07, - "learning_rate": 1.991690504326597e-05, - "loss": 0.9925, + "epoch": 0.13966910350134668, + "grad_norm": 1.0099909313245445, + "learning_rate": 1.9375914766936723e-05, + "loss": 0.626, "step": 363 }, { - "epoch": 0.07, - "learning_rate": 1.9916101381420285e-05, - "loss": 1.0251, + "epoch": 0.1400538668718738, + "grad_norm": 0.9861718446608772, + "learning_rate": 1.937157407489201e-05, + "loss": 0.5873, "step": 364 }, { - "epoch": 0.07, - "learning_rate": 1.9915293868275083e-05, - "loss": 0.9842, + "epoch": 0.1404386302424009, + "grad_norm": 1.1431397814967321, + "learning_rate": 1.936721882937779e-05, + "loss": 0.6286, "step": 365 }, { - "epoch": 0.07, - "learning_rate": 1.9914482504143996e-05, - "loss": 1.0218, + "epoch": 0.14082339361292806, + "grad_norm": 1.1873417059026716, + "learning_rate": 1.93628490371575e-05, + "loss": 0.6505, "step": 366 }, { - "epoch": 0.07, - "learning_rate": 1.9913667289342147e-05, - "loss": 1.0131, + "epoch": 0.14120815698345518, + "grad_norm": 1.0429861887957208, + "learning_rate": 1.9358464705017143e-05, + "loss": 0.5926, "step": 367 }, { - "epoch": 0.07, - "learning_rate": 1.991284822418616e-05, - "loss": 0.9626, + "epoch": 0.1415929203539823, + "grad_norm": 1.0202260124520914, + "learning_rate": 1.9354065839765316e-05, + "loss": 0.6339, "step": 368 }, { - "epoch": 0.07, - "learning_rate": 1.9912025308994146e-05, - "loss": 1.0161, + "epoch": 0.14197768372450942, + "grad_norm": 1.0469072065029588, + "learning_rate": 1.9349652448233187e-05, + "loss": 0.6472, "step": 369 }, { - "epoch": 0.07, - "learning_rate": 1.9911198544085723e-05, - "loss": 1.0663, + "epoch": 0.14236244709503657, + "grad_norm": 1.0391750238337412, + "learning_rate": 1.934522453727447e-05, + "loss": 0.6016, "step": 370 }, { - "epoch": 0.07, - "learning_rate": 1.991036792978199e-05, - "loss": 1.0539, + "epoch": 0.14274721046556368, + "grad_norm": 2.9537737426138864, + "learning_rate": 1.934078211376544e-05, + "loss": 0.7237, "step": 371 }, { - "epoch": 0.07, - "learning_rate": 1.990953346640555e-05, - "loss": 0.9882, + "epoch": 0.1431319738360908, + "grad_norm": 1.1637123301984997, + "learning_rate": 1.93363251846049e-05, + "loss": 0.6013, "step": 372 }, { - "epoch": 0.07, - "learning_rate": 1.9908695154280496e-05, - "loss": 0.9882, + "epoch": 0.14351673720661792, + "grad_norm": 1.0182809458940725, + "learning_rate": 1.9331853756714185e-05, + "loss": 0.5789, "step": 373 }, { - "epoch": 0.07, - "learning_rate": 1.9907852993732425e-05, - "loss": 1.0702, + "epoch": 0.14390150057714507, + "grad_norm": 1.0538009839411813, + "learning_rate": 1.9327367837037142e-05, + "loss": 0.6373, "step": 374 }, { - "epoch": 0.07, - "learning_rate": 1.990700698508842e-05, - "loss": 1.0336, + "epoch": 0.1442862639476722, + "grad_norm": 0.9858519668153444, + "learning_rate": 1.9322867432540126e-05, + "loss": 0.6357, "step": 375 }, { - "epoch": 0.07, - "learning_rate": 1.990615712867706e-05, - "loss": 0.9564, + "epoch": 0.1446710273181993, + "grad_norm": 1.0522789265805077, + "learning_rate": 1.9318352550211986e-05, + "loss": 0.5974, "step": 376 }, { - "epoch": 0.07, - "learning_rate": 1.9905303424828418e-05, - "loss": 1.0957, + "epoch": 0.14505579068872643, + "grad_norm": 1.0722454014217944, + "learning_rate": 1.9313823197064042e-05, + "loss": 0.6016, "step": 377 }, { - "epoch": 0.07, - "learning_rate": 1.9904445873874068e-05, - "loss": 0.9727, + "epoch": 0.14544055405925355, + "grad_norm": 1.1541246443587863, + "learning_rate": 1.9309279380130112e-05, + "loss": 0.6491, "step": 378 }, { - "epoch": 0.07, - "learning_rate": 1.9903584476147066e-05, - "loss": 0.9757, + "epoch": 0.1458253174297807, + "grad_norm": 1.1984827532660347, + "learning_rate": 1.930472110646645e-05, + "loss": 0.6197, "step": 379 }, { - "epoch": 0.07, - "learning_rate": 1.9902719231981975e-05, - "loss": 1.0248, + "epoch": 0.1462100808003078, + "grad_norm": 1.0395427346772779, + "learning_rate": 1.930014838315177e-05, + "loss": 0.6492, "step": 380 }, { - "epoch": 0.07, - "learning_rate": 1.9901850141714843e-05, - "loss": 1.0027, + "epoch": 0.14659484417083493, + "grad_norm": 1.009790924144865, + "learning_rate": 1.9295561217287226e-05, + "loss": 0.5565, "step": 381 }, { - "epoch": 0.07, - "learning_rate": 1.9900977205683213e-05, - "loss": 0.9332, + "epoch": 0.14697960754136205, + "grad_norm": 0.9962724856547507, + "learning_rate": 1.9290959615996407e-05, + "loss": 0.6133, "step": 382 }, { - "epoch": 0.07, - "learning_rate": 1.9900100424226124e-05, - "loss": 0.9918, + "epoch": 0.1473643709118892, + "grad_norm": 0.9897468636083305, + "learning_rate": 1.9286343586425307e-05, + "loss": 0.6075, "step": 383 }, { - "epoch": 0.07, - "learning_rate": 1.9899219797684113e-05, - "loss": 0.9289, + "epoch": 0.14774913428241632, + "grad_norm": 1.07915432002081, + "learning_rate": 1.9281713135742333e-05, + "loss": 0.6388, "step": 384 }, { - "epoch": 0.07, - "learning_rate": 1.98983353263992e-05, - "loss": 1.0084, + "epoch": 0.14813389765294344, + "grad_norm": 0.9536394249562014, + "learning_rate": 1.9277068271138287e-05, + "loss": 0.5796, "step": 385 }, { - "epoch": 0.07, - "learning_rate": 1.9897447010714905e-05, - "loss": 0.9836, + "epoch": 0.14851866102347056, + "grad_norm": 0.8670604180464514, + "learning_rate": 1.927240899982635e-05, + "loss": 0.667, "step": 386 }, { - "epoch": 0.07, - "learning_rate": 1.989655485097624e-05, - "loss": 1.0158, + "epoch": 0.1489034243939977, + "grad_norm": 1.139555756474863, + "learning_rate": 1.9267735329042086e-05, + "loss": 0.6283, "step": 387 }, { - "epoch": 0.07, - "learning_rate": 1.989565884752971e-05, - "loss": 0.9539, + "epoch": 0.14928818776452482, + "grad_norm": 1.095995944784319, + "learning_rate": 1.926304726604341e-05, + "loss": 0.6102, "step": 388 }, { - "epoch": 0.07, - "learning_rate": 1.9894759000723308e-05, - "loss": 1.0215, + "epoch": 0.14967295113505194, + "grad_norm": 1.3630005415325652, + "learning_rate": 1.925834481811059e-05, + "loss": 0.6534, "step": 389 }, { - "epoch": 0.08, - "learning_rate": 1.9893855310906526e-05, - "loss": 1.0317, + "epoch": 0.15005771450557906, + "grad_norm": 0.7374773557175287, + "learning_rate": 1.925362799254623e-05, + "loss": 0.6464, "step": 390 }, { - "epoch": 0.08, - "learning_rate": 1.9892947778430352e-05, - "loss": 1.0507, + "epoch": 0.1504424778761062, + "grad_norm": 1.0710418222364892, + "learning_rate": 1.9248896796675277e-05, + "loss": 0.5942, "step": 391 }, { - "epoch": 0.08, - "learning_rate": 1.9892036403647256e-05, - "loss": 1.059, + "epoch": 0.15082724124663333, + "grad_norm": 1.1143692245370502, + "learning_rate": 1.9244151237844975e-05, + "loss": 0.6812, "step": 392 }, { - "epoch": 0.08, - "learning_rate": 1.9891121186911207e-05, - "loss": 1.008, + "epoch": 0.15121200461716044, + "grad_norm": 1.139009530903883, + "learning_rate": 1.923939132342488e-05, + "loss": 0.6329, "step": 393 }, { - "epoch": 0.08, - "learning_rate": 1.9890202128577664e-05, - "loss": 1.0722, + "epoch": 0.15159676798768756, + "grad_norm": 1.0105317355890624, + "learning_rate": 1.923461706080685e-05, + "loss": 0.6131, "step": 394 }, { - "epoch": 0.08, - "learning_rate": 1.988927922900358e-05, - "loss": 0.9616, + "epoch": 0.1519815313582147, + "grad_norm": 1.127821681130332, + "learning_rate": 1.9229828457405005e-05, + "loss": 0.6636, "step": 395 }, { - "epoch": 0.08, - "learning_rate": 1.9888352488547397e-05, - "loss": 1.0264, + "epoch": 0.15236629472874183, + "grad_norm": 1.0266261419084677, + "learning_rate": 1.922502552065576e-05, + "loss": 0.5822, "step": 396 }, { - "epoch": 0.08, - "learning_rate": 1.988742190756905e-05, - "loss": 0.9467, + "epoch": 0.15275105809926895, + "grad_norm": 1.0967517046894555, + "learning_rate": 1.9220208258017763e-05, + "loss": 0.605, "step": 397 }, { - "epoch": 0.08, - "learning_rate": 1.9886487486429966e-05, - "loss": 1.0142, + "epoch": 0.15313582146979607, + "grad_norm": 1.060678380538241, + "learning_rate": 1.921537667697193e-05, + "loss": 0.6337, "step": 398 }, { - "epoch": 0.08, - "learning_rate": 1.9885549225493064e-05, - "loss": 1.022, + "epoch": 0.1535205848403232, + "grad_norm": 0.970469581597766, + "learning_rate": 1.9210530785021405e-05, + "loss": 0.5514, "step": 399 }, { - "epoch": 0.08, - "learning_rate": 1.9884607125122753e-05, - "loss": 0.9408, + "epoch": 0.15390534821085033, + "grad_norm": 0.9587503923261159, + "learning_rate": 1.920567058969155e-05, + "loss": 0.6519, "step": 400 }, { - "epoch": 0.08, - "learning_rate": 1.988366118568494e-05, - "loss": 0.9511, + "epoch": 0.15429011158137745, + "grad_norm": 1.008548023549872, + "learning_rate": 1.9200796098529956e-05, + "loss": 0.662, "step": 401 }, { - "epoch": 0.08, - "learning_rate": 1.988271140754701e-05, - "loss": 1.0015, + "epoch": 0.15467487495190457, + "grad_norm": 1.044046143448498, + "learning_rate": 1.9195907319106394e-05, + "loss": 0.6701, "step": 402 }, { - "epoch": 0.08, - "learning_rate": 1.9881757791077848e-05, - "loss": 1.0169, + "epoch": 0.1550596383224317, + "grad_norm": 1.070371675616831, + "learning_rate": 1.919100425901283e-05, + "loss": 0.6524, "step": 403 }, { - "epoch": 0.08, - "learning_rate": 1.9880800336647825e-05, - "loss": 1.0117, + "epoch": 0.15544440169295884, + "grad_norm": 1.1196075676371058, + "learning_rate": 1.918608692586342e-05, + "loss": 0.6317, "step": 404 }, { - "epoch": 0.08, - "learning_rate": 1.987983904462881e-05, - "loss": 1.0275, + "epoch": 0.15582916506348596, + "grad_norm": 1.0296821774517884, + "learning_rate": 1.9181155327294468e-05, + "loss": 0.6129, "step": 405 }, { - "epoch": 0.08, - "learning_rate": 1.9878873915394154e-05, - "loss": 1.0243, + "epoch": 0.15621392843401308, + "grad_norm": 0.9857879152317305, + "learning_rate": 1.9176209470964446e-05, + "loss": 0.6192, "step": 406 }, { - "epoch": 0.08, - "learning_rate": 1.9877904949318704e-05, - "loss": 1.0251, + "epoch": 0.1565986918045402, + "grad_norm": 0.7743213255158279, + "learning_rate": 1.9171249364553956e-05, + "loss": 0.6679, "step": 407 }, { - "epoch": 0.08, - "learning_rate": 1.9876932146778796e-05, - "loss": 0.9947, + "epoch": 0.15698345517506734, + "grad_norm": 1.0832112156743108, + "learning_rate": 1.916627501576573e-05, + "loss": 0.6687, "step": 408 }, { - "epoch": 0.08, - "learning_rate": 1.9875955508152254e-05, - "loss": 0.7212, + "epoch": 0.15736821854559446, + "grad_norm": 0.9933306147792358, + "learning_rate": 1.9161286432324628e-05, + "loss": 0.5842, "step": 409 }, { - "epoch": 0.08, - "learning_rate": 1.987497503381839e-05, - "loss": 0.7463, + "epoch": 0.15775298191612158, + "grad_norm": 0.9900354797288368, + "learning_rate": 1.9156283621977603e-05, + "loss": 0.6265, "step": 410 }, { - "epoch": 0.08, - "learning_rate": 1.9873990724158014e-05, - "loss": 0.9364, + "epoch": 0.1581377452866487, + "grad_norm": 0.9969904596584508, + "learning_rate": 1.915126659249371e-05, + "loss": 0.582, "step": 411 }, { - "epoch": 0.08, - "learning_rate": 1.987300257955342e-05, - "loss": 1.0188, + "epoch": 0.15852250865717585, + "grad_norm": 1.0849568562019127, + "learning_rate": 1.914623535166408e-05, + "loss": 0.6032, "step": 412 }, { - "epoch": 0.08, - "learning_rate": 1.987201060038839e-05, - "loss": 0.9958, + "epoch": 0.15890727202770297, + "grad_norm": 1.040830441959859, + "learning_rate": 1.9141189907301922e-05, + "loss": 0.6116, "step": 413 }, { - "epoch": 0.08, - "learning_rate": 1.9871014787048197e-05, - "loss": 1.0324, + "epoch": 0.1592920353982301, + "grad_norm": 0.9911329543149708, + "learning_rate": 1.913613026724249e-05, + "loss": 0.6212, "step": 414 }, { - "epoch": 0.08, - "learning_rate": 1.9870015139919606e-05, - "loss": 1.0062, + "epoch": 0.1596767987687572, + "grad_norm": 1.0562341734252012, + "learning_rate": 1.9131056439343095e-05, + "loss": 0.6329, "step": 415 }, { - "epoch": 0.08, - "learning_rate": 1.9869011659390866e-05, - "loss": 1.074, + "epoch": 0.16006156213928435, + "grad_norm": 1.0384316112429088, + "learning_rate": 1.9125968431483068e-05, + "loss": 0.6026, "step": 416 }, { - "epoch": 0.08, - "learning_rate": 1.9868004345851716e-05, - "loss": 1.0675, + "epoch": 0.16044632550981147, + "grad_norm": 1.0738579219643074, + "learning_rate": 1.912086625156377e-05, + "loss": 0.5909, "step": 417 }, { - "epoch": 0.08, - "learning_rate": 1.9866993199693393e-05, - "loss": 1.0013, + "epoch": 0.1608310888803386, + "grad_norm": 1.0354800877544255, + "learning_rate": 1.911574990750857e-05, + "loss": 0.6155, "step": 418 }, { - "epoch": 0.08, - "learning_rate": 1.98659782213086e-05, - "loss": 1.0646, + "epoch": 0.1612158522508657, + "grad_norm": 1.0987044498684562, + "learning_rate": 1.9110619407262828e-05, + "loss": 0.6062, "step": 419 }, { - "epoch": 0.08, - "learning_rate": 1.986495941109156e-05, - "loss": 0.9574, + "epoch": 0.16160061562139286, + "grad_norm": 1.055313493924488, + "learning_rate": 1.9105474758793897e-05, + "loss": 0.6577, "step": 420 }, { - "epoch": 0.08, - "learning_rate": 1.9863936769437956e-05, - "loss": 1.0685, + "epoch": 0.16198537899191998, + "grad_norm": 1.078811521877104, + "learning_rate": 1.9100315970091088e-05, + "loss": 0.6828, "step": 421 }, { - "epoch": 0.08, - "learning_rate": 1.986291029674497e-05, - "loss": 1.0098, + "epoch": 0.1623701423624471, + "grad_norm": 0.9856857447955188, + "learning_rate": 1.909514304916568e-05, + "loss": 0.603, "step": 422 }, { - "epoch": 0.08, - "learning_rate": 1.986187999341128e-05, - "loss": 1.0702, + "epoch": 0.16275490573297421, + "grad_norm": 1.0647640336723005, + "learning_rate": 1.9089956004050893e-05, + "loss": 0.6415, "step": 423 }, { - "epoch": 0.08, - "learning_rate": 1.9860845859837034e-05, - "loss": 0.9792, + "epoch": 0.16313966910350133, + "grad_norm": 1.1044389425633734, + "learning_rate": 1.908475484280189e-05, + "loss": 0.6378, "step": 424 }, { - "epoch": 0.08, - "learning_rate": 1.985980789642388e-05, - "loss": 0.8752, + "epoch": 0.16352443247402848, + "grad_norm": 1.055349024280057, + "learning_rate": 1.907953957349575e-05, + "loss": 0.6489, "step": 425 }, { - "epoch": 0.08, - "learning_rate": 1.985876610357496e-05, - "loss": 1.0115, + "epoch": 0.1639091958445556, + "grad_norm": 0.9624971201410555, + "learning_rate": 1.9074310204231457e-05, + "loss": 0.631, "step": 426 }, { - "epoch": 0.08, - "learning_rate": 1.9857720481694887e-05, - "loss": 1.0267, + "epoch": 0.16429395921508272, + "grad_norm": 1.1128070429622436, + "learning_rate": 1.9069066743129893e-05, + "loss": 0.6257, "step": 427 }, { - "epoch": 0.08, - "learning_rate": 1.9856671031189765e-05, - "loss": 1.0249, + "epoch": 0.16467872258560984, + "grad_norm": 1.1503238625185563, + "learning_rate": 1.9063809198333832e-05, + "loss": 0.6774, "step": 428 }, { - "epoch": 0.08, - "learning_rate": 1.98556177524672e-05, - "loss": 0.9933, + "epoch": 0.16506348595613698, + "grad_norm": 0.99464668217666, + "learning_rate": 1.905853757800791e-05, + "loss": 0.5543, "step": 429 }, { - "epoch": 0.08, - "learning_rate": 1.9854560645936262e-05, - "loss": 1.0367, + "epoch": 0.1654482493266641, + "grad_norm": 0.9052567355293533, + "learning_rate": 1.905325189033862e-05, + "loss": 0.5755, "step": 430 }, { - "epoch": 0.08, - "learning_rate": 1.9853499712007523e-05, - "loss": 1.0281, + "epoch": 0.16583301269719122, + "grad_norm": 1.0626308722211069, + "learning_rate": 1.904795214353431e-05, + "loss": 0.5999, "step": 431 }, { - "epoch": 0.08, - "learning_rate": 1.9852434951093035e-05, - "loss": 0.9441, + "epoch": 0.16621777606771834, + "grad_norm": 1.1022063000951596, + "learning_rate": 1.9042638345825155e-05, + "loss": 0.6469, "step": 432 }, { - "epoch": 0.08, - "learning_rate": 1.985136636360635e-05, - "loss": 0.9723, + "epoch": 0.1666025394382455, + "grad_norm": 1.1280561500577184, + "learning_rate": 1.9037310505463153e-05, + "loss": 0.6256, "step": 433 }, { - "epoch": 0.08, - "learning_rate": 1.985029394996248e-05, - "loss": 0.7537, + "epoch": 0.1669873028087726, + "grad_norm": 0.9865146408244617, + "learning_rate": 1.9031968630722104e-05, + "loss": 0.6267, "step": 434 }, { - "epoch": 0.08, - "learning_rate": 1.9849217710577945e-05, - "loss": 0.7774, + "epoch": 0.16737206617929973, + "grad_norm": 0.994956170789267, + "learning_rate": 1.902661272989761e-05, + "loss": 0.6186, "step": 435 }, { - "epoch": 0.08, - "learning_rate": 1.9848137645870745e-05, - "loss": 0.957, + "epoch": 0.16775682954982685, + "grad_norm": 1.0749169914901813, + "learning_rate": 1.9021242811307044e-05, + "loss": 0.6494, "step": 436 }, { - "epoch": 0.08, - "learning_rate": 1.9847053756260363e-05, - "loss": 0.9561, + "epoch": 0.168141592920354, + "grad_norm": 0.9933895739150949, + "learning_rate": 1.9015858883289556e-05, + "loss": 0.5865, "step": 437 }, { - "epoch": 0.08, - "learning_rate": 1.984596604216777e-05, - "loss": 0.9859, + "epoch": 0.1685263562908811, + "grad_norm": 0.9900273602311017, + "learning_rate": 1.901046095420606e-05, + "loss": 0.5911, "step": 438 }, { - "epoch": 0.08, - "learning_rate": 1.984487450401542e-05, - "loss": 0.993, + "epoch": 0.16891111966140823, + "grad_norm": 1.1654973641007516, + "learning_rate": 1.9005049032439193e-05, + "loss": 0.6135, "step": 439 }, { - "epoch": 0.08, - "learning_rate": 1.9843779142227258e-05, - "loss": 0.9013, + "epoch": 0.16929588303193535, + "grad_norm": 1.3052605186023023, + "learning_rate": 1.899962312639333e-05, + "loss": 0.6683, "step": 440 }, { - "epoch": 0.08, - "learning_rate": 1.9842679957228706e-05, - "loss": 1.0572, + "epoch": 0.1696806464024625, + "grad_norm": 0.9782705397268895, + "learning_rate": 1.899418324449457e-05, + "loss": 0.5567, "step": 441 }, { - "epoch": 0.09, - "learning_rate": 1.9841576949446675e-05, - "loss": 0.9891, + "epoch": 0.17006540977298962, + "grad_norm": 1.0975238676745338, + "learning_rate": 1.8988729395190712e-05, + "loss": 0.6023, "step": 442 }, { - "epoch": 0.09, - "learning_rate": 1.984047011930956e-05, - "loss": 0.9979, + "epoch": 0.17045017314351674, + "grad_norm": 1.0751012418538337, + "learning_rate": 1.898326158695124e-05, + "loss": 0.6942, "step": 443 }, { - "epoch": 0.09, - "learning_rate": 1.9839359467247243e-05, - "loss": 0.9757, + "epoch": 0.17083493651404386, + "grad_norm": 1.0330437695470474, + "learning_rate": 1.8977779828267314e-05, + "loss": 0.6645, "step": 444 }, { - "epoch": 0.09, - "learning_rate": 1.983824499369109e-05, - "loss": 1.0012, + "epoch": 0.17121969988457098, + "grad_norm": 1.0330939654534583, + "learning_rate": 1.897228412765177e-05, + "loss": 0.6501, "step": 445 }, { - "epoch": 0.09, - "learning_rate": 1.9837126699073948e-05, - "loss": 0.9642, + "epoch": 0.17160446325509812, + "grad_norm": 1.0805736526126277, + "learning_rate": 1.8966774493639084e-05, + "loss": 0.6087, "step": 446 }, { - "epoch": 0.09, - "learning_rate": 1.9836004583830146e-05, - "loss": 1.0114, + "epoch": 0.17198922662562524, + "grad_norm": 0.995646685849914, + "learning_rate": 1.896125093478538e-05, + "loss": 0.626, "step": 447 }, { - "epoch": 0.09, - "learning_rate": 1.9834878648395507e-05, - "loss": 1.097, + "epoch": 0.17237398999615236, + "grad_norm": 1.0081556611920368, + "learning_rate": 1.895571345966839e-05, + "loss": 0.6738, "step": 448 }, { - "epoch": 0.09, - "learning_rate": 1.9833748893207326e-05, - "loss": 1.0387, + "epoch": 0.17275875336667948, + "grad_norm": 1.055568668437573, + "learning_rate": 1.8950162076887477e-05, + "loss": 0.6369, "step": 449 }, { - "epoch": 0.09, - "learning_rate": 1.9832615318704388e-05, - "loss": 0.9194, + "epoch": 0.17314351673720663, + "grad_norm": 0.9369773502239576, + "learning_rate": 1.8944596795063584e-05, + "loss": 0.5856, "step": 450 }, { - "epoch": 0.09, - "learning_rate": 1.9831477925326962e-05, - "loss": 1.0781, + "epoch": 0.17352828010773375, + "grad_norm": 0.6991676890047379, + "learning_rate": 1.8939017622839253e-05, + "loss": 0.6686, "step": 451 }, { - "epoch": 0.09, - "learning_rate": 1.98303367135168e-05, - "loss": 1.0229, + "epoch": 0.17391304347826086, + "grad_norm": 1.017893827041653, + "learning_rate": 1.8933424568878586e-05, + "loss": 0.6029, "step": 452 }, { - "epoch": 0.09, - "learning_rate": 1.9829191683717133e-05, - "loss": 0.9309, + "epoch": 0.17429780684878798, + "grad_norm": 0.9870049361590104, + "learning_rate": 1.8927817641867244e-05, + "loss": 0.607, "step": 453 }, { - "epoch": 0.09, - "learning_rate": 1.9828042836372677e-05, - "loss": 0.9866, + "epoch": 0.17468257021931513, + "grad_norm": 0.9799225414884505, + "learning_rate": 1.8922196850512446e-05, + "loss": 0.5739, "step": 454 }, { - "epoch": 0.09, - "learning_rate": 1.9826890171929634e-05, - "loss": 0.9525, + "epoch": 0.17506733358984225, + "grad_norm": 1.0707456146633925, + "learning_rate": 1.8916562203542916e-05, + "loss": 0.6137, "step": 455 }, { - "epoch": 0.09, - "learning_rate": 1.982573369083568e-05, - "loss": 0.9699, + "epoch": 0.17545209696036937, + "grad_norm": 1.0933895701555099, + "learning_rate": 1.8910913709708918e-05, + "loss": 0.6524, "step": 456 }, { - "epoch": 0.09, - "learning_rate": 1.9824573393539984e-05, - "loss": 1.032, + "epoch": 0.1758368603308965, + "grad_norm": 0.9147655242703298, + "learning_rate": 1.8905251377782206e-05, + "loss": 0.6231, "step": 457 }, { - "epoch": 0.09, - "learning_rate": 1.982340928049319e-05, - "loss": 1.0363, + "epoch": 0.17622162370142364, + "grad_norm": 0.9809265642230655, + "learning_rate": 1.889957521655603e-05, + "loss": 0.5922, "step": 458 }, { - "epoch": 0.09, - "learning_rate": 1.9822241352147426e-05, - "loss": 0.9692, + "epoch": 0.17660638707195075, + "grad_norm": 1.0064367409843633, + "learning_rate": 1.8893885234845117e-05, + "loss": 0.6379, "step": 459 }, { - "epoch": 0.09, - "learning_rate": 1.9821069608956307e-05, - "loss": 0.9932, + "epoch": 0.17699115044247787, + "grad_norm": 0.9452363217731271, + "learning_rate": 1.888818144148565e-05, + "loss": 0.5808, "step": 460 }, { - "epoch": 0.09, - "learning_rate": 1.9819894051374917e-05, - "loss": 0.9539, + "epoch": 0.177375913813005, + "grad_norm": 1.0519947057599823, + "learning_rate": 1.8882463845335263e-05, + "loss": 0.6025, "step": 461 }, { - "epoch": 0.09, - "learning_rate": 1.981871467985983e-05, - "loss": 0.9779, + "epoch": 0.17776067718353214, + "grad_norm": 0.9654471176760567, + "learning_rate": 1.8876732455273022e-05, + "loss": 0.6025, "step": 462 }, { - "epoch": 0.09, - "learning_rate": 1.9817531494869105e-05, - "loss": 1.0128, + "epoch": 0.17814544055405926, + "grad_norm": 1.0516219494086119, + "learning_rate": 1.8870987280199428e-05, + "loss": 0.5859, "step": 463 }, { - "epoch": 0.09, - "learning_rate": 1.9816344496862272e-05, - "loss": 0.9626, + "epoch": 0.17853020392458638, + "grad_norm": 1.0694510930196932, + "learning_rate": 1.8865228329036372e-05, + "loss": 0.6002, "step": 464 }, { - "epoch": 0.09, - "learning_rate": 1.9815153686300352e-05, - "loss": 1.0507, + "epoch": 0.1789149672951135, + "grad_norm": 1.0913803372868829, + "learning_rate": 1.885945561072715e-05, + "loss": 0.6244, "step": 465 }, { - "epoch": 0.09, - "learning_rate": 1.981395906364584e-05, - "loss": 1.0756, + "epoch": 0.17929973066564064, + "grad_norm": 1.0023317878303524, + "learning_rate": 1.885366913423643e-05, + "loss": 0.6036, "step": 466 }, { - "epoch": 0.09, - "learning_rate": 1.9812760629362714e-05, - "loss": 1.0178, + "epoch": 0.17968449403616776, + "grad_norm": 1.0547189326493607, + "learning_rate": 1.8847868908550252e-05, + "loss": 0.6947, "step": 467 }, { - "epoch": 0.09, - "learning_rate": 1.981155838391643e-05, - "loss": 0.9855, + "epoch": 0.18006925740669488, + "grad_norm": 0.9730190481300429, + "learning_rate": 1.8842054942676e-05, + "loss": 0.6331, "step": 468 }, { - "epoch": 0.09, - "learning_rate": 1.9810352327773935e-05, - "loss": 0.9096, + "epoch": 0.180454020777222, + "grad_norm": 1.0429609582054569, + "learning_rate": 1.88362272456424e-05, + "loss": 0.6568, "step": 469 }, { - "epoch": 0.09, - "learning_rate": 1.9809142461403635e-05, - "loss": 0.8963, + "epoch": 0.18083878414774912, + "grad_norm": 0.7212223470782403, + "learning_rate": 1.8830385826499507e-05, + "loss": 0.6824, "step": 470 }, { - "epoch": 0.09, - "learning_rate": 1.9807928785275433e-05, - "loss": 1.0264, + "epoch": 0.18122354751827627, + "grad_norm": 0.9629253943095065, + "learning_rate": 1.8824530694318675e-05, + "loss": 0.5971, "step": 471 }, { - "epoch": 0.09, - "learning_rate": 1.980671129986071e-05, - "loss": 0.9991, + "epoch": 0.1816083108888034, + "grad_norm": 1.087654681860244, + "learning_rate": 1.8818661858192562e-05, + "loss": 0.5986, "step": 472 }, { - "epoch": 0.09, - "learning_rate": 1.9805490005632323e-05, - "loss": 1.0236, + "epoch": 0.1819930742593305, + "grad_norm": 1.0749263207277613, + "learning_rate": 1.8812779327235106e-05, + "loss": 0.6523, "step": 473 }, { - "epoch": 0.09, - "learning_rate": 1.98042649030646e-05, - "loss": 1.0321, + "epoch": 0.18237783762985763, + "grad_norm": 1.0600576965622832, + "learning_rate": 1.88068831105815e-05, + "loss": 0.6042, "step": 474 }, { - "epoch": 0.09, - "learning_rate": 1.9803035992633366e-05, - "loss": 1.0761, + "epoch": 0.18276260100038477, + "grad_norm": 0.9850002908282322, + "learning_rate": 1.8800973217388215e-05, + "loss": 0.6282, "step": 475 }, { - "epoch": 0.09, - "learning_rate": 1.9801803274815915e-05, - "loss": 1.062, + "epoch": 0.1831473643709119, + "grad_norm": 0.955639663538806, + "learning_rate": 1.879504965683294e-05, + "loss": 0.5858, "step": 476 }, { - "epoch": 0.09, - "learning_rate": 1.9800566750091018e-05, - "loss": 0.9667, + "epoch": 0.183532127741439, + "grad_norm": 1.0483283408696717, + "learning_rate": 1.878911243811459e-05, + "loss": 0.6053, "step": 477 }, { - "epoch": 0.09, - "learning_rate": 1.9799326418938924e-05, - "loss": 0.9552, + "epoch": 0.18391689111196613, + "grad_norm": 1.1029883335495858, + "learning_rate": 1.8783161570453295e-05, + "loss": 0.6139, "step": 478 }, { - "epoch": 0.09, - "learning_rate": 1.979808228184137e-05, - "loss": 0.9991, + "epoch": 0.18430165448249328, + "grad_norm": 1.0616322454381932, + "learning_rate": 1.8777197063090394e-05, + "loss": 0.6269, "step": 479 }, { - "epoch": 0.09, - "learning_rate": 1.9796834339281557e-05, - "loss": 0.9703, + "epoch": 0.1846864178530204, + "grad_norm": 1.062279515838876, + "learning_rate": 1.877121892528838e-05, + "loss": 0.6441, "step": 480 }, { - "epoch": 0.09, - "learning_rate": 1.979558259174418e-05, - "loss": 0.984, + "epoch": 0.18507118122354752, + "grad_norm": 1.106164968221549, + "learning_rate": 1.8765227166330933e-05, + "loss": 0.6715, "step": 481 }, { - "epoch": 0.09, - "learning_rate": 1.9794327039715395e-05, - "loss": 0.8934, + "epoch": 0.18545594459407463, + "grad_norm": 1.0161398468262521, + "learning_rate": 1.875922179552288e-05, + "loss": 0.609, "step": 482 }, { - "epoch": 0.09, - "learning_rate": 1.979306768368285e-05, - "loss": 1.0007, + "epoch": 0.18584070796460178, + "grad_norm": 1.1386749873835416, + "learning_rate": 1.875320282219019e-05, + "loss": 0.6354, "step": 483 }, { - "epoch": 0.09, - "learning_rate": 1.9791804524135663e-05, - "loss": 1.0516, + "epoch": 0.1862254713351289, + "grad_norm": 1.0093386253866312, + "learning_rate": 1.874717025567995e-05, + "loss": 0.567, "step": 484 }, { - "epoch": 0.09, - "learning_rate": 1.979053756156443e-05, - "loss": 0.958, + "epoch": 0.18661023470565602, + "grad_norm": 1.1523815613573838, + "learning_rate": 1.8741124105360363e-05, + "loss": 0.66, "step": 485 }, { - "epoch": 0.09, - "learning_rate": 1.9789266796461222e-05, - "loss": 1.0127, + "epoch": 0.18699499807618314, + "grad_norm": 1.1790399424608518, + "learning_rate": 1.8735064380620717e-05, + "loss": 0.6778, "step": 486 }, { - "epoch": 0.09, - "learning_rate": 1.9787992229319594e-05, - "loss": 0.9282, + "epoch": 0.18737976144671029, + "grad_norm": 1.0352928699490684, + "learning_rate": 1.8728991090871387e-05, + "loss": 0.609, "step": 487 }, { - "epoch": 0.09, - "learning_rate": 1.978671386063457e-05, - "loss": 1.0566, + "epoch": 0.1877645248172374, + "grad_norm": 0.9917361210278768, + "learning_rate": 1.8722904245543817e-05, + "loss": 0.614, "step": 488 }, { - "epoch": 0.09, - "learning_rate": 1.978543169090265e-05, - "loss": 0.8264, + "epoch": 0.18814928818776452, + "grad_norm": 1.0433371304678511, + "learning_rate": 1.8716803854090495e-05, + "loss": 0.6016, "step": 489 }, { - "epoch": 0.09, - "learning_rate": 1.9784145720621827e-05, - "loss": 1.0128, + "epoch": 0.18853405155829164, + "grad_norm": 1.026745742353778, + "learning_rate": 1.871068992598495e-05, + "loss": 0.5826, "step": 490 }, { - "epoch": 0.09, - "learning_rate": 1.9782855950291542e-05, - "loss": 1.0344, + "epoch": 0.1889188149288188, + "grad_norm": 1.059258906845605, + "learning_rate": 1.8704562470721728e-05, + "loss": 0.641, "step": 491 }, { - "epoch": 0.09, - "learning_rate": 1.978156238041274e-05, - "loss": 1.0519, + "epoch": 0.1893035782993459, + "grad_norm": 1.1323752323463336, + "learning_rate": 1.8698421497816386e-05, + "loss": 0.6366, "step": 492 }, { - "epoch": 0.09, - "learning_rate": 1.9780265011487822e-05, - "loss": 0.9947, + "epoch": 0.18968834166987303, + "grad_norm": 0.9883783523967515, + "learning_rate": 1.8692267016805473e-05, + "loss": 0.5627, "step": 493 }, { - "epoch": 0.1, - "learning_rate": 1.9778963844020668e-05, - "loss": 0.9776, + "epoch": 0.19007310504040015, + "grad_norm": 1.0377609034106168, + "learning_rate": 1.868609903724651e-05, + "loss": 0.6139, "step": 494 }, { - "epoch": 0.1, - "learning_rate": 1.977765887851664e-05, - "loss": 0.9721, + "epoch": 0.19045786841092727, + "grad_norm": 1.0742647840516806, + "learning_rate": 1.867991756871799e-05, + "loss": 0.6738, "step": 495 }, { - "epoch": 0.1, - "learning_rate": 1.977635011548257e-05, - "loss": 0.756, + "epoch": 0.1908426317814544, + "grad_norm": 1.0869407766437902, + "learning_rate": 1.867372262081934e-05, + "loss": 0.6291, "step": 496 }, { - "epoch": 0.1, - "learning_rate": 1.9775037555426772e-05, - "loss": 0.9836, + "epoch": 0.19122739515198153, + "grad_norm": 0.9412268358712942, + "learning_rate": 1.8667514203170934e-05, + "loss": 0.6018, "step": 497 }, { - "epoch": 0.1, - "learning_rate": 1.9773721198859024e-05, - "loss": 1.0171, + "epoch": 0.19161215852250865, + "grad_norm": 1.0574108086736165, + "learning_rate": 1.8661292325414058e-05, + "loss": 0.6305, "step": 498 }, { - "epoch": 0.1, - "learning_rate": 1.9772401046290584e-05, - "loss": 1.0262, + "epoch": 0.19199692189303577, + "grad_norm": 1.0580270545879111, + "learning_rate": 1.8655056997210893e-05, + "loss": 0.6426, "step": 499 }, { - "epoch": 0.1, - "learning_rate": 1.9771077098234187e-05, - "loss": 1.0734, + "epoch": 0.19238168526356292, + "grad_norm": 0.7839479997199029, + "learning_rate": 1.864880822824452e-05, + "loss": 0.6764, "step": 500 }, { - "epoch": 0.1, - "learning_rate": 1.9769749355204034e-05, - "loss": 1.0859, + "epoch": 0.19276644863409004, + "grad_norm": 1.0443868862685648, + "learning_rate": 1.864254602821888e-05, + "loss": 0.6038, "step": 501 }, { - "epoch": 0.1, - "learning_rate": 1.976841781771581e-05, - "loss": 0.9876, + "epoch": 0.19315121200461716, + "grad_norm": 1.0608048410151527, + "learning_rate": 1.8636270406858786e-05, + "loss": 0.6543, "step": 502 }, { - "epoch": 0.1, - "learning_rate": 1.9767082486286667e-05, - "loss": 0.9664, + "epoch": 0.19353597537514428, + "grad_norm": 0.6163798573142993, + "learning_rate": 1.862998137390989e-05, + "loss": 0.6259, "step": 503 }, { - "epoch": 0.1, - "learning_rate": 1.9765743361435234e-05, - "loss": 0.9353, + "epoch": 0.19392073874567142, + "grad_norm": 0.6419425744254407, + "learning_rate": 1.8623678939138652e-05, + "loss": 0.6566, "step": 504 }, { - "epoch": 0.1, - "learning_rate": 1.9764400443681607e-05, - "loss": 1.0006, + "epoch": 0.19430550211619854, + "grad_norm": 1.069260532306563, + "learning_rate": 1.8617363112332376e-05, + "loss": 0.5993, "step": 505 }, { - "epoch": 0.1, - "learning_rate": 1.9763053733547367e-05, - "loss": 1.0973, + "epoch": 0.19469026548672566, + "grad_norm": 1.1867872578933678, + "learning_rate": 1.8611033903299136e-05, + "loss": 0.6418, "step": 506 }, { - "epoch": 0.1, - "learning_rate": 1.976170323155555e-05, - "loss": 1.0038, + "epoch": 0.19507502885725278, + "grad_norm": 0.9968701576604385, + "learning_rate": 1.8604691321867804e-05, + "loss": 0.611, "step": 507 }, { - "epoch": 0.1, - "learning_rate": 1.976034893823069e-05, - "loss": 0.9556, + "epoch": 0.19545979222777993, + "grad_norm": 1.056451469606898, + "learning_rate": 1.8598335377888012e-05, + "loss": 0.6219, "step": 508 }, { - "epoch": 0.1, - "learning_rate": 1.975899085409876e-05, - "loss": 0.9908, + "epoch": 0.19584455559830705, + "grad_norm": 1.0352273446408082, + "learning_rate": 1.8591966081230142e-05, + "loss": 0.6433, "step": 509 }, { - "epoch": 0.1, - "learning_rate": 1.9757628979687247e-05, - "loss": 1.0483, + "epoch": 0.19622931896883417, + "grad_norm": 1.010181976170659, + "learning_rate": 1.858558344178532e-05, + "loss": 0.5596, "step": 510 }, { - "epoch": 0.1, - "learning_rate": 1.975626331552507e-05, - "loss": 0.9918, + "epoch": 0.19661408233936128, + "grad_norm": 0.971933596962428, + "learning_rate": 1.857918746946538e-05, + "loss": 0.6411, "step": 511 }, { - "epoch": 0.1, - "learning_rate": 1.9754893862142643e-05, - "loss": 0.7518, + "epoch": 0.19699884570988843, + "grad_norm": 1.0351767457608942, + "learning_rate": 1.857277817420287e-05, + "loss": 0.636, "step": 512 }, { - "epoch": 0.1, - "learning_rate": 1.9753520620071846e-05, - "loss": 1.042, + "epoch": 0.19738360908041555, + "grad_norm": 0.71574470318074, + "learning_rate": 1.8566355565951023e-05, + "loss": 0.67, "step": 513 }, { - "epoch": 0.1, - "learning_rate": 1.9752143589846027e-05, - "loss": 1.0537, + "epoch": 0.19776837245094267, + "grad_norm": 0.9933529061468143, + "learning_rate": 1.8559919654683756e-05, + "loss": 0.6073, "step": 514 }, { - "epoch": 0.1, - "learning_rate": 1.9750762772000014e-05, - "loss": 1.0113, + "epoch": 0.1981531358214698, + "grad_norm": 1.001783607153124, + "learning_rate": 1.855347045039563e-05, + "loss": 0.6248, "step": 515 }, { - "epoch": 0.1, - "learning_rate": 1.9749378167070097e-05, - "loss": 0.8888, + "epoch": 0.1985378991919969, + "grad_norm": 1.0023587366397435, + "learning_rate": 1.854700796310186e-05, + "loss": 0.6505, "step": 516 }, { - "epoch": 0.1, - "learning_rate": 1.9747989775594044e-05, - "loss": 0.9708, + "epoch": 0.19892266256252406, + "grad_norm": 0.6511860284233735, + "learning_rate": 1.8540532202838286e-05, + "loss": 0.6671, "step": 517 }, { - "epoch": 0.1, - "learning_rate": 1.974659759811109e-05, - "loss": 0.9299, + "epoch": 0.19930742593305117, + "grad_norm": 1.1500393434170648, + "learning_rate": 1.8534043179661357e-05, + "loss": 0.6261, "step": 518 }, { - "epoch": 0.1, - "learning_rate": 1.9745201635161938e-05, - "loss": 1.0118, + "epoch": 0.1996921893035783, + "grad_norm": 1.0048026016374028, + "learning_rate": 1.8527540903648122e-05, + "loss": 0.6746, "step": 519 }, { - "epoch": 0.1, - "learning_rate": 1.9743801887288762e-05, - "loss": 1.0005, + "epoch": 0.2000769526741054, + "grad_norm": 0.9653853420481484, + "learning_rate": 1.852102538489621e-05, + "loss": 0.5539, "step": 520 }, { - "epoch": 0.1, - "learning_rate": 1.9742398355035212e-05, - "loss": 1.0608, + "epoch": 0.20046171604463256, + "grad_norm": 0.8295394268708312, + "learning_rate": 1.851449663352381e-05, + "loss": 0.5396, "step": 521 }, { - "epoch": 0.1, - "learning_rate": 1.9740991038946404e-05, - "loss": 0.7946, + "epoch": 0.20084647941515968, + "grad_norm": 0.9618086603727722, + "learning_rate": 1.8507954659669677e-05, + "loss": 0.5983, "step": 522 }, { - "epoch": 0.1, - "learning_rate": 1.973957993956892e-05, - "loss": 1.0243, + "epoch": 0.2012312427856868, + "grad_norm": 1.0134382065723386, + "learning_rate": 1.850139947349308e-05, + "loss": 0.5953, "step": 523 }, { - "epoch": 0.1, - "learning_rate": 1.9738165057450817e-05, - "loss": 0.9691, + "epoch": 0.20161600615621392, + "grad_norm": 0.6484310348125228, + "learning_rate": 1.849483108517381e-05, + "loss": 0.6519, "step": 524 }, { - "epoch": 0.1, - "learning_rate": 1.9736746393141617e-05, - "loss": 0.9942, + "epoch": 0.20200076952674106, + "grad_norm": 0.9806405365544044, + "learning_rate": 1.8488249504912173e-05, + "loss": 0.586, "step": 525 }, { - "epoch": 0.1, - "learning_rate": 1.9735323947192317e-05, - "loss": 1.0279, + "epoch": 0.20238553289726818, + "grad_norm": 1.0098487864756704, + "learning_rate": 1.848165474292895e-05, + "loss": 0.6324, "step": 526 }, { - "epoch": 0.1, - "learning_rate": 1.9733897720155377e-05, - "loss": 0.9273, + "epoch": 0.2027702962677953, + "grad_norm": 1.071432075737243, + "learning_rate": 1.847504680946539e-05, + "loss": 0.5716, "step": 527 }, { - "epoch": 0.1, - "learning_rate": 1.9732467712584723e-05, - "loss": 0.9341, + "epoch": 0.20315505963832242, + "grad_norm": 0.9232816935054939, + "learning_rate": 1.8468425714783206e-05, + "loss": 0.6116, "step": 528 }, { - "epoch": 0.1, - "learning_rate": 1.973103392503576e-05, - "loss": 1.0552, + "epoch": 0.20353982300884957, + "grad_norm": 0.9563055851022759, + "learning_rate": 1.846179146916454e-05, + "loss": 0.6117, "step": 529 }, { - "epoch": 0.1, - "learning_rate": 1.9729596358065347e-05, - "loss": 0.9809, + "epoch": 0.2039245863793767, + "grad_norm": 1.017515686294816, + "learning_rate": 1.8455144082911965e-05, + "loss": 0.6563, "step": 530 }, { - "epoch": 0.1, - "learning_rate": 1.9728155012231825e-05, - "loss": 0.9131, + "epoch": 0.2043093497499038, + "grad_norm": 0.9405916832410801, + "learning_rate": 1.8448483566348456e-05, + "loss": 0.6273, "step": 531 }, { - "epoch": 0.1, - "learning_rate": 1.9726709888094994e-05, - "loss": 0.98, + "epoch": 0.20469411312043093, + "grad_norm": 1.0811221583918373, + "learning_rate": 1.8441809929817382e-05, + "loss": 0.6069, "step": 532 }, { - "epoch": 0.1, - "learning_rate": 1.972526098621612e-05, - "loss": 0.9851, + "epoch": 0.20507887649095807, + "grad_norm": 0.6426412359056394, + "learning_rate": 1.8435123183682475e-05, + "loss": 0.6384, "step": 533 }, { - "epoch": 0.1, - "learning_rate": 1.972380830715795e-05, - "loss": 1.0235, + "epoch": 0.2054636398614852, + "grad_norm": 0.9485574670828467, + "learning_rate": 1.8428423338327847e-05, + "loss": 0.6296, "step": 534 }, { - "epoch": 0.1, - "learning_rate": 1.9722351851484677e-05, - "loss": 0.9953, + "epoch": 0.2058484032320123, + "grad_norm": 1.0362438382690966, + "learning_rate": 1.842171040415793e-05, + "loss": 0.6888, "step": 535 }, { - "epoch": 0.1, - "learning_rate": 1.9720891619761974e-05, - "loss": 0.9845, + "epoch": 0.20623316660253943, + "grad_norm": 0.6285734701935441, + "learning_rate": 1.8414984391597492e-05, + "loss": 0.6374, "step": 536 }, { - "epoch": 0.1, - "learning_rate": 1.9719427612556982e-05, - "loss": 1.0769, + "epoch": 0.20661792997306658, + "grad_norm": 1.118925711624862, + "learning_rate": 1.8408245311091618e-05, + "loss": 0.6679, "step": 537 }, { - "epoch": 0.1, - "learning_rate": 1.9717959830438302e-05, - "loss": 0.9143, + "epoch": 0.2070026933435937, + "grad_norm": 1.0295679704099956, + "learning_rate": 1.8401493173105675e-05, + "loss": 0.5698, "step": 538 }, { - "epoch": 0.1, - "learning_rate": 1.9716488273976006e-05, - "loss": 0.9272, + "epoch": 0.20738745671412082, + "grad_norm": 1.0332323195303605, + "learning_rate": 1.8394727988125308e-05, + "loss": 0.5783, "step": 539 }, { - "epoch": 0.1, - "learning_rate": 1.971501294374162e-05, - "loss": 1.0186, + "epoch": 0.20777222008464794, + "grad_norm": 1.0015927147870651, + "learning_rate": 1.8387949766656434e-05, + "loss": 0.6554, "step": 540 }, { - "epoch": 0.1, - "learning_rate": 1.971353384030816e-05, - "loss": 0.9785, + "epoch": 0.20815698345517505, + "grad_norm": 1.0699316029117794, + "learning_rate": 1.8381158519225204e-05, + "loss": 0.6588, "step": 541 }, { - "epoch": 0.1, - "learning_rate": 1.9712050964250083e-05, - "loss": 0.9762, + "epoch": 0.2085417468257022, + "grad_norm": 0.937570391413873, + "learning_rate": 1.8374354256378e-05, + "loss": 0.5666, "step": 542 }, { - "epoch": 0.1, - "learning_rate": 1.9710564316143323e-05, - "loss": 0.9728, + "epoch": 0.20892651019622932, + "grad_norm": 1.0458707063129495, + "learning_rate": 1.8367536988681422e-05, + "loss": 0.6338, "step": 543 }, { - "epoch": 0.1, - "learning_rate": 1.9709073896565276e-05, - "loss": 0.9883, + "epoch": 0.20931127356675644, + "grad_norm": 1.0215066326427427, + "learning_rate": 1.8360706726722253e-05, + "loss": 0.5817, "step": 544 }, { - "epoch": 0.1, - "learning_rate": 1.9707579706094807e-05, - "loss": 0.991, + "epoch": 0.20969603693728356, + "grad_norm": 0.9153132618843477, + "learning_rate": 1.8353863481107473e-05, + "loss": 0.5767, "step": 545 }, { - "epoch": 0.11, - "learning_rate": 1.970608174531224e-05, - "loss": 0.9799, + "epoch": 0.2100808003078107, + "grad_norm": 1.1076438883391282, + "learning_rate": 1.8347007262464206e-05, + "loss": 0.6119, "step": 546 }, { - "epoch": 0.11, - "learning_rate": 1.970458001479937e-05, - "loss": 1.0057, + "epoch": 0.21046556367833782, + "grad_norm": 1.1793667914358656, + "learning_rate": 1.8340138081439743e-05, + "loss": 0.6455, "step": 547 }, { - "epoch": 0.11, - "learning_rate": 1.9703074515139445e-05, - "loss": 1.0208, + "epoch": 0.21085032704886494, + "grad_norm": 0.8642248352958052, + "learning_rate": 1.833325594870148e-05, + "loss": 0.5468, "step": 548 }, { - "epoch": 0.11, - "learning_rate": 1.9701565246917184e-05, - "loss": 1.0497, + "epoch": 0.21123509041939206, + "grad_norm": 0.988777491711932, + "learning_rate": 1.8326360874936952e-05, + "loss": 0.5951, "step": 549 }, { - "epoch": 0.11, - "learning_rate": 1.9700052210718775e-05, - "loss": 1.0442, + "epoch": 0.2116198537899192, + "grad_norm": 1.0436498553317775, + "learning_rate": 1.8319452870853772e-05, + "loss": 0.6308, "step": 550 }, { - "epoch": 0.11, - "learning_rate": 1.969853540713186e-05, - "loss": 1.0361, + "epoch": 0.21200461716044633, + "grad_norm": 1.0632529105206434, + "learning_rate": 1.8312531947179634e-05, + "loss": 0.6157, "step": 551 }, { - "epoch": 0.11, - "learning_rate": 1.9697014836745552e-05, - "loss": 1.0145, + "epoch": 0.21238938053097345, + "grad_norm": 1.1744690346535716, + "learning_rate": 1.8305598114662312e-05, + "loss": 0.6322, "step": 552 }, { - "epoch": 0.11, - "learning_rate": 1.9695490500150418e-05, - "loss": 0.9919, + "epoch": 0.21277414390150057, + "grad_norm": 1.0277831896697394, + "learning_rate": 1.8298651384069605e-05, + "loss": 0.5643, "step": 553 }, { - "epoch": 0.11, - "learning_rate": 1.9693962397938495e-05, - "loss": 1.0268, + "epoch": 0.21315890727202771, + "grad_norm": 0.9837784616661985, + "learning_rate": 1.829169176618936e-05, + "loss": 0.578, "step": 554 }, { - "epoch": 0.11, - "learning_rate": 1.9692430530703282e-05, - "loss": 1.0629, + "epoch": 0.21354367064255483, + "grad_norm": 1.0366694878973723, + "learning_rate": 1.828471927182942e-05, + "loss": 0.6858, "step": 555 }, { - "epoch": 0.11, - "learning_rate": 1.9690894899039735e-05, - "loss": 1.0263, + "epoch": 0.21392843401308195, + "grad_norm": 1.2114253065554537, + "learning_rate": 1.8277733911817642e-05, + "loss": 0.6289, "step": 556 }, { - "epoch": 0.11, - "learning_rate": 1.9689355503544277e-05, - "loss": 0.9104, + "epoch": 0.21431319738360907, + "grad_norm": 1.034713731101644, + "learning_rate": 1.827073569700185e-05, + "loss": 0.5961, "step": 557 }, { - "epoch": 0.11, - "learning_rate": 1.968781234481479e-05, - "loss": 0.9559, + "epoch": 0.21469796075413622, + "grad_norm": 1.013273902016613, + "learning_rate": 1.8263724638249834e-05, + "loss": 0.6286, "step": 558 }, { - "epoch": 0.11, - "learning_rate": 1.9686265423450624e-05, - "loss": 0.9773, + "epoch": 0.21508272412466334, + "grad_norm": 1.0384325045862153, + "learning_rate": 1.825670074644933e-05, + "loss": 0.6164, "step": 559 }, { - "epoch": 0.11, - "learning_rate": 1.9684714740052584e-05, - "loss": 1.0234, + "epoch": 0.21546748749519046, + "grad_norm": 1.0049806645676773, + "learning_rate": 1.824966403250801e-05, + "loss": 0.5837, "step": 560 }, { - "epoch": 0.11, - "learning_rate": 1.9683160295222934e-05, - "loss": 0.9442, + "epoch": 0.21585225086571758, + "grad_norm": 0.9998404715183762, + "learning_rate": 1.8242614507353446e-05, + "loss": 0.5509, "step": 561 }, { - "epoch": 0.11, - "learning_rate": 1.9681602089565403e-05, - "loss": 0.9843, + "epoch": 0.2162370142362447, + "grad_norm": 1.1244844717629057, + "learning_rate": 1.823555218193311e-05, + "loss": 0.5987, "step": 562 }, { - "epoch": 0.11, - "learning_rate": 1.968004012368518e-05, - "loss": 0.9632, + "epoch": 0.21662177760677184, + "grad_norm": 1.05920746914987, + "learning_rate": 1.8228477067214352e-05, + "loss": 0.6278, "step": 563 }, { - "epoch": 0.11, - "learning_rate": 1.967847439818892e-05, - "loss": 0.9345, + "epoch": 0.21700654097729896, + "grad_norm": 1.0081185187035533, + "learning_rate": 1.8221389174184385e-05, + "loss": 0.6017, "step": 564 }, { - "epoch": 0.11, - "learning_rate": 1.9676904913684725e-05, - "loss": 0.9755, + "epoch": 0.21739130434782608, + "grad_norm": 0.9585159177999452, + "learning_rate": 1.8214288513850267e-05, + "loss": 0.5931, "step": 565 }, { - "epoch": 0.11, - "learning_rate": 1.967533167078217e-05, - "loss": 1.0003, + "epoch": 0.2177760677183532, + "grad_norm": 1.0451022870060676, + "learning_rate": 1.820717509723888e-05, + "loss": 0.6606, "step": 566 }, { - "epoch": 0.11, - "learning_rate": 1.9673754670092283e-05, - "loss": 1.0369, + "epoch": 0.21816083108888035, + "grad_norm": 1.0141778629764746, + "learning_rate": 1.8200048935396908e-05, + "loss": 0.5717, "step": 567 }, { - "epoch": 0.11, - "learning_rate": 1.9672173912227556e-05, - "loss": 1.0293, + "epoch": 0.21854559445940747, + "grad_norm": 0.9123020540861806, + "learning_rate": 1.8192910039390844e-05, + "loss": 0.5751, "step": 568 }, { - "epoch": 0.11, - "learning_rate": 1.967058939780193e-05, - "loss": 1.0355, + "epoch": 0.21893035782993459, + "grad_norm": 1.054007746525327, + "learning_rate": 1.8185758420306947e-05, + "loss": 0.6634, "step": 569 }, { - "epoch": 0.11, - "learning_rate": 1.966900112743082e-05, - "loss": 0.9443, + "epoch": 0.2193151212004617, + "grad_norm": 1.0436663445998107, + "learning_rate": 1.817859408925123e-05, + "loss": 0.6636, "step": 570 }, { - "epoch": 0.11, - "learning_rate": 1.966740910173108e-05, - "loss": 0.888, + "epoch": 0.21969988457098885, + "grad_norm": 1.0220757664260887, + "learning_rate": 1.8171417057349457e-05, + "loss": 0.597, "step": 571 }, { - "epoch": 0.11, - "learning_rate": 1.9665813321321054e-05, - "loss": 0.9984, + "epoch": 0.22008464794151597, + "grad_norm": 0.9736369412981524, + "learning_rate": 1.8164227335747108e-05, + "loss": 0.6332, "step": 572 }, { - "epoch": 0.11, - "learning_rate": 1.9664213786820502e-05, - "loss": 1.0409, + "epoch": 0.2204694113120431, + "grad_norm": 0.9792847585759151, + "learning_rate": 1.815702493560937e-05, + "loss": 0.5842, "step": 573 }, { - "epoch": 0.11, - "learning_rate": 1.9662610498850684e-05, - "loss": 0.9217, + "epoch": 0.2208541746825702, + "grad_norm": 0.9878329041815623, + "learning_rate": 1.8149809868121125e-05, + "loss": 0.5698, "step": 574 }, { - "epoch": 0.11, - "learning_rate": 1.9661003458034288e-05, - "loss": 1.0087, + "epoch": 0.22123893805309736, + "grad_norm": 1.0968719799672235, + "learning_rate": 1.814258214448692e-05, + "loss": 0.5678, "step": 575 }, { - "epoch": 0.11, - "learning_rate": 1.965939266499547e-05, - "loss": 0.7735, + "epoch": 0.22162370142362448, + "grad_norm": 1.027190206876489, + "learning_rate": 1.813534177593096e-05, + "loss": 0.621, "step": 576 }, { - "epoch": 0.11, - "learning_rate": 1.9657778120359848e-05, - "loss": 0.7236, + "epoch": 0.2220084647941516, + "grad_norm": 0.9706407677244631, + "learning_rate": 1.8128088773697086e-05, + "loss": 0.6101, "step": 577 }, { - "epoch": 0.11, - "learning_rate": 1.965615982475449e-05, - "loss": 0.9948, + "epoch": 0.2223932281646787, + "grad_norm": 1.0746801314865164, + "learning_rate": 1.8120823149048753e-05, + "loss": 0.6661, "step": 578 }, { - "epoch": 0.11, - "learning_rate": 1.9654537778807924e-05, - "loss": 0.9904, + "epoch": 0.22277799153520586, + "grad_norm": 1.0432359037996817, + "learning_rate": 1.8113544913269025e-05, + "loss": 0.6162, "step": 579 }, { - "epoch": 0.11, - "learning_rate": 1.9652911983150135e-05, - "loss": 1.0234, + "epoch": 0.22316275490573298, + "grad_norm": 1.0106449051558337, + "learning_rate": 1.8106254077660552e-05, + "loss": 0.6167, "step": 580 }, { - "epoch": 0.11, - "learning_rate": 1.965128243841256e-05, - "loss": 0.9371, + "epoch": 0.2235475182762601, + "grad_norm": 1.0146183381793366, + "learning_rate": 1.809895065354554e-05, + "loss": 0.6299, "step": 581 }, { - "epoch": 0.11, - "learning_rate": 1.96496491452281e-05, - "loss": 0.9861, + "epoch": 0.22393228164678722, + "grad_norm": 1.0355655736926492, + "learning_rate": 1.8091634652265755e-05, + "loss": 0.6235, "step": 582 }, { - "epoch": 0.11, - "learning_rate": 1.9648012104231106e-05, - "loss": 0.9893, + "epoch": 0.22431704501731436, + "grad_norm": 1.0715750781370963, + "learning_rate": 1.808430608518249e-05, + "loss": 0.6135, "step": 583 }, { - "epoch": 0.11, - "learning_rate": 1.964637131605738e-05, - "loss": 1.0069, + "epoch": 0.22470180838784148, + "grad_norm": 1.0104991632982039, + "learning_rate": 1.807696496367655e-05, + "loss": 0.5868, "step": 584 }, { - "epoch": 0.11, - "learning_rate": 1.9644726781344197e-05, - "loss": 0.9888, + "epoch": 0.2250865717583686, + "grad_norm": 0.9803933816103925, + "learning_rate": 1.8069611299148236e-05, + "loss": 0.5666, "step": 585 }, { - "epoch": 0.11, - "learning_rate": 1.964307850073026e-05, - "loss": 1.0048, + "epoch": 0.22547133512889572, + "grad_norm": 1.0108257083381489, + "learning_rate": 1.806224510301734e-05, + "loss": 0.6312, "step": 586 }, { - "epoch": 0.11, - "learning_rate": 1.964142647485576e-05, - "loss": 1.0096, + "epoch": 0.22585609849942284, + "grad_norm": 0.9580471338937401, + "learning_rate": 1.8054866386723096e-05, + "loss": 0.5905, "step": 587 }, { - "epoch": 0.11, - "learning_rate": 1.9639770704362305e-05, - "loss": 0.9748, + "epoch": 0.22624086186995, + "grad_norm": 0.7016008157872888, + "learning_rate": 1.804747516172419e-05, + "loss": 0.695, "step": 588 }, { - "epoch": 0.11, - "learning_rate": 1.9638111189892994e-05, - "loss": 0.9388, + "epoch": 0.2266256252404771, + "grad_norm": 1.9657526590875674, + "learning_rate": 1.804007143949874e-05, + "loss": 0.655, "step": 589 }, { - "epoch": 0.11, - "learning_rate": 1.9636447932092354e-05, - "loss": 1.0272, + "epoch": 0.22701038861100423, + "grad_norm": 1.1663105740245914, + "learning_rate": 1.8032655231544253e-05, + "loss": 0.6263, "step": 590 }, { - "epoch": 0.11, - "learning_rate": 1.963478093160638e-05, - "loss": 0.7783, + "epoch": 0.22739515198153135, + "grad_norm": 0.907944908277295, + "learning_rate": 1.8025226549377647e-05, + "loss": 0.5997, "step": 591 }, { - "epoch": 0.11, - "learning_rate": 1.9633110189082515e-05, - "loss": 0.9645, + "epoch": 0.2277799153520585, + "grad_norm": 1.0497058070506826, + "learning_rate": 1.8017785404535198e-05, + "loss": 0.592, "step": 592 }, { - "epoch": 0.11, - "learning_rate": 1.963143570516965e-05, - "loss": 0.9626, + "epoch": 0.2281646787225856, + "grad_norm": 0.9877880734644778, + "learning_rate": 1.801033180857254e-05, + "loss": 0.587, "step": 593 }, { - "epoch": 0.11, - "learning_rate": 1.9629757480518144e-05, - "loss": 0.932, + "epoch": 0.22854944209311273, + "grad_norm": 0.9784329494389811, + "learning_rate": 1.8002865773064644e-05, + "loss": 0.6188, "step": 594 }, { - "epoch": 0.11, - "learning_rate": 1.9628075515779796e-05, - "loss": 0.9755, + "epoch": 0.22893420546363985, + "grad_norm": 0.9687192008721125, + "learning_rate": 1.799538730960579e-05, + "loss": 0.616, "step": 595 }, { - "epoch": 0.11, - "learning_rate": 1.962638981160786e-05, - "loss": 0.9696, + "epoch": 0.229318968834167, + "grad_norm": 0.90615967994333, + "learning_rate": 1.7987896429809573e-05, + "loss": 0.611, "step": 596 }, { - "epoch": 0.11, - "learning_rate": 1.9624700368657045e-05, - "loss": 0.943, + "epoch": 0.22970373220469412, + "grad_norm": 1.028726827561951, + "learning_rate": 1.7980393145308857e-05, + "loss": 0.6007, "step": 597 }, { - "epoch": 0.12, - "learning_rate": 1.9623007187583518e-05, - "loss": 0.9581, + "epoch": 0.23008849557522124, + "grad_norm": 0.9931263523610985, + "learning_rate": 1.7972877467755777e-05, + "loss": 0.5937, "step": 598 }, { - "epoch": 0.12, - "learning_rate": 1.962131026904488e-05, - "loss": 0.9851, + "epoch": 0.23047325894574835, + "grad_norm": 0.663827879773308, + "learning_rate": 1.796534940882171e-05, + "loss": 0.6918, "step": 599 }, { - "epoch": 0.12, - "learning_rate": 1.96196096137002e-05, - "loss": 1.0252, + "epoch": 0.2308580223162755, + "grad_norm": 0.963425901115495, + "learning_rate": 1.795780898019726e-05, + "loss": 0.5795, "step": 600 }, { - "epoch": 0.12, - "learning_rate": 1.9617905222209998e-05, - "loss": 0.8871, + "epoch": 0.23124278568680262, + "grad_norm": 1.0346424870409334, + "learning_rate": 1.7950256193592243e-05, + "loss": 0.6238, "step": 601 }, { - "epoch": 0.12, - "learning_rate": 1.961619709523623e-05, - "loss": 0.9915, + "epoch": 0.23162754905732974, + "grad_norm": 0.963524451013378, + "learning_rate": 1.7942691060735666e-05, + "loss": 0.5854, "step": 602 }, { - "epoch": 0.12, - "learning_rate": 1.9614485233442316e-05, - "loss": 0.9961, + "epoch": 0.23201231242785686, + "grad_norm": 0.8929772405753532, + "learning_rate": 1.7935113593375707e-05, + "loss": 0.5716, "step": 603 }, { - "epoch": 0.12, - "learning_rate": 1.961276963749313e-05, - "loss": 1.003, + "epoch": 0.232397075798384, + "grad_norm": 0.9859863114346746, + "learning_rate": 1.79275238032797e-05, + "loss": 0.5938, "step": 604 }, { - "epoch": 0.12, - "learning_rate": 1.9611050308054982e-05, - "loss": 0.9846, + "epoch": 0.23278183916891113, + "grad_norm": 0.9981988119584122, + "learning_rate": 1.791992170223412e-05, + "loss": 0.6254, "step": 605 }, { - "epoch": 0.12, - "learning_rate": 1.9609327245795642e-05, - "loss": 0.9946, + "epoch": 0.23316660253943824, + "grad_norm": 1.0094355853887458, + "learning_rate": 1.791230730204455e-05, + "loss": 0.6143, "step": 606 }, { - "epoch": 0.12, - "learning_rate": 1.9607600451384327e-05, - "loss": 0.9763, + "epoch": 0.23355136590996536, + "grad_norm": 1.0698915195689152, + "learning_rate": 1.7904680614535675e-05, + "loss": 0.5781, "step": 607 }, { - "epoch": 0.12, - "learning_rate": 1.960586992549171e-05, - "loss": 0.7894, + "epoch": 0.2339361292804925, + "grad_norm": 0.9595114632323791, + "learning_rate": 1.789704165155127e-05, + "loss": 0.5868, "step": 608 }, { - "epoch": 0.12, - "learning_rate": 1.9604135668789897e-05, - "loss": 0.7284, + "epoch": 0.23432089265101963, + "grad_norm": 0.9150164745583255, + "learning_rate": 1.7889390424954168e-05, + "loss": 0.5902, "step": 609 }, { - "epoch": 0.12, - "learning_rate": 1.9602397681952462e-05, - "loss": 0.9288, + "epoch": 0.23470565602154675, + "grad_norm": 1.016873533195387, + "learning_rate": 1.7881726946626244e-05, + "loss": 0.6121, "step": 610 }, { - "epoch": 0.12, - "learning_rate": 1.9600655965654413e-05, - "loss": 0.9497, + "epoch": 0.23509041939207387, + "grad_norm": 1.1165324098134812, + "learning_rate": 1.78740512284684e-05, + "loss": 0.6592, "step": 611 }, { - "epoch": 0.12, - "learning_rate": 1.959891052057222e-05, - "loss": 0.9131, + "epoch": 0.235475182762601, + "grad_norm": 1.0065453218763805, + "learning_rate": 1.7866363282400555e-05, + "loss": 0.6088, "step": 612 }, { - "epoch": 0.12, - "learning_rate": 1.9597161347383783e-05, - "loss": 1.027, + "epoch": 0.23585994613312813, + "grad_norm": 1.001273712715341, + "learning_rate": 1.7858663120361597e-05, + "loss": 0.6486, "step": 613 }, { - "epoch": 0.12, - "learning_rate": 1.959540844676847e-05, - "loss": 0.9272, + "epoch": 0.23624470950365525, + "grad_norm": 0.9931446880483966, + "learning_rate": 1.7850950754309405e-05, + "loss": 0.6093, "step": 614 }, { - "epoch": 0.12, - "learning_rate": 1.9593651819407084e-05, - "loss": 0.9818, + "epoch": 0.23662947287418237, + "grad_norm": 0.9946551749525449, + "learning_rate": 1.7843226196220803e-05, + "loss": 0.6561, "step": 615 }, { - "epoch": 0.12, - "learning_rate": 1.959189146598188e-05, - "loss": 0.961, + "epoch": 0.2370142362447095, + "grad_norm": 1.039188873385458, + "learning_rate": 1.7835489458091544e-05, + "loss": 0.6783, "step": 616 }, { - "epoch": 0.12, - "learning_rate": 1.9590127387176556e-05, - "loss": 0.957, + "epoch": 0.23739899961523664, + "grad_norm": 1.0344780954561983, + "learning_rate": 1.7827740551936296e-05, + "loss": 0.6631, "step": 617 }, { - "epoch": 0.12, - "learning_rate": 1.9588359583676263e-05, - "loss": 0.9818, + "epoch": 0.23778376298576376, + "grad_norm": 1.0307597611428116, + "learning_rate": 1.7819979489788638e-05, + "loss": 0.5606, "step": 618 }, { - "epoch": 0.12, - "learning_rate": 1.9586588056167595e-05, - "loss": 0.9674, + "epoch": 0.23816852635629088, + "grad_norm": 1.0463264981553488, + "learning_rate": 1.7812206283701002e-05, + "loss": 0.6176, "step": 619 }, { - "epoch": 0.12, - "learning_rate": 1.958481280533859e-05, - "loss": 0.9634, + "epoch": 0.238553289726818, + "grad_norm": 0.9809562807013069, + "learning_rate": 1.78044209457447e-05, + "loss": 0.6078, "step": 620 }, { - "epoch": 0.12, - "learning_rate": 1.958303383187874e-05, - "loss": 0.9396, + "epoch": 0.23893805309734514, + "grad_norm": 1.0052914860444955, + "learning_rate": 1.7796623488009875e-05, + "loss": 0.6248, "step": 621 }, { - "epoch": 0.12, - "learning_rate": 1.9581251136478974e-05, - "loss": 0.9709, + "epoch": 0.23932281646787226, + "grad_norm": 1.0485521739892443, + "learning_rate": 1.7788813922605488e-05, + "loss": 0.6137, "step": 622 }, { - "epoch": 0.12, - "learning_rate": 1.9579464719831668e-05, - "loss": 1.032, + "epoch": 0.23970757983839938, + "grad_norm": 0.6965015653155618, + "learning_rate": 1.7780992261659305e-05, + "loss": 0.7128, "step": 623 }, { - "epoch": 0.12, - "learning_rate": 1.9577674582630653e-05, - "loss": 1.0169, + "epoch": 0.2400923432089265, + "grad_norm": 1.0067674632490264, + "learning_rate": 1.777315851731789e-05, + "loss": 0.627, "step": 624 }, { - "epoch": 0.12, - "learning_rate": 1.957588072557119e-05, - "loss": 0.9632, + "epoch": 0.24047710657945365, + "grad_norm": 1.0152946774663856, + "learning_rate": 1.7765312701746543e-05, + "loss": 0.5559, "step": 625 }, { - "epoch": 0.12, - "learning_rate": 1.957408314935e-05, - "loss": 0.9049, + "epoch": 0.24086186994998077, + "grad_norm": 1.0044004865647014, + "learning_rate": 1.7757454827129338e-05, + "loss": 0.593, "step": 626 }, { - "epoch": 0.12, - "learning_rate": 1.9572281854665233e-05, - "loss": 1.0424, + "epoch": 0.24124663332050789, + "grad_norm": 1.0241798178619137, + "learning_rate": 1.7749584905669057e-05, + "loss": 0.6453, "step": 627 }, { - "epoch": 0.12, - "learning_rate": 1.95704768422165e-05, - "loss": 1.006, + "epoch": 0.241631396691035, + "grad_norm": 1.1103517856801473, + "learning_rate": 1.7741702949587196e-05, + "loss": 0.6609, "step": 628 }, { - "epoch": 0.12, - "learning_rate": 1.956866811270484e-05, - "loss": 0.9559, + "epoch": 0.24201616006156215, + "grad_norm": 1.0239999480459139, + "learning_rate": 1.7733808971123946e-05, + "loss": 0.6361, "step": 629 }, { - "epoch": 0.12, - "learning_rate": 1.9566855666832743e-05, - "loss": 1.0516, + "epoch": 0.24240092343208927, + "grad_norm": 0.6623235611546395, + "learning_rate": 1.7725902982538162e-05, + "loss": 0.6682, "step": 630 }, { - "epoch": 0.12, - "learning_rate": 1.9565039505304145e-05, - "loss": 0.987, + "epoch": 0.2427856868026164, + "grad_norm": 0.9901550305494047, + "learning_rate": 1.7717984996107346e-05, + "loss": 0.5702, "step": 631 }, { - "epoch": 0.12, - "learning_rate": 1.956321962882442e-05, - "loss": 1.0588, + "epoch": 0.2431704501731435, + "grad_norm": 1.0246005026553329, + "learning_rate": 1.7710055024127637e-05, + "loss": 0.6104, "step": 632 }, { - "epoch": 0.12, - "learning_rate": 1.956139603810039e-05, - "loss": 0.9573, + "epoch": 0.24355521354367063, + "grad_norm": 1.0885175124501065, + "learning_rate": 1.770211307891379e-05, + "loss": 0.5944, "step": 633 }, { - "epoch": 0.12, - "learning_rate": 1.9559568733840317e-05, - "loss": 0.9285, + "epoch": 0.24393997691419778, + "grad_norm": 1.116177833013039, + "learning_rate": 1.769415917279915e-05, + "loss": 0.6779, "step": 634 }, { - "epoch": 0.12, - "learning_rate": 1.9557737716753898e-05, - "loss": 0.9231, + "epoch": 0.2443247402847249, + "grad_norm": 1.0042831278070996, + "learning_rate": 1.7686193318135635e-05, + "loss": 0.5431, "step": 635 }, { - "epoch": 0.12, - "learning_rate": 1.9555902987552283e-05, - "loss": 0.9939, + "epoch": 0.24470950365525201, + "grad_norm": 1.083820080987718, + "learning_rate": 1.7678215527293724e-05, + "loss": 0.654, "step": 636 }, { - "epoch": 0.12, - "learning_rate": 1.9554064546948064e-05, - "loss": 0.9689, + "epoch": 0.24509426702577913, + "grad_norm": 1.0187061374640627, + "learning_rate": 1.767022581266242e-05, + "loss": 0.633, "step": 637 }, { - "epoch": 0.12, - "learning_rate": 1.9552222395655262e-05, - "loss": 0.9254, + "epoch": 0.24547903039630628, + "grad_norm": 0.6413326486776452, + "learning_rate": 1.766222418664926e-05, + "loss": 0.6919, "step": 638 }, { - "epoch": 0.12, - "learning_rate": 1.9550376534389355e-05, - "loss": 0.923, + "epoch": 0.2458637937668334, + "grad_norm": 0.9296331688822185, + "learning_rate": 1.7654210661680263e-05, + "loss": 0.6321, "step": 639 }, { - "epoch": 0.12, - "learning_rate": 1.9548526963867253e-05, - "loss": 0.9817, + "epoch": 0.24624855713736052, + "grad_norm": 0.9526734206961187, + "learning_rate": 1.7646185250199936e-05, + "loss": 0.5945, "step": 640 }, { - "epoch": 0.12, - "learning_rate": 1.9546673684807303e-05, - "loss": 0.9586, + "epoch": 0.24663332050788764, + "grad_norm": 0.9451821742676015, + "learning_rate": 1.763814796467124e-05, + "loss": 0.5433, "step": 641 }, { - "epoch": 0.12, - "learning_rate": 1.95448166979293e-05, - "loss": 1.0334, + "epoch": 0.24701808387841478, + "grad_norm": 0.9763996640489023, + "learning_rate": 1.7630098817575578e-05, + "loss": 0.6321, "step": 642 }, { - "epoch": 0.12, - "learning_rate": 1.9542956003954477e-05, - "loss": 0.993, + "epoch": 0.2474028472489419, + "grad_norm": 1.0018973295618234, + "learning_rate": 1.7622037821412775e-05, + "loss": 0.6187, "step": 643 }, { - "epoch": 0.12, - "learning_rate": 1.9541091603605508e-05, - "loss": 0.9016, + "epoch": 0.24778761061946902, + "grad_norm": 0.9872997188338046, + "learning_rate": 1.7613964988701057e-05, + "loss": 0.6115, "step": 644 }, { - "epoch": 0.12, - "learning_rate": 1.95392234976065e-05, - "loss": 0.96, + "epoch": 0.24817237398999614, + "grad_norm": 0.9866706946939893, + "learning_rate": 1.7605880331977022e-05, + "loss": 0.623, "step": 645 }, { - "epoch": 0.12, - "learning_rate": 1.9537351686683003e-05, - "loss": 0.9231, + "epoch": 0.2485571373605233, + "grad_norm": 1.0792195216793208, + "learning_rate": 1.7597783863795644e-05, + "loss": 0.6261, "step": 646 }, { - "epoch": 0.12, - "learning_rate": 1.9535476171562014e-05, - "loss": 0.9897, + "epoch": 0.2489419007310504, + "grad_norm": 0.6127839671033437, + "learning_rate": 1.7589675596730233e-05, + "loss": 0.6403, "step": 647 }, { - "epoch": 0.12, - "learning_rate": 1.9533596952971955e-05, - "loss": 0.8802, + "epoch": 0.24932666410157753, + "grad_norm": 0.9912509017863798, + "learning_rate": 1.7581555543372413e-05, + "loss": 0.6179, "step": 648 }, { - "epoch": 0.12, - "learning_rate": 1.9531714031642698e-05, - "loss": 0.9272, + "epoch": 0.24971142747210465, + "grad_norm": 1.2054980844707126, + "learning_rate": 1.7573423716332128e-05, + "loss": 0.5901, "step": 649 }, { - "epoch": 0.13, - "learning_rate": 1.9529827408305542e-05, - "loss": 0.9686, + "epoch": 0.2500961908426318, + "grad_norm": 0.9535373202292964, + "learning_rate": 1.7565280128237595e-05, + "loss": 0.6254, "step": 650 }, { - "epoch": 0.13, - "learning_rate": 1.9527937083693233e-05, - "loss": 1.0324, + "epoch": 0.2504809542131589, + "grad_norm": 1.0498210184898547, + "learning_rate": 1.75571247917353e-05, + "loss": 0.6022, "step": 651 }, { - "epoch": 0.13, - "learning_rate": 1.952604305853995e-05, - "loss": 0.9852, + "epoch": 0.25086571758368603, + "grad_norm": 0.9543872415692135, + "learning_rate": 1.754895771948997e-05, + "loss": 0.576, "step": 652 }, { - "epoch": 0.13, - "learning_rate": 1.9524145333581315e-05, - "loss": 0.9732, + "epoch": 0.25125048095421315, + "grad_norm": 1.1253497789530313, + "learning_rate": 1.7540778924184553e-05, + "loss": 0.6183, "step": 653 }, { - "epoch": 0.13, - "learning_rate": 1.9522243909554375e-05, - "loss": 0.9407, + "epoch": 0.25163524432474027, + "grad_norm": 0.9663742958282615, + "learning_rate": 1.7532588418520215e-05, + "loss": 0.6212, "step": 654 }, { - "epoch": 0.13, - "learning_rate": 1.952033878719763e-05, - "loss": 1.0214, + "epoch": 0.2520200076952674, + "grad_norm": 0.9890248689248791, + "learning_rate": 1.75243862152163e-05, + "loss": 0.6214, "step": 655 }, { - "epoch": 0.13, - "learning_rate": 1.9518429967251e-05, - "loss": 0.7629, + "epoch": 0.2524047710657945, + "grad_norm": 0.9777685861571538, + "learning_rate": 1.7516172327010314e-05, + "loss": 0.5634, "step": 656 }, { - "epoch": 0.13, - "learning_rate": 1.9516517450455853e-05, - "loss": 0.9278, + "epoch": 0.2527895344363217, + "grad_norm": 1.0410122308515757, + "learning_rate": 1.7507946766657914e-05, + "loss": 0.5901, "step": 657 }, { - "epoch": 0.13, - "learning_rate": 1.951460123755499e-05, - "loss": 0.9303, + "epoch": 0.2531742978068488, + "grad_norm": 0.9800761339347797, + "learning_rate": 1.749970954693288e-05, + "loss": 0.5627, "step": 658 }, { - "epoch": 0.13, - "learning_rate": 1.9512681329292635e-05, - "loss": 0.9538, + "epoch": 0.2535590611773759, + "grad_norm": 1.0059785334520728, + "learning_rate": 1.7491460680627105e-05, + "loss": 0.6179, "step": 659 }, { - "epoch": 0.13, - "learning_rate": 1.951075772641447e-05, - "loss": 0.8833, + "epoch": 0.25394382454790304, + "grad_norm": 0.9893374299846757, + "learning_rate": 1.7483200180550554e-05, + "loss": 0.5849, "step": 660 }, { - "epoch": 0.13, - "learning_rate": 1.95088304296676e-05, - "loss": 0.9322, + "epoch": 0.25432858791843016, + "grad_norm": 1.1857134892321888, + "learning_rate": 1.747492805953128e-05, + "loss": 0.6377, "step": 661 }, { - "epoch": 0.13, - "learning_rate": 1.950689943980056e-05, - "loss": 0.9749, + "epoch": 0.2547133512889573, + "grad_norm": 0.8911953671254377, + "learning_rate": 1.7466644330415362e-05, + "loss": 0.5409, "step": 662 }, { - "epoch": 0.13, - "learning_rate": 1.9504964757563322e-05, - "loss": 1.0269, + "epoch": 0.2550981146594844, + "grad_norm": 1.0174870938410525, + "learning_rate": 1.745834900606692e-05, + "loss": 0.6517, "step": 663 }, { - "epoch": 0.13, - "learning_rate": 1.95030263837073e-05, - "loss": 0.9098, + "epoch": 0.2554828780300115, + "grad_norm": 0.9678742267924958, + "learning_rate": 1.7450042099368066e-05, + "loss": 0.6542, "step": 664 }, { - "epoch": 0.13, - "learning_rate": 1.9501084318985335e-05, - "loss": 1.0035, + "epoch": 0.2558676414005387, + "grad_norm": 0.9883777078332283, + "learning_rate": 1.7441723623218917e-05, + "loss": 0.6221, "step": 665 }, { - "epoch": 0.13, - "learning_rate": 1.94991385641517e-05, - "loss": 0.9489, + "epoch": 0.2562524047710658, + "grad_norm": 0.9770582176671307, + "learning_rate": 1.7433393590537543e-05, + "loss": 0.5365, "step": 666 }, { - "epoch": 0.13, - "learning_rate": 1.9497189119962105e-05, - "loss": 1.0332, + "epoch": 0.25663716814159293, + "grad_norm": 0.9421260143744268, + "learning_rate": 1.7425052014259965e-05, + "loss": 0.6195, "step": 667 }, { - "epoch": 0.13, - "learning_rate": 1.9495235987173693e-05, - "loss": 0.9546, + "epoch": 0.25702193151212005, + "grad_norm": 1.2604108977744515, + "learning_rate": 1.7416698907340128e-05, + "loss": 0.6493, "step": 668 }, { - "epoch": 0.13, - "learning_rate": 1.949327916654504e-05, - "loss": 1.0167, + "epoch": 0.25740669488264717, + "grad_norm": 0.9225876466429362, + "learning_rate": 1.740833428274989e-05, + "loss": 0.6048, "step": 669 }, { - "epoch": 0.13, - "learning_rate": 1.949131865883614e-05, - "loss": 0.9073, + "epoch": 0.2577914582531743, + "grad_norm": 1.0076069458509436, + "learning_rate": 1.739995815347899e-05, + "loss": 0.588, "step": 670 }, { - "epoch": 0.13, - "learning_rate": 1.948935446480845e-05, - "loss": 0.9386, + "epoch": 0.2581762216237014, + "grad_norm": 0.9283557429535663, + "learning_rate": 1.739157053253503e-05, + "loss": 0.5512, "step": 671 }, { - "epoch": 0.13, - "learning_rate": 1.948738658522483e-05, - "loss": 0.974, + "epoch": 0.2585609849942285, + "grad_norm": 0.9915296454942223, + "learning_rate": 1.7383171432943466e-05, + "loss": 0.5988, "step": 672 }, { - "epoch": 0.13, - "learning_rate": 1.9485415020849583e-05, - "loss": 0.7766, + "epoch": 0.2589457483647557, + "grad_norm": 1.0463179606534776, + "learning_rate": 1.7374760867747574e-05, + "loss": 0.5651, "step": 673 }, { - "epoch": 0.13, - "learning_rate": 1.9483439772448444e-05, - "loss": 0.8568, + "epoch": 0.2593305117352828, + "grad_norm": 1.070106061960922, + "learning_rate": 1.7366338850008432e-05, + "loss": 0.5876, "step": 674 }, { - "epoch": 0.13, - "learning_rate": 1.9481460840788573e-05, - "loss": 0.9855, + "epoch": 0.25971527510580994, + "grad_norm": 1.0775593522754803, + "learning_rate": 1.7357905392804918e-05, + "loss": 0.6381, "step": 675 }, { - "epoch": 0.13, - "learning_rate": 1.9479478226638565e-05, - "loss": 0.7989, + "epoch": 0.26010003847633706, + "grad_norm": 0.9419090855242344, + "learning_rate": 1.7349460509233654e-05, + "loss": 0.6042, "step": 676 }, { - "epoch": 0.13, - "learning_rate": 1.947749193076845e-05, - "loss": 0.9905, + "epoch": 0.2604848018468642, + "grad_norm": 0.904274037420734, + "learning_rate": 1.7341004212409026e-05, + "loss": 0.6139, "step": 677 }, { - "epoch": 0.13, - "learning_rate": 1.9475501953949674e-05, - "loss": 0.9383, + "epoch": 0.2608695652173913, + "grad_norm": 1.031789748799266, + "learning_rate": 1.7332536515463126e-05, + "loss": 0.5725, "step": 678 }, { - "epoch": 0.13, - "learning_rate": 1.9473508296955126e-05, - "loss": 0.7806, + "epoch": 0.2612543285879184, + "grad_norm": 1.0096973895609505, + "learning_rate": 1.7324057431545768e-05, + "loss": 0.6297, "step": 679 }, { - "epoch": 0.13, - "learning_rate": 1.9471510960559122e-05, - "loss": 0.983, + "epoch": 0.26163909195844554, + "grad_norm": 1.01779581358481, + "learning_rate": 1.7315566973824433e-05, + "loss": 0.643, "step": 680 }, { - "epoch": 0.13, - "learning_rate": 1.9469509945537395e-05, - "loss": 1.0721, + "epoch": 0.26202385532897265, + "grad_norm": 0.9300979313731453, + "learning_rate": 1.730706515548427e-05, + "loss": 0.5942, "step": 681 }, { - "epoch": 0.13, - "learning_rate": 1.9467505252667126e-05, - "loss": 0.999, + "epoch": 0.26240861869949983, + "grad_norm": 1.0211092500403316, + "learning_rate": 1.729855198972808e-05, + "loss": 0.5927, "step": 682 }, { - "epoch": 0.13, - "learning_rate": 1.9465496882726913e-05, - "loss": 1.0216, + "epoch": 0.26279338207002695, + "grad_norm": 1.0670990755064174, + "learning_rate": 1.729002748977628e-05, + "loss": 0.6102, "step": 683 }, { - "epoch": 0.13, - "learning_rate": 1.946348483649678e-05, - "loss": 0.9658, + "epoch": 0.26317814544055407, + "grad_norm": 0.9519793332351464, + "learning_rate": 1.7281491668866874e-05, + "loss": 0.5437, "step": 684 }, { - "epoch": 0.13, - "learning_rate": 1.9461469114758184e-05, - "loss": 0.9744, + "epoch": 0.2635629088110812, + "grad_norm": 1.04596369175055, + "learning_rate": 1.7272944540255468e-05, + "loss": 0.5645, "step": 685 }, { - "epoch": 0.13, - "learning_rate": 1.9459449718294008e-05, - "loss": 0.9352, + "epoch": 0.2639476721816083, + "grad_norm": 0.940224320955648, + "learning_rate": 1.7264386117215216e-05, + "loss": 0.5936, "step": 686 }, { - "epoch": 0.13, - "learning_rate": 1.945742664788856e-05, - "loss": 1.0193, + "epoch": 0.2643324355521354, + "grad_norm": 1.0606892987955536, + "learning_rate": 1.7255816413036818e-05, + "loss": 0.6997, "step": 687 }, { - "epoch": 0.13, - "learning_rate": 1.9455399904327585e-05, - "loss": 1.0227, + "epoch": 0.26471719892266254, + "grad_norm": 1.1192205024204056, + "learning_rate": 1.7247235441028486e-05, + "loss": 0.5901, "step": 688 }, { - "epoch": 0.13, - "learning_rate": 1.945336948839824e-05, - "loss": 0.9465, + "epoch": 0.26510196229318966, + "grad_norm": 0.9283051379164949, + "learning_rate": 1.7238643214515934e-05, + "loss": 0.5775, "step": 689 }, { - "epoch": 0.13, - "learning_rate": 1.9451335400889114e-05, - "loss": 1.0611, + "epoch": 0.26548672566371684, + "grad_norm": 1.0548180599401387, + "learning_rate": 1.7230039746842352e-05, + "loss": 0.5633, "step": 690 }, { - "epoch": 0.13, - "learning_rate": 1.944929764259023e-05, - "loss": 0.7507, + "epoch": 0.26587148903424396, + "grad_norm": 0.7489794350107883, + "learning_rate": 1.7221425051368394e-05, + "loss": 0.7077, "step": 691 }, { - "epoch": 0.13, - "learning_rate": 1.9447256214293026e-05, - "loss": 0.9665, + "epoch": 0.2662562524047711, + "grad_norm": 1.0574161172441945, + "learning_rate": 1.721279914147214e-05, + "loss": 0.5907, "step": 692 }, { - "epoch": 0.13, - "learning_rate": 1.9445211116790365e-05, - "loss": 0.992, + "epoch": 0.2666410157752982, + "grad_norm": 1.0918207535050986, + "learning_rate": 1.7204162030549093e-05, + "loss": 0.6132, "step": 693 }, { - "epoch": 0.13, - "learning_rate": 1.9443162350876544e-05, - "loss": 0.9419, + "epoch": 0.2670257791458253, + "grad_norm": 0.9650960061676707, + "learning_rate": 1.719551373201214e-05, + "loss": 0.6184, "step": 694 }, { - "epoch": 0.13, - "learning_rate": 1.944110991734728e-05, - "loss": 0.9551, + "epoch": 0.26741054251635243, + "grad_norm": 1.0091175304620532, + "learning_rate": 1.7186854259291558e-05, + "loss": 0.6237, "step": 695 }, { - "epoch": 0.13, - "learning_rate": 1.9439053816999717e-05, - "loss": 0.9421, + "epoch": 0.26779530588687955, + "grad_norm": 0.9937381046857249, + "learning_rate": 1.717818362583496e-05, + "loss": 0.5979, "step": 696 }, { - "epoch": 0.13, - "learning_rate": 1.9436994050632415e-05, - "loss": 0.9504, + "epoch": 0.2681800692574067, + "grad_norm": 1.0293852284226077, + "learning_rate": 1.71695018451073e-05, + "loss": 0.6625, "step": 697 }, { - "epoch": 0.13, - "learning_rate": 1.9434930619045367e-05, - "loss": 0.9861, + "epoch": 0.26856483262793385, + "grad_norm": 1.0712580129648237, + "learning_rate": 1.7160808930590845e-05, + "loss": 0.6633, "step": 698 }, { - "epoch": 0.13, - "learning_rate": 1.9432863523039986e-05, - "loss": 0.9616, + "epoch": 0.26894959599846097, + "grad_norm": 0.9807148840879941, + "learning_rate": 1.7152104895785147e-05, + "loss": 0.5624, "step": 699 }, { - "epoch": 0.13, - "learning_rate": 1.9430792763419105e-05, - "loss": 0.9515, + "epoch": 0.2693343593689881, + "grad_norm": 0.954276596699698, + "learning_rate": 1.7143389754207026e-05, + "loss": 0.5851, "step": 700 }, { - "epoch": 0.13, - "learning_rate": 1.942871834098699e-05, - "loss": 0.9867, + "epoch": 0.2697191227395152, + "grad_norm": 1.0382905708223586, + "learning_rate": 1.7134663519390557e-05, + "loss": 0.6373, "step": 701 }, { - "epoch": 0.14, - "learning_rate": 1.9426640256549313e-05, - "loss": 0.7281, + "epoch": 0.2701038861100423, + "grad_norm": 1.0447418459469073, + "learning_rate": 1.7125926204887034e-05, + "loss": 0.6104, "step": 702 }, { - "epoch": 0.14, - "learning_rate": 1.9424558510913186e-05, - "loss": 0.9783, + "epoch": 0.27048864948056944, + "grad_norm": 1.0116926515119515, + "learning_rate": 1.7117177824264962e-05, + "loss": 0.6055, "step": 703 }, { - "epoch": 0.14, - "learning_rate": 1.9422473104887133e-05, - "loss": 0.9933, + "epoch": 0.27087341285109656, + "grad_norm": 1.1313393004520806, + "learning_rate": 1.7108418391110033e-05, + "loss": 0.6439, "step": 704 }, { - "epoch": 0.14, - "learning_rate": 1.9420384039281103e-05, - "loss": 0.9816, + "epoch": 0.2712581762216237, + "grad_norm": 1.1620716183811288, + "learning_rate": 1.7099647919025096e-05, + "loss": 0.634, "step": 705 }, { - "epoch": 0.14, - "learning_rate": 1.941829131490646e-05, - "loss": 0.9742, + "epoch": 0.2716429395921508, + "grad_norm": 0.8393915811045367, + "learning_rate": 1.709086642163015e-05, + "loss": 0.5431, "step": 706 }, { - "epoch": 0.14, - "learning_rate": 1.9416194932576e-05, - "loss": 0.9636, + "epoch": 0.272027702962678, + "grad_norm": 1.065880621242293, + "learning_rate": 1.708207391256231e-05, + "loss": 0.5741, "step": 707 }, { - "epoch": 0.14, - "learning_rate": 1.941409489310393e-05, - "loss": 0.9295, + "epoch": 0.2724124663332051, + "grad_norm": 0.6527558133093259, + "learning_rate": 1.7073270405475796e-05, + "loss": 0.6383, "step": 708 }, { - "epoch": 0.14, - "learning_rate": 1.9411991197305878e-05, - "loss": 0.9476, + "epoch": 0.2727972297037322, + "grad_norm": 1.0125118811070757, + "learning_rate": 1.70644559140419e-05, + "loss": 0.613, "step": 709 }, { - "epoch": 0.14, - "learning_rate": 1.9409883845998905e-05, - "loss": 0.9968, + "epoch": 0.27318199307425933, + "grad_norm": 1.0840657135932987, + "learning_rate": 1.705563045194898e-05, + "loss": 0.6172, "step": 710 }, { - "epoch": 0.14, - "learning_rate": 1.9407772840001473e-05, - "loss": 0.9541, + "epoch": 0.27356675644478645, + "grad_norm": 1.0771975070705357, + "learning_rate": 1.704679403290243e-05, + "loss": 0.627, "step": 711 }, { - "epoch": 0.14, - "learning_rate": 1.9405658180133477e-05, - "loss": 1.028, + "epoch": 0.27395151981531357, + "grad_norm": 1.082144868636492, + "learning_rate": 1.7037946670624652e-05, + "loss": 0.6643, "step": 712 }, { - "epoch": 0.14, - "learning_rate": 1.9403539867216226e-05, - "loss": 1.0312, + "epoch": 0.2743362831858407, + "grad_norm": 0.6838486943112129, + "learning_rate": 1.7029088378855055e-05, + "loss": 0.6983, "step": 713 }, { - "epoch": 0.14, - "learning_rate": 1.9401417902072447e-05, - "loss": 0.9365, + "epoch": 0.2747210465563678, + "grad_norm": 1.009634535092224, + "learning_rate": 1.7020219171350004e-05, + "loss": 0.6158, "step": 714 }, { - "epoch": 0.14, - "learning_rate": 1.9399292285526286e-05, - "loss": 1.0804, + "epoch": 0.275105809926895, + "grad_norm": 0.9818255176667509, + "learning_rate": 1.701133906188283e-05, + "loss": 0.6162, "step": 715 }, { - "epoch": 0.14, - "learning_rate": 1.939716301840331e-05, - "loss": 1.0046, + "epoch": 0.2754905732974221, + "grad_norm": 0.9113635594719018, + "learning_rate": 1.700244806424379e-05, + "loss": 0.5815, "step": 716 }, { - "epoch": 0.14, - "learning_rate": 1.9395030101530504e-05, - "loss": 1.0297, + "epoch": 0.2758753366679492, + "grad_norm": 0.9790759590725917, + "learning_rate": 1.699354619224004e-05, + "loss": 0.5991, "step": 717 }, { - "epoch": 0.14, - "learning_rate": 1.939289353573626e-05, - "loss": 1.012, + "epoch": 0.27626010003847634, + "grad_norm": 0.9945020955728519, + "learning_rate": 1.6984633459695646e-05, + "loss": 0.6104, "step": 718 }, { - "epoch": 0.14, - "learning_rate": 1.9390753321850404e-05, - "loss": 0.961, + "epoch": 0.27664486340900346, + "grad_norm": 1.0547455952335867, + "learning_rate": 1.697570988045151e-05, + "loss": 0.6096, "step": 719 }, { - "epoch": 0.14, - "learning_rate": 1.938860946070417e-05, - "loss": 0.9814, + "epoch": 0.2770296267795306, + "grad_norm": 0.9934342200041285, + "learning_rate": 1.69667754683654e-05, + "loss": 0.5988, "step": 720 }, { - "epoch": 0.14, - "learning_rate": 1.93864619531302e-05, - "loss": 1.0053, + "epoch": 0.2774143901500577, + "grad_norm": 0.67141224371749, + "learning_rate": 1.6957830237311904e-05, + "loss": 0.6918, "step": 721 }, { - "epoch": 0.14, - "learning_rate": 1.9384310799962575e-05, - "loss": 1.0216, + "epoch": 0.2777991535205848, + "grad_norm": 1.2804322250757618, + "learning_rate": 1.6948874201182402e-05, + "loss": 0.6188, "step": 722 }, { - "epoch": 0.14, - "learning_rate": 1.9382156002036764e-05, - "loss": 0.9829, + "epoch": 0.278183916891112, + "grad_norm": 1.1302959494223959, + "learning_rate": 1.6939907373885062e-05, + "loss": 0.6554, "step": 723 }, { - "epoch": 0.14, - "learning_rate": 1.9379997560189677e-05, - "loss": 1.0001, + "epoch": 0.2785686802616391, + "grad_norm": 1.0342213304673609, + "learning_rate": 1.6930929769344807e-05, + "loss": 0.6484, "step": 724 }, { - "epoch": 0.14, - "learning_rate": 1.937783547525962e-05, - "loss": 0.9924, + "epoch": 0.27895344363216623, + "grad_norm": 1.1015096587040252, + "learning_rate": 1.692194140150329e-05, + "loss": 0.6113, "step": 725 }, { - "epoch": 0.14, - "learning_rate": 1.9375669748086326e-05, - "loss": 0.9444, + "epoch": 0.27933820700269335, + "grad_norm": 0.9991417276617662, + "learning_rate": 1.6912942284318898e-05, + "loss": 0.6398, "step": 726 }, { - "epoch": 0.14, - "learning_rate": 1.937350037951094e-05, - "loss": 1.0269, + "epoch": 0.27972297037322047, + "grad_norm": 0.9306912585173606, + "learning_rate": 1.690393243176668e-05, + "loss": 0.5974, "step": 727 }, { - "epoch": 0.14, - "learning_rate": 1.9371327370376018e-05, - "loss": 1.0449, + "epoch": 0.2801077337437476, + "grad_norm": 0.9329360193578702, + "learning_rate": 1.6894911857838394e-05, + "loss": 0.5728, "step": 728 }, { - "epoch": 0.14, - "learning_rate": 1.936915072152553e-05, - "loss": 0.9658, + "epoch": 0.2804924971142747, + "grad_norm": 1.0586177112042972, + "learning_rate": 1.6885880576542417e-05, + "loss": 0.5773, "step": 729 }, { - "epoch": 0.14, - "learning_rate": 1.936697043380486e-05, - "loss": 0.7386, + "epoch": 0.2808772604848018, + "grad_norm": 1.0195476579837952, + "learning_rate": 1.6876838601903765e-05, + "loss": 0.6225, "step": 730 }, { - "epoch": 0.14, - "learning_rate": 1.936478650806081e-05, - "loss": 1.011, + "epoch": 0.28126202385532895, + "grad_norm": 1.022505278595614, + "learning_rate": 1.6867785947964065e-05, + "loss": 0.6196, "step": 731 }, { - "epoch": 0.14, - "learning_rate": 1.936259894514159e-05, - "loss": 0.9907, + "epoch": 0.2816467872258561, + "grad_norm": 0.9166752068397543, + "learning_rate": 1.685872262878152e-05, + "loss": 0.6162, "step": 732 }, { - "epoch": 0.14, - "learning_rate": 1.9360407745896828e-05, - "loss": 1.0081, + "epoch": 0.28203155059638324, + "grad_norm": 1.0631320032943956, + "learning_rate": 1.68496486584309e-05, + "loss": 0.7059, "step": 733 }, { - "epoch": 0.14, - "learning_rate": 1.9358212911177556e-05, - "loss": 0.9645, + "epoch": 0.28241631396691036, + "grad_norm": 0.9895020532885922, + "learning_rate": 1.6840564051003517e-05, + "loss": 0.5881, "step": 734 }, { - "epoch": 0.14, - "learning_rate": 1.935601444183622e-05, - "loss": 1.0177, + "epoch": 0.2828010773374375, + "grad_norm": 1.0855196901251774, + "learning_rate": 1.6831468820607192e-05, + "loss": 0.6032, "step": 735 }, { - "epoch": 0.14, - "learning_rate": 1.935381233872669e-05, - "loss": 0.8948, + "epoch": 0.2831858407079646, + "grad_norm": 1.0316187458951132, + "learning_rate": 1.6822362981366257e-05, + "loss": 0.6224, "step": 736 }, { - "epoch": 0.14, - "learning_rate": 1.935160660270423e-05, - "loss": 0.9769, + "epoch": 0.2835706040784917, + "grad_norm": 1.1347668345567334, + "learning_rate": 1.681324654742151e-05, + "loss": 0.6374, "step": 737 }, { - "epoch": 0.14, - "learning_rate": 1.934939723462552e-05, - "loss": 0.9548, + "epoch": 0.28395536744901884, + "grad_norm": 0.9763728488853611, + "learning_rate": 1.6804119532930202e-05, + "loss": 0.6417, "step": 738 }, { - "epoch": 0.14, - "learning_rate": 1.9347184235348663e-05, - "loss": 0.9704, + "epoch": 0.28434013081954596, + "grad_norm": 0.9798207081775457, + "learning_rate": 1.6794981952066018e-05, + "loss": 0.5901, "step": 739 }, { - "epoch": 0.14, - "learning_rate": 1.9344967605733154e-05, - "loss": 0.945, + "epoch": 0.28472489419007313, + "grad_norm": 0.8925554341360649, + "learning_rate": 1.6785833819019052e-05, + "loss": 0.5447, "step": 740 }, { - "epoch": 0.14, - "learning_rate": 1.934274734663991e-05, - "loss": 0.9391, + "epoch": 0.28510965756060025, + "grad_norm": 1.0554116282790795, + "learning_rate": 1.677667514799578e-05, + "loss": 0.6138, "step": 741 }, { - "epoch": 0.14, - "learning_rate": 1.934052345893125e-05, - "loss": 0.9751, + "epoch": 0.28549442093112737, + "grad_norm": 1.091676207635543, + "learning_rate": 1.676750595321905e-05, + "loss": 0.691, "step": 742 }, { - "epoch": 0.14, - "learning_rate": 1.9338295943470915e-05, - "loss": 0.9623, + "epoch": 0.2858791843016545, + "grad_norm": 1.0096056039423238, + "learning_rate": 1.675832624892805e-05, + "loss": 0.6189, "step": 743 }, { - "epoch": 0.14, - "learning_rate": 1.9336064801124034e-05, - "loss": 0.966, + "epoch": 0.2862639476721816, + "grad_norm": 1.0246113310538654, + "learning_rate": 1.674913604937828e-05, + "loss": 0.619, "step": 744 }, { - "epoch": 0.14, - "learning_rate": 1.933383003275717e-05, - "loss": 0.974, + "epoch": 0.2866487110427087, + "grad_norm": 1.0054766274099256, + "learning_rate": 1.6739935368841555e-05, + "loss": 0.6137, "step": 745 }, { - "epoch": 0.14, - "learning_rate": 1.933159163923827e-05, - "loss": 1.0206, + "epoch": 0.28703347441323585, + "grad_norm": 1.0058769911093366, + "learning_rate": 1.6730724221605955e-05, + "loss": 0.6182, "step": 746 }, { - "epoch": 0.14, - "learning_rate": 1.9329349621436708e-05, - "loss": 0.9334, + "epoch": 0.28741823778376296, + "grad_norm": 1.8207074404681, + "learning_rate": 1.6721502621975813e-05, + "loss": 0.5849, "step": 747 }, { - "epoch": 0.14, - "learning_rate": 1.9327103980223255e-05, - "loss": 0.7681, + "epoch": 0.28780300115429014, + "grad_norm": 0.9531411306639124, + "learning_rate": 1.6712270584271703e-05, + "loss": 0.5471, "step": 748 }, { - "epoch": 0.14, - "learning_rate": 1.932485471647009e-05, - "loss": 0.6891, + "epoch": 0.28818776452481726, + "grad_norm": 0.6299678662586666, + "learning_rate": 1.67030281228304e-05, + "loss": 0.6595, "step": 749 }, { - "epoch": 0.14, - "learning_rate": 1.9322601831050804e-05, - "loss": 1.0307, + "epoch": 0.2885725278953444, + "grad_norm": 1.0613073748230408, + "learning_rate": 1.6693775252004866e-05, + "loss": 0.6352, "step": 750 }, { - "epoch": 0.14, - "learning_rate": 1.9320345324840396e-05, - "loss": 1.0599, + "epoch": 0.2889572912658715, + "grad_norm": 0.9858944139128867, + "learning_rate": 1.668451198616424e-05, + "loss": 0.5793, "step": 751 }, { - "epoch": 0.14, - "learning_rate": 1.9318085198715257e-05, - "loss": 1.0171, + "epoch": 0.2893420546363986, + "grad_norm": 0.9055201092023991, + "learning_rate": 1.667523833969379e-05, + "loss": 0.557, "step": 752 }, { - "epoch": 0.14, - "learning_rate": 1.93158214535532e-05, - "loss": 0.9328, + "epoch": 0.28972681800692573, + "grad_norm": 1.0023686337731668, + "learning_rate": 1.666595432699491e-05, + "loss": 0.6348, "step": 753 }, { - "epoch": 0.15, - "learning_rate": 1.9313554090233436e-05, - "loss": 0.9459, + "epoch": 0.29011158137745285, + "grad_norm": 0.992750914181895, + "learning_rate": 1.6656659962485097e-05, + "loss": 0.6006, "step": 754 }, { - "epoch": 0.15, - "learning_rate": 1.9311283109636586e-05, - "loss": 0.9618, + "epoch": 0.29049634474798, + "grad_norm": 0.9719271195726, + "learning_rate": 1.6647355260597915e-05, + "loss": 0.5938, "step": 755 }, { - "epoch": 0.15, - "learning_rate": 1.9309008512644668e-05, - "loss": 0.9906, + "epoch": 0.2908811081185071, + "grad_norm": 1.0522049897991668, + "learning_rate": 1.6638040235782983e-05, + "loss": 0.6595, "step": 756 }, { - "epoch": 0.15, - "learning_rate": 1.930673030014111e-05, - "loss": 1.0824, + "epoch": 0.29126587148903427, + "grad_norm": 1.001933261629964, + "learning_rate": 1.662871490250596e-05, + "loss": 0.5612, "step": 757 }, { - "epoch": 0.15, - "learning_rate": 1.930444847301075e-05, - "loss": 1.0489, + "epoch": 0.2916506348595614, + "grad_norm": 1.046777985745525, + "learning_rate": 1.66193792752485e-05, + "loss": 0.6342, "step": 758 }, { - "epoch": 0.15, - "learning_rate": 1.9302163032139813e-05, - "loss": 0.9507, + "epoch": 0.2920353982300885, + "grad_norm": 0.916696641994943, + "learning_rate": 1.661003336850825e-05, + "loss": 0.5373, "step": 759 }, { - "epoch": 0.15, - "learning_rate": 1.9299873978415947e-05, - "loss": 0.9777, + "epoch": 0.2924201616006156, + "grad_norm": 1.119368131183829, + "learning_rate": 1.660067719679882e-05, + "loss": 0.6642, "step": 760 }, { - "epoch": 0.15, - "learning_rate": 1.9297581312728187e-05, - "loss": 0.9246, + "epoch": 0.29280492497114274, + "grad_norm": 0.9526156999377315, + "learning_rate": 1.6591310774649766e-05, + "loss": 0.5525, "step": 761 }, { - "epoch": 0.15, - "learning_rate": 1.929528503596698e-05, - "loss": 0.9554, + "epoch": 0.29318968834166986, + "grad_norm": 0.929574046698848, + "learning_rate": 1.6581934116606554e-05, + "loss": 0.581, "step": 762 }, { - "epoch": 0.15, - "learning_rate": 1.929298514902418e-05, - "loss": 1.0437, + "epoch": 0.293574451712197, + "grad_norm": 1.0716533435511868, + "learning_rate": 1.657254723723054e-05, + "loss": 0.6331, "step": 763 }, { - "epoch": 0.15, - "learning_rate": 1.929068165279303e-05, - "loss": 0.8959, + "epoch": 0.2939592150827241, + "grad_norm": 0.9759604742962976, + "learning_rate": 1.6563150151098973e-05, + "loss": 0.5768, "step": 764 }, { - "epoch": 0.15, - "learning_rate": 1.928837454816818e-05, - "loss": 0.9859, + "epoch": 0.2943439784532513, + "grad_norm": 0.9882295541084339, + "learning_rate": 1.655374287280494e-05, + "loss": 0.6384, "step": 765 }, { - "epoch": 0.15, - "learning_rate": 1.9286063836045687e-05, - "loss": 0.9279, + "epoch": 0.2947287418237784, + "grad_norm": 0.9376213205796459, + "learning_rate": 1.654432541695735e-05, + "loss": 0.5941, "step": 766 }, { - "epoch": 0.15, - "learning_rate": 1.9283749517323e-05, - "loss": 0.9376, + "epoch": 0.2951135051943055, + "grad_norm": 0.9848473865912016, + "learning_rate": 1.653489779818093e-05, + "loss": 0.6475, "step": 767 }, { - "epoch": 0.15, - "learning_rate": 1.928143159289898e-05, - "loss": 0.9781, + "epoch": 0.29549826856483263, + "grad_norm": 0.8596454206637315, + "learning_rate": 1.652546003111618e-05, + "loss": 0.5689, "step": 768 }, { - "epoch": 0.15, - "learning_rate": 1.927911006367388e-05, - "loss": 0.9953, + "epoch": 0.29588303193535975, + "grad_norm": 1.1307795868637531, + "learning_rate": 1.6516012130419366e-05, + "loss": 0.6146, "step": 769 }, { - "epoch": 0.15, - "learning_rate": 1.927678493054935e-05, - "loss": 1.0913, + "epoch": 0.29626779530588687, + "grad_norm": 0.9193011729845966, + "learning_rate": 1.6506554110762483e-05, + "loss": 0.563, "step": 770 }, { - "epoch": 0.15, - "learning_rate": 1.9274456194428454e-05, - "loss": 0.9519, + "epoch": 0.296652558676414, + "grad_norm": 0.9005495978288045, + "learning_rate": 1.6497085986833252e-05, + "loss": 0.5709, "step": 771 }, { - "epoch": 0.15, - "learning_rate": 1.9272123856215643e-05, - "loss": 0.9472, + "epoch": 0.2970373220469411, + "grad_norm": 1.1194769295926972, + "learning_rate": 1.6487607773335074e-05, + "loss": 0.6029, "step": 772 }, { - "epoch": 0.15, - "learning_rate": 1.9269787916816764e-05, - "loss": 0.9633, + "epoch": 0.29742208541746823, + "grad_norm": 0.9759729018341151, + "learning_rate": 1.6478119484987026e-05, + "loss": 0.6269, "step": 773 }, { - "epoch": 0.15, - "learning_rate": 1.9267448377139074e-05, - "loss": 1.0105, + "epoch": 0.2978068487879954, + "grad_norm": 0.6759355723867863, + "learning_rate": 1.6468621136523823e-05, + "loss": 0.656, "step": 774 }, { - "epoch": 0.15, - "learning_rate": 1.9265105238091227e-05, - "loss": 0.964, + "epoch": 0.2981916121585225, + "grad_norm": 1.0972022448939385, + "learning_rate": 1.6459112742695807e-05, + "loss": 0.6243, "step": 775 }, { - "epoch": 0.15, - "learning_rate": 1.9262758500583265e-05, - "loss": 0.9976, + "epoch": 0.29857637552904964, + "grad_norm": 0.9617598145933418, + "learning_rate": 1.644959431826893e-05, + "loss": 0.585, "step": 776 }, { - "epoch": 0.15, - "learning_rate": 1.9260408165526638e-05, - "loss": 0.9533, + "epoch": 0.29896113889957676, + "grad_norm": 0.9409414507089534, + "learning_rate": 1.6440065878024697e-05, + "loss": 0.5896, "step": 777 }, { - "epoch": 0.15, - "learning_rate": 1.9258054233834184e-05, - "loss": 0.9226, + "epoch": 0.2993459022701039, + "grad_norm": 1.0134126710391274, + "learning_rate": 1.643052743676019e-05, + "loss": 0.578, "step": 778 }, { - "epoch": 0.15, - "learning_rate": 1.9255696706420147e-05, - "loss": 1.0055, + "epoch": 0.299730665640631, + "grad_norm": 0.9446371543700374, + "learning_rate": 1.642097900928801e-05, + "loss": 0.6615, "step": 779 }, { - "epoch": 0.15, - "learning_rate": 1.9253335584200165e-05, - "loss": 0.9799, + "epoch": 0.3001154290111581, + "grad_norm": 1.1531757976792616, + "learning_rate": 1.641142061043627e-05, + "loss": 0.6385, "step": 780 }, { - "epoch": 0.15, - "learning_rate": 1.9250970868091268e-05, - "loss": 0.9956, + "epoch": 0.30050019238168524, + "grad_norm": 0.9469962516531095, + "learning_rate": 1.6401852255048564e-05, + "loss": 0.6295, "step": 781 }, { - "epoch": 0.15, - "learning_rate": 1.924860255901188e-05, - "loss": 0.9362, + "epoch": 0.3008849557522124, + "grad_norm": 0.9573993174631579, + "learning_rate": 1.6392273957983955e-05, + "loss": 0.6451, "step": 782 }, { - "epoch": 0.15, - "learning_rate": 1.9246230657881834e-05, - "loss": 0.9744, + "epoch": 0.30126971912273953, + "grad_norm": 0.9949326937521563, + "learning_rate": 1.6382685734116934e-05, + "loss": 0.5708, "step": 783 }, { - "epoch": 0.15, - "learning_rate": 1.9243855165622345e-05, - "loss": 0.9612, + "epoch": 0.30165448249326665, + "grad_norm": 0.9972441932158864, + "learning_rate": 1.637308759833742e-05, + "loss": 0.6079, "step": 784 }, { - "epoch": 0.15, - "learning_rate": 1.9241476083156026e-05, - "loss": 0.9608, + "epoch": 0.30203924586379377, + "grad_norm": 0.9784219240590314, + "learning_rate": 1.636347956555072e-05, + "loss": 0.6648, "step": 785 }, { - "epoch": 0.15, - "learning_rate": 1.9239093411406885e-05, - "loss": 1.0359, + "epoch": 0.3024240092343209, + "grad_norm": 1.0112980798801816, + "learning_rate": 1.63538616506775e-05, + "loss": 0.6616, "step": 786 }, { - "epoch": 0.15, - "learning_rate": 1.9236707151300326e-05, - "loss": 0.8541, + "epoch": 0.302808772604848, + "grad_norm": 1.0035380126656397, + "learning_rate": 1.634423386865379e-05, + "loss": 0.5929, "step": 787 }, { - "epoch": 0.15, - "learning_rate": 1.9234317303763145e-05, - "loss": 0.9458, + "epoch": 0.30319353597537513, + "grad_norm": 0.9699549767720443, + "learning_rate": 1.633459623443093e-05, + "loss": 0.6318, "step": 788 }, { - "epoch": 0.15, - "learning_rate": 1.9231923869723528e-05, - "loss": 0.7951, + "epoch": 0.30357829934590225, + "grad_norm": 1.0454703337594917, + "learning_rate": 1.6324948762975567e-05, + "loss": 0.5725, "step": 789 }, { - "epoch": 0.15, - "learning_rate": 1.922952685011106e-05, - "loss": 0.9639, + "epoch": 0.3039630627164294, + "grad_norm": 0.93040842125948, + "learning_rate": 1.6315291469269617e-05, + "loss": 0.602, "step": 790 }, { - "epoch": 0.15, - "learning_rate": 1.9227126245856716e-05, - "loss": 0.7542, + "epoch": 0.30434782608695654, + "grad_norm": 0.5915919398920796, + "learning_rate": 1.6305624368310265e-05, + "loss": 0.6288, "step": 791 }, { - "epoch": 0.15, - "learning_rate": 1.922472205789286e-05, - "loss": 1.0279, + "epoch": 0.30473258945748366, + "grad_norm": 0.9193495558095768, + "learning_rate": 1.6295947475109904e-05, + "loss": 0.6098, "step": 792 }, { - "epoch": 0.15, - "learning_rate": 1.9222314287153255e-05, - "loss": 0.748, + "epoch": 0.3051173528280108, + "grad_norm": 0.9902687069827169, + "learning_rate": 1.628626080469615e-05, + "loss": 0.6218, "step": 793 }, { - "epoch": 0.15, - "learning_rate": 1.9219902934573048e-05, - "loss": 0.9208, + "epoch": 0.3055021161985379, + "grad_norm": 0.9143548214174201, + "learning_rate": 1.6276564372111797e-05, + "loss": 0.5759, "step": 794 }, { - "epoch": 0.15, - "learning_rate": 1.9217488001088784e-05, - "loss": 0.9915, + "epoch": 0.305886879569065, + "grad_norm": 0.8727324357585786, + "learning_rate": 1.62668581924148e-05, + "loss": 0.5589, "step": 795 }, { - "epoch": 0.15, - "learning_rate": 1.9215069487638396e-05, - "loss": 1.0666, + "epoch": 0.30627164293959214, + "grad_norm": 1.0591031166662805, + "learning_rate": 1.6257142280678247e-05, + "loss": 0.652, "step": 796 }, { - "epoch": 0.15, - "learning_rate": 1.92126473951612e-05, - "loss": 0.9688, + "epoch": 0.30665640631011926, + "grad_norm": 1.0305953544375295, + "learning_rate": 1.6247416651990343e-05, + "loss": 0.6514, "step": 797 }, { - "epoch": 0.15, - "learning_rate": 1.921022172459791e-05, - "loss": 1.0448, + "epoch": 0.3070411696806464, + "grad_norm": 1.0042021134075554, + "learning_rate": 1.6237681321454387e-05, + "loss": 0.5932, "step": 798 }, { - "epoch": 0.15, - "learning_rate": 1.920779247689064e-05, - "loss": 0.9199, + "epoch": 0.30742593305117355, + "grad_norm": 1.0502461788278505, + "learning_rate": 1.6227936304188738e-05, + "loss": 0.6404, "step": 799 }, { - "epoch": 0.15, - "learning_rate": 1.9205359652982868e-05, - "loss": 1.0, + "epoch": 0.30781069642170067, + "grad_norm": 0.9820795654855251, + "learning_rate": 1.6218181615326795e-05, + "loss": 0.5873, "step": 800 }, { - "epoch": 0.15, - "learning_rate": 1.9202923253819482e-05, - "loss": 0.9216, + "epoch": 0.3081954597922278, + "grad_norm": 1.0176314240540965, + "learning_rate": 1.620841727001699e-05, + "loss": 0.6701, "step": 801 }, { - "epoch": 0.15, - "learning_rate": 1.920048328034675e-05, - "loss": 0.8961, + "epoch": 0.3085802231627549, + "grad_norm": 0.9285135052576255, + "learning_rate": 1.619864328342273e-05, + "loss": 0.5945, "step": 802 }, { - "epoch": 0.15, - "learning_rate": 1.9198039733512326e-05, - "loss": 0.9419, + "epoch": 0.308964986533282, + "grad_norm": 0.9785426667664, + "learning_rate": 1.6188859670722414e-05, + "loss": 0.5563, "step": 803 }, { - "epoch": 0.15, - "learning_rate": 1.9195592614265262e-05, - "loss": 0.9791, + "epoch": 0.30934974990380915, + "grad_norm": 1.0493274535202173, + "learning_rate": 1.6179066447109376e-05, + "loss": 0.6865, "step": 804 }, { - "epoch": 0.15, - "learning_rate": 1.9193141923555984e-05, - "loss": 0.9181, + "epoch": 0.30973451327433627, + "grad_norm": 1.0442621845193958, + "learning_rate": 1.6169263627791886e-05, + "loss": 0.6401, "step": 805 }, { - "epoch": 0.16, - "learning_rate": 1.919068766233632e-05, - "loss": 0.8486, + "epoch": 0.3101192766448634, + "grad_norm": 1.0650207030891188, + "learning_rate": 1.615945122799311e-05, + "loss": 0.6404, "step": 806 }, { - "epoch": 0.16, - "learning_rate": 1.9188229831559468e-05, - "loss": 1.0046, + "epoch": 0.31050404001539056, + "grad_norm": 0.6669626602020727, + "learning_rate": 1.614962926295109e-05, + "loss": 0.665, "step": 807 }, { - "epoch": 0.16, - "learning_rate": 1.9185768432180026e-05, - "loss": 0.9634, + "epoch": 0.3108888033859177, + "grad_norm": 0.9929774948499741, + "learning_rate": 1.6139797747918725e-05, + "loss": 0.5937, "step": 808 }, { - "epoch": 0.16, - "learning_rate": 1.9183303465153972e-05, - "loss": 0.9661, + "epoch": 0.3112735667564448, + "grad_norm": 1.0981718735108787, + "learning_rate": 1.612995669816375e-05, + "loss": 0.617, "step": 809 }, { - "epoch": 0.16, - "learning_rate": 1.9180834931438673e-05, - "loss": 0.9419, + "epoch": 0.3116583301269719, + "grad_norm": 0.9584983921222681, + "learning_rate": 1.6120106128968686e-05, + "loss": 0.5569, "step": 810 }, { - "epoch": 0.16, - "learning_rate": 1.917836283199288e-05, - "loss": 0.9527, + "epoch": 0.31204309349749904, + "grad_norm": 0.9008785152190121, + "learning_rate": 1.611024605563087e-05, + "loss": 0.6142, "step": 811 }, { - "epoch": 0.16, - "learning_rate": 1.917588716777672e-05, - "loss": 0.9701, + "epoch": 0.31242785686802615, + "grad_norm": 0.9593449271147845, + "learning_rate": 1.6100376493462368e-05, + "loss": 0.6076, "step": 812 }, { - "epoch": 0.16, - "learning_rate": 1.917340793975172e-05, - "loss": 0.9067, + "epoch": 0.3128126202385533, + "grad_norm": 1.1676193977213043, + "learning_rate": 1.609049745779e-05, + "loss": 0.668, "step": 813 }, { - "epoch": 0.16, - "learning_rate": 1.917092514888078e-05, - "loss": 0.9913, + "epoch": 0.3131973836090804, + "grad_norm": 0.9097182180054575, + "learning_rate": 1.608060896395529e-05, + "loss": 0.5815, "step": 814 }, { - "epoch": 0.16, - "learning_rate": 1.9168438796128193e-05, - "loss": 0.9805, + "epoch": 0.31358214697960757, + "grad_norm": 0.9154894490751722, + "learning_rate": 1.6070711027314446e-05, + "loss": 0.5755, "step": 815 }, { - "epoch": 0.16, - "learning_rate": 1.9165948882459623e-05, - "loss": 0.9509, + "epoch": 0.3139669103501347, + "grad_norm": 0.9041528241743848, + "learning_rate": 1.6060803663238357e-05, + "loss": 0.5545, "step": 816 }, { - "epoch": 0.16, - "learning_rate": 1.9163455408842123e-05, - "loss": 1.0248, + "epoch": 0.3143516737206618, + "grad_norm": 0.9491407364524953, + "learning_rate": 1.6050886887112535e-05, + "loss": 0.5911, "step": 817 }, { - "epoch": 0.16, - "learning_rate": 1.9160958376244138e-05, - "loss": 0.9554, + "epoch": 0.3147364370911889, + "grad_norm": 0.962261508100928, + "learning_rate": 1.604096071433711e-05, + "loss": 0.5429, "step": 818 }, { - "epoch": 0.16, - "learning_rate": 1.9158457785635478e-05, - "loss": 1.0009, + "epoch": 0.31512120046171604, + "grad_norm": 1.0547180027124488, + "learning_rate": 1.6031025160326814e-05, + "loss": 0.5771, "step": 819 }, { - "epoch": 0.16, - "learning_rate": 1.915595363798735e-05, - "loss": 0.9396, + "epoch": 0.31550596383224316, + "grad_norm": 1.0688917708592065, + "learning_rate": 1.6021080240510943e-05, + "loss": 0.592, "step": 820 }, { - "epoch": 0.16, - "learning_rate": 1.915344593427233e-05, - "loss": 0.9583, + "epoch": 0.3158907272027703, + "grad_norm": 1.0381482546519782, + "learning_rate": 1.6011125970333333e-05, + "loss": 0.5684, "step": 821 }, { - "epoch": 0.16, - "learning_rate": 1.9150934675464384e-05, - "loss": 0.8924, + "epoch": 0.3162754905732974, + "grad_norm": 0.9962630635863459, + "learning_rate": 1.6001162365252348e-05, + "loss": 0.5978, "step": 822 }, { - "epoch": 0.16, - "learning_rate": 1.9148419862538858e-05, - "loss": 0.9677, + "epoch": 0.3166602539438245, + "grad_norm": 0.9737127522021747, + "learning_rate": 1.5991189440740838e-05, + "loss": 0.5754, "step": 823 }, { - "epoch": 0.16, - "learning_rate": 1.9145901496472474e-05, - "loss": 1.058, + "epoch": 0.3170450173143517, + "grad_norm": 0.981655239875509, + "learning_rate": 1.598120721228614e-05, + "loss": 0.6248, "step": 824 }, { - "epoch": 0.16, - "learning_rate": 1.9143379578243335e-05, - "loss": 0.9586, + "epoch": 0.3174297806848788, + "grad_norm": 0.9377497729787908, + "learning_rate": 1.5971215695390026e-05, + "loss": 0.5864, "step": 825 }, { - "epoch": 0.16, - "learning_rate": 1.914085410883093e-05, - "loss": 1.0024, + "epoch": 0.31781454405540593, + "grad_norm": 1.1347391945408236, + "learning_rate": 1.5961214905568705e-05, + "loss": 0.622, "step": 826 }, { - "epoch": 0.16, - "learning_rate": 1.9138325089216118e-05, - "loss": 0.9468, + "epoch": 0.31819930742593305, + "grad_norm": 1.113696255854578, + "learning_rate": 1.5951204858352772e-05, + "loss": 0.5864, "step": 827 }, { - "epoch": 0.16, - "learning_rate": 1.913579252038114e-05, - "loss": 0.9579, + "epoch": 0.3185840707964602, + "grad_norm": 1.084378180825176, + "learning_rate": 1.5941185569287206e-05, + "loss": 0.6122, "step": 828 }, { - "epoch": 0.16, - "learning_rate": 1.9133256403309627e-05, - "loss": 0.9533, + "epoch": 0.3189688341669873, + "grad_norm": 1.0343481695607295, + "learning_rate": 1.593115705393134e-05, + "loss": 0.5684, "step": 829 }, { - "epoch": 0.16, - "learning_rate": 1.913071673898656e-05, - "loss": 0.9867, + "epoch": 0.3193535975375144, + "grad_norm": 1.0943480043596794, + "learning_rate": 1.5921119327858835e-05, + "loss": 0.6341, "step": 830 }, { - "epoch": 0.16, - "learning_rate": 1.912817352839833e-05, - "loss": 0.8882, + "epoch": 0.31973836090804153, + "grad_norm": 0.9763417385443496, + "learning_rate": 1.5911072406657646e-05, + "loss": 0.6212, "step": 831 }, { - "epoch": 0.16, - "learning_rate": 1.9125626772532683e-05, - "loss": 0.8958, + "epoch": 0.3201231242785687, + "grad_norm": 1.0814754824820747, + "learning_rate": 1.590101630593002e-05, + "loss": 0.6209, "step": 832 }, { - "epoch": 0.16, - "learning_rate": 1.9123076472378753e-05, - "loss": 1.0005, + "epoch": 0.3205078876490958, + "grad_norm": 0.9291199576061675, + "learning_rate": 1.5890951041292453e-05, + "loss": 0.5929, "step": 833 }, { - "epoch": 0.16, - "learning_rate": 1.9120522628927047e-05, - "loss": 1.0583, + "epoch": 0.32089265101962294, + "grad_norm": 1.015675201559871, + "learning_rate": 1.5880876628375668e-05, + "loss": 0.6236, "step": 834 }, { - "epoch": 0.16, - "learning_rate": 1.9117965243169444e-05, - "loss": 0.9431, + "epoch": 0.32127741439015006, + "grad_norm": 1.0871351834089318, + "learning_rate": 1.5870793082824604e-05, + "loss": 0.6133, "step": 835 }, { - "epoch": 0.16, - "learning_rate": 1.9115404316099212e-05, - "loss": 0.9788, + "epoch": 0.3216621777606772, + "grad_norm": 0.9269841362913653, + "learning_rate": 1.5860700420298377e-05, + "loss": 0.5512, "step": 836 }, { - "epoch": 0.16, - "learning_rate": 1.9112839848710978e-05, - "loss": 0.9565, + "epoch": 0.3220469411312043, + "grad_norm": 0.9710739810230657, + "learning_rate": 1.5850598656470265e-05, + "loss": 0.6388, "step": 837 }, { - "epoch": 0.16, - "learning_rate": 1.9110271842000755e-05, - "loss": 0.7492, + "epoch": 0.3224317045017314, + "grad_norm": 0.9635171621117674, + "learning_rate": 1.5840487807027665e-05, + "loss": 0.54, "step": 838 }, { - "epoch": 0.16, - "learning_rate": 1.9107700296965926e-05, - "loss": 1.0554, + "epoch": 0.32281646787225854, + "grad_norm": 0.9221325212680264, + "learning_rate": 1.583036788767211e-05, + "loss": 0.6232, "step": 839 }, { - "epoch": 0.16, - "learning_rate": 1.910512521460525e-05, - "loss": 0.9893, + "epoch": 0.3232012312427857, + "grad_norm": 1.0697557536057518, + "learning_rate": 1.5820238914119195e-05, + "loss": 0.6215, "step": 840 }, { - "epoch": 0.16, - "learning_rate": 1.9102546595918857e-05, - "loss": 1.0504, + "epoch": 0.32358599461331283, + "grad_norm": 0.7111617324205604, + "learning_rate": 1.5810100902098582e-05, + "loss": 0.6611, "step": 841 }, { - "epoch": 0.16, - "learning_rate": 1.9099964441908258e-05, - "loss": 0.975, + "epoch": 0.32397075798383995, + "grad_norm": 0.9971040363466133, + "learning_rate": 1.5799953867353975e-05, + "loss": 0.5747, "step": 842 }, { - "epoch": 0.16, - "learning_rate": 1.9097378753576327e-05, - "loss": 0.9602, + "epoch": 0.32435552135436707, + "grad_norm": 0.9278042661540365, + "learning_rate": 1.5789797825643086e-05, + "loss": 0.619, "step": 843 }, { - "epoch": 0.16, - "learning_rate": 1.9094789531927315e-05, - "loss": 0.9739, + "epoch": 0.3247402847248942, + "grad_norm": 0.961591143589985, + "learning_rate": 1.5779632792737608e-05, + "loss": 0.5746, "step": 844 }, { - "epoch": 0.16, - "learning_rate": 1.909219677796685e-05, - "loss": 0.9925, + "epoch": 0.3251250480954213, + "grad_norm": 0.9475929642952069, + "learning_rate": 1.5769458784423206e-05, + "loss": 0.5414, "step": 845 }, { - "epoch": 0.16, - "learning_rate": 1.9089600492701926e-05, - "loss": 0.7035, + "epoch": 0.32550981146594843, + "grad_norm": 0.9931584627917872, + "learning_rate": 1.575927581649948e-05, + "loss": 0.6108, "step": 846 }, { - "epoch": 0.16, - "learning_rate": 1.908700067714091e-05, - "loss": 0.753, + "epoch": 0.32589457483647555, + "grad_norm": 1.0610444658426947, + "learning_rate": 1.574908390477995e-05, + "loss": 0.6004, "step": 847 }, { - "epoch": 0.16, - "learning_rate": 1.9084397332293537e-05, - "loss": 0.984, + "epoch": 0.32627933820700267, + "grad_norm": 1.0663506153560152, + "learning_rate": 1.5738883065092005e-05, + "loss": 0.6418, "step": 848 }, { - "epoch": 0.16, - "learning_rate": 1.9081790459170926e-05, - "loss": 1.0143, + "epoch": 0.32666410157752984, + "grad_norm": 1.0396638440086352, + "learning_rate": 1.572867331327692e-05, + "loss": 0.6227, "step": 849 }, { - "epoch": 0.16, - "learning_rate": 1.9079180058785547e-05, - "loss": 0.9413, + "epoch": 0.32704886494805696, + "grad_norm": 0.9372811228344775, + "learning_rate": 1.5718454665189806e-05, + "loss": 0.6133, "step": 850 }, { - "epoch": 0.16, - "learning_rate": 1.9076566132151255e-05, - "loss": 0.9701, + "epoch": 0.3274336283185841, + "grad_norm": 1.0799318365655641, + "learning_rate": 1.5708227136699578e-05, + "loss": 0.6133, "step": 851 }, { - "epoch": 0.16, - "learning_rate": 1.907394868028326e-05, - "loss": 1.0083, + "epoch": 0.3278183916891112, + "grad_norm": 1.0853244667341202, + "learning_rate": 1.569799074368895e-05, + "loss": 0.6253, "step": 852 }, { - "epoch": 0.16, - "learning_rate": 1.9071327704198163e-05, - "loss": 0.9898, + "epoch": 0.3282031550596383, + "grad_norm": 0.907859410365573, + "learning_rate": 1.5687745502054407e-05, + "loss": 0.5939, "step": 853 }, { - "epoch": 0.16, - "learning_rate": 1.906870320491391e-05, - "loss": 0.8673, + "epoch": 0.32858791843016544, + "grad_norm": 1.0468020557569218, + "learning_rate": 1.567749142770617e-05, + "loss": 0.5926, "step": 854 }, { - "epoch": 0.16, - "learning_rate": 1.9066075183449835e-05, - "loss": 0.9488, + "epoch": 0.32897268180069256, + "grad_norm": 1.0010865359135492, + "learning_rate": 1.5667228536568167e-05, + "loss": 0.5619, "step": 855 }, { - "epoch": 0.16, - "learning_rate": 1.9063443640826624e-05, - "loss": 0.9469, + "epoch": 0.3293574451712197, + "grad_norm": 1.0044117271579711, + "learning_rate": 1.565695684457803e-05, + "loss": 0.5997, "step": 856 }, { - "epoch": 0.16, - "learning_rate": 1.906080857806634e-05, - "loss": 0.9875, + "epoch": 0.32974220854174685, + "grad_norm": 0.9387826320206298, + "learning_rate": 1.5646676367687067e-05, + "loss": 0.6017, "step": 857 }, { - "epoch": 0.17, - "learning_rate": 1.905816999619242e-05, - "loss": 0.9586, + "epoch": 0.33012697191227397, + "grad_norm": 1.0948355050124836, + "learning_rate": 1.5636387121860207e-05, + "loss": 0.6027, "step": 858 }, { - "epoch": 0.17, - "learning_rate": 1.9055527896229642e-05, - "loss": 0.9674, + "epoch": 0.3305117352828011, + "grad_norm": 1.043592146247732, + "learning_rate": 1.5626089123076004e-05, + "loss": 0.5928, "step": 859 }, { - "epoch": 0.17, - "learning_rate": 1.905288227920418e-05, - "loss": 0.9564, + "epoch": 0.3308964986533282, + "grad_norm": 1.0012366717991514, + "learning_rate": 1.561578238732661e-05, + "loss": 0.5814, "step": 860 }, { - "epoch": 0.17, - "learning_rate": 1.9050233146143554e-05, - "loss": 0.9876, + "epoch": 0.3312812620238553, + "grad_norm": 1.0046208461041362, + "learning_rate": 1.5605466930617747e-05, + "loss": 0.6147, "step": 861 }, { - "epoch": 0.17, - "learning_rate": 1.9047580498076663e-05, - "loss": 1.0343, + "epoch": 0.33166602539438245, + "grad_norm": 1.1057108505898057, + "learning_rate": 1.559514276896867e-05, + "loss": 0.6029, "step": 862 }, { - "epoch": 0.17, - "learning_rate": 1.904492433603376e-05, - "loss": 0.8694, + "epoch": 0.33205078876490957, + "grad_norm": 1.058460058916235, + "learning_rate": 1.5584809918412158e-05, + "loss": 0.5967, "step": 863 }, { - "epoch": 0.17, - "learning_rate": 1.904226466104647e-05, - "loss": 1.0036, + "epoch": 0.3324355521354367, + "grad_norm": 0.9927048397459389, + "learning_rate": 1.5574468394994486e-05, + "loss": 0.6148, "step": 864 }, { - "epoch": 0.17, - "learning_rate": 1.903960147414778e-05, - "loss": 0.9593, + "epoch": 0.33282031550596386, + "grad_norm": 0.916514528123093, + "learning_rate": 1.556411821477539e-05, + "loss": 0.5748, "step": 865 }, { - "epoch": 0.17, - "learning_rate": 1.903693477637204e-05, - "loss": 0.7492, + "epoch": 0.333205078876491, + "grad_norm": 1.1068961933552428, + "learning_rate": 1.5553759393828058e-05, + "loss": 0.5853, "step": 866 }, { - "epoch": 0.17, - "learning_rate": 1.9034264568754967e-05, - "loss": 1.0062, + "epoch": 0.3335898422470181, + "grad_norm": 0.6788609902238475, + "learning_rate": 1.554339194823909e-05, + "loss": 0.656, "step": 867 }, { - "epoch": 0.17, - "learning_rate": 1.9031590852333637e-05, - "loss": 1.0154, + "epoch": 0.3339746056175452, + "grad_norm": 1.0004657492399842, + "learning_rate": 1.553301589410848e-05, + "loss": 0.6286, "step": 868 }, { - "epoch": 0.17, - "learning_rate": 1.9028913628146487e-05, - "loss": 0.9397, + "epoch": 0.33435936898807234, + "grad_norm": 0.9498152357796671, + "learning_rate": 1.5522631247549598e-05, + "loss": 0.5764, "step": 869 }, { - "epoch": 0.17, - "learning_rate": 1.902623289723333e-05, - "loss": 0.9665, + "epoch": 0.33474413235859946, + "grad_norm": 0.957333161527608, + "learning_rate": 1.5512238024689144e-05, + "loss": 0.5744, "step": 870 }, { - "epoch": 0.17, - "learning_rate": 1.902354866063532e-05, - "loss": 0.9443, + "epoch": 0.3351288957291266, + "grad_norm": 0.6354233342991102, + "learning_rate": 1.550183624166715e-05, + "loss": 0.6587, "step": 871 }, { - "epoch": 0.17, - "learning_rate": 1.9020860919394992e-05, - "loss": 0.9199, + "epoch": 0.3355136590996537, + "grad_norm": 0.9762028424441871, + "learning_rate": 1.5491425914636934e-05, + "loss": 0.5909, "step": 872 }, { - "epoch": 0.17, - "learning_rate": 1.9018169674556228e-05, - "loss": 0.9523, + "epoch": 0.3358984224701808, + "grad_norm": 1.0208324687826442, + "learning_rate": 1.548100705976508e-05, + "loss": 0.5969, "step": 873 }, { - "epoch": 0.17, - "learning_rate": 1.901547492716428e-05, - "loss": 0.9504, + "epoch": 0.336283185840708, + "grad_norm": 0.995973043303851, + "learning_rate": 1.5470579693231432e-05, + "loss": 0.5531, "step": 874 }, { - "epoch": 0.17, - "learning_rate": 1.9012776678265756e-05, - "loss": 0.9917, + "epoch": 0.3366679492112351, + "grad_norm": 1.0157166616035114, + "learning_rate": 1.5460143831229026e-05, + "loss": 0.5921, "step": 875 }, { - "epoch": 0.17, - "learning_rate": 1.9010074928908624e-05, - "loss": 0.938, + "epoch": 0.3370527125817622, + "grad_norm": 0.8959715440898841, + "learning_rate": 1.544969948996411e-05, + "loss": 0.5692, "step": 876 }, { - "epoch": 0.17, - "learning_rate": 1.900736968014221e-05, - "loss": 0.9402, + "epoch": 0.33743747595228935, + "grad_norm": 0.8551388949610362, + "learning_rate": 1.5439246685656093e-05, + "loss": 0.5133, "step": 877 }, { - "epoch": 0.17, - "learning_rate": 1.9004660933017208e-05, - "loss": 0.9622, + "epoch": 0.33782223932281646, + "grad_norm": 0.9437526047269974, + "learning_rate": 1.5428785434537527e-05, + "loss": 0.5493, "step": 878 }, { - "epoch": 0.17, - "learning_rate": 1.900194868858566e-05, - "loss": 0.9642, + "epoch": 0.3382070026933436, + "grad_norm": 0.9043779531701643, + "learning_rate": 1.541831575285408e-05, + "loss": 0.5944, "step": 879 }, { - "epoch": 0.17, - "learning_rate": 1.8999232947900968e-05, - "loss": 0.9374, + "epoch": 0.3385917660638707, + "grad_norm": 1.018783818531709, + "learning_rate": 1.540783765686452e-05, + "loss": 0.6566, "step": 880 }, { - "epoch": 0.17, - "learning_rate": 1.89965137120179e-05, - "loss": 0.7193, + "epoch": 0.3389765294343978, + "grad_norm": 0.9824784811856958, + "learning_rate": 1.539735116284067e-05, + "loss": 0.5986, "step": 881 }, { - "epoch": 0.17, - "learning_rate": 1.899379098199257e-05, - "loss": 0.8862, + "epoch": 0.339361292804925, + "grad_norm": 0.9022703912559521, + "learning_rate": 1.53868562870674e-05, + "loss": 0.5392, "step": 882 }, { - "epoch": 0.17, - "learning_rate": 1.899106475888246e-05, - "loss": 0.9945, + "epoch": 0.3397460561754521, + "grad_norm": 0.892089543805682, + "learning_rate": 1.5376353045842604e-05, + "loss": 0.5533, "step": 883 }, { - "epoch": 0.17, - "learning_rate": 1.89883350437464e-05, - "loss": 1.0043, + "epoch": 0.34013081954597923, + "grad_norm": 0.6937753980386002, + "learning_rate": 1.5365841455477158e-05, + "loss": 0.6971, "step": 884 }, { - "epoch": 0.17, - "learning_rate": 1.8985601837644586e-05, - "loss": 1.0377, + "epoch": 0.34051558291650635, + "grad_norm": 1.0708103812345482, + "learning_rate": 1.5355321532294897e-05, + "loss": 0.6255, "step": 885 }, { - "epoch": 0.17, - "learning_rate": 1.8982865141638557e-05, - "loss": 1.0135, + "epoch": 0.3409003462870335, + "grad_norm": 1.1474987276676292, + "learning_rate": 1.5344793292632614e-05, + "loss": 0.6217, "step": 886 }, { - "epoch": 0.17, - "learning_rate": 1.8980124956791216e-05, - "loss": 0.9239, + "epoch": 0.3412851096575606, + "grad_norm": 0.9394146070890443, + "learning_rate": 1.5334256752840007e-05, + "loss": 0.6201, "step": 887 }, { - "epoch": 0.17, - "learning_rate": 1.8977381284166818e-05, - "loss": 1.0181, + "epoch": 0.3416698730280877, + "grad_norm": 0.9440155626631186, + "learning_rate": 1.532371192927966e-05, + "loss": 0.6109, "step": 888 }, { - "epoch": 0.17, - "learning_rate": 1.897463412483098e-05, - "loss": 0.9065, + "epoch": 0.34205463639861483, + "grad_norm": 1.02101518701545, + "learning_rate": 1.531315883832703e-05, + "loss": 0.6392, "step": 889 }, { - "epoch": 0.17, - "learning_rate": 1.897188347985066e-05, - "loss": 0.9762, + "epoch": 0.34243939976914195, + "grad_norm": 0.9775825696133279, + "learning_rate": 1.5302597496370408e-05, + "loss": 0.6493, "step": 890 }, { - "epoch": 0.17, - "learning_rate": 1.896912935029418e-05, - "loss": 0.9245, + "epoch": 0.3428241631396691, + "grad_norm": 0.9478440143898313, + "learning_rate": 1.5292027919810898e-05, + "loss": 0.5765, "step": 891 }, { - "epoch": 0.17, - "learning_rate": 1.896637173723121e-05, - "loss": 0.9991, + "epoch": 0.34320892651019624, + "grad_norm": 0.9225968306293865, + "learning_rate": 1.528145012506239e-05, + "loss": 0.5944, "step": 892 }, { - "epoch": 0.17, - "learning_rate": 1.8963610641732777e-05, - "loss": 1.0052, + "epoch": 0.34359368988072336, + "grad_norm": 1.006320537555887, + "learning_rate": 1.5270864128551542e-05, + "loss": 0.6129, "step": 893 }, { - "epoch": 0.17, - "learning_rate": 1.8960846064871257e-05, - "loss": 0.8797, + "epoch": 0.3439784532512505, + "grad_norm": 1.0230426759505362, + "learning_rate": 1.5260269946717746e-05, + "loss": 0.5951, "step": 894 }, { - "epoch": 0.17, - "learning_rate": 1.8958078007720387e-05, - "loss": 0.9198, + "epoch": 0.3443632166217776, + "grad_norm": 0.9837287543792391, + "learning_rate": 1.5249667596013102e-05, + "loss": 0.5782, "step": 895 }, { - "epoch": 0.17, - "learning_rate": 1.895530647135524e-05, - "loss": 0.8892, + "epoch": 0.3447479799923047, + "grad_norm": 0.6112341721664591, + "learning_rate": 1.5239057092902404e-05, + "loss": 0.6314, "step": 896 }, { - "epoch": 0.17, - "learning_rate": 1.8952531456852248e-05, - "loss": 0.716, + "epoch": 0.34513274336283184, + "grad_norm": 1.0414645027231695, + "learning_rate": 1.5228438453863095e-05, + "loss": 0.627, "step": 897 }, { - "epoch": 0.17, - "learning_rate": 1.8949752965289197e-05, - "loss": 0.9493, + "epoch": 0.34551750673335896, + "grad_norm": 1.0536722603976394, + "learning_rate": 1.5217811695385263e-05, + "loss": 0.6124, "step": 898 }, { - "epoch": 0.17, - "learning_rate": 1.894697099774523e-05, - "loss": 0.8922, + "epoch": 0.34590227010388613, + "grad_norm": 0.6149655993489004, + "learning_rate": 1.5207176833971598e-05, + "loss": 0.6437, "step": 899 }, { - "epoch": 0.17, - "learning_rate": 1.894418555530082e-05, - "loss": 0.9936, + "epoch": 0.34628703347441325, + "grad_norm": 1.0549746310391472, + "learning_rate": 1.5196533886137376e-05, + "loss": 0.6086, "step": 900 }, { - "epoch": 0.17, - "learning_rate": 1.89413966390378e-05, - "loss": 0.9837, + "epoch": 0.34667179684494037, + "grad_norm": 1.0471530227788077, + "learning_rate": 1.5185882868410431e-05, + "loss": 0.6414, "step": 901 }, { - "epoch": 0.17, - "learning_rate": 1.8938604250039362e-05, - "loss": 0.9778, + "epoch": 0.3470565602154675, + "grad_norm": 0.9723233233707275, + "learning_rate": 1.517522379733113e-05, + "loss": 0.6034, "step": 902 }, { - "epoch": 0.17, - "learning_rate": 1.8935808389390032e-05, - "loss": 0.9047, + "epoch": 0.3474413235859946, + "grad_norm": 0.9378513268576659, + "learning_rate": 1.5164556689452346e-05, + "loss": 0.5431, "step": 903 }, { - "epoch": 0.17, - "learning_rate": 1.893300905817569e-05, - "loss": 0.96, + "epoch": 0.34782608695652173, + "grad_norm": 0.9735475750352324, + "learning_rate": 1.5153881561339426e-05, + "loss": 0.594, "step": 904 }, { - "epoch": 0.17, - "learning_rate": 1.8930206257483566e-05, - "loss": 1.0163, + "epoch": 0.34821085032704885, + "grad_norm": 0.9168087445612595, + "learning_rate": 1.5143198429570181e-05, + "loss": 0.5646, "step": 905 }, { - "epoch": 0.17, - "learning_rate": 1.8927399988402233e-05, - "loss": 0.9936, + "epoch": 0.34859561369757597, + "grad_norm": 0.7686208429813068, + "learning_rate": 1.5132507310734847e-05, + "loss": 0.6591, "step": 906 }, { - "epoch": 0.17, - "learning_rate": 1.8924590252021614e-05, - "loss": 0.9158, + "epoch": 0.34898037706810314, + "grad_norm": 1.0935405034624377, + "learning_rate": 1.512180822143607e-05, + "loss": 0.589, "step": 907 }, { - "epoch": 0.17, - "learning_rate": 1.8921777049432985e-05, - "loss": 0.8078, + "epoch": 0.34936514043863026, + "grad_norm": 1.0173631986780833, + "learning_rate": 1.5111101178288858e-05, + "loss": 0.6234, "step": 908 }, { - "epoch": 0.17, - "learning_rate": 1.8918960381728947e-05, - "loss": 1.0311, + "epoch": 0.3497499038091574, + "grad_norm": 1.0498657092541606, + "learning_rate": 1.5100386197920585e-05, + "loss": 0.6297, "step": 909 }, { - "epoch": 0.18, - "learning_rate": 1.8916140250003475e-05, - "loss": 0.9098, + "epoch": 0.3501346671796845, + "grad_norm": 0.6352069702446178, + "learning_rate": 1.5089663296970952e-05, + "loss": 0.6765, "step": 910 }, { - "epoch": 0.18, - "learning_rate": 1.891331665535187e-05, - "loss": 0.9543, + "epoch": 0.3505194305502116, + "grad_norm": 1.0830739359082122, + "learning_rate": 1.5078932492091942e-05, + "loss": 0.6151, "step": 911 }, { - "epoch": 0.18, - "learning_rate": 1.8910489598870784e-05, - "loss": 0.9964, + "epoch": 0.35090419392073874, + "grad_norm": 0.8948284234307127, + "learning_rate": 1.506819379994784e-05, + "loss": 0.5692, "step": 912 }, { - "epoch": 0.18, - "learning_rate": 1.8907659081658214e-05, - "loss": 0.9889, + "epoch": 0.35128895729126586, + "grad_norm": 0.9846336075891183, + "learning_rate": 1.5057447237215152e-05, + "loss": 0.5905, "step": 913 }, { - "epoch": 0.18, - "learning_rate": 1.8904825104813497e-05, - "loss": 0.9177, + "epoch": 0.351673720661793, + "grad_norm": 1.0602149436311796, + "learning_rate": 1.5046692820582625e-05, + "loss": 0.595, "step": 914 }, { - "epoch": 0.18, - "learning_rate": 1.8901987669437322e-05, - "loss": 1.0134, + "epoch": 0.3520584840323201, + "grad_norm": 0.9290629084305116, + "learning_rate": 1.5035930566751198e-05, + "loss": 0.6041, "step": 915 }, { - "epoch": 0.18, - "learning_rate": 1.889914677663171e-05, - "loss": 0.9554, + "epoch": 0.35244324740284727, + "grad_norm": 0.9645537729715606, + "learning_rate": 1.5025160492433976e-05, + "loss": 0.565, "step": 916 }, { - "epoch": 0.18, - "learning_rate": 1.8896302427500042e-05, - "loss": 0.9336, + "epoch": 0.3528280107733744, + "grad_norm": 0.9815820654283561, + "learning_rate": 1.5014382614356213e-05, + "loss": 0.5828, "step": 917 }, { - "epoch": 0.18, - "learning_rate": 1.8893454623147017e-05, - "loss": 0.9923, + "epoch": 0.3532127741439015, + "grad_norm": 0.9273363918482245, + "learning_rate": 1.5003596949255284e-05, + "loss": 0.5634, "step": 918 }, { - "epoch": 0.18, - "learning_rate": 1.88906033646787e-05, - "loss": 0.929, + "epoch": 0.35359753751442863, + "grad_norm": 0.9521534314184353, + "learning_rate": 1.499280351388065e-05, + "loss": 0.6256, "step": 919 }, { - "epoch": 0.18, - "learning_rate": 1.8887748653202478e-05, - "loss": 0.9428, + "epoch": 0.35398230088495575, + "grad_norm": 0.9142695224858879, + "learning_rate": 1.498200232499384e-05, + "loss": 0.5749, "step": 920 }, { - "epoch": 0.18, - "learning_rate": 1.8884890489827097e-05, - "loss": 0.8452, + "epoch": 0.35436706425548287, + "grad_norm": 0.9465260356269594, + "learning_rate": 1.497119339936843e-05, + "loss": 0.6219, "step": 921 }, { - "epoch": 0.18, - "learning_rate": 1.8882028875662627e-05, - "loss": 0.8809, + "epoch": 0.35475182762601, + "grad_norm": 0.6879415953811905, + "learning_rate": 1.496037675379001e-05, + "loss": 0.665, "step": 922 }, { - "epoch": 0.18, - "learning_rate": 1.8879163811820493e-05, - "loss": 0.9353, + "epoch": 0.3551365909965371, + "grad_norm": 0.9901681095089465, + "learning_rate": 1.494955240505615e-05, + "loss": 0.5904, "step": 923 }, { - "epoch": 0.18, - "learning_rate": 1.8876295299413445e-05, - "loss": 0.8892, + "epoch": 0.3555213543670643, + "grad_norm": 1.037291403883846, + "learning_rate": 1.4938720369976385e-05, + "loss": 0.6006, "step": 924 }, { - "epoch": 0.18, - "learning_rate": 1.8873423339555584e-05, - "loss": 1.0012, + "epoch": 0.3559061177375914, + "grad_norm": 0.9507420964284085, + "learning_rate": 1.4927880665372197e-05, + "loss": 0.5966, "step": 925 }, { - "epoch": 0.18, - "learning_rate": 1.8870547933362352e-05, - "loss": 0.7537, + "epoch": 0.3562908811081185, + "grad_norm": 1.090298027806448, + "learning_rate": 1.4917033308076967e-05, + "loss": 0.5992, "step": 926 }, { - "epoch": 0.18, - "learning_rate": 1.886766908195051e-05, - "loss": 0.9761, + "epoch": 0.35667564447864564, + "grad_norm": 1.0104390674705162, + "learning_rate": 1.490617831493596e-05, + "loss": 0.607, "step": 927 }, { - "epoch": 0.18, - "learning_rate": 1.8864786786438187e-05, - "loss": 0.9488, + "epoch": 0.35706040784917276, + "grad_norm": 0.625339608014549, + "learning_rate": 1.489531570280631e-05, + "loss": 0.6321, "step": 928 }, { - "epoch": 0.18, - "learning_rate": 1.886190104794482e-05, - "loss": 1.0093, + "epoch": 0.3574451712196999, + "grad_norm": 0.983985835279662, + "learning_rate": 1.4884445488556972e-05, + "loss": 0.5923, "step": 929 }, { - "epoch": 0.18, - "learning_rate": 1.8859011867591203e-05, - "loss": 0.9488, + "epoch": 0.357829934590227, + "grad_norm": 1.1007749106901559, + "learning_rate": 1.4873567689068708e-05, + "loss": 0.6228, "step": 930 }, { - "epoch": 0.18, - "learning_rate": 1.885611924649946e-05, - "loss": 0.9453, + "epoch": 0.3582146979607541, + "grad_norm": 0.973235874776153, + "learning_rate": 1.4862682321234064e-05, + "loss": 0.5678, "step": 931 }, { - "epoch": 0.18, - "learning_rate": 1.885322318579305e-05, - "loss": 0.9999, + "epoch": 0.3585994613312813, + "grad_norm": 0.925023024477942, + "learning_rate": 1.4851789401957338e-05, + "loss": 0.5803, "step": 932 }, { - "epoch": 0.18, - "learning_rate": 1.8850323686596766e-05, - "loss": 0.9939, + "epoch": 0.3589842247018084, + "grad_norm": 1.0351687094626274, + "learning_rate": 1.484088894815455e-05, + "loss": 0.6064, "step": 933 }, { - "epoch": 0.18, - "learning_rate": 1.8847420750036748e-05, - "loss": 0.8531, + "epoch": 0.3593689880723355, + "grad_norm": 0.9981310990585415, + "learning_rate": 1.4829980976753426e-05, + "loss": 0.6702, "step": 934 }, { - "epoch": 0.18, - "learning_rate": 1.884451437724046e-05, - "loss": 0.889, + "epoch": 0.35975375144286265, + "grad_norm": 0.9653043879294058, + "learning_rate": 1.4819065504693365e-05, + "loss": 0.5769, "step": 935 }, { - "epoch": 0.18, - "learning_rate": 1.8841604569336702e-05, - "loss": 0.9622, + "epoch": 0.36013851481338977, + "grad_norm": 0.9190267822386807, + "learning_rate": 1.4808142548925417e-05, + "loss": 0.5866, "step": 936 }, { - "epoch": 0.18, - "learning_rate": 1.883869132745561e-05, - "loss": 0.8963, + "epoch": 0.3605232781839169, + "grad_norm": 0.943957450883433, + "learning_rate": 1.4797212126412243e-05, + "loss": 0.5546, "step": 937 }, { - "epoch": 0.18, - "learning_rate": 1.883577465272866e-05, - "loss": 0.8991, + "epoch": 0.360908041554444, + "grad_norm": 0.9551640945959476, + "learning_rate": 1.4786274254128112e-05, + "loss": 0.5731, "step": 938 }, { - "epoch": 0.18, - "learning_rate": 1.8832854546288642e-05, - "loss": 0.9435, + "epoch": 0.3612928049249711, + "grad_norm": 0.9137878433417065, + "learning_rate": 1.4775328949058856e-05, + "loss": 0.5642, "step": 939 }, { - "epoch": 0.18, - "learning_rate": 1.8829931009269707e-05, - "loss": 0.9198, + "epoch": 0.36167756829549824, + "grad_norm": 0.9183530372849684, + "learning_rate": 1.4764376228201848e-05, + "loss": 0.5538, "step": 940 }, { - "epoch": 0.18, - "learning_rate": 1.882700404280731e-05, - "loss": 0.9311, + "epoch": 0.3620623316660254, + "grad_norm": 0.9995378547541753, + "learning_rate": 1.4753416108565985e-05, + "loss": 0.5494, "step": 941 }, { - "epoch": 0.18, - "learning_rate": 1.8824073648038258e-05, - "loss": 0.9557, + "epoch": 0.36244709503655254, + "grad_norm": 0.9898253822704316, + "learning_rate": 1.4742448607171644e-05, + "loss": 0.5931, "step": 942 }, { - "epoch": 0.18, - "learning_rate": 1.882113982610068e-05, - "loss": 0.9824, + "epoch": 0.36283185840707965, + "grad_norm": 0.9187946308237913, + "learning_rate": 1.4731473741050673e-05, + "loss": 0.6119, "step": 943 }, { - "epoch": 0.18, - "learning_rate": 1.881820257813404e-05, - "loss": 1.0601, + "epoch": 0.3632166217776068, + "grad_norm": 0.9977680220907835, + "learning_rate": 1.472049152724635e-05, + "loss": 0.5739, "step": 944 }, { - "epoch": 0.18, - "learning_rate": 1.8815261905279133e-05, - "loss": 0.97, + "epoch": 0.3636013851481339, + "grad_norm": 0.8890373577714931, + "learning_rate": 1.470950198281337e-05, + "loss": 0.559, "step": 945 }, { - "epoch": 0.18, - "learning_rate": 1.8812317808678075e-05, - "loss": 0.7527, + "epoch": 0.363986148518661, + "grad_norm": 0.6693570471592357, + "learning_rate": 1.4698505124817811e-05, + "loss": 0.665, "step": 946 }, { - "epoch": 0.18, - "learning_rate": 1.8809370289474327e-05, - "loss": 1.0018, + "epoch": 0.36437091188918813, + "grad_norm": 0.9876757160909431, + "learning_rate": 1.4687500970337103e-05, + "loss": 0.5771, "step": 947 }, { - "epoch": 0.18, - "learning_rate": 1.8806419348812673e-05, - "loss": 0.9946, + "epoch": 0.36475567525971525, + "grad_norm": 1.0387625515143675, + "learning_rate": 1.4676489536460015e-05, + "loss": 0.6368, "step": 948 }, { - "epoch": 0.18, - "learning_rate": 1.8803464987839217e-05, - "loss": 0.9952, + "epoch": 0.3651404386302424, + "grad_norm": 1.043457884402571, + "learning_rate": 1.4665470840286614e-05, + "loss": 0.6092, "step": 949 }, { - "epoch": 0.18, - "learning_rate": 1.88005072077014e-05, - "loss": 0.9393, + "epoch": 0.36552520200076954, + "grad_norm": 1.0460233185938146, + "learning_rate": 1.4654444898928249e-05, + "loss": 0.5831, "step": 950 }, { - "epoch": 0.18, - "learning_rate": 1.8797546009547996e-05, - "loss": 0.9964, + "epoch": 0.36590996537129666, + "grad_norm": 0.9748080273277947, + "learning_rate": 1.4643411729507517e-05, + "loss": 0.6116, "step": 951 }, { - "epoch": 0.18, - "learning_rate": 1.879458139452909e-05, - "loss": 0.9923, + "epoch": 0.3662947287418238, + "grad_norm": 0.954562676712027, + "learning_rate": 1.4632371349158241e-05, + "loss": 0.5716, "step": 952 }, { - "epoch": 0.18, - "learning_rate": 1.8791613363796118e-05, - "loss": 0.9304, + "epoch": 0.3666794921123509, + "grad_norm": 0.6018190011767177, + "learning_rate": 1.4621323775025444e-05, + "loss": 0.653, "step": 953 }, { - "epoch": 0.18, - "learning_rate": 1.8788641918501817e-05, - "loss": 0.8667, + "epoch": 0.367064255482878, + "grad_norm": 0.9525396129521565, + "learning_rate": 1.4610269024265317e-05, + "loss": 0.5169, "step": 954 }, { - "epoch": 0.18, - "learning_rate": 1.8785667059800264e-05, - "loss": 0.8219, + "epoch": 0.36744901885340514, + "grad_norm": 1.0153407693908756, + "learning_rate": 1.4599207114045202e-05, + "loss": 0.604, "step": 955 }, { - "epoch": 0.18, - "learning_rate": 1.8782688788846865e-05, - "loss": 0.8845, + "epoch": 0.36783378222393226, + "grad_norm": 0.9695074946424748, + "learning_rate": 1.4588138061543551e-05, + "loss": 0.5873, "step": 956 }, { - "epoch": 0.18, - "learning_rate": 1.877970710679834e-05, - "loss": 0.9601, + "epoch": 0.36821854559445943, + "grad_norm": 0.9529817464573429, + "learning_rate": 1.4577061883949912e-05, + "loss": 0.6218, "step": 957 }, { - "epoch": 0.18, - "learning_rate": 1.877672201481275e-05, - "loss": 0.9577, + "epoch": 0.36860330896498655, + "grad_norm": 0.9125494445596297, + "learning_rate": 1.4565978598464895e-05, + "loss": 0.562, "step": 958 }, { - "epoch": 0.18, - "learning_rate": 1.877373351404946e-05, - "loss": 0.9258, + "epoch": 0.3689880723355137, + "grad_norm": 1.0108430595571936, + "learning_rate": 1.455488822230016e-05, + "loss": 0.609, "step": 959 }, { - "epoch": 0.18, - "learning_rate": 1.8770741605669173e-05, - "loss": 0.9882, + "epoch": 0.3693728357060408, + "grad_norm": 0.9361811288777656, + "learning_rate": 1.454379077267836e-05, + "loss": 0.598, "step": 960 }, { - "epoch": 0.18, - "learning_rate": 1.876774629083391e-05, - "loss": 0.9231, + "epoch": 0.3697575990765679, + "grad_norm": 0.9379709988889509, + "learning_rate": 1.4532686266833143e-05, + "loss": 0.5626, "step": 961 }, { - "epoch": 0.19, - "learning_rate": 1.8764747570707017e-05, - "loss": 0.9853, + "epoch": 0.37014236244709503, + "grad_norm": 1.047798662627236, + "learning_rate": 1.4521574722009115e-05, + "loss": 0.5808, "step": 962 }, { - "epoch": 0.19, - "learning_rate": 1.8761745446453167e-05, - "loss": 0.9565, + "epoch": 0.37052712581762215, + "grad_norm": 1.0012183276250548, + "learning_rate": 1.4510456155461807e-05, + "loss": 0.6395, "step": 963 }, { - "epoch": 0.19, - "learning_rate": 1.875873991923835e-05, - "loss": 0.9777, + "epoch": 0.37091188918814927, + "grad_norm": 0.9921664621855362, + "learning_rate": 1.4499330584457667e-05, + "loss": 0.5724, "step": 964 }, { - "epoch": 0.19, - "learning_rate": 1.875573099022987e-05, - "loss": 0.9935, + "epoch": 0.3712966525586764, + "grad_norm": 0.9841761996105746, + "learning_rate": 1.4488198026274007e-05, + "loss": 0.5695, "step": 965 }, { - "epoch": 0.19, - "learning_rate": 1.8752718660596367e-05, - "loss": 0.9641, + "epoch": 0.37168141592920356, + "grad_norm": 0.9237151389792289, + "learning_rate": 1.4477058498198993e-05, + "loss": 0.5647, "step": 966 }, { - "epoch": 0.19, - "learning_rate": 1.8749702931507797e-05, - "loss": 0.993, + "epoch": 0.3720661792997307, + "grad_norm": 0.8924465615727354, + "learning_rate": 1.446591201753162e-05, + "loss": 0.5706, "step": 967 }, { - "epoch": 0.19, - "learning_rate": 1.874668380413543e-05, - "loss": 0.9201, + "epoch": 0.3724509426702578, + "grad_norm": 1.0402383179169261, + "learning_rate": 1.4454758601581675e-05, + "loss": 0.6231, "step": 968 }, { - "epoch": 0.19, - "learning_rate": 1.8743661279651856e-05, - "loss": 0.9878, + "epoch": 0.3728357060407849, + "grad_norm": 0.9784279312977849, + "learning_rate": 1.4443598267669723e-05, + "loss": 0.6107, "step": 969 }, { - "epoch": 0.19, - "learning_rate": 1.8740635359231e-05, - "loss": 1.0191, + "epoch": 0.37322046941131204, + "grad_norm": 0.9899635705924709, + "learning_rate": 1.4432431033127056e-05, + "loss": 0.6175, "step": 970 }, { - "epoch": 0.19, - "learning_rate": 1.8737606044048086e-05, - "loss": 0.9158, + "epoch": 0.37360523278183916, + "grad_norm": 1.011742060176989, + "learning_rate": 1.4421256915295697e-05, + "loss": 0.6267, "step": 971 }, { - "epoch": 0.19, - "learning_rate": 1.873457333527967e-05, - "loss": 1.0359, + "epoch": 0.3739899961523663, + "grad_norm": 0.8748862269980235, + "learning_rate": 1.4410075931528356e-05, + "loss": 0.6412, "step": 972 }, { - "epoch": 0.19, - "learning_rate": 1.873153723410362e-05, - "loss": 0.7623, + "epoch": 0.3743747595228934, + "grad_norm": 0.9381591074720641, + "learning_rate": 1.4398888099188396e-05, + "loss": 0.5411, "step": 973 }, { - "epoch": 0.19, - "learning_rate": 1.8728497741699115e-05, - "loss": 0.9706, + "epoch": 0.37475952289342057, + "grad_norm": 0.9915792684221186, + "learning_rate": 1.4387693435649826e-05, + "loss": 0.5763, "step": 974 }, { - "epoch": 0.19, - "learning_rate": 1.872545485924667e-05, - "loss": 0.9513, + "epoch": 0.3751442862639477, + "grad_norm": 1.03984169294167, + "learning_rate": 1.4376491958297263e-05, + "loss": 0.5812, "step": 975 }, { - "epoch": 0.19, - "learning_rate": 1.8722408587928104e-05, - "loss": 1.0909, + "epoch": 0.3755290496344748, + "grad_norm": 1.0099876697950052, + "learning_rate": 1.4365283684525895e-05, + "loss": 0.5718, "step": 976 }, { - "epoch": 0.19, - "learning_rate": 1.8719358928926546e-05, - "loss": 0.9878, + "epoch": 0.37591381300500193, + "grad_norm": 0.9469853592561607, + "learning_rate": 1.4354068631741476e-05, + "loss": 0.5907, "step": 977 }, { - "epoch": 0.19, - "learning_rate": 1.8716305883426456e-05, - "loss": 0.9298, + "epoch": 0.37629857637552905, + "grad_norm": 0.6501232608446648, + "learning_rate": 1.434284681736028e-05, + "loss": 0.6805, "step": 978 }, { - "epoch": 0.19, - "learning_rate": 1.87132494526136e-05, - "loss": 0.9344, + "epoch": 0.37668333974605617, + "grad_norm": 0.9665649822873006, + "learning_rate": 1.433161825880909e-05, + "loss": 0.5721, "step": 979 }, { - "epoch": 0.19, - "learning_rate": 1.8710189637675055e-05, - "loss": 0.9427, + "epoch": 0.3770681031165833, + "grad_norm": 0.9896983402449874, + "learning_rate": 1.4320382973525151e-05, + "loss": 0.5821, "step": 980 }, { - "epoch": 0.19, - "learning_rate": 1.8707126439799225e-05, - "loss": 0.9338, + "epoch": 0.3774528664871104, + "grad_norm": 0.9414818606305622, + "learning_rate": 1.4309140978956161e-05, + "loss": 0.5266, "step": 981 }, { - "epoch": 0.19, - "learning_rate": 1.870405986017582e-05, - "loss": 0.9295, + "epoch": 0.3778376298576376, + "grad_norm": 0.9979190168328879, + "learning_rate": 1.429789229256024e-05, + "loss": 0.5767, "step": 982 }, { - "epoch": 0.19, - "learning_rate": 1.8700989899995857e-05, - "loss": 0.892, + "epoch": 0.3782223932281647, + "grad_norm": 0.8700310917538334, + "learning_rate": 1.4286636931805887e-05, + "loss": 0.5744, "step": 983 }, { - "epoch": 0.19, - "learning_rate": 1.8697916560451682e-05, - "loss": 0.9456, + "epoch": 0.3786071565986918, + "grad_norm": 1.064719538069214, + "learning_rate": 1.427537491417198e-05, + "loss": 0.6262, "step": 984 }, { - "epoch": 0.19, - "learning_rate": 1.869483984273694e-05, - "loss": 0.9484, + "epoch": 0.37899191996921894, + "grad_norm": 0.6710655118514386, + "learning_rate": 1.4264106257147732e-05, + "loss": 0.6398, "step": 985 }, { - "epoch": 0.19, - "learning_rate": 1.8691759748046594e-05, - "loss": 1.0229, + "epoch": 0.37937668333974606, + "grad_norm": 1.0182812345559562, + "learning_rate": 1.4252830978232658e-05, + "loss": 0.6127, "step": 986 }, { - "epoch": 0.19, - "learning_rate": 1.8688676277576916e-05, - "loss": 0.9516, + "epoch": 0.3797614467102732, + "grad_norm": 1.0409894877787191, + "learning_rate": 1.4241549094936567e-05, + "loss": 0.5955, "step": 987 }, { - "epoch": 0.19, - "learning_rate": 1.868558943252549e-05, - "loss": 0.9046, + "epoch": 0.3801462100808003, + "grad_norm": 0.9800890611243236, + "learning_rate": 1.4230260624779512e-05, + "loss": 0.5848, "step": 988 }, { - "epoch": 0.19, - "learning_rate": 1.868249921409122e-05, - "loss": 1.0169, + "epoch": 0.3805309734513274, + "grad_norm": 1.1901238160743828, + "learning_rate": 1.4218965585291792e-05, + "loss": 0.5701, "step": 989 }, { - "epoch": 0.19, - "learning_rate": 1.8679405623474294e-05, - "loss": 0.9768, + "epoch": 0.38091573682185453, + "grad_norm": 0.9668649088821483, + "learning_rate": 1.4207663994013896e-05, + "loss": 0.6047, "step": 990 }, { - "epoch": 0.19, - "learning_rate": 1.8676308661876242e-05, - "loss": 1.0422, + "epoch": 0.3813005001923817, + "grad_norm": 1.1024557177230976, + "learning_rate": 1.4196355868496485e-05, + "loss": 0.6093, "step": 991 }, { - "epoch": 0.19, - "learning_rate": 1.8673208330499884e-05, - "loss": 0.9473, + "epoch": 0.3816852635629088, + "grad_norm": 1.043519502368422, + "learning_rate": 1.4185041226300376e-05, + "loss": 0.6409, "step": 992 }, { - "epoch": 0.19, - "learning_rate": 1.8670104630549348e-05, - "loss": 1.0, + "epoch": 0.38207002693343595, + "grad_norm": 1.0108214424101285, + "learning_rate": 1.4173720084996501e-05, + "loss": 0.6063, "step": 993 }, { - "epoch": 0.19, - "learning_rate": 1.866699756323008e-05, - "loss": 0.8941, + "epoch": 0.38245479030396307, + "grad_norm": 0.9496508914700906, + "learning_rate": 1.4162392462165884e-05, + "loss": 0.5272, "step": 994 }, { - "epoch": 0.19, - "learning_rate": 1.866388712974883e-05, - "loss": 1.0444, + "epoch": 0.3828395536744902, + "grad_norm": 0.6061682579145548, + "learning_rate": 1.415105837539962e-05, + "loss": 0.6463, "step": 995 }, { - "epoch": 0.19, - "learning_rate": 1.866077333131365e-05, - "loss": 0.8769, + "epoch": 0.3832243170450173, + "grad_norm": 0.9688608619696606, + "learning_rate": 1.4139717842298835e-05, + "loss": 0.5708, "step": 996 }, { - "epoch": 0.19, - "learning_rate": 1.8657656169133908e-05, - "loss": 0.9808, + "epoch": 0.3836090804155444, + "grad_norm": 0.9798437530657843, + "learning_rate": 1.4128370880474667e-05, + "loss": 0.5933, "step": 997 }, { - "epoch": 0.19, - "learning_rate": 1.8654535644420276e-05, - "loss": 0.9003, + "epoch": 0.38399384378607154, + "grad_norm": 1.024313390494135, + "learning_rate": 1.4117017507548244e-05, + "loss": 0.6025, "step": 998 }, { - "epoch": 0.19, - "learning_rate": 1.8651411758384718e-05, - "loss": 0.894, + "epoch": 0.3843786071565987, + "grad_norm": 0.928634388897143, + "learning_rate": 1.4105657741150648e-05, + "loss": 0.5233, "step": 999 }, { - "epoch": 0.19, - "learning_rate": 1.8648284512240527e-05, - "loss": 0.941, + "epoch": 0.38476337052712584, + "grad_norm": 1.0298279151003196, + "learning_rate": 1.4094291598922877e-05, + "loss": 0.5919, "step": 1000 }, { - "epoch": 0.19, - "learning_rate": 1.8645153907202285e-05, - "loss": 0.8964, + "epoch": 0.38514813389765296, + "grad_norm": 1.0278768925004698, + "learning_rate": 1.4082919098515846e-05, + "loss": 0.5978, "step": 1001 }, { - "epoch": 0.19, - "learning_rate": 1.8642019944485884e-05, - "loss": 0.7338, + "epoch": 0.3855328972681801, + "grad_norm": 0.969484222231415, + "learning_rate": 1.4071540257590341e-05, + "loss": 0.6079, "step": 1002 }, { - "epoch": 0.19, - "learning_rate": 1.863888262530852e-05, - "loss": 0.9768, + "epoch": 0.3859176606387072, + "grad_norm": 1.1023046763937696, + "learning_rate": 1.4060155093816988e-05, + "loss": 0.5688, "step": 1003 }, { - "epoch": 0.19, - "learning_rate": 1.863574195088869e-05, - "loss": 0.8652, + "epoch": 0.3863024240092343, + "grad_norm": 0.9684836786707451, + "learning_rate": 1.4048763624876233e-05, + "loss": 0.5979, "step": 1004 }, { - "epoch": 0.19, - "learning_rate": 1.8632597922446195e-05, - "loss": 0.9917, + "epoch": 0.38668718737976143, + "grad_norm": 0.8980606564973477, + "learning_rate": 1.4037365868458325e-05, + "loss": 0.5805, "step": 1005 }, { - "epoch": 0.19, - "learning_rate": 1.8629450541202142e-05, - "loss": 1.081, + "epoch": 0.38707195075028855, + "grad_norm": 0.9461416656980055, + "learning_rate": 1.402596184226326e-05, + "loss": 0.6173, "step": 1006 }, { - "epoch": 0.19, - "learning_rate": 1.8626299808378933e-05, - "loss": 0.959, + "epoch": 0.38745671412081567, + "grad_norm": 0.9213379867335817, + "learning_rate": 1.401455156400078e-05, + "loss": 0.5679, "step": 1007 }, { - "epoch": 0.19, - "learning_rate": 1.862314572520028e-05, - "loss": 0.9987, + "epoch": 0.38784147749134285, + "grad_norm": 0.6178490811470693, + "learning_rate": 1.400313505139034e-05, + "loss": 0.6095, "step": 1008 }, { - "epoch": 0.19, - "learning_rate": 1.861998829289119e-05, - "loss": 0.9992, + "epoch": 0.38822624086186996, + "grad_norm": 0.9855122947768239, + "learning_rate": 1.3991712322161065e-05, + "loss": 0.5754, "step": 1009 }, { - "epoch": 0.19, - "learning_rate": 1.861682751267798e-05, - "loss": 0.9278, + "epoch": 0.3886110042323971, + "grad_norm": 0.9242887465176792, + "learning_rate": 1.398028339405174e-05, + "loss": 0.5823, "step": 1010 }, { - "epoch": 0.19, - "learning_rate": 1.861366338578825e-05, - "loss": 0.9703, + "epoch": 0.3889957676029242, + "grad_norm": 1.039363232149176, + "learning_rate": 1.3968848284810785e-05, + "loss": 0.577, "step": 1011 }, { - "epoch": 0.19, - "learning_rate": 1.8610495913450922e-05, - "loss": 0.9187, + "epoch": 0.3893805309734513, + "grad_norm": 1.0693840564066996, + "learning_rate": 1.3957407012196204e-05, + "loss": 0.5795, "step": 1012 }, { - "epoch": 0.19, - "learning_rate": 1.8607325096896197e-05, - "loss": 0.7599, + "epoch": 0.38976529434397844, + "grad_norm": 0.9718433884925407, + "learning_rate": 1.3945959593975582e-05, + "loss": 0.6259, "step": 1013 }, { - "epoch": 0.2, - "learning_rate": 1.8604150937355588e-05, - "loss": 0.9441, + "epoch": 0.39015005771450556, + "grad_norm": 0.9782584999722043, + "learning_rate": 1.3934506047926042e-05, + "loss": 0.611, "step": 1014 }, { - "epoch": 0.2, - "learning_rate": 1.86009734360619e-05, - "loss": 1.0393, + "epoch": 0.3905348210850327, + "grad_norm": 1.1065748011638874, + "learning_rate": 1.3923046391834229e-05, + "loss": 0.5937, "step": 1015 }, { - "epoch": 0.2, - "learning_rate": 1.8597792594249237e-05, - "loss": 1.0475, + "epoch": 0.39091958445555985, + "grad_norm": 1.0873795347646902, + "learning_rate": 1.3911580643496272e-05, + "loss": 0.5922, "step": 1016 }, { - "epoch": 0.2, - "learning_rate": 1.8594608413153e-05, - "loss": 1.007, + "epoch": 0.391304347826087, + "grad_norm": 0.9068442414947984, + "learning_rate": 1.390010882071776e-05, + "loss": 0.5194, "step": 1017 }, { - "epoch": 0.2, - "learning_rate": 1.8591420894009897e-05, - "loss": 0.998, + "epoch": 0.3916891111966141, + "grad_norm": 0.9607335059954972, + "learning_rate": 1.3888630941313728e-05, + "loss": 0.5773, "step": 1018 }, { - "epoch": 0.2, - "learning_rate": 1.8588230038057913e-05, - "loss": 0.9645, + "epoch": 0.3920738745671412, + "grad_norm": 1.1050115555481446, + "learning_rate": 1.3877147023108592e-05, + "loss": 0.6251, "step": 1019 }, { - "epoch": 0.2, - "learning_rate": 1.8585035846536347e-05, - "loss": 0.9392, + "epoch": 0.39245863793766833, + "grad_norm": 0.5852107209499082, + "learning_rate": 1.3865657083936167e-05, + "loss": 0.6258, "step": 1020 }, { - "epoch": 0.2, - "learning_rate": 1.8581838320685782e-05, - "loss": 1.0192, + "epoch": 0.39284340130819545, + "grad_norm": 0.968742387102526, + "learning_rate": 1.385416114163961e-05, + "loss": 0.5413, "step": 1021 }, { - "epoch": 0.2, - "learning_rate": 1.8578637461748105e-05, - "loss": 0.943, + "epoch": 0.39322816467872257, + "grad_norm": 1.0812164902241694, + "learning_rate": 1.3842659214071406e-05, + "loss": 0.5935, "step": 1022 }, { - "epoch": 0.2, - "learning_rate": 1.857543327096649e-05, - "loss": 1.0503, + "epoch": 0.3936129280492497, + "grad_norm": 1.0656409056595197, + "learning_rate": 1.3831151319093323e-05, + "loss": 0.6176, "step": 1023 }, { - "epoch": 0.2, - "learning_rate": 1.85722257495854e-05, - "loss": 0.9322, + "epoch": 0.39399769141977686, + "grad_norm": 0.967248412419409, + "learning_rate": 1.3819637474576411e-05, + "loss": 0.5518, "step": 1024 }, { - "epoch": 0.2, - "learning_rate": 1.856901489885061e-05, - "loss": 1.0351, + "epoch": 0.394382454790304, + "grad_norm": 0.9948192764379216, + "learning_rate": 1.380811769840095e-05, + "loss": 0.6029, "step": 1025 }, { - "epoch": 0.2, - "learning_rate": 1.856580072000918e-05, - "loss": 0.9363, + "epoch": 0.3947672181608311, + "grad_norm": 0.9579034672220184, + "learning_rate": 1.3796592008456427e-05, + "loss": 0.6006, "step": 1026 }, { - "epoch": 0.2, - "learning_rate": 1.8562583214309447e-05, - "loss": 0.9729, + "epoch": 0.3951519815313582, + "grad_norm": 0.9706918750562253, + "learning_rate": 1.3785060422641526e-05, + "loss": 0.5582, "step": 1027 }, { - "epoch": 0.2, - "learning_rate": 1.855936238300106e-05, - "loss": 1.1, + "epoch": 0.39553674490188534, + "grad_norm": 0.9677691482191173, + "learning_rate": 1.3773522958864076e-05, + "loss": 0.6218, "step": 1028 }, { - "epoch": 0.2, - "learning_rate": 1.8556138227334957e-05, - "loss": 0.9783, + "epoch": 0.39592150827241246, + "grad_norm": 0.9959438300055631, + "learning_rate": 1.376197963504104e-05, + "loss": 0.5747, "step": 1029 }, { - "epoch": 0.2, - "learning_rate": 1.855291074856336e-05, - "loss": 0.9743, + "epoch": 0.3963062716429396, + "grad_norm": 0.9870510028797354, + "learning_rate": 1.375043046909848e-05, + "loss": 0.6094, "step": 1030 }, { - "epoch": 0.2, - "learning_rate": 1.8549679947939778e-05, - "loss": 0.9587, + "epoch": 0.3966910350134667, + "grad_norm": 1.007175619604548, + "learning_rate": 1.3738875478971526e-05, + "loss": 0.5926, "step": 1031 }, { - "epoch": 0.2, - "learning_rate": 1.8546445826719023e-05, - "loss": 1.0238, + "epoch": 0.3970757983839938, + "grad_norm": 0.9789030255031183, + "learning_rate": 1.372731468260436e-05, + "loss": 0.5591, "step": 1032 }, { - "epoch": 0.2, - "learning_rate": 1.8543208386157195e-05, - "loss": 1.0031, + "epoch": 0.397460561754521, + "grad_norm": 0.986698658261605, + "learning_rate": 1.3715748097950176e-05, + "loss": 0.6194, "step": 1033 }, { - "epoch": 0.2, - "learning_rate": 1.853996762751167e-05, - "loss": 1.0217, + "epoch": 0.3978453251250481, + "grad_norm": 0.9548514390986464, + "learning_rate": 1.3704175742971158e-05, + "loss": 0.5831, "step": 1034 }, { - "epoch": 0.2, - "learning_rate": 1.8536723552041124e-05, - "loss": 0.9437, + "epoch": 0.39823008849557523, + "grad_norm": 0.9181617294927629, + "learning_rate": 1.3692597635638452e-05, + "loss": 0.5376, "step": 1035 }, { - "epoch": 0.2, - "learning_rate": 1.853347616100552e-05, - "loss": 0.9633, + "epoch": 0.39861485186610235, + "grad_norm": 0.9945920286093195, + "learning_rate": 1.3681013793932132e-05, + "loss": 0.5661, "step": 1036 }, { - "epoch": 0.2, - "learning_rate": 1.8530225455666103e-05, - "loss": 0.9283, + "epoch": 0.39899961523662947, + "grad_norm": 0.9832515847614587, + "learning_rate": 1.3669424235841185e-05, + "loss": 0.5501, "step": 1037 }, { - "epoch": 0.2, - "learning_rate": 1.8526971437285416e-05, - "loss": 1.0186, + "epoch": 0.3993843786071566, + "grad_norm": 1.0234136528586504, + "learning_rate": 1.3657828979363468e-05, + "loss": 0.6164, "step": 1038 }, { - "epoch": 0.2, - "learning_rate": 1.8523714107127278e-05, - "loss": 0.9568, + "epoch": 0.3997691419776837, + "grad_norm": 0.6076865131336912, + "learning_rate": 1.3646228042505694e-05, + "loss": 0.6568, "step": 1039 }, { - "epoch": 0.2, - "learning_rate": 1.8520453466456797e-05, - "loss": 0.9618, + "epoch": 0.4001539053482108, + "grad_norm": 0.9558254304549876, + "learning_rate": 1.3634621443283389e-05, + "loss": 0.5658, "step": 1040 }, { - "epoch": 0.2, - "learning_rate": 1.8517189516540376e-05, - "loss": 1.0732, + "epoch": 0.400538668718738, + "grad_norm": 1.0637258773418374, + "learning_rate": 1.3623009199720882e-05, + "loss": 0.6444, "step": 1041 }, { - "epoch": 0.2, - "learning_rate": 1.8513922258645687e-05, - "loss": 0.9604, + "epoch": 0.4009234320892651, + "grad_norm": 0.9613356746003873, + "learning_rate": 1.3611391329851262e-05, + "loss": 0.5766, "step": 1042 }, { - "epoch": 0.2, - "learning_rate": 1.8510651694041702e-05, - "loss": 0.9504, + "epoch": 0.40130819545979224, + "grad_norm": 1.0496596626548866, + "learning_rate": 1.3599767851716353e-05, + "loss": 0.5697, "step": 1043 }, { - "epoch": 0.2, - "learning_rate": 1.8507377823998664e-05, - "loss": 0.9428, + "epoch": 0.40169295883031936, + "grad_norm": 1.0044750907407405, + "learning_rate": 1.3588138783366692e-05, + "loss": 0.5773, "step": 1044 }, { - "epoch": 0.2, - "learning_rate": 1.850410064978811e-05, - "loss": 0.8585, + "epoch": 0.4020777222008465, + "grad_norm": 1.0371054498837338, + "learning_rate": 1.3576504142861496e-05, + "loss": 0.6319, "step": 1045 }, { - "epoch": 0.2, - "learning_rate": 1.8500820172682858e-05, - "loss": 1.0079, + "epoch": 0.4024624855713736, + "grad_norm": 0.9842508021596283, + "learning_rate": 1.3564863948268631e-05, + "loss": 0.5847, "step": 1046 }, { - "epoch": 0.2, - "learning_rate": 1.8497536393957005e-05, - "loss": 0.7325, + "epoch": 0.4028472489419007, + "grad_norm": 1.0329401476224598, + "learning_rate": 1.3553218217664603e-05, + "loss": 0.5722, "step": 1047 }, { - "epoch": 0.2, - "learning_rate": 1.8494249314885932e-05, - "loss": 0.8797, + "epoch": 0.40323201231242783, + "grad_norm": 0.9266990636001327, + "learning_rate": 1.3541566969134496e-05, + "loss": 0.5963, "step": 1048 }, { - "epoch": 0.2, - "learning_rate": 1.8490958936746304e-05, - "loss": 0.9747, + "epoch": 0.403616775682955, + "grad_norm": 0.9879924569219896, + "learning_rate": 1.3529910220771975e-05, + "loss": 0.561, "step": 1049 }, { - "epoch": 0.2, - "learning_rate": 1.848766526081607e-05, - "loss": 0.9119, + "epoch": 0.40400153905348213, + "grad_norm": 0.9217455020320823, + "learning_rate": 1.3518247990679241e-05, + "loss": 0.5699, "step": 1050 }, { - "epoch": 0.2, - "learning_rate": 1.8484368288374452e-05, - "loss": 0.966, + "epoch": 0.40438630242400925, + "grad_norm": 1.0342099459195506, + "learning_rate": 1.3506580296967011e-05, + "loss": 0.5874, "step": 1051 }, { - "epoch": 0.2, - "learning_rate": 1.8481068020701954e-05, - "loss": 0.9977, + "epoch": 0.40477106579453637, + "grad_norm": 0.9223100677079575, + "learning_rate": 1.3494907157754485e-05, + "loss": 0.5704, "step": 1052 }, { - "epoch": 0.2, - "learning_rate": 1.8477764459080364e-05, - "loss": 0.9694, + "epoch": 0.4051558291650635, + "grad_norm": 0.9593485427613434, + "learning_rate": 1.3483228591169315e-05, + "loss": 0.5983, "step": 1053 }, { - "epoch": 0.2, - "learning_rate": 1.8474457604792746e-05, - "loss": 0.7796, + "epoch": 0.4055405925355906, + "grad_norm": 1.0547643780458376, + "learning_rate": 1.3471544615347591e-05, + "loss": 0.6079, "step": 1054 }, { - "epoch": 0.2, - "learning_rate": 1.8471147459123447e-05, - "loss": 0.9814, + "epoch": 0.4059253559061177, + "grad_norm": 1.065286362177476, + "learning_rate": 1.34598552484338e-05, + "loss": 0.6243, "step": 1055 }, { - "epoch": 0.2, - "learning_rate": 1.8467834023358088e-05, - "loss": 0.8853, + "epoch": 0.40631011927664484, + "grad_norm": 0.9227337793958981, + "learning_rate": 1.3448160508580789e-05, + "loss": 0.5841, "step": 1056 }, { - "epoch": 0.2, - "learning_rate": 1.846451729878357e-05, - "loss": 1.022, + "epoch": 0.40669488264717196, + "grad_norm": 0.979052738318199, + "learning_rate": 1.343646041394977e-05, + "loss": 0.5875, "step": 1057 }, { - "epoch": 0.2, - "learning_rate": 1.846119728668807e-05, - "loss": 1.0306, + "epoch": 0.40707964601769914, + "grad_norm": 0.9580263126551128, + "learning_rate": 1.3424754982710256e-05, + "loss": 0.6056, "step": 1058 }, { - "epoch": 0.2, - "learning_rate": 1.845787398836104e-05, - "loss": 1.0122, + "epoch": 0.40746440938822626, + "grad_norm": 0.9961515588496205, + "learning_rate": 1.3413044233040045e-05, + "loss": 0.595, "step": 1059 }, { - "epoch": 0.2, - "learning_rate": 1.8454547405093212e-05, - "loss": 1.0162, + "epoch": 0.4078491727587534, + "grad_norm": 0.968731626744559, + "learning_rate": 1.3401328183125208e-05, + "loss": 0.5849, "step": 1060 }, { - "epoch": 0.2, - "learning_rate": 1.8451217538176597e-05, - "loss": 0.9437, + "epoch": 0.4082339361292805, + "grad_norm": 0.9848409408533877, + "learning_rate": 1.3389606851160037e-05, + "loss": 0.5636, "step": 1061 }, { - "epoch": 0.2, - "learning_rate": 1.844788438890447e-05, - "loss": 0.9489, + "epoch": 0.4086186994998076, + "grad_norm": 0.9565328411977299, + "learning_rate": 1.3377880255347026e-05, + "loss": 0.5709, "step": 1062 }, { - "epoch": 0.2, - "learning_rate": 1.8444547958571396e-05, - "loss": 0.7416, + "epoch": 0.40900346287033473, + "grad_norm": 0.9519648341724759, + "learning_rate": 1.3366148413896851e-05, + "loss": 0.5516, "step": 1063 }, { - "epoch": 0.2, - "learning_rate": 1.84412082484732e-05, - "loss": 0.7556, + "epoch": 0.40938822624086185, + "grad_norm": 1.0105547708647702, + "learning_rate": 1.3354411345028324e-05, + "loss": 0.5832, "step": 1064 }, { - "epoch": 0.2, - "learning_rate": 1.8437865259906987e-05, - "loss": 0.9998, + "epoch": 0.40977298961138897, + "grad_norm": 1.0349423468299899, + "learning_rate": 1.3342669066968385e-05, + "loss": 0.5937, "step": 1065 }, { - "epoch": 0.21, - "learning_rate": 1.8434518994171136e-05, - "loss": 0.9846, + "epoch": 0.41015775298191615, + "grad_norm": 0.9660233366904522, + "learning_rate": 1.3330921597952056e-05, + "loss": 0.6088, "step": 1066 }, { - "epoch": 0.21, - "learning_rate": 1.84311694525653e-05, - "loss": 0.9342, + "epoch": 0.41054251635244327, + "grad_norm": 0.9788696649346639, + "learning_rate": 1.3319168956222423e-05, + "loss": 0.5277, "step": 1067 }, { - "epoch": 0.21, - "learning_rate": 1.84278166363904e-05, - "loss": 0.8994, + "epoch": 0.4109272797229704, + "grad_norm": 0.9854700571902064, + "learning_rate": 1.3307411160030608e-05, + "loss": 0.6111, "step": 1068 }, { - "epoch": 0.21, - "learning_rate": 1.8424460546948632e-05, - "loss": 0.9479, + "epoch": 0.4113120430934975, + "grad_norm": 0.9999768807436922, + "learning_rate": 1.3295648227635729e-05, + "loss": 0.5613, "step": 1069 }, { - "epoch": 0.21, - "learning_rate": 1.8421101185543463e-05, - "loss": 0.9334, + "epoch": 0.4116968064640246, + "grad_norm": 1.0552401478349673, + "learning_rate": 1.328388017730489e-05, + "loss": 0.5742, "step": 1070 }, { - "epoch": 0.21, - "learning_rate": 1.841773855347963e-05, - "loss": 0.9512, + "epoch": 0.41208156983455174, + "grad_norm": 0.6655174531259913, + "learning_rate": 1.3272107027313142e-05, + "loss": 0.6685, "step": 1071 }, { - "epoch": 0.21, - "learning_rate": 1.841437265206314e-05, - "loss": 0.8737, + "epoch": 0.41246633320507886, + "grad_norm": 1.063202179214444, + "learning_rate": 1.326032879594344e-05, + "loss": 0.5727, "step": 1072 }, { - "epoch": 0.21, - "learning_rate": 1.841100348260127e-05, - "loss": 1.0085, + "epoch": 0.412851096575606, + "grad_norm": 0.9174136198658255, + "learning_rate": 1.3248545501486654e-05, + "loss": 0.5656, "step": 1073 }, { - "epoch": 0.21, - "learning_rate": 1.840763104640257e-05, - "loss": 1.0125, + "epoch": 0.41323585994613315, + "grad_norm": 0.9642493387417168, + "learning_rate": 1.32367571622415e-05, + "loss": 0.6108, "step": 1074 }, { - "epoch": 0.21, - "learning_rate": 1.8404255344776853e-05, - "loss": 0.9674, + "epoch": 0.4136206233166603, + "grad_norm": 1.0056210677197837, + "learning_rate": 1.3224963796514532e-05, + "loss": 0.5842, "step": 1075 }, { - "epoch": 0.21, - "learning_rate": 1.84008763790352e-05, - "loss": 0.8841, + "epoch": 0.4140053866871874, + "grad_norm": 0.9619049624495174, + "learning_rate": 1.3213165422620111e-05, + "loss": 0.5721, "step": 1076 }, { - "epoch": 0.21, - "learning_rate": 1.8397494150489963e-05, - "loss": 1.029, + "epoch": 0.4143901500577145, + "grad_norm": 1.0201688811623655, + "learning_rate": 1.3201362058880375e-05, + "loss": 0.5717, "step": 1077 }, { - "epoch": 0.21, - "learning_rate": 1.8394108660454766e-05, - "loss": 0.8894, + "epoch": 0.41477491342824163, + "grad_norm": 0.9510605825707248, + "learning_rate": 1.3189553723625217e-05, + "loss": 0.6105, "step": 1078 }, { - "epoch": 0.21, - "learning_rate": 1.8390719910244487e-05, - "loss": 0.9822, + "epoch": 0.41515967679876875, + "grad_norm": 0.9585005160558335, + "learning_rate": 1.3177740435192235e-05, + "loss": 0.5833, "step": 1079 }, { - "epoch": 0.21, - "learning_rate": 1.8387327901175286e-05, - "loss": 0.9843, + "epoch": 0.41554444016929587, + "grad_norm": 1.015174794542384, + "learning_rate": 1.3165922211926734e-05, + "loss": 0.6104, "step": 1080 }, { - "epoch": 0.21, - "learning_rate": 1.838393263456457e-05, - "loss": 0.9913, + "epoch": 0.415929203539823, + "grad_norm": 0.6213002419727812, + "learning_rate": 1.3154099072181677e-05, + "loss": 0.6664, "step": 1081 }, { - "epoch": 0.21, - "learning_rate": 1.838053411173103e-05, - "loss": 0.9587, + "epoch": 0.4163139669103501, + "grad_norm": 1.011531841556601, + "learning_rate": 1.314227103431766e-05, + "loss": 0.5493, "step": 1082 }, { - "epoch": 0.21, - "learning_rate": 1.8377132333994606e-05, - "loss": 0.9252, + "epoch": 0.4166987302808773, + "grad_norm": 1.0248754034327239, + "learning_rate": 1.3130438116702888e-05, + "loss": 0.5527, "step": 1083 }, { - "epoch": 0.21, - "learning_rate": 1.837372730267652e-05, - "loss": 0.911, + "epoch": 0.4170834936514044, + "grad_norm": 0.949239833995476, + "learning_rate": 1.3118600337713146e-05, + "loss": 0.5846, "step": 1084 }, { - "epoch": 0.21, - "learning_rate": 1.8370319019099236e-05, - "loss": 0.9352, + "epoch": 0.4174682570219315, + "grad_norm": 1.0859651556833136, + "learning_rate": 1.310675771573176e-05, + "loss": 0.5875, "step": 1085 }, { - "epoch": 0.21, - "learning_rate": 1.8366907484586497e-05, - "loss": 0.976, + "epoch": 0.41785302039245864, + "grad_norm": 0.875858832955551, + "learning_rate": 1.3094910269149587e-05, + "loss": 0.6067, "step": 1086 }, { - "epoch": 0.21, - "learning_rate": 1.83634927004633e-05, - "loss": 0.9751, + "epoch": 0.41823778376298576, + "grad_norm": 0.9454770430601649, + "learning_rate": 1.3083058016364972e-05, + "loss": 0.5664, "step": 1087 }, { - "epoch": 0.21, - "learning_rate": 1.8360074668055915e-05, - "loss": 0.9444, + "epoch": 0.4186225471335129, + "grad_norm": 1.0203312729418248, + "learning_rate": 1.3071200975783725e-05, + "loss": 0.622, "step": 1088 }, { - "epoch": 0.21, - "learning_rate": 1.8356653388691857e-05, - "loss": 0.9107, + "epoch": 0.41900731050404, + "grad_norm": 0.9023391768252292, + "learning_rate": 1.3059339165819082e-05, + "loss": 0.5616, "step": 1089 }, { - "epoch": 0.21, - "learning_rate": 1.8353228863699922e-05, - "loss": 0.9603, + "epoch": 0.4193920738745671, + "grad_norm": 0.6581672811602944, + "learning_rate": 1.3047472604891701e-05, + "loss": 0.6119, "step": 1090 }, { - "epoch": 0.21, - "learning_rate": 1.8349801094410148e-05, - "loss": 0.8936, + "epoch": 0.4197768372450943, + "grad_norm": 0.9632314668481454, + "learning_rate": 1.303560131142961e-05, + "loss": 0.6143, "step": 1091 }, { - "epoch": 0.21, - "learning_rate": 1.8346370082153843e-05, - "loss": 0.8853, + "epoch": 0.4201616006156214, + "grad_norm": 1.0679701640743904, + "learning_rate": 1.3023725303868183e-05, + "loss": 0.6298, "step": 1092 }, { - "epoch": 0.21, - "learning_rate": 1.8342935828263574e-05, - "loss": 1.0094, + "epoch": 0.42054636398614853, + "grad_norm": 0.9753073172535588, + "learning_rate": 1.3011844600650121e-05, + "loss": 0.6486, "step": 1093 }, { - "epoch": 0.21, - "learning_rate": 1.8339498334073166e-05, - "loss": 0.9535, + "epoch": 0.42093112735667565, + "grad_norm": 1.0074161808074784, + "learning_rate": 1.2999959220225416e-05, + "loss": 0.5717, "step": 1094 }, { - "epoch": 0.21, - "learning_rate": 1.83360576009177e-05, - "loss": 0.9125, + "epoch": 0.42131589072720277, + "grad_norm": 1.0386380961618988, + "learning_rate": 1.2988069181051314e-05, + "loss": 0.5961, "step": 1095 }, { - "epoch": 0.21, - "learning_rate": 1.833261363013352e-05, - "loss": 0.9396, + "epoch": 0.4217006540977299, + "grad_norm": 0.6209565606168068, + "learning_rate": 1.2976174501592313e-05, + "loss": 0.6382, "step": 1096 }, { - "epoch": 0.21, - "learning_rate": 1.832916642305822e-05, - "loss": 0.9139, + "epoch": 0.422085417468257, + "grad_norm": 1.0123833903748787, + "learning_rate": 1.2964275200320104e-05, + "loss": 0.5532, "step": 1097 }, { - "epoch": 0.21, - "learning_rate": 1.832571598103066e-05, - "loss": 0.9224, + "epoch": 0.4224701808387841, + "grad_norm": 0.9304954947473683, + "learning_rate": 1.2952371295713558e-05, + "loss": 0.6036, "step": 1098 }, { - "epoch": 0.21, - "learning_rate": 1.8322262305390948e-05, - "loss": 1.0444, + "epoch": 0.4228549442093113, + "grad_norm": 1.0375496236700013, + "learning_rate": 1.2940462806258696e-05, + "loss": 0.611, "step": 1099 }, { - "epoch": 0.21, - "learning_rate": 1.8318805397480455e-05, - "loss": 0.9598, + "epoch": 0.4232397075798384, + "grad_norm": 0.9868239014706708, + "learning_rate": 1.2928549750448661e-05, + "loss": 0.5563, "step": 1100 }, { - "epoch": 0.21, - "learning_rate": 1.8315345258641802e-05, - "loss": 1.0404, + "epoch": 0.42362447095036554, + "grad_norm": 0.9519023777345361, + "learning_rate": 1.2916632146783683e-05, + "loss": 0.5624, "step": 1101 }, { - "epoch": 0.21, - "learning_rate": 1.8311881890218873e-05, - "loss": 0.9215, + "epoch": 0.42400923432089266, + "grad_norm": 0.6904606794867917, + "learning_rate": 1.2904710013771054e-05, + "loss": 0.6456, "step": 1102 }, { - "epoch": 0.21, - "learning_rate": 1.830841529355679e-05, - "loss": 0.9365, + "epoch": 0.4243939976914198, + "grad_norm": 0.9991197957260134, + "learning_rate": 1.2892783369925105e-05, + "loss": 0.5585, "step": 1103 }, { - "epoch": 0.21, - "learning_rate": 1.8304945470001948e-05, - "loss": 0.9683, + "epoch": 0.4247787610619469, + "grad_norm": 0.9673426126994554, + "learning_rate": 1.2880852233767174e-05, + "loss": 0.6037, "step": 1104 }, { - "epoch": 0.21, - "learning_rate": 1.8301472420901985e-05, - "loss": 0.9752, + "epoch": 0.425163524432474, + "grad_norm": 0.9436406747633878, + "learning_rate": 1.2868916623825561e-05, + "loss": 0.5963, "step": 1105 }, { - "epoch": 0.21, - "learning_rate": 1.8297996147605787e-05, - "loss": 0.9592, + "epoch": 0.42554828780300114, + "grad_norm": 0.9627154501436275, + "learning_rate": 1.2856976558635532e-05, + "loss": 0.6296, "step": 1106 }, { - "epoch": 0.21, - "learning_rate": 1.829451665146351e-05, - "loss": 0.9843, + "epoch": 0.42593305117352825, + "grad_norm": 0.6711090499473228, + "learning_rate": 1.2845032056739257e-05, + "loss": 0.6682, "step": 1107 }, { - "epoch": 0.21, - "learning_rate": 1.8291033933826535e-05, - "loss": 0.9664, + "epoch": 0.42631781454405543, + "grad_norm": 1.0498430827537013, + "learning_rate": 1.2833083136685803e-05, + "loss": 0.593, "step": 1108 }, { - "epoch": 0.21, - "learning_rate": 1.8287547996047523e-05, - "loss": 0.9879, + "epoch": 0.42670257791458255, + "grad_norm": 1.0010789129739008, + "learning_rate": 1.2821129817031099e-05, + "loss": 0.6423, "step": 1109 }, { - "epoch": 0.21, - "learning_rate": 1.8284058839480363e-05, - "loss": 0.9, + "epoch": 0.42708734128510967, + "grad_norm": 1.0253499792478094, + "learning_rate": 1.2809172116337903e-05, + "loss": 0.6249, "step": 1110 }, { - "epoch": 0.21, - "learning_rate": 1.8280566465480206e-05, - "loss": 0.9604, + "epoch": 0.4274721046556368, + "grad_norm": 0.9133990345107678, + "learning_rate": 1.2797210053175779e-05, + "loss": 0.6001, "step": 1111 }, { - "epoch": 0.21, - "learning_rate": 1.8277070875403455e-05, - "loss": 0.7576, + "epoch": 0.4278568680261639, + "grad_norm": 1.0738125796498017, + "learning_rate": 1.2785243646121059e-05, + "loss": 0.5903, "step": 1112 }, { - "epoch": 0.21, - "learning_rate": 1.8273572070607756e-05, - "loss": 0.9354, + "epoch": 0.428241631396691, + "grad_norm": 1.0037200507220319, + "learning_rate": 1.2773272913756833e-05, + "loss": 0.5846, "step": 1113 }, { - "epoch": 0.21, - "learning_rate": 1.8270070052451995e-05, - "loss": 0.9508, + "epoch": 0.42862639476721814, + "grad_norm": 0.9563351916811418, + "learning_rate": 1.27612978746729e-05, + "loss": 0.61, "step": 1114 }, { - "epoch": 0.21, - "learning_rate": 1.8266564822296323e-05, - "loss": 0.9235, + "epoch": 0.42901115813774526, + "grad_norm": 0.947683478357291, + "learning_rate": 1.2749318547465742e-05, + "loss": 0.6037, "step": 1115 }, { - "epoch": 0.21, - "learning_rate": 1.826305638150213e-05, - "loss": 0.7877, + "epoch": 0.42939592150827244, + "grad_norm": 0.5995111155302956, + "learning_rate": 1.2737334950738512e-05, + "loss": 0.6478, "step": 1116 }, { - "epoch": 0.21, - "learning_rate": 1.825954473143205e-05, - "loss": 0.7208, + "epoch": 0.42978068487879956, + "grad_norm": 1.0314684567922012, + "learning_rate": 1.272534710310099e-05, + "loss": 0.5855, "step": 1117 }, { - "epoch": 0.22, - "learning_rate": 1.8256029873449976e-05, - "loss": 0.9199, + "epoch": 0.4301654482493267, + "grad_norm": 0.5657528665649021, + "learning_rate": 1.2713355023169547e-05, + "loss": 0.628, "step": 1118 }, { - "epoch": 0.22, - "learning_rate": 1.825251180892103e-05, - "loss": 0.7574, + "epoch": 0.4305502116198538, + "grad_norm": 0.9305135962780334, + "learning_rate": 1.270135872956714e-05, + "loss": 0.5626, "step": 1119 }, { - "epoch": 0.22, - "learning_rate": 1.8248990539211596e-05, - "loss": 0.938, + "epoch": 0.4309349749903809, + "grad_norm": 1.1090078732980393, + "learning_rate": 1.2689358240923264e-05, + "loss": 0.6346, "step": 1120 }, { - "epoch": 0.22, - "learning_rate": 1.8245466065689282e-05, - "loss": 1.0455, + "epoch": 0.43131973836090803, + "grad_norm": 0.9710040349827014, + "learning_rate": 1.2677353575873926e-05, + "loss": 0.5651, "step": 1121 }, { - "epoch": 0.22, - "learning_rate": 1.824193838972297e-05, - "loss": 1.0092, + "epoch": 0.43170450173143515, + "grad_norm": 1.1866783101442708, + "learning_rate": 1.2665344753061622e-05, + "loss": 0.5996, "step": 1122 }, { - "epoch": 0.22, - "learning_rate": 1.823840751268275e-05, - "loss": 0.833, + "epoch": 0.43208926510196227, + "grad_norm": 0.9463279138706393, + "learning_rate": 1.2653331791135308e-05, + "loss": 0.5926, "step": 1123 }, { - "epoch": 0.22, - "learning_rate": 1.8234873435939987e-05, - "loss": 0.9177, + "epoch": 0.4324740284724894, + "grad_norm": 0.9434663241756062, + "learning_rate": 1.264131470875036e-05, + "loss": 0.6353, "step": 1124 }, { - "epoch": 0.22, - "learning_rate": 1.8231336160867275e-05, - "loss": 0.9397, + "epoch": 0.43285879184301657, + "grad_norm": 1.020508386081434, + "learning_rate": 1.2629293524568555e-05, + "loss": 0.6232, "step": 1125 }, { - "epoch": 0.22, - "learning_rate": 1.8227795688838446e-05, - "loss": 0.9467, + "epoch": 0.4332435552135437, + "grad_norm": 0.9882345841648764, + "learning_rate": 1.2617268257258051e-05, + "loss": 0.6611, "step": 1126 }, { - "epoch": 0.22, - "learning_rate": 1.822425202122858e-05, - "loss": 0.894, + "epoch": 0.4336283185840708, + "grad_norm": 0.9302733360130826, + "learning_rate": 1.2605238925493326e-05, + "loss": 0.6699, "step": 1127 }, { - "epoch": 0.22, - "learning_rate": 1.8220705159413996e-05, - "loss": 0.8698, + "epoch": 0.4340130819545979, + "grad_norm": 0.9257498042079934, + "learning_rate": 1.2593205547955185e-05, + "loss": 0.5448, "step": 1128 }, { - "epoch": 0.22, - "learning_rate": 1.8217155104772256e-05, - "loss": 1.0054, + "epoch": 0.43439784532512504, + "grad_norm": 1.0000251031088303, + "learning_rate": 1.2581168143330716e-05, + "loss": 0.6019, "step": 1129 }, { - "epoch": 0.22, - "learning_rate": 1.8213601858682158e-05, - "loss": 0.8102, + "epoch": 0.43478260869565216, + "grad_norm": 0.9334376774754681, + "learning_rate": 1.2569126730313255e-05, + "loss": 0.6517, "step": 1130 }, { - "epoch": 0.22, - "learning_rate": 1.8210045422523744e-05, - "loss": 0.9318, + "epoch": 0.4351673720661793, + "grad_norm": 0.9894164852219632, + "learning_rate": 1.2557081327602361e-05, + "loss": 0.558, "step": 1131 }, { - "epoch": 0.22, - "learning_rate": 1.8206485797678294e-05, - "loss": 0.7472, + "epoch": 0.4355521354367064, + "grad_norm": 1.016570884475846, + "learning_rate": 1.2545031953903796e-05, + "loss": 0.6031, "step": 1132 }, { - "epoch": 0.22, - "learning_rate": 1.820292298552832e-05, - "loss": 1.0124, + "epoch": 0.4359368988072336, + "grad_norm": 0.9078131942961885, + "learning_rate": 1.2532978627929486e-05, + "loss": 0.5383, "step": 1133 }, { - "epoch": 0.22, - "learning_rate": 1.819935698745759e-05, - "loss": 0.8755, + "epoch": 0.4363216621777607, + "grad_norm": 0.9799546104504018, + "learning_rate": 1.2520921368397492e-05, + "loss": 0.5807, "step": 1134 }, { - "epoch": 0.22, - "learning_rate": 1.8195787804851076e-05, - "loss": 0.9422, + "epoch": 0.4367064255482878, + "grad_norm": 1.0006933378642524, + "learning_rate": 1.2508860194031986e-05, + "loss": 0.5712, "step": 1135 }, { - "epoch": 0.22, - "learning_rate": 1.8192215439095025e-05, - "loss": 0.9305, + "epoch": 0.43709118891881493, + "grad_norm": 0.9947704934688096, + "learning_rate": 1.2496795123563218e-05, + "loss": 0.5944, "step": 1136 }, { - "epoch": 0.22, - "learning_rate": 1.8188639891576893e-05, - "loss": 0.8822, + "epoch": 0.43747595228934205, + "grad_norm": 0.8922937389867284, + "learning_rate": 1.248472617572749e-05, + "loss": 0.5484, "step": 1137 }, { - "epoch": 0.22, - "learning_rate": 1.8185061163685386e-05, - "loss": 0.9151, + "epoch": 0.43786071565986917, + "grad_norm": 0.9297885934371464, + "learning_rate": 1.2472653369267122e-05, + "loss": 0.5508, "step": 1138 }, { - "epoch": 0.22, - "learning_rate": 1.818147925681044e-05, - "loss": 1.0419, + "epoch": 0.4382454790303963, + "grad_norm": 0.9606027458530547, + "learning_rate": 1.2460576722930432e-05, + "loss": 0.5859, "step": 1139 }, { - "epoch": 0.22, - "learning_rate": 1.8177894172343227e-05, - "loss": 0.9766, + "epoch": 0.4386302424009234, + "grad_norm": 0.9316645327083906, + "learning_rate": 1.24484962554717e-05, + "loss": 0.598, "step": 1140 }, { - "epoch": 0.22, - "learning_rate": 1.817430591167615e-05, - "loss": 0.9324, + "epoch": 0.4390150057714506, + "grad_norm": 0.8955275191886756, + "learning_rate": 1.2436411985651131e-05, + "loss": 0.5844, "step": 1141 }, { - "epoch": 0.22, - "learning_rate": 1.8170714476202848e-05, - "loss": 0.9603, + "epoch": 0.4393997691419777, + "grad_norm": 0.8738379220219646, + "learning_rate": 1.242432393223485e-05, + "loss": 0.5612, "step": 1142 }, { - "epoch": 0.22, - "learning_rate": 1.8167119867318197e-05, - "loss": 0.7765, + "epoch": 0.4397845325125048, + "grad_norm": 0.9872972149442232, + "learning_rate": 1.2412232113994841e-05, + "loss": 0.6175, "step": 1143 }, { - "epoch": 0.22, - "learning_rate": 1.81635220864183e-05, - "loss": 0.9342, + "epoch": 0.44016929588303194, + "grad_norm": 1.0335330976132246, + "learning_rate": 1.2400136549708945e-05, + "loss": 0.6253, "step": 1144 }, { - "epoch": 0.22, - "learning_rate": 1.8159921134900486e-05, - "loss": 0.9766, + "epoch": 0.44055405925355906, + "grad_norm": 0.9926260463563474, + "learning_rate": 1.2388037258160823e-05, + "loss": 0.6288, "step": 1145 }, { - "epoch": 0.22, - "learning_rate": 1.8156317014163337e-05, - "loss": 1.0026, + "epoch": 0.4409388226240862, + "grad_norm": 0.8572053763141091, + "learning_rate": 1.2375934258139917e-05, + "loss": 0.5315, "step": 1146 }, { - "epoch": 0.22, - "learning_rate": 1.8152709725606642e-05, - "loss": 0.9895, + "epoch": 0.4413235859946133, + "grad_norm": 1.0276002705935736, + "learning_rate": 1.236382756844143e-05, + "loss": 0.5645, "step": 1147 }, { - "epoch": 0.22, - "learning_rate": 1.8149099270631434e-05, - "loss": 0.8908, + "epoch": 0.4417083493651404, + "grad_norm": 0.8935506027957324, + "learning_rate": 1.2351717207866292e-05, + "loss": 0.6268, "step": 1148 }, { - "epoch": 0.22, - "learning_rate": 1.8145485650639973e-05, - "loss": 0.7658, + "epoch": 0.44209311273566754, + "grad_norm": 0.9448177420928556, + "learning_rate": 1.233960319522114e-05, + "loss": 0.6514, "step": 1149 }, { - "epoch": 0.22, - "learning_rate": 1.8141868867035745e-05, - "loss": 0.9783, + "epoch": 0.4424778761061947, + "grad_norm": 0.9617812162979757, + "learning_rate": 1.2327485549318285e-05, + "loss": 0.5665, "step": 1150 }, { - "epoch": 0.22, - "learning_rate": 1.8138248921223465e-05, - "loss": 0.903, + "epoch": 0.44286263947672183, + "grad_norm": 0.9265524412818427, + "learning_rate": 1.2315364288975665e-05, + "loss": 0.5894, "step": 1151 }, { - "epoch": 0.22, - "learning_rate": 1.8134625814609084e-05, - "loss": 0.7725, + "epoch": 0.44324740284724895, + "grad_norm": 0.9619038299952846, + "learning_rate": 1.2303239433016842e-05, + "loss": 0.61, "step": 1152 }, { - "epoch": 0.22, - "learning_rate": 1.8130999548599767e-05, - "loss": 0.9945, + "epoch": 0.44363216621777607, + "grad_norm": 0.9804197592550914, + "learning_rate": 1.229111100027097e-05, + "loss": 0.5905, "step": 1153 }, { - "epoch": 0.22, - "learning_rate": 1.8127370124603927e-05, - "loss": 1.005, + "epoch": 0.4440169295883032, + "grad_norm": 1.1516200350465249, + "learning_rate": 1.2278979009572736e-05, + "loss": 0.6468, "step": 1154 }, { - "epoch": 0.22, - "learning_rate": 1.8123737544031178e-05, - "loss": 0.9216, + "epoch": 0.4444016929588303, + "grad_norm": 1.042877254013063, + "learning_rate": 1.2266843479762372e-05, + "loss": 0.5766, "step": 1155 }, { - "epoch": 0.22, - "learning_rate": 1.8120101808292373e-05, - "loss": 0.9178, + "epoch": 0.4447864563293574, + "grad_norm": 0.9560859155394253, + "learning_rate": 1.2254704429685593e-05, + "loss": 0.5425, "step": 1156 }, { - "epoch": 0.22, - "learning_rate": 1.81164629187996e-05, - "loss": 1.0067, + "epoch": 0.44517121969988455, + "grad_norm": 0.9081704251263675, + "learning_rate": 1.2242561878193589e-05, + "loss": 0.5584, "step": 1157 }, { - "epoch": 0.22, - "learning_rate": 1.811282087696615e-05, - "loss": 0.9355, + "epoch": 0.4455559830704117, + "grad_norm": 0.902606625021449, + "learning_rate": 1.2230415844142984e-05, + "loss": 0.5437, "step": 1158 }, { - "epoch": 0.22, - "learning_rate": 1.8109175684206558e-05, - "loss": 0.9131, + "epoch": 0.44594074644093884, + "grad_norm": 0.9781558650235215, + "learning_rate": 1.2218266346395811e-05, + "loss": 0.5958, "step": 1159 }, { - "epoch": 0.22, - "learning_rate": 1.8105527341936574e-05, - "loss": 1.02, + "epoch": 0.44632550981146596, + "grad_norm": 0.9751134902531522, + "learning_rate": 1.220611340381948e-05, + "loss": 0.6095, "step": 1160 }, { - "epoch": 0.22, - "learning_rate": 1.810187585157317e-05, - "loss": 0.7155, + "epoch": 0.4467102731819931, + "grad_norm": 0.8843341708731247, + "learning_rate": 1.2193957035286757e-05, + "loss": 0.4896, "step": 1161 }, { - "epoch": 0.22, - "learning_rate": 1.8098221214534543e-05, - "loss": 0.7312, + "epoch": 0.4470950365525202, + "grad_norm": 0.9345355587026692, + "learning_rate": 1.2181797259675713e-05, + "loss": 0.5912, "step": 1162 }, { - "epoch": 0.22, - "learning_rate": 1.8094563432240107e-05, - "loss": 0.9999, + "epoch": 0.4474797999230473, + "grad_norm": 0.9666556476804246, + "learning_rate": 1.2169634095869736e-05, + "loss": 0.5512, "step": 1163 }, { - "epoch": 0.22, - "learning_rate": 1.8090902506110513e-05, - "loss": 0.9599, + "epoch": 0.44786456329357444, + "grad_norm": 0.9718580577692861, + "learning_rate": 1.2157467562757443e-05, + "loss": 0.5976, "step": 1164 }, { - "epoch": 0.22, - "learning_rate": 1.8087238437567614e-05, - "loss": 0.9555, + "epoch": 0.44824932666410156, + "grad_norm": 1.0285637004854273, + "learning_rate": 1.214529767923271e-05, + "loss": 0.6042, "step": 1165 }, { - "epoch": 0.22, - "learning_rate": 1.8083571228034498e-05, - "loss": 0.9385, + "epoch": 0.44863409003462873, + "grad_norm": 0.9043257669575927, + "learning_rate": 1.213312446419461e-05, + "loss": 0.5564, "step": 1166 }, { - "epoch": 0.22, - "learning_rate": 1.807990087893546e-05, - "loss": 0.8903, + "epoch": 0.44901885340515585, + "grad_norm": 1.027575107872667, + "learning_rate": 1.2120947936547375e-05, + "loss": 0.6077, "step": 1167 }, { - "epoch": 0.22, - "learning_rate": 1.807622739169603e-05, - "loss": 0.9744, + "epoch": 0.44940361677568297, + "grad_norm": 1.0130515371356847, + "learning_rate": 1.2108768115200405e-05, + "loss": 0.569, "step": 1168 }, { - "epoch": 0.22, - "learning_rate": 1.807255076774294e-05, - "loss": 0.7973, + "epoch": 0.4497883801462101, + "grad_norm": 0.9067547976420515, + "learning_rate": 1.209658501906819e-05, + "loss": 0.6173, "step": 1169 }, { - "epoch": 0.23, - "learning_rate": 1.8068871008504153e-05, - "loss": 1.0076, + "epoch": 0.4501731435167372, + "grad_norm": 1.0113770337805366, + "learning_rate": 1.2084398667070325e-05, + "loss": 0.5987, "step": 1170 }, { - "epoch": 0.23, - "learning_rate": 1.8065188115408844e-05, - "loss": 1.0206, + "epoch": 0.4505579068872643, + "grad_norm": 1.008145004002449, + "learning_rate": 1.2072209078131451e-05, + "loss": 0.591, "step": 1171 }, { - "epoch": 0.23, - "learning_rate": 1.8061502089887406e-05, - "loss": 0.9216, + "epoch": 0.45094267025779144, + "grad_norm": 0.9814440257568853, + "learning_rate": 1.206001627118124e-05, + "loss": 0.5753, "step": 1172 }, { - "epoch": 0.23, - "learning_rate": 1.805781293337145e-05, - "loss": 0.9286, + "epoch": 0.45132743362831856, + "grad_norm": 1.011400677102962, + "learning_rate": 1.2047820265154362e-05, + "loss": 0.6385, "step": 1173 }, { - "epoch": 0.23, - "learning_rate": 1.8054120647293798e-05, - "loss": 0.9302, + "epoch": 0.4517121969988457, + "grad_norm": 1.0090792671299567, + "learning_rate": 1.203562107899045e-05, + "loss": 0.6283, "step": 1174 }, { - "epoch": 0.23, - "learning_rate": 1.8050425233088496e-05, - "loss": 0.7633, + "epoch": 0.45209696036937286, + "grad_norm": 0.6184820154633748, + "learning_rate": 1.2023418731634078e-05, + "loss": 0.6544, "step": 1175 }, { - "epoch": 0.23, - "learning_rate": 1.80467266921908e-05, - "loss": 0.9137, + "epoch": 0.4524817237399, + "grad_norm": 0.9791016743504568, + "learning_rate": 1.2011213242034733e-05, + "loss": 0.6028, "step": 1176 }, { - "epoch": 0.23, - "learning_rate": 1.8043025026037178e-05, - "loss": 0.9435, + "epoch": 0.4528664871104271, + "grad_norm": 1.0946428832841761, + "learning_rate": 1.1999004629146775e-05, + "loss": 0.6245, "step": 1177 }, { - "epoch": 0.23, - "learning_rate": 1.8039320236065314e-05, - "loss": 0.9154, + "epoch": 0.4532512504809542, + "grad_norm": 1.0108975355946759, + "learning_rate": 1.1986792911929418e-05, + "loss": 0.6315, "step": 1178 }, { - "epoch": 0.23, - "learning_rate": 1.803561232371411e-05, - "loss": 0.9685, + "epoch": 0.45363601385148133, + "grad_norm": 1.0258828086178098, + "learning_rate": 1.1974578109346702e-05, + "loss": 0.6512, "step": 1179 }, { - "epoch": 0.23, - "learning_rate": 1.803190129042367e-05, - "loss": 0.9296, + "epoch": 0.45402077722200845, + "grad_norm": 0.8922725443725109, + "learning_rate": 1.1962360240367445e-05, + "loss": 0.5863, "step": 1180 }, { - "epoch": 0.23, - "learning_rate": 1.8028187137635325e-05, - "loss": 0.957, + "epoch": 0.4544055405925356, + "grad_norm": 1.011116369820562, + "learning_rate": 1.195013932396524e-05, + "loss": 0.6178, "step": 1181 }, { - "epoch": 0.23, - "learning_rate": 1.8024469866791602e-05, - "loss": 1.0574, + "epoch": 0.4547903039630627, + "grad_norm": 1.1115765478945283, + "learning_rate": 1.1937915379118406e-05, + "loss": 0.5751, "step": 1182 }, { - "epoch": 0.23, - "learning_rate": 1.802074947933625e-05, - "loss": 0.9158, + "epoch": 0.45517506733358987, + "grad_norm": 1.0057300936229783, + "learning_rate": 1.1925688424809965e-05, + "loss": 0.591, "step": 1183 }, { - "epoch": 0.23, - "learning_rate": 1.801702597671422e-05, - "loss": 0.7373, + "epoch": 0.455559830704117, + "grad_norm": 0.9654680424237339, + "learning_rate": 1.1913458480027614e-05, + "loss": 0.5858, "step": 1184 }, { - "epoch": 0.23, - "learning_rate": 1.8013299360371685e-05, - "loss": 0.8765, + "epoch": 0.4559445940746441, + "grad_norm": 0.9964045053509739, + "learning_rate": 1.1901225563763694e-05, + "loss": 0.6221, "step": 1185 }, { - "epoch": 0.23, - "learning_rate": 1.8009569631756013e-05, - "loss": 0.9983, + "epoch": 0.4563293574451712, + "grad_norm": 0.9527880016633828, + "learning_rate": 1.1888989695015166e-05, + "loss": 0.6403, "step": 1186 }, { - "epoch": 0.23, - "learning_rate": 1.8005836792315793e-05, - "loss": 0.9068, + "epoch": 0.45671412081569834, + "grad_norm": 1.0398146267632802, + "learning_rate": 1.1876750892783558e-05, + "loss": 0.642, "step": 1187 }, { - "epoch": 0.23, - "learning_rate": 1.800210084350081e-05, - "loss": 0.9591, + "epoch": 0.45709888418622546, + "grad_norm": 1.1291556909766802, + "learning_rate": 1.1864509176074974e-05, + "loss": 0.5631, "step": 1188 }, { - "epoch": 0.23, - "learning_rate": 1.799836178676207e-05, - "loss": 0.9093, + "epoch": 0.4574836475567526, + "grad_norm": 0.8783870632013164, + "learning_rate": 1.1852264563900038e-05, + "loss": 0.523, "step": 1189 }, { - "epoch": 0.23, - "learning_rate": 1.799461962355178e-05, - "loss": 0.9758, + "epoch": 0.4578684109272797, + "grad_norm": 1.0230356394172506, + "learning_rate": 1.1840017075273861e-05, + "loss": 0.6267, "step": 1190 }, { - "epoch": 0.23, - "learning_rate": 1.7990874355323345e-05, - "loss": 0.9352, + "epoch": 0.4582531742978069, + "grad_norm": 0.9442538346085079, + "learning_rate": 1.1827766729216035e-05, + "loss": 0.595, "step": 1191 }, { - "epoch": 0.23, - "learning_rate": 1.7987125983531393e-05, - "loss": 0.9007, + "epoch": 0.458637937668334, + "grad_norm": 0.9002283451487824, + "learning_rate": 1.181551354475058e-05, + "loss": 0.5188, "step": 1192 }, { - "epoch": 0.23, - "learning_rate": 1.7983374509631742e-05, - "loss": 0.9503, + "epoch": 0.4590227010388611, + "grad_norm": 1.0436724277897131, + "learning_rate": 1.1803257540905926e-05, + "loss": 0.5984, "step": 1193 }, { - "epoch": 0.23, - "learning_rate": 1.7979619935081424e-05, - "loss": 0.9655, + "epoch": 0.45940746440938823, + "grad_norm": 0.6312786017998998, + "learning_rate": 1.1790998736714882e-05, + "loss": 0.6411, "step": 1194 }, { - "epoch": 0.23, - "learning_rate": 1.797586226133867e-05, - "loss": 0.9357, + "epoch": 0.45979222777991535, + "grad_norm": 0.9712943820707499, + "learning_rate": 1.1778737151214606e-05, + "loss": 0.5577, "step": 1195 }, { - "epoch": 0.23, - "learning_rate": 1.7972101489862924e-05, - "loss": 0.8935, + "epoch": 0.46017699115044247, + "grad_norm": 1.1217317414187056, + "learning_rate": 1.1766472803446577e-05, + "loss": 0.6617, "step": 1196 }, { - "epoch": 0.23, - "learning_rate": 1.7968337622114824e-05, - "loss": 0.9221, + "epoch": 0.4605617545209696, + "grad_norm": 0.9651962875609242, + "learning_rate": 1.1754205712456556e-05, + "loss": 0.5802, "step": 1197 }, { - "epoch": 0.23, - "learning_rate": 1.7964570659556206e-05, - "loss": 0.9322, + "epoch": 0.4609465178914967, + "grad_norm": 0.908916449224402, + "learning_rate": 1.1741935897294572e-05, + "loss": 0.5487, "step": 1198 }, { - "epoch": 0.23, - "learning_rate": 1.796080060365012e-05, - "loss": 0.9718, + "epoch": 0.46133128126202383, + "grad_norm": 0.9758709136508231, + "learning_rate": 1.1729663377014888e-05, + "loss": 0.581, "step": 1199 }, { - "epoch": 0.23, - "learning_rate": 1.7957027455860815e-05, - "loss": 0.9608, + "epoch": 0.461716044632551, + "grad_norm": 0.8524533188542169, + "learning_rate": 1.1717388170675954e-05, + "loss": 0.526, "step": 1200 }, { - "epoch": 0.23, - "learning_rate": 1.795325121765373e-05, - "loss": 0.8461, + "epoch": 0.4621008080030781, + "grad_norm": 0.9603920422117519, + "learning_rate": 1.17051102973404e-05, + "loss": 0.571, "step": 1201 }, { - "epoch": 0.23, - "learning_rate": 1.794947189049552e-05, - "loss": 0.7377, + "epoch": 0.46248557137360524, + "grad_norm": 0.9651699589060333, + "learning_rate": 1.1692829776074999e-05, + "loss": 0.5719, "step": 1202 }, { - "epoch": 0.23, - "learning_rate": 1.7945689475854033e-05, - "loss": 0.9393, + "epoch": 0.46287033474413236, + "grad_norm": 0.9114570570585456, + "learning_rate": 1.1680546625950635e-05, + "loss": 0.5773, "step": 1203 }, { - "epoch": 0.23, - "learning_rate": 1.7941903975198305e-05, - "loss": 0.9035, + "epoch": 0.4632550981146595, + "grad_norm": 0.968373691849325, + "learning_rate": 1.1668260866042271e-05, + "loss": 0.6423, "step": 1204 }, { - "epoch": 0.23, - "learning_rate": 1.7938115389998595e-05, - "loss": 0.9846, + "epoch": 0.4636398614851866, + "grad_norm": 0.9644010136547577, + "learning_rate": 1.1655972515428928e-05, + "loss": 0.5645, "step": 1205 }, { - "epoch": 0.23, - "learning_rate": 1.7934323721726334e-05, - "loss": 0.9437, + "epoch": 0.4640246248557137, + "grad_norm": 0.9338376682208793, + "learning_rate": 1.1643681593193642e-05, + "loss": 0.5987, "step": 1206 }, { - "epoch": 0.23, - "learning_rate": 1.7930528971854166e-05, - "loss": 0.9713, + "epoch": 0.46440938822624084, + "grad_norm": 1.0059928103743148, + "learning_rate": 1.1631388118423457e-05, + "loss": 0.6472, "step": 1207 }, { - "epoch": 0.23, - "learning_rate": 1.792673114185593e-05, - "loss": 0.9324, + "epoch": 0.464794151596768, + "grad_norm": 0.9490438972229435, + "learning_rate": 1.1619092110209361e-05, + "loss": 0.5833, "step": 1208 }, { - "epoch": 0.23, - "learning_rate": 1.7922930233206656e-05, - "loss": 0.9817, + "epoch": 0.46517891496729513, + "grad_norm": 0.9435984997886737, + "learning_rate": 1.1606793587646295e-05, + "loss": 0.5616, "step": 1209 }, { - "epoch": 0.23, - "learning_rate": 1.7919126247382576e-05, - "loss": 0.9459, + "epoch": 0.46556367833782225, + "grad_norm": 1.0267141644982831, + "learning_rate": 1.1594492569833093e-05, + "loss": 0.6199, "step": 1210 }, { - "epoch": 0.23, - "learning_rate": 1.791531918586112e-05, - "loss": 0.9239, + "epoch": 0.46594844170834937, + "grad_norm": 0.9066333923569397, + "learning_rate": 1.1582189075872467e-05, + "loss": 0.6203, "step": 1211 }, { - "epoch": 0.23, - "learning_rate": 1.7911509050120892e-05, - "loss": 1.0274, + "epoch": 0.4663332050788765, + "grad_norm": 0.9134487454767828, + "learning_rate": 1.156988312487098e-05, + "loss": 0.5806, "step": 1212 }, { - "epoch": 0.23, - "learning_rate": 1.7907695841641716e-05, - "loss": 0.8685, + "epoch": 0.4667179684494036, + "grad_norm": 0.9523202474563607, + "learning_rate": 1.1557574735939003e-05, + "loss": 0.5911, "step": 1213 }, { - "epoch": 0.23, - "learning_rate": 1.7903879561904597e-05, - "loss": 0.9373, + "epoch": 0.4671027318199307, + "grad_norm": 0.8830522168962559, + "learning_rate": 1.1545263928190692e-05, + "loss": 0.6058, "step": 1214 }, { - "epoch": 0.23, - "learning_rate": 1.790006021239173e-05, - "loss": 0.8849, + "epoch": 0.46748749519045785, + "grad_norm": 0.9473763525754578, + "learning_rate": 1.153295072074397e-05, + "loss": 0.588, "step": 1215 }, { - "epoch": 0.23, - "learning_rate": 1.789623779458651e-05, - "loss": 0.7658, + "epoch": 0.467872258560985, + "grad_norm": 0.8952063747831213, + "learning_rate": 1.1520635132720475e-05, + "loss": 0.5617, "step": 1216 }, { - "epoch": 0.23, - "learning_rate": 1.789241230997352e-05, - "loss": 1.0675, + "epoch": 0.46825702193151214, + "grad_norm": 1.0036390852214017, + "learning_rate": 1.1508317183245545e-05, + "loss": 0.6233, "step": 1217 }, { - "epoch": 0.23, - "learning_rate": 1.7888583760038534e-05, - "loss": 0.8992, + "epoch": 0.46864178530203926, + "grad_norm": 0.8967179443111969, + "learning_rate": 1.149599689144819e-05, + "loss": 0.5638, "step": 1218 }, { - "epoch": 0.23, - "learning_rate": 1.7884752146268513e-05, - "loss": 0.9084, + "epoch": 0.4690265486725664, + "grad_norm": 0.8902866136878852, + "learning_rate": 1.1483674276461053e-05, + "loss": 0.5733, "step": 1219 }, { - "epoch": 0.23, - "learning_rate": 1.7880917470151614e-05, - "loss": 0.9639, + "epoch": 0.4694113120430935, + "grad_norm": 1.025569332935193, + "learning_rate": 1.1471349357420384e-05, + "loss": 0.5662, "step": 1220 }, { - "epoch": 0.23, - "learning_rate": 1.7877079733177185e-05, - "loss": 0.8697, + "epoch": 0.4697960754136206, + "grad_norm": 1.068895791812618, + "learning_rate": 1.1459022153466016e-05, + "loss": 0.5821, "step": 1221 }, { - "epoch": 0.24, - "learning_rate": 1.7873238936835754e-05, - "loss": 0.9049, + "epoch": 0.47018083878414774, + "grad_norm": 0.9042792049396877, + "learning_rate": 1.1446692683741326e-05, + "loss": 0.5616, "step": 1222 }, { - "epoch": 0.24, - "learning_rate": 1.786939508261904e-05, - "loss": 0.9179, + "epoch": 0.47056560215467486, + "grad_norm": 0.9338499376406343, + "learning_rate": 1.143436096739321e-05, + "loss": 0.5754, "step": 1223 }, { - "epoch": 0.24, - "learning_rate": 1.786554817201996e-05, - "loss": 0.9025, + "epoch": 0.470950365525202, + "grad_norm": 0.902450972825043, + "learning_rate": 1.1422027023572052e-05, + "loss": 0.5829, "step": 1224 }, { - "epoch": 0.24, - "learning_rate": 1.78616982065326e-05, - "loss": 0.7376, + "epoch": 0.47133512889572915, + "grad_norm": 1.0050040135507645, + "learning_rate": 1.14096908714317e-05, + "loss": 0.6135, "step": 1225 }, { - "epoch": 0.24, - "learning_rate": 1.785784518765225e-05, - "loss": 0.9199, + "epoch": 0.47171989226625627, + "grad_norm": 1.1335836568683537, + "learning_rate": 1.1397352530129428e-05, + "loss": 0.5628, "step": 1226 }, { - "epoch": 0.24, - "learning_rate": 1.7853989116875373e-05, - "loss": 0.7615, + "epoch": 0.4721046556367834, + "grad_norm": 0.8711088536267853, + "learning_rate": 1.1385012018825907e-05, + "loss": 0.5641, "step": 1227 }, { - "epoch": 0.24, - "learning_rate": 1.7850129995699626e-05, - "loss": 0.912, + "epoch": 0.4724894190073105, + "grad_norm": 0.9545529705947633, + "learning_rate": 1.1372669356685185e-05, + "loss": 0.6194, "step": 1228 }, { - "epoch": 0.24, - "learning_rate": 1.7846267825623843e-05, - "loss": 1.0531, + "epoch": 0.4728741823778376, + "grad_norm": 1.0464935675557214, + "learning_rate": 1.1360324562874643e-05, + "loss": 0.6153, "step": 1229 }, { - "epoch": 0.24, - "learning_rate": 1.7842402608148053e-05, - "loss": 1.0078, + "epoch": 0.47325894574836475, + "grad_norm": 0.9569793691070418, + "learning_rate": 1.1347977656564974e-05, + "loss": 0.5411, "step": 1230 }, { - "epoch": 0.24, - "learning_rate": 1.7838534344773453e-05, - "loss": 0.9106, + "epoch": 0.47364370911889186, + "grad_norm": 0.9815447671063015, + "learning_rate": 1.1335628656930153e-05, + "loss": 0.6024, "step": 1231 }, { - "epoch": 0.24, - "learning_rate": 1.7834663037002444e-05, - "loss": 0.9759, + "epoch": 0.474028472489419, + "grad_norm": 0.9742010718779374, + "learning_rate": 1.132327758314741e-05, + "loss": 0.6318, "step": 1232 }, { - "epoch": 0.24, - "learning_rate": 1.7830788686338586e-05, - "loss": 0.9356, + "epoch": 0.47441323585994616, + "grad_norm": 1.0476249434882017, + "learning_rate": 1.1310924454397187e-05, + "loss": 0.5945, "step": 1233 }, { - "epoch": 0.24, - "learning_rate": 1.7826911294286636e-05, - "loss": 0.8925, + "epoch": 0.4747979992304733, + "grad_norm": 0.9112661959984337, + "learning_rate": 1.129856928986312e-05, + "loss": 0.5505, "step": 1234 }, { - "epoch": 0.24, - "learning_rate": 1.782303086235253e-05, - "loss": 0.909, + "epoch": 0.4751827626010004, + "grad_norm": 0.9510405891573328, + "learning_rate": 1.1286212108732015e-05, + "loss": 0.5376, "step": 1235 }, { - "epoch": 0.24, - "learning_rate": 1.781914739204338e-05, - "loss": 0.8274, + "epoch": 0.4755675259715275, + "grad_norm": 0.676863429005373, + "learning_rate": 1.1273852930193798e-05, + "loss": 0.662, "step": 1236 }, { - "epoch": 0.24, - "learning_rate": 1.7815260884867486e-05, - "loss": 0.9412, + "epoch": 0.47595228934205464, + "grad_norm": 1.0162641197026765, + "learning_rate": 1.12614917734415e-05, + "loss": 0.5654, "step": 1237 }, { - "epoch": 0.24, - "learning_rate": 1.781137134233432e-05, - "loss": 1.0239, + "epoch": 0.47633705271258175, + "grad_norm": 0.8809176817864317, + "learning_rate": 1.1249128657671233e-05, + "loss": 0.5487, "step": 1238 }, { - "epoch": 0.24, - "learning_rate": 1.7807478765954532e-05, - "loss": 0.9482, + "epoch": 0.4767218160831089, + "grad_norm": 0.9368415133196434, + "learning_rate": 1.1236763602082136e-05, + "loss": 0.6056, "step": 1239 }, { - "epoch": 0.24, - "learning_rate": 1.7803583157239958e-05, - "loss": 0.9661, + "epoch": 0.477106579453636, + "grad_norm": 1.0337938014929193, + "learning_rate": 1.1224396625876375e-05, + "loss": 0.5811, "step": 1240 }, { - "epoch": 0.24, - "learning_rate": 1.7799684517703605e-05, - "loss": 0.8806, + "epoch": 0.4774913428241631, + "grad_norm": 0.9851212931173388, + "learning_rate": 1.1212027748259086e-05, + "loss": 0.6306, "step": 1241 }, { - "epoch": 0.24, - "learning_rate": 1.779578284885966e-05, - "loss": 0.7026, + "epoch": 0.4778761061946903, + "grad_norm": 0.9107409951403038, + "learning_rate": 1.1199656988438373e-05, + "loss": 0.5371, "step": 1242 }, { - "epoch": 0.24, - "learning_rate": 1.779187815222349e-05, - "loss": 0.8873, + "epoch": 0.4782608695652174, + "grad_norm": 0.8733825337371368, + "learning_rate": 1.1187284365625241e-05, + "loss": 0.5335, "step": 1243 }, { - "epoch": 0.24, - "learning_rate": 1.778797042931163e-05, - "loss": 0.894, + "epoch": 0.4786456329357445, + "grad_norm": 0.9316011913244577, + "learning_rate": 1.1174909899033608e-05, + "loss": 0.5887, "step": 1244 }, { - "epoch": 0.24, - "learning_rate": 1.7784059681641798e-05, - "loss": 1.0292, + "epoch": 0.47903039630627164, + "grad_norm": 0.969075358225214, + "learning_rate": 1.1162533607880251e-05, + "loss": 0.5858, "step": 1245 }, { - "epoch": 0.24, - "learning_rate": 1.778014591073288e-05, - "loss": 0.9531, + "epoch": 0.47941515967679876, + "grad_norm": 0.9480796284930808, + "learning_rate": 1.1150155511384772e-05, + "loss": 0.5582, "step": 1246 }, { - "epoch": 0.24, - "learning_rate": 1.777622911810494e-05, - "loss": 0.8743, + "epoch": 0.4797999230473259, + "grad_norm": 0.9579510280165476, + "learning_rate": 1.1137775628769584e-05, + "loss": 0.6501, "step": 1247 }, { - "epoch": 0.24, - "learning_rate": 1.777230930527922e-05, - "loss": 0.9118, + "epoch": 0.480184686417853, + "grad_norm": 1.0191215881683677, + "learning_rate": 1.1125393979259874e-05, + "loss": 0.593, "step": 1248 }, { - "epoch": 0.24, - "learning_rate": 1.7768386473778124e-05, - "loss": 0.9626, + "epoch": 0.4805694497883801, + "grad_norm": 0.964618109546779, + "learning_rate": 1.1113010582083568e-05, + "loss": 0.5614, "step": 1249 }, { - "epoch": 0.24, - "learning_rate": 1.7764460625125236e-05, - "loss": 0.8778, + "epoch": 0.4809542131589073, + "grad_norm": 0.6141018289349331, + "learning_rate": 1.1100625456471307e-05, + "loss": 0.6364, "step": 1250 }, { - "epoch": 0.24, - "learning_rate": 1.776053176084531e-05, - "loss": 0.9628, + "epoch": 0.4813389765294344, + "grad_norm": 0.9553294572644498, + "learning_rate": 1.1088238621656422e-05, + "loss": 0.5341, "step": 1251 }, { - "epoch": 0.24, - "learning_rate": 1.7756599882464274e-05, - "loss": 1.0529, + "epoch": 0.48172373989996153, + "grad_norm": 0.949865151204418, + "learning_rate": 1.1075850096874894e-05, + "loss": 0.5673, "step": 1252 }, { - "epoch": 0.24, - "learning_rate": 1.7752664991509224e-05, - "loss": 0.9001, + "epoch": 0.48210850327048865, + "grad_norm": 0.6266156415782722, + "learning_rate": 1.1063459901365325e-05, + "loss": 0.702, "step": 1253 }, { - "epoch": 0.24, - "learning_rate": 1.7748727089508423e-05, - "loss": 0.9023, + "epoch": 0.48249326664101577, + "grad_norm": 1.120247880538404, + "learning_rate": 1.1051068054368921e-05, + "loss": 0.6074, "step": 1254 }, { - "epoch": 0.24, - "learning_rate": 1.7744786177991307e-05, - "loss": 0.7641, + "epoch": 0.4828780300115429, + "grad_norm": 0.9263182713490261, + "learning_rate": 1.1038674575129442e-05, + "loss": 0.5928, "step": 1255 }, { - "epoch": 0.24, - "learning_rate": 1.774084225848849e-05, - "loss": 0.9588, + "epoch": 0.48326279338207, + "grad_norm": 1.0280749405921419, + "learning_rate": 1.1026279482893187e-05, + "loss": 0.5608, "step": 1256 }, { - "epoch": 0.24, - "learning_rate": 1.773689533253173e-05, - "loss": 0.8955, + "epoch": 0.48364755675259713, + "grad_norm": 1.0442615780615099, + "learning_rate": 1.1013882796908963e-05, + "loss": 0.601, "step": 1257 }, { - "epoch": 0.24, - "learning_rate": 1.7732945401653978e-05, - "loss": 0.942, + "epoch": 0.4840323201231243, + "grad_norm": 0.9738708064626824, + "learning_rate": 1.1001484536428052e-05, + "loss": 0.6151, "step": 1258 }, { - "epoch": 0.24, - "learning_rate": 1.7728992467389342e-05, - "loss": 1.0002, + "epoch": 0.4844170834936514, + "grad_norm": 0.8888048205718987, + "learning_rate": 1.098908472070417e-05, + "loss": 0.5435, "step": 1259 }, { - "epoch": 0.24, - "learning_rate": 1.7725036531273087e-05, - "loss": 0.936, + "epoch": 0.48480184686417854, + "grad_norm": 1.0758253246424843, + "learning_rate": 1.0976683368993464e-05, + "loss": 0.5822, "step": 1260 }, { - "epoch": 0.24, - "learning_rate": 1.7721077594841663e-05, - "loss": 1.0018, + "epoch": 0.48518661023470566, + "grad_norm": 1.0188199897015362, + "learning_rate": 1.0964280500554459e-05, + "loss": 0.6075, "step": 1261 }, { - "epoch": 0.24, - "learning_rate": 1.771711565963267e-05, - "loss": 0.9416, + "epoch": 0.4855713736052328, + "grad_norm": 0.9371647865915378, + "learning_rate": 1.0951876134648032e-05, + "loss": 0.5612, "step": 1262 }, { - "epoch": 0.24, - "learning_rate": 1.7713150727184878e-05, - "loss": 0.9271, + "epoch": 0.4859561369757599, + "grad_norm": 0.9261497329394522, + "learning_rate": 1.0939470290537389e-05, + "loss": 0.5782, "step": 1263 }, { - "epoch": 0.24, - "learning_rate": 1.770918279903822e-05, - "loss": 0.9605, + "epoch": 0.486340900346287, + "grad_norm": 0.9059254335640845, + "learning_rate": 1.0927062987488035e-05, + "loss": 0.5521, "step": 1264 }, { - "epoch": 0.24, - "learning_rate": 1.77052118767338e-05, - "loss": 0.946, + "epoch": 0.48672566371681414, + "grad_norm": 1.0371279254709531, + "learning_rate": 1.0914654244767736e-05, + "loss": 0.5966, "step": 1265 }, { - "epoch": 0.24, - "learning_rate": 1.7701237961813874e-05, - "loss": 0.9316, + "epoch": 0.48711042708734126, + "grad_norm": 0.9895263865869038, + "learning_rate": 1.0902244081646489e-05, + "loss": 0.5722, "step": 1266 }, { - "epoch": 0.24, - "learning_rate": 1.7697261055821864e-05, - "loss": 1.0391, + "epoch": 0.48749519045786843, + "grad_norm": 1.055976757454227, + "learning_rate": 1.0889832517396511e-05, + "loss": 0.6275, "step": 1267 }, { - "epoch": 0.24, - "learning_rate": 1.7693281160302354e-05, - "loss": 0.9434, + "epoch": 0.48787995382839555, + "grad_norm": 0.9555377926785722, + "learning_rate": 1.0877419571292183e-05, + "loss": 0.5753, "step": 1268 }, { - "epoch": 0.24, - "learning_rate": 1.7689298276801095e-05, - "loss": 0.8899, + "epoch": 0.48826471719892267, + "grad_norm": 0.9792097910261011, + "learning_rate": 1.0865005262610033e-05, + "loss": 0.6094, "step": 1269 }, { - "epoch": 0.24, - "learning_rate": 1.7685312406864986e-05, - "loss": 0.9742, + "epoch": 0.4886494805694498, + "grad_norm": 1.0158534843655302, + "learning_rate": 1.085258961062871e-05, + "loss": 0.6123, "step": 1270 }, { - "epoch": 0.24, - "learning_rate": 1.7681323552042094e-05, - "loss": 0.7416, + "epoch": 0.4890342439399769, + "grad_norm": 1.0151728869519805, + "learning_rate": 1.0840172634628948e-05, + "loss": 0.566, "step": 1271 }, { - "epoch": 0.24, - "learning_rate": 1.767733171388165e-05, - "loss": 0.9576, + "epoch": 0.48941900731050403, + "grad_norm": 0.9576086625001471, + "learning_rate": 1.082775435389353e-05, + "loss": 0.5884, "step": 1272 }, { - "epoch": 0.24, - "learning_rate": 1.7673336893934033e-05, - "loss": 0.9332, + "epoch": 0.48980377068103115, + "grad_norm": 1.1607561895939176, + "learning_rate": 1.0815334787707277e-05, + "loss": 0.6375, "step": 1273 }, { - "epoch": 0.25, - "learning_rate": 1.7669339093750786e-05, - "loss": 0.9205, + "epoch": 0.49018853405155827, + "grad_norm": 1.0417861868429699, + "learning_rate": 1.0802913955356998e-05, + "loss": 0.5913, "step": 1274 }, { - "epoch": 0.25, - "learning_rate": 1.766533831488461e-05, - "loss": 0.9522, + "epoch": 0.49057329742208544, + "grad_norm": 1.0125313586229048, + "learning_rate": 1.079049187613147e-05, + "loss": 0.594, "step": 1275 }, { - "epoch": 0.25, - "learning_rate": 1.7661334558889357e-05, - "loss": 0.9341, + "epoch": 0.49095806079261256, + "grad_norm": 1.0609790917024593, + "learning_rate": 1.0778068569321403e-05, + "loss": 0.6108, "step": 1276 }, { - "epoch": 0.25, - "learning_rate": 1.7657327827320046e-05, - "loss": 0.8826, + "epoch": 0.4913428241631397, + "grad_norm": 0.9957511379026509, + "learning_rate": 1.0765644054219422e-05, + "loss": 0.6122, "step": 1277 }, { - "epoch": 0.25, - "learning_rate": 1.765331812173284e-05, - "loss": 1.0185, + "epoch": 0.4917275875336668, + "grad_norm": 0.9406707456307537, + "learning_rate": 1.0753218350120023e-05, + "loss": 0.5328, "step": 1278 }, { - "epoch": 0.25, - "learning_rate": 1.7649305443685068e-05, - "loss": 0.9265, + "epoch": 0.4921123509041939, + "grad_norm": 0.6228436924564613, + "learning_rate": 1.0740791476319543e-05, + "loss": 0.6483, "step": 1279 }, { - "epoch": 0.25, - "learning_rate": 1.76452897947352e-05, - "loss": 0.8785, + "epoch": 0.49249711427472104, + "grad_norm": 0.9433204409907223, + "learning_rate": 1.0728363452116149e-05, + "loss": 0.5847, "step": 1280 }, { - "epoch": 0.25, - "learning_rate": 1.7641271176442876e-05, - "loss": 0.9314, + "epoch": 0.49288187764524816, + "grad_norm": 0.6302372803102911, + "learning_rate": 1.0715934296809782e-05, + "loss": 0.6635, "step": 1281 }, { - "epoch": 0.25, - "learning_rate": 1.7637249590368878e-05, - "loss": 1.0106, + "epoch": 0.4932666410157753, + "grad_norm": 0.9515572416829209, + "learning_rate": 1.0703504029702148e-05, + "loss": 0.5591, "step": 1282 }, { - "epoch": 0.25, - "learning_rate": 1.763322503807514e-05, - "loss": 0.9207, + "epoch": 0.49365140438630245, + "grad_norm": 0.916794903862888, + "learning_rate": 1.0691072670096669e-05, + "loss": 0.5431, "step": 1283 }, { - "epoch": 0.25, - "learning_rate": 1.7629197521124758e-05, - "loss": 0.9221, + "epoch": 0.49403616775682957, + "grad_norm": 1.0584216939380366, + "learning_rate": 1.0678640237298476e-05, + "loss": 0.5705, "step": 1284 }, { - "epoch": 0.25, - "learning_rate": 1.7625167041081967e-05, - "loss": 0.9156, + "epoch": 0.4944209311273567, + "grad_norm": 0.8530122678811783, + "learning_rate": 1.0666206750614363e-05, + "loss": 0.469, "step": 1285 }, { - "epoch": 0.25, - "learning_rate": 1.7621133599512163e-05, - "loss": 0.9514, + "epoch": 0.4948056944978838, + "grad_norm": 1.099602210543576, + "learning_rate": 1.065377222935275e-05, + "loss": 0.6382, "step": 1286 }, { - "epoch": 0.25, - "learning_rate": 1.761709719798189e-05, - "loss": 0.9544, + "epoch": 0.4951904578684109, + "grad_norm": 1.0249956754676326, + "learning_rate": 1.064133669282368e-05, + "loss": 0.5838, "step": 1287 }, { - "epoch": 0.25, - "learning_rate": 1.761305783805883e-05, - "loss": 0.9862, + "epoch": 0.49557522123893805, + "grad_norm": 0.9430916664069405, + "learning_rate": 1.0628900160338764e-05, + "loss": 0.5887, "step": 1288 }, { - "epoch": 0.25, - "learning_rate": 1.7609015521311836e-05, - "loss": 0.7437, + "epoch": 0.49595998460946517, + "grad_norm": 0.8944744754040685, + "learning_rate": 1.0616462651211156e-05, + "loss": 0.5903, "step": 1289 }, { - "epoch": 0.25, - "learning_rate": 1.7604970249310893e-05, - "loss": 0.9617, + "epoch": 0.4963447479799923, + "grad_norm": 0.9678782972583179, + "learning_rate": 1.0604024184755539e-05, + "loss": 0.6093, "step": 1290 }, { - "epoch": 0.25, - "learning_rate": 1.7600922023627137e-05, - "loss": 1.0054, + "epoch": 0.4967295113505194, + "grad_norm": 0.8991766905047878, + "learning_rate": 1.0591584780288069e-05, + "loss": 0.5556, "step": 1291 }, { - "epoch": 0.25, - "learning_rate": 1.759687084583285e-05, - "loss": 0.9758, + "epoch": 0.4971142747210466, + "grad_norm": 0.9398561833546051, + "learning_rate": 1.0579144457126365e-05, + "loss": 0.5902, "step": 1292 }, { - "epoch": 0.25, - "learning_rate": 1.759281671750147e-05, - "loss": 1.0116, + "epoch": 0.4974990380915737, + "grad_norm": 0.9125073283965731, + "learning_rate": 1.0566703234589471e-05, + "loss": 0.5339, "step": 1293 }, { - "epoch": 0.25, - "learning_rate": 1.7588759640207564e-05, - "loss": 0.7631, + "epoch": 0.4978838014621008, + "grad_norm": 0.6276125592798362, + "learning_rate": 1.0554261131997833e-05, + "loss": 0.6388, "step": 1294 }, { - "epoch": 0.25, - "learning_rate": 1.7584699615526857e-05, - "loss": 0.8689, + "epoch": 0.49826856483262794, + "grad_norm": 0.8998296600207127, + "learning_rate": 1.054181816867326e-05, + "loss": 0.5492, "step": 1295 }, { - "epoch": 0.25, - "learning_rate": 1.7580636645036224e-05, - "loss": 0.9947, + "epoch": 0.49865332820315506, + "grad_norm": 1.0075195405767627, + "learning_rate": 1.0529374363938888e-05, + "loss": 0.633, "step": 1296 }, { - "epoch": 0.25, - "learning_rate": 1.757657073031367e-05, - "loss": 1.0337, + "epoch": 0.4990380915736822, + "grad_norm": 0.9363570910744718, + "learning_rate": 1.051692973711918e-05, + "loss": 0.5549, "step": 1297 }, { - "epoch": 0.25, - "learning_rate": 1.7572501872938343e-05, - "loss": 0.7091, + "epoch": 0.4994228549442093, + "grad_norm": 0.9204842777802245, + "learning_rate": 1.0504484307539864e-05, + "loss": 0.5671, "step": 1298 }, { - "epoch": 0.25, - "learning_rate": 1.756843007449055e-05, - "loss": 0.9502, + "epoch": 0.4998076183147364, + "grad_norm": 0.8991967130863284, + "learning_rate": 1.0492038094527907e-05, + "loss": 0.5969, "step": 1299 }, { - "epoch": 0.25, - "learning_rate": 1.7564355336551727e-05, - "loss": 0.9147, + "epoch": 0.5001923816852636, + "grad_norm": 0.9103362946763063, + "learning_rate": 1.047959111741151e-05, + "loss": 0.5555, "step": 1300 }, { - "epoch": 0.25, - "learning_rate": 1.7560277660704455e-05, - "loss": 0.7969, + "epoch": 0.5005771450557907, + "grad_norm": 0.9421671059959295, + "learning_rate": 1.0467143395520044e-05, + "loss": 0.5805, "step": 1301 }, { - "epoch": 0.25, - "learning_rate": 1.755619704853246e-05, - "loss": 0.9154, + "epoch": 0.5009619084263178, + "grad_norm": 0.8463879466607206, + "learning_rate": 1.0454694948184045e-05, + "loss": 0.5028, "step": 1302 }, { - "epoch": 0.25, - "learning_rate": 1.7552113501620595e-05, - "loss": 0.9396, + "epoch": 0.5013466717968449, + "grad_norm": 0.9283787773885035, + "learning_rate": 1.044224579473518e-05, + "loss": 0.5437, "step": 1303 }, { - "epoch": 0.25, - "learning_rate": 1.7548027021554874e-05, - "loss": 0.9647, + "epoch": 0.5017314351673721, + "grad_norm": 0.9271036562856353, + "learning_rate": 1.0429795954506203e-05, + "loss": 0.5871, "step": 1304 }, { - "epoch": 0.25, - "learning_rate": 1.754393760992243e-05, - "loss": 0.9471, + "epoch": 0.5021161985378992, + "grad_norm": 0.9451848989997307, + "learning_rate": 1.0417345446830938e-05, + "loss": 0.611, "step": 1305 }, { - "epoch": 0.25, - "learning_rate": 1.7539845268311548e-05, - "loss": 0.9444, + "epoch": 0.5025009619084263, + "grad_norm": 0.9386062184909925, + "learning_rate": 1.0404894291044247e-05, + "loss": 0.6195, "step": 1306 }, { - "epoch": 0.25, - "learning_rate": 1.7535749998311645e-05, - "loss": 0.9458, + "epoch": 0.5028857252789535, + "grad_norm": 0.6261287625631391, + "learning_rate": 1.0392442506482e-05, + "loss": 0.6499, "step": 1307 }, { - "epoch": 0.25, - "learning_rate": 1.753165180151328e-05, - "loss": 0.7476, + "epoch": 0.5032704886494805, + "grad_norm": 0.9372168580416832, + "learning_rate": 1.037999011248104e-05, + "loss": 0.5878, "step": 1308 }, { - "epoch": 0.25, - "learning_rate": 1.752755067950814e-05, - "loss": 0.9939, + "epoch": 0.5036552520200077, + "grad_norm": 0.9264887191163762, + "learning_rate": 1.0367537128379154e-05, + "loss": 0.6303, "step": 1309 }, { - "epoch": 0.25, - "learning_rate": 1.752344663388906e-05, - "loss": 0.9261, + "epoch": 0.5040400153905348, + "grad_norm": 0.6313295602189505, + "learning_rate": 1.0355083573515052e-05, + "loss": 0.6446, "step": 1310 }, { - "epoch": 0.25, - "learning_rate": 1.7519339666249997e-05, - "loss": 0.9251, + "epoch": 0.504424778761062, + "grad_norm": 1.0856355012388164, + "learning_rate": 1.0342629467228331e-05, + "loss": 0.5887, "step": 1311 }, { - "epoch": 0.25, - "learning_rate": 1.7515229778186052e-05, - "loss": 0.9303, + "epoch": 0.504809542131589, + "grad_norm": 1.1659556477709463, + "learning_rate": 1.0330174828859434e-05, + "loss": 0.6119, "step": 1312 }, { - "epoch": 0.25, - "learning_rate": 1.7511116971293463e-05, - "loss": 0.8878, + "epoch": 0.5051943055021162, + "grad_norm": 0.920529013969653, + "learning_rate": 1.031771967774964e-05, + "loss": 0.5227, "step": 1313 }, { - "epoch": 0.25, - "learning_rate": 1.7507001247169587e-05, - "loss": 0.9157, + "epoch": 0.5055790688726434, + "grad_norm": 0.9388744418033227, + "learning_rate": 1.030526403324102e-05, + "loss": 0.586, "step": 1314 }, { - "epoch": 0.25, - "learning_rate": 1.7502882607412933e-05, - "loss": 0.9135, + "epoch": 0.5059638322431704, + "grad_norm": 1.0022727665194477, + "learning_rate": 1.0292807914676412e-05, + "loss": 0.6288, "step": 1315 }, { - "epoch": 0.25, - "learning_rate": 1.749876105362313e-05, - "loss": 0.9611, + "epoch": 0.5063485956136976, + "grad_norm": 0.6149141552353516, + "learning_rate": 1.0280351341399392e-05, + "loss": 0.6373, "step": 1316 }, { - "epoch": 0.25, - "learning_rate": 1.7494636587400942e-05, - "loss": 1.0268, + "epoch": 0.5067333589842247, + "grad_norm": 1.0346665722318324, + "learning_rate": 1.0267894332754243e-05, + "loss": 0.5812, "step": 1317 }, { - "epoch": 0.25, - "learning_rate": 1.749050921034826e-05, - "loss": 0.936, + "epoch": 0.5071181223547518, + "grad_norm": 1.1115431126040123, + "learning_rate": 1.0255436908085919e-05, + "loss": 0.6304, "step": 1318 }, { - "epoch": 0.25, - "learning_rate": 1.7486378924068123e-05, - "loss": 0.9965, + "epoch": 0.5075028857252789, + "grad_norm": 1.0125221972217346, + "learning_rate": 1.0242979086740019e-05, + "loss": 0.5755, "step": 1319 }, { - "epoch": 0.25, - "learning_rate": 1.748224573016467e-05, - "loss": 0.9699, + "epoch": 0.5078876490958061, + "grad_norm": 1.0377420179782226, + "learning_rate": 1.0230520888062765e-05, + "loss": 0.5953, "step": 1320 }, { - "epoch": 0.25, - "learning_rate": 1.7478109630243195e-05, - "loss": 0.9316, + "epoch": 0.5082724124663333, + "grad_norm": 0.9288023601402551, + "learning_rate": 1.0218062331400969e-05, + "loss": 0.5284, "step": 1321 }, { - "epoch": 0.25, - "learning_rate": 1.747397062591011e-05, - "loss": 0.9318, + "epoch": 0.5086571758368603, + "grad_norm": 0.968454116135532, + "learning_rate": 1.0205603436101978e-05, + "loss": 0.5564, "step": 1322 }, { - "epoch": 0.25, - "learning_rate": 1.746982871877296e-05, - "loss": 0.9583, + "epoch": 0.5090419392073875, + "grad_norm": 1.0215201277443666, + "learning_rate": 1.019314422151369e-05, + "loss": 0.5777, "step": 1323 }, { - "epoch": 0.25, - "learning_rate": 1.7465683910440405e-05, - "loss": 0.7601, + "epoch": 0.5094267025779146, + "grad_norm": 1.0847491762344, + "learning_rate": 1.0180684706984483e-05, + "loss": 0.5727, "step": 1324 }, { - "epoch": 0.25, - "learning_rate": 1.7461536202522248e-05, - "loss": 0.9228, + "epoch": 0.5098114659484417, + "grad_norm": 1.0415954256539601, + "learning_rate": 1.0168224911863205e-05, + "loss": 0.6191, "step": 1325 }, { - "epoch": 0.26, - "learning_rate": 1.745738559662941e-05, - "loss": 0.9276, + "epoch": 0.5101962293189688, + "grad_norm": 0.6890962700579393, + "learning_rate": 1.015576485549914e-05, + "loss": 0.6675, "step": 1326 }, { - "epoch": 0.26, - "learning_rate": 1.7453232094373936e-05, - "loss": 0.9472, + "epoch": 0.510580992689496, + "grad_norm": 0.9770745732330645, + "learning_rate": 1.0143304557241979e-05, + "loss": 0.5736, "step": 1327 }, { - "epoch": 0.26, - "learning_rate": 1.7449075697369005e-05, - "loss": 0.9894, + "epoch": 0.510965756060023, + "grad_norm": 1.2497499608065452, + "learning_rate": 1.0130844036441787e-05, + "loss": 0.5803, "step": 1328 }, { - "epoch": 0.26, - "learning_rate": 1.7444916407228904e-05, - "loss": 0.8964, + "epoch": 0.5113505194305502, + "grad_norm": 1.1083394798140156, + "learning_rate": 1.0118383312448973e-05, + "loss": 0.5704, "step": 1329 }, { - "epoch": 0.26, - "learning_rate": 1.744075422556906e-05, - "loss": 0.8982, + "epoch": 0.5117352828010774, + "grad_norm": 1.0444444337635355, + "learning_rate": 1.0105922404614265e-05, + "loss": 0.6312, "step": 1330 }, { - "epoch": 0.26, - "learning_rate": 1.7436589154006014e-05, - "loss": 1.0721, + "epoch": 0.5121200461716044, + "grad_norm": 0.9229444750162329, + "learning_rate": 1.0093461332288678e-05, + "loss": 0.5581, "step": 1331 }, { - "epoch": 0.26, - "learning_rate": 1.743242119415743e-05, - "loss": 0.9268, + "epoch": 0.5125048095421316, + "grad_norm": 0.8784269740051026, + "learning_rate": 1.0081000114823473e-05, + "loss": 0.5461, "step": 1332 }, { - "epoch": 0.26, - "learning_rate": 1.7428250347642102e-05, - "loss": 0.9143, + "epoch": 0.5128895729126587, + "grad_norm": 0.9970676995778642, + "learning_rate": 1.006853877157015e-05, + "loss": 0.6137, "step": 1333 }, { - "epoch": 0.26, - "learning_rate": 1.7424076616079933e-05, - "loss": 0.9762, + "epoch": 0.5132743362831859, + "grad_norm": 1.1512421630408873, + "learning_rate": 1.0056077321880393e-05, + "loss": 0.6272, "step": 1334 }, { - "epoch": 0.26, - "learning_rate": 1.7419900001091953e-05, - "loss": 0.9634, + "epoch": 0.5136590996537129, + "grad_norm": 1.077662218825046, + "learning_rate": 1.0043615785106051e-05, + "loss": 0.5228, "step": 1335 }, { - "epoch": 0.26, - "learning_rate": 1.7415720504300314e-05, - "loss": 0.9528, + "epoch": 0.5140438630242401, + "grad_norm": 1.0108839618528174, + "learning_rate": 1.0031154180599123e-05, + "loss": 0.6416, "step": 1336 }, { - "epoch": 0.26, - "learning_rate": 1.741153812732828e-05, - "loss": 0.9512, + "epoch": 0.5144286263947672, + "grad_norm": 1.022178330664966, + "learning_rate": 1.0018692527711695e-05, + "loss": 0.6089, "step": 1337 }, { - "epoch": 0.26, - "learning_rate": 1.7407352871800246e-05, - "loss": 0.9427, + "epoch": 0.5148133897652943, + "grad_norm": 0.9025394101042549, + "learning_rate": 1.0006230845795937e-05, + "loss": 0.5926, "step": 1338 }, { - "epoch": 0.26, - "learning_rate": 1.7403164739341708e-05, - "loss": 0.8975, + "epoch": 0.5151981531358215, + "grad_norm": 1.0047845727468478, + "learning_rate": 9.993769154204063e-06, + "loss": 0.534, "step": 1339 }, { - "epoch": 0.26, - "learning_rate": 1.739897373157929e-05, - "loss": 0.9319, + "epoch": 0.5155829165063486, + "grad_norm": 1.0252587207311024, + "learning_rate": 9.981307472288308e-06, + "loss": 0.6216, "step": 1340 }, { - "epoch": 0.26, - "learning_rate": 1.7394779850140736e-05, - "loss": 0.9708, + "epoch": 0.5159676798768758, + "grad_norm": 1.0721969140270957, + "learning_rate": 9.968845819400883e-06, + "loss": 0.5777, "step": 1341 }, { - "epoch": 0.26, - "learning_rate": 1.7390583096654895e-05, - "loss": 0.9093, + "epoch": 0.5163524432474028, + "grad_norm": 0.9611691882514103, + "learning_rate": 9.956384214893949e-06, + "loss": 0.607, "step": 1342 }, { - "epoch": 0.26, - "learning_rate": 1.7386383472751745e-05, - "loss": 0.9156, + "epoch": 0.51673720661793, + "grad_norm": 0.978062666735957, + "learning_rate": 9.94392267811961e-06, + "loss": 0.6456, "step": 1343 }, { - "epoch": 0.26, - "learning_rate": 1.7382180980062365e-05, - "loss": 0.9051, + "epoch": 0.517121969988457, + "grad_norm": 0.9635561879064112, + "learning_rate": 9.931461228429856e-06, + "loss": 0.5763, "step": 1344 }, { - "epoch": 0.26, - "learning_rate": 1.7377975620218954e-05, - "loss": 0.9218, + "epoch": 0.5175067333589842, + "grad_norm": 0.9252436316019189, + "learning_rate": 9.91899988517653e-06, + "loss": 0.5736, "step": 1345 }, { - "epoch": 0.26, - "learning_rate": 1.7373767394854836e-05, - "loss": 0.9875, + "epoch": 0.5178914967295114, + "grad_norm": 0.963022109450362, + "learning_rate": 9.906538667711324e-06, + "loss": 0.5657, "step": 1346 }, { - "epoch": 0.26, - "learning_rate": 1.7369556305604422e-05, - "loss": 0.95, + "epoch": 0.5182762601000385, + "grad_norm": 1.050438957563135, + "learning_rate": 9.894077595385736e-06, + "loss": 0.611, "step": 1347 }, { - "epoch": 0.26, - "learning_rate": 1.736534235410326e-05, - "loss": 0.9281, + "epoch": 0.5186610234705656, + "grad_norm": 1.0176657184669984, + "learning_rate": 9.881616687551032e-06, + "loss": 0.5764, "step": 1348 }, { - "epoch": 0.26, - "learning_rate": 1.7361125541988e-05, - "loss": 0.8832, + "epoch": 0.5190457868410927, + "grad_norm": 0.8860624928650852, + "learning_rate": 9.869155963558215e-06, + "loss": 0.5287, "step": 1349 }, { - "epoch": 0.26, - "learning_rate": 1.7356905870896407e-05, - "loss": 0.9785, + "epoch": 0.5194305502116199, + "grad_norm": 0.9415864420237466, + "learning_rate": 9.856695442758023e-06, + "loss": 0.57, "step": 1350 }, { - "epoch": 0.26, - "learning_rate": 1.735268334246734e-05, - "loss": 0.9412, + "epoch": 0.5198153135821469, + "grad_norm": 0.6644084580887467, + "learning_rate": 9.844235144500865e-06, + "loss": 0.6546, "step": 1351 }, { - "epoch": 0.26, - "learning_rate": 1.7348457958340792e-05, - "loss": 0.6797, + "epoch": 0.5202000769526741, + "grad_norm": 1.0410661709171019, + "learning_rate": 9.831775088136797e-06, + "loss": 0.6158, "step": 1352 }, { - "epoch": 0.26, - "learning_rate": 1.7344229720157846e-05, - "loss": 0.9264, + "epoch": 0.5205848403232012, + "grad_norm": 1.0411566878817453, + "learning_rate": 9.819315293015519e-06, + "loss": 0.585, "step": 1353 }, { - "epoch": 0.26, - "learning_rate": 1.7339998629560705e-05, - "loss": 0.961, + "epoch": 0.5209696036937284, + "grad_norm": 1.068976264135062, + "learning_rate": 9.806855778486314e-06, + "loss": 0.632, "step": 1354 }, { - "epoch": 0.26, - "learning_rate": 1.7335764688192676e-05, - "loss": 0.9748, + "epoch": 0.5213543670642555, + "grad_norm": 1.0310218421836013, + "learning_rate": 9.794396563898022e-06, + "loss": 0.5927, "step": 1355 }, { - "epoch": 0.26, - "learning_rate": 1.733152789769817e-05, - "loss": 1.0181, + "epoch": 0.5217391304347826, + "grad_norm": 0.9939622946294197, + "learning_rate": 9.781937668599035e-06, + "loss": 0.5549, "step": 1356 }, { - "epoch": 0.26, - "learning_rate": 1.7327288259722714e-05, - "loss": 1.0174, + "epoch": 0.5221238938053098, + "grad_norm": 0.9769635001487994, + "learning_rate": 9.769479111937238e-06, + "loss": 0.6004, "step": 1357 }, { - "epoch": 0.26, - "learning_rate": 1.7323045775912927e-05, - "loss": 0.959, + "epoch": 0.5225086571758368, + "grad_norm": 0.9652439203966753, + "learning_rate": 9.757020913259986e-06, + "loss": 0.6057, "step": 1358 }, { - "epoch": 0.26, - "learning_rate": 1.7318800447916543e-05, - "loss": 0.962, + "epoch": 0.522893420546364, + "grad_norm": 0.9659474400634394, + "learning_rate": 9.744563091914085e-06, + "loss": 0.5747, "step": 1359 }, { - "epoch": 0.26, - "learning_rate": 1.7314552277382403e-05, - "loss": 0.9582, + "epoch": 0.5232781839168911, + "grad_norm": 0.9412108421087431, + "learning_rate": 9.732105667245759e-06, + "loss": 0.5309, "step": 1360 }, { - "epoch": 0.26, - "learning_rate": 1.7310301265960446e-05, - "loss": 0.9042, + "epoch": 0.5236629472874182, + "grad_norm": 1.0472647174497431, + "learning_rate": 9.719648658600611e-06, + "loss": 0.6412, "step": 1361 }, { - "epoch": 0.26, - "learning_rate": 1.7306047415301706e-05, - "loss": 0.939, + "epoch": 0.5240477106579453, + "grad_norm": 0.9514833665901985, + "learning_rate": 9.70719208532359e-06, + "loss": 0.6499, "step": 1362 }, { - "epoch": 0.26, - "learning_rate": 1.7301790727058344e-05, - "loss": 0.9205, + "epoch": 0.5244324740284725, + "grad_norm": 1.05949421659902, + "learning_rate": 9.694735966758982e-06, + "loss": 0.5824, "step": 1363 }, { - "epoch": 0.26, - "learning_rate": 1.7297531202883598e-05, - "loss": 0.9489, + "epoch": 0.5248172373989997, + "grad_norm": 0.8862808627429452, + "learning_rate": 9.682280322250365e-06, + "loss": 0.5109, "step": 1364 }, { - "epoch": 0.26, - "learning_rate": 1.7293268844431826e-05, - "loss": 0.9361, + "epoch": 0.5252020007695267, + "grad_norm": 0.9681503460101686, + "learning_rate": 9.669825171140568e-06, + "loss": 0.5443, "step": 1365 }, { - "epoch": 0.26, - "learning_rate": 1.7289003653358472e-05, - "loss": 1.0306, + "epoch": 0.5255867641400539, + "grad_norm": 0.9405237020668291, + "learning_rate": 9.657370532771672e-06, + "loss": 0.5623, "step": 1366 }, { - "epoch": 0.26, - "learning_rate": 1.7284735631320093e-05, - "loss": 0.8952, + "epoch": 0.525971527510581, + "grad_norm": 0.9513589736249594, + "learning_rate": 9.64491642648495e-06, + "loss": 0.5897, "step": 1367 }, { - "epoch": 0.26, - "learning_rate": 1.7280464779974335e-05, - "loss": 0.9588, + "epoch": 0.5263562908811081, + "grad_norm": 1.1104527979237486, + "learning_rate": 9.632462871620847e-06, + "loss": 0.6113, "step": 1368 }, { - "epoch": 0.26, - "learning_rate": 1.7276191100979952e-05, - "loss": 0.9383, + "epoch": 0.5267410542516352, + "grad_norm": 0.958233506899973, + "learning_rate": 9.620009887518963e-06, + "loss": 0.6224, "step": 1369 }, { - "epoch": 0.26, - "learning_rate": 1.7271914595996784e-05, - "loss": 0.9611, + "epoch": 0.5271258176221624, + "grad_norm": 0.9667830636499019, + "learning_rate": 9.607557493518006e-06, + "loss": 0.5233, "step": 1370 }, { - "epoch": 0.26, - "learning_rate": 1.7267635266685782e-05, - "loss": 0.955, + "epoch": 0.5275105809926895, + "grad_norm": 0.8752008855079952, + "learning_rate": 9.595105708955758e-06, + "loss": 0.5257, "step": 1371 }, { - "epoch": 0.26, - "learning_rate": 1.7263353114708993e-05, - "loss": 0.9563, + "epoch": 0.5278953443632166, + "grad_norm": 0.9219793291267582, + "learning_rate": 9.582654553169064e-06, + "loss": 0.5475, "step": 1372 }, { - "epoch": 0.26, - "learning_rate": 1.7259068141729542e-05, - "loss": 0.9053, + "epoch": 0.5282801077337438, + "grad_norm": 1.0072792031573092, + "learning_rate": 9.5702040454938e-06, + "loss": 0.5872, "step": 1373 }, { - "epoch": 0.26, - "learning_rate": 1.7254780349411677e-05, - "loss": 0.8449, + "epoch": 0.5286648711042709, + "grad_norm": 0.9747559764791464, + "learning_rate": 9.557754205264826e-06, + "loss": 0.5487, "step": 1374 }, { - "epoch": 0.26, - "learning_rate": 1.7250489739420718e-05, - "loss": 1.0239, + "epoch": 0.529049634474798, + "grad_norm": 1.0089670907662371, + "learning_rate": 9.545305051815957e-06, + "loss": 0.5585, "step": 1375 }, { - "epoch": 0.26, - "learning_rate": 1.7246196313423095e-05, - "loss": 0.7355, + "epoch": 0.5294343978453251, + "grad_norm": 1.1141146783897313, + "learning_rate": 9.53285660447996e-06, + "loss": 0.5605, "step": 1376 }, { - "epoch": 0.26, - "learning_rate": 1.7241900073086318e-05, - "loss": 0.9154, + "epoch": 0.5298191612158523, + "grad_norm": 0.9964858212173543, + "learning_rate": 9.520408882588497e-06, + "loss": 0.6305, "step": 1377 }, { - "epoch": 0.27, - "learning_rate": 1.7237601020079003e-05, - "loss": 0.9306, + "epoch": 0.5302039245863793, + "grad_norm": 0.9194875996443909, + "learning_rate": 9.507961905472093e-06, + "loss": 0.5351, "step": 1378 }, { - "epoch": 0.27, - "learning_rate": 1.7233299156070852e-05, - "loss": 0.9655, + "epoch": 0.5305886879569065, + "grad_norm": 0.9375287505215805, + "learning_rate": 9.495515692460138e-06, + "loss": 0.5777, "step": 1379 }, { - "epoch": 0.27, - "learning_rate": 1.7228994482732653e-05, - "loss": 0.9878, + "epoch": 0.5309734513274337, + "grad_norm": 0.9771176335499306, + "learning_rate": 9.483070262880823e-06, + "loss": 0.6262, "step": 1380 }, { - "epoch": 0.27, - "learning_rate": 1.72246870017363e-05, - "loss": 0.975, + "epoch": 0.5313582146979607, + "grad_norm": 1.093244194473128, + "learning_rate": 9.47062563606111e-06, + "loss": 0.6267, "step": 1381 }, { - "epoch": 0.27, - "learning_rate": 1.7220376714754766e-05, - "loss": 0.9212, + "epoch": 0.5317429780684879, + "grad_norm": 0.8745906706019505, + "learning_rate": 9.458181831326744e-06, + "loss": 0.5653, "step": 1382 }, { - "epoch": 0.27, - "learning_rate": 1.7216063623462112e-05, - "loss": 0.8741, + "epoch": 0.532127741439015, + "grad_norm": 1.0211807328631908, + "learning_rate": 9.44573886800217e-06, + "loss": 0.5551, "step": 1383 }, { - "epoch": 0.27, - "learning_rate": 1.7211747729533504e-05, - "loss": 1.015, + "epoch": 0.5325125048095422, + "grad_norm": 0.9020890236313851, + "learning_rate": 9.433296765410534e-06, + "loss": 0.5552, "step": 1384 }, { - "epoch": 0.27, - "learning_rate": 1.7207429034645176e-05, - "loss": 0.9218, + "epoch": 0.5328972681800692, + "grad_norm": 0.9663004844317222, + "learning_rate": 9.420855542873638e-06, + "loss": 0.6077, "step": 1385 }, { - "epoch": 0.27, - "learning_rate": 1.720310754047446e-05, - "loss": 0.9596, + "epoch": 0.5332820315505964, + "grad_norm": 0.9316212866127979, + "learning_rate": 9.408415219711934e-06, + "loss": 0.5911, "step": 1386 }, { - "epoch": 0.27, - "learning_rate": 1.719878324869978e-05, - "loss": 0.9186, + "epoch": 0.5336667949211235, + "grad_norm": 1.0304728593814365, + "learning_rate": 9.395975815244468e-06, + "loss": 0.5807, "step": 1387 }, { - "epoch": 0.27, - "learning_rate": 1.7194456161000634e-05, - "loss": 1.0032, + "epoch": 0.5340515582916506, + "grad_norm": 1.0075576440999023, + "learning_rate": 9.383537348788844e-06, + "loss": 0.6021, "step": 1388 }, { - "epoch": 0.27, - "learning_rate": 1.719012627905762e-05, - "loss": 0.8911, + "epoch": 0.5344363216621778, + "grad_norm": 1.0081056988131125, + "learning_rate": 9.371099839661238e-06, + "loss": 0.6013, "step": 1389 }, { - "epoch": 0.27, - "learning_rate": 1.718579360455241e-05, - "loss": 0.7879, + "epoch": 0.5348210850327049, + "grad_norm": 0.9844741390044599, + "learning_rate": 9.358663307176323e-06, + "loss": 0.5466, "step": 1390 }, { - "epoch": 0.27, - "learning_rate": 1.7181458139167767e-05, - "loss": 0.9574, + "epoch": 0.535205848403232, + "grad_norm": 1.0013751115003102, + "learning_rate": 9.346227770647251e-06, + "loss": 0.604, "step": 1391 }, { - "epoch": 0.27, - "learning_rate": 1.7177119884587536e-05, - "loss": 0.9197, + "epoch": 0.5355906117737591, + "grad_norm": 0.9568894045706614, + "learning_rate": 9.33379324938564e-06, + "loss": 0.6641, "step": 1392 }, { - "epoch": 0.27, - "learning_rate": 1.717277884249664e-05, - "loss": 0.9804, + "epoch": 0.5359753751442863, + "grad_norm": 0.6357015479801963, + "learning_rate": 9.321359762701527e-06, + "loss": 0.6414, "step": 1393 }, { - "epoch": 0.27, - "learning_rate": 1.716843501458109e-05, - "loss": 0.7464, + "epoch": 0.5363601385148133, + "grad_norm": 1.0378291768524792, + "learning_rate": 9.308927329903333e-06, + "loss": 0.617, "step": 1394 }, { - "epoch": 0.27, - "learning_rate": 1.716408840252799e-05, - "loss": 0.8998, + "epoch": 0.5367449018853405, + "grad_norm": 0.9717259422360616, + "learning_rate": 9.296495970297855e-06, + "loss": 0.5821, "step": 1395 }, { - "epoch": 0.27, - "learning_rate": 1.7159739008025503e-05, - "loss": 0.998, + "epoch": 0.5371296652558677, + "grad_norm": 0.9579642359081738, + "learning_rate": 9.284065703190221e-06, + "loss": 0.5348, "step": 1396 }, { - "epoch": 0.27, - "learning_rate": 1.7155386832762892e-05, - "loss": 0.7546, + "epoch": 0.5375144286263948, + "grad_norm": 0.8981879904741323, + "learning_rate": 9.271636547883856e-06, + "loss": 0.5156, "step": 1397 }, { - "epoch": 0.27, - "learning_rate": 1.715103187843048e-05, - "loss": 0.9611, + "epoch": 0.5378991919969219, + "grad_norm": 1.0705295225351208, + "learning_rate": 9.259208523680457e-06, + "loss": 0.5875, "step": 1398 }, { - "epoch": 0.27, - "learning_rate": 1.7146674146719688e-05, - "loss": 0.9479, + "epoch": 0.538283955367449, + "grad_norm": 1.146194782563521, + "learning_rate": 9.24678164987998e-06, + "loss": 0.6266, "step": 1399 }, { - "epoch": 0.27, - "learning_rate": 1.7142313639323012e-05, - "loss": 0.9044, + "epoch": 0.5386687187379762, + "grad_norm": 0.9244147291193102, + "learning_rate": 9.234355945780581e-06, + "loss": 0.5573, "step": 1400 }, { - "epoch": 0.27, - "learning_rate": 1.7137950357934017e-05, - "loss": 0.9519, + "epoch": 0.5390534821085032, + "grad_norm": 0.9716019122577356, + "learning_rate": 9.221931430678598e-06, + "loss": 0.6004, "step": 1401 }, { - "epoch": 0.27, - "learning_rate": 1.7133584304247354e-05, - "loss": 0.9152, + "epoch": 0.5394382454790304, + "grad_norm": 0.8651440888150316, + "learning_rate": 9.209508123868534e-06, + "loss": 0.5719, "step": 1402 }, { - "epoch": 0.27, - "learning_rate": 1.7129215479958747e-05, - "loss": 0.858, + "epoch": 0.5398230088495575, + "grad_norm": 0.6064002594585906, + "learning_rate": 9.197086044643004e-06, + "loss": 0.6763, "step": 1403 }, { - "epoch": 0.27, - "learning_rate": 1.7124843886765e-05, - "loss": 0.9367, + "epoch": 0.5402077722200846, + "grad_norm": 0.9623197035196227, + "learning_rate": 9.184665212292723e-06, + "loss": 0.6571, "step": 1404 }, { - "epoch": 0.27, - "learning_rate": 1.712046952636398e-05, - "loss": 0.8813, + "epoch": 0.5405925355906118, + "grad_norm": 1.0058250728533746, + "learning_rate": 9.172245646106471e-06, + "loss": 0.6154, "step": 1405 }, { - "epoch": 0.27, - "learning_rate": 1.7116092400454655e-05, - "loss": 0.9057, + "epoch": 0.5409772989611389, + "grad_norm": 1.0339083994424985, + "learning_rate": 9.159827365371055e-06, + "loss": 0.6317, "step": 1406 }, { - "epoch": 0.27, - "learning_rate": 1.7111712510737035e-05, - "loss": 0.9236, + "epoch": 0.5413620623316661, + "grad_norm": 0.9195448487465863, + "learning_rate": 9.14741038937129e-06, + "loss": 0.5618, "step": 1407 }, { - "epoch": 0.27, - "learning_rate": 1.7107329858912226e-05, - "loss": 0.9127, + "epoch": 0.5417468257021931, + "grad_norm": 0.9882587113814898, + "learning_rate": 9.13499473738997e-06, + "loss": 0.5671, "step": 1408 }, { - "epoch": 0.27, - "learning_rate": 1.7102944446682393e-05, - "loss": 0.94, + "epoch": 0.5421315890727203, + "grad_norm": 0.9373798385716448, + "learning_rate": 9.122580428707822e-06, + "loss": 0.5655, "step": 1409 }, { - "epoch": 0.27, - "learning_rate": 1.709855627575079e-05, - "loss": 0.9318, + "epoch": 0.5425163524432474, + "grad_norm": 0.9830629433610579, + "learning_rate": 9.110167482603494e-06, + "loss": 0.5931, "step": 1410 }, { - "epoch": 0.27, - "learning_rate": 1.7094165347821724e-05, - "loss": 0.9129, + "epoch": 0.5429011158137745, + "grad_norm": 0.9025523996554359, + "learning_rate": 9.097755918353513e-06, + "loss": 0.5582, "step": 1411 }, { - "epoch": 0.27, - "learning_rate": 1.7089771664600584e-05, - "loss": 0.8979, + "epoch": 0.5432858791843016, + "grad_norm": 0.9610957177847527, + "learning_rate": 9.08534575523227e-06, + "loss": 0.5702, "step": 1412 }, { - "epoch": 0.27, - "learning_rate": 1.708537522779382e-05, - "loss": 0.9222, + "epoch": 0.5436706425548288, + "grad_norm": 0.9652144248264015, + "learning_rate": 9.072937012511968e-06, + "loss": 0.5969, "step": 1413 }, { - "epoch": 0.27, - "learning_rate": 1.7080976039108964e-05, - "loss": 0.7765, + "epoch": 0.544055405925356, + "grad_norm": 0.6017474317413107, + "learning_rate": 9.060529709462613e-06, + "loss": 0.6519, "step": 1414 }, { - "epoch": 0.27, - "learning_rate": 1.7076574100254614e-05, - "loss": 0.9498, + "epoch": 0.544440169295883, + "grad_norm": 0.9500637862536985, + "learning_rate": 9.048123865351971e-06, + "loss": 0.5596, "step": 1415 }, { - "epoch": 0.27, - "learning_rate": 1.707216941294042e-05, - "loss": 0.9291, + "epoch": 0.5448249326664102, + "grad_norm": 0.9373866344174334, + "learning_rate": 9.035719499445545e-06, + "loss": 0.5848, "step": 1416 }, { - "epoch": 0.27, - "learning_rate": 1.706776197887712e-05, - "loss": 0.9102, + "epoch": 0.5452096960369373, + "grad_norm": 1.0114799081711168, + "learning_rate": 9.023316631006536e-06, + "loss": 0.6036, "step": 1417 }, { - "epoch": 0.27, - "learning_rate": 1.7063351799776514e-05, - "loss": 0.9541, + "epoch": 0.5455944594074644, + "grad_norm": 1.0336756941273009, + "learning_rate": 9.010915279295833e-06, + "loss": 0.5942, "step": 1418 }, { - "epoch": 0.27, - "learning_rate": 1.7058938877351456e-05, - "loss": 0.9296, + "epoch": 0.5459792227779915, + "grad_norm": 0.8962199697288343, + "learning_rate": 8.998515463571953e-06, + "loss": 0.5763, "step": 1419 }, { - "epoch": 0.27, - "learning_rate": 1.705452321331588e-05, - "loss": 0.9007, + "epoch": 0.5463639861485187, + "grad_norm": 0.9185906342118364, + "learning_rate": 8.986117203091042e-06, + "loss": 0.5982, "step": 1420 }, { - "epoch": 0.27, - "learning_rate": 1.7050104809384774e-05, - "loss": 0.7234, + "epoch": 0.5467487495190458, + "grad_norm": 0.9527209580051119, + "learning_rate": 8.973720517106814e-06, + "loss": 0.5878, "step": 1421 }, { - "epoch": 0.27, - "learning_rate": 1.70456836672742e-05, - "loss": 0.8672, + "epoch": 0.5471335128895729, + "grad_norm": 0.9236440852636413, + "learning_rate": 8.961325424870561e-06, + "loss": 0.6096, "step": 1422 }, { - "epoch": 0.27, - "learning_rate": 1.704125978870128e-05, - "loss": 0.9213, + "epoch": 0.5475182762601001, + "grad_norm": 0.9252511887705451, + "learning_rate": 8.948931945631082e-06, + "loss": 0.5646, "step": 1423 }, { - "epoch": 0.27, - "learning_rate": 1.7036833175384192e-05, - "loss": 0.9667, + "epoch": 0.5479030396306271, + "grad_norm": 1.0162755695471093, + "learning_rate": 8.936540098634675e-06, + "loss": 0.6327, "step": 1424 }, { - "epoch": 0.27, - "learning_rate": 1.7032403829042182e-05, - "loss": 0.7749, + "epoch": 0.5482878030011543, + "grad_norm": 0.914996077073819, + "learning_rate": 8.924149903125108e-06, + "loss": 0.5803, "step": 1425 }, { - "epoch": 0.27, - "learning_rate": 1.7027971751395563e-05, - "loss": 0.7578, + "epoch": 0.5486725663716814, + "grad_norm": 0.9535717265133994, + "learning_rate": 8.91176137834358e-06, + "loss": 0.5522, "step": 1426 }, { - "epoch": 0.27, - "learning_rate": 1.7023536944165697e-05, - "loss": 0.9518, + "epoch": 0.5490573297422086, + "grad_norm": 0.94938451454124, + "learning_rate": 8.899374543528695e-06, + "loss": 0.6533, "step": 1427 }, { - "epoch": 0.27, - "learning_rate": 1.7019099409075014e-05, - "loss": 0.9196, + "epoch": 0.5494420931127356, + "grad_norm": 0.9990452340114857, + "learning_rate": 8.886989417916435e-06, + "loss": 0.6312, "step": 1428 }, { - "epoch": 0.27, - "learning_rate": 1.7014659147847005e-05, - "loss": 0.9017, + "epoch": 0.5498268564832628, + "grad_norm": 1.1468442460998163, + "learning_rate": 8.87460602074013e-06, + "loss": 0.6061, "step": 1429 }, { - "epoch": 0.28, - "learning_rate": 1.701021616220621e-05, - "loss": 0.9664, + "epoch": 0.55021161985379, + "grad_norm": 0.9697659691886739, + "learning_rate": 8.862224371230418e-06, + "loss": 0.548, "step": 1430 }, { - "epoch": 0.28, - "learning_rate": 1.7005770453878234e-05, - "loss": 1.0336, + "epoch": 0.550596383224317, + "grad_norm": 0.7088142503647261, + "learning_rate": 8.84984448861523e-06, + "loss": 0.6318, "step": 1431 }, { - "epoch": 0.28, - "learning_rate": 1.7001322024589742e-05, - "loss": 0.9331, + "epoch": 0.5509811465948442, + "grad_norm": 0.9472234759157095, + "learning_rate": 8.837466392119752e-06, + "loss": 0.5553, "step": 1432 }, { - "epoch": 0.28, - "learning_rate": 1.6996870876068455e-05, - "loss": 0.9669, + "epoch": 0.5513659099653713, + "grad_norm": 0.9493040349649626, + "learning_rate": 8.825090100966396e-06, + "loss": 0.6004, "step": 1433 }, { - "epoch": 0.28, - "learning_rate": 1.6992417010043144e-05, - "loss": 0.8999, + "epoch": 0.5517506733358984, + "grad_norm": 0.6045412347828183, + "learning_rate": 8.81271563437476e-06, + "loss": 0.6657, "step": 1434 }, { - "epoch": 0.28, - "learning_rate": 1.6987960428243637e-05, - "loss": 0.9124, + "epoch": 0.5521354367064255, + "grad_norm": 0.9677580362737557, + "learning_rate": 8.800343011561633e-06, + "loss": 0.5701, "step": 1435 }, { - "epoch": 0.28, - "learning_rate": 1.6983501132400825e-05, - "loss": 0.969, + "epoch": 0.5525202000769527, + "grad_norm": 1.0132236170312463, + "learning_rate": 8.787972251740916e-06, + "loss": 0.5808, "step": 1436 }, { - "epoch": 0.28, - "learning_rate": 1.6979039124246643e-05, - "loss": 1.0539, + "epoch": 0.5529049634474797, + "grad_norm": 0.6142168647157175, + "learning_rate": 8.775603374123627e-06, + "loss": 0.6707, "step": 1437 }, { - "epoch": 0.28, - "learning_rate": 1.6974574405514083e-05, - "loss": 0.7439, + "epoch": 0.5532897268180069, + "grad_norm": 0.9590917099568571, + "learning_rate": 8.763236397917865e-06, + "loss": 0.5952, "step": 1438 }, { - "epoch": 0.28, - "learning_rate": 1.6970106977937192e-05, - "loss": 0.9821, + "epoch": 0.5536744901885341, + "grad_norm": 0.9690954550180992, + "learning_rate": 8.75087134232877e-06, + "loss": 0.6151, "step": 1439 }, { - "epoch": 0.28, - "learning_rate": 1.696563684325107e-05, - "loss": 0.7961, + "epoch": 0.5540592535590612, + "grad_norm": 0.940594863741157, + "learning_rate": 8.738508226558499e-06, + "loss": 0.5944, "step": 1440 }, { - "epoch": 0.28, - "learning_rate": 1.6961164003191862e-05, - "loss": 0.9213, + "epoch": 0.5544440169295883, + "grad_norm": 1.0310319993371182, + "learning_rate": 8.726147069806206e-06, + "loss": 0.5717, "step": 1441 }, { - "epoch": 0.28, - "learning_rate": 1.6956688459496767e-05, - "loss": 1.0347, + "epoch": 0.5548287803001154, + "grad_norm": 0.9329479008398374, + "learning_rate": 8.713787891267988e-06, + "loss": 0.5293, "step": 1442 }, { - "epoch": 0.28, - "learning_rate": 1.695221021390404e-05, - "loss": 0.9469, + "epoch": 0.5552135436706426, + "grad_norm": 0.914757679391358, + "learning_rate": 8.70143071013688e-06, + "loss": 0.5611, "step": 1443 }, { - "epoch": 0.28, - "learning_rate": 1.6947729268152972e-05, - "loss": 0.961, + "epoch": 0.5555983070411696, + "grad_norm": 0.9200053553024535, + "learning_rate": 8.689075545602816e-06, + "loss": 0.5739, "step": 1444 }, { - "epoch": 0.28, - "learning_rate": 1.6943245623983918e-05, - "loss": 1.036, + "epoch": 0.5559830704116968, + "grad_norm": 1.0204020665648343, + "learning_rate": 8.676722416852594e-06, + "loss": 0.566, "step": 1445 }, { - "epoch": 0.28, - "learning_rate": 1.6938759283138268e-05, - "loss": 0.9413, + "epoch": 0.556367833782224, + "grad_norm": 0.9459977663833542, + "learning_rate": 8.66437134306985e-06, + "loss": 0.577, "step": 1446 }, { - "epoch": 0.28, - "learning_rate": 1.693427024735847e-05, - "loss": 0.9771, + "epoch": 0.556752597152751, + "grad_norm": 0.642232995839935, + "learning_rate": 8.652022343435027e-06, + "loss": 0.6595, "step": 1447 }, { - "epoch": 0.28, - "learning_rate": 1.692977851838801e-05, - "loss": 0.9489, + "epoch": 0.5571373605232782, + "grad_norm": 0.9767186001772358, + "learning_rate": 8.63967543712536e-06, + "loss": 0.5259, "step": 1448 }, { - "epoch": 0.28, - "learning_rate": 1.6925284097971427e-05, - "loss": 0.9678, + "epoch": 0.5575221238938053, + "grad_norm": 0.9442476196546111, + "learning_rate": 8.627330643314818e-06, + "loss": 0.6018, "step": 1449 }, { - "epoch": 0.28, - "learning_rate": 1.6920786987854296e-05, - "loss": 0.9914, + "epoch": 0.5579068872643325, + "grad_norm": 0.8906057913934856, + "learning_rate": 8.614987981174093e-06, + "loss": 0.5541, "step": 1450 }, { - "epoch": 0.28, - "learning_rate": 1.691628718978325e-05, - "loss": 0.9532, + "epoch": 0.5582916506348595, + "grad_norm": 1.189305448476274, + "learning_rate": 8.602647469870573e-06, + "loss": 0.5802, "step": 1451 }, { - "epoch": 0.28, - "learning_rate": 1.691178470550596e-05, - "loss": 0.921, + "epoch": 0.5586764140053867, + "grad_norm": 0.9881838353517352, + "learning_rate": 8.590309128568303e-06, + "loss": 0.6144, "step": 1452 }, { - "epoch": 0.28, - "learning_rate": 1.6907279536771127e-05, - "loss": 1.0073, + "epoch": 0.5590611773759138, + "grad_norm": 0.9572912974011906, + "learning_rate": 8.57797297642795e-06, + "loss": 0.5866, "step": 1453 }, { - "epoch": 0.28, - "learning_rate": 1.6902771685328524e-05, - "loss": 0.8535, + "epoch": 0.5594459407464409, + "grad_norm": 1.0057973854776296, + "learning_rate": 8.565639032606794e-06, + "loss": 0.6152, "step": 1454 }, { - "epoch": 0.28, - "learning_rate": 1.6898261152928933e-05, - "loss": 1.0368, + "epoch": 0.5598307041169681, + "grad_norm": 1.0302644506360181, + "learning_rate": 8.553307316258678e-06, + "loss": 0.6107, "step": 1455 }, { - "epoch": 0.28, - "learning_rate": 1.6893747941324197e-05, - "loss": 0.9242, + "epoch": 0.5602154674874952, + "grad_norm": 0.9403263662439455, + "learning_rate": 8.540977846533986e-06, + "loss": 0.5468, "step": 1456 }, { - "epoch": 0.28, - "learning_rate": 1.6889232052267203e-05, - "loss": 0.943, + "epoch": 0.5606002308580224, + "grad_norm": 0.9096889045522819, + "learning_rate": 8.528650642579618e-06, + "loss": 0.5695, "step": 1457 }, { - "epoch": 0.28, - "learning_rate": 1.688471348751186e-05, - "loss": 0.8126, + "epoch": 0.5609849942285494, + "grad_norm": 1.1148452966125522, + "learning_rate": 8.516325723538949e-06, + "loss": 0.6355, "step": 1458 }, { - "epoch": 0.28, - "learning_rate": 1.688019224881313e-05, - "loss": 0.9572, + "epoch": 0.5613697575990766, + "grad_norm": 0.9633088708521582, + "learning_rate": 8.504003108551814e-06, + "loss": 0.6054, "step": 1459 }, { - "epoch": 0.28, - "learning_rate": 1.6875668337927014e-05, - "loss": 0.9655, + "epoch": 0.5617545209696037, + "grad_norm": 1.0965278088808013, + "learning_rate": 8.491682816754456e-06, + "loss": 0.6939, "step": 1460 }, { - "epoch": 0.28, - "learning_rate": 1.6871141756610544e-05, - "loss": 0.8336, + "epoch": 0.5621392843401308, + "grad_norm": 0.9793045328160709, + "learning_rate": 8.479364867279529e-06, + "loss": 0.5548, "step": 1461 }, { - "epoch": 0.28, - "learning_rate": 1.6866612506621788e-05, - "loss": 0.9214, + "epoch": 0.5625240477106579, + "grad_norm": 0.96702984136872, + "learning_rate": 8.467049279256034e-06, + "loss": 0.5597, "step": 1462 }, { - "epoch": 0.28, - "learning_rate": 1.6862080589719863e-05, - "loss": 0.943, + "epoch": 0.5629088110811851, + "grad_norm": 1.027805284835073, + "learning_rate": 8.45473607180931e-06, + "loss": 0.5793, "step": 1463 }, { - "epoch": 0.28, - "learning_rate": 1.6857546007664908e-05, - "loss": 0.9192, + "epoch": 0.5632935744517122, + "grad_norm": 0.8657745520650444, + "learning_rate": 8.442425264061e-06, + "loss": 0.549, "step": 1464 }, { - "epoch": 0.28, - "learning_rate": 1.6853008762218103e-05, - "loss": 1.0108, + "epoch": 0.5636783378222393, + "grad_norm": 0.9364641246548202, + "learning_rate": 8.430116875129023e-06, + "loss": 0.5975, "step": 1465 }, { - "epoch": 0.28, - "learning_rate": 1.684846885514166e-05, - "loss": 0.9991, + "epoch": 0.5640631011927665, + "grad_norm": 1.06463500316044, + "learning_rate": 8.417810924127533e-06, + "loss": 0.6118, "step": 1466 }, { - "epoch": 0.28, - "learning_rate": 1.6843926288198828e-05, - "loss": 0.7426, + "epoch": 0.5644478645632935, + "grad_norm": 0.9754125950353688, + "learning_rate": 8.40550743016691e-06, + "loss": 0.5378, "step": 1467 }, { - "epoch": 0.28, - "learning_rate": 1.683938106315389e-05, - "loss": 0.9148, + "epoch": 0.5648326279338207, + "grad_norm": 0.9837221143508172, + "learning_rate": 8.393206412353709e-06, + "loss": 0.5733, "step": 1468 }, { - "epoch": 0.28, - "learning_rate": 1.683483318177216e-05, - "loss": 0.7783, + "epoch": 0.5652173913043478, + "grad_norm": 0.944849779385138, + "learning_rate": 8.38090788979064e-06, + "loss": 0.545, "step": 1469 }, { - "epoch": 0.28, - "learning_rate": 1.6830282645819974e-05, - "loss": 0.9042, + "epoch": 0.565602154674875, + "grad_norm": 1.0170904687690927, + "learning_rate": 8.368611881576547e-06, + "loss": 0.5773, "step": 1470 }, { - "epoch": 0.28, - "learning_rate": 1.6825729457064718e-05, - "loss": 1.0185, + "epoch": 0.5659869180454021, + "grad_norm": 0.9365494197614428, + "learning_rate": 8.35631840680636e-06, + "loss": 0.6245, "step": 1471 }, { - "epoch": 0.28, - "learning_rate": 1.6821173617274793e-05, - "loss": 0.8716, + "epoch": 0.5663716814159292, + "grad_norm": 0.9804176936880831, + "learning_rate": 8.344027484571075e-06, + "loss": 0.6329, "step": 1472 }, { - "epoch": 0.28, - "learning_rate": 1.6816615128219635e-05, - "loss": 0.9098, + "epoch": 0.5667564447864564, + "grad_norm": 0.9008259998982997, + "learning_rate": 8.331739133957729e-06, + "loss": 0.5336, "step": 1473 }, { - "epoch": 0.28, - "learning_rate": 1.681205399166971e-05, - "loss": 0.915, + "epoch": 0.5671412081569834, + "grad_norm": 1.0241252026362389, + "learning_rate": 8.319453374049367e-06, + "loss": 0.6356, "step": 1474 }, { - "epoch": 0.28, - "learning_rate": 1.6807490209396506e-05, - "loss": 0.9498, + "epoch": 0.5675259715275106, + "grad_norm": 1.0036927351551406, + "learning_rate": 8.307170223925003e-06, + "loss": 0.5691, "step": 1475 }, { - "epoch": 0.28, - "learning_rate": 1.6802923783172553e-05, - "loss": 0.9163, + "epoch": 0.5679107348980377, + "grad_norm": 1.0499801975267546, + "learning_rate": 8.294889702659602e-06, + "loss": 0.595, "step": 1476 }, { - "epoch": 0.28, - "learning_rate": 1.679835471477139e-05, - "loss": 0.9393, + "epoch": 0.5682954982685648, + "grad_norm": 0.8421985440732868, + "learning_rate": 8.282611829324049e-06, + "loss": 0.5403, "step": 1477 }, { - "epoch": 0.28, - "learning_rate": 1.6793783005967593e-05, - "loss": 0.9649, + "epoch": 0.5686802616390919, + "grad_norm": 0.96317987844597, + "learning_rate": 8.270336622985116e-06, + "loss": 0.6171, "step": 1478 }, { - "epoch": 0.28, - "learning_rate": 1.678920865853676e-05, - "loss": 0.7398, + "epoch": 0.5690650250096191, + "grad_norm": 0.9162804174420532, + "learning_rate": 8.258064102705428e-06, + "loss": 0.5569, "step": 1479 }, { - "epoch": 0.28, - "learning_rate": 1.678463167425552e-05, - "loss": 0.9326, + "epoch": 0.5694497883801463, + "grad_norm": 0.9741679375875378, + "learning_rate": 8.245794287543447e-06, + "loss": 0.6333, "step": 1480 }, { - "epoch": 0.28, - "learning_rate": 1.6780052054901512e-05, - "loss": 0.9208, + "epoch": 0.5698345517506733, + "grad_norm": 0.9100423440719114, + "learning_rate": 8.233527196553428e-06, + "loss": 0.5451, "step": 1481 }, { - "epoch": 0.29, - "learning_rate": 1.6775469802253416e-05, - "loss": 0.8805, + "epoch": 0.5702193151212005, + "grad_norm": 0.9590207673142193, + "learning_rate": 8.221262848785395e-06, + "loss": 0.6109, "step": 1482 }, { - "epoch": 0.29, - "learning_rate": 1.6770884918090923e-05, - "loss": 0.9959, + "epoch": 0.5706040784917276, + "grad_norm": 0.9677089849748892, + "learning_rate": 8.20900126328512e-06, + "loss": 0.5631, "step": 1483 }, { - "epoch": 0.29, - "learning_rate": 1.6766297404194745e-05, - "loss": 0.9241, + "epoch": 0.5709888418622547, + "grad_norm": 1.0740825359487336, + "learning_rate": 8.196742459094079e-06, + "loss": 0.5967, "step": 1484 }, { - "epoch": 0.29, - "learning_rate": 1.6761707262346624e-05, - "loss": 0.911, + "epoch": 0.5713736052327818, + "grad_norm": 0.9672126773447882, + "learning_rate": 8.184486455249424e-06, + "loss": 0.5777, "step": 1485 }, { - "epoch": 0.29, - "learning_rate": 1.675711449432932e-05, - "loss": 0.9408, + "epoch": 0.571758368603309, + "grad_norm": 0.92291528059833, + "learning_rate": 8.172233270783966e-06, + "loss": 0.5662, "step": 1486 }, { - "epoch": 0.29, - "learning_rate": 1.6752519101926606e-05, - "loss": 0.8936, + "epoch": 0.572143131973836, + "grad_norm": 0.9621752972709625, + "learning_rate": 8.15998292472614e-06, + "loss": 0.6136, "step": 1487 }, { - "epoch": 0.29, - "learning_rate": 1.6747921086923284e-05, - "loss": 0.9043, + "epoch": 0.5725278953443632, + "grad_norm": 1.0535726184781045, + "learning_rate": 8.147735436099967e-06, + "loss": 0.5682, "step": 1488 }, { - "epoch": 0.29, - "learning_rate": 1.674332045110517e-05, - "loss": 0.9243, + "epoch": 0.5729126587148904, + "grad_norm": 0.9709504095713601, + "learning_rate": 8.135490823925027e-06, + "loss": 0.6197, "step": 1489 }, { - "epoch": 0.29, - "learning_rate": 1.6738717196259092e-05, - "loss": 0.9, + "epoch": 0.5732974220854175, + "grad_norm": 0.9062480355758354, + "learning_rate": 8.123249107216446e-06, + "loss": 0.542, "step": 1490 }, { - "epoch": 0.29, - "learning_rate": 1.673411132417291e-05, - "loss": 0.9458, + "epoch": 0.5736821854559446, + "grad_norm": 1.040876318423481, + "learning_rate": 8.111010304984841e-06, + "loss": 0.5582, "step": 1491 }, { - "epoch": 0.29, - "learning_rate": 1.672950283663548e-05, - "loss": 0.9831, + "epoch": 0.5740669488264717, + "grad_norm": 0.9369476830173759, + "learning_rate": 8.098774436236308e-06, + "loss": 0.5256, "step": 1492 }, { - "epoch": 0.29, - "learning_rate": 1.6724891735436697e-05, - "loss": 0.9357, + "epoch": 0.5744517121969989, + "grad_norm": 0.9571143592694651, + "learning_rate": 8.086541519972388e-06, + "loss": 0.559, "step": 1493 }, { - "epoch": 0.29, - "learning_rate": 1.6720278022367453e-05, - "loss": 0.8634, + "epoch": 0.5748364755675259, + "grad_norm": 1.0328243533280066, + "learning_rate": 8.074311575190039e-06, + "loss": 0.5813, "step": 1494 }, { - "epoch": 0.29, - "learning_rate": 1.6715661699219664e-05, - "loss": 0.9845, + "epoch": 0.5752212389380531, + "grad_norm": 1.0724113473585177, + "learning_rate": 8.062084620881598e-06, + "loss": 0.5895, "step": 1495 }, { - "epoch": 0.29, - "learning_rate": 1.6711042767786257e-05, - "loss": 0.9234, + "epoch": 0.5756060023085803, + "grad_norm": 0.9694991602439562, + "learning_rate": 8.049860676034762e-06, + "loss": 0.6112, "step": 1496 }, { - "epoch": 0.29, - "learning_rate": 1.6706421229861168e-05, - "loss": 0.908, + "epoch": 0.5759907656791073, + "grad_norm": 1.1125972276496021, + "learning_rate": 8.037639759632558e-06, + "loss": 0.558, "step": 1497 }, { - "epoch": 0.29, - "learning_rate": 1.6701797087239354e-05, - "loss": 0.7727, + "epoch": 0.5763755290496345, + "grad_norm": 0.859798549602274, + "learning_rate": 8.025421890653303e-06, + "loss": 0.5669, "step": 1498 }, { - "epoch": 0.29, - "learning_rate": 1.6697170341716772e-05, - "loss": 0.8895, + "epoch": 0.5767602924201616, + "grad_norm": 0.9714449076273433, + "learning_rate": 8.013207088070582e-06, + "loss": 0.5814, "step": 1499 }, { - "epoch": 0.29, - "learning_rate": 1.6692540995090403e-05, - "loss": 0.9615, + "epoch": 0.5771450557906888, + "grad_norm": 0.9876492898543677, + "learning_rate": 8.000995370853227e-06, + "loss": 0.5788, "step": 1500 }, { - "epoch": 0.29, - "learning_rate": 1.668790904915823e-05, - "loss": 0.9531, + "epoch": 0.5775298191612158, + "grad_norm": 0.9719397890700671, + "learning_rate": 7.98878675796527e-06, + "loss": 0.6075, "step": 1501 }, { - "epoch": 0.29, - "learning_rate": 1.6683274505719248e-05, - "loss": 0.9009, + "epoch": 0.577914582531743, + "grad_norm": 0.9918368069652902, + "learning_rate": 7.976581268365924e-06, + "loss": 0.5412, "step": 1502 }, { - "epoch": 0.29, - "learning_rate": 1.6678637366573455e-05, - "loss": 0.9321, + "epoch": 0.5782993459022701, + "grad_norm": 1.0068244085443292, + "learning_rate": 7.964378921009552e-06, + "loss": 0.5949, "step": 1503 }, { - "epoch": 0.29, - "learning_rate": 1.667399763352187e-05, - "loss": 0.9813, + "epoch": 0.5786841092727972, + "grad_norm": 1.056286804091604, + "learning_rate": 7.952179734845642e-06, + "loss": 0.6211, "step": 1504 }, { - "epoch": 0.29, - "learning_rate": 1.666935530836651e-05, - "loss": 0.9694, + "epoch": 0.5790688726433244, + "grad_norm": 0.9436538497221629, + "learning_rate": 7.93998372881876e-06, + "loss": 0.5823, "step": 1505 }, { - "epoch": 0.29, - "learning_rate": 1.6664710392910396e-05, - "loss": 0.8069, + "epoch": 0.5794536360138515, + "grad_norm": 1.0466179376668707, + "learning_rate": 7.92779092186855e-06, + "loss": 0.6208, "step": 1506 }, { - "epoch": 0.29, - "learning_rate": 1.6660062888957564e-05, - "loss": 0.9333, + "epoch": 0.5798383993843786, + "grad_norm": 0.9523303610197328, + "learning_rate": 7.915601332929678e-06, + "loss": 0.554, "step": 1507 }, { - "epoch": 0.29, - "learning_rate": 1.665541279831305e-05, - "loss": 0.9094, + "epoch": 0.5802231627549057, + "grad_norm": 0.9592994244738012, + "learning_rate": 7.903414980931813e-06, + "loss": 0.5247, "step": 1508 }, { - "epoch": 0.29, - "learning_rate": 1.6650760122782898e-05, - "loss": 0.8204, + "epoch": 0.5806079261254329, + "grad_norm": 0.877014664113631, + "learning_rate": 7.8912318847996e-06, + "loss": 0.5457, "step": 1509 }, { - "epoch": 0.29, - "learning_rate": 1.6646104864174147e-05, - "loss": 0.9544, + "epoch": 0.58099268949596, + "grad_norm": 1.0099303128900625, + "learning_rate": 7.879052063452626e-06, + "loss": 0.5992, "step": 1510 }, { - "epoch": 0.29, - "learning_rate": 1.664144702429485e-05, - "loss": 0.9166, + "epoch": 0.5813774528664871, + "grad_norm": 0.9783725442965545, + "learning_rate": 7.866875535805394e-06, + "loss": 0.5474, "step": 1511 }, { - "epoch": 0.29, - "learning_rate": 1.663678660495406e-05, - "loss": 0.8814, + "epoch": 0.5817622162370142, + "grad_norm": 0.9489753124423047, + "learning_rate": 7.85470232076729e-06, + "loss": 0.5786, "step": 1512 }, { - "epoch": 0.29, - "learning_rate": 1.663212360796183e-05, - "loss": 0.9718, + "epoch": 0.5821469796075414, + "grad_norm": 1.019082272415777, + "learning_rate": 7.842532437242559e-06, + "loss": 0.5702, "step": 1513 }, { - "epoch": 0.29, - "learning_rate": 1.662745803512921e-05, - "loss": 0.9709, + "epoch": 0.5825317429780685, + "grad_norm": 1.0060700631892618, + "learning_rate": 7.83036590413027e-06, + "loss": 0.6017, "step": 1514 }, { - "epoch": 0.29, - "learning_rate": 1.662278988826826e-05, - "loss": 0.9703, + "epoch": 0.5829165063485956, + "grad_norm": 1.024037438236854, + "learning_rate": 7.818202740324287e-06, + "loss": 0.6068, "step": 1515 }, { - "epoch": 0.29, - "learning_rate": 1.6618119169192027e-05, - "loss": 0.9923, + "epoch": 0.5833012697191228, + "grad_norm": 0.9108171328169132, + "learning_rate": 7.806042964713248e-06, + "loss": 0.5495, "step": 1516 }, { - "epoch": 0.29, - "learning_rate": 1.661344587971457e-05, - "loss": 1.0293, + "epoch": 0.5836860330896498, + "grad_norm": 1.0092553085527995, + "learning_rate": 7.793886596180521e-06, + "loss": 0.5852, "step": 1517 }, { - "epoch": 0.29, - "learning_rate": 1.6608770021650945e-05, - "loss": 0.958, + "epoch": 0.584070796460177, + "grad_norm": 0.9328356544673627, + "learning_rate": 7.78173365360419e-06, + "loss": 0.6433, "step": 1518 }, { - "epoch": 0.29, - "learning_rate": 1.6604091596817193e-05, - "loss": 0.9294, + "epoch": 0.5844555598307041, + "grad_norm": 1.0232290177167747, + "learning_rate": 7.769584155857019e-06, + "loss": 0.5786, "step": 1519 }, { - "epoch": 0.29, - "learning_rate": 1.6599410607030363e-05, - "loss": 0.9747, + "epoch": 0.5848403232012312, + "grad_norm": 0.8971518623262915, + "learning_rate": 7.757438121806414e-06, + "loss": 0.5342, "step": 1520 }, { - "epoch": 0.29, - "learning_rate": 1.6594727054108498e-05, - "loss": 0.9479, + "epoch": 0.5852250865717584, + "grad_norm": 0.9487737983709945, + "learning_rate": 7.745295570314412e-06, + "loss": 0.5539, "step": 1521 }, { - "epoch": 0.29, - "learning_rate": 1.659004093987064e-05, - "loss": 0.9516, + "epoch": 0.5856098499422855, + "grad_norm": 1.0488449240232716, + "learning_rate": 7.733156520237633e-06, + "loss": 0.611, "step": 1522 }, { - "epoch": 0.29, - "learning_rate": 1.6585352266136814e-05, - "loss": 0.9782, + "epoch": 0.5859946133128127, + "grad_norm": 0.9704923013140451, + "learning_rate": 7.721020990427268e-06, + "loss": 0.5387, "step": 1523 }, { - "epoch": 0.29, - "learning_rate": 1.6580661034728055e-05, - "loss": 0.922, + "epoch": 0.5863793766833397, + "grad_norm": 0.8911377768696978, + "learning_rate": 7.708888999729036e-06, + "loss": 0.5832, "step": 1524 }, { - "epoch": 0.29, - "learning_rate": 1.6575967247466376e-05, - "loss": 0.9479, + "epoch": 0.5867641400538669, + "grad_norm": 0.9519808714183091, + "learning_rate": 7.69676056698316e-06, + "loss": 0.6101, "step": 1525 }, { - "epoch": 0.29, - "learning_rate": 1.657127090617479e-05, - "loss": 0.9382, + "epoch": 0.587148903424394, + "grad_norm": 1.0619310455497668, + "learning_rate": 7.68463571102434e-06, + "loss": 0.5592, "step": 1526 }, { - "epoch": 0.29, - "learning_rate": 1.656657201267731e-05, - "loss": 0.9134, + "epoch": 0.5875336667949211, + "grad_norm": 0.9475302808658681, + "learning_rate": 7.672514450681721e-06, + "loss": 0.6025, "step": 1527 }, { - "epoch": 0.29, - "learning_rate": 1.6561870568798927e-05, - "loss": 0.973, + "epoch": 0.5879184301654482, + "grad_norm": 0.9112495195721528, + "learning_rate": 7.66039680477886e-06, + "loss": 0.539, "step": 1528 }, { - "epoch": 0.29, - "learning_rate": 1.655716657636562e-05, - "loss": 0.8725, + "epoch": 0.5883031935359754, + "grad_norm": 0.9162877447718418, + "learning_rate": 7.648282792133711e-06, + "loss": 0.5155, "step": 1529 }, { - "epoch": 0.29, - "learning_rate": 1.6552460037204382e-05, - "loss": 0.9628, + "epoch": 0.5886879569065026, + "grad_norm": 1.01198622338008, + "learning_rate": 7.636172431558575e-06, + "loss": 0.5746, "step": 1530 }, { - "epoch": 0.29, - "learning_rate": 1.6547750953143168e-05, - "loss": 0.8862, + "epoch": 0.5890727202770296, + "grad_norm": 0.8582976446419343, + "learning_rate": 7.6240657418600846e-06, + "loss": 0.5317, "step": 1531 }, { - "epoch": 0.29, - "learning_rate": 1.654303932601093e-05, - "loss": 0.9697, + "epoch": 0.5894574836475568, + "grad_norm": 1.0626679101809673, + "learning_rate": 7.611962741839178e-06, + "loss": 0.5697, "step": 1532 }, { - "epoch": 0.29, - "learning_rate": 1.6538325157637614e-05, - "loss": 1.0192, + "epoch": 0.5898422470180839, + "grad_norm": 1.0050277780033958, + "learning_rate": 7.599863450291056e-06, + "loss": 0.5815, "step": 1533 }, { - "epoch": 0.3, - "learning_rate": 1.653360844985415e-05, - "loss": 0.7611, + "epoch": 0.590227010388611, + "grad_norm": 1.0310891869254732, + "learning_rate": 7.587767886005164e-06, + "loss": 0.5714, "step": 1534 }, { - "epoch": 0.3, - "learning_rate": 1.652888920449245e-05, - "loss": 0.9518, + "epoch": 0.5906117737591381, + "grad_norm": 0.9453488628731961, + "learning_rate": 7.575676067765154e-06, + "loss": 0.5668, "step": 1535 }, { - "epoch": 0.3, - "learning_rate": 1.6524167423385414e-05, - "loss": 0.9771, + "epoch": 0.5909965371296653, + "grad_norm": 0.9076486473507975, + "learning_rate": 7.563588014348871e-06, + "loss": 0.5587, "step": 1536 }, { - "epoch": 0.3, - "learning_rate": 1.651944310836693e-05, - "loss": 0.9612, + "epoch": 0.5913813005001923, + "grad_norm": 0.8970088004279343, + "learning_rate": 7.551503744528304e-06, + "loss": 0.5741, "step": 1537 }, { - "epoch": 0.3, - "learning_rate": 1.6514716261271866e-05, - "loss": 0.9485, + "epoch": 0.5917660638707195, + "grad_norm": 0.9504470220232004, + "learning_rate": 7.539423277069568e-06, + "loss": 0.5507, "step": 1538 }, { - "epoch": 0.3, - "learning_rate": 1.6509986883936073e-05, - "loss": 0.8889, + "epoch": 0.5921508272412467, + "grad_norm": 1.0126960382319814, + "learning_rate": 7.52734663073288e-06, + "loss": 0.6187, "step": 1539 }, { - "epoch": 0.3, - "learning_rate": 1.650525497819639e-05, - "loss": 0.7782, + "epoch": 0.5925355906117737, + "grad_norm": 0.9646093201899844, + "learning_rate": 7.515273824272516e-06, + "loss": 0.5714, "step": 1540 }, { - "epoch": 0.3, - "learning_rate": 1.6500520545890634e-05, - "loss": 0.883, + "epoch": 0.5929203539823009, + "grad_norm": 0.9314295535966145, + "learning_rate": 7.503204876436785e-06, + "loss": 0.5516, "step": 1541 }, { - "epoch": 0.3, - "learning_rate": 1.6495783588857605e-05, - "loss": 0.9458, + "epoch": 0.593305117352828, + "grad_norm": 0.9613172383619237, + "learning_rate": 7.491139805968018e-06, + "loss": 0.6169, "step": 1542 }, { - "epoch": 0.3, - "learning_rate": 1.649104410893708e-05, - "loss": 0.7313, + "epoch": 0.5936898807233552, + "grad_norm": 1.030383497996588, + "learning_rate": 7.4790786316025125e-06, + "loss": 0.5741, "step": 1543 }, { - "epoch": 0.3, - "learning_rate": 1.648630210796982e-05, - "loss": 0.9331, + "epoch": 0.5940746440938822, + "grad_norm": 0.9402135467541373, + "learning_rate": 7.467021372070515e-06, + "loss": 0.5628, "step": 1544 }, { - "epoch": 0.3, - "learning_rate": 1.6481557587797562e-05, - "loss": 0.953, + "epoch": 0.5944594074644094, + "grad_norm": 1.005799355374073, + "learning_rate": 7.4549680460962044e-06, + "loss": 0.5815, "step": 1545 }, { - "epoch": 0.3, - "learning_rate": 1.6476810550263023e-05, - "loss": 0.947, + "epoch": 0.5948441708349365, + "grad_norm": 0.9550499609223584, + "learning_rate": 7.4429186723976425e-06, + "loss": 0.5463, "step": 1546 }, { - "epoch": 0.3, - "learning_rate": 1.6472060997209898e-05, - "loss": 0.8639, + "epoch": 0.5952289342054636, + "grad_norm": 0.9661865908697149, + "learning_rate": 7.43087326968675e-06, + "loss": 0.5383, "step": 1547 }, { - "epoch": 0.3, - "learning_rate": 1.6467308930482863e-05, - "loss": 0.9366, + "epoch": 0.5956136975759908, + "grad_norm": 1.0541802157677915, + "learning_rate": 7.418831856669286e-06, + "loss": 0.5696, "step": 1548 }, { - "epoch": 0.3, - "learning_rate": 1.6462554351927558e-05, - "loss": 0.8583, + "epoch": 0.5959984609465179, + "grad_norm": 1.0517246527228774, + "learning_rate": 7.406794452044816e-06, + "loss": 0.5559, "step": 1549 }, { - "epoch": 0.3, - "learning_rate": 1.6457797263390613e-05, - "loss": 0.865, + "epoch": 0.596383224317045, + "grad_norm": 0.9295885572002459, + "learning_rate": 7.394761074506679e-06, + "loss": 0.549, "step": 1550 }, { - "epoch": 0.3, - "learning_rate": 1.6453037666719624e-05, - "loss": 0.9608, + "epoch": 0.5967679876875721, + "grad_norm": 0.977889191567664, + "learning_rate": 7.382731742741953e-06, + "loss": 0.5962, "step": 1551 }, { - "epoch": 0.3, - "learning_rate": 1.6448275563763162e-05, - "loss": 0.8683, + "epoch": 0.5971527510580993, + "grad_norm": 0.9390172729555465, + "learning_rate": 7.370706475431446e-06, + "loss": 0.6107, "step": 1552 }, { - "epoch": 0.3, - "learning_rate": 1.644351095637078e-05, - "loss": 0.8994, + "epoch": 0.5975375144286263, + "grad_norm": 0.6204734819504844, + "learning_rate": 7.358685291249644e-06, + "loss": 0.626, "step": 1553 }, { - "epoch": 0.3, - "learning_rate": 1.6438743846392987e-05, - "loss": 0.9756, + "epoch": 0.5979222777991535, + "grad_norm": 0.979930049763565, + "learning_rate": 7.346668208864695e-06, + "loss": 0.5742, "step": 1554 }, { - "epoch": 0.3, - "learning_rate": 1.6433974235681274e-05, - "loss": 1.0314, + "epoch": 0.5983070411696807, + "grad_norm": 0.9423034551982358, + "learning_rate": 7.33465524693838e-06, + "loss": 0.5507, "step": 1555 }, { - "epoch": 0.3, - "learning_rate": 1.6429202126088112e-05, - "loss": 0.929, + "epoch": 0.5986918045402078, + "grad_norm": 1.0019058385019386, + "learning_rate": 7.322646424126079e-06, + "loss": 0.559, "step": 1556 }, { - "epoch": 0.3, - "learning_rate": 1.6424427519466925e-05, - "loss": 0.9086, + "epoch": 0.5990765679107349, + "grad_norm": 1.0722638746693622, + "learning_rate": 7.310641759076742e-06, + "loss": 0.5523, "step": 1557 }, { - "epoch": 0.3, - "learning_rate": 1.641965041767212e-05, - "loss": 0.8515, + "epoch": 0.599461331281262, + "grad_norm": 1.0274403776731222, + "learning_rate": 7.2986412704328625e-06, + "loss": 0.5548, "step": 1558 }, { - "epoch": 0.3, - "learning_rate": 1.6414870822559064e-05, - "loss": 0.976, + "epoch": 0.5998460946517892, + "grad_norm": 0.6427145177490442, + "learning_rate": 7.286644976830457e-06, + "loss": 0.6537, "step": 1559 }, { - "epoch": 0.3, - "learning_rate": 1.6410088735984103e-05, - "loss": 0.9362, + "epoch": 0.6002308580223162, + "grad_norm": 0.5814161153977546, + "learning_rate": 7.274652896899015e-06, + "loss": 0.6315, "step": 1560 }, { - "epoch": 0.3, - "learning_rate": 1.6405304159804534e-05, - "loss": 0.9539, + "epoch": 0.6006156213928434, + "grad_norm": 1.0949055244465278, + "learning_rate": 7.262665049261489e-06, + "loss": 0.6103, "step": 1561 }, { - "epoch": 0.3, - "learning_rate": 1.6400517095878644e-05, - "loss": 0.9557, + "epoch": 0.6010003847633705, + "grad_norm": 0.899852914745554, + "learning_rate": 7.250681452534261e-06, + "loss": 0.5624, "step": 1562 }, { - "epoch": 0.3, - "learning_rate": 1.6395727546065665e-05, - "loss": 0.9724, + "epoch": 0.6013851481338977, + "grad_norm": 1.0459688017410076, + "learning_rate": 7.238702125327106e-06, + "loss": 0.5862, "step": 1563 }, { - "epoch": 0.3, - "learning_rate": 1.6390935512225806e-05, - "loss": 0.9833, + "epoch": 0.6017699115044248, + "grad_norm": 0.9586751440807577, + "learning_rate": 7.226727086243168e-06, + "loss": 0.6137, "step": 1564 }, { - "epoch": 0.3, - "learning_rate": 1.6386140996220232e-05, - "loss": 1.028, + "epoch": 0.6021546748749519, + "grad_norm": 0.645854338222848, + "learning_rate": 7.214756353878942e-06, + "loss": 0.6734, "step": 1565 }, { - "epoch": 0.3, - "learning_rate": 1.6381343999911088e-05, - "loss": 0.9992, + "epoch": 0.6025394382454791, + "grad_norm": 0.9145139875761001, + "learning_rate": 7.202789946824227e-06, + "loss": 0.5689, "step": 1566 }, { - "epoch": 0.3, - "learning_rate": 1.6376544525161463e-05, - "loss": 0.9945, + "epoch": 0.6029242016160061, + "grad_norm": 1.0390651454838415, + "learning_rate": 7.1908278836621e-06, + "loss": 0.5737, "step": 1567 }, { - "epoch": 0.3, - "learning_rate": 1.6371742573835426e-05, - "loss": 0.8575, + "epoch": 0.6033089649865333, + "grad_norm": 0.9246182433589133, + "learning_rate": 7.178870182968904e-06, + "loss": 0.5786, "step": 1568 }, { - "epoch": 0.3, - "learning_rate": 1.636693814779799e-05, - "loss": 1.0388, + "epoch": 0.6036937283570604, + "grad_norm": 0.8981795972886865, + "learning_rate": 7.166916863314199e-06, + "loss": 0.5751, "step": 1569 }, { - "epoch": 0.3, - "learning_rate": 1.6362131248915145e-05, - "loss": 0.9146, + "epoch": 0.6040784917275875, + "grad_norm": 1.0122285073500878, + "learning_rate": 7.154967943260748e-06, + "loss": 0.586, "step": 1570 }, { - "epoch": 0.3, - "learning_rate": 1.6357321879053833e-05, - "loss": 0.9951, + "epoch": 0.6044632550981146, + "grad_norm": 0.6106569368399559, + "learning_rate": 7.143023441364471e-06, + "loss": 0.6272, "step": 1571 }, { - "epoch": 0.3, - "learning_rate": 1.6352510040081962e-05, - "loss": 0.9532, + "epoch": 0.6048480184686418, + "grad_norm": 0.6492669010826267, + "learning_rate": 7.131083376174441e-06, + "loss": 0.6485, "step": 1572 }, { - "epoch": 0.3, - "learning_rate": 1.634769573386839e-05, - "loss": 0.9075, + "epoch": 0.605232781839169, + "grad_norm": 1.0719485766829455, + "learning_rate": 7.119147766232832e-06, + "loss": 0.6182, "step": 1573 }, { - "epoch": 0.3, - "learning_rate": 1.634287896228294e-05, - "loss": 0.9087, + "epoch": 0.605617545209696, + "grad_norm": 0.9344590332350262, + "learning_rate": 7.107216630074895e-06, + "loss": 0.521, "step": 1574 }, { - "epoch": 0.3, - "learning_rate": 1.6338059727196386e-05, - "loss": 0.8947, + "epoch": 0.6060023085802232, + "grad_norm": 0.9505395561984888, + "learning_rate": 7.09528998622895e-06, + "loss": 0.556, "step": 1575 }, { - "epoch": 0.3, - "learning_rate": 1.6333238030480473e-05, - "loss": 0.9242, + "epoch": 0.6063870719507503, + "grad_norm": 0.998148727601241, + "learning_rate": 7.083367853216323e-06, + "loss": 0.5635, "step": 1576 }, { - "epoch": 0.3, - "learning_rate": 1.6328413874007884e-05, - "loss": 0.9552, + "epoch": 0.6067718353212774, + "grad_norm": 0.9978516826883777, + "learning_rate": 7.071450249551342e-06, + "loss": 0.5997, "step": 1577 }, { - "epoch": 0.3, - "learning_rate": 1.6323587259652267e-05, - "loss": 0.8983, + "epoch": 0.6071565986918045, + "grad_norm": 1.0459517410859425, + "learning_rate": 7.059537193741306e-06, + "loss": 0.5509, "step": 1578 }, { - "epoch": 0.3, - "learning_rate": 1.6318758189288227e-05, - "loss": 0.9467, + "epoch": 0.6075413620623317, + "grad_norm": 1.0383331590100593, + "learning_rate": 7.047628704286446e-06, + "loss": 0.5963, "step": 1579 }, { - "epoch": 0.3, - "learning_rate": 1.6313926664791316e-05, - "loss": 0.9451, + "epoch": 0.6079261254328588, + "grad_norm": 1.0754675074601725, + "learning_rate": 7.035724799679898e-06, + "loss": 0.5823, "step": 1580 }, { - "epoch": 0.3, - "learning_rate": 1.6309092688038047e-05, - "loss": 0.9779, + "epoch": 0.6083108888033859, + "grad_norm": 0.9709466962136529, + "learning_rate": 7.023825498407689e-06, + "loss": 0.6282, "step": 1581 }, { - "epoch": 0.3, - "learning_rate": 1.6304256260905872e-05, - "loss": 0.9805, + "epoch": 0.6086956521739131, + "grad_norm": 0.9234372826814818, + "learning_rate": 7.011930818948688e-06, + "loss": 0.5322, "step": 1582 }, { - "epoch": 0.3, - "learning_rate": 1.6299417385273216e-05, - "loss": 0.9245, + "epoch": 0.6090804155444401, + "grad_norm": 1.0831210887773934, + "learning_rate": 7.000040779774591e-06, + "loss": 0.6052, "step": 1583 }, { - "epoch": 0.3, - "learning_rate": 1.629457606301943e-05, - "loss": 0.7595, + "epoch": 0.6094651789149673, + "grad_norm": 0.9764398694422771, + "learning_rate": 6.9881553993498805e-06, + "loss": 0.5602, "step": 1584 }, { - "epoch": 0.3, - "learning_rate": 1.6289732296024837e-05, - "loss": 1.014, + "epoch": 0.6098499422854944, + "grad_norm": 1.012044858942468, + "learning_rate": 6.97627469613182e-06, + "loss": 0.6112, "step": 1585 }, { - "epoch": 0.31, - "learning_rate": 1.6284886086170697e-05, - "loss": 0.885, + "epoch": 0.6102347056560216, + "grad_norm": 1.032860048794341, + "learning_rate": 6.9643986885703955e-06, + "loss": 0.6435, "step": 1586 }, { - "epoch": 0.31, - "learning_rate": 1.628003743533922e-05, - "loss": 0.7466, + "epoch": 0.6106194690265486, + "grad_norm": 0.9027630331021219, + "learning_rate": 6.952527395108302e-06, + "loss": 0.5383, "step": 1587 }, { - "epoch": 0.31, - "learning_rate": 1.6275186345413566e-05, - "loss": 0.9774, + "epoch": 0.6110042323970758, + "grad_norm": 0.881035855640019, + "learning_rate": 6.9406608341809215e-06, + "loss": 0.5961, "step": 1588 }, { - "epoch": 0.31, - "learning_rate": 1.627033281827785e-05, - "loss": 0.754, + "epoch": 0.611388995767603, + "grad_norm": 0.9230484605384653, + "learning_rate": 6.928799024216282e-06, + "loss": 0.6005, "step": 1589 }, { - "epoch": 0.31, - "learning_rate": 1.6265476855817116e-05, - "loss": 0.8224, + "epoch": 0.61177375913813, + "grad_norm": 0.9209176129067688, + "learning_rate": 6.91694198363503e-06, + "loss": 0.5824, "step": 1590 }, { - "epoch": 0.31, - "learning_rate": 1.6260618459917366e-05, - "loss": 0.9681, + "epoch": 0.6121585225086572, + "grad_norm": 0.961101228951713, + "learning_rate": 6.905089730850416e-06, + "loss": 0.6003, "step": 1591 }, { - "epoch": 0.31, - "learning_rate": 1.6255757632465553e-05, - "loss": 0.9843, + "epoch": 0.6125432858791843, + "grad_norm": 1.0753290255688088, + "learning_rate": 6.893242284268244e-06, + "loss": 0.5246, "step": 1592 }, { - "epoch": 0.31, - "learning_rate": 1.625089437534956e-05, - "loss": 0.9104, + "epoch": 0.6129280492497114, + "grad_norm": 1.177686025264356, + "learning_rate": 6.8813996622868584e-06, + "loss": 0.5928, "step": 1593 }, { - "epoch": 0.31, - "learning_rate": 1.624602869045822e-05, - "loss": 0.9223, + "epoch": 0.6133128126202385, + "grad_norm": 0.9348017499792318, + "learning_rate": 6.869561883297116e-06, + "loss": 0.5633, "step": 1594 }, { - "epoch": 0.31, - "learning_rate": 1.624116057968131e-05, - "loss": 0.9375, + "epoch": 0.6136975759907657, + "grad_norm": 0.9736558129367524, + "learning_rate": 6.857728965682344e-06, + "loss": 0.6559, "step": 1595 }, { - "epoch": 0.31, - "learning_rate": 1.6236290044909543e-05, - "loss": 1.0899, + "epoch": 0.6140823393612928, + "grad_norm": 1.1012966134490412, + "learning_rate": 6.8459009278183275e-06, + "loss": 0.6073, "step": 1596 }, { - "epoch": 0.31, - "learning_rate": 1.6231417088034585e-05, - "loss": 0.9399, + "epoch": 0.6144671027318199, + "grad_norm": 0.9591024913370445, + "learning_rate": 6.834077788073268e-06, + "loss": 0.6098, "step": 1597 }, { - "epoch": 0.31, - "learning_rate": 1.622654171094904e-05, - "loss": 0.7645, + "epoch": 0.6148518661023471, + "grad_norm": 1.1372973101368966, + "learning_rate": 6.822259564807768e-06, + "loss": 0.5818, "step": 1598 }, { - "epoch": 0.31, - "learning_rate": 1.6221663915546437e-05, - "loss": 0.9689, + "epoch": 0.6152366294728742, + "grad_norm": 0.9534736073044556, + "learning_rate": 6.810446276374789e-06, + "loss": 0.547, "step": 1599 }, { - "epoch": 0.31, - "learning_rate": 1.6216783703721265e-05, - "loss": 0.9466, + "epoch": 0.6156213928434013, + "grad_norm": 0.9921737105640839, + "learning_rate": 6.7986379411196255e-06, + "loss": 0.6316, "step": 1600 }, { - "epoch": 0.31, - "learning_rate": 1.6211901077368937e-05, - "loss": 0.9041, + "epoch": 0.6160061562139284, + "grad_norm": 1.056425010812889, + "learning_rate": 6.786834577379893e-06, + "loss": 0.6131, "step": 1601 }, { - "epoch": 0.31, - "learning_rate": 1.620701603838581e-05, - "loss": 0.9281, + "epoch": 0.6163909195844556, + "grad_norm": 1.0003659153636808, + "learning_rate": 6.775036203485472e-06, + "loss": 0.5836, "step": 1602 }, { - "epoch": 0.31, - "learning_rate": 1.6202128588669177e-05, - "loss": 0.8996, + "epoch": 0.6167756829549826, + "grad_norm": 0.9727181406153675, + "learning_rate": 6.763242837758504e-06, + "loss": 0.6444, "step": 1603 }, { - "epoch": 0.31, - "learning_rate": 1.619723873011727e-05, - "loss": 0.9338, + "epoch": 0.6171604463255098, + "grad_norm": 1.0280570957956219, + "learning_rate": 6.751454498513349e-06, + "loss": 0.5426, "step": 1604 }, { - "epoch": 0.31, - "learning_rate": 1.6192346464629247e-05, - "loss": 0.8825, + "epoch": 0.617545209696037, + "grad_norm": 1.0074566277592423, + "learning_rate": 6.7396712040565625e-06, + "loss": 0.5796, "step": 1605 }, { - "epoch": 0.31, - "learning_rate": 1.6187451794105212e-05, - "loss": 0.8565, + "epoch": 0.617929973066564, + "grad_norm": 0.6863283797500312, + "learning_rate": 6.727892972686861e-06, + "loss": 0.6449, "step": 1606 }, { - "epoch": 0.31, - "learning_rate": 1.61825547204462e-05, - "loss": 0.9619, + "epoch": 0.6183147364370912, + "grad_norm": 0.986894590366973, + "learning_rate": 6.716119822695111e-06, + "loss": 0.5773, "step": 1607 }, { - "epoch": 0.31, - "learning_rate": 1.6177655245554177e-05, - "loss": 0.8697, + "epoch": 0.6186994998076183, + "grad_norm": 0.8957956195496143, + "learning_rate": 6.704351772364274e-06, + "loss": 0.5974, "step": 1608 }, { - "epoch": 0.31, - "learning_rate": 1.617275337133204e-05, - "loss": 0.8984, + "epoch": 0.6190842631781455, + "grad_norm": 0.9820664609477952, + "learning_rate": 6.692588839969397e-06, + "loss": 0.5387, "step": 1609 }, { - "epoch": 0.31, - "learning_rate": 1.6167849099683623e-05, - "loss": 0.9619, + "epoch": 0.6194690265486725, + "grad_norm": 0.978944837894938, + "learning_rate": 6.680831043777579e-06, + "loss": 0.5629, "step": 1610 }, { - "epoch": 0.31, - "learning_rate": 1.6162942432513687e-05, - "loss": 0.9986, + "epoch": 0.6198537899191997, + "grad_norm": 0.9434175477976118, + "learning_rate": 6.6690784020479484e-06, + "loss": 0.5511, "step": 1611 }, { - "epoch": 0.31, - "learning_rate": 1.6158033371727924e-05, - "loss": 0.9848, + "epoch": 0.6202385532897268, + "grad_norm": 0.6243343258157564, + "learning_rate": 6.657330933031619e-06, + "loss": 0.6512, "step": 1612 }, { - "epoch": 0.31, - "learning_rate": 1.6153121919232962e-05, - "loss": 0.9337, + "epoch": 0.6206233166602539, + "grad_norm": 1.005685965452757, + "learning_rate": 6.645588654971677e-06, + "loss": 0.5544, "step": 1613 }, { - "epoch": 0.31, - "learning_rate": 1.614820807693635e-05, - "loss": 1.0114, + "epoch": 0.6210080800307811, + "grad_norm": 0.9884386448279279, + "learning_rate": 6.633851586103153e-06, + "loss": 0.6014, "step": 1614 }, { - "epoch": 0.31, - "learning_rate": 1.6143291846746563e-05, - "loss": 0.9315, + "epoch": 0.6213928434013082, + "grad_norm": 1.0283391816720802, + "learning_rate": 6.622119744652977e-06, + "loss": 0.5922, "step": 1615 }, { - "epoch": 0.31, - "learning_rate": 1.613837323057301e-05, - "loss": 1.0941, + "epoch": 0.6217776067718354, + "grad_norm": 1.0671812339490976, + "learning_rate": 6.610393148839964e-06, + "loss": 0.6419, "step": 1616 }, { - "epoch": 0.31, - "learning_rate": 1.6133452230326035e-05, - "loss": 0.9613, + "epoch": 0.6221623701423624, + "grad_norm": 0.9956875698673819, + "learning_rate": 6.598671816874794e-06, + "loss": 0.619, "step": 1617 }, { - "epoch": 0.31, - "learning_rate": 1.6128528847916883e-05, - "loss": 0.9845, + "epoch": 0.6225471335128896, + "grad_norm": 1.0591720870453252, + "learning_rate": 6.586955766959958e-06, + "loss": 0.5963, "step": 1618 }, { - "epoch": 0.31, - "learning_rate": 1.6123603085257746e-05, - "loss": 0.9202, + "epoch": 0.6229318968834167, + "grad_norm": 0.9951610424530365, + "learning_rate": 6.5752450172897466e-06, + "loss": 0.5336, "step": 1619 }, { - "epoch": 0.31, - "learning_rate": 1.6118674944261732e-05, - "loss": 0.8768, + "epoch": 0.6233166602539438, + "grad_norm": 1.0063030104798631, + "learning_rate": 6.563539586050233e-06, + "loss": 0.5513, "step": 1620 }, { - "epoch": 0.31, - "learning_rate": 1.6113744426842882e-05, - "loss": 0.8557, + "epoch": 0.6237014236244709, + "grad_norm": 1.0156767526583677, + "learning_rate": 6.551839491419213e-06, + "loss": 0.5669, "step": 1621 }, { - "epoch": 0.31, - "learning_rate": 1.6108811534916137e-05, - "loss": 0.9288, + "epoch": 0.6240861869949981, + "grad_norm": 0.9335336923742056, + "learning_rate": 6.5401447515662065e-06, + "loss": 0.5699, "step": 1622 }, { - "epoch": 0.31, - "learning_rate": 1.6103876270397387e-05, - "loss": 0.9772, + "epoch": 0.6244709503655252, + "grad_norm": 0.6222880846922205, + "learning_rate": 6.52845538465241e-06, + "loss": 0.6608, "step": 1623 }, { - "epoch": 0.31, - "learning_rate": 1.609893863520343e-05, - "loss": 0.9252, + "epoch": 0.6248557137360523, + "grad_norm": 1.011359133485383, + "learning_rate": 6.5167714088306865e-06, + "loss": 0.5578, "step": 1624 }, { - "epoch": 0.31, - "learning_rate": 1.609399863125198e-05, - "loss": 0.9777, + "epoch": 0.6252404771065795, + "grad_norm": 0.9234156543445322, + "learning_rate": 6.505092842245519e-06, + "loss": 0.5238, "step": 1625 }, { - "epoch": 0.31, - "learning_rate": 1.6089056260461687e-05, - "loss": 0.9527, + "epoch": 0.6256252404771065, + "grad_norm": 0.9895140845354052, + "learning_rate": 6.493419703032991e-06, + "loss": 0.5836, "step": 1626 }, { - "epoch": 0.31, - "learning_rate": 1.6084111524752107e-05, - "loss": 0.9034, + "epoch": 0.6260100038476337, + "grad_norm": 0.8910154394118374, + "learning_rate": 6.481752009320761e-06, + "loss": 0.5433, "step": 1627 }, { - "epoch": 0.31, - "learning_rate": 1.607916442604372e-05, - "loss": 0.9193, + "epoch": 0.6263947672181608, + "grad_norm": 0.9291738334794382, + "learning_rate": 6.4700897792280285e-06, + "loss": 0.5927, "step": 1628 }, { - "epoch": 0.31, - "learning_rate": 1.6074214966257914e-05, - "loss": 0.9479, + "epoch": 0.626779530588688, + "grad_norm": 0.9424683964110514, + "learning_rate": 6.458433030865503e-06, + "loss": 0.5654, "step": 1629 }, { - "epoch": 0.31, - "learning_rate": 1.6069263147317015e-05, - "loss": 0.9272, + "epoch": 0.6271642939592151, + "grad_norm": 0.9195993392779599, + "learning_rate": 6.4467817823354005e-06, + "loss": 0.573, "step": 1630 }, { - "epoch": 0.31, - "learning_rate": 1.6064308971144236e-05, - "loss": 0.956, + "epoch": 0.6275490573297422, + "grad_norm": 0.9452126383820607, + "learning_rate": 6.43513605173137e-06, + "loss": 0.5791, "step": 1631 }, { - "epoch": 0.31, - "learning_rate": 1.605935243966374e-05, - "loss": 0.901, + "epoch": 0.6279338207002694, + "grad_norm": 0.9828611997109997, + "learning_rate": 6.4234958571385095e-06, + "loss": 0.5999, "step": 1632 }, { - "epoch": 0.31, - "learning_rate": 1.6054393554800574e-05, - "loss": 0.8951, + "epoch": 0.6283185840707964, + "grad_norm": 0.9149678870997495, + "learning_rate": 6.4118612166333124e-06, + "loss": 0.5418, "step": 1633 }, { - "epoch": 0.31, - "learning_rate": 1.604943231848072e-05, - "loss": 0.8903, + "epoch": 0.6287033474413236, + "grad_norm": 0.9167474143467723, + "learning_rate": 6.400232148283651e-06, + "loss": 0.5692, "step": 1634 }, { - "epoch": 0.31, - "learning_rate": 1.604446873263106e-05, - "loss": 1.009, + "epoch": 0.6290881108118507, + "grad_norm": 0.8613248918176609, + "learning_rate": 6.388608670148741e-06, + "loss": 0.4908, "step": 1635 }, { - "epoch": 0.31, - "learning_rate": 1.6039502799179394e-05, - "loss": 0.9126, + "epoch": 0.6294728741823779, + "grad_norm": 1.0706933149441489, + "learning_rate": 6.376990800279119e-06, + "loss": 0.6019, "step": 1636 }, { - "epoch": 0.31, - "learning_rate": 1.6034534520054435e-05, - "loss": 0.9896, + "epoch": 0.6298576375529049, + "grad_norm": 0.9482412719753828, + "learning_rate": 6.3653785567166125e-06, + "loss": 0.6202, "step": 1637 }, { - "epoch": 0.32, - "learning_rate": 1.60295638971858e-05, - "loss": 0.9825, + "epoch": 0.6302424009234321, + "grad_norm": 1.1360158127392768, + "learning_rate": 6.3537719574943105e-06, + "loss": 0.57, "step": 1638 }, { - "epoch": 0.32, - "learning_rate": 1.602459093250403e-05, - "loss": 0.9722, + "epoch": 0.6306271642939593, + "grad_norm": 0.9532999682920527, + "learning_rate": 6.342171020636533e-06, + "loss": 0.5667, "step": 1639 }, { - "epoch": 0.32, - "learning_rate": 1.601961562794056e-05, - "loss": 0.9523, + "epoch": 0.6310119276644863, + "grad_norm": 1.0493323490127728, + "learning_rate": 6.330575764158819e-06, + "loss": 0.5558, "step": 1640 }, { - "epoch": 0.32, - "learning_rate": 1.601463798542775e-05, - "loss": 0.6696, + "epoch": 0.6313966910350135, + "grad_norm": 1.0387963666248734, + "learning_rate": 6.318986206067872e-06, + "loss": 0.6091, "step": 1641 }, { - "epoch": 0.32, - "learning_rate": 1.6009658006898848e-05, - "loss": 0.8428, + "epoch": 0.6317814544055406, + "grad_norm": 0.9429509557631494, + "learning_rate": 6.30740236436155e-06, + "loss": 0.5588, "step": 1642 }, { - "epoch": 0.32, - "learning_rate": 1.600467569428803e-05, - "loss": 0.8462, + "epoch": 0.6321662177760677, + "grad_norm": 0.961039814551438, + "learning_rate": 6.295824257028844e-06, + "loss": 0.557, "step": 1643 }, { - "epoch": 0.32, - "learning_rate": 1.599969104953036e-05, - "loss": 0.9988, + "epoch": 0.6325509811465948, + "grad_norm": 0.9967235781602125, + "learning_rate": 6.284251902049827e-06, + "loss": 0.5954, "step": 1644 }, { - "epoch": 0.32, - "learning_rate": 1.599470407456182e-05, - "loss": 0.9119, + "epoch": 0.632935744517122, + "grad_norm": 0.9744331373305751, + "learning_rate": 6.272685317395644e-06, + "loss": 0.5944, "step": 1645 }, { - "epoch": 0.32, - "learning_rate": 1.5989714771319297e-05, - "loss": 0.8724, + "epoch": 0.633320507887649, + "grad_norm": 0.910953688968267, + "learning_rate": 6.261124521028477e-06, + "loss": 0.5757, "step": 1646 }, { - "epoch": 0.32, - "learning_rate": 1.5984723141740578e-05, - "loss": 0.9259, + "epoch": 0.6337052712581762, + "grad_norm": 0.8906222488750659, + "learning_rate": 6.249569530901525e-06, + "loss": 0.5945, "step": 1647 }, { - "epoch": 0.32, - "learning_rate": 1.597972918776435e-05, - "loss": 1.0331, + "epoch": 0.6340900346287034, + "grad_norm": 0.990601645671233, + "learning_rate": 6.238020364958964e-06, + "loss": 0.5525, "step": 1648 }, { - "epoch": 0.32, - "learning_rate": 1.5974732911330208e-05, - "loss": 0.9282, + "epoch": 0.6344747979992305, + "grad_norm": 0.9473604811938108, + "learning_rate": 6.2264770411359256e-06, + "loss": 0.5616, "step": 1649 }, { - "epoch": 0.32, - "learning_rate": 1.5969734314378654e-05, - "loss": 0.8727, + "epoch": 0.6348595613697576, + "grad_norm": 0.5981300571907495, + "learning_rate": 6.214939577358479e-06, + "loss": 0.6416, "step": 1650 }, { - "epoch": 0.32, - "learning_rate": 1.5964733398851078e-05, - "loss": 0.9266, + "epoch": 0.6352443247402847, + "grad_norm": 1.0362817279807317, + "learning_rate": 6.203407991543577e-06, + "loss": 0.6003, "step": 1651 }, { - "epoch": 0.32, - "learning_rate": 1.5959730166689783e-05, - "loss": 0.9456, + "epoch": 0.6356290881108119, + "grad_norm": 0.9599987882726416, + "learning_rate": 6.191882301599052e-06, + "loss": 0.5402, "step": 1652 }, { - "epoch": 0.32, - "learning_rate": 1.5954724619837966e-05, - "loss": 0.9418, + "epoch": 0.6360138514813389, + "grad_norm": 0.6254469667463405, + "learning_rate": 6.180362525423591e-06, + "loss": 0.6516, "step": 1653 }, { - "epoch": 0.32, - "learning_rate": 1.5949716760239722e-05, - "loss": 0.879, + "epoch": 0.6363986148518661, + "grad_norm": 0.9309031538706631, + "learning_rate": 6.168848680906678e-06, + "loss": 0.5717, "step": 1654 }, { - "epoch": 0.32, - "learning_rate": 1.5944706589840046e-05, - "loss": 0.8568, + "epoch": 0.6367833782223933, + "grad_norm": 0.9323385846708576, + "learning_rate": 6.157340785928595e-06, + "loss": 0.5636, "step": 1655 }, { - "epoch": 0.32, - "learning_rate": 1.5939694110584833e-05, - "loss": 0.9466, + "epoch": 0.6371681415929203, + "grad_norm": 1.060824948285949, + "learning_rate": 6.145838858360391e-06, + "loss": 0.6425, "step": 1656 }, { - "epoch": 0.32, - "learning_rate": 1.593467932442087e-05, - "loss": 0.9662, + "epoch": 0.6375529049634475, + "grad_norm": 0.9382887821061829, + "learning_rate": 6.134342916063838e-06, + "loss": 0.5913, "step": 1657 }, { - "epoch": 0.32, - "learning_rate": 1.5929662233295846e-05, - "loss": 0.8878, + "epoch": 0.6379376683339746, + "grad_norm": 0.9481919237336738, + "learning_rate": 6.122852976891413e-06, + "loss": 0.556, "step": 1658 }, { - "epoch": 0.32, - "learning_rate": 1.5924642839158334e-05, - "loss": 0.9851, + "epoch": 0.6383224317045018, + "grad_norm": 0.9308874743907009, + "learning_rate": 6.111369058686276e-06, + "loss": 0.5125, "step": 1659 }, { - "epoch": 0.32, - "learning_rate": 1.591962114395781e-05, - "loss": 0.8766, + "epoch": 0.6387071950750288, + "grad_norm": 1.1932311210909237, + "learning_rate": 6.099891179282242e-06, + "loss": 0.5885, "step": 1660 }, { - "epoch": 0.32, - "learning_rate": 1.5914597149644654e-05, - "loss": 0.9636, + "epoch": 0.639091958445556, + "grad_norm": 1.0937675251972827, + "learning_rate": 6.088419356503732e-06, + "loss": 0.6216, "step": 1661 }, { - "epoch": 0.32, - "learning_rate": 1.5909570858170115e-05, - "loss": 0.9307, + "epoch": 0.6394767218160831, + "grad_norm": 0.9654678151383208, + "learning_rate": 6.076953608165772e-06, + "loss": 0.5573, "step": 1662 }, { - "epoch": 0.32, - "learning_rate": 1.5904542271486346e-05, - "loss": 0.8741, + "epoch": 0.6398614851866102, + "grad_norm": 0.921449404272199, + "learning_rate": 6.065493952073961e-06, + "loss": 0.6126, "step": 1663 }, { - "epoch": 0.32, - "learning_rate": 1.5899511391546403e-05, - "loss": 0.9681, + "epoch": 0.6402462485571374, + "grad_norm": 1.0010001660299825, + "learning_rate": 6.054040406024422e-06, + "loss": 0.5995, "step": 1664 }, { - "epoch": 0.32, - "learning_rate": 1.5894478220304215e-05, - "loss": 0.8933, + "epoch": 0.6406310119276645, + "grad_norm": 0.9661016020586151, + "learning_rate": 6.042592987803796e-06, + "loss": 0.5545, "step": 1665 }, { - "epoch": 0.32, - "learning_rate": 1.5889442759714603e-05, - "loss": 0.7407, + "epoch": 0.6410157752981916, + "grad_norm": 1.0527353179629877, + "learning_rate": 6.031151715189217e-06, + "loss": 0.6251, "step": 1666 }, { - "epoch": 0.32, - "learning_rate": 1.5884405011733294e-05, - "loss": 0.9268, + "epoch": 0.6414005386687187, + "grad_norm": 0.9936226517387285, + "learning_rate": 6.019716605948261e-06, + "loss": 0.5596, "step": 1667 }, { - "epoch": 0.32, - "learning_rate": 1.587936497831688e-05, - "loss": 0.8741, + "epoch": 0.6417853020392459, + "grad_norm": 0.9421703748513947, + "learning_rate": 6.008287677838937e-06, + "loss": 0.5396, "step": 1668 }, { - "epoch": 0.32, - "learning_rate": 1.5874322661422856e-05, - "loss": 0.9273, + "epoch": 0.642170065409773, + "grad_norm": 1.005061026933938, + "learning_rate": 5.996864948609662e-06, + "loss": 0.5813, "step": 1669 }, { - "epoch": 0.32, - "learning_rate": 1.5869278063009602e-05, - "loss": 0.961, + "epoch": 0.6425548287803001, + "grad_norm": 0.9720658516316831, + "learning_rate": 5.9854484359992235e-06, + "loss": 0.6013, "step": 1670 }, { - "epoch": 0.32, - "learning_rate": 1.586423118503638e-05, - "loss": 0.8813, + "epoch": 0.6429395921508272, + "grad_norm": 1.0335925462835287, + "learning_rate": 5.974038157736746e-06, + "loss": 0.5687, "step": 1671 }, { - "epoch": 0.32, - "learning_rate": 1.585918202946334e-05, - "loss": 0.9247, + "epoch": 0.6433243555213544, + "grad_norm": 0.9033268914344889, + "learning_rate": 5.962634131541676e-06, + "loss": 0.5615, "step": 1672 }, { - "epoch": 0.32, - "learning_rate": 1.5854130598251514e-05, - "loss": 0.9926, + "epoch": 0.6437091188918815, + "grad_norm": 0.9320964979981102, + "learning_rate": 5.951236375123768e-06, + "loss": 0.5643, "step": 1673 }, { - "epoch": 0.32, - "learning_rate": 1.5849076893362822e-05, - "loss": 0.9169, + "epoch": 0.6440938822624086, + "grad_norm": 0.919538264256571, + "learning_rate": 5.939844906183016e-06, + "loss": 0.5841, "step": 1674 }, { - "epoch": 0.32, - "learning_rate": 1.584402091676006e-05, - "loss": 0.9148, + "epoch": 0.6444786456329358, + "grad_norm": 0.9222992168521441, + "learning_rate": 5.92845974240966e-06, + "loss": 0.5481, "step": 1675 }, { - "epoch": 0.32, - "learning_rate": 1.5838962670406918e-05, - "loss": 0.9623, + "epoch": 0.6448634090034628, + "grad_norm": 1.0474778136751706, + "learning_rate": 5.917080901484156e-06, + "loss": 0.6151, "step": 1676 }, { - "epoch": 0.32, - "learning_rate": 1.5833902156267956e-05, - "loss": 0.967, + "epoch": 0.64524817237399, + "grad_norm": 1.0054465827836325, + "learning_rate": 5.905708401077128e-06, + "loss": 0.5661, "step": 1677 }, { - "epoch": 0.32, - "learning_rate": 1.582883937630862e-05, - "loss": 0.9757, + "epoch": 0.6456329357445171, + "grad_norm": 1.0291711352944757, + "learning_rate": 5.894342258849355e-06, + "loss": 0.5771, "step": 1678 }, { - "epoch": 0.32, - "learning_rate": 1.5823774332495236e-05, - "loss": 0.9339, + "epoch": 0.6460176991150443, + "grad_norm": 1.0222601418024373, + "learning_rate": 5.882982492451757e-06, + "loss": 0.6113, "step": 1679 }, { - "epoch": 0.32, - "learning_rate": 1.581870702679501e-05, - "loss": 0.9376, + "epoch": 0.6464024624855714, + "grad_norm": 0.9700222603296181, + "learning_rate": 5.871629119525335e-06, + "loss": 0.5688, "step": 1680 }, { - "epoch": 0.32, - "learning_rate": 1.581363746117602e-05, - "loss": 0.8653, + "epoch": 0.6467872258560985, + "grad_norm": 0.9543154449612151, + "learning_rate": 5.860282157701167e-06, + "loss": 0.5674, "step": 1681 }, { - "epoch": 0.32, - "learning_rate": 1.580856563760724e-05, - "loss": 0.7237, + "epoch": 0.6471719892266257, + "grad_norm": 0.9285557515550186, + "learning_rate": 5.8489416246003814e-06, + "loss": 0.5542, "step": 1682 }, { - "epoch": 0.32, - "learning_rate": 1.5803491558058486e-05, - "loss": 0.9479, + "epoch": 0.6475567525971527, + "grad_norm": 0.9183053205688118, + "learning_rate": 5.8376075378341194e-06, + "loss": 0.5278, "step": 1683 }, { - "epoch": 0.32, - "learning_rate": 1.579841522450049e-05, - "loss": 0.86, + "epoch": 0.6479415159676799, + "grad_norm": 1.1138294781340996, + "learning_rate": 5.826279915003503e-06, + "loss": 0.5323, "step": 1684 }, { - "epoch": 0.32, - "learning_rate": 1.5793336638904838e-05, - "loss": 0.9742, + "epoch": 0.648326279338207, + "grad_norm": 0.9888409879203782, + "learning_rate": 5.814958773699625e-06, + "loss": 0.5972, "step": 1685 }, { - "epoch": 0.32, - "learning_rate": 1.578825580324399e-05, - "loss": 0.9244, + "epoch": 0.6487110427087341, + "grad_norm": 1.0323524333389873, + "learning_rate": 5.803644131503516e-06, + "loss": 0.6155, "step": 1686 }, { - "epoch": 0.32, - "learning_rate": 1.5783172719491288e-05, - "loss": 0.9085, + "epoch": 0.6490958060792612, + "grad_norm": 1.056575797225963, + "learning_rate": 5.792336005986105e-06, + "loss": 0.5635, "step": 1687 }, { - "epoch": 0.32, - "learning_rate": 1.577808738962094e-05, - "loss": 0.8967, + "epoch": 0.6494805694497884, + "grad_norm": 0.9458935393162997, + "learning_rate": 5.781034414708208e-06, + "loss": 0.5967, "step": 1688 }, { - "epoch": 0.32, - "learning_rate": 1.577299981560803e-05, - "loss": 0.7093, + "epoch": 0.6498653328203156, + "grad_norm": 0.9073530408108605, + "learning_rate": 5.769739375220489e-06, + "loss": 0.5048, "step": 1689 }, { - "epoch": 0.33, - "learning_rate": 1.5767909999428513e-05, - "loss": 0.9929, + "epoch": 0.6502500961908426, + "grad_norm": 1.0156318653552674, + "learning_rate": 5.7584509050634395e-06, + "loss": 0.5733, "step": 1690 }, { - "epoch": 0.33, - "learning_rate": 1.576281794305922e-05, - "loss": 0.9723, + "epoch": 0.6506348595613698, + "grad_norm": 1.0638774318774618, + "learning_rate": 5.747169021767342e-06, + "loss": 0.5338, "step": 1691 }, { - "epoch": 0.33, - "learning_rate": 1.575772364847784e-05, - "loss": 0.9529, + "epoch": 0.6510196229318969, + "grad_norm": 0.8475691025778729, + "learning_rate": 5.73589374285227e-06, + "loss": 0.5631, "step": 1692 }, { - "epoch": 0.33, - "learning_rate": 1.575262711766294e-05, - "loss": 0.9594, + "epoch": 0.651404386302424, + "grad_norm": 1.0049930860846128, + "learning_rate": 5.724625085828022e-06, + "loss": 0.5753, "step": 1693 }, { - "epoch": 0.33, - "learning_rate": 1.5747528352593956e-05, - "loss": 0.9032, + "epoch": 0.6517891496729511, + "grad_norm": 1.0014637603254426, + "learning_rate": 5.713363068194115e-06, + "loss": 0.5971, "step": 1694 }, { - "epoch": 0.33, - "learning_rate": 1.574242735525119e-05, - "loss": 0.9369, + "epoch": 0.6521739130434783, + "grad_norm": 1.0066591933029823, + "learning_rate": 5.702107707439766e-06, + "loss": 0.5689, "step": 1695 }, { - "epoch": 0.33, - "learning_rate": 1.5737324127615808e-05, - "loss": 0.9864, + "epoch": 0.6525586764140053, + "grad_norm": 0.9231227778271333, + "learning_rate": 5.690859021043842e-06, + "loss": 0.5738, "step": 1696 }, { - "epoch": 0.33, - "learning_rate": 1.5732218671669847e-05, - "loss": 0.9966, + "epoch": 0.6529434397845325, + "grad_norm": 0.8880209537075676, + "learning_rate": 5.679617026474853e-06, + "loss": 0.5081, "step": 1697 }, { - "epoch": 0.33, - "learning_rate": 1.5727110989396205e-05, - "loss": 0.8634, + "epoch": 0.6533282031550597, + "grad_norm": 0.878651006452691, + "learning_rate": 5.6683817411909114e-06, + "loss": 0.4938, "step": 1698 }, { - "epoch": 0.33, - "learning_rate": 1.5722001082778645e-05, - "loss": 1.0306, + "epoch": 0.6537129665255867, + "grad_norm": 0.6476112144942279, + "learning_rate": 5.65715318263972e-06, + "loss": 0.6295, "step": 1699 }, { - "epoch": 0.33, - "learning_rate": 1.5716888953801805e-05, - "loss": 0.9485, + "epoch": 0.6540977298961139, + "grad_norm": 1.0134689615049821, + "learning_rate": 5.645931368258527e-06, + "loss": 0.5949, "step": 1700 }, { - "epoch": 0.33, - "learning_rate": 1.5711774604451168e-05, - "loss": 0.8761, + "epoch": 0.654482493266641, + "grad_norm": 0.8957527881176658, + "learning_rate": 5.634716315474109e-06, + "loss": 0.5773, "step": 1701 }, { - "epoch": 0.33, - "learning_rate": 1.5706658036713093e-05, - "loss": 0.7545, + "epoch": 0.6548672566371682, + "grad_norm": 1.008204786042917, + "learning_rate": 5.623508041702743e-06, + "loss": 0.5968, "step": 1702 }, { - "epoch": 0.33, - "learning_rate": 1.5701539252574795e-05, - "loss": 0.7129, + "epoch": 0.6552520200076952, + "grad_norm": 0.9263487813413357, + "learning_rate": 5.612306564350179e-06, + "loss": 0.5502, "step": 1703 }, { - "epoch": 0.33, - "learning_rate": 1.5696418254024344e-05, - "loss": 0.893, + "epoch": 0.6556367833782224, + "grad_norm": 1.0241395010676588, + "learning_rate": 5.601111900811607e-06, + "loss": 0.5543, "step": 1704 }, { - "epoch": 0.33, - "learning_rate": 1.569129504305069e-05, - "loss": 0.944, + "epoch": 0.6560215467487496, + "grad_norm": 0.9745053760782287, + "learning_rate": 5.589924068471648e-06, + "loss": 0.5406, "step": 1705 }, { - "epoch": 0.33, - "learning_rate": 1.568616962164362e-05, - "loss": 0.9125, + "epoch": 0.6564063101192766, + "grad_norm": 0.6518912661604789, + "learning_rate": 5.578743084704306e-06, + "loss": 0.6595, "step": 1706 }, { - "epoch": 0.33, - "learning_rate": 1.5681041991793788e-05, - "loss": 0.9183, + "epoch": 0.6567910734898038, + "grad_norm": 0.9664044241679586, + "learning_rate": 5.567568966872947e-06, + "loss": 0.528, "step": 1707 }, { - "epoch": 0.33, - "learning_rate": 1.567591215549271e-05, - "loss": 0.9238, + "epoch": 0.6571758368603309, + "grad_norm": 0.9439666919681425, + "learning_rate": 5.556401732330281e-06, + "loss": 0.566, "step": 1708 }, { - "epoch": 0.33, - "learning_rate": 1.567078011473276e-05, - "loss": 0.9154, + "epoch": 0.657560600230858, + "grad_norm": 0.8818707410819688, + "learning_rate": 5.545241398418326e-06, + "loss": 0.5237, "step": 1709 }, { - "epoch": 0.33, - "learning_rate": 1.5665645871507152e-05, - "loss": 0.8328, + "epoch": 0.6579453636013851, + "grad_norm": 0.9257657195377806, + "learning_rate": 5.534087982468384e-06, + "loss": 0.5526, "step": 1710 }, { - "epoch": 0.33, - "learning_rate": 1.5660509427809973e-05, - "loss": 0.8712, + "epoch": 0.6583301269719123, + "grad_norm": 1.0050326401244396, + "learning_rate": 5.522941501801008e-06, + "loss": 0.5625, "step": 1711 }, { - "epoch": 0.33, - "learning_rate": 1.565537078563616e-05, - "loss": 0.8107, + "epoch": 0.6587148903424394, + "grad_norm": 0.987340383182925, + "learning_rate": 5.511801973725997e-06, + "loss": 0.5543, "step": 1712 }, { - "epoch": 0.33, - "learning_rate": 1.56502299469815e-05, - "loss": 0.9584, + "epoch": 0.6590996537129665, + "grad_norm": 1.009700013374671, + "learning_rate": 5.500669415542336e-06, + "loss": 0.5668, "step": 1713 }, { - "epoch": 0.33, - "learning_rate": 1.564508691384264e-05, - "loss": 0.9835, + "epoch": 0.6594844170834937, + "grad_norm": 0.9116523114025201, + "learning_rate": 5.4895438445381945e-06, + "loss": 0.5802, "step": 1714 }, { - "epoch": 0.33, - "learning_rate": 1.5639941688217063e-05, - "loss": 0.9501, + "epoch": 0.6598691804540208, + "grad_norm": 0.9177466238220682, + "learning_rate": 5.4784252779908905e-06, + "loss": 0.5748, "step": 1715 }, { - "epoch": 0.33, - "learning_rate": 1.5634794272103126e-05, - "loss": 0.9001, + "epoch": 0.6602539438245479, + "grad_norm": 1.0497358384584967, + "learning_rate": 5.467313733166863e-06, + "loss": 0.6404, "step": 1716 }, { - "epoch": 0.33, - "learning_rate": 1.562964466750003e-05, - "loss": 0.9048, + "epoch": 0.660638707195075, + "grad_norm": 0.9968803336759691, + "learning_rate": 5.456209227321643e-06, + "loss": 0.548, "step": 1717 }, { - "epoch": 0.33, - "learning_rate": 1.562449287640781e-05, - "loss": 0.9482, + "epoch": 0.6610234705656022, + "grad_norm": 1.046449892457246, + "learning_rate": 5.445111777699842e-06, + "loss": 0.5886, "step": 1718 }, { - "epoch": 0.33, - "learning_rate": 1.5619338900827368e-05, - "loss": 0.9183, + "epoch": 0.6614082339361292, + "grad_norm": 0.9987882122901328, + "learning_rate": 5.434021401535105e-06, + "loss": 0.5751, "step": 1719 }, { - "epoch": 0.33, - "learning_rate": 1.5614182742760448e-05, - "loss": 0.9791, + "epoch": 0.6617929973066564, + "grad_norm": 1.0395089117245613, + "learning_rate": 5.422938116050092e-06, + "loss": 0.5597, "step": 1720 }, { - "epoch": 0.33, - "learning_rate": 1.5609024404209643e-05, - "loss": 0.9504, + "epoch": 0.6621777606771835, + "grad_norm": 0.9274789946615977, + "learning_rate": 5.411861938456453e-06, + "loss": 0.5856, "step": 1721 }, { - "epoch": 0.33, - "learning_rate": 1.5603863887178393e-05, - "loss": 0.9285, + "epoch": 0.6625625240477107, + "grad_norm": 0.8695762150215398, + "learning_rate": 5.400792885954802e-06, + "loss": 0.4891, "step": 1722 }, { - "epoch": 0.33, - "learning_rate": 1.5598701193670983e-05, - "loss": 0.8917, + "epoch": 0.6629472874182378, + "grad_norm": 0.9412819091059229, + "learning_rate": 5.389730975734686e-06, + "loss": 0.5556, "step": 1723 }, { - "epoch": 0.33, - "learning_rate": 1.559353632569254e-05, - "loss": 0.9147, + "epoch": 0.6633320507887649, + "grad_norm": 1.0013500564251088, + "learning_rate": 5.378676224974557e-06, + "loss": 0.6141, "step": 1724 }, { - "epoch": 0.33, - "learning_rate": 1.5588369285249048e-05, - "loss": 0.869, + "epoch": 0.6637168141592921, + "grad_norm": 0.9865839160737121, + "learning_rate": 5.367628650841761e-06, + "loss": 0.5707, "step": 1725 }, { - "epoch": 0.33, - "learning_rate": 1.5583200074347318e-05, - "loss": 0.9561, + "epoch": 0.6641015775298191, + "grad_norm": 1.0572622898027162, + "learning_rate": 5.356588270492487e-06, + "loss": 0.5552, "step": 1726 }, { - "epoch": 0.33, - "learning_rate": 1.557802869499501e-05, - "loss": 0.9049, + "epoch": 0.6644863409003463, + "grad_norm": 1.024834424528934, + "learning_rate": 5.3455551010717545e-06, + "loss": 0.5719, "step": 1727 }, { - "epoch": 0.33, - "learning_rate": 1.5572855149200637e-05, - "loss": 0.9226, + "epoch": 0.6648711042708734, + "grad_norm": 0.9651089768484155, + "learning_rate": 5.334529159713389e-06, + "loss": 0.5976, "step": 1728 }, { - "epoch": 0.33, - "learning_rate": 1.5567679438973543e-05, - "loss": 0.9719, + "epoch": 0.6652558676414005, + "grad_norm": 0.9415963764078287, + "learning_rate": 5.323510463539989e-06, + "loss": 0.6006, "step": 1729 }, { - "epoch": 0.33, - "learning_rate": 1.5562501566323906e-05, - "loss": 0.9417, + "epoch": 0.6656406310119277, + "grad_norm": 0.9743974649758395, + "learning_rate": 5.3124990296628974e-06, + "loss": 0.5731, "step": 1730 }, { - "epoch": 0.33, - "learning_rate": 1.555732153326276e-05, - "loss": 0.9914, + "epoch": 0.6660253943824548, + "grad_norm": 0.8756181569754731, + "learning_rate": 5.301494875182192e-06, + "loss": 0.506, "step": 1731 }, { - "epoch": 0.33, - "learning_rate": 1.5552139341801965e-05, - "loss": 0.8983, + "epoch": 0.666410157752982, + "grad_norm": 0.9736152690737813, + "learning_rate": 5.290498017186631e-06, + "loss": 0.5748, "step": 1732 }, { - "epoch": 0.33, - "learning_rate": 1.554695499395423e-05, - "loss": 0.8956, + "epoch": 0.666794921123509, + "grad_norm": 0.9644355554712004, + "learning_rate": 5.279508472753654e-06, + "loss": 0.5629, "step": 1733 }, { - "epoch": 0.33, - "learning_rate": 1.5541768491733092e-05, - "loss": 0.8969, + "epoch": 0.6671796844940362, + "grad_norm": 0.6028692379886013, + "learning_rate": 5.2685262589493314e-06, + "loss": 0.6461, "step": 1734 }, { - "epoch": 0.33, - "learning_rate": 1.5536579837152927e-05, - "loss": 0.859, + "epoch": 0.6675644478645633, + "grad_norm": 1.0331061725361887, + "learning_rate": 5.257551392828359e-06, + "loss": 0.5881, "step": 1735 }, { - "epoch": 0.33, - "learning_rate": 1.5531389032228955e-05, - "loss": 0.934, + "epoch": 0.6679492112350904, + "grad_norm": 1.0633034282383573, + "learning_rate": 5.246583891434018e-06, + "loss": 0.5952, "step": 1736 }, { - "epoch": 0.33, - "learning_rate": 1.552619607897722e-05, - "loss": 0.9279, + "epoch": 0.6683339746056175, + "grad_norm": 1.0206099515339062, + "learning_rate": 5.235623771798151e-06, + "loss": 0.6365, "step": 1737 }, { - "epoch": 0.33, - "learning_rate": 1.55210009794146e-05, - "loss": 0.8974, + "epoch": 0.6687187379761447, + "grad_norm": 0.9764091808060633, + "learning_rate": 5.224671050941146e-06, + "loss": 0.5729, "step": 1738 }, { - "epoch": 0.33, - "learning_rate": 1.5515803735558827e-05, - "loss": 0.9485, + "epoch": 0.6691035013466718, + "grad_norm": 0.5712096222398239, + "learning_rate": 5.213725745871889e-06, + "loss": 0.625, "step": 1739 }, { - "epoch": 0.33, - "learning_rate": 1.5510604349428438e-05, - "loss": 0.9866, + "epoch": 0.6694882647171989, + "grad_norm": 0.9852085812472796, + "learning_rate": 5.20278787358776e-06, + "loss": 0.6069, "step": 1740 }, { - "epoch": 0.33, - "learning_rate": 1.550540282304282e-05, - "loss": 0.9172, + "epoch": 0.6698730280877261, + "grad_norm": 0.9444544438146338, + "learning_rate": 5.1918574510745865e-06, + "loss": 0.5453, "step": 1741 }, { - "epoch": 0.34, - "learning_rate": 1.550019915842218e-05, - "loss": 0.9429, + "epoch": 0.6702577914582531, + "grad_norm": 0.9817945211202368, + "learning_rate": 5.180934495306638e-06, + "loss": 0.546, "step": 1742 }, { - "epoch": 0.34, - "learning_rate": 1.549499335758757e-05, - "loss": 0.9033, + "epoch": 0.6706425548287803, + "grad_norm": 1.0126086266003596, + "learning_rate": 5.170019023246574e-06, + "loss": 0.5713, "step": 1743 }, { - "epoch": 0.34, - "learning_rate": 1.548978542256086e-05, - "loss": 0.9531, + "epoch": 0.6710273181993074, + "grad_norm": 0.6119431505214815, + "learning_rate": 5.159111051845451e-06, + "loss": 0.6519, "step": 1744 }, { - "epoch": 0.34, - "learning_rate": 1.5484575355364744e-05, - "loss": 0.9505, + "epoch": 0.6714120815698346, + "grad_norm": 0.9419160431804533, + "learning_rate": 5.148210598042665e-06, + "loss": 0.5819, "step": 1745 }, { - "epoch": 0.34, - "learning_rate": 1.5479363158022763e-05, - "loss": 0.9298, + "epoch": 0.6717968449403616, + "grad_norm": 0.9513486713305566, + "learning_rate": 5.137317678765939e-06, + "loss": 0.5606, "step": 1746 }, { - "epoch": 0.34, - "learning_rate": 1.547414883255927e-05, - "loss": 0.9093, + "epoch": 0.6721816083108888, + "grad_norm": 1.0008742575616594, + "learning_rate": 5.126432310931295e-06, + "loss": 0.5607, "step": 1747 }, { - "epoch": 0.34, - "learning_rate": 1.546893238099945e-05, - "loss": 0.8661, + "epoch": 0.672566371681416, + "grad_norm": 0.576939205419165, + "learning_rate": 5.115554511443033e-06, + "loss": 0.6499, "step": 1748 }, { - "epoch": 0.34, - "learning_rate": 1.5463713805369312e-05, - "loss": 0.8666, + "epoch": 0.672951135051943, + "grad_norm": 0.9849558253272185, + "learning_rate": 5.104684297193694e-06, + "loss": 0.6017, "step": 1749 }, { - "epoch": 0.34, - "learning_rate": 1.5458493107695688e-05, - "loss": 0.9274, + "epoch": 0.6733358984224702, + "grad_norm": 0.9762564881694822, + "learning_rate": 5.09382168506404e-06, + "loss": 0.572, "step": 1750 }, { - "epoch": 0.34, - "learning_rate": 1.5453270290006237e-05, - "loss": 0.8594, + "epoch": 0.6737206617929973, + "grad_norm": 1.0391475874868656, + "learning_rate": 5.082966691923037e-06, + "loss": 0.6028, "step": 1751 }, { - "epoch": 0.34, - "learning_rate": 1.544804535432945e-05, - "loss": 0.9529, + "epoch": 0.6741054251635245, + "grad_norm": 1.0035559010481825, + "learning_rate": 5.0721193346278066e-06, + "loss": 0.5827, "step": 1752 }, { - "epoch": 0.34, - "learning_rate": 1.544281830269462e-05, - "loss": 0.884, + "epoch": 0.6744901885340515, + "grad_norm": 0.9409611426512595, + "learning_rate": 5.061279630023618e-06, + "loss": 0.5742, "step": 1753 }, { - "epoch": 0.34, - "learning_rate": 1.5437589137131882e-05, - "loss": 0.7327, + "epoch": 0.6748749519045787, + "grad_norm": 0.9691697305718414, + "learning_rate": 5.050447594943856e-06, + "loss": 0.6291, "step": 1754 }, { - "epoch": 0.34, - "learning_rate": 1.5432357859672177e-05, - "loss": 0.6898, + "epoch": 0.6752597152751059, + "grad_norm": 0.9986887516951423, + "learning_rate": 5.0396232462099945e-06, + "loss": 0.5396, "step": 1755 }, { - "epoch": 0.34, - "learning_rate": 1.542712447234728e-05, - "loss": 0.9503, + "epoch": 0.6756444786456329, + "grad_norm": 0.9165090695610844, + "learning_rate": 5.028806600631569e-06, + "loss": 0.6108, "step": 1756 }, { - "epoch": 0.34, - "learning_rate": 1.542188897718977e-05, - "loss": 0.9392, + "epoch": 0.6760292420161601, + "grad_norm": 0.9974037470396901, + "learning_rate": 5.017997675006161e-06, + "loss": 0.6023, "step": 1757 }, { - "epoch": 0.34, - "learning_rate": 1.5416651376233062e-05, - "loss": 0.9462, + "epoch": 0.6764140053866872, + "grad_norm": 0.8853644307551725, + "learning_rate": 5.007196486119355e-06, + "loss": 0.5409, "step": 1758 }, { - "epoch": 0.34, - "learning_rate": 1.5411411671511376e-05, - "loss": 0.7344, + "epoch": 0.6767987687572143, + "grad_norm": 1.0435995905149724, + "learning_rate": 4.996403050744719e-06, + "loss": 0.5687, "step": 1759 }, { - "epoch": 0.34, - "learning_rate": 1.5406169865059747e-05, - "loss": 0.9498, + "epoch": 0.6771835321277414, + "grad_norm": 1.0220165695978982, + "learning_rate": 4.985617385643789e-06, + "loss": 0.5988, "step": 1760 }, { - "epoch": 0.34, - "learning_rate": 1.5400925958914045e-05, - "loss": 0.9321, + "epoch": 0.6775682954982686, + "grad_norm": 0.9335091452280982, + "learning_rate": 4.974839507566027e-06, + "loss": 0.554, "step": 1761 }, { - "epoch": 0.34, - "learning_rate": 1.5395679955110927e-05, - "loss": 0.9199, + "epoch": 0.6779530588687956, + "grad_norm": 0.9169314158407507, + "learning_rate": 4.9640694332488075e-06, + "loss": 0.5778, "step": 1762 }, { - "epoch": 0.34, - "learning_rate": 1.53904318556879e-05, - "loss": 0.9385, + "epoch": 0.6783378222393228, + "grad_norm": 0.5878207045818972, + "learning_rate": 4.953307179417376e-06, + "loss": 0.5994, "step": 1763 }, { - "epoch": 0.34, - "learning_rate": 1.5385181662683244e-05, - "loss": 0.9832, + "epoch": 0.67872258560985, + "grad_norm": 1.0033401003906588, + "learning_rate": 4.94255276278485e-06, + "loss": 0.5972, "step": 1764 }, { - "epoch": 0.34, - "learning_rate": 1.5379929378136088e-05, - "loss": 0.9148, + "epoch": 0.6791073489803771, + "grad_norm": 1.0824617708033302, + "learning_rate": 4.931806200052165e-06, + "loss": 0.5439, "step": 1765 }, { - "epoch": 0.34, - "learning_rate": 1.5374675004086353e-05, - "loss": 0.8753, + "epoch": 0.6794921123509042, + "grad_norm": 1.0311152870852403, + "learning_rate": 4.92106750790806e-06, + "loss": 0.5618, "step": 1766 }, { - "epoch": 0.34, - "learning_rate": 1.5369418542574782e-05, - "loss": 0.9931, + "epoch": 0.6798768757214313, + "grad_norm": 0.9524325230493552, + "learning_rate": 4.910336703029055e-06, + "loss": 0.5696, "step": 1767 }, { - "epoch": 0.34, - "learning_rate": 1.536415999564292e-05, - "loss": 0.8943, + "epoch": 0.6802616390919585, + "grad_norm": 0.9896276378324536, + "learning_rate": 4.899613802079419e-06, + "loss": 0.5491, "step": 1768 }, { - "epoch": 0.34, - "learning_rate": 1.5358899365333123e-05, - "loss": 0.9448, + "epoch": 0.6806464024624855, + "grad_norm": 0.951680620089327, + "learning_rate": 4.888898821711144e-06, + "loss": 0.5792, "step": 1769 }, { - "epoch": 0.34, - "learning_rate": 1.5353636653688563e-05, - "loss": 0.9328, + "epoch": 0.6810311658330127, + "grad_norm": 1.0299614746855026, + "learning_rate": 4.878191778563934e-06, + "loss": 0.6165, "step": 1770 }, { - "epoch": 0.34, - "learning_rate": 1.534837186275322e-05, - "loss": 0.9832, + "epoch": 0.6814159292035398, + "grad_norm": 0.9539399889818537, + "learning_rate": 4.867492689265154e-06, + "loss": 0.5859, "step": 1771 }, { - "epoch": 0.34, - "learning_rate": 1.5343104994571877e-05, - "loss": 0.8885, + "epoch": 0.681800692574067, + "grad_norm": 1.0066794566121489, + "learning_rate": 4.856801570429822e-06, + "loss": 0.5819, "step": 1772 }, { - "epoch": 0.34, - "learning_rate": 1.533783605119012e-05, - "loss": 0.9145, + "epoch": 0.6821854559445941, + "grad_norm": 0.6055580520386267, + "learning_rate": 4.846118438660578e-06, + "loss": 0.6595, "step": 1773 }, { - "epoch": 0.34, - "learning_rate": 1.5332565034654344e-05, - "loss": 0.9366, + "epoch": 0.6825702193151212, + "grad_norm": 1.0545145281066897, + "learning_rate": 4.83544331054766e-06, + "loss": 0.5904, "step": 1774 }, { - "epoch": 0.34, - "learning_rate": 1.5327291947011763e-05, - "loss": 0.8817, + "epoch": 0.6829549826856484, + "grad_norm": 1.0142981134629696, + "learning_rate": 4.824776202668875e-06, + "loss": 0.5967, "step": 1775 }, { - "epoch": 0.34, - "learning_rate": 1.5322016790310373e-05, - "loss": 0.9423, + "epoch": 0.6833397460561754, + "grad_norm": 0.9611349774804838, + "learning_rate": 4.8141171315895694e-06, + "loss": 0.5553, "step": 1776 }, { - "epoch": 0.34, - "learning_rate": 1.5316739566598985e-05, - "loss": 0.8624, + "epoch": 0.6837245094267026, + "grad_norm": 0.9474479804189555, + "learning_rate": 4.803466113862626e-06, + "loss": 0.5518, "step": 1777 }, { - "epoch": 0.34, - "learning_rate": 1.531146027792722e-05, - "loss": 0.9014, + "epoch": 0.6841092727972297, + "grad_norm": 0.9631148973841709, + "learning_rate": 4.792823166028405e-06, + "loss": 0.6271, "step": 1778 }, { - "epoch": 0.34, - "learning_rate": 1.530617892634548e-05, - "loss": 0.9219, + "epoch": 0.6844940361677568, + "grad_norm": 0.5887387572715773, + "learning_rate": 4.7821883046147414e-06, + "loss": 0.6638, "step": 1779 }, { - "epoch": 0.34, - "learning_rate": 1.5300895513904993e-05, - "loss": 0.9093, + "epoch": 0.6848787995382839, + "grad_norm": 1.0102777542411792, + "learning_rate": 4.771561546136908e-06, + "loss": 0.6333, "step": 1780 }, { - "epoch": 0.34, - "learning_rate": 1.529561004265777e-05, - "loss": 0.7429, + "epoch": 0.6852635629088111, + "grad_norm": 1.0402147030162605, + "learning_rate": 4.760942907097601e-06, + "loss": 0.6277, "step": 1781 }, { - "epoch": 0.34, - "learning_rate": 1.5290322514656624e-05, - "loss": 0.9417, + "epoch": 0.6856483262793382, + "grad_norm": 1.0217018823451602, + "learning_rate": 4.750332403986902e-06, + "loss": 0.5933, "step": 1782 }, { - "epoch": 0.34, - "learning_rate": 1.5285032931955177e-05, - "loss": 0.8754, + "epoch": 0.6860330896498653, + "grad_norm": 0.946521317482783, + "learning_rate": 4.739730053282255e-06, + "loss": 0.5552, "step": 1783 }, { - "epoch": 0.34, - "learning_rate": 1.527974129660784e-05, - "loss": 0.9166, + "epoch": 0.6864178530203925, + "grad_norm": 0.8759187425056967, + "learning_rate": 4.7291358714484594e-06, + "loss": 0.5298, "step": 1784 }, { - "epoch": 0.34, - "learning_rate": 1.527444761066982e-05, - "loss": 0.9698, + "epoch": 0.6868026163909196, + "grad_norm": 0.8873107041808451, + "learning_rate": 4.718549874937612e-06, + "loss": 0.5056, "step": 1785 }, { - "epoch": 0.34, - "learning_rate": 1.5269151876197127e-05, - "loss": 0.8701, + "epoch": 0.6871873797614467, + "grad_norm": 1.0929629204948021, + "learning_rate": 4.707972080189106e-06, + "loss": 0.6347, "step": 1786 }, { - "epoch": 0.34, - "learning_rate": 1.5263854095246557e-05, - "loss": 1.0014, + "epoch": 0.6875721431319738, + "grad_norm": 0.9319040093113203, + "learning_rate": 4.697402503629596e-06, + "loss": 0.5878, "step": 1787 }, { - "epoch": 0.34, - "learning_rate": 1.5258554269875716e-05, - "loss": 0.933, + "epoch": 0.687956906502501, + "grad_norm": 1.0043545169033399, + "learning_rate": 4.686841161672974e-06, + "loss": 0.5968, "step": 1788 }, { - "epoch": 0.34, - "learning_rate": 1.5253252402142989e-05, - "loss": 0.9163, + "epoch": 0.6883416698730281, + "grad_norm": 1.0431711964952504, + "learning_rate": 4.67628807072034e-06, + "loss": 0.6468, "step": 1789 }, { - "epoch": 0.34, - "learning_rate": 1.5247948494107566e-05, - "loss": 0.9674, + "epoch": 0.6887264332435552, + "grad_norm": 0.9563920686523799, + "learning_rate": 4.665743247159995e-06, + "loss": 0.6009, "step": 1790 }, { - "epoch": 0.34, - "learning_rate": 1.5242642547829416e-05, - "loss": 0.968, + "epoch": 0.6891111966140824, + "grad_norm": 0.9878009378902047, + "learning_rate": 4.655206707367388e-06, + "loss": 0.5676, "step": 1791 }, { - "epoch": 0.34, - "learning_rate": 1.523733456536931e-05, - "loss": 0.8931, + "epoch": 0.6894959599846094, + "grad_norm": 0.9565213703848053, + "learning_rate": 4.644678467705101e-06, + "loss": 0.5703, "step": 1792 }, { - "epoch": 0.34, - "learning_rate": 1.5232024548788813e-05, - "loss": 1.0136, + "epoch": 0.6898807233551366, + "grad_norm": 0.8998756819482219, + "learning_rate": 4.634158544522849e-06, + "loss": 0.5085, "step": 1793 }, { - "epoch": 0.35, - "learning_rate": 1.5226712500150267e-05, - "loss": 0.8909, + "epoch": 0.6902654867256637, + "grad_norm": 0.9107669375969553, + "learning_rate": 4.623646954157399e-06, + "loss": 0.6079, "step": 1794 }, { - "epoch": 0.35, - "learning_rate": 1.5221398421516816e-05, - "loss": 0.9521, + "epoch": 0.6906502500961909, + "grad_norm": 0.9533695427522734, + "learning_rate": 4.613143712932603e-06, + "loss": 0.5275, "step": 1795 }, { - "epoch": 0.35, - "learning_rate": 1.5216082314952383e-05, - "loss": 0.8698, + "epoch": 0.6910350134667179, + "grad_norm": 0.9177806267084673, + "learning_rate": 4.602648837159333e-06, + "loss": 0.5624, "step": 1796 }, { - "epoch": 0.35, - "learning_rate": 1.521076418252168e-05, - "loss": 0.9966, + "epoch": 0.6914197768372451, + "grad_norm": 0.938418141910206, + "learning_rate": 4.592162343135483e-06, + "loss": 0.6058, "step": 1797 }, { - "epoch": 0.35, - "learning_rate": 1.5205444026290218e-05, - "loss": 0.9423, + "epoch": 0.6918045402077723, + "grad_norm": 1.0994964476995026, + "learning_rate": 4.5816842471459224e-06, + "loss": 0.5623, "step": 1798 }, { - "epoch": 0.35, - "learning_rate": 1.5200121848324276e-05, - "loss": 0.8982, + "epoch": 0.6921893035782993, + "grad_norm": 0.9774692869937253, + "learning_rate": 4.571214565462477e-06, + "loss": 0.6004, "step": 1799 }, { - "epoch": 0.35, - "learning_rate": 1.5194797650690926e-05, - "loss": 0.8829, + "epoch": 0.6925740669488265, + "grad_norm": 0.9981616467593657, + "learning_rate": 4.560753314343912e-06, + "loss": 0.6016, "step": 1800 }, { - "epoch": 0.35, - "learning_rate": 1.5189471435458032e-05, - "loss": 0.9576, + "epoch": 0.6929588303193536, + "grad_norm": 0.5844503967909246, + "learning_rate": 4.5503005100358945e-06, + "loss": 0.6089, "step": 1801 }, { - "epoch": 0.35, - "learning_rate": 1.5184143204694231e-05, - "loss": 0.845, + "epoch": 0.6933435936898807, + "grad_norm": 0.9066842354036336, + "learning_rate": 4.539856168770974e-06, + "loss": 0.5671, "step": 1802 }, { - "epoch": 0.35, - "learning_rate": 1.5178812960468945e-05, - "loss": 0.9599, + "epoch": 0.6937283570604078, + "grad_norm": 0.6571691883049603, + "learning_rate": 4.52942030676857e-06, + "loss": 0.6259, "step": 1803 }, { - "epoch": 0.35, - "learning_rate": 1.5173480704852379e-05, - "loss": 0.8374, + "epoch": 0.694113120430935, + "grad_norm": 0.8926497702666348, + "learning_rate": 4.5189929402349175e-06, + "loss": 0.5412, "step": 1804 }, { - "epoch": 0.35, - "learning_rate": 1.5168146439915525e-05, - "loss": 0.8979, + "epoch": 0.694497883801462, + "grad_norm": 0.9988619476114245, + "learning_rate": 4.508574085363065e-06, + "loss": 0.5281, "step": 1805 }, { - "epoch": 0.35, - "learning_rate": 1.5162810167730144e-05, - "loss": 0.9298, + "epoch": 0.6948826471719892, + "grad_norm": 1.007964371371848, + "learning_rate": 4.498163758332853e-06, + "loss": 0.5866, "step": 1806 }, { - "epoch": 0.35, - "learning_rate": 1.5157471890368785e-05, - "loss": 0.9063, + "epoch": 0.6952674105425164, + "grad_norm": 0.8302926709738292, + "learning_rate": 4.4877619753108605e-06, + "loss": 0.5126, "step": 1807 }, { - "epoch": 0.35, - "learning_rate": 1.5152131609904773e-05, - "loss": 0.9644, + "epoch": 0.6956521739130435, + "grad_norm": 1.0094020201238472, + "learning_rate": 4.477368752450409e-06, + "loss": 0.6095, "step": 1808 }, { - "epoch": 0.35, - "learning_rate": 1.5146789328412213e-05, - "loss": 0.9562, + "epoch": 0.6960369372835706, + "grad_norm": 0.9106976968740014, + "learning_rate": 4.466984105891521e-06, + "loss": 0.5235, "step": 1809 }, { - "epoch": 0.35, - "learning_rate": 1.5141445047965984e-05, - "loss": 0.7047, + "epoch": 0.6964217006540977, + "grad_norm": 0.9465259602527023, + "learning_rate": 4.456608051760914e-06, + "loss": 0.6074, "step": 1810 }, { - "epoch": 0.35, - "learning_rate": 1.5136098770641741e-05, - "loss": 0.8957, + "epoch": 0.6968064640246249, + "grad_norm": 0.9170607408666119, + "learning_rate": 4.446240606171945e-06, + "loss": 0.5391, "step": 1811 }, { - "epoch": 0.35, - "learning_rate": 1.513075049851592e-05, - "loss": 0.7264, + "epoch": 0.6971912273951519, + "grad_norm": 0.9534896219288522, + "learning_rate": 4.4358817852246124e-06, + "loss": 0.5811, "step": 1812 }, { - "epoch": 0.35, - "learning_rate": 1.5125400233665728e-05, - "loss": 0.8856, + "epoch": 0.6975759907656791, + "grad_norm": 0.935625647453271, + "learning_rate": 4.425531605005519e-06, + "loss": 0.5267, "step": 1813 }, { - "epoch": 0.35, - "learning_rate": 1.5120047978169146e-05, - "loss": 0.9012, + "epoch": 0.6979607541362063, + "grad_norm": 1.0290548043683623, + "learning_rate": 4.4151900815878455e-06, + "loss": 0.5792, "step": 1814 }, { - "epoch": 0.35, - "learning_rate": 1.5114693734104926e-05, - "loss": 0.8928, + "epoch": 0.6983455175067333, + "grad_norm": 1.0359638470878285, + "learning_rate": 4.404857231031332e-06, + "loss": 0.601, "step": 1815 }, { - "epoch": 0.35, - "learning_rate": 1.5109337503552594e-05, - "loss": 0.9399, + "epoch": 0.6987302808772605, + "grad_norm": 0.9500117321096713, + "learning_rate": 4.394533069382255e-06, + "loss": 0.5978, "step": 1816 }, { - "epoch": 0.35, - "learning_rate": 1.5103979288592454e-05, - "loss": 0.8931, + "epoch": 0.6991150442477876, + "grad_norm": 0.9408762208637194, + "learning_rate": 4.3842176126733914e-06, + "loss": 0.5673, "step": 1817 }, { - "epoch": 0.35, - "learning_rate": 1.5098619091305571e-05, - "loss": 0.9793, + "epoch": 0.6994998076183148, + "grad_norm": 0.9904529830348353, + "learning_rate": 4.373910876923997e-06, + "loss": 0.5462, "step": 1818 }, { - "epoch": 0.35, - "learning_rate": 1.5093256913773786e-05, - "loss": 0.8004, + "epoch": 0.6998845709888418, + "grad_norm": 0.9758564642306548, + "learning_rate": 4.363612878139799e-06, + "loss": 0.5899, "step": 1819 }, { - "epoch": 0.35, - "learning_rate": 1.50878927580797e-05, - "loss": 0.9288, + "epoch": 0.700269334359369, + "grad_norm": 0.5907508023080977, + "learning_rate": 4.353323632312938e-06, + "loss": 0.6462, "step": 1820 }, { - "epoch": 0.35, - "learning_rate": 1.5082526626306698e-05, - "loss": 0.728, + "epoch": 0.7006540977298961, + "grad_norm": 1.038677723775359, + "learning_rate": 4.343043155421971e-06, + "loss": 0.5701, "step": 1821 }, { - "epoch": 0.35, - "learning_rate": 1.5077158520538921e-05, - "loss": 0.8232, + "epoch": 0.7010388611004232, + "grad_norm": 0.8563750759544907, + "learning_rate": 4.332771463431837e-06, + "loss": 0.5077, "step": 1822 }, { - "epoch": 0.35, - "learning_rate": 1.5071788442861277e-05, - "loss": 0.8576, + "epoch": 0.7014236244709504, + "grad_norm": 0.9307376628377482, + "learning_rate": 4.322508572293836e-06, + "loss": 0.5913, "step": 1823 }, { - "epoch": 0.35, - "learning_rate": 1.5066416395359444e-05, - "loss": 0.8909, + "epoch": 0.7018083878414775, + "grad_norm": 0.9160989592335452, + "learning_rate": 4.312254497945595e-06, + "loss": 0.5455, "step": 1824 }, { - "epoch": 0.35, - "learning_rate": 1.5061042380119864e-05, - "loss": 0.8908, + "epoch": 0.7021931512120047, + "grad_norm": 0.9092427957493121, + "learning_rate": 4.3020092563110485e-06, + "loss": 0.5089, "step": 1825 }, { - "epoch": 0.35, - "learning_rate": 1.5055666399229743e-05, - "loss": 0.9983, + "epoch": 0.7025779145825317, + "grad_norm": 0.5842707407389544, + "learning_rate": 4.291772863300428e-06, + "loss": 0.6202, "step": 1826 }, { - "epoch": 0.35, - "learning_rate": 1.5050288454777047e-05, - "loss": 0.7076, + "epoch": 0.7029626779530589, + "grad_norm": 0.9058515037399106, + "learning_rate": 4.281545334810201e-06, + "loss": 0.5362, "step": 1827 }, { - "epoch": 0.35, - "learning_rate": 1.504490854885051e-05, - "loss": 0.9471, + "epoch": 0.703347441323586, + "grad_norm": 0.9840070993540491, + "learning_rate": 4.27132668672308e-06, + "loss": 0.5401, "step": 1828 }, { - "epoch": 0.35, - "learning_rate": 1.5039526683539627e-05, - "loss": 0.9969, + "epoch": 0.7037322046941131, + "grad_norm": 1.107656132684394, + "learning_rate": 4.2611169349079985e-06, + "loss": 0.6122, "step": 1829 }, { - "epoch": 0.35, - "learning_rate": 1.5034142860934649e-05, - "loss": 0.9043, + "epoch": 0.7041169680646402, + "grad_norm": 0.9878316086776541, + "learning_rate": 4.250916095220056e-06, + "loss": 0.573, "step": 1830 }, { - "epoch": 0.35, - "learning_rate": 1.5028757083126594e-05, - "loss": 0.9026, + "epoch": 0.7045017314351674, + "grad_norm": 0.9413742248695964, + "learning_rate": 4.240724183500518e-06, + "loss": 0.5431, "step": 1831 }, { - "epoch": 0.35, - "learning_rate": 1.5023369352207229e-05, - "loss": 0.8843, + "epoch": 0.7048864948056945, + "grad_norm": 1.0297615858690405, + "learning_rate": 4.230541215576798e-06, + "loss": 0.5887, "step": 1832 }, { - "epoch": 0.35, - "learning_rate": 1.5017979670269096e-05, - "loss": 0.91, + "epoch": 0.7052712581762216, + "grad_norm": 0.9139824795517628, + "learning_rate": 4.220367207262398e-06, + "loss": 0.5828, "step": 1833 }, { - "epoch": 0.35, - "learning_rate": 1.501258803940548e-05, - "loss": 1.0304, + "epoch": 0.7056560215467488, + "grad_norm": 1.0365423717802127, + "learning_rate": 4.210202174356922e-06, + "loss": 0.5832, "step": 1834 }, { - "epoch": 0.35, - "learning_rate": 1.500719446171043e-05, - "loss": 0.8654, + "epoch": 0.7060407849172758, + "grad_norm": 1.0055139626224723, + "learning_rate": 4.2000461326460274e-06, + "loss": 0.5812, "step": 1835 }, { - "epoch": 0.35, - "learning_rate": 1.500179893927875e-05, - "loss": 0.926, + "epoch": 0.706425548287803, + "grad_norm": 1.0392623512119104, + "learning_rate": 4.189899097901421e-06, + "loss": 0.5779, "step": 1836 }, { - "epoch": 0.35, - "learning_rate": 1.4996401474205997e-05, - "loss": 0.6946, + "epoch": 0.7068103116583301, + "grad_norm": 0.9569467918911739, + "learning_rate": 4.179761085880809e-06, + "loss": 0.5588, "step": 1837 }, { - "epoch": 0.35, - "learning_rate": 1.4991002068588484e-05, - "loss": 0.9727, + "epoch": 0.7071950750288573, + "grad_norm": 0.9274665147377722, + "learning_rate": 4.16963211232789e-06, + "loss": 0.5801, "step": 1838 }, { - "epoch": 0.35, - "learning_rate": 1.4985600724523282e-05, - "loss": 0.9166, + "epoch": 0.7075798383993844, + "grad_norm": 0.9206307674328226, + "learning_rate": 4.159512192972337e-06, + "loss": 0.5523, "step": 1839 }, { - "epoch": 0.35, - "learning_rate": 1.4980197444108205e-05, - "loss": 1.0443, + "epoch": 0.7079646017699115, + "grad_norm": 0.9692918135907062, + "learning_rate": 4.149401343529742e-06, + "loss": 0.5274, "step": 1840 }, { - "epoch": 0.35, - "learning_rate": 1.4974792229441826e-05, - "loss": 0.9567, + "epoch": 0.7083493651404387, + "grad_norm": 1.0260353671052063, + "learning_rate": 4.139299579701623e-06, + "loss": 0.5756, "step": 1841 }, { - "epoch": 0.35, - "learning_rate": 1.4969385082623473e-05, - "loss": 0.9592, + "epoch": 0.7087341285109657, + "grad_norm": 1.0489989434818787, + "learning_rate": 4.129206917175397e-06, + "loss": 0.5268, "step": 1842 }, { - "epoch": 0.35, - "learning_rate": 1.4963976005753216e-05, - "loss": 0.8781, + "epoch": 0.7091188918814929, + "grad_norm": 0.9607393478704153, + "learning_rate": 4.119123371624335e-06, + "loss": 0.6248, "step": 1843 }, { - "epoch": 0.35, - "learning_rate": 1.4958565000931877e-05, - "loss": 0.981, + "epoch": 0.70950365525202, + "grad_norm": 0.9129178905897615, + "learning_rate": 4.109048958707552e-06, + "loss": 0.5604, "step": 1844 }, { - "epoch": 0.35, - "learning_rate": 1.4953152070261027e-05, - "loss": 0.9254, + "epoch": 0.7098884186225471, + "grad_norm": 0.9866970590537345, + "learning_rate": 4.09898369406998e-06, + "loss": 0.585, "step": 1845 }, { - "epoch": 0.36, - "learning_rate": 1.494773721584299e-05, - "loss": 0.9238, + "epoch": 0.7102731819930742, + "grad_norm": 1.003762797653451, + "learning_rate": 4.0889275933423576e-06, + "loss": 0.6281, "step": 1846 }, { - "epoch": 0.36, - "learning_rate": 1.4942320439780833e-05, - "loss": 0.8738, + "epoch": 0.7106579453636014, + "grad_norm": 0.5450282174455627, + "learning_rate": 4.078880672141171e-06, + "loss": 0.6016, "step": 1847 }, { - "epoch": 0.36, - "learning_rate": 1.4936901744178367e-05, - "loss": 0.8742, + "epoch": 0.7110427087341286, + "grad_norm": 0.9784158197491888, + "learning_rate": 4.068842946068661e-06, + "loss": 0.5765, "step": 1848 }, { - "epoch": 0.36, - "learning_rate": 1.4931481131140149e-05, - "loss": 0.9609, + "epoch": 0.7114274721046556, + "grad_norm": 1.151908412631802, + "learning_rate": 4.058814430712796e-06, + "loss": 0.6033, "step": 1849 }, { - "epoch": 0.36, - "learning_rate": 1.4926058602771484e-05, - "loss": 0.7606, + "epoch": 0.7118122354751828, + "grad_norm": 0.9972517647119352, + "learning_rate": 4.0487951416472324e-06, + "loss": 0.6118, "step": 1850 }, { - "epoch": 0.36, - "learning_rate": 1.4920634161178424e-05, - "loss": 0.9134, + "epoch": 0.7121969988457099, + "grad_norm": 0.9405896628966383, + "learning_rate": 4.038785094431295e-06, + "loss": 0.5403, "step": 1851 }, { - "epoch": 0.36, - "learning_rate": 1.4915207808467756e-05, - "loss": 0.8632, + "epoch": 0.712581762216237, + "grad_norm": 1.050188435631284, + "learning_rate": 4.0287843046099765e-06, + "loss": 0.6136, "step": 1852 }, { - "epoch": 0.36, - "learning_rate": 1.4909779546747011e-05, - "loss": 0.8916, + "epoch": 0.7129665255867641, + "grad_norm": 0.9642786015277207, + "learning_rate": 4.018792787713865e-06, + "loss": 0.6047, "step": 1853 }, { - "epoch": 0.36, - "learning_rate": 1.4904349378124467e-05, - "loss": 0.8574, + "epoch": 0.7133512889572913, + "grad_norm": 0.9659063359003159, + "learning_rate": 4.008810559259162e-06, + "loss": 0.5712, "step": 1854 }, { - "epoch": 0.36, - "learning_rate": 1.489891730470914e-05, - "loss": 0.9672, + "epoch": 0.7137360523278183, + "grad_norm": 0.9042148611238902, + "learning_rate": 3.998837634747655e-06, + "loss": 0.5514, "step": 1855 }, { - "epoch": 0.36, - "learning_rate": 1.4893483328610778e-05, - "loss": 0.836, + "epoch": 0.7141208156983455, + "grad_norm": 0.5491512301715373, + "learning_rate": 3.98887402966667e-06, + "loss": 0.6123, "step": 1856 }, { - "epoch": 0.36, - "learning_rate": 1.488804745193988e-05, - "loss": 0.8846, + "epoch": 0.7145055790688727, + "grad_norm": 0.5839683219999421, + "learning_rate": 3.97891975948906e-06, + "loss": 0.646, "step": 1857 }, { - "epoch": 0.36, - "learning_rate": 1.4882609676807675e-05, - "loss": 0.8932, + "epoch": 0.7148903424393998, + "grad_norm": 0.9966079168153738, + "learning_rate": 3.968974839673186e-06, + "loss": 0.6401, "step": 1858 }, { - "epoch": 0.36, - "learning_rate": 1.4877170005326136e-05, - "loss": 0.8988, + "epoch": 0.7152751058099269, + "grad_norm": 0.9573584941812288, + "learning_rate": 3.9590392856628946e-06, + "loss": 0.5524, "step": 1859 }, { - "epoch": 0.36, - "learning_rate": 1.4871728439607967e-05, - "loss": 0.9271, + "epoch": 0.715659869180454, + "grad_norm": 0.9936235624494605, + "learning_rate": 3.949113112887471e-06, + "loss": 0.5985, "step": 1860 }, { - "epoch": 0.36, - "learning_rate": 1.4866284981766607e-05, - "loss": 0.9289, + "epoch": 0.7160446325509812, + "grad_norm": 1.0585143794482035, + "learning_rate": 3.939196336761645e-06, + "loss": 0.5812, "step": 1861 }, { - "epoch": 0.36, - "learning_rate": 1.4860839633916236e-05, - "loss": 0.9495, + "epoch": 0.7164293959215082, + "grad_norm": 0.9301412666592186, + "learning_rate": 3.929288972685555e-06, + "loss": 0.5337, "step": 1862 }, { - "epoch": 0.36, - "learning_rate": 1.4855392398171762e-05, - "loss": 0.918, + "epoch": 0.7168141592920354, + "grad_norm": 1.0850128301027582, + "learning_rate": 3.919391036044715e-06, + "loss": 0.5803, "step": 1863 }, { - "epoch": 0.36, - "learning_rate": 1.484994327664883e-05, - "loss": 0.8275, + "epoch": 0.7171989226625626, + "grad_norm": 0.9327879211445063, + "learning_rate": 3.909502542210001e-06, + "loss": 0.5954, "step": 1864 }, { - "epoch": 0.36, - "learning_rate": 1.4844492271463814e-05, - "loss": 0.9238, + "epoch": 0.7175836860330896, + "grad_norm": 1.0242174782458806, + "learning_rate": 3.899623506537635e-06, + "loss": 0.5805, "step": 1865 }, { - "epoch": 0.36, - "learning_rate": 1.4839039384733821e-05, - "loss": 0.8859, + "epoch": 0.7179684494036168, + "grad_norm": 1.0788498582704387, + "learning_rate": 3.8897539443691355e-06, + "loss": 0.5938, "step": 1866 }, { - "epoch": 0.36, - "learning_rate": 1.4833584618576695e-05, - "loss": 0.8889, + "epoch": 0.7183532127741439, + "grad_norm": 1.0240139239932364, + "learning_rate": 3.879893871031314e-06, + "loss": 0.5965, "step": 1867 }, { - "epoch": 0.36, - "learning_rate": 1.4828127975111e-05, - "loss": 0.9263, + "epoch": 0.718737976144671, + "grad_norm": 0.989254247799077, + "learning_rate": 3.870043301836256e-06, + "loss": 0.6042, "step": 1868 }, { - "epoch": 0.36, - "learning_rate": 1.4822669456456031e-05, - "loss": 0.8874, + "epoch": 0.7191227395151981, + "grad_norm": 0.9144383876090636, + "learning_rate": 3.860202252081276e-06, + "loss": 0.5606, "step": 1869 }, { - "epoch": 0.36, - "learning_rate": 1.4817209064731819e-05, - "loss": 0.9605, + "epoch": 0.7195075028857253, + "grad_norm": 0.9588219868074859, + "learning_rate": 3.850370737048913e-06, + "loss": 0.4977, "step": 1870 }, { - "epoch": 0.36, - "learning_rate": 1.4811746802059115e-05, - "loss": 0.9311, + "epoch": 0.7198922662562524, + "grad_norm": 1.0112907226264074, + "learning_rate": 3.840548772006891e-06, + "loss": 0.5593, "step": 1871 }, { - "epoch": 0.36, - "learning_rate": 1.48062826705594e-05, - "loss": 0.9357, + "epoch": 0.7202770296267795, + "grad_norm": 1.0555827547431098, + "learning_rate": 3.830736372208118e-06, + "loss": 0.6176, "step": 1872 }, { - "epoch": 0.36, - "learning_rate": 1.4800816672354876e-05, - "loss": 0.9512, + "epoch": 0.7206617929973067, + "grad_norm": 1.034999767745451, + "learning_rate": 3.820933552890629e-06, + "loss": 0.5781, "step": 1873 }, { - "epoch": 0.36, - "learning_rate": 1.4795348809568477e-05, - "loss": 0.9345, + "epoch": 0.7210465563678338, + "grad_norm": 0.9496751542718201, + "learning_rate": 3.811140329277591e-06, + "loss": 0.5882, "step": 1874 }, { - "epoch": 0.36, - "learning_rate": 1.4789879084323858e-05, - "loss": 0.874, + "epoch": 0.7214313197383609, + "grad_norm": 0.9724001799461602, + "learning_rate": 3.8013567165772735e-06, + "loss": 0.5634, "step": 1875 }, { - "epoch": 0.36, - "learning_rate": 1.4784407498745394e-05, - "loss": 0.9431, + "epoch": 0.721816083108888, + "grad_norm": 0.9837668919578216, + "learning_rate": 3.7915827299830154e-06, + "loss": 0.5499, "step": 1876 }, { - "epoch": 0.36, - "learning_rate": 1.477893405495819e-05, - "loss": 0.871, + "epoch": 0.7222008464794152, + "grad_norm": 1.0087127753098368, + "learning_rate": 3.7818183846732024e-06, + "loss": 0.6025, "step": 1877 }, { - "epoch": 0.36, - "learning_rate": 1.4773458755088068e-05, - "loss": 0.9539, + "epoch": 0.7225856098499422, + "grad_norm": 1.0122024198610877, + "learning_rate": 3.7720636958112623e-06, + "loss": 0.5612, "step": 1878 }, { - "epoch": 0.36, - "learning_rate": 1.4767981601261567e-05, - "loss": 0.9601, + "epoch": 0.7229703732204694, + "grad_norm": 0.839942568514469, + "learning_rate": 3.7623186785456156e-06, + "loss": 0.6434, "step": 1879 }, { - "epoch": 0.36, - "learning_rate": 1.4762502595605957e-05, - "loss": 0.9111, + "epoch": 0.7233551365909965, + "grad_norm": 1.0405461467480457, + "learning_rate": 3.7525833480096575e-06, + "loss": 0.5556, "step": 1880 }, { - "epoch": 0.36, - "learning_rate": 1.4757021740249213e-05, - "loss": 0.9365, + "epoch": 0.7237398999615237, + "grad_norm": 0.9053394345248821, + "learning_rate": 3.7428577193217563e-06, + "loss": 0.5241, "step": 1881 }, { - "epoch": 0.36, - "learning_rate": 1.4751539037320044e-05, - "loss": 0.9024, + "epoch": 0.7241246633320508, + "grad_norm": 0.9864622048647691, + "learning_rate": 3.7331418075852053e-06, + "loss": 0.5903, "step": 1882 }, { - "epoch": 0.36, - "learning_rate": 1.4746054488947863e-05, - "loss": 0.9251, + "epoch": 0.7245094267025779, + "grad_norm": 1.0587027893326817, + "learning_rate": 3.7234356278882076e-06, + "loss": 0.5439, "step": 1883 }, { - "epoch": 0.36, - "learning_rate": 1.4740568097262811e-05, - "loss": 0.9431, + "epoch": 0.7248941900731051, + "grad_norm": 0.9359725097823653, + "learning_rate": 3.7137391953038516e-06, + "loss": 0.5513, "step": 1884 }, { - "epoch": 0.36, - "learning_rate": 1.473507986439573e-05, - "loss": 0.9414, + "epoch": 0.7252789534436321, + "grad_norm": 0.9858732523806819, + "learning_rate": 3.7040525248901003e-06, + "loss": 0.5503, "step": 1885 }, { - "epoch": 0.36, - "learning_rate": 1.4729589792478193e-05, - "loss": 0.9328, + "epoch": 0.7256637168141593, + "grad_norm": 0.9524862232548958, + "learning_rate": 3.6943756316897406e-06, + "loss": 0.5708, "step": 1886 }, { - "epoch": 0.36, - "learning_rate": 1.4724097883642482e-05, - "loss": 0.922, + "epoch": 0.7260484801846864, + "grad_norm": 0.9139614352999907, + "learning_rate": 3.684708530730382e-06, + "loss": 0.5266, "step": 1887 }, { - "epoch": 0.36, - "learning_rate": 1.4718604140021588e-05, - "loss": 0.9443, + "epoch": 0.7264332435552135, + "grad_norm": 0.8640958303601587, + "learning_rate": 3.6750512370244363e-06, + "loss": 0.5271, "step": 1888 }, { - "epoch": 0.36, - "learning_rate": 1.471310856374922e-05, - "loss": 0.9194, + "epoch": 0.7268180069257407, + "grad_norm": 0.9669007547478378, + "learning_rate": 3.6654037655690732e-06, + "loss": 0.5957, "step": 1889 }, { - "epoch": 0.36, - "learning_rate": 1.470761115695979e-05, - "loss": 0.8964, + "epoch": 0.7272027702962678, + "grad_norm": 0.9915797504470046, + "learning_rate": 3.655766131346211e-06, + "loss": 0.5522, "step": 1890 }, { - "epoch": 0.36, - "learning_rate": 1.4702111921788437e-05, - "loss": 0.7331, + "epoch": 0.727587533666795, + "grad_norm": 0.9991612590145086, + "learning_rate": 3.6461383493225012e-06, + "loss": 0.5763, "step": 1891 }, { - "epoch": 0.36, - "learning_rate": 1.4696610860370997e-05, - "loss": 0.8706, + "epoch": 0.727972297037322, + "grad_norm": 1.0322739737897488, + "learning_rate": 3.6365204344492867e-06, + "loss": 0.5506, "step": 1892 }, { - "epoch": 0.36, - "learning_rate": 1.4691107974844015e-05, - "loss": 0.8772, + "epoch": 0.7283570604078492, + "grad_norm": 1.0238234537645174, + "learning_rate": 3.62691240166258e-06, + "loss": 0.5919, "step": 1893 }, { - "epoch": 0.36, - "learning_rate": 1.468560326734475e-05, - "loss": 0.9179, + "epoch": 0.7287418237783763, + "grad_norm": 1.003325307503168, + "learning_rate": 3.617314265883066e-06, + "loss": 0.5795, "step": 1894 }, { - "epoch": 0.36, - "learning_rate": 1.4680096740011172e-05, - "loss": 0.905, + "epoch": 0.7291265871489034, + "grad_norm": 0.9477485984789025, + "learning_rate": 3.6077260420160487e-06, + "loss": 0.5633, "step": 1895 }, { - "epoch": 0.36, - "learning_rate": 1.4674588394981948e-05, - "loss": 0.8976, + "epoch": 0.7295113505194305, + "grad_norm": 0.9274576159627026, + "learning_rate": 3.598147744951438e-06, + "loss": 0.5136, "step": 1896 }, { - "epoch": 0.36, - "learning_rate": 1.4669078234396454e-05, - "loss": 0.9305, + "epoch": 0.7298961138899577, + "grad_norm": 1.027273969446434, + "learning_rate": 3.58857938956373e-06, + "loss": 0.6032, "step": 1897 }, { - "epoch": 0.37, - "learning_rate": 1.4663566260394775e-05, - "loss": 0.945, + "epoch": 0.7302808772604849, + "grad_norm": 0.872792283684926, + "learning_rate": 3.57902099071199e-06, + "loss": 0.5351, "step": 1898 }, { - "epoch": 0.37, - "learning_rate": 1.4658052475117704e-05, - "loss": 0.9786, + "epoch": 0.7306656406310119, + "grad_norm": 0.8781553904706526, + "learning_rate": 3.569472563239814e-06, + "loss": 0.5618, "step": 1899 }, { - "epoch": 0.37, - "learning_rate": 1.4652536880706723e-05, - "loss": 0.9392, + "epoch": 0.7310504040015391, + "grad_norm": 0.9247291296257878, + "learning_rate": 3.559934121975304e-06, + "loss": 0.4992, "step": 1900 }, { - "epoch": 0.37, - "learning_rate": 1.4647019479304028e-05, - "loss": 0.9116, + "epoch": 0.7314351673720662, + "grad_norm": 1.0082635376328732, + "learning_rate": 3.550405681731074e-06, + "loss": 0.6027, "step": 1901 }, { - "epoch": 0.37, - "learning_rate": 1.4641500273052516e-05, - "loss": 0.9531, + "epoch": 0.7318199307425933, + "grad_norm": 0.923801970045176, + "learning_rate": 3.540887257304193e-06, + "loss": 0.5481, "step": 1902 }, { - "epoch": 0.37, - "learning_rate": 1.463597926409578e-05, - "loss": 0.8642, + "epoch": 0.7322046941131204, + "grad_norm": 0.9253580461931376, + "learning_rate": 3.531378863476178e-06, + "loss": 0.5904, "step": 1903 }, { - "epoch": 0.37, - "learning_rate": 1.4630456454578122e-05, - "loss": 0.9595, + "epoch": 0.7325894574836476, + "grad_norm": 0.9983315603349991, + "learning_rate": 3.5218805150129755e-06, + "loss": 0.6089, "step": 1904 }, { - "epoch": 0.37, - "learning_rate": 1.462493184664453e-05, - "loss": 0.9098, + "epoch": 0.7329742208541746, + "grad_norm": 1.001191429191074, + "learning_rate": 3.51239222666493e-06, + "loss": 0.5801, "step": 1905 }, { - "epoch": 0.37, - "learning_rate": 1.4619405442440702e-05, - "loss": 0.8798, + "epoch": 0.7333589842247018, + "grad_norm": 0.9906088130595536, + "learning_rate": 3.5029140131667493e-06, + "loss": 0.6137, "step": 1906 }, { - "epoch": 0.37, - "learning_rate": 1.4613877244113033e-05, - "loss": 0.9222, + "epoch": 0.733743747595229, + "grad_norm": 0.9443438075970022, + "learning_rate": 3.493445889237518e-06, + "loss": 0.5779, "step": 1907 }, { - "epoch": 0.37, - "learning_rate": 1.4608347253808605e-05, - "loss": 0.7812, + "epoch": 0.734128510965756, + "grad_norm": 0.9712102107475222, + "learning_rate": 3.4839878695806385e-06, + "loss": 0.5192, "step": 1908 }, { - "epoch": 0.37, - "learning_rate": 1.460281547367521e-05, - "loss": 0.961, + "epoch": 0.7345132743362832, + "grad_norm": 0.9528455865999759, + "learning_rate": 3.4745399688838243e-06, + "loss": 0.6099, "step": 1909 }, { - "epoch": 0.37, - "learning_rate": 1.4597281905861318e-05, - "loss": 0.8948, + "epoch": 0.7348980377068103, + "grad_norm": 1.0569639486019915, + "learning_rate": 3.4651022018190715e-06, + "loss": 0.619, "step": 1910 }, { - "epoch": 0.37, - "learning_rate": 1.4591746552516109e-05, - "loss": 0.7752, + "epoch": 0.7352828010773375, + "grad_norm": 1.0087246434523038, + "learning_rate": 3.455674583042652e-06, + "loss": 0.6269, "step": 1911 }, { - "epoch": 0.37, - "learning_rate": 1.4586209415789452e-05, - "loss": 0.8994, + "epoch": 0.7356675644478645, + "grad_norm": 0.9055662193863306, + "learning_rate": 3.4462571271950674e-06, + "loss": 0.5293, "step": 1912 }, { - "epoch": 0.37, - "learning_rate": 1.4580670497831904e-05, - "loss": 1.0239, + "epoch": 0.7360523278183917, + "grad_norm": 0.9961762006422132, + "learning_rate": 3.436849848901028e-06, + "loss": 0.6251, "step": 1913 }, { - "epoch": 0.37, - "learning_rate": 1.4575129800794718e-05, - "loss": 0.9427, + "epoch": 0.7364370911889189, + "grad_norm": 1.0075678095032963, + "learning_rate": 3.427452762769462e-06, + "loss": 0.5985, "step": 1914 }, { - "epoch": 0.37, - "learning_rate": 1.4569587326829834e-05, - "loss": 0.9572, + "epoch": 0.7368218545594459, + "grad_norm": 0.912856417402377, + "learning_rate": 3.4180658833934523e-06, + "loss": 0.5529, "step": 1915 }, { - "epoch": 0.37, - "learning_rate": 1.4564043078089891e-05, - "loss": 0.8331, + "epoch": 0.7372066179299731, + "grad_norm": 0.9160113841052815, + "learning_rate": 3.4086892253502344e-06, + "loss": 0.5432, "step": 1916 }, { - "epoch": 0.37, - "learning_rate": 1.4558497056728205e-05, - "loss": 0.9358, + "epoch": 0.7375913813005002, + "grad_norm": 0.8875299979080135, + "learning_rate": 3.3993228032011784e-06, + "loss": 0.5542, "step": 1917 }, { - "epoch": 0.37, - "learning_rate": 1.4552949264898795e-05, - "loss": 0.8331, + "epoch": 0.7379761446710273, + "grad_norm": 1.1189513304738115, + "learning_rate": 3.3899666314917512e-06, + "loss": 0.5731, "step": 1918 }, { - "epoch": 0.37, - "learning_rate": 1.4547399704756348e-05, - "loss": 0.9075, + "epoch": 0.7383609080415544, + "grad_norm": 0.9640870674188529, + "learning_rate": 3.3806207247515068e-06, + "loss": 0.5361, "step": 1919 }, { - "epoch": 0.37, - "learning_rate": 1.4541848378456255e-05, - "loss": 0.8938, + "epoch": 0.7387456714120816, + "grad_norm": 0.9449903025460902, + "learning_rate": 3.3712850974940437e-06, + "loss": 0.5194, "step": 1920 }, { - "epoch": 0.37, - "learning_rate": 1.4536295288154594e-05, - "loss": 0.932, + "epoch": 0.7391304347826086, + "grad_norm": 0.9514400815159572, + "learning_rate": 3.361959764217018e-06, + "loss": 0.5574, "step": 1921 }, { - "epoch": 0.37, - "learning_rate": 1.4530740436008111e-05, - "loss": 0.8973, + "epoch": 0.7395151981531358, + "grad_norm": 0.9742018117179742, + "learning_rate": 3.3526447394020887e-06, + "loss": 0.5832, "step": 1922 }, { - "epoch": 0.37, - "learning_rate": 1.452518382417425e-05, - "loss": 1.0021, + "epoch": 0.739899961523663, + "grad_norm": 1.016337033866378, + "learning_rate": 3.343340037514903e-06, + "loss": 0.6267, "step": 1923 }, { - "epoch": 0.37, - "learning_rate": 1.4519625454811135e-05, - "loss": 0.9088, + "epoch": 0.7402847248941901, + "grad_norm": 0.9869101637712542, + "learning_rate": 3.3340456730050887e-06, + "loss": 0.5958, "step": 1924 }, { - "epoch": 0.37, - "learning_rate": 1.4514065330077575e-05, - "loss": 0.9339, + "epoch": 0.7406694882647172, + "grad_norm": 0.9329949275605486, + "learning_rate": 3.324761660306215e-06, + "loss": 0.5266, "step": 1925 }, { - "epoch": 0.37, - "learning_rate": 1.4508503452133053e-05, - "loss": 0.8649, + "epoch": 0.7410542516352443, + "grad_norm": 0.9474821257516509, + "learning_rate": 3.3154880138357626e-06, + "loss": 0.5467, "step": 1926 }, { - "epoch": 0.37, - "learning_rate": 1.4502939823137744e-05, - "loss": 1.0029, + "epoch": 0.7414390150057715, + "grad_norm": 0.9282609757282041, + "learning_rate": 3.306224747995136e-06, + "loss": 0.6035, "step": 1927 }, { - "epoch": 0.37, - "learning_rate": 1.4497374445252496e-05, - "loss": 0.8814, + "epoch": 0.7418237783762985, + "grad_norm": 1.0253221181577397, + "learning_rate": 3.2969718771696047e-06, + "loss": 0.5614, "step": 1928 }, { - "epoch": 0.37, - "learning_rate": 1.4491807320638835e-05, - "loss": 0.9634, + "epoch": 0.7422085417468257, + "grad_norm": 0.9481838901061963, + "learning_rate": 3.287729415728298e-06, + "loss": 0.5592, "step": 1929 }, { - "epoch": 0.37, - "learning_rate": 1.4486238451458972e-05, - "loss": 0.8546, + "epoch": 0.7425933051173528, + "grad_norm": 0.8874620108814821, + "learning_rate": 3.278497378024187e-06, + "loss": 0.5574, "step": 1930 }, { - "epoch": 0.37, - "learning_rate": 1.4480667839875786e-05, - "loss": 0.9306, + "epoch": 0.74297806848788, + "grad_norm": 1.084822177741577, + "learning_rate": 3.2692757783940467e-06, + "loss": 0.6144, "step": 1931 }, { - "epoch": 0.37, - "learning_rate": 1.4475095488052843e-05, - "loss": 0.8675, + "epoch": 0.7433628318584071, + "grad_norm": 0.8822852704039786, + "learning_rate": 3.2600646311584494e-06, + "loss": 0.5406, "step": 1932 }, { - "epoch": 0.37, - "learning_rate": 1.4469521398154381e-05, - "loss": 0.9259, + "epoch": 0.7437475952289342, + "grad_norm": 1.033441144492799, + "learning_rate": 3.250863950621721e-06, + "loss": 0.5819, "step": 1933 }, { - "epoch": 0.37, - "learning_rate": 1.4463945572345308e-05, - "loss": 0.9267, + "epoch": 0.7441323585994614, + "grad_norm": 0.8808158911222783, + "learning_rate": 3.241673751071954e-06, + "loss": 0.5342, "step": 1934 }, { - "epoch": 0.37, - "learning_rate": 1.4458368012791213e-05, - "loss": 0.734, + "epoch": 0.7445171219699884, + "grad_norm": 0.9078856146254358, + "learning_rate": 3.2324940467809527e-06, + "loss": 0.5001, "step": 1935 }, { - "epoch": 0.37, - "learning_rate": 1.4452788721658355e-05, - "loss": 1.0384, + "epoch": 0.7449018853405156, + "grad_norm": 0.9086214028585684, + "learning_rate": 3.223324852004219e-06, + "loss": 0.5649, "step": 1936 }, { - "epoch": 0.37, - "learning_rate": 1.4447207701113669e-05, - "loss": 0.9312, + "epoch": 0.7452866487110427, + "grad_norm": 0.9409111117583024, + "learning_rate": 3.21416618098095e-06, + "loss": 0.5322, "step": 1937 }, { - "epoch": 0.37, - "learning_rate": 1.4441624953324755e-05, - "loss": 0.9932, + "epoch": 0.7456714120815698, + "grad_norm": 0.9466143457907542, + "learning_rate": 3.2050180479339865e-06, + "loss": 0.5465, "step": 1938 }, { - "epoch": 0.37, - "learning_rate": 1.4436040480459891e-05, - "loss": 0.8806, + "epoch": 0.746056175452097, + "grad_norm": 0.9556085304317968, + "learning_rate": 3.1958804670698008e-06, + "loss": 0.5948, "step": 1939 }, { - "epoch": 0.37, - "learning_rate": 1.443045428468802e-05, - "loss": 0.8763, + "epoch": 0.7464409388226241, + "grad_norm": 0.9287004988061286, + "learning_rate": 3.1867534525784937e-06, + "loss": 0.5517, "step": 1940 }, { - "epoch": 0.37, - "learning_rate": 1.4424866368178761e-05, - "loss": 0.887, + "epoch": 0.7468257021931513, + "grad_norm": 1.0974006764706135, + "learning_rate": 3.177637018633746e-06, + "loss": 0.6163, "step": 1941 }, { - "epoch": 0.37, - "learning_rate": 1.441927673310239e-05, - "loss": 0.9406, + "epoch": 0.7472104655636783, + "grad_norm": 1.0191024614863955, + "learning_rate": 3.1685311793928077e-06, + "loss": 0.5612, "step": 1942 }, { - "epoch": 0.37, - "learning_rate": 1.4413685381629855e-05, - "loss": 0.9437, + "epoch": 0.7475952289342055, + "grad_norm": 1.0234335254323141, + "learning_rate": 3.1594359489964853e-06, + "loss": 0.5772, "step": 1943 }, { - "epoch": 0.37, - "learning_rate": 1.440809231593278e-05, - "loss": 0.8901, + "epoch": 0.7479799923047326, + "grad_norm": 1.029902698697747, + "learning_rate": 3.150351341569101e-06, + "loss": 0.599, "step": 1944 }, { - "epoch": 0.37, - "learning_rate": 1.4402497538183444e-05, - "loss": 0.9055, + "epoch": 0.7483647556752597, + "grad_norm": 0.9866696765584446, + "learning_rate": 3.141277371218484e-06, + "loss": 0.5742, "step": 1945 }, { - "epoch": 0.37, - "learning_rate": 1.4396901050554794e-05, - "loss": 0.9734, + "epoch": 0.7487495190457868, + "grad_norm": 0.955278493250578, + "learning_rate": 3.1322140520359366e-06, + "loss": 0.5675, "step": 1946 }, { - "epoch": 0.37, - "learning_rate": 1.4391302855220442e-05, - "loss": 0.8968, + "epoch": 0.749134282416314, + "grad_norm": 1.007842480092676, + "learning_rate": 3.1231613980962373e-06, + "loss": 0.5751, "step": 1947 }, { - "epoch": 0.37, - "learning_rate": 1.4385702954354662e-05, - "loss": 0.866, + "epoch": 0.7495190457868411, + "grad_norm": 0.9996774437976602, + "learning_rate": 3.1141194234575878e-06, + "loss": 0.572, "step": 1948 }, { - "epoch": 0.37, - "learning_rate": 1.438010135013239e-05, - "loss": 0.9844, + "epoch": 0.7499038091573682, + "grad_norm": 1.027262939252386, + "learning_rate": 3.1050881421616076e-06, + "loss": 0.6021, "step": 1949 }, { - "epoch": 0.38, - "learning_rate": 1.4374498044729225e-05, - "loss": 0.7199, + "epoch": 0.7502885725278954, + "grad_norm": 0.898600770234585, + "learning_rate": 3.0960675682333186e-06, + "loss": 0.5756, "step": 1950 }, { - "epoch": 0.38, - "learning_rate": 1.4368893040321428e-05, - "loss": 0.9024, + "epoch": 0.7506733358984224, + "grad_norm": 0.990416294221106, + "learning_rate": 3.0870577156811077e-06, + "loss": 0.5738, "step": 1951 }, { - "epoch": 0.38, - "learning_rate": 1.4363286339085915e-05, - "loss": 0.9428, + "epoch": 0.7510580992689496, + "grad_norm": 1.1538755213033571, + "learning_rate": 3.0780585984967113e-06, + "loss": 0.5825, "step": 1952 }, { - "epoch": 0.38, - "learning_rate": 1.435767794320027e-05, - "loss": 0.896, + "epoch": 0.7514428626394767, + "grad_norm": 1.0145121999245792, + "learning_rate": 3.069070230655198e-06, + "loss": 0.5724, "step": 1953 }, { - "epoch": 0.38, - "learning_rate": 1.4352067854842724e-05, - "loss": 0.8405, + "epoch": 0.7518276260100039, + "grad_norm": 1.0841381534232097, + "learning_rate": 3.060092626114941e-06, + "loss": 0.6161, "step": 1954 }, { - "epoch": 0.38, - "learning_rate": 1.434645607619217e-05, - "loss": 0.98, + "epoch": 0.7522123893805309, + "grad_norm": 1.0291766872101848, + "learning_rate": 3.051125798817598e-06, + "loss": 0.6085, "step": 1955 }, { - "epoch": 0.38, - "learning_rate": 1.434084260942816e-05, - "loss": 0.9417, + "epoch": 0.7525971527510581, + "grad_norm": 0.9647229154660495, + "learning_rate": 3.042169762688096e-06, + "loss": 0.513, "step": 1956 }, { - "epoch": 0.38, - "learning_rate": 1.4335227456730902e-05, - "loss": 0.7749, + "epoch": 0.7529819161215853, + "grad_norm": 0.9391750427503991, + "learning_rate": 3.0332245316346e-06, + "loss": 0.5854, "step": 1957 }, { - "epoch": 0.38, - "learning_rate": 1.4329610620281253e-05, - "loss": 1.046, + "epoch": 0.7533666794921123, + "grad_norm": 0.9648894885918808, + "learning_rate": 3.024290119548495e-06, + "loss": 0.5335, "step": 1958 }, { - "epoch": 0.38, - "learning_rate": 1.4323992102260733e-05, - "loss": 0.9214, + "epoch": 0.7537514428626395, + "grad_norm": 1.0168360896560233, + "learning_rate": 3.0153665403043586e-06, + "loss": 0.5629, "step": 1959 }, { - "epoch": 0.38, - "learning_rate": 1.4318371904851502e-05, - "loss": 0.8894, + "epoch": 0.7541362062331666, + "grad_norm": 0.941949182730458, + "learning_rate": 3.0064538077599603e-06, + "loss": 0.5137, "step": 1960 }, { - "epoch": 0.38, - "learning_rate": 1.4312750030236382e-05, - "loss": 0.9308, + "epoch": 0.7545209696036937, + "grad_norm": 1.0947807015044355, + "learning_rate": 2.9975519357562155e-06, + "loss": 0.5909, "step": 1961 }, { - "epoch": 0.38, - "learning_rate": 1.4307126480598852e-05, - "loss": 0.8339, + "epoch": 0.7549057329742208, + "grad_norm": 0.9920604585538697, + "learning_rate": 2.9886609381171703e-06, + "loss": 0.5934, "step": 1962 }, { - "epoch": 0.38, - "learning_rate": 1.4301501258123024e-05, - "loss": 0.949, + "epoch": 0.755290496344748, + "grad_norm": 0.9632312330261877, + "learning_rate": 2.9797808286499976e-06, + "loss": 0.5727, "step": 1963 }, { - "epoch": 0.38, - "learning_rate": 1.4295874364993672e-05, - "loss": 0.9257, + "epoch": 0.7556752597152752, + "grad_norm": 1.1209239455077458, + "learning_rate": 2.9709116211449484e-06, + "loss": 0.5787, "step": 1964 }, { - "epoch": 0.38, - "learning_rate": 1.4290245803396221e-05, - "loss": 0.9312, + "epoch": 0.7560600230858022, + "grad_norm": 1.000051429272563, + "learning_rate": 2.9620533293753495e-06, + "loss": 0.5835, "step": 1965 }, { - "epoch": 0.38, - "learning_rate": 1.4284615575516737e-05, - "loss": 0.891, + "epoch": 0.7564447864563294, + "grad_norm": 0.9663894188205748, + "learning_rate": 2.9532059670975732e-06, + "loss": 0.5051, "step": 1966 }, { - "epoch": 0.38, - "learning_rate": 1.4278983683541934e-05, - "loss": 0.7396, + "epoch": 0.7568295498268565, + "grad_norm": 0.95244909207669, + "learning_rate": 2.9443695480510225e-06, + "loss": 0.5816, "step": 1967 }, { - "epoch": 0.38, - "learning_rate": 1.4273350129659173e-05, - "loss": 0.8304, + "epoch": 0.7572143131973836, + "grad_norm": 1.0231557551803332, + "learning_rate": 2.935544085958102e-06, + "loss": 0.5846, "step": 1968 }, { - "epoch": 0.38, - "learning_rate": 1.4267714916056465e-05, - "loss": 0.8944, + "epoch": 0.7575990765679107, + "grad_norm": 0.9381517267221178, + "learning_rate": 2.926729594524207e-06, + "loss": 0.5562, "step": 1969 }, { - "epoch": 0.38, - "learning_rate": 1.426207804492246e-05, - "loss": 0.9375, + "epoch": 0.7579838399384379, + "grad_norm": 1.0919539636162612, + "learning_rate": 2.9179260874376915e-06, + "loss": 0.6093, "step": 1970 }, { - "epoch": 0.38, - "learning_rate": 1.4256439518446456e-05, - "loss": 0.7429, + "epoch": 0.7583686033089649, + "grad_norm": 0.9616294776854274, + "learning_rate": 2.9091335783698517e-06, + "loss": 0.5899, "step": 1971 }, { - "epoch": 0.38, - "learning_rate": 1.4250799338818388e-05, - "loss": 0.8786, + "epoch": 0.7587533666794921, + "grad_norm": 1.0200826301819041, + "learning_rate": 2.9003520809749053e-06, + "loss": 0.5504, "step": 1972 }, { - "epoch": 0.38, - "learning_rate": 1.424515750822884e-05, - "loss": 0.8585, + "epoch": 0.7591381300500193, + "grad_norm": 1.001870725609725, + "learning_rate": 2.8915816088899696e-06, + "loss": 0.597, "step": 1973 }, { - "epoch": 0.38, - "learning_rate": 1.4239514028869032e-05, - "loss": 0.9504, + "epoch": 0.7595228934205464, + "grad_norm": 0.9918429460788176, + "learning_rate": 2.8828221757350406e-06, + "loss": 0.6027, "step": 1974 }, { - "epoch": 0.38, - "learning_rate": 1.4233868902930827e-05, - "loss": 0.8537, + "epoch": 0.7599076567910735, + "grad_norm": 1.0083900609332672, + "learning_rate": 2.874073795112967e-06, + "loss": 0.5781, "step": 1975 }, { - "epoch": 0.38, - "learning_rate": 1.4228222132606729e-05, - "loss": 0.9382, + "epoch": 0.7602924201616006, + "grad_norm": 0.9258643742526984, + "learning_rate": 2.8653364806094454e-06, + "loss": 0.5844, "step": 1976 }, { - "epoch": 0.38, - "learning_rate": 1.4222573720089874e-05, - "loss": 0.9719, + "epoch": 0.7606771835321278, + "grad_norm": 1.0534523132503755, + "learning_rate": 2.856610245792976e-06, + "loss": 0.5529, "step": 1977 }, { - "epoch": 0.38, - "learning_rate": 1.4216923667574042e-05, - "loss": 0.8537, + "epoch": 0.7610619469026548, + "grad_norm": 0.9929816172968349, + "learning_rate": 2.847895104214856e-06, + "loss": 0.5412, "step": 1978 }, { - "epoch": 0.38, - "learning_rate": 1.4211271977253653e-05, - "loss": 0.9884, + "epoch": 0.761446710273182, + "grad_norm": 1.0148058549374095, + "learning_rate": 2.8391910694091584e-06, + "loss": 0.6082, "step": 1979 }, { - "epoch": 0.38, - "learning_rate": 1.4205618651323753e-05, - "loss": 0.9701, + "epoch": 0.7618314736437091, + "grad_norm": 1.0167980488366908, + "learning_rate": 2.8304981548927025e-06, + "loss": 0.58, "step": 1980 }, { - "epoch": 0.38, - "learning_rate": 1.4199963691980027e-05, - "loss": 0.8605, + "epoch": 0.7622162370142362, + "grad_norm": 0.9966300484914428, + "learning_rate": 2.8218163741650415e-06, + "loss": 0.5681, "step": 1981 }, { - "epoch": 0.38, - "learning_rate": 1.4194307101418805e-05, - "loss": 0.9331, + "epoch": 0.7626010003847634, + "grad_norm": 1.02083829999466, + "learning_rate": 2.813145740708445e-06, + "loss": 0.5483, "step": 1982 }, { - "epoch": 0.38, - "learning_rate": 1.4188648881837033e-05, - "loss": 0.9143, + "epoch": 0.7629857637552905, + "grad_norm": 0.8709014104497395, + "learning_rate": 2.8044862679878605e-06, + "loss": 0.5184, "step": 1983 }, { - "epoch": 0.38, - "learning_rate": 1.4182989035432299e-05, - "loss": 0.8609, + "epoch": 0.7633705271258177, + "grad_norm": 1.0532482846080897, + "learning_rate": 2.7958379694509108e-06, + "loss": 0.5747, "step": 1984 }, { - "epoch": 0.38, - "learning_rate": 1.4177327564402825e-05, - "loss": 0.9781, + "epoch": 0.7637552904963447, + "grad_norm": 0.9679247075149837, + "learning_rate": 2.787200858527862e-06, + "loss": 0.5653, "step": 1985 }, { - "epoch": 0.38, - "learning_rate": 1.4171664470947464e-05, - "loss": 0.964, + "epoch": 0.7641400538668719, + "grad_norm": 0.9545732694517383, + "learning_rate": 2.7785749486316085e-06, + "loss": 0.5669, "step": 1986 }, { - "epoch": 0.38, - "learning_rate": 1.416599975726569e-05, - "loss": 1.0024, + "epoch": 0.764524817237399, + "grad_norm": 1.0002666827609774, + "learning_rate": 2.7699602531576496e-06, + "loss": 0.5631, "step": 1987 }, { - "epoch": 0.38, - "learning_rate": 1.4160333425557616e-05, - "loss": 0.8887, + "epoch": 0.7649095806079261, + "grad_norm": 0.974192747364335, + "learning_rate": 2.7613567854840685e-06, + "loss": 0.5454, "step": 1988 }, { - "epoch": 0.38, - "learning_rate": 1.4154665478023977e-05, - "loss": 0.9632, + "epoch": 0.7652943439784533, + "grad_norm": 0.9531007126437687, + "learning_rate": 2.752764558971517e-06, + "loss": 0.5252, "step": 1989 }, { - "epoch": 0.38, - "learning_rate": 1.4148995916866139e-05, - "loss": 0.8954, + "epoch": 0.7656791073489804, + "grad_norm": 1.0424629542353707, + "learning_rate": 2.744183586963185e-06, + "loss": 0.6285, "step": 1990 }, { - "epoch": 0.38, - "learning_rate": 1.41433247442861e-05, - "loss": 0.8515, + "epoch": 0.7660638707195075, + "grad_norm": 0.623035871870837, + "learning_rate": 2.7356138827847856e-06, + "loss": 0.6652, "step": 1991 }, { - "epoch": 0.38, - "learning_rate": 1.4137651962486472e-05, - "loss": 0.7941, + "epoch": 0.7664486340900346, + "grad_norm": 0.9062009225145533, + "learning_rate": 2.7270554597445343e-06, + "loss": 0.51, "step": 1992 }, { - "epoch": 0.38, - "learning_rate": 1.4131977573670499e-05, - "loss": 0.962, + "epoch": 0.7668333974605618, + "grad_norm": 1.063307615409097, + "learning_rate": 2.7185083311331283e-06, + "loss": 0.6218, "step": 1993 }, { - "epoch": 0.38, - "learning_rate": 1.412630158004205e-05, - "loss": 0.9302, + "epoch": 0.7672181608310888, + "grad_norm": 1.0038462199622615, + "learning_rate": 2.709972510223725e-06, + "loss": 0.5878, "step": 1994 }, { - "epoch": 0.38, - "learning_rate": 1.4120623983805617e-05, - "loss": 0.9345, + "epoch": 0.767602924201616, + "grad_norm": 0.973558856010213, + "learning_rate": 2.7014480102719174e-06, + "loss": 0.5771, "step": 1995 }, { - "epoch": 0.38, - "learning_rate": 1.4114944787166307e-05, - "loss": 0.9405, + "epoch": 0.7679876875721431, + "grad_norm": 0.980022033982001, + "learning_rate": 2.692934844515729e-06, + "loss": 0.5608, "step": 1996 }, { - "epoch": 0.38, - "learning_rate": 1.4109263992329858e-05, - "loss": 0.9697, + "epoch": 0.7683724509426703, + "grad_norm": 0.9657030323938196, + "learning_rate": 2.6844330261755715e-06, + "loss": 0.5329, "step": 1997 }, { - "epoch": 0.38, - "learning_rate": 1.4103581601502629e-05, - "loss": 0.699, + "epoch": 0.7687572143131974, + "grad_norm": 1.0025053238345116, + "learning_rate": 2.675942568454236e-06, + "loss": 0.5688, "step": 1998 }, { - "epoch": 0.38, - "learning_rate": 1.409789761689159e-05, - "loss": 0.9473, + "epoch": 0.7691419776837245, + "grad_norm": 1.0112706490901935, + "learning_rate": 2.667463484536876e-06, + "loss": 0.57, "step": 1999 }, { - "epoch": 0.38, - "learning_rate": 1.4092212040704336e-05, - "loss": 0.8968, + "epoch": 0.7695267410542517, + "grad_norm": 0.9691380861621212, + "learning_rate": 2.65899578759098e-06, + "loss": 0.6041, "step": 2000 }, { - "epoch": 0.38, - "learning_rate": 1.408652487514908e-05, - "loss": 0.9069, + "epoch": 0.7699115044247787, + "grad_norm": 1.05944321880058, + "learning_rate": 2.650539490766346e-06, + "loss": 0.5972, "step": 2001 }, { - "epoch": 0.39, - "learning_rate": 1.408083612243465e-05, - "loss": 0.701, + "epoch": 0.7702962677953059, + "grad_norm": 0.9760894626636357, + "learning_rate": 2.642094607195085e-06, + "loss": 0.5355, "step": 2002 }, { - "epoch": 0.39, - "learning_rate": 1.4075145784770496e-05, - "loss": 0.9094, + "epoch": 0.770681031165833, + "grad_norm": 1.040620639456867, + "learning_rate": 2.633661149991569e-06, + "loss": 0.5687, "step": 2003 }, { - "epoch": 0.39, - "learning_rate": 1.4069453864366678e-05, - "loss": 0.9485, + "epoch": 0.7710657945363601, + "grad_norm": 0.9295583741033754, + "learning_rate": 2.6252391322524297e-06, + "loss": 0.5454, "step": 2004 }, { - "epoch": 0.39, - "learning_rate": 1.4063760363433867e-05, - "loss": 0.7746, + "epoch": 0.7714505579068872, + "grad_norm": 0.918636492428104, + "learning_rate": 2.6168285670565374e-06, + "loss": 0.5869, "step": 2005 }, { - "epoch": 0.39, - "learning_rate": 1.405806528418336e-05, - "loss": 0.8968, + "epoch": 0.7718353212774144, + "grad_norm": 0.9712186153891921, + "learning_rate": 2.6084294674649734e-06, + "loss": 0.5805, "step": 2006 }, { - "epoch": 0.39, - "learning_rate": 1.4052368628827057e-05, - "loss": 0.8992, + "epoch": 0.7722200846479416, + "grad_norm": 0.9928832388906447, + "learning_rate": 2.6000418465210143e-06, + "loss": 0.5483, "step": 2007 }, { - "epoch": 0.39, - "learning_rate": 1.4046670399577478e-05, - "loss": 0.9247, + "epoch": 0.7726048480184686, + "grad_norm": 0.9800296965104953, + "learning_rate": 2.5916657172501103e-06, + "loss": 0.558, "step": 2008 }, { - "epoch": 0.39, - "learning_rate": 1.4040970598647742e-05, - "loss": 0.7435, + "epoch": 0.7729896113889958, + "grad_norm": 1.0258187992644974, + "learning_rate": 2.583301092659872e-06, + "loss": 0.5466, "step": 2009 }, { - "epoch": 0.39, - "learning_rate": 1.4035269228251589e-05, - "loss": 0.912, + "epoch": 0.7733743747595229, + "grad_norm": 0.9767857320054621, + "learning_rate": 2.5749479857400383e-06, + "loss": 0.5749, "step": 2010 }, { - "epoch": 0.39, - "learning_rate": 1.4029566290603368e-05, - "loss": 0.8477, + "epoch": 0.77375913813005, + "grad_norm": 0.9638254188611254, + "learning_rate": 2.56660640946246e-06, + "loss": 0.5668, "step": 2011 }, { - "epoch": 0.39, - "learning_rate": 1.4023861787918031e-05, - "loss": 0.9565, + "epoch": 0.7741439015005771, + "grad_norm": 0.9274718866669591, + "learning_rate": 2.558276376781086e-06, + "loss": 0.5628, "step": 2012 }, { - "epoch": 0.39, - "learning_rate": 1.4018155722411144e-05, - "loss": 0.9349, + "epoch": 0.7745286648711043, + "grad_norm": 0.9106260894213841, + "learning_rate": 2.5499579006319365e-06, + "loss": 0.535, "step": 2013 }, { - "epoch": 0.39, - "learning_rate": 1.4012448096298874e-05, - "loss": 0.9243, + "epoch": 0.7749134282416313, + "grad_norm": 0.938960753014566, + "learning_rate": 2.5416509939330836e-06, + "loss": 0.5356, "step": 2014 }, { - "epoch": 0.39, - "learning_rate": 1.4006738911798001e-05, - "loss": 0.9245, + "epoch": 0.7752981916121585, + "grad_norm": 0.9869413029971487, + "learning_rate": 2.5333556695846384e-06, + "loss": 0.5835, "step": 2015 }, { - "epoch": 0.39, - "learning_rate": 1.40010281711259e-05, - "loss": 0.9342, + "epoch": 0.7756829549826857, + "grad_norm": 0.9752759511186649, + "learning_rate": 2.525071940468722e-06, + "loss": 0.5851, "step": 2016 }, { - "epoch": 0.39, - "learning_rate": 1.3995315876500565e-05, - "loss": 0.978, + "epoch": 0.7760677183532128, + "grad_norm": 0.9791804113711832, + "learning_rate": 2.5167998194494468e-06, + "loss": 0.622, "step": 2017 }, { - "epoch": 0.39, - "learning_rate": 1.3989602030140581e-05, - "loss": 0.9658, + "epoch": 0.7764524817237399, + "grad_norm": 0.9994467459003198, + "learning_rate": 2.5085393193729e-06, + "loss": 0.5628, "step": 2018 }, { - "epoch": 0.39, - "learning_rate": 1.398388663426514e-05, - "loss": 0.8993, + "epoch": 0.776837245094267, + "grad_norm": 0.9860328595758336, + "learning_rate": 2.5002904530671236e-06, + "loss": 0.6062, "step": 2019 }, { - "epoch": 0.39, - "learning_rate": 1.3978169691094037e-05, - "loss": 0.8854, + "epoch": 0.7772220084647942, + "grad_norm": 0.9990703014802412, + "learning_rate": 2.492053233342091e-06, + "loss": 0.5585, "step": 2020 }, { - "epoch": 0.39, - "learning_rate": 1.3972451202847665e-05, - "loss": 0.9364, + "epoch": 0.7776067718353212, + "grad_norm": 1.0246112676186077, + "learning_rate": 2.4838276729896884e-06, + "loss": 0.5873, "step": 2021 }, { - "epoch": 0.39, - "learning_rate": 1.3966731171747024e-05, - "loss": 0.8864, + "epoch": 0.7779915352058484, + "grad_norm": 1.0317417800133373, + "learning_rate": 2.4756137847837025e-06, + "loss": 0.5557, "step": 2022 }, { - "epoch": 0.39, - "learning_rate": 1.3961009600013702e-05, - "loss": 0.8993, + "epoch": 0.7783762985763756, + "grad_norm": 0.954406174213891, + "learning_rate": 2.467411581479786e-06, + "loss": 0.556, "step": 2023 }, { - "epoch": 0.39, - "learning_rate": 1.3955286489869894e-05, - "loss": 0.9038, + "epoch": 0.7787610619469026, + "grad_norm": 0.9409974102269181, + "learning_rate": 2.45922107581545e-06, + "loss": 0.5947, "step": 2024 }, { - "epoch": 0.39, - "learning_rate": 1.394956184353839e-05, - "loss": 0.7889, + "epoch": 0.7791458253174298, + "grad_norm": 0.9841056727376677, + "learning_rate": 2.4510422805100366e-06, + "loss": 0.5467, "step": 2025 }, { - "epoch": 0.39, - "learning_rate": 1.3943835663242577e-05, - "loss": 0.9382, + "epoch": 0.7795305886879569, + "grad_norm": 0.9343870415104214, + "learning_rate": 2.4428752082647044e-06, + "loss": 0.5372, "step": 2026 }, { - "epoch": 0.39, - "learning_rate": 1.3938107951206438e-05, - "loss": 0.889, + "epoch": 0.7799153520584841, + "grad_norm": 0.9486818384892137, + "learning_rate": 2.4347198717624054e-06, + "loss": 0.6116, "step": 2027 }, { - "epoch": 0.39, - "learning_rate": 1.3932378709654548e-05, - "loss": 0.9004, + "epoch": 0.7803001154290111, + "grad_norm": 0.9463968543834497, + "learning_rate": 2.426576283667873e-06, + "loss": 0.554, "step": 2028 }, { - "epoch": 0.39, - "learning_rate": 1.3926647940812081e-05, - "loss": 0.8911, + "epoch": 0.7806848787995383, + "grad_norm": 0.6100803394417954, + "learning_rate": 2.418444456627589e-06, + "loss": 0.6344, "step": 2029 }, { - "epoch": 0.39, - "learning_rate": 1.39209156469048e-05, - "loss": 0.8926, + "epoch": 0.7810696421700654, + "grad_norm": 0.9379356325244586, + "learning_rate": 2.4103244032697717e-06, + "loss": 0.5871, "step": 2030 }, { - "epoch": 0.39, - "learning_rate": 1.3915181830159061e-05, - "loss": 0.9911, + "epoch": 0.7814544055405925, + "grad_norm": 0.9267493829798656, + "learning_rate": 2.4022161362043574e-06, + "loss": 0.5141, "step": 2031 }, { - "epoch": 0.39, - "learning_rate": 1.3909446492801819e-05, - "loss": 0.9393, + "epoch": 0.7818391689111197, + "grad_norm": 0.9014609482353283, + "learning_rate": 2.3941196680229794e-06, + "loss": 0.5428, "step": 2032 }, { - "epoch": 0.39, - "learning_rate": 1.3903709637060605e-05, - "loss": 0.9641, + "epoch": 0.7822239322816468, + "grad_norm": 0.9940248041148749, + "learning_rate": 2.3860350112989473e-06, + "loss": 0.5876, "step": 2033 }, { - "epoch": 0.39, - "learning_rate": 1.3897971265163546e-05, - "loss": 0.877, + "epoch": 0.782608695652174, + "grad_norm": 0.922610512767674, + "learning_rate": 2.3779621785872252e-06, + "loss": 0.5404, "step": 2034 }, { - "epoch": 0.39, - "learning_rate": 1.3892231379339369e-05, - "loss": 0.9453, + "epoch": 0.782993459022701, + "grad_norm": 0.8826999054405643, + "learning_rate": 2.3699011824244234e-06, + "loss": 0.5626, "step": 2035 }, { - "epoch": 0.39, - "learning_rate": 1.3886489981817375e-05, - "loss": 0.9693, + "epoch": 0.7833782223932282, + "grad_norm": 1.0364076697552396, + "learning_rate": 2.3618520353287644e-06, + "loss": 0.5714, "step": 2036 }, { - "epoch": 0.39, - "learning_rate": 1.3880747074827454e-05, - "loss": 0.8964, + "epoch": 0.7837629857637552, + "grad_norm": 0.9595550321127287, + "learning_rate": 2.3538147498000695e-06, + "loss": 0.5892, "step": 2037 }, { - "epoch": 0.39, - "learning_rate": 1.3875002660600085e-05, - "loss": 0.8873, + "epoch": 0.7841477491342824, + "grad_norm": 1.0051702624393104, + "learning_rate": 2.3457893383197415e-06, + "loss": 0.5756, "step": 2038 }, { - "epoch": 0.39, - "learning_rate": 1.386925674136634e-05, - "loss": 0.7285, + "epoch": 0.7845325125048095, + "grad_norm": 0.9794237792881144, + "learning_rate": 2.3377758133507455e-06, + "loss": 0.5626, "step": 2039 }, { - "epoch": 0.39, - "learning_rate": 1.3863509319357857e-05, - "loss": 0.9614, + "epoch": 0.7849172758753367, + "grad_norm": 0.935328063884611, + "learning_rate": 2.32977418733758e-06, + "loss": 0.5437, "step": 2040 }, { - "epoch": 0.39, - "learning_rate": 1.3857760396806876e-05, - "loss": 0.6303, + "epoch": 0.7853020392458638, + "grad_norm": 0.5916154878982784, + "learning_rate": 2.321784472706279e-06, + "loss": 0.6481, "step": 2041 }, { - "epoch": 0.39, - "learning_rate": 1.3852009975946209e-05, - "loss": 0.9462, + "epoch": 0.7856868026163909, + "grad_norm": 0.6172230253324921, + "learning_rate": 2.3138066818643647e-06, + "loss": 0.6632, "step": 2042 }, { - "epoch": 0.39, - "learning_rate": 1.3846258059009252e-05, - "loss": 0.922, + "epoch": 0.7860715659869181, + "grad_norm": 0.9450103778311876, + "learning_rate": 2.30584082720085e-06, + "loss": 0.5338, "step": 2043 }, { - "epoch": 0.39, - "learning_rate": 1.384050464822999e-05, - "loss": 0.9561, + "epoch": 0.7864563293574451, + "grad_norm": 0.9108830775455122, + "learning_rate": 2.297886921086211e-06, + "loss": 0.5233, "step": 2044 }, { - "epoch": 0.39, - "learning_rate": 1.383474974584297e-05, - "loss": 0.9307, + "epoch": 0.7868410927279723, + "grad_norm": 0.9444795615453365, + "learning_rate": 2.2899449758723657e-06, + "loss": 0.5075, "step": 2045 }, { - "epoch": 0.39, - "learning_rate": 1.3828993354083342e-05, - "loss": 0.8742, + "epoch": 0.7872258560984994, + "grad_norm": 1.0433643845663307, + "learning_rate": 2.282015003892659e-06, + "loss": 0.5308, "step": 2046 }, { - "epoch": 0.39, - "learning_rate": 1.3823235475186816e-05, - "loss": 0.9343, + "epoch": 0.7876106194690266, + "grad_norm": 0.8988569345456071, + "learning_rate": 2.2740970174618405e-06, + "loss": 0.5639, "step": 2047 }, { - "epoch": 0.39, - "learning_rate": 1.3817476111389685e-05, - "loss": 0.8645, + "epoch": 0.7879953828395537, + "grad_norm": 0.9517599195408954, + "learning_rate": 2.2661910288760545e-06, + "loss": 0.5724, "step": 2048 }, { - "epoch": 0.39, - "learning_rate": 1.3811715264928824e-05, - "loss": 0.9728, + "epoch": 0.7883801462100808, + "grad_norm": 0.9130342093997097, + "learning_rate": 2.258297050412804e-06, + "loss": 0.5937, "step": 2049 }, { - "epoch": 0.39, - "learning_rate": 1.3805952938041674e-05, - "loss": 0.9298, + "epoch": 0.788764909580608, + "grad_norm": 0.5915682427271002, + "learning_rate": 2.2504150943309455e-06, + "loss": 0.651, "step": 2050 }, { - "epoch": 0.39, - "learning_rate": 1.3800189132966257e-05, - "loss": 0.9631, + "epoch": 0.789149672951135, + "grad_norm": 1.0651978363638512, + "learning_rate": 2.242545172870665e-06, + "loss": 0.5359, "step": 2051 }, { - "epoch": 0.39, - "learning_rate": 1.3794423851941174e-05, - "loss": 0.97, + "epoch": 0.7895344363216622, + "grad_norm": 0.9620341583966769, + "learning_rate": 2.2346872982534584e-06, + "loss": 0.5709, "step": 2052 }, { - "epoch": 0.39, - "learning_rate": 1.378865709720559e-05, - "loss": 0.752, + "epoch": 0.7899191996921893, + "grad_norm": 1.0468070543263068, + "learning_rate": 2.2268414826821117e-06, + "loss": 0.5819, "step": 2053 }, { - "epoch": 0.4, - "learning_rate": 1.3782888870999245e-05, - "loss": 0.9461, + "epoch": 0.7903039630627164, + "grad_norm": 1.0695349951590938, + "learning_rate": 2.2190077383406938e-06, + "loss": 0.5488, "step": 2054 }, { - "epoch": 0.4, - "learning_rate": 1.377711917556245e-05, - "loss": 0.918, + "epoch": 0.7906887264332435, + "grad_norm": 0.9153906203430241, + "learning_rate": 2.211186077394516e-06, + "loss": 0.5648, "step": 2055 }, { - "epoch": 0.4, - "learning_rate": 1.3771348013136096e-05, - "loss": 0.9158, + "epoch": 0.7910734898037707, + "grad_norm": 1.0290837263805221, + "learning_rate": 2.2033765119901294e-06, + "loss": 0.5843, "step": 2056 }, { - "epoch": 0.4, - "learning_rate": 1.3765575385961627e-05, - "loss": 0.9044, + "epoch": 0.7914582531742979, + "grad_norm": 0.9301685625321781, + "learning_rate": 2.1955790542553036e-06, + "loss": 0.5676, "step": 2057 }, { - "epoch": 0.4, - "learning_rate": 1.3759801296281072e-05, - "loss": 0.7792, + "epoch": 0.7918430165448249, + "grad_norm": 1.0008226050146725, + "learning_rate": 2.1877937162990015e-06, + "loss": 0.5471, "step": 2058 }, { - "epoch": 0.4, - "learning_rate": 1.3754025746337014e-05, - "loss": 0.9217, + "epoch": 0.7922277799153521, + "grad_norm": 1.0290958316541483, + "learning_rate": 2.180020510211367e-06, + "loss": 0.5991, "step": 2059 }, { - "epoch": 0.4, - "learning_rate": 1.3748248738372616e-05, - "loss": 0.9371, + "epoch": 0.7926125432858792, + "grad_norm": 1.0129279502529276, + "learning_rate": 2.172259448063704e-06, + "loss": 0.5347, "step": 2060 }, { - "epoch": 0.4, - "learning_rate": 1.3742470274631599e-05, - "loss": 0.9544, + "epoch": 0.7929973066564063, + "grad_norm": 0.6024225807474757, + "learning_rate": 2.1645105419084587e-06, + "loss": 0.6494, "step": 2061 }, { - "epoch": 0.4, - "learning_rate": 1.3736690357358253e-05, - "loss": 0.9973, + "epoch": 0.7933820700269334, + "grad_norm": 0.987884824010134, + "learning_rate": 2.1567738037791998e-06, + "loss": 0.5075, "step": 2062 }, { - "epoch": 0.4, - "learning_rate": 1.3730908988797427e-05, - "loss": 0.862, + "epoch": 0.7937668333974606, + "grad_norm": 1.027308159479585, + "learning_rate": 2.1490492456905964e-06, + "loss": 0.5693, "step": 2063 }, { - "epoch": 0.4, - "learning_rate": 1.3725126171194543e-05, - "loss": 0.9398, + "epoch": 0.7941515967679876, + "grad_norm": 0.96921159119536, + "learning_rate": 2.141336879638406e-06, + "loss": 0.5676, "step": 2064 }, { - "epoch": 0.4, - "learning_rate": 1.371934190679558e-05, - "loss": 0.9409, + "epoch": 0.7945363601385148, + "grad_norm": 0.8608224952758883, + "learning_rate": 2.133636717599451e-06, + "loss": 0.5016, "step": 2065 }, { - "epoch": 0.4, - "learning_rate": 1.3713556197847076e-05, - "loss": 0.9135, + "epoch": 0.794921123509042, + "grad_norm": 1.0511355296226712, + "learning_rate": 2.1259487715316e-06, + "loss": 0.5049, "step": 2066 }, { - "epoch": 0.4, - "learning_rate": 1.3707769046596136e-05, - "loss": 0.9284, + "epoch": 0.795305886879569, + "grad_norm": 1.0080719189349723, + "learning_rate": 2.118273053373757e-06, + "loss": 0.5911, "step": 2067 }, { - "epoch": 0.4, - "learning_rate": 1.3701980455290425e-05, - "loss": 0.922, + "epoch": 0.7956906502500962, + "grad_norm": 0.5729955689744612, + "learning_rate": 2.1106095750458332e-06, + "loss": 0.6426, "step": 2068 }, { - "epoch": 0.4, - "learning_rate": 1.3696190426178162e-05, - "loss": 0.8468, + "epoch": 0.7960754136206233, + "grad_norm": 1.0315041447330044, + "learning_rate": 2.1029583484487315e-06, + "loss": 0.5117, "step": 2069 }, { - "epoch": 0.4, - "learning_rate": 1.3690398961508128e-05, - "loss": 0.7747, + "epoch": 0.7964601769911505, + "grad_norm": 1.0030144359445297, + "learning_rate": 2.0953193854643274e-06, + "loss": 0.594, "step": 2070 }, { - "epoch": 0.4, - "learning_rate": 1.3684606063529662e-05, - "loss": 0.7515, + "epoch": 0.7968449403616775, + "grad_norm": 0.9312346591003462, + "learning_rate": 2.0876926979554545e-06, + "loss": 0.5532, "step": 2071 }, { - "epoch": 0.4, - "learning_rate": 1.3678811734492659e-05, - "loss": 0.9474, + "epoch": 0.7972297037322047, + "grad_norm": 1.0014255777277468, + "learning_rate": 2.080078297765884e-06, + "loss": 0.5962, "step": 2072 }, { - "epoch": 0.4, - "learning_rate": 1.367301597664757e-05, - "loss": 0.9271, + "epoch": 0.7976144671027319, + "grad_norm": 0.9972672996077846, + "learning_rate": 2.0724761967202987e-06, + "loss": 0.5775, "step": 2073 }, { - "epoch": 0.4, - "learning_rate": 1.36672187922454e-05, - "loss": 0.9254, + "epoch": 0.7979992304732589, + "grad_norm": 0.9365041497514985, + "learning_rate": 2.0648864066242937e-06, + "loss": 0.551, "step": 2074 }, { - "epoch": 0.4, - "learning_rate": 1.3661420183537705e-05, - "loss": 0.9177, + "epoch": 0.7983839938437861, + "grad_norm": 0.9995154569129021, + "learning_rate": 2.0573089392643362e-06, + "loss": 0.5539, "step": 2075 }, { - "epoch": 0.4, - "learning_rate": 1.3655620152776605e-05, - "loss": 0.847, + "epoch": 0.7987687572143132, + "grad_norm": 0.9108157712113623, + "learning_rate": 2.0497438064077603e-06, + "loss": 0.5322, "step": 2076 }, { - "epoch": 0.4, - "learning_rate": 1.364981870221476e-05, - "loss": 0.8707, + "epoch": 0.7991535205848403, + "grad_norm": 0.9176567807448317, + "learning_rate": 2.0421910198027452e-06, + "loss": 0.5467, "step": 2077 }, { - "epoch": 0.4, - "learning_rate": 1.364401583410539e-05, - "loss": 0.9247, + "epoch": 0.7995382839553674, + "grad_norm": 0.889250791918443, + "learning_rate": 2.0346505911782956e-06, + "loss": 0.518, "step": 2078 }, { - "epoch": 0.4, - "learning_rate": 1.3638211550702256e-05, - "loss": 0.9104, + "epoch": 0.7999230473258946, + "grad_norm": 0.967162389323328, + "learning_rate": 2.0271225322442255e-06, + "loss": 0.6261, "step": 2079 }, { - "epoch": 0.4, - "learning_rate": 1.363240585425968e-05, - "loss": 0.9171, + "epoch": 0.8003078106964217, + "grad_norm": 0.919366836508149, + "learning_rate": 2.019606854691145e-06, + "loss": 0.5524, "step": 2080 }, { - "epoch": 0.4, - "learning_rate": 1.362659874703253e-05, - "loss": 0.8033, + "epoch": 0.8006925740669488, + "grad_norm": 0.9795348319539168, + "learning_rate": 2.01210357019043e-06, + "loss": 0.5893, "step": 2081 }, { - "epoch": 0.4, - "learning_rate": 1.3620790231276213e-05, - "loss": 0.9293, + "epoch": 0.801077337437476, + "grad_norm": 1.0423102712807566, + "learning_rate": 2.004612690394212e-06, + "loss": 0.5773, "step": 2082 }, { - "epoch": 0.4, - "learning_rate": 1.3614980309246692e-05, - "loss": 0.8382, + "epoch": 0.8014621008080031, + "grad_norm": 0.93309032437687, + "learning_rate": 1.997134226935361e-06, + "loss": 0.5596, "step": 2083 }, { - "epoch": 0.4, - "learning_rate": 1.3609168983200474e-05, - "loss": 0.9239, + "epoch": 0.8018468641785302, + "grad_norm": 0.8572773081961972, + "learning_rate": 1.9896681914274616e-06, + "loss": 0.547, "step": 2084 }, { - "epoch": 0.4, - "learning_rate": 1.3603356255394613e-05, - "loss": 0.8899, + "epoch": 0.8022316275490573, + "grad_norm": 0.5986631887637884, + "learning_rate": 1.982214595464804e-06, + "loss": 0.6693, "step": 2085 }, { - "epoch": 0.4, - "learning_rate": 1.3597542128086702e-05, - "loss": 0.8867, + "epoch": 0.8026163909195845, + "grad_norm": 0.9603467775049968, + "learning_rate": 1.9747734506223525e-06, + "loss": 0.5599, "step": 2086 }, { - "epoch": 0.4, - "learning_rate": 1.3591726603534885e-05, - "loss": 0.9923, + "epoch": 0.8030011542901115, + "grad_norm": 0.8913437043037666, + "learning_rate": 1.967344768455747e-06, + "loss": 0.5039, "step": 2087 }, { - "epoch": 0.4, - "learning_rate": 1.3585909683997842e-05, - "loss": 0.9408, + "epoch": 0.8033859176606387, + "grad_norm": 0.9458008805670829, + "learning_rate": 1.9599285605012643e-06, + "loss": 0.5801, "step": 2088 }, { - "epoch": 0.4, - "learning_rate": 1.3580091371734798e-05, - "loss": 0.8948, + "epoch": 0.8037706810311658, + "grad_norm": 0.5860876680236536, + "learning_rate": 1.952524838275811e-06, + "loss": 0.6361, "step": 2089 }, { - "epoch": 0.4, - "learning_rate": 1.357427166900552e-05, - "loss": 0.8663, + "epoch": 0.804155444401693, + "grad_norm": 1.0094426977143176, + "learning_rate": 1.945133613276907e-06, + "loss": 0.5721, "step": 2090 }, { - "epoch": 0.4, - "learning_rate": 1.3568450578070309e-05, - "loss": 0.9496, + "epoch": 0.8045402077722201, + "grad_norm": 0.9930467282171695, + "learning_rate": 1.937754896982663e-06, + "loss": 0.5882, "step": 2091 }, { - "epoch": 0.4, - "learning_rate": 1.3562628101190015e-05, - "loss": 0.9304, + "epoch": 0.8049249711427472, + "grad_norm": 0.5709634010630278, + "learning_rate": 1.9303887008517618e-06, + "loss": 0.6017, "step": 2092 }, { - "epoch": 0.4, - "learning_rate": 1.3556804240626019e-05, - "loss": 0.9214, + "epoch": 0.8053097345132744, + "grad_norm": 1.0683362217447638, + "learning_rate": 1.923035036323452e-06, + "loss": 0.5378, "step": 2093 }, { - "epoch": 0.4, - "learning_rate": 1.3550978998640241e-05, - "loss": 0.8877, + "epoch": 0.8056944978838014, + "grad_norm": 0.9853391942854539, + "learning_rate": 1.9156939148175125e-06, + "loss": 0.5865, "step": 2094 }, { - "epoch": 0.4, - "learning_rate": 1.3545152377495136e-05, - "loss": 0.9345, + "epoch": 0.8060792612543286, + "grad_norm": 1.0004099325040996, + "learning_rate": 1.9083653477342467e-06, + "loss": 0.5073, "step": 2095 }, { - "epoch": 0.4, - "learning_rate": 1.3539324379453698e-05, - "loss": 0.9286, + "epoch": 0.8064640246248557, + "grad_norm": 0.9751491760808162, + "learning_rate": 1.9010493464544621e-06, + "loss": 0.5795, "step": 2096 }, { - "epoch": 0.4, - "learning_rate": 1.3533495006779455e-05, - "loss": 0.8668, + "epoch": 0.8068487879953828, + "grad_norm": 0.9568709554039, + "learning_rate": 1.8937459223394517e-06, + "loss": 0.5487, "step": 2097 }, { - "epoch": 0.4, - "learning_rate": 1.3527664261736471e-05, - "loss": 0.9156, + "epoch": 0.80723355136591, + "grad_norm": 0.9811393411484645, + "learning_rate": 1.8864550867309771e-06, + "loss": 0.5434, "step": 2098 }, { - "epoch": 0.4, - "learning_rate": 1.3521832146589335e-05, - "loss": 1.0116, + "epoch": 0.8076183147364371, + "grad_norm": 1.0334835908096833, + "learning_rate": 1.8791768509512487e-06, + "loss": 0.5907, "step": 2099 }, { - "epoch": 0.4, - "learning_rate": 1.3515998663603174e-05, - "loss": 0.9497, + "epoch": 0.8080030781069643, + "grad_norm": 1.0493208629029973, + "learning_rate": 1.871911226302917e-06, + "loss": 0.5554, "step": 2100 }, { - "epoch": 0.4, - "learning_rate": 1.3510163815043647e-05, - "loss": 0.8574, + "epoch": 0.8083878414774913, + "grad_norm": 1.0145652159680476, + "learning_rate": 1.8646582240690414e-06, + "loss": 0.5681, "step": 2101 }, { - "epoch": 0.4, - "learning_rate": 1.3504327603176943e-05, - "loss": 0.8984, + "epoch": 0.8087726048480185, + "grad_norm": 0.9824680905970112, + "learning_rate": 1.8574178555130818e-06, + "loss": 0.5332, "step": 2102 }, { - "epoch": 0.4, - "learning_rate": 1.3498490030269782e-05, - "loss": 0.8435, + "epoch": 0.8091573682185456, + "grad_norm": 1.083794075304399, + "learning_rate": 1.8501901318788773e-06, + "loss": 0.5894, "step": 2103 }, { - "epoch": 0.4, - "learning_rate": 1.3492651098589398e-05, - "loss": 1.0272, + "epoch": 0.8095421315890727, + "grad_norm": 0.8990936680023011, + "learning_rate": 1.8429750643906331e-06, + "loss": 0.5102, "step": 2104 }, { - "epoch": 0.4, - "learning_rate": 1.3486810810403578e-05, - "loss": 0.9906, + "epoch": 0.8099268949595998, + "grad_norm": 0.9357640991603103, + "learning_rate": 1.835772664252895e-06, + "loss": 0.5491, "step": 2105 }, { - "epoch": 0.41, - "learning_rate": 1.348096916798062e-05, - "loss": 0.8966, + "epoch": 0.810311658330127, + "grad_norm": 0.9398142271885763, + "learning_rate": 1.8285829426505453e-06, + "loss": 0.5643, "step": 2106 }, { - "epoch": 0.41, - "learning_rate": 1.3475126173589343e-05, - "loss": 0.8849, + "epoch": 0.8106964217006541, + "grad_norm": 0.9946018274778488, + "learning_rate": 1.8214059107487726e-06, + "loss": 0.5754, "step": 2107 }, { - "epoch": 0.41, - "learning_rate": 1.3469281829499107e-05, - "loss": 0.8442, + "epoch": 0.8110811850711812, + "grad_norm": 0.9422892363910317, + "learning_rate": 1.8142415796930568e-06, + "loss": 0.5962, "step": 2108 }, { - "epoch": 0.41, - "learning_rate": 1.3463436137979786e-05, - "loss": 0.9435, + "epoch": 0.8114659484417084, + "grad_norm": 0.9216945312328823, + "learning_rate": 1.8070899606091586e-06, + "loss": 0.5433, "step": 2109 }, { - "epoch": 0.41, - "learning_rate": 1.3457589101301776e-05, - "loss": 0.8999, + "epoch": 0.8118507118122354, + "grad_norm": 0.9970572354008035, + "learning_rate": 1.799951064603095e-06, + "loss": 0.5705, "step": 2110 }, { - "epoch": 0.41, - "learning_rate": 1.3451740721736005e-05, - "loss": 0.8741, + "epoch": 0.8122354751827626, + "grad_norm": 1.1258600706517392, + "learning_rate": 1.7928249027611255e-06, + "loss": 0.6218, "step": 2111 }, { - "epoch": 0.41, - "learning_rate": 1.3445891001553905e-05, - "loss": 0.8962, + "epoch": 0.8126202385532897, + "grad_norm": 0.8960096650998449, + "learning_rate": 1.7857114861497337e-06, + "loss": 0.4897, "step": 2112 }, { - "epoch": 0.41, - "learning_rate": 1.3440039943027452e-05, - "loss": 0.7493, + "epoch": 0.8130050019238169, + "grad_norm": 1.115566959364287, + "learning_rate": 1.7786108258156154e-06, + "loss": 0.6097, "step": 2113 }, { - "epoch": 0.41, - "learning_rate": 1.3434187548429126e-05, - "loss": 0.9838, + "epoch": 0.8133897652943439, + "grad_norm": 0.960005795824905, + "learning_rate": 1.7715229327856498e-06, + "loss": 0.5595, "step": 2114 }, { - "epoch": 0.41, - "learning_rate": 1.3428333820031922e-05, - "loss": 0.8776, + "epoch": 0.8137745286648711, + "grad_norm": 1.0011951118127274, + "learning_rate": 1.7644478180668945e-06, + "loss": 0.5399, "step": 2115 }, { - "epoch": 0.41, - "learning_rate": 1.3422478760109371e-05, - "loss": 0.7307, + "epoch": 0.8141592920353983, + "grad_norm": 0.6245208003031331, + "learning_rate": 1.7573854926465582e-06, + "loss": 0.6626, "step": 2116 }, { - "epoch": 0.41, - "learning_rate": 1.3416622370935507e-05, - "loss": 0.9273, + "epoch": 0.8145440554059253, + "grad_norm": 0.9501923388124672, + "learning_rate": 1.7503359674919929e-06, + "loss": 0.5691, "step": 2117 }, { - "epoch": 0.41, - "learning_rate": 1.3410764654784885e-05, - "loss": 0.8984, + "epoch": 0.8149288187764525, + "grad_norm": 0.993291661482522, + "learning_rate": 1.7432992535506687e-06, + "loss": 0.5873, "step": 2118 }, { - "epoch": 0.41, - "learning_rate": 1.3404905613932573e-05, - "loss": 0.8723, + "epoch": 0.8153135821469796, + "grad_norm": 0.994584826937844, + "learning_rate": 1.736275361750167e-06, + "loss": 0.562, "step": 2119 }, { - "epoch": 0.41, - "learning_rate": 1.3399045250654152e-05, - "loss": 0.8444, + "epoch": 0.8156983455175068, + "grad_norm": 0.9889580694041566, + "learning_rate": 1.7292643029981525e-06, + "loss": 0.5802, "step": 2120 }, { - "epoch": 0.41, - "learning_rate": 1.3393183567225724e-05, - "loss": 0.9007, + "epoch": 0.8160831088880338, + "grad_norm": 0.8850713616407012, + "learning_rate": 1.7222660881823594e-06, + "loss": 0.5514, "step": 2121 }, { - "epoch": 0.41, - "learning_rate": 1.3387320565923901e-05, - "loss": 0.9426, + "epoch": 0.816467872258561, + "grad_norm": 0.9965003289992054, + "learning_rate": 1.7152807281705809e-06, + "loss": 0.5795, "step": 2122 }, { - "epoch": 0.41, - "learning_rate": 1.33814562490258e-05, - "loss": 0.9008, + "epoch": 0.8168526356290882, + "grad_norm": 1.096822303941775, + "learning_rate": 1.708308233810644e-06, + "loss": 0.6205, "step": 2123 }, { - "epoch": 0.41, - "learning_rate": 1.3375590618809056e-05, - "loss": 0.8916, + "epoch": 0.8172373989996152, + "grad_norm": 1.0019682044245155, + "learning_rate": 1.701348615930397e-06, + "loss": 0.6064, "step": 2124 }, { - "epoch": 0.41, - "learning_rate": 1.3369723677551813e-05, - "loss": 0.8862, + "epoch": 0.8176221623701424, + "grad_norm": 0.9095140887581772, + "learning_rate": 1.6944018853376898e-06, + "loss": 0.5134, "step": 2125 }, { - "epoch": 0.41, - "learning_rate": 1.3363855427532724e-05, - "loss": 0.703, + "epoch": 0.8180069257406695, + "grad_norm": 0.986308278574375, + "learning_rate": 1.6874680528203657e-06, + "loss": 0.5657, "step": 2126 }, { - "epoch": 0.41, - "learning_rate": 1.3357985871030948e-05, - "loss": 0.9996, + "epoch": 0.8183916891111966, + "grad_norm": 0.9717321747474471, + "learning_rate": 1.6805471291462316e-06, + "loss": 0.6267, "step": 2127 }, { - "epoch": 0.41, - "learning_rate": 1.3352115010326155e-05, - "loss": 0.8499, + "epoch": 0.8187764524817237, + "grad_norm": 0.9509659033342811, + "learning_rate": 1.67363912506305e-06, + "loss": 0.5407, "step": 2128 }, { - "epoch": 0.41, - "learning_rate": 1.3346242847698516e-05, - "loss": 0.9548, + "epoch": 0.8191612158522509, + "grad_norm": 1.0298120225638305, + "learning_rate": 1.66674405129852e-06, + "loss": 0.6356, "step": 2129 }, { - "epoch": 0.41, - "learning_rate": 1.3340369385428713e-05, - "loss": 1.0064, + "epoch": 0.8195459792227779, + "grad_norm": 0.9133116922743225, + "learning_rate": 1.6598619185602616e-06, + "loss": 0.5426, "step": 2130 }, { - "epoch": 0.41, - "learning_rate": 1.3334494625797936e-05, - "loss": 0.9576, + "epoch": 0.8199307425933051, + "grad_norm": 1.0130733060294448, + "learning_rate": 1.6529927375357957e-06, + "loss": 0.5665, "step": 2131 }, { - "epoch": 0.41, - "learning_rate": 1.3328618571087867e-05, - "loss": 0.9078, + "epoch": 0.8203155059638323, + "grad_norm": 0.939877239049408, + "learning_rate": 1.6461365188925304e-06, + "loss": 0.5629, "step": 2132 }, { - "epoch": 0.41, - "learning_rate": 1.33227412235807e-05, - "loss": 0.9414, + "epoch": 0.8207002693343594, + "grad_norm": 0.9791491804150747, + "learning_rate": 1.6392932732777489e-06, + "loss": 0.5996, "step": 2133 }, { - "epoch": 0.41, - "learning_rate": 1.3316862585559132e-05, - "loss": 0.9019, + "epoch": 0.8210850327048865, + "grad_norm": 1.0322670261207054, + "learning_rate": 1.6324630113185835e-06, + "loss": 0.5493, "step": 2134 }, { - "epoch": 0.41, - "learning_rate": 1.3310982659306352e-05, - "loss": 0.9543, + "epoch": 0.8214697960754136, + "grad_norm": 0.9280725432623939, + "learning_rate": 1.625645743622003e-06, + "loss": 0.5095, "step": 2135 }, { - "epoch": 0.41, - "learning_rate": 1.3305101447106064e-05, - "loss": 0.9553, + "epoch": 0.8218545594459408, + "grad_norm": 0.9661658059732262, + "learning_rate": 1.6188414807747999e-06, + "loss": 0.5891, "step": 2136 }, { - "epoch": 0.41, - "learning_rate": 1.3299218951242456e-05, - "loss": 0.9406, + "epoch": 0.8222393228164678, + "grad_norm": 1.0170435530919042, + "learning_rate": 1.6120502333435695e-06, + "loss": 0.5818, "step": 2137 }, { - "epoch": 0.41, - "learning_rate": 1.3293335174000226e-05, - "loss": 1.0174, + "epoch": 0.822624086186995, + "grad_norm": 1.0357965006132126, + "learning_rate": 1.6052720118746923e-06, + "loss": 0.5797, "step": 2138 }, { - "epoch": 0.41, - "learning_rate": 1.328745011766456e-05, - "loss": 0.941, + "epoch": 0.8230088495575221, + "grad_norm": 0.9794517239050398, + "learning_rate": 1.5985068268943283e-06, + "loss": 0.5546, "step": 2139 }, { - "epoch": 0.41, - "learning_rate": 1.3281563784521154e-05, - "loss": 0.8651, + "epoch": 0.8233936129280492, + "grad_norm": 0.9510075743503839, + "learning_rate": 1.5917546889083834e-06, + "loss": 0.5644, "step": 2140 }, { - "epoch": 0.41, - "learning_rate": 1.3275676176856185e-05, - "loss": 0.9344, + "epoch": 0.8237783762985764, + "grad_norm": 0.8798286422294006, + "learning_rate": 1.5850156084025091e-06, + "loss": 0.4963, "step": 2141 }, { - "epoch": 0.41, - "learning_rate": 1.3269787296956333e-05, - "loss": 0.9284, + "epoch": 0.8241631396691035, + "grad_norm": 0.9520455635861426, + "learning_rate": 1.578289595842074e-06, + "loss": 0.5347, "step": 2142 }, { - "epoch": 0.41, - "learning_rate": 1.3263897147108778e-05, - "loss": 0.8978, + "epoch": 0.8245479030396307, + "grad_norm": 0.9627900019257614, + "learning_rate": 1.5715766616721584e-06, + "loss": 0.5483, "step": 2143 }, { - "epoch": 0.41, - "learning_rate": 1.3258005729601178e-05, - "loss": 0.8046, + "epoch": 0.8249326664101577, + "grad_norm": 0.9503614470042245, + "learning_rate": 1.5648768163175277e-06, + "loss": 0.5581, "step": 2144 }, { - "epoch": 0.41, - "learning_rate": 1.3252113046721692e-05, - "loss": 0.8998, + "epoch": 0.8253174297806849, + "grad_norm": 0.9580328468943692, + "learning_rate": 1.5581900701826226e-06, + "loss": 0.5569, "step": 2145 }, { - "epoch": 0.41, - "learning_rate": 1.3246219100758974e-05, - "loss": 0.9059, + "epoch": 0.825702193151212, + "grad_norm": 1.0088152388921652, + "learning_rate": 1.5515164336515465e-06, + "loss": 0.5968, "step": 2146 }, { - "epoch": 0.41, - "learning_rate": 1.3240323894002166e-05, - "loss": 0.9477, + "epoch": 0.8260869565217391, + "grad_norm": 0.8938112640358733, + "learning_rate": 1.5448559170880373e-06, + "loss": 0.5466, "step": 2147 }, { - "epoch": 0.41, - "learning_rate": 1.3234427428740895e-05, - "loss": 0.9907, + "epoch": 0.8264717198922663, + "grad_norm": 0.9750630648620698, + "learning_rate": 1.5382085308354633e-06, + "loss": 0.5941, "step": 2148 }, { - "epoch": 0.41, - "learning_rate": 1.3228529707265279e-05, - "loss": 0.9058, + "epoch": 0.8268564832627934, + "grad_norm": 1.0378981965026466, + "learning_rate": 1.5315742852167992e-06, + "loss": 0.5655, "step": 2149 }, { - "epoch": 0.41, - "learning_rate": 1.322263073186593e-05, - "loss": 0.9281, + "epoch": 0.8272412466333205, + "grad_norm": 0.9554321614049684, + "learning_rate": 1.5249531905346138e-06, + "loss": 0.5528, "step": 2150 }, { - "epoch": 0.41, - "learning_rate": 1.3216730504833938e-05, - "loss": 0.8486, + "epoch": 0.8276260100038476, + "grad_norm": 0.9269079819766781, + "learning_rate": 1.5183452570710522e-06, + "loss": 0.5146, "step": 2151 }, { - "epoch": 0.41, - "learning_rate": 1.3210829028460883e-05, - "loss": 0.9087, + "epoch": 0.8280107733743748, + "grad_norm": 1.0355858858437, + "learning_rate": 1.511750495087827e-06, + "loss": 0.576, "step": 2152 }, { - "epoch": 0.41, - "learning_rate": 1.3204926305038832e-05, - "loss": 0.906, + "epoch": 0.8283955367449019, + "grad_norm": 0.9745752038181736, + "learning_rate": 1.5051689148261895e-06, + "loss": 0.5925, "step": 2153 }, { - "epoch": 0.41, - "learning_rate": 1.3199022336860335e-05, - "loss": 0.896, + "epoch": 0.828780300115429, + "grad_norm": 1.0618885997237744, + "learning_rate": 1.4986005265069204e-06, + "loss": 0.5846, "step": 2154 }, { - "epoch": 0.41, - "learning_rate": 1.3193117126218425e-05, - "loss": 0.86, + "epoch": 0.8291650634859561, + "grad_norm": 0.9642559408672602, + "learning_rate": 1.4920453403303249e-06, + "loss": 0.5739, "step": 2155 }, { - "epoch": 0.41, - "learning_rate": 1.3187210675406617e-05, - "loss": 0.9219, + "epoch": 0.8295498268564833, + "grad_norm": 1.0120051097238525, + "learning_rate": 1.4855033664761898e-06, + "loss": 0.5504, "step": 2156 }, { - "epoch": 0.41, - "learning_rate": 1.318130298671891e-05, - "loss": 0.9629, + "epoch": 0.8299345902270104, + "grad_norm": 0.8953051454120082, + "learning_rate": 1.4789746151037942e-06, + "loss": 0.5863, "step": 2157 }, { - "epoch": 0.42, - "learning_rate": 1.3175394062449777e-05, - "loss": 0.8879, + "epoch": 0.8303193535975375, + "grad_norm": 0.977787294023818, + "learning_rate": 1.4724590963518803e-06, + "loss": 0.603, "step": 2158 }, { - "epoch": 0.42, - "learning_rate": 1.3169483904894185e-05, - "loss": 0.899, + "epoch": 0.8307041169680647, + "grad_norm": 0.6041465093070743, + "learning_rate": 1.4659568203386464e-06, + "loss": 0.6494, "step": 2159 }, { - "epoch": 0.42, - "learning_rate": 1.3163572516347565e-05, - "loss": 0.9175, + "epoch": 0.8310888803385917, + "grad_norm": 1.0134212354661822, + "learning_rate": 1.4594677971617178e-06, + "loss": 0.5832, "step": 2160 }, { - "epoch": 0.42, - "learning_rate": 1.3157659899105835e-05, - "loss": 0.9821, + "epoch": 0.8314736437091189, + "grad_norm": 0.7409409211097707, + "learning_rate": 1.452992036898142e-06, + "loss": 0.6323, "step": 2161 }, { - "epoch": 0.42, - "learning_rate": 1.315174605546538e-05, - "loss": 0.9374, + "epoch": 0.831858407079646, + "grad_norm": 0.5806785137355015, + "learning_rate": 1.446529549604373e-06, + "loss": 0.6402, "step": 2162 }, { - "epoch": 0.42, - "learning_rate": 1.3145830987723081e-05, - "loss": 0.9359, + "epoch": 0.8322431704501732, + "grad_norm": 0.9495294987566152, + "learning_rate": 1.4400803453162482e-06, + "loss": 0.5868, "step": 2163 }, { - "epoch": 0.42, - "learning_rate": 1.3139914698176273e-05, - "loss": 0.8638, + "epoch": 0.8326279338207002, + "grad_norm": 0.981613125105908, + "learning_rate": 1.4336444340489775e-06, + "loss": 0.5158, "step": 2164 }, { - "epoch": 0.42, - "learning_rate": 1.3133997189122777e-05, - "loss": 0.8607, + "epoch": 0.8330126971912274, + "grad_norm": 0.9702363717542078, + "learning_rate": 1.4272218257971327e-06, + "loss": 0.5944, "step": 2165 }, { - "epoch": 0.42, - "learning_rate": 1.3128078462860887e-05, - "loss": 0.8728, + "epoch": 0.8333974605617546, + "grad_norm": 0.9194340407540893, + "learning_rate": 1.4208125305346232e-06, + "loss": 0.553, "step": 2166 }, { - "epoch": 0.42, - "learning_rate": 1.3122158521689367e-05, - "loss": 0.8813, + "epoch": 0.8337822239322816, + "grad_norm": 0.975404418466905, + "learning_rate": 1.4144165582146819e-06, + "loss": 0.5862, "step": 2167 }, { - "epoch": 0.42, - "learning_rate": 1.3116237367907454e-05, - "loss": 0.9091, + "epoch": 0.8341669873028088, + "grad_norm": 0.9557176918246073, + "learning_rate": 1.40803391876986e-06, + "loss": 0.5821, "step": 2168 }, { - "epoch": 0.42, - "learning_rate": 1.3110315003814855e-05, - "loss": 0.9411, + "epoch": 0.8345517506733359, + "grad_norm": 0.9491457751621577, + "learning_rate": 1.4016646221119912e-06, + "loss": 0.5561, "step": 2169 }, { - "epoch": 0.42, - "learning_rate": 1.3104391431711748e-05, - "loss": 0.9192, + "epoch": 0.834936514043863, + "grad_norm": 0.9611028480133492, + "learning_rate": 1.395308678132199e-06, + "loss": 0.5603, "step": 2170 }, { - "epoch": 0.42, - "learning_rate": 1.309846665389878e-05, - "loss": 0.8266, + "epoch": 0.8353212774143901, + "grad_norm": 0.9986553190788584, + "learning_rate": 1.3889660967008656e-06, + "loss": 0.5356, "step": 2171 }, { - "epoch": 0.42, - "learning_rate": 1.309254067267707e-05, - "loss": 0.9061, + "epoch": 0.8357060407849173, + "grad_norm": 0.8917256749375093, + "learning_rate": 1.3826368876676278e-06, + "loss": 0.5433, "step": 2172 }, { - "epoch": 0.42, - "learning_rate": 1.3086613490348198e-05, - "loss": 0.944, + "epoch": 0.8360908041554445, + "grad_norm": 0.9107594485315667, + "learning_rate": 1.3763210608613497e-06, + "loss": 0.5723, "step": 2173 }, { - "epoch": 0.42, - "learning_rate": 1.3080685109214208e-05, - "loss": 0.9268, + "epoch": 0.8364755675259715, + "grad_norm": 1.0283035584044318, + "learning_rate": 1.370018626090116e-06, + "loss": 0.5984, "step": 2174 }, { - "epoch": 0.42, - "learning_rate": 1.3074755531577628e-05, - "loss": 0.9066, + "epoch": 0.8368603308964987, + "grad_norm": 0.5764404694246344, + "learning_rate": 1.3637295931412153e-06, + "loss": 0.6358, "step": 2175 }, { - "epoch": 0.42, - "learning_rate": 1.3068824759741428e-05, - "loss": 0.8557, + "epoch": 0.8372450942670258, + "grad_norm": 0.9347155879711636, + "learning_rate": 1.3574539717811231e-06, + "loss": 0.5717, "step": 2176 }, { - "epoch": 0.42, - "learning_rate": 1.306289279600905e-05, - "loss": 0.9234, + "epoch": 0.8376298576375529, + "grad_norm": 0.9082882016847791, + "learning_rate": 1.3511917717554846e-06, + "loss": 0.5224, "step": 2177 }, { - "epoch": 0.42, - "learning_rate": 1.3056959642684404e-05, - "loss": 1.0065, + "epoch": 0.83801462100808, + "grad_norm": 0.9932869537439577, + "learning_rate": 1.3449430027891096e-06, + "loss": 0.6169, "step": 2178 }, { - "epoch": 0.42, - "learning_rate": 1.305102530207186e-05, - "loss": 0.8734, + "epoch": 0.8383993843786072, + "grad_norm": 0.9991218154590285, + "learning_rate": 1.338707674585945e-06, + "loss": 0.5741, "step": 2179 }, { - "epoch": 0.42, - "learning_rate": 1.3045089776476246e-05, - "loss": 0.9654, + "epoch": 0.8387841477491342, + "grad_norm": 1.012242409030064, + "learning_rate": 1.332485796829065e-06, + "loss": 0.5644, "step": 2180 }, { - "epoch": 0.42, - "learning_rate": 1.3039153068202853e-05, - "loss": 0.8872, + "epoch": 0.8391689111196614, + "grad_norm": 1.0385279650522403, + "learning_rate": 1.3262773791806617e-06, + "loss": 0.5514, "step": 2181 }, { - "epoch": 0.42, - "learning_rate": 1.3033215179557424e-05, - "loss": 0.8845, + "epoch": 0.8395536744901886, + "grad_norm": 0.9544612535499719, + "learning_rate": 1.3200824312820137e-06, + "loss": 0.493, "step": 2182 }, { - "epoch": 0.42, - "learning_rate": 1.3027276112846172e-05, - "loss": 0.9184, + "epoch": 0.8399384378607156, + "grad_norm": 0.9395863133365806, + "learning_rate": 1.3139009627534927e-06, + "loss": 0.4967, "step": 2183 }, { - "epoch": 0.42, - "learning_rate": 1.3021335870375763e-05, - "loss": 0.9266, + "epoch": 0.8403232012312428, + "grad_norm": 0.976485680177933, + "learning_rate": 1.3077329831945295e-06, + "loss": 0.5623, "step": 2184 }, { - "epoch": 0.42, - "learning_rate": 1.3015394454453316e-05, - "loss": 0.9315, + "epoch": 0.8407079646017699, + "grad_norm": 0.9324303191677762, + "learning_rate": 1.3015785021836159e-06, + "loss": 0.5396, "step": 2185 }, { - "epoch": 0.42, - "learning_rate": 1.3009451867386411e-05, - "loss": 0.9458, + "epoch": 0.8410927279722971, + "grad_norm": 0.9688895127061932, + "learning_rate": 1.295437529278275e-06, + "loss": 0.5745, "step": 2186 }, { - "epoch": 0.42, - "learning_rate": 1.3003508111483077e-05, - "loss": 0.9366, + "epoch": 0.8414774913428241, + "grad_norm": 0.8903772651495003, + "learning_rate": 1.2893100740150522e-06, + "loss": 0.5494, "step": 2187 }, { - "epoch": 0.42, - "learning_rate": 1.29975631890518e-05, - "loss": 0.9353, + "epoch": 0.8418622547133513, + "grad_norm": 0.9565588355552361, + "learning_rate": 1.2831961459095088e-06, + "loss": 0.5693, "step": 2188 }, { - "epoch": 0.42, - "learning_rate": 1.2991617102401524e-05, - "loss": 0.9049, + "epoch": 0.8422470180838784, + "grad_norm": 0.929100119387537, + "learning_rate": 1.2770957544561868e-06, + "loss": 0.5415, "step": 2189 }, { - "epoch": 0.42, - "learning_rate": 1.2985669853841635e-05, - "loss": 0.9307, + "epoch": 0.8426317814544055, + "grad_norm": 0.9052175317299197, + "learning_rate": 1.2710089091286148e-06, + "loss": 0.5649, "step": 2190 }, { - "epoch": 0.42, - "learning_rate": 1.297972144568198e-05, - "loss": 0.8443, + "epoch": 0.8430165448249327, + "grad_norm": 0.9705787862989591, + "learning_rate": 1.2649356193792873e-06, + "loss": 0.5782, "step": 2191 }, { - "epoch": 0.42, - "learning_rate": 1.2973771880232853e-05, - "loss": 0.9282, + "epoch": 0.8434013081954598, + "grad_norm": 0.917276003758711, + "learning_rate": 1.2588758946396417e-06, + "loss": 0.5293, "step": 2192 }, { - "epoch": 0.42, - "learning_rate": 1.2967821159804994e-05, - "loss": 0.8294, + "epoch": 0.843786071565987, + "grad_norm": 0.9739083778682102, + "learning_rate": 1.2528297443200489e-06, + "loss": 0.5947, "step": 2193 }, { - "epoch": 0.42, - "learning_rate": 1.2961869286709594e-05, - "loss": 0.8587, + "epoch": 0.844170834936514, + "grad_norm": 1.1091465146604234, + "learning_rate": 1.246797177809812e-06, + "loss": 0.6635, "step": 2194 }, { - "epoch": 0.42, - "learning_rate": 1.295591626325829e-05, - "loss": 0.9341, + "epoch": 0.8445555983070412, + "grad_norm": 1.0863542274537445, + "learning_rate": 1.2407782044771222e-06, + "loss": 0.5928, "step": 2195 }, { - "epoch": 0.42, - "learning_rate": 1.2949962091763174e-05, - "loss": 0.9068, + "epoch": 0.8449403616775683, + "grad_norm": 1.0748935060096365, + "learning_rate": 1.23477283366907e-06, + "loss": 0.5668, "step": 2196 }, { - "epoch": 0.42, - "learning_rate": 1.2944006774536773e-05, - "loss": 0.957, + "epoch": 0.8453251250480954, + "grad_norm": 0.9242608898342136, + "learning_rate": 1.2287810747116224e-06, + "loss": 0.5597, "step": 2197 }, { - "epoch": 0.42, - "learning_rate": 1.2938050313892062e-05, - "loss": 0.7509, + "epoch": 0.8457098884186226, + "grad_norm": 0.9992381483694528, + "learning_rate": 1.2228029369096094e-06, + "loss": 0.5391, "step": 2198 }, { - "epoch": 0.42, - "learning_rate": 1.2932092712142468e-05, - "loss": 0.901, + "epoch": 0.8460946517891497, + "grad_norm": 0.9547397241358518, + "learning_rate": 1.216838429546704e-06, + "loss": 0.5925, "step": 2199 }, { - "epoch": 0.42, - "learning_rate": 1.292613397160185e-05, - "loss": 0.7486, + "epoch": 0.8464794151596768, + "grad_norm": 0.9146235721471933, + "learning_rate": 1.2108875618854122e-06, + "loss": 0.5565, "step": 2200 }, { - "epoch": 0.42, - "learning_rate": 1.2920174094584514e-05, - "loss": 0.8337, + "epoch": 0.8468641785302039, + "grad_norm": 0.9454578451614936, + "learning_rate": 1.204950343167065e-06, + "loss": 0.52, "step": 2201 }, { - "epoch": 0.42, - "learning_rate": 1.2914213083405211e-05, - "loss": 0.9609, + "epoch": 0.8472489419007311, + "grad_norm": 1.0026367433826298, + "learning_rate": 1.1990267826117874e-06, + "loss": 0.5782, "step": 2202 }, { - "epoch": 0.42, - "learning_rate": 1.2908250940379124e-05, - "loss": 0.888, + "epoch": 0.8476337052712581, + "grad_norm": 0.8531297364070357, + "learning_rate": 1.1931168894184974e-06, + "loss": 0.4957, "step": 2203 }, { - "epoch": 0.42, - "learning_rate": 1.2902287667821885e-05, - "loss": 0.8738, + "epoch": 0.8480184686417853, + "grad_norm": 0.8920305996432838, + "learning_rate": 1.187220672764897e-06, + "loss": 0.5241, "step": 2204 }, { - "epoch": 0.42, - "learning_rate": 1.289632326804956e-05, - "loss": 0.9598, + "epoch": 0.8484032320123124, + "grad_norm": 1.018276793306794, + "learning_rate": 1.1813381418074388e-06, + "loss": 0.5864, "step": 2205 }, { - "epoch": 0.42, - "learning_rate": 1.2890357743378649e-05, - "loss": 0.8961, + "epoch": 0.8487879953828396, + "grad_norm": 1.0836240131289108, + "learning_rate": 1.1754693056813272e-06, + "loss": 0.5373, "step": 2206 }, { - "epoch": 0.42, - "learning_rate": 1.2884391096126098e-05, - "loss": 0.7347, + "epoch": 0.8491727587533667, + "grad_norm": 1.0009070130410842, + "learning_rate": 1.169614173500494e-06, + "loss": 0.5547, "step": 2207 }, { - "epoch": 0.42, - "learning_rate": 1.2878423328609281e-05, - "loss": 0.8658, + "epoch": 0.8495575221238938, + "grad_norm": 0.9103602225219811, + "learning_rate": 1.1637727543576027e-06, + "loss": 0.5158, "step": 2208 }, { - "epoch": 0.42, - "learning_rate": 1.2872454443146015e-05, - "loss": 0.9021, + "epoch": 0.849942285494421, + "grad_norm": 0.9874822016900371, + "learning_rate": 1.1579450573240058e-06, + "loss": 0.5738, "step": 2209 }, { - "epoch": 0.43, - "learning_rate": 1.286648444205454e-05, - "loss": 0.8692, + "epoch": 0.850327048864948, + "grad_norm": 0.9122035577086606, + "learning_rate": 1.1521310914497518e-06, + "loss": 0.5671, "step": 2210 }, { - "epoch": 0.43, - "learning_rate": 1.2860513327653537e-05, - "loss": 0.7283, + "epoch": 0.8507118122354752, + "grad_norm": 0.9537650943162822, + "learning_rate": 1.1463308657635718e-06, + "loss": 0.5693, "step": 2211 }, { - "epoch": 0.43, - "learning_rate": 1.2854541102262119e-05, - "loss": 0.8901, + "epoch": 0.8510965756060023, + "grad_norm": 1.026312848174914, + "learning_rate": 1.140544389272853e-06, + "loss": 0.5926, "step": 2212 }, { - "epoch": 0.43, - "learning_rate": 1.284856776819983e-05, - "loss": 0.8433, + "epoch": 0.8514813389765294, + "grad_norm": 1.038178997815406, + "learning_rate": 1.1347716709636282e-06, + "loss": 0.5508, "step": 2213 }, { - "epoch": 0.43, - "learning_rate": 1.2842593327786649e-05, - "loss": 0.9383, + "epoch": 0.8518661023470565, + "grad_norm": 1.0208331313143728, + "learning_rate": 1.129012719800575e-06, + "loss": 0.568, "step": 2214 }, { - "epoch": 0.43, - "learning_rate": 1.2836617783342968e-05, - "loss": 0.8989, + "epoch": 0.8522508657175837, + "grad_norm": 0.9357099700175558, + "learning_rate": 1.1232675447269803e-06, + "loss": 0.5485, "step": 2215 }, { - "epoch": 0.43, - "learning_rate": 1.2830641137189628e-05, - "loss": 0.8463, + "epoch": 0.8526356290881109, + "grad_norm": 0.9125600811866593, + "learning_rate": 1.1175361546647413e-06, + "loss": 0.6208, "step": 2216 }, { - "epoch": 0.43, - "learning_rate": 1.282466339164789e-05, - "loss": 0.8837, + "epoch": 0.8530203924586379, + "grad_norm": 0.9837854262151443, + "learning_rate": 1.1118185585143536e-06, + "loss": 0.5861, "step": 2217 }, { - "epoch": 0.43, - "learning_rate": 1.2818684549039437e-05, - "loss": 0.9594, + "epoch": 0.8534051558291651, + "grad_norm": 0.5890740072299072, + "learning_rate": 1.1061147651548855e-06, + "loss": 0.6216, "step": 2218 }, { - "epoch": 0.43, - "learning_rate": 1.2812704611686386e-05, - "loss": 1.0263, + "epoch": 0.8537899191996922, + "grad_norm": 1.0558169332237624, + "learning_rate": 1.1004247834439697e-06, + "loss": 0.5826, "step": 2219 }, { - "epoch": 0.43, - "learning_rate": 1.2806723581911274e-05, - "loss": 0.9018, + "epoch": 0.8541746825702193, + "grad_norm": 0.9697118637335257, + "learning_rate": 1.0947486222177928e-06, + "loss": 0.5303, "step": 2220 }, { - "epoch": 0.43, - "learning_rate": 1.2800741462037065e-05, - "loss": 0.9113, + "epoch": 0.8545594459407464, + "grad_norm": 0.9493300615642107, + "learning_rate": 1.0890862902910849e-06, + "loss": 0.5625, "step": 2221 }, { - "epoch": 0.43, - "learning_rate": 1.2794758254387147e-05, - "loss": 0.7192, + "epoch": 0.8549442093112736, + "grad_norm": 1.0625156646072613, + "learning_rate": 1.0834377964570863e-06, + "loss": 0.5855, "step": 2222 }, { - "epoch": 0.43, - "learning_rate": 1.2788773961285323e-05, - "loss": 0.8778, + "epoch": 0.8553289726818007, + "grad_norm": 0.9707270614771907, + "learning_rate": 1.0778031494875574e-06, + "loss": 0.5464, "step": 2223 }, { - "epoch": 0.43, - "learning_rate": 1.2782788585055829e-05, - "loss": 0.9613, + "epoch": 0.8557137360523278, + "grad_norm": 0.9758017631565047, + "learning_rate": 1.072182358132755e-06, + "loss": 0.563, "step": 2224 }, { - "epoch": 0.43, - "learning_rate": 1.2776802128023317e-05, - "loss": 0.885, + "epoch": 0.856098499422855, + "grad_norm": 0.9621785657419004, + "learning_rate": 1.066575431121417e-06, + "loss": 0.5869, "step": 2225 }, { - "epoch": 0.43, - "learning_rate": 1.2770814592512853e-05, - "loss": 0.895, + "epoch": 0.856483262793382, + "grad_norm": 0.9798432427551076, + "learning_rate": 1.0609823771607487e-06, + "loss": 0.5419, "step": 2226 }, { - "epoch": 0.43, - "learning_rate": 1.2764825980849931e-05, - "loss": 0.9338, + "epoch": 0.8568680261639092, + "grad_norm": 0.9869867988639779, + "learning_rate": 1.055403204936416e-06, + "loss": 0.6435, "step": 2227 }, { - "epoch": 0.43, - "learning_rate": 1.2758836295360455e-05, - "loss": 0.9341, + "epoch": 0.8572527895344363, + "grad_norm": 1.0323902192633538, + "learning_rate": 1.0498379231125278e-06, + "loss": 0.5144, "step": 2228 }, { - "epoch": 0.43, - "learning_rate": 1.2752845538370752e-05, - "loss": 0.8305, + "epoch": 0.8576375529049635, + "grad_norm": 0.9530391775416417, + "learning_rate": 1.0442865403316117e-06, + "loss": 0.6139, "step": 2229 }, { - "epoch": 0.43, - "learning_rate": 1.2746853712207567e-05, - "loss": 0.8836, + "epoch": 0.8580223162754905, + "grad_norm": 0.9393924785707544, + "learning_rate": 1.0387490652146236e-06, + "loss": 0.5169, "step": 2230 }, { - "epoch": 0.43, - "learning_rate": 1.274086081919805e-05, - "loss": 0.9138, + "epoch": 0.8584070796460177, + "grad_norm": 0.9728569959055703, + "learning_rate": 1.0332255063609177e-06, + "loss": 0.5589, "step": 2231 }, { - "epoch": 0.43, - "learning_rate": 1.273486686166977e-05, - "loss": 0.9354, + "epoch": 0.8587918430165449, + "grad_norm": 0.8999147414472509, + "learning_rate": 1.027715872348234e-06, + "loss": 0.5048, "step": 2232 }, { - "epoch": 0.43, - "learning_rate": 1.2728871841950719e-05, - "loss": 0.9036, + "epoch": 0.8591766063870719, + "grad_norm": 0.991620530046533, + "learning_rate": 1.0222201717326885e-06, + "loss": 0.5832, "step": 2233 }, { - "epoch": 0.43, - "learning_rate": 1.2722875762369288e-05, - "loss": 0.927, + "epoch": 0.8595613697575991, + "grad_norm": 0.9598728952590616, + "learning_rate": 1.0167384130487667e-06, + "loss": 0.5541, "step": 2234 }, { - "epoch": 0.43, - "learning_rate": 1.2716878625254287e-05, - "loss": 0.8879, + "epoch": 0.8599461331281262, + "grad_norm": 0.9309325185859749, + "learning_rate": 1.0112706048092924e-06, + "loss": 0.5712, "step": 2235 }, { - "epoch": 0.43, - "learning_rate": 1.2710880432934934e-05, - "loss": 0.937, + "epoch": 0.8603308964986534, + "grad_norm": 0.8837426592872529, + "learning_rate": 1.00581675550543e-06, + "loss": 0.5158, "step": 2236 }, { - "epoch": 0.43, - "learning_rate": 1.270488118774086e-05, - "loss": 0.9222, + "epoch": 0.8607156598691804, + "grad_norm": 0.5687894947546177, + "learning_rate": 1.0003768736066722e-06, + "loss": 0.6451, "step": 2237 }, { - "epoch": 0.43, - "learning_rate": 1.26988808920021e-05, - "loss": 0.9433, + "epoch": 0.8611004232397076, + "grad_norm": 0.9818113403190611, + "learning_rate": 9.949509675608115e-07, + "loss": 0.5554, "step": 2238 }, { - "epoch": 0.43, - "learning_rate": 1.26928795480491e-05, - "loss": 0.9596, + "epoch": 0.8614851866102347, + "grad_norm": 1.0502535845635863, + "learning_rate": 9.895390457939414e-07, + "loss": 0.5878, "step": 2239 }, { - "epoch": 0.43, - "learning_rate": 1.2686877158212715e-05, - "loss": 0.9323, + "epoch": 0.8618699499807618, + "grad_norm": 0.9098533431592243, + "learning_rate": 9.84141116710442e-07, + "loss": 0.5474, "step": 2240 }, { - "epoch": 0.43, - "learning_rate": 1.26808737248242e-05, - "loss": 0.8855, + "epoch": 0.862254713351289, + "grad_norm": 0.9879900552287485, + "learning_rate": 9.787571886929604e-07, + "loss": 0.5544, "step": 2241 }, { - "epoch": 0.43, - "learning_rate": 1.2674869250215225e-05, - "loss": 0.8036, + "epoch": 0.8626394767218161, + "grad_norm": 1.0153506296174477, + "learning_rate": 9.733872701023938e-07, + "loss": 0.5671, "step": 2242 }, { - "epoch": 0.43, - "learning_rate": 1.2668863736717855e-05, - "loss": 0.8862, + "epoch": 0.8630242400923432, + "grad_norm": 1.0664682809196848, + "learning_rate": 9.680313692778976e-07, + "loss": 0.5749, "step": 2243 }, { - "epoch": 0.43, - "learning_rate": 1.2662857186664558e-05, - "loss": 0.8485, + "epoch": 0.8634090034628703, + "grad_norm": 0.938701861940246, + "learning_rate": 9.626894945368492e-07, + "loss": 0.5139, "step": 2244 }, { - "epoch": 0.43, - "learning_rate": 1.2656849602388222e-05, - "loss": 0.9041, + "epoch": 0.8637937668333975, + "grad_norm": 0.8848650560164751, + "learning_rate": 9.573616541748464e-07, + "loss": 0.4816, "step": 2245 }, { - "epoch": 0.43, - "learning_rate": 1.2650840986222111e-05, - "loss": 0.8774, + "epoch": 0.8641785302039245, + "grad_norm": 0.9914337663402641, + "learning_rate": 9.520478564656898e-07, + "loss": 0.5497, "step": 2246 }, { - "epoch": 0.43, - "learning_rate": 1.2644831340499906e-05, - "loss": 0.9134, + "epoch": 0.8645632935744517, + "grad_norm": 0.5714345199447183, + "learning_rate": 9.467481096613829e-07, + "loss": 0.6264, "step": 2247 }, { - "epoch": 0.43, - "learning_rate": 1.2638820667555685e-05, - "loss": 0.9561, + "epoch": 0.8649480569449788, + "grad_norm": 1.0123695678011442, + "learning_rate": 9.414624219920953e-07, + "loss": 0.6027, "step": 2248 }, { - "epoch": 0.43, - "learning_rate": 1.2632808969723927e-05, - "loss": 0.8577, + "epoch": 0.865332820315506, + "grad_norm": 0.9630366794814329, + "learning_rate": 9.361908016661703e-07, + "loss": 0.5684, "step": 2249 }, { - "epoch": 0.43, - "learning_rate": 1.26267962493395e-05, - "loss": 0.9668, + "epoch": 0.8657175836860331, + "grad_norm": 0.9599897403576285, + "learning_rate": 9.309332568701079e-07, + "loss": 0.573, "step": 2250 }, { - "epoch": 0.43, - "learning_rate": 1.2620782508737678e-05, - "loss": 0.9017, + "epoch": 0.8661023470565602, + "grad_norm": 1.0228152210059642, + "learning_rate": 9.256897957685463e-07, + "loss": 0.5974, "step": 2251 }, { - "epoch": 0.43, - "learning_rate": 1.2614767750254129e-05, - "loss": 0.7278, + "epoch": 0.8664871104270874, + "grad_norm": 0.9206150308534283, + "learning_rate": 9.204604265042505e-07, + "loss": 0.5332, "step": 2252 }, { - "epoch": 0.43, - "learning_rate": 1.2608751976224916e-05, - "loss": 0.8742, + "epoch": 0.8668718737976144, + "grad_norm": 1.0005067456569876, + "learning_rate": 9.15245157198108e-07, + "loss": 0.572, "step": 2253 }, { - "epoch": 0.43, - "learning_rate": 1.2602735188986498e-05, - "loss": 0.9517, + "epoch": 0.8672566371681416, + "grad_norm": 0.9021470267169375, + "learning_rate": 9.10043995949108e-07, + "loss": 0.5455, "step": 2254 }, { - "epoch": 0.43, - "learning_rate": 1.2596717390875721e-05, - "loss": 0.8949, + "epoch": 0.8676414005386687, + "grad_norm": 1.0675081913085267, + "learning_rate": 9.04856950834323e-07, + "loss": 0.5553, "step": 2255 }, { - "epoch": 0.43, - "learning_rate": 1.2590698584229834e-05, - "loss": 0.9307, + "epoch": 0.8680261639091958, + "grad_norm": 0.5961141100103023, + "learning_rate": 8.996840299089149e-07, + "loss": 0.6638, "step": 2256 }, { - "epoch": 0.43, - "learning_rate": 1.2584678771386467e-05, - "loss": 0.9678, + "epoch": 0.868410927279723, + "grad_norm": 0.8577973432337829, + "learning_rate": 8.945252412061056e-07, + "loss": 0.5177, "step": 2257 }, { - "epoch": 0.43, - "learning_rate": 1.2578657954683651e-05, - "loss": 0.8795, + "epoch": 0.8687956906502501, + "grad_norm": 1.0432575128527706, + "learning_rate": 8.893805927371724e-07, + "loss": 0.5737, "step": 2258 }, { - "epoch": 0.43, - "learning_rate": 1.2572636136459799e-05, - "loss": 0.9787, + "epoch": 0.8691804540207773, + "grad_norm": 0.9346896707634492, + "learning_rate": 8.842500924914299e-07, + "loss": 0.5782, "step": 2259 }, { - "epoch": 0.43, - "learning_rate": 1.2566613319053713e-05, - "loss": 0.9169, + "epoch": 0.8695652173913043, + "grad_norm": 0.9598682514628613, + "learning_rate": 8.791337484362305e-07, + "loss": 0.562, "step": 2260 }, { - "epoch": 0.43, - "learning_rate": 1.2560589504804592e-05, - "loss": 0.9499, + "epoch": 0.8699499807618315, + "grad_norm": 1.0045440170674582, + "learning_rate": 8.740315685169364e-07, + "loss": 0.5879, "step": 2261 }, { - "epoch": 0.44, - "learning_rate": 1.2554564696052011e-05, - "loss": 0.9495, + "epoch": 0.8703347441323586, + "grad_norm": 0.9822556732326482, + "learning_rate": 8.689435606569086e-07, + "loss": 0.5349, "step": 2262 }, { - "epoch": 0.44, - "learning_rate": 1.2548538895135942e-05, - "loss": 0.7037, + "epoch": 0.8707195075028857, + "grad_norm": 0.895417205363757, + "learning_rate": 8.638697327575108e-07, + "loss": 0.5378, "step": 2263 }, { - "epoch": 0.44, - "learning_rate": 1.254251210439673e-05, - "loss": 0.8887, + "epoch": 0.8711042708734128, + "grad_norm": 1.0308415859202689, + "learning_rate": 8.588100926980802e-07, + "loss": 0.5105, "step": 2264 }, { - "epoch": 0.44, - "learning_rate": 1.2536484326175114e-05, - "loss": 0.9389, + "epoch": 0.87148903424394, + "grad_norm": 0.9347919741708954, + "learning_rate": 8.537646483359185e-07, + "loss": 0.54, "step": 2265 }, { - "epoch": 0.44, - "learning_rate": 1.2530455562812214e-05, - "loss": 0.8137, + "epoch": 0.8718737976144671, + "grad_norm": 0.936566799616806, + "learning_rate": 8.487334075062914e-07, + "loss": 0.5902, "step": 2266 }, { - "epoch": 0.44, - "learning_rate": 1.252442581664953e-05, - "loss": 0.9112, + "epoch": 0.8722585609849942, + "grad_norm": 0.9576282869147945, + "learning_rate": 8.437163780224011e-07, + "loss": 0.5803, "step": 2267 }, { - "epoch": 0.44, - "learning_rate": 1.2518395090028952e-05, - "loss": 1.0166, + "epoch": 0.8726433243555214, + "grad_norm": 0.975637131219978, + "learning_rate": 8.387135676753755e-07, + "loss": 0.5634, "step": 2268 }, { - "epoch": 0.44, - "learning_rate": 1.2512363385292739e-05, - "loss": 0.903, + "epoch": 0.8730280877260485, + "grad_norm": 1.00417556716445, + "learning_rate": 8.337249842342721e-07, + "loss": 0.5796, "step": 2269 }, { - "epoch": 0.44, - "learning_rate": 1.2506330704783533e-05, - "loss": 0.9039, + "epoch": 0.8734128510965756, + "grad_norm": 0.9502445674927483, + "learning_rate": 8.287506354460484e-07, + "loss": 0.5665, "step": 2270 }, { - "epoch": 0.44, - "learning_rate": 1.2500297050844367e-05, - "loss": 0.9523, + "epoch": 0.8737976144671027, + "grad_norm": 1.0408352510269336, + "learning_rate": 8.237905290355563e-07, + "loss": 0.5561, "step": 2271 }, { - "epoch": 0.44, - "learning_rate": 1.2494262425818637e-05, - "loss": 0.9409, + "epoch": 0.8741823778376299, + "grad_norm": 0.9143209039633214, + "learning_rate": 8.188446727055311e-07, + "loss": 0.5138, "step": 2272 }, { - "epoch": 0.44, - "learning_rate": 1.2488226832050116e-05, - "loss": 0.974, + "epoch": 0.8745671412081569, + "grad_norm": 1.0161309503102538, + "learning_rate": 8.139130741365819e-07, + "loss": 0.577, "step": 2273 }, { - "epoch": 0.44, - "learning_rate": 1.2482190271882973e-05, - "loss": 0.8782, + "epoch": 0.8749519045786841, + "grad_norm": 0.9389371514159602, + "learning_rate": 8.08995740987173e-07, + "loss": 0.562, "step": 2274 }, { - "epoch": 0.44, - "learning_rate": 1.2476152747661727e-05, - "loss": 0.8908, + "epoch": 0.8753366679492113, + "grad_norm": 0.8311807149760604, + "learning_rate": 8.040926808936112e-07, + "loss": 0.5029, "step": 2275 }, { - "epoch": 0.44, - "learning_rate": 1.2470114261731288e-05, - "loss": 0.8534, + "epoch": 0.8757214313197383, + "grad_norm": 0.9765768016876651, + "learning_rate": 7.99203901470047e-07, + "loss": 0.5512, "step": 2276 }, { - "epoch": 0.44, - "learning_rate": 1.246407481643693e-05, - "loss": 0.8787, + "epoch": 0.8761061946902655, + "grad_norm": 0.9593338597208799, + "learning_rate": 7.943294103084487e-07, + "loss": 0.5758, "step": 2277 }, { - "epoch": 0.44, - "learning_rate": 1.245803441412431e-05, - "loss": 0.897, + "epoch": 0.8764909580607926, + "grad_norm": 0.9403528835964469, + "learning_rate": 7.894692149785954e-07, + "loss": 0.565, "step": 2278 }, { - "epoch": 0.44, - "learning_rate": 1.2451993057139445e-05, - "loss": 0.9427, + "epoch": 0.8768757214313198, + "grad_norm": 1.0761766726193975, + "learning_rate": 7.846233230280698e-07, + "loss": 0.5925, "step": 2279 }, { - "epoch": 0.44, - "learning_rate": 1.2445950747828732e-05, - "loss": 0.8793, + "epoch": 0.8772604848018468, + "grad_norm": 1.046820429540986, + "learning_rate": 7.797917419822377e-07, + "loss": 0.5734, "step": 2280 }, { - "epoch": 0.44, - "learning_rate": 1.2439907488538934e-05, - "loss": 0.9514, + "epoch": 0.877645248172374, + "grad_norm": 0.9300275550830976, + "learning_rate": 7.749744793442448e-07, + "loss": 0.5107, "step": 2281 }, { - "epoch": 0.44, - "learning_rate": 1.243386328161718e-05, - "loss": 1.0151, + "epoch": 0.8780300115429012, + "grad_norm": 0.5762627597880069, + "learning_rate": 7.701715425949952e-07, + "loss": 0.6473, "step": 2282 }, { - "epoch": 0.44, - "learning_rate": 1.2427818129410975e-05, - "loss": 0.9421, + "epoch": 0.8784147749134282, + "grad_norm": 0.9948907052044671, + "learning_rate": 7.653829391931533e-07, + "loss": 0.567, "step": 2283 }, { - "epoch": 0.44, - "learning_rate": 1.2421772034268187e-05, - "loss": 0.8378, + "epoch": 0.8787995382839554, + "grad_norm": 0.9594776661927084, + "learning_rate": 7.606086765751209e-07, + "loss": 0.5563, "step": 2284 }, { - "epoch": 0.44, - "learning_rate": 1.2415724998537042e-05, - "loss": 0.9112, + "epoch": 0.8791843016544825, + "grad_norm": 0.9887394311892638, + "learning_rate": 7.55848762155027e-07, + "loss": 0.5769, "step": 2285 }, { - "epoch": 0.44, - "learning_rate": 1.2409677024566145e-05, - "loss": 1.0162, + "epoch": 0.8795690650250096, + "grad_norm": 1.015859265687764, + "learning_rate": 7.511032033247256e-07, + "loss": 0.6096, "step": 2286 }, { - "epoch": 0.44, - "learning_rate": 1.240362811470446e-05, - "loss": 0.9503, + "epoch": 0.8799538283955367, + "grad_norm": 0.8323444153263105, + "learning_rate": 7.463720074537728e-07, + "loss": 0.4599, "step": 2287 }, { - "epoch": 0.44, - "learning_rate": 1.2397578271301312e-05, - "loss": 0.932, + "epoch": 0.8803385917660639, + "grad_norm": 0.9484904454686517, + "learning_rate": 7.416551818894158e-07, + "loss": 0.5096, "step": 2288 }, { - "epoch": 0.44, - "learning_rate": 1.2391527496706389e-05, - "loss": 0.9052, + "epoch": 0.880723355136591, + "grad_norm": 0.5983013291320777, + "learning_rate": 7.369527339565951e-07, + "loss": 0.6187, "step": 2289 }, { - "epoch": 0.44, - "learning_rate": 1.2385475793269744e-05, - "loss": 0.8869, + "epoch": 0.8811081185071181, + "grad_norm": 0.9492228620037688, + "learning_rate": 7.322646709579173e-07, + "loss": 0.5507, "step": 2290 }, { - "epoch": 0.44, - "learning_rate": 1.2379423163341791e-05, - "loss": 0.9793, + "epoch": 0.8814928818776453, + "grad_norm": 0.5938928408757417, + "learning_rate": 7.275910001736497e-07, + "loss": 0.6471, "step": 2291 }, { - "epoch": 0.44, - "learning_rate": 1.2373369609273299e-05, - "loss": 0.8295, + "epoch": 0.8818776452481724, + "grad_norm": 0.9273000158055648, + "learning_rate": 7.229317288617144e-07, + "loss": 0.5436, "step": 2292 }, { - "epoch": 0.44, - "learning_rate": 1.2367315133415396e-05, - "loss": 0.9244, + "epoch": 0.8822624086186995, + "grad_norm": 0.9920779242140841, + "learning_rate": 7.182868642576679e-07, + "loss": 0.5624, "step": 2293 }, { - "epoch": 0.44, - "learning_rate": 1.2361259738119575e-05, - "loss": 0.844, + "epoch": 0.8826471719892266, + "grad_norm": 0.5847043143379148, + "learning_rate": 7.13656413574696e-07, + "loss": 0.6375, "step": 2294 }, { - "epoch": 0.44, - "learning_rate": 1.2355203425737683e-05, - "loss": 0.8237, + "epoch": 0.8830319353597538, + "grad_norm": 0.8527566328305394, + "learning_rate": 7.090403840035942e-07, + "loss": 0.4619, "step": 2295 }, { - "epoch": 0.44, - "learning_rate": 1.2349146198621917e-05, - "loss": 0.8952, + "epoch": 0.8834166987302808, + "grad_norm": 0.9470897084534993, + "learning_rate": 7.044387827127752e-07, + "loss": 0.5887, "step": 2296 }, { - "epoch": 0.44, - "learning_rate": 1.2343088059124839e-05, - "loss": 0.8346, + "epoch": 0.883801462100808, + "grad_norm": 0.9628683579373889, + "learning_rate": 6.99851616848235e-07, + "loss": 0.5787, "step": 2297 }, { - "epoch": 0.44, - "learning_rate": 1.2337029009599357e-05, - "loss": 0.7106, + "epoch": 0.8841862254713351, + "grad_norm": 0.9342281009238881, + "learning_rate": 6.952788935335541e-07, + "loss": 0.5343, "step": 2298 }, { - "epoch": 0.44, - "learning_rate": 1.2330969052398735e-05, - "loss": 0.998, + "epoch": 0.8845709888418622, + "grad_norm": 0.9694640038959959, + "learning_rate": 6.907206198698912e-07, + "loss": 0.5316, "step": 2299 }, { - "epoch": 0.44, - "learning_rate": 1.2324908189876597e-05, - "loss": 0.853, + "epoch": 0.8849557522123894, + "grad_norm": 0.8762076150169696, + "learning_rate": 6.861768029359595e-07, + "loss": 0.5441, "step": 2300 }, { - "epoch": 0.44, - "learning_rate": 1.2318846424386907e-05, - "loss": 0.9636, + "epoch": 0.8853405155829165, + "grad_norm": 0.9865479003630175, + "learning_rate": 6.816474497880177e-07, + "loss": 0.5371, "step": 2301 }, { - "epoch": 0.44, - "learning_rate": 1.2312783758283981e-05, - "loss": 0.9826, + "epoch": 0.8857252789534437, + "grad_norm": 1.01785546462695, + "learning_rate": 6.77132567459875e-07, + "loss": 0.5642, "step": 2302 }, { - "epoch": 0.44, - "learning_rate": 1.23067201939225e-05, - "loss": 1.0428, + "epoch": 0.8861100423239707, + "grad_norm": 1.0032179423889367, + "learning_rate": 6.726321629628585e-07, + "loss": 0.5598, "step": 2303 }, { - "epoch": 0.44, - "learning_rate": 1.2300655733657475e-05, - "loss": 0.8669, + "epoch": 0.8864948056944979, + "grad_norm": 0.9946687786994566, + "learning_rate": 6.681462432858154e-07, + "loss": 0.5393, "step": 2304 }, { - "epoch": 0.44, - "learning_rate": 1.2294590379844268e-05, - "loss": 0.9146, + "epoch": 0.886879569065025, + "grad_norm": 0.5825856986902961, + "learning_rate": 6.636748153951e-07, + "loss": 0.6489, "step": 2305 }, { - "epoch": 0.44, - "learning_rate": 1.2288524134838602e-05, - "loss": 0.9362, + "epoch": 0.8872643324355521, + "grad_norm": 0.9068152401332193, + "learning_rate": 6.592178862345622e-07, + "loss": 0.53, "step": 2306 }, { - "epoch": 0.44, - "learning_rate": 1.2282457000996533e-05, - "loss": 0.9265, + "epoch": 0.8876490958060793, + "grad_norm": 1.0175972688861668, + "learning_rate": 6.547754627255332e-07, + "loss": 0.5523, "step": 2307 }, { - "epoch": 0.44, - "learning_rate": 1.2276388980674465e-05, - "loss": 0.8547, + "epoch": 0.8880338591766064, + "grad_norm": 0.932663740434546, + "learning_rate": 6.503475517668168e-07, + "loss": 0.5692, "step": 2308 }, { - "epoch": 0.44, - "learning_rate": 1.227032007622915e-05, - "loss": 0.883, + "epoch": 0.8884186225471336, + "grad_norm": 0.9239282410359164, + "learning_rate": 6.459341602346858e-07, + "loss": 0.5082, "step": 2309 }, { - "epoch": 0.44, - "learning_rate": 1.2264250290017675e-05, - "loss": 0.8593, + "epoch": 0.8888033859176606, + "grad_norm": 0.9269856681837106, + "learning_rate": 6.415352949828601e-07, + "loss": 0.5263, "step": 2310 }, { - "epoch": 0.44, - "learning_rate": 1.2258179624397477e-05, - "loss": 0.8681, + "epoch": 0.8891881492881878, + "grad_norm": 0.937778026235724, + "learning_rate": 6.371509628425021e-07, + "loss": 0.5357, "step": 2311 }, { - "epoch": 0.44, - "learning_rate": 1.2252108081726337e-05, - "loss": 0.8957, + "epoch": 0.8895729126587149, + "grad_norm": 0.8335352988578929, + "learning_rate": 6.327811706222097e-07, + "loss": 0.5223, "step": 2312 }, { - "epoch": 0.45, - "learning_rate": 1.224603566436237e-05, - "loss": 0.9496, + "epoch": 0.889957676029242, + "grad_norm": 0.911403679258441, + "learning_rate": 6.284259251079939e-07, + "loss": 0.5464, "step": 2313 }, { - "epoch": 0.45, - "learning_rate": 1.2239962374664029e-05, - "loss": 0.9229, + "epoch": 0.8903424393997691, + "grad_norm": 1.0530487052532278, + "learning_rate": 6.240852330632796e-07, + "loss": 0.5855, "step": 2314 }, { - "epoch": 0.45, - "learning_rate": 1.2233888214990113e-05, - "loss": 0.7323, + "epoch": 0.8907272027702963, + "grad_norm": 0.9253511431770243, + "learning_rate": 6.197591012288918e-07, + "loss": 0.5282, "step": 2315 }, { - "epoch": 0.45, - "learning_rate": 1.2227813187699757e-05, - "loss": 0.9467, + "epoch": 0.8911119661408234, + "grad_norm": 0.9263059144784703, + "learning_rate": 6.154475363230417e-07, + "loss": 0.5529, "step": 2316 }, { - "epoch": 0.45, - "learning_rate": 1.222173729515243e-05, - "loss": 0.9314, + "epoch": 0.8914967295113505, + "grad_norm": 1.0139268403323378, + "learning_rate": 6.111505450413202e-07, + "loss": 0.6033, "step": 2317 }, { - "epoch": 0.45, - "learning_rate": 1.2215660539707936e-05, - "loss": 0.9781, + "epoch": 0.8918814928818777, + "grad_norm": 0.9392318259893848, + "learning_rate": 6.068681340566896e-07, + "loss": 0.5513, "step": 2318 }, { - "epoch": 0.45, - "learning_rate": 1.2209582923726424e-05, - "loss": 0.9048, + "epoch": 0.8922662562524047, + "grad_norm": 0.8614339297053727, + "learning_rate": 6.026003100194633e-07, + "loss": 0.5015, "step": 2319 }, { - "epoch": 0.45, - "learning_rate": 1.2203504449568361e-05, - "loss": 0.9242, + "epoch": 0.8926510196229319, + "grad_norm": 0.9403060191585343, + "learning_rate": 5.983470795573088e-07, + "loss": 0.5271, "step": 2320 }, { - "epoch": 0.45, - "learning_rate": 1.2197425119594563e-05, - "loss": 0.9238, + "epoch": 0.893035782993459, + "grad_norm": 0.9232470284904319, + "learning_rate": 5.941084492752236e-07, + "loss": 0.5775, "step": 2321 }, { - "epoch": 0.45, - "learning_rate": 1.219134493616617e-05, - "loss": 0.8912, + "epoch": 0.8934205463639862, + "grad_norm": 0.9426172564350922, + "learning_rate": 5.898844257555392e-07, + "loss": 0.536, "step": 2322 }, { - "epoch": 0.45, - "learning_rate": 1.2185263901644653e-05, - "loss": 0.9071, + "epoch": 0.8938053097345132, + "grad_norm": 0.9305356858648866, + "learning_rate": 5.856750155578983e-07, + "loss": 0.5415, "step": 2323 }, { - "epoch": 0.45, - "learning_rate": 1.217918201839182e-05, - "loss": 0.9655, + "epoch": 0.8941900731050404, + "grad_norm": 0.9446972965488135, + "learning_rate": 5.81480225219252e-07, + "loss": 0.5816, "step": 2324 }, { - "epoch": 0.45, - "learning_rate": 1.2173099288769799e-05, - "loss": 0.9086, + "epoch": 0.8945748364755676, + "grad_norm": 0.8721978150223165, + "learning_rate": 5.773000612538505e-07, + "loss": 0.5328, "step": 2325 }, { - "epoch": 0.45, - "learning_rate": 1.2167015715141057e-05, - "loss": 0.9427, + "epoch": 0.8949595998460946, + "grad_norm": 1.0315203823601828, + "learning_rate": 5.731345301532265e-07, + "loss": 0.5661, "step": 2326 }, { - "epoch": 0.45, - "learning_rate": 1.216093129986838e-05, - "loss": 0.8205, + "epoch": 0.8953443632166218, + "grad_norm": 1.0067215897578643, + "learning_rate": 5.68983638386188e-07, + "loss": 0.5863, "step": 2327 }, { - "epoch": 0.45, - "learning_rate": 1.2154846045314885e-05, - "loss": 0.8372, + "epoch": 0.8957291265871489, + "grad_norm": 0.9661517864797351, + "learning_rate": 5.648473923988129e-07, + "loss": 0.5293, "step": 2328 }, { - "epoch": 0.45, - "learning_rate": 1.214875995384402e-05, - "loss": 0.8953, + "epoch": 0.896113889957676, + "grad_norm": 0.9727696386346942, + "learning_rate": 5.607257986144321e-07, + "loss": 0.5448, "step": 2329 }, { - "epoch": 0.45, - "learning_rate": 1.214267302781955e-05, - "loss": 0.8846, + "epoch": 0.8964986533282031, + "grad_norm": 0.9705292856427484, + "learning_rate": 5.566188634336212e-07, + "loss": 0.6032, "step": 2330 }, { - "epoch": 0.45, - "learning_rate": 1.2136585269605558e-05, - "loss": 0.9454, + "epoch": 0.8968834166987303, + "grad_norm": 0.9132708725828145, + "learning_rate": 5.525265932341984e-07, + "loss": 0.5462, "step": 2331 }, { - "epoch": 0.45, - "learning_rate": 1.2130496681566475e-05, - "loss": 0.8702, + "epoch": 0.8972681800692575, + "grad_norm": 1.0084468032438618, + "learning_rate": 5.484489943712013e-07, + "loss": 0.5634, "step": 2332 }, { - "epoch": 0.45, - "learning_rate": 1.212440726606703e-05, - "loss": 0.89, + "epoch": 0.8976529434397845, + "grad_norm": 0.581200136707664, + "learning_rate": 5.443860731768869e-07, + "loss": 0.6445, "step": 2333 }, { - "epoch": 0.45, - "learning_rate": 1.211831702547228e-05, - "loss": 0.9661, + "epoch": 0.8980377068103117, + "grad_norm": 0.9577169607005912, + "learning_rate": 5.403378359607181e-07, + "loss": 0.5823, "step": 2334 }, { - "epoch": 0.45, - "learning_rate": 1.2112225962147605e-05, - "loss": 0.8556, + "epoch": 0.8984224701808388, + "grad_norm": 1.025079744338712, + "learning_rate": 5.36304289009355e-07, + "loss": 0.571, "step": 2335 }, { - "epoch": 0.45, - "learning_rate": 1.210613407845871e-05, - "loss": 0.6779, + "epoch": 0.8988072335513659, + "grad_norm": 0.9934941183621226, + "learning_rate": 5.322854385866439e-07, + "loss": 0.5539, "step": 2336 }, { - "epoch": 0.45, - "learning_rate": 1.2100041376771605e-05, - "loss": 0.8873, + "epoch": 0.899191996921893, + "grad_norm": 0.8984464214082862, + "learning_rate": 5.282812909336077e-07, + "loss": 0.5189, "step": 2337 }, { - "epoch": 0.45, - "learning_rate": 1.209394785945263e-05, - "loss": 0.7654, + "epoch": 0.8995767602924202, + "grad_norm": 0.9395528977331445, + "learning_rate": 5.242918522684392e-07, + "loss": 0.592, "step": 2338 }, { - "epoch": 0.45, - "learning_rate": 1.2087853528868432e-05, - "loss": 0.8926, + "epoch": 0.8999615236629472, + "grad_norm": 0.981742140822768, + "learning_rate": 5.203171287864872e-07, + "loss": 0.566, "step": 2339 }, { - "epoch": 0.45, - "learning_rate": 1.2081758387385982e-05, - "loss": 0.8379, + "epoch": 0.9003462870334744, + "grad_norm": 1.0046179278481708, + "learning_rate": 5.163571266602485e-07, + "loss": 0.5487, "step": 2340 }, { - "epoch": 0.45, - "learning_rate": 1.2075662437372567e-05, - "loss": 0.9273, + "epoch": 0.9007310504040016, + "grad_norm": 0.9314277312278729, + "learning_rate": 5.124118520393606e-07, + "loss": 0.4836, "step": 2341 }, { - "epoch": 0.45, - "learning_rate": 1.2069565681195776e-05, - "loss": 0.7402, + "epoch": 0.9011158137745287, + "grad_norm": 1.0013315425071572, + "learning_rate": 5.084813110505871e-07, + "loss": 0.603, "step": 2342 }, { - "epoch": 0.45, - "learning_rate": 1.206346812122352e-05, - "loss": 0.9057, + "epoch": 0.9015005771450558, + "grad_norm": 0.8920367774809554, + "learning_rate": 5.045655097978131e-07, + "loss": 0.5637, "step": 2343 }, { - "epoch": 0.45, - "learning_rate": 1.2057369759824025e-05, - "loss": 0.9256, + "epoch": 0.9018853405155829, + "grad_norm": 1.0109025375833829, + "learning_rate": 5.006644543620342e-07, + "loss": 0.6565, "step": 2344 }, { - "epoch": 0.45, - "learning_rate": 1.2051270599365825e-05, - "loss": 0.9312, + "epoch": 0.9022701038861101, + "grad_norm": 1.0584196165101287, + "learning_rate": 4.967781508013459e-07, + "loss": 0.5907, "step": 2345 }, { - "epoch": 0.45, - "learning_rate": 1.2045170642217756e-05, - "loss": 0.9263, + "epoch": 0.9026548672566371, + "grad_norm": 0.9451393741817894, + "learning_rate": 4.929066051509346e-07, + "loss": 0.5398, "step": 2346 }, { - "epoch": 0.45, - "learning_rate": 1.2039069890748978e-05, - "loss": 0.9734, + "epoch": 0.9030396306271643, + "grad_norm": 1.0078026941130995, + "learning_rate": 4.890498234230689e-07, + "loss": 0.6033, "step": 2347 }, { - "epoch": 0.45, - "learning_rate": 1.2032968347328952e-05, - "loss": 0.8684, + "epoch": 0.9034243939976914, + "grad_norm": 0.9629152175659278, + "learning_rate": 4.852078116070902e-07, + "loss": 0.5183, "step": 2348 }, { - "epoch": 0.45, - "learning_rate": 1.2026866014327446e-05, - "loss": 0.9257, + "epoch": 0.9038091573682185, + "grad_norm": 0.8751393475776088, + "learning_rate": 4.813805756694035e-07, + "loss": 0.553, "step": 2349 }, { - "epoch": 0.45, - "learning_rate": 1.2020762894114535e-05, - "loss": 0.8729, + "epoch": 0.9041939207387457, + "grad_norm": 0.9652416208216412, + "learning_rate": 4.775681215534656e-07, + "loss": 0.5248, "step": 2350 }, { - "epoch": 0.45, - "learning_rate": 1.20146589890606e-05, - "loss": 0.9016, + "epoch": 0.9045786841092728, + "grad_norm": 0.9807209326322125, + "learning_rate": 4.7377045517978173e-07, + "loss": 0.5758, "step": 2351 }, { - "epoch": 0.45, - "learning_rate": 1.2008554301536328e-05, - "loss": 0.9043, + "epoch": 0.9049634474798, + "grad_norm": 0.9868908541243337, + "learning_rate": 4.6998758244588995e-07, + "loss": 0.5464, "step": 2352 }, { - "epoch": 0.45, - "learning_rate": 1.2002448833912712e-05, - "loss": 0.8876, + "epoch": 0.905348210850327, + "grad_norm": 0.9109868274431026, + "learning_rate": 4.662195092263566e-07, + "loss": 0.5201, "step": 2353 }, { - "epoch": 0.45, - "learning_rate": 1.1996342588561042e-05, - "loss": 0.9065, + "epoch": 0.9057329742208542, + "grad_norm": 0.9747477117584985, + "learning_rate": 4.6246624137276206e-07, + "loss": 0.5558, "step": 2354 }, { - "epoch": 0.45, - "learning_rate": 1.1990235567852917e-05, - "loss": 0.9439, + "epoch": 0.9061177375913813, + "grad_norm": 0.9304561972337929, + "learning_rate": 4.587277847136984e-07, + "loss": 0.5469, "step": 2355 }, { - "epoch": 0.45, - "learning_rate": 1.1984127774160226e-05, - "loss": 0.9119, + "epoch": 0.9065025009619084, + "grad_norm": 1.0256007302810355, + "learning_rate": 4.550041450547549e-07, + "loss": 0.5825, "step": 2356 }, { - "epoch": 0.45, - "learning_rate": 1.1978019209855174e-05, - "loss": 0.9684, + "epoch": 0.9068872643324356, + "grad_norm": 0.5814480741612442, + "learning_rate": 4.512953281785104e-07, + "loss": 0.6502, "step": 2357 }, { - "epoch": 0.45, - "learning_rate": 1.1971909877310253e-05, - "loss": 0.8654, + "epoch": 0.9072720277029627, + "grad_norm": 0.9785172435110316, + "learning_rate": 4.476013398445289e-07, + "loss": 0.5713, "step": 2358 }, { - "epoch": 0.45, - "learning_rate": 1.1965799778898258e-05, - "loss": 0.9111, + "epoch": 0.9076567910734898, + "grad_norm": 0.9818991233810478, + "learning_rate": 4.4392218578934164e-07, + "loss": 0.5698, "step": 2359 }, { - "epoch": 0.45, - "learning_rate": 1.1959688916992279e-05, - "loss": 0.9324, + "epoch": 0.9080415544440169, + "grad_norm": 0.9165823634372543, + "learning_rate": 4.4025787172644495e-07, + "loss": 0.5242, "step": 2360 }, { - "epoch": 0.45, - "learning_rate": 1.1953577293965707e-05, - "loss": 0.977, + "epoch": 0.9084263178145441, + "grad_norm": 0.9961728976519675, + "learning_rate": 4.366084033462914e-07, + "loss": 0.6163, "step": 2361 }, { - "epoch": 0.45, - "learning_rate": 1.1947464912192228e-05, - "loss": 0.986, + "epoch": 0.9088110811850711, + "grad_norm": 0.9638056181583311, + "learning_rate": 4.329737863162753e-07, + "loss": 0.5646, "step": 2362 }, { - "epoch": 0.45, - "learning_rate": 1.1941351774045815e-05, - "loss": 0.7919, + "epoch": 0.9091958445555983, + "grad_norm": 0.8814052259536912, + "learning_rate": 4.2935402628073166e-07, + "loss": 0.5585, "step": 2363 }, { - "epoch": 0.45, - "learning_rate": 1.1935237881900743e-05, - "loss": 0.868, + "epoch": 0.9095806079261254, + "grad_norm": 1.026995883434525, + "learning_rate": 4.2574912886092166e-07, + "loss": 0.5451, "step": 2364 }, { - "epoch": 0.46, - "learning_rate": 1.1929123238131579e-05, - "loss": 0.9177, + "epoch": 0.9099653712966526, + "grad_norm": 1.0982153816398865, + "learning_rate": 4.221590996550251e-07, + "loss": 0.5328, "step": 2365 }, { - "epoch": 0.46, - "learning_rate": 1.1923007845113178e-05, - "loss": 0.9558, + "epoch": 0.9103501346671797, + "grad_norm": 0.9607740756108617, + "learning_rate": 4.1858394423813563e-07, + "loss": 0.5554, "step": 2366 }, { - "epoch": 0.46, - "learning_rate": 1.1916891705220689e-05, - "loss": 0.7957, + "epoch": 0.9107348980377068, + "grad_norm": 0.967222096244933, + "learning_rate": 4.1502366816224327e-07, + "loss": 0.5344, "step": 2367 }, { - "epoch": 0.46, - "learning_rate": 1.191077482082955e-05, - "loss": 0.8914, + "epoch": 0.911119661408234, + "grad_norm": 0.9253157972340882, + "learning_rate": 4.1147827695623643e-07, + "loss": 0.5575, "step": 2368 }, { - "epoch": 0.46, - "learning_rate": 1.1904657194315486e-05, - "loss": 0.9309, + "epoch": 0.911504424778761, + "grad_norm": 0.9026837444510366, + "learning_rate": 4.0794777612588543e-07, + "loss": 0.4801, "step": 2369 }, { - "epoch": 0.46, - "learning_rate": 1.1898538828054517e-05, - "loss": 0.9408, + "epoch": 0.9118891881492882, + "grad_norm": 1.073463076208279, + "learning_rate": 4.044321711538368e-07, + "loss": 0.5375, "step": 2370 }, { - "epoch": 0.46, - "learning_rate": 1.1892419724422946e-05, - "loss": 0.8667, + "epoch": 0.9122739515198153, + "grad_norm": 1.0276425033461005, + "learning_rate": 4.00931467499609e-07, + "loss": 0.6026, "step": 2371 }, { - "epoch": 0.46, - "learning_rate": 1.1886299885797357e-05, - "loss": 0.8377, + "epoch": 0.9126587148903424, + "grad_norm": 1.0511011473806682, + "learning_rate": 3.974456705995733e-07, + "loss": 0.6301, "step": 2372 }, { - "epoch": 0.46, - "learning_rate": 1.1880179314554629e-05, - "loss": 0.9064, + "epoch": 0.9130434782608695, + "grad_norm": 0.9855482443108753, + "learning_rate": 3.9397478586695513e-07, + "loss": 0.5699, "step": 2373 }, { - "epoch": 0.46, - "learning_rate": 1.1874058013071923e-05, - "loss": 0.9517, + "epoch": 0.9134282416313967, + "grad_norm": 0.5579522758737274, + "learning_rate": 3.90518818691823e-07, + "loss": 0.6328, "step": 2374 }, { - "epoch": 0.46, - "learning_rate": 1.1867935983726676e-05, - "loss": 0.8924, + "epoch": 0.9138130050019239, + "grad_norm": 0.9136343680249246, + "learning_rate": 3.8707777444107697e-07, + "loss": 0.5381, "step": 2375 }, { - "epoch": 0.46, - "learning_rate": 1.186181322889662e-05, - "loss": 0.8796, + "epoch": 0.9141977683724509, + "grad_norm": 1.0158813047543156, + "learning_rate": 3.8365165845844266e-07, + "loss": 0.5932, "step": 2376 }, { - "epoch": 0.46, - "learning_rate": 1.1855689750959759e-05, - "loss": 0.8724, + "epoch": 0.9145825317429781, + "grad_norm": 0.9570696660979296, + "learning_rate": 3.8024047606446736e-07, + "loss": 0.5698, "step": 2377 }, { - "epoch": 0.46, - "learning_rate": 1.1849565552294379e-05, - "loss": 0.7114, + "epoch": 0.9149672951135052, + "grad_norm": 1.0723454200351203, + "learning_rate": 3.768442325565036e-07, + "loss": 0.5744, "step": 2378 }, { - "epoch": 0.46, - "learning_rate": 1.1843440635279056e-05, - "loss": 0.9472, + "epoch": 0.9153520584840323, + "grad_norm": 1.0051260357364857, + "learning_rate": 3.7346293320870363e-07, + "loss": 0.5381, "step": 2379 }, { - "epoch": 0.46, - "learning_rate": 1.1837315002292629e-05, - "loss": 0.8981, + "epoch": 0.9157368218545594, + "grad_norm": 0.9316948440166639, + "learning_rate": 3.700965832720171e-07, + "loss": 0.5514, "step": 2380 }, { - "epoch": 0.46, - "learning_rate": 1.1831188655714225e-05, - "loss": 0.8705, + "epoch": 0.9161215852250866, + "grad_norm": 0.9252796701163073, + "learning_rate": 3.6674518797417236e-07, + "loss": 0.5578, "step": 2381 }, { - "epoch": 0.46, - "learning_rate": 1.182506159792325e-05, - "loss": 0.9389, + "epoch": 0.9165063485956138, + "grad_norm": 0.9537621731089124, + "learning_rate": 3.6340875251967946e-07, + "loss": 0.5501, "step": 2382 }, { - "epoch": 0.46, - "learning_rate": 1.1818933831299381e-05, - "loss": 0.7024, + "epoch": 0.9168911119661408, + "grad_norm": 0.9182127690078434, + "learning_rate": 3.6008728208981157e-07, + "loss": 0.4988, "step": 2383 }, { - "epoch": 0.46, - "learning_rate": 1.1812805358222571e-05, - "loss": 0.8889, + "epoch": 0.917275875336668, + "grad_norm": 0.9997492752755271, + "learning_rate": 3.5678078184260834e-07, + "loss": 0.5617, "step": 2384 }, { - "epoch": 0.46, - "learning_rate": 1.180667618107305e-05, - "loss": 0.8783, + "epoch": 0.917660638707195, + "grad_norm": 0.9782343660626003, + "learning_rate": 3.5348925691285675e-07, + "loss": 0.5508, "step": 2385 }, { - "epoch": 0.46, - "learning_rate": 1.1800546302231317e-05, - "loss": 0.8881, + "epoch": 0.9180454020777222, + "grad_norm": 0.9357892193971725, + "learning_rate": 3.502127124120891e-07, + "loss": 0.5681, "step": 2386 }, { - "epoch": 0.46, - "learning_rate": 1.1794415724078147e-05, - "loss": 0.8749, + "epoch": 0.9184301654482493, + "grad_norm": 0.9323127377505398, + "learning_rate": 3.4695115342857524e-07, + "loss": 0.562, "step": 2387 }, { - "epoch": 0.46, - "learning_rate": 1.1788284448994588e-05, - "loss": 0.9019, + "epoch": 0.9188149288187765, + "grad_norm": 0.9519817003468568, + "learning_rate": 3.437045850273113e-07, + "loss": 0.5778, "step": 2388 }, { - "epoch": 0.46, - "learning_rate": 1.1782152479361956e-05, - "loss": 0.8269, + "epoch": 0.9191996921893035, + "grad_norm": 0.9833139092401583, + "learning_rate": 3.404730122500155e-07, + "loss": 0.5581, "step": 2389 }, { - "epoch": 0.46, - "learning_rate": 1.1776019817561834e-05, - "loss": 0.9094, + "epoch": 0.9195844555598307, + "grad_norm": 0.9074580866011454, + "learning_rate": 3.3725644011512125e-07, + "loss": 0.5458, "step": 2390 }, { - "epoch": 0.46, - "learning_rate": 1.1769886465976086e-05, - "loss": 0.8418, + "epoch": 0.9199692189303579, + "grad_norm": 1.0142679019194938, + "learning_rate": 3.3405487361776177e-07, + "loss": 0.5793, "step": 2391 }, { - "epoch": 0.46, - "learning_rate": 1.1763752426986823e-05, - "loss": 0.9219, + "epoch": 0.9203539823008849, + "grad_norm": 0.9673160792447684, + "learning_rate": 3.308683177297711e-07, + "loss": 0.5881, "step": 2392 }, { - "epoch": 0.46, - "learning_rate": 1.1757617702976443e-05, - "loss": 0.8584, + "epoch": 0.9207387456714121, + "grad_norm": 0.9907206579172364, + "learning_rate": 3.2769677739966975e-07, + "loss": 0.5721, "step": 2393 }, { - "epoch": 0.46, - "learning_rate": 1.17514822963276e-05, - "loss": 0.8963, + "epoch": 0.9211235090419392, + "grad_norm": 0.6296360371556724, + "learning_rate": 3.245402575526646e-07, + "loss": 0.6463, "step": 2394 }, { - "epoch": 0.46, - "learning_rate": 1.1745346209423216e-05, - "loss": 0.8272, + "epoch": 0.9215082724124664, + "grad_norm": 0.9277300167974788, + "learning_rate": 3.2139876309063233e-07, + "loss": 0.5949, "step": 2395 }, { - "epoch": 0.46, - "learning_rate": 1.1739209444646479e-05, - "loss": 0.7077, + "epoch": 0.9218930357829934, + "grad_norm": 0.9331767724124044, + "learning_rate": 3.182722988921161e-07, + "loss": 0.5734, "step": 2396 }, { - "epoch": 0.46, - "learning_rate": 1.1733072004380827e-05, - "loss": 0.9086, + "epoch": 0.9222777991535206, + "grad_norm": 0.9521854089451881, + "learning_rate": 3.151608698123232e-07, + "loss": 0.5952, "step": 2397 }, { - "epoch": 0.46, - "learning_rate": 1.1726933891009985e-05, - "loss": 0.8505, + "epoch": 0.9226625625240477, + "grad_norm": 0.9934286548281223, + "learning_rate": 3.1206448068310635e-07, + "loss": 0.6269, "step": 2398 }, { - "epoch": 0.46, - "learning_rate": 1.1720795106917917e-05, - "loss": 0.9529, + "epoch": 0.9230473258945748, + "grad_norm": 0.9363257083806247, + "learning_rate": 3.0898313631296586e-07, + "loss": 0.5297, "step": 2399 }, { - "epoch": 0.46, - "learning_rate": 1.171465565448886e-05, - "loss": 0.9325, + "epoch": 0.923432089265102, + "grad_norm": 1.0059538516151814, + "learning_rate": 3.0591684148703617e-07, + "loss": 0.5342, "step": 2400 }, { - "epoch": 0.46, - "learning_rate": 1.1708515536107299e-05, - "loss": 0.8685, + "epoch": 0.9238168526356291, + "grad_norm": 0.965749263311831, + "learning_rate": 3.0286560096708275e-07, + "loss": 0.5929, "step": 2401 }, { - "epoch": 0.46, - "learning_rate": 1.1702374754157998e-05, - "loss": 0.9923, + "epoch": 0.9242016160061562, + "grad_norm": 0.9319179115022181, + "learning_rate": 2.998294194914897e-07, + "loss": 0.5515, "step": 2402 }, { - "epoch": 0.46, - "learning_rate": 1.1696233311025957e-05, - "loss": 0.8644, + "epoch": 0.9245863793766833, + "grad_norm": 0.9716088427884352, + "learning_rate": 2.968083017752599e-07, + "loss": 0.5923, "step": 2403 }, { - "epoch": 0.46, - "learning_rate": 1.1690091209096441e-05, - "loss": 0.9011, + "epoch": 0.9249711427472105, + "grad_norm": 0.8562414022044039, + "learning_rate": 2.938022525099982e-07, + "loss": 0.4878, "step": 2404 }, { - "epoch": 0.46, - "learning_rate": 1.1683948450754976e-05, - "loss": 0.9175, + "epoch": 0.9253559061177375, + "grad_norm": 0.5969048719923384, + "learning_rate": 2.908112763639137e-07, + "loss": 0.6451, "step": 2405 }, { - "epoch": 0.46, - "learning_rate": 1.1677805038387337e-05, - "loss": 0.881, + "epoch": 0.9257406694882647, + "grad_norm": 0.9967031473809232, + "learning_rate": 2.878353779818044e-07, + "loss": 0.6064, "step": 2406 }, { - "epoch": 0.46, - "learning_rate": 1.1671660974379554e-05, - "loss": 0.9308, + "epoch": 0.9261254328587919, + "grad_norm": 0.9937910347931432, + "learning_rate": 2.848745619850546e-07, + "loss": 0.5376, "step": 2407 }, { - "epoch": 0.46, - "learning_rate": 1.1665516261117914e-05, - "loss": 0.8608, + "epoch": 0.926510196229319, + "grad_norm": 0.9466785187942308, + "learning_rate": 2.8192883297162634e-07, + "loss": 0.5458, "step": 2408 }, { - "epoch": 0.46, - "learning_rate": 1.1659370900988946e-05, - "loss": 0.9475, + "epoch": 0.9268949595998461, + "grad_norm": 0.9511487459018984, + "learning_rate": 2.7899819551605256e-07, + "loss": 0.5689, "step": 2409 }, { - "epoch": 0.46, - "learning_rate": 1.165322489637944e-05, - "loss": 0.8752, + "epoch": 0.9272797229703732, + "grad_norm": 0.9925734205201574, + "learning_rate": 2.760826541694328e-07, + "loss": 0.5988, "step": 2410 }, { - "epoch": 0.46, - "learning_rate": 1.164707824967644e-05, - "loss": 0.6829, + "epoch": 0.9276644863409004, + "grad_norm": 1.003808776162742, + "learning_rate": 2.7318221345941865e-07, + "loss": 0.5243, "step": 2411 }, { - "epoch": 0.46, - "learning_rate": 1.1640930963267226e-05, - "loss": 0.9154, + "epoch": 0.9280492497114274, + "grad_norm": 0.9362452434566728, + "learning_rate": 2.7029687789021377e-07, + "loss": 0.5332, "step": 2412 }, { - "epoch": 0.46, - "learning_rate": 1.1634783039539328e-05, - "loss": 0.9355, + "epoch": 0.9284340130819546, + "grad_norm": 0.9065876475250806, + "learning_rate": 2.67426651942565e-07, + "loss": 0.5158, "step": 2413 }, { - "epoch": 0.46, - "learning_rate": 1.162863448088054e-05, - "loss": 0.8705, + "epoch": 0.9288187764524817, + "grad_norm": 1.0097444169795147, + "learning_rate": 2.645715400737536e-07, + "loss": 0.5493, "step": 2414 }, { - "epoch": 0.46, - "learning_rate": 1.1622485289678886e-05, - "loss": 0.8786, + "epoch": 0.9292035398230089, + "grad_norm": 1.0485610091610649, + "learning_rate": 2.6173154671758847e-07, + "loss": 0.5859, "step": 2415 }, { - "epoch": 0.46, - "learning_rate": 1.1616335468322641e-05, - "loss": 0.9635, + "epoch": 0.929588303193536, + "grad_norm": 0.9939174688047393, + "learning_rate": 2.589066762844039e-07, + "loss": 0.6049, "step": 2416 }, { - "epoch": 0.47, - "learning_rate": 1.1610185019200324e-05, - "loss": 0.8686, + "epoch": 0.9299730665640631, + "grad_norm": 0.9502338638141001, + "learning_rate": 2.5609693316104745e-07, + "loss": 0.5562, "step": 2417 }, { - "epoch": 0.47, - "learning_rate": 1.1604033944700701e-05, - "loss": 0.8589, + "epoch": 0.9303578299345903, + "grad_norm": 0.943993353040941, + "learning_rate": 2.5330232171087433e-07, + "loss": 0.5116, "step": 2418 }, { - "epoch": 0.47, - "learning_rate": 1.1597882247212776e-05, - "loss": 0.9036, + "epoch": 0.9307425933051173, + "grad_norm": 0.9415839178607306, + "learning_rate": 2.5052284627374077e-07, + "loss": 0.5436, "step": 2419 }, { - "epoch": 0.47, - "learning_rate": 1.15917299291258e-05, - "loss": 0.9309, + "epoch": 0.9311273566756445, + "grad_norm": 0.932586220982553, + "learning_rate": 2.477585111659997e-07, + "loss": 0.5863, "step": 2420 }, { - "epoch": 0.47, - "learning_rate": 1.1585576992829261e-05, - "loss": 0.8943, + "epoch": 0.9315121200461716, + "grad_norm": 1.0047124316380434, + "learning_rate": 2.4500932068049046e-07, + "loss": 0.5913, "step": 2421 }, { - "epoch": 0.47, - "learning_rate": 1.1579423440712887e-05, - "loss": 0.8767, + "epoch": 0.9318968834166987, + "grad_norm": 1.009245593535221, + "learning_rate": 2.422752790865346e-07, + "loss": 0.5355, "step": 2422 }, { - "epoch": 0.47, - "learning_rate": 1.1573269275166652e-05, - "loss": 0.8464, + "epoch": 0.9322816467872258, + "grad_norm": 0.9141894572991723, + "learning_rate": 2.3955639062992696e-07, + "loss": 0.5948, "step": 2423 }, { - "epoch": 0.47, - "learning_rate": 1.1567114498580758e-05, - "loss": 0.9548, + "epoch": 0.932666410157753, + "grad_norm": 0.990832642191648, + "learning_rate": 2.3685265953293345e-07, + "loss": 0.5781, "step": 2424 }, { - "epoch": 0.47, - "learning_rate": 1.1560959113345649e-05, - "loss": 0.9526, + "epoch": 0.9330511735282802, + "grad_norm": 0.8839520239435691, + "learning_rate": 2.3416408999427876e-07, + "loss": 0.5111, "step": 2425 }, { - "epoch": 0.47, - "learning_rate": 1.1554803121852005e-05, - "loss": 0.8779, + "epoch": 0.9334359368988072, + "grad_norm": 0.8607701941832979, + "learning_rate": 2.3149068618914417e-07, + "loss": 0.5179, "step": 2426 }, { - "epoch": 0.47, - "learning_rate": 1.1548646526490749e-05, - "loss": 0.9078, + "epoch": 0.9338207002693344, + "grad_norm": 0.9378181597006409, + "learning_rate": 2.2883245226915652e-07, + "loss": 0.5267, "step": 2427 }, { - "epoch": 0.47, - "learning_rate": 1.1542489329653024e-05, - "loss": 0.9567, + "epoch": 0.9342054636398615, + "grad_norm": 1.1294172375393698, + "learning_rate": 2.2618939236238924e-07, + "loss": 0.5674, "step": 2428 }, { - "epoch": 0.47, - "learning_rate": 1.153633153373022e-05, - "loss": 0.7328, + "epoch": 0.9345902270103886, + "grad_norm": 1.0910056589943973, + "learning_rate": 2.2356151057334908e-07, + "loss": 0.5964, "step": 2429 }, { - "epoch": 0.47, - "learning_rate": 1.1530173141113947e-05, - "loss": 0.8814, + "epoch": 0.9349749903809157, + "grad_norm": 0.9938468434102637, + "learning_rate": 2.209488109829727e-07, + "loss": 0.5365, "step": 2430 }, { - "epoch": 0.47, - "learning_rate": 1.1524014154196063e-05, - "loss": 0.9297, + "epoch": 0.9353597537514429, + "grad_norm": 0.9362046484560679, + "learning_rate": 2.1835129764861907e-07, + "loss": 0.5662, "step": 2431 }, { - "epoch": 0.47, - "learning_rate": 1.1517854575368644e-05, - "loss": 0.9399, + "epoch": 0.93574451712197, + "grad_norm": 0.985268002909651, + "learning_rate": 2.1576897460406477e-07, + "loss": 0.5759, "step": 2432 }, { - "epoch": 0.47, - "learning_rate": 1.1511694407023994e-05, - "loss": 0.8541, + "epoch": 0.9361292804924971, + "grad_norm": 1.023932783223243, + "learning_rate": 2.1320184585949532e-07, + "loss": 0.5733, "step": 2433 }, { - "epoch": 0.47, - "learning_rate": 1.1505533651554654e-05, - "loss": 0.9818, + "epoch": 0.9365140438630243, + "grad_norm": 0.9936384041019983, + "learning_rate": 2.106499154015018e-07, + "loss": 0.582, "step": 2434 }, { - "epoch": 0.47, - "learning_rate": 1.1499372311353398e-05, - "loss": 0.8266, + "epoch": 0.9368988072335513, + "grad_norm": 1.0258676049672566, + "learning_rate": 2.0811318719307194e-07, + "loss": 0.5465, "step": 2435 }, { - "epoch": 0.47, - "learning_rate": 1.149321038881321e-05, - "loss": 0.9242, + "epoch": 0.9372835706040785, + "grad_norm": 0.9118460575719739, + "learning_rate": 2.0559166517358787e-07, + "loss": 0.5698, "step": 2436 }, { - "epoch": 0.47, - "learning_rate": 1.1487047886327314e-05, - "loss": 0.8499, + "epoch": 0.9376683339746056, + "grad_norm": 0.9900909698296422, + "learning_rate": 2.0308535325881616e-07, + "loss": 0.5811, "step": 2437 }, { - "epoch": 0.47, - "learning_rate": 1.1480884806289151e-05, - "loss": 0.8852, + "epoch": 0.9380530973451328, + "grad_norm": 0.9745529287717064, + "learning_rate": 2.0059425534090128e-07, + "loss": 0.5685, "step": 2438 }, { - "epoch": 0.47, - "learning_rate": 1.1474721151092397e-05, - "loss": 0.8984, + "epoch": 0.9384378607156598, + "grad_norm": 0.9828874590153357, + "learning_rate": 1.981183752883631e-07, + "loss": 0.5312, "step": 2439 }, { - "epoch": 0.47, - "learning_rate": 1.1468556923130943e-05, - "loss": 0.7416, + "epoch": 0.938822624086187, + "grad_norm": 0.964894594104015, + "learning_rate": 1.9565771694608937e-07, + "loss": 0.5814, "step": 2440 }, { - "epoch": 0.47, - "learning_rate": 1.14623921247989e-05, - "loss": 0.826, + "epoch": 0.9392073874567142, + "grad_norm": 0.9158363978701964, + "learning_rate": 1.9321228413532788e-07, + "loss": 0.4908, "step": 2441 }, { - "epoch": 0.47, - "learning_rate": 1.1456226758490603e-05, - "loss": 0.888, + "epoch": 0.9395921508272412, + "grad_norm": 0.9995871364758443, + "learning_rate": 1.907820806536842e-07, + "loss": 0.5392, "step": 2442 }, { - "epoch": 0.47, - "learning_rate": 1.1450060826600618e-05, - "loss": 0.8029, + "epoch": 0.9399769141977684, + "grad_norm": 0.9937849630397988, + "learning_rate": 1.883671102751128e-07, + "loss": 0.6031, "step": 2443 }, { - "epoch": 0.47, - "learning_rate": 1.1443894331523718e-05, - "loss": 0.896, + "epoch": 0.9403616775682955, + "grad_norm": 0.9345996677902778, + "learning_rate": 1.859673767499115e-07, + "loss": 0.549, "step": 2444 }, { - "epoch": 0.47, - "learning_rate": 1.1437727275654893e-05, - "loss": 0.9425, + "epoch": 0.9407464409388226, + "grad_norm": 1.0082930388534317, + "learning_rate": 1.83582883804716e-07, + "loss": 0.5336, "step": 2445 }, { - "epoch": 0.47, - "learning_rate": 1.1431559661389362e-05, - "loss": 0.8706, + "epoch": 0.9411312043093497, + "grad_norm": 0.9756209301161831, + "learning_rate": 1.8121363514249534e-07, + "loss": 0.4854, "step": 2446 }, { - "epoch": 0.47, - "learning_rate": 1.1425391491122557e-05, - "loss": 0.795, + "epoch": 0.9415159676798769, + "grad_norm": 0.9907728309331885, + "learning_rate": 1.7885963444254528e-07, + "loss": 0.5468, "step": 2447 }, { - "epoch": 0.47, - "learning_rate": 1.141922276725012e-05, - "loss": 0.942, + "epoch": 0.941900731050404, + "grad_norm": 1.016956156571003, + "learning_rate": 1.7652088536048052e-07, + "loss": 0.527, "step": 2448 }, { - "epoch": 0.47, - "learning_rate": 1.1413053492167915e-05, - "loss": 0.8904, + "epoch": 0.9422854944209311, + "grad_norm": 0.8890642294641641, + "learning_rate": 1.7419739152823468e-07, + "loss": 0.5128, "step": 2449 }, { - "epoch": 0.47, - "learning_rate": 1.1406883668272015e-05, - "loss": 0.8778, + "epoch": 0.9426702577914583, + "grad_norm": 0.9602461030706585, + "learning_rate": 1.7188915655404814e-07, + "loss": 0.5605, "step": 2450 }, { - "epoch": 0.47, - "learning_rate": 1.140071329795871e-05, - "loss": 0.8087, + "epoch": 0.9430550211619854, + "grad_norm": 0.9548840616885846, + "learning_rate": 1.695961840224636e-07, + "loss": 0.5244, "step": 2451 }, { - "epoch": 0.47, - "learning_rate": 1.13945423836245e-05, - "loss": 0.8917, + "epoch": 0.9434397845325125, + "grad_norm": 0.937949571322601, + "learning_rate": 1.6731847749432705e-07, + "loss": 0.5148, "step": 2452 }, { - "epoch": 0.47, - "learning_rate": 1.1388370927666102e-05, - "loss": 0.8746, + "epoch": 0.9438245479030396, + "grad_norm": 0.961750126643927, + "learning_rate": 1.6505604050677249e-07, + "loss": 0.5714, "step": 2453 }, { - "epoch": 0.47, - "learning_rate": 1.1382198932480429e-05, - "loss": 0.8692, + "epoch": 0.9442093112735668, + "grad_norm": 0.9688255961171185, + "learning_rate": 1.6280887657322276e-07, + "loss": 0.5595, "step": 2454 }, { - "epoch": 0.47, - "learning_rate": 1.1376026400464616e-05, - "loss": 0.8537, + "epoch": 0.9445940746440938, + "grad_norm": 0.9980302471925951, + "learning_rate": 1.6057698918338526e-07, + "loss": 0.6014, "step": 2455 }, { - "epoch": 0.47, - "learning_rate": 1.136985333401601e-05, - "loss": 0.9619, + "epoch": 0.944978838014621, + "grad_norm": 0.9495825393763954, + "learning_rate": 1.5836038180324198e-07, + "loss": 0.577, "step": 2456 }, { - "epoch": 0.47, - "learning_rate": 1.1363679735532151e-05, - "loss": 0.8853, + "epoch": 0.9453636013851482, + "grad_norm": 0.9069473820422234, + "learning_rate": 1.561590578750438e-07, + "loss": 0.5627, "step": 2457 }, { - "epoch": 0.47, - "learning_rate": 1.1357505607410797e-05, - "loss": 0.9406, + "epoch": 0.9457483647556753, + "grad_norm": 0.9402516718573974, + "learning_rate": 1.5397302081731069e-07, + "loss": 0.5459, "step": 2458 }, { - "epoch": 0.47, - "learning_rate": 1.1351330952049908e-05, - "loss": 0.6931, + "epoch": 0.9461331281262024, + "grad_norm": 0.9834533515416687, + "learning_rate": 1.518022740248215e-07, + "loss": 0.5216, "step": 2459 }, { - "epoch": 0.47, - "learning_rate": 1.1345155771847646e-05, - "loss": 0.9287, + "epoch": 0.9465178914967295, + "grad_norm": 1.0126691744035865, + "learning_rate": 1.4964682086861082e-07, + "loss": 0.532, "step": 2460 }, { - "epoch": 0.47, - "learning_rate": 1.1338980069202388e-05, - "loss": 0.8678, + "epoch": 0.9469026548672567, + "grad_norm": 0.9347440206392564, + "learning_rate": 1.475066646959611e-07, + "loss": 0.5332, "step": 2461 }, { - "epoch": 0.47, - "learning_rate": 1.1332803846512697e-05, - "loss": 0.8436, + "epoch": 0.9472874182377837, + "grad_norm": 0.9388689614590471, + "learning_rate": 1.4538180883040264e-07, + "loss": 0.5743, "step": 2462 }, { - "epoch": 0.47, - "learning_rate": 1.1326627106177348e-05, - "loss": 0.7999, + "epoch": 0.9476721816083109, + "grad_norm": 0.9870105655434643, + "learning_rate": 1.4327225657170485e-07, + "loss": 0.5707, "step": 2463 }, { - "epoch": 0.47, - "learning_rate": 1.132044985059532e-05, - "loss": 0.8815, + "epoch": 0.948056944978838, + "grad_norm": 0.9125617439265533, + "learning_rate": 1.411780111958694e-07, + "loss": 0.5271, "step": 2464 }, { - "epoch": 0.47, - "learning_rate": 1.1314272082165785e-05, - "loss": 0.9056, + "epoch": 0.9484417083493651, + "grad_norm": 0.9452312491160781, + "learning_rate": 1.3909907595512806e-07, + "loss": 0.515, "step": 2465 }, { - "epoch": 0.47, - "learning_rate": 1.1308093803288119e-05, - "loss": 0.9523, + "epoch": 0.9488264717198923, + "grad_norm": 0.8920626831998674, + "learning_rate": 1.3703545407793951e-07, + "loss": 0.5248, "step": 2466 }, { - "epoch": 0.47, - "learning_rate": 1.130191501636189e-05, - "loss": 0.8918, + "epoch": 0.9492112350904194, + "grad_norm": 0.9572139397237391, + "learning_rate": 1.34987148768978e-07, + "loss": 0.5378, "step": 2467 }, { - "epoch": 0.47, - "learning_rate": 1.1295735723786872e-05, - "loss": 0.8795, + "epoch": 0.9495959984609466, + "grad_norm": 0.9822711014766814, + "learning_rate": 1.3295416320913357e-07, + "loss": 0.5975, "step": 2468 }, { - "epoch": 0.48, - "learning_rate": 1.1289555927963032e-05, - "loss": 0.9198, + "epoch": 0.9499807618314736, + "grad_norm": 0.9742273692841893, + "learning_rate": 1.3093650055550855e-07, + "loss": 0.5522, "step": 2469 }, { - "epoch": 0.48, - "learning_rate": 1.1283375631290528e-05, - "loss": 0.9326, + "epoch": 0.9503655252020008, + "grad_norm": 0.9648234929159051, + "learning_rate": 1.2893416394140323e-07, + "loss": 0.5675, "step": 2470 }, { - "epoch": 0.48, - "learning_rate": 1.1277194836169714e-05, - "loss": 0.8788, + "epoch": 0.9507502885725279, + "grad_norm": 1.0402190222687415, + "learning_rate": 1.269471564763247e-07, + "loss": 0.5853, "step": 2471 }, { - "epoch": 0.48, - "learning_rate": 1.1271013545001144e-05, - "loss": 0.9854, + "epoch": 0.951135051943055, + "grad_norm": 0.9548842934790226, + "learning_rate": 1.2497548124597026e-07, + "loss": 0.5305, "step": 2472 }, { - "epoch": 0.48, - "learning_rate": 1.1264831760185562e-05, - "loss": 0.9807, + "epoch": 0.9515198153135821, + "grad_norm": 1.01264789016436, + "learning_rate": 1.2301914131222726e-07, + "loss": 0.5501, "step": 2473 }, { - "epoch": 0.48, - "learning_rate": 1.1258649484123895e-05, - "loss": 0.8556, + "epoch": 0.9519045786841093, + "grad_norm": 0.8348287767840127, + "learning_rate": 1.2107813971317106e-07, + "loss": 0.4978, "step": 2474 }, { - "epoch": 0.48, - "learning_rate": 1.1252466719217274e-05, - "loss": 0.91, + "epoch": 0.9522893420546364, + "grad_norm": 1.0393966283584164, + "learning_rate": 1.1915247946305498e-07, + "loss": 0.5979, "step": 2475 }, { - "epoch": 0.48, - "learning_rate": 1.1246283467867012e-05, - "loss": 0.8964, + "epoch": 0.9526741054251635, + "grad_norm": 0.9280316193439552, + "learning_rate": 1.1724216355231022e-07, + "loss": 0.5274, "step": 2476 }, { - "epoch": 0.48, - "learning_rate": 1.1240099732474613e-05, - "loss": 0.8497, + "epoch": 0.9530588687956907, + "grad_norm": 0.9103203003487964, + "learning_rate": 1.1534719494753821e-07, + "loss": 0.5269, "step": 2477 }, { - "epoch": 0.48, - "learning_rate": 1.1233915515441765e-05, - "loss": 0.8549, + "epoch": 0.9534436321662177, + "grad_norm": 0.9784446779536516, + "learning_rate": 1.1346757659150498e-07, + "loss": 0.5343, "step": 2478 }, { - "epoch": 0.48, - "learning_rate": 1.1227730819170349e-05, - "loss": 0.8415, + "epoch": 0.9538283955367449, + "grad_norm": 0.9151540775122403, + "learning_rate": 1.116033114031434e-07, + "loss": 0.547, "step": 2479 }, { - "epoch": 0.48, - "learning_rate": 1.1221545646062431e-05, - "loss": 0.8549, + "epoch": 0.954213158907272, + "grad_norm": 1.0247938427397418, + "learning_rate": 1.0975440227753764e-07, + "loss": 0.5928, "step": 2480 }, { - "epoch": 0.48, - "learning_rate": 1.121535999852026e-05, - "loss": 0.928, + "epoch": 0.9545979222777992, + "grad_norm": 0.9942357803708224, + "learning_rate": 1.0792085208593095e-07, + "loss": 0.5435, "step": 2481 }, { - "epoch": 0.48, - "learning_rate": 1.1209173878946271e-05, - "loss": 0.8048, + "epoch": 0.9549826856483262, + "grad_norm": 0.9712505184941882, + "learning_rate": 1.061026636757101e-07, + "loss": 0.5678, "step": 2482 }, { - "epoch": 0.48, - "learning_rate": 1.1202987289743078e-05, - "loss": 0.8088, + "epoch": 0.9553674490188534, + "grad_norm": 0.9315370668667267, + "learning_rate": 1.0429983987041092e-07, + "loss": 0.5026, "step": 2483 }, { - "epoch": 0.48, - "learning_rate": 1.1196800233313488e-05, - "loss": 0.782, + "epoch": 0.9557522123893806, + "grad_norm": 0.927628758273562, + "learning_rate": 1.0251238346970393e-07, + "loss": 0.4985, "step": 2484 }, { - "epoch": 0.48, - "learning_rate": 1.1190612712060475e-05, - "loss": 1.0286, + "epoch": 0.9561369757599076, + "grad_norm": 1.0392094447343838, + "learning_rate": 1.007402972493976e-07, + "loss": 0.6133, "step": 2485 }, { - "epoch": 0.48, - "learning_rate": 1.1184424728387204e-05, - "loss": 0.8416, + "epoch": 0.9565217391304348, + "grad_norm": 0.9160447837565115, + "learning_rate": 9.898358396143171e-08, + "loss": 0.5005, "step": 2486 }, { - "epoch": 0.48, - "learning_rate": 1.1178236284697017e-05, - "loss": 0.8944, + "epoch": 0.9569065025009619, + "grad_norm": 0.9878949744436316, + "learning_rate": 9.72422463338718e-08, + "loss": 0.5798, "step": 2487 }, { - "epoch": 0.48, - "learning_rate": 1.1172047383393434e-05, - "loss": 0.9575, + "epoch": 0.957291265871489, + "grad_norm": 0.8892132284159903, + "learning_rate": 9.55162870709081e-08, + "loss": 0.532, "step": 2488 }, { - "epoch": 0.48, - "learning_rate": 1.1165858026880151e-05, - "loss": 0.858, + "epoch": 0.9576760292420161, + "grad_norm": 0.9303910660521667, + "learning_rate": 9.380570885284546e-08, + "loss": 0.543, "step": 2489 }, { - "epoch": 0.48, - "learning_rate": 1.1159668217561048e-05, - "loss": 0.8785, + "epoch": 0.9580607926125433, + "grad_norm": 0.9959429605824407, + "learning_rate": 9.211051433610674e-08, + "loss": 0.5631, "step": 2490 }, { - "epoch": 0.48, - "learning_rate": 1.115347795784017e-05, - "loss": 0.9034, + "epoch": 0.9584455559830705, + "grad_norm": 0.9660850412990554, + "learning_rate": 9.04307061532217e-08, + "loss": 0.5538, "step": 2491 }, { - "epoch": 0.48, - "learning_rate": 1.1147287250121745e-05, - "loss": 0.8644, + "epoch": 0.9588303193535975, + "grad_norm": 1.030694195085526, + "learning_rate": 8.876628691282918e-08, + "loss": 0.6092, "step": 2492 }, { - "epoch": 0.48, - "learning_rate": 1.1141096096810174e-05, - "loss": 0.9071, + "epoch": 0.9592150827241247, + "grad_norm": 0.9977092375822906, + "learning_rate": 8.711725919966718e-08, + "loss": 0.5595, "step": 2493 }, { - "epoch": 0.48, - "learning_rate": 1.1134904500310029e-05, - "loss": 0.941, + "epoch": 0.9595998460946518, + "grad_norm": 0.9451152218442164, + "learning_rate": 8.54836255745728e-08, + "loss": 0.5837, "step": 2494 }, { - "epoch": 0.48, - "learning_rate": 1.1128712463026048e-05, - "loss": 0.9271, + "epoch": 0.9599846094651789, + "grad_norm": 0.9551630629830747, + "learning_rate": 8.386538857447779e-08, + "loss": 0.5645, "step": 2495 }, { - "epoch": 0.48, - "learning_rate": 1.1122519987363156e-05, - "loss": 0.9136, + "epoch": 0.960369372835706, + "grad_norm": 0.9799378741410755, + "learning_rate": 8.226255071240308e-08, + "loss": 0.5817, "step": 2496 }, { - "epoch": 0.48, - "learning_rate": 1.1116327075726436e-05, - "loss": 0.8441, + "epoch": 0.9607541362062332, + "grad_norm": 1.0652537433452924, + "learning_rate": 8.067511447745535e-08, + "loss": 0.6236, "step": 2497 }, { - "epoch": 0.48, - "learning_rate": 1.1110133730521142e-05, - "loss": 1.0055, + "epoch": 0.9611388995767602, + "grad_norm": 1.0022138275023702, + "learning_rate": 7.910308233482488e-08, + "loss": 0.5945, "step": 2498 }, { - "epoch": 0.48, - "learning_rate": 1.11039399541527e-05, - "loss": 0.852, + "epoch": 0.9615236629472874, + "grad_norm": 0.95444133089293, + "learning_rate": 7.754645672577776e-08, + "loss": 0.5534, "step": 2499 }, { - "epoch": 0.48, - "learning_rate": 1.10977457490267e-05, - "loss": 0.9152, + "epoch": 0.9619084263178146, + "grad_norm": 1.0002645768630563, + "learning_rate": 7.600524006765808e-08, + "loss": 0.563, "step": 2500 }, { - "epoch": 0.48, - "learning_rate": 1.10915511175489e-05, - "loss": 0.8298, + "epoch": 0.9622931896883417, + "grad_norm": 0.9331606038138555, + "learning_rate": 7.447943475387797e-08, + "loss": 0.5324, "step": 2501 }, { - "epoch": 0.48, - "learning_rate": 1.1085356062125225e-05, - "loss": 0.9304, + "epoch": 0.9626779530588688, + "grad_norm": 0.9174088273331613, + "learning_rate": 7.296904315391873e-08, + "loss": 0.5435, "step": 2502 }, { - "epoch": 0.48, - "learning_rate": 1.1079160585161759e-05, - "loss": 0.801, + "epoch": 0.9630627164293959, + "grad_norm": 0.9624397610148372, + "learning_rate": 7.147406761332298e-08, + "loss": 0.5648, "step": 2503 }, { - "epoch": 0.48, - "learning_rate": 1.107296468906476e-05, - "loss": 0.8892, + "epoch": 0.9634474797999231, + "grad_norm": 0.9763373895948856, + "learning_rate": 6.999451045369587e-08, + "loss": 0.5682, "step": 2504 }, { - "epoch": 0.48, - "learning_rate": 1.106676837624064e-05, - "loss": 0.8979, + "epoch": 0.9638322431704501, + "grad_norm": 1.036710204355769, + "learning_rate": 6.853037397269724e-08, + "loss": 0.5534, "step": 2505 }, { - "epoch": 0.48, - "learning_rate": 1.1060571649095972e-05, - "loss": 0.737, + "epoch": 0.9642170065409773, + "grad_norm": 0.9802597230736333, + "learning_rate": 6.70816604440383e-08, + "loss": 0.5718, "step": 2506 }, { - "epoch": 0.48, - "learning_rate": 1.10543745100375e-05, - "loss": 0.9054, + "epoch": 0.9646017699115044, + "grad_norm": 0.988169432507897, + "learning_rate": 6.564837211748054e-08, + "loss": 0.5333, "step": 2507 }, { - "epoch": 0.48, - "learning_rate": 1.1048176961472114e-05, - "loss": 0.8123, + "epoch": 0.9649865332820315, + "grad_norm": 1.0833364373960177, + "learning_rate": 6.42305112188335e-08, + "loss": 0.5627, "step": 2508 }, { - "epoch": 0.48, - "learning_rate": 1.1041979005806876e-05, - "loss": 0.9744, + "epoch": 0.9653712966525587, + "grad_norm": 0.9567273837845938, + "learning_rate": 6.282807994994477e-08, + "loss": 0.5544, "step": 2509 }, { - "epoch": 0.48, - "learning_rate": 1.1035780645449001e-05, - "loss": 0.9225, + "epoch": 0.9657560600230858, + "grad_norm": 0.9546998686032789, + "learning_rate": 6.144108048870335e-08, + "loss": 0.5486, "step": 2510 }, { - "epoch": 0.48, - "learning_rate": 1.1029581882805857e-05, - "loss": 0.988, + "epoch": 0.966140823393613, + "grad_norm": 0.9167816936092369, + "learning_rate": 6.00695149890329e-08, + "loss": 0.5723, "step": 2511 }, { - "epoch": 0.48, - "learning_rate": 1.1023382720284973e-05, - "loss": 0.958, + "epoch": 0.96652558676414, + "grad_norm": 0.9301785915032098, + "learning_rate": 5.871338558088857e-08, + "loss": 0.538, "step": 2512 }, { - "epoch": 0.48, - "learning_rate": 1.1017183160294033e-05, - "loss": 0.9405, + "epoch": 0.9669103501346672, + "grad_norm": 0.8558487586104943, + "learning_rate": 5.7372694370254614e-08, + "loss": 0.4827, "step": 2513 }, { - "epoch": 0.48, - "learning_rate": 1.1010983205240878e-05, - "loss": 0.9489, + "epoch": 0.9672951135051943, + "grad_norm": 0.9959112014253435, + "learning_rate": 5.6047443439141146e-08, + "loss": 0.5775, "step": 2514 }, { - "epoch": 0.48, - "learning_rate": 1.1004782857533488e-05, - "loss": 0.9447, + "epoch": 0.9676798768757214, + "grad_norm": 0.9391270387455319, + "learning_rate": 5.47376348455797e-08, + "loss": 0.574, "step": 2515 }, { - "epoch": 0.48, - "learning_rate": 1.099858211958002e-05, - "loss": 0.8933, + "epoch": 0.9680646402462486, + "grad_norm": 0.9964908238547098, + "learning_rate": 5.344327062362098e-08, + "loss": 0.5576, "step": 2516 }, { - "epoch": 0.48, - "learning_rate": 1.0992380993788763e-05, - "loss": 0.869, + "epoch": 0.9684494036167757, + "grad_norm": 0.9495653131108978, + "learning_rate": 5.216435278333376e-08, + "loss": 0.5244, "step": 2517 }, { - "epoch": 0.48, - "learning_rate": 1.0986179482568162e-05, - "loss": 0.9235, + "epoch": 0.9688341669873028, + "grad_norm": 0.9644613405746275, + "learning_rate": 5.0900883310794903e-08, + "loss": 0.5948, "step": 2518 }, { - "epoch": 0.48, - "learning_rate": 1.0979977588326815e-05, - "loss": 0.9627, + "epoch": 0.9692189303578299, + "grad_norm": 0.9399674319529813, + "learning_rate": 4.9652864168096e-08, + "loss": 0.5746, "step": 2519 }, { - "epoch": 0.48, - "learning_rate": 1.0973775313473465e-05, - "loss": 0.895, + "epoch": 0.9696036937283571, + "grad_norm": 0.9242622765111859, + "learning_rate": 4.84202972933312e-08, + "loss": 0.5948, "step": 2520 }, { - "epoch": 0.49, - "learning_rate": 1.0967572660417001e-05, - "loss": 0.7618, + "epoch": 0.9699884570988841, + "grad_norm": 0.9886141318234315, + "learning_rate": 4.720318460060047e-08, + "loss": 0.5604, "step": 2521 }, { - "epoch": 0.49, - "learning_rate": 1.0961369631566468e-05, - "loss": 0.9295, + "epoch": 0.9703732204694113, + "grad_norm": 0.9501817158025843, + "learning_rate": 4.6001527980004125e-08, + "loss": 0.5749, "step": 2522 }, { - "epoch": 0.49, - "learning_rate": 1.0955166229331048e-05, - "loss": 0.8043, + "epoch": 0.9707579838399384, + "grad_norm": 0.9177390133542773, + "learning_rate": 4.4815329297639434e-08, + "loss": 0.5137, "step": 2523 }, { - "epoch": 0.49, - "learning_rate": 1.0948962456120068e-05, - "loss": 0.8703, + "epoch": 0.9711427472104656, + "grad_norm": 0.9466367052800572, + "learning_rate": 4.364459039559843e-08, + "loss": 0.5771, "step": 2524 }, { - "epoch": 0.49, - "learning_rate": 1.0942758314343007e-05, - "loss": 0.872, + "epoch": 0.9715275105809927, + "grad_norm": 0.931400397469538, + "learning_rate": 4.248931309196791e-08, + "loss": 0.5792, "step": 2525 }, { - "epoch": 0.49, - "learning_rate": 1.0936553806409482e-05, - "loss": 0.9321, + "epoch": 0.9719122739515198, + "grad_norm": 0.8815526929260644, + "learning_rate": 4.134949918081832e-08, + "loss": 0.5454, "step": 2526 }, { - "epoch": 0.49, - "learning_rate": 1.0930348934729249e-05, - "loss": 0.9532, + "epoch": 0.972297037322047, + "grad_norm": 0.9135022420743383, + "learning_rate": 4.022515043221154e-08, + "loss": 0.6054, "step": 2527 }, { - "epoch": 0.49, - "learning_rate": 1.0924143701712211e-05, - "loss": 0.9672, + "epoch": 0.972681800692574, + "grad_norm": 1.000062785306744, + "learning_rate": 3.9116268592189755e-08, + "loss": 0.585, "step": 2528 }, { - "epoch": 0.49, - "learning_rate": 1.0917938109768404e-05, - "loss": 0.9489, + "epoch": 0.9730665640631012, + "grad_norm": 0.9712621324985554, + "learning_rate": 3.802285538277772e-08, + "loss": 0.5853, "step": 2529 }, { - "epoch": 0.49, - "learning_rate": 1.0911732161308014e-05, - "loss": 0.8795, + "epoch": 0.9734513274336283, + "grad_norm": 0.9158482246439126, + "learning_rate": 3.69449125019794e-08, + "loss": 0.5491, "step": 2530 }, { - "epoch": 0.49, - "learning_rate": 1.0905525858741364e-05, - "loss": 0.9509, + "epoch": 0.9738360908041555, + "grad_norm": 0.9387718123008233, + "learning_rate": 3.588244162377019e-08, + "loss": 0.5579, "step": 2531 }, { - "epoch": 0.49, - "learning_rate": 1.08993192044789e-05, - "loss": 0.9157, + "epoch": 0.9742208541746825, + "grad_norm": 0.5516989490763526, + "learning_rate": 3.483544439810249e-08, + "loss": 0.6279, "step": 2532 }, { - "epoch": 0.49, - "learning_rate": 1.089311220093122e-05, - "loss": 0.9226, + "epoch": 0.9746056175452097, + "grad_norm": 0.9527036709186201, + "learning_rate": 3.3803922450897917e-08, + "loss": 0.5831, "step": 2533 }, { - "epoch": 0.49, - "learning_rate": 1.0886904850509052e-05, - "loss": 0.9142, + "epoch": 0.9749903809157369, + "grad_norm": 0.9861948381823205, + "learning_rate": 3.2787877384045095e-08, + "loss": 0.5005, "step": 2534 }, { - "epoch": 0.49, - "learning_rate": 1.0880697155623264e-05, - "loss": 0.9483, + "epoch": 0.9753751442862639, + "grad_norm": 0.9248127947696587, + "learning_rate": 3.178731077539743e-08, + "loss": 0.6038, "step": 2535 }, { - "epoch": 0.49, - "learning_rate": 1.0874489118684846e-05, - "loss": 0.8627, + "epoch": 0.9757599076567911, + "grad_norm": 1.1386800672125237, + "learning_rate": 3.080222417877421e-08, + "loss": 0.5836, "step": 2536 }, { - "epoch": 0.49, - "learning_rate": 1.086828074210493e-05, - "loss": 0.9211, + "epoch": 0.9761446710273182, + "grad_norm": 1.7447500521989872, + "learning_rate": 2.983261912395397e-08, + "loss": 0.6411, "step": 2537 }, { - "epoch": 0.49, - "learning_rate": 1.0862072028294777e-05, - "loss": 0.9706, + "epoch": 0.9765294343978453, + "grad_norm": 0.9599933634188371, + "learning_rate": 2.8878497116671124e-08, + "loss": 0.5211, "step": 2538 }, { - "epoch": 0.49, - "learning_rate": 1.0855862979665788e-05, - "loss": 0.8589, + "epoch": 0.9769141977683724, + "grad_norm": 0.9885209812695472, + "learning_rate": 2.7939859638617118e-08, + "loss": 0.5824, "step": 2539 }, { - "epoch": 0.49, - "learning_rate": 1.0849653598629477e-05, - "loss": 0.8515, + "epoch": 0.9772989611388996, + "grad_norm": 0.9009848474855839, + "learning_rate": 2.7016708147439285e-08, + "loss": 0.5484, "step": 2540 }, { - "epoch": 0.49, - "learning_rate": 1.0843443887597495e-05, - "loss": 0.9333, + "epoch": 0.9776837245094268, + "grad_norm": 0.9578691745800283, + "learning_rate": 2.6109044076733092e-08, + "loss": 0.5889, "step": 2541 }, { - "epoch": 0.49, - "learning_rate": 1.0837233848981632e-05, - "loss": 0.8934, + "epoch": 0.9780684878799538, + "grad_norm": 1.0371294246112601, + "learning_rate": 2.5216868836043242e-08, + "loss": 0.5157, "step": 2542 }, { - "epoch": 0.49, - "learning_rate": 1.0831023485193787e-05, - "loss": 0.9063, + "epoch": 0.978453251250481, + "grad_norm": 0.9348288294789964, + "learning_rate": 2.434018381086589e-08, + "loss": 0.5924, "step": 2543 }, { - "epoch": 0.49, - "learning_rate": 1.0824812798645997e-05, - "loss": 0.9464, + "epoch": 0.9788380146210081, + "grad_norm": 0.9391018409456195, + "learning_rate": 2.3478990362634235e-08, + "loss": 0.5444, "step": 2544 }, { - "epoch": 0.49, - "learning_rate": 1.0818601791750418e-05, - "loss": 0.9116, + "epoch": 0.9792227779915352, + "grad_norm": 0.9966931004061198, + "learning_rate": 2.263328982872959e-08, + "loss": 0.55, "step": 2545 }, { - "epoch": 0.49, - "learning_rate": 1.0812390466919337e-05, - "loss": 0.8928, + "epoch": 0.9796075413620623, + "grad_norm": 0.9766657009711533, + "learning_rate": 2.1803083522471402e-08, + "loss": 0.5859, "step": 2546 }, { - "epoch": 0.49, - "learning_rate": 1.0806178826565162e-05, - "loss": 0.9354, + "epoch": 0.9799923047325895, + "grad_norm": 1.064087215695503, + "learning_rate": 2.098837273311838e-08, + "loss": 0.6189, "step": 2547 }, { - "epoch": 0.49, - "learning_rate": 1.0799966873100419e-05, - "loss": 0.9032, + "epoch": 0.9803770681031165, + "grad_norm": 1.0299033382948126, + "learning_rate": 2.0189158725867353e-08, + "loss": 0.5238, "step": 2548 }, { - "epoch": 0.49, - "learning_rate": 1.0793754608937758e-05, - "loss": 0.8513, + "epoch": 0.9807618314736437, + "grad_norm": 0.9874427979646306, + "learning_rate": 1.9405442741844415e-08, + "loss": 0.4939, "step": 2549 }, { - "epoch": 0.49, - "learning_rate": 1.0787542036489955e-05, - "loss": 0.9263, + "epoch": 0.9811465948441709, + "grad_norm": 0.9345698132550467, + "learning_rate": 1.8637225998114904e-08, + "loss": 0.557, "step": 2550 }, { - "epoch": 0.49, - "learning_rate": 1.0781329158169902e-05, - "loss": 0.906, + "epoch": 0.981531358214698, + "grad_norm": 0.9727496303836601, + "learning_rate": 1.7884509687668972e-08, + "loss": 0.5907, "step": 2551 }, { - "epoch": 0.49, - "learning_rate": 1.0775115976390607e-05, - "loss": 0.9531, + "epoch": 0.9819161215852251, + "grad_norm": 0.8943064749361685, + "learning_rate": 1.714729497942935e-08, + "loss": 0.5742, "step": 2552 }, { - "epoch": 0.49, - "learning_rate": 1.0768902493565197e-05, - "loss": 0.789, + "epoch": 0.9823008849557522, + "grad_norm": 0.9479115342496424, + "learning_rate": 1.6425583018244706e-08, + "loss": 0.5857, "step": 2553 }, { - "epoch": 0.49, - "learning_rate": 1.0762688712106918e-05, - "loss": 0.9149, + "epoch": 0.9826856483262794, + "grad_norm": 1.0797448738055673, + "learning_rate": 1.57193749248874e-08, + "loss": 0.5223, "step": 2554 }, { - "epoch": 0.49, - "learning_rate": 1.0756474634429133e-05, - "loss": 0.9444, + "epoch": 0.9830704116968064, + "grad_norm": 0.5810670738407322, + "learning_rate": 1.5028671796055715e-08, + "loss": 0.6582, "step": 2555 }, { - "epoch": 0.49, - "learning_rate": 1.0750260262945314e-05, - "loss": 0.9201, + "epoch": 0.9834551750673336, + "grad_norm": 0.9131383721538058, + "learning_rate": 1.435347470436832e-08, + "loss": 0.5872, "step": 2556 }, { - "epoch": 0.49, - "learning_rate": 1.0744045600069055e-05, - "loss": 0.7132, + "epoch": 0.9838399384378607, + "grad_norm": 0.9546732547427607, + "learning_rate": 1.3693784698363133e-08, + "loss": 0.4958, "step": 2557 }, { - "epoch": 0.49, - "learning_rate": 1.0737830648214063e-05, - "loss": 0.9004, + "epoch": 0.9842247018083878, + "grad_norm": 0.9661106940598758, + "learning_rate": 1.3049602802498451e-08, + "loss": 0.5838, "step": 2558 }, { - "epoch": 0.49, - "learning_rate": 1.0731615409794144e-05, - "loss": 0.8789, + "epoch": 0.984609465178915, + "grad_norm": 1.0908579737622115, + "learning_rate": 1.2420930017148503e-08, + "loss": 0.5875, "step": 2559 }, { - "epoch": 0.49, - "learning_rate": 1.0725399887223234e-05, - "loss": 0.9207, + "epoch": 0.9849942285494421, + "grad_norm": 0.9971515274723757, + "learning_rate": 1.1807767318602337e-08, + "loss": 0.5975, "step": 2560 }, { - "epoch": 0.49, - "learning_rate": 1.0719184082915364e-05, - "loss": 0.8507, + "epoch": 0.9853789919199692, + "grad_norm": 0.967042023565226, + "learning_rate": 1.1210115659063825e-08, + "loss": 0.5978, "step": 2561 }, { - "epoch": 0.49, - "learning_rate": 1.0712967999284682e-05, - "loss": 0.9276, + "epoch": 0.9857637552904963, + "grad_norm": 0.9864991921117112, + "learning_rate": 1.0627975966649439e-08, + "loss": 0.5114, "step": 2562 }, { - "epoch": 0.49, - "learning_rate": 1.0706751638745448e-05, - "loss": 0.7971, + "epoch": 0.9861485186610235, + "grad_norm": 0.9786355932023123, + "learning_rate": 1.0061349145383814e-08, + "loss": 0.5711, "step": 2563 }, { - "epoch": 0.49, - "learning_rate": 1.0700535003712023e-05, - "loss": 0.8737, + "epoch": 0.9865332820315506, + "grad_norm": 1.0716691024677867, + "learning_rate": 9.510236075205292e-09, + "loss": 0.617, "step": 2564 }, { - "epoch": 0.49, - "learning_rate": 1.069431809659887e-05, - "loss": 0.8442, + "epoch": 0.9869180454020777, + "grad_norm": 0.9668047564861693, + "learning_rate": 8.974637611955939e-09, + "loss": 0.5599, "step": 2565 }, { - "epoch": 0.49, - "learning_rate": 1.068810091982057e-05, - "loss": 0.8865, + "epoch": 0.9873028087726049, + "grad_norm": 0.9239837932871595, + "learning_rate": 8.454554587388198e-09, + "loss": 0.5488, "step": 2566 }, { - "epoch": 0.49, - "learning_rate": 1.0681883475791803e-05, - "loss": 0.9736, + "epoch": 0.987687572143132, + "grad_norm": 0.9401705996589803, + "learning_rate": 7.949987809158232e-09, + "loss": 0.5337, "step": 2567 }, { - "epoch": 0.49, - "learning_rate": 1.067566576692735e-05, - "loss": 0.865, + "epoch": 0.9880723355136591, + "grad_norm": 0.9424227709345321, + "learning_rate": 7.460938060825929e-09, + "loss": 0.5756, "step": 2568 }, { - "epoch": 0.49, - "learning_rate": 1.0669447795642103e-05, - "loss": 0.9312, + "epoch": 0.9884570988841862, + "grad_norm": 0.5784671376713784, + "learning_rate": 6.987406101855998e-09, + "loss": 0.6428, "step": 2569 }, { - "epoch": 0.49, - "learning_rate": 1.066322956435104e-05, - "loss": 0.8829, + "epoch": 0.9888418622547134, + "grad_norm": 0.983717299309872, + "learning_rate": 6.5293926676135434e-09, + "loss": 0.631, "step": 2570 }, { - "epoch": 0.49, - "learning_rate": 1.065701107546926e-05, - "loss": 0.8815, + "epoch": 0.9892266256252404, + "grad_norm": 0.9655066204096425, + "learning_rate": 6.086898469365166e-09, + "loss": 0.591, "step": 2571 }, { - "epoch": 0.49, - "learning_rate": 1.065079233141195e-05, - "loss": 0.9021, + "epoch": 0.9896113889957676, + "grad_norm": 0.9849729118965784, + "learning_rate": 5.6599241942767445e-09, + "loss": 0.5561, "step": 2572 }, { - "epoch": 0.5, - "learning_rate": 1.0644573334594395e-05, - "loss": 0.8829, + "epoch": 0.9899961523662947, + "grad_norm": 0.9747954223145545, + "learning_rate": 5.248470505412328e-09, + "loss": 0.595, "step": 2573 }, { - "epoch": 0.5, - "learning_rate": 1.0638354087431986e-05, - "loss": 0.8737, + "epoch": 0.9903809157368219, + "grad_norm": 1.0337515072456762, + "learning_rate": 4.8525380417330234e-09, + "loss": 0.5435, "step": 2574 }, { - "epoch": 0.5, - "learning_rate": 1.0632134592340204e-05, - "loss": 0.8719, + "epoch": 0.990765679107349, + "grad_norm": 0.6585800561913944, + "learning_rate": 4.472127418099215e-09, + "loss": 0.6634, "step": 2575 }, { - "epoch": 0.5, - "learning_rate": 1.0625914851734632e-05, - "loss": 0.8458, + "epoch": 0.9911504424778761, + "grad_norm": 0.94201695094983, + "learning_rate": 4.1072392252639034e-09, + "loss": 0.5728, "step": 2576 }, { - "epoch": 0.5, - "learning_rate": 1.0619694868030943e-05, - "loss": 0.9642, + "epoch": 0.9915352058484033, + "grad_norm": 1.1288486416664785, + "learning_rate": 3.757874029874931e-09, + "loss": 0.6016, "step": 2577 }, { - "epoch": 0.5, - "learning_rate": 1.0613474643644907e-05, - "loss": 0.8763, + "epoch": 0.9919199692189303, + "grad_norm": 0.9824014232632274, + "learning_rate": 3.424032374476083e-09, + "loss": 0.5739, "step": 2578 }, { - "epoch": 0.5, - "learning_rate": 1.0607254180992391e-05, - "loss": 0.8641, + "epoch": 0.9923047325894575, + "grad_norm": 0.5929040842369449, + "learning_rate": 3.105714777501545e-09, + "loss": 0.6382, "step": 2579 }, { - "epoch": 0.5, - "learning_rate": 1.0601033482489346e-05, - "loss": 0.8262, + "epoch": 0.9926894959599846, + "grad_norm": 0.9777079978640617, + "learning_rate": 2.802921733278119e-09, + "loss": 0.5375, "step": 2580 }, { - "epoch": 0.5, - "learning_rate": 1.0594812550551826e-05, - "loss": 0.8871, + "epoch": 0.9930742593305117, + "grad_norm": 0.9131127070107397, + "learning_rate": 2.5156537120263335e-09, + "loss": 0.567, "step": 2581 }, { - "epoch": 0.5, - "learning_rate": 1.058859138759596e-05, - "loss": 0.8811, + "epoch": 0.9934590227010388, + "grad_norm": 0.9214472559606114, + "learning_rate": 2.2439111598537844e-09, + "loss": 0.5826, "step": 2582 }, { - "epoch": 0.5, - "learning_rate": 1.0582369996037985e-05, - "loss": 0.8303, + "epoch": 0.993843786071566, + "grad_norm": 0.8973570465735903, + "learning_rate": 1.987694498760684e-09, + "loss": 0.532, "step": 2583 }, { - "epoch": 0.5, - "learning_rate": 1.0576148378294213e-05, - "loss": 0.9375, + "epoch": 0.9942285494420932, + "grad_norm": 0.9176639752914193, + "learning_rate": 1.747004126635421e-09, + "loss": 0.5718, "step": 2584 }, { - "epoch": 0.5, - "learning_rate": 1.056992653678105e-05, - "loss": 0.9382, + "epoch": 0.9946133128126202, + "grad_norm": 0.9305526867694586, + "learning_rate": 1.5218404172545609e-09, + "loss": 0.5595, "step": 2585 }, { - "epoch": 0.5, - "learning_rate": 1.0563704473914986e-05, - "loss": 0.9126, + "epoch": 0.9949980761831474, + "grad_norm": 0.973482240615476, + "learning_rate": 1.3122037202828452e-09, + "loss": 0.5601, "step": 2586 }, { - "epoch": 0.5, - "learning_rate": 1.0557482192112603e-05, - "loss": 0.874, + "epoch": 0.9953828395536745, + "grad_norm": 0.9630643200164869, + "learning_rate": 1.1180943612754124e-09, + "loss": 0.5677, "step": 2587 }, { - "epoch": 0.5, - "learning_rate": 1.0551259693790556e-05, - "loss": 0.9294, + "epoch": 0.9957676029242016, + "grad_norm": 0.9980392563615073, + "learning_rate": 9.39512641668916e-10, + "loss": 0.5626, "step": 2588 }, { - "epoch": 0.5, - "learning_rate": 1.0545036981365601e-05, - "loss": 0.8191, + "epoch": 0.9961523662947287, + "grad_norm": 0.9873847173258852, + "learning_rate": 7.764588387915161e-10, + "loss": 0.5686, "step": 2589 }, { - "epoch": 0.5, - "learning_rate": 1.053881405725456e-05, - "loss": 0.8146, + "epoch": 0.9965371296652559, + "grad_norm": 0.9000346801702849, + "learning_rate": 6.289332058551089e-10, + "loss": 0.5136, "step": 2590 }, { - "epoch": 0.5, - "learning_rate": 1.0532590923874349e-05, - "loss": 0.8885, + "epoch": 0.996921893035783, + "grad_norm": 0.9073281554688536, + "learning_rate": 4.969359719586563e-10, + "loss": 0.5564, "step": 2591 }, { - "epoch": 0.5, - "learning_rate": 1.0526367583641958e-05, - "loss": 0.9843, + "epoch": 0.9973066564063101, + "grad_norm": 0.9919749874532774, + "learning_rate": 3.804673420837457e-10, + "loss": 0.5749, "step": 2592 }, { - "epoch": 0.5, - "learning_rate": 1.0520144038974468e-05, - "loss": 0.8852, + "epoch": 0.9976914197768373, + "grad_norm": 0.9248065562127321, + "learning_rate": 2.795274971001405e-10, + "loss": 0.5228, "step": 2593 }, { - "epoch": 0.5, - "learning_rate": 1.0513920292289021e-05, - "loss": 0.9148, + "epoch": 0.9980761831473643, + "grad_norm": 0.9526152166900259, + "learning_rate": 1.941165937602296e-10, + "loss": 0.5528, "step": 2594 }, { - "epoch": 0.5, - "learning_rate": 1.0507696346002857e-05, - "loss": 0.9313, + "epoch": 0.9984609465178915, + "grad_norm": 1.0143783746533086, + "learning_rate": 1.2423476470346808e-10, + "loss": 0.5643, "step": 2595 }, { - "epoch": 0.5, - "learning_rate": 1.0501472202533285e-05, - "loss": 0.9493, + "epoch": 0.9988457098884186, + "grad_norm": 0.9223740266372356, + "learning_rate": 6.988211845082582e-11, + "loss": 0.5213, "step": 2596 }, { - "epoch": 0.5, - "learning_rate": 1.0495247864297684e-05, - "loss": 0.8774, + "epoch": 0.9992304732589458, + "grad_norm": 0.9992313144738935, + "learning_rate": 3.105873940811854e-11, + "loss": 0.5608, "step": 2597 }, { - "epoch": 0.5, - "learning_rate": 1.0489023333713522e-05, - "loss": 0.8998, + "epoch": 0.9996152366294728, + "grad_norm": 0.825503170967095, + "learning_rate": 7.764687866007592e-12, + "loss": 0.4516, "step": 2598 }, { - "epoch": 0.5, - "learning_rate": 1.0482798613198328e-05, - "loss": 0.8051, + "epoch": 1.0, + "grad_norm": 0.7219915244027095, + "learning_rate": 0.0, + "loss": 0.6351, "step": 2599 }, - { - "epoch": 0.5, - "learning_rate": 1.047657370516972e-05, - "loss": 0.9107, - "step": 2600 - }, - { - "epoch": 0.5, - "learning_rate": 1.0470348612045376e-05, - "loss": 0.8879, - "step": 2601 - }, - { - "epoch": 0.5, - "learning_rate": 1.0464123336243049e-05, - "loss": 0.7396, - "step": 2602 - }, - { - "epoch": 0.5, - "learning_rate": 1.0457897880180566e-05, - "loss": 0.961, - "step": 2603 - }, - { - "epoch": 0.5, - "learning_rate": 1.0451672246275826e-05, - "loss": 0.8591, - "step": 2604 - }, - { - "epoch": 0.5, - "learning_rate": 1.0445446436946788e-05, - "loss": 0.8529, - "step": 2605 - }, - { - "epoch": 0.5, - "learning_rate": 1.0439220454611486e-05, - "loss": 0.9136, - "step": 2606 - }, - { - "epoch": 0.5, - "learning_rate": 1.0432994301688021e-05, - "loss": 0.8898, - "step": 2607 - }, - { - "epoch": 0.5, - "learning_rate": 1.0426767980594559e-05, - "loss": 0.8746, - "step": 2608 - }, - { - "epoch": 0.5, - "learning_rate": 1.0420541493749332e-05, - "loss": 0.9712, - "step": 2609 - }, - { - "epoch": 0.5, - "learning_rate": 1.0414314843570634e-05, - "loss": 0.7965, - "step": 2610 - }, - { - "epoch": 0.5, - "learning_rate": 1.0408088032476822e-05, - "loss": 0.8847, - "step": 2611 - }, - { - "epoch": 0.5, - "learning_rate": 1.0401861062886324e-05, - "loss": 0.8674, - "step": 2612 - }, - { - "epoch": 0.5, - "learning_rate": 1.0395633937217622e-05, - "loss": 0.8603, - "step": 2613 - }, - { - "epoch": 0.5, - "learning_rate": 1.0389406657889254e-05, - "loss": 0.8807, - "step": 2614 - }, - { - "epoch": 0.5, - "learning_rate": 1.0383179227319826e-05, - "loss": 0.6413, - "step": 2615 - }, - { - "epoch": 0.5, - "learning_rate": 1.0376951647928007e-05, - "loss": 0.9321, - "step": 2616 - }, - { - "epoch": 0.5, - "learning_rate": 1.0370723922132506e-05, - "loss": 0.9008, - "step": 2617 - }, - { - "epoch": 0.5, - "learning_rate": 1.036449605235211e-05, - "loss": 0.9228, - "step": 2618 - }, - { - "epoch": 0.5, - "learning_rate": 1.0358268041005644e-05, - "loss": 0.9537, - "step": 2619 - }, - { - "epoch": 0.5, - "learning_rate": 1.0352039890511997e-05, - "loss": 0.8095, - "step": 2620 - }, - { - "epoch": 0.5, - "learning_rate": 1.034581160329012e-05, - "loss": 0.8423, - "step": 2621 - }, - { - "epoch": 0.5, - "learning_rate": 1.0339583181758997e-05, - "loss": 0.8943, - "step": 2622 - }, - { - "epoch": 0.5, - "learning_rate": 1.033335462833768e-05, - "loss": 0.9486, - "step": 2623 - }, - { - "epoch": 0.5, - "learning_rate": 1.0327125945445265e-05, - "loss": 0.9055, - "step": 2624 - }, - { - "epoch": 0.51, - "learning_rate": 1.0320897135500904e-05, - "loss": 0.8582, - "step": 2625 - }, - { - "epoch": 0.51, - "learning_rate": 1.0314668200923791e-05, - "loss": 0.8645, - "step": 2626 - }, - { - "epoch": 0.51, - "learning_rate": 1.0308439144133177e-05, - "loss": 0.7424, - "step": 2627 - }, - { - "epoch": 0.51, - "learning_rate": 1.0302209967548354e-05, - "loss": 0.8723, - "step": 2628 - }, - { - "epoch": 0.51, - "learning_rate": 1.029598067358866e-05, - "loss": 0.8841, - "step": 2629 - }, - { - "epoch": 0.51, - "learning_rate": 1.0289751264673485e-05, - "loss": 0.6669, - "step": 2630 - }, - { - "epoch": 0.51, - "learning_rate": 1.0283521743222256e-05, - "loss": 0.8632, - "step": 2631 - }, - { - "epoch": 0.51, - "learning_rate": 1.0277292111654447e-05, - "loss": 0.9546, - "step": 2632 - }, - { - "epoch": 0.51, - "learning_rate": 1.0271062372389582e-05, - "loss": 0.8899, - "step": 2633 - }, - { - "epoch": 0.51, - "learning_rate": 1.0264832527847212e-05, - "loss": 0.899, - "step": 2634 - }, - { - "epoch": 0.51, - "learning_rate": 1.0258602580446941e-05, - "loss": 0.8708, - "step": 2635 - }, - { - "epoch": 0.51, - "learning_rate": 1.0252372532608405e-05, - "loss": 0.8359, - "step": 2636 - }, - { - "epoch": 0.51, - "learning_rate": 1.024614238675129e-05, - "loss": 0.9684, - "step": 2637 - }, - { - "epoch": 0.51, - "learning_rate": 1.0239912145295303e-05, - "loss": 0.8886, - "step": 2638 - }, - { - "epoch": 0.51, - "learning_rate": 1.0233681810660207e-05, - "loss": 0.8984, - "step": 2639 - }, - { - "epoch": 0.51, - "learning_rate": 1.0227451385265788e-05, - "loss": 0.8524, - "step": 2640 - }, - { - "epoch": 0.51, - "learning_rate": 1.022122087153187e-05, - "loss": 0.9784, - "step": 2641 - }, - { - "epoch": 0.51, - "learning_rate": 1.0214990271878319e-05, - "loss": 0.9221, - "step": 2642 - }, - { - "epoch": 0.51, - "learning_rate": 1.0208759588725016e-05, - "loss": 0.9515, - "step": 2643 - }, - { - "epoch": 0.51, - "learning_rate": 1.0202528824491899e-05, - "loss": 0.8817, - "step": 2644 - }, - { - "epoch": 0.51, - "learning_rate": 1.0196297981598921e-05, - "loss": 0.8611, - "step": 2645 - }, - { - "epoch": 0.51, - "learning_rate": 1.019006706246607e-05, - "loss": 0.8295, - "step": 2646 - }, - { - "epoch": 0.51, - "learning_rate": 1.018383606951336e-05, - "loss": 0.9246, - "step": 2647 - }, - { - "epoch": 0.51, - "learning_rate": 1.0177605005160837e-05, - "loss": 0.8336, - "step": 2648 - }, - { - "epoch": 0.51, - "learning_rate": 1.0171373871828578e-05, - "loss": 0.8877, - "step": 2649 - }, - { - "epoch": 0.51, - "learning_rate": 1.0165142671936685e-05, - "loss": 0.8724, - "step": 2650 - }, - { - "epoch": 0.51, - "learning_rate": 1.0158911407905279e-05, - "loss": 0.8884, - "step": 2651 - }, - { - "epoch": 0.51, - "learning_rate": 1.0152680082154514e-05, - "loss": 0.8977, - "step": 2652 - }, - { - "epoch": 0.51, - "learning_rate": 1.0146448697104561e-05, - "loss": 0.9207, - "step": 2653 - }, - { - "epoch": 0.51, - "learning_rate": 1.0140217255175626e-05, - "loss": 0.8673, - "step": 2654 - }, - { - "epoch": 0.51, - "learning_rate": 1.013398575878792e-05, - "loss": 0.674, - "step": 2655 - }, - { - "epoch": 0.51, - "learning_rate": 1.0127754210361694e-05, - "loss": 0.8563, - "step": 2656 - }, - { - "epoch": 0.51, - "learning_rate": 1.0121522612317204e-05, - "loss": 0.9287, - "step": 2657 - }, - { - "epoch": 0.51, - "learning_rate": 1.011529096707473e-05, - "loss": 0.8385, - "step": 2658 - }, - { - "epoch": 0.51, - "learning_rate": 1.0109059277054574e-05, - "loss": 0.6725, - "step": 2659 - }, - { - "epoch": 0.51, - "learning_rate": 1.010282754467705e-05, - "loss": 0.9646, - "step": 2660 - }, - { - "epoch": 0.51, - "learning_rate": 1.0096595772362492e-05, - "loss": 0.8682, - "step": 2661 - }, - { - "epoch": 0.51, - "learning_rate": 1.0090363962531251e-05, - "loss": 0.7097, - "step": 2662 - }, - { - "epoch": 0.51, - "learning_rate": 1.0084132117603689e-05, - "loss": 0.8809, - "step": 2663 - }, - { - "epoch": 0.51, - "learning_rate": 1.0077900240000181e-05, - "loss": 0.8489, - "step": 2664 - }, - { - "epoch": 0.51, - "learning_rate": 1.0071668332141115e-05, - "loss": 0.9233, - "step": 2665 - }, - { - "epoch": 0.51, - "learning_rate": 1.0065436396446899e-05, - "loss": 0.8542, - "step": 2666 - }, - { - "epoch": 0.51, - "learning_rate": 1.0059204435337938e-05, - "loss": 0.9515, - "step": 2667 - }, - { - "epoch": 0.51, - "learning_rate": 1.0052972451234656e-05, - "loss": 0.8501, - "step": 2668 - }, - { - "epoch": 0.51, - "learning_rate": 1.0046740446557485e-05, - "loss": 0.8843, - "step": 2669 - }, - { - "epoch": 0.51, - "learning_rate": 1.0040508423726865e-05, - "loss": 0.8746, - "step": 2670 - }, - { - "epoch": 0.51, - "learning_rate": 1.0034276385163238e-05, - "loss": 0.8681, - "step": 2671 - }, - { - "epoch": 0.51, - "learning_rate": 1.0028044333287056e-05, - "loss": 0.7283, - "step": 2672 - }, - { - "epoch": 0.51, - "learning_rate": 1.002181227051878e-05, - "loss": 0.8585, - "step": 2673 - }, - { - "epoch": 0.51, - "learning_rate": 1.0015580199278873e-05, - "loss": 0.89, - "step": 2674 - }, - { - "epoch": 0.51, - "learning_rate": 1.0009348121987795e-05, - "loss": 0.9339, - "step": 2675 - }, - { - "epoch": 0.51, - "learning_rate": 1.000311604106601e-05, - "loss": 0.9516, - "step": 2676 - }, - { - "epoch": 0.52, - "learning_rate": 9.996883958933993e-06, - "loss": 0.8818, - "step": 2677 - }, - { - "epoch": 0.52, - "learning_rate": 9.99065187801221e-06, - "loss": 0.9021, - "step": 2678 - }, - { - "epoch": 0.52, - "learning_rate": 9.984419800721132e-06, - "loss": 0.8039, - "step": 2679 - }, - { - "epoch": 0.52, - "learning_rate": 9.978187729481218e-06, - "loss": 0.8162, - "step": 2680 - }, - { - "epoch": 0.52, - "learning_rate": 9.971955666712945e-06, - "loss": 1.0008, - "step": 2681 - }, - { - "epoch": 0.52, - "learning_rate": 9.965723614836764e-06, - "loss": 0.7998, - "step": 2682 - }, - { - "epoch": 0.52, - "learning_rate": 9.959491576273139e-06, - "loss": 0.9975, - "step": 2683 - }, - { - "epoch": 0.52, - "learning_rate": 9.95325955344252e-06, - "loss": 0.9062, - "step": 2684 - }, - { - "epoch": 0.52, - "learning_rate": 9.947027548765347e-06, - "loss": 0.8812, - "step": 2685 - }, - { - "epoch": 0.52, - "learning_rate": 9.940795564662064e-06, - "loss": 0.9445, - "step": 2686 - }, - { - "epoch": 0.52, - "learning_rate": 9.934563603553103e-06, - "loss": 0.8419, - "step": 2687 - }, - { - "epoch": 0.52, - "learning_rate": 9.928331667858886e-06, - "loss": 0.8824, - "step": 2688 - }, - { - "epoch": 0.52, - "learning_rate": 9.922099759999822e-06, - "loss": 0.8727, - "step": 2689 - }, - { - "epoch": 0.52, - "learning_rate": 9.915867882396314e-06, - "loss": 0.9598, - "step": 2690 - }, - { - "epoch": 0.52, - "learning_rate": 9.909636037468754e-06, - "loss": 0.8901, - "step": 2691 - }, - { - "epoch": 0.52, - "learning_rate": 9.90340422763751e-06, - "loss": 0.9542, - "step": 2692 - }, - { - "epoch": 0.52, - "learning_rate": 9.897172455322953e-06, - "loss": 0.8927, - "step": 2693 - }, - { - "epoch": 0.52, - "learning_rate": 9.890940722945429e-06, - "loss": 0.8445, - "step": 2694 - }, - { - "epoch": 0.52, - "learning_rate": 9.884709032925274e-06, - "loss": 0.914, - "step": 2695 - }, - { - "epoch": 0.52, - "learning_rate": 9.878477387682801e-06, - "loss": 0.8054, - "step": 2696 - }, - { - "epoch": 0.52, - "learning_rate": 9.872245789638308e-06, - "loss": 0.7825, - "step": 2697 - }, - { - "epoch": 0.52, - "learning_rate": 9.866014241212078e-06, - "loss": 0.9566, - "step": 2698 - }, - { - "epoch": 0.52, - "learning_rate": 9.859782744824376e-06, - "loss": 0.9201, - "step": 2699 - }, - { - "epoch": 0.52, - "learning_rate": 9.85355130289544e-06, - "loss": 0.9173, - "step": 2700 - }, - { - "epoch": 0.52, - "learning_rate": 9.84731991784549e-06, - "loss": 0.9392, - "step": 2701 - }, - { - "epoch": 0.52, - "learning_rate": 9.841088592094726e-06, - "loss": 0.8972, - "step": 2702 - }, - { - "epoch": 0.52, - "learning_rate": 9.834857328063316e-06, - "loss": 0.8721, - "step": 2703 - }, - { - "epoch": 0.52, - "learning_rate": 9.828626128171422e-06, - "loss": 0.8911, - "step": 2704 - }, - { - "epoch": 0.52, - "learning_rate": 9.822394994839164e-06, - "loss": 0.8963, - "step": 2705 - }, - { - "epoch": 0.52, - "learning_rate": 9.816163930486643e-06, - "loss": 0.9469, - "step": 2706 - }, - { - "epoch": 0.52, - "learning_rate": 9.809932937533935e-06, - "loss": 0.8921, - "step": 2707 - }, - { - "epoch": 0.52, - "learning_rate": 9.803702018401084e-06, - "loss": 0.861, - "step": 2708 - }, - { - "epoch": 0.52, - "learning_rate": 9.797471175508101e-06, - "loss": 0.9071, - "step": 2709 - }, - { - "epoch": 0.52, - "learning_rate": 9.791240411274982e-06, - "loss": 0.887, - "step": 2710 - }, - { - "epoch": 0.52, - "learning_rate": 9.785009728121686e-06, - "loss": 0.9132, - "step": 2711 - }, - { - "epoch": 0.52, - "learning_rate": 9.778779128468133e-06, - "loss": 0.8712, - "step": 2712 - }, - { - "epoch": 0.52, - "learning_rate": 9.772548614734217e-06, - "loss": 0.8778, - "step": 2713 - }, - { - "epoch": 0.52, - "learning_rate": 9.766318189339798e-06, - "loss": 0.895, - "step": 2714 - }, - { - "epoch": 0.52, - "learning_rate": 9.760087854704697e-06, - "loss": 0.8739, - "step": 2715 - }, - { - "epoch": 0.52, - "learning_rate": 9.753857613248714e-06, - "loss": 0.8819, - "step": 2716 - }, - { - "epoch": 0.52, - "learning_rate": 9.747627467391596e-06, - "loss": 0.8943, - "step": 2717 - }, - { - "epoch": 0.52, - "learning_rate": 9.741397419553062e-06, - "loss": 0.8469, - "step": 2718 - }, - { - "epoch": 0.52, - "learning_rate": 9.735167472152793e-06, - "loss": 0.9168, - "step": 2719 - }, - { - "epoch": 0.52, - "learning_rate": 9.728937627610425e-06, - "loss": 0.8941, - "step": 2720 - }, - { - "epoch": 0.52, - "learning_rate": 9.722707888345553e-06, - "loss": 0.9213, - "step": 2721 - }, - { - "epoch": 0.52, - "learning_rate": 9.716478256777749e-06, - "loss": 0.897, - "step": 2722 - }, - { - "epoch": 0.52, - "learning_rate": 9.710248735326519e-06, - "loss": 0.8588, - "step": 2723 - }, - { - "epoch": 0.52, - "learning_rate": 9.704019326411344e-06, - "loss": 0.8661, - "step": 2724 - }, - { - "epoch": 0.52, - "learning_rate": 9.697790032451651e-06, - "loss": 0.8369, - "step": 2725 - }, - { - "epoch": 0.52, - "learning_rate": 9.691560855866826e-06, - "loss": 0.9247, - "step": 2726 - }, - { - "epoch": 0.52, - "learning_rate": 9.685331799076208e-06, - "loss": 0.9336, - "step": 2727 - }, - { - "epoch": 0.52, - "learning_rate": 9.6791028644991e-06, - "loss": 0.8304, - "step": 2728 - }, - { - "epoch": 0.53, - "learning_rate": 9.672874054554738e-06, - "loss": 1.0099, - "step": 2729 - }, - { - "epoch": 0.53, - "learning_rate": 9.666645371662324e-06, - "loss": 0.906, - "step": 2730 - }, - { - "epoch": 0.53, - "learning_rate": 9.660416818241007e-06, - "loss": 0.8936, - "step": 2731 - }, - { - "epoch": 0.53, - "learning_rate": 9.654188396709882e-06, - "loss": 0.8925, - "step": 2732 - }, - { - "epoch": 0.53, - "learning_rate": 9.647960109488003e-06, - "loss": 0.9398, - "step": 2733 - }, - { - "epoch": 0.53, - "learning_rate": 9.64173195899436e-06, - "loss": 0.8658, - "step": 2734 - }, - { - "epoch": 0.53, - "learning_rate": 9.635503947647894e-06, - "loss": 0.8659, - "step": 2735 - }, - { - "epoch": 0.53, - "learning_rate": 9.629276077867497e-06, - "loss": 0.8376, - "step": 2736 - }, - { - "epoch": 0.53, - "learning_rate": 9.623048352071998e-06, - "loss": 0.8392, - "step": 2737 - }, - { - "epoch": 0.53, - "learning_rate": 9.616820772680174e-06, - "loss": 0.8864, - "step": 2738 - }, - { - "epoch": 0.53, - "learning_rate": 9.610593342110746e-06, - "loss": 0.8997, - "step": 2739 - }, - { - "epoch": 0.53, - "learning_rate": 9.604366062782381e-06, - "loss": 0.8648, - "step": 2740 - }, - { - "epoch": 0.53, - "learning_rate": 9.598138937113677e-06, - "loss": 0.8886, - "step": 2741 - }, - { - "epoch": 0.53, - "learning_rate": 9.59191196752318e-06, - "loss": 0.7243, - "step": 2742 - }, - { - "epoch": 0.53, - "learning_rate": 9.58568515642937e-06, - "loss": 0.8927, - "step": 2743 - }, - { - "epoch": 0.53, - "learning_rate": 9.579458506250668e-06, - "loss": 0.9678, - "step": 2744 - }, - { - "epoch": 0.53, - "learning_rate": 9.573232019405441e-06, - "loss": 0.8657, - "step": 2745 - }, - { - "epoch": 0.53, - "learning_rate": 9.567005698311982e-06, - "loss": 0.9077, - "step": 2746 - }, - { - "epoch": 0.53, - "learning_rate": 9.560779545388517e-06, - "loss": 0.9189, - "step": 2747 - }, - { - "epoch": 0.53, - "learning_rate": 9.554553563053217e-06, - "loss": 0.8968, - "step": 2748 - }, - { - "epoch": 0.53, - "learning_rate": 9.548327753724181e-06, - "loss": 0.9075, - "step": 2749 - }, - { - "epoch": 0.53, - "learning_rate": 9.542102119819436e-06, - "loss": 0.8177, - "step": 2750 - }, - { - "epoch": 0.53, - "learning_rate": 9.535876663756955e-06, - "loss": 0.9079, - "step": 2751 - }, - { - "epoch": 0.53, - "learning_rate": 9.529651387954628e-06, - "loss": 0.8411, - "step": 2752 - }, - { - "epoch": 0.53, - "learning_rate": 9.523426294830284e-06, - "loss": 0.8543, - "step": 2753 - }, - { - "epoch": 0.53, - "learning_rate": 9.517201386801675e-06, - "loss": 0.9628, - "step": 2754 - }, - { - "epoch": 0.53, - "learning_rate": 9.510976666286484e-06, - "loss": 0.9891, - "step": 2755 - }, - { - "epoch": 0.53, - "learning_rate": 9.504752135702318e-06, - "loss": 0.9222, - "step": 2756 - }, - { - "epoch": 0.53, - "learning_rate": 9.498527797466718e-06, - "loss": 0.8467, - "step": 2757 - }, - { - "epoch": 0.53, - "learning_rate": 9.492303653997146e-06, - "loss": 0.868, - "step": 2758 - }, - { - "epoch": 0.53, - "learning_rate": 9.48607970771098e-06, - "loss": 0.9158, - "step": 2759 - }, - { - "epoch": 0.53, - "learning_rate": 9.479855961025538e-06, - "loss": 0.8577, - "step": 2760 - }, - { - "epoch": 0.53, - "learning_rate": 9.473632416358045e-06, - "loss": 0.7542, - "step": 2761 - }, - { - "epoch": 0.53, - "learning_rate": 9.467409076125653e-06, - "loss": 0.9052, - "step": 2762 - }, - { - "epoch": 0.53, - "learning_rate": 9.461185942745443e-06, - "loss": 0.9538, - "step": 2763 - }, - { - "epoch": 0.53, - "learning_rate": 9.454963018634402e-06, - "loss": 0.7992, - "step": 2764 - }, - { - "epoch": 0.53, - "learning_rate": 9.448740306209447e-06, - "loss": 0.8866, - "step": 2765 - }, - { - "epoch": 0.53, - "learning_rate": 9.442517807887402e-06, - "loss": 0.8274, - "step": 2766 - }, - { - "epoch": 0.53, - "learning_rate": 9.436295526085016e-06, - "loss": 0.9239, - "step": 2767 - }, - { - "epoch": 0.53, - "learning_rate": 9.430073463218952e-06, - "loss": 0.9071, - "step": 2768 - }, - { - "epoch": 0.53, - "learning_rate": 9.423851621705789e-06, - "loss": 0.9016, - "step": 2769 - }, - { - "epoch": 0.53, - "learning_rate": 9.41763000396202e-06, - "loss": 0.8718, - "step": 2770 - }, - { - "epoch": 0.53, - "learning_rate": 9.411408612404043e-06, - "loss": 0.8777, - "step": 2771 - }, - { - "epoch": 0.53, - "learning_rate": 9.40518744944818e-06, - "loss": 0.8686, - "step": 2772 - }, - { - "epoch": 0.53, - "learning_rate": 9.398966517510654e-06, - "loss": 0.9818, - "step": 2773 - }, - { - "epoch": 0.53, - "learning_rate": 9.39274581900761e-06, - "loss": 0.9567, - "step": 2774 - }, - { - "epoch": 0.53, - "learning_rate": 9.386525356355095e-06, - "loss": 0.7123, - "step": 2775 - }, - { - "epoch": 0.53, - "learning_rate": 9.380305131969059e-06, - "loss": 0.8598, - "step": 2776 - }, - { - "epoch": 0.53, - "learning_rate": 9.374085148265372e-06, - "loss": 0.8527, - "step": 2777 - }, - { - "epoch": 0.53, - "learning_rate": 9.3678654076598e-06, - "loss": 0.8844, - "step": 2778 - }, - { - "epoch": 0.53, - "learning_rate": 9.361645912568015e-06, - "loss": 0.9215, - "step": 2779 - }, - { - "epoch": 0.53, - "learning_rate": 9.355426665405607e-06, - "loss": 0.9418, - "step": 2780 - }, - { - "epoch": 0.54, - "learning_rate": 9.349207668588053e-06, - "loss": 0.8861, - "step": 2781 - }, - { - "epoch": 0.54, - "learning_rate": 9.342988924530742e-06, - "loss": 0.9162, - "step": 2782 - }, - { - "epoch": 0.54, - "learning_rate": 9.336770435648963e-06, - "loss": 0.8395, - "step": 2783 - }, - { - "epoch": 0.54, - "learning_rate": 9.330552204357904e-06, - "loss": 0.9056, - "step": 2784 - }, - { - "epoch": 0.54, - "learning_rate": 9.32433423307265e-06, - "loss": 0.8869, - "step": 2785 - }, - { - "epoch": 0.54, - "learning_rate": 9.318116524208198e-06, - "loss": 0.8357, - "step": 2786 - }, - { - "epoch": 0.54, - "learning_rate": 9.311899080179433e-06, - "loss": 0.8458, - "step": 2787 - }, - { - "epoch": 0.54, - "learning_rate": 9.305681903401133e-06, - "loss": 0.924, - "step": 2788 - }, - { - "epoch": 0.54, - "learning_rate": 9.299464996287984e-06, - "loss": 0.9205, - "step": 2789 - }, - { - "epoch": 0.54, - "learning_rate": 9.293248361254557e-06, - "loss": 0.917, - "step": 2790 - }, - { - "epoch": 0.54, - "learning_rate": 9.287032000715318e-06, - "loss": 0.8749, - "step": 2791 - }, - { - "epoch": 0.54, - "learning_rate": 9.28081591708464e-06, - "loss": 0.8772, - "step": 2792 - }, - { - "epoch": 0.54, - "learning_rate": 9.27460011277677e-06, - "loss": 0.8814, - "step": 2793 - }, - { - "epoch": 0.54, - "learning_rate": 9.268384590205858e-06, - "loss": 0.8791, - "step": 2794 - }, - { - "epoch": 0.54, - "learning_rate": 9.262169351785944e-06, - "loss": 0.8979, - "step": 2795 - }, - { - "epoch": 0.54, - "learning_rate": 9.255954399930948e-06, - "loss": 0.8984, - "step": 2796 - }, - { - "epoch": 0.54, - "learning_rate": 9.249739737054686e-06, - "loss": 0.9047, - "step": 2797 - }, - { - "epoch": 0.54, - "learning_rate": 9.24352536557087e-06, - "loss": 0.8663, - "step": 2798 - }, - { - "epoch": 0.54, - "learning_rate": 9.237311287893086e-06, - "loss": 0.8627, - "step": 2799 - }, - { - "epoch": 0.54, - "learning_rate": 9.231097506434808e-06, - "loss": 0.8207, - "step": 2800 - }, - { - "epoch": 0.54, - "learning_rate": 9.224884023609398e-06, - "loss": 0.914, - "step": 2801 - }, - { - "epoch": 0.54, - "learning_rate": 9.218670841830098e-06, - "loss": 0.8906, - "step": 2802 - }, - { - "epoch": 0.54, - "learning_rate": 9.212457963510045e-06, - "loss": 0.8757, - "step": 2803 - }, - { - "epoch": 0.54, - "learning_rate": 9.206245391062243e-06, - "loss": 0.9124, - "step": 2804 - }, - { - "epoch": 0.54, - "learning_rate": 9.200033126899585e-06, - "loss": 0.902, - "step": 2805 - }, - { - "epoch": 0.54, - "learning_rate": 9.193821173434843e-06, - "loss": 0.7157, - "step": 2806 - }, - { - "epoch": 0.54, - "learning_rate": 9.187609533080668e-06, - "loss": 0.7833, - "step": 2807 - }, - { - "epoch": 0.54, - "learning_rate": 9.181398208249583e-06, - "loss": 0.8718, - "step": 2808 - }, - { - "epoch": 0.54, - "learning_rate": 9.175187201354005e-06, - "loss": 0.9114, - "step": 2809 - }, - { - "epoch": 0.54, - "learning_rate": 9.168976514806216e-06, - "loss": 0.928, - "step": 2810 - }, - { - "epoch": 0.54, - "learning_rate": 9.162766151018372e-06, - "loss": 0.8623, - "step": 2811 - }, - { - "epoch": 0.54, - "learning_rate": 9.156556112402508e-06, - "loss": 0.8601, - "step": 2812 - }, - { - "epoch": 0.54, - "learning_rate": 9.150346401370528e-06, - "loss": 0.8166, - "step": 2813 - }, - { - "epoch": 0.54, - "learning_rate": 9.144137020334214e-06, - "loss": 0.8855, - "step": 2814 - }, - { - "epoch": 0.54, - "learning_rate": 9.137927971705223e-06, - "loss": 0.9371, - "step": 2815 - }, - { - "epoch": 0.54, - "learning_rate": 9.131719257895074e-06, - "loss": 0.9003, - "step": 2816 - }, - { - "epoch": 0.54, - "learning_rate": 9.125510881315159e-06, - "loss": 0.8882, - "step": 2817 - }, - { - "epoch": 0.54, - "learning_rate": 9.119302844376741e-06, - "loss": 0.8732, - "step": 2818 - }, - { - "epoch": 0.54, - "learning_rate": 9.113095149490951e-06, - "loss": 0.8494, - "step": 2819 - }, - { - "epoch": 0.54, - "learning_rate": 9.106887799068782e-06, - "loss": 0.9318, - "step": 2820 - }, - { - "epoch": 0.54, - "learning_rate": 9.100680795521104e-06, - "loss": 0.8829, - "step": 2821 - }, - { - "epoch": 0.54, - "learning_rate": 9.09447414125864e-06, - "loss": 1.016, - "step": 2822 - }, - { - "epoch": 0.54, - "learning_rate": 9.088267838691987e-06, - "loss": 0.9303, - "step": 2823 - }, - { - "epoch": 0.54, - "learning_rate": 9.0820618902316e-06, - "loss": 0.9525, - "step": 2824 - }, - { - "epoch": 0.54, - "learning_rate": 9.075856298287796e-06, - "loss": 0.6934, - "step": 2825 - }, - { - "epoch": 0.54, - "learning_rate": 9.069651065270753e-06, - "loss": 0.8335, - "step": 2826 - }, - { - "epoch": 0.54, - "learning_rate": 9.06344619359052e-06, - "loss": 0.8381, - "step": 2827 - }, - { - "epoch": 0.54, - "learning_rate": 9.057241685656995e-06, - "loss": 0.8901, - "step": 2828 - }, - { - "epoch": 0.54, - "learning_rate": 9.051037543879933e-06, - "loss": 0.8342, - "step": 2829 - }, - { - "epoch": 0.54, - "learning_rate": 9.044833770668957e-06, - "loss": 0.8338, - "step": 2830 - }, - { - "epoch": 0.54, - "learning_rate": 9.038630368433537e-06, - "loss": 0.8681, - "step": 2831 - }, - { - "epoch": 0.54, - "learning_rate": 9.032427339583e-06, - "loss": 0.8704, - "step": 2832 - }, - { - "epoch": 0.55, - "learning_rate": 9.026224686526539e-06, - "loss": 0.9693, - "step": 2833 - }, - { - "epoch": 0.55, - "learning_rate": 9.020022411673186e-06, - "loss": 0.8965, - "step": 2834 - }, - { - "epoch": 0.55, - "learning_rate": 9.013820517431841e-06, - "loss": 0.8865, - "step": 2835 - }, - { - "epoch": 0.55, - "learning_rate": 9.00761900621124e-06, - "loss": 0.926, - "step": 2836 - }, - { - "epoch": 0.55, - "learning_rate": 9.00141788041998e-06, - "loss": 0.9149, - "step": 2837 - }, - { - "epoch": 0.55, - "learning_rate": 8.99521714246651e-06, - "loss": 0.867, - "step": 2838 - }, - { - "epoch": 0.55, - "learning_rate": 8.989016794759127e-06, - "loss": 0.8655, - "step": 2839 - }, - { - "epoch": 0.55, - "learning_rate": 8.98281683970597e-06, - "loss": 0.9834, - "step": 2840 - }, - { - "epoch": 0.55, - "learning_rate": 8.97661727971503e-06, - "loss": 0.9519, - "step": 2841 - }, - { - "epoch": 0.55, - "learning_rate": 8.970418117194146e-06, - "loss": 0.8617, - "step": 2842 - }, - { - "epoch": 0.55, - "learning_rate": 8.964219354550999e-06, - "loss": 0.8837, - "step": 2843 - }, - { - "epoch": 0.55, - "learning_rate": 8.958020994193124e-06, - "loss": 0.8956, - "step": 2844 - }, - { - "epoch": 0.55, - "learning_rate": 8.951823038527887e-06, - "loss": 0.8535, - "step": 2845 - }, - { - "epoch": 0.55, - "learning_rate": 8.945625489962503e-06, - "loss": 0.8964, - "step": 2846 - }, - { - "epoch": 0.55, - "learning_rate": 8.93942835090403e-06, - "loss": 0.8995, - "step": 2847 - }, - { - "epoch": 0.55, - "learning_rate": 8.933231623759365e-06, - "loss": 0.8665, - "step": 2848 - }, - { - "epoch": 0.55, - "learning_rate": 8.927035310935241e-06, - "loss": 0.9004, - "step": 2849 - }, - { - "epoch": 0.55, - "learning_rate": 8.920839414838243e-06, - "loss": 0.8689, - "step": 2850 - }, - { - "epoch": 0.55, - "learning_rate": 8.914643937874778e-06, - "loss": 0.8742, - "step": 2851 - }, - { - "epoch": 0.55, - "learning_rate": 8.908448882451104e-06, - "loss": 0.8873, - "step": 2852 - }, - { - "epoch": 0.55, - "learning_rate": 8.902254250973306e-06, - "loss": 0.8764, - "step": 2853 - }, - { - "epoch": 0.55, - "learning_rate": 8.896060045847305e-06, - "loss": 0.8288, - "step": 2854 - }, - { - "epoch": 0.55, - "learning_rate": 8.88986626947886e-06, - "loss": 0.8829, - "step": 2855 - }, - { - "epoch": 0.55, - "learning_rate": 8.883672924273566e-06, - "loss": 0.8643, - "step": 2856 - }, - { - "epoch": 0.55, - "learning_rate": 8.877480012636847e-06, - "loss": 0.9121, - "step": 2857 - }, - { - "epoch": 0.55, - "learning_rate": 8.871287536973954e-06, - "loss": 0.8241, - "step": 2858 - }, - { - "epoch": 0.55, - "learning_rate": 8.865095499689978e-06, - "loss": 0.709, - "step": 2859 - }, - { - "epoch": 0.55, - "learning_rate": 8.85890390318983e-06, - "loss": 1.0138, - "step": 2860 - }, - { - "epoch": 0.55, - "learning_rate": 8.852712749878255e-06, - "loss": 0.8159, - "step": 2861 - }, - { - "epoch": 0.55, - "learning_rate": 8.846522042159833e-06, - "loss": 0.9126, - "step": 2862 - }, - { - "epoch": 0.55, - "learning_rate": 8.840331782438954e-06, - "loss": 0.8445, - "step": 2863 - }, - { - "epoch": 0.55, - "learning_rate": 8.83414197311985e-06, - "loss": 0.9031, - "step": 2864 - }, - { - "epoch": 0.55, - "learning_rate": 8.82795261660657e-06, - "loss": 0.823, - "step": 2865 - }, - { - "epoch": 0.55, - "learning_rate": 8.821763715302986e-06, - "loss": 0.9386, - "step": 2866 - }, - { - "epoch": 0.55, - "learning_rate": 8.815575271612798e-06, - "loss": 0.9207, - "step": 2867 - }, - { - "epoch": 0.55, - "learning_rate": 8.809387287939528e-06, - "loss": 0.9357, - "step": 2868 - }, - { - "epoch": 0.55, - "learning_rate": 8.803199766686517e-06, - "loss": 0.8521, - "step": 2869 - }, - { - "epoch": 0.55, - "learning_rate": 8.797012710256923e-06, - "loss": 0.8663, - "step": 2870 - }, - { - "epoch": 0.55, - "learning_rate": 8.790826121053732e-06, - "loss": 0.8356, - "step": 2871 - }, - { - "epoch": 0.55, - "learning_rate": 8.784640001479741e-06, - "loss": 0.9574, - "step": 2872 - }, - { - "epoch": 0.55, - "learning_rate": 8.77845435393757e-06, - "loss": 0.8245, - "step": 2873 - }, - { - "epoch": 0.55, - "learning_rate": 8.772269180829653e-06, - "loss": 0.9071, - "step": 2874 - }, - { - "epoch": 0.55, - "learning_rate": 8.766084484558237e-06, - "loss": 0.8945, - "step": 2875 - }, - { - "epoch": 0.55, - "learning_rate": 8.759900267525393e-06, - "loss": 0.858, - "step": 2876 - }, - { - "epoch": 0.55, - "learning_rate": 8.753716532132992e-06, - "loss": 0.8684, - "step": 2877 - }, - { - "epoch": 0.55, - "learning_rate": 8.747533280782725e-06, - "loss": 0.8725, - "step": 2878 - }, - { - "epoch": 0.55, - "learning_rate": 8.741350515876103e-06, - "loss": 0.8347, - "step": 2879 - }, - { - "epoch": 0.55, - "learning_rate": 8.73516823981444e-06, - "loss": 0.904, - "step": 2880 - }, - { - "epoch": 0.55, - "learning_rate": 8.728986454998858e-06, - "loss": 0.9158, - "step": 2881 - }, - { - "epoch": 0.55, - "learning_rate": 8.72280516383029e-06, - "loss": 0.8358, - "step": 2882 - }, - { - "epoch": 0.55, - "learning_rate": 8.716624368709477e-06, - "loss": 0.8845, - "step": 2883 - }, - { - "epoch": 0.55, - "learning_rate": 8.71044407203697e-06, - "loss": 0.9431, - "step": 2884 - }, - { - "epoch": 0.56, - "learning_rate": 8.70426427621313e-06, - "loss": 0.8259, - "step": 2885 - }, - { - "epoch": 0.56, - "learning_rate": 8.698084983638111e-06, - "loss": 0.8902, - "step": 2886 - }, - { - "epoch": 0.56, - "learning_rate": 8.691906196711884e-06, - "loss": 0.9371, - "step": 2887 - }, - { - "epoch": 0.56, - "learning_rate": 8.685727917834218e-06, - "loss": 0.9184, - "step": 2888 - }, - { - "epoch": 0.56, - "learning_rate": 8.679550149404685e-06, - "loss": 0.8774, - "step": 2889 - }, - { - "epoch": 0.56, - "learning_rate": 8.673372893822653e-06, - "loss": 0.8534, - "step": 2890 - }, - { - "epoch": 0.56, - "learning_rate": 8.667196153487308e-06, - "loss": 0.8734, - "step": 2891 - }, - { - "epoch": 0.56, - "learning_rate": 8.661019930797615e-06, - "loss": 0.8866, - "step": 2892 - }, - { - "epoch": 0.56, - "learning_rate": 8.654844228152355e-06, - "loss": 0.9869, - "step": 2893 - }, - { - "epoch": 0.56, - "learning_rate": 8.648669047950097e-06, - "loss": 0.7781, - "step": 2894 - }, - { - "epoch": 0.56, - "learning_rate": 8.642494392589206e-06, - "loss": 0.9205, - "step": 2895 - }, - { - "epoch": 0.56, - "learning_rate": 8.63632026446785e-06, - "loss": 0.8542, - "step": 2896 - }, - { - "epoch": 0.56, - "learning_rate": 8.630146665983993e-06, - "loss": 0.8507, - "step": 2897 - }, - { - "epoch": 0.56, - "learning_rate": 8.623973599535385e-06, - "loss": 0.8464, - "step": 2898 - }, - { - "epoch": 0.56, - "learning_rate": 8.617801067519575e-06, - "loss": 0.8702, - "step": 2899 - }, - { - "epoch": 0.56, - "learning_rate": 8.611629072333905e-06, - "loss": 0.8997, - "step": 2900 - }, - { - "epoch": 0.56, - "learning_rate": 8.605457616375503e-06, - "loss": 0.9129, - "step": 2901 - }, - { - "epoch": 0.56, - "learning_rate": 8.599286702041292e-06, - "loss": 0.8584, - "step": 2902 - }, - { - "epoch": 0.56, - "learning_rate": 8.593116331727987e-06, - "loss": 0.9126, - "step": 2903 - }, - { - "epoch": 0.56, - "learning_rate": 8.586946507832088e-06, - "loss": 0.7176, - "step": 2904 - }, - { - "epoch": 0.56, - "learning_rate": 8.580777232749883e-06, - "loss": 0.8948, - "step": 2905 - }, - { - "epoch": 0.56, - "learning_rate": 8.574608508877448e-06, - "loss": 0.8578, - "step": 2906 - }, - { - "epoch": 0.56, - "learning_rate": 8.568440338610638e-06, - "loss": 0.914, - "step": 2907 - }, - { - "epoch": 0.56, - "learning_rate": 8.562272724345108e-06, - "loss": 0.894, - "step": 2908 - }, - { - "epoch": 0.56, - "learning_rate": 8.556105668476287e-06, - "loss": 0.9597, - "step": 2909 - }, - { - "epoch": 0.56, - "learning_rate": 8.549939173399385e-06, - "loss": 0.9142, - "step": 2910 - }, - { - "epoch": 0.56, - "learning_rate": 8.5437732415094e-06, - "loss": 0.9667, - "step": 2911 - }, - { - "epoch": 0.56, - "learning_rate": 8.537607875201106e-06, - "loss": 0.9382, - "step": 2912 - }, - { - "epoch": 0.56, - "learning_rate": 8.531443076869058e-06, - "loss": 0.8632, - "step": 2913 - }, - { - "epoch": 0.56, - "learning_rate": 8.525278848907603e-06, - "loss": 0.864, - "step": 2914 - }, - { - "epoch": 0.56, - "learning_rate": 8.51911519371085e-06, - "loss": 0.8847, - "step": 2915 - }, - { - "epoch": 0.56, - "learning_rate": 8.512952113672689e-06, - "loss": 0.8774, - "step": 2916 - }, - { - "epoch": 0.56, - "learning_rate": 8.506789611186794e-06, - "loss": 0.8424, - "step": 2917 - }, - { - "epoch": 0.56, - "learning_rate": 8.500627688646607e-06, - "loss": 0.8572, - "step": 2918 - }, - { - "epoch": 0.56, - "learning_rate": 8.494466348445345e-06, - "loss": 0.9032, - "step": 2919 - }, - { - "epoch": 0.56, - "learning_rate": 8.48830559297601e-06, - "loss": 0.951, - "step": 2920 - }, - { - "epoch": 0.56, - "learning_rate": 8.48214542463136e-06, - "loss": 0.8567, - "step": 2921 - }, - { - "epoch": 0.56, - "learning_rate": 8.475985845803938e-06, - "loss": 0.8834, - "step": 2922 - }, - { - "epoch": 0.56, - "learning_rate": 8.469826858886054e-06, - "loss": 0.8529, - "step": 2923 - }, - { - "epoch": 0.56, - "learning_rate": 8.463668466269785e-06, - "loss": 0.9363, - "step": 2924 - }, - { - "epoch": 0.56, - "learning_rate": 8.457510670346976e-06, - "loss": 0.937, - "step": 2925 - }, - { - "epoch": 0.56, - "learning_rate": 8.451353473509254e-06, - "loss": 0.899, - "step": 2926 - }, - { - "epoch": 0.56, - "learning_rate": 8.445196878147997e-06, - "loss": 0.9155, - "step": 2927 - }, - { - "epoch": 0.56, - "learning_rate": 8.439040886654354e-06, - "loss": 0.8261, - "step": 2928 - }, - { - "epoch": 0.56, - "learning_rate": 8.432885501419248e-06, - "loss": 0.9042, - "step": 2929 - }, - { - "epoch": 0.56, - "learning_rate": 8.426730724833354e-06, - "loss": 0.8686, - "step": 2930 - }, - { - "epoch": 0.56, - "learning_rate": 8.420576559287112e-06, - "loss": 0.9237, - "step": 2931 - }, - { - "epoch": 0.56, - "learning_rate": 8.414423007170742e-06, - "loss": 0.8999, - "step": 2932 - }, - { - "epoch": 0.56, - "learning_rate": 8.408270070874201e-06, - "loss": 0.9149, - "step": 2933 - }, - { - "epoch": 0.56, - "learning_rate": 8.402117752787225e-06, - "loss": 0.8446, - "step": 2934 - }, - { - "epoch": 0.56, - "learning_rate": 8.395966055299302e-06, - "loss": 0.9141, - "step": 2935 - }, - { - "epoch": 0.56, - "learning_rate": 8.389814980799679e-06, - "loss": 0.9372, - "step": 2936 - }, - { - "epoch": 0.57, - "learning_rate": 8.38366453167736e-06, - "loss": 0.8827, - "step": 2937 - }, - { - "epoch": 0.57, - "learning_rate": 8.377514710321117e-06, - "loss": 0.9194, - "step": 2938 - }, - { - "epoch": 0.57, - "learning_rate": 8.371365519119463e-06, - "loss": 0.9693, - "step": 2939 - }, - { - "epoch": 0.57, - "learning_rate": 8.365216960460675e-06, - "loss": 0.9076, - "step": 2940 - }, - { - "epoch": 0.57, - "learning_rate": 8.359069036732781e-06, - "loss": 0.8496, - "step": 2941 - }, - { - "epoch": 0.57, - "learning_rate": 8.352921750323562e-06, - "loss": 1.0168, - "step": 2942 - }, - { - "epoch": 0.57, - "learning_rate": 8.346775103620559e-06, - "loss": 0.8361, - "step": 2943 - }, - { - "epoch": 0.57, - "learning_rate": 8.340629099011057e-06, - "loss": 0.9277, - "step": 2944 - }, - { - "epoch": 0.57, - "learning_rate": 8.33448373888209e-06, - "loss": 0.7121, - "step": 2945 - }, - { - "epoch": 0.57, - "learning_rate": 8.328339025620449e-06, - "loss": 0.931, - "step": 2946 - }, - { - "epoch": 0.57, - "learning_rate": 8.322194961612668e-06, - "loss": 0.8861, - "step": 2947 - }, - { - "epoch": 0.57, - "learning_rate": 8.316051549245026e-06, - "loss": 0.9486, - "step": 2948 - }, - { - "epoch": 0.57, - "learning_rate": 8.309908790903562e-06, - "loss": 0.8892, - "step": 2949 - }, - { - "epoch": 0.57, - "learning_rate": 8.303766688974047e-06, - "loss": 0.9362, - "step": 2950 - }, - { - "epoch": 0.57, - "learning_rate": 8.297625245842006e-06, - "loss": 0.7058, - "step": 2951 - }, - { - "epoch": 0.57, - "learning_rate": 8.291484463892703e-06, - "loss": 0.8541, - "step": 2952 - }, - { - "epoch": 0.57, - "learning_rate": 8.285344345511147e-06, - "loss": 0.8668, - "step": 2953 - }, - { - "epoch": 0.57, - "learning_rate": 8.279204893082083e-06, - "loss": 0.912, - "step": 2954 - }, - { - "epoch": 0.57, - "learning_rate": 8.273066108990017e-06, - "loss": 0.8294, - "step": 2955 - }, - { - "epoch": 0.57, - "learning_rate": 8.266927995619175e-06, - "loss": 0.9406, - "step": 2956 - }, - { - "epoch": 0.57, - "learning_rate": 8.260790555353526e-06, - "loss": 0.929, - "step": 2957 - }, - { - "epoch": 0.57, - "learning_rate": 8.254653790576787e-06, - "loss": 0.8755, - "step": 2958 - }, - { - "epoch": 0.57, - "learning_rate": 8.248517703672405e-06, - "loss": 0.8798, - "step": 2959 - }, - { - "epoch": 0.57, - "learning_rate": 8.242382297023558e-06, - "loss": 0.9165, - "step": 2960 - }, - { - "epoch": 0.57, - "learning_rate": 8.23624757301318e-06, - "loss": 0.8755, - "step": 2961 - }, - { - "epoch": 0.57, - "learning_rate": 8.230113534023917e-06, - "loss": 0.9117, - "step": 2962 - }, - { - "epoch": 0.57, - "learning_rate": 8.223980182438167e-06, - "loss": 0.8649, - "step": 2963 - }, - { - "epoch": 0.57, - "learning_rate": 8.217847520638049e-06, - "loss": 0.8574, - "step": 2964 - }, - { - "epoch": 0.57, - "learning_rate": 8.211715551005414e-06, - "loss": 0.9064, - "step": 2965 - }, - { - "epoch": 0.57, - "learning_rate": 8.205584275921854e-06, - "loss": 0.7794, - "step": 2966 - }, - { - "epoch": 0.57, - "learning_rate": 8.199453697768686e-06, - "loss": 0.8339, - "step": 2967 - }, - { - "epoch": 0.57, - "learning_rate": 8.193323818926955e-06, - "loss": 0.8459, - "step": 2968 - }, - { - "epoch": 0.57, - "learning_rate": 8.187194641777432e-06, - "loss": 0.832, - "step": 2969 - }, - { - "epoch": 0.57, - "learning_rate": 8.181066168700622e-06, - "loss": 0.9032, - "step": 2970 - }, - { - "epoch": 0.57, - "learning_rate": 8.174938402076754e-06, - "loss": 0.9012, - "step": 2971 - }, - { - "epoch": 0.57, - "learning_rate": 8.168811344285776e-06, - "loss": 0.8203, - "step": 2972 - }, - { - "epoch": 0.57, - "learning_rate": 8.162684997707374e-06, - "loss": 0.7979, - "step": 2973 - }, - { - "epoch": 0.57, - "learning_rate": 8.156559364720947e-06, - "loss": 0.8686, - "step": 2974 - }, - { - "epoch": 0.57, - "learning_rate": 8.150434447705623e-06, - "loss": 0.8953, - "step": 2975 - }, - { - "epoch": 0.57, - "learning_rate": 8.144310249040246e-06, - "loss": 0.9312, - "step": 2976 - }, - { - "epoch": 0.57, - "learning_rate": 8.138186771103382e-06, - "loss": 0.839, - "step": 2977 - }, - { - "epoch": 0.57, - "learning_rate": 8.132064016273325e-06, - "loss": 0.8369, - "step": 2978 - }, - { - "epoch": 0.57, - "learning_rate": 8.12594198692808e-06, - "loss": 0.8402, - "step": 2979 - }, - { - "epoch": 0.57, - "learning_rate": 8.119820685445373e-06, - "loss": 0.8613, - "step": 2980 - }, - { - "epoch": 0.57, - "learning_rate": 8.113700114202647e-06, - "loss": 0.7888, - "step": 2981 - }, - { - "epoch": 0.57, - "learning_rate": 8.107580275577059e-06, - "loss": 0.8354, - "step": 2982 - }, - { - "epoch": 0.57, - "learning_rate": 8.101461171945483e-06, - "loss": 0.8517, - "step": 2983 - }, - { - "epoch": 0.57, - "learning_rate": 8.095342805684516e-06, - "loss": 0.866, - "step": 2984 - }, - { - "epoch": 0.57, - "learning_rate": 8.089225179170454e-06, - "loss": 0.8563, - "step": 2985 - }, - { - "epoch": 0.57, - "learning_rate": 8.083108294779313e-06, - "loss": 0.8944, - "step": 2986 - }, - { - "epoch": 0.57, - "learning_rate": 8.076992154886826e-06, - "loss": 0.8197, - "step": 2987 - }, - { - "epoch": 0.57, - "learning_rate": 8.070876761868426e-06, - "loss": 0.7401, - "step": 2988 - }, - { - "epoch": 0.58, - "learning_rate": 8.064762118099258e-06, - "loss": 0.8735, - "step": 2989 - }, - { - "epoch": 0.58, - "learning_rate": 8.058648225954188e-06, - "loss": 0.8916, - "step": 2990 - }, - { - "epoch": 0.58, - "learning_rate": 8.052535087807774e-06, - "loss": 0.847, - "step": 2991 - }, - { - "epoch": 0.58, - "learning_rate": 8.046422706034294e-06, - "loss": 0.8532, - "step": 2992 - }, - { - "epoch": 0.58, - "learning_rate": 8.040311083007725e-06, - "loss": 0.9462, - "step": 2993 - }, - { - "epoch": 0.58, - "learning_rate": 8.034200221101746e-06, - "loss": 0.8414, - "step": 2994 - }, - { - "epoch": 0.58, - "learning_rate": 8.028090122689747e-06, - "loss": 0.8043, - "step": 2995 - }, - { - "epoch": 0.58, - "learning_rate": 8.021980790144828e-06, - "loss": 0.9639, - "step": 2996 - }, - { - "epoch": 0.58, - "learning_rate": 8.015872225839776e-06, - "loss": 0.83, - "step": 2997 - }, - { - "epoch": 0.58, - "learning_rate": 8.009764432147086e-06, - "loss": 0.8983, - "step": 2998 - }, - { - "epoch": 0.58, - "learning_rate": 8.003657411438961e-06, - "loss": 0.9214, - "step": 2999 - }, - { - "epoch": 0.58, - "learning_rate": 7.997551166087293e-06, - "loss": 0.8816, - "step": 3000 - }, - { - "epoch": 0.58, - "learning_rate": 7.991445698463672e-06, - "loss": 0.8881, - "step": 3001 - }, - { - "epoch": 0.58, - "learning_rate": 7.985341010939402e-06, - "loss": 0.8669, - "step": 3002 - }, - { - "epoch": 0.58, - "learning_rate": 7.979237105885467e-06, - "loss": 0.8966, - "step": 3003 - }, - { - "epoch": 0.58, - "learning_rate": 7.973133985672558e-06, - "loss": 0.8769, - "step": 3004 - }, - { - "epoch": 0.58, - "learning_rate": 7.967031652671051e-06, - "loss": 0.8906, - "step": 3005 - }, - { - "epoch": 0.58, - "learning_rate": 7.960930109251023e-06, - "loss": 0.8569, - "step": 3006 - }, - { - "epoch": 0.58, - "learning_rate": 7.954829357782243e-06, - "loss": 0.6983, - "step": 3007 - }, - { - "epoch": 0.58, - "learning_rate": 7.948729400634178e-06, - "loss": 0.8071, - "step": 3008 - }, - { - "epoch": 0.58, - "learning_rate": 7.942630240175977e-06, - "loss": 0.9143, - "step": 3009 - }, - { - "epoch": 0.58, - "learning_rate": 7.936531878776484e-06, - "loss": 0.8619, - "step": 3010 - }, - { - "epoch": 0.58, - "learning_rate": 7.930434318804229e-06, - "loss": 0.8008, - "step": 3011 - }, - { - "epoch": 0.58, - "learning_rate": 7.924337562627435e-06, - "loss": 0.82, - "step": 3012 - }, - { - "epoch": 0.58, - "learning_rate": 7.918241612614016e-06, - "loss": 0.9441, - "step": 3013 - }, - { - "epoch": 0.58, - "learning_rate": 7.91214647113157e-06, - "loss": 0.8592, - "step": 3014 - }, - { - "epoch": 0.58, - "learning_rate": 7.906052140547373e-06, - "loss": 0.8935, - "step": 3015 - }, - { - "epoch": 0.58, - "learning_rate": 7.899958623228398e-06, - "loss": 0.8218, - "step": 3016 - }, - { - "epoch": 0.58, - "learning_rate": 7.893865921541294e-06, - "loss": 0.8463, - "step": 3017 - }, - { - "epoch": 0.58, - "learning_rate": 7.887774037852395e-06, - "loss": 0.8596, - "step": 3018 - }, - { - "epoch": 0.58, - "learning_rate": 7.881682974527723e-06, - "loss": 0.8395, - "step": 3019 - }, - { - "epoch": 0.58, - "learning_rate": 7.875592733932972e-06, - "loss": 0.8297, - "step": 3020 - }, - { - "epoch": 0.58, - "learning_rate": 7.869503318433529e-06, - "loss": 0.8745, - "step": 3021 - }, - { - "epoch": 0.58, - "learning_rate": 7.863414730394444e-06, - "loss": 0.8827, - "step": 3022 - }, - { - "epoch": 0.58, - "learning_rate": 7.857326972180455e-06, - "loss": 0.8714, - "step": 3023 - }, - { - "epoch": 0.58, - "learning_rate": 7.85124004615598e-06, - "loss": 0.9092, - "step": 3024 - }, - { - "epoch": 0.58, - "learning_rate": 7.845153954685114e-06, - "loss": 0.9399, - "step": 3025 - }, - { - "epoch": 0.58, - "learning_rate": 7.839068700131623e-06, - "loss": 0.8427, - "step": 3026 - }, - { - "epoch": 0.58, - "learning_rate": 7.832984284858946e-06, - "loss": 0.8649, - "step": 3027 - }, - { - "epoch": 0.58, - "learning_rate": 7.826900711230204e-06, - "loss": 0.928, - "step": 3028 - }, - { - "epoch": 0.58, - "learning_rate": 7.820817981608185e-06, - "loss": 0.8588, - "step": 3029 - }, - { - "epoch": 0.58, - "learning_rate": 7.814736098355348e-06, - "loss": 0.8832, - "step": 3030 - }, - { - "epoch": 0.58, - "learning_rate": 7.808655063833832e-06, - "loss": 0.8898, - "step": 3031 - }, - { - "epoch": 0.58, - "learning_rate": 7.802574880405438e-06, - "loss": 0.8777, - "step": 3032 - }, - { - "epoch": 0.58, - "learning_rate": 7.79649555043164e-06, - "loss": 0.8559, - "step": 3033 - }, - { - "epoch": 0.58, - "learning_rate": 7.790417076273581e-06, - "loss": 0.8749, - "step": 3034 - }, - { - "epoch": 0.58, - "learning_rate": 7.784339460292065e-06, - "loss": 0.6824, - "step": 3035 - }, - { - "epoch": 0.58, - "learning_rate": 7.77826270484757e-06, - "loss": 0.9686, - "step": 3036 - }, - { - "epoch": 0.58, - "learning_rate": 7.772186812300244e-06, - "loss": 0.8637, - "step": 3037 - }, - { - "epoch": 0.58, - "learning_rate": 7.766111785009888e-06, - "loss": 0.9239, - "step": 3038 - }, - { - "epoch": 0.58, - "learning_rate": 7.760037625335973e-06, - "loss": 0.8414, - "step": 3039 - }, - { - "epoch": 0.58, - "learning_rate": 7.753964335637634e-06, - "loss": 0.9097, - "step": 3040 - }, - { - "epoch": 0.59, - "learning_rate": 7.747891918273668e-06, - "loss": 0.813, - "step": 3041 - }, - { - "epoch": 0.59, - "learning_rate": 7.741820375602524e-06, - "loss": 0.9231, - "step": 3042 - }, - { - "epoch": 0.59, - "learning_rate": 7.735749709982329e-06, - "loss": 0.8814, - "step": 3043 - }, - { - "epoch": 0.59, - "learning_rate": 7.729679923770855e-06, - "loss": 0.872, - "step": 3044 - }, - { - "epoch": 0.59, - "learning_rate": 7.723611019325538e-06, - "loss": 0.8439, - "step": 3045 - }, - { - "epoch": 0.59, - "learning_rate": 7.71754299900347e-06, - "loss": 0.8698, - "step": 3046 - }, - { - "epoch": 0.59, - "learning_rate": 7.7114758651614e-06, - "loss": 0.8714, - "step": 3047 - }, - { - "epoch": 0.59, - "learning_rate": 7.705409620155733e-06, - "loss": 0.8779, - "step": 3048 - }, - { - "epoch": 0.59, - "learning_rate": 7.699344266342529e-06, - "loss": 0.6782, - "step": 3049 - }, - { - "epoch": 0.59, - "learning_rate": 7.693279806077504e-06, - "loss": 0.8834, - "step": 3050 - }, - { - "epoch": 0.59, - "learning_rate": 7.68721624171602e-06, - "loss": 0.8713, - "step": 3051 - }, - { - "epoch": 0.59, - "learning_rate": 7.681153575613098e-06, - "loss": 0.8113, - "step": 3052 - }, - { - "epoch": 0.59, - "learning_rate": 7.675091810123404e-06, - "loss": 0.8417, - "step": 3053 - }, - { - "epoch": 0.59, - "learning_rate": 7.669030947601265e-06, - "loss": 0.9036, - "step": 3054 - }, - { - "epoch": 0.59, - "learning_rate": 7.662970990400647e-06, - "loss": 0.8869, - "step": 3055 - }, - { - "epoch": 0.59, - "learning_rate": 7.656911940875163e-06, - "loss": 0.8453, - "step": 3056 - }, - { - "epoch": 0.59, - "learning_rate": 7.650853801378084e-06, - "loss": 0.9213, - "step": 3057 - }, - { - "epoch": 0.59, - "learning_rate": 7.644796574262322e-06, - "loss": 0.8061, - "step": 3058 - }, - { - "epoch": 0.59, - "learning_rate": 7.638740261880423e-06, - "loss": 0.9001, - "step": 3059 - }, - { - "epoch": 0.59, - "learning_rate": 7.632684866584606e-06, - "loss": 0.985, - "step": 3060 - }, - { - "epoch": 0.59, - "learning_rate": 7.626630390726704e-06, - "loss": 0.7365, - "step": 3061 - }, - { - "epoch": 0.59, - "learning_rate": 7.620576836658212e-06, - "loss": 0.7185, - "step": 3062 - }, - { - "epoch": 0.59, - "learning_rate": 7.61452420673026e-06, - "loss": 0.8597, - "step": 3063 - }, - { - "epoch": 0.59, - "learning_rate": 7.608472503293615e-06, - "loss": 0.9113, - "step": 3064 - }, - { - "epoch": 0.59, - "learning_rate": 7.60242172869869e-06, - "loss": 0.8778, - "step": 3065 - }, - { - "epoch": 0.59, - "learning_rate": 7.596371885295542e-06, - "loss": 0.9069, - "step": 3066 - }, - { - "epoch": 0.59, - "learning_rate": 7.590322975433857e-06, - "loss": 0.9077, - "step": 3067 - }, - { - "epoch": 0.59, - "learning_rate": 7.584275001462961e-06, - "loss": 0.8761, - "step": 3068 - }, - { - "epoch": 0.59, - "learning_rate": 7.578227965731819e-06, - "loss": 0.9254, - "step": 3069 - }, - { - "epoch": 0.59, - "learning_rate": 7.572181870589028e-06, - "loss": 0.9228, - "step": 3070 - }, - { - "epoch": 0.59, - "learning_rate": 7.566136718382821e-06, - "loss": 0.899, - "step": 3071 - }, - { - "epoch": 0.59, - "learning_rate": 7.560092511461069e-06, - "loss": 0.9224, - "step": 3072 - }, - { - "epoch": 0.59, - "learning_rate": 7.55404925217127e-06, - "loss": 0.9121, - "step": 3073 - }, - { - "epoch": 0.59, - "learning_rate": 7.548006942860557e-06, - "loss": 0.8154, - "step": 3074 - }, - { - "epoch": 0.59, - "learning_rate": 7.541965585875695e-06, - "loss": 0.9155, - "step": 3075 - }, - { - "epoch": 0.59, - "learning_rate": 7.535925183563073e-06, - "loss": 0.7054, - "step": 3076 - }, - { - "epoch": 0.59, - "learning_rate": 7.529885738268714e-06, - "loss": 0.7812, - "step": 3077 - }, - { - "epoch": 0.59, - "learning_rate": 7.523847252338274e-06, - "loss": 0.8162, - "step": 3078 - }, - { - "epoch": 0.59, - "learning_rate": 7.51780972811703e-06, - "loss": 0.9716, - "step": 3079 - }, - { - "epoch": 0.59, - "learning_rate": 7.511773167949885e-06, - "loss": 0.929, - "step": 3080 - }, - { - "epoch": 0.59, - "learning_rate": 7.5057375741813685e-06, - "loss": 0.7348, - "step": 3081 - }, - { - "epoch": 0.59, - "learning_rate": 7.499702949155634e-06, - "loss": 0.9218, - "step": 3082 - }, - { - "epoch": 0.59, - "learning_rate": 7.493669295216467e-06, - "loss": 0.9199, - "step": 3083 - }, - { - "epoch": 0.59, - "learning_rate": 7.487636614707265e-06, - "loss": 0.8455, - "step": 3084 - }, - { - "epoch": 0.59, - "learning_rate": 7.48160490997105e-06, - "loss": 0.9017, - "step": 3085 - }, - { - "epoch": 0.59, - "learning_rate": 7.475574183350471e-06, - "loss": 0.9005, - "step": 3086 - }, - { - "epoch": 0.59, - "learning_rate": 7.46954443718779e-06, - "loss": 0.8849, - "step": 3087 - }, - { - "epoch": 0.59, - "learning_rate": 7.463515673824888e-06, - "loss": 0.7823, - "step": 3088 - }, - { - "epoch": 0.59, - "learning_rate": 7.457487895603273e-06, - "loss": 0.6884, - "step": 3089 - }, - { - "epoch": 0.59, - "learning_rate": 7.451461104864061e-06, - "loss": 0.9499, - "step": 3090 - }, - { - "epoch": 0.59, - "learning_rate": 7.44543530394799e-06, - "loss": 0.6776, - "step": 3091 - }, - { - "epoch": 0.59, - "learning_rate": 7.439410495195411e-06, - "loss": 0.8443, - "step": 3092 - }, - { - "epoch": 0.6, - "learning_rate": 7.433386680946288e-06, - "loss": 0.8264, - "step": 3093 - }, - { - "epoch": 0.6, - "learning_rate": 7.427363863540202e-06, - "loss": 0.8823, - "step": 3094 - }, - { - "epoch": 0.6, - "learning_rate": 7.421342045316351e-06, - "loss": 0.8808, - "step": 3095 - }, - { - "epoch": 0.6, - "learning_rate": 7.415321228613534e-06, - "loss": 0.8989, - "step": 3096 - }, - { - "epoch": 0.6, - "learning_rate": 7.409301415770168e-06, - "loss": 0.871, - "step": 3097 - }, - { - "epoch": 0.6, - "learning_rate": 7.403282609124281e-06, - "loss": 0.8985, - "step": 3098 - }, - { - "epoch": 0.6, - "learning_rate": 7.397264811013507e-06, - "loss": 0.901, - "step": 3099 - }, - { - "epoch": 0.6, - "learning_rate": 7.391248023775084e-06, - "loss": 0.9518, - "step": 3100 - }, - { - "epoch": 0.6, - "learning_rate": 7.385232249745873e-06, - "loss": 0.8897, - "step": 3101 - }, - { - "epoch": 0.6, - "learning_rate": 7.379217491262325e-06, - "loss": 0.8539, - "step": 3102 - }, - { - "epoch": 0.6, - "learning_rate": 7.373203750660505e-06, - "loss": 0.865, - "step": 3103 - }, - { - "epoch": 0.6, - "learning_rate": 7.36719103027608e-06, - "loss": 0.9112, - "step": 3104 - }, - { - "epoch": 0.6, - "learning_rate": 7.361179332444318e-06, - "loss": 0.8769, - "step": 3105 - }, - { - "epoch": 0.6, - "learning_rate": 7.355168659500094e-06, - "loss": 0.9667, - "step": 3106 - }, - { - "epoch": 0.6, - "learning_rate": 7.3491590137778915e-06, - "loss": 0.8715, - "step": 3107 - }, - { - "epoch": 0.6, - "learning_rate": 7.343150397611782e-06, - "loss": 0.9168, - "step": 3108 - }, - { - "epoch": 0.6, - "learning_rate": 7.3371428133354435e-06, - "loss": 0.794, - "step": 3109 - }, - { - "epoch": 0.6, - "learning_rate": 7.33113626328215e-06, - "loss": 0.8425, - "step": 3110 - }, - { - "epoch": 0.6, - "learning_rate": 7.325130749784781e-06, - "loss": 0.8522, - "step": 3111 - }, - { - "epoch": 0.6, - "learning_rate": 7.3191262751758005e-06, - "loss": 0.8595, - "step": 3112 - }, - { - "epoch": 0.6, - "learning_rate": 7.3131228417872905e-06, - "loss": 0.8415, - "step": 3113 - }, - { - "epoch": 0.6, - "learning_rate": 7.307120451950902e-06, - "loss": 0.8491, - "step": 3114 - }, - { - "epoch": 0.6, - "learning_rate": 7.301119107997905e-06, - "loss": 0.8894, - "step": 3115 - }, - { - "epoch": 0.6, - "learning_rate": 7.295118812259145e-06, - "loss": 0.8771, - "step": 3116 - }, - { - "epoch": 0.6, - "learning_rate": 7.289119567065068e-06, - "loss": 0.8953, - "step": 3117 - }, - { - "epoch": 0.6, - "learning_rate": 7.2831213747457155e-06, - "loss": 0.885, - "step": 3118 - }, - { - "epoch": 0.6, - "learning_rate": 7.2771242376307125e-06, - "loss": 0.8035, - "step": 3119 - }, - { - "epoch": 0.6, - "learning_rate": 7.271128158049283e-06, - "loss": 1.0118, - "step": 3120 - }, - { - "epoch": 0.6, - "learning_rate": 7.2651331383302326e-06, - "loss": 0.8093, - "step": 3121 - }, - { - "epoch": 0.6, - "learning_rate": 7.2591391808019555e-06, - "loss": 0.948, - "step": 3122 - }, - { - "epoch": 0.6, - "learning_rate": 7.253146287792434e-06, - "loss": 0.8714, - "step": 3123 - }, - { - "epoch": 0.6, - "learning_rate": 7.247154461629248e-06, - "loss": 0.9238, - "step": 3124 - }, - { - "epoch": 0.6, - "learning_rate": 7.241163704639547e-06, - "loss": 0.8504, - "step": 3125 - }, - { - "epoch": 0.6, - "learning_rate": 7.235174019150071e-06, - "loss": 0.9289, - "step": 3126 - }, - { - "epoch": 0.6, - "learning_rate": 7.229185407487149e-06, - "loss": 0.679, - "step": 3127 - }, - { - "epoch": 0.6, - "learning_rate": 7.2231978719766884e-06, - "loss": 0.8617, - "step": 3128 - }, - { - "epoch": 0.6, - "learning_rate": 7.217211414944171e-06, - "loss": 0.9091, - "step": 3129 - }, - { - "epoch": 0.6, - "learning_rate": 7.2112260387146784e-06, - "loss": 0.8271, - "step": 3130 - }, - { - "epoch": 0.6, - "learning_rate": 7.2052417456128565e-06, - "loss": 0.875, - "step": 3131 - }, - { - "epoch": 0.6, - "learning_rate": 7.199258537962936e-06, - "loss": 0.947, - "step": 3132 - }, - { - "epoch": 0.6, - "learning_rate": 7.193276418088729e-06, - "loss": 0.8265, - "step": 3133 - }, - { - "epoch": 0.6, - "learning_rate": 7.187295388313618e-06, - "loss": 0.9132, - "step": 3134 - }, - { - "epoch": 0.6, - "learning_rate": 7.181315450960562e-06, - "loss": 0.9571, - "step": 3135 - }, - { - "epoch": 0.6, - "learning_rate": 7.175336608352113e-06, - "loss": 0.9114, - "step": 3136 - }, - { - "epoch": 0.6, - "learning_rate": 7.169358862810374e-06, - "loss": 0.9099, - "step": 3137 - }, - { - "epoch": 0.6, - "learning_rate": 7.163382216657033e-06, - "loss": 0.8914, - "step": 3138 - }, - { - "epoch": 0.6, - "learning_rate": 7.1574066722133565e-06, - "loss": 0.836, - "step": 3139 - }, - { - "epoch": 0.6, - "learning_rate": 7.151432231800173e-06, - "loss": 0.8157, - "step": 3140 - }, - { - "epoch": 0.6, - "learning_rate": 7.145458897737882e-06, - "loss": 0.8662, - "step": 3141 - }, - { - "epoch": 0.6, - "learning_rate": 7.139486672346466e-06, - "loss": 0.8307, - "step": 3142 - }, - { - "epoch": 0.6, - "learning_rate": 7.133515557945463e-06, - "loss": 0.9195, - "step": 3143 - }, - { - "epoch": 0.6, - "learning_rate": 7.12754555685399e-06, - "loss": 0.8797, - "step": 3144 - }, - { - "epoch": 0.61, - "learning_rate": 7.121576671390722e-06, - "loss": 0.8837, - "step": 3145 - }, - { - "epoch": 0.61, - "learning_rate": 7.115608903873905e-06, - "loss": 0.9033, - "step": 3146 - }, - { - "epoch": 0.61, - "learning_rate": 7.109642256621353e-06, - "loss": 0.9742, - "step": 3147 - }, - { - "epoch": 0.61, - "learning_rate": 7.103676731950443e-06, - "loss": 0.8231, - "step": 3148 - }, - { - "epoch": 0.61, - "learning_rate": 7.0977123321781176e-06, - "loss": 0.9419, - "step": 3149 - }, - { - "epoch": 0.61, - "learning_rate": 7.091749059620881e-06, - "loss": 0.8312, - "step": 3150 - }, - { - "epoch": 0.61, - "learning_rate": 7.0857869165947945e-06, - "loss": 0.9108, - "step": 3151 - }, - { - "epoch": 0.61, - "learning_rate": 7.079825905415491e-06, - "loss": 0.8885, - "step": 3152 - }, - { - "epoch": 0.61, - "learning_rate": 7.073866028398153e-06, - "loss": 0.926, - "step": 3153 - }, - { - "epoch": 0.61, - "learning_rate": 7.067907287857535e-06, - "loss": 0.885, - "step": 3154 - }, - { - "epoch": 0.61, - "learning_rate": 7.061949686107938e-06, - "loss": 0.8853, - "step": 3155 - }, - { - "epoch": 0.61, - "learning_rate": 7.0559932254632315e-06, - "loss": 0.8813, - "step": 3156 - }, - { - "epoch": 0.61, - "learning_rate": 7.0500379082368305e-06, - "loss": 0.872, - "step": 3157 - }, - { - "epoch": 0.61, - "learning_rate": 7.044083736741711e-06, - "loss": 0.886, - "step": 3158 - }, - { - "epoch": 0.61, - "learning_rate": 7.03813071329041e-06, - "loss": 0.9342, - "step": 3159 - }, - { - "epoch": 0.61, - "learning_rate": 7.032178840195009e-06, - "loss": 0.8479, - "step": 3160 - }, - { - "epoch": 0.61, - "learning_rate": 7.026228119767149e-06, - "loss": 0.844, - "step": 3161 - }, - { - "epoch": 0.61, - "learning_rate": 7.020278554318023e-06, - "loss": 0.9592, - "step": 3162 - }, - { - "epoch": 0.61, - "learning_rate": 7.014330146158367e-06, - "loss": 0.8259, - "step": 3163 - }, - { - "epoch": 0.61, - "learning_rate": 7.008382897598477e-06, - "loss": 0.7254, - "step": 3164 - }, - { - "epoch": 0.61, - "learning_rate": 7.002436810948201e-06, - "loss": 0.8544, - "step": 3165 - }, - { - "epoch": 0.61, - "learning_rate": 6.996491888516927e-06, - "loss": 0.9462, - "step": 3166 - }, - { - "epoch": 0.61, - "learning_rate": 6.990548132613592e-06, - "loss": 0.9448, - "step": 3167 - }, - { - "epoch": 0.61, - "learning_rate": 6.984605545546686e-06, - "loss": 0.7423, - "step": 3168 - }, - { - "epoch": 0.61, - "learning_rate": 6.978664129624241e-06, - "loss": 0.8521, - "step": 3169 - }, - { - "epoch": 0.61, - "learning_rate": 6.972723887153828e-06, - "loss": 0.8847, - "step": 3170 - }, - { - "epoch": 0.61, - "learning_rate": 6.9667848204425785e-06, - "loss": 0.8356, - "step": 3171 - }, - { - "epoch": 0.61, - "learning_rate": 6.960846931797152e-06, - "loss": 0.9197, - "step": 3172 - }, - { - "epoch": 0.61, - "learning_rate": 6.9549102235237565e-06, - "loss": 0.7449, - "step": 3173 - }, - { - "epoch": 0.61, - "learning_rate": 6.948974697928144e-06, - "loss": 0.9142, - "step": 3174 - }, - { - "epoch": 0.61, - "learning_rate": 6.943040357315598e-06, - "loss": 0.8335, - "step": 3175 - }, - { - "epoch": 0.61, - "learning_rate": 6.9371072039909515e-06, - "loss": 0.9555, - "step": 3176 - }, - { - "epoch": 0.61, - "learning_rate": 6.931175240258576e-06, - "loss": 0.8719, - "step": 3177 - }, - { - "epoch": 0.61, - "learning_rate": 6.9252444684223765e-06, - "loss": 0.7977, - "step": 3178 - }, - { - "epoch": 0.61, - "learning_rate": 6.919314890785793e-06, - "loss": 0.9191, - "step": 3179 - }, - { - "epoch": 0.61, - "learning_rate": 6.913386509651807e-06, - "loss": 0.8103, - "step": 3180 - }, - { - "epoch": 0.61, - "learning_rate": 6.907459327322934e-06, - "loss": 0.8245, - "step": 3181 - }, - { - "epoch": 0.61, - "learning_rate": 6.90153334610122e-06, - "loss": 0.8543, - "step": 3182 - }, - { - "epoch": 0.61, - "learning_rate": 6.895608568288255e-06, - "loss": 0.8402, - "step": 3183 - }, - { - "epoch": 0.61, - "learning_rate": 6.889684996185148e-06, - "loss": 0.8808, - "step": 3184 - }, - { - "epoch": 0.61, - "learning_rate": 6.88376263209255e-06, - "loss": 0.9025, - "step": 3185 - }, - { - "epoch": 0.61, - "learning_rate": 6.877841478310639e-06, - "loss": 0.9383, - "step": 3186 - }, - { - "epoch": 0.61, - "learning_rate": 6.871921537139117e-06, - "loss": 0.8526, - "step": 3187 - }, - { - "epoch": 0.61, - "learning_rate": 6.866002810877224e-06, - "loss": 0.8474, - "step": 3188 - }, - { - "epoch": 0.61, - "learning_rate": 6.860085301823729e-06, - "loss": 0.8614, - "step": 3189 - }, - { - "epoch": 0.61, - "learning_rate": 6.854169012276923e-06, - "loss": 0.9258, - "step": 3190 - }, - { - "epoch": 0.61, - "learning_rate": 6.848253944534622e-06, - "loss": 0.8471, - "step": 3191 - }, - { - "epoch": 0.61, - "learning_rate": 6.84234010089417e-06, - "loss": 0.8652, - "step": 3192 - }, - { - "epoch": 0.61, - "learning_rate": 6.836427483652436e-06, - "loss": 0.6994, - "step": 3193 - }, - { - "epoch": 0.61, - "learning_rate": 6.830516095105817e-06, - "loss": 0.8856, - "step": 3194 - }, - { - "epoch": 0.61, - "learning_rate": 6.824605937550224e-06, - "loss": 0.965, - "step": 3195 - }, - { - "epoch": 0.61, - "learning_rate": 6.818697013281093e-06, - "loss": 0.916, - "step": 3196 - }, - { - "epoch": 0.62, - "learning_rate": 6.8127893245933864e-06, - "loss": 0.9063, - "step": 3197 - }, - { - "epoch": 0.62, - "learning_rate": 6.806882873781579e-06, - "loss": 0.9315, - "step": 3198 - }, - { - "epoch": 0.62, - "learning_rate": 6.800977663139666e-06, - "loss": 0.7226, - "step": 3199 - }, - { - "epoch": 0.62, - "learning_rate": 6.795073694961171e-06, - "loss": 0.8625, - "step": 3200 - }, - { - "epoch": 0.62, - "learning_rate": 6.789170971539119e-06, - "loss": 0.8614, - "step": 3201 - }, - { - "epoch": 0.62, - "learning_rate": 6.783269495166066e-06, - "loss": 0.9322, - "step": 3202 - }, - { - "epoch": 0.62, - "learning_rate": 6.777369268134076e-06, - "loss": 0.9223, - "step": 3203 - }, - { - "epoch": 0.62, - "learning_rate": 6.771470292734723e-06, - "loss": 0.9015, - "step": 3204 - }, - { - "epoch": 0.62, - "learning_rate": 6.7655725712591055e-06, - "loss": 0.8655, - "step": 3205 - }, - { - "epoch": 0.62, - "learning_rate": 6.759676105997834e-06, - "loss": 0.8469, - "step": 3206 - }, - { - "epoch": 0.62, - "learning_rate": 6.753780899241027e-06, - "loss": 0.8959, - "step": 3207 - }, - { - "epoch": 0.62, - "learning_rate": 6.747886953278311e-06, - "loss": 0.7899, - "step": 3208 - }, - { - "epoch": 0.62, - "learning_rate": 6.741994270398826e-06, - "loss": 0.8779, - "step": 3209 - }, - { - "epoch": 0.62, - "learning_rate": 6.736102852891227e-06, - "loss": 0.8419, - "step": 3210 - }, - { - "epoch": 0.62, - "learning_rate": 6.730212703043666e-06, - "loss": 0.8585, - "step": 3211 - }, - { - "epoch": 0.62, - "learning_rate": 6.7243238231438176e-06, - "loss": 0.8863, - "step": 3212 - }, - { - "epoch": 0.62, - "learning_rate": 6.718436215478849e-06, - "loss": 0.8747, - "step": 3213 - }, - { - "epoch": 0.62, - "learning_rate": 6.712549882335442e-06, - "loss": 0.9413, - "step": 3214 - }, - { - "epoch": 0.62, - "learning_rate": 6.70666482599978e-06, - "loss": 0.8779, - "step": 3215 - }, - { - "epoch": 0.62, - "learning_rate": 6.7007810487575475e-06, - "loss": 0.9123, - "step": 3216 - }, - { - "epoch": 0.62, - "learning_rate": 6.694898552893941e-06, - "loss": 0.8323, - "step": 3217 - }, - { - "epoch": 0.62, - "learning_rate": 6.6890173406936485e-06, - "loss": 0.8682, - "step": 3218 - }, - { - "epoch": 0.62, - "learning_rate": 6.683137414440872e-06, - "loss": 0.8272, - "step": 3219 - }, - { - "epoch": 0.62, - "learning_rate": 6.677258776419304e-06, - "loss": 0.8782, - "step": 3220 - }, - { - "epoch": 0.62, - "learning_rate": 6.671381428912138e-06, - "loss": 0.8477, - "step": 3221 - }, - { - "epoch": 0.62, - "learning_rate": 6.66550537420207e-06, - "loss": 0.759, - "step": 3222 - }, - { - "epoch": 0.62, - "learning_rate": 6.659630614571287e-06, - "loss": 0.8607, - "step": 3223 - }, - { - "epoch": 0.62, - "learning_rate": 6.653757152301488e-06, - "loss": 0.8763, - "step": 3224 - }, - { - "epoch": 0.62, - "learning_rate": 6.647884989673849e-06, - "loss": 0.6657, - "step": 3225 - }, - { - "epoch": 0.62, - "learning_rate": 6.642014128969055e-06, - "loss": 0.8868, - "step": 3226 - }, - { - "epoch": 0.62, - "learning_rate": 6.63614457246728e-06, - "loss": 0.8906, - "step": 3227 - }, - { - "epoch": 0.62, - "learning_rate": 6.630276322448188e-06, - "loss": 0.8969, - "step": 3228 - }, - { - "epoch": 0.62, - "learning_rate": 6.624409381190946e-06, - "loss": 0.8657, - "step": 3229 - }, - { - "epoch": 0.62, - "learning_rate": 6.618543750974202e-06, - "loss": 0.8128, - "step": 3230 - }, - { - "epoch": 0.62, - "learning_rate": 6.6126794340761025e-06, - "loss": 0.8548, - "step": 3231 - }, - { - "epoch": 0.62, - "learning_rate": 6.606816432774279e-06, - "loss": 0.8911, - "step": 3232 - }, - { - "epoch": 0.62, - "learning_rate": 6.600954749345851e-06, - "loss": 0.9662, - "step": 3233 - }, - { - "epoch": 0.62, - "learning_rate": 6.595094386067428e-06, - "loss": 0.9747, - "step": 3234 - }, - { - "epoch": 0.62, - "learning_rate": 6.589235345215117e-06, - "loss": 0.8771, - "step": 3235 - }, - { - "epoch": 0.62, - "learning_rate": 6.583377629064494e-06, - "loss": 0.9333, - "step": 3236 - }, - { - "epoch": 0.62, - "learning_rate": 6.5775212398906295e-06, - "loss": 0.9244, - "step": 3237 - }, - { - "epoch": 0.62, - "learning_rate": 6.571666179968079e-06, - "loss": 0.838, - "step": 3238 - }, - { - "epoch": 0.62, - "learning_rate": 6.565812451570881e-06, - "loss": 0.8762, - "step": 3239 - }, - { - "epoch": 0.62, - "learning_rate": 6.5599600569725495e-06, - "loss": 0.8799, - "step": 3240 - }, - { - "epoch": 0.62, - "learning_rate": 6.554108998446096e-06, - "loss": 0.9338, - "step": 3241 - }, - { - "epoch": 0.62, - "learning_rate": 6.548259278263999e-06, - "loss": 0.8844, - "step": 3242 - }, - { - "epoch": 0.62, - "learning_rate": 6.542410898698226e-06, - "loss": 0.8682, - "step": 3243 - }, - { - "epoch": 0.62, - "learning_rate": 6.536563862020218e-06, - "loss": 0.8472, - "step": 3244 - }, - { - "epoch": 0.62, - "learning_rate": 6.530718170500896e-06, - "loss": 0.9428, - "step": 3245 - }, - { - "epoch": 0.62, - "learning_rate": 6.524873826410658e-06, - "loss": 0.8258, - "step": 3246 - }, - { - "epoch": 0.62, - "learning_rate": 6.519030832019383e-06, - "loss": 0.8413, - "step": 3247 - }, - { - "epoch": 0.62, - "learning_rate": 6.513189189596422e-06, - "loss": 0.9257, - "step": 3248 - }, - { - "epoch": 0.63, - "learning_rate": 6.507348901410604e-06, - "loss": 0.862, - "step": 3249 - }, - { - "epoch": 0.63, - "learning_rate": 6.501509969730224e-06, - "loss": 0.7737, - "step": 3250 - }, - { - "epoch": 0.63, - "learning_rate": 6.495672396823061e-06, - "loss": 0.8848, - "step": 3251 - }, - { - "epoch": 0.63, - "learning_rate": 6.489836184956353e-06, - "loss": 0.9005, - "step": 3252 - }, - { - "epoch": 0.63, - "learning_rate": 6.484001336396828e-06, - "loss": 0.6919, - "step": 3253 - }, - { - "epoch": 0.63, - "learning_rate": 6.478167853410668e-06, - "loss": 0.9317, - "step": 3254 - }, - { - "epoch": 0.63, - "learning_rate": 6.472335738263534e-06, - "loss": 0.8888, - "step": 3255 - }, - { - "epoch": 0.63, - "learning_rate": 6.466504993220548e-06, - "loss": 0.8767, - "step": 3256 - }, - { - "epoch": 0.63, - "learning_rate": 6.460675620546305e-06, - "loss": 0.8391, - "step": 3257 - }, - { - "epoch": 0.63, - "learning_rate": 6.454847622504867e-06, - "loss": 0.8902, - "step": 3258 - }, - { - "epoch": 0.63, - "learning_rate": 6.4490210013597635e-06, - "loss": 0.8752, - "step": 3259 - }, - { - "epoch": 0.63, - "learning_rate": 6.4431957593739845e-06, - "loss": 0.9058, - "step": 3260 - }, - { - "epoch": 0.63, - "learning_rate": 6.4373718988099896e-06, - "loss": 0.9237, - "step": 3261 - }, - { - "epoch": 0.63, - "learning_rate": 6.431549421929694e-06, - "loss": 0.8913, - "step": 3262 - }, - { - "epoch": 0.63, - "learning_rate": 6.4257283309944804e-06, - "loss": 0.9349, - "step": 3263 - }, - { - "epoch": 0.63, - "learning_rate": 6.419908628265203e-06, - "loss": 0.7567, - "step": 3264 - }, - { - "epoch": 0.63, - "learning_rate": 6.414090316002161e-06, - "loss": 0.9852, - "step": 3265 - }, - { - "epoch": 0.63, - "learning_rate": 6.4082733964651166e-06, - "loss": 0.8274, - "step": 3266 - }, - { - "epoch": 0.63, - "learning_rate": 6.4024578719133e-06, - "loss": 0.826, - "step": 3267 - }, - { - "epoch": 0.63, - "learning_rate": 6.396643744605391e-06, - "loss": 0.7291, - "step": 3268 - }, - { - "epoch": 0.63, - "learning_rate": 6.390831016799527e-06, - "loss": 0.852, - "step": 3269 - }, - { - "epoch": 0.63, - "learning_rate": 6.385019690753311e-06, - "loss": 0.8411, - "step": 3270 - }, - { - "epoch": 0.63, - "learning_rate": 6.379209768723791e-06, - "loss": 0.8685, - "step": 3271 - }, - { - "epoch": 0.63, - "learning_rate": 6.373401252967475e-06, - "loss": 0.9221, - "step": 3272 - }, - { - "epoch": 0.63, - "learning_rate": 6.367594145740324e-06, - "loss": 0.8583, - "step": 3273 - }, - { - "epoch": 0.63, - "learning_rate": 6.361788449297748e-06, - "loss": 0.8588, - "step": 3274 - }, - { - "epoch": 0.63, - "learning_rate": 6.355984165894613e-06, - "loss": 0.88, - "step": 3275 - }, - { - "epoch": 0.63, - "learning_rate": 6.350181297785242e-06, - "loss": 0.9008, - "step": 3276 - }, - { - "epoch": 0.63, - "learning_rate": 6.344379847223398e-06, - "loss": 0.8856, - "step": 3277 - }, - { - "epoch": 0.63, - "learning_rate": 6.338579816462298e-06, - "loss": 0.8436, - "step": 3278 - }, - { - "epoch": 0.63, - "learning_rate": 6.332781207754605e-06, - "loss": 0.9714, - "step": 3279 - }, - { - "epoch": 0.63, - "learning_rate": 6.326984023352435e-06, - "loss": 0.7938, - "step": 3280 - }, - { - "epoch": 0.63, - "learning_rate": 6.321188265507342e-06, - "loss": 0.9164, - "step": 3281 - }, - { - "epoch": 0.63, - "learning_rate": 6.31539393647034e-06, - "loss": 0.897, - "step": 3282 - }, - { - "epoch": 0.63, - "learning_rate": 6.309601038491874e-06, - "loss": 0.8548, - "step": 3283 - }, - { - "epoch": 0.63, - "learning_rate": 6.303809573821842e-06, - "loss": 0.9489, - "step": 3284 - }, - { - "epoch": 0.63, - "learning_rate": 6.298019544709579e-06, - "loss": 0.9153, - "step": 3285 - }, - { - "epoch": 0.63, - "learning_rate": 6.292230953403866e-06, - "loss": 0.8736, - "step": 3286 - }, - { - "epoch": 0.63, - "learning_rate": 6.286443802152926e-06, - "loss": 0.865, - "step": 3287 - }, - { - "epoch": 0.63, - "learning_rate": 6.280658093204422e-06, - "loss": 0.8851, - "step": 3288 - }, - { - "epoch": 0.63, - "learning_rate": 6.274873828805459e-06, - "loss": 0.8698, - "step": 3289 - }, - { - "epoch": 0.63, - "learning_rate": 6.269091011202576e-06, - "loss": 0.8975, - "step": 3290 - }, - { - "epoch": 0.63, - "learning_rate": 6.263309642641751e-06, - "loss": 0.8947, - "step": 3291 - }, - { - "epoch": 0.63, - "learning_rate": 6.257529725368405e-06, - "loss": 0.8312, - "step": 3292 - }, - { - "epoch": 0.63, - "learning_rate": 6.251751261627386e-06, - "loss": 0.8929, - "step": 3293 - }, - { - "epoch": 0.63, - "learning_rate": 6.245974253662988e-06, - "loss": 0.9073, - "step": 3294 - }, - { - "epoch": 0.63, - "learning_rate": 6.240198703718932e-06, - "loss": 0.9108, - "step": 3295 - }, - { - "epoch": 0.63, - "learning_rate": 6.234424614038375e-06, - "loss": 0.8607, - "step": 3296 - }, - { - "epoch": 0.63, - "learning_rate": 6.2286519868639095e-06, - "loss": 0.8321, - "step": 3297 - }, - { - "epoch": 0.63, - "learning_rate": 6.222880824437549e-06, - "loss": 0.8987, - "step": 3298 - }, - { - "epoch": 0.63, - "learning_rate": 6.217111129000759e-06, - "loss": 0.9236, - "step": 3299 - }, - { - "epoch": 0.63, - "learning_rate": 6.211342902794413e-06, - "loss": 0.9541, - "step": 3300 - }, - { - "epoch": 0.64, - "learning_rate": 6.205576148058828e-06, - "loss": 0.8866, - "step": 3301 - }, - { - "epoch": 0.64, - "learning_rate": 6.199810867033745e-06, - "loss": 0.8975, - "step": 3302 - }, - { - "epoch": 0.64, - "learning_rate": 6.19404706195833e-06, - "loss": 0.8846, - "step": 3303 - }, - { - "epoch": 0.64, - "learning_rate": 6.188284735071177e-06, - "loss": 0.8482, - "step": 3304 - }, - { - "epoch": 0.64, - "learning_rate": 6.182523888610316e-06, - "loss": 0.8779, - "step": 3305 - }, - { - "epoch": 0.64, - "learning_rate": 6.176764524813187e-06, - "loss": 0.8456, - "step": 3306 - }, - { - "epoch": 0.64, - "learning_rate": 6.171006645916662e-06, - "loss": 0.9011, - "step": 3307 - }, - { - "epoch": 0.64, - "learning_rate": 6.165250254157032e-06, - "loss": 0.9143, - "step": 3308 - }, - { - "epoch": 0.64, - "learning_rate": 6.159495351770017e-06, - "loss": 0.894, - "step": 3309 - }, - { - "epoch": 0.64, - "learning_rate": 6.153741940990749e-06, - "loss": 0.9481, - "step": 3310 - }, - { - "epoch": 0.64, - "learning_rate": 6.1479900240537956e-06, - "loss": 0.95, - "step": 3311 - }, - { - "epoch": 0.64, - "learning_rate": 6.142239603193128e-06, - "loss": 0.8665, - "step": 3312 - }, - { - "epoch": 0.64, - "learning_rate": 6.136490680642146e-06, - "loss": 0.9131, - "step": 3313 - }, - { - "epoch": 0.64, - "learning_rate": 6.130743258633667e-06, - "loss": 0.6627, - "step": 3314 - }, - { - "epoch": 0.64, - "learning_rate": 6.124997339399916e-06, - "loss": 0.8999, - "step": 3315 - }, - { - "epoch": 0.64, - "learning_rate": 6.119252925172549e-06, - "loss": 0.8613, - "step": 3316 - }, - { - "epoch": 0.64, - "learning_rate": 6.113510018182628e-06, - "loss": 0.8623, - "step": 3317 - }, - { - "epoch": 0.64, - "learning_rate": 6.107768620660633e-06, - "loss": 0.9019, - "step": 3318 - }, - { - "epoch": 0.64, - "learning_rate": 6.102028734836456e-06, - "loss": 0.8543, - "step": 3319 - }, - { - "epoch": 0.64, - "learning_rate": 6.0962903629394e-06, - "loss": 0.8306, - "step": 3320 - }, - { - "epoch": 0.64, - "learning_rate": 6.090553507198187e-06, - "loss": 0.9682, - "step": 3321 - }, - { - "epoch": 0.64, - "learning_rate": 6.0848181698409384e-06, - "loss": 0.8169, - "step": 3322 - }, - { - "epoch": 0.64, - "learning_rate": 6.079084353095202e-06, - "loss": 0.8753, - "step": 3323 - }, - { - "epoch": 0.64, - "learning_rate": 6.07335205918792e-06, - "loss": 0.9053, - "step": 3324 - }, - { - "epoch": 0.64, - "learning_rate": 6.067621290345455e-06, - "loss": 0.9145, - "step": 3325 - }, - { - "epoch": 0.64, - "learning_rate": 6.061892048793568e-06, - "loss": 0.7894, - "step": 3326 - }, - { - "epoch": 0.64, - "learning_rate": 6.056164336757426e-06, - "loss": 0.88, - "step": 3327 - }, - { - "epoch": 0.64, - "learning_rate": 6.050438156461613e-06, - "loss": 0.9735, - "step": 3328 - }, - { - "epoch": 0.64, - "learning_rate": 6.044713510130108e-06, - "loss": 0.8226, - "step": 3329 - }, - { - "epoch": 0.64, - "learning_rate": 6.038990399986302e-06, - "loss": 0.9498, - "step": 3330 - }, - { - "epoch": 0.64, - "learning_rate": 6.03326882825298e-06, - "loss": 0.7733, - "step": 3331 - }, - { - "epoch": 0.64, - "learning_rate": 6.027548797152336e-06, - "loss": 0.8588, - "step": 3332 - }, - { - "epoch": 0.64, - "learning_rate": 6.021830308905963e-06, - "loss": 0.8368, - "step": 3333 - }, - { - "epoch": 0.64, - "learning_rate": 6.016113365734861e-06, - "loss": 0.9115, - "step": 3334 - }, - { - "epoch": 0.64, - "learning_rate": 6.0103979698594215e-06, - "loss": 0.8869, - "step": 3335 - }, - { - "epoch": 0.64, - "learning_rate": 6.004684123499436e-06, - "loss": 0.7611, - "step": 3336 - }, - { - "epoch": 0.64, - "learning_rate": 5.998971828874102e-06, - "loss": 0.8452, - "step": 3337 - }, - { - "epoch": 0.64, - "learning_rate": 5.993261088202005e-06, - "loss": 0.7587, - "step": 3338 - }, - { - "epoch": 0.64, - "learning_rate": 5.987551903701128e-06, - "loss": 0.6805, - "step": 3339 - }, - { - "epoch": 0.64, - "learning_rate": 5.9818442775888595e-06, - "loss": 0.8556, - "step": 3340 - }, - { - "epoch": 0.64, - "learning_rate": 5.97613821208197e-06, - "loss": 0.8635, - "step": 3341 - }, - { - "epoch": 0.64, - "learning_rate": 5.970433709396635e-06, - "loss": 0.9317, - "step": 3342 - }, - { - "epoch": 0.64, - "learning_rate": 5.964730771748415e-06, - "loss": 0.83, - "step": 3343 - }, - { - "epoch": 0.64, - "learning_rate": 5.959029401352262e-06, - "loss": 0.8693, - "step": 3344 - }, - { - "epoch": 0.64, - "learning_rate": 5.953329600422524e-06, - "loss": 0.8635, - "step": 3345 - }, - { - "epoch": 0.64, - "learning_rate": 5.947631371172943e-06, - "loss": 0.7288, - "step": 3346 - }, - { - "epoch": 0.64, - "learning_rate": 5.941934715816642e-06, - "loss": 0.8163, - "step": 3347 - }, - { - "epoch": 0.64, - "learning_rate": 5.936239636566137e-06, - "loss": 0.9157, - "step": 3348 - }, - { - "epoch": 0.64, - "learning_rate": 5.930546135633327e-06, - "loss": 0.9573, - "step": 3349 - }, - { - "epoch": 0.64, - "learning_rate": 5.924854215229509e-06, - "loss": 0.8809, - "step": 3350 - }, - { - "epoch": 0.64, - "learning_rate": 5.919163877565351e-06, - "loss": 0.8678, - "step": 3351 - }, - { - "epoch": 0.64, - "learning_rate": 5.9134751248509236e-06, - "loss": 0.903, - "step": 3352 - }, - { - "epoch": 0.65, - "learning_rate": 5.9077879592956675e-06, - "loss": 0.8575, - "step": 3353 - }, - { - "epoch": 0.65, - "learning_rate": 5.902102383108415e-06, - "loss": 0.8512, - "step": 3354 - }, - { - "epoch": 0.65, - "learning_rate": 5.896418398497377e-06, - "loss": 0.9659, - "step": 3355 - }, - { - "epoch": 0.65, - "learning_rate": 5.890736007670144e-06, - "loss": 0.8293, - "step": 3356 - }, - { - "epoch": 0.65, - "learning_rate": 5.885055212833696e-06, - "loss": 0.8737, - "step": 3357 - }, - { - "epoch": 0.65, - "learning_rate": 5.879376016194387e-06, - "loss": 0.6522, - "step": 3358 - }, - { - "epoch": 0.65, - "learning_rate": 5.873698419957952e-06, - "loss": 0.8843, - "step": 3359 - }, - { - "epoch": 0.65, - "learning_rate": 5.8680224263295045e-06, - "loss": 0.7709, - "step": 3360 - }, - { - "epoch": 0.65, - "learning_rate": 5.862348037513533e-06, - "loss": 0.892, - "step": 3361 - }, - { - "epoch": 0.65, - "learning_rate": 5.856675255713905e-06, - "loss": 0.9165, - "step": 3362 - }, - { - "epoch": 0.65, - "learning_rate": 5.851004083133862e-06, - "loss": 0.9047, - "step": 3363 - }, - { - "epoch": 0.65, - "learning_rate": 5.8453345219760275e-06, - "loss": 0.826, - "step": 3364 - }, - { - "epoch": 0.65, - "learning_rate": 5.839666574442389e-06, - "loss": 0.7208, - "step": 3365 - }, - { - "epoch": 0.65, - "learning_rate": 5.834000242734317e-06, - "loss": 0.8939, - "step": 3366 - }, - { - "epoch": 0.65, - "learning_rate": 5.828335529052541e-06, - "loss": 0.8737, - "step": 3367 - }, - { - "epoch": 0.65, - "learning_rate": 5.822672435597172e-06, - "loss": 0.8985, - "step": 3368 - }, - { - "epoch": 0.65, - "learning_rate": 5.817010964567702e-06, - "loss": 0.8743, - "step": 3369 - }, - { - "epoch": 0.65, - "learning_rate": 5.811351118162969e-06, - "loss": 0.7145, - "step": 3370 - }, - { - "epoch": 0.65, - "learning_rate": 5.805692898581196e-06, - "loss": 0.9047, - "step": 3371 - }, - { - "epoch": 0.65, - "learning_rate": 5.800036308019974e-06, - "loss": 0.9071, - "step": 3372 - }, - { - "epoch": 0.65, - "learning_rate": 5.79438134867625e-06, - "loss": 0.968, - "step": 3373 - }, - { - "epoch": 0.65, - "learning_rate": 5.788728022746348e-06, - "loss": 0.9785, - "step": 3374 - }, - { - "epoch": 0.65, - "learning_rate": 5.783076332425957e-06, - "loss": 0.8974, - "step": 3375 - }, - { - "epoch": 0.65, - "learning_rate": 5.777426279910125e-06, - "loss": 0.9414, - "step": 3376 - }, - { - "epoch": 0.65, - "learning_rate": 5.771777867393275e-06, - "loss": 0.7054, - "step": 3377 - }, - { - "epoch": 0.65, - "learning_rate": 5.766131097069174e-06, - "loss": 0.8539, - "step": 3378 - }, - { - "epoch": 0.65, - "learning_rate": 5.760485971130969e-06, - "loss": 0.8977, - "step": 3379 - }, - { - "epoch": 0.65, - "learning_rate": 5.7548424917711596e-06, - "loss": 0.7908, - "step": 3380 - }, - { - "epoch": 0.65, - "learning_rate": 5.749200661181611e-06, - "loss": 0.893, - "step": 3381 - }, - { - "epoch": 0.65, - "learning_rate": 5.7435604815535475e-06, - "loss": 0.8218, - "step": 3382 - }, - { - "epoch": 0.65, - "learning_rate": 5.7379219550775415e-06, - "loss": 0.8898, - "step": 3383 - }, - { - "epoch": 0.65, - "learning_rate": 5.732285083943537e-06, - "loss": 0.9061, - "step": 3384 - }, - { - "epoch": 0.65, - "learning_rate": 5.726649870340833e-06, - "loss": 0.8622, - "step": 3385 - }, - { - "epoch": 0.65, - "learning_rate": 5.721016316458068e-06, - "loss": 0.8416, - "step": 3386 - }, - { - "epoch": 0.65, - "learning_rate": 5.715384424483268e-06, - "loss": 0.8886, - "step": 3387 - }, - { - "epoch": 0.65, - "learning_rate": 5.709754196603781e-06, - "loss": 0.9764, - "step": 3388 - }, - { - "epoch": 0.65, - "learning_rate": 5.704125635006329e-06, - "loss": 0.9488, - "step": 3389 - }, - { - "epoch": 0.65, - "learning_rate": 5.6984987418769825e-06, - "loss": 0.8255, - "step": 3390 - }, - { - "epoch": 0.65, - "learning_rate": 5.692873519401154e-06, - "loss": 0.8415, - "step": 3391 - }, - { - "epoch": 0.65, - "learning_rate": 5.6872499697636195e-06, - "loss": 0.8785, - "step": 3392 - }, - { - "epoch": 0.65, - "learning_rate": 5.681628095148502e-06, - "loss": 0.8294, - "step": 3393 - }, - { - "epoch": 0.65, - "learning_rate": 5.6760078977392706e-06, - "loss": 0.8842, - "step": 3394 - }, - { - "epoch": 0.65, - "learning_rate": 5.67038937971875e-06, - "loss": 0.8878, - "step": 3395 - }, - { - "epoch": 0.65, - "learning_rate": 5.664772543269101e-06, - "loss": 0.7258, - "step": 3396 - }, - { - "epoch": 0.65, - "learning_rate": 5.659157390571842e-06, - "loss": 0.902, - "step": 3397 - }, - { - "epoch": 0.65, - "learning_rate": 5.653543923807833e-06, - "loss": 0.7385, - "step": 3398 - }, - { - "epoch": 0.65, - "learning_rate": 5.6479321451572785e-06, - "loss": 0.9391, - "step": 3399 - }, - { - "epoch": 0.65, - "learning_rate": 5.642322056799732e-06, - "loss": 0.8337, - "step": 3400 - }, - { - "epoch": 0.65, - "learning_rate": 5.636713660914087e-06, - "loss": 0.9222, - "step": 3401 - }, - { - "epoch": 0.65, - "learning_rate": 5.631106959678575e-06, - "loss": 0.8966, - "step": 3402 - }, - { - "epoch": 0.65, - "learning_rate": 5.625501955270777e-06, - "loss": 0.8797, - "step": 3403 - }, - { - "epoch": 0.65, - "learning_rate": 5.619898649867612e-06, - "loss": 0.908, - "step": 3404 - }, - { - "epoch": 0.66, - "learning_rate": 5.614297045645339e-06, - "loss": 0.7465, - "step": 3405 - }, - { - "epoch": 0.66, - "learning_rate": 5.6086971447795625e-06, - "loss": 0.656, - "step": 3406 - }, - { - "epoch": 0.66, - "learning_rate": 5.603098949445209e-06, - "loss": 0.885, - "step": 3407 - }, - { - "epoch": 0.66, - "learning_rate": 5.597502461816557e-06, - "loss": 0.6651, - "step": 3408 - }, - { - "epoch": 0.66, - "learning_rate": 5.5919076840672215e-06, - "loss": 0.809, - "step": 3409 - }, - { - "epoch": 0.66, - "learning_rate": 5.5863146183701454e-06, - "loss": 0.8328, - "step": 3410 - }, - { - "epoch": 0.66, - "learning_rate": 5.580723266897616e-06, - "loss": 0.88, - "step": 3411 - }, - { - "epoch": 0.66, - "learning_rate": 5.575133631821243e-06, - "loss": 0.803, - "step": 3412 - }, - { - "epoch": 0.66, - "learning_rate": 5.5695457153119806e-06, - "loss": 0.8589, - "step": 3413 - }, - { - "epoch": 0.66, - "learning_rate": 5.563959519540114e-06, - "loss": 0.8304, - "step": 3414 - }, - { - "epoch": 0.66, - "learning_rate": 5.558375046675244e-06, - "loss": 0.933, - "step": 3415 - }, - { - "epoch": 0.66, - "learning_rate": 5.552792298886335e-06, - "loss": 0.9321, - "step": 3416 - }, - { - "epoch": 0.66, - "learning_rate": 5.547211278341646e-06, - "loss": 0.7498, - "step": 3417 - }, - { - "epoch": 0.66, - "learning_rate": 5.541631987208789e-06, - "loss": 0.8368, - "step": 3418 - }, - { - "epoch": 0.66, - "learning_rate": 5.536054427654698e-06, - "loss": 0.797, - "step": 3419 - }, - { - "epoch": 0.66, - "learning_rate": 5.530478601845624e-06, - "loss": 0.7334, - "step": 3420 - }, - { - "epoch": 0.66, - "learning_rate": 5.52490451194716e-06, - "loss": 0.8332, - "step": 3421 - }, - { - "epoch": 0.66, - "learning_rate": 5.519332160124215e-06, - "loss": 0.7323, - "step": 3422 - }, - { - "epoch": 0.66, - "learning_rate": 5.513761548541032e-06, - "loss": 0.9174, - "step": 3423 - }, - { - "epoch": 0.66, - "learning_rate": 5.508192679361169e-06, - "loss": 0.8729, - "step": 3424 - }, - { - "epoch": 0.66, - "learning_rate": 5.502625554747508e-06, - "loss": 0.7946, - "step": 3425 - }, - { - "epoch": 0.66, - "learning_rate": 5.497060176862259e-06, - "loss": 0.8298, - "step": 3426 - }, - { - "epoch": 0.66, - "learning_rate": 5.491496547866948e-06, - "loss": 0.7491, - "step": 3427 - }, - { - "epoch": 0.66, - "learning_rate": 5.485934669922428e-06, - "loss": 0.8107, - "step": 3428 - }, - { - "epoch": 0.66, - "learning_rate": 5.480374545188866e-06, - "loss": 0.7118, - "step": 3429 - }, - { - "epoch": 0.66, - "learning_rate": 5.474816175825754e-06, - "loss": 0.8815, - "step": 3430 - }, - { - "epoch": 0.66, - "learning_rate": 5.469259563991894e-06, - "loss": 0.8537, - "step": 3431 - }, - { - "epoch": 0.66, - "learning_rate": 5.46370471184541e-06, - "loss": 0.8679, - "step": 3432 - }, - { - "epoch": 0.66, - "learning_rate": 5.458151621543744e-06, - "loss": 0.9358, - "step": 3433 - }, - { - "epoch": 0.66, - "learning_rate": 5.452600295243653e-06, - "loss": 0.9375, - "step": 3434 - }, - { - "epoch": 0.66, - "learning_rate": 5.4470507351012116e-06, - "loss": 0.9342, - "step": 3435 - }, - { - "epoch": 0.66, - "learning_rate": 5.441502943271797e-06, - "loss": 0.9601, - "step": 3436 - }, - { - "epoch": 0.66, - "learning_rate": 5.4359569219101115e-06, - "loss": 0.9006, - "step": 3437 - }, - { - "epoch": 0.66, - "learning_rate": 5.430412673170167e-06, - "loss": 0.7894, - "step": 3438 - }, - { - "epoch": 0.66, - "learning_rate": 5.424870199205283e-06, - "loss": 0.8682, - "step": 3439 - }, - { - "epoch": 0.66, - "learning_rate": 5.4193295021681e-06, - "loss": 0.93, - "step": 3440 - }, - { - "epoch": 0.66, - "learning_rate": 5.413790584210551e-06, - "loss": 0.8829, - "step": 3441 - }, - { - "epoch": 0.66, - "learning_rate": 5.408253447483892e-06, - "loss": 0.862, - "step": 3442 - }, - { - "epoch": 0.66, - "learning_rate": 5.402718094138688e-06, - "loss": 0.9585, - "step": 3443 - }, - { - "epoch": 0.66, - "learning_rate": 5.397184526324792e-06, - "loss": 0.899, - "step": 3444 - }, - { - "epoch": 0.66, - "learning_rate": 5.391652746191398e-06, - "loss": 0.8556, - "step": 3445 - }, - { - "epoch": 0.66, - "learning_rate": 5.38612275588697e-06, - "loss": 0.7897, - "step": 3446 - }, - { - "epoch": 0.66, - "learning_rate": 5.380594557559298e-06, - "loss": 0.9103, - "step": 3447 - }, - { - "epoch": 0.66, - "learning_rate": 5.375068153355474e-06, - "loss": 0.8238, - "step": 3448 - }, - { - "epoch": 0.66, - "learning_rate": 5.369543545421883e-06, - "loss": 0.891, - "step": 3449 - }, - { - "epoch": 0.66, - "learning_rate": 5.364020735904223e-06, - "loss": 0.8113, - "step": 3450 - }, - { - "epoch": 0.66, - "learning_rate": 5.358499726947488e-06, - "loss": 0.9078, - "step": 3451 - }, - { - "epoch": 0.66, - "learning_rate": 5.352980520695974e-06, - "loss": 0.7697, - "step": 3452 - }, - { - "epoch": 0.66, - "learning_rate": 5.347463119293283e-06, - "loss": 0.8568, - "step": 3453 - }, - { - "epoch": 0.66, - "learning_rate": 5.341947524882301e-06, - "loss": 0.8858, - "step": 3454 - }, - { - "epoch": 0.66, - "learning_rate": 5.336433739605227e-06, - "loss": 0.8988, - "step": 3455 - }, - { - "epoch": 0.66, - "learning_rate": 5.330921765603549e-06, - "loss": 0.8316, - "step": 3456 - }, - { - "epoch": 0.67, - "learning_rate": 5.325411605018056e-06, - "loss": 0.9104, - "step": 3457 - }, - { - "epoch": 0.67, - "learning_rate": 5.31990325998883e-06, - "loss": 0.8739, - "step": 3458 - }, - { - "epoch": 0.67, - "learning_rate": 5.314396732655253e-06, - "loss": 0.8591, - "step": 3459 - }, - { - "epoch": 0.67, - "learning_rate": 5.308892025155989e-06, - "loss": 0.8701, - "step": 3460 - }, - { - "epoch": 0.67, - "learning_rate": 5.303389139629007e-06, - "loss": 0.7471, - "step": 3461 - }, - { - "epoch": 0.67, - "learning_rate": 5.297888078211564e-06, - "loss": 0.8851, - "step": 3462 - }, - { - "epoch": 0.67, - "learning_rate": 5.2923888430402085e-06, - "loss": 0.9039, - "step": 3463 - }, - { - "epoch": 0.67, - "learning_rate": 5.286891436250785e-06, - "loss": 0.9115, - "step": 3464 - }, - { - "epoch": 0.67, - "learning_rate": 5.281395859978414e-06, - "loss": 0.8882, - "step": 3465 - }, - { - "epoch": 0.67, - "learning_rate": 5.2759021163575184e-06, - "loss": 0.8886, - "step": 3466 - }, - { - "epoch": 0.67, - "learning_rate": 5.27041020752181e-06, - "loss": 0.6834, - "step": 3467 - }, - { - "epoch": 0.67, - "learning_rate": 5.26492013560427e-06, - "loss": 0.8955, - "step": 3468 - }, - { - "epoch": 0.67, - "learning_rate": 5.259431902737195e-06, - "loss": 0.8938, - "step": 3469 - }, - { - "epoch": 0.67, - "learning_rate": 5.2539455110521385e-06, - "loss": 0.8951, - "step": 3470 - }, - { - "epoch": 0.67, - "learning_rate": 5.248460962679958e-06, - "loss": 0.9199, - "step": 3471 - }, - { - "epoch": 0.67, - "learning_rate": 5.24297825975079e-06, - "loss": 0.8425, - "step": 3472 - }, - { - "epoch": 0.67, - "learning_rate": 5.237497404394044e-06, - "loss": 0.8511, - "step": 3473 - }, - { - "epoch": 0.67, - "learning_rate": 5.232018398738436e-06, - "loss": 0.6829, - "step": 3474 - }, - { - "epoch": 0.67, - "learning_rate": 5.226541244911936e-06, - "loss": 0.7927, - "step": 3475 - }, - { - "epoch": 0.67, - "learning_rate": 5.221065945041811e-06, - "loss": 0.8729, - "step": 3476 - }, - { - "epoch": 0.67, - "learning_rate": 5.215592501254609e-06, - "loss": 0.8107, - "step": 3477 - }, - { - "epoch": 0.67, - "learning_rate": 5.210120915676147e-06, - "loss": 0.843, - "step": 3478 - }, - { - "epoch": 0.67, - "learning_rate": 5.2046511904315265e-06, - "loss": 0.8555, - "step": 3479 - }, - { - "epoch": 0.67, - "learning_rate": 5.199183327645128e-06, - "loss": 0.8752, - "step": 3480 - }, - { - "epoch": 0.67, - "learning_rate": 5.193717329440604e-06, - "loss": 0.7753, - "step": 3481 - }, - { - "epoch": 0.67, - "learning_rate": 5.188253197940889e-06, - "loss": 0.8891, - "step": 3482 - }, - { - "epoch": 0.67, - "learning_rate": 5.182790935268185e-06, - "loss": 0.8157, - "step": 3483 - }, - { - "epoch": 0.67, - "learning_rate": 5.177330543543971e-06, - "loss": 0.9036, - "step": 3484 - }, - { - "epoch": 0.67, - "learning_rate": 5.171872024889004e-06, - "loss": 0.8752, - "step": 3485 - }, - { - "epoch": 0.67, - "learning_rate": 5.166415381423306e-06, - "loss": 0.9144, - "step": 3486 - }, - { - "epoch": 0.67, - "learning_rate": 5.160960615266179e-06, - "loss": 0.8113, - "step": 3487 - }, - { - "epoch": 0.67, - "learning_rate": 5.155507728536191e-06, - "loss": 0.8643, - "step": 3488 - }, - { - "epoch": 0.67, - "learning_rate": 5.150056723351173e-06, - "loss": 0.878, - "step": 3489 - }, - { - "epoch": 0.67, - "learning_rate": 5.14460760182824e-06, - "loss": 0.8512, - "step": 3490 - }, - { - "epoch": 0.67, - "learning_rate": 5.139160366083765e-06, - "loss": 0.851, - "step": 3491 - }, - { - "epoch": 0.67, - "learning_rate": 5.133715018233393e-06, - "loss": 0.8983, - "step": 3492 - }, - { - "epoch": 0.67, - "learning_rate": 5.128271560392037e-06, - "loss": 0.89, - "step": 3493 - }, - { - "epoch": 0.67, - "learning_rate": 5.122829994673866e-06, - "loss": 0.8762, - "step": 3494 - }, - { - "epoch": 0.67, - "learning_rate": 5.117390323192326e-06, - "loss": 0.7842, - "step": 3495 - }, - { - "epoch": 0.67, - "learning_rate": 5.111952548060126e-06, - "loss": 0.8934, - "step": 3496 - }, - { - "epoch": 0.67, - "learning_rate": 5.106516671389224e-06, - "loss": 0.8377, - "step": 3497 - }, - { - "epoch": 0.67, - "learning_rate": 5.101082695290866e-06, - "loss": 0.8428, - "step": 3498 - }, - { - "epoch": 0.67, - "learning_rate": 5.0956506218755344e-06, - "loss": 0.8928, - "step": 3499 - }, - { - "epoch": 0.67, - "learning_rate": 5.09022045325299e-06, - "loss": 0.9227, - "step": 3500 - }, - { - "epoch": 0.67, - "learning_rate": 5.0847921915322486e-06, - "loss": 0.904, - "step": 3501 - }, - { - "epoch": 0.67, - "learning_rate": 5.07936583882158e-06, - "loss": 0.8141, - "step": 3502 - }, - { - "epoch": 0.67, - "learning_rate": 5.073941397228518e-06, - "loss": 0.9055, - "step": 3503 - }, - { - "epoch": 0.67, - "learning_rate": 5.068518868859854e-06, - "loss": 0.8965, - "step": 3504 - }, - { - "epoch": 0.67, - "learning_rate": 5.063098255821637e-06, - "loss": 0.978, - "step": 3505 - }, - { - "epoch": 0.67, - "learning_rate": 5.0576795602191734e-06, - "loss": 0.8867, - "step": 3506 - }, - { - "epoch": 0.67, - "learning_rate": 5.052262784157014e-06, - "loss": 0.8611, - "step": 3507 - }, - { - "epoch": 0.67, - "learning_rate": 5.046847929738971e-06, - "loss": 0.8607, - "step": 3508 - }, - { - "epoch": 0.68, - "learning_rate": 5.041434999068127e-06, - "loss": 0.8358, - "step": 3509 - }, - { - "epoch": 0.68, - "learning_rate": 5.036023994246787e-06, - "loss": 0.8905, - "step": 3510 - }, - { - "epoch": 0.68, - "learning_rate": 5.030614917376532e-06, - "loss": 0.8912, - "step": 3511 - }, - { - "epoch": 0.68, - "learning_rate": 5.025207770558176e-06, - "loss": 0.7294, - "step": 3512 - }, - { - "epoch": 0.68, - "learning_rate": 5.0198025558917985e-06, - "loss": 0.8494, - "step": 3513 - }, - { - "epoch": 0.68, - "learning_rate": 5.014399275476721e-06, - "loss": 0.8458, - "step": 3514 - }, - { - "epoch": 0.68, - "learning_rate": 5.008997931411517e-06, - "loss": 0.9709, - "step": 3515 - }, - { - "epoch": 0.68, - "learning_rate": 5.003598525794002e-06, - "loss": 0.8581, - "step": 3516 - }, - { - "epoch": 0.68, - "learning_rate": 4.998201060721253e-06, - "loss": 0.88, - "step": 3517 - }, - { - "epoch": 0.68, - "learning_rate": 4.992805538289571e-06, - "loss": 0.8727, - "step": 3518 - }, - { - "epoch": 0.68, - "learning_rate": 4.987411960594521e-06, - "loss": 0.8639, - "step": 3519 - }, - { - "epoch": 0.68, - "learning_rate": 4.982020329730904e-06, - "loss": 0.8382, - "step": 3520 - }, - { - "epoch": 0.68, - "learning_rate": 4.976630647792771e-06, - "loss": 0.7721, - "step": 3521 - }, - { - "epoch": 0.68, - "learning_rate": 4.971242916873412e-06, - "loss": 0.6481, - "step": 3522 - }, - { - "epoch": 0.68, - "learning_rate": 4.965857139065354e-06, - "loss": 0.9673, - "step": 3523 - }, - { - "epoch": 0.68, - "learning_rate": 4.9604733164603755e-06, - "loss": 0.919, - "step": 3524 - }, - { - "epoch": 0.68, - "learning_rate": 4.955091451149495e-06, - "loss": 0.8358, - "step": 3525 - }, - { - "epoch": 0.68, - "learning_rate": 4.9497115452229535e-06, - "loss": 0.8282, - "step": 3526 - }, - { - "epoch": 0.68, - "learning_rate": 4.9443336007702614e-06, - "loss": 0.8295, - "step": 3527 - }, - { - "epoch": 0.68, - "learning_rate": 4.938957619880138e-06, - "loss": 0.9119, - "step": 3528 - }, - { - "epoch": 0.68, - "learning_rate": 4.9335836046405575e-06, - "loss": 0.7973, - "step": 3529 - }, - { - "epoch": 0.68, - "learning_rate": 4.928211557138728e-06, - "loss": 0.7927, - "step": 3530 - }, - { - "epoch": 0.68, - "learning_rate": 4.922841479461083e-06, - "loss": 0.8561, - "step": 3531 - }, - { - "epoch": 0.68, - "learning_rate": 4.917473373693305e-06, - "loss": 0.8652, - "step": 3532 - }, - { - "epoch": 0.68, - "learning_rate": 4.9121072419203016e-06, - "loss": 0.889, - "step": 3533 - }, - { - "epoch": 0.68, - "learning_rate": 4.906743086226218e-06, - "loss": 0.7014, - "step": 3534 - }, - { - "epoch": 0.68, - "learning_rate": 4.901380908694434e-06, - "loss": 0.8961, - "step": 3535 - }, - { - "epoch": 0.68, - "learning_rate": 4.8960207114075495e-06, - "loss": 0.7733, - "step": 3536 - }, - { - "epoch": 0.68, - "learning_rate": 4.890662496447407e-06, - "loss": 0.8617, - "step": 3537 - }, - { - "epoch": 0.68, - "learning_rate": 4.8853062658950765e-06, - "loss": 0.9035, - "step": 3538 - }, - { - "epoch": 0.68, - "learning_rate": 4.879952021830856e-06, - "loss": 0.8432, - "step": 3539 - }, - { - "epoch": 0.68, - "learning_rate": 4.874599766334276e-06, - "loss": 0.8627, - "step": 3540 - }, - { - "epoch": 0.68, - "learning_rate": 4.8692495014840825e-06, - "loss": 0.8721, - "step": 3541 - }, - { - "epoch": 0.68, - "learning_rate": 4.863901229358261e-06, - "loss": 0.7655, - "step": 3542 - }, - { - "epoch": 0.68, - "learning_rate": 4.858554952034019e-06, - "loss": 0.6968, - "step": 3543 - }, - { - "epoch": 0.68, - "learning_rate": 4.853210671587789e-06, - "loss": 0.8626, - "step": 3544 - }, - { - "epoch": 0.68, - "learning_rate": 4.847868390095227e-06, - "loss": 0.8739, - "step": 3545 - }, - { - "epoch": 0.68, - "learning_rate": 4.842528109631218e-06, - "loss": 0.837, - "step": 3546 - }, - { - "epoch": 0.68, - "learning_rate": 4.837189832269858e-06, - "loss": 0.7581, - "step": 3547 - }, - { - "epoch": 0.68, - "learning_rate": 4.8318535600844775e-06, - "loss": 0.8131, - "step": 3548 - }, - { - "epoch": 0.68, - "learning_rate": 4.8265192951476206e-06, - "loss": 0.8569, - "step": 3549 - }, - { - "epoch": 0.68, - "learning_rate": 4.8211870395310556e-06, - "loss": 0.8965, - "step": 3550 - }, - { - "epoch": 0.68, - "learning_rate": 4.815856795305772e-06, - "loss": 0.7947, - "step": 3551 - }, - { - "epoch": 0.68, - "learning_rate": 4.81052856454197e-06, - "loss": 0.8873, - "step": 3552 - }, - { - "epoch": 0.68, - "learning_rate": 4.805202349309074e-06, - "loss": 0.8052, - "step": 3553 - }, - { - "epoch": 0.68, - "learning_rate": 4.7998781516757295e-06, - "loss": 0.9178, - "step": 3554 - }, - { - "epoch": 0.68, - "learning_rate": 4.794555973709783e-06, - "loss": 0.8791, - "step": 3555 - }, - { - "epoch": 0.68, - "learning_rate": 4.789235817478322e-06, - "loss": 0.9253, - "step": 3556 - }, - { - "epoch": 0.68, - "learning_rate": 4.783917685047621e-06, - "loss": 0.8567, - "step": 3557 - }, - { - "epoch": 0.68, - "learning_rate": 4.778601578483187e-06, - "loss": 0.8227, - "step": 3558 - }, - { - "epoch": 0.68, - "learning_rate": 4.773287499849737e-06, - "loss": 0.8393, - "step": 3559 - }, - { - "epoch": 0.68, - "learning_rate": 4.767975451211191e-06, - "loss": 0.8972, - "step": 3560 - }, - { - "epoch": 0.69, - "learning_rate": 4.762665434630692e-06, - "loss": 0.8136, - "step": 3561 - }, - { - "epoch": 0.69, - "learning_rate": 4.757357452170588e-06, - "loss": 0.8132, - "step": 3562 - }, - { - "epoch": 0.69, - "learning_rate": 4.752051505892438e-06, - "loss": 0.8348, - "step": 3563 - }, - { - "epoch": 0.69, - "learning_rate": 4.746747597857014e-06, - "loss": 0.8929, - "step": 3564 - }, - { - "epoch": 0.69, - "learning_rate": 4.741445730124287e-06, - "loss": 0.8333, - "step": 3565 - }, - { - "epoch": 0.69, - "learning_rate": 4.736145904753445e-06, - "loss": 0.8801, - "step": 3566 - }, - { - "epoch": 0.69, - "learning_rate": 4.730848123802877e-06, - "loss": 0.8979, - "step": 3567 - }, - { - "epoch": 0.69, - "learning_rate": 4.725552389330183e-06, - "loss": 0.8779, - "step": 3568 - }, - { - "epoch": 0.69, - "learning_rate": 4.720258703392161e-06, - "loss": 0.9208, - "step": 3569 - }, - { - "epoch": 0.69, - "learning_rate": 4.714967068044826e-06, - "loss": 0.9008, - "step": 3570 - }, - { - "epoch": 0.69, - "learning_rate": 4.7096774853433765e-06, - "loss": 0.8749, - "step": 3571 - }, - { - "epoch": 0.69, - "learning_rate": 4.704389957342237e-06, - "loss": 0.7302, - "step": 3572 - }, - { - "epoch": 0.69, - "learning_rate": 4.699104486095008e-06, - "loss": 0.8615, - "step": 3573 - }, - { - "epoch": 0.69, - "learning_rate": 4.69382107365452e-06, - "loss": 0.936, - "step": 3574 - }, - { - "epoch": 0.69, - "learning_rate": 4.6885397220727855e-06, - "loss": 0.8253, - "step": 3575 - }, - { - "epoch": 0.69, - "learning_rate": 4.683260433401016e-06, - "loss": 0.8821, - "step": 3576 - }, - { - "epoch": 0.69, - "learning_rate": 4.677983209689631e-06, - "loss": 0.8546, - "step": 3577 - }, - { - "epoch": 0.69, - "learning_rate": 4.6727080529882394e-06, - "loss": 0.8169, - "step": 3578 - }, - { - "epoch": 0.69, - "learning_rate": 4.667434965345654e-06, - "loss": 0.8339, - "step": 3579 - }, - { - "epoch": 0.69, - "learning_rate": 4.6621639488098856e-06, - "loss": 0.831, - "step": 3580 - }, - { - "epoch": 0.69, - "learning_rate": 4.656895005428127e-06, - "loss": 0.9205, - "step": 3581 - }, - { - "epoch": 0.69, - "learning_rate": 4.651628137246781e-06, - "loss": 0.8168, - "step": 3582 - }, - { - "epoch": 0.69, - "learning_rate": 4.6463633463114395e-06, - "loss": 0.8852, - "step": 3583 - }, - { - "epoch": 0.69, - "learning_rate": 4.641100634666877e-06, - "loss": 0.8504, - "step": 3584 - }, - { - "epoch": 0.69, - "learning_rate": 4.635840004357086e-06, - "loss": 0.714, - "step": 3585 - }, - { - "epoch": 0.69, - "learning_rate": 4.630581457425222e-06, - "loss": 0.7793, - "step": 3586 - }, - { - "epoch": 0.69, - "learning_rate": 4.625324995913648e-06, - "loss": 0.8602, - "step": 3587 - }, - { - "epoch": 0.69, - "learning_rate": 4.620070621863917e-06, - "loss": 0.6495, - "step": 3588 - }, - { - "epoch": 0.69, - "learning_rate": 4.614818337316759e-06, - "loss": 0.8678, - "step": 3589 - }, - { - "epoch": 0.69, - "learning_rate": 4.609568144312107e-06, - "loss": 0.8865, - "step": 3590 - }, - { - "epoch": 0.69, - "learning_rate": 4.6043200448890724e-06, - "loss": 0.9271, - "step": 3591 - }, - { - "epoch": 0.69, - "learning_rate": 4.599074041085958e-06, - "loss": 0.8416, - "step": 3592 - }, - { - "epoch": 0.69, - "learning_rate": 4.593830134940256e-06, - "loss": 0.8398, - "step": 3593 - }, - { - "epoch": 0.69, - "learning_rate": 4.588588328488629e-06, - "loss": 0.8973, - "step": 3594 - }, - { - "epoch": 0.69, - "learning_rate": 4.5833486237669414e-06, - "loss": 0.8463, - "step": 3595 - }, - { - "epoch": 0.69, - "learning_rate": 4.578111022810231e-06, - "loss": 0.7862, - "step": 3596 - }, - { - "epoch": 0.69, - "learning_rate": 4.5728755276527225e-06, - "loss": 0.8946, - "step": 3597 - }, - { - "epoch": 0.69, - "learning_rate": 4.567642140327823e-06, - "loss": 0.8635, - "step": 3598 - }, - { - "epoch": 0.69, - "learning_rate": 4.562410862868123e-06, - "loss": 0.8266, - "step": 3599 - }, - { - "epoch": 0.69, - "learning_rate": 4.557181697305383e-06, - "loss": 0.8645, - "step": 3600 - }, - { - "epoch": 0.69, - "learning_rate": 4.551954645670557e-06, - "loss": 0.7257, - "step": 3601 - }, - { - "epoch": 0.69, - "learning_rate": 4.546729709993762e-06, - "loss": 0.9304, - "step": 3602 - }, - { - "epoch": 0.69, - "learning_rate": 4.541506892304314e-06, - "loss": 0.8804, - "step": 3603 - }, - { - "epoch": 0.69, - "learning_rate": 4.536286194630694e-06, - "loss": 0.9103, - "step": 3604 - }, - { - "epoch": 0.69, - "learning_rate": 4.531067619000553e-06, - "loss": 0.8245, - "step": 3605 - }, - { - "epoch": 0.69, - "learning_rate": 4.525851167440731e-06, - "loss": 0.8414, - "step": 3606 - }, - { - "epoch": 0.69, - "learning_rate": 4.52063684197724e-06, - "loss": 0.8657, - "step": 3607 - }, - { - "epoch": 0.69, - "learning_rate": 4.515424644635254e-06, - "loss": 0.8236, - "step": 3608 - }, - { - "epoch": 0.69, - "learning_rate": 4.510214577439146e-06, - "loss": 0.8685, - "step": 3609 - }, - { - "epoch": 0.69, - "learning_rate": 4.5050066424124324e-06, - "loss": 0.8882, - "step": 3610 - }, - { - "epoch": 0.69, - "learning_rate": 4.49980084157782e-06, - "loss": 0.9005, - "step": 3611 - }, - { - "epoch": 0.69, - "learning_rate": 4.494597176957186e-06, - "loss": 0.892, - "step": 3612 - }, - { - "epoch": 0.7, - "learning_rate": 4.489395650571562e-06, - "loss": 0.8581, - "step": 3613 - }, - { - "epoch": 0.7, - "learning_rate": 4.4841962644411765e-06, - "loss": 0.8812, - "step": 3614 - }, - { - "epoch": 0.7, - "learning_rate": 4.4789990205854e-06, - "loss": 0.8803, - "step": 3615 - }, - { - "epoch": 0.7, - "learning_rate": 4.473803921022784e-06, - "loss": 0.8782, - "step": 3616 - }, - { - "epoch": 0.7, - "learning_rate": 4.468610967771051e-06, - "loss": 0.8956, - "step": 3617 - }, - { - "epoch": 0.7, - "learning_rate": 4.4634201628470766e-06, - "loss": 0.8299, - "step": 3618 - }, - { - "epoch": 0.7, - "learning_rate": 4.458231508266912e-06, - "loss": 0.7576, - "step": 3619 - }, - { - "epoch": 0.7, - "learning_rate": 4.453045006045773e-06, - "loss": 0.8856, - "step": 3620 - }, - { - "epoch": 0.7, - "learning_rate": 4.447860658198035e-06, - "loss": 0.9262, - "step": 3621 - }, - { - "epoch": 0.7, - "learning_rate": 4.442678466737245e-06, - "loss": 0.8394, - "step": 3622 - }, - { - "epoch": 0.7, - "learning_rate": 4.4374984336760975e-06, - "loss": 0.8671, - "step": 3623 - }, - { - "epoch": 0.7, - "learning_rate": 4.432320561026461e-06, - "loss": 0.9103, - "step": 3624 - }, - { - "epoch": 0.7, - "learning_rate": 4.427144850799363e-06, - "loss": 0.7932, - "step": 3625 - }, - { - "epoch": 0.7, - "learning_rate": 4.421971305004989e-06, - "loss": 0.7777, - "step": 3626 - }, - { - "epoch": 0.7, - "learning_rate": 4.416799925652684e-06, - "loss": 0.6478, - "step": 3627 - }, - { - "epoch": 0.7, - "learning_rate": 4.411630714750956e-06, - "loss": 0.9357, - "step": 3628 - }, - { - "epoch": 0.7, - "learning_rate": 4.4064636743074605e-06, - "loss": 0.7234, - "step": 3629 - }, - { - "epoch": 0.7, - "learning_rate": 4.40129880632902e-06, - "loss": 0.8376, - "step": 3630 - }, - { - "epoch": 0.7, - "learning_rate": 4.396136112821608e-06, - "loss": 0.9268, - "step": 3631 - }, - { - "epoch": 0.7, - "learning_rate": 4.390975595790358e-06, - "loss": 0.8615, - "step": 3632 - }, - { - "epoch": 0.7, - "learning_rate": 4.385817257239556e-06, - "loss": 0.7405, - "step": 3633 - }, - { - "epoch": 0.7, - "learning_rate": 4.380661099172636e-06, - "loss": 0.9095, - "step": 3634 - }, - { - "epoch": 0.7, - "learning_rate": 4.375507123592194e-06, - "loss": 0.9163, - "step": 3635 - }, - { - "epoch": 0.7, - "learning_rate": 4.370355332499977e-06, - "loss": 0.8605, - "step": 3636 - }, - { - "epoch": 0.7, - "learning_rate": 4.365205727896872e-06, - "loss": 0.9291, - "step": 3637 - }, - { - "epoch": 0.7, - "learning_rate": 4.36005831178294e-06, - "loss": 0.9075, - "step": 3638 - }, - { - "epoch": 0.7, - "learning_rate": 4.354913086157367e-06, - "loss": 0.7932, - "step": 3639 - }, - { - "epoch": 0.7, - "learning_rate": 4.349770053018502e-06, - "loss": 0.8165, - "step": 3640 - }, - { - "epoch": 0.7, - "learning_rate": 4.344629214363845e-06, - "loss": 0.8156, - "step": 3641 - }, - { - "epoch": 0.7, - "learning_rate": 4.339490572190031e-06, - "loss": 0.8374, - "step": 3642 - }, - { - "epoch": 0.7, - "learning_rate": 4.334354128492851e-06, - "loss": 0.8964, - "step": 3643 - }, - { - "epoch": 0.7, - "learning_rate": 4.329219885267244e-06, - "loss": 0.9151, - "step": 3644 - }, - { - "epoch": 0.7, - "learning_rate": 4.324087844507289e-06, - "loss": 0.8638, - "step": 3645 - }, - { - "epoch": 0.7, - "learning_rate": 4.318958008206214e-06, - "loss": 0.8488, - "step": 3646 - }, - { - "epoch": 0.7, - "learning_rate": 4.313830378356384e-06, - "loss": 0.8589, - "step": 3647 - }, - { - "epoch": 0.7, - "learning_rate": 4.3087049569493136e-06, - "loss": 0.8755, - "step": 3648 - }, - { - "epoch": 0.7, - "learning_rate": 4.303581745975656e-06, - "loss": 0.8153, - "step": 3649 - }, - { - "epoch": 0.7, - "learning_rate": 4.2984607474252084e-06, - "loss": 0.9465, - "step": 3650 - }, - { - "epoch": 0.7, - "learning_rate": 4.293341963286912e-06, - "loss": 0.7658, - "step": 3651 - }, - { - "epoch": 0.7, - "learning_rate": 4.288225395548835e-06, - "loss": 0.8093, - "step": 3652 - }, - { - "epoch": 0.7, - "learning_rate": 4.283111046198198e-06, - "loss": 0.8373, - "step": 3653 - }, - { - "epoch": 0.7, - "learning_rate": 4.277998917221354e-06, - "loss": 0.7246, - "step": 3654 - }, - { - "epoch": 0.7, - "learning_rate": 4.272889010603798e-06, - "loss": 0.8603, - "step": 3655 - }, - { - "epoch": 0.7, - "learning_rate": 4.267781328330155e-06, - "loss": 0.7914, - "step": 3656 - }, - { - "epoch": 0.7, - "learning_rate": 4.262675872384197e-06, - "loss": 0.8784, - "step": 3657 - }, - { - "epoch": 0.7, - "learning_rate": 4.257572644748813e-06, - "loss": 0.8421, - "step": 3658 - }, - { - "epoch": 0.7, - "learning_rate": 4.252471647406045e-06, - "loss": 0.8302, - "step": 3659 - }, - { - "epoch": 0.7, - "learning_rate": 4.2473728823370605e-06, - "loss": 0.7443, - "step": 3660 - }, - { - "epoch": 0.7, - "learning_rate": 4.242276351522161e-06, - "loss": 0.814, - "step": 3661 - }, - { - "epoch": 0.7, - "learning_rate": 4.237182056940784e-06, - "loss": 0.8285, - "step": 3662 - }, - { - "epoch": 0.7, - "learning_rate": 4.232090000571488e-06, - "loss": 0.8669, - "step": 3663 - }, - { - "epoch": 0.7, - "learning_rate": 4.2270001843919714e-06, - "loss": 0.8515, - "step": 3664 - }, - { - "epoch": 0.71, - "learning_rate": 4.221912610379065e-06, - "loss": 0.8345, - "step": 3665 - }, - { - "epoch": 0.71, - "learning_rate": 4.216827280508712e-06, - "loss": 0.7761, - "step": 3666 - }, - { - "epoch": 0.71, - "learning_rate": 4.211744196756011e-06, - "loss": 0.9473, - "step": 3667 - }, - { - "epoch": 0.71, - "learning_rate": 4.206663361095164e-06, - "loss": 0.8997, - "step": 3668 - }, - { - "epoch": 0.71, - "learning_rate": 4.201584775499509e-06, - "loss": 0.8612, - "step": 3669 - }, - { - "epoch": 0.71, - "learning_rate": 4.196508441941516e-06, - "loss": 0.7902, - "step": 3670 - }, - { - "epoch": 0.71, - "learning_rate": 4.191434362392768e-06, - "loss": 0.8194, - "step": 3671 - }, - { - "epoch": 0.71, - "learning_rate": 4.186362538823981e-06, - "loss": 0.7513, - "step": 3672 - }, - { - "epoch": 0.71, - "learning_rate": 4.181292973204992e-06, - "loss": 0.7996, - "step": 3673 - }, - { - "epoch": 0.71, - "learning_rate": 4.1762256675047655e-06, - "loss": 0.7178, - "step": 3674 - }, - { - "epoch": 0.71, - "learning_rate": 4.171160623691384e-06, - "loss": 0.8339, - "step": 3675 - }, - { - "epoch": 0.71, - "learning_rate": 4.166097843732048e-06, - "loss": 0.8205, - "step": 3676 - }, - { - "epoch": 0.71, - "learning_rate": 4.161037329593085e-06, - "loss": 0.7984, - "step": 3677 - }, - { - "epoch": 0.71, - "learning_rate": 4.155979083239942e-06, - "loss": 0.8895, - "step": 3678 - }, - { - "epoch": 0.71, - "learning_rate": 4.1509231066371815e-06, - "loss": 0.861, - "step": 3679 - }, - { - "epoch": 0.71, - "learning_rate": 4.1458694017484915e-06, - "loss": 0.8174, - "step": 3680 - }, - { - "epoch": 0.71, - "learning_rate": 4.140817970536664e-06, - "loss": 0.8245, - "step": 3681 - }, - { - "epoch": 0.71, - "learning_rate": 4.135768814963622e-06, - "loss": 0.7861, - "step": 3682 - }, - { - "epoch": 0.71, - "learning_rate": 4.130721936990399e-06, - "loss": 0.9048, - "step": 3683 - }, - { - "epoch": 0.71, - "learning_rate": 4.1256773385771444e-06, - "loss": 0.8666, - "step": 3684 - }, - { - "epoch": 0.71, - "learning_rate": 4.120635021683122e-06, - "loss": 0.8341, - "step": 3685 - }, - { - "epoch": 0.71, - "learning_rate": 4.115594988266711e-06, - "loss": 0.8218, - "step": 3686 - }, - { - "epoch": 0.71, - "learning_rate": 4.1105572402853976e-06, - "loss": 0.8778, - "step": 3687 - }, - { - "epoch": 0.71, - "learning_rate": 4.1055217796957895e-06, - "loss": 0.7652, - "step": 3688 - }, - { - "epoch": 0.71, - "learning_rate": 4.100488608453599e-06, - "loss": 0.7891, - "step": 3689 - }, - { - "epoch": 0.71, - "learning_rate": 4.095457728513652e-06, - "loss": 0.8513, - "step": 3690 - }, - { - "epoch": 0.71, - "learning_rate": 4.09042914182989e-06, - "loss": 0.9277, - "step": 3691 - }, - { - "epoch": 0.71, - "learning_rate": 4.08540285035535e-06, - "loss": 0.9249, - "step": 3692 - }, - { - "epoch": 0.71, - "learning_rate": 4.0803788560421885e-06, - "loss": 0.9168, - "step": 3693 - }, - { - "epoch": 0.71, - "learning_rate": 4.075357160841671e-06, - "loss": 0.7441, - "step": 3694 - }, - { - "epoch": 0.71, - "learning_rate": 4.070337766704155e-06, - "loss": 0.8692, - "step": 3695 - }, - { - "epoch": 0.71, - "learning_rate": 4.065320675579132e-06, - "loss": 0.9392, - "step": 3696 - }, - { - "epoch": 0.71, - "learning_rate": 4.0603058894151685e-06, - "loss": 0.7648, - "step": 3697 - }, - { - "epoch": 0.71, - "learning_rate": 4.055293410159954e-06, - "loss": 0.8702, - "step": 3698 - }, - { - "epoch": 0.71, - "learning_rate": 4.050283239760282e-06, - "loss": 0.8854, - "step": 3699 - }, - { - "epoch": 0.71, - "learning_rate": 4.045275380162038e-06, - "loss": 0.8752, - "step": 3700 - }, - { - "epoch": 0.71, - "learning_rate": 4.04026983331022e-06, - "loss": 0.8931, - "step": 3701 - }, - { - "epoch": 0.71, - "learning_rate": 4.035266601148924e-06, - "loss": 0.8852, - "step": 3702 - }, - { - "epoch": 0.71, - "learning_rate": 4.03026568562135e-06, - "loss": 0.8668, - "step": 3703 - }, - { - "epoch": 0.71, - "learning_rate": 4.025267088669797e-06, - "loss": 0.9099, - "step": 3704 - }, - { - "epoch": 0.71, - "learning_rate": 4.020270812235656e-06, - "loss": 0.8606, - "step": 3705 - }, - { - "epoch": 0.71, - "learning_rate": 4.015276858259427e-06, - "loss": 0.8944, - "step": 3706 - }, - { - "epoch": 0.71, - "learning_rate": 4.010285228680705e-06, - "loss": 0.8584, - "step": 3707 - }, - { - "epoch": 0.71, - "learning_rate": 4.005295925438181e-06, - "loss": 0.8555, - "step": 3708 - }, - { - "epoch": 0.71, - "learning_rate": 4.000308950469646e-06, - "loss": 0.8557, - "step": 3709 - }, - { - "epoch": 0.71, - "learning_rate": 3.995324305711976e-06, - "loss": 0.8529, - "step": 3710 - }, - { - "epoch": 0.71, - "learning_rate": 3.990341993101154e-06, - "loss": 0.914, - "step": 3711 - }, - { - "epoch": 0.71, - "learning_rate": 3.985362014572256e-06, - "loss": 0.8552, - "step": 3712 - }, - { - "epoch": 0.71, - "learning_rate": 3.9803843720594385e-06, - "loss": 0.8662, - "step": 3713 - }, - { - "epoch": 0.71, - "learning_rate": 3.97540906749597e-06, - "loss": 0.8625, - "step": 3714 - }, - { - "epoch": 0.71, - "learning_rate": 3.970436102814203e-06, - "loss": 0.8995, - "step": 3715 - }, - { - "epoch": 0.71, - "learning_rate": 3.965465479945569e-06, - "loss": 0.9364, - "step": 3716 - }, - { - "epoch": 0.72, - "learning_rate": 3.9604972008206085e-06, - "loss": 0.8372, - "step": 3717 - }, - { - "epoch": 0.72, - "learning_rate": 3.955531267368942e-06, - "loss": 0.8366, - "step": 3718 - }, - { - "epoch": 0.72, - "learning_rate": 3.950567681519279e-06, - "loss": 0.9605, - "step": 3719 - }, - { - "epoch": 0.72, - "learning_rate": 3.945606445199427e-06, - "loss": 0.8606, - "step": 3720 - }, - { - "epoch": 0.72, - "learning_rate": 3.940647560336262e-06, - "loss": 0.8809, - "step": 3721 - }, - { - "epoch": 0.72, - "learning_rate": 3.935691028855763e-06, - "loss": 0.8945, - "step": 3722 - }, - { - "epoch": 0.72, - "learning_rate": 3.930736852682993e-06, - "loss": 0.901, - "step": 3723 - }, - { - "epoch": 0.72, - "learning_rate": 3.9257850337420856e-06, - "loss": 0.9024, - "step": 3724 - }, - { - "epoch": 0.72, - "learning_rate": 3.920835573956285e-06, - "loss": 0.875, - "step": 3725 - }, - { - "epoch": 0.72, - "learning_rate": 3.915888475247894e-06, - "loss": 0.8281, - "step": 3726 - }, - { - "epoch": 0.72, - "learning_rate": 3.910943739538313e-06, - "loss": 0.8263, - "step": 3727 - }, - { - "epoch": 0.72, - "learning_rate": 3.906001368748023e-06, - "loss": 0.8286, - "step": 3728 - }, - { - "epoch": 0.72, - "learning_rate": 3.901061364796574e-06, - "loss": 0.8658, - "step": 3729 - }, - { - "epoch": 0.72, - "learning_rate": 3.8961237296026155e-06, - "loss": 0.7959, - "step": 3730 - }, - { - "epoch": 0.72, - "learning_rate": 3.891188465083865e-06, - "loss": 0.7009, - "step": 3731 - }, - { - "epoch": 0.72, - "learning_rate": 3.886255573157121e-06, - "loss": 0.8924, - "step": 3732 - }, - { - "epoch": 0.72, - "learning_rate": 3.88132505573827e-06, - "loss": 0.8569, - "step": 3733 - }, - { - "epoch": 0.72, - "learning_rate": 3.876396914742258e-06, - "loss": 0.94, - "step": 3734 - }, - { - "epoch": 0.72, - "learning_rate": 3.871471152083121e-06, - "loss": 0.8946, - "step": 3735 - }, - { - "epoch": 0.72, - "learning_rate": 3.866547769673968e-06, - "loss": 0.8781, - "step": 3736 - }, - { - "epoch": 0.72, - "learning_rate": 3.861626769426988e-06, - "loss": 0.7981, - "step": 3737 - }, - { - "epoch": 0.72, - "learning_rate": 3.8567081532534374e-06, - "loss": 0.8162, - "step": 3738 - }, - { - "epoch": 0.72, - "learning_rate": 3.851791923063655e-06, - "loss": 0.8312, - "step": 3739 - }, - { - "epoch": 0.72, - "learning_rate": 3.846878080767039e-06, - "loss": 0.9281, - "step": 3740 - }, - { - "epoch": 0.72, - "learning_rate": 3.841966628272079e-06, - "loss": 0.8261, - "step": 3741 - }, - { - "epoch": 0.72, - "learning_rate": 3.837057567486314e-06, - "loss": 0.8975, - "step": 3742 - }, - { - "epoch": 0.72, - "learning_rate": 3.832150900316377e-06, - "loss": 0.9282, - "step": 3743 - }, - { - "epoch": 0.72, - "learning_rate": 3.827246628667962e-06, - "loss": 0.892, - "step": 3744 - }, - { - "epoch": 0.72, - "learning_rate": 3.822344754445826e-06, - "loss": 0.811, - "step": 3745 - }, - { - "epoch": 0.72, - "learning_rate": 3.817445279553801e-06, - "loss": 0.7816, - "step": 3746 - }, - { - "epoch": 0.72, - "learning_rate": 3.8125482058947905e-06, - "loss": 0.863, - "step": 3747 - }, - { - "epoch": 0.72, - "learning_rate": 3.8076535353707523e-06, - "loss": 0.8537, - "step": 3748 - }, - { - "epoch": 0.72, - "learning_rate": 3.8027612698827344e-06, - "loss": 0.8453, - "step": 3749 - }, - { - "epoch": 0.72, - "learning_rate": 3.7978714113308246e-06, - "loss": 0.8956, - "step": 3750 - }, - { - "epoch": 0.72, - "learning_rate": 3.7929839616141917e-06, - "loss": 0.8654, - "step": 3751 - }, - { - "epoch": 0.72, - "learning_rate": 3.788098922631067e-06, - "loss": 0.8992, - "step": 3752 - }, - { - "epoch": 0.72, - "learning_rate": 3.7832162962787355e-06, - "loss": 0.907, - "step": 3753 - }, - { - "epoch": 0.72, - "learning_rate": 3.7783360844535653e-06, - "loss": 0.8046, - "step": 3754 - }, - { - "epoch": 0.72, - "learning_rate": 3.773458289050963e-06, - "loss": 0.9165, - "step": 3755 - }, - { - "epoch": 0.72, - "learning_rate": 3.768582911965414e-06, - "loss": 0.9276, - "step": 3756 - }, - { - "epoch": 0.72, - "learning_rate": 3.763709955090461e-06, - "loss": 0.8977, - "step": 3757 - }, - { - "epoch": 0.72, - "learning_rate": 3.7588394203186963e-06, - "loss": 0.9198, - "step": 3758 - }, - { - "epoch": 0.72, - "learning_rate": 3.753971309541784e-06, - "loss": 0.8641, - "step": 3759 - }, - { - "epoch": 0.72, - "learning_rate": 3.7491056246504433e-06, - "loss": 0.7088, - "step": 3760 - }, - { - "epoch": 0.72, - "learning_rate": 3.7442423675344474e-06, - "loss": 0.8288, - "step": 3761 - }, - { - "epoch": 0.72, - "learning_rate": 3.739381540082635e-06, - "loss": 0.8794, - "step": 3762 - }, - { - "epoch": 0.72, - "learning_rate": 3.7345231441828876e-06, - "loss": 0.8779, - "step": 3763 - }, - { - "epoch": 0.72, - "learning_rate": 3.729667181722154e-06, - "loss": 0.8716, - "step": 3764 - }, - { - "epoch": 0.72, - "learning_rate": 3.7248136545864345e-06, - "loss": 0.8716, - "step": 3765 - }, - { - "epoch": 0.72, - "learning_rate": 3.719962564660783e-06, - "loss": 0.8845, - "step": 3766 - }, - { - "epoch": 0.72, - "learning_rate": 3.7151139138293056e-06, - "loss": 0.9308, - "step": 3767 - }, - { - "epoch": 0.72, - "learning_rate": 3.7102677039751667e-06, - "loss": 0.8967, - "step": 3768 - }, - { - "epoch": 0.73, - "learning_rate": 3.705423936980572e-06, - "loss": 0.7104, - "step": 3769 - }, - { - "epoch": 0.73, - "learning_rate": 3.700582614726791e-06, - "loss": 0.8845, - "step": 3770 - }, - { - "epoch": 0.73, - "learning_rate": 3.6957437390941274e-06, - "loss": 0.8183, - "step": 3771 - }, - { - "epoch": 0.73, - "learning_rate": 3.6909073119619555e-06, - "loss": 0.8186, - "step": 3772 - }, - { - "epoch": 0.73, - "learning_rate": 3.6860733352086866e-06, - "loss": 0.8574, - "step": 3773 - }, - { - "epoch": 0.73, - "learning_rate": 3.6812418107117765e-06, - "loss": 0.8516, - "step": 3774 - }, - { - "epoch": 0.73, - "learning_rate": 3.6764127403477347e-06, - "loss": 0.79, - "step": 3775 - }, - { - "epoch": 0.73, - "learning_rate": 3.6715861259921226e-06, - "loss": 0.8421, - "step": 3776 - }, - { - "epoch": 0.73, - "learning_rate": 3.6667619695195287e-06, - "loss": 0.9003, - "step": 3777 - }, - { - "epoch": 0.73, - "learning_rate": 3.6619402728036157e-06, - "loss": 0.9286, - "step": 3778 - }, - { - "epoch": 0.73, - "learning_rate": 3.657121037717064e-06, - "loss": 0.8663, - "step": 3779 - }, - { - "epoch": 0.73, - "learning_rate": 3.652304266131612e-06, - "loss": 0.8927, - "step": 3780 - }, - { - "epoch": 0.73, - "learning_rate": 3.6474899599180426e-06, - "loss": 0.8098, - "step": 3781 - }, - { - "epoch": 0.73, - "learning_rate": 3.642678120946168e-06, - "loss": 0.9281, - "step": 3782 - }, - { - "epoch": 0.73, - "learning_rate": 3.6378687510848576e-06, - "loss": 0.8153, - "step": 3783 - }, - { - "epoch": 0.73, - "learning_rate": 3.6330618522020124e-06, - "loss": 0.887, - "step": 3784 - }, - { - "epoch": 0.73, - "learning_rate": 3.6282574261645776e-06, - "loss": 0.8896, - "step": 3785 - }, - { - "epoch": 0.73, - "learning_rate": 3.62345547483854e-06, - "loss": 0.8826, - "step": 3786 - }, - { - "epoch": 0.73, - "learning_rate": 3.618656000088916e-06, - "loss": 0.9064, - "step": 3787 - }, - { - "epoch": 0.73, - "learning_rate": 3.6138590037797695e-06, - "loss": 0.8582, - "step": 3788 - }, - { - "epoch": 0.73, - "learning_rate": 3.6090644877741986e-06, - "loss": 0.8521, - "step": 3789 - }, - { - "epoch": 0.73, - "learning_rate": 3.6042724539343378e-06, - "loss": 0.9493, - "step": 3790 - }, - { - "epoch": 0.73, - "learning_rate": 3.599482904121361e-06, - "loss": 0.8485, - "step": 3791 - }, - { - "epoch": 0.73, - "learning_rate": 3.594695840195468e-06, - "loss": 0.8547, - "step": 3792 - }, - { - "epoch": 0.73, - "learning_rate": 3.5899112640159017e-06, - "loss": 0.7838, - "step": 3793 - }, - { - "epoch": 0.73, - "learning_rate": 3.585129177440938e-06, - "loss": 0.9209, - "step": 3794 - }, - { - "epoch": 0.73, - "learning_rate": 3.580349582327882e-06, - "loss": 0.9052, - "step": 3795 - }, - { - "epoch": 0.73, - "learning_rate": 3.575572480533076e-06, - "loss": 0.8281, - "step": 3796 - }, - { - "epoch": 0.73, - "learning_rate": 3.570797873911892e-06, - "loss": 0.8719, - "step": 3797 - }, - { - "epoch": 0.73, - "learning_rate": 3.566025764318728e-06, - "loss": 0.8041, - "step": 3798 - }, - { - "epoch": 0.73, - "learning_rate": 3.5612561536070213e-06, - "loss": 0.9006, - "step": 3799 - }, - { - "epoch": 0.73, - "learning_rate": 3.5564890436292243e-06, - "loss": 0.8799, - "step": 3800 - }, - { - "epoch": 0.73, - "learning_rate": 3.5517244362368363e-06, - "loss": 0.8869, - "step": 3801 - }, - { - "epoch": 0.73, - "learning_rate": 3.5469623332803795e-06, - "loss": 0.9376, - "step": 3802 - }, - { - "epoch": 0.73, - "learning_rate": 3.5422027366093893e-06, - "loss": 0.8774, - "step": 3803 - }, - { - "epoch": 0.73, - "learning_rate": 3.5374456480724427e-06, - "loss": 0.8908, - "step": 3804 - }, - { - "epoch": 0.73, - "learning_rate": 3.532691069517142e-06, - "loss": 0.7935, - "step": 3805 - }, - { - "epoch": 0.73, - "learning_rate": 3.5279390027901004e-06, - "loss": 0.7925, - "step": 3806 - }, - { - "epoch": 0.73, - "learning_rate": 3.5231894497369802e-06, - "loss": 0.8453, - "step": 3807 - }, - { - "epoch": 0.73, - "learning_rate": 3.5184424122024406e-06, - "loss": 0.8561, - "step": 3808 - }, - { - "epoch": 0.73, - "learning_rate": 3.5136978920301822e-06, - "loss": 0.686, - "step": 3809 - }, - { - "epoch": 0.73, - "learning_rate": 3.508955891062924e-06, - "loss": 0.8386, - "step": 3810 - }, - { - "epoch": 0.73, - "learning_rate": 3.5042164111423983e-06, - "loss": 0.8854, - "step": 3811 - }, - { - "epoch": 0.73, - "learning_rate": 3.4994794541093667e-06, - "loss": 0.8991, - "step": 3812 - }, - { - "epoch": 0.73, - "learning_rate": 3.4947450218036106e-06, - "loss": 0.8877, - "step": 3813 - }, - { - "epoch": 0.73, - "learning_rate": 3.4900131160639283e-06, - "loss": 0.7601, - "step": 3814 - }, - { - "epoch": 0.73, - "learning_rate": 3.485283738728139e-06, - "loss": 0.7834, - "step": 3815 - }, - { - "epoch": 0.73, - "learning_rate": 3.4805568916330747e-06, - "loss": 0.9084, - "step": 3816 - }, - { - "epoch": 0.73, - "learning_rate": 3.4758325766145896e-06, - "loss": 0.8222, - "step": 3817 - }, - { - "epoch": 0.73, - "learning_rate": 3.471110795507554e-06, - "loss": 0.8361, - "step": 3818 - }, - { - "epoch": 0.73, - "learning_rate": 3.4663915501458523e-06, - "loss": 0.9073, - "step": 3819 - }, - { - "epoch": 0.73, - "learning_rate": 3.4616748423623893e-06, - "loss": 0.7633, - "step": 3820 - }, - { - "epoch": 0.74, - "learning_rate": 3.4569606739890737e-06, - "loss": 0.7924, - "step": 3821 - }, - { - "epoch": 0.74, - "learning_rate": 3.452249046856836e-06, - "loss": 0.7628, - "step": 3822 - }, - { - "epoch": 0.74, - "learning_rate": 3.4475399627956197e-06, - "loss": 0.7982, - "step": 3823 - }, - { - "epoch": 0.74, - "learning_rate": 3.4428334236343774e-06, - "loss": 0.7976, - "step": 3824 - }, - { - "epoch": 0.74, - "learning_rate": 3.438129431201075e-06, - "loss": 0.8122, - "step": 3825 - }, - { - "epoch": 0.74, - "learning_rate": 3.433427987322693e-06, - "loss": 0.9234, - "step": 3826 - }, - { - "epoch": 0.74, - "learning_rate": 3.4287290938252103e-06, - "loss": 0.8758, - "step": 3827 - }, - { - "epoch": 0.74, - "learning_rate": 3.424032752533627e-06, - "loss": 0.8874, - "step": 3828 - }, - { - "epoch": 0.74, - "learning_rate": 3.4193389652719478e-06, - "loss": 0.84, - "step": 3829 - }, - { - "epoch": 0.74, - "learning_rate": 3.4146477338631856e-06, - "loss": 0.8143, - "step": 3830 - }, - { - "epoch": 0.74, - "learning_rate": 3.4099590601293632e-06, - "loss": 0.8887, - "step": 3831 - }, - { - "epoch": 0.74, - "learning_rate": 3.4052729458915024e-06, - "loss": 0.9063, - "step": 3832 - }, - { - "epoch": 0.74, - "learning_rate": 3.4005893929696377e-06, - "loss": 0.8953, - "step": 3833 - }, - { - "epoch": 0.74, - "learning_rate": 3.3959084031828114e-06, - "loss": 0.8653, - "step": 3834 - }, - { - "epoch": 0.74, - "learning_rate": 3.3912299783490567e-06, - "loss": 0.8548, - "step": 3835 - }, - { - "epoch": 0.74, - "learning_rate": 3.3865541202854314e-06, - "loss": 0.9265, - "step": 3836 - }, - { - "epoch": 0.74, - "learning_rate": 3.3818808308079753e-06, - "loss": 0.8749, - "step": 3837 - }, - { - "epoch": 0.74, - "learning_rate": 3.3772101117317437e-06, - "loss": 0.8278, - "step": 3838 - }, - { - "epoch": 0.74, - "learning_rate": 3.372541964870795e-06, - "loss": 0.8521, - "step": 3839 - }, - { - "epoch": 0.74, - "learning_rate": 3.367876392038174e-06, - "loss": 0.7844, - "step": 3840 - }, - { - "epoch": 0.74, - "learning_rate": 3.363213395045941e-06, - "loss": 0.788, - "step": 3841 - }, - { - "epoch": 0.74, - "learning_rate": 3.3585529757051504e-06, - "loss": 0.8594, - "step": 3842 - }, - { - "epoch": 0.74, - "learning_rate": 3.353895135825854e-06, - "loss": 0.7541, - "step": 3843 - }, - { - "epoch": 0.74, - "learning_rate": 3.3492398772171074e-06, - "loss": 0.8077, - "step": 3844 - }, - { - "epoch": 0.74, - "learning_rate": 3.344587201686952e-06, - "loss": 0.9221, - "step": 3845 - }, - { - "epoch": 0.74, - "learning_rate": 3.3399371110424372e-06, - "loss": 0.8628, - "step": 3846 - }, - { - "epoch": 0.74, - "learning_rate": 3.3352896070896057e-06, - "loss": 0.8382, - "step": 3847 - }, - { - "epoch": 0.74, - "learning_rate": 3.330644691633492e-06, - "loss": 0.8793, - "step": 3848 - }, - { - "epoch": 0.74, - "learning_rate": 3.3260023664781326e-06, - "loss": 0.858, - "step": 3849 - }, - { - "epoch": 0.74, - "learning_rate": 3.321362633426547e-06, - "loss": 0.8607, - "step": 3850 - }, - { - "epoch": 0.74, - "learning_rate": 3.316725494280757e-06, - "loss": 0.8336, - "step": 3851 - }, - { - "epoch": 0.74, - "learning_rate": 3.3120909508417754e-06, - "loss": 0.8993, - "step": 3852 - }, - { - "epoch": 0.74, - "learning_rate": 3.307459004909599e-06, - "loss": 0.8862, - "step": 3853 - }, - { - "epoch": 0.74, - "learning_rate": 3.3028296582832285e-06, - "loss": 0.9149, - "step": 3854 - }, - { - "epoch": 0.74, - "learning_rate": 3.2982029127606517e-06, - "loss": 0.8217, - "step": 3855 - }, - { - "epoch": 0.74, - "learning_rate": 3.2935787701388346e-06, - "loss": 0.892, - "step": 3856 - }, - { - "epoch": 0.74, - "learning_rate": 3.2889572322137454e-06, - "loss": 0.8966, - "step": 3857 - }, - { - "epoch": 0.74, - "learning_rate": 3.2843383007803364e-06, - "loss": 0.8455, - "step": 3858 - }, - { - "epoch": 0.74, - "learning_rate": 3.279721977632546e-06, - "loss": 0.8795, - "step": 3859 - }, - { - "epoch": 0.74, - "learning_rate": 3.275108264563306e-06, - "loss": 0.8757, - "step": 3860 - }, - { - "epoch": 0.74, - "learning_rate": 3.270497163364521e-06, - "loss": 0.871, - "step": 3861 - }, - { - "epoch": 0.74, - "learning_rate": 3.2658886758270947e-06, - "loss": 0.8626, - "step": 3862 - }, - { - "epoch": 0.74, - "learning_rate": 3.2612828037409116e-06, - "loss": 0.8252, - "step": 3863 - }, - { - "epoch": 0.74, - "learning_rate": 3.256679548894831e-06, - "loss": 0.8117, - "step": 3864 - }, - { - "epoch": 0.74, - "learning_rate": 3.252078913076718e-06, - "loss": 0.852, - "step": 3865 - }, - { - "epoch": 0.74, - "learning_rate": 3.247480898073395e-06, - "loss": 0.8236, - "step": 3866 - }, - { - "epoch": 0.74, - "learning_rate": 3.242885505670681e-06, - "loss": 0.8705, - "step": 3867 - }, - { - "epoch": 0.74, - "learning_rate": 3.238292737653379e-06, - "loss": 0.8145, - "step": 3868 - }, - { - "epoch": 0.74, - "learning_rate": 3.233702595805258e-06, - "loss": 0.9319, - "step": 3869 - }, - { - "epoch": 0.74, - "learning_rate": 3.229115081909082e-06, - "loss": 0.845, - "step": 3870 - }, - { - "epoch": 0.74, - "learning_rate": 3.224530197746587e-06, - "loss": 0.7764, - "step": 3871 - }, - { - "epoch": 0.74, - "learning_rate": 3.2199479450984892e-06, - "loss": 0.8915, - "step": 3872 - }, - { - "epoch": 0.75, - "learning_rate": 3.2153683257444856e-06, - "loss": 0.8223, - "step": 3873 - }, - { - "epoch": 0.75, - "learning_rate": 3.210791341463243e-06, - "loss": 0.8808, - "step": 3874 - }, - { - "epoch": 0.75, - "learning_rate": 3.206216994032411e-06, - "loss": 0.6898, - "step": 3875 - }, - { - "epoch": 0.75, - "learning_rate": 3.2016452852286127e-06, - "loss": 0.8305, - "step": 3876 - }, - { - "epoch": 0.75, - "learning_rate": 3.1970762168274495e-06, - "loss": 0.8371, - "step": 3877 - }, - { - "epoch": 0.75, - "learning_rate": 3.1925097906034962e-06, - "loss": 0.7076, - "step": 3878 - }, - { - "epoch": 0.75, - "learning_rate": 3.187946008330295e-06, - "loss": 0.8787, - "step": 3879 - }, - { - "epoch": 0.75, - "learning_rate": 3.1833848717803674e-06, - "loss": 0.7182, - "step": 3880 - }, - { - "epoch": 0.75, - "learning_rate": 3.178826382725212e-06, - "loss": 0.8785, - "step": 3881 - }, - { - "epoch": 0.75, - "learning_rate": 3.1742705429352827e-06, - "loss": 0.9387, - "step": 3882 - }, - { - "epoch": 0.75, - "learning_rate": 3.1697173541800254e-06, - "loss": 0.8696, - "step": 3883 - }, - { - "epoch": 0.75, - "learning_rate": 3.165166818227845e-06, - "loss": 0.8624, - "step": 3884 - }, - { - "epoch": 0.75, - "learning_rate": 3.1606189368461117e-06, - "loss": 0.8628, - "step": 3885 - }, - { - "epoch": 0.75, - "learning_rate": 3.156073711801172e-06, - "loss": 0.8315, - "step": 3886 - }, - { - "epoch": 0.75, - "learning_rate": 3.151531144858344e-06, - "loss": 0.8489, - "step": 3887 - }, - { - "epoch": 0.75, - "learning_rate": 3.146991237781899e-06, - "loss": 0.7355, - "step": 3888 - }, - { - "epoch": 0.75, - "learning_rate": 3.142453992335096e-06, - "loss": 0.8492, - "step": 3889 - }, - { - "epoch": 0.75, - "learning_rate": 3.137919410280139e-06, - "loss": 0.8609, - "step": 3890 - }, - { - "epoch": 0.75, - "learning_rate": 3.1333874933782114e-06, - "loss": 0.7311, - "step": 3891 - }, - { - "epoch": 0.75, - "learning_rate": 3.128858243389461e-06, - "loss": 0.9022, - "step": 3892 - }, - { - "epoch": 0.75, - "learning_rate": 3.124331662072987e-06, - "loss": 0.8604, - "step": 3893 - }, - { - "epoch": 0.75, - "learning_rate": 3.119807751186872e-06, - "loss": 0.8889, - "step": 3894 - }, - { - "epoch": 0.75, - "learning_rate": 3.1152865124881436e-06, - "loss": 0.8368, - "step": 3895 - }, - { - "epoch": 0.75, - "learning_rate": 3.110767947732801e-06, - "loss": 0.8187, - "step": 3896 - }, - { - "epoch": 0.75, - "learning_rate": 3.106252058675806e-06, - "loss": 0.927, - "step": 3897 - }, - { - "epoch": 0.75, - "learning_rate": 3.101738847071072e-06, - "loss": 0.8302, - "step": 3898 - }, - { - "epoch": 0.75, - "learning_rate": 3.097228314671481e-06, - "loss": 0.88, - "step": 3899 - }, - { - "epoch": 0.75, - "learning_rate": 3.092720463228872e-06, - "loss": 0.8209, - "step": 3900 - }, - { - "epoch": 0.75, - "learning_rate": 3.0882152944940423e-06, - "loss": 0.8761, - "step": 3901 - }, - { - "epoch": 0.75, - "learning_rate": 3.0837128102167514e-06, - "loss": 0.8255, - "step": 3902 - }, - { - "epoch": 0.75, - "learning_rate": 3.079213012145705e-06, - "loss": 0.7905, - "step": 3903 - }, - { - "epoch": 0.75, - "learning_rate": 3.0747159020285766e-06, - "loss": 0.8931, - "step": 3904 - }, - { - "epoch": 0.75, - "learning_rate": 3.0702214816119925e-06, - "loss": 0.8196, - "step": 3905 - }, - { - "epoch": 0.75, - "learning_rate": 3.065729752641532e-06, - "loss": 0.8328, - "step": 3906 - }, - { - "epoch": 0.75, - "learning_rate": 3.0612407168617352e-06, - "loss": 0.7728, - "step": 3907 - }, - { - "epoch": 0.75, - "learning_rate": 3.0567543760160866e-06, - "loss": 0.8658, - "step": 3908 - }, - { - "epoch": 0.75, - "learning_rate": 3.05227073184703e-06, - "loss": 0.8754, - "step": 3909 - }, - { - "epoch": 0.75, - "learning_rate": 3.047789786095967e-06, - "loss": 0.8947, - "step": 3910 - }, - { - "epoch": 0.75, - "learning_rate": 3.0433115405032334e-06, - "loss": 0.8712, - "step": 3911 - }, - { - "epoch": 0.75, - "learning_rate": 3.0388359968081395e-06, - "loss": 0.8388, - "step": 3912 - }, - { - "epoch": 0.75, - "learning_rate": 3.034363156748933e-06, - "loss": 0.6832, - "step": 3913 - }, - { - "epoch": 0.75, - "learning_rate": 3.0298930220628086e-06, - "loss": 0.7932, - "step": 3914 - }, - { - "epoch": 0.75, - "learning_rate": 3.025425594485919e-06, - "loss": 0.9582, - "step": 3915 - }, - { - "epoch": 0.75, - "learning_rate": 3.0209608757533626e-06, - "loss": 0.8462, - "step": 3916 - }, - { - "epoch": 0.75, - "learning_rate": 3.0164988675991768e-06, - "loss": 0.8306, - "step": 3917 - }, - { - "epoch": 0.75, - "learning_rate": 3.0120395717563655e-06, - "loss": 0.8443, - "step": 3918 - }, - { - "epoch": 0.75, - "learning_rate": 3.00758298995686e-06, - "loss": 0.8538, - "step": 3919 - }, - { - "epoch": 0.75, - "learning_rate": 3.0031291239315473e-06, - "loss": 0.868, - "step": 3920 - }, - { - "epoch": 0.75, - "learning_rate": 2.9986779754102613e-06, - "loss": 0.8728, - "step": 3921 - }, - { - "epoch": 0.75, - "learning_rate": 2.9942295461217698e-06, - "loss": 0.9306, - "step": 3922 - }, - { - "epoch": 0.75, - "learning_rate": 2.9897838377937947e-06, - "loss": 0.9021, - "step": 3923 - }, - { - "epoch": 0.75, - "learning_rate": 2.985340852152999e-06, - "loss": 0.8431, - "step": 3924 - }, - { - "epoch": 0.76, - "learning_rate": 2.9809005909249866e-06, - "loss": 0.8694, - "step": 3925 - }, - { - "epoch": 0.76, - "learning_rate": 2.9764630558343064e-06, - "loss": 0.9412, - "step": 3926 - }, - { - "epoch": 0.76, - "learning_rate": 2.9720282486044407e-06, - "loss": 0.7591, - "step": 3927 - }, - { - "epoch": 0.76, - "learning_rate": 2.9675961709578194e-06, - "loss": 0.8238, - "step": 3928 - }, - { - "epoch": 0.76, - "learning_rate": 2.9631668246158105e-06, - "loss": 0.8402, - "step": 3929 - }, - { - "epoch": 0.76, - "learning_rate": 2.958740211298722e-06, - "loss": 0.8919, - "step": 3930 - }, - { - "epoch": 0.76, - "learning_rate": 2.9543163327258016e-06, - "loss": 0.8148, - "step": 3931 - }, - { - "epoch": 0.76, - "learning_rate": 2.949895190615227e-06, - "loss": 0.865, - "step": 3932 - }, - { - "epoch": 0.76, - "learning_rate": 2.9454767866841225e-06, - "loss": 0.8464, - "step": 3933 - }, - { - "epoch": 0.76, - "learning_rate": 2.941061122648545e-06, - "loss": 0.8784, - "step": 3934 - }, - { - "epoch": 0.76, - "learning_rate": 2.9366482002234874e-06, - "loss": 0.8114, - "step": 3935 - }, - { - "epoch": 0.76, - "learning_rate": 2.932238021122877e-06, - "loss": 0.8523, - "step": 3936 - }, - { - "epoch": 0.76, - "learning_rate": 2.9278305870595814e-06, - "loss": 0.7427, - "step": 3937 - }, - { - "epoch": 0.76, - "learning_rate": 2.92342589974539e-06, - "loss": 0.8859, - "step": 3938 - }, - { - "epoch": 0.76, - "learning_rate": 2.919023960891039e-06, - "loss": 0.934, - "step": 3939 - }, - { - "epoch": 0.76, - "learning_rate": 2.9146247722061806e-06, - "loss": 0.8874, - "step": 3940 - }, - { - "epoch": 0.76, - "learning_rate": 2.910228335399419e-06, - "loss": 0.8101, - "step": 3941 - }, - { - "epoch": 0.76, - "learning_rate": 2.90583465217828e-06, - "loss": 0.692, - "step": 3942 - }, - { - "epoch": 0.76, - "learning_rate": 2.9014437242492133e-06, - "loss": 0.8912, - "step": 3943 - }, - { - "epoch": 0.76, - "learning_rate": 2.897055553317607e-06, - "loss": 0.9149, - "step": 3944 - }, - { - "epoch": 0.76, - "learning_rate": 2.89267014108778e-06, - "loss": 0.8615, - "step": 3945 - }, - { - "epoch": 0.76, - "learning_rate": 2.8882874892629654e-06, - "loss": 0.8359, - "step": 3946 - }, - { - "epoch": 0.76, - "learning_rate": 2.883907599545348e-06, - "loss": 0.8093, - "step": 3947 - }, - { - "epoch": 0.76, - "learning_rate": 2.8795304736360184e-06, - "loss": 0.8945, - "step": 3948 - }, - { - "epoch": 0.76, - "learning_rate": 2.8751561132350025e-06, - "loss": 0.8203, - "step": 3949 - }, - { - "epoch": 0.76, - "learning_rate": 2.8707845200412567e-06, - "loss": 0.8916, - "step": 3950 - }, - { - "epoch": 0.76, - "learning_rate": 2.866415695752649e-06, - "loss": 0.9093, - "step": 3951 - }, - { - "epoch": 0.76, - "learning_rate": 2.862049642065986e-06, - "loss": 0.739, - "step": 3952 - }, - { - "epoch": 0.76, - "learning_rate": 2.857686360676991e-06, - "loss": 0.8884, - "step": 3953 - }, - { - "epoch": 0.76, - "learning_rate": 2.853325853280312e-06, - "loss": 0.8142, - "step": 3954 - }, - { - "epoch": 0.76, - "learning_rate": 2.8489681215695242e-06, - "loss": 0.9571, - "step": 3955 - }, - { - "epoch": 0.76, - "learning_rate": 2.8446131672371136e-06, - "loss": 0.8797, - "step": 3956 - }, - { - "epoch": 0.76, - "learning_rate": 2.840260991974497e-06, - "loss": 0.8284, - "step": 3957 - }, - { - "epoch": 0.76, - "learning_rate": 2.83591159747201e-06, - "loss": 0.8732, - "step": 3958 - }, - { - "epoch": 0.76, - "learning_rate": 2.8315649854189066e-06, - "loss": 0.8795, - "step": 3959 - }, - { - "epoch": 0.76, - "learning_rate": 2.8272211575033635e-06, - "loss": 0.6704, - "step": 3960 - }, - { - "epoch": 0.76, - "learning_rate": 2.8228801154124687e-06, - "loss": 0.8682, - "step": 3961 - }, - { - "epoch": 0.76, - "learning_rate": 2.8185418608322344e-06, - "loss": 0.9179, - "step": 3962 - }, - { - "epoch": 0.76, - "learning_rate": 2.814206395447593e-06, - "loss": 0.9353, - "step": 3963 - }, - { - "epoch": 0.76, - "learning_rate": 2.80987372094238e-06, - "loss": 0.8074, - "step": 3964 - }, - { - "epoch": 0.76, - "learning_rate": 2.805543838999364e-06, - "loss": 0.9021, - "step": 3965 - }, - { - "epoch": 0.76, - "learning_rate": 2.801216751300223e-06, - "loss": 0.8135, - "step": 3966 - }, - { - "epoch": 0.76, - "learning_rate": 2.796892459525541e-06, - "loss": 0.8398, - "step": 3967 - }, - { - "epoch": 0.76, - "learning_rate": 2.7925709653548295e-06, - "loss": 0.8813, - "step": 3968 - }, - { - "epoch": 0.76, - "learning_rate": 2.788252270466497e-06, - "loss": 0.8185, - "step": 3969 - }, - { - "epoch": 0.76, - "learning_rate": 2.783936376537886e-06, - "loss": 0.8927, - "step": 3970 - }, - { - "epoch": 0.76, - "learning_rate": 2.7796232852452378e-06, - "loss": 0.8549, - "step": 3971 - }, - { - "epoch": 0.76, - "learning_rate": 2.775312998263703e-06, - "loss": 0.8898, - "step": 3972 - }, - { - "epoch": 0.76, - "learning_rate": 2.771005517267349e-06, - "loss": 0.8818, - "step": 3973 - }, - { - "epoch": 0.76, - "learning_rate": 2.7667008439291552e-06, - "loss": 0.9128, - "step": 3974 - }, - { - "epoch": 0.76, - "learning_rate": 2.762398979920998e-06, - "loss": 0.9259, - "step": 3975 - }, - { - "epoch": 0.76, - "learning_rate": 2.7580999269136854e-06, - "loss": 0.892, - "step": 3976 - }, - { - "epoch": 0.77, - "learning_rate": 2.7538036865769093e-06, - "loss": 0.834, - "step": 3977 - }, - { - "epoch": 0.77, - "learning_rate": 2.749510260579282e-06, - "loss": 0.8161, - "step": 3978 - }, - { - "epoch": 0.77, - "learning_rate": 2.7452196505883265e-06, - "loss": 0.8966, - "step": 3979 - }, - { - "epoch": 0.77, - "learning_rate": 2.7409318582704594e-06, - "loss": 0.8877, - "step": 3980 - }, - { - "epoch": 0.77, - "learning_rate": 2.736646885291011e-06, - "loss": 0.8501, - "step": 3981 - }, - { - "epoch": 0.77, - "learning_rate": 2.7323647333142176e-06, - "loss": 0.8144, - "step": 3982 - }, - { - "epoch": 0.77, - "learning_rate": 2.728085404003217e-06, - "loss": 0.8004, - "step": 3983 - }, - { - "epoch": 0.77, - "learning_rate": 2.7238088990200538e-06, - "loss": 0.7461, - "step": 3984 - }, - { - "epoch": 0.77, - "learning_rate": 2.7195352200256675e-06, - "loss": 0.8447, - "step": 3985 - }, - { - "epoch": 0.77, - "learning_rate": 2.7152643686799095e-06, - "loss": 0.8706, - "step": 3986 - }, - { - "epoch": 0.77, - "learning_rate": 2.710996346641528e-06, - "loss": 0.8357, - "step": 3987 - }, - { - "epoch": 0.77, - "learning_rate": 2.706731155568175e-06, - "loss": 0.778, - "step": 3988 - }, - { - "epoch": 0.77, - "learning_rate": 2.7024687971164032e-06, - "loss": 0.8267, - "step": 3989 - }, - { - "epoch": 0.77, - "learning_rate": 2.698209272941659e-06, - "loss": 0.9071, - "step": 3990 - }, - { - "epoch": 0.77, - "learning_rate": 2.693952584698294e-06, - "loss": 0.792, - "step": 3991 - }, - { - "epoch": 0.77, - "learning_rate": 2.689698734039561e-06, - "loss": 0.905, - "step": 3992 - }, - { - "epoch": 0.77, - "learning_rate": 2.685447722617597e-06, - "loss": 0.8221, - "step": 3993 - }, - { - "epoch": 0.77, - "learning_rate": 2.6811995520834543e-06, - "loss": 0.8251, - "step": 3994 - }, - { - "epoch": 0.77, - "learning_rate": 2.676954224087075e-06, - "loss": 0.8221, - "step": 3995 - }, - { - "epoch": 0.77, - "learning_rate": 2.6727117402772886e-06, - "loss": 0.8666, - "step": 3996 - }, - { - "epoch": 0.77, - "learning_rate": 2.668472102301829e-06, - "loss": 0.8538, - "step": 3997 - }, - { - "epoch": 0.77, - "learning_rate": 2.664235311807327e-06, - "loss": 0.8503, - "step": 3998 - }, - { - "epoch": 0.77, - "learning_rate": 2.6600013704392946e-06, - "loss": 0.7552, - "step": 3999 - }, - { - "epoch": 0.77, - "learning_rate": 2.655770279842157e-06, - "loss": 0.7985, - "step": 4000 - }, - { - "epoch": 0.77, - "learning_rate": 2.651542041659211e-06, - "loss": 0.7918, - "step": 4001 - }, - { - "epoch": 0.77, - "learning_rate": 2.6473166575326603e-06, - "loss": 0.8005, - "step": 4002 - }, - { - "epoch": 0.77, - "learning_rate": 2.6430941291035984e-06, - "loss": 0.8172, - "step": 4003 - }, - { - "epoch": 0.77, - "learning_rate": 2.6388744580119975e-06, - "loss": 0.8411, - "step": 4004 - }, - { - "epoch": 0.77, - "learning_rate": 2.6346576458967397e-06, - "loss": 0.9105, - "step": 4005 - }, - { - "epoch": 0.77, - "learning_rate": 2.630443694395579e-06, - "loss": 0.8251, - "step": 4006 - }, - { - "epoch": 0.77, - "learning_rate": 2.6262326051451683e-06, - "loss": 0.8967, - "step": 4007 - }, - { - "epoch": 0.77, - "learning_rate": 2.6220243797810483e-06, - "loss": 0.8926, - "step": 4008 - }, - { - "epoch": 0.77, - "learning_rate": 2.6178190199376394e-06, - "loss": 0.8165, - "step": 4009 - }, - { - "epoch": 0.77, - "learning_rate": 2.6136165272482596e-06, - "loss": 0.858, - "step": 4010 - }, - { - "epoch": 0.77, - "learning_rate": 2.6094169033451066e-06, - "loss": 0.7844, - "step": 4011 - }, - { - "epoch": 0.77, - "learning_rate": 2.605220149859267e-06, - "loss": 0.9067, - "step": 4012 - }, - { - "epoch": 0.77, - "learning_rate": 2.6010262684207134e-06, - "loss": 0.8147, - "step": 4013 - }, - { - "epoch": 0.77, - "learning_rate": 2.596835260658297e-06, - "loss": 0.8516, - "step": 4014 - }, - { - "epoch": 0.77, - "learning_rate": 2.5926471281997577e-06, - "loss": 0.8739, - "step": 4015 - }, - { - "epoch": 0.77, - "learning_rate": 2.588461872671719e-06, - "loss": 0.8095, - "step": 4016 - }, - { - "epoch": 0.77, - "learning_rate": 2.5842794956996863e-06, - "loss": 0.7979, - "step": 4017 - }, - { - "epoch": 0.77, - "learning_rate": 2.580099998908049e-06, - "loss": 0.8832, - "step": 4018 - }, - { - "epoch": 0.77, - "learning_rate": 2.575923383920069e-06, - "loss": 0.888, - "step": 4019 - }, - { - "epoch": 0.77, - "learning_rate": 2.5717496523578998e-06, - "loss": 0.8896, - "step": 4020 - }, - { - "epoch": 0.77, - "learning_rate": 2.5675788058425723e-06, - "loss": 0.9184, - "step": 4021 - }, - { - "epoch": 0.77, - "learning_rate": 2.563410845993988e-06, - "loss": 0.874, - "step": 4022 - }, - { - "epoch": 0.77, - "learning_rate": 2.5592457744309405e-06, - "loss": 0.9055, - "step": 4023 - }, - { - "epoch": 0.77, - "learning_rate": 2.5550835927710982e-06, - "loss": 0.79, - "step": 4024 - }, - { - "epoch": 0.77, - "learning_rate": 2.5509243026309983e-06, - "loss": 0.8613, - "step": 4025 - }, - { - "epoch": 0.77, - "learning_rate": 2.546767905626063e-06, - "loss": 0.8959, - "step": 4026 - }, - { - "epoch": 0.77, - "learning_rate": 2.5426144033705937e-06, - "loss": 0.8953, - "step": 4027 - }, - { - "epoch": 0.77, - "learning_rate": 2.5384637974777513e-06, - "loss": 0.6787, - "step": 4028 - }, - { - "epoch": 0.78, - "learning_rate": 2.5343160895595977e-06, - "loss": 0.8668, - "step": 4029 - }, - { - "epoch": 0.78, - "learning_rate": 2.530171281227044e-06, - "loss": 0.8404, - "step": 4030 - }, - { - "epoch": 0.78, - "learning_rate": 2.52602937408989e-06, - "loss": 0.8333, - "step": 4031 - }, - { - "epoch": 0.78, - "learning_rate": 2.5218903697568075e-06, - "loss": 0.8641, - "step": 4032 - }, - { - "epoch": 0.78, - "learning_rate": 2.517754269835332e-06, - "loss": 0.7283, - "step": 4033 - }, - { - "epoch": 0.78, - "learning_rate": 2.5136210759318814e-06, - "loss": 0.8291, - "step": 4034 - }, - { - "epoch": 0.78, - "learning_rate": 2.5094907896517383e-06, - "loss": 0.8711, - "step": 4035 - }, - { - "epoch": 0.78, - "learning_rate": 2.505363412599059e-06, - "loss": 0.896, - "step": 4036 - }, - { - "epoch": 0.78, - "learning_rate": 2.5012389463768737e-06, - "loss": 0.8066, - "step": 4037 - }, - { - "epoch": 0.78, - "learning_rate": 2.4971173925870694e-06, - "loss": 0.891, - "step": 4038 - }, - { - "epoch": 0.78, - "learning_rate": 2.4929987528304144e-06, - "loss": 0.8649, - "step": 4039 - }, - { - "epoch": 0.78, - "learning_rate": 2.4888830287065414e-06, - "loss": 0.8632, - "step": 4040 - }, - { - "epoch": 0.78, - "learning_rate": 2.4847702218139493e-06, - "loss": 0.8003, - "step": 4041 - }, - { - "epoch": 0.78, - "learning_rate": 2.480660333750007e-06, - "loss": 0.7037, - "step": 4042 - }, - { - "epoch": 0.78, - "learning_rate": 2.476553366110944e-06, - "loss": 0.8001, - "step": 4043 - }, - { - "epoch": 0.78, - "learning_rate": 2.4724493204918598e-06, - "loss": 0.8116, - "step": 4044 - }, - { - "epoch": 0.78, - "learning_rate": 2.4683481984867207e-06, - "loss": 0.9281, - "step": 4045 - }, - { - "epoch": 0.78, - "learning_rate": 2.4642500016883532e-06, - "loss": 0.8484, - "step": 4046 - }, - { - "epoch": 0.78, - "learning_rate": 2.4601547316884544e-06, - "loss": 0.8894, - "step": 4047 - }, - { - "epoch": 0.78, - "learning_rate": 2.4560623900775728e-06, - "loss": 0.8658, - "step": 4048 - }, - { - "epoch": 0.78, - "learning_rate": 2.45197297844513e-06, - "loss": 0.8577, - "step": 4049 - }, - { - "epoch": 0.78, - "learning_rate": 2.447886498379409e-06, - "loss": 0.8934, - "step": 4050 - }, - { - "epoch": 0.78, - "learning_rate": 2.4438029514675444e-06, - "loss": 0.8797, - "step": 4051 - }, - { - "epoch": 0.78, - "learning_rate": 2.439722339295545e-06, - "loss": 0.7883, - "step": 4052 - }, - { - "epoch": 0.78, - "learning_rate": 2.4356446634482756e-06, - "loss": 0.8051, - "step": 4053 - }, - { - "epoch": 0.78, - "learning_rate": 2.4315699255094516e-06, - "loss": 0.8244, - "step": 4054 - }, - { - "epoch": 0.78, - "learning_rate": 2.427498127061658e-06, - "loss": 0.8132, - "step": 4055 - }, - { - "epoch": 0.78, - "learning_rate": 2.423429269686336e-06, - "loss": 0.8594, - "step": 4056 - }, - { - "epoch": 0.78, - "learning_rate": 2.4193633549637765e-06, - "loss": 0.8844, - "step": 4057 - }, - { - "epoch": 0.78, - "learning_rate": 2.4153003844731425e-06, - "loss": 0.858, - "step": 4058 - }, - { - "epoch": 0.78, - "learning_rate": 2.411240359792438e-06, - "loss": 0.6981, - "step": 4059 - }, - { - "epoch": 0.78, - "learning_rate": 2.407183282498534e-06, - "loss": 0.8941, - "step": 4060 - }, - { - "epoch": 0.78, - "learning_rate": 2.403129154167153e-06, - "loss": 0.9595, - "step": 4061 - }, - { - "epoch": 0.78, - "learning_rate": 2.3990779763728666e-06, - "loss": 0.8971, - "step": 4062 - }, - { - "epoch": 0.78, - "learning_rate": 2.3950297506891084e-06, - "loss": 0.8104, - "step": 4063 - }, - { - "epoch": 0.78, - "learning_rate": 2.390984478688164e-06, - "loss": 0.8341, - "step": 4064 - }, - { - "epoch": 0.78, - "learning_rate": 2.386942161941169e-06, - "loss": 0.8828, - "step": 4065 - }, - { - "epoch": 0.78, - "learning_rate": 2.3829028020181154e-06, - "loss": 0.8715, - "step": 4066 - }, - { - "epoch": 0.78, - "learning_rate": 2.3788664004878405e-06, - "loss": 0.8876, - "step": 4067 - }, - { - "epoch": 0.78, - "learning_rate": 2.374832958918035e-06, - "loss": 0.8187, - "step": 4068 - }, - { - "epoch": 0.78, - "learning_rate": 2.3708024788752448e-06, - "loss": 0.9473, - "step": 4069 - }, - { - "epoch": 0.78, - "learning_rate": 2.3667749619248614e-06, - "loss": 0.8366, - "step": 4070 - }, - { - "epoch": 0.78, - "learning_rate": 2.3627504096311273e-06, - "loss": 0.8769, - "step": 4071 - }, - { - "epoch": 0.78, - "learning_rate": 2.3587288235571258e-06, - "loss": 0.8648, - "step": 4072 - }, - { - "epoch": 0.78, - "learning_rate": 2.354710205264801e-06, - "loss": 0.928, - "step": 4073 - }, - { - "epoch": 0.78, - "learning_rate": 2.350694556314934e-06, - "loss": 0.822, - "step": 4074 - }, - { - "epoch": 0.78, - "learning_rate": 2.3466818782671597e-06, - "loss": 0.8902, - "step": 4075 - }, - { - "epoch": 0.78, - "learning_rate": 2.3426721726799573e-06, - "loss": 0.835, - "step": 4076 - }, - { - "epoch": 0.78, - "learning_rate": 2.3386654411106446e-06, - "loss": 0.8756, - "step": 4077 - }, - { - "epoch": 0.78, - "learning_rate": 2.3346616851153935e-06, - "loss": 0.8905, - "step": 4078 - }, - { - "epoch": 0.78, - "learning_rate": 2.330660906249218e-06, - "loss": 0.898, - "step": 4079 - }, - { - "epoch": 0.78, - "learning_rate": 2.3266631060659685e-06, - "loss": 0.8437, - "step": 4080 - }, - { - "epoch": 0.79, - "learning_rate": 2.32266828611835e-06, - "loss": 0.8647, - "step": 4081 - }, - { - "epoch": 0.79, - "learning_rate": 2.318676447957907e-06, - "loss": 0.9114, - "step": 4082 - }, - { - "epoch": 0.79, - "learning_rate": 2.3146875931350165e-06, - "loss": 0.8895, - "step": 4083 - }, - { - "epoch": 0.79, - "learning_rate": 2.310701723198908e-06, - "loss": 0.8735, - "step": 4084 - }, - { - "epoch": 0.79, - "learning_rate": 2.3067188396976482e-06, - "loss": 0.8667, - "step": 4085 - }, - { - "epoch": 0.79, - "learning_rate": 2.3027389441781368e-06, - "loss": 0.8706, - "step": 4086 - }, - { - "epoch": 0.79, - "learning_rate": 2.2987620381861288e-06, - "loss": 0.883, - "step": 4087 - }, - { - "epoch": 0.79, - "learning_rate": 2.2947881232662007e-06, - "loss": 0.8388, - "step": 4088 - }, - { - "epoch": 0.79, - "learning_rate": 2.290817200961779e-06, - "loss": 0.9405, - "step": 4089 - }, - { - "epoch": 0.79, - "learning_rate": 2.286849272815126e-06, - "loss": 0.7789, - "step": 4090 - }, - { - "epoch": 0.79, - "learning_rate": 2.2828843403673338e-06, - "loss": 0.8742, - "step": 4091 - }, - { - "epoch": 0.79, - "learning_rate": 2.2789224051583403e-06, - "loss": 0.8411, - "step": 4092 - }, - { - "epoch": 0.79, - "learning_rate": 2.274963468726914e-06, - "loss": 0.8843, - "step": 4093 - }, - { - "epoch": 0.79, - "learning_rate": 2.2710075326106618e-06, - "loss": 0.7872, - "step": 4094 - }, - { - "epoch": 0.79, - "learning_rate": 2.2670545983460245e-06, - "loss": 0.8471, - "step": 4095 - }, - { - "epoch": 0.79, - "learning_rate": 2.263104667468272e-06, - "loss": 0.8485, - "step": 4096 - }, - { - "epoch": 0.79, - "learning_rate": 2.259157741511515e-06, - "loss": 0.9353, - "step": 4097 - }, - { - "epoch": 0.79, - "learning_rate": 2.2552138220086927e-06, - "loss": 0.7017, - "step": 4098 - }, - { - "epoch": 0.79, - "learning_rate": 2.2512729104915787e-06, - "loss": 0.8413, - "step": 4099 - }, - { - "epoch": 0.79, - "learning_rate": 2.2473350084907806e-06, - "loss": 0.8878, - "step": 4100 - }, - { - "epoch": 0.79, - "learning_rate": 2.243400117535729e-06, - "loss": 0.7804, - "step": 4101 - }, - { - "epoch": 0.79, - "learning_rate": 2.2394682391546928e-06, - "loss": 0.8521, - "step": 4102 - }, - { - "epoch": 0.79, - "learning_rate": 2.2355393748747702e-06, - "loss": 0.7216, - "step": 4103 - }, - { - "epoch": 0.79, - "learning_rate": 2.2316135262218787e-06, - "loss": 0.8056, - "step": 4104 - }, - { - "epoch": 0.79, - "learning_rate": 2.227690694720784e-06, - "loss": 0.777, - "step": 4105 - }, - { - "epoch": 0.79, - "learning_rate": 2.223770881895061e-06, - "loss": 0.9381, - "step": 4106 - }, - { - "epoch": 0.79, - "learning_rate": 2.2198540892671215e-06, - "loss": 0.8631, - "step": 4107 - }, - { - "epoch": 0.79, - "learning_rate": 2.215940318358206e-06, - "loss": 0.849, - "step": 4108 - }, - { - "epoch": 0.79, - "learning_rate": 2.2120295706883698e-06, - "loss": 0.8213, - "step": 4109 - }, - { - "epoch": 0.79, - "learning_rate": 2.2081218477765097e-06, - "loss": 0.9264, - "step": 4110 - }, - { - "epoch": 0.79, - "learning_rate": 2.204217151140342e-06, - "loss": 0.7984, - "step": 4111 - }, - { - "epoch": 0.79, - "learning_rate": 2.200315482296398e-06, - "loss": 0.81, - "step": 4112 - }, - { - "epoch": 0.79, - "learning_rate": 2.1964168427600462e-06, - "loss": 0.8775, - "step": 4113 - }, - { - "epoch": 0.79, - "learning_rate": 2.1925212340454737e-06, - "loss": 0.92, - "step": 4114 - }, - { - "epoch": 0.79, - "learning_rate": 2.1886286576656834e-06, - "loss": 0.8387, - "step": 4115 - }, - { - "epoch": 0.79, - "learning_rate": 2.184739115132517e-06, - "loss": 0.9113, - "step": 4116 - }, - { - "epoch": 0.79, - "learning_rate": 2.1808526079566215e-06, - "loss": 0.8823, - "step": 4117 - }, - { - "epoch": 0.79, - "learning_rate": 2.1769691376474722e-06, - "loss": 0.8374, - "step": 4118 - }, - { - "epoch": 0.79, - "learning_rate": 2.1730887057133678e-06, - "loss": 0.8912, - "step": 4119 - }, - { - "epoch": 0.79, - "learning_rate": 2.1692113136614177e-06, - "loss": 0.8548, - "step": 4120 - }, - { - "epoch": 0.79, - "learning_rate": 2.1653369629975595e-06, - "loss": 0.8553, - "step": 4121 - }, - { - "epoch": 0.79, - "learning_rate": 2.1614656552265457e-06, - "loss": 0.8864, - "step": 4122 - }, - { - "epoch": 0.79, - "learning_rate": 2.157597391851949e-06, - "loss": 0.6757, - "step": 4123 - }, - { - "epoch": 0.79, - "learning_rate": 2.1537321743761587e-06, - "loss": 0.7597, - "step": 4124 - }, - { - "epoch": 0.79, - "learning_rate": 2.1498700043003773e-06, - "loss": 0.7965, - "step": 4125 - }, - { - "epoch": 0.79, - "learning_rate": 2.1460108831246295e-06, - "loss": 0.8601, - "step": 4126 - }, - { - "epoch": 0.79, - "learning_rate": 2.142154812347753e-06, - "loss": 0.9165, - "step": 4127 - }, - { - "epoch": 0.79, - "learning_rate": 2.1383017934674012e-06, - "loss": 0.875, - "step": 4128 - }, - { - "epoch": 0.79, - "learning_rate": 2.1344518279800454e-06, - "loss": 0.8907, - "step": 4129 - }, - { - "epoch": 0.79, - "learning_rate": 2.130604917380962e-06, - "loss": 0.7727, - "step": 4130 - }, - { - "epoch": 0.79, - "learning_rate": 2.12676106316425e-06, - "loss": 0.7279, - "step": 4131 - }, - { - "epoch": 0.79, - "learning_rate": 2.1229202668228197e-06, - "loss": 0.7554, - "step": 4132 - }, - { - "epoch": 0.8, - "learning_rate": 2.1190825298483855e-06, - "loss": 0.7949, - "step": 4133 - }, - { - "epoch": 0.8, - "learning_rate": 2.115247853731488e-06, - "loss": 0.7974, - "step": 4134 - }, - { - "epoch": 0.8, - "learning_rate": 2.11141623996147e-06, - "loss": 0.8389, - "step": 4135 - }, - { - "epoch": 0.8, - "learning_rate": 2.107587690026481e-06, - "loss": 0.8264, - "step": 4136 - }, - { - "epoch": 0.8, - "learning_rate": 2.103762205413493e-06, - "loss": 0.8493, - "step": 4137 - }, - { - "epoch": 0.8, - "learning_rate": 2.0999397876082726e-06, - "loss": 0.8814, - "step": 4138 - }, - { - "epoch": 0.8, - "learning_rate": 2.096120438095404e-06, - "loss": 0.8298, - "step": 4139 - }, - { - "epoch": 0.8, - "learning_rate": 2.092304158358286e-06, - "loss": 0.8508, - "step": 4140 - }, - { - "epoch": 0.8, - "learning_rate": 2.0884909498791106e-06, - "loss": 0.8824, - "step": 4141 - }, - { - "epoch": 0.8, - "learning_rate": 2.0846808141388852e-06, - "loss": 0.8403, - "step": 4142 - }, - { - "epoch": 0.8, - "learning_rate": 2.080873752617426e-06, - "loss": 0.8274, - "step": 4143 - }, - { - "epoch": 0.8, - "learning_rate": 2.0770697667933436e-06, - "loss": 0.8556, - "step": 4144 - }, - { - "epoch": 0.8, - "learning_rate": 2.073268858144074e-06, - "loss": 0.7993, - "step": 4145 - }, - { - "epoch": 0.8, - "learning_rate": 2.0694710281458372e-06, - "loss": 0.8036, - "step": 4146 - }, - { - "epoch": 0.8, - "learning_rate": 2.0656762782736693e-06, - "loss": 0.9405, - "step": 4147 - }, - { - "epoch": 0.8, - "learning_rate": 2.061884610001411e-06, - "loss": 0.7175, - "step": 4148 - }, - { - "epoch": 0.8, - "learning_rate": 2.0580960248016966e-06, - "loss": 0.8778, - "step": 4149 - }, - { - "epoch": 0.8, - "learning_rate": 2.0543105241459713e-06, - "loss": 0.896, - "step": 4150 - }, - { - "epoch": 0.8, - "learning_rate": 2.0505281095044804e-06, - "loss": 0.7545, - "step": 4151 - }, - { - "epoch": 0.8, - "learning_rate": 2.0467487823462696e-06, - "loss": 0.8486, - "step": 4152 - }, - { - "epoch": 0.8, - "learning_rate": 2.042972544139189e-06, - "loss": 0.88, - "step": 4153 - }, - { - "epoch": 0.8, - "learning_rate": 2.039199396349881e-06, - "loss": 0.8764, - "step": 4154 - }, - { - "epoch": 0.8, - "learning_rate": 2.0354293404437963e-06, - "loss": 0.896, - "step": 4155 - }, - { - "epoch": 0.8, - "learning_rate": 2.0316623778851784e-06, - "loss": 0.7974, - "step": 4156 - }, - { - "epoch": 0.8, - "learning_rate": 2.0278985101370753e-06, - "loss": 0.8577, - "step": 4157 - }, - { - "epoch": 0.8, - "learning_rate": 2.024137738661329e-06, - "loss": 0.859, - "step": 4158 - }, - { - "epoch": 0.8, - "learning_rate": 2.020380064918579e-06, - "loss": 0.9123, - "step": 4159 - }, - { - "epoch": 0.8, - "learning_rate": 2.01662549036826e-06, - "loss": 0.891, - "step": 4160 - }, - { - "epoch": 0.8, - "learning_rate": 2.0128740164686134e-06, - "loss": 0.8088, - "step": 4161 - }, - { - "epoch": 0.8, - "learning_rate": 2.009125644676656e-06, - "loss": 0.8289, - "step": 4162 - }, - { - "epoch": 0.8, - "learning_rate": 2.0053803764482226e-06, - "loss": 0.8463, - "step": 4163 - }, - { - "epoch": 0.8, - "learning_rate": 2.001638213237932e-06, - "loss": 0.8698, - "step": 4164 - }, - { - "epoch": 0.8, - "learning_rate": 1.997899156499191e-06, - "loss": 0.8874, - "step": 4165 - }, - { - "epoch": 0.8, - "learning_rate": 1.994163207684212e-06, - "loss": 0.897, - "step": 4166 - }, - { - "epoch": 0.8, - "learning_rate": 1.9904303682439896e-06, - "loss": 0.9267, - "step": 4167 - }, - { - "epoch": 0.8, - "learning_rate": 1.986700639628316e-06, - "loss": 0.8976, - "step": 4168 - }, - { - "epoch": 0.8, - "learning_rate": 1.9829740232857807e-06, - "loss": 0.9064, - "step": 4169 - }, - { - "epoch": 0.8, - "learning_rate": 1.9792505206637523e-06, - "loss": 0.8994, - "step": 4170 - }, - { - "epoch": 0.8, - "learning_rate": 1.9755301332083997e-06, - "loss": 0.8493, - "step": 4171 - }, - { - "epoch": 0.8, - "learning_rate": 1.9718128623646792e-06, - "loss": 0.847, - "step": 4172 - }, - { - "epoch": 0.8, - "learning_rate": 1.9680987095763315e-06, - "loss": 0.7824, - "step": 4173 - }, - { - "epoch": 0.8, - "learning_rate": 1.964387676285894e-06, - "loss": 0.8557, - "step": 4174 - }, - { - "epoch": 0.8, - "learning_rate": 1.9606797639346874e-06, - "loss": 0.8765, - "step": 4175 - }, - { - "epoch": 0.8, - "learning_rate": 1.9569749739628243e-06, - "loss": 0.7271, - "step": 4176 - }, - { - "epoch": 0.8, - "learning_rate": 1.9532733078092034e-06, - "loss": 0.9126, - "step": 4177 - }, - { - "epoch": 0.8, - "learning_rate": 1.9495747669115062e-06, - "loss": 0.9148, - "step": 4178 - }, - { - "epoch": 0.8, - "learning_rate": 1.9458793527062035e-06, - "loss": 0.8287, - "step": 4179 - }, - { - "epoch": 0.8, - "learning_rate": 1.9421870666285523e-06, - "loss": 0.8438, - "step": 4180 - }, - { - "epoch": 0.8, - "learning_rate": 1.9384979101125944e-06, - "loss": 0.8984, - "step": 4181 - }, - { - "epoch": 0.8, - "learning_rate": 1.934811884591159e-06, - "loss": 0.7854, - "step": 4182 - }, - { - "epoch": 0.8, - "learning_rate": 1.93112899149585e-06, - "loss": 0.9635, - "step": 4183 - }, - { - "epoch": 0.8, - "learning_rate": 1.9274492322570616e-06, - "loss": 0.8743, - "step": 4184 - }, - { - "epoch": 0.81, - "learning_rate": 1.923772608303972e-06, - "loss": 0.7903, - "step": 4185 - }, - { - "epoch": 0.81, - "learning_rate": 1.9200991210645394e-06, - "loss": 0.7978, - "step": 4186 - }, - { - "epoch": 0.81, - "learning_rate": 1.916428771965506e-06, - "loss": 0.8243, - "step": 4187 - }, - { - "epoch": 0.81, - "learning_rate": 1.912761562432388e-06, - "loss": 0.8393, - "step": 4188 - }, - { - "epoch": 0.81, - "learning_rate": 1.9090974938894902e-06, - "loss": 0.8134, - "step": 4189 - }, - { - "epoch": 0.81, - "learning_rate": 1.9054365677598963e-06, - "loss": 0.7918, - "step": 4190 - }, - { - "epoch": 0.81, - "learning_rate": 1.9017787854654613e-06, - "loss": 0.9231, - "step": 4191 - }, - { - "epoch": 0.81, - "learning_rate": 1.898124148426832e-06, - "loss": 0.835, - "step": 4192 - }, - { - "epoch": 0.81, - "learning_rate": 1.8944726580634287e-06, - "loss": 0.6989, - "step": 4193 - }, - { - "epoch": 0.81, - "learning_rate": 1.8908243157934424e-06, - "loss": 0.8005, - "step": 4194 - }, - { - "epoch": 0.81, - "learning_rate": 1.8871791230338499e-06, - "loss": 0.8228, - "step": 4195 - }, - { - "epoch": 0.81, - "learning_rate": 1.883537081200404e-06, - "loss": 0.8222, - "step": 4196 - }, - { - "epoch": 0.81, - "learning_rate": 1.8798981917076254e-06, - "loss": 0.9165, - "step": 4197 - }, - { - "epoch": 0.81, - "learning_rate": 1.876262455968826e-06, - "loss": 0.8137, - "step": 4198 - }, - { - "epoch": 0.81, - "learning_rate": 1.872629875396076e-06, - "loss": 0.8901, - "step": 4199 - }, - { - "epoch": 0.81, - "learning_rate": 1.8690004514002314e-06, - "loss": 0.9251, - "step": 4200 - }, - { - "epoch": 0.81, - "learning_rate": 1.8653741853909201e-06, - "loss": 0.8093, - "step": 4201 - }, - { - "epoch": 0.81, - "learning_rate": 1.861751078776538e-06, - "loss": 0.932, - "step": 4202 - }, - { - "epoch": 0.81, - "learning_rate": 1.8581311329642592e-06, - "loss": 0.7859, - "step": 4203 - }, - { - "epoch": 0.81, - "learning_rate": 1.8545143493600293e-06, - "loss": 0.8344, - "step": 4204 - }, - { - "epoch": 0.81, - "learning_rate": 1.8509007293685666e-06, - "loss": 0.8269, - "step": 4205 - }, - { - "epoch": 0.81, - "learning_rate": 1.8472902743933608e-06, - "loss": 0.7568, - "step": 4206 - }, - { - "epoch": 0.81, - "learning_rate": 1.8436829858366655e-06, - "loss": 0.8492, - "step": 4207 - }, - { - "epoch": 0.81, - "learning_rate": 1.8400788650995137e-06, - "loss": 0.8274, - "step": 4208 - }, - { - "epoch": 0.81, - "learning_rate": 1.8364779135817045e-06, - "loss": 0.8148, - "step": 4209 - }, - { - "epoch": 0.81, - "learning_rate": 1.8328801326818045e-06, - "loss": 0.8335, - "step": 4210 - }, - { - "epoch": 0.81, - "learning_rate": 1.829285523797155e-06, - "loss": 0.7751, - "step": 4211 - }, - { - "epoch": 0.81, - "learning_rate": 1.8256940883238538e-06, - "loss": 0.8311, - "step": 4212 - }, - { - "epoch": 0.81, - "learning_rate": 1.822105827656776e-06, - "loss": 0.8507, - "step": 4213 - }, - { - "epoch": 0.81, - "learning_rate": 1.8185207431895613e-06, - "loss": 0.8665, - "step": 4214 - }, - { - "epoch": 0.81, - "learning_rate": 1.8149388363146148e-06, - "loss": 0.8851, - "step": 4215 - }, - { - "epoch": 0.81, - "learning_rate": 1.8113601084231091e-06, - "loss": 0.8594, - "step": 4216 - }, - { - "epoch": 0.81, - "learning_rate": 1.8077845609049782e-06, - "loss": 0.8418, - "step": 4217 - }, - { - "epoch": 0.81, - "learning_rate": 1.8042121951489254e-06, - "loss": 0.8594, - "step": 4218 - }, - { - "epoch": 0.81, - "learning_rate": 1.800643012542418e-06, - "loss": 0.8504, - "step": 4219 - }, - { - "epoch": 0.81, - "learning_rate": 1.7970770144716777e-06, - "loss": 0.7888, - "step": 4220 - }, - { - "epoch": 0.81, - "learning_rate": 1.7935142023217056e-06, - "loss": 0.8075, - "step": 4221 - }, - { - "epoch": 0.81, - "learning_rate": 1.7899545774762573e-06, - "loss": 0.7691, - "step": 4222 - }, - { - "epoch": 0.81, - "learning_rate": 1.7863981413178433e-06, - "loss": 0.8953, - "step": 4223 - }, - { - "epoch": 0.81, - "learning_rate": 1.7828448952277456e-06, - "loss": 0.8582, - "step": 4224 - }, - { - "epoch": 0.81, - "learning_rate": 1.7792948405860079e-06, - "loss": 0.9031, - "step": 4225 - }, - { - "epoch": 0.81, - "learning_rate": 1.7757479787714217e-06, - "loss": 0.8219, - "step": 4226 - }, - { - "epoch": 0.81, - "learning_rate": 1.7722043111615572e-06, - "loss": 0.8812, - "step": 4227 - }, - { - "epoch": 0.81, - "learning_rate": 1.768663839132727e-06, - "loss": 0.7951, - "step": 4228 - }, - { - "epoch": 0.81, - "learning_rate": 1.7651265640600113e-06, - "loss": 0.8705, - "step": 4229 - }, - { - "epoch": 0.81, - "learning_rate": 1.7615924873172506e-06, - "loss": 0.8244, - "step": 4230 - }, - { - "epoch": 0.81, - "learning_rate": 1.7580616102770353e-06, - "loss": 0.8038, - "step": 4231 - }, - { - "epoch": 0.81, - "learning_rate": 1.754533934310717e-06, - "loss": 0.7519, - "step": 4232 - }, - { - "epoch": 0.81, - "learning_rate": 1.7510094607884075e-06, - "loss": 0.8922, - "step": 4233 - }, - { - "epoch": 0.81, - "learning_rate": 1.7474881910789698e-06, - "loss": 0.8202, - "step": 4234 - }, - { - "epoch": 0.81, - "learning_rate": 1.7439701265500274e-06, - "loss": 0.8233, - "step": 4235 - }, - { - "epoch": 0.81, - "learning_rate": 1.740455268567951e-06, - "loss": 0.8522, - "step": 4236 - }, - { - "epoch": 0.82, - "learning_rate": 1.7369436184978738e-06, - "loss": 0.8793, - "step": 4237 - }, - { - "epoch": 0.82, - "learning_rate": 1.7334351777036807e-06, - "loss": 0.8978, - "step": 4238 - }, - { - "epoch": 0.82, - "learning_rate": 1.729929947548008e-06, - "loss": 0.9106, - "step": 4239 - }, - { - "epoch": 0.82, - "learning_rate": 1.7264279293922503e-06, - "loss": 0.8782, - "step": 4240 - }, - { - "epoch": 0.82, - "learning_rate": 1.7229291245965463e-06, - "loss": 0.8062, - "step": 4241 - }, - { - "epoch": 0.82, - "learning_rate": 1.7194335345197933e-06, - "loss": 0.8465, - "step": 4242 - }, - { - "epoch": 0.82, - "learning_rate": 1.7159411605196407e-06, - "loss": 0.8022, - "step": 4243 - }, - { - "epoch": 0.82, - "learning_rate": 1.7124520039524805e-06, - "loss": 0.8631, - "step": 4244 - }, - { - "epoch": 0.82, - "learning_rate": 1.7089660661734685e-06, - "loss": 0.8597, - "step": 4245 - }, - { - "epoch": 0.82, - "learning_rate": 1.7054833485364962e-06, - "loss": 0.7891, - "step": 4246 - }, - { - "epoch": 0.82, - "learning_rate": 1.702003852394214e-06, - "loss": 0.8865, - "step": 4247 - }, - { - "epoch": 0.82, - "learning_rate": 1.6985275790980205e-06, - "loss": 0.8786, - "step": 4248 - }, - { - "epoch": 0.82, - "learning_rate": 1.6950545299980526e-06, - "loss": 0.7698, - "step": 4249 - }, - { - "epoch": 0.82, - "learning_rate": 1.691584706443209e-06, - "loss": 0.7506, - "step": 4250 - }, - { - "epoch": 0.82, - "learning_rate": 1.6881181097811305e-06, - "loss": 0.8786, - "step": 4251 - }, - { - "epoch": 0.82, - "learning_rate": 1.6846547413581981e-06, - "loss": 0.8368, - "step": 4252 - }, - { - "epoch": 0.82, - "learning_rate": 1.6811946025195459e-06, - "loss": 0.8435, - "step": 4253 - }, - { - "epoch": 0.82, - "learning_rate": 1.6777376946090552e-06, - "loss": 0.9141, - "step": 4254 - }, - { - "epoch": 0.82, - "learning_rate": 1.674284018969342e-06, - "loss": 0.8242, - "step": 4255 - }, - { - "epoch": 0.82, - "learning_rate": 1.6708335769417827e-06, - "loss": 0.8941, - "step": 4256 - }, - { - "epoch": 0.82, - "learning_rate": 1.667386369866484e-06, - "loss": 0.8709, - "step": 4257 - }, - { - "epoch": 0.82, - "learning_rate": 1.6639423990823011e-06, - "loss": 0.8635, - "step": 4258 - }, - { - "epoch": 0.82, - "learning_rate": 1.660501665926838e-06, - "loss": 0.7601, - "step": 4259 - }, - { - "epoch": 0.82, - "learning_rate": 1.6570641717364277e-06, - "loss": 0.6305, - "step": 4260 - }, - { - "epoch": 0.82, - "learning_rate": 1.653629917846159e-06, - "loss": 0.8263, - "step": 4261 - }, - { - "epoch": 0.82, - "learning_rate": 1.6501989055898537e-06, - "loss": 0.8596, - "step": 4262 - }, - { - "epoch": 0.82, - "learning_rate": 1.6467711363000794e-06, - "loss": 0.8388, - "step": 4263 - }, - { - "epoch": 0.82, - "learning_rate": 1.6433466113081442e-06, - "loss": 0.8411, - "step": 4264 - }, - { - "epoch": 0.82, - "learning_rate": 1.6399253319440888e-06, - "loss": 0.8834, - "step": 4265 - }, - { - "epoch": 0.82, - "learning_rate": 1.6365072995367004e-06, - "loss": 0.9617, - "step": 4266 - }, - { - "epoch": 0.82, - "learning_rate": 1.6330925154135057e-06, - "loss": 0.8464, - "step": 4267 - }, - { - "epoch": 0.82, - "learning_rate": 1.6296809809007652e-06, - "loss": 0.8034, - "step": 4268 - }, - { - "epoch": 0.82, - "learning_rate": 1.6262726973234844e-06, - "loss": 0.9183, - "step": 4269 - }, - { - "epoch": 0.82, - "learning_rate": 1.6228676660053932e-06, - "loss": 0.8587, - "step": 4270 - }, - { - "epoch": 0.82, - "learning_rate": 1.6194658882689718e-06, - "loss": 0.8912, - "step": 4271 - }, - { - "epoch": 0.82, - "learning_rate": 1.6160673654354331e-06, - "loss": 0.8699, - "step": 4272 - }, - { - "epoch": 0.82, - "learning_rate": 1.6126720988247168e-06, - "loss": 0.7016, - "step": 4273 - }, - { - "epoch": 0.82, - "learning_rate": 1.609280089755515e-06, - "loss": 0.8327, - "step": 4274 - }, - { - "epoch": 0.82, - "learning_rate": 1.605891339545237e-06, - "loss": 0.893, - "step": 4275 - }, - { - "epoch": 0.82, - "learning_rate": 1.6025058495100388e-06, - "loss": 0.8508, - "step": 4276 - }, - { - "epoch": 0.82, - "learning_rate": 1.5991236209648052e-06, - "loss": 0.7878, - "step": 4277 - }, - { - "epoch": 0.82, - "learning_rate": 1.5957446552231526e-06, - "loss": 0.8793, - "step": 4278 - }, - { - "epoch": 0.82, - "learning_rate": 1.5923689535974307e-06, - "loss": 0.8657, - "step": 4279 - }, - { - "epoch": 0.82, - "learning_rate": 1.588996517398731e-06, - "loss": 0.8155, - "step": 4280 - }, - { - "epoch": 0.82, - "learning_rate": 1.5856273479368611e-06, - "loss": 0.8877, - "step": 4281 - }, - { - "epoch": 0.82, - "learning_rate": 1.582261446520371e-06, - "loss": 0.9698, - "step": 4282 - }, - { - "epoch": 0.82, - "learning_rate": 1.5788988144565397e-06, - "loss": 0.7899, - "step": 4283 - }, - { - "epoch": 0.82, - "learning_rate": 1.575539453051369e-06, - "loss": 0.9117, - "step": 4284 - }, - { - "epoch": 0.82, - "learning_rate": 1.572183363609603e-06, - "loss": 0.8019, - "step": 4285 - }, - { - "epoch": 0.82, - "learning_rate": 1.568830547434703e-06, - "loss": 0.8664, - "step": 4286 - }, - { - "epoch": 0.82, - "learning_rate": 1.5654810058288661e-06, - "loss": 0.8504, - "step": 4287 - }, - { - "epoch": 0.82, - "learning_rate": 1.5621347400930176e-06, - "loss": 0.7201, - "step": 4288 - }, - { - "epoch": 0.83, - "learning_rate": 1.5587917515268048e-06, - "loss": 0.8771, - "step": 4289 - }, - { - "epoch": 0.83, - "learning_rate": 1.5554520414286067e-06, - "loss": 0.9088, - "step": 4290 - }, - { - "epoch": 0.83, - "learning_rate": 1.5521156110955293e-06, - "loss": 0.8767, - "step": 4291 - }, - { - "epoch": 0.83, - "learning_rate": 1.5487824618234049e-06, - "loss": 0.8749, - "step": 4292 - }, - { - "epoch": 0.83, - "learning_rate": 1.54545259490679e-06, - "loss": 0.8483, - "step": 4293 - }, - { - "epoch": 0.83, - "learning_rate": 1.5421260116389636e-06, - "loss": 0.9187, - "step": 4294 - }, - { - "epoch": 0.83, - "learning_rate": 1.5388027133119343e-06, - "loss": 0.8371, - "step": 4295 - }, - { - "epoch": 0.83, - "learning_rate": 1.535482701216433e-06, - "loss": 0.8753, - "step": 4296 - }, - { - "epoch": 0.83, - "learning_rate": 1.5321659766419129e-06, - "loss": 0.8551, - "step": 4297 - }, - { - "epoch": 0.83, - "learning_rate": 1.5288525408765564e-06, - "loss": 0.9211, - "step": 4298 - }, - { - "epoch": 0.83, - "learning_rate": 1.5255423952072567e-06, - "loss": 0.9574, - "step": 4299 - }, - { - "epoch": 0.83, - "learning_rate": 1.52223554091964e-06, - "loss": 0.8612, - "step": 4300 - }, - { - "epoch": 0.83, - "learning_rate": 1.5189319792980517e-06, - "loss": 0.7743, - "step": 4301 - }, - { - "epoch": 0.83, - "learning_rate": 1.5156317116255515e-06, - "loss": 0.7091, - "step": 4302 - }, - { - "epoch": 0.83, - "learning_rate": 1.5123347391839305e-06, - "loss": 0.8774, - "step": 4303 - }, - { - "epoch": 0.83, - "learning_rate": 1.5090410632536968e-06, - "loss": 0.9411, - "step": 4304 - }, - { - "epoch": 0.83, - "learning_rate": 1.5057506851140701e-06, - "loss": 0.8752, - "step": 4305 - }, - { - "epoch": 0.83, - "learning_rate": 1.5024636060429998e-06, - "loss": 0.7356, - "step": 4306 - }, - { - "epoch": 0.83, - "learning_rate": 1.4991798273171465e-06, - "loss": 0.8743, - "step": 4307 - }, - { - "epoch": 0.83, - "learning_rate": 1.49589935021189e-06, - "loss": 0.8765, - "step": 4308 - }, - { - "epoch": 0.83, - "learning_rate": 1.4926221760013393e-06, - "loss": 0.8053, - "step": 4309 - }, - { - "epoch": 0.83, - "learning_rate": 1.4893483059583014e-06, - "loss": 0.6988, - "step": 4310 - }, - { - "epoch": 0.83, - "learning_rate": 1.4860777413543138e-06, - "loss": 0.7301, - "step": 4311 - }, - { - "epoch": 0.83, - "learning_rate": 1.4828104834596268e-06, - "loss": 0.9022, - "step": 4312 - }, - { - "epoch": 0.83, - "learning_rate": 1.4795465335432036e-06, - "loss": 0.8264, - "step": 4313 - }, - { - "epoch": 0.83, - "learning_rate": 1.4762858928727241e-06, - "loss": 0.7175, - "step": 4314 - }, - { - "epoch": 0.83, - "learning_rate": 1.4730285627145858e-06, - "loss": 0.7979, - "step": 4315 - }, - { - "epoch": 0.83, - "learning_rate": 1.4697745443338984e-06, - "loss": 0.777, - "step": 4316 - }, - { - "epoch": 0.83, - "learning_rate": 1.4665238389944859e-06, - "loss": 0.8584, - "step": 4317 - }, - { - "epoch": 0.83, - "learning_rate": 1.46327644795888e-06, - "loss": 0.8551, - "step": 4318 - }, - { - "epoch": 0.83, - "learning_rate": 1.4600323724883337e-06, - "loss": 0.9006, - "step": 4319 - }, - { - "epoch": 0.83, - "learning_rate": 1.4567916138428072e-06, - "loss": 0.8899, - "step": 4320 - }, - { - "epoch": 0.83, - "learning_rate": 1.4535541732809755e-06, - "loss": 0.7885, - "step": 4321 - }, - { - "epoch": 0.83, - "learning_rate": 1.4503200520602245e-06, - "loss": 0.793, - "step": 4322 - }, - { - "epoch": 0.83, - "learning_rate": 1.4470892514366442e-06, - "loss": 0.8741, - "step": 4323 - }, - { - "epoch": 0.83, - "learning_rate": 1.443861772665044e-06, - "loss": 0.8658, - "step": 4324 - }, - { - "epoch": 0.83, - "learning_rate": 1.4406376169989389e-06, - "loss": 0.7611, - "step": 4325 - }, - { - "epoch": 0.83, - "learning_rate": 1.4374167856905542e-06, - "loss": 0.9491, - "step": 4326 - }, - { - "epoch": 0.83, - "learning_rate": 1.4341992799908255e-06, - "loss": 0.7283, - "step": 4327 - }, - { - "epoch": 0.83, - "learning_rate": 1.4309851011493903e-06, - "loss": 0.853, - "step": 4328 - }, - { - "epoch": 0.83, - "learning_rate": 1.427774250414601e-06, - "loss": 0.8923, - "step": 4329 - }, - { - "epoch": 0.83, - "learning_rate": 1.4245667290335175e-06, - "loss": 0.867, - "step": 4330 - }, - { - "epoch": 0.83, - "learning_rate": 1.421362538251897e-06, - "loss": 0.8259, - "step": 4331 - }, - { - "epoch": 0.83, - "learning_rate": 1.4181616793142173e-06, - "loss": 0.7949, - "step": 4332 - }, - { - "epoch": 0.83, - "learning_rate": 1.414964153463655e-06, - "loss": 0.8383, - "step": 4333 - }, - { - "epoch": 0.83, - "learning_rate": 1.4117699619420878e-06, - "loss": 0.8146, - "step": 4334 - }, - { - "epoch": 0.83, - "learning_rate": 1.4085791059901077e-06, - "loss": 0.7572, - "step": 4335 - }, - { - "epoch": 0.83, - "learning_rate": 1.4053915868470013e-06, - "loss": 0.8506, - "step": 4336 - }, - { - "epoch": 0.83, - "learning_rate": 1.402207405750765e-06, - "loss": 0.8647, - "step": 4337 - }, - { - "epoch": 0.83, - "learning_rate": 1.399026563938105e-06, - "loss": 0.8346, - "step": 4338 - }, - { - "epoch": 0.83, - "learning_rate": 1.3958490626444154e-06, - "loss": 0.9488, - "step": 4339 - }, - { - "epoch": 0.83, - "learning_rate": 1.3926749031038055e-06, - "loss": 0.9132, - "step": 4340 - }, - { - "epoch": 0.84, - "learning_rate": 1.3895040865490817e-06, - "loss": 0.8833, - "step": 4341 - }, - { - "epoch": 0.84, - "learning_rate": 1.3863366142117506e-06, - "loss": 0.8352, - "step": 4342 - }, - { - "epoch": 0.84, - "learning_rate": 1.383172487322023e-06, - "loss": 0.8065, - "step": 4343 - }, - { - "epoch": 0.84, - "learning_rate": 1.3800117071088104e-06, - "loss": 0.8293, - "step": 4344 - }, - { - "epoch": 0.84, - "learning_rate": 1.3768542747997215e-06, - "loss": 0.8284, - "step": 4345 - }, - { - "epoch": 0.84, - "learning_rate": 1.3737001916210713e-06, - "loss": 0.7935, - "step": 4346 - }, - { - "epoch": 0.84, - "learning_rate": 1.370549458797863e-06, - "loss": 0.8021, - "step": 4347 - }, - { - "epoch": 0.84, - "learning_rate": 1.3674020775538078e-06, - "loss": 0.8073, - "step": 4348 - }, - { - "epoch": 0.84, - "learning_rate": 1.3642580491113122e-06, - "loss": 0.8304, - "step": 4349 - }, - { - "epoch": 0.84, - "learning_rate": 1.3611173746914797e-06, - "loss": 0.8385, - "step": 4350 - }, - { - "epoch": 0.84, - "learning_rate": 1.3579800555141165e-06, - "loss": 0.813, - "step": 4351 - }, - { - "epoch": 0.84, - "learning_rate": 1.3548460927977158e-06, - "loss": 0.8934, - "step": 4352 - }, - { - "epoch": 0.84, - "learning_rate": 1.351715487759474e-06, - "loss": 0.8495, - "step": 4353 - }, - { - "epoch": 0.84, - "learning_rate": 1.3485882416152819e-06, - "loss": 0.769, - "step": 4354 - }, - { - "epoch": 0.84, - "learning_rate": 1.3454643555797276e-06, - "loss": 0.7679, - "step": 4355 - }, - { - "epoch": 0.84, - "learning_rate": 1.3423438308660929e-06, - "loss": 0.8845, - "step": 4356 - }, - { - "epoch": 0.84, - "learning_rate": 1.3392266686863508e-06, - "loss": 0.862, - "step": 4357 - }, - { - "epoch": 0.84, - "learning_rate": 1.3361128702511716e-06, - "loss": 0.8655, - "step": 4358 - }, - { - "epoch": 0.84, - "learning_rate": 1.3330024367699224e-06, - "loss": 0.7969, - "step": 4359 - }, - { - "epoch": 0.84, - "learning_rate": 1.3298953694506522e-06, - "loss": 0.8825, - "step": 4360 - }, - { - "epoch": 0.84, - "learning_rate": 1.3267916695001172e-06, - "loss": 0.782, - "step": 4361 - }, - { - "epoch": 0.84, - "learning_rate": 1.3236913381237592e-06, - "loss": 0.8854, - "step": 4362 - }, - { - "epoch": 0.84, - "learning_rate": 1.3205943765257057e-06, - "loss": 0.7946, - "step": 4363 - }, - { - "epoch": 0.84, - "learning_rate": 1.317500785908783e-06, - "loss": 0.7592, - "step": 4364 - }, - { - "epoch": 0.84, - "learning_rate": 1.31441056747451e-06, - "loss": 0.855, - "step": 4365 - }, - { - "epoch": 0.84, - "learning_rate": 1.3113237224230836e-06, - "loss": 0.9589, - "step": 4366 - }, - { - "epoch": 0.84, - "learning_rate": 1.3082402519534076e-06, - "loss": 0.8439, - "step": 4367 - }, - { - "epoch": 0.84, - "learning_rate": 1.3051601572630611e-06, - "loss": 0.8127, - "step": 4368 - }, - { - "epoch": 0.84, - "learning_rate": 1.3020834395483195e-06, - "loss": 0.789, - "step": 4369 - }, - { - "epoch": 0.84, - "learning_rate": 1.2990101000041445e-06, - "loss": 0.8429, - "step": 4370 - }, - { - "epoch": 0.84, - "learning_rate": 1.2959401398241844e-06, - "loss": 0.8383, - "step": 4371 - }, - { - "epoch": 0.84, - "learning_rate": 1.2928735602007768e-06, - "loss": 0.9096, - "step": 4372 - }, - { - "epoch": 0.84, - "learning_rate": 1.2898103623249458e-06, - "loss": 0.7264, - "step": 4373 - }, - { - "epoch": 0.84, - "learning_rate": 1.2867505473864029e-06, - "loss": 0.8375, - "step": 4374 - }, - { - "epoch": 0.84, - "learning_rate": 1.283694116573546e-06, - "loss": 0.8201, - "step": 4375 - }, - { - "epoch": 0.84, - "learning_rate": 1.2806410710734552e-06, - "loss": 0.8792, - "step": 4376 - }, - { - "epoch": 0.84, - "learning_rate": 1.2775914120718992e-06, - "loss": 0.8293, - "step": 4377 - }, - { - "epoch": 0.84, - "learning_rate": 1.2745451407533294e-06, - "loss": 0.8897, - "step": 4378 - }, - { - "epoch": 0.84, - "learning_rate": 1.2715022583008851e-06, - "loss": 0.8158, - "step": 4379 - }, - { - "epoch": 0.84, - "learning_rate": 1.2684627658963865e-06, - "loss": 0.8263, - "step": 4380 - }, - { - "epoch": 0.84, - "learning_rate": 1.265426664720334e-06, - "loss": 0.9262, - "step": 4381 - }, - { - "epoch": 0.84, - "learning_rate": 1.2623939559519161e-06, - "loss": 0.8785, - "step": 4382 - }, - { - "epoch": 0.84, - "learning_rate": 1.2593646407690051e-06, - "loss": 0.7306, - "step": 4383 - }, - { - "epoch": 0.84, - "learning_rate": 1.2563387203481447e-06, - "loss": 0.851, - "step": 4384 - }, - { - "epoch": 0.84, - "learning_rate": 1.2533161958645755e-06, - "loss": 0.9086, - "step": 4385 - }, - { - "epoch": 0.84, - "learning_rate": 1.2502970684922067e-06, - "loss": 0.8778, - "step": 4386 - }, - { - "epoch": 0.84, - "learning_rate": 1.2472813394036344e-06, - "loss": 0.8406, - "step": 4387 - }, - { - "epoch": 0.84, - "learning_rate": 1.2442690097701327e-06, - "loss": 0.8104, - "step": 4388 - }, - { - "epoch": 0.84, - "learning_rate": 1.2412600807616526e-06, - "loss": 0.796, - "step": 4389 - }, - { - "epoch": 0.84, - "learning_rate": 1.2382545535468316e-06, - "loss": 0.7997, - "step": 4390 - }, - { - "epoch": 0.84, - "learning_rate": 1.2352524292929823e-06, - "loss": 0.8498, - "step": 4391 - }, - { - "epoch": 0.84, - "learning_rate": 1.2322537091660912e-06, - "loss": 0.828, - "step": 4392 - }, - { - "epoch": 0.85, - "learning_rate": 1.22925839433083e-06, - "loss": 0.7569, - "step": 4393 - }, - { - "epoch": 0.85, - "learning_rate": 1.2262664859505434e-06, - "loss": 0.8584, - "step": 4394 - }, - { - "epoch": 0.85, - "learning_rate": 1.2232779851872511e-06, - "loss": 0.7341, - "step": 4395 - }, - { - "epoch": 0.85, - "learning_rate": 1.2202928932016588e-06, - "loss": 0.7411, - "step": 4396 - }, - { - "epoch": 0.85, - "learning_rate": 1.217311211153137e-06, - "loss": 0.7728, - "step": 4397 - }, - { - "epoch": 0.85, - "learning_rate": 1.2143329401997372e-06, - "loss": 0.8192, - "step": 4398 - }, - { - "epoch": 0.85, - "learning_rate": 1.2113580814981884e-06, - "loss": 0.774, - "step": 4399 - }, - { - "epoch": 0.85, - "learning_rate": 1.2083866362038865e-06, - "loss": 0.8879, - "step": 4400 - }, - { - "epoch": 0.85, - "learning_rate": 1.2054186054709105e-06, - "loss": 0.8801, - "step": 4401 - }, - { - "epoch": 0.85, - "learning_rate": 1.2024539904520072e-06, - "loss": 0.7116, - "step": 4402 - }, - { - "epoch": 0.85, - "learning_rate": 1.1994927922985999e-06, - "loss": 0.799, - "step": 4403 - }, - { - "epoch": 0.85, - "learning_rate": 1.1965350121607866e-06, - "loss": 0.8478, - "step": 4404 - }, - { - "epoch": 0.85, - "learning_rate": 1.1935806511873306e-06, - "loss": 0.7875, - "step": 4405 - }, - { - "epoch": 0.85, - "learning_rate": 1.1906297105256725e-06, - "loss": 0.8405, - "step": 4406 - }, - { - "epoch": 0.85, - "learning_rate": 1.187682191321925e-06, - "loss": 0.8778, - "step": 4407 - }, - { - "epoch": 0.85, - "learning_rate": 1.1847380947208697e-06, - "loss": 0.6952, - "step": 4408 - }, - { - "epoch": 0.85, - "learning_rate": 1.1817974218659621e-06, - "loss": 0.7061, - "step": 4409 - }, - { - "epoch": 0.85, - "learning_rate": 1.178860173899321e-06, - "loss": 0.8921, - "step": 4410 - }, - { - "epoch": 0.85, - "learning_rate": 1.1759263519617437e-06, - "loss": 0.9134, - "step": 4411 - }, - { - "epoch": 0.85, - "learning_rate": 1.172995957192693e-06, - "loss": 0.9203, - "step": 4412 - }, - { - "epoch": 0.85, - "learning_rate": 1.1700689907302953e-06, - "loss": 0.8025, - "step": 4413 - }, - { - "epoch": 0.85, - "learning_rate": 1.167145453711358e-06, - "loss": 0.8164, - "step": 4414 - }, - { - "epoch": 0.85, - "learning_rate": 1.1642253472713427e-06, - "loss": 0.7818, - "step": 4415 - }, - { - "epoch": 0.85, - "learning_rate": 1.161308672544389e-06, - "loss": 0.8161, - "step": 4416 - }, - { - "epoch": 0.85, - "learning_rate": 1.1583954306633004e-06, - "loss": 0.8291, - "step": 4417 - }, - { - "epoch": 0.85, - "learning_rate": 1.1554856227595435e-06, - "loss": 0.8547, - "step": 4418 - }, - { - "epoch": 0.85, - "learning_rate": 1.1525792499632526e-06, - "loss": 0.8075, - "step": 4419 - }, - { - "epoch": 0.85, - "learning_rate": 1.1496763134032363e-06, - "loss": 0.8629, - "step": 4420 - }, - { - "epoch": 0.85, - "learning_rate": 1.1467768142069546e-06, - "loss": 0.8998, - "step": 4421 - }, - { - "epoch": 0.85, - "learning_rate": 1.1438807535005437e-06, - "loss": 0.8767, - "step": 4422 - }, - { - "epoch": 0.85, - "learning_rate": 1.1409881324088013e-06, - "loss": 0.8456, - "step": 4423 - }, - { - "epoch": 0.85, - "learning_rate": 1.138098952055181e-06, - "loss": 0.8227, - "step": 4424 - }, - { - "epoch": 0.85, - "learning_rate": 1.1352132135618165e-06, - "loss": 0.7975, - "step": 4425 - }, - { - "epoch": 0.85, - "learning_rate": 1.13233091804949e-06, - "loss": 0.8445, - "step": 4426 - }, - { - "epoch": 0.85, - "learning_rate": 1.1294520666376518e-06, - "loss": 0.9219, - "step": 4427 - }, - { - "epoch": 0.85, - "learning_rate": 1.1265766604444172e-06, - "loss": 0.8384, - "step": 4428 - }, - { - "epoch": 0.85, - "learning_rate": 1.1237047005865576e-06, - "loss": 0.7807, - "step": 4429 - }, - { - "epoch": 0.85, - "learning_rate": 1.1208361881795116e-06, - "loss": 0.9027, - "step": 4430 - }, - { - "epoch": 0.85, - "learning_rate": 1.1179711243373736e-06, - "loss": 0.8421, - "step": 4431 - }, - { - "epoch": 0.85, - "learning_rate": 1.1151095101729047e-06, - "loss": 0.836, - "step": 4432 - }, - { - "epoch": 0.85, - "learning_rate": 1.1122513467975237e-06, - "loss": 0.9106, - "step": 4433 - }, - { - "epoch": 0.85, - "learning_rate": 1.1093966353213036e-06, - "loss": 0.8179, - "step": 4434 - }, - { - "epoch": 0.85, - "learning_rate": 1.1065453768529844e-06, - "loss": 0.9167, - "step": 4435 - }, - { - "epoch": 0.85, - "learning_rate": 1.103697572499961e-06, - "loss": 0.8681, - "step": 4436 - }, - { - "epoch": 0.85, - "learning_rate": 1.1008532233682878e-06, - "loss": 0.8292, - "step": 4437 - }, - { - "epoch": 0.85, - "learning_rate": 1.0980123305626812e-06, - "loss": 0.6738, - "step": 4438 - }, - { - "epoch": 0.85, - "learning_rate": 1.0951748951865048e-06, - "loss": 0.7946, - "step": 4439 - }, - { - "epoch": 0.85, - "learning_rate": 1.0923409183417887e-06, - "loss": 0.7111, - "step": 4440 - }, - { - "epoch": 0.85, - "learning_rate": 1.0895104011292202e-06, - "loss": 0.8231, - "step": 4441 - }, - { - "epoch": 0.85, - "learning_rate": 1.0866833446481317e-06, - "loss": 0.793, - "step": 4442 - }, - { - "epoch": 0.85, - "learning_rate": 1.0838597499965276e-06, - "loss": 0.8832, - "step": 4443 - }, - { - "epoch": 0.85, - "learning_rate": 1.0810396182710535e-06, - "loss": 0.8375, - "step": 4444 - }, - { - "epoch": 0.86, - "learning_rate": 1.0782229505670195e-06, - "loss": 0.8906, - "step": 4445 - }, - { - "epoch": 0.86, - "learning_rate": 1.0754097479783876e-06, - "loss": 0.816, - "step": 4446 - }, - { - "epoch": 0.86, - "learning_rate": 1.0726000115977696e-06, - "loss": 0.8741, - "step": 4447 - }, - { - "epoch": 0.86, - "learning_rate": 1.069793742516435e-06, - "loss": 0.8512, - "step": 4448 - }, - { - "epoch": 0.86, - "learning_rate": 1.066990941824312e-06, - "loss": 0.8199, - "step": 4449 - }, - { - "epoch": 0.86, - "learning_rate": 1.0641916106099691e-06, - "loss": 0.8388, - "step": 4450 - }, - { - "epoch": 0.86, - "learning_rate": 1.0613957499606388e-06, - "loss": 0.8468, - "step": 4451 - }, - { - "epoch": 0.86, - "learning_rate": 1.0586033609622004e-06, - "loss": 0.8977, - "step": 4452 - }, - { - "epoch": 0.86, - "learning_rate": 1.0558144446991836e-06, - "loss": 0.8064, - "step": 4453 - }, - { - "epoch": 0.86, - "learning_rate": 1.053029002254773e-06, - "loss": 0.8947, - "step": 4454 - }, - { - "epoch": 0.86, - "learning_rate": 1.0502470347108017e-06, - "loss": 0.8493, - "step": 4455 - }, - { - "epoch": 0.86, - "learning_rate": 1.0474685431477537e-06, - "loss": 0.7785, - "step": 4456 - }, - { - "epoch": 0.86, - "learning_rate": 1.0446935286447657e-06, - "loss": 0.8419, - "step": 4457 - }, - { - "epoch": 0.86, - "learning_rate": 1.0419219922796175e-06, - "loss": 0.784, - "step": 4458 - }, - { - "epoch": 0.86, - "learning_rate": 1.039153935128744e-06, - "loss": 0.8125, - "step": 4459 - }, - { - "epoch": 0.86, - "learning_rate": 1.0363893582672246e-06, - "loss": 0.8097, - "step": 4460 - }, - { - "epoch": 0.86, - "learning_rate": 1.033628262768792e-06, - "loss": 0.8791, - "step": 4461 - }, - { - "epoch": 0.86, - "learning_rate": 1.0308706497058252e-06, - "loss": 0.8583, - "step": 4462 - }, - { - "epoch": 0.86, - "learning_rate": 1.0281165201493437e-06, - "loss": 0.8371, - "step": 4463 - }, - { - "epoch": 0.86, - "learning_rate": 1.0253658751690232e-06, - "loss": 0.924, - "step": 4464 - }, - { - "epoch": 0.86, - "learning_rate": 1.0226187158331825e-06, - "loss": 0.6665, - "step": 4465 - }, - { - "epoch": 0.86, - "learning_rate": 1.0198750432087855e-06, - "loss": 0.8505, - "step": 4466 - }, - { - "epoch": 0.86, - "learning_rate": 1.017134858361446e-06, - "loss": 0.9197, - "step": 4467 - }, - { - "epoch": 0.86, - "learning_rate": 1.0143981623554155e-06, - "loss": 0.9474, - "step": 4468 - }, - { - "epoch": 0.86, - "learning_rate": 1.0116649562535984e-06, - "loss": 0.8265, - "step": 4469 - }, - { - "epoch": 0.86, - "learning_rate": 1.0089352411175424e-06, - "loss": 0.8456, - "step": 4470 - }, - { - "epoch": 0.86, - "learning_rate": 1.006209018007429e-06, - "loss": 0.8197, - "step": 4471 - }, - { - "epoch": 0.86, - "learning_rate": 1.0034862879821029e-06, - "loss": 0.8384, - "step": 4472 - }, - { - "epoch": 0.86, - "learning_rate": 1.0007670520990331e-06, - "loss": 0.8326, - "step": 4473 - }, - { - "epoch": 0.86, - "learning_rate": 9.98051311414342e-07, - "loss": 0.9024, - "step": 4474 - }, - { - "epoch": 0.86, - "learning_rate": 9.953390669827944e-07, - "loss": 0.8359, - "step": 4475 - }, - { - "epoch": 0.86, - "learning_rate": 9.926303198577913e-07, - "loss": 0.7781, - "step": 4476 - }, - { - "epoch": 0.86, - "learning_rate": 9.899250710913767e-07, - "loss": 0.7758, - "step": 4477 - }, - { - "epoch": 0.86, - "learning_rate": 9.872233217342463e-07, - "loss": 0.9121, - "step": 4478 - }, - { - "epoch": 0.86, - "learning_rate": 9.845250728357214e-07, - "loss": 0.8623, - "step": 4479 - }, - { - "epoch": 0.86, - "learning_rate": 9.818303254437723e-07, - "loss": 0.8892, - "step": 4480 - }, - { - "epoch": 0.86, - "learning_rate": 9.791390806050117e-07, - "loss": 0.9272, - "step": 4481 - }, - { - "epoch": 0.86, - "learning_rate": 9.764513393646812e-07, - "loss": 0.8372, - "step": 4482 - }, - { - "epoch": 0.86, - "learning_rate": 9.737671027666728e-07, - "loss": 0.8171, - "step": 4483 - }, - { - "epoch": 0.86, - "learning_rate": 9.710863718535135e-07, - "loss": 0.8778, - "step": 4484 - }, - { - "epoch": 0.86, - "learning_rate": 9.684091476663659e-07, - "loss": 0.8879, - "step": 4485 - }, - { - "epoch": 0.86, - "learning_rate": 9.657354312450363e-07, - "loss": 0.8287, - "step": 4486 - }, - { - "epoch": 0.86, - "learning_rate": 9.630652236279626e-07, - "loss": 0.7783, - "step": 4487 - }, - { - "epoch": 0.86, - "learning_rate": 9.603985258522219e-07, - "loss": 0.8533, - "step": 4488 - }, - { - "epoch": 0.86, - "learning_rate": 9.577353389535315e-07, - "loss": 0.7866, - "step": 4489 - }, - { - "epoch": 0.86, - "learning_rate": 9.550756639662417e-07, - "loss": 0.8617, - "step": 4490 - }, - { - "epoch": 0.86, - "learning_rate": 9.524195019233407e-07, - "loss": 0.8821, - "step": 4491 - }, - { - "epoch": 0.86, - "learning_rate": 9.497668538564475e-07, - "loss": 0.8541, - "step": 4492 - }, - { - "epoch": 0.86, - "learning_rate": 9.471177207958238e-07, - "loss": 0.711, - "step": 4493 - }, - { - "epoch": 0.86, - "learning_rate": 9.444721037703597e-07, - "loss": 0.8795, - "step": 4494 - }, - { - "epoch": 0.86, - "learning_rate": 9.418300038075845e-07, - "loss": 0.8745, - "step": 4495 - }, - { - "epoch": 0.86, - "learning_rate": 9.391914219336606e-07, - "loss": 0.8588, - "step": 4496 - }, - { - "epoch": 0.87, - "learning_rate": 9.365563591733784e-07, - "loss": 0.8204, - "step": 4497 - }, - { - "epoch": 0.87, - "learning_rate": 9.33924816550168e-07, - "loss": 0.9041, - "step": 4498 - }, - { - "epoch": 0.87, - "learning_rate": 9.31296795086093e-07, - "loss": 0.813, - "step": 4499 - }, - { - "epoch": 0.87, - "learning_rate": 9.286722958018391e-07, - "loss": 0.8294, - "step": 4500 - }, - { - "epoch": 0.87, - "learning_rate": 9.260513197167398e-07, - "loss": 0.8519, - "step": 4501 - }, - { - "epoch": 0.87, - "learning_rate": 9.234338678487509e-07, - "loss": 0.8475, - "step": 4502 - }, - { - "epoch": 0.87, - "learning_rate": 9.208199412144559e-07, - "loss": 0.8842, - "step": 4503 - }, - { - "epoch": 0.87, - "learning_rate": 9.182095408290781e-07, - "loss": 0.8476, - "step": 4504 - }, - { - "epoch": 0.87, - "learning_rate": 9.156026677064633e-07, - "loss": 0.8609, - "step": 4505 - }, - { - "epoch": 0.87, - "learning_rate": 9.129993228590917e-07, - "loss": 0.8418, - "step": 4506 - }, - { - "epoch": 0.87, - "learning_rate": 9.103995072980765e-07, - "loss": 0.6609, - "step": 4507 - }, - { - "epoch": 0.87, - "learning_rate": 9.078032220331523e-07, - "loss": 0.8999, - "step": 4508 - }, - { - "epoch": 0.87, - "learning_rate": 9.052104680726859e-07, - "loss": 0.8298, - "step": 4509 - }, - { - "epoch": 0.87, - "learning_rate": 9.026212464236772e-07, - "loss": 0.8273, - "step": 4510 - }, - { - "epoch": 0.87, - "learning_rate": 9.000355580917464e-07, - "loss": 0.8546, - "step": 4511 - }, - { - "epoch": 0.87, - "learning_rate": 8.974534040811444e-07, - "loss": 0.8479, - "step": 4512 - }, - { - "epoch": 0.87, - "learning_rate": 8.948747853947526e-07, - "loss": 0.7852, - "step": 4513 - }, - { - "epoch": 0.87, - "learning_rate": 8.922997030340752e-07, - "loss": 0.9148, - "step": 4514 - }, - { - "epoch": 0.87, - "learning_rate": 8.897281579992467e-07, - "loss": 0.8106, - "step": 4515 - }, - { - "epoch": 0.87, - "learning_rate": 8.871601512890238e-07, - "loss": 0.7778, - "step": 4516 - }, - { - "epoch": 0.87, - "learning_rate": 8.845956839007897e-07, - "loss": 0.7944, - "step": 4517 - }, - { - "epoch": 0.87, - "learning_rate": 8.820347568305543e-07, - "loss": 0.8276, - "step": 4518 - }, - { - "epoch": 0.87, - "learning_rate": 8.794773710729543e-07, - "loss": 0.7686, - "step": 4519 - }, - { - "epoch": 0.87, - "learning_rate": 8.769235276212496e-07, - "loss": 0.7786, - "step": 4520 - }, - { - "epoch": 0.87, - "learning_rate": 8.743732274673189e-07, - "loss": 0.7762, - "step": 4521 - }, - { - "epoch": 0.87, - "learning_rate": 8.718264716016722e-07, - "loss": 0.7963, - "step": 4522 - }, - { - "epoch": 0.87, - "learning_rate": 8.692832610134428e-07, - "loss": 0.8422, - "step": 4523 - }, - { - "epoch": 0.87, - "learning_rate": 8.66743596690377e-07, - "loss": 0.8513, - "step": 4524 - }, - { - "epoch": 0.87, - "learning_rate": 8.642074796188594e-07, - "loss": 0.7935, - "step": 4525 - }, - { - "epoch": 0.87, - "learning_rate": 8.61674910783884e-07, - "loss": 0.7876, - "step": 4526 - }, - { - "epoch": 0.87, - "learning_rate": 8.59145891169072e-07, - "loss": 0.8356, - "step": 4527 - }, - { - "epoch": 0.87, - "learning_rate": 8.566204217566664e-07, - "loss": 0.7975, - "step": 4528 - }, - { - "epoch": 0.87, - "learning_rate": 8.540985035275273e-07, - "loss": 0.8293, - "step": 4529 - }, - { - "epoch": 0.87, - "learning_rate": 8.515801374611432e-07, - "loss": 0.826, - "step": 4530 - }, - { - "epoch": 0.87, - "learning_rate": 8.490653245356184e-07, - "loss": 0.7823, - "step": 4531 - }, - { - "epoch": 0.87, - "learning_rate": 8.465540657276728e-07, - "loss": 0.8675, - "step": 4532 - }, - { - "epoch": 0.87, - "learning_rate": 8.44046362012656e-07, - "loss": 0.8393, - "step": 4533 - }, - { - "epoch": 0.87, - "learning_rate": 8.415422143645247e-07, - "loss": 0.95, - "step": 4534 - }, - { - "epoch": 0.87, - "learning_rate": 8.390416237558641e-07, - "loss": 0.8567, - "step": 4535 - }, - { - "epoch": 0.87, - "learning_rate": 8.365445911578785e-07, - "loss": 0.7825, - "step": 4536 - }, - { - "epoch": 0.87, - "learning_rate": 8.340511175403809e-07, - "loss": 0.8656, - "step": 4537 - }, - { - "epoch": 0.87, - "learning_rate": 8.315612038718101e-07, - "loss": 0.8533, - "step": 4538 - }, - { - "epoch": 0.87, - "learning_rate": 8.290748511192214e-07, - "loss": 0.8786, - "step": 4539 - }, - { - "epoch": 0.87, - "learning_rate": 8.265920602482825e-07, - "loss": 0.8705, - "step": 4540 - }, - { - "epoch": 0.87, - "learning_rate": 8.241128322232816e-07, - "loss": 0.828, - "step": 4541 - }, - { - "epoch": 0.87, - "learning_rate": 8.216371680071244e-07, - "loss": 0.844, - "step": 4542 - }, - { - "epoch": 0.87, - "learning_rate": 8.191650685613273e-07, - "loss": 0.8589, - "step": 4543 - }, - { - "epoch": 0.87, - "learning_rate": 8.166965348460298e-07, - "loss": 0.7925, - "step": 4544 - }, - { - "epoch": 0.87, - "learning_rate": 8.142315678199764e-07, - "loss": 0.8188, - "step": 4545 - }, - { - "epoch": 0.87, - "learning_rate": 8.117701684405343e-07, - "loss": 0.7775, - "step": 4546 - }, - { - "epoch": 0.87, - "learning_rate": 8.093123376636836e-07, - "loss": 0.8167, - "step": 4547 - }, - { - "epoch": 0.87, - "learning_rate": 8.06858076444017e-07, - "loss": 0.8278, - "step": 4548 - }, - { - "epoch": 0.88, - "learning_rate": 8.044073857347423e-07, - "loss": 0.8809, - "step": 4549 - }, - { - "epoch": 0.88, - "learning_rate": 8.019602664876758e-07, - "loss": 0.8659, - "step": 4550 - }, - { - "epoch": 0.88, - "learning_rate": 7.995167196532527e-07, - "loss": 0.8979, - "step": 4551 - }, - { - "epoch": 0.88, - "learning_rate": 7.970767461805218e-07, - "loss": 0.8312, - "step": 4552 - }, - { - "epoch": 0.88, - "learning_rate": 7.946403470171326e-07, - "loss": 0.8301, - "step": 4553 - }, - { - "epoch": 0.88, - "learning_rate": 7.922075231093628e-07, - "loss": 0.8627, - "step": 4554 - }, - { - "epoch": 0.88, - "learning_rate": 7.897782754020889e-07, - "loss": 0.8778, - "step": 4555 - }, - { - "epoch": 0.88, - "learning_rate": 7.873526048388025e-07, - "loss": 0.799, - "step": 4556 - }, - { - "epoch": 0.88, - "learning_rate": 7.849305123616091e-07, - "loss": 0.9133, - "step": 4557 - }, - { - "epoch": 0.88, - "learning_rate": 7.825119989112173e-07, - "loss": 0.8178, - "step": 4558 - }, - { - "epoch": 0.88, - "learning_rate": 7.800970654269513e-07, - "loss": 0.8483, - "step": 4559 - }, - { - "epoch": 0.88, - "learning_rate": 7.776857128467464e-07, - "loss": 0.8182, - "step": 4560 - }, - { - "epoch": 0.88, - "learning_rate": 7.75277942107141e-07, - "loss": 0.9151, - "step": 4561 - }, - { - "epoch": 0.88, - "learning_rate": 7.728737541432862e-07, - "loss": 0.7403, - "step": 4562 - }, - { - "epoch": 0.88, - "learning_rate": 7.704731498889428e-07, - "loss": 0.7566, - "step": 4563 - }, - { - "epoch": 0.88, - "learning_rate": 7.680761302764727e-07, - "loss": 0.8678, - "step": 4564 - }, - { - "epoch": 0.88, - "learning_rate": 7.65682696236858e-07, - "loss": 0.8799, - "step": 4565 - }, - { - "epoch": 0.88, - "learning_rate": 7.632928486996749e-07, - "loss": 0.8632, - "step": 4566 - }, - { - "epoch": 0.88, - "learning_rate": 7.609065885931155e-07, - "loss": 0.8817, - "step": 4567 - }, - { - "epoch": 0.88, - "learning_rate": 7.585239168439762e-07, - "loss": 0.7023, - "step": 4568 - }, - { - "epoch": 0.88, - "learning_rate": 7.561448343776567e-07, - "loss": 0.8163, - "step": 4569 - }, - { - "epoch": 0.88, - "learning_rate": 7.537693421181658e-07, - "loss": 0.8215, - "step": 4570 - }, - { - "epoch": 0.88, - "learning_rate": 7.513974409881186e-07, - "loss": 0.8812, - "step": 4571 - }, - { - "epoch": 0.88, - "learning_rate": 7.49029131908734e-07, - "loss": 0.9046, - "step": 4572 - }, - { - "epoch": 0.88, - "learning_rate": 7.466644157998371e-07, - "loss": 0.8571, - "step": 4573 - }, - { - "epoch": 0.88, - "learning_rate": 7.443032935798533e-07, - "loss": 0.8607, - "step": 4574 - }, - { - "epoch": 0.88, - "learning_rate": 7.419457661658169e-07, - "loss": 0.8151, - "step": 4575 - }, - { - "epoch": 0.88, - "learning_rate": 7.395918344733644e-07, - "loss": 0.8329, - "step": 4576 - }, - { - "epoch": 0.88, - "learning_rate": 7.372414994167354e-07, - "loss": 0.8756, - "step": 4577 - }, - { - "epoch": 0.88, - "learning_rate": 7.348947619087754e-07, - "loss": 0.8925, - "step": 4578 - }, - { - "epoch": 0.88, - "learning_rate": 7.325516228609264e-07, - "loss": 0.8743, - "step": 4579 - }, - { - "epoch": 0.88, - "learning_rate": 7.302120831832382e-07, - "loss": 0.7751, - "step": 4580 - }, - { - "epoch": 0.88, - "learning_rate": 7.278761437843629e-07, - "loss": 0.821, - "step": 4581 - }, - { - "epoch": 0.88, - "learning_rate": 7.255438055715469e-07, - "loss": 0.9183, - "step": 4582 - }, - { - "epoch": 0.88, - "learning_rate": 7.232150694506512e-07, - "loss": 0.8538, - "step": 4583 - }, - { - "epoch": 0.88, - "learning_rate": 7.208899363261234e-07, - "loss": 0.7672, - "step": 4584 - }, - { - "epoch": 0.88, - "learning_rate": 7.185684071010224e-07, - "loss": 0.8216, - "step": 4585 - }, - { - "epoch": 0.88, - "learning_rate": 7.162504826770033e-07, - "loss": 0.8953, - "step": 4586 - }, - { - "epoch": 0.88, - "learning_rate": 7.139361639543185e-07, - "loss": 0.8988, - "step": 4587 - }, - { - "epoch": 0.88, - "learning_rate": 7.116254518318222e-07, - "loss": 0.9298, - "step": 4588 - }, - { - "epoch": 0.88, - "learning_rate": 7.093183472069753e-07, - "loss": 0.8585, - "step": 4589 - }, - { - "epoch": 0.88, - "learning_rate": 7.070148509758223e-07, - "loss": 0.8339, - "step": 4590 - }, - { - "epoch": 0.88, - "learning_rate": 7.047149640330197e-07, - "loss": 0.9072, - "step": 4591 - }, - { - "epoch": 0.88, - "learning_rate": 7.024186872718164e-07, - "loss": 0.7736, - "step": 4592 - }, - { - "epoch": 0.88, - "learning_rate": 7.001260215840567e-07, - "loss": 0.8242, - "step": 4593 - }, - { - "epoch": 0.88, - "learning_rate": 6.978369678601892e-07, - "loss": 0.8713, - "step": 4594 - }, - { - "epoch": 0.88, - "learning_rate": 6.955515269892533e-07, - "loss": 0.814, - "step": 4595 - }, - { - "epoch": 0.88, - "learning_rate": 6.932696998588895e-07, - "loss": 0.868, - "step": 4596 - }, - { - "epoch": 0.88, - "learning_rate": 6.909914873553347e-07, - "loss": 0.8419, - "step": 4597 - }, - { - "epoch": 0.88, - "learning_rate": 6.887168903634178e-07, - "loss": 0.7485, - "step": 4598 - }, - { - "epoch": 0.88, - "learning_rate": 6.864459097665654e-07, - "loss": 0.8504, - "step": 4599 - }, - { - "epoch": 0.88, - "learning_rate": 6.84178546446802e-07, - "loss": 0.8054, - "step": 4600 - }, - { - "epoch": 0.89, - "learning_rate": 6.819148012847454e-07, - "loss": 0.8461, - "step": 4601 - }, - { - "epoch": 0.89, - "learning_rate": 6.796546751596089e-07, - "loss": 0.7001, - "step": 4602 - }, - { - "epoch": 0.89, - "learning_rate": 6.77398168949196e-07, - "loss": 0.8427, - "step": 4603 - }, - { - "epoch": 0.89, - "learning_rate": 6.751452835299111e-07, - "loss": 0.8193, - "step": 4604 - }, - { - "epoch": 0.89, - "learning_rate": 6.728960197767475e-07, - "loss": 0.868, - "step": 4605 - }, - { - "epoch": 0.89, - "learning_rate": 6.706503785632934e-07, - "loss": 0.8242, - "step": 4606 - }, - { - "epoch": 0.89, - "learning_rate": 6.68408360761732e-07, - "loss": 0.8257, - "step": 4607 - }, - { - "epoch": 0.89, - "learning_rate": 6.661699672428334e-07, - "loss": 0.8182, - "step": 4608 - }, - { - "epoch": 0.89, - "learning_rate": 6.639351988759657e-07, - "loss": 0.729, - "step": 4609 - }, - { - "epoch": 0.89, - "learning_rate": 6.6170405652909e-07, - "loss": 0.8665, - "step": 4610 - }, - { - "epoch": 0.89, - "learning_rate": 6.594765410687487e-07, - "loss": 0.8486, - "step": 4611 - }, - { - "epoch": 0.89, - "learning_rate": 6.57252653360092e-07, - "loss": 0.8438, - "step": 4612 - }, - { - "epoch": 0.89, - "learning_rate": 6.550323942668469e-07, - "loss": 0.7427, - "step": 4613 - }, - { - "epoch": 0.89, - "learning_rate": 6.528157646513378e-07, - "loss": 0.7823, - "step": 4614 - }, - { - "epoch": 0.89, - "learning_rate": 6.506027653744796e-07, - "loss": 0.8706, - "step": 4615 - }, - { - "epoch": 0.89, - "learning_rate": 6.483933972957734e-07, - "loss": 0.8574, - "step": 4616 - }, - { - "epoch": 0.89, - "learning_rate": 6.461876612733109e-07, - "loss": 0.8076, - "step": 4617 - }, - { - "epoch": 0.89, - "learning_rate": 6.4398555816378e-07, - "loss": 0.8337, - "step": 4618 - }, - { - "epoch": 0.89, - "learning_rate": 6.417870888224476e-07, - "loss": 0.8342, - "step": 4619 - }, - { - "epoch": 0.89, - "learning_rate": 6.395922541031741e-07, - "loss": 0.8932, - "step": 4620 - }, - { - "epoch": 0.89, - "learning_rate": 6.374010548584119e-07, - "loss": 0.8151, - "step": 4621 - }, - { - "epoch": 0.89, - "learning_rate": 6.352134919391928e-07, - "loss": 0.8873, - "step": 4622 - }, - { - "epoch": 0.89, - "learning_rate": 6.330295661951436e-07, - "loss": 0.7796, - "step": 4623 - }, - { - "epoch": 0.89, - "learning_rate": 6.308492784744746e-07, - "loss": 0.8317, - "step": 4624 - }, - { - "epoch": 0.89, - "learning_rate": 6.286726296239854e-07, - "loss": 0.8205, - "step": 4625 - }, - { - "epoch": 0.89, - "learning_rate": 6.264996204890628e-07, - "loss": 0.8747, - "step": 4626 - }, - { - "epoch": 0.89, - "learning_rate": 6.24330251913674e-07, - "loss": 0.8933, - "step": 4627 - }, - { - "epoch": 0.89, - "learning_rate": 6.221645247403807e-07, - "loss": 0.8485, - "step": 4628 - }, - { - "epoch": 0.89, - "learning_rate": 6.200024398103255e-07, - "loss": 0.8336, - "step": 4629 - }, - { - "epoch": 0.89, - "learning_rate": 6.178439979632367e-07, - "loss": 0.6981, - "step": 4630 - }, - { - "epoch": 0.89, - "learning_rate": 6.156892000374293e-07, - "loss": 0.9058, - "step": 4631 - }, - { - "epoch": 0.89, - "learning_rate": 6.135380468698004e-07, - "loss": 0.823, - "step": 4632 - }, - { - "epoch": 0.89, - "learning_rate": 6.113905392958342e-07, - "loss": 0.8248, - "step": 4633 - }, - { - "epoch": 0.89, - "learning_rate": 6.092466781495976e-07, - "loss": 0.9123, - "step": 4634 - }, - { - "epoch": 0.89, - "learning_rate": 6.071064642637404e-07, - "loss": 0.8421, - "step": 4635 - }, - { - "epoch": 0.89, - "learning_rate": 6.049698984695007e-07, - "loss": 0.6918, - "step": 4636 - }, - { - "epoch": 0.89, - "learning_rate": 6.028369815966917e-07, - "loss": 0.8114, - "step": 4637 - }, - { - "epoch": 0.89, - "learning_rate": 6.007077144737161e-07, - "loss": 0.8417, - "step": 4638 - }, - { - "epoch": 0.89, - "learning_rate": 5.985820979275569e-07, - "loss": 0.8579, - "step": 4639 - }, - { - "epoch": 0.89, - "learning_rate": 5.96460132783776e-07, - "loss": 0.8188, - "step": 4640 - }, - { - "epoch": 0.89, - "learning_rate": 5.943418198665251e-07, - "loss": 0.7971, - "step": 4641 - }, - { - "epoch": 0.89, - "learning_rate": 5.922271599985286e-07, - "loss": 0.8457, - "step": 4642 - }, - { - "epoch": 0.89, - "learning_rate": 5.901161540010969e-07, - "loss": 0.8481, - "step": 4643 - }, - { - "epoch": 0.89, - "learning_rate": 5.880088026941233e-07, - "loss": 0.8087, - "step": 4644 - }, - { - "epoch": 0.89, - "learning_rate": 5.859051068960741e-07, - "loss": 0.8306, - "step": 4645 - }, - { - "epoch": 0.89, - "learning_rate": 5.838050674240025e-07, - "loss": 0.8285, - "step": 4646 - }, - { - "epoch": 0.89, - "learning_rate": 5.817086850935416e-07, - "loss": 0.8327, - "step": 4647 - }, - { - "epoch": 0.89, - "learning_rate": 5.796159607189001e-07, - "loss": 0.8074, - "step": 4648 - }, - { - "epoch": 0.89, - "learning_rate": 5.775268951128676e-07, - "loss": 0.8058, - "step": 4649 - }, - { - "epoch": 0.89, - "learning_rate": 5.754414890868154e-07, - "loss": 0.8044, - "step": 4650 - }, - { - "epoch": 0.89, - "learning_rate": 5.733597434506877e-07, - "loss": 0.9057, - "step": 4651 - }, - { - "epoch": 0.9, - "learning_rate": 5.712816590130133e-07, - "loss": 0.8372, - "step": 4652 - }, - { - "epoch": 0.9, - "learning_rate": 5.692072365808954e-07, - "loss": 0.9003, - "step": 4653 - }, - { - "epoch": 0.9, - "learning_rate": 5.671364769600162e-07, - "loss": 0.8879, - "step": 4654 - }, - { - "epoch": 0.9, - "learning_rate": 5.650693809546348e-07, - "loss": 0.8443, - "step": 4655 - }, - { - "epoch": 0.9, - "learning_rate": 5.630059493675866e-07, - "loss": 0.8647, - "step": 4656 - }, - { - "epoch": 0.9, - "learning_rate": 5.60946183000285e-07, - "loss": 0.8171, - "step": 4657 - }, - { - "epoch": 0.9, - "learning_rate": 5.5889008265272e-07, - "loss": 0.8511, - "step": 4658 - }, - { - "epoch": 0.9, - "learning_rate": 5.568376491234562e-07, - "loss": 0.8319, - "step": 4659 - }, - { - "epoch": 0.9, - "learning_rate": 5.547888832096382e-07, - "loss": 0.7998, - "step": 4660 - }, - { - "epoch": 0.9, - "learning_rate": 5.527437857069784e-07, - "loss": 0.6558, - "step": 4661 - }, - { - "epoch": 0.9, - "learning_rate": 5.507023574097725e-07, - "loss": 0.8431, - "step": 4662 - }, - { - "epoch": 0.9, - "learning_rate": 5.486645991108875e-07, - "loss": 0.8609, - "step": 4663 - }, - { - "epoch": 0.9, - "learning_rate": 5.466305116017623e-07, - "loss": 0.8629, - "step": 4664 - }, - { - "epoch": 0.9, - "learning_rate": 5.446000956724174e-07, - "loss": 0.8608, - "step": 4665 - }, - { - "epoch": 0.9, - "learning_rate": 5.425733521114396e-07, - "loss": 0.8141, - "step": 4666 - }, - { - "epoch": 0.9, - "learning_rate": 5.405502817059937e-07, - "loss": 0.8628, - "step": 4667 - }, - { - "epoch": 0.9, - "learning_rate": 5.385308852418191e-07, - "loss": 0.8761, - "step": 4668 - }, - { - "epoch": 0.9, - "learning_rate": 5.365151635032218e-07, - "loss": 0.7737, - "step": 4669 - }, - { - "epoch": 0.9, - "learning_rate": 5.345031172730875e-07, - "loss": 0.846, - "step": 4670 - }, - { - "epoch": 0.9, - "learning_rate": 5.324947473328735e-07, - "loss": 0.7912, - "step": 4671 - }, - { - "epoch": 0.9, - "learning_rate": 5.304900544626046e-07, - "loss": 0.9209, - "step": 4672 - }, - { - "epoch": 0.9, - "learning_rate": 5.284890394408826e-07, - "loss": 0.741, - "step": 4673 - }, - { - "epoch": 0.9, - "learning_rate": 5.264917030448757e-07, - "loss": 0.798, - "step": 4674 - }, - { - "epoch": 0.9, - "learning_rate": 5.244980460503268e-07, - "loss": 0.8502, - "step": 4675 - }, - { - "epoch": 0.9, - "learning_rate": 5.225080692315532e-07, - "loss": 0.8668, - "step": 4676 - }, - { - "epoch": 0.9, - "learning_rate": 5.205217733614353e-07, - "loss": 0.7638, - "step": 4677 - }, - { - "epoch": 0.9, - "learning_rate": 5.185391592114286e-07, - "loss": 0.8182, - "step": 4678 - }, - { - "epoch": 0.9, - "learning_rate": 5.165602275515592e-07, - "loss": 0.8207, - "step": 4679 - }, - { - "epoch": 0.9, - "learning_rate": 5.145849791504187e-07, - "loss": 0.9076, - "step": 4680 - }, - { - "epoch": 0.9, - "learning_rate": 5.126134147751716e-07, - "loss": 0.826, - "step": 4681 - }, - { - "epoch": 0.9, - "learning_rate": 5.106455351915507e-07, - "loss": 0.749, - "step": 4682 - }, - { - "epoch": 0.9, - "learning_rate": 5.086813411638581e-07, - "loss": 0.7449, - "step": 4683 - }, - { - "epoch": 0.9, - "learning_rate": 5.067208334549656e-07, - "loss": 0.6836, - "step": 4684 - }, - { - "epoch": 0.9, - "learning_rate": 5.047640128263087e-07, - "loss": 0.8223, - "step": 4685 - }, - { - "epoch": 0.9, - "learning_rate": 5.028108800378961e-07, - "loss": 0.6643, - "step": 4686 - }, - { - "epoch": 0.9, - "learning_rate": 5.008614358483021e-07, - "loss": 0.7654, - "step": 4687 - }, - { - "epoch": 0.9, - "learning_rate": 4.989156810146667e-07, - "loss": 0.83, - "step": 4688 - }, - { - "epoch": 0.9, - "learning_rate": 4.969736162927019e-07, - "loss": 0.7829, - "step": 4689 - }, - { - "epoch": 0.9, - "learning_rate": 4.9503524243668e-07, - "loss": 0.6952, - "step": 4690 - }, - { - "epoch": 0.9, - "learning_rate": 4.931005601994432e-07, - "loss": 0.8326, - "step": 4691 - }, - { - "epoch": 0.9, - "learning_rate": 4.911695703324038e-07, - "loss": 0.8038, - "step": 4692 - }, - { - "epoch": 0.9, - "learning_rate": 4.892422735855284e-07, - "loss": 0.816, - "step": 4693 - }, - { - "epoch": 0.9, - "learning_rate": 4.873186707073663e-07, - "loss": 0.8563, - "step": 4694 - }, - { - "epoch": 0.9, - "learning_rate": 4.853987624450151e-07, - "loss": 0.8348, - "step": 4695 - }, - { - "epoch": 0.9, - "learning_rate": 4.834825495441475e-07, - "loss": 0.8654, - "step": 4696 - }, - { - "epoch": 0.9, - "learning_rate": 4.815700327490014e-07, - "loss": 0.8267, - "step": 4697 - }, - { - "epoch": 0.9, - "learning_rate": 4.796612128023726e-07, - "loss": 0.8574, - "step": 4698 - }, - { - "epoch": 0.9, - "learning_rate": 4.777560904456236e-07, - "loss": 0.844, - "step": 4699 - }, - { - "epoch": 0.9, - "learning_rate": 4.7585466641868696e-07, - "loss": 0.8238, - "step": 4700 - }, - { - "epoch": 0.9, - "learning_rate": 4.7395694146004976e-07, - "loss": 0.8955, - "step": 4701 - }, - { - "epoch": 0.9, - "learning_rate": 4.7206291630677024e-07, - "loss": 0.6771, - "step": 4702 - }, - { - "epoch": 0.9, - "learning_rate": 4.7017259169446104e-07, - "loss": 0.8494, - "step": 4703 - }, - { - "epoch": 0.91, - "learning_rate": 4.6828596835730487e-07, - "loss": 0.877, - "step": 4704 - }, - { - "epoch": 0.91, - "learning_rate": 4.664030470280467e-07, - "loss": 0.7805, - "step": 4705 - }, - { - "epoch": 0.91, - "learning_rate": 4.645238284379883e-07, - "loss": 0.7002, - "step": 4706 - }, - { - "epoch": 0.91, - "learning_rate": 4.626483133169968e-07, - "loss": 0.8064, - "step": 4707 - }, - { - "epoch": 0.91, - "learning_rate": 4.60776502393504e-07, - "loss": 0.8134, - "step": 4708 - }, - { - "epoch": 0.91, - "learning_rate": 4.5890839639449514e-07, - "loss": 0.777, - "step": 4709 - }, - { - "epoch": 0.91, - "learning_rate": 4.5704399604552417e-07, - "loss": 0.6637, - "step": 4710 - }, - { - "epoch": 0.91, - "learning_rate": 4.551833020707008e-07, - "loss": 0.8582, - "step": 4711 - }, - { - "epoch": 0.91, - "learning_rate": 4.533263151926981e-07, - "loss": 0.8728, - "step": 4712 - }, - { - "epoch": 0.91, - "learning_rate": 4.514730361327502e-07, - "loss": 0.8865, - "step": 4713 - }, - { - "epoch": 0.91, - "learning_rate": 4.4962346561064574e-07, - "loss": 0.8027, - "step": 4714 - }, - { - "epoch": 0.91, - "learning_rate": 4.4777760434473796e-07, - "loss": 0.877, - "step": 4715 - }, - { - "epoch": 0.91, - "learning_rate": 4.4593545305193774e-07, - "loss": 0.8486, - "step": 4716 - }, - { - "epoch": 0.91, - "learning_rate": 4.440970124477173e-07, - "loss": 0.6736, - "step": 4717 - }, - { - "epoch": 0.91, - "learning_rate": 4.4226228324610544e-07, - "loss": 0.8838, - "step": 4718 - }, - { - "epoch": 0.91, - "learning_rate": 4.404312661596877e-07, - "loss": 0.8202, - "step": 4719 - }, - { - "epoch": 0.91, - "learning_rate": 4.386039618996119e-07, - "loss": 0.8905, - "step": 4720 - }, - { - "epoch": 0.91, - "learning_rate": 4.3678037117558135e-07, - "loss": 0.7667, - "step": 4721 - }, - { - "epoch": 0.91, - "learning_rate": 4.349604946958563e-07, - "loss": 0.7489, - "step": 4722 - }, - { - "epoch": 0.91, - "learning_rate": 4.331443331672591e-07, - "loss": 0.9196, - "step": 4723 - }, - { - "epoch": 0.91, - "learning_rate": 4.313318872951633e-07, - "loss": 0.8102, - "step": 4724 - }, - { - "epoch": 0.91, - "learning_rate": 4.295231577835024e-07, - "loss": 0.8133, - "step": 4725 - }, - { - "epoch": 0.91, - "learning_rate": 4.2771814533476787e-07, - "loss": 0.8176, - "step": 4726 - }, - { - "epoch": 0.91, - "learning_rate": 4.2591685065000223e-07, - "loss": 0.8544, - "step": 4727 - }, - { - "epoch": 0.91, - "learning_rate": 4.241192744288092e-07, - "loss": 0.9002, - "step": 4728 - }, - { - "epoch": 0.91, - "learning_rate": 4.223254173693492e-07, - "loss": 0.8282, - "step": 4729 - }, - { - "epoch": 0.91, - "learning_rate": 4.2053528016833267e-07, - "loss": 0.7877, - "step": 4730 - }, - { - "epoch": 0.91, - "learning_rate": 4.1874886352103015e-07, - "loss": 0.9181, - "step": 4731 - }, - { - "epoch": 0.91, - "learning_rate": 4.1696616812126333e-07, - "loss": 0.8534, - "step": 4732 - }, - { - "epoch": 0.91, - "learning_rate": 4.1518719466141165e-07, - "loss": 0.8084, - "step": 4733 - }, - { - "epoch": 0.91, - "learning_rate": 4.13411943832408e-07, - "loss": 0.8571, - "step": 4734 - }, - { - "epoch": 0.91, - "learning_rate": 4.116404163237386e-07, - "loss": 0.8528, - "step": 4735 - }, - { - "epoch": 0.91, - "learning_rate": 4.0987261282344425e-07, - "loss": 0.8522, - "step": 4736 - }, - { - "epoch": 0.91, - "learning_rate": 4.081085340181223e-07, - "loss": 0.9021, - "step": 4737 - }, - { - "epoch": 0.91, - "learning_rate": 4.06348180592917e-07, - "loss": 0.711, - "step": 4738 - }, - { - "epoch": 0.91, - "learning_rate": 4.0459155323153034e-07, - "loss": 0.824, - "step": 4739 - }, - { - "epoch": 0.91, - "learning_rate": 4.028386526162176e-07, - "loss": 0.8861, - "step": 4740 - }, - { - "epoch": 0.91, - "learning_rate": 4.010894794277831e-07, - "loss": 0.812, - "step": 4741 - }, - { - "epoch": 0.91, - "learning_rate": 3.993440343455879e-07, - "loss": 0.8844, - "step": 4742 - }, - { - "epoch": 0.91, - "learning_rate": 3.976023180475397e-07, - "loss": 0.7791, - "step": 4743 - }, - { - "epoch": 0.91, - "learning_rate": 3.95864331210104e-07, - "loss": 0.8377, - "step": 4744 - }, - { - "epoch": 0.91, - "learning_rate": 3.941300745082932e-07, - "loss": 0.7904, - "step": 4745 - }, - { - "epoch": 0.91, - "learning_rate": 3.9239954861567177e-07, - "loss": 0.8927, - "step": 4746 - }, - { - "epoch": 0.91, - "learning_rate": 3.906727542043598e-07, - "loss": 0.8656, - "step": 4747 - }, - { - "epoch": 0.91, - "learning_rate": 3.8894969194502083e-07, - "loss": 0.7256, - "step": 4748 - }, - { - "epoch": 0.91, - "learning_rate": 3.87230362506873e-07, - "loss": 0.8797, - "step": 4749 - }, - { - "epoch": 0.91, - "learning_rate": 3.8551476655768527e-07, - "loss": 0.8325, - "step": 4750 - }, - { - "epoch": 0.91, - "learning_rate": 3.8380290476377255e-07, - "loss": 0.8941, - "step": 4751 - }, - { - "epoch": 0.91, - "learning_rate": 3.8209477779000637e-07, - "loss": 0.8468, - "step": 4752 - }, - { - "epoch": 0.91, - "learning_rate": 3.803903862998004e-07, - "loss": 0.7241, - "step": 4753 - }, - { - "epoch": 0.91, - "learning_rate": 3.7868973095512185e-07, - "loss": 0.8131, - "step": 4754 - }, - { - "epoch": 0.91, - "learning_rate": 3.7699281241648565e-07, - "loss": 0.848, - "step": 4755 - }, - { - "epoch": 0.92, - "learning_rate": 3.7529963134295466e-07, - "loss": 0.8405, - "step": 4756 - }, - { - "epoch": 0.92, - "learning_rate": 3.736101883921406e-07, - "loss": 0.9219, - "step": 4757 - }, - { - "epoch": 0.92, - "learning_rate": 3.719244842202074e-07, - "loss": 0.8174, - "step": 4758 - }, - { - "epoch": 0.92, - "learning_rate": 3.702425194818582e-07, - "loss": 0.7905, - "step": 4759 - }, - { - "epoch": 0.92, - "learning_rate": 3.685642948303503e-07, - "loss": 0.848, - "step": 4760 - }, - { - "epoch": 0.92, - "learning_rate": 3.66889810917489e-07, - "loss": 0.8065, - "step": 4761 - }, - { - "epoch": 0.92, - "learning_rate": 3.6521906839362187e-07, - "loss": 0.887, - "step": 4762 - }, - { - "epoch": 0.92, - "learning_rate": 3.635520679076465e-07, - "loss": 0.7684, - "step": 4763 - }, - { - "epoch": 0.92, - "learning_rate": 3.6188881010700725e-07, - "loss": 0.7775, - "step": 4764 - }, - { - "epoch": 0.92, - "learning_rate": 3.6022929563769513e-07, - "loss": 0.8383, - "step": 4765 - }, - { - "epoch": 0.92, - "learning_rate": 3.5857352514424573e-07, - "loss": 0.8265, - "step": 4766 - }, - { - "epoch": 0.92, - "learning_rate": 3.5692149926974006e-07, - "loss": 0.893, - "step": 4767 - }, - { - "epoch": 0.92, - "learning_rate": 3.552732186558072e-07, - "loss": 0.8339, - "step": 4768 - }, - { - "epoch": 0.92, - "learning_rate": 3.536286839426195e-07, - "loss": 0.7638, - "step": 4769 - }, - { - "epoch": 0.92, - "learning_rate": 3.51987895768896e-07, - "loss": 0.7811, - "step": 4770 - }, - { - "epoch": 0.92, - "learning_rate": 3.5035085477190143e-07, - "loss": 0.9027, - "step": 4771 - }, - { - "epoch": 0.92, - "learning_rate": 3.4871756158744054e-07, - "loss": 0.8595, - "step": 4772 - }, - { - "epoch": 0.92, - "learning_rate": 3.4708801684986693e-07, - "loss": 0.7998, - "step": 4773 - }, - { - "epoch": 0.92, - "learning_rate": 3.454622211920766e-07, - "loss": 0.8423, - "step": 4774 - }, - { - "epoch": 0.92, - "learning_rate": 3.4384017524551116e-07, - "loss": 0.7933, - "step": 4775 - }, - { - "epoch": 0.92, - "learning_rate": 3.422218796401544e-07, - "loss": 0.822, - "step": 4776 - }, - { - "epoch": 0.92, - "learning_rate": 3.4060733500453247e-07, - "loss": 0.8231, - "step": 4777 - }, - { - "epoch": 0.92, - "learning_rate": 3.38996541965716e-07, - "loss": 0.77, - "step": 4778 - }, - { - "epoch": 0.92, - "learning_rate": 3.3738950114932e-07, - "loss": 0.8849, - "step": 4779 - }, - { - "epoch": 0.92, - "learning_rate": 3.3578621317949755e-07, - "loss": 0.7881, - "step": 4780 - }, - { - "epoch": 0.92, - "learning_rate": 3.341866786789505e-07, - "loss": 0.8693, - "step": 4781 - }, - { - "epoch": 0.92, - "learning_rate": 3.325908982689185e-07, - "loss": 0.7989, - "step": 4782 - }, - { - "epoch": 0.92, - "learning_rate": 3.309988725691837e-07, - "loss": 0.7927, - "step": 4783 - }, - { - "epoch": 0.92, - "learning_rate": 3.294106021980714e-07, - "loss": 0.7804, - "step": 4784 - }, - { - "epoch": 0.92, - "learning_rate": 3.278260877724471e-07, - "loss": 0.7882, - "step": 4785 - }, - { - "epoch": 0.92, - "learning_rate": 3.2624532990771507e-07, - "loss": 0.8698, - "step": 4786 - }, - { - "epoch": 0.92, - "learning_rate": 3.2466832921782986e-07, - "loss": 0.8369, - "step": 4787 - }, - { - "epoch": 0.92, - "learning_rate": 3.2309508631527486e-07, - "loss": 0.8542, - "step": 4788 - }, - { - "epoch": 0.92, - "learning_rate": 3.215256018110824e-07, - "loss": 0.8012, - "step": 4789 - }, - { - "epoch": 0.92, - "learning_rate": 3.199598763148215e-07, - "loss": 0.7715, - "step": 4790 - }, - { - "epoch": 0.92, - "learning_rate": 3.183979104346002e-07, - "loss": 0.8644, - "step": 4791 - }, - { - "epoch": 0.92, - "learning_rate": 3.1683970477706994e-07, - "loss": 0.768, - "step": 4792 - }, - { - "epoch": 0.92, - "learning_rate": 3.1528525994741876e-07, - "loss": 0.8034, - "step": 4793 - }, - { - "epoch": 0.92, - "learning_rate": 3.13734576549376e-07, - "loss": 0.8964, - "step": 4794 - }, - { - "epoch": 0.92, - "learning_rate": 3.121876551852099e-07, - "loss": 0.8396, - "step": 4795 - }, - { - "epoch": 0.92, - "learning_rate": 3.1064449645572536e-07, - "loss": 0.8659, - "step": 4796 - }, - { - "epoch": 0.92, - "learning_rate": 3.091051009602675e-07, - "loss": 0.8238, - "step": 4797 - }, - { - "epoch": 0.92, - "learning_rate": 3.0756946929672017e-07, - "loss": 0.8338, - "step": 4798 - }, - { - "epoch": 0.92, - "learning_rate": 3.060376020615052e-07, - "loss": 0.932, - "step": 4799 - }, - { - "epoch": 0.92, - "learning_rate": 3.0450949984958347e-07, - "loss": 0.8651, - "step": 4800 - }, - { - "epoch": 0.92, - "learning_rate": 3.0298516325444893e-07, - "loss": 0.8593, - "step": 4801 - }, - { - "epoch": 0.92, - "learning_rate": 3.0146459286813924e-07, - "loss": 0.7331, - "step": 4802 - }, - { - "epoch": 0.92, - "learning_rate": 2.999477892812264e-07, - "loss": 0.901, - "step": 4803 - }, - { - "epoch": 0.92, - "learning_rate": 2.984347530828158e-07, - "loss": 0.8623, - "step": 4804 - }, - { - "epoch": 0.92, - "learning_rate": 2.969254848605585e-07, - "loss": 0.8136, - "step": 4805 - }, - { - "epoch": 0.92, - "learning_rate": 2.9541998520063344e-07, - "loss": 0.8833, - "step": 4806 - }, - { - "epoch": 0.92, - "learning_rate": 2.9391825468775946e-07, - "loss": 0.8959, - "step": 4807 - }, - { - "epoch": 0.93, - "learning_rate": 2.9242029390519454e-07, - "loss": 0.6976, - "step": 4808 - }, - { - "epoch": 0.93, - "learning_rate": 2.909261034347255e-07, - "loss": 0.9073, - "step": 4809 - }, - { - "epoch": 0.93, - "learning_rate": 2.894356838566792e-07, - "loss": 0.858, - "step": 4810 - }, - { - "epoch": 0.93, - "learning_rate": 2.879490357499204e-07, - "loss": 0.8143, - "step": 4811 - }, - { - "epoch": 0.93, - "learning_rate": 2.864661596918428e-07, - "loss": 0.7764, - "step": 4812 - }, - { - "epoch": 0.93, - "learning_rate": 2.849870562583812e-07, - "loss": 0.7891, - "step": 4813 - }, - { - "epoch": 0.93, - "learning_rate": 2.8351172602399945e-07, - "loss": 0.8301, - "step": 4814 - }, - { - "epoch": 0.93, - "learning_rate": 2.8204016956169924e-07, - "loss": 0.8667, - "step": 4815 - }, - { - "epoch": 0.93, - "learning_rate": 2.8057238744301994e-07, - "loss": 0.8091, - "step": 4816 - }, - { - "epoch": 0.93, - "learning_rate": 2.7910838023802676e-07, - "loss": 0.826, - "step": 4817 - }, - { - "epoch": 0.93, - "learning_rate": 2.7764814851532485e-07, - "loss": 0.8272, - "step": 4818 - }, - { - "epoch": 0.93, - "learning_rate": 2.761916928420527e-07, - "loss": 0.8318, - "step": 4819 - }, - { - "epoch": 0.93, - "learning_rate": 2.74739013783879e-07, - "loss": 0.8861, - "step": 4820 - }, - { - "epoch": 0.93, - "learning_rate": 2.7329011190500797e-07, - "loss": 0.8335, - "step": 4821 - }, - { - "epoch": 0.93, - "learning_rate": 2.7184498776817615e-07, - "loss": 0.8893, - "step": 4822 - }, - { - "epoch": 0.93, - "learning_rate": 2.704036419346534e-07, - "loss": 0.7343, - "step": 4823 - }, - { - "epoch": 0.93, - "learning_rate": 2.689660749642442e-07, - "loss": 0.8136, - "step": 4824 - }, - { - "epoch": 0.93, - "learning_rate": 2.675322874152786e-07, - "loss": 0.8262, - "step": 4825 - }, - { - "epoch": 0.93, - "learning_rate": 2.6610227984462556e-07, - "loss": 0.826, - "step": 4826 - }, - { - "epoch": 0.93, - "learning_rate": 2.646760528076842e-07, - "loss": 0.8687, - "step": 4827 - }, - { - "epoch": 0.93, - "learning_rate": 2.6325360685838243e-07, - "loss": 0.8865, - "step": 4828 - }, - { - "epoch": 0.93, - "learning_rate": 2.618349425491851e-07, - "loss": 0.765, - "step": 4829 - }, - { - "epoch": 0.93, - "learning_rate": 2.604200604310825e-07, - "loss": 0.8067, - "step": 4830 - }, - { - "epoch": 0.93, - "learning_rate": 2.590089610535984e-07, - "loss": 0.8457, - "step": 4831 - }, - { - "epoch": 0.93, - "learning_rate": 2.5760164496479e-07, - "loss": 0.8895, - "step": 4832 - }, - { - "epoch": 0.93, - "learning_rate": 2.5619811271123897e-07, - "loss": 0.7858, - "step": 4833 - }, - { - "epoch": 0.93, - "learning_rate": 2.5479836483806586e-07, - "loss": 0.8472, - "step": 4834 - }, - { - "epoch": 0.93, - "learning_rate": 2.5340240188891143e-07, - "loss": 0.8415, - "step": 4835 - }, - { - "epoch": 0.93, - "learning_rate": 2.520102244059552e-07, - "loss": 0.8336, - "step": 4836 - }, - { - "epoch": 0.93, - "learning_rate": 2.506218329299026e-07, - "loss": 0.8906, - "step": 4837 - }, - { - "epoch": 0.93, - "learning_rate": 2.4923722799998664e-07, - "loss": 0.8045, - "step": 4838 - }, - { - "epoch": 0.93, - "learning_rate": 2.4785641015397375e-07, - "loss": 0.833, - "step": 4839 - }, - { - "epoch": 0.93, - "learning_rate": 2.464793799281573e-07, - "loss": 0.8206, - "step": 4840 - }, - { - "epoch": 0.93, - "learning_rate": 2.4510613785735936e-07, - "loss": 0.8779, - "step": 4841 - }, - { - "epoch": 0.93, - "learning_rate": 2.4373668447493225e-07, - "loss": 0.7211, - "step": 4842 - }, - { - "epoch": 0.93, - "learning_rate": 2.423710203127561e-07, - "loss": 0.813, - "step": 4843 - }, - { - "epoch": 0.93, - "learning_rate": 2.410091459012376e-07, - "loss": 0.8608, - "step": 4844 - }, - { - "epoch": 0.93, - "learning_rate": 2.3965106176931375e-07, - "loss": 0.9164, - "step": 4845 - }, - { - "epoch": 0.93, - "learning_rate": 2.3829676844444926e-07, - "loss": 0.7351, - "step": 4846 - }, - { - "epoch": 0.93, - "learning_rate": 2.3694626645263675e-07, - "loss": 0.8125, - "step": 4847 - }, - { - "epoch": 0.93, - "learning_rate": 2.3559955631839436e-07, - "loss": 0.8566, - "step": 4848 - }, - { - "epoch": 0.93, - "learning_rate": 2.3425663856476932e-07, - "loss": 0.825, - "step": 4849 - }, - { - "epoch": 0.93, - "learning_rate": 2.3291751371333438e-07, - "loss": 0.8434, - "step": 4850 - }, - { - "epoch": 0.93, - "learning_rate": 2.3158218228419127e-07, - "loss": 0.8782, - "step": 4851 - }, - { - "epoch": 0.93, - "learning_rate": 2.3025064479596625e-07, - "loss": 0.6949, - "step": 4852 - }, - { - "epoch": 0.93, - "learning_rate": 2.2892290176581678e-07, - "loss": 0.8884, - "step": 4853 - }, - { - "epoch": 0.93, - "learning_rate": 2.2759895370941809e-07, - "loss": 0.8238, - "step": 4854 - }, - { - "epoch": 0.93, - "learning_rate": 2.2627880114097779e-07, - "loss": 0.8546, - "step": 4855 - }, - { - "epoch": 0.93, - "learning_rate": 2.2496244457323013e-07, - "loss": 0.838, - "step": 4856 - }, - { - "epoch": 0.93, - "learning_rate": 2.2364988451742953e-07, - "loss": 0.8371, - "step": 4857 - }, - { - "epoch": 0.93, - "learning_rate": 2.2234112148336373e-07, - "loss": 0.8647, - "step": 4858 - }, - { - "epoch": 0.93, - "learning_rate": 2.2103615597933613e-07, - "loss": 0.8739, - "step": 4859 - }, - { - "epoch": 0.94, - "learning_rate": 2.1973498851218244e-07, - "loss": 0.7764, - "step": 4860 - }, - { - "epoch": 0.94, - "learning_rate": 2.1843761958726283e-07, - "loss": 0.7606, - "step": 4861 - }, - { - "epoch": 0.94, - "learning_rate": 2.1714404970845647e-07, - "loss": 0.8368, - "step": 4862 - }, - { - "epoch": 0.94, - "learning_rate": 2.1585427937817594e-07, - "loss": 0.8024, - "step": 4863 - }, - { - "epoch": 0.94, - "learning_rate": 2.145683090973494e-07, - "loss": 0.8005, - "step": 4864 - }, - { - "epoch": 0.94, - "learning_rate": 2.1328613936543396e-07, - "loss": 0.8079, - "step": 4865 - }, - { - "epoch": 0.94, - "learning_rate": 2.1200777068041134e-07, - "loss": 0.8992, - "step": 4866 - }, - { - "epoch": 0.94, - "learning_rate": 2.1073320353878102e-07, - "loss": 0.8153, - "step": 4867 - }, - { - "epoch": 0.94, - "learning_rate": 2.0946243843557367e-07, - "loss": 0.9479, - "step": 4868 - }, - { - "epoch": 0.94, - "learning_rate": 2.0819547586434008e-07, - "loss": 0.8981, - "step": 4869 - }, - { - "epoch": 0.94, - "learning_rate": 2.0693231631715105e-07, - "loss": 0.8951, - "step": 4870 - }, - { - "epoch": 0.94, - "learning_rate": 2.0567296028460638e-07, - "loss": 0.8415, - "step": 4871 - }, - { - "epoch": 0.94, - "learning_rate": 2.0441740825582258e-07, - "loss": 0.7861, - "step": 4872 - }, - { - "epoch": 0.94, - "learning_rate": 2.0316566071844402e-07, - "loss": 0.8196, - "step": 4873 - }, - { - "epoch": 0.94, - "learning_rate": 2.0191771815863292e-07, - "loss": 0.8462, - "step": 4874 - }, - { - "epoch": 0.94, - "learning_rate": 2.0067358106107714e-07, - "loss": 0.9297, - "step": 4875 - }, - { - "epoch": 0.94, - "learning_rate": 1.994332499089846e-07, - "loss": 0.848, - "step": 4876 - }, - { - "epoch": 0.94, - "learning_rate": 1.9819672518408662e-07, - "loss": 0.9131, - "step": 4877 - }, - { - "epoch": 0.94, - "learning_rate": 1.9696400736663457e-07, - "loss": 0.8335, - "step": 4878 - }, - { - "epoch": 0.94, - "learning_rate": 1.9573509693540104e-07, - "loss": 0.7961, - "step": 4879 - }, - { - "epoch": 0.94, - "learning_rate": 1.9450999436768093e-07, - "loss": 0.7644, - "step": 4880 - }, - { - "epoch": 0.94, - "learning_rate": 1.9328870013929134e-07, - "loss": 0.8692, - "step": 4881 - }, - { - "epoch": 0.94, - "learning_rate": 1.9207121472456846e-07, - "loss": 0.7706, - "step": 4882 - }, - { - "epoch": 0.94, - "learning_rate": 1.9085753859636736e-07, - "loss": 0.8719, - "step": 4883 - }, - { - "epoch": 0.94, - "learning_rate": 1.8964767222606873e-07, - "loss": 0.8852, - "step": 4884 - }, - { - "epoch": 0.94, - "learning_rate": 1.8844161608356782e-07, - "loss": 0.7854, - "step": 4885 - }, - { - "epoch": 0.94, - "learning_rate": 1.872393706372866e-07, - "loss": 0.6609, - "step": 4886 - }, - { - "epoch": 0.94, - "learning_rate": 1.8604093635416155e-07, - "loss": 0.8769, - "step": 4887 - }, - { - "epoch": 0.94, - "learning_rate": 1.8484631369964922e-07, - "loss": 0.7793, - "step": 4888 - }, - { - "epoch": 0.94, - "learning_rate": 1.8365550313772852e-07, - "loss": 0.9143, - "step": 4889 - }, - { - "epoch": 0.94, - "learning_rate": 1.8246850513089832e-07, - "loss": 0.6911, - "step": 4890 - }, - { - "epoch": 0.94, - "learning_rate": 1.8128532014017098e-07, - "loss": 0.7518, - "step": 4891 - }, - { - "epoch": 0.94, - "learning_rate": 1.8010594862508669e-07, - "loss": 0.8373, - "step": 4892 - }, - { - "epoch": 0.94, - "learning_rate": 1.789303910436968e-07, - "loss": 0.8078, - "step": 4893 - }, - { - "epoch": 0.94, - "learning_rate": 1.777586478525739e-07, - "loss": 0.8201, - "step": 4894 - }, - { - "epoch": 0.94, - "learning_rate": 1.7659071950681172e-07, - "loss": 0.8609, - "step": 4895 - }, - { - "epoch": 0.94, - "learning_rate": 1.754266064600174e-07, - "loss": 0.6541, - "step": 4896 - }, - { - "epoch": 0.94, - "learning_rate": 1.742663091643204e-07, - "loss": 0.8092, - "step": 4897 - }, - { - "epoch": 0.94, - "learning_rate": 1.7310982807036915e-07, - "loss": 0.7647, - "step": 4898 - }, - { - "epoch": 0.94, - "learning_rate": 1.719571636273243e-07, - "loss": 0.8944, - "step": 4899 - }, - { - "epoch": 0.94, - "learning_rate": 1.7080831628286886e-07, - "loss": 0.8738, - "step": 4900 - }, - { - "epoch": 0.94, - "learning_rate": 1.6966328648320152e-07, - "loss": 0.7837, - "step": 4901 - }, - { - "epoch": 0.94, - "learning_rate": 1.685220746730387e-07, - "loss": 0.8532, - "step": 4902 - }, - { - "epoch": 0.94, - "learning_rate": 1.673846812956137e-07, - "loss": 0.8555, - "step": 4903 - }, - { - "epoch": 0.94, - "learning_rate": 1.6625110679267642e-07, - "loss": 0.8474, - "step": 4904 - }, - { - "epoch": 0.94, - "learning_rate": 1.6512135160449583e-07, - "loss": 0.8317, - "step": 4905 - }, - { - "epoch": 0.94, - "learning_rate": 1.6399541616985648e-07, - "loss": 0.966, - "step": 4906 - }, - { - "epoch": 0.94, - "learning_rate": 1.6287330092605525e-07, - "loss": 0.8188, - "step": 4907 - }, - { - "epoch": 0.94, - "learning_rate": 1.6175500630891128e-07, - "loss": 0.8692, - "step": 4908 - }, - { - "epoch": 0.94, - "learning_rate": 1.6064053275275716e-07, - "loss": 0.8785, - "step": 4909 - }, - { - "epoch": 0.94, - "learning_rate": 1.5952988069044105e-07, - "loss": 0.8888, - "step": 4910 - }, - { - "epoch": 0.94, - "learning_rate": 1.5842305055332796e-07, - "loss": 0.8306, - "step": 4911 - }, - { - "epoch": 0.95, - "learning_rate": 1.573200427712973e-07, - "loss": 0.81, - "step": 4912 - }, - { - "epoch": 0.95, - "learning_rate": 1.562208577727442e-07, - "loss": 0.8689, - "step": 4913 - }, - { - "epoch": 0.95, - "learning_rate": 1.551254959845805e-07, - "loss": 0.7736, - "step": 4914 - }, - { - "epoch": 0.95, - "learning_rate": 1.540339578322314e-07, - "loss": 0.9076, - "step": 4915 - }, - { - "epoch": 0.95, - "learning_rate": 1.5294624373963894e-07, - "loss": 0.7688, - "step": 4916 - }, - { - "epoch": 0.95, - "learning_rate": 1.5186235412925744e-07, - "loss": 0.8169, - "step": 4917 - }, - { - "epoch": 0.95, - "learning_rate": 1.5078228942205674e-07, - "loss": 0.8218, - "step": 4918 - }, - { - "epoch": 0.95, - "learning_rate": 1.4970605003752359e-07, - "loss": 0.7976, - "step": 4919 - }, - { - "epoch": 0.95, - "learning_rate": 1.4863363639365357e-07, - "loss": 0.8373, - "step": 4920 - }, - { - "epoch": 0.95, - "learning_rate": 1.4756504890696466e-07, - "loss": 0.8898, - "step": 4921 - }, - { - "epoch": 0.95, - "learning_rate": 1.4650028799247928e-07, - "loss": 0.8793, - "step": 4922 - }, - { - "epoch": 0.95, - "learning_rate": 1.454393540637411e-07, - "loss": 0.8842, - "step": 4923 - }, - { - "epoch": 0.95, - "learning_rate": 1.4438224753280384e-07, - "loss": 0.8581, - "step": 4924 - }, - { - "epoch": 0.95, - "learning_rate": 1.4332896881023462e-07, - "loss": 0.8388, - "step": 4925 - }, - { - "epoch": 0.95, - "learning_rate": 1.422795183051151e-07, - "loss": 0.7958, - "step": 4926 - }, - { - "epoch": 0.95, - "learning_rate": 1.4123389642504148e-07, - "loss": 0.896, - "step": 4927 - }, - { - "epoch": 0.95, - "learning_rate": 1.401921035761189e-07, - "loss": 0.6259, - "step": 4928 - }, - { - "epoch": 0.95, - "learning_rate": 1.3915414016296925e-07, - "loss": 0.9077, - "step": 4929 - }, - { - "epoch": 0.95, - "learning_rate": 1.381200065887256e-07, - "loss": 0.8386, - "step": 4930 - }, - { - "epoch": 0.95, - "learning_rate": 1.3708970325503222e-07, - "loss": 0.8582, - "step": 4931 - }, - { - "epoch": 0.95, - "learning_rate": 1.3606323056204795e-07, - "loss": 0.8326, - "step": 4932 - }, - { - "epoch": 0.95, - "learning_rate": 1.3504058890844274e-07, - "loss": 0.8532, - "step": 4933 - }, - { - "epoch": 0.95, - "learning_rate": 1.3402177869139887e-07, - "loss": 0.8525, - "step": 4934 - }, - { - "epoch": 0.95, - "learning_rate": 1.3300680030661096e-07, - "loss": 0.8261, - "step": 4935 - }, - { - "epoch": 0.95, - "learning_rate": 1.3199565414828363e-07, - "loss": 0.8521, - "step": 4936 - }, - { - "epoch": 0.95, - "learning_rate": 1.3098834060913612e-07, - "loss": 0.7853, - "step": 4937 - }, - { - "epoch": 0.95, - "learning_rate": 1.2998486008039545e-07, - "loss": 0.82, - "step": 4938 - }, - { - "epoch": 0.95, - "learning_rate": 1.289852129518032e-07, - "loss": 0.8563, - "step": 4939 - }, - { - "epoch": 0.95, - "learning_rate": 1.2798939961161217e-07, - "loss": 0.8502, - "step": 4940 - }, - { - "epoch": 0.95, - "learning_rate": 1.269974204465818e-07, - "loss": 0.8654, - "step": 4941 - }, - { - "epoch": 0.95, - "learning_rate": 1.2600927584198618e-07, - "loss": 0.8253, - "step": 4942 - }, - { - "epoch": 0.95, - "learning_rate": 1.2502496618161165e-07, - "loss": 0.8187, - "step": 4943 - }, - { - "epoch": 0.95, - "learning_rate": 1.2404449184774904e-07, - "loss": 0.9223, - "step": 4944 - }, - { - "epoch": 0.95, - "learning_rate": 1.2306785322120596e-07, - "loss": 0.8057, - "step": 4945 - }, - { - "epoch": 0.95, - "learning_rate": 1.220950506812968e-07, - "loss": 0.837, - "step": 4946 - }, - { - "epoch": 0.95, - "learning_rate": 1.2112608460584707e-07, - "loss": 0.8528, - "step": 4947 - }, - { - "epoch": 0.95, - "learning_rate": 1.2016095537119242e-07, - "loss": 0.8767, - "step": 4948 - }, - { - "epoch": 0.95, - "learning_rate": 1.1919966335217636e-07, - "loss": 0.8315, - "step": 4949 - }, - { - "epoch": 0.95, - "learning_rate": 1.1824220892215465e-07, - "loss": 0.7735, - "step": 4950 - }, - { - "epoch": 0.95, - "learning_rate": 1.1728859245299207e-07, - "loss": 0.8597, - "step": 4951 - }, - { - "epoch": 0.95, - "learning_rate": 1.1633881431506122e-07, - "loss": 0.8059, - "step": 4952 - }, - { - "epoch": 0.95, - "learning_rate": 1.1539287487724594e-07, - "loss": 0.8815, - "step": 4953 - }, - { - "epoch": 0.95, - "learning_rate": 1.1445077450693786e-07, - "loss": 0.8576, - "step": 4954 - }, - { - "epoch": 0.95, - "learning_rate": 1.1351251357003656e-07, - "loss": 0.9237, - "step": 4955 - }, - { - "epoch": 0.95, - "learning_rate": 1.1257809243095385e-07, - "loss": 0.8442, - "step": 4956 - }, - { - "epoch": 0.95, - "learning_rate": 1.1164751145260722e-07, - "loss": 0.8447, - "step": 4957 - }, - { - "epoch": 0.95, - "learning_rate": 1.107207709964242e-07, - "loss": 0.7842, - "step": 4958 - }, - { - "epoch": 0.95, - "learning_rate": 1.0979787142233911e-07, - "loss": 0.8967, - "step": 4959 - }, - { - "epoch": 0.95, - "learning_rate": 1.0887881308879633e-07, - "loss": 0.8386, - "step": 4960 - }, - { - "epoch": 0.95, - "learning_rate": 1.0796359635274701e-07, - "loss": 0.7337, - "step": 4961 - }, - { - "epoch": 0.95, - "learning_rate": 1.0705222156965011e-07, - "loss": 0.8548, - "step": 4962 - }, - { - "epoch": 0.95, - "learning_rate": 1.0614468909347476e-07, - "loss": 0.8234, - "step": 4963 - }, - { - "epoch": 0.96, - "learning_rate": 1.0524099927669563e-07, - "loss": 0.8303, - "step": 4964 - }, - { - "epoch": 0.96, - "learning_rate": 1.0434115247029419e-07, - "loss": 0.7661, - "step": 4965 - }, - { - "epoch": 0.96, - "learning_rate": 1.0344514902376201e-07, - "loss": 0.8203, - "step": 4966 - }, - { - "epoch": 0.96, - "learning_rate": 1.0255298928509627e-07, - "loss": 0.837, - "step": 4967 - }, - { - "epoch": 0.96, - "learning_rate": 1.0166467360079979e-07, - "loss": 0.8471, - "step": 4968 - }, - { - "epoch": 0.96, - "learning_rate": 1.007802023158877e-07, - "loss": 0.814, - "step": 4969 - }, - { - "epoch": 0.96, - "learning_rate": 9.989957577387521e-08, - "loss": 0.8631, - "step": 4970 - }, - { - "epoch": 0.96, - "learning_rate": 9.902279431678874e-08, - "loss": 0.7983, - "step": 4971 - }, - { - "epoch": 0.96, - "learning_rate": 9.814985828516033e-08, - "loss": 0.8256, - "step": 4972 - }, - { - "epoch": 0.96, - "learning_rate": 9.728076801802656e-08, - "loss": 0.8591, - "step": 4973 - }, - { - "epoch": 0.96, - "learning_rate": 9.641552385293518e-08, - "loss": 0.6809, - "step": 4974 - }, - { - "epoch": 0.96, - "learning_rate": 9.555412612593518e-08, - "loss": 0.8756, - "step": 4975 - }, - { - "epoch": 0.96, - "learning_rate": 9.469657517158226e-08, - "loss": 0.8841, - "step": 4976 - }, - { - "epoch": 0.96, - "learning_rate": 9.384287132294223e-08, - "loss": 0.7833, - "step": 4977 - }, - { - "epoch": 0.96, - "learning_rate": 9.299301491158207e-08, - "loss": 0.8459, - "step": 4978 - }, - { - "epoch": 0.96, - "learning_rate": 9.214700626757667e-08, - "loss": 0.8361, - "step": 4979 - }, - { - "epoch": 0.96, - "learning_rate": 9.130484571950538e-08, - "loss": 0.935, - "step": 4980 - }, - { - "epoch": 0.96, - "learning_rate": 9.046653359445323e-08, - "loss": 0.8622, - "step": 4981 - }, - { - "epoch": 0.96, - "learning_rate": 8.963207021801423e-08, - "loss": 0.7989, - "step": 4982 - }, - { - "epoch": 0.96, - "learning_rate": 8.880145591428024e-08, - "loss": 0.8526, - "step": 4983 - }, - { - "epoch": 0.96, - "learning_rate": 8.797469100585432e-08, - "loss": 0.8358, - "step": 4984 - }, - { - "epoch": 0.96, - "learning_rate": 8.715177581384182e-08, - "loss": 0.9225, - "step": 4985 - }, - { - "epoch": 0.96, - "learning_rate": 8.633271065785486e-08, - "loss": 0.6883, - "step": 4986 - }, - { - "epoch": 0.96, - "learning_rate": 8.551749585600678e-08, - "loss": 0.9095, - "step": 4987 - }, - { - "epoch": 0.96, - "learning_rate": 8.470613172491981e-08, - "loss": 0.7715, - "step": 4988 - }, - { - "epoch": 0.96, - "learning_rate": 8.389861857971748e-08, - "loss": 0.8349, - "step": 4989 - }, - { - "epoch": 0.96, - "learning_rate": 8.309495673402778e-08, - "loss": 0.8559, - "step": 4990 - }, - { - "epoch": 0.96, - "learning_rate": 8.229514649998438e-08, - "loss": 0.797, - "step": 4991 - }, - { - "epoch": 0.96, - "learning_rate": 8.149918818822433e-08, - "loss": 0.8271, - "step": 4992 - }, - { - "epoch": 0.96, - "learning_rate": 8.070708210788925e-08, - "loss": 0.8293, - "step": 4993 - }, - { - "epoch": 0.96, - "learning_rate": 7.991882856662303e-08, - "loss": 0.8024, - "step": 4994 - }, - { - "epoch": 0.96, - "learning_rate": 7.913442787057523e-08, - "loss": 0.6988, - "step": 4995 - }, - { - "epoch": 0.96, - "learning_rate": 7.835388032439661e-08, - "loss": 0.8801, - "step": 4996 - }, - { - "epoch": 0.96, - "learning_rate": 7.757718623124466e-08, - "loss": 0.8324, - "step": 4997 - }, - { - "epoch": 0.96, - "learning_rate": 7.680434589277696e-08, - "loss": 0.8796, - "step": 4998 - }, - { - "epoch": 0.96, - "learning_rate": 7.603535960915675e-08, - "loss": 0.8226, - "step": 4999 - }, - { - "epoch": 0.96, - "learning_rate": 7.527022767904957e-08, - "loss": 0.7937, - "step": 5000 - }, - { - "epoch": 0.96, - "learning_rate": 7.450895039962214e-08, - "loss": 0.8451, - "step": 5001 - }, - { - "epoch": 0.96, - "learning_rate": 7.375152806654685e-08, - "loss": 0.8581, - "step": 5002 - }, - { - "epoch": 0.96, - "learning_rate": 7.299796097399947e-08, - "loss": 0.7871, - "step": 5003 - }, - { - "epoch": 0.96, - "learning_rate": 7.224824941465369e-08, - "loss": 0.8675, - "step": 5004 - }, - { - "epoch": 0.96, - "learning_rate": 7.150239367969102e-08, - "loss": 0.8673, - "step": 5005 - }, - { - "epoch": 0.96, - "learning_rate": 7.076039405879309e-08, - "loss": 0.8496, - "step": 5006 - }, - { - "epoch": 0.96, - "learning_rate": 7.002225084014269e-08, - "loss": 0.8344, - "step": 5007 - }, - { - "epoch": 0.96, - "learning_rate": 6.928796431042717e-08, - "loss": 0.9233, - "step": 5008 - }, - { - "epoch": 0.96, - "learning_rate": 6.855753475483507e-08, - "loss": 0.7978, - "step": 5009 - }, - { - "epoch": 0.96, - "learning_rate": 6.783096245705612e-08, - "loss": 1.0111, - "step": 5010 - }, - { - "epoch": 0.96, - "learning_rate": 6.710824769928349e-08, - "loss": 0.8839, - "step": 5011 - }, - { - "epoch": 0.96, - "learning_rate": 6.63893907622104e-08, - "loss": 0.829, - "step": 5012 - }, - { - "epoch": 0.96, - "learning_rate": 6.567439192503244e-08, - "loss": 0.84, - "step": 5013 - }, - { - "epoch": 0.96, - "learning_rate": 6.496325146544746e-08, - "loss": 0.7124, - "step": 5014 - }, - { - "epoch": 0.96, - "learning_rate": 6.425596965965453e-08, - "loss": 0.8442, - "step": 5015 - }, - { - "epoch": 0.97, - "learning_rate": 6.35525467823539e-08, - "loss": 0.8497, - "step": 5016 - }, - { - "epoch": 0.97, - "learning_rate": 6.285298310674703e-08, - "loss": 0.9303, - "step": 5017 - }, - { - "epoch": 0.97, - "learning_rate": 6.215727890453438e-08, - "loss": 0.8988, - "step": 5018 - }, - { - "epoch": 0.97, - "learning_rate": 6.14654344459209e-08, - "loss": 0.8372, - "step": 5019 - }, - { - "epoch": 0.97, - "learning_rate": 6.077744999961166e-08, - "loss": 0.7898, - "step": 5020 - }, - { - "epoch": 0.97, - "learning_rate": 6.009332583281069e-08, - "loss": 0.8863, - "step": 5021 - }, - { - "epoch": 0.97, - "learning_rate": 5.941306221122545e-08, - "loss": 0.8383, - "step": 5022 - }, - { - "epoch": 0.97, - "learning_rate": 5.873665939906015e-08, - "loss": 0.8154, - "step": 5023 - }, - { - "epoch": 0.97, - "learning_rate": 5.806411765902353e-08, - "loss": 0.8034, - "step": 5024 - }, - { - "epoch": 0.97, - "learning_rate": 5.739543725232333e-08, - "loss": 0.8889, - "step": 5025 - }, - { - "epoch": 0.97, - "learning_rate": 5.673061843866623e-08, - "loss": 0.8619, - "step": 5026 - }, - { - "epoch": 0.97, - "learning_rate": 5.606966147626125e-08, - "loss": 0.8741, - "step": 5027 - }, - { - "epoch": 0.97, - "learning_rate": 5.5412566621815266e-08, - "loss": 0.7622, - "step": 5028 - }, - { - "epoch": 0.97, - "learning_rate": 5.4759334130536355e-08, - "loss": 0.8339, - "step": 5029 - }, - { - "epoch": 0.97, - "learning_rate": 5.41099642561338e-08, - "loss": 0.8192, - "step": 5030 - }, - { - "epoch": 0.97, - "learning_rate": 5.3464457250814726e-08, - "loss": 0.8399, - "step": 5031 - }, - { - "epoch": 0.97, - "learning_rate": 5.2822813365286386e-08, - "loss": 0.7045, - "step": 5032 - }, - { - "epoch": 0.97, - "learning_rate": 5.218503284875609e-08, - "loss": 0.846, - "step": 5033 - }, - { - "epoch": 0.97, - "learning_rate": 5.155111594893014e-08, - "loss": 0.8118, - "step": 5034 - }, - { - "epoch": 0.97, - "learning_rate": 5.092106291201604e-08, - "loss": 0.8565, - "step": 5035 - }, - { - "epoch": 0.97, - "learning_rate": 5.029487398271693e-08, - "loss": 0.8166, - "step": 5036 - }, - { - "epoch": 0.97, - "learning_rate": 4.967254940423716e-08, - "loss": 0.863, - "step": 5037 - }, - { - "epoch": 0.97, - "learning_rate": 4.905408941828338e-08, - "loss": 0.8216, - "step": 5038 - }, - { - "epoch": 0.97, - "learning_rate": 4.8439494265055674e-08, - "loss": 0.7955, - "step": 5039 - }, - { - "epoch": 0.97, - "learning_rate": 4.7828764183257545e-08, - "loss": 0.8485, - "step": 5040 - }, - { - "epoch": 0.97, - "learning_rate": 4.722189941008703e-08, - "loss": 0.82, - "step": 5041 - }, - { - "epoch": 0.97, - "learning_rate": 4.66189001812456e-08, - "loss": 0.8451, - "step": 5042 - }, - { - "epoch": 0.97, - "learning_rate": 4.6019766730930336e-08, - "loss": 0.8805, - "step": 5043 - }, - { - "epoch": 0.97, - "learning_rate": 4.542449929183845e-08, - "loss": 0.83, - "step": 5044 - }, - { - "epoch": 0.97, - "learning_rate": 4.4833098095165005e-08, - "loss": 0.8963, - "step": 5045 - }, - { - "epoch": 0.97, - "learning_rate": 4.424556337060182e-08, - "loss": 0.8495, - "step": 5046 - }, - { - "epoch": 0.97, - "learning_rate": 4.366189534634191e-08, - "loss": 0.8554, - "step": 5047 - }, - { - "epoch": 0.97, - "learning_rate": 4.308209424907506e-08, - "loss": 0.8794, - "step": 5048 - }, - { - "epoch": 0.97, - "learning_rate": 4.250616030399002e-08, - "loss": 0.8017, - "step": 5049 - }, - { - "epoch": 0.97, - "learning_rate": 4.1934093734771194e-08, - "loss": 0.8699, - "step": 5050 - }, - { - "epoch": 0.97, - "learning_rate": 4.13658947636042e-08, - "loss": 0.7987, - "step": 5051 - }, - { - "epoch": 0.97, - "learning_rate": 4.080156361117027e-08, - "loss": 0.8809, - "step": 5052 - }, - { - "epoch": 0.97, - "learning_rate": 4.024110049664853e-08, - "loss": 0.8309, - "step": 5053 - }, - { - "epoch": 0.97, - "learning_rate": 3.9684505637718194e-08, - "loss": 0.8658, - "step": 5054 - }, - { - "epoch": 0.97, - "learning_rate": 3.913177925055189e-08, - "loss": 0.8161, - "step": 5055 - }, - { - "epoch": 0.97, - "learning_rate": 3.858292154982457e-08, - "loss": 0.8102, - "step": 5056 - }, - { - "epoch": 0.97, - "learning_rate": 3.8037932748704596e-08, - "loss": 0.9177, - "step": 5057 - }, - { - "epoch": 0.97, - "learning_rate": 3.7496813058859325e-08, - "loss": 0.8113, - "step": 5058 - }, - { - "epoch": 0.97, - "learning_rate": 3.695956269045509e-08, - "loss": 0.8237, - "step": 5059 - }, - { - "epoch": 0.97, - "learning_rate": 3.642618185215163e-08, - "loss": 0.7779, - "step": 5060 - }, - { - "epoch": 0.97, - "learning_rate": 3.589667075110992e-08, - "loss": 0.7974, - "step": 5061 - }, - { - "epoch": 0.97, - "learning_rate": 3.537102959298322e-08, - "loss": 0.8899, - "step": 5062 - }, - { - "epoch": 0.97, - "learning_rate": 3.4849258581928224e-08, - "loss": 0.8921, - "step": 5063 - }, - { - "epoch": 0.97, - "learning_rate": 3.4331357920591726e-08, - "loss": 0.8395, - "step": 5064 - }, - { - "epoch": 0.97, - "learning_rate": 3.3817327810121706e-08, - "loss": 0.794, - "step": 5065 - }, - { - "epoch": 0.97, - "learning_rate": 3.3307168450160685e-08, - "loss": 0.8874, - "step": 5066 - }, - { - "epoch": 0.97, - "learning_rate": 3.280088003885018e-08, - "loss": 0.8131, - "step": 5067 - }, - { - "epoch": 0.98, - "learning_rate": 3.229846277282511e-08, - "loss": 0.8353, - "step": 5068 - }, - { - "epoch": 0.98, - "learning_rate": 3.1799916847220505e-08, - "loss": 0.8999, - "step": 5069 - }, - { - "epoch": 0.98, - "learning_rate": 3.130524245566369e-08, - "loss": 0.8693, - "step": 5070 - }, - { - "epoch": 0.98, - "learning_rate": 3.0814439790280984e-08, - "loss": 0.8067, - "step": 5071 - }, - { - "epoch": 0.98, - "learning_rate": 3.032750904169546e-08, - "loss": 0.8701, - "step": 5072 - }, - { - "epoch": 0.98, - "learning_rate": 2.9844450399024725e-08, - "loss": 0.823, - "step": 5073 - }, - { - "epoch": 0.98, - "learning_rate": 2.936526404988427e-08, - "loss": 0.8121, - "step": 5074 - }, - { - "epoch": 0.98, - "learning_rate": 2.8889950180382985e-08, - "loss": 0.872, - "step": 5075 - }, - { - "epoch": 0.98, - "learning_rate": 2.8418508975127655e-08, - "loss": 0.8566, - "step": 5076 - }, - { - "epoch": 0.98, - "learning_rate": 2.7950940617221812e-08, - "loss": 0.7031, - "step": 5077 - }, - { - "epoch": 0.98, - "learning_rate": 2.7487245288261298e-08, - "loss": 0.8577, - "step": 5078 - }, - { - "epoch": 0.98, - "learning_rate": 2.702742316834206e-08, - "loss": 0.856, - "step": 5079 - }, - { - "epoch": 0.98, - "learning_rate": 2.657147443605457e-08, - "loss": 0.848, - "step": 5080 - }, - { - "epoch": 0.98, - "learning_rate": 2.6119399268480505e-08, - "loss": 0.8534, - "step": 5081 - }, - { - "epoch": 0.98, - "learning_rate": 2.5671197841203867e-08, - "loss": 0.8282, - "step": 5082 - }, - { - "epoch": 0.98, - "learning_rate": 2.5226870328299845e-08, - "loss": 0.8733, - "step": 5083 - }, - { - "epoch": 0.98, - "learning_rate": 2.4786416902339293e-08, - "loss": 0.8591, - "step": 5084 - }, - { - "epoch": 0.98, - "learning_rate": 2.4349837734390923e-08, - "loss": 0.8908, - "step": 5085 - }, - { - "epoch": 0.98, - "learning_rate": 2.3917132994016878e-08, - "loss": 0.8116, - "step": 5086 - }, - { - "epoch": 0.98, - "learning_rate": 2.3488302849272726e-08, - "loss": 0.827, - "step": 5087 - }, - { - "epoch": 0.98, - "learning_rate": 2.3063347466713014e-08, - "loss": 0.7435, - "step": 5088 - }, - { - "epoch": 0.98, - "learning_rate": 2.264226701138461e-08, - "loss": 0.8589, - "step": 5089 - }, - { - "epoch": 0.98, - "learning_rate": 2.2225061646830025e-08, - "loss": 0.824, - "step": 5090 - }, - { - "epoch": 0.98, - "learning_rate": 2.181173153508853e-08, - "loss": 0.8315, - "step": 5091 - }, - { - "epoch": 0.98, - "learning_rate": 2.1402276836691715e-08, - "loss": 0.7647, - "step": 5092 - }, - { - "epoch": 0.98, - "learning_rate": 2.0996697710666812e-08, - "loss": 0.8081, - "step": 5093 - }, - { - "epoch": 0.98, - "learning_rate": 2.059499431453671e-08, - "loss": 0.8797, - "step": 5094 - }, - { - "epoch": 0.98, - "learning_rate": 2.0197166804317713e-08, - "loss": 0.8283, - "step": 5095 - }, - { - "epoch": 0.98, - "learning_rate": 1.9803215334522895e-08, - "loss": 0.7885, - "step": 5096 - }, - { - "epoch": 0.98, - "learning_rate": 1.9413140058156533e-08, - "loss": 0.9292, - "step": 5097 - }, - { - "epoch": 0.98, - "learning_rate": 1.9026941126721876e-08, - "loss": 0.879, - "step": 5098 - }, - { - "epoch": 0.98, - "learning_rate": 1.864461869021117e-08, - "loss": 0.8182, - "step": 5099 - }, - { - "epoch": 0.98, - "learning_rate": 1.826617289711563e-08, - "loss": 0.8158, - "step": 5100 - }, - { - "epoch": 0.98, - "learning_rate": 1.7891603894418797e-08, - "loss": 0.8569, - "step": 5101 - }, - { - "epoch": 0.98, - "learning_rate": 1.7520911827598742e-08, - "loss": 0.8493, - "step": 5102 - }, - { - "epoch": 0.98, - "learning_rate": 1.7154096840629186e-08, - "loss": 0.9004, - "step": 5103 - }, - { - "epoch": 0.98, - "learning_rate": 1.679115907597617e-08, - "loss": 0.9502, - "step": 5104 - }, - { - "epoch": 0.98, - "learning_rate": 1.6432098674600272e-08, - "loss": 0.8291, - "step": 5105 - }, - { - "epoch": 0.98, - "learning_rate": 1.6076915775956604e-08, - "loss": 0.8292, - "step": 5106 - }, - { - "epoch": 0.98, - "learning_rate": 1.5725610517994816e-08, - "loss": 0.8983, - "step": 5107 - }, - { - "epoch": 0.98, - "learning_rate": 1.537818303715688e-08, - "loss": 0.7809, - "step": 5108 - }, - { - "epoch": 0.98, - "learning_rate": 1.50346334683793e-08, - "loss": 0.8616, - "step": 5109 - }, - { - "epoch": 0.98, - "learning_rate": 1.4694961945093122e-08, - "loss": 0.7943, - "step": 5110 - }, - { - "epoch": 0.98, - "learning_rate": 1.4359168599223926e-08, - "loss": 0.9287, - "step": 5111 - }, - { - "epoch": 0.98, - "learning_rate": 1.4027253561188502e-08, - "loss": 0.8438, - "step": 5112 - }, - { - "epoch": 0.98, - "learning_rate": 1.3699216959899287e-08, - "loss": 0.7506, - "step": 5113 - }, - { - "epoch": 0.98, - "learning_rate": 1.337505892276325e-08, - "loss": 0.8537, - "step": 5114 - }, - { - "epoch": 0.98, - "learning_rate": 1.3054779575677468e-08, - "loss": 0.7568, - "step": 5115 - }, - { - "epoch": 0.98, - "learning_rate": 1.2738379043035765e-08, - "loss": 0.8279, - "step": 5116 - }, - { - "epoch": 0.98, - "learning_rate": 1.2425857447725397e-08, - "loss": 0.815, - "step": 5117 - }, - { - "epoch": 0.98, - "learning_rate": 1.211721491112372e-08, - "loss": 0.8265, - "step": 5118 - }, - { - "epoch": 0.98, - "learning_rate": 1.1812451553107063e-08, - "loss": 0.8459, - "step": 5119 - }, - { - "epoch": 0.99, - "learning_rate": 1.1511567492038522e-08, - "loss": 0.9024, - "step": 5120 - }, - { - "epoch": 0.99, - "learning_rate": 1.1214562844781285e-08, - "loss": 0.8195, - "step": 5121 - }, - { - "epoch": 0.99, - "learning_rate": 1.0921437726686413e-08, - "loss": 0.811, - "step": 5122 - }, - { - "epoch": 0.99, - "learning_rate": 1.0632192251601726e-08, - "loss": 0.7804, - "step": 5123 - }, - { - "epoch": 0.99, - "learning_rate": 1.0346826531865139e-08, - "loss": 0.744, - "step": 5124 - }, - { - "epoch": 0.99, - "learning_rate": 1.006534067831022e-08, - "loss": 0.8722, - "step": 5125 - }, - { - "epoch": 0.99, - "learning_rate": 9.78773480026396e-09, - "loss": 0.8902, - "step": 5126 - }, - { - "epoch": 0.99, - "learning_rate": 9.514009005543445e-09, - "loss": 0.8719, - "step": 5127 - }, - { - "epoch": 0.99, - "learning_rate": 9.244163400462525e-09, - "loss": 0.905, - "step": 5128 - }, - { - "epoch": 0.99, - "learning_rate": 8.97819808982403e-09, - "loss": 0.8686, - "step": 5129 - }, - { - "epoch": 0.99, - "learning_rate": 8.716113176927554e-09, - "loss": 0.7773, - "step": 5130 - }, - { - "epoch": 0.99, - "learning_rate": 8.457908763562783e-09, - "loss": 0.9107, - "step": 5131 - }, - { - "epoch": 0.99, - "learning_rate": 8.203584950013942e-09, - "loss": 0.8113, - "step": 5132 - }, - { - "epoch": 0.99, - "learning_rate": 7.953141835057576e-09, - "loss": 0.8845, - "step": 5133 - }, - { - "epoch": 0.99, - "learning_rate": 7.706579515962542e-09, - "loss": 0.847, - "step": 5134 - }, - { - "epoch": 0.99, - "learning_rate": 7.463898088490019e-09, - "loss": 0.9193, - "step": 5135 - }, - { - "epoch": 0.99, - "learning_rate": 7.225097646895718e-09, - "loss": 0.9325, - "step": 5136 - }, - { - "epoch": 0.99, - "learning_rate": 6.990178283927673e-09, - "loss": 0.821, - "step": 5137 - }, - { - "epoch": 0.99, - "learning_rate": 6.759140090824012e-09, - "loss": 0.8243, - "step": 5138 - }, - { - "epoch": 0.99, - "learning_rate": 6.531983157318511e-09, - "loss": 0.9859, - "step": 5139 - }, - { - "epoch": 0.99, - "learning_rate": 6.308707571636152e-09, - "loss": 0.7831, - "step": 5140 - }, - { - "epoch": 0.99, - "learning_rate": 6.089313420494236e-09, - "loss": 0.8176, - "step": 5141 - }, - { - "epoch": 0.99, - "learning_rate": 5.87380078910349e-09, - "loss": 0.8663, - "step": 5142 - }, - { - "epoch": 0.99, - "learning_rate": 5.662169761165848e-09, - "loss": 0.8721, - "step": 5143 - }, - { - "epoch": 0.99, - "learning_rate": 5.4544204188777815e-09, - "loss": 0.88, - "step": 5144 - }, - { - "epoch": 0.99, - "learning_rate": 5.25055284292475e-09, - "loss": 0.9074, - "step": 5145 - }, - { - "epoch": 0.99, - "learning_rate": 5.050567112487858e-09, - "loss": 0.7517, - "step": 5146 - }, - { - "epoch": 0.99, - "learning_rate": 4.85446330523831e-09, - "loss": 0.8522, - "step": 5147 - }, - { - "epoch": 0.99, - "learning_rate": 4.662241497341846e-09, - "loss": 0.823, - "step": 5148 - }, - { - "epoch": 0.99, - "learning_rate": 4.473901763454302e-09, - "loss": 0.876, - "step": 5149 - }, - { - "epoch": 0.99, - "learning_rate": 4.289444176724944e-09, - "loss": 0.8964, - "step": 5150 - }, - { - "epoch": 0.99, - "learning_rate": 4.108868808794242e-09, - "loss": 0.7601, - "step": 5151 - }, - { - "epoch": 0.99, - "learning_rate": 3.932175729797205e-09, - "loss": 0.7874, - "step": 5152 - }, - { - "epoch": 0.99, - "learning_rate": 3.759365008357829e-09, - "loss": 0.8523, - "step": 5153 - }, - { - "epoch": 0.99, - "learning_rate": 3.590436711594647e-09, - "loss": 0.874, - "step": 5154 - }, - { - "epoch": 0.99, - "learning_rate": 3.4253909051173985e-09, - "loss": 0.8609, - "step": 5155 - }, - { - "epoch": 0.99, - "learning_rate": 3.2642276530281404e-09, - "loss": 0.9054, - "step": 5156 - }, - { - "epoch": 0.99, - "learning_rate": 3.1069470179201365e-09, - "loss": 0.7538, - "step": 5157 - }, - { - "epoch": 0.99, - "learning_rate": 2.9535490608789687e-09, - "loss": 0.7899, - "step": 5158 - }, - { - "epoch": 0.99, - "learning_rate": 2.8040338414847545e-09, - "loss": 0.8235, - "step": 5159 - }, - { - "epoch": 0.99, - "learning_rate": 2.6584014178054894e-09, - "loss": 0.8625, - "step": 5160 - }, - { - "epoch": 0.99, - "learning_rate": 2.516651846403706e-09, - "loss": 0.75, - "step": 5161 - }, - { - "epoch": 0.99, - "learning_rate": 2.378785182333143e-09, - "loss": 0.8382, - "step": 5162 - }, - { - "epoch": 0.99, - "learning_rate": 2.2448014791398574e-09, - "loss": 0.8935, - "step": 5163 - }, - { - "epoch": 0.99, - "learning_rate": 2.1147007888622227e-09, - "loss": 0.6783, - "step": 5164 - }, - { - "epoch": 0.99, - "learning_rate": 1.9884831620287094e-09, - "loss": 0.8414, - "step": 5165 - }, - { - "epoch": 0.99, - "learning_rate": 1.8661486476612144e-09, - "loss": 0.8932, - "step": 5166 - }, - { - "epoch": 0.99, - "learning_rate": 1.747697293272843e-09, - "loss": 0.9082, - "step": 5167 - }, - { - "epoch": 0.99, - "learning_rate": 1.633129144870127e-09, - "loss": 0.7914, - "step": 5168 - }, - { - "epoch": 0.99, - "learning_rate": 1.5224442469474743e-09, - "loss": 0.8048, - "step": 5169 - }, - { - "epoch": 0.99, - "learning_rate": 1.4156426424960513e-09, - "loss": 0.8378, - "step": 5170 - }, - { - "epoch": 0.99, - "learning_rate": 1.3127243729949003e-09, - "loss": 0.8286, - "step": 5171 - }, - { - "epoch": 1.0, - "learning_rate": 1.2136894784176011e-09, - "loss": 0.8562, - "step": 5172 - }, - { - "epoch": 1.0, - "learning_rate": 1.1185379972256105e-09, - "loss": 0.9048, - "step": 5173 - }, - { - "epoch": 1.0, - "learning_rate": 1.0272699663782525e-09, - "loss": 0.8124, - "step": 5174 - }, - { - "epoch": 1.0, - "learning_rate": 9.398854213193976e-10, - "loss": 0.7947, - "step": 5175 - }, - { - "epoch": 1.0, - "learning_rate": 8.563843959907836e-10, - "loss": 0.797, - "step": 5176 - }, - { - "epoch": 1.0, - "learning_rate": 7.767669228231356e-10, - "loss": 0.8328, - "step": 5177 - }, - { - "epoch": 1.0, - "learning_rate": 7.010330327372749e-10, - "loss": 0.9072, - "step": 5178 - }, - { - "epoch": 1.0, - "learning_rate": 6.291827551474505e-10, - "loss": 0.859, - "step": 5179 - }, - { - "epoch": 1.0, - "learning_rate": 5.612161179613385e-10, - "loss": 0.7899, - "step": 5180 - }, - { - "epoch": 1.0, - "learning_rate": 4.971331475756013e-10, - "loss": 0.8118, - "step": 5181 - }, - { - "epoch": 1.0, - "learning_rate": 4.369338688781088e-10, - "loss": 0.8617, - "step": 5182 - }, - { - "epoch": 1.0, - "learning_rate": 3.8061830525126797e-10, - "loss": 0.8898, - "step": 5183 - }, - { - "epoch": 1.0, - "learning_rate": 3.2818647856647234e-10, - "loss": 0.8209, - "step": 5184 - }, - { - "epoch": 1.0, - "learning_rate": 2.796384091885429e-10, - "loss": 0.8346, - "step": 5185 - }, - { - "epoch": 1.0, - "learning_rate": 2.3497411597128705e-10, - "loss": 0.8924, - "step": 5186 - }, - { - "epoch": 1.0, - "learning_rate": 1.9419361626416e-10, - "loss": 0.8506, - "step": 5187 - }, { "epoch": 1.0, - "learning_rate": 1.5729692590338296e-10, - "loss": 0.8278, - "step": 5188 - }, - { - "epoch": 1.0, - "learning_rate": 1.2428405921971476e-10, - "loss": 0.8309, - "step": 5189 - }, - { - "epoch": 1.0, - "learning_rate": 9.515502903734153e-11, - "loss": 0.8824, - "step": 5190 - }, - { - "epoch": 1.0, - "learning_rate": 6.990984666610523e-11, - "loss": 0.6772, - "step": 5191 - }, - { - "epoch": 1.0, - "learning_rate": 4.854852191371606e-11, - "loss": 0.8527, - "step": 5192 - }, - { - "epoch": 1.0, - "learning_rate": 3.1071063075760466e-11, - "loss": 0.9254, - "step": 5193 - }, - { - "epoch": 1.0, - "learning_rate": 1.7477476940142013e-11, - "loss": 0.8892, - "step": 5194 - }, - { - "epoch": 1.0, - "learning_rate": 7.767768785971185e-12, - "loss": 0.8313, - "step": 5195 - }, - { - "epoch": 1.0, - "learning_rate": 1.9419423846755993e-12, - "loss": 0.8945, - "step": 5196 - }, - { - "epoch": 1.0, - "learning_rate": 0.0, - "loss": 0.6747, - "step": 5197 - }, - { - "epoch": 1.0, - "step": 5197, - "total_flos": 8203109849497600.0, - "train_loss": 0.9009988033957314, - "train_runtime": 69301.2495, - "train_samples_per_second": 9.6, - "train_steps_per_second": 0.075 + "step": 2599, + "total_flos": 8414358899261440.0, + "train_loss": 0.621661503952987, + "train_runtime": 57192.6099, + "train_samples_per_second": 11.633, + "train_steps_per_second": 0.045 } ], "logging_steps": 1.0, - "max_steps": 5197, + "max_steps": 2599, "num_input_tokens_seen": 0, "num_train_epochs": 1, - "save_steps": 10000, - "total_flos": 8203109849497600.0, - "train_batch_size": 4, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 8414358899261440.0, + "train_batch_size": 8, "trial_name": null, "trial_params": null }