{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 1.1737089201877935,
  "eval_steps": 500,
  "global_step": 1000,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0011737089201877935,
      "grad_norm": 0.27773135900497437,
      "learning_rate": 4e-05,
      "loss": 1.1957,
      "step": 1
    },
    {
      "epoch": 0.002347417840375587,
      "grad_norm": 0.26547771692276,
      "learning_rate": 8e-05,
      "loss": 1.1284,
      "step": 2
    },
    {
      "epoch": 0.0035211267605633804,
      "grad_norm": 0.236787810921669,
      "learning_rate": 0.00012,
      "loss": 1.1823,
      "step": 3
    },
    {
      "epoch": 0.004694835680751174,
      "grad_norm": 0.2459038347005844,
      "learning_rate": 0.00016,
      "loss": 1.1409,
      "step": 4
    },
    {
      "epoch": 0.005868544600938967,
      "grad_norm": 0.2526487410068512,
      "learning_rate": 0.0002,
      "loss": 1.12,
      "step": 5
    },
    {
      "epoch": 0.007042253521126761,
      "grad_norm": 0.2795103192329407,
      "learning_rate": 0.00019976387249114524,
      "loss": 1.1579,
      "step": 6
    },
    {
      "epoch": 0.008215962441314555,
      "grad_norm": 0.2876183092594147,
      "learning_rate": 0.00019952774498229045,
      "loss": 1.1211,
      "step": 7
    },
    {
      "epoch": 0.009389671361502348,
      "grad_norm": 0.3014296293258667,
      "learning_rate": 0.00019929161747343565,
      "loss": 1.1118,
      "step": 8
    },
    {
      "epoch": 0.01056338028169014,
      "grad_norm": 0.29106494784355164,
      "learning_rate": 0.00019905548996458088,
      "loss": 1.1787,
      "step": 9
    },
    {
      "epoch": 0.011737089201877934,
      "grad_norm": 0.3211474120616913,
      "learning_rate": 0.00019881936245572609,
      "loss": 1.1004,
      "step": 10
    },
    {
      "epoch": 0.012910798122065728,
      "grad_norm": 0.3358176350593567,
      "learning_rate": 0.00019858323494687132,
      "loss": 1.1099,
      "step": 11
    },
    {
      "epoch": 0.014084507042253521,
      "grad_norm": 0.3236922323703766,
      "learning_rate": 0.00019834710743801655,
      "loss": 1.048,
      "step": 12
    },
    {
      "epoch": 0.015258215962441314,
      "grad_norm": 0.31388312578201294,
      "learning_rate": 0.00019811097992916175,
      "loss": 1.0532,
      "step": 13
    },
    {
      "epoch": 0.01643192488262911,
      "grad_norm": 0.320402055978775,
      "learning_rate": 0.00019787485242030696,
      "loss": 1.0757,
      "step": 14
    },
    {
      "epoch": 0.017605633802816902,
      "grad_norm": 0.32999494671821594,
      "learning_rate": 0.0001976387249114522,
      "loss": 1.122,
      "step": 15
    },
    {
      "epoch": 0.018779342723004695,
      "grad_norm": 0.30936214327812195,
      "learning_rate": 0.00019740259740259742,
      "loss": 1.1156,
      "step": 16
    },
    {
      "epoch": 0.01995305164319249,
      "grad_norm": 0.2863931953907013,
      "learning_rate": 0.00019716646989374263,
      "loss": 1.0414,
      "step": 17
    },
    {
      "epoch": 0.02112676056338028,
      "grad_norm": 0.29143351316452026,
      "learning_rate": 0.00019693034238488786,
      "loss": 1.0379,
      "step": 18
    },
    {
      "epoch": 0.022300469483568074,
      "grad_norm": 0.28874626755714417,
      "learning_rate": 0.0001966942148760331,
      "loss": 1.0388,
      "step": 19
    },
    {
      "epoch": 0.023474178403755867,
      "grad_norm": 0.30588293075561523,
      "learning_rate": 0.00019645808736717827,
      "loss": 1.0515,
      "step": 20
    },
    {
      "epoch": 0.02464788732394366,
      "grad_norm": 0.29231536388397217,
      "learning_rate": 0.0001962219598583235,
      "loss": 1.0472,
      "step": 21
    },
    {
      "epoch": 0.025821596244131457,
      "grad_norm": 0.2783581614494324,
      "learning_rate": 0.00019598583234946873,
      "loss": 1.0608,
      "step": 22
    },
    {
      "epoch": 0.02699530516431925,
      "grad_norm": 0.29816293716430664,
      "learning_rate": 0.00019574970484061393,
      "loss": 1.0986,
      "step": 23
    },
    {
      "epoch": 0.028169014084507043,
      "grad_norm": 0.27919578552246094,
      "learning_rate": 0.00019551357733175916,
      "loss": 1.0265,
      "step": 24
    },
    {
      "epoch": 0.029342723004694836,
      "grad_norm": 0.3144524097442627,
      "learning_rate": 0.00019527744982290437,
      "loss": 1.0699,
      "step": 25
    },
    {
      "epoch": 0.03051643192488263,
      "grad_norm": 0.3090282678604126,
      "learning_rate": 0.0001950413223140496,
      "loss": 1.0601,
      "step": 26
    },
    {
      "epoch": 0.03169014084507042,
      "grad_norm": 0.30304697155952454,
      "learning_rate": 0.0001948051948051948,
      "loss": 1.0926,
      "step": 27
    },
    {
      "epoch": 0.03286384976525822,
      "grad_norm": 0.29015883803367615,
      "learning_rate": 0.00019456906729634004,
      "loss": 1.0,
      "step": 28
    },
    {
      "epoch": 0.03403755868544601,
      "grad_norm": 0.29359501600265503,
      "learning_rate": 0.00019433293978748527,
      "loss": 0.988,
      "step": 29
    },
    {
      "epoch": 0.035211267605633804,
      "grad_norm": 0.2772333323955536,
      "learning_rate": 0.00019409681227863047,
      "loss": 0.9758,
      "step": 30
    },
    {
      "epoch": 0.036384976525821594,
      "grad_norm": 0.2761421799659729,
      "learning_rate": 0.00019386068476977568,
      "loss": 0.9926,
      "step": 31
    },
    {
      "epoch": 0.03755868544600939,
      "grad_norm": NaN,
      "learning_rate": 0.00019386068476977568,
      "loss": 1.0944,
      "step": 32
    },
    {
      "epoch": 0.03873239436619718,
      "grad_norm": 0.2766799330711365,
      "learning_rate": 0.0001936245572609209,
      "loss": 0.9813,
      "step": 33
    },
    {
      "epoch": 0.03990610328638498,
      "grad_norm": 0.28922533988952637,
      "learning_rate": 0.0001933884297520661,
      "loss": 0.9839,
      "step": 34
    },
    {
      "epoch": 0.04107981220657277,
      "grad_norm": 0.28271371126174927,
      "learning_rate": 0.00019315230224321134,
      "loss": 1.0125,
      "step": 35
    },
    {
      "epoch": 0.04225352112676056,
      "grad_norm": 0.2955509424209595,
      "learning_rate": 0.00019291617473435658,
      "loss": 1.0049,
      "step": 36
    },
    {
      "epoch": 0.04342723004694836,
      "grad_norm": 0.2909109592437744,
      "learning_rate": 0.00019268004722550178,
      "loss": 1.0015,
      "step": 37
    },
    {
      "epoch": 0.04460093896713615,
      "grad_norm": 0.29657021164894104,
      "learning_rate": 0.00019244391971664698,
      "loss": 1.0107,
      "step": 38
    },
    {
      "epoch": 0.045774647887323945,
      "grad_norm": 0.29010507464408875,
      "learning_rate": 0.00019220779220779222,
      "loss": 0.9918,
      "step": 39
    },
    {
      "epoch": 0.046948356807511735,
      "grad_norm": 0.2906627058982849,
      "learning_rate": 0.00019197166469893745,
      "loss": 0.9843,
      "step": 40
    },
    {
      "epoch": 0.04812206572769953,
      "grad_norm": 0.2919193208217621,
      "learning_rate": 0.00019173553719008265,
      "loss": 0.9889,
      "step": 41
    },
    {
      "epoch": 0.04929577464788732,
      "grad_norm": 0.3219091296195984,
      "learning_rate": 0.00019149940968122788,
      "loss": 0.9979,
      "step": 42
    },
    {
      "epoch": 0.05046948356807512,
      "grad_norm": 0.29512301087379456,
      "learning_rate": 0.0001912632821723731,
      "loss": 0.989,
      "step": 43
    },
    {
      "epoch": 0.051643192488262914,
      "grad_norm": 0.3190619647502899,
      "learning_rate": 0.0001910271546635183,
      "loss": 0.9563,
      "step": 44
    },
    {
      "epoch": 0.0528169014084507,
      "grad_norm": 0.310253381729126,
      "learning_rate": 0.00019079102715466352,
      "loss": 1.037,
      "step": 45
    },
    {
      "epoch": 0.0539906103286385,
      "grad_norm": 0.3140093684196472,
      "learning_rate": 0.00019055489964580876,
      "loss": 0.9687,
      "step": 46
    },
    {
      "epoch": 0.05516431924882629,
      "grad_norm": 0.2816644310951233,
      "learning_rate": 0.00019031877213695396,
      "loss": 0.9372,
      "step": 47
    },
    {
      "epoch": 0.056338028169014086,
      "grad_norm": 0.3012441396713257,
      "learning_rate": 0.0001900826446280992,
      "loss": 0.9968,
      "step": 48
    },
    {
      "epoch": 0.057511737089201875,
      "grad_norm": 0.29789185523986816,
      "learning_rate": 0.0001898465171192444,
      "loss": 0.9143,
      "step": 49
    },
    {
      "epoch": 0.05868544600938967,
      "grad_norm": 0.29454007744789124,
      "learning_rate": 0.00018961038961038963,
      "loss": 0.9837,
      "step": 50
    },
    {
      "epoch": 0.05985915492957746,
      "grad_norm": 0.321218341588974,
      "learning_rate": 0.00018937426210153483,
      "loss": 1.0135,
      "step": 51
    },
    {
      "epoch": 0.06103286384976526,
      "grad_norm": 0.30039164423942566,
      "learning_rate": 0.00018913813459268006,
      "loss": 0.9639,
      "step": 52
    },
    {
      "epoch": 0.062206572769953054,
      "grad_norm": 0.3052615225315094,
      "learning_rate": 0.0001889020070838253,
      "loss": 0.9401,
      "step": 53
    },
    {
      "epoch": 0.06338028169014084,
      "grad_norm": 0.3177138864994049,
      "learning_rate": 0.00018866587957497047,
      "loss": 0.9626,
      "step": 54
    },
    {
      "epoch": 0.06455399061032864,
      "grad_norm": 0.3098903298377991,
      "learning_rate": 0.0001884297520661157,
      "loss": 0.9535,
      "step": 55
    },
    {
      "epoch": 0.06572769953051644,
      "grad_norm": 0.33165299892425537,
      "learning_rate": 0.00018819362455726094,
      "loss": 1.0475,
      "step": 56
    },
    {
      "epoch": 0.06690140845070422,
      "grad_norm": 0.3054540455341339,
      "learning_rate": 0.00018795749704840614,
      "loss": 0.988,
      "step": 57
    },
    {
      "epoch": 0.06807511737089202,
      "grad_norm": 0.3412969708442688,
      "learning_rate": 0.00018772136953955137,
      "loss": 0.9531,
      "step": 58
    },
    {
      "epoch": 0.06924882629107981,
      "grad_norm": 0.3173505663871765,
      "learning_rate": 0.0001874852420306966,
      "loss": 1.0037,
      "step": 59
    },
    {
      "epoch": 0.07042253521126761,
      "grad_norm": 0.29377281665802,
      "learning_rate": 0.0001872491145218418,
      "loss": 0.9205,
      "step": 60
    },
    {
      "epoch": 0.0715962441314554,
      "grad_norm": 0.2970433831214905,
      "learning_rate": 0.000187012987012987,
      "loss": 0.8902,
      "step": 61
    },
    {
      "epoch": 0.07276995305164319,
      "grad_norm": 0.3081493675708771,
      "learning_rate": 0.00018677685950413224,
      "loss": 0.9498,
      "step": 62
    },
    {
      "epoch": 0.07394366197183098,
      "grad_norm": 0.31438371539115906,
      "learning_rate": 0.00018654073199527747,
      "loss": 0.9406,
      "step": 63
    },
    {
      "epoch": 0.07511737089201878,
      "grad_norm": 0.29640915989875793,
      "learning_rate": 0.00018630460448642268,
      "loss": 0.8948,
      "step": 64
    },
    {
      "epoch": 0.07629107981220658,
      "grad_norm": 0.33342233300209045,
      "learning_rate": 0.00018606847697756788,
      "loss": 0.941,
      "step": 65
    },
    {
      "epoch": 0.07746478873239436,
      "grad_norm": 0.31546634435653687,
      "learning_rate": 0.00018583234946871312,
      "loss": 0.9392,
      "step": 66
    },
    {
      "epoch": 0.07863849765258216,
      "grad_norm": 0.31528937816619873,
      "learning_rate": 0.00018559622195985832,
      "loss": 0.9293,
      "step": 67
    },
    {
      "epoch": 0.07981220657276995,
      "grad_norm": 0.33473101258277893,
      "learning_rate": 0.00018536009445100355,
      "loss": 0.9214,
      "step": 68
    },
    {
      "epoch": 0.08098591549295775,
      "grad_norm": 0.6588060259819031,
      "learning_rate": 0.00018512396694214878,
      "loss": 0.944,
      "step": 69
    },
    {
      "epoch": 0.08215962441314555,
      "grad_norm": 0.30120280385017395,
      "learning_rate": 0.000184887839433294,
      "loss": 0.9171,
      "step": 70
    },
    {
      "epoch": 0.08333333333333333,
      "grad_norm": 0.3417011499404907,
      "learning_rate": 0.0001846517119244392,
      "loss": 0.9382,
      "step": 71
    },
    {
      "epoch": 0.08450704225352113,
      "grad_norm": 0.3202987313270569,
      "learning_rate": 0.00018441558441558442,
      "loss": 0.931,
      "step": 72
    },
    {
      "epoch": 0.08568075117370892,
      "grad_norm": 0.3390517234802246,
      "learning_rate": 0.00018417945690672965,
      "loss": 0.9218,
      "step": 73
    },
    {
      "epoch": 0.08685446009389672,
      "grad_norm": 0.32109472155570984,
      "learning_rate": 0.00018394332939787486,
      "loss": 0.9226,
      "step": 74
    },
    {
      "epoch": 0.0880281690140845,
      "grad_norm": 0.3435365855693817,
      "learning_rate": 0.0001837072018890201,
      "loss": 0.9402,
      "step": 75
    },
    {
      "epoch": 0.0892018779342723,
      "grad_norm": 0.3335697054862976,
      "learning_rate": 0.00018347107438016532,
      "loss": 0.9385,
      "step": 76
    },
    {
      "epoch": 0.0903755868544601,
      "grad_norm": 0.32050758600234985,
      "learning_rate": 0.0001832349468713105,
      "loss": 0.8992,
      "step": 77
    },
    {
      "epoch": 0.09154929577464789,
      "grad_norm": 0.32620421051979065,
      "learning_rate": 0.00018299881936245573,
      "loss": 0.9476,
      "step": 78
    },
    {
      "epoch": 0.09272300469483569,
      "grad_norm": 0.33306750655174255,
      "learning_rate": 0.00018276269185360096,
      "loss": 0.9458,
      "step": 79
    },
    {
      "epoch": 0.09389671361502347,
      "grad_norm": 0.3500649034976959,
      "learning_rate": 0.00018252656434474617,
      "loss": 0.9612,
      "step": 80
    },
    {
      "epoch": 0.09507042253521127,
      "grad_norm": 0.3186359405517578,
      "learning_rate": 0.0001822904368358914,
      "loss": 0.9527,
      "step": 81
    },
    {
      "epoch": 0.09624413145539906,
      "grad_norm": 0.3317716717720032,
      "learning_rate": 0.0001820543093270366,
      "loss": 0.9648,
      "step": 82
    },
    {
      "epoch": 0.09741784037558686,
      "grad_norm": 0.3196907639503479,
      "learning_rate": 0.00018181818181818183,
      "loss": 0.9643,
      "step": 83
    },
    {
      "epoch": 0.09859154929577464,
      "grad_norm": 0.3195818066596985,
      "learning_rate": 0.00018158205430932704,
      "loss": 0.9121,
      "step": 84
    },
    {
      "epoch": 0.09976525821596244,
      "grad_norm": 0.33151793479919434,
      "learning_rate": 0.00018134592680047227,
      "loss": 0.9051,
      "step": 85
    },
    {
      "epoch": 0.10093896713615023,
      "grad_norm": 0.3110804259777069,
      "learning_rate": 0.00018110979929161747,
      "loss": 0.9241,
      "step": 86
    },
    {
      "epoch": 0.10211267605633803,
      "grad_norm": 0.34278568625450134,
      "learning_rate": 0.0001808736717827627,
      "loss": 0.9634,
      "step": 87
    },
    {
      "epoch": 0.10328638497652583,
      "grad_norm": 0.34013500809669495,
      "learning_rate": 0.0001806375442739079,
      "loss": 0.8822,
      "step": 88
    },
    {
      "epoch": 0.10446009389671361,
      "grad_norm": 0.3449755012989044,
      "learning_rate": 0.00018040141676505314,
      "loss": 0.969,
      "step": 89
    },
    {
      "epoch": 0.1056338028169014,
      "grad_norm": 0.3166862726211548,
      "learning_rate": 0.00018016528925619835,
      "loss": 0.885,
      "step": 90
    },
    {
      "epoch": 0.1068075117370892,
      "grad_norm": 0.3260084092617035,
      "learning_rate": 0.00017992916174734358,
      "loss": 0.8908,
      "step": 91
    },
    {
      "epoch": 0.107981220657277,
      "grad_norm": 0.32791605591773987,
      "learning_rate": 0.0001796930342384888,
      "loss": 0.8822,
      "step": 92
    },
    {
      "epoch": 0.10915492957746478,
      "grad_norm": 0.31909653544425964,
      "learning_rate": 0.000179456906729634,
      "loss": 0.8463,
      "step": 93
    },
    {
      "epoch": 0.11032863849765258,
      "grad_norm": 0.3413308262825012,
      "learning_rate": 0.00017922077922077922,
      "loss": 0.9232,
      "step": 94
    },
    {
      "epoch": 0.11150234741784038,
      "grad_norm": 0.32644134759902954,
      "learning_rate": 0.00017898465171192445,
      "loss": 0.9113,
      "step": 95
    },
    {
      "epoch": 0.11267605633802817,
      "grad_norm": 0.33090126514434814,
      "learning_rate": 0.00017874852420306965,
      "loss": 0.9286,
      "step": 96
    },
    {
      "epoch": 0.11384976525821597,
      "grad_norm": 0.37200361490249634,
      "learning_rate": 0.00017851239669421489,
      "loss": 0.9239,
      "step": 97
    },
    {
      "epoch": 0.11502347417840375,
      "grad_norm": 0.3274000585079193,
      "learning_rate": 0.00017827626918536012,
      "loss": 0.9038,
      "step": 98
    },
    {
      "epoch": 0.11619718309859155,
      "grad_norm": 0.3768482506275177,
      "learning_rate": 0.00017804014167650532,
      "loss": 0.8558,
      "step": 99
    },
    {
      "epoch": 0.11737089201877934,
      "grad_norm": 0.32970595359802246,
      "learning_rate": 0.00017780401416765053,
      "loss": 0.9057,
      "step": 100
    },
    {
      "epoch": 0.11854460093896714,
      "grad_norm": 0.37230944633483887,
      "learning_rate": 0.00017756788665879576,
      "loss": 0.9211,
      "step": 101
    },
    {
      "epoch": 0.11971830985915492,
      "grad_norm": 0.352201372385025,
      "learning_rate": 0.000177331759149941,
      "loss": 0.9497,
      "step": 102
    },
    {
      "epoch": 0.12089201877934272,
      "grad_norm": 0.363364577293396,
      "learning_rate": 0.0001770956316410862,
      "loss": 0.9535,
      "step": 103
    },
    {
      "epoch": 0.12206572769953052,
      "grad_norm": 0.3388724625110626,
      "learning_rate": 0.00017685950413223143,
      "loss": 0.8908,
      "step": 104
    },
    {
      "epoch": 0.12323943661971831,
      "grad_norm": 0.34684258699417114,
      "learning_rate": 0.00017662337662337663,
      "loss": 0.8981,
      "step": 105
    },
    {
      "epoch": 0.12441314553990611,
      "grad_norm": 0.31892621517181396,
      "learning_rate": 0.00017638724911452183,
      "loss": 0.8461,
      "step": 106
    },
    {
      "epoch": 0.1255868544600939,
      "grad_norm": 0.32913845777511597,
      "learning_rate": 0.00017615112160566707,
      "loss": 0.9087,
      "step": 107
    },
    {
      "epoch": 0.1267605633802817,
      "grad_norm": 0.3695410490036011,
      "learning_rate": 0.0001759149940968123,
      "loss": 0.8899,
      "step": 108
    },
    {
      "epoch": 0.12793427230046947,
      "grad_norm": 0.3455798923969269,
      "learning_rate": 0.0001756788665879575,
      "loss": 0.9045,
      "step": 109
    },
    {
      "epoch": 0.12910798122065728,
      "grad_norm": 0.3612275719642639,
      "learning_rate": 0.0001754427390791027,
      "loss": 0.8861,
      "step": 110
    },
    {
      "epoch": 0.13028169014084506,
      "grad_norm": 0.4106651544570923,
      "learning_rate": 0.00017520661157024794,
      "loss": 0.9152,
      "step": 111
    },
    {
      "epoch": 0.13145539906103287,
      "grad_norm": 0.3604993224143982,
      "learning_rate": 0.00017497048406139317,
      "loss": 0.9141,
      "step": 112
    },
    {
      "epoch": 0.13262910798122066,
      "grad_norm": 0.3496919870376587,
      "learning_rate": 0.00017473435655253837,
      "loss": 0.9061,
      "step": 113
    },
    {
      "epoch": 0.13380281690140844,
      "grad_norm": 0.33643972873687744,
      "learning_rate": 0.0001744982290436836,
      "loss": 0.8877,
      "step": 114
    },
    {
      "epoch": 0.13497652582159625,
      "grad_norm": 0.33064204454421997,
      "learning_rate": 0.00017426210153482884,
      "loss": 0.8967,
      "step": 115
    },
    {
      "epoch": 0.13615023474178403,
      "grad_norm": 0.37868356704711914,
      "learning_rate": 0.00017402597402597401,
      "loss": 0.8957,
      "step": 116
    },
    {
      "epoch": 0.13732394366197184,
      "grad_norm": 0.34379109740257263,
      "learning_rate": 0.00017378984651711925,
      "loss": 0.9332,
      "step": 117
    },
    {
      "epoch": 0.13849765258215962,
      "grad_norm": 0.37193912267684937,
      "learning_rate": 0.00017355371900826448,
      "loss": 0.9513,
      "step": 118
    },
    {
      "epoch": 0.1396713615023474,
      "grad_norm": 0.33701232075691223,
      "learning_rate": 0.00017331759149940968,
      "loss": 0.8946,
      "step": 119
    },
    {
      "epoch": 0.14084507042253522,
      "grad_norm": 0.35765206813812256,
      "learning_rate": 0.0001730814639905549,
      "loss": 0.8931,
      "step": 120
    },
    {
      "epoch": 0.142018779342723,
      "grad_norm": 0.3511311411857605,
      "learning_rate": 0.00017284533648170012,
      "loss": 0.9042,
      "step": 121
    },
    {
      "epoch": 0.1431924882629108,
      "grad_norm": 0.33516445755958557,
      "learning_rate": 0.00017260920897284535,
      "loss": 0.8564,
      "step": 122
    },
    {
      "epoch": 0.1443661971830986,
      "grad_norm": 0.385959267616272,
      "learning_rate": 0.00017237308146399055,
      "loss": 0.963,
      "step": 123
    },
    {
      "epoch": 0.14553990610328638,
      "grad_norm": 0.34608641266822815,
      "learning_rate": 0.00017213695395513578,
      "loss": 0.8666,
      "step": 124
    },
    {
      "epoch": 0.1467136150234742,
      "grad_norm": 0.3705556392669678,
      "learning_rate": 0.00017190082644628102,
      "loss": 0.7783,
      "step": 125
    },
    {
      "epoch": 0.14788732394366197,
      "grad_norm": 0.3213210701942444,
      "learning_rate": 0.00017166469893742622,
      "loss": 0.8428,
      "step": 126
    },
    {
      "epoch": 0.14906103286384975,
      "grad_norm": 0.3903498351573944,
      "learning_rate": 0.00017142857142857143,
      "loss": 0.8418,
      "step": 127
    },
    {
      "epoch": 0.15023474178403756,
      "grad_norm": 0.3556365668773651,
      "learning_rate": 0.00017119244391971666,
      "loss": 0.8612,
      "step": 128
    },
    {
      "epoch": 0.15140845070422534,
      "grad_norm": 0.3734995424747467,
      "learning_rate": 0.00017095631641086186,
      "loss": 0.8845,
      "step": 129
    },
    {
      "epoch": 0.15258215962441316,
      "grad_norm": 0.33735260367393494,
      "learning_rate": 0.0001707201889020071,
      "loss": 0.8752,
      "step": 130
    },
    {
      "epoch": 0.15375586854460094,
      "grad_norm": 0.38340267539024353,
      "learning_rate": 0.00017048406139315232,
      "loss": 0.8847,
      "step": 131
    },
    {
      "epoch": 0.15492957746478872,
      "grad_norm": 0.3654419779777527,
      "learning_rate": 0.00017024793388429753,
      "loss": 0.8448,
      "step": 132
    },
    {
      "epoch": 0.15610328638497653,
      "grad_norm": 0.3601568341255188,
      "learning_rate": 0.00017001180637544273,
      "loss": 0.8981,
      "step": 133
    },
    {
      "epoch": 0.1572769953051643,
      "grad_norm": 0.40733832120895386,
      "learning_rate": 0.00016977567886658796,
      "loss": 0.9135,
      "step": 134
    },
    {
      "epoch": 0.15845070422535212,
      "grad_norm": 0.34627673029899597,
      "learning_rate": 0.0001695395513577332,
      "loss": 0.9164,
      "step": 135
    },
    {
      "epoch": 0.1596244131455399,
      "grad_norm": 0.3865872621536255,
      "learning_rate": 0.0001693034238488784,
      "loss": 0.9222,
      "step": 136
    },
    {
      "epoch": 0.1607981220657277,
      "grad_norm": 0.4011456072330475,
      "learning_rate": 0.00016906729634002363,
      "loss": 0.8843,
      "step": 137
    },
    {
      "epoch": 0.1619718309859155,
      "grad_norm": 0.32259878516197205,
      "learning_rate": 0.00016883116883116884,
      "loss": 0.8427,
      "step": 138
    },
    {
      "epoch": 0.16314553990610328,
      "grad_norm": 0.3807618319988251,
      "learning_rate": 0.00016859504132231404,
      "loss": 0.8684,
      "step": 139
    },
    {
      "epoch": 0.1643192488262911,
      "grad_norm": 0.3658106327056885,
      "learning_rate": 0.00016835891381345927,
      "loss": 0.9024,
      "step": 140
    },
    {
      "epoch": 0.16549295774647887,
      "grad_norm": 0.3638787865638733,
      "learning_rate": 0.0001681227863046045,
      "loss": 0.8582,
      "step": 141
    },
    {
      "epoch": 0.16666666666666666,
      "grad_norm": 0.3839091360569,
      "learning_rate": 0.0001678866587957497,
      "loss": 0.8543,
      "step": 142
    },
    {
      "epoch": 0.16784037558685447,
      "grad_norm": 0.33579927682876587,
      "learning_rate": 0.00016765053128689494,
      "loss": 0.8765,
      "step": 143
    },
    {
      "epoch": 0.16901408450704225,
      "grad_norm": 0.35091203451156616,
      "learning_rate": 0.00016741440377804014,
      "loss": 0.8504,
      "step": 144
    },
    {
      "epoch": 0.17018779342723006,
      "grad_norm": 0.35823047161102295,
      "learning_rate": 0.00016717827626918538,
      "loss": 0.8534,
      "step": 145
    },
    {
      "epoch": 0.17136150234741784,
      "grad_norm": 0.37154486775398254,
      "learning_rate": 0.00016694214876033058,
      "loss": 0.851,
      "step": 146
    },
    {
      "epoch": 0.17253521126760563,
      "grad_norm": 0.33140066266059875,
      "learning_rate": 0.0001667060212514758,
      "loss": 0.8136,
      "step": 147
    },
    {
      "epoch": 0.17370892018779344,
      "grad_norm": 0.37408292293548584,
      "learning_rate": 0.00016646989374262104,
      "loss": 0.8933,
      "step": 148
    },
    {
      "epoch": 0.17488262910798122,
      "grad_norm": 0.36203357577323914,
      "learning_rate": 0.00016623376623376625,
      "loss": 0.8747,
      "step": 149
    },
    {
      "epoch": 0.176056338028169,
      "grad_norm": 0.35033532977104187,
      "learning_rate": 0.00016599763872491145,
      "loss": 0.8273,
      "step": 150
    },
    {
      "epoch": 0.1772300469483568,
      "grad_norm": 0.345048189163208,
      "learning_rate": 0.00016576151121605668,
      "loss": 0.8698,
      "step": 151
    },
    {
      "epoch": 0.1784037558685446,
      "grad_norm": 0.3592989146709442,
      "learning_rate": 0.0001655253837072019,
      "loss": 0.8483,
      "step": 152
    },
    {
      "epoch": 0.1795774647887324,
      "grad_norm": 0.3685864806175232,
      "learning_rate": 0.00016528925619834712,
      "loss": 0.915,
      "step": 153
    },
    {
      "epoch": 0.1807511737089202,
      "grad_norm": 0.3427909314632416,
      "learning_rate": 0.00016505312868949235,
      "loss": 0.8321,
      "step": 154
    },
    {
      "epoch": 0.18192488262910797,
      "grad_norm": 0.34697192907333374,
      "learning_rate": 0.00016481700118063756,
      "loss": 0.8801,
      "step": 155
    },
    {
      "epoch": 0.18309859154929578,
      "grad_norm": 0.3387276530265808,
      "learning_rate": 0.00016458087367178276,
      "loss": 0.8237,
      "step": 156
    },
    {
      "epoch": 0.18427230046948356,
      "grad_norm": 0.3547775447368622,
      "learning_rate": 0.000164344746162928,
      "loss": 0.8645,
      "step": 157
    },
    {
      "epoch": 0.18544600938967137,
      "grad_norm": 0.3342725932598114,
      "learning_rate": 0.00016410861865407322,
      "loss": 0.82,
      "step": 158
    },
    {
      "epoch": 0.18661971830985916,
      "grad_norm": 0.4317960739135742,
      "learning_rate": 0.00016387249114521843,
      "loss": 0.8614,
      "step": 159
    },
    {
      "epoch": 0.18779342723004694,
      "grad_norm": 0.35031062364578247,
      "learning_rate": 0.00016363636363636366,
      "loss": 0.8193,
      "step": 160
    },
    {
      "epoch": 0.18896713615023475,
      "grad_norm": 0.3616986572742462,
      "learning_rate": 0.00016340023612750886,
      "loss": 0.8571,
      "step": 161
    },
    {
      "epoch": 0.19014084507042253,
      "grad_norm": 0.36284518241882324,
      "learning_rate": 0.00016316410861865407,
      "loss": 0.8555,
      "step": 162
    },
    {
      "epoch": 0.19131455399061034,
      "grad_norm": 0.42962291836738586,
      "learning_rate": 0.0001629279811097993,
      "loss": 0.8574,
      "step": 163
    },
    {
      "epoch": 0.19248826291079812,
      "grad_norm": 0.330268532037735,
      "learning_rate": 0.00016269185360094453,
      "loss": 0.8952,
      "step": 164
    },
    {
      "epoch": 0.1936619718309859,
      "grad_norm": 0.33917295932769775,
      "learning_rate": 0.00016245572609208974,
      "loss": 0.8588,
      "step": 165
    },
    {
      "epoch": 0.19483568075117372,
      "grad_norm": 0.3963412046432495,
      "learning_rate": 0.00016221959858323494,
      "loss": 0.8451,
      "step": 166
    },
    {
      "epoch": 0.1960093896713615,
      "grad_norm": 0.33864182233810425,
      "learning_rate": 0.00016198347107438017,
      "loss": 0.8734,
      "step": 167
    },
    {
      "epoch": 0.19718309859154928,
      "grad_norm": 0.3751653730869293,
      "learning_rate": 0.00016174734356552538,
      "loss": 0.8786,
      "step": 168
    },
    {
      "epoch": 0.1983568075117371,
      "grad_norm": 0.4138842821121216,
      "learning_rate": 0.0001615112160566706,
      "loss": 0.8608,
      "step": 169
    },
    {
      "epoch": 0.19953051643192488,
      "grad_norm": 0.3747748136520386,
      "learning_rate": 0.00016127508854781584,
      "loss": 0.8901,
      "step": 170
    },
    {
      "epoch": 0.2007042253521127,
      "grad_norm": 0.3302014172077179,
      "learning_rate": 0.00016103896103896104,
      "loss": 0.8538,
      "step": 171
    },
    {
      "epoch": 0.20187793427230047,
      "grad_norm": 0.36144372820854187,
      "learning_rate": 0.00016080283353010625,
      "loss": 0.8634,
      "step": 172
    },
    {
      "epoch": 0.20305164319248825,
      "grad_norm": 0.3579455018043518,
      "learning_rate": 0.00016056670602125148,
      "loss": 0.8536,
      "step": 173
    },
    {
      "epoch": 0.20422535211267606,
      "grad_norm": 0.3475671410560608,
      "learning_rate": 0.0001603305785123967,
      "loss": 0.8304,
      "step": 174
    },
    {
      "epoch": 0.20539906103286384,
      "grad_norm": 0.34114810824394226,
      "learning_rate": 0.00016009445100354192,
      "loss": 0.8276,
      "step": 175
    },
    {
      "epoch": 0.20657276995305165,
      "grad_norm": 0.32198190689086914,
      "learning_rate": 0.00015985832349468715,
      "loss": 0.815,
      "step": 176
    },
    {
      "epoch": 0.20774647887323944,
      "grad_norm": 0.4003874361515045,
      "learning_rate": 0.00015962219598583238,
      "loss": 0.8523,
      "step": 177
    },
    {
      "epoch": 0.20892018779342722,
      "grad_norm": 0.32290229201316833,
      "learning_rate": 0.00015938606847697756,
      "loss": 0.8465,
      "step": 178
    },
    {
      "epoch": 0.21009389671361503,
      "grad_norm": 0.35729506611824036,
      "learning_rate": 0.0001591499409681228,
      "loss": 0.8437,
      "step": 179
    },
    {
      "epoch": 0.2112676056338028,
      "grad_norm": 0.33743324875831604,
      "learning_rate": 0.00015891381345926802,
      "loss": 0.8351,
      "step": 180
    },
    {
      "epoch": 0.21244131455399062,
      "grad_norm": 0.34673774242401123,
      "learning_rate": 0.00015867768595041322,
      "loss": 0.8146,
      "step": 181
    },
    {
      "epoch": 0.2136150234741784,
      "grad_norm": 0.37883323431015015,
      "learning_rate": 0.00015844155844155845,
      "loss": 0.8889,
      "step": 182
    },
    {
      "epoch": 0.2147887323943662,
      "grad_norm": 0.34172534942626953,
      "learning_rate": 0.00015820543093270366,
      "loss": 0.8479,
      "step": 183
    },
    {
      "epoch": 0.215962441314554,
      "grad_norm": 0.39948219060897827,
      "learning_rate": 0.0001579693034238489,
      "loss": 0.8383,
      "step": 184
    },
    {
      "epoch": 0.21713615023474178,
      "grad_norm": 0.33746814727783203,
      "learning_rate": 0.0001577331759149941,
      "loss": 0.8713,
      "step": 185
    },
    {
      "epoch": 0.21830985915492956,
      "grad_norm": 0.34141069650650024,
      "learning_rate": 0.00015749704840613933,
      "loss": 0.8303,
      "step": 186
    },
    {
      "epoch": 0.21948356807511737,
      "grad_norm": 0.35994264483451843,
      "learning_rate": 0.00015726092089728456,
      "loss": 0.7919,
      "step": 187
    },
    {
      "epoch": 0.22065727699530516,
      "grad_norm": 0.34234684705734253,
      "learning_rate": 0.00015702479338842976,
      "loss": 0.8225,
      "step": 188
    },
    {
      "epoch": 0.22183098591549297,
      "grad_norm": 0.3601793050765991,
      "learning_rate": 0.00015678866587957497,
      "loss": 0.8395,
      "step": 189
    },
    {
      "epoch": 0.22300469483568075,
      "grad_norm": 0.3154338002204895,
      "learning_rate": 0.0001565525383707202,
      "loss": 0.7735,
      "step": 190
    },
    {
      "epoch": 0.22417840375586853,
      "grad_norm": 0.3758296072483063,
      "learning_rate": 0.0001563164108618654,
      "loss": 0.8241,
      "step": 191
    },
    {
      "epoch": 0.22535211267605634,
      "grad_norm": 0.3732200264930725,
      "learning_rate": 0.00015608028335301063,
      "loss": 0.8116,
      "step": 192
    },
    {
      "epoch": 0.22652582159624413,
      "grad_norm": 0.3601556718349457,
      "learning_rate": 0.00015584415584415587,
      "loss": 0.8242,
      "step": 193
    },
    {
      "epoch": 0.22769953051643194,
      "grad_norm": 0.360442191362381,
      "learning_rate": 0.00015560802833530107,
      "loss": 0.832,
      "step": 194
    },
    {
      "epoch": 0.22887323943661972,
      "grad_norm": 0.35598254203796387,
      "learning_rate": 0.00015537190082644627,
      "loss": 0.8938,
      "step": 195
    },
    {
      "epoch": 0.2300469483568075,
      "grad_norm": 0.3962613046169281,
      "learning_rate": 0.0001551357733175915,
      "loss": 0.8409,
      "step": 196
    },
    {
      "epoch": 0.2312206572769953,
      "grad_norm": 0.3521510064601898,
      "learning_rate": 0.00015489964580873674,
      "loss": 0.8298,
      "step": 197
    },
    {
      "epoch": 0.2323943661971831,
      "grad_norm": 0.34407946467399597,
      "learning_rate": 0.00015466351829988194,
      "loss": 0.7921,
      "step": 198
    },
    {
      "epoch": 0.2335680751173709,
      "grad_norm": 0.3572155237197876,
      "learning_rate": 0.00015442739079102717,
      "loss": 0.8997,
      "step": 199
    },
    {
      "epoch": 0.2347417840375587,
      "grad_norm": 0.345745712518692,
      "learning_rate": 0.00015419126328217238,
      "loss": 0.8563,
      "step": 200
    },
    {
      "epoch": 0.23591549295774647,
      "grad_norm": 0.3741077780723572,
      "learning_rate": 0.00015395513577331758,
      "loss": 0.8334,
      "step": 201
    },
    {
      "epoch": 0.23708920187793428,
      "grad_norm": 0.36866459250450134,
      "learning_rate": 0.00015371900826446281,
      "loss": 0.8398,
      "step": 202
    },
    {
      "epoch": 0.23826291079812206,
      "grad_norm": 0.3834739625453949,
      "learning_rate": 0.00015348288075560805,
      "loss": 0.8181,
      "step": 203
    },
    {
      "epoch": 0.23943661971830985,
      "grad_norm": 0.373045951128006,
      "learning_rate": 0.00015324675324675325,
      "loss": 0.8044,
      "step": 204
    },
    {
      "epoch": 0.24061032863849766,
      "grad_norm": 0.3418562412261963,
      "learning_rate": 0.00015301062573789848,
      "loss": 0.8454,
      "step": 205
    },
    {
      "epoch": 0.24178403755868544,
      "grad_norm": 0.36289098858833313,
      "learning_rate": 0.00015277449822904369,
      "loss": 0.8478,
      "step": 206
    },
    {
      "epoch": 0.24295774647887325,
      "grad_norm": 0.38806968927383423,
      "learning_rate": 0.00015253837072018892,
      "loss": 0.804,
      "step": 207
    },
    {
      "epoch": 0.24413145539906103,
      "grad_norm": 0.34217599034309387,
      "learning_rate": 0.00015230224321133412,
      "loss": 0.8391,
      "step": 208
    },
    {
      "epoch": 0.24530516431924881,
      "grad_norm": 0.3738957643508911,
      "learning_rate": 0.00015206611570247935,
      "loss": 0.9026,
      "step": 209
    },
    {
      "epoch": 0.24647887323943662,
      "grad_norm": 0.3481609523296356,
      "learning_rate": 0.00015182998819362458,
      "loss": 0.8674,
      "step": 210
    },
    {
      "epoch": 0.2476525821596244,
      "grad_norm": 0.38967254757881165,
      "learning_rate": 0.00015159386068476976,
      "loss": 0.8796,
      "step": 211
    },
    {
      "epoch": 0.24882629107981222,
      "grad_norm": 0.34841835498809814,
      "learning_rate": 0.000151357733175915,
      "loss": 0.7913,
      "step": 212
    },
    {
      "epoch": 0.25,
      "grad_norm": 0.33826395869255066,
      "learning_rate": 0.00015112160566706023,
      "loss": 0.8539,
      "step": 213
    },
    {
      "epoch": 0.2511737089201878,
      "grad_norm": 0.35131266713142395,
      "learning_rate": 0.00015088547815820543,
      "loss": 0.8072,
      "step": 214
    },
    {
      "epoch": 0.25234741784037557,
      "grad_norm": 0.3298250734806061,
      "learning_rate": 0.00015064935064935066,
      "loss": 0.7688,
      "step": 215
    },
    {
      "epoch": 0.2535211267605634,
      "grad_norm": 0.33808133006095886,
      "learning_rate": 0.0001504132231404959,
      "loss": 0.7609,
      "step": 216
    },
    {
      "epoch": 0.2546948356807512,
      "grad_norm": 0.37146687507629395,
      "learning_rate": 0.0001501770956316411,
      "loss": 0.843,
      "step": 217
    },
    {
      "epoch": 0.25586854460093894,
      "grad_norm": 0.33817118406295776,
      "learning_rate": 0.0001499409681227863,
      "loss": 0.7828,
      "step": 218
    },
    {
      "epoch": 0.25704225352112675,
      "grad_norm": 0.35203686356544495,
      "learning_rate": 0.00014970484061393153,
      "loss": 0.8236,
      "step": 219
    },
    {
      "epoch": 0.25821596244131456,
      "grad_norm": 0.34176716208457947,
      "learning_rate": 0.00014946871310507676,
      "loss": 0.8191,
      "step": 220
    },
    {
      "epoch": 0.25938967136150237,
      "grad_norm": 0.34649035334587097,
      "learning_rate": 0.00014923258559622197,
      "loss": 0.8284,
      "step": 221
    },
    {
      "epoch": 0.2605633802816901,
      "grad_norm": 0.35891467332839966,
      "learning_rate": 0.00014899645808736717,
      "loss": 0.8149,
      "step": 222
    },
    {
      "epoch": 0.26173708920187794,
      "grad_norm": 0.3408451974391937,
      "learning_rate": 0.0001487603305785124,
      "loss": 0.8049,
      "step": 223
    },
    {
      "epoch": 0.26291079812206575,
      "grad_norm": 0.36554664373397827,
      "learning_rate": 0.0001485242030696576,
      "loss": 0.8478,
      "step": 224
    },
    {
      "epoch": 0.2640845070422535,
      "grad_norm": 0.3355228304862976,
      "learning_rate": 0.00014828807556080284,
      "loss": 0.815,
      "step": 225
    },
    {
      "epoch": 0.2652582159624413,
      "grad_norm": 0.3500598669052124,
      "learning_rate": 0.00014805194805194807,
      "loss": 0.8571,
      "step": 226
    },
    {
      "epoch": 0.2664319248826291,
      "grad_norm": 0.3362652659416199,
      "learning_rate": 0.00014781582054309328,
      "loss": 0.8363,
      "step": 227
    },
    {
      "epoch": 0.2676056338028169,
      "grad_norm": 0.34258243441581726,
      "learning_rate": 0.00014757969303423848,
      "loss": 0.7648,
      "step": 228
    },
    {
      "epoch": 0.2687793427230047,
      "grad_norm": 0.34023317694664,
      "learning_rate": 0.0001473435655253837,
      "loss": 0.8373,
      "step": 229
    },
    {
      "epoch": 0.2699530516431925,
      "grad_norm": 0.35829535126686096,
      "learning_rate": 0.00014710743801652894,
      "loss": 0.8255,
      "step": 230
    },
    {
      "epoch": 0.2711267605633803,
      "grad_norm": 0.3499360978603363,
      "learning_rate": 0.00014687131050767415,
      "loss": 0.8514,
      "step": 231
    },
    {
      "epoch": 0.27230046948356806,
      "grad_norm": 0.3703480362892151,
      "learning_rate": 0.00014663518299881938,
      "loss": 0.8615,
      "step": 232
    },
    {
      "epoch": 0.2734741784037559,
      "grad_norm": 0.3460928499698639,
      "learning_rate": 0.0001463990554899646,
      "loss": 0.7891,
      "step": 233
    },
    {
      "epoch": 0.2746478873239437,
      "grad_norm": 0.34184372425079346,
      "learning_rate": 0.0001461629279811098,
      "loss": 0.8168,
      "step": 234
    },
    {
      "epoch": 0.27582159624413144,
      "grad_norm": 0.34520068764686584,
      "learning_rate": 0.00014592680047225502,
      "loss": 0.8271,
      "step": 235
    },
    {
      "epoch": 0.27699530516431925,
      "grad_norm": 0.3415423631668091,
      "learning_rate": 0.00014569067296340025,
      "loss": 0.783,
      "step": 236
    },
    {
      "epoch": 0.27816901408450706,
      "grad_norm": 0.34584441781044006,
      "learning_rate": 0.00014545454545454546,
      "loss": 0.8488,
      "step": 237
    },
    {
      "epoch": 0.2793427230046948,
      "grad_norm": 0.33898866176605225,
      "learning_rate": 0.0001452184179456907,
      "loss": 0.8786,
      "step": 238
    },
    {
      "epoch": 0.2805164319248826,
      "grad_norm": 0.3591814339160919,
      "learning_rate": 0.0001449822904368359,
      "loss": 0.8081,
      "step": 239
    },
    {
      "epoch": 0.28169014084507044,
      "grad_norm": 0.34305432438850403,
      "learning_rate": 0.0001447461629279811,
      "loss": 0.7911,
      "step": 240
    },
    {
      "epoch": 0.2828638497652582,
      "grad_norm": 0.35866865515708923,
      "learning_rate": 0.00014451003541912633,
      "loss": 0.8393,
      "step": 241
    },
    {
      "epoch": 0.284037558685446,
      "grad_norm": 0.3422331213951111,
      "learning_rate": 0.00014427390791027156,
      "loss": 0.848,
      "step": 242
    },
    {
      "epoch": 0.2852112676056338,
      "grad_norm": 0.33504337072372437,
      "learning_rate": 0.00014403778040141676,
      "loss": 0.7782,
      "step": 243
    },
    {
      "epoch": 0.2863849765258216,
      "grad_norm": 0.3509252667427063,
      "learning_rate": 0.000143801652892562,
      "loss": 0.8535,
      "step": 244
    },
    {
      "epoch": 0.2875586854460094,
      "grad_norm": 0.3254059851169586,
      "learning_rate": 0.0001435655253837072,
      "loss": 0.7642,
      "step": 245
    },
    {
      "epoch": 0.2887323943661972,
      "grad_norm": 0.33594879508018494,
      "learning_rate": 0.00014332939787485243,
      "loss": 0.814,
      "step": 246
    },
    {
      "epoch": 0.289906103286385,
      "grad_norm": 0.3620656132698059,
      "learning_rate": 0.00014309327036599764,
      "loss": 0.8248,
      "step": 247
    },
    {
      "epoch": 0.29107981220657275,
      "grad_norm": 0.3325202167034149,
      "learning_rate": 0.00014285714285714287,
      "loss": 0.7408,
      "step": 248
    },
    {
      "epoch": 0.29225352112676056,
      "grad_norm": 0.33905264735221863,
      "learning_rate": 0.0001426210153482881,
      "loss": 0.8446,
      "step": 249
    },
    {
      "epoch": 0.2934272300469484,
      "grad_norm": 0.3577309548854828,
      "learning_rate": 0.0001423848878394333,
      "loss": 0.784,
      "step": 250
    },
    {
      "epoch": 0.29460093896713613,
      "grad_norm": 0.3840247392654419,
      "learning_rate": 0.0001421487603305785,
      "loss": 0.8068,
      "step": 251
    },
    {
      "epoch": 0.29577464788732394,
      "grad_norm": 0.3539847433567047,
      "learning_rate": 0.00014191263282172374,
      "loss": 0.8232,
      "step": 252
    },
    {
      "epoch": 0.29694835680751175,
      "grad_norm": 0.33225932717323303,
      "learning_rate": 0.00014167650531286894,
      "loss": 0.7946,
      "step": 253
    },
    {
      "epoch": 0.2981220657276995,
      "grad_norm": 0.3429291546344757,
      "learning_rate": 0.00014144037780401418,
      "loss": 0.816,
      "step": 254
    },
    {
      "epoch": 0.2992957746478873,
      "grad_norm": 0.3584197163581848,
      "learning_rate": 0.0001412042502951594,
      "loss": 0.8351,
      "step": 255
    },
    {
      "epoch": 0.3004694835680751,
      "grad_norm": 0.35585007071495056,
      "learning_rate": 0.0001409681227863046,
      "loss": 0.8255,
      "step": 256
    },
    {
      "epoch": 0.30164319248826293,
      "grad_norm": 0.3510012924671173,
      "learning_rate": 0.00014073199527744982,
      "loss": 0.7889,
      "step": 257
    },
    {
      "epoch": 0.3028169014084507,
      "grad_norm": 0.36646419763565063,
      "learning_rate": 0.00014049586776859505,
      "loss": 0.8161,
      "step": 258
    },
    {
      "epoch": 0.3039906103286385,
      "grad_norm": 0.35207659006118774,
      "learning_rate": 0.00014025974025974028,
      "loss": 0.8151,
      "step": 259
    },
    {
      "epoch": 0.3051643192488263,
      "grad_norm": 0.33348143100738525,
      "learning_rate": 0.00014002361275088548,
      "loss": 0.8108,
      "step": 260
    },
    {
      "epoch": 0.30633802816901406,
      "grad_norm": 0.3474767506122589,
      "learning_rate": 0.00013978748524203072,
      "loss": 0.8105,
      "step": 261
    },
    {
      "epoch": 0.3075117370892019,
      "grad_norm": 0.37046462297439575,
      "learning_rate": 0.00013955135773317592,
      "loss": 0.867,
      "step": 262
    },
    {
      "epoch": 0.3086854460093897,
      "grad_norm": 0.3426377475261688,
      "learning_rate": 0.00013931523022432112,
      "loss": 0.8281,
      "step": 263
    },
    {
      "epoch": 0.30985915492957744,
      "grad_norm": 0.3340952694416046,
      "learning_rate": 0.00013907910271546636,
      "loss": 0.7805,
      "step": 264
    },
    {
      "epoch": 0.31103286384976525,
      "grad_norm": 0.3546634316444397,
      "learning_rate": 0.0001388429752066116,
      "loss": 0.824,
      "step": 265
    },
    {
      "epoch": 0.31220657276995306,
      "grad_norm": 0.3211507499217987,
      "learning_rate": 0.0001386068476977568,
      "loss": 0.7572,
      "step": 266
    },
    {
      "epoch": 0.31338028169014087,
      "grad_norm": 0.3440265357494354,
      "learning_rate": 0.000138370720188902,
      "loss": 0.8247,
      "step": 267
    },
    {
      "epoch": 0.3145539906103286,
      "grad_norm": 0.34174132347106934,
      "learning_rate": 0.00013813459268004723,
      "loss": 0.7939,
      "step": 268
    },
    {
      "epoch": 0.31572769953051644,
      "grad_norm": 0.3415057361125946,
      "learning_rate": 0.00013789846517119246,
      "loss": 0.8184,
      "step": 269
    },
    {
      "epoch": 0.31690140845070425,
      "grad_norm": 0.3313206732273102,
      "learning_rate": 0.00013766233766233766,
      "loss": 0.7936,
      "step": 270
    },
    {
      "epoch": 0.318075117370892,
      "grad_norm": 0.35693395137786865,
      "learning_rate": 0.0001374262101534829,
      "loss": 0.7738,
      "step": 271
    },
    {
      "epoch": 0.3192488262910798,
      "grad_norm": 0.3530910313129425,
      "learning_rate": 0.00013719008264462813,
      "loss": 0.7901,
      "step": 272
    },
    {
      "epoch": 0.3204225352112676,
      "grad_norm": 0.34867924451828003,
      "learning_rate": 0.0001369539551357733,
      "loss": 0.8281,
      "step": 273
    },
    {
      "epoch": 0.3215962441314554,
      "grad_norm": 0.34141889214515686,
      "learning_rate": 0.00013671782762691854,
      "loss": 0.7987,
      "step": 274
    },
    {
      "epoch": 0.3227699530516432,
      "grad_norm": 0.3511849045753479,
      "learning_rate": 0.00013648170011806377,
      "loss": 0.8306,
      "step": 275
    },
    {
      "epoch": 0.323943661971831,
      "grad_norm": 0.343523770570755,
      "learning_rate": 0.00013624557260920897,
      "loss": 0.7813,
      "step": 276
    },
    {
      "epoch": 0.32511737089201875,
      "grad_norm": 0.3539726138114929,
      "learning_rate": 0.0001360094451003542,
      "loss": 0.8258,
      "step": 277
    },
    {
      "epoch": 0.32629107981220656,
      "grad_norm": 0.35628989338874817,
      "learning_rate": 0.00013577331759149943,
      "loss": 0.829,
      "step": 278
    },
    {
      "epoch": 0.3274647887323944,
      "grad_norm": 0.3531114459037781,
      "learning_rate": 0.00013553719008264464,
      "loss": 0.8475,
      "step": 279
    },
    {
      "epoch": 0.3286384976525822,
      "grad_norm": 0.35344576835632324,
      "learning_rate": 0.00013530106257378984,
      "loss": 0.8343,
      "step": 280
    },
    {
      "epoch": 0.32981220657276994,
      "grad_norm": 0.37604016065597534,
      "learning_rate": 0.00013506493506493507,
      "loss": 0.7598,
      "step": 281
    },
    {
      "epoch": 0.33098591549295775,
      "grad_norm": 0.35646241903305054,
      "learning_rate": 0.0001348288075560803,
      "loss": 0.83,
      "step": 282
    },
    {
      "epoch": 0.33215962441314556,
      "grad_norm": 0.36084675788879395,
      "learning_rate": 0.0001345926800472255,
      "loss": 0.7465,
      "step": 283
    },
    {
      "epoch": 0.3333333333333333,
      "grad_norm": 0.3514406085014343,
      "learning_rate": 0.00013435655253837071,
      "loss": 0.7979,
      "step": 284
    },
    {
      "epoch": 0.3345070422535211,
      "grad_norm": 0.3554603159427643,
      "learning_rate": 0.00013412042502951595,
      "loss": 0.8487,
      "step": 285
    },
    {
      "epoch": 0.33568075117370894,
      "grad_norm": 0.3360341787338257,
      "learning_rate": 0.00013388429752066115,
      "loss": 0.7787,
      "step": 286
    },
    {
      "epoch": 0.3368544600938967,
      "grad_norm": 0.35026323795318604,
      "learning_rate": 0.00013364817001180638,
      "loss": 0.7845,
      "step": 287
    },
    {
      "epoch": 0.3380281690140845,
      "grad_norm": 0.3419228494167328,
      "learning_rate": 0.00013341204250295161,
      "loss": 0.7971,
      "step": 288
    },
    {
      "epoch": 0.3392018779342723,
      "grad_norm": 0.3314400315284729,
      "learning_rate": 0.00013317591499409682,
      "loss": 0.7899,
      "step": 289
    },
    {
      "epoch": 0.3403755868544601,
      "grad_norm": 0.3434331715106964,
      "learning_rate": 0.00013293978748524202,
      "loss": 0.827,
      "step": 290
    },
    {
      "epoch": 0.3415492957746479,
      "grad_norm": 0.34718382358551025,
      "learning_rate": 0.00013270365997638725,
      "loss": 0.7835,
      "step": 291
    },
    {
      "epoch": 0.3427230046948357,
      "grad_norm": 0.3585168421268463,
      "learning_rate": 0.00013246753246753249,
      "loss": 0.8728,
      "step": 292
    },
    {
      "epoch": 0.3438967136150235,
      "grad_norm": 0.3508673906326294,
      "learning_rate": 0.0001322314049586777,
      "loss": 0.836,
      "step": 293
    },
    {
      "epoch": 0.34507042253521125,
      "grad_norm": 0.40241560339927673,
      "learning_rate": 0.00013199527744982292,
      "loss": 0.8043,
      "step": 294
    },
    {
      "epoch": 0.34624413145539906,
      "grad_norm": 0.33775267004966736,
      "learning_rate": 0.00013175914994096813,
      "loss": 0.8047,
      "step": 295
    },
    {
      "epoch": 0.3474178403755869,
      "grad_norm": 0.3423898220062256,
      "learning_rate": 0.00013152302243211333,
      "loss": 0.7894,
      "step": 296
    },
    {
      "epoch": 0.3485915492957746,
      "grad_norm": 0.3472992479801178,
      "learning_rate": 0.00013128689492325856,
      "loss": 0.8198,
      "step": 297
    },
    {
      "epoch": 0.34976525821596244,
      "grad_norm": 0.3425481915473938,
      "learning_rate": 0.0001310507674144038,
      "loss": 0.8178,
      "step": 298
    },
    {
      "epoch": 0.35093896713615025,
      "grad_norm": 0.3459112048149109,
      "learning_rate": 0.000130814639905549,
      "loss": 0.7749,
      "step": 299
    },
    {
      "epoch": 0.352112676056338,
      "grad_norm": 0.353595495223999,
      "learning_rate": 0.00013057851239669423,
      "loss": 0.7886,
      "step": 300
    },
    {
      "epoch": 0.3532863849765258,
      "grad_norm": 0.35495465993881226,
      "learning_rate": 0.00013034238488783943,
      "loss": 0.771,
      "step": 301
    },
    {
      "epoch": 0.3544600938967136,
      "grad_norm": 0.34812483191490173,
      "learning_rate": 0.00013010625737898467,
      "loss": 0.8335,
      "step": 302
    },
    {
      "epoch": 0.35563380281690143,
      "grad_norm": 0.3655085861682892,
      "learning_rate": 0.00012987012987012987,
      "loss": 0.8117,
      "step": 303
    },
    {
      "epoch": 0.3568075117370892,
      "grad_norm": 0.35925915837287903,
      "learning_rate": 0.0001296340023612751,
      "loss": 0.8147,
      "step": 304
    },
    {
      "epoch": 0.357981220657277,
      "grad_norm": 0.3293222486972809,
      "learning_rate": 0.00012939787485242033,
      "loss": 0.7602,
      "step": 305
    },
    {
      "epoch": 0.3591549295774648,
      "grad_norm": 0.3486446738243103,
      "learning_rate": 0.00012916174734356554,
      "loss": 0.7857,
      "step": 306
    },
    {
      "epoch": 0.36032863849765256,
      "grad_norm": 0.382565975189209,
      "learning_rate": 0.00012892561983471074,
      "loss": 0.863,
      "step": 307
    },
    {
      "epoch": 0.3615023474178404,
      "grad_norm": 0.32544344663619995,
      "learning_rate": 0.00012868949232585597,
      "loss": 0.781,
      "step": 308
    },
    {
      "epoch": 0.3626760563380282,
      "grad_norm": 0.38700491189956665,
      "learning_rate": 0.00012845336481700118,
      "loss": 0.8102,
      "step": 309
    },
    {
      "epoch": 0.36384976525821594,
      "grad_norm": 0.3503759503364563,
      "learning_rate": 0.0001282172373081464,
      "loss": 0.7699,
      "step": 310
    },
    {
      "epoch": 0.36502347417840375,
      "grad_norm": 0.3323630094528198,
      "learning_rate": 0.00012798110979929164,
      "loss": 0.7511,
      "step": 311
    },
    {
      "epoch": 0.36619718309859156,
      "grad_norm": 0.3668995797634125,
      "learning_rate": 0.00012774498229043685,
      "loss": 0.7374,
      "step": 312
    },
    {
      "epoch": 0.3673708920187793,
      "grad_norm": 0.37373387813568115,
      "learning_rate": 0.00012750885478158205,
      "loss": 0.8077,
      "step": 313
    },
    {
      "epoch": 0.3685446009389671,
      "grad_norm": 0.3601135015487671,
      "learning_rate": 0.00012727272727272728,
      "loss": 0.7991,
      "step": 314
    },
    {
      "epoch": 0.36971830985915494,
      "grad_norm": 0.3527435064315796,
      "learning_rate": 0.00012703659976387249,
      "loss": 0.7971,
      "step": 315
    },
    {
      "epoch": 0.37089201877934275,
      "grad_norm": 0.3584372401237488,
      "learning_rate": 0.00012680047225501772,
      "loss": 0.7513,
      "step": 316
    },
    {
      "epoch": 0.3720657276995305,
      "grad_norm": 0.3517726957798004,
      "learning_rate": 0.00012656434474616295,
      "loss": 0.8206,
      "step": 317
    },
    {
      "epoch": 0.3732394366197183,
      "grad_norm": 0.3655302822589874,
      "learning_rate": 0.00012632821723730815,
      "loss": 0.771,
      "step": 318
    },
    {
      "epoch": 0.3744131455399061,
      "grad_norm": 0.3659893274307251,
      "learning_rate": 0.00012609208972845336,
      "loss": 0.8048,
      "step": 319
    },
    {
      "epoch": 0.3755868544600939,
      "grad_norm": 0.36364591121673584,
      "learning_rate": 0.0001258559622195986,
      "loss": 0.7832,
      "step": 320
    },
    {
      "epoch": 0.3767605633802817,
      "grad_norm": 0.37528395652770996,
      "learning_rate": 0.00012561983471074382,
      "loss": 0.7926,
      "step": 321
    },
    {
      "epoch": 0.3779342723004695,
      "grad_norm": 0.37137654423713684,
      "learning_rate": 0.00012538370720188903,
      "loss": 0.8486,
      "step": 322
    },
    {
      "epoch": 0.37910798122065725,
      "grad_norm": 0.3466728925704956,
      "learning_rate": 0.00012514757969303423,
      "loss": 0.7961,
      "step": 323
    },
    {
      "epoch": 0.38028169014084506,
      "grad_norm": 0.38629114627838135,
      "learning_rate": 0.00012491145218417946,
      "loss": 0.8071,
      "step": 324
    },
    {
      "epoch": 0.3814553990610329,
      "grad_norm": 0.34686383605003357,
      "learning_rate": 0.00012467532467532467,
      "loss": 0.7698,
      "step": 325
    },
    {
      "epoch": 0.3826291079812207,
      "grad_norm": 0.36625292897224426,
      "learning_rate": 0.0001244391971664699,
      "loss": 0.8486,
      "step": 326
    },
    {
      "epoch": 0.38380281690140844,
      "grad_norm": 0.38903650641441345,
      "learning_rate": 0.00012420306965761513,
      "loss": 0.8031,
      "step": 327
    },
    {
      "epoch": 0.38497652582159625,
      "grad_norm": 0.3456287980079651,
      "learning_rate": 0.00012396694214876033,
      "loss": 0.7887,
      "step": 328
    },
    {
      "epoch": 0.38615023474178406,
      "grad_norm": 0.36374613642692566,
      "learning_rate": 0.00012373081463990554,
      "loss": 0.7588,
      "step": 329
    },
    {
      "epoch": 0.3873239436619718,
      "grad_norm": 0.360626220703125,
      "learning_rate": 0.00012349468713105077,
      "loss": 0.8239,
      "step": 330
    },
    {
      "epoch": 0.3884976525821596,
      "grad_norm": 0.40213796496391296,
      "learning_rate": 0.000123258559622196,
      "loss": 0.8029,
      "step": 331
    },
    {
      "epoch": 0.38967136150234744,
      "grad_norm": 0.3273613750934601,
      "learning_rate": 0.0001230224321133412,
      "loss": 0.7567,
      "step": 332
    },
    {
      "epoch": 0.3908450704225352,
      "grad_norm": 0.34953057765960693,
      "learning_rate": 0.00012278630460448644,
      "loss": 0.7512,
      "step": 333
    },
    {
      "epoch": 0.392018779342723,
      "grad_norm": 0.34772762656211853,
      "learning_rate": 0.00012255017709563167,
      "loss": 0.7551,
      "step": 334
    },
    {
      "epoch": 0.3931924882629108,
      "grad_norm": 0.34170207381248474,
      "learning_rate": 0.00012231404958677685,
      "loss": 0.7884,
      "step": 335
    },
    {
      "epoch": 0.39436619718309857,
      "grad_norm": 0.3696103096008301,
      "learning_rate": 0.00012207792207792208,
      "loss": 0.8658,
      "step": 336
    },
    {
      "epoch": 0.3955399061032864,
      "grad_norm": 0.3513827621936798,
      "learning_rate": 0.00012184179456906731,
      "loss": 0.8199,
      "step": 337
    },
    {
      "epoch": 0.3967136150234742,
      "grad_norm": 0.3454856872558594,
      "learning_rate": 0.00012160566706021253,
      "loss": 0.7627,
      "step": 338
    },
    {
      "epoch": 0.397887323943662,
      "grad_norm": 0.3246639370918274,
      "learning_rate": 0.00012136953955135774,
      "loss": 0.7454,
      "step": 339
    },
    {
      "epoch": 0.39906103286384975,
      "grad_norm": 0.33567938208580017,
      "learning_rate": 0.00012113341204250295,
      "loss": 0.7611,
      "step": 340
    },
    {
      "epoch": 0.40023474178403756,
      "grad_norm": 0.33728334307670593,
      "learning_rate": 0.00012089728453364817,
      "loss": 0.7575,
      "step": 341
    },
    {
      "epoch": 0.4014084507042254,
      "grad_norm": 0.35161352157592773,
      "learning_rate": 0.0001206611570247934,
      "loss": 0.8117,
      "step": 342
    },
    {
      "epoch": 0.4025821596244131,
      "grad_norm": 0.3425585925579071,
      "learning_rate": 0.00012042502951593862,
      "loss": 0.8019,
      "step": 343
    },
    {
      "epoch": 0.40375586854460094,
      "grad_norm": 0.3406507968902588,
      "learning_rate": 0.00012018890200708383,
      "loss": 0.8235,
      "step": 344
    },
    {
      "epoch": 0.40492957746478875,
      "grad_norm": 0.37840309739112854,
      "learning_rate": 0.00011995277449822907,
      "loss": 0.7866,
      "step": 345
    },
    {
      "epoch": 0.4061032863849765,
      "grad_norm": 0.35816213488578796,
      "learning_rate": 0.00011971664698937426,
      "loss": 0.8425,
      "step": 346
    },
    {
      "epoch": 0.4072769953051643,
      "grad_norm": 0.3441546559333801,
      "learning_rate": 0.00011948051948051949,
      "loss": 0.8094,
      "step": 347
    },
    {
      "epoch": 0.4084507042253521,
      "grad_norm": 0.34275054931640625,
      "learning_rate": 0.0001192443919716647,
      "loss": 0.7244,
      "step": 348
    },
    {
      "epoch": 0.4096244131455399,
      "grad_norm": 0.33207401633262634,
      "learning_rate": 0.00011900826446280992,
      "loss": 0.8108,
      "step": 349
    },
    {
      "epoch": 0.4107981220657277,
      "grad_norm": 0.3412252962589264,
      "learning_rate": 0.00011877213695395516,
      "loss": 0.7818,
      "step": 350
    },
    {
      "epoch": 0.4119718309859155,
      "grad_norm": 0.36701643466949463,
      "learning_rate": 0.00011853600944510035,
      "loss": 0.8293,
      "step": 351
    },
    {
      "epoch": 0.4131455399061033,
      "grad_norm": 0.34462520480155945,
      "learning_rate": 0.00011829988193624558,
      "loss": 0.7603,
      "step": 352
    },
    {
      "epoch": 0.41431924882629106,
      "grad_norm": 0.35232508182525635,
      "learning_rate": 0.0001180637544273908,
      "loss": 0.7616,
      "step": 353
    },
    {
      "epoch": 0.4154929577464789,
      "grad_norm": 0.37428373098373413,
      "learning_rate": 0.00011782762691853601,
      "loss": 0.7919,
      "step": 354
    },
    {
      "epoch": 0.4166666666666667,
      "grad_norm": 0.3429507911205292,
      "learning_rate": 0.00011759149940968123,
      "loss": 0.7859,
      "step": 355
    },
    {
      "epoch": 0.41784037558685444,
      "grad_norm": 0.3584844470024109,
      "learning_rate": 0.00011735537190082646,
      "loss": 0.7934,
      "step": 356
    },
    {
      "epoch": 0.41901408450704225,
      "grad_norm": 0.356391578912735,
      "learning_rate": 0.00011711924439197165,
      "loss": 0.8222,
      "step": 357
    },
    {
      "epoch": 0.42018779342723006,
      "grad_norm": 0.3663417100906372,
      "learning_rate": 0.00011688311688311689,
      "loss": 0.7507,
      "step": 358
    },
    {
      "epoch": 0.4213615023474178,
      "grad_norm": 0.3388553559780121,
      "learning_rate": 0.0001166469893742621,
      "loss": 0.8263,
      "step": 359
    },
    {
      "epoch": 0.4225352112676056,
      "grad_norm": 0.34876593947410583,
      "learning_rate": 0.00011641086186540732,
      "loss": 0.7969,
      "step": 360
    },
    {
      "epoch": 0.42370892018779344,
      "grad_norm": 0.3500271737575531,
      "learning_rate": 0.00011617473435655255,
      "loss": 0.7789,
      "step": 361
    },
    {
      "epoch": 0.42488262910798125,
      "grad_norm": 0.3554798662662506,
      "learning_rate": 0.00011593860684769777,
      "loss": 0.7681,
      "step": 362
    },
    {
      "epoch": 0.426056338028169,
      "grad_norm": 0.34559762477874756,
      "learning_rate": 0.00011570247933884298,
      "loss": 0.7676,
      "step": 363
    },
    {
      "epoch": 0.4272300469483568,
      "grad_norm": 0.3520505726337433,
      "learning_rate": 0.0001154663518299882,
      "loss": 0.7494,
      "step": 364
    },
    {
      "epoch": 0.4284037558685446,
      "grad_norm": 0.35454803705215454,
      "learning_rate": 0.00011523022432113341,
      "loss": 0.7516,
      "step": 365
    },
    {
      "epoch": 0.4295774647887324,
      "grad_norm": 0.36526602506637573,
      "learning_rate": 0.00011499409681227864,
      "loss": 0.7789,
      "step": 366
    },
    {
      "epoch": 0.4307511737089202,
      "grad_norm": 0.34084445238113403,
      "learning_rate": 0.00011475796930342386,
      "loss": 0.7446,
      "step": 367
    },
    {
      "epoch": 0.431924882629108,
      "grad_norm": 0.3405500054359436,
      "learning_rate": 0.00011452184179456907,
      "loss": 0.8217,
      "step": 368
    },
    {
      "epoch": 0.43309859154929575,
      "grad_norm": 0.3523256182670593,
      "learning_rate": 0.00011428571428571428,
      "loss": 0.7311,
      "step": 369
    },
    {
      "epoch": 0.43427230046948356,
      "grad_norm": 0.3336530327796936,
      "learning_rate": 0.0001140495867768595,
      "loss": 0.7806,
      "step": 370
    },
    {
      "epoch": 0.4354460093896714,
      "grad_norm": 0.3268769383430481,
      "learning_rate": 0.00011381345926800473,
      "loss": 0.7945,
      "step": 371
    },
    {
      "epoch": 0.43661971830985913,
      "grad_norm": 0.35258617997169495,
      "learning_rate": 0.00011357733175914995,
      "loss": 0.7468,
      "step": 372
    },
    {
      "epoch": 0.43779342723004694,
      "grad_norm": 0.3546913266181946,
      "learning_rate": 0.00011334120425029517,
      "loss": 0.7921,
      "step": 373
    },
    {
      "epoch": 0.43896713615023475,
      "grad_norm": 0.36266180872917175,
      "learning_rate": 0.00011310507674144037,
      "loss": 0.7623,
      "step": 374
    },
    {
      "epoch": 0.44014084507042256,
      "grad_norm": 0.3355543613433838,
      "learning_rate": 0.00011286894923258559,
      "loss": 0.7436,
      "step": 375
    },
    {
      "epoch": 0.4413145539906103,
      "grad_norm": 0.33666127920150757,
      "learning_rate": 0.00011263282172373082,
      "loss": 0.7609,
      "step": 376
    },
    {
      "epoch": 0.4424882629107981,
      "grad_norm": 0.3505670428276062,
      "learning_rate": 0.00011239669421487604,
      "loss": 0.7868,
      "step": 377
    },
    {
      "epoch": 0.44366197183098594,
      "grad_norm": 0.3446255028247833,
      "learning_rate": 0.00011216056670602126,
      "loss": 0.765,
      "step": 378
    },
    {
      "epoch": 0.4448356807511737,
      "grad_norm": 0.3761040270328522,
      "learning_rate": 0.00011192443919716649,
      "loss": 0.8104,
      "step": 379
    },
    {
      "epoch": 0.4460093896713615,
      "grad_norm": 0.35692986845970154,
      "learning_rate": 0.00011168831168831168,
      "loss": 0.7896,
      "step": 380
    },
    {
      "epoch": 0.4471830985915493,
      "grad_norm": 0.34384050965309143,
      "learning_rate": 0.00011145218417945691,
      "loss": 0.7716,
      "step": 381
    },
    {
      "epoch": 0.44835680751173707,
      "grad_norm": 0.3477395176887512,
      "learning_rate": 0.00011121605667060213,
      "loss": 0.8146,
      "step": 382
    },
    {
      "epoch": 0.4495305164319249,
      "grad_norm": 0.35172998905181885,
      "learning_rate": 0.00011097992916174735,
      "loss": 0.7844,
      "step": 383
    },
    {
      "epoch": 0.4507042253521127,
      "grad_norm": 0.33881857991218567,
      "learning_rate": 0.00011074380165289258,
      "loss": 0.7528,
      "step": 384
    },
    {
      "epoch": 0.4518779342723005,
      "grad_norm": 0.3429534137248993,
      "learning_rate": 0.00011050767414403777,
      "loss": 0.7826,
      "step": 385
    },
    {
      "epoch": 0.45305164319248825,
      "grad_norm": 0.34472665190696716,
      "learning_rate": 0.000110271546635183,
      "loss": 0.7153,
      "step": 386
    },
    {
      "epoch": 0.45422535211267606,
      "grad_norm": 0.3572479486465454,
      "learning_rate": 0.00011003541912632822,
      "loss": 0.7811,
      "step": 387
    },
    {
      "epoch": 0.45539906103286387,
      "grad_norm": 0.3531682789325714,
      "learning_rate": 0.00010979929161747344,
      "loss": 0.8016,
      "step": 388
    },
    {
      "epoch": 0.4565727699530516,
      "grad_norm": 0.3845299780368805,
      "learning_rate": 0.00010956316410861867,
      "loss": 0.7817,
      "step": 389
    },
    {
      "epoch": 0.45774647887323944,
      "grad_norm": 0.35217660665512085,
      "learning_rate": 0.00010932703659976389,
      "loss": 0.7495,
      "step": 390
    },
    {
      "epoch": 0.45892018779342725,
      "grad_norm": 0.35103702545166016,
      "learning_rate": 0.00010909090909090909,
      "loss": 0.7602,
      "step": 391
    },
    {
      "epoch": 0.460093896713615,
      "grad_norm": 0.3511259853839874,
      "learning_rate": 0.00010885478158205431,
      "loss": 0.7923,
      "step": 392
    },
    {
      "epoch": 0.4612676056338028,
      "grad_norm": 0.33732983469963074,
      "learning_rate": 0.00010861865407319953,
      "loss": 0.7875,
      "step": 393
    },
    {
      "epoch": 0.4624413145539906,
      "grad_norm": 0.35035955905914307,
      "learning_rate": 0.00010838252656434476,
      "loss": 0.7737,
      "step": 394
    },
    {
      "epoch": 0.4636150234741784,
      "grad_norm": 0.3277076482772827,
      "learning_rate": 0.00010814639905548998,
      "loss": 0.7619,
      "step": 395
    },
    {
      "epoch": 0.4647887323943662,
      "grad_norm": 0.34461456537246704,
      "learning_rate": 0.00010791027154663518,
      "loss": 0.7394,
      "step": 396
    },
    {
      "epoch": 0.465962441314554,
      "grad_norm": 0.36000820994377136,
      "learning_rate": 0.0001076741440377804,
      "loss": 0.8004,
      "step": 397
    },
    {
      "epoch": 0.4671361502347418,
      "grad_norm": 0.3291054666042328,
      "learning_rate": 0.00010743801652892562,
      "loss": 0.721,
      "step": 398
    },
    {
      "epoch": 0.46830985915492956,
      "grad_norm": 0.37541574239730835,
      "learning_rate": 0.00010720188902007085,
      "loss": 0.7673,
      "step": 399
    },
    {
      "epoch": 0.4694835680751174,
      "grad_norm": 0.33268067240715027,
      "learning_rate": 0.00010696576151121607,
      "loss": 0.7439,
      "step": 400
    },
    {
      "epoch": 0.4706572769953052,
      "grad_norm": 0.34383484721183777,
      "learning_rate": 0.00010672963400236129,
      "loss": 0.7453,
      "step": 401
    },
    {
      "epoch": 0.47183098591549294,
      "grad_norm": 0.3543702960014343,
      "learning_rate": 0.00010649350649350649,
      "loss": 0.7544,
      "step": 402
    },
    {
      "epoch": 0.47300469483568075,
      "grad_norm": 0.34553685784339905,
      "learning_rate": 0.00010625737898465171,
      "loss": 0.7656,
      "step": 403
    },
    {
      "epoch": 0.47417840375586856,
      "grad_norm": 0.3437071144580841,
      "learning_rate": 0.00010602125147579694,
      "loss": 0.773,
      "step": 404
    },
    {
      "epoch": 0.4753521126760563,
      "grad_norm": 0.34917253255844116,
      "learning_rate": 0.00010578512396694216,
      "loss": 0.7607,
      "step": 405
    },
    {
      "epoch": 0.4765258215962441,
      "grad_norm": 0.33429262042045593,
      "learning_rate": 0.00010554899645808738,
      "loss": 0.768,
      "step": 406
    },
    {
      "epoch": 0.47769953051643194,
      "grad_norm": 0.33842045068740845,
      "learning_rate": 0.00010531286894923261,
      "loss": 0.7665,
      "step": 407
    },
    {
      "epoch": 0.4788732394366197,
      "grad_norm": 0.3419265151023865,
      "learning_rate": 0.0001050767414403778,
      "loss": 0.7717,
      "step": 408
    },
    {
      "epoch": 0.4800469483568075,
      "grad_norm": 0.3458483815193176,
      "learning_rate": 0.00010484061393152303,
      "loss": 0.8031,
      "step": 409
    },
    {
      "epoch": 0.4812206572769953,
      "grad_norm": 0.37077274918556213,
      "learning_rate": 0.00010460448642266825,
      "loss": 0.8009,
      "step": 410
    },
    {
      "epoch": 0.4823943661971831,
      "grad_norm": 0.35040315985679626,
      "learning_rate": 0.00010436835891381347,
      "loss": 0.7545,
      "step": 411
    },
    {
      "epoch": 0.4835680751173709,
      "grad_norm": 0.3503456115722656,
      "learning_rate": 0.0001041322314049587,
      "loss": 0.8515,
      "step": 412
    },
    {
      "epoch": 0.4847417840375587,
      "grad_norm": 0.34627342224121094,
      "learning_rate": 0.00010389610389610389,
      "loss": 0.716,
      "step": 413
    },
    {
      "epoch": 0.4859154929577465,
      "grad_norm": 0.3596992790699005,
      "learning_rate": 0.00010365997638724912,
      "loss": 0.7636,
      "step": 414
    },
    {
      "epoch": 0.48708920187793425,
      "grad_norm": 0.3346829116344452,
      "learning_rate": 0.00010342384887839434,
      "loss": 0.7635,
      "step": 415
    },
    {
      "epoch": 0.48826291079812206,
      "grad_norm": 0.37179237604141235,
      "learning_rate": 0.00010318772136953956,
      "loss": 0.7642,
      "step": 416
    },
    {
      "epoch": 0.4894366197183099,
      "grad_norm": 0.34897381067276,
      "learning_rate": 0.00010295159386068479,
      "loss": 0.7792,
      "step": 417
    },
    {
      "epoch": 0.49061032863849763,
      "grad_norm": 0.3820830285549164,
      "learning_rate": 0.00010271546635183,
      "loss": 0.7722,
      "step": 418
    },
    {
      "epoch": 0.49178403755868544,
      "grad_norm": 0.3688552677631378,
      "learning_rate": 0.00010247933884297521,
      "loss": 0.7927,
      "step": 419
    },
    {
      "epoch": 0.49295774647887325,
      "grad_norm": 0.35100415349006653,
      "learning_rate": 0.00010224321133412043,
      "loss": 0.7848,
      "step": 420
    },
    {
      "epoch": 0.49413145539906106,
      "grad_norm": 0.3596225082874298,
      "learning_rate": 0.00010200708382526565,
      "loss": 0.7383,
      "step": 421
    },
    {
      "epoch": 0.4953051643192488,
      "grad_norm": 0.36203423142433167,
      "learning_rate": 0.00010177095631641088,
      "loss": 0.769,
      "step": 422
    },
    {
      "epoch": 0.4964788732394366,
      "grad_norm": 0.3776590824127197,
      "learning_rate": 0.0001015348288075561,
      "loss": 0.8007,
      "step": 423
    },
    {
      "epoch": 0.49765258215962443,
      "grad_norm": 0.36009421944618225,
      "learning_rate": 0.0001012987012987013,
      "loss": 0.7557,
      "step": 424
    },
    {
      "epoch": 0.4988262910798122,
      "grad_norm": 0.3442706763744354,
      "learning_rate": 0.00010106257378984652,
      "loss": 0.7488,
      "step": 425
    },
    {
      "epoch": 0.5,
      "grad_norm": 0.3635407090187073,
      "learning_rate": 0.00010082644628099174,
      "loss": 0.7922,
      "step": 426
    },
    {
      "epoch": 0.5011737089201878,
      "grad_norm": 0.3766370117664337,
      "learning_rate": 0.00010059031877213697,
      "loss": 0.7818,
      "step": 427
    },
    {
      "epoch": 0.5023474178403756,
      "grad_norm": 0.34344202280044556,
      "learning_rate": 0.00010035419126328218,
      "loss": 0.8308,
      "step": 428
    },
    {
      "epoch": 0.5035211267605634,
      "grad_norm": 0.3495674133300781,
      "learning_rate": 0.0001001180637544274,
      "loss": 0.799,
      "step": 429
    },
    {
      "epoch": 0.5046948356807511,
      "grad_norm": 0.36545464396476746,
      "learning_rate": 9.988193624557262e-05,
      "loss": 0.7453,
      "step": 430
    },
    {
      "epoch": 0.505868544600939,
      "grad_norm": 0.3482630252838135,
      "learning_rate": 9.964580873671782e-05,
      "loss": 0.7422,
      "step": 431
    },
    {
      "epoch": 0.5070422535211268,
      "grad_norm": 0.3745418190956116,
      "learning_rate": 9.940968122786304e-05,
      "loss": 0.7333,
      "step": 432
    },
    {
      "epoch": 0.5082159624413145,
      "grad_norm": 0.3470025062561035,
      "learning_rate": 9.917355371900827e-05,
      "loss": 0.7907,
      "step": 433
    },
    {
      "epoch": 0.5093896713615024,
      "grad_norm": 0.38251325488090515,
      "learning_rate": 9.893742621015348e-05,
      "loss": 0.7629,
      "step": 434
    },
    {
      "epoch": 0.5105633802816901,
      "grad_norm": 0.3829626739025116,
      "learning_rate": 9.870129870129871e-05,
      "loss": 0.7939,
      "step": 435
    },
    {
      "epoch": 0.5117370892018779,
      "grad_norm": 0.35726287961006165,
      "learning_rate": 9.846517119244393e-05,
      "loss": 0.755,
      "step": 436
    },
    {
      "epoch": 0.5129107981220657,
      "grad_norm": 0.38168108463287354,
      "learning_rate": 9.822904368358913e-05,
      "loss": 0.7396,
      "step": 437
    },
    {
      "epoch": 0.5140845070422535,
      "grad_norm": 0.35728660225868225,
      "learning_rate": 9.799291617473436e-05,
      "loss": 0.7568,
      "step": 438
    },
    {
      "epoch": 0.5152582159624414,
      "grad_norm": 0.37819668650627136,
      "learning_rate": 9.775678866587958e-05,
      "loss": 0.8046,
      "step": 439
    },
    {
      "epoch": 0.5164319248826291,
      "grad_norm": 0.4106784760951996,
      "learning_rate": 9.75206611570248e-05,
      "loss": 0.7116,
      "step": 440
    },
    {
      "epoch": 0.5176056338028169,
      "grad_norm": 0.3476578891277313,
      "learning_rate": 9.728453364817002e-05,
      "loss": 0.7824,
      "step": 441
    },
    {
      "epoch": 0.5187793427230047,
      "grad_norm": 0.36705800890922546,
      "learning_rate": 9.704840613931524e-05,
      "loss": 0.7631,
      "step": 442
    },
    {
      "epoch": 0.5199530516431925,
      "grad_norm": 0.3880864977836609,
      "learning_rate": 9.681227863046045e-05,
      "loss": 0.7608,
      "step": 443
    },
    {
      "epoch": 0.5211267605633803,
      "grad_norm": 0.3610959053039551,
      "learning_rate": 9.657615112160567e-05,
      "loss": 0.7909,
      "step": 444
    },
    {
      "epoch": 0.5223004694835681,
      "grad_norm": 0.33494657278060913,
      "learning_rate": 9.634002361275089e-05,
      "loss": 0.7108,
      "step": 445
    },
    {
      "epoch": 0.5234741784037559,
      "grad_norm": 0.352055162191391,
      "learning_rate": 9.610389610389611e-05,
      "loss": 0.7177,
      "step": 446
    },
    {
      "epoch": 0.5246478873239436,
      "grad_norm": 0.35466742515563965,
      "learning_rate": 9.586776859504133e-05,
      "loss": 0.7762,
      "step": 447
    },
    {
      "epoch": 0.5258215962441315,
      "grad_norm": 0.34477657079696655,
      "learning_rate": 9.563164108618654e-05,
      "loss": 0.7583,
      "step": 448
    },
    {
      "epoch": 0.5269953051643192,
      "grad_norm": 0.37008315324783325,
      "learning_rate": 9.539551357733176e-05,
      "loss": 0.7954,
      "step": 449
    },
    {
      "epoch": 0.528169014084507,
      "grad_norm": 0.34141793847084045,
      "learning_rate": 9.515938606847698e-05,
      "loss": 0.7444,
      "step": 450
    },
    {
      "epoch": 0.5293427230046949,
      "grad_norm": 0.3429400622844696,
      "learning_rate": 9.49232585596222e-05,
      "loss": 0.7499,
      "step": 451
    },
    {
      "epoch": 0.5305164319248826,
      "grad_norm": 0.3666730225086212,
      "learning_rate": 9.468713105076742e-05,
      "loss": 0.7704,
      "step": 452
    },
    {
      "epoch": 0.5316901408450704,
      "grad_norm": 0.34185874462127686,
      "learning_rate": 9.445100354191265e-05,
      "loss": 0.7446,
      "step": 453
    },
    {
      "epoch": 0.5328638497652582,
      "grad_norm": 0.3718375861644745,
      "learning_rate": 9.421487603305785e-05,
      "loss": 0.7316,
      "step": 454
    },
    {
      "epoch": 0.534037558685446,
      "grad_norm": 0.35064697265625,
      "learning_rate": 9.397874852420307e-05,
      "loss": 0.7651,
      "step": 455
    },
    {
      "epoch": 0.5352112676056338,
      "grad_norm": 0.3724139630794525,
      "learning_rate": 9.37426210153483e-05,
      "loss": 0.7639,
      "step": 456
    },
    {
      "epoch": 0.5363849765258216,
      "grad_norm": 0.3420800566673279,
      "learning_rate": 9.35064935064935e-05,
      "loss": 0.7578,
      "step": 457
    },
    {
      "epoch": 0.5375586854460094,
      "grad_norm": 0.3437943160533905,
      "learning_rate": 9.327036599763874e-05,
      "loss": 0.7898,
      "step": 458
    },
    {
      "epoch": 0.5387323943661971,
      "grad_norm": 0.3799413740634918,
      "learning_rate": 9.303423848878394e-05,
      "loss": 0.7216,
      "step": 459
    },
    {
      "epoch": 0.539906103286385,
      "grad_norm": 0.35702013969421387,
      "learning_rate": 9.279811097992916e-05,
      "loss": 0.7509,
      "step": 460
    },
    {
      "epoch": 0.5410798122065728,
      "grad_norm": 0.36074140667915344,
      "learning_rate": 9.256198347107439e-05,
      "loss": 0.7448,
      "step": 461
    },
    {
      "epoch": 0.5422535211267606,
      "grad_norm": 0.34211182594299316,
      "learning_rate": 9.23258559622196e-05,
      "loss": 0.7143,
      "step": 462
    },
    {
      "epoch": 0.5434272300469484,
      "grad_norm": 0.3816893398761749,
      "learning_rate": 9.208972845336483e-05,
      "loss": 0.7178,
      "step": 463
    },
    {
      "epoch": 0.5446009389671361,
      "grad_norm": 0.36033767461776733,
      "learning_rate": 9.185360094451005e-05,
      "loss": 0.7406,
      "step": 464
    },
    {
      "epoch": 0.545774647887324,
      "grad_norm": 0.38050010800361633,
      "learning_rate": 9.161747343565525e-05,
      "loss": 0.7528,
      "step": 465
    },
    {
      "epoch": 0.5469483568075117,
      "grad_norm": 0.3648395240306854,
      "learning_rate": 9.138134592680048e-05,
      "loss": 0.7802,
      "step": 466
    },
    {
      "epoch": 0.5481220657276995,
      "grad_norm": 0.35185542702674866,
      "learning_rate": 9.11452184179457e-05,
      "loss": 0.7489,
      "step": 467
    },
    {
      "epoch": 0.5492957746478874,
      "grad_norm": 0.3487717807292938,
      "learning_rate": 9.090909090909092e-05,
      "loss": 0.7742,
      "step": 468
    },
    {
      "epoch": 0.5504694835680751,
      "grad_norm": 0.36121654510498047,
      "learning_rate": 9.067296340023614e-05,
      "loss": 0.7974,
      "step": 469
    },
    {
      "epoch": 0.5516431924882629,
      "grad_norm": 0.3470339775085449,
      "learning_rate": 9.043683589138135e-05,
      "loss": 0.723,
      "step": 470
    },
    {
      "epoch": 0.5528169014084507,
      "grad_norm": 0.33549764752388,
      "learning_rate": 9.020070838252657e-05,
      "loss": 0.7334,
      "step": 471
    },
    {
      "epoch": 0.5539906103286385,
      "grad_norm": 0.36101868748664856,
      "learning_rate": 8.996458087367179e-05,
      "loss": 0.6817,
      "step": 472
    },
    {
      "epoch": 0.5551643192488263,
      "grad_norm": 0.36847153306007385,
      "learning_rate": 8.9728453364817e-05,
      "loss": 0.7942,
      "step": 473
    },
    {
      "epoch": 0.5563380281690141,
      "grad_norm": 0.3564891815185547,
      "learning_rate": 8.949232585596222e-05,
      "loss": 0.7071,
      "step": 474
    },
    {
      "epoch": 0.5575117370892019,
      "grad_norm": 0.36866652965545654,
      "learning_rate": 8.925619834710744e-05,
      "loss": 0.7685,
      "step": 475
    },
    {
      "epoch": 0.5586854460093896,
      "grad_norm": 0.370924711227417,
      "learning_rate": 8.902007083825266e-05,
      "loss": 0.7313,
      "step": 476
    },
    {
      "epoch": 0.5598591549295775,
      "grad_norm": 0.3611142039299011,
      "learning_rate": 8.878394332939788e-05,
      "loss": 0.7666,
      "step": 477
    },
    {
      "epoch": 0.5610328638497653,
      "grad_norm": 0.3418121635913849,
      "learning_rate": 8.85478158205431e-05,
      "loss": 0.7194,
      "step": 478
    },
    {
      "epoch": 0.562206572769953,
      "grad_norm": 0.3478650748729706,
      "learning_rate": 8.831168831168831e-05,
      "loss": 0.7145,
      "step": 479
    },
    {
      "epoch": 0.5633802816901409,
      "grad_norm": 0.3567008078098297,
      "learning_rate": 8.807556080283353e-05,
      "loss": 0.7591,
      "step": 480
    },
    {
      "epoch": 0.5645539906103286,
      "grad_norm": 0.3629607558250427,
      "learning_rate": 8.783943329397875e-05,
      "loss": 0.7856,
      "step": 481
    },
    {
      "epoch": 0.5657276995305164,
      "grad_norm": 0.37257978320121765,
      "learning_rate": 8.760330578512397e-05,
      "loss": 0.709,
      "step": 482
    },
    {
      "epoch": 0.5669014084507042,
      "grad_norm": 0.3570626676082611,
      "learning_rate": 8.736717827626919e-05,
      "loss": 0.7639,
      "step": 483
    },
    {
      "epoch": 0.568075117370892,
      "grad_norm": 0.34790506958961487,
      "learning_rate": 8.713105076741442e-05,
      "loss": 0.7375,
      "step": 484
    },
    {
      "epoch": 0.5692488262910798,
      "grad_norm": 0.3525756895542145,
      "learning_rate": 8.689492325855962e-05,
      "loss": 0.7274,
      "step": 485
    },
    {
      "epoch": 0.5704225352112676,
      "grad_norm": 0.3545394837856293,
      "learning_rate": 8.665879574970484e-05,
      "loss": 0.7531,
      "step": 486
    },
    {
      "epoch": 0.5715962441314554,
      "grad_norm": 0.35677066445350647,
      "learning_rate": 8.642266824085006e-05,
      "loss": 0.7682,
      "step": 487
    },
    {
      "epoch": 0.5727699530516432,
      "grad_norm": 0.3439461290836334,
      "learning_rate": 8.618654073199528e-05,
      "loss": 0.7176,
      "step": 488
    },
    {
      "epoch": 0.573943661971831,
      "grad_norm": 0.3622515797615051,
      "learning_rate": 8.595041322314051e-05,
      "loss": 0.7004,
      "step": 489
    },
    {
      "epoch": 0.5751173708920188,
      "grad_norm": 0.36056646704673767,
      "learning_rate": 8.571428571428571e-05,
      "loss": 0.74,
      "step": 490
    },
    {
      "epoch": 0.5762910798122066,
      "grad_norm": 0.3509630262851715,
      "learning_rate": 8.547815820543093e-05,
      "loss": 0.8006,
      "step": 491
    },
    {
      "epoch": 0.5774647887323944,
      "grad_norm": 0.3422422707080841,
      "learning_rate": 8.524203069657616e-05,
      "loss": 0.7162,
      "step": 492
    },
    {
      "epoch": 0.5786384976525821,
      "grad_norm": 0.35553744435310364,
      "learning_rate": 8.500590318772137e-05,
      "loss": 0.7554,
      "step": 493
    },
    {
      "epoch": 0.57981220657277,
      "grad_norm": 0.3443603813648224,
      "learning_rate": 8.47697756788666e-05,
      "loss": 0.7128,
      "step": 494
    },
    {
      "epoch": 0.5809859154929577,
      "grad_norm": 0.3314555883407593,
      "learning_rate": 8.453364817001182e-05,
      "loss": 0.7123,
      "step": 495
    },
    {
      "epoch": 0.5821596244131455,
      "grad_norm": 0.33951112627983093,
      "learning_rate": 8.429752066115702e-05,
      "loss": 0.7501,
      "step": 496
    },
    {
      "epoch": 0.5833333333333334,
      "grad_norm": 0.327809602022171,
      "learning_rate": 8.406139315230225e-05,
      "loss": 0.7543,
      "step": 497
    },
    {
      "epoch": 0.5845070422535211,
      "grad_norm": 0.33205023407936096,
      "learning_rate": 8.382526564344747e-05,
      "loss": 0.7395,
      "step": 498
    },
    {
      "epoch": 0.5856807511737089,
      "grad_norm": 0.3762659430503845,
      "learning_rate": 8.358913813459269e-05,
      "loss": 0.7424,
      "step": 499
    },
    {
      "epoch": 0.5868544600938967,
      "grad_norm": 0.3421575427055359,
      "learning_rate": 8.33530106257379e-05,
      "loss": 0.7167,
      "step": 500
    },
    {
      "epoch": 0.5880281690140845,
      "grad_norm": 0.3560996353626251,
      "learning_rate": 8.311688311688312e-05,
      "loss": 0.7464,
      "step": 501
    },
    {
      "epoch": 0.5892018779342723,
      "grad_norm": 0.3566039800643921,
      "learning_rate": 8.288075560802834e-05,
      "loss": 0.715,
      "step": 502
    },
    {
      "epoch": 0.5903755868544601,
      "grad_norm": 0.3481593430042267,
      "learning_rate": 8.264462809917356e-05,
      "loss": 0.7506,
      "step": 503
    },
    {
      "epoch": 0.5915492957746479,
      "grad_norm": 0.34428590536117554,
      "learning_rate": 8.240850059031878e-05,
      "loss": 0.7272,
      "step": 504
    },
    {
      "epoch": 0.5927230046948356,
      "grad_norm": 0.35629555583000183,
      "learning_rate": 8.2172373081464e-05,
      "loss": 0.7334,
      "step": 505
    },
    {
      "epoch": 0.5938967136150235,
      "grad_norm": 0.37292811274528503,
      "learning_rate": 8.193624557260921e-05,
      "loss": 0.7505,
      "step": 506
    },
    {
      "epoch": 0.5950704225352113,
      "grad_norm": 0.359614759683609,
      "learning_rate": 8.170011806375443e-05,
      "loss": 0.8006,
      "step": 507
    },
    {
      "epoch": 0.596244131455399,
      "grad_norm": 0.3388945460319519,
      "learning_rate": 8.146399055489965e-05,
      "loss": 0.7542,
      "step": 508
    },
    {
      "epoch": 0.5974178403755869,
      "grad_norm": 0.3528054356575012,
      "learning_rate": 8.122786304604487e-05,
      "loss": 0.7412,
      "step": 509
    },
    {
      "epoch": 0.5985915492957746,
      "grad_norm": 0.3354608416557312,
      "learning_rate": 8.099173553719009e-05,
      "loss": 0.7062,
      "step": 510
    },
    {
      "epoch": 0.5997652582159625,
      "grad_norm": 0.35168859362602234,
      "learning_rate": 8.07556080283353e-05,
      "loss": 0.7653,
      "step": 511
    },
    {
      "epoch": 0.6009389671361502,
      "grad_norm": 0.33843398094177246,
      "learning_rate": 8.051948051948052e-05,
      "loss": 0.7339,
      "step": 512
    },
    {
      "epoch": 0.602112676056338,
      "grad_norm": 0.32910212874412537,
      "learning_rate": 8.028335301062574e-05,
      "loss": 0.6966,
      "step": 513
    },
    {
      "epoch": 0.6032863849765259,
      "grad_norm": 0.3462936580181122,
      "learning_rate": 8.004722550177096e-05,
      "loss": 0.7386,
      "step": 514
    },
    {
      "epoch": 0.6044600938967136,
      "grad_norm": 0.3483426868915558,
      "learning_rate": 7.981109799291619e-05,
      "loss": 0.7548,
      "step": 515
    },
    {
      "epoch": 0.6056338028169014,
      "grad_norm": 0.3555918335914612,
      "learning_rate": 7.95749704840614e-05,
      "loss": 0.7144,
      "step": 516
    },
    {
      "epoch": 0.6068075117370892,
      "grad_norm": 0.3545628786087036,
      "learning_rate": 7.933884297520661e-05,
      "loss": 0.7601,
      "step": 517
    },
    {
      "epoch": 0.607981220657277,
      "grad_norm": 0.3554907441139221,
      "learning_rate": 7.910271546635183e-05,
      "loss": 0.7464,
      "step": 518
    },
    {
      "epoch": 0.6091549295774648,
      "grad_norm": 0.3457619547843933,
      "learning_rate": 7.886658795749705e-05,
      "loss": 0.7372,
      "step": 519
    },
    {
      "epoch": 0.6103286384976526,
      "grad_norm": 0.3450148105621338,
      "learning_rate": 7.863046044864228e-05,
      "loss": 0.7265,
      "step": 520
    },
    {
      "epoch": 0.6115023474178404,
      "grad_norm": 0.3475225567817688,
      "learning_rate": 7.839433293978748e-05,
      "loss": 0.798,
      "step": 521
    },
    {
      "epoch": 0.6126760563380281,
      "grad_norm": 0.34560921788215637,
      "learning_rate": 7.81582054309327e-05,
      "loss": 0.7583,
      "step": 522
    },
    {
      "epoch": 0.613849765258216,
      "grad_norm": 0.33480820059776306,
      "learning_rate": 7.792207792207793e-05,
      "loss": 0.7658,
      "step": 523
    },
    {
      "epoch": 0.6150234741784038,
      "grad_norm": 0.34581395983695984,
      "learning_rate": 7.768595041322314e-05,
      "loss": 0.7368,
      "step": 524
    },
    {
      "epoch": 0.6161971830985915,
      "grad_norm": 0.35383906960487366,
      "learning_rate": 7.744982290436837e-05,
      "loss": 0.7963,
      "step": 525
    },
    {
      "epoch": 0.6173708920187794,
      "grad_norm": 0.352117121219635,
      "learning_rate": 7.721369539551359e-05,
      "loss": 0.7589,
      "step": 526
    },
    {
      "epoch": 0.6185446009389671,
      "grad_norm": 0.34420257806777954,
      "learning_rate": 7.697756788665879e-05,
      "loss": 0.7209,
      "step": 527
    },
    {
      "epoch": 0.6197183098591549,
      "grad_norm": 0.3449562191963196,
      "learning_rate": 7.674144037780402e-05,
      "loss": 0.7526,
      "step": 528
    },
    {
      "epoch": 0.6208920187793427,
      "grad_norm": 0.37377694249153137,
      "learning_rate": 7.650531286894924e-05,
      "loss": 0.7348,
      "step": 529
    },
    {
      "epoch": 0.6220657276995305,
      "grad_norm": 0.32662031054496765,
      "learning_rate": 7.626918536009446e-05,
      "loss": 0.7125,
      "step": 530
    },
    {
      "epoch": 0.6232394366197183,
      "grad_norm": 0.3551415801048279,
      "learning_rate": 7.603305785123968e-05,
      "loss": 0.7497,
      "step": 531
    },
    {
      "epoch": 0.6244131455399061,
      "grad_norm": 0.3519802689552307,
      "learning_rate": 7.579693034238488e-05,
      "loss": 0.7864,
      "step": 532
    },
    {
      "epoch": 0.6255868544600939,
      "grad_norm": 0.3773750364780426,
      "learning_rate": 7.556080283353011e-05,
      "loss": 0.7681,
      "step": 533
    },
    {
      "epoch": 0.6267605633802817,
      "grad_norm": 0.3558037281036377,
      "learning_rate": 7.532467532467533e-05,
      "loss": 0.7392,
      "step": 534
    },
    {
      "epoch": 0.6279342723004695,
      "grad_norm": 0.33910447359085083,
      "learning_rate": 7.508854781582055e-05,
      "loss": 0.7036,
      "step": 535
    },
    {
      "epoch": 0.6291079812206573,
      "grad_norm": 0.35620275139808655,
      "learning_rate": 7.485242030696577e-05,
      "loss": 0.7272,
      "step": 536
    },
    {
      "epoch": 0.6302816901408451,
      "grad_norm": 0.3377542495727539,
      "learning_rate": 7.461629279811098e-05,
      "loss": 0.7244,
      "step": 537
    },
    {
      "epoch": 0.6314553990610329,
      "grad_norm": 0.35217198729515076,
      "learning_rate": 7.43801652892562e-05,
      "loss": 0.7655,
      "step": 538
    },
    {
      "epoch": 0.6326291079812206,
      "grad_norm": 0.34656718373298645,
      "learning_rate": 7.414403778040142e-05,
      "loss": 0.7474,
      "step": 539
    },
    {
      "epoch": 0.6338028169014085,
      "grad_norm": 0.34429579973220825,
      "learning_rate": 7.390791027154664e-05,
      "loss": 0.7333,
      "step": 540
    },
    {
      "epoch": 0.6349765258215962,
      "grad_norm": 0.374262273311615,
      "learning_rate": 7.367178276269186e-05,
      "loss": 0.7876,
      "step": 541
    },
    {
      "epoch": 0.636150234741784,
      "grad_norm": 0.363299161195755,
      "learning_rate": 7.343565525383707e-05,
      "loss": 0.7784,
      "step": 542
    },
    {
      "epoch": 0.6373239436619719,
      "grad_norm": 0.36767125129699707,
      "learning_rate": 7.31995277449823e-05,
      "loss": 0.7329,
      "step": 543
    },
    {
      "epoch": 0.6384976525821596,
      "grad_norm": 0.3338686525821686,
      "learning_rate": 7.296340023612751e-05,
      "loss": 0.7737,
      "step": 544
    },
    {
      "epoch": 0.6396713615023474,
      "grad_norm": 0.3493046164512634,
      "learning_rate": 7.272727272727273e-05,
      "loss": 0.7461,
      "step": 545
    },
    {
      "epoch": 0.6408450704225352,
      "grad_norm": 0.3691573441028595,
      "learning_rate": 7.249114521841795e-05,
      "loss": 0.765,
      "step": 546
    },
    {
      "epoch": 0.642018779342723,
      "grad_norm": 0.3573099374771118,
      "learning_rate": 7.225501770956316e-05,
      "loss": 0.7589,
      "step": 547
    },
    {
      "epoch": 0.6431924882629108,
      "grad_norm": 0.36218926310539246,
      "learning_rate": 7.201889020070838e-05,
      "loss": 0.7314,
      "step": 548
    },
    {
      "epoch": 0.6443661971830986,
      "grad_norm": 0.35753628611564636,
      "learning_rate": 7.17827626918536e-05,
      "loss": 0.7564,
      "step": 549
    },
    {
      "epoch": 0.6455399061032864,
      "grad_norm": 0.3394756615161896,
      "learning_rate": 7.154663518299882e-05,
      "loss": 0.7162,
      "step": 550
    },
    {
      "epoch": 0.6467136150234741,
      "grad_norm": 0.350090891122818,
      "learning_rate": 7.131050767414405e-05,
      "loss": 0.7561,
      "step": 551
    },
    {
      "epoch": 0.647887323943662,
      "grad_norm": 0.328924298286438,
      "learning_rate": 7.107438016528925e-05,
      "loss": 0.7143,
      "step": 552
    },
    {
      "epoch": 0.6490610328638498,
      "grad_norm": 0.3552818298339844,
      "learning_rate": 7.083825265643447e-05,
      "loss": 0.7264,
      "step": 553
    },
    {
      "epoch": 0.6502347417840375,
      "grad_norm": 0.3504960536956787,
      "learning_rate": 7.06021251475797e-05,
      "loss": 0.7512,
      "step": 554
    },
    {
      "epoch": 0.6514084507042254,
      "grad_norm": 0.33755823969841003,
      "learning_rate": 7.036599763872491e-05,
      "loss": 0.7621,
      "step": 555
    },
    {
      "epoch": 0.6525821596244131,
      "grad_norm": 0.35977354645729065,
      "learning_rate": 7.012987012987014e-05,
      "loss": 0.776,
      "step": 556
    },
    {
      "epoch": 0.653755868544601,
      "grad_norm": 0.37304726243019104,
      "learning_rate": 6.989374262101536e-05,
      "loss": 0.7601,
      "step": 557
    },
    {
      "epoch": 0.6549295774647887,
      "grad_norm": 0.3569071590900421,
      "learning_rate": 6.965761511216056e-05,
      "loss": 0.7303,
      "step": 558
    },
    {
      "epoch": 0.6561032863849765,
      "grad_norm": 0.348264217376709,
      "learning_rate": 6.94214876033058e-05,
      "loss": 0.759,
      "step": 559
    },
    {
      "epoch": 0.6572769953051644,
      "grad_norm": 0.3501366674900055,
      "learning_rate": 6.9185360094451e-05,
      "loss": 0.7588,
      "step": 560
    },
    {
      "epoch": 0.6584507042253521,
      "grad_norm": 0.3633224666118622,
      "learning_rate": 6.894923258559623e-05,
      "loss": 0.7741,
      "step": 561
    },
    {
      "epoch": 0.6596244131455399,
      "grad_norm": 0.35944506525993347,
      "learning_rate": 6.871310507674145e-05,
      "loss": 0.756,
      "step": 562
    },
    {
      "epoch": 0.6607981220657277,
      "grad_norm": 0.3479359745979309,
      "learning_rate": 6.847697756788665e-05,
      "loss": 0.7292,
      "step": 563
    },
    {
      "epoch": 0.6619718309859155,
      "grad_norm": 0.37013959884643555,
      "learning_rate": 6.824085005903188e-05,
      "loss": 0.7618,
      "step": 564
    },
    {
      "epoch": 0.6631455399061033,
      "grad_norm": 0.36679190397262573,
      "learning_rate": 6.80047225501771e-05,
      "loss": 0.7797,
      "step": 565
    },
    {
      "epoch": 0.6643192488262911,
      "grad_norm": 0.35092490911483765,
      "learning_rate": 6.776859504132232e-05,
      "loss": 0.705,
      "step": 566
    },
    {
      "epoch": 0.6654929577464789,
      "grad_norm": 0.3594275712966919,
      "learning_rate": 6.753246753246754e-05,
      "loss": 0.7215,
      "step": 567
    },
    {
      "epoch": 0.6666666666666666,
      "grad_norm": 0.3503059148788452,
      "learning_rate": 6.729634002361276e-05,
      "loss": 0.7248,
      "step": 568
    },
    {
      "epoch": 0.6678403755868545,
      "grad_norm": 0.35919633507728577,
      "learning_rate": 6.706021251475797e-05,
      "loss": 0.7718,
      "step": 569
    },
    {
      "epoch": 0.6690140845070423,
      "grad_norm": 0.36752262711524963,
      "learning_rate": 6.682408500590319e-05,
      "loss": 0.7738,
      "step": 570
    },
    {
      "epoch": 0.67018779342723,
      "grad_norm": 0.33812567591667175,
      "learning_rate": 6.658795749704841e-05,
      "loss": 0.7846,
      "step": 571
    },
    {
      "epoch": 0.6713615023474179,
      "grad_norm": 0.3429810404777527,
      "learning_rate": 6.635182998819363e-05,
      "loss": 0.7371,
      "step": 572
    },
    {
      "epoch": 0.6725352112676056,
      "grad_norm": 0.3457571864128113,
      "learning_rate": 6.611570247933885e-05,
      "loss": 0.7318,
      "step": 573
    },
    {
      "epoch": 0.6737089201877934,
      "grad_norm": 0.3476294279098511,
      "learning_rate": 6.587957497048406e-05,
      "loss": 0.7344,
      "step": 574
    },
    {
      "epoch": 0.6748826291079812,
      "grad_norm": 0.34464409947395325,
      "learning_rate": 6.564344746162928e-05,
      "loss": 0.7429,
      "step": 575
    },
    {
      "epoch": 0.676056338028169,
      "grad_norm": 0.34444373846054077,
      "learning_rate": 6.54073199527745e-05,
      "loss": 0.7663,
      "step": 576
    },
    {
      "epoch": 0.6772300469483568,
      "grad_norm": 0.3656728267669678,
      "learning_rate": 6.517119244391972e-05,
      "loss": 0.7068,
      "step": 577
    },
    {
      "epoch": 0.6784037558685446,
      "grad_norm": 0.3591727614402771,
      "learning_rate": 6.493506493506494e-05,
      "loss": 0.7481,
      "step": 578
    },
    {
      "epoch": 0.6795774647887324,
      "grad_norm": 0.38865676522254944,
      "learning_rate": 6.469893742621017e-05,
      "loss": 0.7659,
      "step": 579
    },
    {
      "epoch": 0.6807511737089202,
      "grad_norm": 0.3438194990158081,
      "learning_rate": 6.446280991735537e-05,
      "loss": 0.6748,
      "step": 580
    },
    {
      "epoch": 0.681924882629108,
      "grad_norm": 0.34979990124702454,
      "learning_rate": 6.422668240850059e-05,
      "loss": 0.7529,
      "step": 581
    },
    {
      "epoch": 0.6830985915492958,
      "grad_norm": 0.37309062480926514,
      "learning_rate": 6.399055489964582e-05,
      "loss": 0.7417,
      "step": 582
    },
    {
      "epoch": 0.6842723004694836,
      "grad_norm": 0.3737837076187134,
      "learning_rate": 6.375442739079102e-05,
      "loss": 0.773,
      "step": 583
    },
    {
      "epoch": 0.6854460093896714,
      "grad_norm": 0.3397013247013092,
      "learning_rate": 6.351829988193624e-05,
      "loss": 0.7093,
      "step": 584
    },
    {
      "epoch": 0.6866197183098591,
      "grad_norm": 0.37165701389312744,
      "learning_rate": 6.328217237308147e-05,
      "loss": 0.7078,
      "step": 585
    },
    {
      "epoch": 0.687793427230047,
      "grad_norm": 0.3533116579055786,
      "learning_rate": 6.304604486422668e-05,
      "loss": 0.7105,
      "step": 586
    },
    {
      "epoch": 0.6889671361502347,
      "grad_norm": 0.35352569818496704,
      "learning_rate": 6.280991735537191e-05,
      "loss": 0.7282,
      "step": 587
    },
    {
      "epoch": 0.6901408450704225,
      "grad_norm": 0.3754810094833374,
      "learning_rate": 6.257378984651711e-05,
      "loss": 0.7364,
      "step": 588
    },
    {
      "epoch": 0.6913145539906104,
      "grad_norm": 0.36235493421554565,
      "learning_rate": 6.233766233766233e-05,
      "loss": 0.7024,
      "step": 589
    },
    {
      "epoch": 0.6924882629107981,
      "grad_norm": 0.3446933627128601,
      "learning_rate": 6.210153482880756e-05,
      "loss": 0.7392,
      "step": 590
    },
    {
      "epoch": 0.6936619718309859,
      "grad_norm": 0.34918078780174255,
      "learning_rate": 6.186540731995277e-05,
      "loss": 0.6716,
      "step": 591
    },
    {
      "epoch": 0.6948356807511737,
      "grad_norm": 0.3438567519187927,
      "learning_rate": 6.1629279811098e-05,
      "loss": 0.7812,
      "step": 592
    },
    {
      "epoch": 0.6960093896713615,
      "grad_norm": 0.346626341342926,
      "learning_rate": 6.139315230224322e-05,
      "loss": 0.7538,
      "step": 593
    },
    {
      "epoch": 0.6971830985915493,
      "grad_norm": 0.3506343960762024,
      "learning_rate": 6.115702479338842e-05,
      "loss": 0.7434,
      "step": 594
    },
    {
      "epoch": 0.6983568075117371,
      "grad_norm": 0.35403555631637573,
      "learning_rate": 6.0920897284533654e-05,
      "loss": 0.7333,
      "step": 595
    },
    {
      "epoch": 0.6995305164319249,
      "grad_norm": 0.3391430377960205,
      "learning_rate": 6.068476977567887e-05,
      "loss": 0.7486,
      "step": 596
    },
    {
      "epoch": 0.7007042253521126,
      "grad_norm": 0.33783578872680664,
      "learning_rate": 6.044864226682408e-05,
      "loss": 0.7588,
      "step": 597
    },
    {
      "epoch": 0.7018779342723005,
      "grad_norm": 0.3333738446235657,
      "learning_rate": 6.021251475796931e-05,
      "loss": 0.7268,
      "step": 598
    },
    {
      "epoch": 0.7030516431924883,
      "grad_norm": 0.3494018316268921,
      "learning_rate": 5.997638724911453e-05,
      "loss": 0.7363,
      "step": 599
    },
    {
      "epoch": 0.704225352112676,
      "grad_norm": 0.34416642785072327,
      "learning_rate": 5.9740259740259744e-05,
      "loss": 0.7322,
      "step": 600
    },
    {
      "epoch": 0.7053990610328639,
      "grad_norm": 0.3523387312889099,
      "learning_rate": 5.950413223140496e-05,
      "loss": 0.6986,
      "step": 601
    },
    {
      "epoch": 0.7065727699530516,
      "grad_norm": 0.33000361919403076,
      "learning_rate": 5.926800472255017e-05,
      "loss": 0.7535,
      "step": 602
    },
    {
      "epoch": 0.7077464788732394,
      "grad_norm": 0.33932214975357056,
      "learning_rate": 5.90318772136954e-05,
      "loss": 0.7051,
      "step": 603
    },
    {
      "epoch": 0.7089201877934272,
      "grad_norm": 0.3373797833919525,
      "learning_rate": 5.8795749704840616e-05,
      "loss": 0.7022,
      "step": 604
    },
    {
      "epoch": 0.710093896713615,
      "grad_norm": 0.35239875316619873,
      "learning_rate": 5.855962219598583e-05,
      "loss": 0.7893,
      "step": 605
    },
    {
      "epoch": 0.7112676056338029,
      "grad_norm": 0.36973506212234497,
      "learning_rate": 5.832349468713105e-05,
      "loss": 0.7157,
      "step": 606
    },
    {
      "epoch": 0.7124413145539906,
      "grad_norm": 0.3447434604167938,
      "learning_rate": 5.8087367178276277e-05,
      "loss": 0.7306,
      "step": 607
    },
    {
      "epoch": 0.7136150234741784,
      "grad_norm": 0.36380118131637573,
      "learning_rate": 5.785123966942149e-05,
      "loss": 0.7238,
      "step": 608
    },
    {
      "epoch": 0.7147887323943662,
      "grad_norm": 0.33784252405166626,
      "learning_rate": 5.7615112160566706e-05,
      "loss": 0.6792,
      "step": 609
    },
    {
      "epoch": 0.715962441314554,
      "grad_norm": 0.34995025396347046,
      "learning_rate": 5.737898465171193e-05,
      "loss": 0.7158,
      "step": 610
    },
    {
      "epoch": 0.7171361502347418,
      "grad_norm": 0.3586655259132385,
      "learning_rate": 5.714285714285714e-05,
      "loss": 0.7345,
      "step": 611
    },
    {
      "epoch": 0.7183098591549296,
      "grad_norm": 0.3490711450576782,
      "learning_rate": 5.6906729634002366e-05,
      "loss": 0.759,
      "step": 612
    },
    {
      "epoch": 0.7194835680751174,
      "grad_norm": 0.3405636250972748,
      "learning_rate": 5.6670602125147584e-05,
      "loss": 0.7069,
      "step": 613
    },
    {
      "epoch": 0.7206572769953051,
      "grad_norm": 0.3362460136413574,
      "learning_rate": 5.6434474616292796e-05,
      "loss": 0.7413,
      "step": 614
    },
    {
      "epoch": 0.721830985915493,
      "grad_norm": 0.3571033775806427,
      "learning_rate": 5.619834710743802e-05,
      "loss": 0.7138,
      "step": 615
    },
    {
      "epoch": 0.7230046948356808,
      "grad_norm": 0.33801379799842834,
      "learning_rate": 5.5962219598583245e-05,
      "loss": 0.7004,
      "step": 616
    },
    {
      "epoch": 0.7241784037558685,
      "grad_norm": 0.350063294172287,
      "learning_rate": 5.5726092089728456e-05,
      "loss": 0.7342,
      "step": 617
    },
    {
      "epoch": 0.7253521126760564,
      "grad_norm": 0.3471220135688782,
      "learning_rate": 5.5489964580873674e-05,
      "loss": 0.7591,
      "step": 618
    },
    {
      "epoch": 0.7265258215962441,
      "grad_norm": 0.3600592613220215,
      "learning_rate": 5.5253837072018886e-05,
      "loss": 0.7427,
      "step": 619
    },
    {
      "epoch": 0.7276995305164319,
      "grad_norm": 0.34294822812080383,
      "learning_rate": 5.501770956316411e-05,
      "loss": 0.7085,
      "step": 620
    },
    {
      "epoch": 0.7288732394366197,
      "grad_norm": 0.3481101989746094,
      "learning_rate": 5.4781582054309335e-05,
      "loss": 0.7465,
      "step": 621
    },
    {
      "epoch": 0.7300469483568075,
      "grad_norm": 0.3402861952781677,
      "learning_rate": 5.4545454545454546e-05,
      "loss": 0.7613,
      "step": 622
    },
    {
      "epoch": 0.7312206572769953,
      "grad_norm": 0.3475019335746765,
      "learning_rate": 5.4309327036599764e-05,
      "loss": 0.775,
      "step": 623
    },
    {
      "epoch": 0.7323943661971831,
      "grad_norm": 0.34003034234046936,
      "learning_rate": 5.407319952774499e-05,
      "loss": 0.6817,
      "step": 624
    },
    {
      "epoch": 0.7335680751173709,
      "grad_norm": 0.33620044589042664,
      "learning_rate": 5.38370720188902e-05,
      "loss": 0.7392,
      "step": 625
    },
    {
      "epoch": 0.7347417840375586,
      "grad_norm": 0.34645119309425354,
      "learning_rate": 5.3600944510035425e-05,
      "loss": 0.717,
      "step": 626
    },
    {
      "epoch": 0.7359154929577465,
      "grad_norm": 0.3485560417175293,
      "learning_rate": 5.336481700118064e-05,
      "loss": 0.7361,
      "step": 627
    },
    {
      "epoch": 0.7370892018779343,
      "grad_norm": 0.36997392773628235,
      "learning_rate": 5.3128689492325854e-05,
      "loss": 0.7264,
      "step": 628
    },
    {
      "epoch": 0.7382629107981221,
      "grad_norm": 0.3379404842853546,
      "learning_rate": 5.289256198347108e-05,
      "loss": 0.7303,
      "step": 629
    },
    {
      "epoch": 0.7394366197183099,
      "grad_norm": 0.3385223150253296,
      "learning_rate": 5.2656434474616304e-05,
      "loss": 0.7174,
      "step": 630
    },
    {
      "epoch": 0.7406103286384976,
      "grad_norm": 0.38303306698799133,
      "learning_rate": 5.2420306965761515e-05,
      "loss": 0.7539,
      "step": 631
    },
    {
      "epoch": 0.7417840375586855,
      "grad_norm": 0.3544706404209137,
      "learning_rate": 5.218417945690673e-05,
      "loss": 0.7108,
      "step": 632
    },
    {
      "epoch": 0.7429577464788732,
      "grad_norm": 0.35137131810188293,
      "learning_rate": 5.1948051948051944e-05,
      "loss": 0.7184,
      "step": 633
    },
    {
      "epoch": 0.744131455399061,
      "grad_norm": 0.35326629877090454,
      "learning_rate": 5.171192443919717e-05,
      "loss": 0.7114,
      "step": 634
    },
    {
      "epoch": 0.7453051643192489,
      "grad_norm": 0.35051414370536804,
      "learning_rate": 5.1475796930342393e-05,
      "loss": 0.6966,
      "step": 635
    },
    {
      "epoch": 0.7464788732394366,
      "grad_norm": 0.37491628527641296,
      "learning_rate": 5.1239669421487605e-05,
      "loss": 0.7061,
      "step": 636
    },
    {
      "epoch": 0.7476525821596244,
      "grad_norm": 0.37242433428764343,
      "learning_rate": 5.100354191263282e-05,
      "loss": 0.6904,
      "step": 637
    },
    {
      "epoch": 0.7488262910798122,
      "grad_norm": 0.376429945230484,
      "learning_rate": 5.076741440377805e-05,
      "loss": 0.7203,
      "step": 638
    },
    {
      "epoch": 0.75,
      "grad_norm": 0.34106218814849854,
      "learning_rate": 5.053128689492326e-05,
      "loss": 0.6878,
      "step": 639
    },
    {
      "epoch": 0.7511737089201878,
      "grad_norm": 0.37987956404685974,
      "learning_rate": 5.029515938606848e-05,
      "loss": 0.7835,
      "step": 640
    },
    {
      "epoch": 0.7523474178403756,
      "grad_norm": 0.355932354927063,
      "learning_rate": 5.00590318772137e-05,
      "loss": 0.7382,
      "step": 641
    },
    {
      "epoch": 0.7535211267605634,
      "grad_norm": 0.33495378494262695,
      "learning_rate": 4.982290436835891e-05,
      "loss": 0.7244,
      "step": 642
    },
    {
      "epoch": 0.7546948356807511,
      "grad_norm": 0.36573663353919983,
      "learning_rate": 4.958677685950414e-05,
      "loss": 0.7339,
      "step": 643
    },
    {
      "epoch": 0.755868544600939,
      "grad_norm": 0.34233418107032776,
      "learning_rate": 4.9350649350649355e-05,
      "loss": 0.7303,
      "step": 644
    },
    {
      "epoch": 0.7570422535211268,
      "grad_norm": 0.36358365416526794,
      "learning_rate": 4.9114521841794566e-05,
      "loss": 0.7169,
      "step": 645
    },
    {
      "epoch": 0.7582159624413145,
      "grad_norm": 0.3423750400543213,
      "learning_rate": 4.887839433293979e-05,
      "loss": 0.7413,
      "step": 646
    },
    {
      "epoch": 0.7593896713615024,
      "grad_norm": 0.34080007672309875,
      "learning_rate": 4.864226682408501e-05,
      "loss": 0.7319,
      "step": 647
    },
    {
      "epoch": 0.7605633802816901,
      "grad_norm": 0.35408544540405273,
      "learning_rate": 4.840613931523023e-05,
      "loss": 0.6895,
      "step": 648
    },
    {
      "epoch": 0.7617370892018779,
      "grad_norm": 0.34515753388404846,
      "learning_rate": 4.8170011806375445e-05,
      "loss": 0.7181,
      "step": 649
    },
    {
      "epoch": 0.7629107981220657,
      "grad_norm": 0.3446560502052307,
      "learning_rate": 4.793388429752066e-05,
      "loss": 0.7156,
      "step": 650
    },
    {
      "epoch": 0.7640845070422535,
      "grad_norm": 0.3451150357723236,
      "learning_rate": 4.769775678866588e-05,
      "loss": 0.7232,
      "step": 651
    },
    {
      "epoch": 0.7652582159624414,
      "grad_norm": 0.357740193605423,
      "learning_rate": 4.74616292798111e-05,
      "loss": 0.6872,
      "step": 652
    },
    {
      "epoch": 0.7664319248826291,
      "grad_norm": 0.3685015141963959,
      "learning_rate": 4.7225501770956324e-05,
      "loss": 0.735,
      "step": 653
    },
    {
      "epoch": 0.7676056338028169,
      "grad_norm": 0.3503192961215973,
      "learning_rate": 4.6989374262101535e-05,
      "loss": 0.7336,
      "step": 654
    },
    {
      "epoch": 0.7687793427230047,
      "grad_norm": 0.33453887701034546,
      "learning_rate": 4.675324675324675e-05,
      "loss": 0.7101,
      "step": 655
    },
    {
      "epoch": 0.7699530516431925,
      "grad_norm": 0.3708442747592926,
      "learning_rate": 4.651711924439197e-05,
      "loss": 0.7153,
      "step": 656
    },
    {
      "epoch": 0.7711267605633803,
      "grad_norm": 0.3736172318458557,
      "learning_rate": 4.6280991735537196e-05,
      "loss": 0.7071,
      "step": 657
    },
    {
      "epoch": 0.7723004694835681,
      "grad_norm": 0.35988256335258484,
      "learning_rate": 4.6044864226682414e-05,
      "loss": 0.7285,
      "step": 658
    },
    {
      "epoch": 0.7734741784037559,
      "grad_norm": 0.34314337372779846,
      "learning_rate": 4.5808736717827625e-05,
      "loss": 0.7137,
      "step": 659
    },
    {
      "epoch": 0.7746478873239436,
      "grad_norm": 0.3723309338092804,
      "learning_rate": 4.557260920897285e-05,
      "loss": 0.7391,
      "step": 660
    },
    {
      "epoch": 0.7758215962441315,
      "grad_norm": 0.3581268787384033,
      "learning_rate": 4.533648170011807e-05,
      "loss": 0.7157,
      "step": 661
    },
    {
      "epoch": 0.7769953051643192,
      "grad_norm": 0.36784443259239197,
      "learning_rate": 4.5100354191263286e-05,
      "loss": 0.6865,
      "step": 662
    },
    {
      "epoch": 0.778169014084507,
      "grad_norm": 0.36377546191215515,
      "learning_rate": 4.48642266824085e-05,
      "loss": 0.7437,
      "step": 663
    },
    {
      "epoch": 0.7793427230046949,
      "grad_norm": 0.349101722240448,
      "learning_rate": 4.462809917355372e-05,
      "loss": 0.7226,
      "step": 664
    },
    {
      "epoch": 0.7805164319248826,
      "grad_norm": 0.36608216166496277,
      "learning_rate": 4.439197166469894e-05,
      "loss": 0.7543,
      "step": 665
    },
    {
      "epoch": 0.7816901408450704,
      "grad_norm": 0.3495696783065796,
      "learning_rate": 4.415584415584416e-05,
      "loss": 0.708,
      "step": 666
    },
    {
      "epoch": 0.7828638497652582,
      "grad_norm": 0.3664140999317169,
      "learning_rate": 4.3919716646989375e-05,
      "loss": 0.7225,
      "step": 667
    },
    {
      "epoch": 0.784037558685446,
      "grad_norm": 0.3560849726200104,
      "learning_rate": 4.368358913813459e-05,
      "loss": 0.6972,
      "step": 668
    },
    {
      "epoch": 0.7852112676056338,
      "grad_norm": 0.3571857511997223,
      "learning_rate": 4.344746162927981e-05,
      "loss": 0.694,
      "step": 669
    },
    {
      "epoch": 0.7863849765258216,
      "grad_norm": 0.37072160840034485,
      "learning_rate": 4.321133412042503e-05,
      "loss": 0.7202,
      "step": 670
    },
    {
      "epoch": 0.7875586854460094,
      "grad_norm": 0.354948490858078,
      "learning_rate": 4.2975206611570254e-05,
      "loss": 0.7481,
      "step": 671
    },
    {
      "epoch": 0.7887323943661971,
      "grad_norm": 0.3736347258090973,
      "learning_rate": 4.2739079102715465e-05,
      "loss": 0.7261,
      "step": 672
    },
    {
      "epoch": 0.789906103286385,
      "grad_norm": 0.3690294623374939,
      "learning_rate": 4.250295159386068e-05,
      "loss": 0.7529,
      "step": 673
    },
    {
      "epoch": 0.7910798122065728,
      "grad_norm": 0.354192316532135,
      "learning_rate": 4.226682408500591e-05,
      "loss": 0.7176,
      "step": 674
    },
    {
      "epoch": 0.7922535211267606,
      "grad_norm": 0.355185866355896,
      "learning_rate": 4.2030696576151126e-05,
      "loss": 0.7099,
      "step": 675
    },
    {
      "epoch": 0.7934272300469484,
      "grad_norm": 0.3503565490245819,
      "learning_rate": 4.1794569067296344e-05,
      "loss": 0.7072,
      "step": 676
    },
    {
      "epoch": 0.7946009389671361,
      "grad_norm": 0.3727845549583435,
      "learning_rate": 4.155844155844156e-05,
      "loss": 0.7334,
      "step": 677
    },
    {
      "epoch": 0.795774647887324,
      "grad_norm": 0.33894312381744385,
      "learning_rate": 4.132231404958678e-05,
      "loss": 0.6946,
      "step": 678
    },
    {
      "epoch": 0.7969483568075117,
      "grad_norm": 0.3385523855686188,
      "learning_rate": 4.1086186540732e-05,
      "loss": 0.7096,
      "step": 679
    },
    {
      "epoch": 0.7981220657276995,
      "grad_norm": 0.3488437235355377,
      "learning_rate": 4.0850059031877216e-05,
      "loss": 0.6942,
      "step": 680
    },
    {
      "epoch": 0.7992957746478874,
      "grad_norm": 0.34666576981544495,
      "learning_rate": 4.0613931523022434e-05,
      "loss": 0.7329,
      "step": 681
    },
    {
      "epoch": 0.8004694835680751,
      "grad_norm": 0.3557136356830597,
      "learning_rate": 4.037780401416765e-05,
      "loss": 0.7655,
      "step": 682
    },
    {
      "epoch": 0.8016431924882629,
      "grad_norm": 0.3647683262825012,
      "learning_rate": 4.014167650531287e-05,
      "loss": 0.7578,
      "step": 683
    },
    {
      "epoch": 0.8028169014084507,
      "grad_norm": 0.3452191650867462,
      "learning_rate": 3.9905548996458095e-05,
      "loss": 0.7145,
      "step": 684
    },
    {
      "epoch": 0.8039906103286385,
      "grad_norm": 0.3540481925010681,
      "learning_rate": 3.9669421487603306e-05,
      "loss": 0.7347,
      "step": 685
    },
    {
      "epoch": 0.8051643192488263,
      "grad_norm": 0.3536418378353119,
      "learning_rate": 3.9433293978748524e-05,
      "loss": 0.7103,
      "step": 686
    },
    {
      "epoch": 0.8063380281690141,
      "grad_norm": 0.34728798270225525,
      "learning_rate": 3.919716646989374e-05,
      "loss": 0.7376,
      "step": 687
    },
    {
      "epoch": 0.8075117370892019,
      "grad_norm": 0.354643851518631,
      "learning_rate": 3.8961038961038966e-05,
      "loss": 0.7223,
      "step": 688
    },
    {
      "epoch": 0.8086854460093896,
      "grad_norm": 0.3438583016395569,
      "learning_rate": 3.8724911452184184e-05,
      "loss": 0.6906,
      "step": 689
    },
    {
      "epoch": 0.8098591549295775,
      "grad_norm": 0.34713107347488403,
      "learning_rate": 3.8488783943329396e-05,
      "loss": 0.7361,
      "step": 690
    },
    {
      "epoch": 0.8110328638497653,
      "grad_norm": 0.3483150005340576,
      "learning_rate": 3.825265643447462e-05,
      "loss": 0.7016,
      "step": 691
    },
    {
      "epoch": 0.812206572769953,
      "grad_norm": 0.34848445653915405,
      "learning_rate": 3.801652892561984e-05,
      "loss": 0.6966,
      "step": 692
    },
    {
      "epoch": 0.8133802816901409,
      "grad_norm": 0.34223318099975586,
      "learning_rate": 3.7780401416765056e-05,
      "loss": 0.7088,
      "step": 693
    },
    {
      "epoch": 0.8145539906103286,
      "grad_norm": 0.33693239092826843,
      "learning_rate": 3.7544273907910274e-05,
      "loss": 0.7108,
      "step": 694
    },
    {
      "epoch": 0.8157276995305164,
      "grad_norm": 0.34613272547721863,
      "learning_rate": 3.730814639905549e-05,
      "loss": 0.7075,
      "step": 695
    },
    {
      "epoch": 0.8169014084507042,
      "grad_norm": 0.3430733382701874,
      "learning_rate": 3.707201889020071e-05,
      "loss": 0.7246,
      "step": 696
    },
    {
      "epoch": 0.818075117370892,
      "grad_norm": 0.35237351059913635,
      "learning_rate": 3.683589138134593e-05,
      "loss": 0.6918,
      "step": 697
    },
    {
      "epoch": 0.8192488262910798,
      "grad_norm": 0.3375650644302368,
      "learning_rate": 3.659976387249115e-05,
      "loss": 0.6978,
      "step": 698
    },
    {
      "epoch": 0.8204225352112676,
      "grad_norm": 0.3585062026977539,
      "learning_rate": 3.6363636363636364e-05,
      "loss": 0.7241,
      "step": 699
    },
    {
      "epoch": 0.8215962441314554,
      "grad_norm": 0.35660460591316223,
      "learning_rate": 3.612750885478158e-05,
      "loss": 0.6946,
      "step": 700
    },
    {
      "epoch": 0.8227699530516432,
      "grad_norm": 0.3468845784664154,
      "learning_rate": 3.58913813459268e-05,
      "loss": 0.7535,
      "step": 701
    },
    {
      "epoch": 0.823943661971831,
      "grad_norm": 0.365291029214859,
      "learning_rate": 3.5655253837072025e-05,
      "loss": 0.7438,
      "step": 702
    },
    {
      "epoch": 0.8251173708920188,
      "grad_norm": 0.353506863117218,
      "learning_rate": 3.5419126328217236e-05,
      "loss": 0.7359,
      "step": 703
    },
    {
      "epoch": 0.8262910798122066,
      "grad_norm": 0.381610244512558,
      "learning_rate": 3.5182998819362454e-05,
      "loss": 0.7821,
      "step": 704
    },
    {
      "epoch": 0.8274647887323944,
      "grad_norm": 0.37710806727409363,
      "learning_rate": 3.494687131050768e-05,
      "loss": 0.7349,
      "step": 705
    },
    {
      "epoch": 0.8286384976525821,
      "grad_norm": 0.361545592546463,
      "learning_rate": 3.47107438016529e-05,
      "loss": 0.7229,
      "step": 706
    },
    {
      "epoch": 0.82981220657277,
      "grad_norm": 0.3615299463272095,
      "learning_rate": 3.4474616292798115e-05,
      "loss": 0.748,
      "step": 707
    },
    {
      "epoch": 0.8309859154929577,
      "grad_norm": 0.3437252342700958,
      "learning_rate": 3.4238488783943326e-05,
      "loss": 0.7165,
      "step": 708
    },
    {
      "epoch": 0.8321596244131455,
      "grad_norm": 0.35603129863739014,
      "learning_rate": 3.400236127508855e-05,
      "loss": 0.7373,
      "step": 709
    },
    {
      "epoch": 0.8333333333333334,
      "grad_norm": 0.3586898446083069,
      "learning_rate": 3.376623376623377e-05,
      "loss": 0.7056,
      "step": 710
    },
    {
      "epoch": 0.8345070422535211,
      "grad_norm": 0.3558507263660431,
      "learning_rate": 3.353010625737899e-05,
      "loss": 0.742,
      "step": 711
    },
    {
      "epoch": 0.8356807511737089,
      "grad_norm": 0.3359735608100891,
      "learning_rate": 3.3293978748524205e-05,
      "loss": 0.6994,
      "step": 712
    },
    {
      "epoch": 0.8368544600938967,
      "grad_norm": 0.34250345826148987,
      "learning_rate": 3.305785123966942e-05,
      "loss": 0.6762,
      "step": 713
    },
    {
      "epoch": 0.8380281690140845,
      "grad_norm": 0.38417667150497437,
      "learning_rate": 3.282172373081464e-05,
      "loss": 0.7213,
      "step": 714
    },
    {
      "epoch": 0.8392018779342723,
      "grad_norm": 0.3643978536128998,
      "learning_rate": 3.258559622195986e-05,
      "loss": 0.6884,
      "step": 715
    },
    {
      "epoch": 0.8403755868544601,
      "grad_norm": 0.3544299602508545,
      "learning_rate": 3.234946871310508e-05,
      "loss": 0.712,
      "step": 716
    },
    {
      "epoch": 0.8415492957746479,
      "grad_norm": 0.36903661489486694,
      "learning_rate": 3.2113341204250294e-05,
      "loss": 0.7227,
      "step": 717
    },
    {
      "epoch": 0.8427230046948356,
      "grad_norm": 0.3557377755641937,
      "learning_rate": 3.187721369539551e-05,
      "loss": 0.6904,
      "step": 718
    },
    {
      "epoch": 0.8438967136150235,
      "grad_norm": 0.36762547492980957,
      "learning_rate": 3.164108618654074e-05,
      "loss": 0.7469,
      "step": 719
    },
    {
      "epoch": 0.8450704225352113,
      "grad_norm": 0.35805556178092957,
      "learning_rate": 3.1404958677685955e-05,
      "loss": 0.7443,
      "step": 720
    },
    {
      "epoch": 0.846244131455399,
      "grad_norm": 0.38130536675453186,
      "learning_rate": 3.1168831168831166e-05,
      "loss": 0.7664,
      "step": 721
    },
    {
      "epoch": 0.8474178403755869,
      "grad_norm": 0.3599521219730377,
      "learning_rate": 3.0932703659976384e-05,
      "loss": 0.7065,
      "step": 722
    },
    {
      "epoch": 0.8485915492957746,
      "grad_norm": 0.3531062602996826,
      "learning_rate": 3.069657615112161e-05,
      "loss": 0.7451,
      "step": 723
    },
    {
      "epoch": 0.8497652582159625,
      "grad_norm": 0.36916878819465637,
      "learning_rate": 3.0460448642266827e-05,
      "loss": 0.7077,
      "step": 724
    },
    {
      "epoch": 0.8509389671361502,
      "grad_norm": 0.38139578700065613,
      "learning_rate": 3.022432113341204e-05,
      "loss": 0.7452,
      "step": 725
    },
    {
      "epoch": 0.852112676056338,
      "grad_norm": 0.337944358587265,
      "learning_rate": 2.9988193624557266e-05,
      "loss": 0.6596,
      "step": 726
    },
    {
      "epoch": 0.8532863849765259,
      "grad_norm": 0.36196213960647583,
      "learning_rate": 2.975206611570248e-05,
      "loss": 0.7081,
      "step": 727
    },
    {
      "epoch": 0.8544600938967136,
      "grad_norm": 0.34913602471351624,
      "learning_rate": 2.95159386068477e-05,
      "loss": 0.6901,
      "step": 728
    },
    {
      "epoch": 0.8556338028169014,
      "grad_norm": 0.343414843082428,
      "learning_rate": 2.9279811097992914e-05,
      "loss": 0.675,
      "step": 729
    },
    {
      "epoch": 0.8568075117370892,
      "grad_norm": 0.3704102039337158,
      "learning_rate": 2.9043683589138138e-05,
      "loss": 0.7566,
      "step": 730
    },
    {
      "epoch": 0.857981220657277,
      "grad_norm": 0.3464911878108978,
      "learning_rate": 2.8807556080283353e-05,
      "loss": 0.6872,
      "step": 731
    },
    {
      "epoch": 0.8591549295774648,
      "grad_norm": 0.3615940511226654,
      "learning_rate": 2.857142857142857e-05,
      "loss": 0.7755,
      "step": 732
    },
    {
      "epoch": 0.8603286384976526,
      "grad_norm": 0.35284191370010376,
      "learning_rate": 2.8335301062573792e-05,
      "loss": 0.7483,
      "step": 733
    },
    {
      "epoch": 0.8615023474178404,
      "grad_norm": 0.3469059467315674,
      "learning_rate": 2.809917355371901e-05,
      "loss": 0.6902,
      "step": 734
    },
    {
      "epoch": 0.8626760563380281,
      "grad_norm": 0.35148003697395325,
      "learning_rate": 2.7863046044864228e-05,
      "loss": 0.732,
      "step": 735
    },
    {
      "epoch": 0.863849765258216,
      "grad_norm": 0.3533206880092621,
      "learning_rate": 2.7626918536009443e-05,
      "loss": 0.7287,
      "step": 736
    },
    {
      "epoch": 0.8650234741784038,
      "grad_norm": 0.383095383644104,
      "learning_rate": 2.7390791027154668e-05,
      "loss": 0.8017,
      "step": 737
    },
    {
      "epoch": 0.8661971830985915,
      "grad_norm": 0.3541397452354431,
      "learning_rate": 2.7154663518299882e-05,
      "loss": 0.7291,
      "step": 738
    },
    {
      "epoch": 0.8673708920187794,
      "grad_norm": 0.35989582538604736,
      "learning_rate": 2.69185360094451e-05,
      "loss": 0.7211,
      "step": 739
    },
    {
      "epoch": 0.8685446009389671,
      "grad_norm": 0.34245404601097107,
      "learning_rate": 2.668240850059032e-05,
      "loss": 0.7062,
      "step": 740
    },
    {
      "epoch": 0.8697183098591549,
      "grad_norm": 0.3396112024784088,
      "learning_rate": 2.644628099173554e-05,
      "loss": 0.6946,
      "step": 741
    },
    {
      "epoch": 0.8708920187793427,
      "grad_norm": 0.34901162981987,
      "learning_rate": 2.6210153482880757e-05,
      "loss": 0.7742,
      "step": 742
    },
    {
      "epoch": 0.8720657276995305,
      "grad_norm": 0.3654363453388214,
      "learning_rate": 2.5974025974025972e-05,
      "loss": 0.7894,
      "step": 743
    },
    {
      "epoch": 0.8732394366197183,
      "grad_norm": 0.3478833734989166,
      "learning_rate": 2.5737898465171197e-05,
      "loss": 0.6909,
      "step": 744
    },
    {
      "epoch": 0.8744131455399061,
      "grad_norm": 0.3447161912918091,
      "learning_rate": 2.550177095631641e-05,
      "loss": 0.7166,
      "step": 745
    },
    {
      "epoch": 0.8755868544600939,
      "grad_norm": 0.35436901450157166,
      "learning_rate": 2.526564344746163e-05,
      "loss": 0.6962,
      "step": 746
    },
    {
      "epoch": 0.8767605633802817,
      "grad_norm": 0.3359661400318146,
      "learning_rate": 2.502951593860685e-05,
      "loss": 0.7345,
      "step": 747
    },
    {
      "epoch": 0.8779342723004695,
      "grad_norm": 0.35876211524009705,
      "learning_rate": 2.479338842975207e-05,
      "loss": 0.6723,
      "step": 748
    },
    {
      "epoch": 0.8791079812206573,
      "grad_norm": 0.35507625341415405,
      "learning_rate": 2.4557260920897283e-05,
      "loss": 0.6744,
      "step": 749
    },
    {
      "epoch": 0.8802816901408451,
      "grad_norm": 0.3504907786846161,
      "learning_rate": 2.4321133412042505e-05,
      "loss": 0.7281,
      "step": 750
    },
    {
      "epoch": 0.8814553990610329,
      "grad_norm": 0.3498130440711975,
      "learning_rate": 2.4085005903187723e-05,
      "loss": 0.7079,
      "step": 751
    },
    {
      "epoch": 0.8826291079812206,
      "grad_norm": 0.36793026328086853,
      "learning_rate": 2.384887839433294e-05,
      "loss": 0.747,
      "step": 752
    },
    {
      "epoch": 0.8838028169014085,
      "grad_norm": 0.3484232723712921,
      "learning_rate": 2.3612750885478162e-05,
      "loss": 0.7347,
      "step": 753
    },
    {
      "epoch": 0.8849765258215962,
      "grad_norm": 0.34402692317962646,
      "learning_rate": 2.3376623376623376e-05,
      "loss": 0.6717,
      "step": 754
    },
    {
      "epoch": 0.886150234741784,
      "grad_norm": 0.377380907535553,
      "learning_rate": 2.3140495867768598e-05,
      "loss": 0.7642,
      "step": 755
    },
    {
      "epoch": 0.8873239436619719,
      "grad_norm": 0.361382395029068,
      "learning_rate": 2.2904368358913812e-05,
      "loss": 0.7081,
      "step": 756
    },
    {
      "epoch": 0.8884976525821596,
      "grad_norm": 0.3643784821033478,
      "learning_rate": 2.2668240850059034e-05,
      "loss": 0.7219,
      "step": 757
    },
    {
      "epoch": 0.8896713615023474,
      "grad_norm": 0.3974801301956177,
      "learning_rate": 2.243211334120425e-05,
      "loss": 0.712,
      "step": 758
    },
    {
      "epoch": 0.8908450704225352,
      "grad_norm": 0.35573598742485046,
      "learning_rate": 2.219598583234947e-05,
      "loss": 0.7335,
      "step": 759
    },
    {
      "epoch": 0.892018779342723,
      "grad_norm": 0.3532857596874237,
      "learning_rate": 2.1959858323494688e-05,
      "loss": 0.7013,
      "step": 760
    },
    {
      "epoch": 0.8931924882629108,
      "grad_norm": 0.33362728357315063,
      "learning_rate": 2.1723730814639906e-05,
      "loss": 0.6739,
      "step": 761
    },
    {
      "epoch": 0.8943661971830986,
      "grad_norm": 0.3325813412666321,
      "learning_rate": 2.1487603305785127e-05,
      "loss": 0.7099,
      "step": 762
    },
    {
      "epoch": 0.8955399061032864,
      "grad_norm": 0.3451225459575653,
      "learning_rate": 2.125147579693034e-05,
      "loss": 0.6959,
      "step": 763
    },
    {
      "epoch": 0.8967136150234741,
      "grad_norm": 0.3604796528816223,
      "learning_rate": 2.1015348288075563e-05,
      "loss": 0.737,
      "step": 764
    },
    {
      "epoch": 0.897887323943662,
      "grad_norm": 0.34980282187461853,
      "learning_rate": 2.077922077922078e-05,
      "loss": 0.7206,
      "step": 765
    },
    {
      "epoch": 0.8990610328638498,
      "grad_norm": 0.35130617022514343,
      "learning_rate": 2.0543093270366e-05,
      "loss": 0.7153,
      "step": 766
    },
    {
      "epoch": 0.9002347417840375,
      "grad_norm": 0.34524810314178467,
      "learning_rate": 2.0306965761511217e-05,
      "loss": 0.7237,
      "step": 767
    },
    {
      "epoch": 0.9014084507042254,
      "grad_norm": 0.35661572217941284,
      "learning_rate": 2.0070838252656435e-05,
      "loss": 0.6831,
      "step": 768
    },
    {
      "epoch": 0.9025821596244131,
      "grad_norm": 0.35206255316734314,
      "learning_rate": 1.9834710743801653e-05,
      "loss": 0.7721,
      "step": 769
    },
    {
      "epoch": 0.903755868544601,
      "grad_norm": 0.35439351201057434,
      "learning_rate": 1.959858323494687e-05,
      "loss": 0.7142,
      "step": 770
    },
    {
      "epoch": 0.9049295774647887,
      "grad_norm": 0.33722493052482605,
      "learning_rate": 1.9362455726092092e-05,
      "loss": 0.6968,
      "step": 771
    },
    {
      "epoch": 0.9061032863849765,
      "grad_norm": 0.3573172092437744,
      "learning_rate": 1.912632821723731e-05,
      "loss": 0.7301,
      "step": 772
    },
    {
      "epoch": 0.9072769953051644,
      "grad_norm": 0.3347008526325226,
      "learning_rate": 1.8890200708382528e-05,
      "loss": 0.6721,
      "step": 773
    },
    {
      "epoch": 0.9084507042253521,
      "grad_norm": 0.3563063144683838,
      "learning_rate": 1.8654073199527746e-05,
      "loss": 0.7233,
      "step": 774
    },
    {
      "epoch": 0.9096244131455399,
      "grad_norm": 0.35159915685653687,
      "learning_rate": 1.8417945690672964e-05,
      "loss": 0.7184,
      "step": 775
    },
    {
      "epoch": 0.9107981220657277,
      "grad_norm": 0.35826948285102844,
      "learning_rate": 1.8181818181818182e-05,
      "loss": 0.7301,
      "step": 776
    },
    {
      "epoch": 0.9119718309859155,
      "grad_norm": 0.3533133268356323,
      "learning_rate": 1.79456906729634e-05,
      "loss": 0.7373,
      "step": 777
    },
    {
      "epoch": 0.9131455399061033,
      "grad_norm": 0.3495820164680481,
      "learning_rate": 1.7709563164108618e-05,
      "loss": 0.7379,
      "step": 778
    },
    {
      "epoch": 0.9143192488262911,
      "grad_norm": 0.33082085847854614,
      "learning_rate": 1.747343565525384e-05,
      "loss": 0.6789,
      "step": 779
    },
    {
      "epoch": 0.9154929577464789,
      "grad_norm": 0.34669029712677,
      "learning_rate": 1.7237308146399057e-05,
      "loss": 0.6962,
      "step": 780
    },
    {
      "epoch": 0.9166666666666666,
      "grad_norm": 0.3401969373226166,
      "learning_rate": 1.7001180637544275e-05,
      "loss": 0.717,
      "step": 781
    },
    {
      "epoch": 0.9178403755868545,
      "grad_norm": 0.3488728702068329,
      "learning_rate": 1.6765053128689493e-05,
      "loss": 0.7087,
      "step": 782
    },
    {
      "epoch": 0.9190140845070423,
      "grad_norm": 0.39244547486305237,
      "learning_rate": 1.652892561983471e-05,
      "loss": 0.7331,
      "step": 783
    },
    {
      "epoch": 0.92018779342723,
      "grad_norm": 0.33185505867004395,
      "learning_rate": 1.629279811097993e-05,
      "loss": 0.6821,
      "step": 784
    },
    {
      "epoch": 0.9213615023474179,
      "grad_norm": 0.34186288714408875,
      "learning_rate": 1.6056670602125147e-05,
      "loss": 0.6766,
      "step": 785
    },
    {
      "epoch": 0.9225352112676056,
      "grad_norm": 0.34512627124786377,
      "learning_rate": 1.582054309327037e-05,
      "loss": 0.6837,
      "step": 786
    },
    {
      "epoch": 0.9237089201877934,
      "grad_norm": 0.34042122960090637,
      "learning_rate": 1.5584415584415583e-05,
      "loss": 0.7266,
      "step": 787
    },
    {
      "epoch": 0.9248826291079812,
      "grad_norm": 0.34173402190208435,
      "learning_rate": 1.5348288075560805e-05,
      "loss": 0.6998,
      "step": 788
    },
    {
      "epoch": 0.926056338028169,
      "grad_norm": 0.34008073806762695,
      "learning_rate": 1.511216056670602e-05,
      "loss": 0.7211,
      "step": 789
    },
    {
      "epoch": 0.9272300469483568,
      "grad_norm": 0.3400252163410187,
      "learning_rate": 1.487603305785124e-05,
      "loss": 0.6771,
      "step": 790
    },
    {
      "epoch": 0.9284037558685446,
      "grad_norm": 0.3393029570579529,
      "learning_rate": 1.4639905548996457e-05,
      "loss": 0.7274,
      "step": 791
    },
    {
      "epoch": 0.9295774647887324,
      "grad_norm": 0.3489772379398346,
      "learning_rate": 1.4403778040141676e-05,
      "loss": 0.7195,
      "step": 792
    },
    {
      "epoch": 0.9307511737089202,
      "grad_norm": 0.3434072732925415,
      "learning_rate": 1.4167650531286896e-05,
      "loss": 0.6806,
      "step": 793
    },
    {
      "epoch": 0.931924882629108,
      "grad_norm": 0.35593146085739136,
      "learning_rate": 1.3931523022432114e-05,
      "loss": 0.7026,
      "step": 794
    },
    {
      "epoch": 0.9330985915492958,
      "grad_norm": 0.33654287457466125,
      "learning_rate": 1.3695395513577334e-05,
      "loss": 0.6655,
      "step": 795
    },
    {
      "epoch": 0.9342723004694836,
      "grad_norm": 0.35049983859062195,
      "learning_rate": 1.345926800472255e-05,
      "loss": 0.686,
      "step": 796
    },
    {
      "epoch": 0.9354460093896714,
      "grad_norm": 0.3442087471485138,
      "learning_rate": 1.322314049586777e-05,
      "loss": 0.7048,
      "step": 797
    },
    {
      "epoch": 0.9366197183098591,
      "grad_norm": 0.3569439649581909,
      "learning_rate": 1.2987012987012986e-05,
      "loss": 0.7271,
      "step": 798
    },
    {
      "epoch": 0.937793427230047,
      "grad_norm": 0.3418942391872406,
      "learning_rate": 1.2750885478158206e-05,
      "loss": 0.7132,
      "step": 799
    },
    {
      "epoch": 0.9389671361502347,
      "grad_norm": 0.3399513363838196,
      "learning_rate": 1.2514757969303425e-05,
      "loss": 0.7046,
      "step": 800
    },
    {
      "epoch": 0.9401408450704225,
      "grad_norm": 0.34055379033088684,
      "learning_rate": 0.00010641553855208948,
      "loss": 0.7293,
      "step": 801
    },
    {
      "epoch": 0.9413145539906104,
      "grad_norm": 0.3299119770526886,
      "learning_rate": 0.0001062978222483814,
      "loss": 0.6779,
      "step": 802
    },
    {
      "epoch": 0.9424882629107981,
      "grad_norm": 0.3833242356777191,
      "learning_rate": 0.00010618010594467334,
      "loss": 0.6909,
      "step": 803
    },
    {
      "epoch": 0.9436619718309859,
      "grad_norm": 0.39958855509757996,
      "learning_rate": 0.00010606238964096529,
      "loss": 0.7307,
      "step": 804
    },
    {
      "epoch": 0.9448356807511737,
      "grad_norm": 0.38618725538253784,
      "learning_rate": 0.00010594467333725722,
      "loss": 0.6984,
      "step": 805
    },
    {
      "epoch": 0.9460093896713615,
      "grad_norm": 0.4084942936897278,
      "learning_rate": 0.00010582695703354914,
      "loss": 0.7456,
      "step": 806
    },
    {
      "epoch": 0.9471830985915493,
      "grad_norm": 0.4109421372413635,
      "learning_rate": 0.00010570924072984109,
      "loss": 0.6991,
      "step": 807
    },
    {
      "epoch": 0.9483568075117371,
      "grad_norm": 0.382415771484375,
      "learning_rate": 0.00010559152442613303,
      "loss": 0.726,
      "step": 808
    },
    {
      "epoch": 0.9495305164319249,
      "grad_norm": 0.4036392867565155,
      "learning_rate": 0.00010547380812242496,
      "loss": 0.7264,
      "step": 809
    },
    {
      "epoch": 0.9507042253521126,
      "grad_norm": 0.38903331756591797,
      "learning_rate": 0.00010535609181871691,
      "loss": 0.691,
      "step": 810
    },
    {
      "epoch": 0.9518779342723005,
      "grad_norm": 0.3803318440914154,
      "learning_rate": 0.00010523837551500883,
      "loss": 0.7271,
      "step": 811
    },
    {
      "epoch": 0.9530516431924883,
      "grad_norm": 0.3850460350513458,
      "learning_rate": 0.00010512065921130076,
      "loss": 0.7111,
      "step": 812
    },
    {
      "epoch": 0.954225352112676,
      "grad_norm": 0.4110994040966034,
      "learning_rate": 0.00010500294290759271,
      "loss": 0.7282,
      "step": 813
    },
    {
      "epoch": 0.9553990610328639,
      "grad_norm": 0.3853722810745239,
      "learning_rate": 0.00010488522660388465,
      "loss": 0.7194,
      "step": 814
    },
    {
      "epoch": 0.9565727699530516,
      "grad_norm": 0.37440797686576843,
      "learning_rate": 0.00010476751030017658,
      "loss": 0.7116,
      "step": 815
    },
    {
      "epoch": 0.9577464788732394,
      "grad_norm": 0.42637899518013,
      "learning_rate": 0.00010464979399646853,
      "loss": 0.7189,
      "step": 816
    },
    {
      "epoch": 0.9589201877934272,
      "grad_norm": 0.4067356288433075,
      "learning_rate": 0.00010453207769276045,
      "loss": 0.7509,
      "step": 817
    },
    {
      "epoch": 0.960093896713615,
      "grad_norm": 0.3854503929615021,
      "learning_rate": 0.00010441436138905238,
      "loss": 0.7426,
      "step": 818
    },
    {
      "epoch": 0.9612676056338029,
      "grad_norm": 0.4298991858959198,
      "learning_rate": 0.00010429664508534433,
      "loss": 0.7528,
      "step": 819
    },
    {
      "epoch": 0.9624413145539906,
      "grad_norm": 0.3748774826526642,
      "learning_rate": 0.00010417892878163627,
      "loss": 0.6512,
      "step": 820
    },
    {
      "epoch": 0.9636150234741784,
      "grad_norm": 0.38448989391326904,
      "learning_rate": 0.00010406121247792819,
      "loss": 0.6929,
      "step": 821
    },
    {
      "epoch": 0.9647887323943662,
      "grad_norm": 0.42416030168533325,
      "learning_rate": 0.00010394349617422015,
      "loss": 0.7312,
      "step": 822
    },
    {
      "epoch": 0.965962441314554,
      "grad_norm": 0.3875625729560852,
      "learning_rate": 0.00010382577987051207,
      "loss": 0.7121,
      "step": 823
    },
    {
      "epoch": 0.9671361502347418,
      "grad_norm": 0.4241638481616974,
      "learning_rate": 0.000103708063566804,
      "loss": 0.7248,
      "step": 824
    },
    {
      "epoch": 0.9683098591549296,
      "grad_norm": 0.4026165306568146,
      "learning_rate": 0.00010359034726309595,
      "loss": 0.7224,
      "step": 825
    },
    {
      "epoch": 0.9694835680751174,
      "grad_norm": 0.39895206689834595,
      "learning_rate": 0.00010347263095938789,
      "loss": 0.7193,
      "step": 826
    },
    {
      "epoch": 0.9706572769953051,
      "grad_norm": 0.395463228225708,
      "learning_rate": 0.00010335491465567981,
      "loss": 0.7673,
      "step": 827
    },
    {
      "epoch": 0.971830985915493,
      "grad_norm": 0.4351494312286377,
      "learning_rate": 0.00010323719835197174,
      "loss": 0.7684,
      "step": 828
    },
    {
      "epoch": 0.9730046948356808,
      "grad_norm": 0.4378681182861328,
      "learning_rate": 0.00010311948204826369,
      "loss": 0.7277,
      "step": 829
    },
    {
      "epoch": 0.9741784037558685,
      "grad_norm": 0.4214630722999573,
      "learning_rate": 0.00010300176574455563,
      "loss": 0.7107,
      "step": 830
    },
    {
      "epoch": 0.9753521126760564,
      "grad_norm": 0.41999107599258423,
      "learning_rate": 0.00010288404944084755,
      "loss": 0.7328,
      "step": 831
    },
    {
      "epoch": 0.9765258215962441,
      "grad_norm": 0.49026909470558167,
      "learning_rate": 0.00010276633313713951,
      "loss": 0.7345,
      "step": 832
    },
    {
      "epoch": 0.9776995305164319,
      "grad_norm": 0.4068211317062378,
      "learning_rate": 0.00010264861683343143,
      "loss": 0.701,
      "step": 833
    },
    {
      "epoch": 0.9788732394366197,
      "grad_norm": 0.42514288425445557,
      "learning_rate": 0.00010253090052972336,
      "loss": 0.729,
      "step": 834
    },
    {
      "epoch": 0.9800469483568075,
      "grad_norm": 0.4883005619049072,
      "learning_rate": 0.00010241318422601531,
      "loss": 0.7183,
      "step": 835
    },
    {
      "epoch": 0.9812206572769953,
      "grad_norm": 0.38146787881851196,
      "learning_rate": 0.00010229546792230725,
      "loss": 0.6977,
      "step": 836
    },
    {
      "epoch": 0.9823943661971831,
      "grad_norm": 0.3898909389972687,
      "learning_rate": 0.00010217775161859917,
      "loss": 0.7131,
      "step": 837
    },
    {
      "epoch": 0.9835680751173709,
      "grad_norm": 0.39693424105644226,
      "learning_rate": 0.00010206003531489112,
      "loss": 0.7184,
      "step": 838
    },
    {
      "epoch": 0.9847417840375586,
      "grad_norm": 0.3968975841999054,
      "learning_rate": 0.00010194231901118305,
      "loss": 0.7536,
      "step": 839
    },
    {
      "epoch": 0.9859154929577465,
      "grad_norm": 0.4030087888240814,
      "learning_rate": 0.00010182460270747499,
      "loss": 0.7156,
      "step": 840
    },
    {
      "epoch": 0.9870892018779343,
      "grad_norm": 0.37477344274520874,
      "learning_rate": 0.00010170688640376693,
      "loss": 0.6815,
      "step": 841
    },
    {
      "epoch": 0.9882629107981221,
      "grad_norm": 0.40929409861564636,
      "learning_rate": 0.00010158917010005887,
      "loss": 0.6827,
      "step": 842
    },
    {
      "epoch": 0.9894366197183099,
      "grad_norm": 0.36350882053375244,
      "learning_rate": 0.00010147145379635079,
      "loss": 0.6927,
      "step": 843
    },
    {
      "epoch": 0.9906103286384976,
      "grad_norm": 0.3828059434890747,
      "learning_rate": 0.00010135373749264274,
      "loss": 0.7254,
      "step": 844
    },
    {
      "epoch": 0.9917840375586855,
      "grad_norm": 0.4095743000507355,
      "learning_rate": 0.00010123602118893467,
      "loss": 0.719,
      "step": 845
    },
    {
      "epoch": 0.9929577464788732,
      "grad_norm": 0.37418296933174133,
      "learning_rate": 0.0001011183048852266,
      "loss": 0.682,
      "step": 846
    },
    {
      "epoch": 0.994131455399061,
      "grad_norm": 0.39427751302719116,
      "learning_rate": 0.00010100058858151855,
      "loss": 0.7742,
      "step": 847
    },
    {
      "epoch": 0.9953051643192489,
      "grad_norm": 0.3696395754814148,
      "learning_rate": 0.00010088287227781048,
      "loss": 0.7377,
      "step": 848
    },
    {
      "epoch": 0.9964788732394366,
      "grad_norm": 0.36249879002571106,
      "learning_rate": 0.00010076515597410241,
      "loss": 0.7237,
      "step": 849
    },
    {
      "epoch": 0.9976525821596244,
      "grad_norm": 0.3712272047996521,
      "learning_rate": 0.00010064743967039436,
      "loss": 0.6737,
      "step": 850
    },
    {
      "epoch": 0.9988262910798122,
      "grad_norm": 0.37550613284111023,
      "learning_rate": 0.00010052972336668629,
      "loss": 0.7147,
      "step": 851
    },
    {
      "epoch": 1.0,
      "grad_norm": 0.405351459980011,
      "learning_rate": 0.00010041200706297821,
      "loss": 0.7364,
      "step": 852
    },
    {
      "epoch": 1.0011737089201878,
      "grad_norm": 0.39747750759124756,
      "learning_rate": 0.00010029429075927018,
      "loss": 0.6934,
      "step": 853
    },
    {
      "epoch": 1.0023474178403755,
      "grad_norm": 0.3695623576641083,
      "learning_rate": 0.0001001765744555621,
      "loss": 0.6971,
      "step": 854
    },
    {
      "epoch": 1.0035211267605635,
      "grad_norm": 0.3880208134651184,
      "learning_rate": 0.00010005885815185403,
      "loss": 0.7219,
      "step": 855
    },
    {
      "epoch": 1.0046948356807512,
      "grad_norm": 0.40131011605262756,
      "learning_rate": 9.994114184814597e-05,
      "loss": 0.6925,
      "step": 856
    },
    {
      "epoch": 1.005868544600939,
      "grad_norm": 0.38630256056785583,
      "learning_rate": 9.982342554443791e-05,
      "loss": 0.7412,
      "step": 857
    },
    {
      "epoch": 1.0070422535211268,
      "grad_norm": 0.39141979813575745,
      "learning_rate": 9.970570924072985e-05,
      "loss": 0.7089,
      "step": 858
    },
    {
      "epoch": 1.0082159624413145,
      "grad_norm": 0.3811167776584625,
      "learning_rate": 9.958799293702178e-05,
      "loss": 0.6979,
      "step": 859
    },
    {
      "epoch": 1.0093896713615023,
      "grad_norm": 0.38177528977394104,
      "learning_rate": 9.947027663331372e-05,
      "loss": 0.7181,
      "step": 860
    },
    {
      "epoch": 1.0105633802816902,
      "grad_norm": 0.36225804686546326,
      "learning_rate": 9.935256032960567e-05,
      "loss": 0.6495,
      "step": 861
    },
    {
      "epoch": 1.011737089201878,
      "grad_norm": 0.3796376585960388,
      "learning_rate": 9.923484402589759e-05,
      "loss": 0.6661,
      "step": 862
    },
    {
      "epoch": 1.0129107981220657,
      "grad_norm": 0.3896029591560364,
      "learning_rate": 9.911712772218953e-05,
      "loss": 0.6705,
      "step": 863
    },
    {
      "epoch": 1.0140845070422535,
      "grad_norm": 0.35688912868499756,
      "learning_rate": 9.899941141848147e-05,
      "loss": 0.6835,
      "step": 864
    },
    {
      "epoch": 1.0152582159624413,
      "grad_norm": 0.3919657766819,
      "learning_rate": 9.88816951147734e-05,
      "loss": 0.6771,
      "step": 865
    },
    {
      "epoch": 1.016431924882629,
      "grad_norm": 0.390311062335968,
      "learning_rate": 9.876397881106534e-05,
      "loss": 0.7208,
      "step": 866
    },
    {
      "epoch": 1.017605633802817,
      "grad_norm": 0.3857402205467224,
      "learning_rate": 9.864626250735727e-05,
      "loss": 0.7321,
      "step": 867
    },
    {
      "epoch": 1.0187793427230047,
      "grad_norm": 0.3688738942146301,
      "learning_rate": 9.852854620364921e-05,
      "loss": 0.6853,
      "step": 868
    },
    {
      "epoch": 1.0199530516431925,
      "grad_norm": 0.3814820647239685,
      "learning_rate": 9.841082989994114e-05,
      "loss": 0.664,
      "step": 869
    },
    {
      "epoch": 1.0211267605633803,
      "grad_norm": 0.3849344253540039,
      "learning_rate": 9.829311359623309e-05,
      "loss": 0.6844,
      "step": 870
    },
    {
      "epoch": 1.022300469483568,
      "grad_norm": 0.36203038692474365,
      "learning_rate": 9.817539729252502e-05,
      "loss": 0.7201,
      "step": 871
    },
    {
      "epoch": 1.0234741784037558,
      "grad_norm": 0.36614471673965454,
      "learning_rate": 9.805768098881696e-05,
      "loss": 0.659,
      "step": 872
    },
    {
      "epoch": 1.0246478873239437,
      "grad_norm": 0.3908173143863678,
      "learning_rate": 9.79399646851089e-05,
      "loss": 0.6638,
      "step": 873
    },
    {
      "epoch": 1.0258215962441315,
      "grad_norm": 0.35966452956199646,
      "learning_rate": 9.782224838140083e-05,
      "loss": 0.7187,
      "step": 874
    },
    {
      "epoch": 1.0269953051643192,
      "grad_norm": 0.40878093242645264,
      "learning_rate": 9.770453207769276e-05,
      "loss": 0.691,
      "step": 875
    },
    {
      "epoch": 1.028169014084507,
      "grad_norm": 0.38903382420539856,
      "learning_rate": 9.75868157739847e-05,
      "loss": 0.718,
      "step": 876
    },
    {
      "epoch": 1.0293427230046948,
      "grad_norm": 0.3865324556827545,
      "learning_rate": 9.746909947027663e-05,
      "loss": 0.7331,
      "step": 877
    },
    {
      "epoch": 1.0305164319248827,
      "grad_norm": 0.37417513132095337,
      "learning_rate": 9.735138316656858e-05,
      "loss": 0.677,
      "step": 878
    },
    {
      "epoch": 1.0316901408450705,
      "grad_norm": 0.38043439388275146,
      "learning_rate": 9.72336668628605e-05,
      "loss": 0.6932,
      "step": 879
    },
    {
      "epoch": 1.0328638497652582,
      "grad_norm": 0.37418729066848755,
      "learning_rate": 9.711595055915245e-05,
      "loss": 0.7119,
      "step": 880
    },
    {
      "epoch": 1.034037558685446,
      "grad_norm": 0.4013047218322754,
      "learning_rate": 9.699823425544438e-05,
      "loss": 0.7041,
      "step": 881
    },
    {
      "epoch": 1.0352112676056338,
      "grad_norm": 0.38462570309638977,
      "learning_rate": 9.688051795173632e-05,
      "loss": 0.6861,
      "step": 882
    },
    {
      "epoch": 1.0363849765258215,
      "grad_norm": 0.3900148868560791,
      "learning_rate": 9.676280164802825e-05,
      "loss": 0.6382,
      "step": 883
    },
    {
      "epoch": 1.0375586854460095,
      "grad_norm": 0.3882652819156647,
      "learning_rate": 9.66450853443202e-05,
      "loss": 0.6948,
      "step": 884
    },
    {
      "epoch": 1.0387323943661972,
      "grad_norm": 0.36546608805656433,
      "learning_rate": 9.652736904061212e-05,
      "loss": 0.7064,
      "step": 885
    },
    {
      "epoch": 1.039906103286385,
      "grad_norm": 0.3788559138774872,
      "learning_rate": 9.640965273690407e-05,
      "loss": 0.7129,
      "step": 886
    },
    {
      "epoch": 1.0410798122065728,
      "grad_norm": 0.3979467451572418,
      "learning_rate": 9.6291936433196e-05,
      "loss": 0.7196,
      "step": 887
    },
    {
      "epoch": 1.0422535211267605,
      "grad_norm": 0.3777488172054291,
      "learning_rate": 9.617422012948794e-05,
      "loss": 0.6922,
      "step": 888
    },
    {
      "epoch": 1.0434272300469483,
      "grad_norm": 0.39730504155158997,
      "learning_rate": 9.605650382577987e-05,
      "loss": 0.6529,
      "step": 889
    },
    {
      "epoch": 1.0446009389671362,
      "grad_norm": 0.39619576930999756,
      "learning_rate": 9.593878752207182e-05,
      "loss": 0.6505,
      "step": 890
    },
    {
      "epoch": 1.045774647887324,
      "grad_norm": 0.3763888478279114,
      "learning_rate": 9.582107121836374e-05,
      "loss": 0.638,
      "step": 891
    },
    {
      "epoch": 1.0469483568075117,
      "grad_norm": 0.3947450518608093,
      "learning_rate": 9.570335491465569e-05,
      "loss": 0.7099,
      "step": 892
    },
    {
      "epoch": 1.0481220657276995,
      "grad_norm": 0.43239885568618774,
      "learning_rate": 9.558563861094763e-05,
      "loss": 0.7112,
      "step": 893
    },
    {
      "epoch": 1.0492957746478873,
      "grad_norm": 0.37725165486335754,
      "learning_rate": 9.546792230723956e-05,
      "loss": 0.6775,
      "step": 894
    },
    {
      "epoch": 1.050469483568075,
      "grad_norm": 0.3807140290737152,
      "learning_rate": 9.53502060035315e-05,
      "loss": 0.7201,
      "step": 895
    },
    {
      "epoch": 1.051643192488263,
      "grad_norm": 0.40270236134529114,
      "learning_rate": 9.523248969982343e-05,
      "loss": 0.6908,
      "step": 896
    },
    {
      "epoch": 1.0528169014084507,
      "grad_norm": 0.38907137513160706,
      "learning_rate": 9.511477339611536e-05,
      "loss": 0.7274,
      "step": 897
    },
    {
      "epoch": 1.0539906103286385,
      "grad_norm": 0.35074397921562195,
      "learning_rate": 9.49970570924073e-05,
      "loss": 0.6765,
      "step": 898
    },
    {
      "epoch": 1.0551643192488263,
      "grad_norm": 0.37548649311065674,
      "learning_rate": 9.487934078869925e-05,
      "loss": 0.7258,
      "step": 899
    },
    {
      "epoch": 1.056338028169014,
      "grad_norm": 0.3947518467903137,
      "learning_rate": 9.476162448499118e-05,
      "loss": 0.7142,
      "step": 900
    },
    {
      "epoch": 1.057511737089202,
      "grad_norm": 0.36888387799263,
      "learning_rate": 9.464390818128312e-05,
      "loss": 0.664,
      "step": 901
    },
    {
      "epoch": 1.0586854460093897,
      "grad_norm": 0.3735831379890442,
      "learning_rate": 9.452619187757505e-05,
      "loss": 0.6914,
      "step": 902
    },
    {
      "epoch": 1.0598591549295775,
      "grad_norm": 0.3840358257293701,
      "learning_rate": 9.440847557386698e-05,
      "loss": 0.663,
      "step": 903
    },
    {
      "epoch": 1.0610328638497653,
      "grad_norm": 0.408840537071228,
      "learning_rate": 9.429075927015892e-05,
      "loss": 0.7225,
      "step": 904
    },
    {
      "epoch": 1.062206572769953,
      "grad_norm": 0.36408165097236633,
      "learning_rate": 9.417304296645085e-05,
      "loss": 0.6744,
      "step": 905
    },
    {
      "epoch": 1.0633802816901408,
      "grad_norm": 0.4005196690559387,
      "learning_rate": 9.405532666274279e-05,
      "loss": 0.7285,
      "step": 906
    },
    {
      "epoch": 1.0645539906103287,
      "grad_norm": 0.3824830949306488,
      "learning_rate": 9.393761035903474e-05,
      "loss": 0.6978,
      "step": 907
    },
    {
      "epoch": 1.0657276995305165,
      "grad_norm": 0.38410818576812744,
      "learning_rate": 9.381989405532666e-05,
      "loss": 0.6725,
      "step": 908
    },
    {
      "epoch": 1.0669014084507042,
      "grad_norm": 0.37026217579841614,
      "learning_rate": 9.37021777516186e-05,
      "loss": 0.6908,
      "step": 909
    },
    {
      "epoch": 1.068075117370892,
      "grad_norm": 0.37652963399887085,
      "learning_rate": 9.358446144791054e-05,
      "loss": 0.6674,
      "step": 910
    },
    {
      "epoch": 1.0692488262910798,
      "grad_norm": 0.40584585070610046,
      "learning_rate": 9.346674514420247e-05,
      "loss": 0.7087,
      "step": 911
    },
    {
      "epoch": 1.0704225352112675,
      "grad_norm": 0.3777616620063782,
      "learning_rate": 9.334902884049441e-05,
      "loss": 0.6633,
      "step": 912
    },
    {
      "epoch": 1.0715962441314555,
      "grad_norm": 0.35584181547164917,
      "learning_rate": 9.323131253678636e-05,
      "loss": 0.6484,
      "step": 913
    },
    {
      "epoch": 1.0727699530516432,
      "grad_norm": 0.40920573472976685,
      "learning_rate": 9.311359623307828e-05,
      "loss": 0.6781,
      "step": 914
    },
    {
      "epoch": 1.073943661971831,
      "grad_norm": 0.37617766857147217,
      "learning_rate": 9.299587992937023e-05,
      "loss": 0.6785,
      "step": 915
    },
    {
      "epoch": 1.0751173708920188,
      "grad_norm": 0.36854755878448486,
      "learning_rate": 9.287816362566216e-05,
      "loss": 0.6805,
      "step": 916
    },
    {
      "epoch": 1.0762910798122065,
      "grad_norm": 0.3820021152496338,
      "learning_rate": 9.27604473219541e-05,
      "loss": 0.7413,
      "step": 917
    },
    {
      "epoch": 1.0774647887323943,
      "grad_norm": 0.3654205799102783,
      "learning_rate": 9.264273101824603e-05,
      "loss": 0.6996,
      "step": 918
    },
    {
      "epoch": 1.0786384976525822,
      "grad_norm": 0.36847448348999023,
      "learning_rate": 9.252501471453798e-05,
      "loss": 0.6593,
      "step": 919
    },
    {
      "epoch": 1.07981220657277,
      "grad_norm": 0.4072454571723938,
      "learning_rate": 9.24072984108299e-05,
      "loss": 0.7062,
      "step": 920
    },
    {
      "epoch": 1.0809859154929577,
      "grad_norm": 0.37201663851737976,
      "learning_rate": 9.228958210712185e-05,
      "loss": 0.7188,
      "step": 921
    },
    {
      "epoch": 1.0821596244131455,
      "grad_norm": 0.40708494186401367,
      "learning_rate": 9.217186580341378e-05,
      "loss": 0.6984,
      "step": 922
    },
    {
      "epoch": 1.0833333333333333,
      "grad_norm": 0.37668758630752563,
      "learning_rate": 9.205414949970572e-05,
      "loss": 0.6856,
      "step": 923
    },
    {
      "epoch": 1.084507042253521,
      "grad_norm": 0.41518712043762207,
      "learning_rate": 9.193643319599765e-05,
      "loss": 0.7093,
      "step": 924
    },
    {
      "epoch": 1.085680751173709,
      "grad_norm": 0.3661474883556366,
      "learning_rate": 9.181871689228958e-05,
      "loss": 0.6765,
      "step": 925
    },
    {
      "epoch": 1.0868544600938967,
      "grad_norm": 0.3910673260688782,
      "learning_rate": 9.170100058858152e-05,
      "loss": 0.6778,
      "step": 926
    },
    {
      "epoch": 1.0880281690140845,
      "grad_norm": 0.3851100206375122,
      "learning_rate": 9.158328428487345e-05,
      "loss": 0.7188,
      "step": 927
    },
    {
      "epoch": 1.0892018779342723,
      "grad_norm": 0.36254799365997314,
      "learning_rate": 9.14655679811654e-05,
      "loss": 0.7182,
      "step": 928
    },
    {
      "epoch": 1.09037558685446,
      "grad_norm": 0.39364567399024963,
      "learning_rate": 9.134785167745734e-05,
      "loss": 0.7208,
      "step": 929
    },
    {
      "epoch": 1.091549295774648,
      "grad_norm": 0.3755466639995575,
      "learning_rate": 9.123013537374927e-05,
      "loss": 0.6771,
      "step": 930
    },
    {
      "epoch": 1.0927230046948357,
      "grad_norm": 0.361087828874588,
      "learning_rate": 9.11124190700412e-05,
      "loss": 0.6541,
      "step": 931
    },
    {
      "epoch": 1.0938967136150235,
      "grad_norm": 0.37327754497528076,
      "learning_rate": 9.099470276633314e-05,
      "loss": 0.698,
      "step": 932
    },
    {
      "epoch": 1.0950704225352113,
      "grad_norm": 0.38413748145103455,
      "learning_rate": 9.087698646262507e-05,
      "loss": 0.6933,
      "step": 933
    },
    {
      "epoch": 1.096244131455399,
      "grad_norm": 0.4182147681713104,
      "learning_rate": 9.075927015891701e-05,
      "loss": 0.6776,
      "step": 934
    },
    {
      "epoch": 1.0974178403755868,
      "grad_norm": 0.3987724483013153,
      "learning_rate": 9.064155385520894e-05,
      "loss": 0.694,
      "step": 935
    },
    {
      "epoch": 1.0985915492957747,
      "grad_norm": 0.37629225850105286,
      "learning_rate": 9.052383755150089e-05,
      "loss": 0.6565,
      "step": 936
    },
    {
      "epoch": 1.0997652582159625,
      "grad_norm": 0.38973352313041687,
      "learning_rate": 9.040612124779281e-05,
      "loss": 0.6739,
      "step": 937
    },
    {
      "epoch": 1.1009389671361502,
      "grad_norm": 0.3845914900302887,
      "learning_rate": 9.028840494408476e-05,
      "loss": 0.6788,
      "step": 938
    },
    {
      "epoch": 1.102112676056338,
      "grad_norm": 0.3861023485660553,
      "learning_rate": 9.01706886403767e-05,
      "loss": 0.6763,
      "step": 939
    },
    {
      "epoch": 1.1032863849765258,
      "grad_norm": 0.37565183639526367,
      "learning_rate": 9.005297233666863e-05,
      "loss": 0.6478,
      "step": 940
    },
    {
      "epoch": 1.1044600938967135,
      "grad_norm": 0.4068315029144287,
      "learning_rate": 8.993525603296056e-05,
      "loss": 0.6752,
      "step": 941
    },
    {
      "epoch": 1.1056338028169015,
      "grad_norm": 0.37796974182128906,
      "learning_rate": 8.981753972925251e-05,
      "loss": 0.7355,
      "step": 942
    },
    {
      "epoch": 1.1068075117370892,
      "grad_norm": 0.4024117887020111,
      "learning_rate": 8.969982342554443e-05,
      "loss": 0.6648,
      "step": 943
    },
    {
      "epoch": 1.107981220657277,
      "grad_norm": 0.404442697763443,
      "learning_rate": 8.958210712183638e-05,
      "loss": 0.7,
      "step": 944
    },
    {
      "epoch": 1.1091549295774648,
      "grad_norm": 0.35948899388313293,
      "learning_rate": 8.946439081812832e-05,
      "loss": 0.6859,
      "step": 945
    },
    {
      "epoch": 1.1103286384976525,
      "grad_norm": 0.4014012813568115,
      "learning_rate": 8.934667451442025e-05,
      "loss": 0.7294,
      "step": 946
    },
    {
      "epoch": 1.1115023474178405,
      "grad_norm": 0.38261109590530396,
      "learning_rate": 8.922895821071219e-05,
      "loss": 0.6965,
      "step": 947
    },
    {
      "epoch": 1.1126760563380282,
      "grad_norm": 0.39297208189964294,
      "learning_rate": 8.911124190700413e-05,
      "loss": 0.7153,
      "step": 948
    },
    {
      "epoch": 1.113849765258216,
      "grad_norm": 0.3710176348686218,
      "learning_rate": 8.899352560329605e-05,
      "loss": 0.7085,
      "step": 949
    },
    {
      "epoch": 1.1150234741784038,
      "grad_norm": 0.3750080168247223,
      "learning_rate": 8.8875809299588e-05,
      "loss": 0.6739,
      "step": 950
    },
    {
      "epoch": 1.1161971830985915,
      "grad_norm": 0.3672105073928833,
      "learning_rate": 8.875809299587994e-05,
      "loss": 0.7097,
      "step": 951
    },
    {
      "epoch": 1.1173708920187793,
      "grad_norm": 0.3663265109062195,
      "learning_rate": 8.864037669217187e-05,
      "loss": 0.6594,
      "step": 952
    },
    {
      "epoch": 1.1185446009389672,
      "grad_norm": 0.4023442268371582,
      "learning_rate": 8.85226603884638e-05,
      "loss": 0.7186,
      "step": 953
    },
    {
      "epoch": 1.119718309859155,
      "grad_norm": 0.36602139472961426,
      "learning_rate": 8.840494408475574e-05,
      "loss": 0.67,
      "step": 954
    },
    {
      "epoch": 1.1208920187793427,
      "grad_norm": 0.36866381764411926,
      "learning_rate": 8.828722778104768e-05,
      "loss": 0.6954,
      "step": 955
    },
    {
      "epoch": 1.1220657276995305,
      "grad_norm": 0.38905832171440125,
      "learning_rate": 8.816951147733961e-05,
      "loss": 0.7214,
      "step": 956
    },
    {
      "epoch": 1.1232394366197183,
      "grad_norm": 0.3806670010089874,
      "learning_rate": 8.805179517363156e-05,
      "loss": 0.6679,
      "step": 957
    },
    {
      "epoch": 1.124413145539906,
      "grad_norm": 0.3796343505382538,
      "learning_rate": 8.793407886992349e-05,
      "loss": 0.6334,
      "step": 958
    },
    {
      "epoch": 1.125586854460094,
      "grad_norm": 0.4143288731575012,
      "learning_rate": 8.781636256621543e-05,
      "loss": 0.7484,
      "step": 959
    },
    {
      "epoch": 1.1267605633802817,
      "grad_norm": 0.3692832887172699,
      "learning_rate": 8.769864626250736e-05,
      "loss": 0.6581,
      "step": 960
    },
    {
      "epoch": 1.1279342723004695,
      "grad_norm": 0.39971667528152466,
      "learning_rate": 8.75809299587993e-05,
      "loss": 0.7252,
      "step": 961
    },
    {
      "epoch": 1.1291079812206573,
      "grad_norm": 0.391924113035202,
      "learning_rate": 8.746321365509123e-05,
      "loss": 0.673,
      "step": 962
    },
    {
      "epoch": 1.130281690140845,
      "grad_norm": 0.39626866579055786,
      "learning_rate": 8.734549735138317e-05,
      "loss": 0.7161,
      "step": 963
    },
    {
      "epoch": 1.131455399061033,
      "grad_norm": 0.3812800347805023,
      "learning_rate": 8.72277810476751e-05,
      "loss": 0.6735,
      "step": 964
    },
    {
      "epoch": 1.1326291079812207,
      "grad_norm": 0.36054447293281555,
      "learning_rate": 8.711006474396705e-05,
      "loss": 0.6861,
      "step": 965
    },
    {
      "epoch": 1.1338028169014085,
      "grad_norm": 0.41179588437080383,
      "learning_rate": 8.699234844025897e-05,
      "loss": 0.7151,
      "step": 966
    },
    {
      "epoch": 1.1349765258215962,
      "grad_norm": 0.3688051998615265,
      "learning_rate": 8.687463213655092e-05,
      "loss": 0.6608,
      "step": 967
    },
    {
      "epoch": 1.136150234741784,
      "grad_norm": 0.3877013325691223,
      "learning_rate": 8.675691583284285e-05,
      "loss": 0.6826,
      "step": 968
    },
    {
      "epoch": 1.1373239436619718,
      "grad_norm": 0.38986387848854065,
      "learning_rate": 8.663919952913479e-05,
      "loss": 0.6915,
      "step": 969
    },
    {
      "epoch": 1.1384976525821595,
      "grad_norm": 0.41986656188964844,
      "learning_rate": 8.652148322542672e-05,
      "loss": 0.7471,
      "step": 970
    },
    {
      "epoch": 1.1396713615023475,
      "grad_norm": 0.3977747857570648,
      "learning_rate": 8.640376692171867e-05,
      "loss": 0.6844,
      "step": 971
    },
    {
      "epoch": 1.1408450704225352,
      "grad_norm": 0.3956218361854553,
      "learning_rate": 8.628605061801059e-05,
      "loss": 0.6586,
      "step": 972
    },
    {
      "epoch": 1.142018779342723,
      "grad_norm": 0.3789028227329254,
      "learning_rate": 8.616833431430254e-05,
      "loss": 0.7415,
      "step": 973
    },
    {
      "epoch": 1.1431924882629108,
      "grad_norm": 0.3878764808177948,
      "learning_rate": 8.605061801059447e-05,
      "loss": 0.6559,
      "step": 974
    },
    {
      "epoch": 1.1443661971830985,
      "grad_norm": 0.37901559472084045,
      "learning_rate": 8.593290170688641e-05,
      "loss": 0.6685,
      "step": 975
    },
    {
      "epoch": 1.1455399061032865,
      "grad_norm": 0.40399041771888733,
      "learning_rate": 8.581518540317834e-05,
      "loss": 0.6602,
      "step": 976
    },
    {
      "epoch": 1.1467136150234742,
      "grad_norm": 0.38144391775131226,
      "learning_rate": 8.569746909947029e-05,
      "loss": 0.6683,
      "step": 977
    },
    {
      "epoch": 1.147887323943662,
      "grad_norm": 0.3610433042049408,
      "learning_rate": 8.557975279576221e-05,
      "loss": 0.6579,
      "step": 978
    },
    {
      "epoch": 1.1490610328638498,
      "grad_norm": 0.42147722840309143,
      "learning_rate": 8.546203649205416e-05,
      "loss": 0.6997,
      "step": 979
    },
    {
      "epoch": 1.1502347417840375,
      "grad_norm": 0.3799455761909485,
      "learning_rate": 8.53443201883461e-05,
      "loss": 0.7096,
      "step": 980
    },
    {
      "epoch": 1.1514084507042253,
      "grad_norm": 0.4173739552497864,
      "learning_rate": 8.522660388463803e-05,
      "loss": 0.6708,
      "step": 981
    },
    {
      "epoch": 1.1525821596244132,
      "grad_norm": 0.3997640013694763,
      "learning_rate": 8.510888758092996e-05,
      "loss": 0.6514,
      "step": 982
    },
    {
      "epoch": 1.153755868544601,
      "grad_norm": 0.3758656978607178,
      "learning_rate": 8.49911712772219e-05,
      "loss": 0.6442,
      "step": 983
    },
    {
      "epoch": 1.1549295774647887,
      "grad_norm": 0.37429675459861755,
      "learning_rate": 8.487345497351383e-05,
      "loss": 0.6619,
      "step": 984
    },
    {
      "epoch": 1.1561032863849765,
      "grad_norm": 0.3747265934944153,
      "learning_rate": 8.475573866980577e-05,
      "loss": 0.7107,
      "step": 985
    },
    {
      "epoch": 1.1572769953051643,
      "grad_norm": 0.37782514095306396,
      "learning_rate": 8.463802236609771e-05,
      "loss": 0.7241,
      "step": 986
    },
    {
      "epoch": 1.158450704225352,
      "grad_norm": 0.3703122138977051,
      "learning_rate": 8.452030606238965e-05,
      "loss": 0.6952,
      "step": 987
    },
    {
      "epoch": 1.15962441314554,
      "grad_norm": 0.37990477681159973,
      "learning_rate": 8.440258975868158e-05,
      "loss": 0.7364,
      "step": 988
    },
    {
      "epoch": 1.1607981220657277,
      "grad_norm": 0.42046844959259033,
      "learning_rate": 8.428487345497352e-05,
      "loss": 0.695,
      "step": 989
    },
    {
      "epoch": 1.1619718309859155,
      "grad_norm": 0.3745966851711273,
      "learning_rate": 8.416715715126545e-05,
      "loss": 0.6875,
      "step": 990
    },
    {
      "epoch": 1.1631455399061033,
      "grad_norm": 0.3496320843696594,
      "learning_rate": 8.404944084755739e-05,
      "loss": 0.6826,
      "step": 991
    },
    {
      "epoch": 1.164319248826291,
      "grad_norm": 0.39181873202323914,
      "learning_rate": 8.393172454384934e-05,
      "loss": 0.6937,
      "step": 992
    },
    {
      "epoch": 1.165492957746479,
      "grad_norm": 0.3910543620586395,
      "learning_rate": 8.381400824014126e-05,
      "loss": 0.749,
      "step": 993
    },
    {
      "epoch": 1.1666666666666667,
      "grad_norm": 0.3770748972892761,
      "learning_rate": 8.36962919364332e-05,
      "loss": 0.6743,
      "step": 994
    },
    {
      "epoch": 1.1678403755868545,
      "grad_norm": 0.3675018846988678,
      "learning_rate": 8.357857563272513e-05,
      "loss": 0.6499,
      "step": 995
    },
    {
      "epoch": 1.1690140845070423,
      "grad_norm": 0.36867639422416687,
      "learning_rate": 8.346085932901707e-05,
      "loss": 0.6642,
      "step": 996
    },
    {
      "epoch": 1.17018779342723,
      "grad_norm": 0.3860320746898651,
      "learning_rate": 8.334314302530901e-05,
      "loss": 0.6947,
      "step": 997
    },
    {
      "epoch": 1.1713615023474178,
      "grad_norm": 0.36680731177330017,
      "learning_rate": 8.322542672160094e-05,
      "loss": 0.7111,
      "step": 998
    },
    {
      "epoch": 1.1725352112676055,
      "grad_norm": 0.38997524976730347,
      "learning_rate": 8.310771041789288e-05,
      "loss": 0.6842,
      "step": 999
    },
    {
      "epoch": 1.1737089201877935,
      "grad_norm": 0.3883102834224701,
      "learning_rate": 8.298999411418483e-05,
      "loss": 0.6655,
      "step": 1000
    }
  ],
  "logging_steps": 1,
  "max_steps": 1704,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 2,
  "save_steps": 100,
  "total_flos": 1.4844237184940114e+18,
  "train_batch_size": 4,
  "trial_name": null,
  "trial_params": null
}