random_btK13FgDAVNMY44K / trainer_state.json
cutelemonlili's picture
Add files using upload-large-folder tool
967979c verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 1284,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.001557632398753894,
"grad_norm": 1.7232097279069922,
"learning_rate": 9.999985033870806e-06,
"loss": 0.6355,
"step": 1
},
{
"epoch": 0.003115264797507788,
"grad_norm": 0.7663661005911214,
"learning_rate": 9.999940135572817e-06,
"loss": 0.2544,
"step": 2
},
{
"epoch": 0.004672897196261682,
"grad_norm": 0.8839734168626316,
"learning_rate": 9.999865305374812e-06,
"loss": 0.3767,
"step": 3
},
{
"epoch": 0.006230529595015576,
"grad_norm": 0.8042784289646617,
"learning_rate": 9.999760543724761e-06,
"loss": 0.184,
"step": 4
},
{
"epoch": 0.00778816199376947,
"grad_norm": 0.6765931399234855,
"learning_rate": 9.999625851249816e-06,
"loss": 0.2941,
"step": 5
},
{
"epoch": 0.009345794392523364,
"grad_norm": 0.7193614710776666,
"learning_rate": 9.999461228756304e-06,
"loss": 0.2705,
"step": 6
},
{
"epoch": 0.010903426791277258,
"grad_norm": 0.7344615984114217,
"learning_rate": 9.99926667722973e-06,
"loss": 0.3259,
"step": 7
},
{
"epoch": 0.012461059190031152,
"grad_norm": 0.73017739071341,
"learning_rate": 9.999042197834769e-06,
"loss": 0.2852,
"step": 8
},
{
"epoch": 0.014018691588785047,
"grad_norm": 0.592507619306345,
"learning_rate": 9.998787791915254e-06,
"loss": 0.2649,
"step": 9
},
{
"epoch": 0.01557632398753894,
"grad_norm": 0.9399500241178238,
"learning_rate": 9.998503460994176e-06,
"loss": 0.2521,
"step": 10
},
{
"epoch": 0.017133956386292833,
"grad_norm": 0.7765731965837874,
"learning_rate": 9.998189206773665e-06,
"loss": 0.3556,
"step": 11
},
{
"epoch": 0.018691588785046728,
"grad_norm": 0.6693444621163488,
"learning_rate": 9.997845031134992e-06,
"loss": 0.2239,
"step": 12
},
{
"epoch": 0.020249221183800622,
"grad_norm": 1.4936804808243616,
"learning_rate": 9.997470936138547e-06,
"loss": 0.8968,
"step": 13
},
{
"epoch": 0.021806853582554516,
"grad_norm": 0.7687864170878725,
"learning_rate": 9.997066924023832e-06,
"loss": 0.294,
"step": 14
},
{
"epoch": 0.02336448598130841,
"grad_norm": 1.0982871522503537,
"learning_rate": 9.996632997209444e-06,
"loss": 0.3007,
"step": 15
},
{
"epoch": 0.024922118380062305,
"grad_norm": 0.9371896720117513,
"learning_rate": 9.996169158293066e-06,
"loss": 0.2438,
"step": 16
},
{
"epoch": 0.0264797507788162,
"grad_norm": 0.7798940457815309,
"learning_rate": 9.995675410051448e-06,
"loss": 0.2212,
"step": 17
},
{
"epoch": 0.028037383177570093,
"grad_norm": 0.7355604168049059,
"learning_rate": 9.99515175544039e-06,
"loss": 0.2793,
"step": 18
},
{
"epoch": 0.029595015576323987,
"grad_norm": 0.8219475835131775,
"learning_rate": 9.994598197594723e-06,
"loss": 0.2868,
"step": 19
},
{
"epoch": 0.03115264797507788,
"grad_norm": 0.8165917677127098,
"learning_rate": 9.994014739828298e-06,
"loss": 0.2703,
"step": 20
},
{
"epoch": 0.03271028037383177,
"grad_norm": 0.733767269375867,
"learning_rate": 9.993401385633951e-06,
"loss": 0.2059,
"step": 21
},
{
"epoch": 0.03426791277258567,
"grad_norm": 1.2694949029078932,
"learning_rate": 9.992758138683502e-06,
"loss": 0.9151,
"step": 22
},
{
"epoch": 0.03582554517133956,
"grad_norm": 0.7777986747788719,
"learning_rate": 9.992085002827719e-06,
"loss": 0.2599,
"step": 23
},
{
"epoch": 0.037383177570093455,
"grad_norm": 0.610655735988334,
"learning_rate": 9.991381982096293e-06,
"loss": 0.2347,
"step": 24
},
{
"epoch": 0.03894080996884735,
"grad_norm": 0.8562287986067203,
"learning_rate": 9.990649080697825e-06,
"loss": 0.2359,
"step": 25
},
{
"epoch": 0.040498442367601244,
"grad_norm": 0.6673172477309072,
"learning_rate": 9.989886303019793e-06,
"loss": 0.2271,
"step": 26
},
{
"epoch": 0.04205607476635514,
"grad_norm": 0.8629676432514921,
"learning_rate": 9.989093653628532e-06,
"loss": 0.2304,
"step": 27
},
{
"epoch": 0.04361370716510903,
"grad_norm": 0.6898071747588274,
"learning_rate": 9.988271137269197e-06,
"loss": 0.3192,
"step": 28
},
{
"epoch": 0.045171339563862926,
"grad_norm": 0.5010462374255644,
"learning_rate": 9.987418758865743e-06,
"loss": 0.2406,
"step": 29
},
{
"epoch": 0.04672897196261682,
"grad_norm": 0.8259320779398918,
"learning_rate": 9.98653652352089e-06,
"loss": 0.3266,
"step": 30
},
{
"epoch": 0.048286604361370715,
"grad_norm": 0.6076733459531145,
"learning_rate": 9.9856244365161e-06,
"loss": 0.2997,
"step": 31
},
{
"epoch": 0.04984423676012461,
"grad_norm": 0.7298955355144913,
"learning_rate": 9.984682503311536e-06,
"loss": 0.2967,
"step": 32
},
{
"epoch": 0.0514018691588785,
"grad_norm": 0.7346554751645953,
"learning_rate": 9.983710729546038e-06,
"loss": 0.3105,
"step": 33
},
{
"epoch": 0.0529595015576324,
"grad_norm": 0.668130630437843,
"learning_rate": 9.98270912103708e-06,
"loss": 0.3535,
"step": 34
},
{
"epoch": 0.05451713395638629,
"grad_norm": 0.6379634068346577,
"learning_rate": 9.981677683780743e-06,
"loss": 0.3709,
"step": 35
},
{
"epoch": 0.056074766355140186,
"grad_norm": 0.798860986544702,
"learning_rate": 9.98061642395168e-06,
"loss": 0.2243,
"step": 36
},
{
"epoch": 0.05763239875389408,
"grad_norm": 0.6593735789582216,
"learning_rate": 9.979525347903067e-06,
"loss": 0.2541,
"step": 37
},
{
"epoch": 0.059190031152647975,
"grad_norm": 0.7224987244711324,
"learning_rate": 9.978404462166582e-06,
"loss": 0.2736,
"step": 38
},
{
"epoch": 0.06074766355140187,
"grad_norm": 0.6440291059790796,
"learning_rate": 9.977253773452349e-06,
"loss": 0.2764,
"step": 39
},
{
"epoch": 0.06230529595015576,
"grad_norm": 0.6222848807943162,
"learning_rate": 9.976073288648913e-06,
"loss": 0.2698,
"step": 40
},
{
"epoch": 0.06386292834890965,
"grad_norm": 0.7394026438854199,
"learning_rate": 9.97486301482319e-06,
"loss": 0.2522,
"step": 41
},
{
"epoch": 0.06542056074766354,
"grad_norm": 0.7520093536738988,
"learning_rate": 9.973622959220426e-06,
"loss": 0.2137,
"step": 42
},
{
"epoch": 0.06697819314641744,
"grad_norm": 0.744961609793494,
"learning_rate": 9.97235312926415e-06,
"loss": 0.2328,
"step": 43
},
{
"epoch": 0.06853582554517133,
"grad_norm": 0.7656967225499874,
"learning_rate": 9.971053532556143e-06,
"loss": 0.2669,
"step": 44
},
{
"epoch": 0.07009345794392523,
"grad_norm": 0.6382236962718162,
"learning_rate": 9.969724176876373e-06,
"loss": 0.2192,
"step": 45
},
{
"epoch": 0.07165109034267912,
"grad_norm": 0.6472374848660734,
"learning_rate": 9.968365070182967e-06,
"loss": 0.3144,
"step": 46
},
{
"epoch": 0.07320872274143302,
"grad_norm": 0.6391395623819942,
"learning_rate": 9.966976220612151e-06,
"loss": 0.2048,
"step": 47
},
{
"epoch": 0.07476635514018691,
"grad_norm": 0.7757553970772143,
"learning_rate": 9.965557636478203e-06,
"loss": 0.2579,
"step": 48
},
{
"epoch": 0.0763239875389408,
"grad_norm": 0.8133958395260938,
"learning_rate": 9.964109326273411e-06,
"loss": 0.3226,
"step": 49
},
{
"epoch": 0.0778816199376947,
"grad_norm": 0.4637065889612604,
"learning_rate": 9.962631298668015e-06,
"loss": 0.185,
"step": 50
},
{
"epoch": 0.0794392523364486,
"grad_norm": 0.6273381176066575,
"learning_rate": 9.961123562510153e-06,
"loss": 0.3509,
"step": 51
},
{
"epoch": 0.08099688473520249,
"grad_norm": 0.6419234115207709,
"learning_rate": 9.959586126825818e-06,
"loss": 0.2806,
"step": 52
},
{
"epoch": 0.08255451713395638,
"grad_norm": 0.7160133507575562,
"learning_rate": 9.95801900081879e-06,
"loss": 0.1838,
"step": 53
},
{
"epoch": 0.08411214953271028,
"grad_norm": 0.7848165329694581,
"learning_rate": 9.956422193870597e-06,
"loss": 0.255,
"step": 54
},
{
"epoch": 0.08566978193146417,
"grad_norm": 0.6040970248107416,
"learning_rate": 9.954795715540447e-06,
"loss": 0.2142,
"step": 55
},
{
"epoch": 0.08722741433021806,
"grad_norm": 0.6115559611935754,
"learning_rate": 9.953139575565169e-06,
"loss": 0.2668,
"step": 56
},
{
"epoch": 0.08878504672897196,
"grad_norm": 0.6298402847496053,
"learning_rate": 9.95145378385917e-06,
"loss": 0.2881,
"step": 57
},
{
"epoch": 0.09034267912772585,
"grad_norm": 0.5535053280930566,
"learning_rate": 9.949738350514358e-06,
"loss": 0.2202,
"step": 58
},
{
"epoch": 0.09190031152647975,
"grad_norm": 0.6352050541969153,
"learning_rate": 9.947993285800093e-06,
"loss": 0.3398,
"step": 59
},
{
"epoch": 0.09345794392523364,
"grad_norm": 0.826557642193822,
"learning_rate": 9.94621860016312e-06,
"loss": 0.3364,
"step": 60
},
{
"epoch": 0.09501557632398754,
"grad_norm": 0.7298038729692221,
"learning_rate": 9.944414304227508e-06,
"loss": 0.2634,
"step": 61
},
{
"epoch": 0.09657320872274143,
"grad_norm": 1.1507218720730263,
"learning_rate": 9.94258040879459e-06,
"loss": 0.8284,
"step": 62
},
{
"epoch": 0.09813084112149532,
"grad_norm": 0.6885391578463873,
"learning_rate": 9.94071692484289e-06,
"loss": 0.2336,
"step": 63
},
{
"epoch": 0.09968847352024922,
"grad_norm": 0.7334194656077329,
"learning_rate": 9.938823863528065e-06,
"loss": 0.2652,
"step": 64
},
{
"epoch": 0.10124610591900311,
"grad_norm": 1.0514064266169973,
"learning_rate": 9.936901236182836e-06,
"loss": 0.835,
"step": 65
},
{
"epoch": 0.102803738317757,
"grad_norm": 0.6334750887159478,
"learning_rate": 9.934949054316917e-06,
"loss": 0.2812,
"step": 66
},
{
"epoch": 0.1043613707165109,
"grad_norm": 0.4857684609448803,
"learning_rate": 9.932967329616953e-06,
"loss": 0.202,
"step": 67
},
{
"epoch": 0.1059190031152648,
"grad_norm": 0.690411540980916,
"learning_rate": 9.930956073946442e-06,
"loss": 0.2828,
"step": 68
},
{
"epoch": 0.10747663551401869,
"grad_norm": 0.6862713871911896,
"learning_rate": 9.928915299345669e-06,
"loss": 0.3246,
"step": 69
},
{
"epoch": 0.10903426791277258,
"grad_norm": 0.6617081291471549,
"learning_rate": 9.926845018031632e-06,
"loss": 0.2621,
"step": 70
},
{
"epoch": 0.11059190031152648,
"grad_norm": 0.5745643134324515,
"learning_rate": 9.924745242397968e-06,
"loss": 0.2224,
"step": 71
},
{
"epoch": 0.11214953271028037,
"grad_norm": 0.7300598195282793,
"learning_rate": 9.922615985014887e-06,
"loss": 0.2297,
"step": 72
},
{
"epoch": 0.11370716510903427,
"grad_norm": 0.6648229017872759,
"learning_rate": 9.920457258629081e-06,
"loss": 0.2461,
"step": 73
},
{
"epoch": 0.11526479750778816,
"grad_norm": 0.6868885744716628,
"learning_rate": 9.918269076163664e-06,
"loss": 0.2728,
"step": 74
},
{
"epoch": 0.11682242990654206,
"grad_norm": 0.6982895714335172,
"learning_rate": 9.916051450718085e-06,
"loss": 0.257,
"step": 75
},
{
"epoch": 0.11838006230529595,
"grad_norm": 0.5677598559231926,
"learning_rate": 9.91380439556805e-06,
"loss": 0.2614,
"step": 76
},
{
"epoch": 0.11993769470404984,
"grad_norm": 0.5071293538430349,
"learning_rate": 9.911527924165445e-06,
"loss": 0.2793,
"step": 77
},
{
"epoch": 0.12149532710280374,
"grad_norm": 0.7281461877976642,
"learning_rate": 9.909222050138259e-06,
"loss": 0.2861,
"step": 78
},
{
"epoch": 0.12305295950155763,
"grad_norm": 0.6491525810286177,
"learning_rate": 9.906886787290492e-06,
"loss": 0.2799,
"step": 79
},
{
"epoch": 0.12461059190031153,
"grad_norm": 0.9626139969649712,
"learning_rate": 9.904522149602087e-06,
"loss": 0.2251,
"step": 80
},
{
"epoch": 0.1261682242990654,
"grad_norm": 0.705551656345552,
"learning_rate": 9.902128151228827e-06,
"loss": 0.2899,
"step": 81
},
{
"epoch": 0.1277258566978193,
"grad_norm": 0.5796094318767273,
"learning_rate": 9.899704806502272e-06,
"loss": 0.3023,
"step": 82
},
{
"epoch": 0.1292834890965732,
"grad_norm": 0.5778885783153778,
"learning_rate": 9.89725212992966e-06,
"loss": 0.2596,
"step": 83
},
{
"epoch": 0.1308411214953271,
"grad_norm": 0.6407326481544323,
"learning_rate": 9.894770136193814e-06,
"loss": 0.3182,
"step": 84
},
{
"epoch": 0.13239875389408098,
"grad_norm": 0.8761396102635524,
"learning_rate": 9.892258840153075e-06,
"loss": 0.3261,
"step": 85
},
{
"epoch": 0.13395638629283488,
"grad_norm": 0.8248387465515045,
"learning_rate": 9.889718256841195e-06,
"loss": 0.3105,
"step": 86
},
{
"epoch": 0.13551401869158877,
"grad_norm": 2.168823178408131,
"learning_rate": 9.88714840146725e-06,
"loss": 0.7869,
"step": 87
},
{
"epoch": 0.13707165109034267,
"grad_norm": 0.6163273279750748,
"learning_rate": 9.884549289415556e-06,
"loss": 0.2249,
"step": 88
},
{
"epoch": 0.13862928348909656,
"grad_norm": 1.5331833228458356,
"learning_rate": 9.881920936245577e-06,
"loss": 0.7778,
"step": 89
},
{
"epoch": 0.14018691588785046,
"grad_norm": 1.5428867178521768,
"learning_rate": 9.879263357691815e-06,
"loss": 0.8951,
"step": 90
},
{
"epoch": 0.14174454828660435,
"grad_norm": 0.6177160477898084,
"learning_rate": 9.876576569663739e-06,
"loss": 0.224,
"step": 91
},
{
"epoch": 0.14330218068535824,
"grad_norm": 0.7981644114495157,
"learning_rate": 9.873860588245675e-06,
"loss": 0.2892,
"step": 92
},
{
"epoch": 0.14485981308411214,
"grad_norm": 0.600801453450908,
"learning_rate": 9.871115429696716e-06,
"loss": 0.2782,
"step": 93
},
{
"epoch": 0.14641744548286603,
"grad_norm": 0.6547888169420532,
"learning_rate": 9.868341110450618e-06,
"loss": 0.3347,
"step": 94
},
{
"epoch": 0.14797507788161993,
"grad_norm": 0.5823853177715245,
"learning_rate": 9.865537647115713e-06,
"loss": 0.2778,
"step": 95
},
{
"epoch": 0.14953271028037382,
"grad_norm": 0.9116578300075924,
"learning_rate": 9.862705056474795e-06,
"loss": 0.3503,
"step": 96
},
{
"epoch": 0.15109034267912771,
"grad_norm": 0.5426782446240797,
"learning_rate": 9.859843355485033e-06,
"loss": 0.2386,
"step": 97
},
{
"epoch": 0.1526479750778816,
"grad_norm": 0.5543482391953797,
"learning_rate": 9.856952561277862e-06,
"loss": 0.2344,
"step": 98
},
{
"epoch": 0.1542056074766355,
"grad_norm": 0.7660668413572433,
"learning_rate": 9.854032691158881e-06,
"loss": 0.2692,
"step": 99
},
{
"epoch": 0.1557632398753894,
"grad_norm": 0.7745787606896338,
"learning_rate": 9.851083762607752e-06,
"loss": 0.2374,
"step": 100
},
{
"epoch": 0.1573208722741433,
"grad_norm": 0.8283533162770704,
"learning_rate": 9.848105793278092e-06,
"loss": 0.2699,
"step": 101
},
{
"epoch": 0.1588785046728972,
"grad_norm": 0.5323847631668214,
"learning_rate": 9.84509880099737e-06,
"loss": 0.2177,
"step": 102
},
{
"epoch": 0.16043613707165108,
"grad_norm": 0.7022062850923283,
"learning_rate": 9.842062803766804e-06,
"loss": 0.2258,
"step": 103
},
{
"epoch": 0.16199376947040497,
"grad_norm": 1.9739756098890826,
"learning_rate": 9.838997819761241e-06,
"loss": 0.9263,
"step": 104
},
{
"epoch": 0.16355140186915887,
"grad_norm": 0.7651110334163559,
"learning_rate": 9.835903867329061e-06,
"loss": 0.2654,
"step": 105
},
{
"epoch": 0.16510903426791276,
"grad_norm": 0.6634416278579489,
"learning_rate": 9.832780964992059e-06,
"loss": 0.2959,
"step": 106
},
{
"epoch": 0.16666666666666666,
"grad_norm": 0.62947915393871,
"learning_rate": 9.829629131445342e-06,
"loss": 0.2608,
"step": 107
},
{
"epoch": 0.16822429906542055,
"grad_norm": 1.55385845539561,
"learning_rate": 9.826448385557208e-06,
"loss": 0.842,
"step": 108
},
{
"epoch": 0.16978193146417445,
"grad_norm": 0.6067362647236724,
"learning_rate": 9.823238746369038e-06,
"loss": 0.3388,
"step": 109
},
{
"epoch": 0.17133956386292834,
"grad_norm": 0.7095716481506402,
"learning_rate": 9.820000233095179e-06,
"loss": 0.2953,
"step": 110
},
{
"epoch": 0.17289719626168223,
"grad_norm": 1.4374378842104667,
"learning_rate": 9.81673286512284e-06,
"loss": 0.7775,
"step": 111
},
{
"epoch": 0.17445482866043613,
"grad_norm": 0.6761770106689733,
"learning_rate": 9.813436662011958e-06,
"loss": 0.2929,
"step": 112
},
{
"epoch": 0.17601246105919002,
"grad_norm": 0.7954842040051252,
"learning_rate": 9.810111643495095e-06,
"loss": 0.2734,
"step": 113
},
{
"epoch": 0.17757009345794392,
"grad_norm": 1.271851983971207,
"learning_rate": 9.806757829477313e-06,
"loss": 0.7987,
"step": 114
},
{
"epoch": 0.1791277258566978,
"grad_norm": 0.526209190777812,
"learning_rate": 9.803375240036057e-06,
"loss": 0.3054,
"step": 115
},
{
"epoch": 0.1806853582554517,
"grad_norm": 0.6440505513615228,
"learning_rate": 9.799963895421036e-06,
"loss": 0.2416,
"step": 116
},
{
"epoch": 0.1822429906542056,
"grad_norm": 0.9367999131179932,
"learning_rate": 9.7965238160541e-06,
"loss": 0.3409,
"step": 117
},
{
"epoch": 0.1838006230529595,
"grad_norm": 0.8415567444030029,
"learning_rate": 9.793055022529116e-06,
"loss": 0.2847,
"step": 118
},
{
"epoch": 0.1853582554517134,
"grad_norm": 0.5786248320276434,
"learning_rate": 9.789557535611853e-06,
"loss": 0.2644,
"step": 119
},
{
"epoch": 0.18691588785046728,
"grad_norm": 0.687463433264138,
"learning_rate": 9.786031376239842e-06,
"loss": 0.2569,
"step": 120
},
{
"epoch": 0.18847352024922118,
"grad_norm": 0.45347828103414345,
"learning_rate": 9.78247656552227e-06,
"loss": 0.2415,
"step": 121
},
{
"epoch": 0.19003115264797507,
"grad_norm": 0.5650726061282849,
"learning_rate": 9.778893124739836e-06,
"loss": 0.2547,
"step": 122
},
{
"epoch": 0.19158878504672897,
"grad_norm": 0.5103395837306397,
"learning_rate": 9.775281075344639e-06,
"loss": 0.3065,
"step": 123
},
{
"epoch": 0.19314641744548286,
"grad_norm": 0.5183636563691485,
"learning_rate": 9.771640438960037e-06,
"loss": 0.2922,
"step": 124
},
{
"epoch": 0.19470404984423675,
"grad_norm": 0.5026042818767394,
"learning_rate": 9.76797123738052e-06,
"loss": 0.2588,
"step": 125
},
{
"epoch": 0.19626168224299065,
"grad_norm": 0.6794540572358105,
"learning_rate": 9.76427349257159e-06,
"loss": 0.3041,
"step": 126
},
{
"epoch": 0.19781931464174454,
"grad_norm": 0.6283325050944261,
"learning_rate": 9.76054722666962e-06,
"loss": 0.2991,
"step": 127
},
{
"epoch": 0.19937694704049844,
"grad_norm": 0.7184815540245492,
"learning_rate": 9.756792461981713e-06,
"loss": 0.2599,
"step": 128
},
{
"epoch": 0.20093457943925233,
"grad_norm": 0.701839843872524,
"learning_rate": 9.753009220985593e-06,
"loss": 0.2711,
"step": 129
},
{
"epoch": 0.20249221183800623,
"grad_norm": 0.6505643462960924,
"learning_rate": 9.749197526329446e-06,
"loss": 0.2947,
"step": 130
},
{
"epoch": 0.20404984423676012,
"grad_norm": 0.6104336914925119,
"learning_rate": 9.745357400831799e-06,
"loss": 0.2331,
"step": 131
},
{
"epoch": 0.205607476635514,
"grad_norm": 0.908932761888463,
"learning_rate": 9.741488867481377e-06,
"loss": 0.2377,
"step": 132
},
{
"epoch": 0.2071651090342679,
"grad_norm": 0.5980236644974856,
"learning_rate": 9.737591949436969e-06,
"loss": 0.1601,
"step": 133
},
{
"epoch": 0.2087227414330218,
"grad_norm": 0.5937884904455855,
"learning_rate": 9.733666670027288e-06,
"loss": 0.2542,
"step": 134
},
{
"epoch": 0.2102803738317757,
"grad_norm": 0.7666646448496817,
"learning_rate": 9.729713052750827e-06,
"loss": 0.2465,
"step": 135
},
{
"epoch": 0.2118380062305296,
"grad_norm": 0.5386370084617528,
"learning_rate": 9.725731121275725e-06,
"loss": 0.3172,
"step": 136
},
{
"epoch": 0.21339563862928349,
"grad_norm": 0.5517332858518383,
"learning_rate": 9.721720899439621e-06,
"loss": 0.2864,
"step": 137
},
{
"epoch": 0.21495327102803738,
"grad_norm": 0.60004714901346,
"learning_rate": 9.71768241124952e-06,
"loss": 0.3101,
"step": 138
},
{
"epoch": 0.21651090342679127,
"grad_norm": 0.7525697267510544,
"learning_rate": 9.71361568088163e-06,
"loss": 0.2235,
"step": 139
},
{
"epoch": 0.21806853582554517,
"grad_norm": 0.5929490381364206,
"learning_rate": 9.709520732681238e-06,
"loss": 0.231,
"step": 140
},
{
"epoch": 0.21962616822429906,
"grad_norm": 0.7027523851155623,
"learning_rate": 9.705397591162556e-06,
"loss": 0.2379,
"step": 141
},
{
"epoch": 0.22118380062305296,
"grad_norm": 0.6504000514836087,
"learning_rate": 9.701246281008568e-06,
"loss": 0.2462,
"step": 142
},
{
"epoch": 0.22274143302180685,
"grad_norm": 0.5770728592720746,
"learning_rate": 9.697066827070894e-06,
"loss": 0.2859,
"step": 143
},
{
"epoch": 0.22429906542056074,
"grad_norm": 0.5798579381876525,
"learning_rate": 9.692859254369631e-06,
"loss": 0.2663,
"step": 144
},
{
"epoch": 0.22585669781931464,
"grad_norm": 0.5033709259076921,
"learning_rate": 9.68862358809321e-06,
"loss": 0.1913,
"step": 145
},
{
"epoch": 0.22741433021806853,
"grad_norm": 0.5493196423085126,
"learning_rate": 9.684359853598245e-06,
"loss": 0.2197,
"step": 146
},
{
"epoch": 0.22897196261682243,
"grad_norm": 0.5652921319820055,
"learning_rate": 9.680068076409373e-06,
"loss": 0.2601,
"step": 147
},
{
"epoch": 0.23052959501557632,
"grad_norm": 0.6783593909741826,
"learning_rate": 9.675748282219114e-06,
"loss": 0.2922,
"step": 148
},
{
"epoch": 0.23208722741433022,
"grad_norm": 0.7088765017570264,
"learning_rate": 9.671400496887704e-06,
"loss": 0.2997,
"step": 149
},
{
"epoch": 0.2336448598130841,
"grad_norm": 0.6064540912244738,
"learning_rate": 9.667024746442953e-06,
"loss": 0.3283,
"step": 150
},
{
"epoch": 0.235202492211838,
"grad_norm": 0.6900794118386824,
"learning_rate": 9.662621057080077e-06,
"loss": 0.2447,
"step": 151
},
{
"epoch": 0.2367601246105919,
"grad_norm": 0.8898873494327856,
"learning_rate": 9.65818945516155e-06,
"loss": 0.2857,
"step": 152
},
{
"epoch": 0.2383177570093458,
"grad_norm": 0.7596388046727889,
"learning_rate": 9.653729967216944e-06,
"loss": 0.2765,
"step": 153
},
{
"epoch": 0.2398753894080997,
"grad_norm": 0.5897245550654492,
"learning_rate": 9.64924261994277e-06,
"loss": 0.2629,
"step": 154
},
{
"epoch": 0.24143302180685358,
"grad_norm": 0.6261944121589039,
"learning_rate": 9.644727440202308e-06,
"loss": 0.3554,
"step": 155
},
{
"epoch": 0.24299065420560748,
"grad_norm": 0.7471727372694826,
"learning_rate": 9.640184455025472e-06,
"loss": 0.3595,
"step": 156
},
{
"epoch": 0.24454828660436137,
"grad_norm": 0.6539521944383881,
"learning_rate": 9.635613691608619e-06,
"loss": 0.2351,
"step": 157
},
{
"epoch": 0.24610591900311526,
"grad_norm": 0.7372613573398835,
"learning_rate": 9.631015177314402e-06,
"loss": 0.3061,
"step": 158
},
{
"epoch": 0.24766355140186916,
"grad_norm": 0.802899194753761,
"learning_rate": 9.62638893967161e-06,
"loss": 0.3606,
"step": 159
},
{
"epoch": 0.24922118380062305,
"grad_norm": 0.754935748177354,
"learning_rate": 9.621735006374984e-06,
"loss": 0.223,
"step": 160
},
{
"epoch": 0.2507788161993769,
"grad_norm": 0.4173684214144698,
"learning_rate": 9.617053405285077e-06,
"loss": 0.2319,
"step": 161
},
{
"epoch": 0.2523364485981308,
"grad_norm": 0.6435417325636267,
"learning_rate": 9.612344164428063e-06,
"loss": 0.3996,
"step": 162
},
{
"epoch": 0.2538940809968847,
"grad_norm": 0.8830545123878781,
"learning_rate": 9.607607311995587e-06,
"loss": 0.3724,
"step": 163
},
{
"epoch": 0.2554517133956386,
"grad_norm": 2.2533700279001554,
"learning_rate": 9.602842876344589e-06,
"loss": 0.8261,
"step": 164
},
{
"epoch": 0.2570093457943925,
"grad_norm": 0.6641726752237889,
"learning_rate": 9.59805088599713e-06,
"loss": 0.3339,
"step": 165
},
{
"epoch": 0.2585669781931464,
"grad_norm": 0.6276989828906703,
"learning_rate": 9.59323136964023e-06,
"loss": 0.219,
"step": 166
},
{
"epoch": 0.2601246105919003,
"grad_norm": 0.8210416456426778,
"learning_rate": 9.588384356125691e-06,
"loss": 0.282,
"step": 167
},
{
"epoch": 0.2616822429906542,
"grad_norm": 0.751594963703605,
"learning_rate": 9.583509874469924e-06,
"loss": 0.3373,
"step": 168
},
{
"epoch": 0.2632398753894081,
"grad_norm": 0.7261273311893066,
"learning_rate": 9.578607953853777e-06,
"loss": 0.28,
"step": 169
},
{
"epoch": 0.26479750778816197,
"grad_norm": 0.6922505407762392,
"learning_rate": 9.573678623622364e-06,
"loss": 0.3412,
"step": 170
},
{
"epoch": 0.26635514018691586,
"grad_norm": 0.6419868376151807,
"learning_rate": 9.568721913284879e-06,
"loss": 0.2588,
"step": 171
},
{
"epoch": 0.26791277258566976,
"grad_norm": 0.49324052267306684,
"learning_rate": 9.563737852514432e-06,
"loss": 0.2332,
"step": 172
},
{
"epoch": 0.26947040498442365,
"grad_norm": 0.5770500149070207,
"learning_rate": 9.558726471147862e-06,
"loss": 0.2937,
"step": 173
},
{
"epoch": 0.27102803738317754,
"grad_norm": 0.7329648250931949,
"learning_rate": 9.553687799185556e-06,
"loss": 0.3309,
"step": 174
},
{
"epoch": 0.27258566978193144,
"grad_norm": 0.4864393923841514,
"learning_rate": 9.548621866791286e-06,
"loss": 0.2043,
"step": 175
},
{
"epoch": 0.27414330218068533,
"grad_norm": 1.136074014302222,
"learning_rate": 9.54352870429201e-06,
"loss": 0.7893,
"step": 176
},
{
"epoch": 0.2757009345794392,
"grad_norm": 0.6443586968700755,
"learning_rate": 9.538408342177699e-06,
"loss": 0.3025,
"step": 177
},
{
"epoch": 0.2772585669781931,
"grad_norm": 0.6403852564269055,
"learning_rate": 9.533260811101152e-06,
"loss": 0.2629,
"step": 178
},
{
"epoch": 0.278816199376947,
"grad_norm": 0.6869290500301357,
"learning_rate": 9.528086141877817e-06,
"loss": 0.301,
"step": 179
},
{
"epoch": 0.2803738317757009,
"grad_norm": 0.9338061428659953,
"learning_rate": 9.522884365485599e-06,
"loss": 0.7925,
"step": 180
},
{
"epoch": 0.2819314641744548,
"grad_norm": 0.7304512334819939,
"learning_rate": 9.517655513064682e-06,
"loss": 0.44,
"step": 181
},
{
"epoch": 0.2834890965732087,
"grad_norm": 0.6207928517482323,
"learning_rate": 9.512399615917342e-06,
"loss": 0.228,
"step": 182
},
{
"epoch": 0.2850467289719626,
"grad_norm": 0.8847617104487071,
"learning_rate": 9.507116705507748e-06,
"loss": 0.4137,
"step": 183
},
{
"epoch": 0.2866043613707165,
"grad_norm": 0.6631070285038642,
"learning_rate": 9.50180681346179e-06,
"loss": 0.3438,
"step": 184
},
{
"epoch": 0.2881619937694704,
"grad_norm": 1.0647904328891042,
"learning_rate": 9.49646997156688e-06,
"loss": 0.2452,
"step": 185
},
{
"epoch": 0.2897196261682243,
"grad_norm": 0.4959301062995678,
"learning_rate": 9.491106211771765e-06,
"loss": 0.2915,
"step": 186
},
{
"epoch": 0.29127725856697817,
"grad_norm": 0.705003606824088,
"learning_rate": 9.485715566186333e-06,
"loss": 0.2945,
"step": 187
},
{
"epoch": 0.29283489096573206,
"grad_norm": 1.045178016891709,
"learning_rate": 9.480298067081422e-06,
"loss": 0.5356,
"step": 188
},
{
"epoch": 0.29439252336448596,
"grad_norm": 0.6198122538394608,
"learning_rate": 9.474853746888631e-06,
"loss": 0.2744,
"step": 189
},
{
"epoch": 0.29595015576323985,
"grad_norm": 0.631725313682397,
"learning_rate": 9.469382638200119e-06,
"loss": 0.3064,
"step": 190
},
{
"epoch": 0.29750778816199375,
"grad_norm": 0.49405112326099454,
"learning_rate": 9.463884773768413e-06,
"loss": 0.2016,
"step": 191
},
{
"epoch": 0.29906542056074764,
"grad_norm": 0.9255242318461628,
"learning_rate": 9.458360186506212e-06,
"loss": 0.318,
"step": 192
},
{
"epoch": 0.30062305295950154,
"grad_norm": 0.5172541239349757,
"learning_rate": 9.452808909486195e-06,
"loss": 0.2182,
"step": 193
},
{
"epoch": 0.30218068535825543,
"grad_norm": 0.6548225129226617,
"learning_rate": 9.447230975940808e-06,
"loss": 0.3501,
"step": 194
},
{
"epoch": 0.3037383177570093,
"grad_norm": 0.6373263585258429,
"learning_rate": 9.441626419262084e-06,
"loss": 0.2331,
"step": 195
},
{
"epoch": 0.3052959501557632,
"grad_norm": 0.8451103465765266,
"learning_rate": 9.43599527300143e-06,
"loss": 0.2379,
"step": 196
},
{
"epoch": 0.3068535825545171,
"grad_norm": 0.6257655537331657,
"learning_rate": 9.430337570869432e-06,
"loss": 0.262,
"step": 197
},
{
"epoch": 0.308411214953271,
"grad_norm": 0.7885047351755374,
"learning_rate": 9.424653346735649e-06,
"loss": 0.2783,
"step": 198
},
{
"epoch": 0.3099688473520249,
"grad_norm": 1.2716968661440529,
"learning_rate": 9.418942634628413e-06,
"loss": 0.3266,
"step": 199
},
{
"epoch": 0.3115264797507788,
"grad_norm": 0.709734211796566,
"learning_rate": 9.413205468734628e-06,
"loss": 0.298,
"step": 200
},
{
"epoch": 0.3130841121495327,
"grad_norm": 0.6966111704003821,
"learning_rate": 9.40744188339956e-06,
"loss": 0.3872,
"step": 201
},
{
"epoch": 0.3146417445482866,
"grad_norm": 0.6328947624746845,
"learning_rate": 9.401651913126634e-06,
"loss": 0.2855,
"step": 202
},
{
"epoch": 0.3161993769470405,
"grad_norm": 0.6782499058815041,
"learning_rate": 9.395835592577228e-06,
"loss": 0.2555,
"step": 203
},
{
"epoch": 0.3177570093457944,
"grad_norm": 0.6111146053952768,
"learning_rate": 9.389992956570463e-06,
"loss": 0.2119,
"step": 204
},
{
"epoch": 0.31931464174454827,
"grad_norm": 2.1014154014060646,
"learning_rate": 9.384124040082996e-06,
"loss": 1.3393,
"step": 205
},
{
"epoch": 0.32087227414330216,
"grad_norm": 0.6937660819701476,
"learning_rate": 9.378228878248814e-06,
"loss": 0.3003,
"step": 206
},
{
"epoch": 0.32242990654205606,
"grad_norm": 0.7798606308310687,
"learning_rate": 9.372307506359019e-06,
"loss": 0.2394,
"step": 207
},
{
"epoch": 0.32398753894080995,
"grad_norm": 0.8064220607757047,
"learning_rate": 9.366359959861615e-06,
"loss": 0.2457,
"step": 208
},
{
"epoch": 0.32554517133956384,
"grad_norm": 0.8663909019847741,
"learning_rate": 9.360386274361305e-06,
"loss": 0.2363,
"step": 209
},
{
"epoch": 0.32710280373831774,
"grad_norm": 0.7765065772183143,
"learning_rate": 9.354386485619264e-06,
"loss": 0.2519,
"step": 210
},
{
"epoch": 0.32866043613707163,
"grad_norm": 0.6924828379419826,
"learning_rate": 9.348360629552941e-06,
"loss": 0.2237,
"step": 211
},
{
"epoch": 0.3302180685358255,
"grad_norm": 0.6564177523605338,
"learning_rate": 9.342308742235831e-06,
"loss": 0.2369,
"step": 212
},
{
"epoch": 0.3317757009345794,
"grad_norm": 0.6166208484170821,
"learning_rate": 9.336230859897266e-06,
"loss": 0.2531,
"step": 213
},
{
"epoch": 0.3333333333333333,
"grad_norm": 0.6325656097098429,
"learning_rate": 9.330127018922195e-06,
"loss": 0.228,
"step": 214
},
{
"epoch": 0.3348909657320872,
"grad_norm": 0.9005060609426595,
"learning_rate": 9.323997255850965e-06,
"loss": 0.2068,
"step": 215
},
{
"epoch": 0.3364485981308411,
"grad_norm": 0.6030970162930338,
"learning_rate": 9.317841607379106e-06,
"loss": 0.2269,
"step": 216
},
{
"epoch": 0.338006230529595,
"grad_norm": 0.5757510215326483,
"learning_rate": 9.311660110357116e-06,
"loss": 0.2277,
"step": 217
},
{
"epoch": 0.3395638629283489,
"grad_norm": 0.48332154427815277,
"learning_rate": 9.30545280179022e-06,
"loss": 0.2088,
"step": 218
},
{
"epoch": 0.3411214953271028,
"grad_norm": 1.6587632436476218,
"learning_rate": 9.299219718838174e-06,
"loss": 0.8341,
"step": 219
},
{
"epoch": 0.3426791277258567,
"grad_norm": 0.9279443473544912,
"learning_rate": 9.292960898815032e-06,
"loss": 0.2415,
"step": 220
},
{
"epoch": 0.3442367601246106,
"grad_norm": 1.6924178170865158,
"learning_rate": 9.286676379188913e-06,
"loss": 0.9856,
"step": 221
},
{
"epoch": 0.34579439252336447,
"grad_norm": 1.8214244233005599,
"learning_rate": 9.280366197581792e-06,
"loss": 0.826,
"step": 222
},
{
"epoch": 0.34735202492211836,
"grad_norm": 0.783390501917877,
"learning_rate": 9.274030391769264e-06,
"loss": 0.8049,
"step": 223
},
{
"epoch": 0.34890965732087226,
"grad_norm": 0.5859949211179467,
"learning_rate": 9.267668999680328e-06,
"loss": 0.2939,
"step": 224
},
{
"epoch": 0.35046728971962615,
"grad_norm": 0.7610274820950346,
"learning_rate": 9.261282059397145e-06,
"loss": 0.2581,
"step": 225
},
{
"epoch": 0.35202492211838005,
"grad_norm": 0.6066786542121007,
"learning_rate": 9.25486960915483e-06,
"loss": 0.2758,
"step": 226
},
{
"epoch": 0.35358255451713394,
"grad_norm": 1.1073512044256302,
"learning_rate": 9.248431687341203e-06,
"loss": 0.237,
"step": 227
},
{
"epoch": 0.35514018691588783,
"grad_norm": 1.0141799395602682,
"learning_rate": 9.241968332496576e-06,
"loss": 0.2568,
"step": 228
},
{
"epoch": 0.35669781931464173,
"grad_norm": 0.5568072088382595,
"learning_rate": 9.235479583313504e-06,
"loss": 0.2334,
"step": 229
},
{
"epoch": 0.3582554517133956,
"grad_norm": 0.7930065901392893,
"learning_rate": 9.228965478636575e-06,
"loss": 0.2793,
"step": 230
},
{
"epoch": 0.3598130841121495,
"grad_norm": 0.7048596889355911,
"learning_rate": 9.222426057462162e-06,
"loss": 0.3062,
"step": 231
},
{
"epoch": 0.3613707165109034,
"grad_norm": 0.8914753088496622,
"learning_rate": 9.215861358938191e-06,
"loss": 0.3289,
"step": 232
},
{
"epoch": 0.3629283489096573,
"grad_norm": 0.8233950771824091,
"learning_rate": 9.209271422363915e-06,
"loss": 0.3021,
"step": 233
},
{
"epoch": 0.3644859813084112,
"grad_norm": 0.8572614177235147,
"learning_rate": 9.20265628718967e-06,
"loss": 0.3435,
"step": 234
},
{
"epoch": 0.3660436137071651,
"grad_norm": 0.6533204687228193,
"learning_rate": 9.196015993016642e-06,
"loss": 0.2537,
"step": 235
},
{
"epoch": 0.367601246105919,
"grad_norm": 0.676657852907951,
"learning_rate": 9.189350579596635e-06,
"loss": 0.273,
"step": 236
},
{
"epoch": 0.3691588785046729,
"grad_norm": 0.8090061594796749,
"learning_rate": 9.182660086831819e-06,
"loss": 0.3786,
"step": 237
},
{
"epoch": 0.3707165109034268,
"grad_norm": 0.6195251333212162,
"learning_rate": 9.175944554774507e-06,
"loss": 0.2799,
"step": 238
},
{
"epoch": 0.37227414330218067,
"grad_norm": 0.5501170288605118,
"learning_rate": 9.169204023626911e-06,
"loss": 0.2577,
"step": 239
},
{
"epoch": 0.37383177570093457,
"grad_norm": 0.6087726744520703,
"learning_rate": 9.162438533740891e-06,
"loss": 0.2578,
"step": 240
},
{
"epoch": 0.37538940809968846,
"grad_norm": 0.6423864822269919,
"learning_rate": 9.15564812561773e-06,
"loss": 0.2865,
"step": 241
},
{
"epoch": 0.37694704049844235,
"grad_norm": 0.6443057475504381,
"learning_rate": 9.148832839907871e-06,
"loss": 0.2898,
"step": 242
},
{
"epoch": 0.37850467289719625,
"grad_norm": 0.5688149258379418,
"learning_rate": 9.141992717410697e-06,
"loss": 0.239,
"step": 243
},
{
"epoch": 0.38006230529595014,
"grad_norm": 0.6629923621745606,
"learning_rate": 9.135127799074271e-06,
"loss": 0.2614,
"step": 244
},
{
"epoch": 0.38161993769470404,
"grad_norm": 0.6300463035343699,
"learning_rate": 9.128238125995093e-06,
"loss": 0.24,
"step": 245
},
{
"epoch": 0.38317757009345793,
"grad_norm": 0.643027164394109,
"learning_rate": 9.121323739417858e-06,
"loss": 0.2365,
"step": 246
},
{
"epoch": 0.3847352024922118,
"grad_norm": 2.8455942968073282,
"learning_rate": 9.114384680735211e-06,
"loss": 0.7814,
"step": 247
},
{
"epoch": 0.3862928348909657,
"grad_norm": 0.630104563255545,
"learning_rate": 9.107420991487488e-06,
"loss": 0.2234,
"step": 248
},
{
"epoch": 0.3878504672897196,
"grad_norm": 0.905027542232206,
"learning_rate": 9.100432713362477e-06,
"loss": 0.3142,
"step": 249
},
{
"epoch": 0.3894080996884735,
"grad_norm": 0.6960898261331736,
"learning_rate": 9.09341988819517e-06,
"loss": 0.3195,
"step": 250
},
{
"epoch": 0.3909657320872274,
"grad_norm": 0.7350311907562319,
"learning_rate": 9.086382557967507e-06,
"loss": 0.2479,
"step": 251
},
{
"epoch": 0.3925233644859813,
"grad_norm": 0.7212085527913289,
"learning_rate": 9.07932076480812e-06,
"loss": 0.2334,
"step": 252
},
{
"epoch": 0.3940809968847352,
"grad_norm": 0.6340762358232386,
"learning_rate": 9.072234550992099e-06,
"loss": 0.282,
"step": 253
},
{
"epoch": 0.3956386292834891,
"grad_norm": 0.5688729860657064,
"learning_rate": 9.065123958940716e-06,
"loss": 0.2928,
"step": 254
},
{
"epoch": 0.397196261682243,
"grad_norm": 0.8922598360101682,
"learning_rate": 9.057989031221187e-06,
"loss": 0.281,
"step": 255
},
{
"epoch": 0.3987538940809969,
"grad_norm": 0.7114134585553089,
"learning_rate": 9.050829810546416e-06,
"loss": 0.2635,
"step": 256
},
{
"epoch": 0.40031152647975077,
"grad_norm": 0.6966492509810489,
"learning_rate": 9.043646339774726e-06,
"loss": 0.3394,
"step": 257
},
{
"epoch": 0.40186915887850466,
"grad_norm": 0.5958512506208158,
"learning_rate": 9.036438661909624e-06,
"loss": 0.2718,
"step": 258
},
{
"epoch": 0.40342679127725856,
"grad_norm": 0.9116531361266647,
"learning_rate": 9.029206820099518e-06,
"loss": 0.2785,
"step": 259
},
{
"epoch": 0.40498442367601245,
"grad_norm": 0.5887944110529719,
"learning_rate": 9.021950857637486e-06,
"loss": 0.2668,
"step": 260
},
{
"epoch": 0.40654205607476634,
"grad_norm": 0.6108632031180715,
"learning_rate": 9.014670817960993e-06,
"loss": 0.1969,
"step": 261
},
{
"epoch": 0.40809968847352024,
"grad_norm": 0.6907915569936867,
"learning_rate": 9.007366744651646e-06,
"loss": 0.2268,
"step": 262
},
{
"epoch": 0.40965732087227413,
"grad_norm": 0.5793225985517833,
"learning_rate": 9.000038681434925e-06,
"loss": 0.1931,
"step": 263
},
{
"epoch": 0.411214953271028,
"grad_norm": 1.0784967522591073,
"learning_rate": 8.99268667217993e-06,
"loss": 0.2329,
"step": 264
},
{
"epoch": 0.4127725856697819,
"grad_norm": 0.602627598917929,
"learning_rate": 8.985310760899108e-06,
"loss": 0.2574,
"step": 265
},
{
"epoch": 0.4143302180685358,
"grad_norm": 0.6724448335681737,
"learning_rate": 8.977910991747993e-06,
"loss": 0.2781,
"step": 266
},
{
"epoch": 0.4158878504672897,
"grad_norm": 0.713294514234468,
"learning_rate": 8.970487409024949e-06,
"loss": 0.2284,
"step": 267
},
{
"epoch": 0.4174454828660436,
"grad_norm": 0.8427860833648895,
"learning_rate": 8.96304005717089e-06,
"loss": 0.2953,
"step": 268
},
{
"epoch": 0.4190031152647975,
"grad_norm": 0.5840913639256132,
"learning_rate": 8.955568980769033e-06,
"loss": 0.2974,
"step": 269
},
{
"epoch": 0.4205607476635514,
"grad_norm": 0.6131987426518342,
"learning_rate": 8.948074224544615e-06,
"loss": 0.2773,
"step": 270
},
{
"epoch": 0.4221183800623053,
"grad_norm": 0.5947515631497642,
"learning_rate": 8.94055583336463e-06,
"loss": 0.275,
"step": 271
},
{
"epoch": 0.4236760124610592,
"grad_norm": 0.5946955459161034,
"learning_rate": 8.933013852237564e-06,
"loss": 0.3307,
"step": 272
},
{
"epoch": 0.4252336448598131,
"grad_norm": 0.7451920245650221,
"learning_rate": 8.925448326313125e-06,
"loss": 0.286,
"step": 273
},
{
"epoch": 0.42679127725856697,
"grad_norm": 0.924856917547545,
"learning_rate": 8.917859300881965e-06,
"loss": 0.2879,
"step": 274
},
{
"epoch": 0.42834890965732086,
"grad_norm": 1.0327909293428048,
"learning_rate": 8.91024682137542e-06,
"loss": 0.2408,
"step": 275
},
{
"epoch": 0.42990654205607476,
"grad_norm": 0.880531058290941,
"learning_rate": 8.90261093336523e-06,
"loss": 0.3045,
"step": 276
},
{
"epoch": 0.43146417445482865,
"grad_norm": 0.580559025082547,
"learning_rate": 8.89495168256327e-06,
"loss": 0.2805,
"step": 277
},
{
"epoch": 0.43302180685358255,
"grad_norm": 0.5475610266945151,
"learning_rate": 8.887269114821275e-06,
"loss": 0.2338,
"step": 278
},
{
"epoch": 0.43457943925233644,
"grad_norm": 0.6043881248930308,
"learning_rate": 8.879563276130567e-06,
"loss": 0.2516,
"step": 279
},
{
"epoch": 0.43613707165109034,
"grad_norm": 0.6621273162261064,
"learning_rate": 8.871834212621773e-06,
"loss": 0.2485,
"step": 280
},
{
"epoch": 0.43769470404984423,
"grad_norm": 0.6108881149273678,
"learning_rate": 8.86408197056456e-06,
"loss": 0.2082,
"step": 281
},
{
"epoch": 0.4392523364485981,
"grad_norm": 0.6626887696196829,
"learning_rate": 8.856306596367351e-06,
"loss": 0.2741,
"step": 282
},
{
"epoch": 0.440809968847352,
"grad_norm": 0.6517640250027363,
"learning_rate": 8.84850813657705e-06,
"loss": 0.3733,
"step": 283
},
{
"epoch": 0.4423676012461059,
"grad_norm": 0.5979891922361272,
"learning_rate": 8.840686637878756e-06,
"loss": 0.2615,
"step": 284
},
{
"epoch": 0.4439252336448598,
"grad_norm": 0.5447033448285953,
"learning_rate": 8.832842147095495e-06,
"loss": 0.2204,
"step": 285
},
{
"epoch": 0.4454828660436137,
"grad_norm": 0.6176926760203407,
"learning_rate": 8.82497471118793e-06,
"loss": 0.259,
"step": 286
},
{
"epoch": 0.4470404984423676,
"grad_norm": 0.6803871655362536,
"learning_rate": 8.817084377254089e-06,
"loss": 0.3406,
"step": 287
},
{
"epoch": 0.4485981308411215,
"grad_norm": 1.0924030353117906,
"learning_rate": 8.809171192529074e-06,
"loss": 0.7712,
"step": 288
},
{
"epoch": 0.4501557632398754,
"grad_norm": 0.7261898907998104,
"learning_rate": 8.80123520438478e-06,
"loss": 0.2439,
"step": 289
},
{
"epoch": 0.4517133956386293,
"grad_norm": 0.5278888030187282,
"learning_rate": 8.79327646032962e-06,
"loss": 0.2838,
"step": 290
},
{
"epoch": 0.4532710280373832,
"grad_norm": 0.7336946619047033,
"learning_rate": 8.785295008008227e-06,
"loss": 0.2863,
"step": 291
},
{
"epoch": 0.45482866043613707,
"grad_norm": 0.5663680560662904,
"learning_rate": 8.777290895201186e-06,
"loss": 0.2487,
"step": 292
},
{
"epoch": 0.45638629283489096,
"grad_norm": 0.6674793012574268,
"learning_rate": 8.769264169824725e-06,
"loss": 0.2115,
"step": 293
},
{
"epoch": 0.45794392523364486,
"grad_norm": 0.629476332991769,
"learning_rate": 8.761214879930452e-06,
"loss": 0.2698,
"step": 294
},
{
"epoch": 0.45950155763239875,
"grad_norm": 0.5733049643694853,
"learning_rate": 8.753143073705048e-06,
"loss": 0.2755,
"step": 295
},
{
"epoch": 0.46105919003115264,
"grad_norm": 0.7496857759281685,
"learning_rate": 8.745048799469996e-06,
"loss": 0.2066,
"step": 296
},
{
"epoch": 0.46261682242990654,
"grad_norm": 0.5833814287468523,
"learning_rate": 8.736932105681272e-06,
"loss": 0.2233,
"step": 297
},
{
"epoch": 0.46417445482866043,
"grad_norm": 0.7333726802931715,
"learning_rate": 8.728793040929075e-06,
"loss": 0.2529,
"step": 298
},
{
"epoch": 0.4657320872274143,
"grad_norm": 0.6752352954357181,
"learning_rate": 8.720631653937522e-06,
"loss": 0.2487,
"step": 299
},
{
"epoch": 0.4672897196261682,
"grad_norm": 0.6281610479983413,
"learning_rate": 8.712447993564362e-06,
"loss": 0.2513,
"step": 300
},
{
"epoch": 0.4688473520249221,
"grad_norm": 1.719855082212449,
"learning_rate": 8.70424210880068e-06,
"loss": 1.2841,
"step": 301
},
{
"epoch": 0.470404984423676,
"grad_norm": 0.8697991461292468,
"learning_rate": 8.696014048770611e-06,
"loss": 0.1752,
"step": 302
},
{
"epoch": 0.4719626168224299,
"grad_norm": 0.7388848487578813,
"learning_rate": 8.68776386273104e-06,
"loss": 0.3559,
"step": 303
},
{
"epoch": 0.4735202492211838,
"grad_norm": 0.6740217209253545,
"learning_rate": 8.679491600071304e-06,
"loss": 0.2506,
"step": 304
},
{
"epoch": 0.4750778816199377,
"grad_norm": 0.8324393939489836,
"learning_rate": 8.671197310312905e-06,
"loss": 0.2951,
"step": 305
},
{
"epoch": 0.4766355140186916,
"grad_norm": 0.5859142473050233,
"learning_rate": 8.662881043109208e-06,
"loss": 0.2737,
"step": 306
},
{
"epoch": 0.4781931464174455,
"grad_norm": 0.5144143552921643,
"learning_rate": 8.654542848245146e-06,
"loss": 0.2555,
"step": 307
},
{
"epoch": 0.4797507788161994,
"grad_norm": 0.9870060592323339,
"learning_rate": 8.646182775636917e-06,
"loss": 0.3268,
"step": 308
},
{
"epoch": 0.48130841121495327,
"grad_norm": 1.22444825315241,
"learning_rate": 8.637800875331693e-06,
"loss": 0.8927,
"step": 309
},
{
"epoch": 0.48286604361370716,
"grad_norm": 0.6219488136095235,
"learning_rate": 8.629397197507315e-06,
"loss": 0.1925,
"step": 310
},
{
"epoch": 0.48442367601246106,
"grad_norm": 0.6663733783513853,
"learning_rate": 8.620971792471994e-06,
"loss": 0.2525,
"step": 311
},
{
"epoch": 0.48598130841121495,
"grad_norm": 0.7779309034351384,
"learning_rate": 8.612524710664012e-06,
"loss": 0.2857,
"step": 312
},
{
"epoch": 0.48753894080996885,
"grad_norm": 0.5815264215390993,
"learning_rate": 8.604056002651414e-06,
"loss": 0.2132,
"step": 313
},
{
"epoch": 0.48909657320872274,
"grad_norm": 0.639985193258881,
"learning_rate": 8.595565719131711e-06,
"loss": 0.2927,
"step": 314
},
{
"epoch": 0.49065420560747663,
"grad_norm": 0.5843630349664094,
"learning_rate": 8.587053910931576e-06,
"loss": 0.2073,
"step": 315
},
{
"epoch": 0.49221183800623053,
"grad_norm": 0.5718073343739106,
"learning_rate": 8.578520629006537e-06,
"loss": 0.2701,
"step": 316
},
{
"epoch": 0.4937694704049844,
"grad_norm": 0.8093573764947887,
"learning_rate": 8.569965924440675e-06,
"loss": 0.3701,
"step": 317
},
{
"epoch": 0.4953271028037383,
"grad_norm": 0.45875585137104996,
"learning_rate": 8.561389848446314e-06,
"loss": 0.2535,
"step": 318
},
{
"epoch": 0.4968847352024922,
"grad_norm": 0.6800533406672229,
"learning_rate": 8.55279245236372e-06,
"loss": 0.2542,
"step": 319
},
{
"epoch": 0.4984423676012461,
"grad_norm": 0.5791491500061517,
"learning_rate": 8.544173787660788e-06,
"loss": 0.2233,
"step": 320
},
{
"epoch": 0.5,
"grad_norm": 0.6853895685598861,
"learning_rate": 8.535533905932739e-06,
"loss": 0.1762,
"step": 321
},
{
"epoch": 0.5015576323987538,
"grad_norm": 0.5400182911842227,
"learning_rate": 8.526872858901806e-06,
"loss": 0.2565,
"step": 322
},
{
"epoch": 0.5031152647975078,
"grad_norm": 0.7099762938056243,
"learning_rate": 8.518190698416929e-06,
"loss": 0.2901,
"step": 323
},
{
"epoch": 0.5046728971962616,
"grad_norm": 0.7351812911105055,
"learning_rate": 8.509487476453442e-06,
"loss": 0.2829,
"step": 324
},
{
"epoch": 0.5062305295950156,
"grad_norm": 0.6208369210464704,
"learning_rate": 8.500763245112763e-06,
"loss": 0.2916,
"step": 325
},
{
"epoch": 0.5077881619937694,
"grad_norm": 0.7390752110429559,
"learning_rate": 8.492018056622083e-06,
"loss": 0.3912,
"step": 326
},
{
"epoch": 0.5093457943925234,
"grad_norm": 0.5930641088378493,
"learning_rate": 8.483251963334047e-06,
"loss": 0.2193,
"step": 327
},
{
"epoch": 0.5109034267912772,
"grad_norm": 0.6344017285315496,
"learning_rate": 8.474465017726452e-06,
"loss": 0.2191,
"step": 328
},
{
"epoch": 0.5124610591900312,
"grad_norm": 0.7250776947740573,
"learning_rate": 8.465657272401921e-06,
"loss": 0.2435,
"step": 329
},
{
"epoch": 0.514018691588785,
"grad_norm": 0.6175903219629659,
"learning_rate": 8.456828780087598e-06,
"loss": 0.2005,
"step": 330
},
{
"epoch": 0.5155763239875389,
"grad_norm": 0.6186407482326058,
"learning_rate": 8.447979593634823e-06,
"loss": 0.2969,
"step": 331
},
{
"epoch": 0.5171339563862928,
"grad_norm": 0.6221498367824214,
"learning_rate": 8.439109766018825e-06,
"loss": 0.2532,
"step": 332
},
{
"epoch": 0.5186915887850467,
"grad_norm": 0.5991537960888588,
"learning_rate": 8.430219350338398e-06,
"loss": 0.2029,
"step": 333
},
{
"epoch": 0.5202492211838006,
"grad_norm": 0.5074549508649242,
"learning_rate": 8.421308399815586e-06,
"loss": 0.2452,
"step": 334
},
{
"epoch": 0.5218068535825545,
"grad_norm": 1.0447367430008974,
"learning_rate": 8.412376967795362e-06,
"loss": 0.7821,
"step": 335
},
{
"epoch": 0.5233644859813084,
"grad_norm": 0.6715190449661054,
"learning_rate": 8.403425107745315e-06,
"loss": 0.1873,
"step": 336
},
{
"epoch": 0.5249221183800623,
"grad_norm": 0.8046073365053287,
"learning_rate": 8.394452873255321e-06,
"loss": 0.2459,
"step": 337
},
{
"epoch": 0.5264797507788161,
"grad_norm": 0.6185327399453182,
"learning_rate": 8.385460318037228e-06,
"loss": 0.2347,
"step": 338
},
{
"epoch": 0.5280373831775701,
"grad_norm": 0.6523073370428902,
"learning_rate": 8.376447495924533e-06,
"loss": 0.2494,
"step": 339
},
{
"epoch": 0.5295950155763239,
"grad_norm": 0.8230020627407991,
"learning_rate": 8.367414460872064e-06,
"loss": 0.7898,
"step": 340
},
{
"epoch": 0.5311526479750779,
"grad_norm": 0.6739918439315772,
"learning_rate": 8.358361266955641e-06,
"loss": 0.3159,
"step": 341
},
{
"epoch": 0.5327102803738317,
"grad_norm": 0.7519773686193855,
"learning_rate": 8.34928796837178e-06,
"loss": 0.2593,
"step": 342
},
{
"epoch": 0.5342679127725857,
"grad_norm": 0.5445971280446563,
"learning_rate": 8.34019461943734e-06,
"loss": 0.2923,
"step": 343
},
{
"epoch": 0.5358255451713395,
"grad_norm": 0.7071155258569332,
"learning_rate": 8.331081274589217e-06,
"loss": 0.2598,
"step": 344
},
{
"epoch": 0.5373831775700935,
"grad_norm": 0.719502132159577,
"learning_rate": 8.321947988384006e-06,
"loss": 0.276,
"step": 345
},
{
"epoch": 0.5389408099688473,
"grad_norm": 0.5170866766320925,
"learning_rate": 8.312794815497688e-06,
"loss": 0.2236,
"step": 346
},
{
"epoch": 0.5404984423676013,
"grad_norm": 0.6654228108717598,
"learning_rate": 8.303621810725287e-06,
"loss": 0.2666,
"step": 347
},
{
"epoch": 0.5420560747663551,
"grad_norm": 0.6117403127181333,
"learning_rate": 8.294429028980555e-06,
"loss": 0.2767,
"step": 348
},
{
"epoch": 0.543613707165109,
"grad_norm": 0.8584300691210696,
"learning_rate": 8.285216525295636e-06,
"loss": 0.2521,
"step": 349
},
{
"epoch": 0.5451713395638629,
"grad_norm": 0.7222215782352946,
"learning_rate": 8.275984354820736e-06,
"loss": 0.2499,
"step": 350
},
{
"epoch": 0.5467289719626168,
"grad_norm": 0.643947565152816,
"learning_rate": 8.266732572823799e-06,
"loss": 0.2092,
"step": 351
},
{
"epoch": 0.5482866043613707,
"grad_norm": 0.6231517894213087,
"learning_rate": 8.25746123469017e-06,
"loss": 0.2598,
"step": 352
},
{
"epoch": 0.5498442367601246,
"grad_norm": 0.5983797307315736,
"learning_rate": 8.248170395922266e-06,
"loss": 0.3241,
"step": 353
},
{
"epoch": 0.5514018691588785,
"grad_norm": 0.6611331737862588,
"learning_rate": 8.238860112139246e-06,
"loss": 0.2292,
"step": 354
},
{
"epoch": 0.5529595015576324,
"grad_norm": 0.667013942050513,
"learning_rate": 8.229530439076674e-06,
"loss": 0.2911,
"step": 355
},
{
"epoch": 0.5545171339563862,
"grad_norm": 0.5711025724816864,
"learning_rate": 8.220181432586187e-06,
"loss": 0.3828,
"step": 356
},
{
"epoch": 0.5560747663551402,
"grad_norm": 0.7326983314478233,
"learning_rate": 8.210813148635158e-06,
"loss": 0.4075,
"step": 357
},
{
"epoch": 0.557632398753894,
"grad_norm": 0.5806269356738779,
"learning_rate": 8.201425643306367e-06,
"loss": 0.2283,
"step": 358
},
{
"epoch": 0.559190031152648,
"grad_norm": 0.8188821554767335,
"learning_rate": 8.192018972797665e-06,
"loss": 0.2549,
"step": 359
},
{
"epoch": 0.5607476635514018,
"grad_norm": 0.6809831648038962,
"learning_rate": 8.182593193421625e-06,
"loss": 0.2543,
"step": 360
},
{
"epoch": 0.5623052959501558,
"grad_norm": 0.6203982930897892,
"learning_rate": 8.173148361605224e-06,
"loss": 0.3205,
"step": 361
},
{
"epoch": 0.5638629283489096,
"grad_norm": 0.4686220806935477,
"learning_rate": 8.163684533889489e-06,
"loss": 0.2356,
"step": 362
},
{
"epoch": 0.5654205607476636,
"grad_norm": 0.7261908034014758,
"learning_rate": 8.154201766929167e-06,
"loss": 0.2763,
"step": 363
},
{
"epoch": 0.5669781931464174,
"grad_norm": 0.6301811228833201,
"learning_rate": 8.144700117492386e-06,
"loss": 0.2369,
"step": 364
},
{
"epoch": 0.5685358255451713,
"grad_norm": 0.7734460446532929,
"learning_rate": 8.135179642460308e-06,
"loss": 0.2722,
"step": 365
},
{
"epoch": 0.5700934579439252,
"grad_norm": 0.6997558366317673,
"learning_rate": 8.125640398826803e-06,
"loss": 0.2543,
"step": 366
},
{
"epoch": 0.5716510903426791,
"grad_norm": 0.8988136993440748,
"learning_rate": 8.116082443698085e-06,
"loss": 0.2835,
"step": 367
},
{
"epoch": 0.573208722741433,
"grad_norm": 0.8141909161490042,
"learning_rate": 8.106505834292396e-06,
"loss": 0.2538,
"step": 368
},
{
"epoch": 0.5747663551401869,
"grad_norm": 1.3911734406138567,
"learning_rate": 8.09691062793964e-06,
"loss": 0.7141,
"step": 369
},
{
"epoch": 0.5763239875389408,
"grad_norm": 0.6359597508417911,
"learning_rate": 8.087296882081062e-06,
"loss": 0.2794,
"step": 370
},
{
"epoch": 0.5778816199376947,
"grad_norm": 0.5841003451023182,
"learning_rate": 8.077664654268883e-06,
"loss": 0.2973,
"step": 371
},
{
"epoch": 0.5794392523364486,
"grad_norm": 0.6441301027816297,
"learning_rate": 8.06801400216597e-06,
"loss": 0.2477,
"step": 372
},
{
"epoch": 0.5809968847352025,
"grad_norm": 0.6922555735413835,
"learning_rate": 8.058344983545486e-06,
"loss": 0.2524,
"step": 373
},
{
"epoch": 0.5825545171339563,
"grad_norm": 0.6083842427437397,
"learning_rate": 8.048657656290545e-06,
"loss": 0.2837,
"step": 374
},
{
"epoch": 0.5841121495327103,
"grad_norm": 0.5803164555571954,
"learning_rate": 8.03895207839386e-06,
"loss": 0.3084,
"step": 375
},
{
"epoch": 0.5856697819314641,
"grad_norm": 0.8102916914102428,
"learning_rate": 8.029228307957408e-06,
"loss": 0.2462,
"step": 376
},
{
"epoch": 0.5872274143302181,
"grad_norm": 0.7061482090328073,
"learning_rate": 8.019486403192069e-06,
"loss": 0.2487,
"step": 377
},
{
"epoch": 0.5887850467289719,
"grad_norm": 0.5113763283575798,
"learning_rate": 8.009726422417286e-06,
"loss": 0.2931,
"step": 378
},
{
"epoch": 0.5903426791277259,
"grad_norm": 0.6897001371312232,
"learning_rate": 7.99994842406071e-06,
"loss": 0.2186,
"step": 379
},
{
"epoch": 0.5919003115264797,
"grad_norm": 0.9072221405351328,
"learning_rate": 7.99015246665786e-06,
"loss": 0.8348,
"step": 380
},
{
"epoch": 0.5934579439252337,
"grad_norm": 0.6531934878590409,
"learning_rate": 7.980338608851756e-06,
"loss": 0.2228,
"step": 381
},
{
"epoch": 0.5950155763239875,
"grad_norm": 0.7327462885118111,
"learning_rate": 7.970506909392588e-06,
"loss": 0.2433,
"step": 382
},
{
"epoch": 0.5965732087227414,
"grad_norm": 1.1643513095965004,
"learning_rate": 7.960657427137347e-06,
"loss": 0.3147,
"step": 383
},
{
"epoch": 0.5981308411214953,
"grad_norm": 0.6452843748972524,
"learning_rate": 7.950790221049485e-06,
"loss": 0.3303,
"step": 384
},
{
"epoch": 0.5996884735202492,
"grad_norm": 0.7274077390061954,
"learning_rate": 7.940905350198553e-06,
"loss": 0.3347,
"step": 385
},
{
"epoch": 0.6012461059190031,
"grad_norm": 0.6178112722471735,
"learning_rate": 7.931002873759852e-06,
"loss": 0.2276,
"step": 386
},
{
"epoch": 0.602803738317757,
"grad_norm": 0.7461711392201733,
"learning_rate": 7.921082851014079e-06,
"loss": 0.2634,
"step": 387
},
{
"epoch": 0.6043613707165109,
"grad_norm": 0.8795743223985983,
"learning_rate": 7.911145341346972e-06,
"loss": 0.2896,
"step": 388
},
{
"epoch": 0.6059190031152648,
"grad_norm": 0.8319649658997385,
"learning_rate": 7.901190404248952e-06,
"loss": 0.7066,
"step": 389
},
{
"epoch": 0.6074766355140186,
"grad_norm": 0.5267541152010679,
"learning_rate": 7.89121809931477e-06,
"loss": 0.2416,
"step": 390
},
{
"epoch": 0.6090342679127726,
"grad_norm": 0.46625213633043144,
"learning_rate": 7.881228486243144e-06,
"loss": 0.2482,
"step": 391
},
{
"epoch": 0.6105919003115264,
"grad_norm": 0.6924717742437231,
"learning_rate": 7.871221624836414e-06,
"loss": 0.3523,
"step": 392
},
{
"epoch": 0.6121495327102804,
"grad_norm": 0.5838406566254681,
"learning_rate": 7.861197575000168e-06,
"loss": 0.228,
"step": 393
},
{
"epoch": 0.6137071651090342,
"grad_norm": 0.8869116056189287,
"learning_rate": 7.8511563967429e-06,
"loss": 0.2496,
"step": 394
},
{
"epoch": 0.6152647975077882,
"grad_norm": 0.45593356473166835,
"learning_rate": 7.841098150175636e-06,
"loss": 0.2643,
"step": 395
},
{
"epoch": 0.616822429906542,
"grad_norm": 0.8435131653971922,
"learning_rate": 7.831022895511586e-06,
"loss": 0.2496,
"step": 396
},
{
"epoch": 0.618380062305296,
"grad_norm": 0.4987004271913618,
"learning_rate": 7.820930693065771e-06,
"loss": 0.2063,
"step": 397
},
{
"epoch": 0.6199376947040498,
"grad_norm": 0.7942215586211049,
"learning_rate": 7.810821603254677e-06,
"loss": 0.2181,
"step": 398
},
{
"epoch": 0.6214953271028038,
"grad_norm": 0.6794367411830937,
"learning_rate": 7.800695686595879e-06,
"loss": 0.3009,
"step": 399
},
{
"epoch": 0.6230529595015576,
"grad_norm": 0.7273130133089402,
"learning_rate": 7.790553003707691e-06,
"loss": 0.2469,
"step": 400
},
{
"epoch": 0.6246105919003115,
"grad_norm": 0.6902330948327355,
"learning_rate": 7.780393615308787e-06,
"loss": 0.2508,
"step": 401
},
{
"epoch": 0.6261682242990654,
"grad_norm": 0.7084362576430354,
"learning_rate": 7.770217582217863e-06,
"loss": 0.2551,
"step": 402
},
{
"epoch": 0.6277258566978193,
"grad_norm": 0.8029205384416798,
"learning_rate": 7.760024965353246e-06,
"loss": 0.2333,
"step": 403
},
{
"epoch": 0.6292834890965732,
"grad_norm": 0.6112029983652504,
"learning_rate": 7.749815825732543e-06,
"loss": 0.298,
"step": 404
},
{
"epoch": 0.6308411214953271,
"grad_norm": 0.7494581341489577,
"learning_rate": 7.739590224472275e-06,
"loss": 0.3462,
"step": 405
},
{
"epoch": 0.632398753894081,
"grad_norm": 0.7347669711126691,
"learning_rate": 7.729348222787514e-06,
"loss": 0.3149,
"step": 406
},
{
"epoch": 0.6339563862928349,
"grad_norm": 0.6796064286407987,
"learning_rate": 7.719089881991503e-06,
"loss": 0.2873,
"step": 407
},
{
"epoch": 0.6355140186915887,
"grad_norm": 0.7425509324765857,
"learning_rate": 7.708815263495307e-06,
"loss": 0.3278,
"step": 408
},
{
"epoch": 0.6370716510903427,
"grad_norm": 0.609414275478013,
"learning_rate": 7.698524428807431e-06,
"loss": 0.2708,
"step": 409
},
{
"epoch": 0.6386292834890965,
"grad_norm": 0.7757117977400942,
"learning_rate": 7.68821743953346e-06,
"loss": 0.2555,
"step": 410
},
{
"epoch": 0.6401869158878505,
"grad_norm": 0.6642687790623766,
"learning_rate": 7.677894357375689e-06,
"loss": 0.3625,
"step": 411
},
{
"epoch": 0.6417445482866043,
"grad_norm": 0.5791966784356082,
"learning_rate": 7.667555244132749e-06,
"loss": 0.2661,
"step": 412
},
{
"epoch": 0.6433021806853583,
"grad_norm": 0.5594732951892226,
"learning_rate": 7.65720016169924e-06,
"loss": 0.2995,
"step": 413
},
{
"epoch": 0.6448598130841121,
"grad_norm": 0.6021900759219545,
"learning_rate": 7.646829172065367e-06,
"loss": 0.3099,
"step": 414
},
{
"epoch": 0.6464174454828661,
"grad_norm": 0.5562483872284556,
"learning_rate": 7.636442337316555e-06,
"loss": 0.2376,
"step": 415
},
{
"epoch": 0.6479750778816199,
"grad_norm": 0.5829741964791303,
"learning_rate": 7.6260397196330895e-06,
"loss": 0.2774,
"step": 416
},
{
"epoch": 0.6495327102803738,
"grad_norm": 0.7958468559537486,
"learning_rate": 7.615621381289737e-06,
"loss": 0.2316,
"step": 417
},
{
"epoch": 0.6510903426791277,
"grad_norm": 0.6088648202059304,
"learning_rate": 7.6051873846553795e-06,
"loss": 0.33,
"step": 418
},
{
"epoch": 0.6526479750778816,
"grad_norm": 0.523142097250351,
"learning_rate": 7.594737792192629e-06,
"loss": 0.2589,
"step": 419
},
{
"epoch": 0.6542056074766355,
"grad_norm": 0.6233483869502079,
"learning_rate": 7.584272666457471e-06,
"loss": 0.2409,
"step": 420
},
{
"epoch": 0.6557632398753894,
"grad_norm": 0.8831745439168878,
"learning_rate": 7.573792070098873e-06,
"loss": 0.3156,
"step": 421
},
{
"epoch": 0.6573208722741433,
"grad_norm": 0.646939395915981,
"learning_rate": 7.5632960658584184e-06,
"loss": 0.1882,
"step": 422
},
{
"epoch": 0.6588785046728972,
"grad_norm": 0.7493677482681486,
"learning_rate": 7.5527847165699295e-06,
"loss": 0.2533,
"step": 423
},
{
"epoch": 0.660436137071651,
"grad_norm": 0.6895089900125264,
"learning_rate": 7.542258085159091e-06,
"loss": 0.2239,
"step": 424
},
{
"epoch": 0.661993769470405,
"grad_norm": 0.8218899831192643,
"learning_rate": 7.531716234643071e-06,
"loss": 0.3025,
"step": 425
},
{
"epoch": 0.6635514018691588,
"grad_norm": 0.8511691627825192,
"learning_rate": 7.5211592281301525e-06,
"loss": 0.2081,
"step": 426
},
{
"epoch": 0.6651090342679128,
"grad_norm": 0.6202374314769092,
"learning_rate": 7.510587128819341e-06,
"loss": 0.2159,
"step": 427
},
{
"epoch": 0.6666666666666666,
"grad_norm": 0.8637856332283039,
"learning_rate": 7.500000000000001e-06,
"loss": 0.6963,
"step": 428
},
{
"epoch": 0.6682242990654206,
"grad_norm": 0.7070722564579784,
"learning_rate": 7.489397905051465e-06,
"loss": 0.3265,
"step": 429
},
{
"epoch": 0.6697819314641744,
"grad_norm": 1.0912368137134154,
"learning_rate": 7.478780907442665e-06,
"loss": 0.3064,
"step": 430
},
{
"epoch": 0.6713395638629284,
"grad_norm": 0.4993170737898787,
"learning_rate": 7.468149070731742e-06,
"loss": 0.3532,
"step": 431
},
{
"epoch": 0.6728971962616822,
"grad_norm": 0.6200661171236782,
"learning_rate": 7.457502458565673e-06,
"loss": 0.2325,
"step": 432
},
{
"epoch": 0.6744548286604362,
"grad_norm": 0.6152606478186087,
"learning_rate": 7.446841134679888e-06,
"loss": 0.2538,
"step": 433
},
{
"epoch": 0.67601246105919,
"grad_norm": 0.5515844149625706,
"learning_rate": 7.436165162897886e-06,
"loss": 0.2619,
"step": 434
},
{
"epoch": 0.677570093457944,
"grad_norm": 0.7008609334925875,
"learning_rate": 7.425474607130858e-06,
"loss": 0.3168,
"step": 435
},
{
"epoch": 0.6791277258566978,
"grad_norm": 0.6379535213002501,
"learning_rate": 7.414769531377298e-06,
"loss": 0.268,
"step": 436
},
{
"epoch": 0.6806853582554517,
"grad_norm": 0.5317208819493666,
"learning_rate": 7.4040499997226245e-06,
"loss": 0.2193,
"step": 437
},
{
"epoch": 0.6822429906542056,
"grad_norm": 0.7290549365391932,
"learning_rate": 7.393316076338798e-06,
"loss": 0.3694,
"step": 438
},
{
"epoch": 0.6838006230529595,
"grad_norm": 0.8546888899097251,
"learning_rate": 7.382567825483929e-06,
"loss": 0.2822,
"step": 439
},
{
"epoch": 0.6853582554517134,
"grad_norm": 0.7390434139959143,
"learning_rate": 7.371805311501905e-06,
"loss": 0.24,
"step": 440
},
{
"epoch": 0.6869158878504673,
"grad_norm": 0.7021761813392882,
"learning_rate": 7.361028598821993e-06,
"loss": 0.3065,
"step": 441
},
{
"epoch": 0.6884735202492211,
"grad_norm": 0.5340954968447894,
"learning_rate": 7.350237751958466e-06,
"loss": 0.221,
"step": 442
},
{
"epoch": 0.6900311526479751,
"grad_norm": 0.5761626294301733,
"learning_rate": 7.339432835510203e-06,
"loss": 0.2345,
"step": 443
},
{
"epoch": 0.6915887850467289,
"grad_norm": 0.662018391928594,
"learning_rate": 7.328613914160319e-06,
"loss": 0.3171,
"step": 444
},
{
"epoch": 0.6931464174454829,
"grad_norm": 0.8130781056088618,
"learning_rate": 7.3177810526757594e-06,
"loss": 0.2909,
"step": 445
},
{
"epoch": 0.6947040498442367,
"grad_norm": 0.7219882547975953,
"learning_rate": 7.3069343159069296e-06,
"loss": 0.2481,
"step": 446
},
{
"epoch": 0.6962616822429907,
"grad_norm": 0.6369674341462834,
"learning_rate": 7.296073768787293e-06,
"loss": 0.3649,
"step": 447
},
{
"epoch": 0.6978193146417445,
"grad_norm": 0.7223244796104977,
"learning_rate": 7.285199476332991e-06,
"loss": 0.3488,
"step": 448
},
{
"epoch": 0.6993769470404985,
"grad_norm": 0.9091117254585579,
"learning_rate": 7.27431150364245e-06,
"loss": 0.3168,
"step": 449
},
{
"epoch": 0.7009345794392523,
"grad_norm": 0.7868966044967969,
"learning_rate": 7.263409915895992e-06,
"loss": 0.259,
"step": 450
},
{
"epoch": 0.7024922118380063,
"grad_norm": 0.6563385278402535,
"learning_rate": 7.252494778355444e-06,
"loss": 0.25,
"step": 451
},
{
"epoch": 0.7040498442367601,
"grad_norm": 0.641132207138942,
"learning_rate": 7.2415661563637506e-06,
"loss": 0.3307,
"step": 452
},
{
"epoch": 0.705607476635514,
"grad_norm": 0.7073578438725788,
"learning_rate": 7.23062411534458e-06,
"loss": 0.2261,
"step": 453
},
{
"epoch": 0.7071651090342679,
"grad_norm": 1.040324988179143,
"learning_rate": 7.2196687208019315e-06,
"loss": 0.2057,
"step": 454
},
{
"epoch": 0.7087227414330218,
"grad_norm": 0.6121914696936145,
"learning_rate": 7.208700038319744e-06,
"loss": 0.3199,
"step": 455
},
{
"epoch": 0.7102803738317757,
"grad_norm": 0.48700575482675645,
"learning_rate": 7.1977181335615085e-06,
"loss": 0.2259,
"step": 456
},
{
"epoch": 0.7118380062305296,
"grad_norm": 0.545525370035186,
"learning_rate": 7.186723072269863e-06,
"loss": 0.268,
"step": 457
},
{
"epoch": 0.7133956386292835,
"grad_norm": 0.846012722177333,
"learning_rate": 7.175714920266214e-06,
"loss": 0.7256,
"step": 458
},
{
"epoch": 0.7149532710280374,
"grad_norm": 0.5989442738821008,
"learning_rate": 7.164693743450329e-06,
"loss": 0.3005,
"step": 459
},
{
"epoch": 0.7165109034267912,
"grad_norm": 0.6556232944526054,
"learning_rate": 7.153659607799952e-06,
"loss": 0.2745,
"step": 460
},
{
"epoch": 0.7180685358255452,
"grad_norm": 0.6984028449665124,
"learning_rate": 7.142612579370402e-06,
"loss": 0.2272,
"step": 461
},
{
"epoch": 0.719626168224299,
"grad_norm": 0.5324547293774875,
"learning_rate": 7.131552724294181e-06,
"loss": 0.2518,
"step": 462
},
{
"epoch": 0.721183800623053,
"grad_norm": 0.5943008852162496,
"learning_rate": 7.1204801087805765e-06,
"loss": 0.2663,
"step": 463
},
{
"epoch": 0.7227414330218068,
"grad_norm": 0.5954388212811877,
"learning_rate": 7.109394799115268e-06,
"loss": 0.25,
"step": 464
},
{
"epoch": 0.7242990654205608,
"grad_norm": 0.5237590199785461,
"learning_rate": 7.098296861659925e-06,
"loss": 0.2451,
"step": 465
},
{
"epoch": 0.7258566978193146,
"grad_norm": 0.6108790141608955,
"learning_rate": 7.0871863628518136e-06,
"loss": 0.2782,
"step": 466
},
{
"epoch": 0.7274143302180686,
"grad_norm": 0.5784066614984076,
"learning_rate": 7.0760633692033975e-06,
"loss": 0.2588,
"step": 467
},
{
"epoch": 0.7289719626168224,
"grad_norm": 0.5736026805586273,
"learning_rate": 7.064927947301942e-06,
"loss": 0.3319,
"step": 468
},
{
"epoch": 0.7305295950155763,
"grad_norm": 0.660786534496975,
"learning_rate": 7.0537801638091116e-06,
"loss": 0.3207,
"step": 469
},
{
"epoch": 0.7320872274143302,
"grad_norm": 0.6164260678789174,
"learning_rate": 7.042620085460574e-06,
"loss": 0.2759,
"step": 470
},
{
"epoch": 0.7336448598130841,
"grad_norm": 1.0298833742062845,
"learning_rate": 7.0314477790656e-06,
"loss": 0.2769,
"step": 471
},
{
"epoch": 0.735202492211838,
"grad_norm": 0.7848249743313419,
"learning_rate": 7.020263311506659e-06,
"loss": 0.3963,
"step": 472
},
{
"epoch": 0.7367601246105919,
"grad_norm": 0.5488287365596327,
"learning_rate": 7.009066749739026e-06,
"loss": 0.2244,
"step": 473
},
{
"epoch": 0.7383177570093458,
"grad_norm": 0.5479634675942974,
"learning_rate": 6.99785816079038e-06,
"loss": 0.277,
"step": 474
},
{
"epoch": 0.7398753894080997,
"grad_norm": 0.632641842156797,
"learning_rate": 6.986637611760394e-06,
"loss": 0.2948,
"step": 475
},
{
"epoch": 0.7414330218068536,
"grad_norm": 0.8957280703439034,
"learning_rate": 6.975405169820344e-06,
"loss": 0.353,
"step": 476
},
{
"epoch": 0.7429906542056075,
"grad_norm": 0.575855558736389,
"learning_rate": 6.9641609022127e-06,
"loss": 0.2667,
"step": 477
},
{
"epoch": 0.7445482866043613,
"grad_norm": 0.6675031465700932,
"learning_rate": 6.952904876250729e-06,
"loss": 0.239,
"step": 478
},
{
"epoch": 0.7461059190031153,
"grad_norm": 0.5488652674770181,
"learning_rate": 6.941637159318083e-06,
"loss": 0.2605,
"step": 479
},
{
"epoch": 0.7476635514018691,
"grad_norm": 0.765674305969199,
"learning_rate": 6.9303578188684085e-06,
"loss": 0.2668,
"step": 480
},
{
"epoch": 0.7492211838006231,
"grad_norm": 0.6239515815918181,
"learning_rate": 6.919066922424931e-06,
"loss": 0.2883,
"step": 481
},
{
"epoch": 0.7507788161993769,
"grad_norm": 0.8869720139234101,
"learning_rate": 6.907764537580053e-06,
"loss": 0.2726,
"step": 482
},
{
"epoch": 0.7523364485981309,
"grad_norm": 0.7182796918869947,
"learning_rate": 6.896450731994959e-06,
"loss": 0.2575,
"step": 483
},
{
"epoch": 0.7538940809968847,
"grad_norm": 0.689714104473123,
"learning_rate": 6.8851255733992006e-06,
"loss": 0.2548,
"step": 484
},
{
"epoch": 0.7554517133956387,
"grad_norm": 0.8752253075858156,
"learning_rate": 6.873789129590287e-06,
"loss": 0.2598,
"step": 485
},
{
"epoch": 0.7570093457943925,
"grad_norm": 0.6547980788626615,
"learning_rate": 6.862441468433298e-06,
"loss": 0.274,
"step": 486
},
{
"epoch": 0.7585669781931464,
"grad_norm": 0.6955009265885427,
"learning_rate": 6.851082657860453e-06,
"loss": 0.286,
"step": 487
},
{
"epoch": 0.7601246105919003,
"grad_norm": 0.6057981135550708,
"learning_rate": 6.839712765870725e-06,
"loss": 0.3072,
"step": 488
},
{
"epoch": 0.7616822429906542,
"grad_norm": 0.5562050274960125,
"learning_rate": 6.828331860529422e-06,
"loss": 0.2765,
"step": 489
},
{
"epoch": 0.7632398753894081,
"grad_norm": 0.9242326126038012,
"learning_rate": 6.816940009967787e-06,
"loss": 0.8322,
"step": 490
},
{
"epoch": 0.764797507788162,
"grad_norm": 0.7207873437414208,
"learning_rate": 6.805537282382581e-06,
"loss": 0.2175,
"step": 491
},
{
"epoch": 0.7663551401869159,
"grad_norm": 0.5928431496932391,
"learning_rate": 6.79412374603568e-06,
"loss": 0.24,
"step": 492
},
{
"epoch": 0.7679127725856698,
"grad_norm": 0.8086943486132299,
"learning_rate": 6.782699469253671e-06,
"loss": 0.3252,
"step": 493
},
{
"epoch": 0.7694704049844237,
"grad_norm": 0.6500702055304157,
"learning_rate": 6.771264520427432e-06,
"loss": 0.2831,
"step": 494
},
{
"epoch": 0.7710280373831776,
"grad_norm": 0.5699110226071109,
"learning_rate": 6.759818968011731e-06,
"loss": 0.2604,
"step": 495
},
{
"epoch": 0.7725856697819314,
"grad_norm": 0.5580360203832775,
"learning_rate": 6.748362880524819e-06,
"loss": 0.2684,
"step": 496
},
{
"epoch": 0.7741433021806854,
"grad_norm": 0.7886971673525824,
"learning_rate": 6.736896326548006e-06,
"loss": 0.2123,
"step": 497
},
{
"epoch": 0.7757009345794392,
"grad_norm": 0.6054794222872896,
"learning_rate": 6.7254193747252645e-06,
"loss": 0.3127,
"step": 498
},
{
"epoch": 0.7772585669781932,
"grad_norm": 0.7462264966697667,
"learning_rate": 6.713932093762811e-06,
"loss": 0.3051,
"step": 499
},
{
"epoch": 0.778816199376947,
"grad_norm": 0.6730107352048917,
"learning_rate": 6.702434552428702e-06,
"loss": 0.3007,
"step": 500
},
{
"epoch": 0.778816199376947,
"eval_loss": 0.366039514541626,
"eval_runtime": 2.8278,
"eval_samples_per_second": 9.194,
"eval_steps_per_second": 2.475,
"step": 500
},
{
"epoch": 0.780373831775701,
"grad_norm": 0.6861562261314369,
"learning_rate": 6.690926819552408e-06,
"loss": 0.287,
"step": 501
},
{
"epoch": 0.7819314641744548,
"grad_norm": 0.9308842751027873,
"learning_rate": 6.679408964024419e-06,
"loss": 0.8811,
"step": 502
},
{
"epoch": 0.7834890965732088,
"grad_norm": 0.6729340583401545,
"learning_rate": 6.667881054795818e-06,
"loss": 0.2304,
"step": 503
},
{
"epoch": 0.7850467289719626,
"grad_norm": 0.7848693380567189,
"learning_rate": 6.65634316087788e-06,
"loss": 0.2965,
"step": 504
},
{
"epoch": 0.7866043613707165,
"grad_norm": 0.7035209610164758,
"learning_rate": 6.6447953513416474e-06,
"loss": 0.2589,
"step": 505
},
{
"epoch": 0.7881619937694704,
"grad_norm": 0.5912497530045528,
"learning_rate": 6.633237695317523e-06,
"loss": 0.2566,
"step": 506
},
{
"epoch": 0.7897196261682243,
"grad_norm": 0.7247030582803601,
"learning_rate": 6.621670261994857e-06,
"loss": 0.2726,
"step": 507
},
{
"epoch": 0.7912772585669782,
"grad_norm": 0.6656199289111854,
"learning_rate": 6.610093120621532e-06,
"loss": 0.2999,
"step": 508
},
{
"epoch": 0.7928348909657321,
"grad_norm": 0.8158049430019846,
"learning_rate": 6.598506340503541e-06,
"loss": 0.2453,
"step": 509
},
{
"epoch": 0.794392523364486,
"grad_norm": 0.6984808487227331,
"learning_rate": 6.586909991004587e-06,
"loss": 0.3149,
"step": 510
},
{
"epoch": 0.7959501557632399,
"grad_norm": 0.6126995511808185,
"learning_rate": 6.575304141545653e-06,
"loss": 0.2666,
"step": 511
},
{
"epoch": 0.7975077881619937,
"grad_norm": 0.5505067863178127,
"learning_rate": 6.5636888616046e-06,
"loss": 0.2998,
"step": 512
},
{
"epoch": 0.7990654205607477,
"grad_norm": 0.7685023837309144,
"learning_rate": 6.552064220715737e-06,
"loss": 0.1876,
"step": 513
},
{
"epoch": 0.8006230529595015,
"grad_norm": 0.6186176349687995,
"learning_rate": 6.5404302884694145e-06,
"loss": 0.2823,
"step": 514
},
{
"epoch": 0.8021806853582555,
"grad_norm": 0.6245560858038731,
"learning_rate": 6.528787134511608e-06,
"loss": 0.2063,
"step": 515
},
{
"epoch": 0.8037383177570093,
"grad_norm": 0.6027333943405707,
"learning_rate": 6.5171348285434965e-06,
"loss": 0.3079,
"step": 516
},
{
"epoch": 0.8052959501557633,
"grad_norm": 0.7334879584524152,
"learning_rate": 6.505473440321044e-06,
"loss": 0.2906,
"step": 517
},
{
"epoch": 0.8068535825545171,
"grad_norm": 0.5611150628723894,
"learning_rate": 6.493803039654589e-06,
"loss": 0.2437,
"step": 518
},
{
"epoch": 0.8084112149532711,
"grad_norm": 0.5709098533110826,
"learning_rate": 6.48212369640842e-06,
"loss": 0.2695,
"step": 519
},
{
"epoch": 0.8099688473520249,
"grad_norm": 0.7469745364074795,
"learning_rate": 6.4704354805003626e-06,
"loss": 0.2828,
"step": 520
},
{
"epoch": 0.8115264797507789,
"grad_norm": 0.7580041865120294,
"learning_rate": 6.458738461901354e-06,
"loss": 0.2456,
"step": 521
},
{
"epoch": 0.8130841121495327,
"grad_norm": 0.7356278600835281,
"learning_rate": 6.447032710635035e-06,
"loss": 0.2325,
"step": 522
},
{
"epoch": 0.8146417445482866,
"grad_norm": 0.5694798902062997,
"learning_rate": 6.435318296777316e-06,
"loss": 0.2763,
"step": 523
},
{
"epoch": 0.8161993769470405,
"grad_norm": 0.6227419240061058,
"learning_rate": 6.423595290455971e-06,
"loss": 0.2871,
"step": 524
},
{
"epoch": 0.8177570093457944,
"grad_norm": 0.7005734890264759,
"learning_rate": 6.41186376185021e-06,
"loss": 0.3003,
"step": 525
},
{
"epoch": 0.8193146417445483,
"grad_norm": 0.8909957470129115,
"learning_rate": 6.400123781190265e-06,
"loss": 0.3328,
"step": 526
},
{
"epoch": 0.8208722741433022,
"grad_norm": 0.5534613081412874,
"learning_rate": 6.388375418756959e-06,
"loss": 0.2816,
"step": 527
},
{
"epoch": 0.822429906542056,
"grad_norm": 0.7546735553270361,
"learning_rate": 6.3766187448813e-06,
"loss": 0.249,
"step": 528
},
{
"epoch": 0.82398753894081,
"grad_norm": 1.08724358600162,
"learning_rate": 6.3648538299440444e-06,
"loss": 0.2978,
"step": 529
},
{
"epoch": 0.8255451713395638,
"grad_norm": 0.5510668935985298,
"learning_rate": 6.35308074437529e-06,
"loss": 0.2586,
"step": 530
},
{
"epoch": 0.8271028037383178,
"grad_norm": 0.9295747471096518,
"learning_rate": 6.341299558654042e-06,
"loss": 0.4423,
"step": 531
},
{
"epoch": 0.8286604361370716,
"grad_norm": 0.6078601432381958,
"learning_rate": 6.329510343307801e-06,
"loss": 0.3089,
"step": 532
},
{
"epoch": 0.8302180685358256,
"grad_norm": 0.7009464939112006,
"learning_rate": 6.3177131689121325e-06,
"loss": 0.276,
"step": 533
},
{
"epoch": 0.8317757009345794,
"grad_norm": 0.6055023924718476,
"learning_rate": 6.305908106090255e-06,
"loss": 0.289,
"step": 534
},
{
"epoch": 0.8333333333333334,
"grad_norm": 1.0146910918408667,
"learning_rate": 6.294095225512604e-06,
"loss": 0.2117,
"step": 535
},
{
"epoch": 0.8348909657320872,
"grad_norm": 0.5684609592585866,
"learning_rate": 6.282274597896421e-06,
"loss": 0.268,
"step": 536
},
{
"epoch": 0.8364485981308412,
"grad_norm": 0.6324240721524141,
"learning_rate": 6.2704462940053165e-06,
"loss": 0.2348,
"step": 537
},
{
"epoch": 0.838006230529595,
"grad_norm": 0.582281483203043,
"learning_rate": 6.2586103846488654e-06,
"loss": 0.2975,
"step": 538
},
{
"epoch": 0.839563862928349,
"grad_norm": 0.9776397911217686,
"learning_rate": 6.246766940682165e-06,
"loss": 0.7799,
"step": 539
},
{
"epoch": 0.8411214953271028,
"grad_norm": 0.5174311636719064,
"learning_rate": 6.234916033005421e-06,
"loss": 0.1973,
"step": 540
},
{
"epoch": 0.8426791277258567,
"grad_norm": 0.6331348424871293,
"learning_rate": 6.22305773256352e-06,
"loss": 0.2749,
"step": 541
},
{
"epoch": 0.8442367601246106,
"grad_norm": 0.5229024799327089,
"learning_rate": 6.211192110345603e-06,
"loss": 0.2811,
"step": 542
},
{
"epoch": 0.8457943925233645,
"grad_norm": 0.5575336291274628,
"learning_rate": 6.199319237384645e-06,
"loss": 0.2534,
"step": 543
},
{
"epoch": 0.8473520249221184,
"grad_norm": 0.7289277957152529,
"learning_rate": 6.187439184757025e-06,
"loss": 0.7677,
"step": 544
},
{
"epoch": 0.8489096573208723,
"grad_norm": 0.540862000313681,
"learning_rate": 6.1755520235821055e-06,
"loss": 0.3294,
"step": 545
},
{
"epoch": 0.8504672897196262,
"grad_norm": 0.7035928510596402,
"learning_rate": 6.163657825021802e-06,
"loss": 0.3147,
"step": 546
},
{
"epoch": 0.8520249221183801,
"grad_norm": 0.5438019081147566,
"learning_rate": 6.1517566602801596e-06,
"loss": 0.2003,
"step": 547
},
{
"epoch": 0.8535825545171339,
"grad_norm": 0.8025432437697821,
"learning_rate": 6.139848600602926e-06,
"loss": 0.2756,
"step": 548
},
{
"epoch": 0.8551401869158879,
"grad_norm": 0.654724311621637,
"learning_rate": 6.127933717277123e-06,
"loss": 0.2934,
"step": 549
},
{
"epoch": 0.8566978193146417,
"grad_norm": 1.328494349119985,
"learning_rate": 6.116012081630629e-06,
"loss": 0.2731,
"step": 550
},
{
"epoch": 0.8582554517133957,
"grad_norm": 0.5435085655801555,
"learning_rate": 6.104083765031734e-06,
"loss": 0.1934,
"step": 551
},
{
"epoch": 0.8598130841121495,
"grad_norm": 0.7875228878328122,
"learning_rate": 6.0921488388887315e-06,
"loss": 0.7651,
"step": 552
},
{
"epoch": 0.8613707165109035,
"grad_norm": 0.9138141754922854,
"learning_rate": 6.080207374649482e-06,
"loss": 0.6927,
"step": 553
},
{
"epoch": 0.8629283489096573,
"grad_norm": 0.658537276564411,
"learning_rate": 6.068259443800981e-06,
"loss": 0.3088,
"step": 554
},
{
"epoch": 0.8644859813084113,
"grad_norm": 0.8424646536545584,
"learning_rate": 6.0563051178689395e-06,
"loss": 0.6504,
"step": 555
},
{
"epoch": 0.8660436137071651,
"grad_norm": 0.5487578856405264,
"learning_rate": 6.0443444684173524e-06,
"loss": 0.2504,
"step": 556
},
{
"epoch": 0.867601246105919,
"grad_norm": 0.577742497246078,
"learning_rate": 6.032377567048071e-06,
"loss": 0.2724,
"step": 557
},
{
"epoch": 0.8691588785046729,
"grad_norm": 0.5720861179090082,
"learning_rate": 6.0204044854003705e-06,
"loss": 0.2494,
"step": 558
},
{
"epoch": 0.8707165109034268,
"grad_norm": 0.614579567677496,
"learning_rate": 6.008425295150526e-06,
"loss": 0.2431,
"step": 559
},
{
"epoch": 0.8722741433021807,
"grad_norm": 0.7053311054530548,
"learning_rate": 5.996440068011383e-06,
"loss": 0.3007,
"step": 560
},
{
"epoch": 0.8738317757009346,
"grad_norm": 0.6676390464189279,
"learning_rate": 5.9844488757319205e-06,
"loss": 0.2309,
"step": 561
},
{
"epoch": 0.8753894080996885,
"grad_norm": 0.7749136512043995,
"learning_rate": 5.972451790096837e-06,
"loss": 0.3327,
"step": 562
},
{
"epoch": 0.8769470404984424,
"grad_norm": 0.6384638289071073,
"learning_rate": 5.960448882926101e-06,
"loss": 0.3447,
"step": 563
},
{
"epoch": 0.8785046728971962,
"grad_norm": 0.5992691951009588,
"learning_rate": 5.948440226074539e-06,
"loss": 0.2181,
"step": 564
},
{
"epoch": 0.8800623052959502,
"grad_norm": 0.7553595588283479,
"learning_rate": 5.936425891431394e-06,
"loss": 0.2307,
"step": 565
},
{
"epoch": 0.881619937694704,
"grad_norm": 0.7852167594459105,
"learning_rate": 5.924405950919902e-06,
"loss": 0.3119,
"step": 566
},
{
"epoch": 0.883177570093458,
"grad_norm": 0.5241767306732831,
"learning_rate": 5.91238047649685e-06,
"loss": 0.2293,
"step": 567
},
{
"epoch": 0.8847352024922118,
"grad_norm": 0.7471757161407457,
"learning_rate": 5.900349540152167e-06,
"loss": 0.3251,
"step": 568
},
{
"epoch": 0.8862928348909658,
"grad_norm": 0.7843562558921093,
"learning_rate": 5.888313213908468e-06,
"loss": 0.2868,
"step": 569
},
{
"epoch": 0.8878504672897196,
"grad_norm": 0.7503757117692468,
"learning_rate": 5.876271569820638e-06,
"loss": 0.2555,
"step": 570
},
{
"epoch": 0.8894080996884736,
"grad_norm": 0.7275473381141497,
"learning_rate": 5.864224679975399e-06,
"loss": 0.2945,
"step": 571
},
{
"epoch": 0.8909657320872274,
"grad_norm": 0.7108035276261796,
"learning_rate": 5.852172616490875e-06,
"loss": 0.2826,
"step": 572
},
{
"epoch": 0.8925233644859814,
"grad_norm": 0.7292991034746364,
"learning_rate": 5.84011545151616e-06,
"loss": 0.2052,
"step": 573
},
{
"epoch": 0.8940809968847352,
"grad_norm": 0.7701552350107903,
"learning_rate": 5.828053257230893e-06,
"loss": 0.272,
"step": 574
},
{
"epoch": 0.8956386292834891,
"grad_norm": 0.5385532832886137,
"learning_rate": 5.815986105844813e-06,
"loss": 0.2859,
"step": 575
},
{
"epoch": 0.897196261682243,
"grad_norm": 0.4784130573714912,
"learning_rate": 5.803914069597342e-06,
"loss": 0.2385,
"step": 576
},
{
"epoch": 0.8987538940809969,
"grad_norm": 0.6661872649899948,
"learning_rate": 5.791837220757139e-06,
"loss": 0.2601,
"step": 577
},
{
"epoch": 0.9003115264797508,
"grad_norm": 0.6423784935357807,
"learning_rate": 5.779755631621679e-06,
"loss": 0.2861,
"step": 578
},
{
"epoch": 0.9018691588785047,
"grad_norm": 0.5564393840451491,
"learning_rate": 5.767669374516807e-06,
"loss": 0.2247,
"step": 579
},
{
"epoch": 0.9034267912772586,
"grad_norm": 1.0831258133460262,
"learning_rate": 5.755578521796321e-06,
"loss": 0.7525,
"step": 580
},
{
"epoch": 0.9049844236760125,
"grad_norm": 0.7513349003919912,
"learning_rate": 5.743483145841525e-06,
"loss": 0.2417,
"step": 581
},
{
"epoch": 0.9065420560747663,
"grad_norm": 0.6795049942946425,
"learning_rate": 5.731383319060805e-06,
"loss": 0.3177,
"step": 582
},
{
"epoch": 0.9080996884735203,
"grad_norm": 0.7390882203257174,
"learning_rate": 5.719279113889184e-06,
"loss": 0.3581,
"step": 583
},
{
"epoch": 0.9096573208722741,
"grad_norm": 0.582238701325137,
"learning_rate": 5.707170602787908e-06,
"loss": 0.2755,
"step": 584
},
{
"epoch": 0.9112149532710281,
"grad_norm": 0.5600730304265147,
"learning_rate": 5.695057858243989e-06,
"loss": 0.2745,
"step": 585
},
{
"epoch": 0.9127725856697819,
"grad_norm": 0.638261445819756,
"learning_rate": 5.682940952769788e-06,
"loss": 0.177,
"step": 586
},
{
"epoch": 0.9143302180685359,
"grad_norm": 0.6191376433403027,
"learning_rate": 5.670819958902576e-06,
"loss": 0.2447,
"step": 587
},
{
"epoch": 0.9158878504672897,
"grad_norm": 0.8780779667116613,
"learning_rate": 5.658694949204094e-06,
"loss": 0.2438,
"step": 588
},
{
"epoch": 0.9174454828660437,
"grad_norm": 0.7198384217051569,
"learning_rate": 5.646565996260129e-06,
"loss": 0.2408,
"step": 589
},
{
"epoch": 0.9190031152647975,
"grad_norm": 1.0167707545321327,
"learning_rate": 5.634433172680072e-06,
"loss": 0.7316,
"step": 590
},
{
"epoch": 0.9205607476635514,
"grad_norm": 0.6546640817603999,
"learning_rate": 5.622296551096481e-06,
"loss": 0.3253,
"step": 591
},
{
"epoch": 0.9221183800623053,
"grad_norm": 0.8382914937621136,
"learning_rate": 5.61015620416466e-06,
"loss": 0.2981,
"step": 592
},
{
"epoch": 0.9236760124610592,
"grad_norm": 0.6120676244562511,
"learning_rate": 5.598012204562204e-06,
"loss": 0.2647,
"step": 593
},
{
"epoch": 0.9252336448598131,
"grad_norm": 0.700304920368459,
"learning_rate": 5.5858646249885855e-06,
"loss": 0.2249,
"step": 594
},
{
"epoch": 0.926791277258567,
"grad_norm": 0.6840607666164885,
"learning_rate": 5.573713538164698e-06,
"loss": 0.2795,
"step": 595
},
{
"epoch": 0.9283489096573209,
"grad_norm": 0.8513439406827232,
"learning_rate": 5.561559016832438e-06,
"loss": 0.2931,
"step": 596
},
{
"epoch": 0.9299065420560748,
"grad_norm": 0.7770854828276103,
"learning_rate": 5.549401133754259e-06,
"loss": 0.2819,
"step": 597
},
{
"epoch": 0.9314641744548287,
"grad_norm": 0.5874428152071298,
"learning_rate": 5.5372399617127415e-06,
"loss": 0.4483,
"step": 598
},
{
"epoch": 0.9330218068535826,
"grad_norm": 0.6488291930752155,
"learning_rate": 5.525075573510154e-06,
"loss": 0.3873,
"step": 599
},
{
"epoch": 0.9345794392523364,
"grad_norm": 0.5286049009581415,
"learning_rate": 5.512908041968018e-06,
"loss": 0.3113,
"step": 600
},
{
"epoch": 0.9361370716510904,
"grad_norm": 0.5929114341675884,
"learning_rate": 5.500737439926674e-06,
"loss": 0.334,
"step": 601
},
{
"epoch": 0.9376947040498442,
"grad_norm": 0.847988678139219,
"learning_rate": 5.488563840244843e-06,
"loss": 0.3026,
"step": 602
},
{
"epoch": 0.9392523364485982,
"grad_norm": 0.5635536494534855,
"learning_rate": 5.476387315799189e-06,
"loss": 0.2146,
"step": 603
},
{
"epoch": 0.940809968847352,
"grad_norm": 0.5570339586064699,
"learning_rate": 5.464207939483891e-06,
"loss": 0.2407,
"step": 604
},
{
"epoch": 0.942367601246106,
"grad_norm": 0.5996622051462993,
"learning_rate": 5.452025784210193e-06,
"loss": 0.2301,
"step": 605
},
{
"epoch": 0.9439252336448598,
"grad_norm": 0.6134433053870486,
"learning_rate": 5.439840922905982e-06,
"loss": 0.2881,
"step": 606
},
{
"epoch": 0.9454828660436138,
"grad_norm": 0.6534399078221188,
"learning_rate": 5.42765342851534e-06,
"loss": 0.2991,
"step": 607
},
{
"epoch": 0.9470404984423676,
"grad_norm": 0.690575160461211,
"learning_rate": 5.415463373998112e-06,
"loss": 0.3353,
"step": 608
},
{
"epoch": 0.9485981308411215,
"grad_norm": 0.8921403858615625,
"learning_rate": 5.403270832329473e-06,
"loss": 0.2008,
"step": 609
},
{
"epoch": 0.9501557632398754,
"grad_norm": 0.7025072704858318,
"learning_rate": 5.391075876499483e-06,
"loss": 0.2621,
"step": 610
},
{
"epoch": 0.9517133956386293,
"grad_norm": 0.6757376481036776,
"learning_rate": 5.3788785795126554e-06,
"loss": 0.2469,
"step": 611
},
{
"epoch": 0.9532710280373832,
"grad_norm": 0.6875817981934039,
"learning_rate": 5.36667901438752e-06,
"loss": 0.236,
"step": 612
},
{
"epoch": 0.9548286604361371,
"grad_norm": 0.7529907066389188,
"learning_rate": 5.354477254156184e-06,
"loss": 0.2755,
"step": 613
},
{
"epoch": 0.956386292834891,
"grad_norm": 0.5896914001022201,
"learning_rate": 5.342273371863895e-06,
"loss": 0.2634,
"step": 614
},
{
"epoch": 0.9579439252336449,
"grad_norm": 0.7027739875415374,
"learning_rate": 5.330067440568605e-06,
"loss": 0.2829,
"step": 615
},
{
"epoch": 0.9595015576323987,
"grad_norm": 0.7667507882778263,
"learning_rate": 5.317859533340532e-06,
"loss": 0.3506,
"step": 616
},
{
"epoch": 0.9610591900311527,
"grad_norm": 0.8456849495117152,
"learning_rate": 5.30564972326172e-06,
"loss": 0.3054,
"step": 617
},
{
"epoch": 0.9626168224299065,
"grad_norm": 0.51248054747711,
"learning_rate": 5.293438083425611e-06,
"loss": 0.2301,
"step": 618
},
{
"epoch": 0.9641744548286605,
"grad_norm": 0.7467447701014356,
"learning_rate": 5.281224686936594e-06,
"loss": 0.3769,
"step": 619
},
{
"epoch": 0.9657320872274143,
"grad_norm": 0.6414852824849853,
"learning_rate": 5.26900960690958e-06,
"loss": 0.2779,
"step": 620
},
{
"epoch": 0.9672897196261683,
"grad_norm": 0.6295561918474766,
"learning_rate": 5.256792916469552e-06,
"loss": 0.2586,
"step": 621
},
{
"epoch": 0.9688473520249221,
"grad_norm": 0.6049915483456579,
"learning_rate": 5.244574688751138e-06,
"loss": 0.2195,
"step": 622
},
{
"epoch": 0.9704049844236761,
"grad_norm": 1.0443726591172307,
"learning_rate": 5.23235499689817e-06,
"loss": 0.7974,
"step": 623
},
{
"epoch": 0.9719626168224299,
"grad_norm": 0.4947295180592083,
"learning_rate": 5.220133914063239e-06,
"loss": 0.2173,
"step": 624
},
{
"epoch": 0.9735202492211839,
"grad_norm": 0.774384784856486,
"learning_rate": 5.20791151340727e-06,
"loss": 0.2196,
"step": 625
},
{
"epoch": 0.9750778816199377,
"grad_norm": 0.6848166382044595,
"learning_rate": 5.195687868099073e-06,
"loss": 0.2784,
"step": 626
},
{
"epoch": 0.9766355140186916,
"grad_norm": 1.5114886266488685,
"learning_rate": 5.1834630513149086e-06,
"loss": 0.8531,
"step": 627
},
{
"epoch": 0.9781931464174455,
"grad_norm": 1.0107267590162416,
"learning_rate": 5.171237136238054e-06,
"loss": 0.2692,
"step": 628
},
{
"epoch": 0.9797507788161994,
"grad_norm": 0.7292686375997546,
"learning_rate": 5.159010196058356e-06,
"loss": 0.2305,
"step": 629
},
{
"epoch": 0.9813084112149533,
"grad_norm": 0.565137268345777,
"learning_rate": 5.1467823039718046e-06,
"loss": 0.3076,
"step": 630
},
{
"epoch": 0.9828660436137072,
"grad_norm": 1.0043606696953649,
"learning_rate": 5.134553533180082e-06,
"loss": 0.7515,
"step": 631
},
{
"epoch": 0.9844236760124611,
"grad_norm": 0.6233222851344706,
"learning_rate": 5.122323956890136e-06,
"loss": 0.3019,
"step": 632
},
{
"epoch": 0.985981308411215,
"grad_norm": 0.730841019679576,
"learning_rate": 5.110093648313732e-06,
"loss": 0.3483,
"step": 633
},
{
"epoch": 0.9875389408099688,
"grad_norm": 0.6632089778706831,
"learning_rate": 5.097862680667024e-06,
"loss": 0.3608,
"step": 634
},
{
"epoch": 0.9890965732087228,
"grad_norm": 0.6691349603071851,
"learning_rate": 5.085631127170106e-06,
"loss": 0.2524,
"step": 635
},
{
"epoch": 0.9906542056074766,
"grad_norm": 0.73951871814352,
"learning_rate": 5.073399061046584e-06,
"loss": 0.3134,
"step": 636
},
{
"epoch": 0.9922118380062306,
"grad_norm": 0.6079224831882191,
"learning_rate": 5.061166555523129e-06,
"loss": 0.308,
"step": 637
},
{
"epoch": 0.9937694704049844,
"grad_norm": 0.6166501321049228,
"learning_rate": 5.048933683829046e-06,
"loss": 0.3073,
"step": 638
},
{
"epoch": 0.9953271028037384,
"grad_norm": 1.003680827370098,
"learning_rate": 5.0367005191958275e-06,
"loss": 0.3229,
"step": 639
},
{
"epoch": 0.9968847352024922,
"grad_norm": 0.7517097037555416,
"learning_rate": 5.024467134856725e-06,
"loss": 0.2744,
"step": 640
},
{
"epoch": 0.9984423676012462,
"grad_norm": 0.5399606823979543,
"learning_rate": 5.012233604046303e-06,
"loss": 0.247,
"step": 641
},
{
"epoch": 1.0,
"grad_norm": 0.602288219366613,
"learning_rate": 5e-06,
"loss": 0.2401,
"step": 642
},
{
"epoch": 1.0015576323987538,
"grad_norm": 0.5865025697495287,
"learning_rate": 4.987766395953699e-06,
"loss": 0.1907,
"step": 643
},
{
"epoch": 1.0031152647975077,
"grad_norm": 0.6271078194072086,
"learning_rate": 4.975532865143277e-06,
"loss": 0.1837,
"step": 644
},
{
"epoch": 1.0046728971962617,
"grad_norm": 0.5433097442594456,
"learning_rate": 4.963299480804173e-06,
"loss": 0.2122,
"step": 645
},
{
"epoch": 1.0062305295950156,
"grad_norm": 0.6939024837652152,
"learning_rate": 4.951066316170956e-06,
"loss": 0.2464,
"step": 646
},
{
"epoch": 1.0077881619937694,
"grad_norm": 0.6120591075726564,
"learning_rate": 4.938833444476873e-06,
"loss": 0.2943,
"step": 647
},
{
"epoch": 1.0093457943925233,
"grad_norm": 0.5279986678775014,
"learning_rate": 4.926600938953418e-06,
"loss": 0.2139,
"step": 648
},
{
"epoch": 1.0109034267912773,
"grad_norm": 0.5465124083851354,
"learning_rate": 4.9143688728298946e-06,
"loss": 0.2946,
"step": 649
},
{
"epoch": 1.0124610591900312,
"grad_norm": 0.598599797464655,
"learning_rate": 4.9021373193329775e-06,
"loss": 0.2467,
"step": 650
},
{
"epoch": 1.014018691588785,
"grad_norm": 0.5911973059402644,
"learning_rate": 4.889906351686269e-06,
"loss": 0.2439,
"step": 651
},
{
"epoch": 1.0155763239875388,
"grad_norm": 0.5007183880780492,
"learning_rate": 4.8776760431098665e-06,
"loss": 0.2448,
"step": 652
},
{
"epoch": 1.017133956386293,
"grad_norm": 0.6198547585957133,
"learning_rate": 4.865446466819918e-06,
"loss": 0.2107,
"step": 653
},
{
"epoch": 1.0186915887850467,
"grad_norm": 0.6141060329229784,
"learning_rate": 4.853217696028197e-06,
"loss": 0.2124,
"step": 654
},
{
"epoch": 1.0202492211838006,
"grad_norm": 0.5500970083546799,
"learning_rate": 4.840989803941645e-06,
"loss": 0.2413,
"step": 655
},
{
"epoch": 1.0218068535825544,
"grad_norm": 0.5785491153029504,
"learning_rate": 4.828762863761948e-06,
"loss": 0.322,
"step": 656
},
{
"epoch": 1.0233644859813085,
"grad_norm": 0.5158799094788992,
"learning_rate": 4.816536948685091e-06,
"loss": 0.2458,
"step": 657
},
{
"epoch": 1.0249221183800623,
"grad_norm": 0.7110301489590548,
"learning_rate": 4.804312131900929e-06,
"loss": 0.1491,
"step": 658
},
{
"epoch": 1.0264797507788161,
"grad_norm": 0.49142641785128816,
"learning_rate": 4.792088486592731e-06,
"loss": 0.2403,
"step": 659
},
{
"epoch": 1.02803738317757,
"grad_norm": 0.6365446048004576,
"learning_rate": 4.779866085936762e-06,
"loss": 0.2723,
"step": 660
},
{
"epoch": 1.029595015576324,
"grad_norm": 0.5622051291501104,
"learning_rate": 4.767645003101831e-06,
"loss": 0.1508,
"step": 661
},
{
"epoch": 1.0311526479750779,
"grad_norm": 0.7978535815313595,
"learning_rate": 4.755425311248863e-06,
"loss": 0.7126,
"step": 662
},
{
"epoch": 1.0327102803738317,
"grad_norm": 0.49688606974260013,
"learning_rate": 4.7432070835304494e-06,
"loss": 0.1821,
"step": 663
},
{
"epoch": 1.0342679127725856,
"grad_norm": 0.9843736648935775,
"learning_rate": 4.730990393090422e-06,
"loss": 0.6915,
"step": 664
},
{
"epoch": 1.0358255451713396,
"grad_norm": 0.5197585726631249,
"learning_rate": 4.718775313063406e-06,
"loss": 0.1985,
"step": 665
},
{
"epoch": 1.0373831775700935,
"grad_norm": 0.7127374001801975,
"learning_rate": 4.70656191657439e-06,
"loss": 0.2209,
"step": 666
},
{
"epoch": 1.0389408099688473,
"grad_norm": 0.5805515887865611,
"learning_rate": 4.6943502767382815e-06,
"loss": 0.1781,
"step": 667
},
{
"epoch": 1.0404984423676011,
"grad_norm": 0.5812301249853751,
"learning_rate": 4.6821404666594715e-06,
"loss": 0.2362,
"step": 668
},
{
"epoch": 1.0420560747663552,
"grad_norm": 0.5137436961715455,
"learning_rate": 4.669932559431396e-06,
"loss": 0.2381,
"step": 669
},
{
"epoch": 1.043613707165109,
"grad_norm": 0.49254347681935357,
"learning_rate": 4.657726628136105e-06,
"loss": 0.2335,
"step": 670
},
{
"epoch": 1.0451713395638629,
"grad_norm": 0.807702133830089,
"learning_rate": 4.645522745843817e-06,
"loss": 0.1636,
"step": 671
},
{
"epoch": 1.0467289719626167,
"grad_norm": 0.6598785992415666,
"learning_rate": 4.6333209856124814e-06,
"loss": 0.2039,
"step": 672
},
{
"epoch": 1.0482866043613708,
"grad_norm": 0.5735476926058175,
"learning_rate": 4.621121420487345e-06,
"loss": 0.2344,
"step": 673
},
{
"epoch": 1.0498442367601246,
"grad_norm": 0.5406292279451508,
"learning_rate": 4.608924123500519e-06,
"loss": 0.1802,
"step": 674
},
{
"epoch": 1.0514018691588785,
"grad_norm": 0.5983081571916266,
"learning_rate": 4.596729167670529e-06,
"loss": 0.1676,
"step": 675
},
{
"epoch": 1.0529595015576323,
"grad_norm": 0.5810265835484077,
"learning_rate": 4.58453662600189e-06,
"loss": 0.2822,
"step": 676
},
{
"epoch": 1.0545171339563864,
"grad_norm": 0.6855576677633736,
"learning_rate": 4.572346571484661e-06,
"loss": 0.1977,
"step": 677
},
{
"epoch": 1.0560747663551402,
"grad_norm": 0.5335540200996262,
"learning_rate": 4.5601590770940195e-06,
"loss": 0.2152,
"step": 678
},
{
"epoch": 1.057632398753894,
"grad_norm": 0.6453393789551976,
"learning_rate": 4.547974215789808e-06,
"loss": 0.1886,
"step": 679
},
{
"epoch": 1.0591900311526479,
"grad_norm": 0.5915758123605618,
"learning_rate": 4.535792060516112e-06,
"loss": 0.2209,
"step": 680
},
{
"epoch": 1.060747663551402,
"grad_norm": 0.6302498485071517,
"learning_rate": 4.523612684200811e-06,
"loss": 0.239,
"step": 681
},
{
"epoch": 1.0623052959501558,
"grad_norm": 0.7064325822119844,
"learning_rate": 4.511436159755159e-06,
"loss": 0.1988,
"step": 682
},
{
"epoch": 1.0638629283489096,
"grad_norm": 0.5229568055974967,
"learning_rate": 4.499262560073328e-06,
"loss": 0.2219,
"step": 683
},
{
"epoch": 1.0654205607476634,
"grad_norm": 0.848853210404809,
"learning_rate": 4.487091958031984e-06,
"loss": 0.1813,
"step": 684
},
{
"epoch": 1.0669781931464175,
"grad_norm": 0.62716757784117,
"learning_rate": 4.474924426489847e-06,
"loss": 0.2433,
"step": 685
},
{
"epoch": 1.0685358255451713,
"grad_norm": 0.5961449452514619,
"learning_rate": 4.46276003828726e-06,
"loss": 0.2576,
"step": 686
},
{
"epoch": 1.0700934579439252,
"grad_norm": 0.5336164214991632,
"learning_rate": 4.450598866245743e-06,
"loss": 0.2094,
"step": 687
},
{
"epoch": 1.071651090342679,
"grad_norm": 0.5350296348658355,
"learning_rate": 4.438440983167564e-06,
"loss": 0.2883,
"step": 688
},
{
"epoch": 1.073208722741433,
"grad_norm": 0.53810152034016,
"learning_rate": 4.426286461835303e-06,
"loss": 0.1968,
"step": 689
},
{
"epoch": 1.074766355140187,
"grad_norm": 0.4674678946148812,
"learning_rate": 4.414135375011416e-06,
"loss": 0.1468,
"step": 690
},
{
"epoch": 1.0763239875389408,
"grad_norm": 0.5672398628672646,
"learning_rate": 4.401987795437797e-06,
"loss": 0.1796,
"step": 691
},
{
"epoch": 1.0778816199376946,
"grad_norm": 0.6409062495534954,
"learning_rate": 4.3898437958353435e-06,
"loss": 0.1536,
"step": 692
},
{
"epoch": 1.0794392523364487,
"grad_norm": 0.7979975112099554,
"learning_rate": 4.377703448903519e-06,
"loss": 0.7814,
"step": 693
},
{
"epoch": 1.0809968847352025,
"grad_norm": 0.6128214407671699,
"learning_rate": 4.3655668273199305e-06,
"loss": 0.249,
"step": 694
},
{
"epoch": 1.0825545171339563,
"grad_norm": 0.6671186706045426,
"learning_rate": 4.353434003739872e-06,
"loss": 0.162,
"step": 695
},
{
"epoch": 1.0841121495327102,
"grad_norm": 0.5657453591665842,
"learning_rate": 4.341305050795907e-06,
"loss": 0.2355,
"step": 696
},
{
"epoch": 1.0856697819314642,
"grad_norm": 0.5876907165277706,
"learning_rate": 4.329180041097425e-06,
"loss": 0.2898,
"step": 697
},
{
"epoch": 1.087227414330218,
"grad_norm": 0.5418335332942893,
"learning_rate": 4.3170590472302125e-06,
"loss": 0.2049,
"step": 698
},
{
"epoch": 1.088785046728972,
"grad_norm": 0.6194906159202879,
"learning_rate": 4.304942141756012e-06,
"loss": 0.2716,
"step": 699
},
{
"epoch": 1.0903426791277258,
"grad_norm": 0.6321743236263448,
"learning_rate": 4.292829397212094e-06,
"loss": 0.1858,
"step": 700
},
{
"epoch": 1.0919003115264798,
"grad_norm": 0.5884478735032194,
"learning_rate": 4.280720886110815e-06,
"loss": 0.2181,
"step": 701
},
{
"epoch": 1.0934579439252337,
"grad_norm": 0.512888102374138,
"learning_rate": 4.268616680939197e-06,
"loss": 0.2495,
"step": 702
},
{
"epoch": 1.0950155763239875,
"grad_norm": 0.6806148269647997,
"learning_rate": 4.256516854158476e-06,
"loss": 0.1811,
"step": 703
},
{
"epoch": 1.0965732087227413,
"grad_norm": 0.590183243846604,
"learning_rate": 4.244421478203681e-06,
"loss": 0.1759,
"step": 704
},
{
"epoch": 1.0981308411214954,
"grad_norm": 0.7141876070652246,
"learning_rate": 4.232330625483194e-06,
"loss": 0.2155,
"step": 705
},
{
"epoch": 1.0996884735202492,
"grad_norm": 0.5606859986060387,
"learning_rate": 4.220244368378324e-06,
"loss": 0.1748,
"step": 706
},
{
"epoch": 1.101246105919003,
"grad_norm": 0.5374363163472021,
"learning_rate": 4.208162779242862e-06,
"loss": 0.15,
"step": 707
},
{
"epoch": 1.102803738317757,
"grad_norm": 0.6337046437481759,
"learning_rate": 4.19608593040266e-06,
"loss": 0.2094,
"step": 708
},
{
"epoch": 1.104361370716511,
"grad_norm": 0.6441679702023525,
"learning_rate": 4.184013894155187e-06,
"loss": 0.1347,
"step": 709
},
{
"epoch": 1.1059190031152648,
"grad_norm": 0.5105399247363097,
"learning_rate": 4.171946742769109e-06,
"loss": 0.2492,
"step": 710
},
{
"epoch": 1.1074766355140186,
"grad_norm": 0.835104618126293,
"learning_rate": 4.1598845484838405e-06,
"loss": 0.6552,
"step": 711
},
{
"epoch": 1.1090342679127725,
"grad_norm": 0.5091044357624278,
"learning_rate": 4.147827383509127e-06,
"loss": 0.2459,
"step": 712
},
{
"epoch": 1.1105919003115265,
"grad_norm": 0.4173053058204635,
"learning_rate": 4.135775320024601e-06,
"loss": 0.1834,
"step": 713
},
{
"epoch": 1.1121495327102804,
"grad_norm": 0.516073956820414,
"learning_rate": 4.123728430179363e-06,
"loss": 0.2096,
"step": 714
},
{
"epoch": 1.1137071651090342,
"grad_norm": 0.5229385439265322,
"learning_rate": 4.111686786091534e-06,
"loss": 0.1619,
"step": 715
},
{
"epoch": 1.115264797507788,
"grad_norm": 0.697243292089673,
"learning_rate": 4.099650459847835e-06,
"loss": 0.2181,
"step": 716
},
{
"epoch": 1.1168224299065421,
"grad_norm": 0.7935991971149668,
"learning_rate": 4.087619523503149e-06,
"loss": 0.7976,
"step": 717
},
{
"epoch": 1.118380062305296,
"grad_norm": 0.7468568843345416,
"learning_rate": 4.0755940490801e-06,
"loss": 0.1844,
"step": 718
},
{
"epoch": 1.1199376947040498,
"grad_norm": 0.655225509433663,
"learning_rate": 4.0635741085686065e-06,
"loss": 0.2016,
"step": 719
},
{
"epoch": 1.1214953271028036,
"grad_norm": 0.6590392753571896,
"learning_rate": 4.051559773925462e-06,
"loss": 0.2175,
"step": 720
},
{
"epoch": 1.1230529595015577,
"grad_norm": 0.5673140068671504,
"learning_rate": 4.039551117073899e-06,
"loss": 0.1377,
"step": 721
},
{
"epoch": 1.1246105919003115,
"grad_norm": 0.5308315587462934,
"learning_rate": 4.027548209903165e-06,
"loss": 0.1628,
"step": 722
},
{
"epoch": 1.1261682242990654,
"grad_norm": 0.5915750734678845,
"learning_rate": 4.01555112426808e-06,
"loss": 0.1005,
"step": 723
},
{
"epoch": 1.1277258566978192,
"grad_norm": 0.6605084387776982,
"learning_rate": 4.00355993198862e-06,
"loss": 0.1835,
"step": 724
},
{
"epoch": 1.1292834890965733,
"grad_norm": 0.5150684098328038,
"learning_rate": 3.991574704849474e-06,
"loss": 0.2452,
"step": 725
},
{
"epoch": 1.1308411214953271,
"grad_norm": 0.5762133251543371,
"learning_rate": 3.97959551459963e-06,
"loss": 0.7117,
"step": 726
},
{
"epoch": 1.132398753894081,
"grad_norm": 0.5992236027329398,
"learning_rate": 3.967622432951931e-06,
"loss": 0.2487,
"step": 727
},
{
"epoch": 1.1339563862928348,
"grad_norm": 0.6328864728376592,
"learning_rate": 3.955655531582649e-06,
"loss": 0.237,
"step": 728
},
{
"epoch": 1.1355140186915889,
"grad_norm": 0.5851714493036804,
"learning_rate": 3.943694882131061e-06,
"loss": 0.2293,
"step": 729
},
{
"epoch": 1.1370716510903427,
"grad_norm": 0.6659466040289207,
"learning_rate": 3.931740556199021e-06,
"loss": 0.7144,
"step": 730
},
{
"epoch": 1.1386292834890965,
"grad_norm": 0.8181017630361279,
"learning_rate": 3.91979262535052e-06,
"loss": 0.4389,
"step": 731
},
{
"epoch": 1.1401869158878504,
"grad_norm": 0.6310519685249386,
"learning_rate": 3.907851161111269e-06,
"loss": 0.1969,
"step": 732
},
{
"epoch": 1.1417445482866044,
"grad_norm": 0.5414429317292473,
"learning_rate": 3.895916234968267e-06,
"loss": 0.1969,
"step": 733
},
{
"epoch": 1.1433021806853583,
"grad_norm": 0.5694107952685924,
"learning_rate": 3.883987918369373e-06,
"loss": 0.2409,
"step": 734
},
{
"epoch": 1.144859813084112,
"grad_norm": 0.5303093457895504,
"learning_rate": 3.8720662827228774e-06,
"loss": 0.1599,
"step": 735
},
{
"epoch": 1.146417445482866,
"grad_norm": 0.5798145908998529,
"learning_rate": 3.860151399397077e-06,
"loss": 0.2864,
"step": 736
},
{
"epoch": 1.14797507788162,
"grad_norm": 0.7216318366333795,
"learning_rate": 3.848243339719841e-06,
"loss": 0.1274,
"step": 737
},
{
"epoch": 1.1495327102803738,
"grad_norm": 0.5363385976748041,
"learning_rate": 3.836342174978199e-06,
"loss": 0.1679,
"step": 738
},
{
"epoch": 1.1510903426791277,
"grad_norm": 0.5904714591263907,
"learning_rate": 3.824447976417897e-06,
"loss": 0.1803,
"step": 739
},
{
"epoch": 1.1526479750778815,
"grad_norm": 0.7957820087907058,
"learning_rate": 3.8125608152429777e-06,
"loss": 0.2525,
"step": 740
},
{
"epoch": 1.1542056074766356,
"grad_norm": 0.6852233460884398,
"learning_rate": 3.8006807626153565e-06,
"loss": 0.6679,
"step": 741
},
{
"epoch": 1.1557632398753894,
"grad_norm": 0.7058769394932962,
"learning_rate": 3.7888078896543984e-06,
"loss": 0.1596,
"step": 742
},
{
"epoch": 1.1573208722741433,
"grad_norm": 0.5829336602902025,
"learning_rate": 3.776942267436482e-06,
"loss": 0.1256,
"step": 743
},
{
"epoch": 1.158878504672897,
"grad_norm": 0.4936789586579176,
"learning_rate": 3.7650839669945804e-06,
"loss": 0.2213,
"step": 744
},
{
"epoch": 1.1604361370716512,
"grad_norm": 0.6225558777529572,
"learning_rate": 3.7532330593178356e-06,
"loss": 0.1667,
"step": 745
},
{
"epoch": 1.161993769470405,
"grad_norm": 0.5614342757375034,
"learning_rate": 3.741389615351136e-06,
"loss": 0.1611,
"step": 746
},
{
"epoch": 1.1635514018691588,
"grad_norm": 0.5754099281842366,
"learning_rate": 3.729553705994685e-06,
"loss": 0.2156,
"step": 747
},
{
"epoch": 1.1651090342679127,
"grad_norm": 0.5573995209639809,
"learning_rate": 3.7177254021035824e-06,
"loss": 0.1777,
"step": 748
},
{
"epoch": 1.1666666666666667,
"grad_norm": 0.6243355246035951,
"learning_rate": 3.705904774487396e-06,
"loss": 0.2092,
"step": 749
},
{
"epoch": 1.1682242990654206,
"grad_norm": 0.7072395196568583,
"learning_rate": 3.694091893909746e-06,
"loss": 0.1916,
"step": 750
},
{
"epoch": 1.1697819314641744,
"grad_norm": 0.5876756814093886,
"learning_rate": 3.6822868310878683e-06,
"loss": 0.1514,
"step": 751
},
{
"epoch": 1.1713395638629283,
"grad_norm": 0.51963445472172,
"learning_rate": 3.670489656692202e-06,
"loss": 0.2446,
"step": 752
},
{
"epoch": 1.1728971962616823,
"grad_norm": 0.6338437719274572,
"learning_rate": 3.658700441345959e-06,
"loss": 0.2041,
"step": 753
},
{
"epoch": 1.1744548286604362,
"grad_norm": 0.7029067603645407,
"learning_rate": 3.646919255624711e-06,
"loss": 0.1852,
"step": 754
},
{
"epoch": 1.17601246105919,
"grad_norm": 2.8451831521816353,
"learning_rate": 3.6351461700559564e-06,
"loss": 0.6003,
"step": 755
},
{
"epoch": 1.1775700934579438,
"grad_norm": 0.5800825083133982,
"learning_rate": 3.623381255118702e-06,
"loss": 0.1983,
"step": 756
},
{
"epoch": 1.179127725856698,
"grad_norm": 0.7108249246289187,
"learning_rate": 3.6116245812430404e-06,
"loss": 0.1904,
"step": 757
},
{
"epoch": 1.1806853582554517,
"grad_norm": 0.5537550002098082,
"learning_rate": 3.5998762188097364e-06,
"loss": 0.2697,
"step": 758
},
{
"epoch": 1.1822429906542056,
"grad_norm": 0.5354252171867416,
"learning_rate": 3.588136238149791e-06,
"loss": 0.2256,
"step": 759
},
{
"epoch": 1.1838006230529594,
"grad_norm": 0.5746205607851972,
"learning_rate": 3.5764047095440313e-06,
"loss": 0.2216,
"step": 760
},
{
"epoch": 1.1853582554517135,
"grad_norm": 0.6474421569647941,
"learning_rate": 3.5646817032226855e-06,
"loss": 0.18,
"step": 761
},
{
"epoch": 1.1869158878504673,
"grad_norm": 0.6499250531970322,
"learning_rate": 3.552967289364967e-06,
"loss": 0.1564,
"step": 762
},
{
"epoch": 1.1884735202492211,
"grad_norm": 0.6915421275577466,
"learning_rate": 3.541261538098647e-06,
"loss": 0.1937,
"step": 763
},
{
"epoch": 1.190031152647975,
"grad_norm": 0.7484332653622916,
"learning_rate": 3.529564519499641e-06,
"loss": 0.8181,
"step": 764
},
{
"epoch": 1.191588785046729,
"grad_norm": 0.5487998303234818,
"learning_rate": 3.517876303591581e-06,
"loss": 0.2187,
"step": 765
},
{
"epoch": 1.1931464174454829,
"grad_norm": 0.5468867601412329,
"learning_rate": 3.506196960345413e-06,
"loss": 0.2741,
"step": 766
},
{
"epoch": 1.1947040498442367,
"grad_norm": 0.5333599222192844,
"learning_rate": 3.494526559678958e-06,
"loss": 0.196,
"step": 767
},
{
"epoch": 1.1962616822429906,
"grad_norm": 0.6163139809793089,
"learning_rate": 3.4828651714565056e-06,
"loss": 0.1689,
"step": 768
},
{
"epoch": 1.1978193146417446,
"grad_norm": 0.6118788241013663,
"learning_rate": 3.4712128654883915e-06,
"loss": 0.6684,
"step": 769
},
{
"epoch": 1.1993769470404985,
"grad_norm": 0.7588288371409719,
"learning_rate": 3.459569711530586e-06,
"loss": 0.1844,
"step": 770
},
{
"epoch": 1.2009345794392523,
"grad_norm": 0.512022978994329,
"learning_rate": 3.447935779284265e-06,
"loss": 0.2842,
"step": 771
},
{
"epoch": 1.2024922118380061,
"grad_norm": 0.5591196557510058,
"learning_rate": 3.436311138395402e-06,
"loss": 0.1664,
"step": 772
},
{
"epoch": 1.2040498442367602,
"grad_norm": 0.8115349813699498,
"learning_rate": 3.424695858454347e-06,
"loss": 0.307,
"step": 773
},
{
"epoch": 1.205607476635514,
"grad_norm": 0.5218970414061956,
"learning_rate": 3.4130900089954142e-06,
"loss": 0.2348,
"step": 774
},
{
"epoch": 1.2071651090342679,
"grad_norm": 0.6595592104442388,
"learning_rate": 3.4014936594964608e-06,
"loss": 0.2381,
"step": 775
},
{
"epoch": 1.2087227414330217,
"grad_norm": 0.5724303041429597,
"learning_rate": 3.3899068793784717e-06,
"loss": 0.2087,
"step": 776
},
{
"epoch": 1.2102803738317758,
"grad_norm": 0.5855841777284334,
"learning_rate": 3.378329738005144e-06,
"loss": 0.2215,
"step": 777
},
{
"epoch": 1.2118380062305296,
"grad_norm": 0.5794484038490832,
"learning_rate": 3.3667623046824783e-06,
"loss": 0.1275,
"step": 778
},
{
"epoch": 1.2133956386292835,
"grad_norm": 0.8252092993882152,
"learning_rate": 3.3552046486583547e-06,
"loss": 0.2076,
"step": 779
},
{
"epoch": 1.2149532710280373,
"grad_norm": 0.5054517090774381,
"learning_rate": 3.3436568391221215e-06,
"loss": 0.2269,
"step": 780
},
{
"epoch": 1.2165109034267914,
"grad_norm": 0.6361245982706095,
"learning_rate": 3.3321189452041814e-06,
"loss": 0.1765,
"step": 781
},
{
"epoch": 1.2180685358255452,
"grad_norm": 0.6359846065805195,
"learning_rate": 3.3205910359755823e-06,
"loss": 0.213,
"step": 782
},
{
"epoch": 1.219626168224299,
"grad_norm": 0.575827824012367,
"learning_rate": 3.309073180447593e-06,
"loss": 0.2128,
"step": 783
},
{
"epoch": 1.2211838006230529,
"grad_norm": 0.594472470089166,
"learning_rate": 3.2975654475713005e-06,
"loss": 0.1867,
"step": 784
},
{
"epoch": 1.222741433021807,
"grad_norm": 0.6841555235014364,
"learning_rate": 3.286067906237188e-06,
"loss": 0.1615,
"step": 785
},
{
"epoch": 1.2242990654205608,
"grad_norm": 0.5190015178190842,
"learning_rate": 3.274580625274737e-06,
"loss": 0.2139,
"step": 786
},
{
"epoch": 1.2258566978193146,
"grad_norm": 0.675548730024154,
"learning_rate": 3.263103673451996e-06,
"loss": 0.2535,
"step": 787
},
{
"epoch": 1.2274143302180685,
"grad_norm": 0.7466550657712839,
"learning_rate": 3.2516371194751838e-06,
"loss": 0.2706,
"step": 788
},
{
"epoch": 1.2289719626168225,
"grad_norm": 0.6666523188287753,
"learning_rate": 3.24018103198827e-06,
"loss": 0.1727,
"step": 789
},
{
"epoch": 1.2305295950155763,
"grad_norm": 0.6337742576262074,
"learning_rate": 3.22873547957257e-06,
"loss": 0.1979,
"step": 790
},
{
"epoch": 1.2320872274143302,
"grad_norm": 0.5592619596660579,
"learning_rate": 3.217300530746331e-06,
"loss": 0.2195,
"step": 791
},
{
"epoch": 1.233644859813084,
"grad_norm": 0.6098769962820866,
"learning_rate": 3.2058762539643214e-06,
"loss": 0.231,
"step": 792
},
{
"epoch": 1.235202492211838,
"grad_norm": 0.5644158815873026,
"learning_rate": 3.1944627176174204e-06,
"loss": 0.2373,
"step": 793
},
{
"epoch": 1.236760124610592,
"grad_norm": 0.9387436364990357,
"learning_rate": 3.1830599900322135e-06,
"loss": 0.8066,
"step": 794
},
{
"epoch": 1.2383177570093458,
"grad_norm": 0.5202721107932754,
"learning_rate": 3.1716681394705783e-06,
"loss": 0.2198,
"step": 795
},
{
"epoch": 1.2398753894080996,
"grad_norm": 0.5496594883977267,
"learning_rate": 3.1602872341292772e-06,
"loss": 0.1961,
"step": 796
},
{
"epoch": 1.2414330218068537,
"grad_norm": 0.6098209049046375,
"learning_rate": 3.148917342139548e-06,
"loss": 0.2209,
"step": 797
},
{
"epoch": 1.2429906542056075,
"grad_norm": 0.6346192193491992,
"learning_rate": 3.1375585315667047e-06,
"loss": 0.1802,
"step": 798
},
{
"epoch": 1.2445482866043613,
"grad_norm": 0.765060316147878,
"learning_rate": 3.1262108704097137e-06,
"loss": 0.7635,
"step": 799
},
{
"epoch": 1.2461059190031152,
"grad_norm": 0.5914294283180993,
"learning_rate": 3.1148744266008024e-06,
"loss": 0.1632,
"step": 800
},
{
"epoch": 1.2476635514018692,
"grad_norm": 0.6570214224554458,
"learning_rate": 3.1035492680050416e-06,
"loss": 0.2406,
"step": 801
},
{
"epoch": 1.249221183800623,
"grad_norm": 0.5452895252692876,
"learning_rate": 3.0922354624199487e-06,
"loss": 0.2458,
"step": 802
},
{
"epoch": 1.250778816199377,
"grad_norm": 0.8021393101552716,
"learning_rate": 3.080933077575071e-06,
"loss": 0.2422,
"step": 803
},
{
"epoch": 1.2523364485981308,
"grad_norm": 0.5980931498074656,
"learning_rate": 3.0696421811315923e-06,
"loss": 0.3071,
"step": 804
},
{
"epoch": 1.2538940809968846,
"grad_norm": 0.7151706000107999,
"learning_rate": 3.0583628406819167e-06,
"loss": 0.1693,
"step": 805
},
{
"epoch": 1.2554517133956387,
"grad_norm": 0.7557667483815892,
"learning_rate": 3.0470951237492724e-06,
"loss": 0.6472,
"step": 806
},
{
"epoch": 1.2570093457943925,
"grad_norm": 0.5383018776710425,
"learning_rate": 3.035839097787301e-06,
"loss": 0.2179,
"step": 807
},
{
"epoch": 1.2585669781931463,
"grad_norm": 0.5591749963571674,
"learning_rate": 3.024594830179658e-06,
"loss": 0.2537,
"step": 808
},
{
"epoch": 1.2601246105919004,
"grad_norm": 0.5434411022961728,
"learning_rate": 3.0133623882396067e-06,
"loss": 0.1933,
"step": 809
},
{
"epoch": 1.2616822429906542,
"grad_norm": 0.6201199414157127,
"learning_rate": 3.0021418392096215e-06,
"loss": 0.2193,
"step": 810
},
{
"epoch": 1.263239875389408,
"grad_norm": 0.669872737304252,
"learning_rate": 2.9909332502609744e-06,
"loss": 0.1446,
"step": 811
},
{
"epoch": 1.264797507788162,
"grad_norm": 0.5828212724381946,
"learning_rate": 2.979736688493343e-06,
"loss": 0.253,
"step": 812
},
{
"epoch": 1.2663551401869158,
"grad_norm": 0.5644246073635343,
"learning_rate": 2.968552220934402e-06,
"loss": 0.2241,
"step": 813
},
{
"epoch": 1.2679127725856698,
"grad_norm": 0.5607183966563147,
"learning_rate": 2.9573799145394266e-06,
"loss": 0.2745,
"step": 814
},
{
"epoch": 1.2694704049844237,
"grad_norm": 0.474278573645314,
"learning_rate": 2.9462198361908893e-06,
"loss": 0.1337,
"step": 815
},
{
"epoch": 1.2710280373831775,
"grad_norm": 1.3740104676242695,
"learning_rate": 2.9350720526980592e-06,
"loss": 0.7275,
"step": 816
},
{
"epoch": 1.2725856697819315,
"grad_norm": 0.711919465609596,
"learning_rate": 2.9239366307966033e-06,
"loss": 0.7703,
"step": 817
},
{
"epoch": 1.2741433021806854,
"grad_norm": 0.6630095165519386,
"learning_rate": 2.912813637148187e-06,
"loss": 0.1362,
"step": 818
},
{
"epoch": 1.2757009345794392,
"grad_norm": 0.602978479954925,
"learning_rate": 2.9017031383400772e-06,
"loss": 0.1538,
"step": 819
},
{
"epoch": 1.277258566978193,
"grad_norm": 0.7257094178132808,
"learning_rate": 2.890605200884733e-06,
"loss": 0.1533,
"step": 820
},
{
"epoch": 1.278816199376947,
"grad_norm": 0.5041273431174441,
"learning_rate": 2.879519891219424e-06,
"loss": 0.195,
"step": 821
},
{
"epoch": 1.280373831775701,
"grad_norm": 0.5315316010733497,
"learning_rate": 2.868447275705821e-06,
"loss": 0.1634,
"step": 822
},
{
"epoch": 1.2819314641744548,
"grad_norm": 0.6131393193633405,
"learning_rate": 2.8573874206296005e-06,
"loss": 0.2044,
"step": 823
},
{
"epoch": 1.2834890965732086,
"grad_norm": 0.5970183436040595,
"learning_rate": 2.846340392200051e-06,
"loss": 0.2539,
"step": 824
},
{
"epoch": 1.2850467289719627,
"grad_norm": 0.5065607828613443,
"learning_rate": 2.8353062565496715e-06,
"loss": 0.166,
"step": 825
},
{
"epoch": 1.2866043613707165,
"grad_norm": 0.6837585450327539,
"learning_rate": 2.824285079733788e-06,
"loss": 0.2406,
"step": 826
},
{
"epoch": 1.2881619937694704,
"grad_norm": 0.5599124367538377,
"learning_rate": 2.8132769277301374e-06,
"loss": 0.157,
"step": 827
},
{
"epoch": 1.2897196261682242,
"grad_norm": 0.49511182258884123,
"learning_rate": 2.8022818664384945e-06,
"loss": 0.1824,
"step": 828
},
{
"epoch": 1.291277258566978,
"grad_norm": 0.5896227768783965,
"learning_rate": 2.791299961680255e-06,
"loss": 0.258,
"step": 829
},
{
"epoch": 1.2928348909657321,
"grad_norm": 0.5730839746888057,
"learning_rate": 2.7803312791980697e-06,
"loss": 0.2379,
"step": 830
},
{
"epoch": 1.294392523364486,
"grad_norm": 0.6736027436106719,
"learning_rate": 2.769375884655421e-06,
"loss": 0.2515,
"step": 831
},
{
"epoch": 1.2959501557632398,
"grad_norm": 0.6242264709025079,
"learning_rate": 2.758433843636252e-06,
"loss": 0.1923,
"step": 832
},
{
"epoch": 1.2975077881619939,
"grad_norm": 0.6286501649370759,
"learning_rate": 2.7475052216445588e-06,
"loss": 0.1842,
"step": 833
},
{
"epoch": 1.2990654205607477,
"grad_norm": 0.5914028874575505,
"learning_rate": 2.7365900841040104e-06,
"loss": 0.1473,
"step": 834
},
{
"epoch": 1.3006230529595015,
"grad_norm": 0.6185390608342917,
"learning_rate": 2.7256884963575536e-06,
"loss": 0.2347,
"step": 835
},
{
"epoch": 1.3021806853582554,
"grad_norm": 0.5490349592608532,
"learning_rate": 2.714800523667011e-06,
"loss": 0.1875,
"step": 836
},
{
"epoch": 1.3037383177570092,
"grad_norm": 0.619509184599072,
"learning_rate": 2.703926231212708e-06,
"loss": 0.2472,
"step": 837
},
{
"epoch": 1.3052959501557633,
"grad_norm": 0.7158852617243849,
"learning_rate": 2.6930656840930713e-06,
"loss": 0.2468,
"step": 838
},
{
"epoch": 1.3068535825545171,
"grad_norm": 0.723996693315863,
"learning_rate": 2.6822189473242422e-06,
"loss": 0.2276,
"step": 839
},
{
"epoch": 1.308411214953271,
"grad_norm": 0.5632679246702721,
"learning_rate": 2.671386085839682e-06,
"loss": 0.2215,
"step": 840
},
{
"epoch": 1.309968847352025,
"grad_norm": 0.5985281299233187,
"learning_rate": 2.6605671644897967e-06,
"loss": 0.3572,
"step": 841
},
{
"epoch": 1.3115264797507789,
"grad_norm": 0.686365038362172,
"learning_rate": 2.6497622480415346e-06,
"loss": 0.3266,
"step": 842
},
{
"epoch": 1.3130841121495327,
"grad_norm": 0.6582167035193466,
"learning_rate": 2.6389714011780078e-06,
"loss": 0.2669,
"step": 843
},
{
"epoch": 1.3146417445482865,
"grad_norm": 0.5788593198598405,
"learning_rate": 2.628194688498096e-06,
"loss": 0.2572,
"step": 844
},
{
"epoch": 1.3161993769470404,
"grad_norm": 0.5457859405003289,
"learning_rate": 2.6174321745160707e-06,
"loss": 0.2016,
"step": 845
},
{
"epoch": 1.3177570093457944,
"grad_norm": 0.5093352344419979,
"learning_rate": 2.606683923661204e-06,
"loss": 0.2152,
"step": 846
},
{
"epoch": 1.3193146417445483,
"grad_norm": 0.6332315738079737,
"learning_rate": 2.5959500002773763e-06,
"loss": 0.1449,
"step": 847
},
{
"epoch": 1.320872274143302,
"grad_norm": 0.4685817262067038,
"learning_rate": 2.585230468622705e-06,
"loss": 0.2152,
"step": 848
},
{
"epoch": 1.3224299065420562,
"grad_norm": 0.49981674536826676,
"learning_rate": 2.574525392869143e-06,
"loss": 0.1783,
"step": 849
},
{
"epoch": 1.32398753894081,
"grad_norm": 1.0724140185603979,
"learning_rate": 2.563834837102115e-06,
"loss": 0.6521,
"step": 850
},
{
"epoch": 1.3255451713395638,
"grad_norm": 0.5582924411201017,
"learning_rate": 2.5531588653201134e-06,
"loss": 0.1696,
"step": 851
},
{
"epoch": 1.3271028037383177,
"grad_norm": 0.5064036048508646,
"learning_rate": 2.542497541434329e-06,
"loss": 0.1616,
"step": 852
},
{
"epoch": 1.3286604361370715,
"grad_norm": 0.7424483742643145,
"learning_rate": 2.531850929268258e-06,
"loss": 0.3,
"step": 853
},
{
"epoch": 1.3302180685358256,
"grad_norm": 0.6230465177616844,
"learning_rate": 2.5212190925573366e-06,
"loss": 0.638,
"step": 854
},
{
"epoch": 1.3317757009345794,
"grad_norm": 0.5918692120070366,
"learning_rate": 2.510602094948535e-06,
"loss": 0.1742,
"step": 855
},
{
"epoch": 1.3333333333333333,
"grad_norm": 0.6010294004442239,
"learning_rate": 2.5000000000000015e-06,
"loss": 0.2177,
"step": 856
},
{
"epoch": 1.3348909657320873,
"grad_norm": 0.8792131655034126,
"learning_rate": 2.4894128711806603e-06,
"loss": 0.1769,
"step": 857
},
{
"epoch": 1.3364485981308412,
"grad_norm": 0.5439058673773991,
"learning_rate": 2.4788407718698487e-06,
"loss": 0.1423,
"step": 858
},
{
"epoch": 1.338006230529595,
"grad_norm": 0.6698324814866782,
"learning_rate": 2.468283765356931e-06,
"loss": 0.2016,
"step": 859
},
{
"epoch": 1.3395638629283488,
"grad_norm": 0.7209231410413558,
"learning_rate": 2.4577419148409123e-06,
"loss": 0.2526,
"step": 860
},
{
"epoch": 1.3411214953271027,
"grad_norm": 0.5554513132189719,
"learning_rate": 2.447215283430072e-06,
"loss": 0.2054,
"step": 861
},
{
"epoch": 1.3426791277258567,
"grad_norm": 0.6204181777681318,
"learning_rate": 2.436703934141583e-06,
"loss": 0.2055,
"step": 862
},
{
"epoch": 1.3442367601246106,
"grad_norm": 0.6174847596208328,
"learning_rate": 2.4262079299011287e-06,
"loss": 0.2214,
"step": 863
},
{
"epoch": 1.3457943925233644,
"grad_norm": 0.596347448095447,
"learning_rate": 2.4157273335425296e-06,
"loss": 0.2396,
"step": 864
},
{
"epoch": 1.3473520249221185,
"grad_norm": 0.6945784756475979,
"learning_rate": 2.4052622078073714e-06,
"loss": 0.6505,
"step": 865
},
{
"epoch": 1.3489096573208723,
"grad_norm": 0.6484352021895371,
"learning_rate": 2.394812615344622e-06,
"loss": 0.2738,
"step": 866
},
{
"epoch": 1.3504672897196262,
"grad_norm": 0.5351329498045785,
"learning_rate": 2.384378618710265e-06,
"loss": 0.1667,
"step": 867
},
{
"epoch": 1.35202492211838,
"grad_norm": 0.5884847134807437,
"learning_rate": 2.3739602803669126e-06,
"loss": 0.2267,
"step": 868
},
{
"epoch": 1.3535825545171338,
"grad_norm": 0.7327148616338638,
"learning_rate": 2.363557662683446e-06,
"loss": 0.7327,
"step": 869
},
{
"epoch": 1.355140186915888,
"grad_norm": 0.712258388986809,
"learning_rate": 2.3531708279346347e-06,
"loss": 0.2281,
"step": 870
},
{
"epoch": 1.3566978193146417,
"grad_norm": 0.598978404959667,
"learning_rate": 2.3427998383007605e-06,
"loss": 0.2433,
"step": 871
},
{
"epoch": 1.3582554517133956,
"grad_norm": 0.6295698566133894,
"learning_rate": 2.3324447558672543e-06,
"loss": 0.2569,
"step": 872
},
{
"epoch": 1.3598130841121496,
"grad_norm": 0.5986998801181967,
"learning_rate": 2.3221056426243112e-06,
"loss": 0.1962,
"step": 873
},
{
"epoch": 1.3613707165109035,
"grad_norm": 0.5674067782697227,
"learning_rate": 2.3117825604665405e-06,
"loss": 0.2158,
"step": 874
},
{
"epoch": 1.3629283489096573,
"grad_norm": 0.6884839665928056,
"learning_rate": 2.3014755711925695e-06,
"loss": 0.2329,
"step": 875
},
{
"epoch": 1.3644859813084111,
"grad_norm": 0.6885485571502845,
"learning_rate": 2.291184736504695e-06,
"loss": 0.1955,
"step": 876
},
{
"epoch": 1.366043613707165,
"grad_norm": 0.5473669303308986,
"learning_rate": 2.2809101180084963e-06,
"loss": 0.2171,
"step": 877
},
{
"epoch": 1.367601246105919,
"grad_norm": 0.5860469422390239,
"learning_rate": 2.2706517772124875e-06,
"loss": 0.3142,
"step": 878
},
{
"epoch": 1.3691588785046729,
"grad_norm": 0.4589130720603424,
"learning_rate": 2.2604097755277244e-06,
"loss": 0.1892,
"step": 879
},
{
"epoch": 1.3707165109034267,
"grad_norm": 0.56817784386682,
"learning_rate": 2.250184174267459e-06,
"loss": 0.2104,
"step": 880
},
{
"epoch": 1.3722741433021808,
"grad_norm": 2.9170440438070244,
"learning_rate": 2.239975034646756e-06,
"loss": 0.4075,
"step": 881
},
{
"epoch": 1.3738317757009346,
"grad_norm": 0.5936408877749505,
"learning_rate": 2.2297824177821374e-06,
"loss": 0.1725,
"step": 882
},
{
"epoch": 1.3753894080996885,
"grad_norm": 0.5570669743525561,
"learning_rate": 2.2196063846912142e-06,
"loss": 0.1089,
"step": 883
},
{
"epoch": 1.3769470404984423,
"grad_norm": 0.7043574775467037,
"learning_rate": 2.209446996292312e-06,
"loss": 0.2412,
"step": 884
},
{
"epoch": 1.3785046728971961,
"grad_norm": 0.6356878135461869,
"learning_rate": 2.1993043134041213e-06,
"loss": 0.2314,
"step": 885
},
{
"epoch": 1.3800623052959502,
"grad_norm": 0.6088173529635026,
"learning_rate": 2.1891783967453235e-06,
"loss": 0.1986,
"step": 886
},
{
"epoch": 1.381619937694704,
"grad_norm": 0.6505134414780895,
"learning_rate": 2.1790693069342293e-06,
"loss": 0.2621,
"step": 887
},
{
"epoch": 1.3831775700934579,
"grad_norm": 0.5229082442758375,
"learning_rate": 2.168977104488415e-06,
"loss": 0.1857,
"step": 888
},
{
"epoch": 1.384735202492212,
"grad_norm": 0.6219752894683416,
"learning_rate": 2.158901849824363e-06,
"loss": 0.2393,
"step": 889
},
{
"epoch": 1.3862928348909658,
"grad_norm": 0.6712717762459158,
"learning_rate": 2.1488436032571e-06,
"loss": 0.1975,
"step": 890
},
{
"epoch": 1.3878504672897196,
"grad_norm": 0.5882444001428154,
"learning_rate": 2.138802424999833e-06,
"loss": 0.1938,
"step": 891
},
{
"epoch": 1.3894080996884735,
"grad_norm": 0.7887785551105917,
"learning_rate": 2.128778375163588e-06,
"loss": 0.2438,
"step": 892
},
{
"epoch": 1.3909657320872273,
"grad_norm": 0.6726478463164812,
"learning_rate": 2.1187715137568564e-06,
"loss": 0.1609,
"step": 893
},
{
"epoch": 1.3925233644859814,
"grad_norm": 1.1772556659919378,
"learning_rate": 2.1087819006852327e-06,
"loss": 0.1874,
"step": 894
},
{
"epoch": 1.3940809968847352,
"grad_norm": 0.6141735195326121,
"learning_rate": 2.098809595751049e-06,
"loss": 0.1771,
"step": 895
},
{
"epoch": 1.395638629283489,
"grad_norm": 0.574616048911299,
"learning_rate": 2.08885465865303e-06,
"loss": 0.1973,
"step": 896
},
{
"epoch": 1.397196261682243,
"grad_norm": 0.7291041551854346,
"learning_rate": 2.078917148985921e-06,
"loss": 0.7534,
"step": 897
},
{
"epoch": 1.398753894080997,
"grad_norm": 0.5907252824302475,
"learning_rate": 2.06899712624015e-06,
"loss": 0.2802,
"step": 898
},
{
"epoch": 1.4003115264797508,
"grad_norm": 0.5328199626427012,
"learning_rate": 2.0590946498014493e-06,
"loss": 0.2303,
"step": 899
},
{
"epoch": 1.4018691588785046,
"grad_norm": 0.6150130691322249,
"learning_rate": 2.049209778950518e-06,
"loss": 0.2424,
"step": 900
},
{
"epoch": 1.4034267912772584,
"grad_norm": 0.6127344165171512,
"learning_rate": 2.0393425728626526e-06,
"loss": 0.1674,
"step": 901
},
{
"epoch": 1.4049844236760125,
"grad_norm": 0.6455010665412071,
"learning_rate": 2.029493090607413e-06,
"loss": 0.201,
"step": 902
},
{
"epoch": 1.4065420560747663,
"grad_norm": 0.6680633935331085,
"learning_rate": 2.0196613911482445e-06,
"loss": 0.1652,
"step": 903
},
{
"epoch": 1.4080996884735202,
"grad_norm": 0.5230495421575583,
"learning_rate": 2.0098475333421426e-06,
"loss": 0.1869,
"step": 904
},
{
"epoch": 1.4096573208722742,
"grad_norm": 0.6484127593933029,
"learning_rate": 2.0000515759392904e-06,
"loss": 0.2167,
"step": 905
},
{
"epoch": 1.411214953271028,
"grad_norm": 0.6594541288417016,
"learning_rate": 1.9902735775827146e-06,
"loss": 0.3135,
"step": 906
},
{
"epoch": 1.412772585669782,
"grad_norm": 0.6546159891601314,
"learning_rate": 1.980513596807932e-06,
"loss": 0.1546,
"step": 907
},
{
"epoch": 1.4143302180685358,
"grad_norm": 0.6853904172880749,
"learning_rate": 1.9707716920425923e-06,
"loss": 0.2063,
"step": 908
},
{
"epoch": 1.4158878504672896,
"grad_norm": 0.6131385223592029,
"learning_rate": 1.9610479216061393e-06,
"loss": 0.2792,
"step": 909
},
{
"epoch": 1.4174454828660437,
"grad_norm": 0.47832488985489935,
"learning_rate": 1.951342343709456e-06,
"loss": 0.2819,
"step": 910
},
{
"epoch": 1.4190031152647975,
"grad_norm": 0.7260584028544175,
"learning_rate": 1.941655016454515e-06,
"loss": 0.1664,
"step": 911
},
{
"epoch": 1.4205607476635513,
"grad_norm": 0.531050945261339,
"learning_rate": 1.9319859978340312e-06,
"loss": 0.218,
"step": 912
},
{
"epoch": 1.4221183800623054,
"grad_norm": 0.7077461606190044,
"learning_rate": 1.9223353457311178e-06,
"loss": 0.2129,
"step": 913
},
{
"epoch": 1.4236760124610592,
"grad_norm": 0.6914074130890262,
"learning_rate": 1.912703117918938e-06,
"loss": 0.1846,
"step": 914
},
{
"epoch": 1.425233644859813,
"grad_norm": 0.7199368018705896,
"learning_rate": 1.9030893720603604e-06,
"loss": 0.2253,
"step": 915
},
{
"epoch": 1.426791277258567,
"grad_norm": 0.5583301531155793,
"learning_rate": 1.8934941657076057e-06,
"loss": 0.2059,
"step": 916
},
{
"epoch": 1.4283489096573208,
"grad_norm": 0.5646047767722963,
"learning_rate": 1.8839175563019145e-06,
"loss": 0.2514,
"step": 917
},
{
"epoch": 1.4299065420560748,
"grad_norm": 0.6817330899431177,
"learning_rate": 1.8743596011731984e-06,
"loss": 0.2018,
"step": 918
},
{
"epoch": 1.4314641744548287,
"grad_norm": 0.6478720709719429,
"learning_rate": 1.8648203575396912e-06,
"loss": 0.1172,
"step": 919
},
{
"epoch": 1.4330218068535825,
"grad_norm": 0.6481753359929265,
"learning_rate": 1.8552998825076163e-06,
"loss": 0.2396,
"step": 920
},
{
"epoch": 1.4345794392523366,
"grad_norm": 0.6132125740119907,
"learning_rate": 1.845798233070833e-06,
"loss": 0.2065,
"step": 921
},
{
"epoch": 1.4361370716510904,
"grad_norm": 0.5448210616281429,
"learning_rate": 1.8363154661105131e-06,
"loss": 0.2022,
"step": 922
},
{
"epoch": 1.4376947040498442,
"grad_norm": 0.5931140592473454,
"learning_rate": 1.826851638394777e-06,
"loss": 0.2241,
"step": 923
},
{
"epoch": 1.439252336448598,
"grad_norm": 0.5732273344982183,
"learning_rate": 1.8174068065783768e-06,
"loss": 0.238,
"step": 924
},
{
"epoch": 1.440809968847352,
"grad_norm": 0.6031290862741796,
"learning_rate": 1.8079810272023373e-06,
"loss": 0.2065,
"step": 925
},
{
"epoch": 1.442367601246106,
"grad_norm": 0.7027323074692515,
"learning_rate": 1.7985743566936338e-06,
"loss": 0.1353,
"step": 926
},
{
"epoch": 1.4439252336448598,
"grad_norm": 0.679661751499936,
"learning_rate": 1.7891868513648436e-06,
"loss": 0.1636,
"step": 927
},
{
"epoch": 1.4454828660436136,
"grad_norm": 0.6123098696997167,
"learning_rate": 1.7798185674138163e-06,
"loss": 0.7362,
"step": 928
},
{
"epoch": 1.4470404984423677,
"grad_norm": 0.6533459270958951,
"learning_rate": 1.7704695609233275e-06,
"loss": 0.2479,
"step": 929
},
{
"epoch": 1.4485981308411215,
"grad_norm": 0.6018511814849575,
"learning_rate": 1.7611398878607544e-06,
"loss": 0.2135,
"step": 930
},
{
"epoch": 1.4501557632398754,
"grad_norm": 0.6980529317039801,
"learning_rate": 1.7518296040777355e-06,
"loss": 0.1639,
"step": 931
},
{
"epoch": 1.4517133956386292,
"grad_norm": 0.6062016861868827,
"learning_rate": 1.742538765309832e-06,
"loss": 0.2538,
"step": 932
},
{
"epoch": 1.453271028037383,
"grad_norm": 0.6539624285807091,
"learning_rate": 1.7332674271762018e-06,
"loss": 0.1811,
"step": 933
},
{
"epoch": 1.4548286604361371,
"grad_norm": 0.7122687298587204,
"learning_rate": 1.724015645179264e-06,
"loss": 0.1891,
"step": 934
},
{
"epoch": 1.456386292834891,
"grad_norm": 0.7014287750312252,
"learning_rate": 1.7147834747043651e-06,
"loss": 0.1866,
"step": 935
},
{
"epoch": 1.4579439252336448,
"grad_norm": 0.6319275863012598,
"learning_rate": 1.7055709710194452e-06,
"loss": 0.1921,
"step": 936
},
{
"epoch": 1.4595015576323989,
"grad_norm": 0.6620029301577026,
"learning_rate": 1.6963781892747128e-06,
"loss": 0.1889,
"step": 937
},
{
"epoch": 1.4610591900311527,
"grad_norm": 0.6613697258053355,
"learning_rate": 1.6872051845023124e-06,
"loss": 0.2277,
"step": 938
},
{
"epoch": 1.4626168224299065,
"grad_norm": 0.6709827699093153,
"learning_rate": 1.6780520116159954e-06,
"loss": 0.2345,
"step": 939
},
{
"epoch": 1.4641744548286604,
"grad_norm": 0.6308994122097046,
"learning_rate": 1.6689187254107852e-06,
"loss": 0.2384,
"step": 940
},
{
"epoch": 1.4657320872274142,
"grad_norm": 0.49997735693518447,
"learning_rate": 1.6598053805626607e-06,
"loss": 0.2885,
"step": 941
},
{
"epoch": 1.4672897196261683,
"grad_norm": 0.5769855416317906,
"learning_rate": 1.6507120316282222e-06,
"loss": 0.2262,
"step": 942
},
{
"epoch": 1.4688473520249221,
"grad_norm": 0.7766043611879518,
"learning_rate": 1.6416387330443594e-06,
"loss": 0.2012,
"step": 943
},
{
"epoch": 1.470404984423676,
"grad_norm": 0.6306011227357949,
"learning_rate": 1.6325855391279399e-06,
"loss": 0.1672,
"step": 944
},
{
"epoch": 1.47196261682243,
"grad_norm": 0.6252880509484167,
"learning_rate": 1.623552504075467e-06,
"loss": 0.21,
"step": 945
},
{
"epoch": 1.4735202492211839,
"grad_norm": 0.5733165288371812,
"learning_rate": 1.6145396819627734e-06,
"loss": 0.203,
"step": 946
},
{
"epoch": 1.4750778816199377,
"grad_norm": 0.7571650504143702,
"learning_rate": 1.6055471267446804e-06,
"loss": 0.2244,
"step": 947
},
{
"epoch": 1.4766355140186915,
"grad_norm": 0.7478203882298783,
"learning_rate": 1.5965748922546876e-06,
"loss": 0.1799,
"step": 948
},
{
"epoch": 1.4781931464174454,
"grad_norm": 0.6868312970173284,
"learning_rate": 1.5876230322046393e-06,
"loss": 0.6746,
"step": 949
},
{
"epoch": 1.4797507788161994,
"grad_norm": 0.6037632401087822,
"learning_rate": 1.578691600184416e-06,
"loss": 0.2057,
"step": 950
},
{
"epoch": 1.4813084112149533,
"grad_norm": 0.6046826777481786,
"learning_rate": 1.569780649661603e-06,
"loss": 0.2926,
"step": 951
},
{
"epoch": 1.482866043613707,
"grad_norm": 0.6848611609536812,
"learning_rate": 1.5608902339811765e-06,
"loss": 0.2328,
"step": 952
},
{
"epoch": 1.4844236760124612,
"grad_norm": 0.5462582273781476,
"learning_rate": 1.5520204063651784e-06,
"loss": 0.2703,
"step": 953
},
{
"epoch": 1.485981308411215,
"grad_norm": 0.5459862014182448,
"learning_rate": 1.5431712199124033e-06,
"loss": 0.2268,
"step": 954
},
{
"epoch": 1.4875389408099688,
"grad_norm": 0.5909228089011523,
"learning_rate": 1.5343427275980804e-06,
"loss": 0.1918,
"step": 955
},
{
"epoch": 1.4890965732087227,
"grad_norm": 0.6922392218824316,
"learning_rate": 1.5255349822735494e-06,
"loss": 0.2065,
"step": 956
},
{
"epoch": 1.4906542056074765,
"grad_norm": 0.6347770175659823,
"learning_rate": 1.5167480366659538e-06,
"loss": 0.1683,
"step": 957
},
{
"epoch": 1.4922118380062306,
"grad_norm": 0.5516070303093964,
"learning_rate": 1.507981943377918e-06,
"loss": 0.2173,
"step": 958
},
{
"epoch": 1.4937694704049844,
"grad_norm": 0.6071049706243482,
"learning_rate": 1.4992367548872383e-06,
"loss": 0.2458,
"step": 959
},
{
"epoch": 1.4953271028037383,
"grad_norm": 0.6720642183082266,
"learning_rate": 1.490512523546559e-06,
"loss": 0.1915,
"step": 960
},
{
"epoch": 1.4968847352024923,
"grad_norm": 0.5147991401453758,
"learning_rate": 1.481809301583072e-06,
"loss": 0.2315,
"step": 961
},
{
"epoch": 1.4984423676012462,
"grad_norm": 0.5193622138782028,
"learning_rate": 1.4731271410981961e-06,
"loss": 0.2994,
"step": 962
},
{
"epoch": 1.5,
"grad_norm": 0.6483302843412619,
"learning_rate": 1.4644660940672628e-06,
"loss": 0.1486,
"step": 963
},
{
"epoch": 1.5015576323987538,
"grad_norm": 0.6211856895290482,
"learning_rate": 1.4558262123392125e-06,
"loss": 0.1979,
"step": 964
},
{
"epoch": 1.5031152647975077,
"grad_norm": 0.6267158569891346,
"learning_rate": 1.4472075476362819e-06,
"loss": 0.226,
"step": 965
},
{
"epoch": 1.5046728971962615,
"grad_norm": 0.6878081812868871,
"learning_rate": 1.4386101515536865e-06,
"loss": 0.215,
"step": 966
},
{
"epoch": 1.5062305295950156,
"grad_norm": 0.6013152247422209,
"learning_rate": 1.4300340755593256e-06,
"loss": 0.1853,
"step": 967
},
{
"epoch": 1.5077881619937694,
"grad_norm": 0.5961333640534994,
"learning_rate": 1.4214793709934644e-06,
"loss": 0.2134,
"step": 968
},
{
"epoch": 1.5093457943925235,
"grad_norm": 0.6714185068729337,
"learning_rate": 1.4129460890684255e-06,
"loss": 0.2285,
"step": 969
},
{
"epoch": 1.5109034267912773,
"grad_norm": 0.5287976544223729,
"learning_rate": 1.4044342808682904e-06,
"loss": 0.226,
"step": 970
},
{
"epoch": 1.5124610591900312,
"grad_norm": 0.7244263964282488,
"learning_rate": 1.3959439973485872e-06,
"loss": 0.1919,
"step": 971
},
{
"epoch": 1.514018691588785,
"grad_norm": 0.7411473501185608,
"learning_rate": 1.38747528933599e-06,
"loss": 0.1573,
"step": 972
},
{
"epoch": 1.5155763239875388,
"grad_norm": 0.7535346088046625,
"learning_rate": 1.3790282075280064e-06,
"loss": 0.1939,
"step": 973
},
{
"epoch": 1.5171339563862927,
"grad_norm": 0.7535256340431424,
"learning_rate": 1.3706028024926855e-06,
"loss": 0.1736,
"step": 974
},
{
"epoch": 1.5186915887850467,
"grad_norm": 0.5709920085521653,
"learning_rate": 1.3621991246683093e-06,
"loss": 0.222,
"step": 975
},
{
"epoch": 1.5202492211838006,
"grad_norm": 0.7297081050093409,
"learning_rate": 1.3538172243630848e-06,
"loss": 0.1788,
"step": 976
},
{
"epoch": 1.5218068535825546,
"grad_norm": 0.6180657113624616,
"learning_rate": 1.3454571517548554e-06,
"loss": 0.177,
"step": 977
},
{
"epoch": 1.5233644859813085,
"grad_norm": 0.591364040153287,
"learning_rate": 1.3371189568907915e-06,
"loss": 0.2128,
"step": 978
},
{
"epoch": 1.5249221183800623,
"grad_norm": 0.5022261849806481,
"learning_rate": 1.3288026896870964e-06,
"loss": 0.2267,
"step": 979
},
{
"epoch": 1.5264797507788161,
"grad_norm": 0.6203532747393247,
"learning_rate": 1.3205083999286973e-06,
"loss": 0.1955,
"step": 980
},
{
"epoch": 1.52803738317757,
"grad_norm": 0.6729761565050358,
"learning_rate": 1.3122361372689624e-06,
"loss": 0.1657,
"step": 981
},
{
"epoch": 1.5295950155763238,
"grad_norm": 0.6536059366129146,
"learning_rate": 1.3039859512293885e-06,
"loss": 0.2198,
"step": 982
},
{
"epoch": 1.5311526479750779,
"grad_norm": 0.695539877357984,
"learning_rate": 1.2957578911993212e-06,
"loss": 0.2023,
"step": 983
},
{
"epoch": 1.5327102803738317,
"grad_norm": 0.4424609062997814,
"learning_rate": 1.28755200643564e-06,
"loss": 0.2238,
"step": 984
},
{
"epoch": 1.5342679127725858,
"grad_norm": 0.5843664425742057,
"learning_rate": 1.27936834606248e-06,
"loss": 0.2197,
"step": 985
},
{
"epoch": 1.5358255451713396,
"grad_norm": 0.6148367721669108,
"learning_rate": 1.2712069590709265e-06,
"loss": 0.2004,
"step": 986
},
{
"epoch": 1.5373831775700935,
"grad_norm": 0.562472309572901,
"learning_rate": 1.2630678943187292e-06,
"loss": 0.1835,
"step": 987
},
{
"epoch": 1.5389408099688473,
"grad_norm": 0.7987485306921659,
"learning_rate": 1.2549512005300068e-06,
"loss": 0.1995,
"step": 988
},
{
"epoch": 1.5404984423676011,
"grad_norm": 0.6656097374020112,
"learning_rate": 1.246856926294953e-06,
"loss": 0.1874,
"step": 989
},
{
"epoch": 1.542056074766355,
"grad_norm": 0.7982996720144556,
"learning_rate": 1.23878512006955e-06,
"loss": 0.351,
"step": 990
},
{
"epoch": 1.543613707165109,
"grad_norm": 0.7237322332450993,
"learning_rate": 1.2307358301752753e-06,
"loss": 0.6736,
"step": 991
},
{
"epoch": 1.5451713395638629,
"grad_norm": 0.6606058936315011,
"learning_rate": 1.222709104798816e-06,
"loss": 0.2446,
"step": 992
},
{
"epoch": 1.546728971962617,
"grad_norm": 0.7491652827965601,
"learning_rate": 1.214704991991773e-06,
"loss": 0.1912,
"step": 993
},
{
"epoch": 1.5482866043613708,
"grad_norm": 0.5316071822548012,
"learning_rate": 1.2067235396703819e-06,
"loss": 0.2006,
"step": 994
},
{
"epoch": 1.5498442367601246,
"grad_norm": 0.4870862919772457,
"learning_rate": 1.198764795615221e-06,
"loss": 0.1721,
"step": 995
},
{
"epoch": 1.5514018691588785,
"grad_norm": 0.45786723638909316,
"learning_rate": 1.190828807470929e-06,
"loss": 0.2847,
"step": 996
},
{
"epoch": 1.5529595015576323,
"grad_norm": 0.5220478168315765,
"learning_rate": 1.182915622745912e-06,
"loss": 0.1942,
"step": 997
},
{
"epoch": 1.5545171339563861,
"grad_norm": 1.0355443091689227,
"learning_rate": 1.17502528881207e-06,
"loss": 0.6648,
"step": 998
},
{
"epoch": 1.5560747663551402,
"grad_norm": 0.6645917404798355,
"learning_rate": 1.1671578529045075e-06,
"loss": 0.1686,
"step": 999
},
{
"epoch": 1.557632398753894,
"grad_norm": 0.5772980090930082,
"learning_rate": 1.1593133621212454e-06,
"loss": 0.1601,
"step": 1000
},
{
"epoch": 1.557632398753894,
"eval_loss": 0.36974087357521057,
"eval_runtime": 2.7835,
"eval_samples_per_second": 9.341,
"eval_steps_per_second": 2.515,
"step": 1000
},
{
"epoch": 1.559190031152648,
"grad_norm": 0.6634169549696792,
"learning_rate": 1.151491863422951e-06,
"loss": 0.178,
"step": 1001
},
{
"epoch": 1.560747663551402,
"grad_norm": 0.6293088334490837,
"learning_rate": 1.1436934036326485e-06,
"loss": 0.2054,
"step": 1002
},
{
"epoch": 1.5623052959501558,
"grad_norm": 0.6633410484590091,
"learning_rate": 1.135918029435441e-06,
"loss": 0.2654,
"step": 1003
},
{
"epoch": 1.5638629283489096,
"grad_norm": 0.6622964143754947,
"learning_rate": 1.1281657873782287e-06,
"loss": 0.2793,
"step": 1004
},
{
"epoch": 1.5654205607476634,
"grad_norm": 0.6119616717415093,
"learning_rate": 1.1204367238694357e-06,
"loss": 0.2711,
"step": 1005
},
{
"epoch": 1.5669781931464173,
"grad_norm": 0.47949709495438925,
"learning_rate": 1.1127308851787244e-06,
"loss": 0.2021,
"step": 1006
},
{
"epoch": 1.5685358255451713,
"grad_norm": 0.6934647914219084,
"learning_rate": 1.1050483174367304e-06,
"loss": 0.203,
"step": 1007
},
{
"epoch": 1.5700934579439252,
"grad_norm": 0.5481726220781126,
"learning_rate": 1.0973890666347703e-06,
"loss": 0.2368,
"step": 1008
},
{
"epoch": 1.5716510903426792,
"grad_norm": 0.6509857110936647,
"learning_rate": 1.0897531786245819e-06,
"loss": 0.2125,
"step": 1009
},
{
"epoch": 1.573208722741433,
"grad_norm": 0.5617926695948265,
"learning_rate": 1.0821406991180367e-06,
"loss": 0.2398,
"step": 1010
},
{
"epoch": 1.574766355140187,
"grad_norm": 0.6111199895380683,
"learning_rate": 1.0745516736868766e-06,
"loss": 0.2485,
"step": 1011
},
{
"epoch": 1.5763239875389408,
"grad_norm": 0.7124396857915882,
"learning_rate": 1.0669861477624376e-06,
"loss": 0.2517,
"step": 1012
},
{
"epoch": 1.5778816199376946,
"grad_norm": 0.5990035454998449,
"learning_rate": 1.0594441666353722e-06,
"loss": 0.151,
"step": 1013
},
{
"epoch": 1.5794392523364484,
"grad_norm": 0.6520260923864126,
"learning_rate": 1.0519257754553868e-06,
"loss": 0.1888,
"step": 1014
},
{
"epoch": 1.5809968847352025,
"grad_norm": 0.7169934380455352,
"learning_rate": 1.0444310192309675e-06,
"loss": 0.2318,
"step": 1015
},
{
"epoch": 1.5825545171339563,
"grad_norm": 0.5993914903355098,
"learning_rate": 1.0369599428291116e-06,
"loss": 0.2124,
"step": 1016
},
{
"epoch": 1.5841121495327104,
"grad_norm": 0.49513044936794753,
"learning_rate": 1.0295125909750537e-06,
"loss": 0.2402,
"step": 1017
},
{
"epoch": 1.5856697819314642,
"grad_norm": 0.7226758243950167,
"learning_rate": 1.0220890082520074e-06,
"loss": 0.2398,
"step": 1018
},
{
"epoch": 1.587227414330218,
"grad_norm": 0.6288423009834784,
"learning_rate": 1.0146892391008929e-06,
"loss": 0.1948,
"step": 1019
},
{
"epoch": 1.588785046728972,
"grad_norm": 0.596604250578145,
"learning_rate": 1.0073133278200702e-06,
"loss": 0.232,
"step": 1020
},
{
"epoch": 1.5903426791277258,
"grad_norm": 0.6135147035860976,
"learning_rate": 9.999613185650748e-07,
"loss": 0.2614,
"step": 1021
},
{
"epoch": 1.5919003115264796,
"grad_norm": 0.5404711008426897,
"learning_rate": 9.926332553483547e-07,
"loss": 0.1885,
"step": 1022
},
{
"epoch": 1.5934579439252337,
"grad_norm": 0.7031398847787177,
"learning_rate": 9.853291820390087e-07,
"loss": 0.2325,
"step": 1023
},
{
"epoch": 1.5950155763239875,
"grad_norm": 0.6348294111355727,
"learning_rate": 9.780491423625154e-07,
"loss": 0.2385,
"step": 1024
},
{
"epoch": 1.5965732087227416,
"grad_norm": 0.5521768442130486,
"learning_rate": 9.70793179900484e-07,
"loss": 0.1988,
"step": 1025
},
{
"epoch": 1.5981308411214954,
"grad_norm": 0.613980141544231,
"learning_rate": 9.63561338090378e-07,
"loss": 0.1027,
"step": 1026
},
{
"epoch": 1.5996884735202492,
"grad_norm": 0.7720687456198224,
"learning_rate": 9.563536602252749e-07,
"loss": 0.2905,
"step": 1027
},
{
"epoch": 1.601246105919003,
"grad_norm": 0.6319851908391722,
"learning_rate": 9.49170189453586e-07,
"loss": 0.2617,
"step": 1028
},
{
"epoch": 1.602803738317757,
"grad_norm": 0.6736383488893641,
"learning_rate": 9.420109687788148e-07,
"loss": 0.2641,
"step": 1029
},
{
"epoch": 1.6043613707165107,
"grad_norm": 0.6656334468054872,
"learning_rate": 9.348760410592855e-07,
"loss": 0.2171,
"step": 1030
},
{
"epoch": 1.6059190031152648,
"grad_norm": 0.6957076685070199,
"learning_rate": 9.277654490079035e-07,
"loss": 0.236,
"step": 1031
},
{
"epoch": 1.6074766355140186,
"grad_norm": 0.6288117317451123,
"learning_rate": 9.206792351918809e-07,
"loss": 0.1562,
"step": 1032
},
{
"epoch": 1.6090342679127727,
"grad_norm": 0.6616920441316214,
"learning_rate": 9.136174420324962e-07,
"loss": 0.218,
"step": 1033
},
{
"epoch": 1.6105919003115265,
"grad_norm": 0.6342551825259773,
"learning_rate": 9.065801118048312e-07,
"loss": 0.2252,
"step": 1034
},
{
"epoch": 1.6121495327102804,
"grad_norm": 0.5589154301426593,
"learning_rate": 8.995672866375237e-07,
"loss": 0.1551,
"step": 1035
},
{
"epoch": 1.6137071651090342,
"grad_norm": 0.4635682927354973,
"learning_rate": 8.925790085125152e-07,
"loss": 0.1833,
"step": 1036
},
{
"epoch": 1.615264797507788,
"grad_norm": 0.5543844541586774,
"learning_rate": 8.856153192647909e-07,
"loss": 0.2067,
"step": 1037
},
{
"epoch": 1.616822429906542,
"grad_norm": 0.6110967305529611,
"learning_rate": 8.786762605821419e-07,
"loss": 0.2331,
"step": 1038
},
{
"epoch": 1.618380062305296,
"grad_norm": 0.6930362377044883,
"learning_rate": 8.717618740049083e-07,
"loss": 0.2451,
"step": 1039
},
{
"epoch": 1.6199376947040498,
"grad_norm": 0.8137005439911986,
"learning_rate": 8.648722009257315e-07,
"loss": 0.1626,
"step": 1040
},
{
"epoch": 1.6214953271028039,
"grad_norm": 0.7004708747608873,
"learning_rate": 8.580072825893049e-07,
"loss": 0.1831,
"step": 1041
},
{
"epoch": 1.6230529595015577,
"grad_norm": 0.607273962654228,
"learning_rate": 8.511671600921306e-07,
"loss": 0.161,
"step": 1042
},
{
"epoch": 1.6246105919003115,
"grad_norm": 0.5132546282344246,
"learning_rate": 8.443518743822726e-07,
"loss": 0.2764,
"step": 1043
},
{
"epoch": 1.6261682242990654,
"grad_norm": 0.6893277741210059,
"learning_rate": 8.375614662591097e-07,
"loss": 0.1793,
"step": 1044
},
{
"epoch": 1.6277258566978192,
"grad_norm": 0.46872986327742755,
"learning_rate": 8.307959763730899e-07,
"loss": 0.1761,
"step": 1045
},
{
"epoch": 1.629283489096573,
"grad_norm": 0.5290839127332908,
"learning_rate": 8.240554452254929e-07,
"loss": 0.2329,
"step": 1046
},
{
"epoch": 1.6308411214953271,
"grad_norm": 0.6109039006881162,
"learning_rate": 8.173399131681831e-07,
"loss": 0.2121,
"step": 1047
},
{
"epoch": 1.632398753894081,
"grad_norm": 0.6091367601010639,
"learning_rate": 8.106494204033677e-07,
"loss": 0.1873,
"step": 1048
},
{
"epoch": 1.633956386292835,
"grad_norm": 0.5867126273845332,
"learning_rate": 8.039840069833594e-07,
"loss": 0.7795,
"step": 1049
},
{
"epoch": 1.6355140186915889,
"grad_norm": 0.5685700476106482,
"learning_rate": 7.973437128103306e-07,
"loss": 0.2459,
"step": 1050
},
{
"epoch": 1.6370716510903427,
"grad_norm": 0.5135498312661578,
"learning_rate": 7.907285776360863e-07,
"loss": 0.255,
"step": 1051
},
{
"epoch": 1.6386292834890965,
"grad_norm": 0.6667857649972504,
"learning_rate": 7.841386410618096e-07,
"loss": 0.2476,
"step": 1052
},
{
"epoch": 1.6401869158878504,
"grad_norm": 0.5960262952580946,
"learning_rate": 7.775739425378398e-07,
"loss": 0.2462,
"step": 1053
},
{
"epoch": 1.6417445482866042,
"grad_norm": 0.8148918825910239,
"learning_rate": 7.710345213634247e-07,
"loss": 0.2342,
"step": 1054
},
{
"epoch": 1.6433021806853583,
"grad_norm": 0.6117412041332031,
"learning_rate": 7.645204166864967e-07,
"loss": 0.1889,
"step": 1055
},
{
"epoch": 1.644859813084112,
"grad_norm": 0.6194608199292435,
"learning_rate": 7.580316675034255e-07,
"loss": 0.1807,
"step": 1056
},
{
"epoch": 1.6464174454828662,
"grad_norm": 0.9323712553721532,
"learning_rate": 7.515683126587975e-07,
"loss": 0.1879,
"step": 1057
},
{
"epoch": 1.64797507788162,
"grad_norm": 0.5493635025344051,
"learning_rate": 7.451303908451707e-07,
"loss": 0.2106,
"step": 1058
},
{
"epoch": 1.6495327102803738,
"grad_norm": 0.7158277912716375,
"learning_rate": 7.387179406028549e-07,
"loss": 0.2354,
"step": 1059
},
{
"epoch": 1.6510903426791277,
"grad_norm": 0.5571552725370883,
"learning_rate": 7.323310003196749e-07,
"loss": 0.2145,
"step": 1060
},
{
"epoch": 1.6526479750778815,
"grad_norm": 0.7538580273599934,
"learning_rate": 7.259696082307372e-07,
"loss": 0.2482,
"step": 1061
},
{
"epoch": 1.6542056074766354,
"grad_norm": 0.5327326528906995,
"learning_rate": 7.196338024182098e-07,
"loss": 0.2354,
"step": 1062
},
{
"epoch": 1.6557632398753894,
"grad_norm": 0.557287992992955,
"learning_rate": 7.133236208110878e-07,
"loss": 0.1693,
"step": 1063
},
{
"epoch": 1.6573208722741433,
"grad_norm": 0.5104215870170742,
"learning_rate": 7.070391011849698e-07,
"loss": 0.1717,
"step": 1064
},
{
"epoch": 1.6588785046728973,
"grad_norm": 0.5917806139945608,
"learning_rate": 7.007802811618258e-07,
"loss": 0.2085,
"step": 1065
},
{
"epoch": 1.6604361370716512,
"grad_norm": 0.6306946462467933,
"learning_rate": 6.945471982097818e-07,
"loss": 0.2145,
"step": 1066
},
{
"epoch": 1.661993769470405,
"grad_norm": 0.7196660102146798,
"learning_rate": 6.883398896428867e-07,
"loss": 0.2321,
"step": 1067
},
{
"epoch": 1.6635514018691588,
"grad_norm": 0.532187998071398,
"learning_rate": 6.821583926208947e-07,
"loss": 0.2366,
"step": 1068
},
{
"epoch": 1.6651090342679127,
"grad_norm": 0.5827926596867945,
"learning_rate": 6.760027441490369e-07,
"loss": 0.285,
"step": 1069
},
{
"epoch": 1.6666666666666665,
"grad_norm": 0.7057955872693081,
"learning_rate": 6.698729810778065e-07,
"loss": 0.6118,
"step": 1070
},
{
"epoch": 1.6682242990654206,
"grad_norm": 0.7121732240730015,
"learning_rate": 6.637691401027352e-07,
"loss": 0.8544,
"step": 1071
},
{
"epoch": 1.6697819314641744,
"grad_norm": 0.49184886655898746,
"learning_rate": 6.576912577641697e-07,
"loss": 0.2325,
"step": 1072
},
{
"epoch": 1.6713395638629285,
"grad_norm": 0.6760569677497837,
"learning_rate": 6.516393704470608e-07,
"loss": 0.8096,
"step": 1073
},
{
"epoch": 1.6728971962616823,
"grad_norm": 0.7478627429100672,
"learning_rate": 6.456135143807369e-07,
"loss": 0.1594,
"step": 1074
},
{
"epoch": 1.6744548286604362,
"grad_norm": 0.5379337402997096,
"learning_rate": 6.396137256386975e-07,
"loss": 0.1994,
"step": 1075
},
{
"epoch": 1.67601246105919,
"grad_norm": 0.5032927663415819,
"learning_rate": 6.336400401383857e-07,
"loss": 0.2132,
"step": 1076
},
{
"epoch": 1.6775700934579438,
"grad_norm": 0.47582414487022096,
"learning_rate": 6.276924936409829e-07,
"loss": 0.2478,
"step": 1077
},
{
"epoch": 1.6791277258566977,
"grad_norm": 0.6434383172973982,
"learning_rate": 6.217711217511857e-07,
"loss": 0.2143,
"step": 1078
},
{
"epoch": 1.6806853582554517,
"grad_norm": 0.6604513508530203,
"learning_rate": 6.158759599170045e-07,
"loss": 0.19,
"step": 1079
},
{
"epoch": 1.6822429906542056,
"grad_norm": 0.5696994691557421,
"learning_rate": 6.100070434295379e-07,
"loss": 0.2234,
"step": 1080
},
{
"epoch": 1.6838006230529596,
"grad_norm": 0.668728026470969,
"learning_rate": 6.041644074227738e-07,
"loss": 0.1627,
"step": 1081
},
{
"epoch": 1.6853582554517135,
"grad_norm": 0.5959783592005928,
"learning_rate": 5.983480868733666e-07,
"loss": 0.2239,
"step": 1082
},
{
"epoch": 1.6869158878504673,
"grad_norm": 0.7134998370022319,
"learning_rate": 5.925581166004407e-07,
"loss": 0.1782,
"step": 1083
},
{
"epoch": 1.6884735202492211,
"grad_norm": 0.5805335392567044,
"learning_rate": 5.867945312653733e-07,
"loss": 0.1725,
"step": 1084
},
{
"epoch": 1.690031152647975,
"grad_norm": 0.938179646578548,
"learning_rate": 5.810573653715884e-07,
"loss": 0.6935,
"step": 1085
},
{
"epoch": 1.6915887850467288,
"grad_norm": 0.6255817452393061,
"learning_rate": 5.753466532643531e-07,
"loss": 0.2057,
"step": 1086
},
{
"epoch": 1.6931464174454829,
"grad_norm": 0.6164273521801443,
"learning_rate": 5.696624291305692e-07,
"loss": 0.2354,
"step": 1087
},
{
"epoch": 1.6947040498442367,
"grad_norm": 0.6024358987476546,
"learning_rate": 5.64004726998571e-07,
"loss": 0.1628,
"step": 1088
},
{
"epoch": 1.6962616822429908,
"grad_norm": 0.583734062156969,
"learning_rate": 5.583735807379165e-07,
"loss": 0.258,
"step": 1089
},
{
"epoch": 1.6978193146417446,
"grad_norm": 0.5381732861463082,
"learning_rate": 5.527690240591927e-07,
"loss": 0.259,
"step": 1090
},
{
"epoch": 1.6993769470404985,
"grad_norm": 0.6432738953861825,
"learning_rate": 5.471910905138062e-07,
"loss": 0.2233,
"step": 1091
},
{
"epoch": 1.7009345794392523,
"grad_norm": 0.7405741920054726,
"learning_rate": 5.416398134937878e-07,
"loss": 0.2732,
"step": 1092
},
{
"epoch": 1.7024922118380061,
"grad_norm": 0.5681851173547261,
"learning_rate": 5.361152262315883e-07,
"loss": 0.1976,
"step": 1093
},
{
"epoch": 1.70404984423676,
"grad_norm": 0.5473658208153401,
"learning_rate": 5.306173617998823e-07,
"loss": 0.2139,
"step": 1094
},
{
"epoch": 1.705607476635514,
"grad_norm": 0.5255770004268006,
"learning_rate": 5.251462531113705e-07,
"loss": 0.1707,
"step": 1095
},
{
"epoch": 1.7071651090342679,
"grad_norm": 0.5374196236910275,
"learning_rate": 5.197019329185787e-07,
"loss": 0.2281,
"step": 1096
},
{
"epoch": 1.708722741433022,
"grad_norm": 0.5260272455503919,
"learning_rate": 5.142844338136693e-07,
"loss": 0.2471,
"step": 1097
},
{
"epoch": 1.7102803738317758,
"grad_norm": 0.5646648215018458,
"learning_rate": 5.088937882282358e-07,
"loss": 0.1951,
"step": 1098
},
{
"epoch": 1.7118380062305296,
"grad_norm": 0.5458034326374251,
"learning_rate": 5.035300284331213e-07,
"loss": 0.2432,
"step": 1099
},
{
"epoch": 1.7133956386292835,
"grad_norm": 0.6037542349280097,
"learning_rate": 4.981931865382111e-07,
"loss": 0.2618,
"step": 1100
},
{
"epoch": 1.7149532710280373,
"grad_norm": 0.5948181143891571,
"learning_rate": 4.92883294492254e-07,
"loss": 0.1635,
"step": 1101
},
{
"epoch": 1.7165109034267911,
"grad_norm": 0.5776080511589572,
"learning_rate": 4.876003840826588e-07,
"loss": 0.2451,
"step": 1102
},
{
"epoch": 1.7180685358255452,
"grad_norm": 0.5026789603534472,
"learning_rate": 4.823444869353177e-07,
"loss": 0.1562,
"step": 1103
},
{
"epoch": 1.719626168224299,
"grad_norm": 0.5367686440727194,
"learning_rate": 4.771156345144018e-07,
"loss": 0.2041,
"step": 1104
},
{
"epoch": 1.721183800623053,
"grad_norm": 0.658875304880407,
"learning_rate": 4.7191385812218547e-07,
"loss": 0.7385,
"step": 1105
},
{
"epoch": 1.722741433021807,
"grad_norm": 0.6189314088368518,
"learning_rate": 4.6673918889884985e-07,
"loss": 0.269,
"step": 1106
},
{
"epoch": 1.7242990654205608,
"grad_norm": 0.7000827321303451,
"learning_rate": 4.615916578223029e-07,
"loss": 0.219,
"step": 1107
},
{
"epoch": 1.7258566978193146,
"grad_norm": 0.6218605150183268,
"learning_rate": 4.5647129570799196e-07,
"loss": 0.2765,
"step": 1108
},
{
"epoch": 1.7274143302180685,
"grad_norm": 0.7349959501432963,
"learning_rate": 4.513781332087153e-07,
"loss": 0.2457,
"step": 1109
},
{
"epoch": 1.7289719626168223,
"grad_norm": 0.7008089188891375,
"learning_rate": 4.463122008144449e-07,
"loss": 0.2083,
"step": 1110
},
{
"epoch": 1.7305295950155763,
"grad_norm": 0.5607384923036217,
"learning_rate": 4.4127352885214026e-07,
"loss": 0.2166,
"step": 1111
},
{
"epoch": 1.7320872274143302,
"grad_norm": 0.645885067353582,
"learning_rate": 4.362621474855688e-07,
"loss": 0.2312,
"step": 1112
},
{
"epoch": 1.7336448598130842,
"grad_norm": 1.0208563939368382,
"learning_rate": 4.3127808671512114e-07,
"loss": 0.8117,
"step": 1113
},
{
"epoch": 1.735202492211838,
"grad_norm": 0.6252228702704483,
"learning_rate": 4.263213763776369e-07,
"loss": 0.2111,
"step": 1114
},
{
"epoch": 1.736760124610592,
"grad_norm": 0.5295697965398305,
"learning_rate": 4.213920461462234e-07,
"loss": 0.1693,
"step": 1115
},
{
"epoch": 1.7383177570093458,
"grad_norm": 0.5321708104559851,
"learning_rate": 4.1649012553007795e-07,
"loss": 0.1598,
"step": 1116
},
{
"epoch": 1.7398753894080996,
"grad_norm": 0.6297092340917954,
"learning_rate": 4.116156438743102e-07,
"loss": 0.6563,
"step": 1117
},
{
"epoch": 1.7414330218068534,
"grad_norm": 0.6502646751207796,
"learning_rate": 4.067686303597701e-07,
"loss": 0.828,
"step": 1118
},
{
"epoch": 1.7429906542056075,
"grad_norm": 0.6299291575701418,
"learning_rate": 4.0194911400287084e-07,
"loss": 0.1618,
"step": 1119
},
{
"epoch": 1.7445482866043613,
"grad_norm": 0.6763449045114679,
"learning_rate": 3.971571236554117e-07,
"loss": 0.1755,
"step": 1120
},
{
"epoch": 1.7461059190031154,
"grad_norm": 0.6097387148482865,
"learning_rate": 3.923926880044132e-07,
"loss": 0.2144,
"step": 1121
},
{
"epoch": 1.7476635514018692,
"grad_norm": 0.6555185956249271,
"learning_rate": 3.876558355719373e-07,
"loss": 0.7322,
"step": 1122
},
{
"epoch": 1.749221183800623,
"grad_norm": 0.529186428154948,
"learning_rate": 3.8294659471492466e-07,
"loss": 0.2299,
"step": 1123
},
{
"epoch": 1.750778816199377,
"grad_norm": 0.5210926587424783,
"learning_rate": 3.7826499362501657e-07,
"loss": 0.7365,
"step": 1124
},
{
"epoch": 1.7523364485981308,
"grad_norm": 0.5705242040491778,
"learning_rate": 3.7361106032839266e-07,
"loss": 0.1769,
"step": 1125
},
{
"epoch": 1.7538940809968846,
"grad_norm": 0.6221568867702569,
"learning_rate": 3.6898482268559745e-07,
"loss": 0.2641,
"step": 1126
},
{
"epoch": 1.7554517133956387,
"grad_norm": 0.5761942177039651,
"learning_rate": 3.643863083913829e-07,
"loss": 0.1756,
"step": 1127
},
{
"epoch": 1.7570093457943925,
"grad_norm": 0.5598040118881634,
"learning_rate": 3.5981554497452886e-07,
"loss": 0.1968,
"step": 1128
},
{
"epoch": 1.7585669781931466,
"grad_norm": 0.6115168676953736,
"learning_rate": 3.552725597976925e-07,
"loss": 0.1656,
"step": 1129
},
{
"epoch": 1.7601246105919004,
"grad_norm": 0.5618155080059914,
"learning_rate": 3.507573800572328e-07,
"loss": 0.2284,
"step": 1130
},
{
"epoch": 1.7616822429906542,
"grad_norm": 0.5746401068264577,
"learning_rate": 3.462700327830559e-07,
"loss": 0.2942,
"step": 1131
},
{
"epoch": 1.763239875389408,
"grad_norm": 0.5812252313415237,
"learning_rate": 3.418105448384507e-07,
"loss": 0.2478,
"step": 1132
},
{
"epoch": 1.764797507788162,
"grad_norm": 0.5205912713872596,
"learning_rate": 3.3737894291992426e-07,
"loss": 0.1793,
"step": 1133
},
{
"epoch": 1.7663551401869158,
"grad_norm": 0.59827648712384,
"learning_rate": 3.32975253557048e-07,
"loss": 0.1643,
"step": 1134
},
{
"epoch": 1.7679127725856698,
"grad_norm": 0.5476307417528222,
"learning_rate": 3.2859950311229625e-07,
"loss": 0.2551,
"step": 1135
},
{
"epoch": 1.7694704049844237,
"grad_norm": 0.5912924193469346,
"learning_rate": 3.242517177808874e-07,
"loss": 0.2958,
"step": 1136
},
{
"epoch": 1.7710280373831777,
"grad_norm": 0.6149100468721329,
"learning_rate": 3.199319235906273e-07,
"loss": 0.2391,
"step": 1137
},
{
"epoch": 1.7725856697819315,
"grad_norm": 0.5135930360411095,
"learning_rate": 3.1564014640175613e-07,
"loss": 0.1651,
"step": 1138
},
{
"epoch": 1.7741433021806854,
"grad_norm": 0.545428012439618,
"learning_rate": 3.1137641190678967e-07,
"loss": 0.1866,
"step": 1139
},
{
"epoch": 1.7757009345794392,
"grad_norm": 0.7534351814752701,
"learning_rate": 3.0714074563037043e-07,
"loss": 0.1752,
"step": 1140
},
{
"epoch": 1.777258566978193,
"grad_norm": 0.6368630911991421,
"learning_rate": 3.029331729291074e-07,
"loss": 0.2049,
"step": 1141
},
{
"epoch": 1.778816199376947,
"grad_norm": 0.594879170975228,
"learning_rate": 2.9875371899143246e-07,
"loss": 0.6515,
"step": 1142
},
{
"epoch": 1.780373831775701,
"grad_norm": 0.5107491983811178,
"learning_rate": 2.9460240883744496e-07,
"loss": 0.1768,
"step": 1143
},
{
"epoch": 1.7819314641744548,
"grad_norm": 0.5417424283009235,
"learning_rate": 2.9047926731876177e-07,
"loss": 0.2087,
"step": 1144
},
{
"epoch": 1.7834890965732089,
"grad_norm": 0.6798250278910997,
"learning_rate": 2.8638431911837126e-07,
"loss": 0.28,
"step": 1145
},
{
"epoch": 1.7850467289719627,
"grad_norm": 1.0039697477055525,
"learning_rate": 2.8231758875048087e-07,
"loss": 0.2178,
"step": 1146
},
{
"epoch": 1.7866043613707165,
"grad_norm": 0.6262879884675681,
"learning_rate": 2.7827910056037857e-07,
"loss": 0.1775,
"step": 1147
},
{
"epoch": 1.7881619937694704,
"grad_norm": 0.6215897703245564,
"learning_rate": 2.742688787242764e-07,
"loss": 0.1726,
"step": 1148
},
{
"epoch": 1.7897196261682242,
"grad_norm": 0.7119013996583601,
"learning_rate": 2.702869472491748e-07,
"loss": 0.6972,
"step": 1149
},
{
"epoch": 1.791277258566978,
"grad_norm": 0.5300668663016045,
"learning_rate": 2.6633332997271277e-07,
"loss": 0.1797,
"step": 1150
},
{
"epoch": 1.7928348909657321,
"grad_norm": 0.6010057607669876,
"learning_rate": 2.624080505630311e-07,
"loss": 0.2206,
"step": 1151
},
{
"epoch": 1.794392523364486,
"grad_norm": 0.5566227596606641,
"learning_rate": 2.585111325186235e-07,
"loss": 0.2991,
"step": 1152
},
{
"epoch": 1.79595015576324,
"grad_norm": 0.7373120599532028,
"learning_rate": 2.5464259916820276e-07,
"loss": 0.2155,
"step": 1153
},
{
"epoch": 1.7975077881619939,
"grad_norm": 0.5444732937470245,
"learning_rate": 2.508024736705561e-07,
"loss": 0.2394,
"step": 1154
},
{
"epoch": 1.7990654205607477,
"grad_norm": 0.6187155054450917,
"learning_rate": 2.469907790144088e-07,
"loss": 0.1835,
"step": 1155
},
{
"epoch": 1.8006230529595015,
"grad_norm": 0.45905420748321907,
"learning_rate": 2.4320753801828853e-07,
"loss": 0.1604,
"step": 1156
},
{
"epoch": 1.8021806853582554,
"grad_norm": 0.43879992123704614,
"learning_rate": 2.394527733303831e-07,
"loss": 0.1723,
"step": 1157
},
{
"epoch": 1.8037383177570092,
"grad_norm": 0.5693210977954104,
"learning_rate": 2.3572650742840985e-07,
"loss": 0.1148,
"step": 1158
},
{
"epoch": 1.8052959501557633,
"grad_norm": 0.5824148680871707,
"learning_rate": 2.3202876261948004e-07,
"loss": 0.2304,
"step": 1159
},
{
"epoch": 1.8068535825545171,
"grad_norm": 0.692163156120945,
"learning_rate": 2.2835956103996525e-07,
"loss": 0.1877,
"step": 1160
},
{
"epoch": 1.8084112149532712,
"grad_norm": 0.627351603757966,
"learning_rate": 2.247189246553616e-07,
"loss": 0.2133,
"step": 1161
},
{
"epoch": 1.809968847352025,
"grad_norm": 0.6447153050268107,
"learning_rate": 2.2110687526016418e-07,
"loss": 0.1753,
"step": 1162
},
{
"epoch": 1.8115264797507789,
"grad_norm": 0.6498742455536043,
"learning_rate": 2.1752343447773105e-07,
"loss": 0.2801,
"step": 1163
},
{
"epoch": 1.8130841121495327,
"grad_norm": 0.5556829005135521,
"learning_rate": 2.1396862376015904e-07,
"loss": 0.2613,
"step": 1164
},
{
"epoch": 1.8146417445482865,
"grad_norm": 0.5643429928213758,
"learning_rate": 2.1044246438814918e-07,
"loss": 0.2529,
"step": 1165
},
{
"epoch": 1.8161993769470404,
"grad_norm": 0.747941354496335,
"learning_rate": 2.0694497747088428e-07,
"loss": 0.2321,
"step": 1166
},
{
"epoch": 1.8177570093457944,
"grad_norm": 0.6102806034003578,
"learning_rate": 2.034761839459015e-07,
"loss": 0.218,
"step": 1167
},
{
"epoch": 1.8193146417445483,
"grad_norm": 0.7249346044063573,
"learning_rate": 2.0003610457896506e-07,
"loss": 0.2317,
"step": 1168
},
{
"epoch": 1.8208722741433023,
"grad_norm": 0.5858879975789372,
"learning_rate": 1.966247599639448e-07,
"loss": 0.1886,
"step": 1169
},
{
"epoch": 1.8224299065420562,
"grad_norm": 0.458859760014875,
"learning_rate": 1.9324217052268835e-07,
"loss": 0.2037,
"step": 1170
},
{
"epoch": 1.82398753894081,
"grad_norm": 0.6009222197974706,
"learning_rate": 1.8988835650490645e-07,
"loss": 0.1845,
"step": 1171
},
{
"epoch": 1.8255451713395638,
"grad_norm": 0.7980290452291322,
"learning_rate": 1.865633379880427e-07,
"loss": 0.2341,
"step": 1172
},
{
"epoch": 1.8271028037383177,
"grad_norm": 0.5432836689095761,
"learning_rate": 1.832671348771614e-07,
"loss": 0.2282,
"step": 1173
},
{
"epoch": 1.8286604361370715,
"grad_norm": 0.628743562125201,
"learning_rate": 1.799997669048209e-07,
"loss": 0.261,
"step": 1174
},
{
"epoch": 1.8302180685358256,
"grad_norm": 0.5832048925707848,
"learning_rate": 1.7676125363096431e-07,
"loss": 0.2157,
"step": 1175
},
{
"epoch": 1.8317757009345794,
"grad_norm": 0.6281683541540782,
"learning_rate": 1.7355161444279346e-07,
"loss": 0.2799,
"step": 1176
},
{
"epoch": 1.8333333333333335,
"grad_norm": 0.6990945601104308,
"learning_rate": 1.7037086855465902e-07,
"loss": 0.2248,
"step": 1177
},
{
"epoch": 1.8348909657320873,
"grad_norm": 0.48096314961131165,
"learning_rate": 1.6721903500794112e-07,
"loss": 0.2014,
"step": 1178
},
{
"epoch": 1.8364485981308412,
"grad_norm": 0.6720925885316696,
"learning_rate": 1.6409613267094006e-07,
"loss": 0.1839,
"step": 1179
},
{
"epoch": 1.838006230529595,
"grad_norm": 0.5023236174498488,
"learning_rate": 1.6100218023876013e-07,
"loss": 0.1729,
"step": 1180
},
{
"epoch": 1.8395638629283488,
"grad_norm": 0.6472364425599902,
"learning_rate": 1.5793719623319713e-07,
"loss": 0.2147,
"step": 1181
},
{
"epoch": 1.8411214953271027,
"grad_norm": 0.5750750734641096,
"learning_rate": 1.5490119900263e-07,
"loss": 0.1734,
"step": 1182
},
{
"epoch": 1.8426791277258567,
"grad_norm": 0.640400103725089,
"learning_rate": 1.5189420672190924e-07,
"loss": 0.2149,
"step": 1183
},
{
"epoch": 1.8442367601246106,
"grad_norm": 0.6159833044318871,
"learning_rate": 1.489162373922498e-07,
"loss": 0.2426,
"step": 1184
},
{
"epoch": 1.8457943925233646,
"grad_norm": 0.5891969287939168,
"learning_rate": 1.4596730884112008e-07,
"loss": 0.2164,
"step": 1185
},
{
"epoch": 1.8473520249221185,
"grad_norm": 0.6060235472721057,
"learning_rate": 1.4304743872213868e-07,
"loss": 0.2567,
"step": 1186
},
{
"epoch": 1.8489096573208723,
"grad_norm": 0.6985533236531171,
"learning_rate": 1.4015664451496713e-07,
"loss": 0.1688,
"step": 1187
},
{
"epoch": 1.8504672897196262,
"grad_norm": 0.5859326752612632,
"learning_rate": 1.372949435252058e-07,
"loss": 0.2311,
"step": 1188
},
{
"epoch": 1.85202492211838,
"grad_norm": 0.7294577853076925,
"learning_rate": 1.3446235288428867e-07,
"loss": 0.2023,
"step": 1189
},
{
"epoch": 1.8535825545171338,
"grad_norm": 0.6888020410981194,
"learning_rate": 1.31658889549382e-07,
"loss": 0.2295,
"step": 1190
},
{
"epoch": 1.855140186915888,
"grad_norm": 0.6480785185138133,
"learning_rate": 1.288845703032854e-07,
"loss": 0.2359,
"step": 1191
},
{
"epoch": 1.8566978193146417,
"grad_norm": 0.7015546454899397,
"learning_rate": 1.2613941175432577e-07,
"loss": 0.2293,
"step": 1192
},
{
"epoch": 1.8582554517133958,
"grad_norm": 0.5363209042476926,
"learning_rate": 1.2342343033626248e-07,
"loss": 0.18,
"step": 1193
},
{
"epoch": 1.8598130841121496,
"grad_norm": 0.6459401688040521,
"learning_rate": 1.2073664230818571e-07,
"loss": 0.1598,
"step": 1194
},
{
"epoch": 1.8613707165109035,
"grad_norm": 0.638563320329426,
"learning_rate": 1.180790637544249e-07,
"loss": 0.2476,
"step": 1195
},
{
"epoch": 1.8629283489096573,
"grad_norm": 0.6002853300206833,
"learning_rate": 1.1545071058444324e-07,
"loss": 0.2199,
"step": 1196
},
{
"epoch": 1.8644859813084111,
"grad_norm": 0.5764195151127395,
"learning_rate": 1.1285159853275107e-07,
"loss": 0.1227,
"step": 1197
},
{
"epoch": 1.866043613707165,
"grad_norm": 0.6123103816053441,
"learning_rate": 1.1028174315880657e-07,
"loss": 0.815,
"step": 1198
},
{
"epoch": 1.867601246105919,
"grad_norm": 0.8123905890394445,
"learning_rate": 1.0774115984692523e-07,
"loss": 0.7364,
"step": 1199
},
{
"epoch": 1.8691588785046729,
"grad_norm": 0.5718174425382163,
"learning_rate": 1.0522986380618606e-07,
"loss": 0.2227,
"step": 1200
},
{
"epoch": 1.870716510903427,
"grad_norm": 0.5280345967170433,
"learning_rate": 1.0274787007034215e-07,
"loss": 0.232,
"step": 1201
},
{
"epoch": 1.8722741433021808,
"grad_norm": 0.6572165953398976,
"learning_rate": 1.0029519349772754e-07,
"loss": 0.2738,
"step": 1202
},
{
"epoch": 1.8738317757009346,
"grad_norm": 0.5244811712323645,
"learning_rate": 9.787184877117328e-08,
"loss": 0.2594,
"step": 1203
},
{
"epoch": 1.8753894080996885,
"grad_norm": 0.6676225172095591,
"learning_rate": 9.547785039791535e-08,
"loss": 0.1547,
"step": 1204
},
{
"epoch": 1.8769470404984423,
"grad_norm": 0.5476881051496193,
"learning_rate": 9.311321270950801e-08,
"loss": 0.2128,
"step": 1205
},
{
"epoch": 1.8785046728971961,
"grad_norm": 0.5345787062303978,
"learning_rate": 9.077794986174226e-08,
"loss": 0.229,
"step": 1206
},
{
"epoch": 1.8800623052959502,
"grad_norm": 0.6008670505158255,
"learning_rate": 8.84720758345553e-08,
"loss": 0.2073,
"step": 1207
},
{
"epoch": 1.881619937694704,
"grad_norm": 0.6417543491732723,
"learning_rate": 8.619560443195174e-08,
"loss": 0.2276,
"step": 1208
},
{
"epoch": 1.883177570093458,
"grad_norm": 0.5552051709339759,
"learning_rate": 8.394854928191587e-08,
"loss": 0.2693,
"step": 1209
},
{
"epoch": 1.884735202492212,
"grad_norm": 0.6694794375263459,
"learning_rate": 8.173092383633563e-08,
"loss": 0.1767,
"step": 1210
},
{
"epoch": 1.8862928348909658,
"grad_norm": 0.5551294915003306,
"learning_rate": 7.954274137091877e-08,
"loss": 0.147,
"step": 1211
},
{
"epoch": 1.8878504672897196,
"grad_norm": 0.534343522912461,
"learning_rate": 7.738401498511406e-08,
"loss": 0.1891,
"step": 1212
},
{
"epoch": 1.8894080996884735,
"grad_norm": 0.6889315198186039,
"learning_rate": 7.525475760203239e-08,
"loss": 0.2832,
"step": 1213
},
{
"epoch": 1.8909657320872273,
"grad_norm": 0.6872767665873238,
"learning_rate": 7.315498196836968e-08,
"loss": 0.1472,
"step": 1214
},
{
"epoch": 1.8925233644859814,
"grad_norm": 0.6612078267914279,
"learning_rate": 7.108470065433193e-08,
"loss": 0.2704,
"step": 1215
},
{
"epoch": 1.8940809968847352,
"grad_norm": 0.560144578359172,
"learning_rate": 6.904392605355803e-08,
"loss": 0.2943,
"step": 1216
},
{
"epoch": 1.8956386292834893,
"grad_norm": 0.590357610353342,
"learning_rate": 6.703267038304706e-08,
"loss": 0.229,
"step": 1217
},
{
"epoch": 1.897196261682243,
"grad_norm": 0.496234632076301,
"learning_rate": 6.505094568308223e-08,
"loss": 0.2807,
"step": 1218
},
{
"epoch": 1.898753894080997,
"grad_norm": 0.6744039849106092,
"learning_rate": 6.309876381716429e-08,
"loss": 0.1744,
"step": 1219
},
{
"epoch": 1.9003115264797508,
"grad_norm": 0.6282540133425786,
"learning_rate": 6.117613647193543e-08,
"loss": 0.1896,
"step": 1220
},
{
"epoch": 1.9018691588785046,
"grad_norm": 0.545913837904785,
"learning_rate": 5.928307515711107e-08,
"loss": 0.1798,
"step": 1221
},
{
"epoch": 1.9034267912772584,
"grad_norm": 0.5750165821891594,
"learning_rate": 5.741959120541096e-08,
"loss": 0.2328,
"step": 1222
},
{
"epoch": 1.9049844236760125,
"grad_norm": 0.6138340557718204,
"learning_rate": 5.558569577249207e-08,
"loss": 0.1179,
"step": 1223
},
{
"epoch": 1.9065420560747663,
"grad_norm": 0.5250776885173793,
"learning_rate": 5.378139983688135e-08,
"loss": 0.178,
"step": 1224
},
{
"epoch": 1.9080996884735204,
"grad_norm": 0.5353000767702344,
"learning_rate": 5.2006714199908106e-08,
"loss": 0.1881,
"step": 1225
},
{
"epoch": 1.9096573208722742,
"grad_norm": 0.6835275834635214,
"learning_rate": 5.026164948564283e-08,
"loss": 0.2075,
"step": 1226
},
{
"epoch": 1.911214953271028,
"grad_norm": 0.48910598318494225,
"learning_rate": 4.854621614083122e-08,
"loss": 0.1864,
"step": 1227
},
{
"epoch": 1.912772585669782,
"grad_norm": 0.5349299444820935,
"learning_rate": 4.686042443483196e-08,
"loss": 0.297,
"step": 1228
},
{
"epoch": 1.9143302180685358,
"grad_norm": 0.42493561251864087,
"learning_rate": 4.520428445955571e-08,
"loss": 0.1656,
"step": 1229
},
{
"epoch": 1.9158878504672896,
"grad_norm": 0.5786034487132414,
"learning_rate": 4.357780612940343e-08,
"loss": 0.2243,
"step": 1230
},
{
"epoch": 1.9174454828660437,
"grad_norm": 0.6628360349913834,
"learning_rate": 4.1980999181210344e-08,
"loss": 0.237,
"step": 1231
},
{
"epoch": 1.9190031152647975,
"grad_norm": 0.5726232559997532,
"learning_rate": 4.041387317418377e-08,
"loss": 0.19,
"step": 1232
},
{
"epoch": 1.9205607476635516,
"grad_norm": 0.5957569462038291,
"learning_rate": 3.88764374898476e-08,
"loss": 0.1922,
"step": 1233
},
{
"epoch": 1.9221183800623054,
"grad_norm": 0.5981083547680797,
"learning_rate": 3.736870133198622e-08,
"loss": 0.1861,
"step": 1234
},
{
"epoch": 1.9236760124610592,
"grad_norm": 0.5039034286835927,
"learning_rate": 3.589067372658961e-08,
"loss": 0.1587,
"step": 1235
},
{
"epoch": 1.925233644859813,
"grad_norm": 0.5946390323111295,
"learning_rate": 3.444236352179831e-08,
"loss": 0.1658,
"step": 1236
},
{
"epoch": 1.926791277258567,
"grad_norm": 0.595459693543413,
"learning_rate": 3.302377938785128e-08,
"loss": 0.1509,
"step": 1237
},
{
"epoch": 1.9283489096573208,
"grad_norm": 0.6496361440227659,
"learning_rate": 3.163492981703431e-08,
"loss": 0.693,
"step": 1238
},
{
"epoch": 1.9299065420560748,
"grad_norm": 0.6473442377574977,
"learning_rate": 3.027582312362776e-08,
"loss": 0.2082,
"step": 1239
},
{
"epoch": 1.9314641744548287,
"grad_norm": 0.6541508019021366,
"learning_rate": 2.8946467443858873e-08,
"loss": 0.1911,
"step": 1240
},
{
"epoch": 1.9330218068535827,
"grad_norm": 0.5343926005691645,
"learning_rate": 2.764687073585015e-08,
"loss": 0.1724,
"step": 1241
},
{
"epoch": 1.9345794392523366,
"grad_norm": 0.6477434345058528,
"learning_rate": 2.6377040779574924e-08,
"loss": 0.2109,
"step": 1242
},
{
"epoch": 1.9361370716510904,
"grad_norm": 0.4788618278691479,
"learning_rate": 2.5136985176809625e-08,
"loss": 0.2257,
"step": 1243
},
{
"epoch": 1.9376947040498442,
"grad_norm": 0.7161454071348646,
"learning_rate": 2.3926711351086594e-08,
"loss": 0.2201,
"step": 1244
},
{
"epoch": 1.939252336448598,
"grad_norm": 0.6162622985277744,
"learning_rate": 2.2746226547651352e-08,
"loss": 0.2353,
"step": 1245
},
{
"epoch": 1.940809968847352,
"grad_norm": 0.6572303875322422,
"learning_rate": 2.1595537833419276e-08,
"loss": 0.1238,
"step": 1246
},
{
"epoch": 1.942367601246106,
"grad_norm": 0.4905959486856584,
"learning_rate": 2.047465209693289e-08,
"loss": 0.1913,
"step": 1247
},
{
"epoch": 1.9439252336448598,
"grad_norm": 0.6262698157072888,
"learning_rate": 1.9383576048320752e-08,
"loss": 0.2135,
"step": 1248
},
{
"epoch": 1.9454828660436139,
"grad_norm": 0.5977256427085367,
"learning_rate": 1.8322316219256396e-08,
"loss": 0.3103,
"step": 1249
},
{
"epoch": 1.9470404984423677,
"grad_norm": 0.5597696638121235,
"learning_rate": 1.7290878962920587e-08,
"loss": 0.2003,
"step": 1250
},
{
"epoch": 1.9485981308411215,
"grad_norm": 0.7414334515835818,
"learning_rate": 1.6289270453963e-08,
"loss": 0.133,
"step": 1251
},
{
"epoch": 1.9501557632398754,
"grad_norm": 0.6577156096873663,
"learning_rate": 1.531749668846394e-08,
"loss": 0.1938,
"step": 1252
},
{
"epoch": 1.9517133956386292,
"grad_norm": 0.44233229164345883,
"learning_rate": 1.4375563483901011e-08,
"loss": 0.2196,
"step": 1253
},
{
"epoch": 1.953271028037383,
"grad_norm": 0.5494650237391789,
"learning_rate": 1.3463476479110837e-08,
"loss": 0.239,
"step": 1254
},
{
"epoch": 1.9548286604361371,
"grad_norm": 0.4597714878638595,
"learning_rate": 1.2581241134258515e-08,
"loss": 0.1569,
"step": 1255
},
{
"epoch": 1.956386292834891,
"grad_norm": 0.5344018379047227,
"learning_rate": 1.1728862730803759e-08,
"loss": 0.2506,
"step": 1256
},
{
"epoch": 1.957943925233645,
"grad_norm": 0.6170756968114266,
"learning_rate": 1.0906346371468147e-08,
"loss": 0.1756,
"step": 1257
},
{
"epoch": 1.9595015576323989,
"grad_norm": 0.7298638061579327,
"learning_rate": 1.011369698020681e-08,
"loss": 1.2315,
"step": 1258
},
{
"epoch": 1.9610591900311527,
"grad_norm": 0.7726214050156128,
"learning_rate": 9.350919302176242e-09,
"loss": 0.2132,
"step": 1259
},
{
"epoch": 1.9626168224299065,
"grad_norm": 0.7662101654548159,
"learning_rate": 8.618017903708198e-09,
"loss": 0.1979,
"step": 1260
},
{
"epoch": 1.9641744548286604,
"grad_norm": 0.7759433714405526,
"learning_rate": 7.914997172282502e-09,
"loss": 0.7825,
"step": 1261
},
{
"epoch": 1.9657320872274142,
"grad_norm": 0.5561454562523979,
"learning_rate": 7.241861316497623e-09,
"loss": 0.1787,
"step": 1262
},
{
"epoch": 1.9672897196261683,
"grad_norm": 0.6088898793952118,
"learning_rate": 6.598614366049027e-09,
"loss": 0.1525,
"step": 1263
},
{
"epoch": 1.9688473520249221,
"grad_norm": 0.6388680112674057,
"learning_rate": 5.985260171704199e-09,
"loss": 0.2538,
"step": 1264
},
{
"epoch": 1.9704049844236762,
"grad_norm": 1.2537539847150645,
"learning_rate": 5.401802405277656e-09,
"loss": 0.208,
"step": 1265
},
{
"epoch": 1.97196261682243,
"grad_norm": 0.5173311007016309,
"learning_rate": 4.848244559610971e-09,
"loss": 0.1927,
"step": 1266
},
{
"epoch": 1.9735202492211839,
"grad_norm": 0.7196153604341586,
"learning_rate": 4.324589948552227e-09,
"loss": 0.1292,
"step": 1267
},
{
"epoch": 1.9750778816199377,
"grad_norm": 0.8257531526111838,
"learning_rate": 3.830841706934374e-09,
"loss": 0.2268,
"step": 1268
},
{
"epoch": 1.9766355140186915,
"grad_norm": 0.6450834847941774,
"learning_rate": 3.367002790556906e-09,
"loss": 0.8063,
"step": 1269
},
{
"epoch": 1.9781931464174454,
"grad_norm": 0.5873649988075496,
"learning_rate": 2.9330759761692086e-09,
"loss": 0.1333,
"step": 1270
},
{
"epoch": 1.9797507788161994,
"grad_norm": 0.5338211522410232,
"learning_rate": 2.5290638614533516e-09,
"loss": 0.1651,
"step": 1271
},
{
"epoch": 1.9813084112149533,
"grad_norm": 0.5778154838677994,
"learning_rate": 2.154968865007989e-09,
"loss": 0.1375,
"step": 1272
},
{
"epoch": 1.9828660436137073,
"grad_norm": 0.6159025357834766,
"learning_rate": 1.8107932263350392e-09,
"loss": 0.2551,
"step": 1273
},
{
"epoch": 1.9844236760124612,
"grad_norm": 0.5276126425838753,
"learning_rate": 1.4965390058258035e-09,
"loss": 0.2857,
"step": 1274
},
{
"epoch": 1.985981308411215,
"grad_norm": 0.578702087814479,
"learning_rate": 1.2122080847470907e-09,
"loss": 0.1644,
"step": 1275
},
{
"epoch": 1.9875389408099688,
"grad_norm": 0.6156843300297827,
"learning_rate": 9.578021652323356e-10,
"loss": 0.203,
"step": 1276
},
{
"epoch": 1.9890965732087227,
"grad_norm": 0.5234412510622443,
"learning_rate": 7.3332277027105e-10,
"loss": 0.2086,
"step": 1277
},
{
"epoch": 1.9906542056074765,
"grad_norm": 0.6889557509874759,
"learning_rate": 5.38771243697167e-10,
"loss": 0.2412,
"step": 1278
},
{
"epoch": 1.9922118380062306,
"grad_norm": 0.5571525262541802,
"learning_rate": 3.7414875018515483e-10,
"loss": 0.314,
"step": 1279
},
{
"epoch": 1.9937694704049844,
"grad_norm": 0.48476554367462193,
"learning_rate": 2.3945627523891403e-10,
"loss": 0.1878,
"step": 1280
},
{
"epoch": 1.9953271028037385,
"grad_norm": 0.5759369634737167,
"learning_rate": 1.3469462518844733e-10,
"loss": 0.181,
"step": 1281
},
{
"epoch": 1.9968847352024923,
"grad_norm": 0.6640918787171276,
"learning_rate": 5.986442718486363e-11,
"loss": 0.1886,
"step": 1282
},
{
"epoch": 1.9984423676012462,
"grad_norm": 0.6150458129744069,
"learning_rate": 1.496612919482665e-11,
"loss": 0.2397,
"step": 1283
},
{
"epoch": 2.0,
"grad_norm": 0.6727485887005468,
"learning_rate": 0.0,
"loss": 0.2332,
"step": 1284
},
{
"epoch": 2.0,
"step": 1284,
"total_flos": 17255618273280.0,
"train_loss": 0.27858535300162723,
"train_runtime": 1476.77,
"train_samples_per_second": 3.477,
"train_steps_per_second": 0.869
}
],
"logging_steps": 1,
"max_steps": 1284,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 70000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 17255618273280.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}