{
  "best_metric": 2.435030460357666,
  "best_model_checkpoint": "zhtw-en/checkpoint-92500",
  "epoch": 3.0,
  "eval_steps": 2500,
  "global_step": 93276,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.016081307088640164,
      "grad_norm": 8.550692558288574,
      "learning_rate": 4.919593464556799e-05,
      "loss": 3.4707,
      "num_input_tokens_seen": 301408,
      "step": 500
    },
    {
      "epoch": 0.03216261417728033,
      "grad_norm": 9.999076843261719,
      "learning_rate": 4.8391869291135987e-05,
      "loss": 3.3664,
      "num_input_tokens_seen": 597672,
      "step": 1000
    },
    {
      "epoch": 0.048243921265920496,
      "grad_norm": 8.839406967163086,
      "learning_rate": 4.7587803936703975e-05,
      "loss": 3.3089,
      "num_input_tokens_seen": 896520,
      "step": 1500
    },
    {
      "epoch": 0.06432522835456066,
      "grad_norm": 7.105090618133545,
      "learning_rate": 4.678373858227197e-05,
      "loss": 3.267,
      "num_input_tokens_seen": 1194832,
      "step": 2000
    },
    {
      "epoch": 0.08040653544320082,
      "grad_norm": 7.992733001708984,
      "learning_rate": 4.597967322783996e-05,
      "loss": 3.2254,
      "num_input_tokens_seen": 1493088,
      "step": 2500
    },
    {
      "epoch": 0.08040653544320082,
      "eval_loss": 2.910461664199829,
      "eval_runtime": 2.5005,
      "eval_samples_per_second": 999.788,
      "eval_steps_per_second": 125.173,
      "num_input_tokens_seen": 1493088,
      "step": 2500
    },
    {
      "epoch": 0.09648784253184099,
      "grad_norm": 8.68657112121582,
      "learning_rate": 4.5175607873407955e-05,
      "loss": 3.1691,
      "num_input_tokens_seen": 1793976,
      "step": 3000
    },
    {
      "epoch": 0.11256914962048115,
      "grad_norm": 7.246800899505615,
      "learning_rate": 4.4371542518975943e-05,
      "loss": 3.1685,
      "num_input_tokens_seen": 2095352,
      "step": 3500
    },
    {
      "epoch": 0.1286504567091213,
      "grad_norm": 9.030860900878906,
      "learning_rate": 4.356747716454393e-05,
      "loss": 3.1333,
      "num_input_tokens_seen": 2393856,
      "step": 4000
    },
    {
      "epoch": 0.14473176379776148,
      "grad_norm": 7.463845252990723,
      "learning_rate": 4.276341181011193e-05,
      "loss": 3.1295,
      "num_input_tokens_seen": 2694496,
      "step": 4500
    },
    {
      "epoch": 0.16081307088640165,
      "grad_norm": 8.482089042663574,
      "learning_rate": 4.195934645567992e-05,
      "loss": 3.0946,
      "num_input_tokens_seen": 2990968,
      "step": 5000
    },
    {
      "epoch": 0.16081307088640165,
      "eval_loss": 2.830476999282837,
      "eval_runtime": 2.5201,
      "eval_samples_per_second": 992.009,
      "eval_steps_per_second": 124.199,
      "num_input_tokens_seen": 2990968,
      "step": 5000
    },
    {
      "epoch": 0.17689437797504182,
      "grad_norm": 8.570518493652344,
      "learning_rate": 4.115528110124791e-05,
      "loss": 3.112,
      "num_input_tokens_seen": 3289488,
      "step": 5500
    },
    {
      "epoch": 0.19297568506368198,
      "grad_norm": 9.759325981140137,
      "learning_rate": 4.03512157468159e-05,
      "loss": 3.0933,
      "num_input_tokens_seen": 3590264,
      "step": 6000
    },
    {
      "epoch": 0.20905699215232215,
      "grad_norm": 6.518988609313965,
      "learning_rate": 3.9547150392383896e-05,
      "loss": 3.0858,
      "num_input_tokens_seen": 3885160,
      "step": 6500
    },
    {
      "epoch": 0.2251382992409623,
      "grad_norm": 6.913475036621094,
      "learning_rate": 3.8743085037951885e-05,
      "loss": 3.0543,
      "num_input_tokens_seen": 4184600,
      "step": 7000
    },
    {
      "epoch": 0.24121960632960246,
      "grad_norm": 8.485562324523926,
      "learning_rate": 3.793901968351988e-05,
      "loss": 3.0473,
      "num_input_tokens_seen": 4477792,
      "step": 7500
    },
    {
      "epoch": 0.24121960632960246,
      "eval_loss": 2.773728847503662,
      "eval_runtime": 2.5738,
      "eval_samples_per_second": 971.323,
      "eval_steps_per_second": 121.61,
      "num_input_tokens_seen": 4477792,
      "step": 7500
    },
    {
      "epoch": 0.2573009134182426,
      "grad_norm": 7.89262056350708,
      "learning_rate": 3.713495432908787e-05,
      "loss": 3.029,
      "num_input_tokens_seen": 4779520,
      "step": 8000
    },
    {
      "epoch": 0.2733822205068828,
      "grad_norm": 6.879751205444336,
      "learning_rate": 3.6330888974655864e-05,
      "loss": 3.0127,
      "num_input_tokens_seen": 5078952,
      "step": 8500
    },
    {
      "epoch": 0.28946352759552296,
      "grad_norm": 8.109273910522461,
      "learning_rate": 3.552682362022385e-05,
      "loss": 3.0078,
      "num_input_tokens_seen": 5376128,
      "step": 9000
    },
    {
      "epoch": 0.3055448346841631,
      "grad_norm": 8.074146270751953,
      "learning_rate": 3.472275826579184e-05,
      "loss": 2.9988,
      "num_input_tokens_seen": 5671664,
      "step": 9500
    },
    {
      "epoch": 0.3216261417728033,
      "grad_norm": 6.523529529571533,
      "learning_rate": 3.391869291135984e-05,
      "loss": 2.9633,
      "num_input_tokens_seen": 5967560,
      "step": 10000
    },
    {
      "epoch": 0.3216261417728033,
      "eval_loss": 2.7306864261627197,
      "eval_runtime": 2.505,
      "eval_samples_per_second": 997.986,
      "eval_steps_per_second": 124.948,
      "num_input_tokens_seen": 5967560,
      "step": 10000
    },
    {
      "epoch": 0.33770744886144344,
      "grad_norm": 6.974269866943359,
      "learning_rate": 3.311462755692783e-05,
      "loss": 2.9732,
      "num_input_tokens_seen": 6265312,
      "step": 10500
    },
    {
      "epoch": 0.35378875595008363,
      "grad_norm": 7.644798278808594,
      "learning_rate": 3.231056220249582e-05,
      "loss": 2.9729,
      "num_input_tokens_seen": 6563632,
      "step": 11000
    },
    {
      "epoch": 0.3698700630387238,
      "grad_norm": 7.96437406539917,
      "learning_rate": 3.150649684806381e-05,
      "loss": 2.9484,
      "num_input_tokens_seen": 6865528,
      "step": 11500
    },
    {
      "epoch": 0.38595137012736397,
      "grad_norm": 7.939519882202148,
      "learning_rate": 3.0702431493631805e-05,
      "loss": 2.9387,
      "num_input_tokens_seen": 7165632,
      "step": 12000
    },
    {
      "epoch": 0.4020326772160041,
      "grad_norm": 7.698306083679199,
      "learning_rate": 2.9898366139199797e-05,
      "loss": 2.9355,
      "num_input_tokens_seen": 7463192,
      "step": 12500
    },
    {
      "epoch": 0.4020326772160041,
      "eval_loss": 2.684298515319824,
      "eval_runtime": 2.5214,
      "eval_samples_per_second": 991.52,
      "eval_steps_per_second": 124.138,
      "num_input_tokens_seen": 7463192,
      "step": 12500
    },
    {
      "epoch": 0.4181139843046443,
      "grad_norm": 7.21583890914917,
      "learning_rate": 2.9094300784767786e-05,
      "loss": 2.9418,
      "num_input_tokens_seen": 7758024,
      "step": 13000
    },
    {
      "epoch": 0.43419529139328444,
      "grad_norm": 7.767180919647217,
      "learning_rate": 2.8290235430335778e-05,
      "loss": 2.923,
      "num_input_tokens_seen": 8052032,
      "step": 13500
    },
    {
      "epoch": 0.4502765984819246,
      "grad_norm": 7.057159423828125,
      "learning_rate": 2.7486170075903773e-05,
      "loss": 2.9016,
      "num_input_tokens_seen": 8347768,
      "step": 14000
    },
    {
      "epoch": 0.4663579055705648,
      "grad_norm": 7.320003032684326,
      "learning_rate": 2.6682104721471762e-05,
      "loss": 2.885,
      "num_input_tokens_seen": 8646192,
      "step": 14500
    },
    {
      "epoch": 0.4824392126592049,
      "grad_norm": 7.630561828613281,
      "learning_rate": 2.587803936703975e-05,
      "loss": 2.9076,
      "num_input_tokens_seen": 8950264,
      "step": 15000
    },
    {
      "epoch": 0.4824392126592049,
      "eval_loss": 2.658709764480591,
      "eval_runtime": 2.6566,
      "eval_samples_per_second": 941.035,
      "eval_steps_per_second": 117.818,
      "num_input_tokens_seen": 8950264,
      "step": 15000
    },
    {
      "epoch": 0.4985205197478451,
      "grad_norm": 6.413602828979492,
      "learning_rate": 2.507397401260775e-05,
      "loss": 2.8818,
      "num_input_tokens_seen": 9247272,
      "step": 15500
    },
    {
      "epoch": 0.5146018268364853,
      "grad_norm": 6.833747863769531,
      "learning_rate": 2.426990865817574e-05,
      "loss": 2.8998,
      "num_input_tokens_seen": 9547432,
      "step": 16000
    },
    {
      "epoch": 0.5306831339251254,
      "grad_norm": 7.930506706237793,
      "learning_rate": 2.3465843303743727e-05,
      "loss": 2.8543,
      "num_input_tokens_seen": 9844072,
      "step": 16500
    },
    {
      "epoch": 0.5467644410137656,
      "grad_norm": 6.841344356536865,
      "learning_rate": 2.2661777949311722e-05,
      "loss": 2.8669,
      "num_input_tokens_seen": 10142344,
      "step": 17000
    },
    {
      "epoch": 0.5628457481024057,
      "grad_norm": 6.899343967437744,
      "learning_rate": 2.185771259487971e-05,
      "loss": 2.8714,
      "num_input_tokens_seen": 10443344,
      "step": 17500
    },
    {
      "epoch": 0.5628457481024057,
      "eval_loss": 2.6303601264953613,
      "eval_runtime": 2.5354,
      "eval_samples_per_second": 986.046,
      "eval_steps_per_second": 123.453,
      "num_input_tokens_seen": 10443344,
      "step": 17500
    },
    {
      "epoch": 0.5789270551910459,
      "grad_norm": 6.256689071655273,
      "learning_rate": 2.1053647240447703e-05,
      "loss": 2.8418,
      "num_input_tokens_seen": 10744312,
      "step": 18000
    },
    {
      "epoch": 0.5950083622796861,
      "grad_norm": 7.627821445465088,
      "learning_rate": 2.0249581886015695e-05,
      "loss": 2.8462,
      "num_input_tokens_seen": 11048208,
      "step": 18500
    },
    {
      "epoch": 0.6110896693683262,
      "grad_norm": 7.331953525543213,
      "learning_rate": 1.9445516531583687e-05,
      "loss": 2.8345,
      "num_input_tokens_seen": 11347880,
      "step": 19000
    },
    {
      "epoch": 0.6271709764569664,
      "grad_norm": 6.463207244873047,
      "learning_rate": 1.864145117715168e-05,
      "loss": 2.8531,
      "num_input_tokens_seen": 11650144,
      "step": 19500
    },
    {
      "epoch": 0.6432522835456066,
      "grad_norm": 7.423746109008789,
      "learning_rate": 1.783738582271967e-05,
      "loss": 2.8716,
      "num_input_tokens_seen": 11951096,
      "step": 20000
    },
    {
      "epoch": 0.6432522835456066,
      "eval_loss": 2.6024744510650635,
      "eval_runtime": 2.5182,
      "eval_samples_per_second": 992.755,
      "eval_steps_per_second": 124.293,
      "num_input_tokens_seen": 11951096,
      "step": 20000
    },
    {
      "epoch": 0.6593335906342468,
      "grad_norm": 7.352589130401611,
      "learning_rate": 1.7033320468287664e-05,
      "loss": 2.8243,
      "num_input_tokens_seen": 12252592,
      "step": 20500
    },
    {
      "epoch": 0.6754148977228869,
      "grad_norm": 7.22981071472168,
      "learning_rate": 1.6229255113855656e-05,
      "loss": 2.8454,
      "num_input_tokens_seen": 12546792,
      "step": 21000
    },
    {
      "epoch": 0.6914962048115271,
      "grad_norm": 6.819567680358887,
      "learning_rate": 1.5425189759423648e-05,
      "loss": 2.8047,
      "num_input_tokens_seen": 12838728,
      "step": 21500
    },
    {
      "epoch": 0.7075775119001673,
      "grad_norm": 8.716426849365234,
      "learning_rate": 1.4621124404991638e-05,
      "loss": 2.8144,
      "num_input_tokens_seen": 13137688,
      "step": 22000
    },
    {
      "epoch": 0.7236588189888075,
      "grad_norm": 7.324875831604004,
      "learning_rate": 1.381705905055963e-05,
      "loss": 2.7989,
      "num_input_tokens_seen": 13432464,
      "step": 22500
    },
    {
      "epoch": 0.7236588189888075,
      "eval_loss": 2.5822224617004395,
      "eval_runtime": 2.5158,
      "eval_samples_per_second": 993.705,
      "eval_steps_per_second": 124.412,
      "num_input_tokens_seen": 13432464,
      "step": 22500
    },
    {
      "epoch": 0.7397401260774475,
      "grad_norm": 7.962778568267822,
      "learning_rate": 1.301299369612762e-05,
      "loss": 2.7653,
      "num_input_tokens_seen": 13730024,
      "step": 23000
    },
    {
      "epoch": 0.7558214331660877,
      "grad_norm": 6.807019233703613,
      "learning_rate": 1.2208928341695614e-05,
      "loss": 2.7933,
      "num_input_tokens_seen": 14026400,
      "step": 23500
    },
    {
      "epoch": 0.7719027402547279,
      "grad_norm": 8.716556549072266,
      "learning_rate": 1.1404862987263605e-05,
      "loss": 2.7988,
      "num_input_tokens_seen": 14326608,
      "step": 24000
    },
    {
      "epoch": 0.787984047343368,
      "grad_norm": 7.388988018035889,
      "learning_rate": 1.0600797632831597e-05,
      "loss": 2.7928,
      "num_input_tokens_seen": 14623864,
      "step": 24500
    },
    {
      "epoch": 0.8040653544320082,
      "grad_norm": 7.011099815368652,
      "learning_rate": 9.796732278399589e-06,
      "loss": 2.7941,
      "num_input_tokens_seen": 14919424,
      "step": 25000
    },
    {
      "epoch": 0.8040653544320082,
      "eval_loss": 2.5630149841308594,
      "eval_runtime": 2.5601,
      "eval_samples_per_second": 976.534,
      "eval_steps_per_second": 122.262,
      "num_input_tokens_seen": 14919424,
      "step": 25000
    },
    {
      "epoch": 0.8201466615206484,
      "grad_norm": 6.740393161773682,
      "learning_rate": 8.992666923967581e-06,
      "loss": 2.8089,
      "num_input_tokens_seen": 15216136,
      "step": 25500
    },
    {
      "epoch": 0.8362279686092886,
      "grad_norm": 7.124479293823242,
      "learning_rate": 8.188601569535573e-06,
      "loss": 2.7704,
      "num_input_tokens_seen": 15515592,
      "step": 26000
    },
    {
      "epoch": 0.8523092756979287,
      "grad_norm": 7.781102180480957,
      "learning_rate": 7.384536215103564e-06,
      "loss": 2.8022,
      "num_input_tokens_seen": 15818560,
      "step": 26500
    },
    {
      "epoch": 0.8683905827865689,
      "grad_norm": 6.861135005950928,
      "learning_rate": 6.580470860671556e-06,
      "loss": 2.7891,
      "num_input_tokens_seen": 16114056,
      "step": 27000
    },
    {
      "epoch": 0.8844718898752091,
      "grad_norm": 7.128973484039307,
      "learning_rate": 5.776405506239547e-06,
      "loss": 2.7692,
      "num_input_tokens_seen": 16415080,
      "step": 27500
    },
    {
      "epoch": 0.8844718898752091,
      "eval_loss": 2.5496785640716553,
      "eval_runtime": 2.6422,
      "eval_samples_per_second": 946.175,
      "eval_steps_per_second": 118.461,
      "num_input_tokens_seen": 16415080,
      "step": 27500
    },
    {
      "epoch": 0.9005531969638492,
      "grad_norm": 8.560084342956543,
      "learning_rate": 4.9723401518075395e-06,
      "loss": 2.7627,
      "num_input_tokens_seen": 16711136,
      "step": 28000
    },
    {
      "epoch": 0.9166345040524894,
      "grad_norm": 7.5000224113464355,
      "learning_rate": 4.168274797375531e-06,
      "loss": 2.7687,
      "num_input_tokens_seen": 17005880,
      "step": 28500
    },
    {
      "epoch": 0.9327158111411296,
      "grad_norm": 6.699025630950928,
      "learning_rate": 3.3642094429435228e-06,
      "loss": 2.779,
      "num_input_tokens_seen": 17307064,
      "step": 29000
    },
    {
      "epoch": 0.9487971182297698,
      "grad_norm": 6.6417131423950195,
      "learning_rate": 2.560144088511514e-06,
      "loss": 2.7493,
      "num_input_tokens_seen": 17602296,
      "step": 29500
    },
    {
      "epoch": 0.9648784253184098,
      "grad_norm": 6.775792121887207,
      "learning_rate": 1.756078734079506e-06,
      "loss": 2.757,
      "num_input_tokens_seen": 17897832,
      "step": 30000
    },
    {
      "epoch": 0.9648784253184098,
      "eval_loss": 2.5388031005859375,
      "eval_runtime": 2.6123,
      "eval_samples_per_second": 957.018,
      "eval_steps_per_second": 119.819,
      "num_input_tokens_seen": 17897832,
      "step": 30000
    },
    {
      "epoch": 0.98095973240705,
      "grad_norm": 7.619235038757324,
      "learning_rate": 9.520133796474978e-07,
      "loss": 2.7433,
      "num_input_tokens_seen": 18195568,
      "step": 30500
    },
    {
      "epoch": 0.9970410394956902,
      "grad_norm": 6.682379722595215,
      "learning_rate": 1.479480252154895e-07,
      "loss": 2.7266,
      "num_input_tokens_seen": 18491904,
      "step": 31000
    },
    {
      "epoch": 1.0131223465843304,
      "grad_norm": 6.5948309898376465,
      "learning_rate": 3.311462755692783e-05,
      "loss": 2.6645,
      "num_input_tokens_seen": 18790628,
      "step": 31500
    },
    {
      "epoch": 1.0292036536729705,
      "grad_norm": 6.920671463012695,
      "learning_rate": 3.2846605772117164e-05,
      "loss": 2.6881,
      "num_input_tokens_seen": 19090780,
      "step": 32000
    },
    {
      "epoch": 1.0452849607616108,
      "grad_norm": 6.296219348907471,
      "learning_rate": 3.257858398730649e-05,
      "loss": 2.7024,
      "num_input_tokens_seen": 19384812,
      "step": 32500
    },
    {
      "epoch": 1.0452849607616108,
      "eval_loss": 2.6005640029907227,
      "eval_runtime": 2.5084,
      "eval_samples_per_second": 996.636,
      "eval_steps_per_second": 124.779,
      "num_input_tokens_seen": 19384812,
      "step": 32500
    },
    {
      "epoch": 1.061366267850251,
      "grad_norm": 7.068648815155029,
      "learning_rate": 3.231056220249582e-05,
      "loss": 2.6939,
      "num_input_tokens_seen": 19683060,
      "step": 33000
    },
    {
      "epoch": 1.077447574938891,
      "grad_norm": 5.753154754638672,
      "learning_rate": 3.204254041768515e-05,
      "loss": 2.6977,
      "num_input_tokens_seen": 19979196,
      "step": 33500
    },
    {
      "epoch": 1.0935288820275313,
      "grad_norm": 8.155505180358887,
      "learning_rate": 3.1774518632874485e-05,
      "loss": 2.7048,
      "num_input_tokens_seen": 20278524,
      "step": 34000
    },
    {
      "epoch": 1.1096101891161714,
      "grad_norm": 7.031659126281738,
      "learning_rate": 3.150649684806381e-05,
      "loss": 2.7237,
      "num_input_tokens_seen": 20577572,
      "step": 34500
    },
    {
      "epoch": 1.1256914962048115,
      "grad_norm": 7.90298318862915,
      "learning_rate": 3.123847506325314e-05,
      "loss": 2.7248,
      "num_input_tokens_seen": 20876844,
      "step": 35000
    },
    {
      "epoch": 1.1256914962048115,
      "eval_loss": 2.6041972637176514,
      "eval_runtime": 2.7564,
      "eval_samples_per_second": 906.969,
      "eval_steps_per_second": 113.552,
      "num_input_tokens_seen": 20876844,
      "step": 35000
    },
    {
      "epoch": 1.1417728032934518,
      "grad_norm": 6.368433475494385,
      "learning_rate": 3.0970453278442473e-05,
      "loss": 2.7246,
      "num_input_tokens_seen": 21179820,
      "step": 35500
    },
    {
      "epoch": 1.1578541103820919,
      "grad_norm": 7.143220901489258,
      "learning_rate": 3.0702431493631805e-05,
      "loss": 2.7211,
      "num_input_tokens_seen": 21476172,
      "step": 36000
    },
    {
      "epoch": 1.173935417470732,
      "grad_norm": 7.216341972351074,
      "learning_rate": 3.0434409708821134e-05,
      "loss": 2.7088,
      "num_input_tokens_seen": 21774292,
      "step": 36500
    },
    {
      "epoch": 1.1900167245593722,
      "grad_norm": 6.958596706390381,
      "learning_rate": 3.0166387924010465e-05,
      "loss": 2.7166,
      "num_input_tokens_seen": 22070908,
      "step": 37000
    },
    {
      "epoch": 1.2060980316480123,
      "grad_norm": 7.161530494689941,
      "learning_rate": 2.9898366139199797e-05,
      "loss": 2.6764,
      "num_input_tokens_seen": 22372340,
      "step": 37500
    },
    {
      "epoch": 1.2060980316480123,
      "eval_loss": 2.5923423767089844,
      "eval_runtime": 2.6849,
      "eval_samples_per_second": 931.145,
      "eval_steps_per_second": 116.579,
      "num_input_tokens_seen": 22372340,
      "step": 37500
    },
    {
      "epoch": 1.2221793387366526,
      "grad_norm": 7.448612213134766,
      "learning_rate": 2.963034435438913e-05,
      "loss": 2.7098,
      "num_input_tokens_seen": 22672932,
      "step": 38000
    },
    {
      "epoch": 1.2382606458252927,
      "grad_norm": 8.339189529418945,
      "learning_rate": 2.9362322569578454e-05,
      "loss": 2.702,
      "num_input_tokens_seen": 22971844,
      "step": 38500
    },
    {
      "epoch": 1.2543419529139328,
      "grad_norm": 6.795124053955078,
      "learning_rate": 2.9094300784767786e-05,
      "loss": 2.7007,
      "num_input_tokens_seen": 23266964,
      "step": 39000
    },
    {
      "epoch": 1.2704232600025729,
      "grad_norm": 6.3036298751831055,
      "learning_rate": 2.8826278999957118e-05,
      "loss": 2.71,
      "num_input_tokens_seen": 23564068,
      "step": 39500
    },
    {
      "epoch": 1.2865045670912132,
      "grad_norm": 8.75069808959961,
      "learning_rate": 2.855825721514645e-05,
      "loss": 2.6854,
      "num_input_tokens_seen": 23866100,
      "step": 40000
    },
    {
      "epoch": 1.2865045670912132,
      "eval_loss": 2.5792863368988037,
      "eval_runtime": 2.5776,
      "eval_samples_per_second": 969.878,
      "eval_steps_per_second": 121.429,
      "num_input_tokens_seen": 23866100,
      "step": 40000
    },
    {
      "epoch": 1.3025858741798533,
      "grad_norm": 6.966170310974121,
      "learning_rate": 2.8290235430335778e-05,
      "loss": 2.697,
      "num_input_tokens_seen": 24162356,
      "step": 40500
    },
    {
      "epoch": 1.3186671812684936,
      "grad_norm": 7.854964733123779,
      "learning_rate": 2.802221364552511e-05,
      "loss": 2.6954,
      "num_input_tokens_seen": 24458980,
      "step": 41000
    },
    {
      "epoch": 1.3347484883571337,
      "grad_norm": 7.1461944580078125,
      "learning_rate": 2.775419186071444e-05,
      "loss": 2.6839,
      "num_input_tokens_seen": 24757828,
      "step": 41500
    },
    {
      "epoch": 1.3508297954457738,
      "grad_norm": 8.25295639038086,
      "learning_rate": 2.7486170075903773e-05,
      "loss": 2.7035,
      "num_input_tokens_seen": 25052236,
      "step": 42000
    },
    {
      "epoch": 1.366911102534414,
      "grad_norm": 6.336223602294922,
      "learning_rate": 2.7218148291093105e-05,
      "loss": 2.683,
      "num_input_tokens_seen": 25348084,
      "step": 42500
    },
    {
      "epoch": 1.366911102534414,
      "eval_loss": 2.5722219944000244,
      "eval_runtime": 2.7384,
      "eval_samples_per_second": 912.958,
      "eval_steps_per_second": 114.302,
      "num_input_tokens_seen": 25348084,
      "step": 42500
    },
    {
      "epoch": 1.3829924096230541,
      "grad_norm": 9.477555274963379,
      "learning_rate": 2.695012650628243e-05,
      "loss": 2.6877,
      "num_input_tokens_seen": 25642372,
      "step": 43000
    },
    {
      "epoch": 1.3990737167116944,
      "grad_norm": 8.233431816101074,
      "learning_rate": 2.6682104721471762e-05,
      "loss": 2.6927,
      "num_input_tokens_seen": 25939652,
      "step": 43500
    },
    {
      "epoch": 1.4151550238003345,
      "grad_norm": 5.860446929931641,
      "learning_rate": 2.6414082936661094e-05,
      "loss": 2.6819,
      "num_input_tokens_seen": 26248940,
      "step": 44000
    },
    {
      "epoch": 1.4312363308889746,
      "grad_norm": 6.748124599456787,
      "learning_rate": 2.6146061151850426e-05,
      "loss": 2.6893,
      "num_input_tokens_seen": 26552860,
      "step": 44500
    },
    {
      "epoch": 1.4473176379776147,
      "grad_norm": 6.038182258605957,
      "learning_rate": 2.587803936703975e-05,
      "loss": 2.6871,
      "num_input_tokens_seen": 26854100,
      "step": 45000
    },
    {
      "epoch": 1.4473176379776147,
      "eval_loss": 2.5538456439971924,
      "eval_runtime": 2.6078,
      "eval_samples_per_second": 958.67,
      "eval_steps_per_second": 120.026,
      "num_input_tokens_seen": 26854100,
      "step": 45000
    },
    {
      "epoch": 1.463398945066255,
      "grad_norm": 7.815784454345703,
      "learning_rate": 2.5610017582229086e-05,
      "loss": 2.6709,
      "num_input_tokens_seen": 27148148,
      "step": 45500
    },
    {
      "epoch": 1.479480252154895,
      "grad_norm": 7.8851094245910645,
      "learning_rate": 2.5341995797418418e-05,
      "loss": 2.6698,
      "num_input_tokens_seen": 27445020,
      "step": 46000
    },
    {
      "epoch": 1.4955615592435354,
      "grad_norm": 7.389246940612793,
      "learning_rate": 2.507397401260775e-05,
      "loss": 2.6787,
      "num_input_tokens_seen": 27742908,
      "step": 46500
    },
    {
      "epoch": 1.5116428663321755,
      "grad_norm": 7.621913909912109,
      "learning_rate": 2.4805952227797078e-05,
      "loss": 2.6713,
      "num_input_tokens_seen": 28037284,
      "step": 47000
    },
    {
      "epoch": 1.5277241734208156,
      "grad_norm": 7.889066219329834,
      "learning_rate": 2.4537930442986407e-05,
      "loss": 2.6551,
      "num_input_tokens_seen": 28332612,
      "step": 47500
    },
    {
      "epoch": 1.5277241734208156,
      "eval_loss": 2.5442593097686768,
      "eval_runtime": 2.6341,
      "eval_samples_per_second": 949.086,
      "eval_steps_per_second": 118.826,
      "num_input_tokens_seen": 28332612,
      "step": 47500
    },
    {
      "epoch": 1.5438054805094557,
      "grad_norm": 7.912906646728516,
      "learning_rate": 2.426990865817574e-05,
      "loss": 2.6881,
      "num_input_tokens_seen": 28630948,
      "step": 48000
    },
    {
      "epoch": 1.559886787598096,
      "grad_norm": 6.370878219604492,
      "learning_rate": 2.4001886873365067e-05,
      "loss": 2.6424,
      "num_input_tokens_seen": 28928732,
      "step": 48500
    },
    {
      "epoch": 1.5759680946867363,
      "grad_norm": 7.0892653465271,
      "learning_rate": 2.37338650885544e-05,
      "loss": 2.6626,
      "num_input_tokens_seen": 29224436,
      "step": 49000
    },
    {
      "epoch": 1.5920494017753763,
      "grad_norm": 6.357864856719971,
      "learning_rate": 2.3465843303743727e-05,
      "loss": 2.6546,
      "num_input_tokens_seen": 29520148,
      "step": 49500
    },
    {
      "epoch": 1.6081307088640164,
      "grad_norm": 8.4866943359375,
      "learning_rate": 2.319782151893306e-05,
      "loss": 2.661,
      "num_input_tokens_seen": 29822156,
      "step": 50000
    },
    {
      "epoch": 1.6081307088640164,
      "eval_loss": 2.527804374694824,
      "eval_runtime": 2.6771,
      "eval_samples_per_second": 933.855,
      "eval_steps_per_second": 116.919,
      "num_input_tokens_seen": 29822156,
      "step": 50000
    },
    {
      "epoch": 1.6242120159526565,
      "grad_norm": 6.843733787536621,
      "learning_rate": 2.292979973412239e-05,
      "loss": 2.6521,
      "num_input_tokens_seen": 30122052,
      "step": 50500
    },
    {
      "epoch": 1.6402933230412968,
      "grad_norm": 6.88835334777832,
      "learning_rate": 2.2661777949311722e-05,
      "loss": 2.6614,
      "num_input_tokens_seen": 30422196,
      "step": 51000
    },
    {
      "epoch": 1.656374630129937,
      "grad_norm": 5.855214595794678,
      "learning_rate": 2.239375616450105e-05,
      "loss": 2.6454,
      "num_input_tokens_seen": 30722660,
      "step": 51500
    },
    {
      "epoch": 1.6724559372185772,
      "grad_norm": 6.58035135269165,
      "learning_rate": 2.2125734379690383e-05,
      "loss": 2.6524,
      "num_input_tokens_seen": 31018220,
      "step": 52000
    },
    {
      "epoch": 1.6885372443072173,
      "grad_norm": 6.767495155334473,
      "learning_rate": 2.185771259487971e-05,
      "loss": 2.6497,
      "num_input_tokens_seen": 31319476,
      "step": 52500
    },
    {
      "epoch": 1.6885372443072173,
      "eval_loss": 2.526638984680176,
      "eval_runtime": 2.5213,
      "eval_samples_per_second": 991.545,
      "eval_steps_per_second": 124.141,
      "num_input_tokens_seen": 31319476,
      "step": 52500
    },
    {
      "epoch": 1.7046185513958574,
      "grad_norm": 7.022729873657227,
      "learning_rate": 2.1589690810069043e-05,
      "loss": 2.6437,
      "num_input_tokens_seen": 31621308,
      "step": 53000
    },
    {
      "epoch": 1.7206998584844975,
      "grad_norm": 6.241069793701172,
      "learning_rate": 2.132166902525837e-05,
      "loss": 2.6447,
      "num_input_tokens_seen": 31917460,
      "step": 53500
    },
    {
      "epoch": 1.7367811655731378,
      "grad_norm": 7.7204084396362305,
      "learning_rate": 2.1053647240447703e-05,
      "loss": 2.6448,
      "num_input_tokens_seen": 32217596,
      "step": 54000
    },
    {
      "epoch": 1.752862472661778,
      "grad_norm": 6.703210830688477,
      "learning_rate": 2.0785625455637035e-05,
      "loss": 2.6366,
      "num_input_tokens_seen": 32513884,
      "step": 54500
    },
    {
      "epoch": 1.7689437797504182,
      "grad_norm": 6.371466159820557,
      "learning_rate": 2.0517603670826367e-05,
      "loss": 2.6281,
      "num_input_tokens_seen": 32813220,
      "step": 55000
    },
    {
      "epoch": 1.7689437797504182,
      "eval_loss": 2.5115973949432373,
      "eval_runtime": 2.5216,
      "eval_samples_per_second": 991.443,
      "eval_steps_per_second": 124.129,
      "num_input_tokens_seen": 32813220,
      "step": 55000
    },
    {
      "epoch": 1.7850250868390583,
      "grad_norm": 7.277946949005127,
      "learning_rate": 2.0249581886015695e-05,
      "loss": 2.6536,
      "num_input_tokens_seen": 33110188,
      "step": 55500
    },
    {
      "epoch": 1.8011063939276983,
      "grad_norm": 7.93104887008667,
      "learning_rate": 1.9981560101205027e-05,
      "loss": 2.5981,
      "num_input_tokens_seen": 33405980,
      "step": 56000
    },
    {
      "epoch": 1.8171877010163386,
      "grad_norm": 7.782486438751221,
      "learning_rate": 1.971353831639436e-05,
      "loss": 2.635,
      "num_input_tokens_seen": 33700596,
      "step": 56500
    },
    {
      "epoch": 1.8332690081049787,
      "grad_norm": 8.59358024597168,
      "learning_rate": 1.9445516531583687e-05,
      "loss": 2.6269,
      "num_input_tokens_seen": 33997724,
      "step": 57000
    },
    {
      "epoch": 1.849350315193619,
      "grad_norm": 6.669950485229492,
      "learning_rate": 1.917749474677302e-05,
      "loss": 2.6067,
      "num_input_tokens_seen": 34298052,
      "step": 57500
    },
    {
      "epoch": 1.849350315193619,
      "eval_loss": 2.5047078132629395,
      "eval_runtime": 2.5169,
      "eval_samples_per_second": 993.288,
      "eval_steps_per_second": 124.36,
      "num_input_tokens_seen": 34298052,
      "step": 57500
    },
    {
      "epoch": 1.8654316222822591,
      "grad_norm": 6.265903949737549,
      "learning_rate": 1.8909472961962348e-05,
      "loss": 2.5966,
      "num_input_tokens_seen": 34593980,
      "step": 58000
    },
    {
      "epoch": 1.8815129293708992,
      "grad_norm": 7.943974018096924,
      "learning_rate": 1.864145117715168e-05,
      "loss": 2.6303,
      "num_input_tokens_seen": 34894428,
      "step": 58500
    },
    {
      "epoch": 1.8975942364595393,
      "grad_norm": 8.290629386901855,
      "learning_rate": 1.837342939234101e-05,
      "loss": 2.6303,
      "num_input_tokens_seen": 35193236,
      "step": 59000
    },
    {
      "epoch": 1.9136755435481796,
      "grad_norm": 7.974947929382324,
      "learning_rate": 1.8105407607530343e-05,
      "loss": 2.6272,
      "num_input_tokens_seen": 35486796,
      "step": 59500
    },
    {
      "epoch": 1.92975685063682,
      "grad_norm": 5.827637195587158,
      "learning_rate": 1.783738582271967e-05,
      "loss": 2.6112,
      "num_input_tokens_seen": 35783604,
      "step": 60000
    },
    {
      "epoch": 1.92975685063682,
      "eval_loss": 2.4935405254364014,
      "eval_runtime": 2.5107,
      "eval_samples_per_second": 995.741,
      "eval_steps_per_second": 124.667,
      "num_input_tokens_seen": 35783604,
      "step": 60000
    },
    {
      "epoch": 1.94583815772546,
      "grad_norm": 6.535378456115723,
      "learning_rate": 1.7569364037909003e-05,
      "loss": 2.6135,
      "num_input_tokens_seen": 36086620,
      "step": 60500
    },
    {
      "epoch": 1.9619194648141,
      "grad_norm": 6.398725986480713,
      "learning_rate": 1.7301342253098332e-05,
      "loss": 2.602,
      "num_input_tokens_seen": 36386140,
      "step": 61000
    },
    {
      "epoch": 1.9780007719027402,
      "grad_norm": 6.332113265991211,
      "learning_rate": 1.7033320468287664e-05,
      "loss": 2.6258,
      "num_input_tokens_seen": 36684308,
      "step": 61500
    },
    {
      "epoch": 1.9940820789913805,
      "grad_norm": 7.8002753257751465,
      "learning_rate": 1.6765298683476992e-05,
      "loss": 2.6226,
      "num_input_tokens_seen": 36984724,
      "step": 62000
    },
    {
      "epoch": 2.0101633860800208,
      "grad_norm": 6.9457011222839355,
      "learning_rate": 1.6497276898666324e-05,
      "loss": 2.5207,
      "num_input_tokens_seen": 37281092,
      "step": 62500
    },
    {
      "epoch": 2.0101633860800208,
      "eval_loss": 2.4945950508117676,
      "eval_runtime": 2.5094,
      "eval_samples_per_second": 996.26,
      "eval_steps_per_second": 124.732,
      "num_input_tokens_seen": 37281092,
      "step": 62500
    },
    {
      "epoch": 2.026244693168661,
      "grad_norm": 7.541498184204102,
      "learning_rate": 1.6229255113855656e-05,
      "loss": 2.4728,
      "num_input_tokens_seen": 37582300,
      "step": 63000
    },
    {
      "epoch": 2.042326000257301,
      "grad_norm": 6.7798027992248535,
      "learning_rate": 1.5961233329044987e-05,
      "loss": 2.4539,
      "num_input_tokens_seen": 37880828,
      "step": 63500
    },
    {
      "epoch": 2.058407307345941,
      "grad_norm": 7.033351898193359,
      "learning_rate": 1.5693211544234316e-05,
      "loss": 2.4467,
      "num_input_tokens_seen": 38181276,
      "step": 64000
    },
    {
      "epoch": 2.074488614434581,
      "grad_norm": 6.487890720367432,
      "learning_rate": 1.5425189759423648e-05,
      "loss": 2.4764,
      "num_input_tokens_seen": 38473348,
      "step": 64500
    },
    {
      "epoch": 2.0905699215232216,
      "grad_norm": 6.955127716064453,
      "learning_rate": 1.5157167974612976e-05,
      "loss": 2.4799,
      "num_input_tokens_seen": 38768588,
      "step": 65000
    },
    {
      "epoch": 2.0905699215232216,
      "eval_loss": 2.491555690765381,
      "eval_runtime": 2.5076,
      "eval_samples_per_second": 996.967,
      "eval_steps_per_second": 124.82,
      "num_input_tokens_seen": 38768588,
      "step": 65000
    },
    {
      "epoch": 2.1066512286118617,
      "grad_norm": 6.78762674331665,
      "learning_rate": 1.4889146189802308e-05,
      "loss": 2.4726,
      "num_input_tokens_seen": 39067460,
      "step": 65500
    },
    {
      "epoch": 2.122732535700502,
      "grad_norm": 7.199331283569336,
      "learning_rate": 1.4621124404991638e-05,
      "loss": 2.4562,
      "num_input_tokens_seen": 39360244,
      "step": 66000
    },
    {
      "epoch": 2.138813842789142,
      "grad_norm": 7.353775501251221,
      "learning_rate": 1.435310262018097e-05,
      "loss": 2.4629,
      "num_input_tokens_seen": 39660020,
      "step": 66500
    },
    {
      "epoch": 2.154895149877782,
      "grad_norm": 6.827337265014648,
      "learning_rate": 1.4085080835370298e-05,
      "loss": 2.4817,
      "num_input_tokens_seen": 39960476,
      "step": 67000
    },
    {
      "epoch": 2.170976456966422,
      "grad_norm": 6.532020092010498,
      "learning_rate": 1.381705905055963e-05,
      "loss": 2.4727,
      "num_input_tokens_seen": 40252972,
      "step": 67500
    },
    {
      "epoch": 2.170976456966422,
      "eval_loss": 2.4865615367889404,
      "eval_runtime": 2.5372,
      "eval_samples_per_second": 985.335,
      "eval_steps_per_second": 123.364,
      "num_input_tokens_seen": 40252972,
      "step": 67500
    },
    {
      "epoch": 2.1870577640550626,
      "grad_norm": 6.601158142089844,
      "learning_rate": 1.354903726574896e-05,
      "loss": 2.4666,
      "num_input_tokens_seen": 40553732,
      "step": 68000
    },
    {
      "epoch": 2.2031390711437027,
      "grad_norm": 7.200645446777344,
      "learning_rate": 1.3281015480938292e-05,
      "loss": 2.4657,
      "num_input_tokens_seen": 40851196,
      "step": 68500
    },
    {
      "epoch": 2.2192203782323427,
      "grad_norm": 8.067240715026855,
      "learning_rate": 1.301299369612762e-05,
      "loss": 2.4801,
      "num_input_tokens_seen": 41149276,
      "step": 69000
    },
    {
      "epoch": 2.235301685320983,
      "grad_norm": 7.724194526672363,
      "learning_rate": 1.2744971911316952e-05,
      "loss": 2.4766,
      "num_input_tokens_seen": 41448540,
      "step": 69500
    },
    {
      "epoch": 2.251382992409623,
      "grad_norm": 6.999200344085693,
      "learning_rate": 1.2476950126506282e-05,
      "loss": 2.4719,
      "num_input_tokens_seen": 41746300,
      "step": 70000
    },
    {
      "epoch": 2.251382992409623,
      "eval_loss": 2.476020097732544,
      "eval_runtime": 2.5166,
      "eval_samples_per_second": 993.407,
      "eval_steps_per_second": 124.375,
      "num_input_tokens_seen": 41746300,
      "step": 70000
    },
    {
      "epoch": 2.267464299498263,
      "grad_norm": 6.666884899139404,
      "learning_rate": 1.2208928341695614e-05,
      "loss": 2.4771,
      "num_input_tokens_seen": 42042532,
      "step": 70500
    },
    {
      "epoch": 2.2835456065869035,
      "grad_norm": 8.354509353637695,
      "learning_rate": 1.1940906556884944e-05,
      "loss": 2.4679,
      "num_input_tokens_seen": 42341628,
      "step": 71000
    },
    {
      "epoch": 2.2996269136755436,
      "grad_norm": 8.39284610748291,
      "learning_rate": 1.1672884772074275e-05,
      "loss": 2.4597,
      "num_input_tokens_seen": 42642948,
      "step": 71500
    },
    {
      "epoch": 2.3157082207641837,
      "grad_norm": 7.233700275421143,
      "learning_rate": 1.1404862987263605e-05,
      "loss": 2.4592,
      "num_input_tokens_seen": 42941820,
      "step": 72000
    },
    {
      "epoch": 2.331789527852824,
      "grad_norm": 7.843503475189209,
      "learning_rate": 1.1136841202452935e-05,
      "loss": 2.4738,
      "num_input_tokens_seen": 43241188,
      "step": 72500
    },
    {
      "epoch": 2.331789527852824,
      "eval_loss": 2.47127103805542,
      "eval_runtime": 2.5277,
      "eval_samples_per_second": 989.05,
      "eval_steps_per_second": 123.829,
      "num_input_tokens_seen": 43241188,
      "step": 72500
    },
    {
      "epoch": 2.347870834941464,
      "grad_norm": 6.37482213973999,
      "learning_rate": 1.0868819417642267e-05,
      "loss": 2.4576,
      "num_input_tokens_seen": 43535900,
      "step": 73000
    },
    {
      "epoch": 2.3639521420301044,
      "grad_norm": 6.642532825469971,
      "learning_rate": 1.0600797632831597e-05,
      "loss": 2.467,
      "num_input_tokens_seen": 43833516,
      "step": 73500
    },
    {
      "epoch": 2.3800334491187445,
      "grad_norm": 6.606197357177734,
      "learning_rate": 1.0332775848020927e-05,
      "loss": 2.4752,
      "num_input_tokens_seen": 44134084,
      "step": 74000
    },
    {
      "epoch": 2.3961147562073846,
      "grad_norm": 6.338978290557861,
      "learning_rate": 1.0064754063210257e-05,
      "loss": 2.4473,
      "num_input_tokens_seen": 44432540,
      "step": 74500
    },
    {
      "epoch": 2.4121960632960247,
      "grad_norm": 7.172792434692383,
      "learning_rate": 9.796732278399589e-06,
      "loss": 2.4629,
      "num_input_tokens_seen": 44730244,
      "step": 75000
    },
    {
      "epoch": 2.4121960632960247,
      "eval_loss": 2.4629955291748047,
      "eval_runtime": 2.521,
      "eval_samples_per_second": 991.68,
      "eval_steps_per_second": 124.158,
      "num_input_tokens_seen": 44730244,
      "step": 75000
    },
    {
      "epoch": 2.4282773703846647,
      "grad_norm": 6.65930700302124,
      "learning_rate": 9.528710493588919e-06,
      "loss": 2.4512,
      "num_input_tokens_seen": 45031884,
      "step": 75500
    },
    {
      "epoch": 2.4443586774733053,
      "grad_norm": 7.209745407104492,
      "learning_rate": 9.260688708778249e-06,
      "loss": 2.4557,
      "num_input_tokens_seen": 45334924,
      "step": 76000
    },
    {
      "epoch": 2.4604399845619453,
      "grad_norm": 6.847073078155518,
      "learning_rate": 8.992666923967581e-06,
      "loss": 2.4512,
      "num_input_tokens_seen": 45633532,
      "step": 76500
    },
    {
      "epoch": 2.4765212916505854,
      "grad_norm": 7.705162525177002,
      "learning_rate": 8.724645139156911e-06,
      "loss": 2.4568,
      "num_input_tokens_seen": 45933804,
      "step": 77000
    },
    {
      "epoch": 2.4926025987392255,
      "grad_norm": 7.5681962966918945,
      "learning_rate": 8.456623354346243e-06,
      "loss": 2.4524,
      "num_input_tokens_seen": 46231060,
      "step": 77500
    },
    {
      "epoch": 2.4926025987392255,
      "eval_loss": 2.457481861114502,
      "eval_runtime": 2.5921,
      "eval_samples_per_second": 964.467,
      "eval_steps_per_second": 120.751,
      "num_input_tokens_seen": 46231060,
      "step": 77500
    },
    {
      "epoch": 2.5086839058278656,
      "grad_norm": 6.8857269287109375,
      "learning_rate": 8.188601569535573e-06,
      "loss": 2.4622,
      "num_input_tokens_seen": 46525772,
      "step": 78000
    },
    {
      "epoch": 2.524765212916506,
      "grad_norm": 6.347681522369385,
      "learning_rate": 7.920579784724903e-06,
      "loss": 2.4528,
      "num_input_tokens_seen": 46822532,
      "step": 78500
    },
    {
      "epoch": 2.5408465200051458,
      "grad_norm": 6.935575008392334,
      "learning_rate": 7.652557999914233e-06,
      "loss": 2.4414,
      "num_input_tokens_seen": 47122964,
      "step": 79000
    },
    {
      "epoch": 2.5569278270937863,
      "grad_norm": 6.603360652923584,
      "learning_rate": 7.384536215103564e-06,
      "loss": 2.4655,
      "num_input_tokens_seen": 47423300,
      "step": 79500
    },
    {
      "epoch": 2.5730091341824264,
      "grad_norm": 7.182071208953857,
      "learning_rate": 7.116514430292895e-06,
      "loss": 2.435,
      "num_input_tokens_seen": 47718964,
      "step": 80000
    },
    {
      "epoch": 2.5730091341824264,
      "eval_loss": 2.455320358276367,
      "eval_runtime": 2.5065,
      "eval_samples_per_second": 997.4,
      "eval_steps_per_second": 124.874,
      "num_input_tokens_seen": 47718964,
      "step": 80000
    },
    {
      "epoch": 2.5890904412710665,
      "grad_norm": 7.3647260665893555,
      "learning_rate": 6.848492645482225e-06,
      "loss": 2.4356,
      "num_input_tokens_seen": 48015996,
      "step": 80500
    },
    {
      "epoch": 2.6051717483597066,
      "grad_norm": 7.950341701507568,
      "learning_rate": 6.580470860671556e-06,
      "loss": 2.4453,
      "num_input_tokens_seen": 48316420,
      "step": 81000
    },
    {
      "epoch": 2.6212530554483466,
      "grad_norm": 6.016787052154541,
      "learning_rate": 6.312449075860886e-06,
      "loss": 2.45,
      "num_input_tokens_seen": 48611452,
      "step": 81500
    },
    {
      "epoch": 2.637334362536987,
      "grad_norm": 7.281980514526367,
      "learning_rate": 6.044427291050217e-06,
      "loss": 2.4687,
      "num_input_tokens_seen": 48913668,
      "step": 82000
    },
    {
      "epoch": 2.6534156696256272,
      "grad_norm": 6.644787311553955,
      "learning_rate": 5.776405506239547e-06,
      "loss": 2.4621,
      "num_input_tokens_seen": 49209724,
      "step": 82500
    },
    {
      "epoch": 2.6534156696256272,
      "eval_loss": 2.4475488662719727,
      "eval_runtime": 2.517,
      "eval_samples_per_second": 993.239,
      "eval_steps_per_second": 124.354,
      "num_input_tokens_seen": 49209724,
      "step": 82500
    },
    {
      "epoch": 2.6694969767142673,
      "grad_norm": 6.181220054626465,
      "learning_rate": 5.508383721428878e-06,
      "loss": 2.4343,
      "num_input_tokens_seen": 49505772,
      "step": 83000
    },
    {
      "epoch": 2.6855782838029074,
      "grad_norm": 6.418393135070801,
      "learning_rate": 5.2403619366182085e-06,
      "loss": 2.4329,
      "num_input_tokens_seen": 49809956,
      "step": 83500
    },
    {
      "epoch": 2.7016595908915475,
      "grad_norm": 6.279716491699219,
      "learning_rate": 4.9723401518075395e-06,
      "loss": 2.4481,
      "num_input_tokens_seen": 50112060,
      "step": 84000
    },
    {
      "epoch": 2.717740897980188,
      "grad_norm": 6.502873420715332,
      "learning_rate": 4.70431836699687e-06,
      "loss": 2.4464,
      "num_input_tokens_seen": 50414356,
      "step": 84500
    },
    {
      "epoch": 2.733822205068828,
      "grad_norm": 6.15990686416626,
      "learning_rate": 4.4362965821862e-06,
      "loss": 2.4492,
      "num_input_tokens_seen": 50712980,
      "step": 85000
    },
    {
      "epoch": 2.733822205068828,
      "eval_loss": 2.4440150260925293,
      "eval_runtime": 2.529,
      "eval_samples_per_second": 988.544,
      "eval_steps_per_second": 123.766,
      "num_input_tokens_seen": 50712980,
      "step": 85000
    },
    {
      "epoch": 2.749903512157468,
      "grad_norm": 6.876352310180664,
      "learning_rate": 4.168274797375531e-06,
      "loss": 2.4514,
      "num_input_tokens_seen": 51012460,
      "step": 85500
    },
    {
      "epoch": 2.7659848192461083,
      "grad_norm": 7.305426597595215,
      "learning_rate": 3.900253012564861e-06,
      "loss": 2.4317,
      "num_input_tokens_seen": 51308524,
      "step": 86000
    },
    {
      "epoch": 2.7820661263347484,
      "grad_norm": 6.460892677307129,
      "learning_rate": 3.632231227754192e-06,
      "loss": 2.4559,
      "num_input_tokens_seen": 51610700,
      "step": 86500
    },
    {
      "epoch": 2.798147433423389,
      "grad_norm": 8.062651634216309,
      "learning_rate": 3.3642094429435228e-06,
      "loss": 2.4535,
      "num_input_tokens_seen": 51910236,
      "step": 87000
    },
    {
      "epoch": 2.814228740512029,
      "grad_norm": 7.140311241149902,
      "learning_rate": 3.0961876581328533e-06,
      "loss": 2.4536,
      "num_input_tokens_seen": 52204380,
      "step": 87500
    },
    {
      "epoch": 2.814228740512029,
      "eval_loss": 2.4393906593322754,
      "eval_runtime": 2.5312,
      "eval_samples_per_second": 987.685,
      "eval_steps_per_second": 123.658,
      "num_input_tokens_seen": 52204380,
      "step": 87500
    },
    {
      "epoch": 2.830310047600669,
      "grad_norm": 6.569787502288818,
      "learning_rate": 2.8281658733221834e-06,
      "loss": 2.4379,
      "num_input_tokens_seen": 52504668,
      "step": 88000
    },
    {
      "epoch": 2.846391354689309,
      "grad_norm": 7.735711097717285,
      "learning_rate": 2.560144088511514e-06,
      "loss": 2.4239,
      "num_input_tokens_seen": 52798740,
      "step": 88500
    },
    {
      "epoch": 2.8624726617779492,
      "grad_norm": 7.504124641418457,
      "learning_rate": 2.292122303700845e-06,
      "loss": 2.4427,
      "num_input_tokens_seen": 53097716,
      "step": 89000
    },
    {
      "epoch": 2.8785539688665893,
      "grad_norm": 6.647756099700928,
      "learning_rate": 2.0241005188901755e-06,
      "loss": 2.4682,
      "num_input_tokens_seen": 53397564,
      "step": 89500
    },
    {
      "epoch": 2.8946352759552294,
      "grad_norm": 6.640815734863281,
      "learning_rate": 1.756078734079506e-06,
      "loss": 2.4148,
      "num_input_tokens_seen": 53695620,
      "step": 90000
    },
    {
      "epoch": 2.8946352759552294,
      "eval_loss": 2.43597674369812,
      "eval_runtime": 2.5128,
      "eval_samples_per_second": 994.891,
      "eval_steps_per_second": 124.56,
      "num_input_tokens_seen": 53695620,
      "step": 90000
    },
    {
      "epoch": 2.91071658304387,
      "grad_norm": 7.346447467803955,
      "learning_rate": 1.4880569492688366e-06,
      "loss": 2.4352,
      "num_input_tokens_seen": 53991180,
      "step": 90500
    },
    {
      "epoch": 2.92679789013251,
      "grad_norm": 6.777767658233643,
      "learning_rate": 1.2200351644581672e-06,
      "loss": 2.4664,
      "num_input_tokens_seen": 54288348,
      "step": 91000
    },
    {
      "epoch": 2.94287919722115,
      "grad_norm": 6.908254623413086,
      "learning_rate": 9.520133796474978e-07,
      "loss": 2.4474,
      "num_input_tokens_seen": 54590740,
      "step": 91500
    },
    {
      "epoch": 2.95896050430979,
      "grad_norm": 7.04544734954834,
      "learning_rate": 6.839915948368284e-07,
      "loss": 2.4554,
      "num_input_tokens_seen": 54889220,
      "step": 92000
    },
    {
      "epoch": 2.9750418113984303,
      "grad_norm": 9.98161792755127,
      "learning_rate": 4.159698100261589e-07,
      "loss": 2.4243,
      "num_input_tokens_seen": 55190020,
      "step": 92500
    },
    {
      "epoch": 2.9750418113984303,
      "eval_loss": 2.435030460357666,
      "eval_runtime": 2.5128,
      "eval_samples_per_second": 994.908,
      "eval_steps_per_second": 124.563,
      "num_input_tokens_seen": 55190020,
      "step": 92500
    },
    {
      "epoch": 2.991123118487071,
      "grad_norm": 6.586206912994385,
      "learning_rate": 1.479480252154895e-07,
      "loss": 2.44,
      "num_input_tokens_seen": 55490868,
      "step": 93000
    },
    {
      "epoch": 3.0,
      "num_input_tokens_seen": 55653732,
      "step": 93276,
      "total_flos": 1.4738832163602432e+16,
      "train_loss": 1.7073542784139526,
      "train_runtime": 2504.3889,
      "train_samples_per_second": 297.955,
      "train_steps_per_second": 37.245,
      "train_tokens_per_second": 22216.164
    }
  ],
  "logging_steps": 500,
  "max_steps": 93276,
  "num_input_tokens_seen": 55653732,
  "num_train_epochs": 3,
  "save_steps": 2500,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 1.4738832163602432e+16,
  "train_batch_size": 8,
  "trial_name": null,
  "trial_params": null
}