{ "best_metric": 2.435030460357666, "best_model_checkpoint": "zhtw-en/checkpoint-92500", "epoch": 3.0, "eval_steps": 2500, "global_step": 93276, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.016081307088640164, "grad_norm": 8.550692558288574, "learning_rate": 4.919593464556799e-05, "loss": 3.4707, "num_input_tokens_seen": 301408, "step": 500 }, { "epoch": 0.03216261417728033, "grad_norm": 9.999076843261719, "learning_rate": 4.8391869291135987e-05, "loss": 3.3664, "num_input_tokens_seen": 597672, "step": 1000 }, { "epoch": 0.048243921265920496, "grad_norm": 8.839406967163086, "learning_rate": 4.7587803936703975e-05, "loss": 3.3089, "num_input_tokens_seen": 896520, "step": 1500 }, { "epoch": 0.06432522835456066, "grad_norm": 7.105090618133545, "learning_rate": 4.678373858227197e-05, "loss": 3.267, "num_input_tokens_seen": 1194832, "step": 2000 }, { "epoch": 0.08040653544320082, "grad_norm": 7.992733001708984, "learning_rate": 4.597967322783996e-05, "loss": 3.2254, "num_input_tokens_seen": 1493088, "step": 2500 }, { "epoch": 0.08040653544320082, "eval_loss": 2.910461664199829, "eval_runtime": 2.5005, "eval_samples_per_second": 999.788, "eval_steps_per_second": 125.173, "num_input_tokens_seen": 1493088, "step": 2500 }, { "epoch": 0.09648784253184099, "grad_norm": 8.68657112121582, "learning_rate": 4.5175607873407955e-05, "loss": 3.1691, "num_input_tokens_seen": 1793976, "step": 3000 }, { "epoch": 0.11256914962048115, "grad_norm": 7.246800899505615, "learning_rate": 4.4371542518975943e-05, "loss": 3.1685, "num_input_tokens_seen": 2095352, "step": 3500 }, { "epoch": 0.1286504567091213, "grad_norm": 9.030860900878906, "learning_rate": 4.356747716454393e-05, "loss": 3.1333, "num_input_tokens_seen": 2393856, "step": 4000 }, { "epoch": 0.14473176379776148, "grad_norm": 7.463845252990723, "learning_rate": 4.276341181011193e-05, "loss": 3.1295, "num_input_tokens_seen": 2694496, "step": 4500 }, { "epoch": 0.16081307088640165, "grad_norm": 8.482089042663574, "learning_rate": 4.195934645567992e-05, "loss": 3.0946, "num_input_tokens_seen": 2990968, "step": 5000 }, { "epoch": 0.16081307088640165, "eval_loss": 2.830476999282837, "eval_runtime": 2.5201, "eval_samples_per_second": 992.009, "eval_steps_per_second": 124.199, "num_input_tokens_seen": 2990968, "step": 5000 }, { "epoch": 0.17689437797504182, "grad_norm": 8.570518493652344, "learning_rate": 4.115528110124791e-05, "loss": 3.112, "num_input_tokens_seen": 3289488, "step": 5500 }, { "epoch": 0.19297568506368198, "grad_norm": 9.759325981140137, "learning_rate": 4.03512157468159e-05, "loss": 3.0933, "num_input_tokens_seen": 3590264, "step": 6000 }, { "epoch": 0.20905699215232215, "grad_norm": 6.518988609313965, "learning_rate": 3.9547150392383896e-05, "loss": 3.0858, "num_input_tokens_seen": 3885160, "step": 6500 }, { "epoch": 0.2251382992409623, "grad_norm": 6.913475036621094, "learning_rate": 3.8743085037951885e-05, "loss": 3.0543, "num_input_tokens_seen": 4184600, "step": 7000 }, { "epoch": 0.24121960632960246, "grad_norm": 8.485562324523926, "learning_rate": 3.793901968351988e-05, "loss": 3.0473, "num_input_tokens_seen": 4477792, "step": 7500 }, { "epoch": 0.24121960632960246, "eval_loss": 2.773728847503662, "eval_runtime": 2.5738, "eval_samples_per_second": 971.323, "eval_steps_per_second": 121.61, "num_input_tokens_seen": 4477792, "step": 7500 }, { "epoch": 0.2573009134182426, "grad_norm": 7.89262056350708, "learning_rate": 3.713495432908787e-05, "loss": 3.029, "num_input_tokens_seen": 4779520, "step": 8000 }, { "epoch": 0.2733822205068828, "grad_norm": 6.879751205444336, "learning_rate": 3.6330888974655864e-05, "loss": 3.0127, "num_input_tokens_seen": 5078952, "step": 8500 }, { "epoch": 0.28946352759552296, "grad_norm": 8.109273910522461, "learning_rate": 3.552682362022385e-05, "loss": 3.0078, "num_input_tokens_seen": 5376128, "step": 9000 }, { "epoch": 0.3055448346841631, "grad_norm": 8.074146270751953, "learning_rate": 3.472275826579184e-05, "loss": 2.9988, "num_input_tokens_seen": 5671664, "step": 9500 }, { "epoch": 0.3216261417728033, "grad_norm": 6.523529529571533, "learning_rate": 3.391869291135984e-05, "loss": 2.9633, "num_input_tokens_seen": 5967560, "step": 10000 }, { "epoch": 0.3216261417728033, "eval_loss": 2.7306864261627197, "eval_runtime": 2.505, "eval_samples_per_second": 997.986, "eval_steps_per_second": 124.948, "num_input_tokens_seen": 5967560, "step": 10000 }, { "epoch": 0.33770744886144344, "grad_norm": 6.974269866943359, "learning_rate": 3.311462755692783e-05, "loss": 2.9732, "num_input_tokens_seen": 6265312, "step": 10500 }, { "epoch": 0.35378875595008363, "grad_norm": 7.644798278808594, "learning_rate": 3.231056220249582e-05, "loss": 2.9729, "num_input_tokens_seen": 6563632, "step": 11000 }, { "epoch": 0.3698700630387238, "grad_norm": 7.96437406539917, "learning_rate": 3.150649684806381e-05, "loss": 2.9484, "num_input_tokens_seen": 6865528, "step": 11500 }, { "epoch": 0.38595137012736397, "grad_norm": 7.939519882202148, "learning_rate": 3.0702431493631805e-05, "loss": 2.9387, "num_input_tokens_seen": 7165632, "step": 12000 }, { "epoch": 0.4020326772160041, "grad_norm": 7.698306083679199, "learning_rate": 2.9898366139199797e-05, "loss": 2.9355, "num_input_tokens_seen": 7463192, "step": 12500 }, { "epoch": 0.4020326772160041, "eval_loss": 2.684298515319824, "eval_runtime": 2.5214, "eval_samples_per_second": 991.52, "eval_steps_per_second": 124.138, "num_input_tokens_seen": 7463192, "step": 12500 }, { "epoch": 0.4181139843046443, "grad_norm": 7.21583890914917, "learning_rate": 2.9094300784767786e-05, "loss": 2.9418, "num_input_tokens_seen": 7758024, "step": 13000 }, { "epoch": 0.43419529139328444, "grad_norm": 7.767180919647217, "learning_rate": 2.8290235430335778e-05, "loss": 2.923, "num_input_tokens_seen": 8052032, "step": 13500 }, { "epoch": 0.4502765984819246, "grad_norm": 7.057159423828125, "learning_rate": 2.7486170075903773e-05, "loss": 2.9016, "num_input_tokens_seen": 8347768, "step": 14000 }, { "epoch": 0.4663579055705648, "grad_norm": 7.320003032684326, "learning_rate": 2.6682104721471762e-05, "loss": 2.885, "num_input_tokens_seen": 8646192, "step": 14500 }, { "epoch": 0.4824392126592049, "grad_norm": 7.630561828613281, "learning_rate": 2.587803936703975e-05, "loss": 2.9076, "num_input_tokens_seen": 8950264, "step": 15000 }, { "epoch": 0.4824392126592049, "eval_loss": 2.658709764480591, "eval_runtime": 2.6566, "eval_samples_per_second": 941.035, "eval_steps_per_second": 117.818, "num_input_tokens_seen": 8950264, "step": 15000 }, { "epoch": 0.4985205197478451, "grad_norm": 6.413602828979492, "learning_rate": 2.507397401260775e-05, "loss": 2.8818, "num_input_tokens_seen": 9247272, "step": 15500 }, { "epoch": 0.5146018268364853, "grad_norm": 6.833747863769531, "learning_rate": 2.426990865817574e-05, "loss": 2.8998, "num_input_tokens_seen": 9547432, "step": 16000 }, { "epoch": 0.5306831339251254, "grad_norm": 7.930506706237793, "learning_rate": 2.3465843303743727e-05, "loss": 2.8543, "num_input_tokens_seen": 9844072, "step": 16500 }, { "epoch": 0.5467644410137656, "grad_norm": 6.841344356536865, "learning_rate": 2.2661777949311722e-05, "loss": 2.8669, "num_input_tokens_seen": 10142344, "step": 17000 }, { "epoch": 0.5628457481024057, "grad_norm": 6.899343967437744, "learning_rate": 2.185771259487971e-05, "loss": 2.8714, "num_input_tokens_seen": 10443344, "step": 17500 }, { "epoch": 0.5628457481024057, "eval_loss": 2.6303601264953613, "eval_runtime": 2.5354, "eval_samples_per_second": 986.046, "eval_steps_per_second": 123.453, "num_input_tokens_seen": 10443344, "step": 17500 }, { "epoch": 0.5789270551910459, "grad_norm": 6.256689071655273, "learning_rate": 2.1053647240447703e-05, "loss": 2.8418, "num_input_tokens_seen": 10744312, "step": 18000 }, { "epoch": 0.5950083622796861, "grad_norm": 7.627821445465088, "learning_rate": 2.0249581886015695e-05, "loss": 2.8462, "num_input_tokens_seen": 11048208, "step": 18500 }, { "epoch": 0.6110896693683262, "grad_norm": 7.331953525543213, "learning_rate": 1.9445516531583687e-05, "loss": 2.8345, "num_input_tokens_seen": 11347880, "step": 19000 }, { "epoch": 0.6271709764569664, "grad_norm": 6.463207244873047, "learning_rate": 1.864145117715168e-05, "loss": 2.8531, "num_input_tokens_seen": 11650144, "step": 19500 }, { "epoch": 0.6432522835456066, "grad_norm": 7.423746109008789, "learning_rate": 1.783738582271967e-05, "loss": 2.8716, "num_input_tokens_seen": 11951096, "step": 20000 }, { "epoch": 0.6432522835456066, "eval_loss": 2.6024744510650635, "eval_runtime": 2.5182, "eval_samples_per_second": 992.755, "eval_steps_per_second": 124.293, "num_input_tokens_seen": 11951096, "step": 20000 }, { "epoch": 0.6593335906342468, "grad_norm": 7.352589130401611, "learning_rate": 1.7033320468287664e-05, "loss": 2.8243, "num_input_tokens_seen": 12252592, "step": 20500 }, { "epoch": 0.6754148977228869, "grad_norm": 7.22981071472168, "learning_rate": 1.6229255113855656e-05, "loss": 2.8454, "num_input_tokens_seen": 12546792, "step": 21000 }, { "epoch": 0.6914962048115271, "grad_norm": 6.819567680358887, "learning_rate": 1.5425189759423648e-05, "loss": 2.8047, "num_input_tokens_seen": 12838728, "step": 21500 }, { "epoch": 0.7075775119001673, "grad_norm": 8.716426849365234, "learning_rate": 1.4621124404991638e-05, "loss": 2.8144, "num_input_tokens_seen": 13137688, "step": 22000 }, { "epoch": 0.7236588189888075, "grad_norm": 7.324875831604004, "learning_rate": 1.381705905055963e-05, "loss": 2.7989, "num_input_tokens_seen": 13432464, "step": 22500 }, { "epoch": 0.7236588189888075, "eval_loss": 2.5822224617004395, "eval_runtime": 2.5158, "eval_samples_per_second": 993.705, "eval_steps_per_second": 124.412, "num_input_tokens_seen": 13432464, "step": 22500 }, { "epoch": 0.7397401260774475, "grad_norm": 7.962778568267822, "learning_rate": 1.301299369612762e-05, "loss": 2.7653, "num_input_tokens_seen": 13730024, "step": 23000 }, { "epoch": 0.7558214331660877, "grad_norm": 6.807019233703613, "learning_rate": 1.2208928341695614e-05, "loss": 2.7933, "num_input_tokens_seen": 14026400, "step": 23500 }, { "epoch": 0.7719027402547279, "grad_norm": 8.716556549072266, "learning_rate": 1.1404862987263605e-05, "loss": 2.7988, "num_input_tokens_seen": 14326608, "step": 24000 }, { "epoch": 0.787984047343368, "grad_norm": 7.388988018035889, "learning_rate": 1.0600797632831597e-05, "loss": 2.7928, "num_input_tokens_seen": 14623864, "step": 24500 }, { "epoch": 0.8040653544320082, "grad_norm": 7.011099815368652, "learning_rate": 9.796732278399589e-06, "loss": 2.7941, "num_input_tokens_seen": 14919424, "step": 25000 }, { "epoch": 0.8040653544320082, "eval_loss": 2.5630149841308594, "eval_runtime": 2.5601, "eval_samples_per_second": 976.534, "eval_steps_per_second": 122.262, "num_input_tokens_seen": 14919424, "step": 25000 }, { "epoch": 0.8201466615206484, "grad_norm": 6.740393161773682, "learning_rate": 8.992666923967581e-06, "loss": 2.8089, "num_input_tokens_seen": 15216136, "step": 25500 }, { "epoch": 0.8362279686092886, "grad_norm": 7.124479293823242, "learning_rate": 8.188601569535573e-06, "loss": 2.7704, "num_input_tokens_seen": 15515592, "step": 26000 }, { "epoch": 0.8523092756979287, "grad_norm": 7.781102180480957, "learning_rate": 7.384536215103564e-06, "loss": 2.8022, "num_input_tokens_seen": 15818560, "step": 26500 }, { "epoch": 0.8683905827865689, "grad_norm": 6.861135005950928, "learning_rate": 6.580470860671556e-06, "loss": 2.7891, "num_input_tokens_seen": 16114056, "step": 27000 }, { "epoch": 0.8844718898752091, "grad_norm": 7.128973484039307, "learning_rate": 5.776405506239547e-06, "loss": 2.7692, "num_input_tokens_seen": 16415080, "step": 27500 }, { "epoch": 0.8844718898752091, "eval_loss": 2.5496785640716553, "eval_runtime": 2.6422, "eval_samples_per_second": 946.175, "eval_steps_per_second": 118.461, "num_input_tokens_seen": 16415080, "step": 27500 }, { "epoch": 0.9005531969638492, "grad_norm": 8.560084342956543, "learning_rate": 4.9723401518075395e-06, "loss": 2.7627, "num_input_tokens_seen": 16711136, "step": 28000 }, { "epoch": 0.9166345040524894, "grad_norm": 7.5000224113464355, "learning_rate": 4.168274797375531e-06, "loss": 2.7687, "num_input_tokens_seen": 17005880, "step": 28500 }, { "epoch": 0.9327158111411296, "grad_norm": 6.699025630950928, "learning_rate": 3.3642094429435228e-06, "loss": 2.779, "num_input_tokens_seen": 17307064, "step": 29000 }, { "epoch": 0.9487971182297698, "grad_norm": 6.6417131423950195, "learning_rate": 2.560144088511514e-06, "loss": 2.7493, "num_input_tokens_seen": 17602296, "step": 29500 }, { "epoch": 0.9648784253184098, "grad_norm": 6.775792121887207, "learning_rate": 1.756078734079506e-06, "loss": 2.757, "num_input_tokens_seen": 17897832, "step": 30000 }, { "epoch": 0.9648784253184098, "eval_loss": 2.5388031005859375, "eval_runtime": 2.6123, "eval_samples_per_second": 957.018, "eval_steps_per_second": 119.819, "num_input_tokens_seen": 17897832, "step": 30000 }, { "epoch": 0.98095973240705, "grad_norm": 7.619235038757324, "learning_rate": 9.520133796474978e-07, "loss": 2.7433, "num_input_tokens_seen": 18195568, "step": 30500 }, { "epoch": 0.9970410394956902, "grad_norm": 6.682379722595215, "learning_rate": 1.479480252154895e-07, "loss": 2.7266, "num_input_tokens_seen": 18491904, "step": 31000 }, { "epoch": 1.0131223465843304, "grad_norm": 6.5948309898376465, "learning_rate": 3.311462755692783e-05, "loss": 2.6645, "num_input_tokens_seen": 18790628, "step": 31500 }, { "epoch": 1.0292036536729705, "grad_norm": 6.920671463012695, "learning_rate": 3.2846605772117164e-05, "loss": 2.6881, "num_input_tokens_seen": 19090780, "step": 32000 }, { "epoch": 1.0452849607616108, "grad_norm": 6.296219348907471, "learning_rate": 3.257858398730649e-05, "loss": 2.7024, "num_input_tokens_seen": 19384812, "step": 32500 }, { "epoch": 1.0452849607616108, "eval_loss": 2.6005640029907227, "eval_runtime": 2.5084, "eval_samples_per_second": 996.636, "eval_steps_per_second": 124.779, "num_input_tokens_seen": 19384812, "step": 32500 }, { "epoch": 1.061366267850251, "grad_norm": 7.068648815155029, "learning_rate": 3.231056220249582e-05, "loss": 2.6939, "num_input_tokens_seen": 19683060, "step": 33000 }, { "epoch": 1.077447574938891, "grad_norm": 5.753154754638672, "learning_rate": 3.204254041768515e-05, "loss": 2.6977, "num_input_tokens_seen": 19979196, "step": 33500 }, { "epoch": 1.0935288820275313, "grad_norm": 8.155505180358887, "learning_rate": 3.1774518632874485e-05, "loss": 2.7048, "num_input_tokens_seen": 20278524, "step": 34000 }, { "epoch": 1.1096101891161714, "grad_norm": 7.031659126281738, "learning_rate": 3.150649684806381e-05, "loss": 2.7237, "num_input_tokens_seen": 20577572, "step": 34500 }, { "epoch": 1.1256914962048115, "grad_norm": 7.90298318862915, "learning_rate": 3.123847506325314e-05, "loss": 2.7248, "num_input_tokens_seen": 20876844, "step": 35000 }, { "epoch": 1.1256914962048115, "eval_loss": 2.6041972637176514, "eval_runtime": 2.7564, "eval_samples_per_second": 906.969, "eval_steps_per_second": 113.552, "num_input_tokens_seen": 20876844, "step": 35000 }, { "epoch": 1.1417728032934518, "grad_norm": 6.368433475494385, "learning_rate": 3.0970453278442473e-05, "loss": 2.7246, "num_input_tokens_seen": 21179820, "step": 35500 }, { "epoch": 1.1578541103820919, "grad_norm": 7.143220901489258, "learning_rate": 3.0702431493631805e-05, "loss": 2.7211, "num_input_tokens_seen": 21476172, "step": 36000 }, { "epoch": 1.173935417470732, "grad_norm": 7.216341972351074, "learning_rate": 3.0434409708821134e-05, "loss": 2.7088, "num_input_tokens_seen": 21774292, "step": 36500 }, { "epoch": 1.1900167245593722, "grad_norm": 6.958596706390381, "learning_rate": 3.0166387924010465e-05, "loss": 2.7166, "num_input_tokens_seen": 22070908, "step": 37000 }, { "epoch": 1.2060980316480123, "grad_norm": 7.161530494689941, "learning_rate": 2.9898366139199797e-05, "loss": 2.6764, "num_input_tokens_seen": 22372340, "step": 37500 }, { "epoch": 1.2060980316480123, "eval_loss": 2.5923423767089844, "eval_runtime": 2.6849, "eval_samples_per_second": 931.145, "eval_steps_per_second": 116.579, "num_input_tokens_seen": 22372340, "step": 37500 }, { "epoch": 1.2221793387366526, "grad_norm": 7.448612213134766, "learning_rate": 2.963034435438913e-05, "loss": 2.7098, "num_input_tokens_seen": 22672932, "step": 38000 }, { "epoch": 1.2382606458252927, "grad_norm": 8.339189529418945, "learning_rate": 2.9362322569578454e-05, "loss": 2.702, "num_input_tokens_seen": 22971844, "step": 38500 }, { "epoch": 1.2543419529139328, "grad_norm": 6.795124053955078, "learning_rate": 2.9094300784767786e-05, "loss": 2.7007, "num_input_tokens_seen": 23266964, "step": 39000 }, { "epoch": 1.2704232600025729, "grad_norm": 6.3036298751831055, "learning_rate": 2.8826278999957118e-05, "loss": 2.71, "num_input_tokens_seen": 23564068, "step": 39500 }, { "epoch": 1.2865045670912132, "grad_norm": 8.75069808959961, "learning_rate": 2.855825721514645e-05, "loss": 2.6854, "num_input_tokens_seen": 23866100, "step": 40000 }, { "epoch": 1.2865045670912132, "eval_loss": 2.5792863368988037, "eval_runtime": 2.5776, "eval_samples_per_second": 969.878, "eval_steps_per_second": 121.429, "num_input_tokens_seen": 23866100, "step": 40000 }, { "epoch": 1.3025858741798533, "grad_norm": 6.966170310974121, "learning_rate": 2.8290235430335778e-05, "loss": 2.697, "num_input_tokens_seen": 24162356, "step": 40500 }, { "epoch": 1.3186671812684936, "grad_norm": 7.854964733123779, "learning_rate": 2.802221364552511e-05, "loss": 2.6954, "num_input_tokens_seen": 24458980, "step": 41000 }, { "epoch": 1.3347484883571337, "grad_norm": 7.1461944580078125, "learning_rate": 2.775419186071444e-05, "loss": 2.6839, "num_input_tokens_seen": 24757828, "step": 41500 }, { "epoch": 1.3508297954457738, "grad_norm": 8.25295639038086, "learning_rate": 2.7486170075903773e-05, "loss": 2.7035, "num_input_tokens_seen": 25052236, "step": 42000 }, { "epoch": 1.366911102534414, "grad_norm": 6.336223602294922, "learning_rate": 2.7218148291093105e-05, "loss": 2.683, "num_input_tokens_seen": 25348084, "step": 42500 }, { "epoch": 1.366911102534414, "eval_loss": 2.5722219944000244, "eval_runtime": 2.7384, "eval_samples_per_second": 912.958, "eval_steps_per_second": 114.302, "num_input_tokens_seen": 25348084, "step": 42500 }, { "epoch": 1.3829924096230541, "grad_norm": 9.477555274963379, "learning_rate": 2.695012650628243e-05, "loss": 2.6877, "num_input_tokens_seen": 25642372, "step": 43000 }, { "epoch": 1.3990737167116944, "grad_norm": 8.233431816101074, "learning_rate": 2.6682104721471762e-05, "loss": 2.6927, "num_input_tokens_seen": 25939652, "step": 43500 }, { "epoch": 1.4151550238003345, "grad_norm": 5.860446929931641, "learning_rate": 2.6414082936661094e-05, "loss": 2.6819, "num_input_tokens_seen": 26248940, "step": 44000 }, { "epoch": 1.4312363308889746, "grad_norm": 6.748124599456787, "learning_rate": 2.6146061151850426e-05, "loss": 2.6893, "num_input_tokens_seen": 26552860, "step": 44500 }, { "epoch": 1.4473176379776147, "grad_norm": 6.038182258605957, "learning_rate": 2.587803936703975e-05, "loss": 2.6871, "num_input_tokens_seen": 26854100, "step": 45000 }, { "epoch": 1.4473176379776147, "eval_loss": 2.5538456439971924, "eval_runtime": 2.6078, "eval_samples_per_second": 958.67, "eval_steps_per_second": 120.026, "num_input_tokens_seen": 26854100, "step": 45000 }, { "epoch": 1.463398945066255, "grad_norm": 7.815784454345703, "learning_rate": 2.5610017582229086e-05, "loss": 2.6709, "num_input_tokens_seen": 27148148, "step": 45500 }, { "epoch": 1.479480252154895, "grad_norm": 7.8851094245910645, "learning_rate": 2.5341995797418418e-05, "loss": 2.6698, "num_input_tokens_seen": 27445020, "step": 46000 }, { "epoch": 1.4955615592435354, "grad_norm": 7.389246940612793, "learning_rate": 2.507397401260775e-05, "loss": 2.6787, "num_input_tokens_seen": 27742908, "step": 46500 }, { "epoch": 1.5116428663321755, "grad_norm": 7.621913909912109, "learning_rate": 2.4805952227797078e-05, "loss": 2.6713, "num_input_tokens_seen": 28037284, "step": 47000 }, { "epoch": 1.5277241734208156, "grad_norm": 7.889066219329834, "learning_rate": 2.4537930442986407e-05, "loss": 2.6551, "num_input_tokens_seen": 28332612, "step": 47500 }, { "epoch": 1.5277241734208156, "eval_loss": 2.5442593097686768, "eval_runtime": 2.6341, "eval_samples_per_second": 949.086, "eval_steps_per_second": 118.826, "num_input_tokens_seen": 28332612, "step": 47500 }, { "epoch": 1.5438054805094557, "grad_norm": 7.912906646728516, "learning_rate": 2.426990865817574e-05, "loss": 2.6881, "num_input_tokens_seen": 28630948, "step": 48000 }, { "epoch": 1.559886787598096, "grad_norm": 6.370878219604492, "learning_rate": 2.4001886873365067e-05, "loss": 2.6424, "num_input_tokens_seen": 28928732, "step": 48500 }, { "epoch": 1.5759680946867363, "grad_norm": 7.0892653465271, "learning_rate": 2.37338650885544e-05, "loss": 2.6626, "num_input_tokens_seen": 29224436, "step": 49000 }, { "epoch": 1.5920494017753763, "grad_norm": 6.357864856719971, "learning_rate": 2.3465843303743727e-05, "loss": 2.6546, "num_input_tokens_seen": 29520148, "step": 49500 }, { "epoch": 1.6081307088640164, "grad_norm": 8.4866943359375, "learning_rate": 2.319782151893306e-05, "loss": 2.661, "num_input_tokens_seen": 29822156, "step": 50000 }, { "epoch": 1.6081307088640164, "eval_loss": 2.527804374694824, "eval_runtime": 2.6771, "eval_samples_per_second": 933.855, "eval_steps_per_second": 116.919, "num_input_tokens_seen": 29822156, "step": 50000 }, { "epoch": 1.6242120159526565, "grad_norm": 6.843733787536621, "learning_rate": 2.292979973412239e-05, "loss": 2.6521, "num_input_tokens_seen": 30122052, "step": 50500 }, { "epoch": 1.6402933230412968, "grad_norm": 6.88835334777832, "learning_rate": 2.2661777949311722e-05, "loss": 2.6614, "num_input_tokens_seen": 30422196, "step": 51000 }, { "epoch": 1.656374630129937, "grad_norm": 5.855214595794678, "learning_rate": 2.239375616450105e-05, "loss": 2.6454, "num_input_tokens_seen": 30722660, "step": 51500 }, { "epoch": 1.6724559372185772, "grad_norm": 6.58035135269165, "learning_rate": 2.2125734379690383e-05, "loss": 2.6524, "num_input_tokens_seen": 31018220, "step": 52000 }, { "epoch": 1.6885372443072173, "grad_norm": 6.767495155334473, "learning_rate": 2.185771259487971e-05, "loss": 2.6497, "num_input_tokens_seen": 31319476, "step": 52500 }, { "epoch": 1.6885372443072173, "eval_loss": 2.526638984680176, "eval_runtime": 2.5213, "eval_samples_per_second": 991.545, "eval_steps_per_second": 124.141, "num_input_tokens_seen": 31319476, "step": 52500 }, { "epoch": 1.7046185513958574, "grad_norm": 7.022729873657227, "learning_rate": 2.1589690810069043e-05, "loss": 2.6437, "num_input_tokens_seen": 31621308, "step": 53000 }, { "epoch": 1.7206998584844975, "grad_norm": 6.241069793701172, "learning_rate": 2.132166902525837e-05, "loss": 2.6447, "num_input_tokens_seen": 31917460, "step": 53500 }, { "epoch": 1.7367811655731378, "grad_norm": 7.7204084396362305, "learning_rate": 2.1053647240447703e-05, "loss": 2.6448, "num_input_tokens_seen": 32217596, "step": 54000 }, { "epoch": 1.752862472661778, "grad_norm": 6.703210830688477, "learning_rate": 2.0785625455637035e-05, "loss": 2.6366, "num_input_tokens_seen": 32513884, "step": 54500 }, { "epoch": 1.7689437797504182, "grad_norm": 6.371466159820557, "learning_rate": 2.0517603670826367e-05, "loss": 2.6281, "num_input_tokens_seen": 32813220, "step": 55000 }, { "epoch": 1.7689437797504182, "eval_loss": 2.5115973949432373, "eval_runtime": 2.5216, "eval_samples_per_second": 991.443, "eval_steps_per_second": 124.129, "num_input_tokens_seen": 32813220, "step": 55000 }, { "epoch": 1.7850250868390583, "grad_norm": 7.277946949005127, "learning_rate": 2.0249581886015695e-05, "loss": 2.6536, "num_input_tokens_seen": 33110188, "step": 55500 }, { "epoch": 1.8011063939276983, "grad_norm": 7.93104887008667, "learning_rate": 1.9981560101205027e-05, "loss": 2.5981, "num_input_tokens_seen": 33405980, "step": 56000 }, { "epoch": 1.8171877010163386, "grad_norm": 7.782486438751221, "learning_rate": 1.971353831639436e-05, "loss": 2.635, "num_input_tokens_seen": 33700596, "step": 56500 }, { "epoch": 1.8332690081049787, "grad_norm": 8.59358024597168, "learning_rate": 1.9445516531583687e-05, "loss": 2.6269, "num_input_tokens_seen": 33997724, "step": 57000 }, { "epoch": 1.849350315193619, "grad_norm": 6.669950485229492, "learning_rate": 1.917749474677302e-05, "loss": 2.6067, "num_input_tokens_seen": 34298052, "step": 57500 }, { "epoch": 1.849350315193619, "eval_loss": 2.5047078132629395, "eval_runtime": 2.5169, "eval_samples_per_second": 993.288, "eval_steps_per_second": 124.36, "num_input_tokens_seen": 34298052, "step": 57500 }, { "epoch": 1.8654316222822591, "grad_norm": 6.265903949737549, "learning_rate": 1.8909472961962348e-05, "loss": 2.5966, "num_input_tokens_seen": 34593980, "step": 58000 }, { "epoch": 1.8815129293708992, "grad_norm": 7.943974018096924, "learning_rate": 1.864145117715168e-05, "loss": 2.6303, "num_input_tokens_seen": 34894428, "step": 58500 }, { "epoch": 1.8975942364595393, "grad_norm": 8.290629386901855, "learning_rate": 1.837342939234101e-05, "loss": 2.6303, "num_input_tokens_seen": 35193236, "step": 59000 }, { "epoch": 1.9136755435481796, "grad_norm": 7.974947929382324, "learning_rate": 1.8105407607530343e-05, "loss": 2.6272, "num_input_tokens_seen": 35486796, "step": 59500 }, { "epoch": 1.92975685063682, "grad_norm": 5.827637195587158, "learning_rate": 1.783738582271967e-05, "loss": 2.6112, "num_input_tokens_seen": 35783604, "step": 60000 }, { "epoch": 1.92975685063682, "eval_loss": 2.4935405254364014, "eval_runtime": 2.5107, "eval_samples_per_second": 995.741, "eval_steps_per_second": 124.667, "num_input_tokens_seen": 35783604, "step": 60000 }, { "epoch": 1.94583815772546, "grad_norm": 6.535378456115723, "learning_rate": 1.7569364037909003e-05, "loss": 2.6135, "num_input_tokens_seen": 36086620, "step": 60500 }, { "epoch": 1.9619194648141, "grad_norm": 6.398725986480713, "learning_rate": 1.7301342253098332e-05, "loss": 2.602, "num_input_tokens_seen": 36386140, "step": 61000 }, { "epoch": 1.9780007719027402, "grad_norm": 6.332113265991211, "learning_rate": 1.7033320468287664e-05, "loss": 2.6258, "num_input_tokens_seen": 36684308, "step": 61500 }, { "epoch": 1.9940820789913805, "grad_norm": 7.8002753257751465, "learning_rate": 1.6765298683476992e-05, "loss": 2.6226, "num_input_tokens_seen": 36984724, "step": 62000 }, { "epoch": 2.0101633860800208, "grad_norm": 6.9457011222839355, "learning_rate": 1.6497276898666324e-05, "loss": 2.5207, "num_input_tokens_seen": 37281092, "step": 62500 }, { "epoch": 2.0101633860800208, "eval_loss": 2.4945950508117676, "eval_runtime": 2.5094, "eval_samples_per_second": 996.26, "eval_steps_per_second": 124.732, "num_input_tokens_seen": 37281092, "step": 62500 }, { "epoch": 2.026244693168661, "grad_norm": 7.541498184204102, "learning_rate": 1.6229255113855656e-05, "loss": 2.4728, "num_input_tokens_seen": 37582300, "step": 63000 }, { "epoch": 2.042326000257301, "grad_norm": 6.7798027992248535, "learning_rate": 1.5961233329044987e-05, "loss": 2.4539, "num_input_tokens_seen": 37880828, "step": 63500 }, { "epoch": 2.058407307345941, "grad_norm": 7.033351898193359, "learning_rate": 1.5693211544234316e-05, "loss": 2.4467, "num_input_tokens_seen": 38181276, "step": 64000 }, { "epoch": 2.074488614434581, "grad_norm": 6.487890720367432, "learning_rate": 1.5425189759423648e-05, "loss": 2.4764, "num_input_tokens_seen": 38473348, "step": 64500 }, { "epoch": 2.0905699215232216, "grad_norm": 6.955127716064453, "learning_rate": 1.5157167974612976e-05, "loss": 2.4799, "num_input_tokens_seen": 38768588, "step": 65000 }, { "epoch": 2.0905699215232216, "eval_loss": 2.491555690765381, "eval_runtime": 2.5076, "eval_samples_per_second": 996.967, "eval_steps_per_second": 124.82, "num_input_tokens_seen": 38768588, "step": 65000 }, { "epoch": 2.1066512286118617, "grad_norm": 6.78762674331665, "learning_rate": 1.4889146189802308e-05, "loss": 2.4726, "num_input_tokens_seen": 39067460, "step": 65500 }, { "epoch": 2.122732535700502, "grad_norm": 7.199331283569336, "learning_rate": 1.4621124404991638e-05, "loss": 2.4562, "num_input_tokens_seen": 39360244, "step": 66000 }, { "epoch": 2.138813842789142, "grad_norm": 7.353775501251221, "learning_rate": 1.435310262018097e-05, "loss": 2.4629, "num_input_tokens_seen": 39660020, "step": 66500 }, { "epoch": 2.154895149877782, "grad_norm": 6.827337265014648, "learning_rate": 1.4085080835370298e-05, "loss": 2.4817, "num_input_tokens_seen": 39960476, "step": 67000 }, { "epoch": 2.170976456966422, "grad_norm": 6.532020092010498, "learning_rate": 1.381705905055963e-05, "loss": 2.4727, "num_input_tokens_seen": 40252972, "step": 67500 }, { "epoch": 2.170976456966422, "eval_loss": 2.4865615367889404, "eval_runtime": 2.5372, "eval_samples_per_second": 985.335, "eval_steps_per_second": 123.364, "num_input_tokens_seen": 40252972, "step": 67500 }, { "epoch": 2.1870577640550626, "grad_norm": 6.601158142089844, "learning_rate": 1.354903726574896e-05, "loss": 2.4666, "num_input_tokens_seen": 40553732, "step": 68000 }, { "epoch": 2.2031390711437027, "grad_norm": 7.200645446777344, "learning_rate": 1.3281015480938292e-05, "loss": 2.4657, "num_input_tokens_seen": 40851196, "step": 68500 }, { "epoch": 2.2192203782323427, "grad_norm": 8.067240715026855, "learning_rate": 1.301299369612762e-05, "loss": 2.4801, "num_input_tokens_seen": 41149276, "step": 69000 }, { "epoch": 2.235301685320983, "grad_norm": 7.724194526672363, "learning_rate": 1.2744971911316952e-05, "loss": 2.4766, "num_input_tokens_seen": 41448540, "step": 69500 }, { "epoch": 2.251382992409623, "grad_norm": 6.999200344085693, "learning_rate": 1.2476950126506282e-05, "loss": 2.4719, "num_input_tokens_seen": 41746300, "step": 70000 }, { "epoch": 2.251382992409623, "eval_loss": 2.476020097732544, "eval_runtime": 2.5166, "eval_samples_per_second": 993.407, "eval_steps_per_second": 124.375, "num_input_tokens_seen": 41746300, "step": 70000 }, { "epoch": 2.267464299498263, "grad_norm": 6.666884899139404, "learning_rate": 1.2208928341695614e-05, "loss": 2.4771, "num_input_tokens_seen": 42042532, "step": 70500 }, { "epoch": 2.2835456065869035, "grad_norm": 8.354509353637695, "learning_rate": 1.1940906556884944e-05, "loss": 2.4679, "num_input_tokens_seen": 42341628, "step": 71000 }, { "epoch": 2.2996269136755436, "grad_norm": 8.39284610748291, "learning_rate": 1.1672884772074275e-05, "loss": 2.4597, "num_input_tokens_seen": 42642948, "step": 71500 }, { "epoch": 2.3157082207641837, "grad_norm": 7.233700275421143, "learning_rate": 1.1404862987263605e-05, "loss": 2.4592, "num_input_tokens_seen": 42941820, "step": 72000 }, { "epoch": 2.331789527852824, "grad_norm": 7.843503475189209, "learning_rate": 1.1136841202452935e-05, "loss": 2.4738, "num_input_tokens_seen": 43241188, "step": 72500 }, { "epoch": 2.331789527852824, "eval_loss": 2.47127103805542, "eval_runtime": 2.5277, "eval_samples_per_second": 989.05, "eval_steps_per_second": 123.829, "num_input_tokens_seen": 43241188, "step": 72500 }, { "epoch": 2.347870834941464, "grad_norm": 6.37482213973999, "learning_rate": 1.0868819417642267e-05, "loss": 2.4576, "num_input_tokens_seen": 43535900, "step": 73000 }, { "epoch": 2.3639521420301044, "grad_norm": 6.642532825469971, "learning_rate": 1.0600797632831597e-05, "loss": 2.467, "num_input_tokens_seen": 43833516, "step": 73500 }, { "epoch": 2.3800334491187445, "grad_norm": 6.606197357177734, "learning_rate": 1.0332775848020927e-05, "loss": 2.4752, "num_input_tokens_seen": 44134084, "step": 74000 }, { "epoch": 2.3961147562073846, "grad_norm": 6.338978290557861, "learning_rate": 1.0064754063210257e-05, "loss": 2.4473, "num_input_tokens_seen": 44432540, "step": 74500 }, { "epoch": 2.4121960632960247, "grad_norm": 7.172792434692383, "learning_rate": 9.796732278399589e-06, "loss": 2.4629, "num_input_tokens_seen": 44730244, "step": 75000 }, { "epoch": 2.4121960632960247, "eval_loss": 2.4629955291748047, "eval_runtime": 2.521, "eval_samples_per_second": 991.68, "eval_steps_per_second": 124.158, "num_input_tokens_seen": 44730244, "step": 75000 }, { "epoch": 2.4282773703846647, "grad_norm": 6.65930700302124, "learning_rate": 9.528710493588919e-06, "loss": 2.4512, "num_input_tokens_seen": 45031884, "step": 75500 }, { "epoch": 2.4443586774733053, "grad_norm": 7.209745407104492, "learning_rate": 9.260688708778249e-06, "loss": 2.4557, "num_input_tokens_seen": 45334924, "step": 76000 }, { "epoch": 2.4604399845619453, "grad_norm": 6.847073078155518, "learning_rate": 8.992666923967581e-06, "loss": 2.4512, "num_input_tokens_seen": 45633532, "step": 76500 }, { "epoch": 2.4765212916505854, "grad_norm": 7.705162525177002, "learning_rate": 8.724645139156911e-06, "loss": 2.4568, "num_input_tokens_seen": 45933804, "step": 77000 }, { "epoch": 2.4926025987392255, "grad_norm": 7.5681962966918945, "learning_rate": 8.456623354346243e-06, "loss": 2.4524, "num_input_tokens_seen": 46231060, "step": 77500 }, { "epoch": 2.4926025987392255, "eval_loss": 2.457481861114502, "eval_runtime": 2.5921, "eval_samples_per_second": 964.467, "eval_steps_per_second": 120.751, "num_input_tokens_seen": 46231060, "step": 77500 }, { "epoch": 2.5086839058278656, "grad_norm": 6.8857269287109375, "learning_rate": 8.188601569535573e-06, "loss": 2.4622, "num_input_tokens_seen": 46525772, "step": 78000 }, { "epoch": 2.524765212916506, "grad_norm": 6.347681522369385, "learning_rate": 7.920579784724903e-06, "loss": 2.4528, "num_input_tokens_seen": 46822532, "step": 78500 }, { "epoch": 2.5408465200051458, "grad_norm": 6.935575008392334, "learning_rate": 7.652557999914233e-06, "loss": 2.4414, "num_input_tokens_seen": 47122964, "step": 79000 }, { "epoch": 2.5569278270937863, "grad_norm": 6.603360652923584, "learning_rate": 7.384536215103564e-06, "loss": 2.4655, "num_input_tokens_seen": 47423300, "step": 79500 }, { "epoch": 2.5730091341824264, "grad_norm": 7.182071208953857, "learning_rate": 7.116514430292895e-06, "loss": 2.435, "num_input_tokens_seen": 47718964, "step": 80000 }, { "epoch": 2.5730091341824264, "eval_loss": 2.455320358276367, "eval_runtime": 2.5065, "eval_samples_per_second": 997.4, "eval_steps_per_second": 124.874, "num_input_tokens_seen": 47718964, "step": 80000 }, { "epoch": 2.5890904412710665, "grad_norm": 7.3647260665893555, "learning_rate": 6.848492645482225e-06, "loss": 2.4356, "num_input_tokens_seen": 48015996, "step": 80500 }, { "epoch": 2.6051717483597066, "grad_norm": 7.950341701507568, "learning_rate": 6.580470860671556e-06, "loss": 2.4453, "num_input_tokens_seen": 48316420, "step": 81000 }, { "epoch": 2.6212530554483466, "grad_norm": 6.016787052154541, "learning_rate": 6.312449075860886e-06, "loss": 2.45, "num_input_tokens_seen": 48611452, "step": 81500 }, { "epoch": 2.637334362536987, "grad_norm": 7.281980514526367, "learning_rate": 6.044427291050217e-06, "loss": 2.4687, "num_input_tokens_seen": 48913668, "step": 82000 }, { "epoch": 2.6534156696256272, "grad_norm": 6.644787311553955, "learning_rate": 5.776405506239547e-06, "loss": 2.4621, "num_input_tokens_seen": 49209724, "step": 82500 }, { "epoch": 2.6534156696256272, "eval_loss": 2.4475488662719727, "eval_runtime": 2.517, "eval_samples_per_second": 993.239, "eval_steps_per_second": 124.354, "num_input_tokens_seen": 49209724, "step": 82500 }, { "epoch": 2.6694969767142673, "grad_norm": 6.181220054626465, "learning_rate": 5.508383721428878e-06, "loss": 2.4343, "num_input_tokens_seen": 49505772, "step": 83000 }, { "epoch": 2.6855782838029074, "grad_norm": 6.418393135070801, "learning_rate": 5.2403619366182085e-06, "loss": 2.4329, "num_input_tokens_seen": 49809956, "step": 83500 }, { "epoch": 2.7016595908915475, "grad_norm": 6.279716491699219, "learning_rate": 4.9723401518075395e-06, "loss": 2.4481, "num_input_tokens_seen": 50112060, "step": 84000 }, { "epoch": 2.717740897980188, "grad_norm": 6.502873420715332, "learning_rate": 4.70431836699687e-06, "loss": 2.4464, "num_input_tokens_seen": 50414356, "step": 84500 }, { "epoch": 2.733822205068828, "grad_norm": 6.15990686416626, "learning_rate": 4.4362965821862e-06, "loss": 2.4492, "num_input_tokens_seen": 50712980, "step": 85000 }, { "epoch": 2.733822205068828, "eval_loss": 2.4440150260925293, "eval_runtime": 2.529, "eval_samples_per_second": 988.544, "eval_steps_per_second": 123.766, "num_input_tokens_seen": 50712980, "step": 85000 }, { "epoch": 2.749903512157468, "grad_norm": 6.876352310180664, "learning_rate": 4.168274797375531e-06, "loss": 2.4514, "num_input_tokens_seen": 51012460, "step": 85500 }, { "epoch": 2.7659848192461083, "grad_norm": 7.305426597595215, "learning_rate": 3.900253012564861e-06, "loss": 2.4317, "num_input_tokens_seen": 51308524, "step": 86000 }, { "epoch": 2.7820661263347484, "grad_norm": 6.460892677307129, "learning_rate": 3.632231227754192e-06, "loss": 2.4559, "num_input_tokens_seen": 51610700, "step": 86500 }, { "epoch": 2.798147433423389, "grad_norm": 8.062651634216309, "learning_rate": 3.3642094429435228e-06, "loss": 2.4535, "num_input_tokens_seen": 51910236, "step": 87000 }, { "epoch": 2.814228740512029, "grad_norm": 7.140311241149902, "learning_rate": 3.0961876581328533e-06, "loss": 2.4536, "num_input_tokens_seen": 52204380, "step": 87500 }, { "epoch": 2.814228740512029, "eval_loss": 2.4393906593322754, "eval_runtime": 2.5312, "eval_samples_per_second": 987.685, "eval_steps_per_second": 123.658, "num_input_tokens_seen": 52204380, "step": 87500 }, { "epoch": 2.830310047600669, "grad_norm": 6.569787502288818, "learning_rate": 2.8281658733221834e-06, "loss": 2.4379, "num_input_tokens_seen": 52504668, "step": 88000 }, { "epoch": 2.846391354689309, "grad_norm": 7.735711097717285, "learning_rate": 2.560144088511514e-06, "loss": 2.4239, "num_input_tokens_seen": 52798740, "step": 88500 }, { "epoch": 2.8624726617779492, "grad_norm": 7.504124641418457, "learning_rate": 2.292122303700845e-06, "loss": 2.4427, "num_input_tokens_seen": 53097716, "step": 89000 }, { "epoch": 2.8785539688665893, "grad_norm": 6.647756099700928, "learning_rate": 2.0241005188901755e-06, "loss": 2.4682, "num_input_tokens_seen": 53397564, "step": 89500 }, { "epoch": 2.8946352759552294, "grad_norm": 6.640815734863281, "learning_rate": 1.756078734079506e-06, "loss": 2.4148, "num_input_tokens_seen": 53695620, "step": 90000 }, { "epoch": 2.8946352759552294, "eval_loss": 2.43597674369812, "eval_runtime": 2.5128, "eval_samples_per_second": 994.891, "eval_steps_per_second": 124.56, "num_input_tokens_seen": 53695620, "step": 90000 }, { "epoch": 2.91071658304387, "grad_norm": 7.346447467803955, "learning_rate": 1.4880569492688366e-06, "loss": 2.4352, "num_input_tokens_seen": 53991180, "step": 90500 }, { "epoch": 2.92679789013251, "grad_norm": 6.777767658233643, "learning_rate": 1.2200351644581672e-06, "loss": 2.4664, "num_input_tokens_seen": 54288348, "step": 91000 }, { "epoch": 2.94287919722115, "grad_norm": 6.908254623413086, "learning_rate": 9.520133796474978e-07, "loss": 2.4474, "num_input_tokens_seen": 54590740, "step": 91500 }, { "epoch": 2.95896050430979, "grad_norm": 7.04544734954834, "learning_rate": 6.839915948368284e-07, "loss": 2.4554, "num_input_tokens_seen": 54889220, "step": 92000 }, { "epoch": 2.9750418113984303, "grad_norm": 9.98161792755127, "learning_rate": 4.159698100261589e-07, "loss": 2.4243, "num_input_tokens_seen": 55190020, "step": 92500 }, { "epoch": 2.9750418113984303, "eval_loss": 2.435030460357666, "eval_runtime": 2.5128, "eval_samples_per_second": 994.908, "eval_steps_per_second": 124.563, "num_input_tokens_seen": 55190020, "step": 92500 }, { "epoch": 2.991123118487071, "grad_norm": 6.586206912994385, "learning_rate": 1.479480252154895e-07, "loss": 2.44, "num_input_tokens_seen": 55490868, "step": 93000 }, { "epoch": 3.0, "num_input_tokens_seen": 55653732, "step": 93276, "total_flos": 1.4738832163602432e+16, "train_loss": 1.7073542784139526, "train_runtime": 2504.3889, "train_samples_per_second": 297.955, "train_steps_per_second": 37.245, "train_tokens_per_second": 22216.164 } ], "logging_steps": 500, "max_steps": 93276, "num_input_tokens_seen": 55653732, "num_train_epochs": 3, "save_steps": 2500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.4738832163602432e+16, "train_batch_size": 8, "trial_name": null, "trial_params": null }