|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 750, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 8.695652173913044e-06, |
|
"loss": 1.5895, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 1.739130434782609e-05, |
|
"loss": 1.3485, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"learning_rate": 2.608695652173913e-05, |
|
"loss": 1.479, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 3.478260869565218e-05, |
|
"loss": 1.4631, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 4.347826086956522e-05, |
|
"loss": 1.7193, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 5.217391304347826e-05, |
|
"loss": 1.3891, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.086956521739131e-05, |
|
"loss": 1.4848, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 6.956521739130436e-05, |
|
"loss": 1.6442, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 7.82608695652174e-05, |
|
"loss": 1.5395, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 8.695652173913044e-05, |
|
"loss": 1.4763, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"learning_rate": 9.565217391304348e-05, |
|
"loss": 1.3408, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00010434782608695653, |
|
"loss": 1.4619, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00011304347826086956, |
|
"loss": 1.5815, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00012173913043478263, |
|
"loss": 1.2742, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00013043478260869567, |
|
"loss": 1.4163, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0001391304347826087, |
|
"loss": 1.3146, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.00014782608695652173, |
|
"loss": 1.301, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"learning_rate": 0.0001565217391304348, |
|
"loss": 1.3672, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00016521739130434784, |
|
"loss": 1.5228, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00017391304347826088, |
|
"loss": 1.2977, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00018260869565217392, |
|
"loss": 1.6576, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00019130434782608697, |
|
"loss": 1.2855, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0002, |
|
"loss": 1.3101, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.0001999990663152786, |
|
"loss": 1.335, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00019999626527854967, |
|
"loss": 1.2408, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"learning_rate": 0.00019999159694211892, |
|
"loss": 1.1966, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0001999850613931615, |
|
"loss": 1.2114, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.0001999766587537202, |
|
"loss": 1.4837, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00019996638918070336, |
|
"loss": 1.3792, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00019995425286588184, |
|
"loss": 1.4196, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00019994025003588548, |
|
"loss": 1.5315, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00019992438095219886, |
|
"loss": 1.276, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"learning_rate": 0.00019990664591115634, |
|
"loss": 1.3932, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00019988704524393676, |
|
"loss": 1.4347, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00019986557931655688, |
|
"loss": 1.3393, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00019984224852986494, |
|
"loss": 1.3243, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00019981705331953293, |
|
"loss": 1.5642, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00019978999415604847, |
|
"loss": 1.4884, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.0001997610715447061, |
|
"loss": 1.4377, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00019973028602559786, |
|
"loss": 1.2243, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 0.00019969763817360314, |
|
"loss": 1.3491, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00019966312859837787, |
|
"loss": 1.452, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0001996267579443434, |
|
"loss": 1.395, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0001995885268906742, |
|
"loss": 1.4748, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0001995484361512853, |
|
"loss": 1.5524, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.0001995064864748188, |
|
"loss": 1.4089, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00019946267864463026, |
|
"loss": 1.4285, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"learning_rate": 0.00019941701347877367, |
|
"loss": 1.4041, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0001993694918299864, |
|
"loss": 1.4757, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00019932011458567315, |
|
"loss": 1.3526, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00019926888266788955, |
|
"loss": 1.3727, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00019921579703332474, |
|
"loss": 1.3149, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.0001991608586732837, |
|
"loss": 1.3851, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00019910406861366848, |
|
"loss": 1.3576, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00019904542791495937, |
|
"loss": 1.5287, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"learning_rate": 0.00019898493767219483, |
|
"loss": 1.4147, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0001989225990149512, |
|
"loss": 1.3346, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0001988584131073215, |
|
"loss": 1.301, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00019879238114789373, |
|
"loss": 1.2017, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00019872450436972855, |
|
"loss": 1.4269, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00019865478404033617, |
|
"loss": 1.2938, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.0001985832214616527, |
|
"loss": 1.2991, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"learning_rate": 0.00019850981797001592, |
|
"loss": 1.2597, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00019843457493614014, |
|
"loss": 1.3653, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00019835749376509082, |
|
"loss": 1.5105, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00019827857589625817, |
|
"loss": 1.4437, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00019819782280333038, |
|
"loss": 1.4416, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00019811523599426602, |
|
"loss": 1.3771, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0001980308170112659, |
|
"loss": 1.2417, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.0001979445674307444, |
|
"loss": 1.4953, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"learning_rate": 0.00019785648886329972, |
|
"loss": 1.2115, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0001977665829536842, |
|
"loss": 1.3111, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019767485138077326, |
|
"loss": 1.5339, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.0001975812958575343, |
|
"loss": 1.388, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019748591813099456, |
|
"loss": 1.4833, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019738871998220857, |
|
"loss": 1.2342, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019728970322622485, |
|
"loss": 1.2687, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 0.00019718886971205205, |
|
"loss": 1.2532, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0001970862213226244, |
|
"loss": 1.2687, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019698175997476657, |
|
"loss": 1.7166, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.0001968754876191578, |
|
"loss": 1.2961, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019676740624029567, |
|
"loss": 1.2892, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019665751785645874, |
|
"loss": 1.3508, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019654582451966913, |
|
"loss": 1.6626, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019643232831565414, |
|
"loss": 1.2956, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"learning_rate": 0.00019631703136380714, |
|
"loss": 1.316, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019619993581714823, |
|
"loss": 1.4281, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019608104386228378, |
|
"loss": 1.3362, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019596035771936592, |
|
"loss": 1.2399, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019583787964205074, |
|
"loss": 1.4614, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019571361191745646, |
|
"loss": 1.5927, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019558755686612057, |
|
"loss": 1.3433, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"learning_rate": 0.00019545971684195663, |
|
"loss": 1.2615, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019533009423221013, |
|
"loss": 1.4105, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0001951986914574141, |
|
"loss": 1.2109, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019506551097134384, |
|
"loss": 1.4404, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.000194930555260971, |
|
"loss": 1.2278, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0001947938268464173, |
|
"loss": 1.2697, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019465532828090735, |
|
"loss": 1.3165, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.00019451506215072107, |
|
"loss": 1.4462, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"learning_rate": 0.0001943730310751453, |
|
"loss": 1.1843, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019422923770642494, |
|
"loss": 1.2396, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019408368472971345, |
|
"loss": 1.4353, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019393637486302256, |
|
"loss": 1.3102, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.0001937873108571718, |
|
"loss": 1.4087, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019363649549573682, |
|
"loss": 1.6766, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019348393159499759, |
|
"loss": 1.3699, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 0.00019332962200388577, |
|
"loss": 1.2109, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019317356960393156, |
|
"loss": 1.2772, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019301577730920975, |
|
"loss": 1.4304, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019285624806628543, |
|
"loss": 1.3526, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019269498485415895, |
|
"loss": 1.4856, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019253199068421028, |
|
"loss": 1.3627, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019236726860014268, |
|
"loss": 1.221, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019220082167792607, |
|
"loss": 1.1005, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"learning_rate": 0.00019203265302573938, |
|
"loss": 1.3417, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019186276578391265, |
|
"loss": 1.5484, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019169116312486834, |
|
"loss": 1.1668, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019151784825306205, |
|
"loss": 1.207, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019134282440492272, |
|
"loss": 1.3144, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.0001911660948487922, |
|
"loss": 1.0671, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019098766288486426, |
|
"loss": 1.3509, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"learning_rate": 0.00019080753184512282, |
|
"loss": 1.2516, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019062570509327992, |
|
"loss": 1.487, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019044218602471275, |
|
"loss": 1.3777, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019025697806640033, |
|
"loss": 1.1699, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00019007008467685944, |
|
"loss": 1.1989, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00018988150934608013, |
|
"loss": 1.3678, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.00018969125559546054, |
|
"loss": 1.6616, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0001894993269777411, |
|
"loss": 1.4135, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"learning_rate": 0.0001893057270769381, |
|
"loss": 1.3604, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00018911045950827693, |
|
"loss": 1.4385, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00018891352791812452, |
|
"loss": 1.8386, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00018871493598392124, |
|
"loss": 1.1961, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0001885146874141121, |
|
"loss": 1.138, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.00018831278594807783, |
|
"loss": 1.1199, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.0001881092353560646, |
|
"loss": 1.3006, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"learning_rate": 0.000187904039439114, |
|
"loss": 1.1866, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00018769720202899194, |
|
"loss": 1.1488, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00018748872698811693, |
|
"loss": 1.3549, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0001872786182094882, |
|
"loss": 1.203, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.0001870668796166129, |
|
"loss": 1.446, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00018685351516343278, |
|
"loss": 1.5952, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00018663852883425046, |
|
"loss": 1.3839, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00018642192464365497, |
|
"loss": 1.458, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 0.00018620370663644678, |
|
"loss": 1.3013, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00018598387888756223, |
|
"loss": 1.4674, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00018576244550199758, |
|
"loss": 1.364, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00018553941061473218, |
|
"loss": 1.3133, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.0001853147783906514, |
|
"loss": 1.5333, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00018508855302446867, |
|
"loss": 1.2932, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.00018486073874064745, |
|
"loss": 1.3218, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"learning_rate": 0.000184631339793322, |
|
"loss": 1.5033, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00018440036046621816, |
|
"loss": 1.1661, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00018416780507257333, |
|
"loss": 1.2916, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00018393367795505587, |
|
"loss": 1.402, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.000183697983485684, |
|
"loss": 1.262, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0001834607260657443, |
|
"loss": 1.412, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.00018322191012570919, |
|
"loss": 1.4023, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0001829815401251547, |
|
"loss": 1.5077, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"learning_rate": 0.0001827396205526767, |
|
"loss": 1.3016, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00018249615592580732, |
|
"loss": 1.7402, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0001822511507909307, |
|
"loss": 1.4275, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00018200460972319786, |
|
"loss": 1.3369, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.0001817565373264413, |
|
"loss": 1.5247, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00018150693823308913, |
|
"loss": 1.1656, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.00018125581710407864, |
|
"loss": 1.3228, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"learning_rate": 0.000181003178628769, |
|
"loss": 1.3855, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.0001807490275248539, |
|
"loss": 1.3271, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018049336853827343, |
|
"loss": 1.2697, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00018023620644312539, |
|
"loss": 1.2532, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00017997754604157605, |
|
"loss": 1.2072, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00017971739216377089, |
|
"loss": 1.3372, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00017945574966774376, |
|
"loss": 1.4446, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00017919262343932678, |
|
"loss": 1.293, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"learning_rate": 0.00017892801839205867, |
|
"loss": 1.4083, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00017866193946709327, |
|
"loss": 1.2635, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00017839439163310714, |
|
"loss": 1.1442, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00017812537988620675, |
|
"loss": 1.3182, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00017785490924983525, |
|
"loss": 1.1219, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00017758298477467865, |
|
"loss": 1.1947, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00017730961153857155, |
|
"loss": 1.3136, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 0.00017703479464640216, |
|
"loss": 1.3152, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001767585392300172, |
|
"loss": 1.2263, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001764808504481259, |
|
"loss": 1.1573, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00017620173348620368, |
|
"loss": 1.3633, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00017592119355639544, |
|
"loss": 1.3713, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00017563923589741806, |
|
"loss": 1.3472, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00017535586577446276, |
|
"loss": 1.3634, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.00017507108847909656, |
|
"loss": 1.5102, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"learning_rate": 0.0001747849093291637, |
|
"loss": 1.7271, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00017449733366868618, |
|
"loss": 1.5071, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.000174208366867764, |
|
"loss": 1.3833, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00017391801432247488, |
|
"loss": 1.3213, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00017362628145477354, |
|
"loss": 1.2756, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00017333317371239044, |
|
"loss": 1.283, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00017303869656872995, |
|
"loss": 1.2728, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"learning_rate": 0.00017274285552276828, |
|
"loss": 1.3388, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00017244565609895074, |
|
"loss": 1.2645, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001721471038470885, |
|
"loss": 1.3977, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00017184720434225518, |
|
"loss": 1.3328, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001715459631846824, |
|
"loss": 1.3337, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001712433859996555, |
|
"loss": 1.3793, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001709394784374084, |
|
"loss": 1.264, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.00017063424617301805, |
|
"loss": 1.5134, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"learning_rate": 0.0001703276949062985, |
|
"loss": 1.5133, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0001700198303616944, |
|
"loss": 1.2208, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00016971065828817424, |
|
"loss": 1.1971, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00016940018445912272, |
|
"loss": 1.2801, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00016908841467223336, |
|
"loss": 1.4235, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00016877535474939988, |
|
"loss": 1.2821, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.00016846101053660762, |
|
"loss": 1.3171, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"learning_rate": 0.0001681453879038243, |
|
"loss": 1.2901, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00016782849274489054, |
|
"loss": 1.5555, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00016751033097740976, |
|
"loss": 1.1944, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00016719090854263753, |
|
"loss": 1.2592, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.0001668702314053708, |
|
"loss": 1.2844, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00016654830555383647, |
|
"loss": 1.2734, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00016622513699957948, |
|
"loss": 1.4152, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00016590073177735065, |
|
"loss": 1.3154, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 0.00016557509594499403, |
|
"loss": 1.2228, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00016524823558333363, |
|
"loss": 1.3587, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00016492015679605993, |
|
"loss": 1.1756, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00016459086570961594, |
|
"loss": 1.3158, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00016426036847308286, |
|
"loss": 1.2782, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00016392867125806504, |
|
"loss": 1.3629, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00016359578025857495, |
|
"loss": 1.2868, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"learning_rate": 0.00016326170169091733, |
|
"loss": 1.4121, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00016292644179357336, |
|
"loss": 1.3111, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00016259000682708385, |
|
"loss": 1.3113, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00016225240307393257, |
|
"loss": 1.5472, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00016191363683842883, |
|
"loss": 1.3063, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0001615737144465898, |
|
"loss": 1.2559, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00016123264224602245, |
|
"loss": 1.1525, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.00016089042660580468, |
|
"loss": 1.3345, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"learning_rate": 0.0001605470739163669, |
|
"loss": 1.0383, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00016020259058937227, |
|
"loss": 1.3177, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00015985698305759713, |
|
"loss": 1.308, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00015951025777481096, |
|
"loss": 1.5034, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.0001591624212156558, |
|
"loss": 1.1636, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00015881347987552517, |
|
"loss": 1.4084, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00015846344027044306, |
|
"loss": 1.2837, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"learning_rate": 0.00015811230893694213, |
|
"loss": 1.3305, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00015776009243194158, |
|
"loss": 1.3823, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001574067973326248, |
|
"loss": 1.3252, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00015705243023631652, |
|
"loss": 1.0895, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00015669699776035958, |
|
"loss": 1.2711, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00015634050654199147, |
|
"loss": 1.1413, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00015598296323822024, |
|
"loss": 1.1686, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.0001556243745257003, |
|
"loss": 1.3666, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"learning_rate": 0.00015526474710060768, |
|
"loss": 1.344, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00015490408767851506, |
|
"loss": 1.3171, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00015454240299426627, |
|
"loss": 1.1675, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00015417969980185053, |
|
"loss": 1.2013, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0001538159848742765, |
|
"loss": 1.2159, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00015345126500344554, |
|
"loss": 1.1951, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.0001530855470000251, |
|
"loss": 1.3548, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 0.00015271883769332143, |
|
"loss": 1.2713, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00015235114393115202, |
|
"loss": 1.1563, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00015198247257971787, |
|
"loss": 1.3486, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00015161283052347516, |
|
"loss": 1.0929, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00015124222466500665, |
|
"loss": 1.411, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00015087066192489288, |
|
"loss": 1.121, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00015049814924158298, |
|
"loss": 0.9785, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00015012469357126494, |
|
"loss": 1.1973, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"learning_rate": 0.00014975030188773585, |
|
"loss": 1.6289, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00014937498118227155, |
|
"loss": 1.3897, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00014899873846349626, |
|
"loss": 1.2387, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.0001486215807572515, |
|
"loss": 1.1199, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00014824351510646507, |
|
"loss": 1.2615, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00014786454857101939, |
|
"loss": 1.3599, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00014748468822761975, |
|
"loss": 1.2769, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"learning_rate": 0.00014710394116966205, |
|
"loss": 1.05, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014672231450710066, |
|
"loss": 1.3054, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014633981536631512, |
|
"loss": 1.2249, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014595645088997757, |
|
"loss": 1.2833, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014557222823691912, |
|
"loss": 1.3083, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.0001451871545819961, |
|
"loss": 1.2982, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014480123711595636, |
|
"loss": 1.3815, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014441448304530467, |
|
"loss": 1.1667, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"learning_rate": 0.00014402689959216845, |
|
"loss": 1.4142, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00014363849399416255, |
|
"loss": 1.5461, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0001432492735042545, |
|
"loss": 1.345, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.00014285924539062878, |
|
"loss": 1.2894, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0001424684169365512, |
|
"loss": 1.2761, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0001420767954402329, |
|
"loss": 1.3956, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.000141684388214694, |
|
"loss": 1.5331, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"learning_rate": 0.0001412912025876272, |
|
"loss": 1.411, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0001408972459012606, |
|
"loss": 1.169, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00014050252551222115, |
|
"loss": 1.1902, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0001401070487913967, |
|
"loss": 1.2907, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00013971082312379864, |
|
"loss": 1.4225, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00013931385590842412, |
|
"loss": 1.3166, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.0001389161545581175, |
|
"loss": 1.2402, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00013851772649943237, |
|
"loss": 1.4516, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 0.00013811857917249253, |
|
"loss": 1.2131, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00013771872003085316, |
|
"loss": 1.2261, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00013731815654136168, |
|
"loss": 1.4692, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00013691689618401835, |
|
"loss": 1.1822, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.0001365149464518364, |
|
"loss": 1.2183, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00013611231485070231, |
|
"loss": 1.1812, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00013570900889923565, |
|
"loss": 1.3421, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"learning_rate": 0.00013530503612864847, |
|
"loss": 1.2209, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.0001349004040826048, |
|
"loss": 1.3287, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00013449512031707986, |
|
"loss": 1.7032, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00013408919240021888, |
|
"loss": 1.1261, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00013368262791219567, |
|
"loss": 1.1894, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00013327543444507134, |
|
"loss": 1.4187, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00013286761960265214, |
|
"loss": 1.333, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00013245919100034794, |
|
"loss": 1.1201, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"learning_rate": 0.00013205015626502957, |
|
"loss": 1.1145, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00013164052303488673, |
|
"loss": 1.1974, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00013123029895928515, |
|
"loss": 1.2703, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00013081949169862396, |
|
"loss": 1.0925, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.0001304081089241923, |
|
"loss": 1.2201, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00012999615831802647, |
|
"loss": 1.3681, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00012958364757276614, |
|
"loss": 1.2453, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"learning_rate": 0.00012917058439151102, |
|
"loss": 1.1403, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00012875697648767663, |
|
"loss": 1.0894, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00012834283158485063, |
|
"loss": 1.2346, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00012792815741664837, |
|
"loss": 1.4033, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00012751296172656862, |
|
"loss": 1.1793, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00012709725226784873, |
|
"loss": 1.0227, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00012668103680332012, |
|
"loss": 1.6327, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.0001262643231052632, |
|
"loss": 1.1799, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 0.00012584711895526226, |
|
"loss": 1.2269, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001254294321440601, |
|
"loss": 1.2257, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001250112704714126, |
|
"loss": 1.2054, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00012459264174594304, |
|
"loss": 1.2643, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001241735537849963, |
|
"loss": 1.403, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.00012375401441449294, |
|
"loss": 1.3265, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.000123334031468783, |
|
"loss": 1.1098, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"learning_rate": 0.0001229136127904996, |
|
"loss": 1.334, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00012249276623041267, |
|
"loss": 1.1986, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00012207149964728236, |
|
"loss": 1.4485, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00012164982090771202, |
|
"loss": 1.3588, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00012122773788600163, |
|
"loss": 1.3638, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00012080525846400055, |
|
"loss": 1.0629, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00012038239053096038, |
|
"loss": 1.3022, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00011995914198338776, |
|
"loss": 1.1267, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"learning_rate": 0.00011953552072489666, |
|
"loss": 1.3345, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00011911153466606104, |
|
"loss": 1.2134, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00011868719172426703, |
|
"loss": 1.4767, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00011826249982356501, |
|
"loss": 1.1404, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00011783746689452177, |
|
"loss": 1.2737, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0001174121008740724, |
|
"loss": 1.2809, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.00011698640970537193, |
|
"loss": 1.313, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"learning_rate": 0.0001165604013376472, |
|
"loss": 1.2174, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00011613408372604825, |
|
"loss": 1.3712, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00011570746483149997, |
|
"loss": 1.1396, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00011528055262055317, |
|
"loss": 1.348, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.000114853355065236, |
|
"loss": 1.2176, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00011442588014290512, |
|
"loss": 1.2583, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.0001139981358360966, |
|
"loss": 1.2568, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00011357013013237689, |
|
"loss": 1.1061, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.47, |
|
"learning_rate": 0.00011314187102419374, |
|
"loss": 1.2565, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00011271336650872685, |
|
"loss": 1.2792, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00011228462458773865, |
|
"loss": 1.211, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00011185565326742473, |
|
"loss": 1.1221, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00011142646055826442, |
|
"loss": 1.4592, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00011099705447487128, |
|
"loss": 1.1723, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00011056744303584321, |
|
"loss": 1.0701, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 0.00011013763426361303, |
|
"loss": 1.3934, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0001097076361842984, |
|
"loss": 1.2439, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00010927745682755201, |
|
"loss": 1.1544, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0001088471042264118, |
|
"loss": 1.2123, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00010841658641715065, |
|
"loss": 1.207, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.0001079859114391266, |
|
"loss": 1.0921, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00010755508733463265, |
|
"loss": 1.3459, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00010712412214874639, |
|
"loss": 1.2615, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.49, |
|
"learning_rate": 0.00010669302392918005, |
|
"loss": 1.283, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.0001062618007261301, |
|
"loss": 1.2935, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00010583046059212678, |
|
"loss": 1.2122, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00010539901158188398, |
|
"loss": 1.1683, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00010496746175214868, |
|
"loss": 1.1848, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00010453581916155059, |
|
"loss": 1.1766, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00010410409187045145, |
|
"loss": 1.2522, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"learning_rate": 0.00010367228794079482, |
|
"loss": 1.3705, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00010324041543595535, |
|
"loss": 1.2845, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00010280848242058819, |
|
"loss": 1.2278, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.0001023764969604785, |
|
"loss": 1.075, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00010194446712239076, |
|
"loss": 1.2883, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00010151240097391819, |
|
"loss": 1.2658, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00010108030658333192, |
|
"loss": 1.3066, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00010064819201943066, |
|
"loss": 1.0891, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.51, |
|
"learning_rate": 0.00010021606535138965, |
|
"loss": 1.297, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.978393464861035e-05, |
|
"loss": 1.2637, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.935180798056935e-05, |
|
"loss": 1.2524, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.891969341666809e-05, |
|
"loss": 1.1895, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.848759902608186e-05, |
|
"loss": 1.1293, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.805553287760922e-05, |
|
"loss": 1.3118, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.76235030395215e-05, |
|
"loss": 1.4023, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"learning_rate": 9.719151757941184e-05, |
|
"loss": 1.1086, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.675958456404467e-05, |
|
"loss": 1.3357, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.632771205920519e-05, |
|
"loss": 1.1231, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.589590812954857e-05, |
|
"loss": 1.141, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.546418083844943e-05, |
|
"loss": 1.3655, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.503253824785133e-05, |
|
"loss": 1.3134, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.460098841811601e-05, |
|
"loss": 1.0862, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.416953940787324e-05, |
|
"loss": 1.5868, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 9.373819927386995e-05, |
|
"loss": 1.2349, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.330697607081994e-05, |
|
"loss": 1.1996, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.287587785125364e-05, |
|
"loss": 1.3662, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.24449126653674e-05, |
|
"loss": 1.3599, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.20140885608734e-05, |
|
"loss": 1.0668, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.158341358284939e-05, |
|
"loss": 1.0857, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.115289577358825e-05, |
|
"loss": 1.2289, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"learning_rate": 9.072254317244801e-05, |
|
"loss": 1.2852, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 9.029236381570161e-05, |
|
"loss": 1.2121, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.986236573638696e-05, |
|
"loss": 1.2368, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.94325569641568e-05, |
|
"loss": 1.4359, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.900294552512877e-05, |
|
"loss": 1.2492, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.857353944173559e-05, |
|
"loss": 1.0758, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.81443467325753e-05, |
|
"loss": 1.2528, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.771537541226139e-05, |
|
"loss": 1.3388, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"learning_rate": 8.728663349127314e-05, |
|
"loss": 1.1169, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.685812897580629e-05, |
|
"loss": 1.1608, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.642986986762315e-05, |
|
"loss": 1.0971, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.600186416390342e-05, |
|
"loss": 1.0571, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.55741198570949e-05, |
|
"loss": 1.114, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.514664493476401e-05, |
|
"loss": 1.2416, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.471944737944686e-05, |
|
"loss": 1.2167, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.56, |
|
"learning_rate": 8.429253516850005e-05, |
|
"loss": 1.22, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.386591627395173e-05, |
|
"loss": 1.2406, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.343959866235283e-05, |
|
"loss": 1.1654, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.301359029462809e-05, |
|
"loss": 1.1381, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.25878991259276e-05, |
|
"loss": 1.348, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.216253310547824e-05, |
|
"loss": 1.1045, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.173750017643504e-05, |
|
"loss": 1.0247, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.1312808275733e-05, |
|
"loss": 1.2314, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"learning_rate": 8.088846533393899e-05, |
|
"loss": 1.3811, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.046447927510335e-05, |
|
"loss": 1.1985, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 8.004085801661227e-05, |
|
"loss": 1.1293, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.961760946903963e-05, |
|
"loss": 1.0805, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.919474153599949e-05, |
|
"loss": 1.1358, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.877226211399839e-05, |
|
"loss": 1.5697, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.835017909228802e-05, |
|
"loss": 1.1928, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 7.792850035271768e-05, |
|
"loss": 1.2692, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.750723376958733e-05, |
|
"loss": 1.2783, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.708638720950044e-05, |
|
"loss": 1.2121, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.666596853121702e-05, |
|
"loss": 1.1913, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.624598558550706e-05, |
|
"loss": 1.2804, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.582644621500368e-05, |
|
"loss": 1.1179, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.540735825405699e-05, |
|
"loss": 1.2151, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.498872952858744e-05, |
|
"loss": 1.4533, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"learning_rate": 7.45705678559399e-05, |
|
"loss": 1.4376, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.415288104473774e-05, |
|
"loss": 1.1821, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.373567689473683e-05, |
|
"loss": 1.1519, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.33189631966799e-05, |
|
"loss": 1.2709, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.290274773215132e-05, |
|
"loss": 1.1006, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.248703827343142e-05, |
|
"loss": 1.2009, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.207184258335162e-05, |
|
"loss": 1.2132, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.6, |
|
"learning_rate": 7.165716841514939e-05, |
|
"loss": 0.9096, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.124302351232336e-05, |
|
"loss": 1.221, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.0829415608489e-05, |
|
"loss": 1.2712, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.041635242723385e-05, |
|
"loss": 1.1966, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 7.000384168197354e-05, |
|
"loss": 1.1948, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.95918910758077e-05, |
|
"loss": 1.2142, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.918050830137609e-05, |
|
"loss": 1.1955, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.876970104071482e-05, |
|
"loss": 1.1356, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"learning_rate": 6.83594769651133e-05, |
|
"loss": 1.2463, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.794984373497048e-05, |
|
"loss": 1.3856, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.754080899965208e-05, |
|
"loss": 1.1603, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.713238039734788e-05, |
|
"loss": 1.0642, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.67245655549287e-05, |
|
"loss": 1.2829, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.631737208780433e-05, |
|
"loss": 1.2689, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.591080759978113e-05, |
|
"loss": 1.2362, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.62, |
|
"learning_rate": 6.550487968292012e-05, |
|
"loss": 1.1748, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.509959591739522e-05, |
|
"loss": 1.0881, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.469496387135158e-05, |
|
"loss": 1.0945, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.429099110076435e-05, |
|
"loss": 1.4631, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.388768514929768e-05, |
|
"loss": 1.2458, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.348505354816364e-05, |
|
"loss": 1.1905, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.308310381598168e-05, |
|
"loss": 0.9306, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.268184345863835e-05, |
|
"loss": 1.207, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 6.228127996914686e-05, |
|
"loss": 1.1058, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.18814208275075e-05, |
|
"loss": 1.269, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.148227350056763e-05, |
|
"loss": 1.2571, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.10838454418825e-05, |
|
"loss": 1.4694, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.068614409157591e-05, |
|
"loss": 1.4105, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 6.028917687620138e-05, |
|
"loss": 1.0538, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 5.989295120860334e-05, |
|
"loss": 1.0992, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 0.64, |
|
"learning_rate": 5.949747448777889e-05, |
|
"loss": 1.2507, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.9102754098739424e-05, |
|
"loss": 0.9888, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.870879741237285e-05, |
|
"loss": 1.154, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.831561178530602e-05, |
|
"loss": 0.988, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.792320455976714e-05, |
|
"loss": 1.1021, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.7531583063448816e-05, |
|
"loss": 1.0548, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.714075460937125e-05, |
|
"loss": 1.2447, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.675072649574551e-05, |
|
"loss": 1.1437, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"learning_rate": 5.6361506005837475e-05, |
|
"loss": 1.003, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.597310040783161e-05, |
|
"loss": 1.319, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.5585516954695316e-05, |
|
"loss": 1.0514, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.5198762884043666e-05, |
|
"loss": 1.2173, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.481284541800391e-05, |
|
"loss": 1.2133, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.442777176308089e-05, |
|
"loss": 1.0926, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.404354911002243e-05, |
|
"loss": 1.2688, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"learning_rate": 5.36601846336849e-05, |
|
"loss": 1.1891, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.327768549289934e-05, |
|
"loss": 1.2581, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.289605883033792e-05, |
|
"loss": 1.1357, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.251531177238029e-05, |
|
"loss": 1.0797, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.213545142898061e-05, |
|
"loss": 1.2987, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.175648489353493e-05, |
|
"loss": 1.463, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.137841924274851e-05, |
|
"loss": 1.0267, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.1001261536503787e-05, |
|
"loss": 1.0952, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 5.06250188177285e-05, |
|
"loss": 1.2898, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 5.024969811226419e-05, |
|
"loss": 1.0503, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.9875306428735094e-05, |
|
"loss": 1.2685, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.9501850758417056e-05, |
|
"loss": 1.2042, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.912933807510713e-05, |
|
"loss": 1.4269, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.875777533499339e-05, |
|
"loss": 1.2585, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.838716947652485e-05, |
|
"loss": 1.0463, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"learning_rate": 4.8017527420282135e-05, |
|
"loss": 1.1023, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.7648856068848e-05, |
|
"loss": 1.0081, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.7281162306678585e-05, |
|
"loss": 1.2722, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.691445299997491e-05, |
|
"loss": 1.2418, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.6548734996554486e-05, |
|
"loss": 1.1642, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.618401512572351e-05, |
|
"loss": 1.3041, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.582030019814948e-05, |
|
"loss": 1.0348, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.545759700573378e-05, |
|
"loss": 1.1782, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.69, |
|
"learning_rate": 4.509591232148495e-05, |
|
"loss": 1.3283, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.473525289939233e-05, |
|
"loss": 1.0124, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.437562547429971e-05, |
|
"loss": 0.9797, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.4017036761779787e-05, |
|
"loss": 1.1973, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.365949345800856e-05, |
|
"loss": 1.2428, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.330300223964042e-05, |
|
"loss": 1.2354, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.2947569763683506e-05, |
|
"loss": 1.1127, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"learning_rate": 4.259320266737522e-05, |
|
"loss": 1.3324, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.223990756805841e-05, |
|
"loss": 1.2347, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.188769106305787e-05, |
|
"loss": 1.0754, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.1536559729556944e-05, |
|
"loss": 1.1642, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.118652012447486e-05, |
|
"loss": 1.2139, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.083757878434422e-05, |
|
"loss": 1.4011, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.048974222518904e-05, |
|
"loss": 1.2346, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 4.01430169424029e-05, |
|
"loss": 1.1561, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.71, |
|
"learning_rate": 3.9797409410627794e-05, |
|
"loss": 1.1374, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.945292608363313e-05, |
|
"loss": 1.2548, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.910957339419533e-05, |
|
"loss": 1.09, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.8767357753977596e-05, |
|
"loss": 1.1598, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.842628555341018e-05, |
|
"loss": 1.2341, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.808636316157119e-05, |
|
"loss": 1.2293, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.7747596926067484e-05, |
|
"loss": 1.1257, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 3.740999317291618e-05, |
|
"loss": 1.1656, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.7073558206426675e-05, |
|
"loss": 0.9748, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.673829830908266e-05, |
|
"loss": 1.1949, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.640421974142508e-05, |
|
"loss": 1.4119, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.607132874193499e-05, |
|
"loss": 1.1605, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.573963152691715e-05, |
|
"loss": 1.3131, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.540913429038407e-05, |
|
"loss": 1.2644, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.507984320394012e-05, |
|
"loss": 1.1329, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.73, |
|
"learning_rate": 3.47517644166664e-05, |
|
"loss": 1.2342, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.442490405500598e-05, |
|
"loss": 1.0007, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.409926822264937e-05, |
|
"loss": 1.3068, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.377486300042054e-05, |
|
"loss": 1.2366, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.3451694446163553e-05, |
|
"loss": 1.148, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.3129768594629186e-05, |
|
"loss": 1.0449, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.280909145736246e-05, |
|
"loss": 1.1513, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"learning_rate": 3.2489669022590244e-05, |
|
"loss": 1.0452, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.217150725510946e-05, |
|
"loss": 1.2184, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.185461209617571e-05, |
|
"loss": 0.9621, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.153898946339241e-05, |
|
"loss": 1.2505, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.122464525060013e-05, |
|
"loss": 1.1572, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.091158532776666e-05, |
|
"loss": 1.0986, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.059981554087732e-05, |
|
"loss": 1.2599, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 3.0289341711825815e-05, |
|
"loss": 1.261, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.75, |
|
"learning_rate": 2.998016963830562e-05, |
|
"loss": 0.9793, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.9672305093701503e-05, |
|
"loss": 1.2321, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.9365753826981947e-05, |
|
"loss": 1.2587, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.9060521562591624e-05, |
|
"loss": 1.2269, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.875661400034452e-05, |
|
"loss": 1.1654, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.8454036815317642e-05, |
|
"loss": 1.1827, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.815279565774488e-05, |
|
"loss": 1.0381, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"learning_rate": 2.78528961529115e-05, |
|
"loss": 1.2321, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.7554343901049294e-05, |
|
"loss": 1.1796, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.7257144477231756e-05, |
|
"loss": 1.2575, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.6961303431270068e-05, |
|
"loss": 1.0709, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.666682628760958e-05, |
|
"loss": 1.0956, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.6373718545226445e-05, |
|
"loss": 1.2598, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.608198567752512e-05, |
|
"loss": 1.1718, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.5791633132236025e-05, |
|
"loss": 1.0968, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 2.550266633131382e-05, |
|
"loss": 1.2986, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.5215090670836306e-05, |
|
"loss": 1.2091, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.4928911520903465e-05, |
|
"loss": 1.0521, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.4644134225537264e-05, |
|
"loss": 1.1918, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.4360764102581945e-05, |
|
"loss": 1.1098, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.4078806443604596e-05, |
|
"loss": 1.3272, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.3798266513796317e-05, |
|
"loss": 1.1128, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"learning_rate": 2.351914955187412e-05, |
|
"loss": 0.9976, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.3241460769982814e-05, |
|
"loss": 1.0915, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.2965205353597863e-05, |
|
"loss": 1.3458, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.2690388461428468e-05, |
|
"loss": 1.093, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.241701522532136e-05, |
|
"loss": 1.2252, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.2145090750164777e-05, |
|
"loss": 1.272, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.1874620113793286e-05, |
|
"loss": 1.1345, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.160560836689286e-05, |
|
"loss": 1.2129, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"learning_rate": 2.1338060532906736e-05, |
|
"loss": 1.4181, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.107198160794136e-05, |
|
"loss": 0.9967, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.0807376560673254e-05, |
|
"loss": 1.1963, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.0544250332256276e-05, |
|
"loss": 1.1214, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.028260783622914e-05, |
|
"loss": 1.0757, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 2.0022453958423938e-05, |
|
"loss": 1.1913, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.9763793556874655e-05, |
|
"loss": 1.1523, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 0.8, |
|
"learning_rate": 1.9506631461726566e-05, |
|
"loss": 0.9793, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.925097247514609e-05, |
|
"loss": 1.4055, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8996821371231022e-05, |
|
"loss": 1.0458, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8744182895921368e-05, |
|
"loss": 1.4917, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.849306176691088e-05, |
|
"loss": 1.2117, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.8243462673558755e-05, |
|
"loss": 1.2286, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.799539027680216e-05, |
|
"loss": 1.2713, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.7748849209069286e-05, |
|
"loss": 1.3255, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"learning_rate": 1.7503844074192653e-05, |
|
"loss": 1.0577, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.7260379447323328e-05, |
|
"loss": 1.1908, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.7018459874845326e-05, |
|
"loss": 1.2434, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.677808987429079e-05, |
|
"loss": 1.192, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.6539273934255727e-05, |
|
"loss": 1.1859, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.6302016514315998e-05, |
|
"loss": 1.1361, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.6066322044944126e-05, |
|
"loss": 1.1442, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 1.5832194927426668e-05, |
|
"loss": 1.342, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.5599639533781853e-05, |
|
"loss": 1.2013, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.5368660206678033e-05, |
|
"loss": 1.3602, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.5139261259352589e-05, |
|
"loss": 1.3237, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4911446975531328e-05, |
|
"loss": 1.0208, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4685221609348632e-05, |
|
"loss": 1.0632, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4460589385267842e-05, |
|
"loss": 1.0564, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4237554498002425e-05, |
|
"loss": 1.1912, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"learning_rate": 1.4016121112437786e-05, |
|
"loss": 1.4941, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.379629336355326e-05, |
|
"loss": 1.1406, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3578075356345044e-05, |
|
"loss": 1.1407, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3361471165749562e-05, |
|
"loss": 1.1103, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.3146484836567264e-05, |
|
"loss": 1.2449, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.2933120383387132e-05, |
|
"loss": 1.1988, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.2721381790511832e-05, |
|
"loss": 1.2136, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 0.84, |
|
"learning_rate": 1.2511273011883096e-05, |
|
"loss": 1.0404, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2302797971008085e-05, |
|
"loss": 1.2031, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.2095960560886e-05, |
|
"loss": 1.1924, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1890764643935393e-05, |
|
"loss": 1.3458, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.168721405192218e-05, |
|
"loss": 1.2207, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1485312585887886e-05, |
|
"loss": 0.9582, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1285064016078784e-05, |
|
"loss": 1.3107, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.1086472081875487e-05, |
|
"loss": 1.1833, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"learning_rate": 1.0889540491723105e-05, |
|
"loss": 1.0454, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.0694272923061932e-05, |
|
"loss": 1.1076, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.0500673022258922e-05, |
|
"loss": 1.1454, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.030874440453944e-05, |
|
"loss": 1.1487, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 1.0118490653919855e-05, |
|
"loss": 0.9731, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.929915323140571e-06, |
|
"loss": 0.8489, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.743021933599695e-06, |
|
"loss": 1.2174, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 0.86, |
|
"learning_rate": 9.557813975287267e-06, |
|
"loss": 1.3401, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.374294906720082e-06, |
|
"loss": 1.1344, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.192468154877187e-06, |
|
"loss": 1.1999, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 9.012337115135772e-06, |
|
"loss": 1.3144, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.833905151207833e-06, |
|
"loss": 1.1487, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.657175595077316e-06, |
|
"loss": 1.0052, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.482151746937982e-06, |
|
"loss": 1.1675, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.308836875131665e-06, |
|
"loss": 1.1484, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 8.137234216087353e-06, |
|
"loss": 0.9894, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.967346974260625e-06, |
|
"loss": 0.9718, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.79917832207394e-06, |
|
"loss": 1.2581, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.63273139985733e-06, |
|
"loss": 1.1366, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.468009315789748e-06, |
|
"loss": 1.3317, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.305015145841054e-06, |
|
"loss": 0.993, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 7.143751933714582e-06, |
|
"loss": 1.242, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 0.88, |
|
"learning_rate": 6.9842226907902766e-06, |
|
"loss": 1.2982, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.8264303960684506e-06, |
|
"loss": 1.1203, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.67037799611423e-06, |
|
"loss": 0.9316, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.51606840500244e-06, |
|
"loss": 1.1395, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.3635045042632066e-06, |
|
"loss": 1.0749, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.212689142828232e-06, |
|
"loss": 1.2441, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 6.063625136977447e-06, |
|
"loss": 1.2737, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 5.916315270286588e-06, |
|
"loss": 1.1665, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"learning_rate": 5.770762293575083e-06, |
|
"loss": 1.1282, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.626968924854714e-06, |
|
"loss": 1.1662, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.4849378492789374e-06, |
|
"loss": 1.1101, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.344671719092664e-06, |
|
"loss": 1.3308, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.206173153582705e-06, |
|
"loss": 1.1049, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 5.069444739029006e-06, |
|
"loss": 1.0428, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.934489028656164e-06, |
|
"loss": 1.0741, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"learning_rate": 4.801308542585892e-06, |
|
"loss": 1.0412, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.669905767789884e-06, |
|
"loss": 1.0229, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.540283158043391e-06, |
|
"loss": 1.1564, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.41244313387944e-06, |
|
"loss": 1.2456, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.286388082543569e-06, |
|
"loss": 1.0691, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.162120357949284e-06, |
|
"loss": 1.2539, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 4.0396422806341036e-06, |
|
"loss": 1.1281, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.918956137716235e-06, |
|
"loss": 1.0582, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.91, |
|
"learning_rate": 3.8000641828518036e-06, |
|
"loss": 1.1231, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.682968636192863e-06, |
|
"loss": 1.2181, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.567671684345875e-06, |
|
"loss": 1.1241, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.454175480330857e-06, |
|
"loss": 1.3235, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.342482143541281e-06, |
|
"loss": 1.1104, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.23259375970435e-06, |
|
"loss": 1.215, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.124512380842204e-06, |
|
"loss": 0.9921, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 3.018240025233465e-06, |
|
"loss": 1.3698, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.9137786773756117e-06, |
|
"loss": 1.14, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.8111302879479717e-06, |
|
"loss": 1.0937, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.710296773775167e-06, |
|
"loss": 1.077, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.6112800177914398e-06, |
|
"loss": 1.3246, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.514081869005458e-06, |
|
"loss": 1.1759, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.418704142465722e-06, |
|
"loss": 1.1861, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.3251486192267578e-06, |
|
"loss": 0.9643, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.93, |
|
"learning_rate": 2.233417046315822e-06, |
|
"loss": 1.1879, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.1435111367002824e-06, |
|
"loss": 1.0398, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 2.055432569255622e-06, |
|
"loss": 1.0496, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.9691829887340864e-06, |
|
"loss": 0.9831, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.8847640057339965e-06, |
|
"loss": 1.0613, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.8021771966696189e-06, |
|
"loss": 1.1937, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.721424103741831e-06, |
|
"loss": 1.1883, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"learning_rate": 1.6425062349091913e-06, |
|
"loss": 1.0838, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.56542506385986e-06, |
|
"loss": 1.3727, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4901820299840997e-06, |
|
"loss": 1.2732, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.4167785383472854e-06, |
|
"loss": 1.1294, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.345215959663837e-06, |
|
"loss": 1.0111, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.2754956302714615e-06, |
|
"loss": 1.1271, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.2076188521062847e-06, |
|
"loss": 1.2032, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.1415868926785256e-06, |
|
"loss": 1.2941, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.95, |
|
"learning_rate": 1.0774009850488153e-06, |
|
"loss": 1.2416, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 1.0150623278051718e-06, |
|
"loss": 1.1924, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 9.545720850406504e-07, |
|
"loss": 1.1774, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.959313863315389e-07, |
|
"loss": 1.1971, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 8.391413267163417e-07, |
|
"loss": 1.2369, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.842029666752626e-07, |
|
"loss": 1.0584, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 7.311173321104647e-07, |
|
"loss": 1.3607, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 6.798854143268641e-07, |
|
"loss": 1.3582, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 6.305081700136328e-07, |
|
"loss": 1.0977, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.829865212263474e-07, |
|
"loss": 1.1176, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 5.373213553697576e-07, |
|
"loss": 1.0638, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.935135251811996e-07, |
|
"loss": 1.2728, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.515638487147311e-07, |
|
"loss": 1.2765, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 4.114731093257884e-07, |
|
"loss": 1.2579, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.732420556565752e-07, |
|
"loss": 1.0823, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.97, |
|
"learning_rate": 3.368714016221186e-07, |
|
"loss": 1.1057, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 3.023618263968797e-07, |
|
"loss": 1.0939, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.697139744021415e-07, |
|
"loss": 1.265, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.389284552939075e-07, |
|
"loss": 1.1871, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 2.1000584395155509e-07, |
|
"loss": 0.9003, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.8294668046708873e-07, |
|
"loss": 1.0993, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.577514701350591e-07, |
|
"loss": 1.228, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"learning_rate": 1.3442068344312608e-07, |
|
"loss": 1.1979, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.129547560632771e-07, |
|
"loss": 0.9901, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 9.33540888436668e-08, |
|
"loss": 1.1744, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 7.561904780116758e-08, |
|
"loss": 1.0345, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 5.97499641145416e-08, |
|
"loss": 1.0167, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 4.574713411816811e-08, |
|
"loss": 1.1209, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 3.361081929664778e-08, |
|
"loss": 1.1664, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 2.3341246279806607e-08, |
|
"loss": 1.1198, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.99, |
|
"learning_rate": 1.4938606838510448e-08, |
|
"loss": 1.1858, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 8.403057881067877e-09, |
|
"loss": 1.2325, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 3.734721450343592e-09, |
|
"loss": 1.1638, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 9.336847214269639e-10, |
|
"loss": 1.2065, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"learning_rate": 0.0, |
|
"loss": 1.2331, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 750, |
|
"total_flos": 1.2117523283994214e+17, |
|
"train_loss": 1.253919720808665, |
|
"train_runtime": 5931.6255, |
|
"train_samples_per_second": 2.023, |
|
"train_steps_per_second": 0.126 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 750, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 50000, |
|
"total_flos": 1.2117523283994214e+17, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|