lesso's picture
Training in progress, step 200, checkpoint
0d6764d verified
{
"best_metric": 0.8263623714447021,
"best_model_checkpoint": "miner_id_24/checkpoint-200",
"epoch": 0.4434589800443459,
"eval_steps": 50,
"global_step": 200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0022172949002217295,
"grad_norm": 43.49946975708008,
"learning_rate": 1.004e-05,
"loss": 2.8648,
"step": 1
},
{
"epoch": 0.0022172949002217295,
"eval_loss": 2.7855024337768555,
"eval_runtime": 20.2353,
"eval_samples_per_second": 9.39,
"eval_steps_per_second": 2.372,
"step": 1
},
{
"epoch": 0.004434589800443459,
"grad_norm": 49.54320526123047,
"learning_rate": 2.008e-05,
"loss": 3.479,
"step": 2
},
{
"epoch": 0.0066518847006651885,
"grad_norm": 43.55087661743164,
"learning_rate": 3.012e-05,
"loss": 3.174,
"step": 3
},
{
"epoch": 0.008869179600886918,
"grad_norm": 34.47139358520508,
"learning_rate": 4.016e-05,
"loss": 2.9292,
"step": 4
},
{
"epoch": 0.011086474501108648,
"grad_norm": 33.695804595947266,
"learning_rate": 5.02e-05,
"loss": 2.9573,
"step": 5
},
{
"epoch": 0.013303769401330377,
"grad_norm": 35.9599609375,
"learning_rate": 6.024e-05,
"loss": 3.0289,
"step": 6
},
{
"epoch": 0.015521064301552107,
"grad_norm": 30.871747970581055,
"learning_rate": 7.028e-05,
"loss": 2.7663,
"step": 7
},
{
"epoch": 0.017738359201773836,
"grad_norm": 21.825246810913086,
"learning_rate": 8.032e-05,
"loss": 2.2641,
"step": 8
},
{
"epoch": 0.019955654101995565,
"grad_norm": 26.116636276245117,
"learning_rate": 9.036000000000001e-05,
"loss": 2.573,
"step": 9
},
{
"epoch": 0.022172949002217297,
"grad_norm": 22.22761344909668,
"learning_rate": 0.0001004,
"loss": 2.4794,
"step": 10
},
{
"epoch": 0.024390243902439025,
"grad_norm": 26.01641082763672,
"learning_rate": 9.987157894736842e-05,
"loss": 2.447,
"step": 11
},
{
"epoch": 0.026607538802660754,
"grad_norm": 21.385372161865234,
"learning_rate": 9.934315789473684e-05,
"loss": 2.0634,
"step": 12
},
{
"epoch": 0.028824833702882482,
"grad_norm": 34.55962371826172,
"learning_rate": 9.881473684210525e-05,
"loss": 2.431,
"step": 13
},
{
"epoch": 0.031042128603104215,
"grad_norm": 29.898178100585938,
"learning_rate": 9.828631578947369e-05,
"loss": 2.6574,
"step": 14
},
{
"epoch": 0.03325942350332594,
"grad_norm": 27.43971824645996,
"learning_rate": 9.77578947368421e-05,
"loss": 1.9155,
"step": 15
},
{
"epoch": 0.03547671840354767,
"grad_norm": 27.087106704711914,
"learning_rate": 9.722947368421052e-05,
"loss": 2.1793,
"step": 16
},
{
"epoch": 0.037694013303769404,
"grad_norm": 30.93043327331543,
"learning_rate": 9.670105263157895e-05,
"loss": 2.4952,
"step": 17
},
{
"epoch": 0.03991130820399113,
"grad_norm": 27.66971206665039,
"learning_rate": 9.617263157894737e-05,
"loss": 2.2107,
"step": 18
},
{
"epoch": 0.04212860310421286,
"grad_norm": 37.52210235595703,
"learning_rate": 9.564421052631579e-05,
"loss": 2.235,
"step": 19
},
{
"epoch": 0.04434589800443459,
"grad_norm": 23.63651466369629,
"learning_rate": 9.511578947368421e-05,
"loss": 2.1324,
"step": 20
},
{
"epoch": 0.04656319290465632,
"grad_norm": 24.720478057861328,
"learning_rate": 9.458736842105264e-05,
"loss": 2.5987,
"step": 21
},
{
"epoch": 0.04878048780487805,
"grad_norm": 23.012195587158203,
"learning_rate": 9.405894736842106e-05,
"loss": 2.3515,
"step": 22
},
{
"epoch": 0.050997782705099776,
"grad_norm": 24.576961517333984,
"learning_rate": 9.353052631578947e-05,
"loss": 2.2289,
"step": 23
},
{
"epoch": 0.05321507760532151,
"grad_norm": 24.782127380371094,
"learning_rate": 9.300210526315789e-05,
"loss": 2.2446,
"step": 24
},
{
"epoch": 0.05543237250554324,
"grad_norm": 29.085662841796875,
"learning_rate": 9.247368421052631e-05,
"loss": 2.4364,
"step": 25
},
{
"epoch": 0.057649667405764965,
"grad_norm": 26.606853485107422,
"learning_rate": 9.194526315789473e-05,
"loss": 2.2234,
"step": 26
},
{
"epoch": 0.0598669623059867,
"grad_norm": 30.282766342163086,
"learning_rate": 9.141684210526316e-05,
"loss": 2.5798,
"step": 27
},
{
"epoch": 0.06208425720620843,
"grad_norm": 29.041223526000977,
"learning_rate": 9.088842105263158e-05,
"loss": 2.4436,
"step": 28
},
{
"epoch": 0.06430155210643015,
"grad_norm": 32.80604553222656,
"learning_rate": 9.036000000000001e-05,
"loss": 2.8015,
"step": 29
},
{
"epoch": 0.06651884700665188,
"grad_norm": 28.190860748291016,
"learning_rate": 8.983157894736843e-05,
"loss": 2.359,
"step": 30
},
{
"epoch": 0.06873614190687362,
"grad_norm": 35.894405364990234,
"learning_rate": 8.930315789473684e-05,
"loss": 1.9933,
"step": 31
},
{
"epoch": 0.07095343680709534,
"grad_norm": 27.286930084228516,
"learning_rate": 8.877473684210526e-05,
"loss": 2.4784,
"step": 32
},
{
"epoch": 0.07317073170731707,
"grad_norm": 31.852373123168945,
"learning_rate": 8.824631578947368e-05,
"loss": 2.2698,
"step": 33
},
{
"epoch": 0.07538802660753881,
"grad_norm": 30.357845306396484,
"learning_rate": 8.771789473684211e-05,
"loss": 2.2334,
"step": 34
},
{
"epoch": 0.07760532150776053,
"grad_norm": 34.58502197265625,
"learning_rate": 8.718947368421053e-05,
"loss": 2.6314,
"step": 35
},
{
"epoch": 0.07982261640798226,
"grad_norm": 29.154176712036133,
"learning_rate": 8.666105263157895e-05,
"loss": 2.1524,
"step": 36
},
{
"epoch": 0.082039911308204,
"grad_norm": 28.38983917236328,
"learning_rate": 8.613263157894737e-05,
"loss": 1.7586,
"step": 37
},
{
"epoch": 0.08425720620842572,
"grad_norm": 22.648962020874023,
"learning_rate": 8.560421052631578e-05,
"loss": 1.4079,
"step": 38
},
{
"epoch": 0.08647450110864745,
"grad_norm": 43.15308380126953,
"learning_rate": 8.50757894736842e-05,
"loss": 2.1336,
"step": 39
},
{
"epoch": 0.08869179600886919,
"grad_norm": 41.409969329833984,
"learning_rate": 8.454736842105263e-05,
"loss": 2.5629,
"step": 40
},
{
"epoch": 0.09090909090909091,
"grad_norm": 32.42283248901367,
"learning_rate": 8.401894736842106e-05,
"loss": 2.0355,
"step": 41
},
{
"epoch": 0.09312638580931264,
"grad_norm": 64.17134094238281,
"learning_rate": 8.349052631578948e-05,
"loss": 2.8994,
"step": 42
},
{
"epoch": 0.09534368070953436,
"grad_norm": 51.92913055419922,
"learning_rate": 8.29621052631579e-05,
"loss": 1.9425,
"step": 43
},
{
"epoch": 0.0975609756097561,
"grad_norm": 60.37343215942383,
"learning_rate": 8.243368421052632e-05,
"loss": 2.5374,
"step": 44
},
{
"epoch": 0.09977827050997783,
"grad_norm": 56.61979675292969,
"learning_rate": 8.190526315789474e-05,
"loss": 2.7696,
"step": 45
},
{
"epoch": 0.10199556541019955,
"grad_norm": 45.54286575317383,
"learning_rate": 8.137684210526315e-05,
"loss": 2.2621,
"step": 46
},
{
"epoch": 0.10421286031042129,
"grad_norm": 48.23731231689453,
"learning_rate": 8.084842105263157e-05,
"loss": 3.2684,
"step": 47
},
{
"epoch": 0.10643015521064302,
"grad_norm": 52.1467399597168,
"learning_rate": 8.032e-05,
"loss": 3.0954,
"step": 48
},
{
"epoch": 0.10864745011086474,
"grad_norm": 57.875450134277344,
"learning_rate": 7.979157894736842e-05,
"loss": 2.946,
"step": 49
},
{
"epoch": 0.11086474501108648,
"grad_norm": 88.02739715576172,
"learning_rate": 7.926315789473684e-05,
"loss": 4.3883,
"step": 50
},
{
"epoch": 0.11086474501108648,
"eval_loss": 1.270225167274475,
"eval_runtime": 20.2194,
"eval_samples_per_second": 9.397,
"eval_steps_per_second": 2.374,
"step": 50
},
{
"epoch": 0.1130820399113082,
"grad_norm": 26.53325653076172,
"learning_rate": 7.873473684210526e-05,
"loss": 3.0032,
"step": 51
},
{
"epoch": 0.11529933481152993,
"grad_norm": 17.52619743347168,
"learning_rate": 7.820631578947369e-05,
"loss": 2.7827,
"step": 52
},
{
"epoch": 0.11751662971175167,
"grad_norm": 14.779261589050293,
"learning_rate": 7.76778947368421e-05,
"loss": 2.2784,
"step": 53
},
{
"epoch": 0.1197339246119734,
"grad_norm": 12.997217178344727,
"learning_rate": 7.714947368421052e-05,
"loss": 2.2414,
"step": 54
},
{
"epoch": 0.12195121951219512,
"grad_norm": 14.368338584899902,
"learning_rate": 7.662105263157896e-05,
"loss": 2.2074,
"step": 55
},
{
"epoch": 0.12416851441241686,
"grad_norm": 15.52922534942627,
"learning_rate": 7.609263157894737e-05,
"loss": 2.1736,
"step": 56
},
{
"epoch": 0.12638580931263857,
"grad_norm": 18.17425537109375,
"learning_rate": 7.556421052631579e-05,
"loss": 2.3588,
"step": 57
},
{
"epoch": 0.1286031042128603,
"grad_norm": 16.153467178344727,
"learning_rate": 7.503578947368421e-05,
"loss": 2.4205,
"step": 58
},
{
"epoch": 0.13082039911308205,
"grad_norm": 15.452371597290039,
"learning_rate": 7.450736842105263e-05,
"loss": 2.2599,
"step": 59
},
{
"epoch": 0.13303769401330376,
"grad_norm": 18.882034301757812,
"learning_rate": 7.397894736842105e-05,
"loss": 1.9239,
"step": 60
},
{
"epoch": 0.1352549889135255,
"grad_norm": 15.408071517944336,
"learning_rate": 7.345052631578948e-05,
"loss": 2.0407,
"step": 61
},
{
"epoch": 0.13747228381374724,
"grad_norm": 18.57291603088379,
"learning_rate": 7.29221052631579e-05,
"loss": 2.1355,
"step": 62
},
{
"epoch": 0.13968957871396895,
"grad_norm": 14.49044418334961,
"learning_rate": 7.239368421052631e-05,
"loss": 1.6433,
"step": 63
},
{
"epoch": 0.1419068736141907,
"grad_norm": 21.082347869873047,
"learning_rate": 7.186526315789474e-05,
"loss": 2.5801,
"step": 64
},
{
"epoch": 0.14412416851441243,
"grad_norm": 18.55680274963379,
"learning_rate": 7.133684210526316e-05,
"loss": 1.7756,
"step": 65
},
{
"epoch": 0.14634146341463414,
"grad_norm": 17.898927688598633,
"learning_rate": 7.080842105263158e-05,
"loss": 1.6243,
"step": 66
},
{
"epoch": 0.14855875831485588,
"grad_norm": 17.05344581604004,
"learning_rate": 7.028e-05,
"loss": 1.7034,
"step": 67
},
{
"epoch": 0.15077605321507762,
"grad_norm": 19.20818328857422,
"learning_rate": 6.975157894736843e-05,
"loss": 1.8613,
"step": 68
},
{
"epoch": 0.15299334811529933,
"grad_norm": 14.604522705078125,
"learning_rate": 6.922315789473685e-05,
"loss": 1.3704,
"step": 69
},
{
"epoch": 0.15521064301552107,
"grad_norm": 18.43328285217285,
"learning_rate": 6.869473684210527e-05,
"loss": 1.9582,
"step": 70
},
{
"epoch": 0.1574279379157428,
"grad_norm": 19.990154266357422,
"learning_rate": 6.816631578947368e-05,
"loss": 1.7685,
"step": 71
},
{
"epoch": 0.15964523281596452,
"grad_norm": 17.83521842956543,
"learning_rate": 6.76378947368421e-05,
"loss": 2.1192,
"step": 72
},
{
"epoch": 0.16186252771618626,
"grad_norm": 21.896644592285156,
"learning_rate": 6.710947368421052e-05,
"loss": 1.6801,
"step": 73
},
{
"epoch": 0.164079822616408,
"grad_norm": 17.964820861816406,
"learning_rate": 6.658105263157894e-05,
"loss": 1.7591,
"step": 74
},
{
"epoch": 0.1662971175166297,
"grad_norm": 22.256559371948242,
"learning_rate": 6.605263157894737e-05,
"loss": 2.5407,
"step": 75
},
{
"epoch": 0.16851441241685144,
"grad_norm": 19.16011619567871,
"learning_rate": 6.55242105263158e-05,
"loss": 1.7553,
"step": 76
},
{
"epoch": 0.17073170731707318,
"grad_norm": 24.629257202148438,
"learning_rate": 6.499578947368422e-05,
"loss": 2.3935,
"step": 77
},
{
"epoch": 0.1729490022172949,
"grad_norm": 19.385765075683594,
"learning_rate": 6.446736842105264e-05,
"loss": 1.8258,
"step": 78
},
{
"epoch": 0.17516629711751663,
"grad_norm": 20.863401412963867,
"learning_rate": 6.393894736842105e-05,
"loss": 1.9962,
"step": 79
},
{
"epoch": 0.17738359201773837,
"grad_norm": 24.57994842529297,
"learning_rate": 6.341052631578947e-05,
"loss": 2.5787,
"step": 80
},
{
"epoch": 0.17960088691796008,
"grad_norm": 23.363492965698242,
"learning_rate": 6.288210526315789e-05,
"loss": 1.577,
"step": 81
},
{
"epoch": 0.18181818181818182,
"grad_norm": 22.762670516967773,
"learning_rate": 6.235368421052632e-05,
"loss": 2.2144,
"step": 82
},
{
"epoch": 0.18403547671840353,
"grad_norm": 19.128597259521484,
"learning_rate": 6.182526315789474e-05,
"loss": 1.6379,
"step": 83
},
{
"epoch": 0.18625277161862527,
"grad_norm": 20.36613655090332,
"learning_rate": 6.129684210526316e-05,
"loss": 2.1041,
"step": 84
},
{
"epoch": 0.188470066518847,
"grad_norm": 30.103256225585938,
"learning_rate": 6.076842105263158e-05,
"loss": 1.8926,
"step": 85
},
{
"epoch": 0.19068736141906872,
"grad_norm": 25.986953735351562,
"learning_rate": 6.024e-05,
"loss": 1.9419,
"step": 86
},
{
"epoch": 0.19290465631929046,
"grad_norm": 29.100027084350586,
"learning_rate": 5.971157894736842e-05,
"loss": 2.3349,
"step": 87
},
{
"epoch": 0.1951219512195122,
"grad_norm": 27.938316345214844,
"learning_rate": 5.9183157894736835e-05,
"loss": 1.8313,
"step": 88
},
{
"epoch": 0.1973392461197339,
"grad_norm": 36.821903228759766,
"learning_rate": 5.8654736842105267e-05,
"loss": 2.5949,
"step": 89
},
{
"epoch": 0.19955654101995565,
"grad_norm": 28.251705169677734,
"learning_rate": 5.8126315789473684e-05,
"loss": 2.038,
"step": 90
},
{
"epoch": 0.2017738359201774,
"grad_norm": 20.87628936767578,
"learning_rate": 5.759789473684211e-05,
"loss": 1.4964,
"step": 91
},
{
"epoch": 0.2039911308203991,
"grad_norm": 36.82621383666992,
"learning_rate": 5.706947368421053e-05,
"loss": 2.1786,
"step": 92
},
{
"epoch": 0.20620842572062084,
"grad_norm": 34.26662826538086,
"learning_rate": 5.6541052631578945e-05,
"loss": 2.3302,
"step": 93
},
{
"epoch": 0.20842572062084258,
"grad_norm": 29.61810874938965,
"learning_rate": 5.601263157894736e-05,
"loss": 2.0587,
"step": 94
},
{
"epoch": 0.2106430155210643,
"grad_norm": 31.78813934326172,
"learning_rate": 5.5484210526315794e-05,
"loss": 2.5443,
"step": 95
},
{
"epoch": 0.21286031042128603,
"grad_norm": 37.09660339355469,
"learning_rate": 5.495578947368421e-05,
"loss": 1.9665,
"step": 96
},
{
"epoch": 0.21507760532150777,
"grad_norm": 24.321231842041016,
"learning_rate": 5.442736842105264e-05,
"loss": 2.1167,
"step": 97
},
{
"epoch": 0.21729490022172948,
"grad_norm": 28.0638484954834,
"learning_rate": 5.3898947368421055e-05,
"loss": 1.2507,
"step": 98
},
{
"epoch": 0.21951219512195122,
"grad_norm": 43.129005432128906,
"learning_rate": 5.337052631578947e-05,
"loss": 3.1423,
"step": 99
},
{
"epoch": 0.22172949002217296,
"grad_norm": 90.48893737792969,
"learning_rate": 5.284210526315789e-05,
"loss": 2.6743,
"step": 100
},
{
"epoch": 0.22172949002217296,
"eval_loss": 1.0482261180877686,
"eval_runtime": 20.243,
"eval_samples_per_second": 9.386,
"eval_steps_per_second": 2.371,
"step": 100
},
{
"epoch": 0.22394678492239467,
"grad_norm": 17.120187759399414,
"learning_rate": 5.231368421052631e-05,
"loss": 2.3941,
"step": 101
},
{
"epoch": 0.2261640798226164,
"grad_norm": 16.493423461914062,
"learning_rate": 5.178526315789474e-05,
"loss": 2.3262,
"step": 102
},
{
"epoch": 0.22838137472283815,
"grad_norm": 12.841573715209961,
"learning_rate": 5.1256842105263165e-05,
"loss": 2.0664,
"step": 103
},
{
"epoch": 0.23059866962305986,
"grad_norm": 12.463582992553711,
"learning_rate": 5.072842105263158e-05,
"loss": 2.0229,
"step": 104
},
{
"epoch": 0.2328159645232816,
"grad_norm": 13.76339054107666,
"learning_rate": 5.02e-05,
"loss": 2.0176,
"step": 105
},
{
"epoch": 0.23503325942350334,
"grad_norm": 12.230353355407715,
"learning_rate": 4.967157894736842e-05,
"loss": 1.6626,
"step": 106
},
{
"epoch": 0.23725055432372505,
"grad_norm": 14.192876815795898,
"learning_rate": 4.914315789473684e-05,
"loss": 2.0265,
"step": 107
},
{
"epoch": 0.2394678492239468,
"grad_norm": 13.952037811279297,
"learning_rate": 4.861473684210526e-05,
"loss": 1.7241,
"step": 108
},
{
"epoch": 0.24168514412416853,
"grad_norm": 14.577288627624512,
"learning_rate": 4.8086315789473686e-05,
"loss": 1.7365,
"step": 109
},
{
"epoch": 0.24390243902439024,
"grad_norm": 19.91164779663086,
"learning_rate": 4.7557894736842104e-05,
"loss": 1.8842,
"step": 110
},
{
"epoch": 0.24611973392461198,
"grad_norm": 14.9652681350708,
"learning_rate": 4.702947368421053e-05,
"loss": 1.7975,
"step": 111
},
{
"epoch": 0.24833702882483372,
"grad_norm": 13.924230575561523,
"learning_rate": 4.6501052631578946e-05,
"loss": 1.9894,
"step": 112
},
{
"epoch": 0.25055432372505543,
"grad_norm": 13.089156150817871,
"learning_rate": 4.5972631578947364e-05,
"loss": 1.758,
"step": 113
},
{
"epoch": 0.25277161862527714,
"grad_norm": 12.60175895690918,
"learning_rate": 4.544421052631579e-05,
"loss": 1.6666,
"step": 114
},
{
"epoch": 0.2549889135254989,
"grad_norm": 17.490257263183594,
"learning_rate": 4.4915789473684213e-05,
"loss": 1.9925,
"step": 115
},
{
"epoch": 0.2572062084257206,
"grad_norm": 12.952994346618652,
"learning_rate": 4.438736842105263e-05,
"loss": 1.6062,
"step": 116
},
{
"epoch": 0.25942350332594233,
"grad_norm": 15.763453483581543,
"learning_rate": 4.3858947368421056e-05,
"loss": 2.0694,
"step": 117
},
{
"epoch": 0.2616407982261641,
"grad_norm": 14.717920303344727,
"learning_rate": 4.3330526315789474e-05,
"loss": 1.7913,
"step": 118
},
{
"epoch": 0.2638580931263858,
"grad_norm": 16.53207778930664,
"learning_rate": 4.280210526315789e-05,
"loss": 1.7959,
"step": 119
},
{
"epoch": 0.2660753880266075,
"grad_norm": 13.253047943115234,
"learning_rate": 4.2273684210526317e-05,
"loss": 1.6143,
"step": 120
},
{
"epoch": 0.2682926829268293,
"grad_norm": 15.786408424377441,
"learning_rate": 4.174526315789474e-05,
"loss": 1.6721,
"step": 121
},
{
"epoch": 0.270509977827051,
"grad_norm": 12.185956001281738,
"learning_rate": 4.121684210526316e-05,
"loss": 1.5245,
"step": 122
},
{
"epoch": 0.2727272727272727,
"grad_norm": 21.057693481445312,
"learning_rate": 4.068842105263158e-05,
"loss": 2.1359,
"step": 123
},
{
"epoch": 0.2749445676274945,
"grad_norm": 13.455188751220703,
"learning_rate": 4.016e-05,
"loss": 2.0307,
"step": 124
},
{
"epoch": 0.2771618625277162,
"grad_norm": 15.558540344238281,
"learning_rate": 3.963157894736842e-05,
"loss": 1.7351,
"step": 125
},
{
"epoch": 0.2793791574279379,
"grad_norm": 16.883926391601562,
"learning_rate": 3.9103157894736844e-05,
"loss": 2.1596,
"step": 126
},
{
"epoch": 0.28159645232815966,
"grad_norm": 20.178613662719727,
"learning_rate": 3.857473684210526e-05,
"loss": 1.8029,
"step": 127
},
{
"epoch": 0.2838137472283814,
"grad_norm": 18.442106246948242,
"learning_rate": 3.804631578947369e-05,
"loss": 2.0837,
"step": 128
},
{
"epoch": 0.2860310421286031,
"grad_norm": 19.457265853881836,
"learning_rate": 3.7517894736842105e-05,
"loss": 1.6284,
"step": 129
},
{
"epoch": 0.28824833702882485,
"grad_norm": 26.571035385131836,
"learning_rate": 3.698947368421052e-05,
"loss": 1.7662,
"step": 130
},
{
"epoch": 0.29046563192904656,
"grad_norm": 19.368440628051758,
"learning_rate": 3.646105263157895e-05,
"loss": 1.8661,
"step": 131
},
{
"epoch": 0.2926829268292683,
"grad_norm": 22.966577529907227,
"learning_rate": 3.593263157894737e-05,
"loss": 2.619,
"step": 132
},
{
"epoch": 0.29490022172949004,
"grad_norm": 24.9805965423584,
"learning_rate": 3.540421052631579e-05,
"loss": 1.5081,
"step": 133
},
{
"epoch": 0.29711751662971175,
"grad_norm": 19.5284366607666,
"learning_rate": 3.4875789473684215e-05,
"loss": 1.7447,
"step": 134
},
{
"epoch": 0.29933481152993346,
"grad_norm": 21.54673194885254,
"learning_rate": 3.434736842105263e-05,
"loss": 1.6988,
"step": 135
},
{
"epoch": 0.30155210643015523,
"grad_norm": 23.316730499267578,
"learning_rate": 3.381894736842105e-05,
"loss": 2.2736,
"step": 136
},
{
"epoch": 0.30376940133037694,
"grad_norm": 22.004526138305664,
"learning_rate": 3.329052631578947e-05,
"loss": 1.9549,
"step": 137
},
{
"epoch": 0.30598669623059865,
"grad_norm": 23.052682876586914,
"learning_rate": 3.27621052631579e-05,
"loss": 2.0963,
"step": 138
},
{
"epoch": 0.3082039911308204,
"grad_norm": 20.599315643310547,
"learning_rate": 3.223368421052632e-05,
"loss": 1.2949,
"step": 139
},
{
"epoch": 0.31042128603104213,
"grad_norm": 20.93349266052246,
"learning_rate": 3.1705263157894736e-05,
"loss": 1.9277,
"step": 140
},
{
"epoch": 0.31263858093126384,
"grad_norm": 27.307830810546875,
"learning_rate": 3.117684210526316e-05,
"loss": 1.122,
"step": 141
},
{
"epoch": 0.3148558758314856,
"grad_norm": 17.325700759887695,
"learning_rate": 3.064842105263158e-05,
"loss": 1.2689,
"step": 142
},
{
"epoch": 0.3170731707317073,
"grad_norm": 30.8470458984375,
"learning_rate": 3.012e-05,
"loss": 2.1764,
"step": 143
},
{
"epoch": 0.31929046563192903,
"grad_norm": 19.84284210205078,
"learning_rate": 2.9591578947368418e-05,
"loss": 1.059,
"step": 144
},
{
"epoch": 0.3215077605321508,
"grad_norm": 23.532106399536133,
"learning_rate": 2.9063157894736842e-05,
"loss": 1.9666,
"step": 145
},
{
"epoch": 0.3237250554323725,
"grad_norm": 17.46700668334961,
"learning_rate": 2.8534736842105264e-05,
"loss": 1.5324,
"step": 146
},
{
"epoch": 0.3259423503325942,
"grad_norm": 23.52845001220703,
"learning_rate": 2.800631578947368e-05,
"loss": 1.3784,
"step": 147
},
{
"epoch": 0.328159645232816,
"grad_norm": 27.91365623474121,
"learning_rate": 2.7477894736842106e-05,
"loss": 1.6899,
"step": 148
},
{
"epoch": 0.3303769401330377,
"grad_norm": 32.38272476196289,
"learning_rate": 2.6949473684210527e-05,
"loss": 1.6625,
"step": 149
},
{
"epoch": 0.3325942350332594,
"grad_norm": 51.69731140136719,
"learning_rate": 2.6421052631578945e-05,
"loss": 1.9981,
"step": 150
},
{
"epoch": 0.3325942350332594,
"eval_loss": 0.8841075897216797,
"eval_runtime": 20.2428,
"eval_samples_per_second": 9.386,
"eval_steps_per_second": 2.371,
"step": 150
},
{
"epoch": 0.3348115299334812,
"grad_norm": 9.736676216125488,
"learning_rate": 2.589263157894737e-05,
"loss": 1.6967,
"step": 151
},
{
"epoch": 0.3370288248337029,
"grad_norm": 10.335311889648438,
"learning_rate": 2.536421052631579e-05,
"loss": 1.6221,
"step": 152
},
{
"epoch": 0.3392461197339246,
"grad_norm": 10.948195457458496,
"learning_rate": 2.483578947368421e-05,
"loss": 1.891,
"step": 153
},
{
"epoch": 0.34146341463414637,
"grad_norm": 10.740894317626953,
"learning_rate": 2.430736842105263e-05,
"loss": 1.5576,
"step": 154
},
{
"epoch": 0.3436807095343681,
"grad_norm": 10.873139381408691,
"learning_rate": 2.3778947368421052e-05,
"loss": 1.5878,
"step": 155
},
{
"epoch": 0.3458980044345898,
"grad_norm": 14.580072402954102,
"learning_rate": 2.3250526315789473e-05,
"loss": 1.9496,
"step": 156
},
{
"epoch": 0.34811529933481156,
"grad_norm": 12.087544441223145,
"learning_rate": 2.2722105263157894e-05,
"loss": 1.8294,
"step": 157
},
{
"epoch": 0.35033259423503327,
"grad_norm": 14.1340913772583,
"learning_rate": 2.2193684210526316e-05,
"loss": 1.8763,
"step": 158
},
{
"epoch": 0.352549889135255,
"grad_norm": 11.832399368286133,
"learning_rate": 2.1665263157894737e-05,
"loss": 1.771,
"step": 159
},
{
"epoch": 0.35476718403547675,
"grad_norm": 14.125495910644531,
"learning_rate": 2.1136842105263158e-05,
"loss": 2.1433,
"step": 160
},
{
"epoch": 0.35698447893569846,
"grad_norm": 16.098318099975586,
"learning_rate": 2.060842105263158e-05,
"loss": 1.5098,
"step": 161
},
{
"epoch": 0.35920177383592017,
"grad_norm": 11.606100082397461,
"learning_rate": 2.008e-05,
"loss": 1.531,
"step": 162
},
{
"epoch": 0.3614190687361419,
"grad_norm": 14.126666069030762,
"learning_rate": 1.9551578947368422e-05,
"loss": 2.1956,
"step": 163
},
{
"epoch": 0.36363636363636365,
"grad_norm": 13.657944679260254,
"learning_rate": 1.9023157894736843e-05,
"loss": 2.0216,
"step": 164
},
{
"epoch": 0.36585365853658536,
"grad_norm": 16.890853881835938,
"learning_rate": 1.849473684210526e-05,
"loss": 2.0006,
"step": 165
},
{
"epoch": 0.36807095343680707,
"grad_norm": 12.908790588378906,
"learning_rate": 1.7966315789473686e-05,
"loss": 1.6173,
"step": 166
},
{
"epoch": 0.37028824833702884,
"grad_norm": 17.241039276123047,
"learning_rate": 1.7437894736842107e-05,
"loss": 1.9784,
"step": 167
},
{
"epoch": 0.37250554323725055,
"grad_norm": 15.207599639892578,
"learning_rate": 1.6909473684210525e-05,
"loss": 1.5988,
"step": 168
},
{
"epoch": 0.37472283813747226,
"grad_norm": 15.888737678527832,
"learning_rate": 1.638105263157895e-05,
"loss": 1.4648,
"step": 169
},
{
"epoch": 0.376940133037694,
"grad_norm": 12.339811325073242,
"learning_rate": 1.5852631578947368e-05,
"loss": 1.5447,
"step": 170
},
{
"epoch": 0.37915742793791574,
"grad_norm": 12.044404983520508,
"learning_rate": 1.532421052631579e-05,
"loss": 1.3339,
"step": 171
},
{
"epoch": 0.38137472283813745,
"grad_norm": 11.459868431091309,
"learning_rate": 1.4795789473684209e-05,
"loss": 1.4563,
"step": 172
},
{
"epoch": 0.3835920177383592,
"grad_norm": 14.248141288757324,
"learning_rate": 1.4267368421052632e-05,
"loss": 1.5893,
"step": 173
},
{
"epoch": 0.3858093126385809,
"grad_norm": 19.118242263793945,
"learning_rate": 1.3738947368421053e-05,
"loss": 1.9648,
"step": 174
},
{
"epoch": 0.38802660753880264,
"grad_norm": 15.491469383239746,
"learning_rate": 1.3210526315789473e-05,
"loss": 1.5081,
"step": 175
},
{
"epoch": 0.3902439024390244,
"grad_norm": 13.472190856933594,
"learning_rate": 1.2682105263157896e-05,
"loss": 1.2014,
"step": 176
},
{
"epoch": 0.3924611973392461,
"grad_norm": 16.698389053344727,
"learning_rate": 1.2153684210526315e-05,
"loss": 1.6185,
"step": 177
},
{
"epoch": 0.3946784922394678,
"grad_norm": 14.444684028625488,
"learning_rate": 1.1625263157894737e-05,
"loss": 1.663,
"step": 178
},
{
"epoch": 0.3968957871396896,
"grad_norm": 12.130651473999023,
"learning_rate": 1.1096842105263158e-05,
"loss": 1.1374,
"step": 179
},
{
"epoch": 0.3991130820399113,
"grad_norm": 13.900895118713379,
"learning_rate": 1.0568421052631579e-05,
"loss": 1.3243,
"step": 180
},
{
"epoch": 0.401330376940133,
"grad_norm": 15.653491020202637,
"learning_rate": 1.004e-05,
"loss": 1.7645,
"step": 181
},
{
"epoch": 0.4035476718403548,
"grad_norm": 15.8237886428833,
"learning_rate": 9.511578947368422e-06,
"loss": 1.6201,
"step": 182
},
{
"epoch": 0.4057649667405765,
"grad_norm": 32.045345306396484,
"learning_rate": 8.983157894736843e-06,
"loss": 1.7167,
"step": 183
},
{
"epoch": 0.4079822616407982,
"grad_norm": 13.554454803466797,
"learning_rate": 8.454736842105263e-06,
"loss": 1.1857,
"step": 184
},
{
"epoch": 0.41019955654101997,
"grad_norm": 13.749832153320312,
"learning_rate": 7.926315789473684e-06,
"loss": 1.0891,
"step": 185
},
{
"epoch": 0.4124168514412417,
"grad_norm": 16.378047943115234,
"learning_rate": 7.397894736842104e-06,
"loss": 1.244,
"step": 186
},
{
"epoch": 0.4146341463414634,
"grad_norm": 20.462013244628906,
"learning_rate": 6.8694736842105265e-06,
"loss": 1.659,
"step": 187
},
{
"epoch": 0.41685144124168516,
"grad_norm": 18.432472229003906,
"learning_rate": 6.341052631578948e-06,
"loss": 1.8659,
"step": 188
},
{
"epoch": 0.4190687361419069,
"grad_norm": 21.487030029296875,
"learning_rate": 5.812631578947368e-06,
"loss": 1.7793,
"step": 189
},
{
"epoch": 0.4212860310421286,
"grad_norm": 27.553062438964844,
"learning_rate": 5.2842105263157896e-06,
"loss": 2.1433,
"step": 190
},
{
"epoch": 0.42350332594235035,
"grad_norm": 20.972902297973633,
"learning_rate": 4.755789473684211e-06,
"loss": 1.3501,
"step": 191
},
{
"epoch": 0.42572062084257206,
"grad_norm": 21.430740356445312,
"learning_rate": 4.227368421052631e-06,
"loss": 1.2867,
"step": 192
},
{
"epoch": 0.4279379157427938,
"grad_norm": 21.622007369995117,
"learning_rate": 3.698947368421052e-06,
"loss": 1.8264,
"step": 193
},
{
"epoch": 0.43015521064301554,
"grad_norm": 18.00780487060547,
"learning_rate": 3.170526315789474e-06,
"loss": 1.5396,
"step": 194
},
{
"epoch": 0.43237250554323725,
"grad_norm": 21.456939697265625,
"learning_rate": 2.6421052631578948e-06,
"loss": 0.8305,
"step": 195
},
{
"epoch": 0.43458980044345896,
"grad_norm": 37.9657096862793,
"learning_rate": 2.1136842105263157e-06,
"loss": 2.433,
"step": 196
},
{
"epoch": 0.43680709534368073,
"grad_norm": 31.422948837280273,
"learning_rate": 1.585263157894737e-06,
"loss": 1.5014,
"step": 197
},
{
"epoch": 0.43902439024390244,
"grad_norm": 22.746129989624023,
"learning_rate": 1.0568421052631578e-06,
"loss": 1.9673,
"step": 198
},
{
"epoch": 0.44124168514412415,
"grad_norm": 45.2837028503418,
"learning_rate": 5.284210526315789e-07,
"loss": 1.036,
"step": 199
},
{
"epoch": 0.4434589800443459,
"grad_norm": 56.693782806396484,
"learning_rate": 0.0,
"loss": 2.1921,
"step": 200
},
{
"epoch": 0.4434589800443459,
"eval_loss": 0.8263623714447021,
"eval_runtime": 20.2153,
"eval_samples_per_second": 9.399,
"eval_steps_per_second": 2.374,
"step": 200
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 7.15499809800192e+16,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}