|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 500, |
|
"global_step": 9013, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0022190169754798626, |
|
"grad_norm": 7.625, |
|
"learning_rate": 3.3277870216306157e-07, |
|
"loss": 1.2617, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.004438033950959725, |
|
"grad_norm": 5.5625, |
|
"learning_rate": 6.655574043261231e-07, |
|
"loss": 1.2121, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.006657050926439588, |
|
"grad_norm": 4.78125, |
|
"learning_rate": 9.983361064891848e-07, |
|
"loss": 1.2737, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.00887606790191945, |
|
"grad_norm": 5.46875, |
|
"learning_rate": 1.3311148086522463e-06, |
|
"loss": 1.1634, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.011095084877399313, |
|
"grad_norm": 6.46875, |
|
"learning_rate": 1.6638935108153078e-06, |
|
"loss": 1.1582, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.013314101852879175, |
|
"grad_norm": 6.09375, |
|
"learning_rate": 1.9966722129783695e-06, |
|
"loss": 1.1071, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.015533118828359036, |
|
"grad_norm": 5.5625, |
|
"learning_rate": 2.329450915141431e-06, |
|
"loss": 1.0944, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.0177521358038389, |
|
"grad_norm": 3.3125, |
|
"learning_rate": 2.6622296173044925e-06, |
|
"loss": 0.9622, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.01997115277931876, |
|
"grad_norm": 5.65625, |
|
"learning_rate": 2.995008319467554e-06, |
|
"loss": 0.7219, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.022190169754798626, |
|
"grad_norm": 3.234375, |
|
"learning_rate": 3.3277870216306156e-06, |
|
"loss": 0.6194, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.024409186730278486, |
|
"grad_norm": 3.359375, |
|
"learning_rate": 3.6605657237936775e-06, |
|
"loss": 0.5223, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.02662820370575835, |
|
"grad_norm": 2.234375, |
|
"learning_rate": 3.993344425956739e-06, |
|
"loss": 0.5673, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.02884722068123821, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 4.326123128119801e-06, |
|
"loss": 0.5215, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.031066237656718072, |
|
"grad_norm": 2.765625, |
|
"learning_rate": 4.658901830282862e-06, |
|
"loss": 0.5166, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.03328525463219793, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 4.991680532445923e-06, |
|
"loss": 0.5023, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.0355042716076778, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 5.324459234608985e-06, |
|
"loss": 0.4611, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.03772328858315766, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 5.657237936772047e-06, |
|
"loss": 0.5203, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.03994230555863752, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 5.990016638935108e-06, |
|
"loss": 0.5498, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.04216132253411738, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 6.32279534109817e-06, |
|
"loss": 0.5597, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.04438033950959725, |
|
"grad_norm": 2.453125, |
|
"learning_rate": 6.655574043261231e-06, |
|
"loss": 0.5133, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.04659935648507711, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 6.988352745424292e-06, |
|
"loss": 0.4843, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.04881837346055697, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 7.321131447587355e-06, |
|
"loss": 0.5358, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.051037390436036834, |
|
"grad_norm": 2.6875, |
|
"learning_rate": 7.653910149750416e-06, |
|
"loss": 0.3936, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.0532564074115167, |
|
"grad_norm": 2.125, |
|
"learning_rate": 7.986688851913478e-06, |
|
"loss": 0.5245, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.05547542438699656, |
|
"grad_norm": 2.796875, |
|
"learning_rate": 8.319467554076538e-06, |
|
"loss": 0.5094, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.05769444136247642, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 8.652246256239602e-06, |
|
"loss": 0.4476, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.059913458337956284, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 8.985024958402662e-06, |
|
"loss": 0.4836, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.062132475313436145, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 9.317803660565724e-06, |
|
"loss": 0.4805, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.06435149228891601, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 9.650582362728786e-06, |
|
"loss": 0.425, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.06657050926439587, |
|
"grad_norm": 2.6875, |
|
"learning_rate": 9.983361064891846e-06, |
|
"loss": 0.5223, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.06878952623987573, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 1.031613976705491e-05, |
|
"loss": 0.5, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.0710085432153556, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 1.064891846921797e-05, |
|
"loss": 0.5271, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.07322756019083546, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.0981697171381032e-05, |
|
"loss": 0.4981, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.07544657716631532, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 1.1314475873544094e-05, |
|
"loss": 0.5133, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.07766559414179519, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 1.1647254575707154e-05, |
|
"loss": 0.4693, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.07988461111727505, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 1.1980033277870216e-05, |
|
"loss": 0.4844, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.08210362809275491, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 1.2312811980033278e-05, |
|
"loss": 0.4943, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.08432264506823477, |
|
"grad_norm": 2.75, |
|
"learning_rate": 1.264559068219634e-05, |
|
"loss": 0.5, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.08654166204371463, |
|
"grad_norm": 1.46875, |
|
"learning_rate": 1.2978369384359402e-05, |
|
"loss": 0.4318, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.0887606790191945, |
|
"grad_norm": 2.484375, |
|
"learning_rate": 1.3311148086522462e-05, |
|
"loss": 0.4745, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.09097969599467436, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 1.3643926788685524e-05, |
|
"loss": 0.5459, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.09319871297015422, |
|
"grad_norm": 2.078125, |
|
"learning_rate": 1.3976705490848584e-05, |
|
"loss": 0.5021, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.09541772994563408, |
|
"grad_norm": 0.97265625, |
|
"learning_rate": 1.4309484193011648e-05, |
|
"loss": 0.4271, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.09763674692111395, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 1.464226289517471e-05, |
|
"loss": 0.4924, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.09985576389659381, |
|
"grad_norm": 2.25, |
|
"learning_rate": 1.497504159733777e-05, |
|
"loss": 0.514, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.10207478087207367, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 1.5307820299500832e-05, |
|
"loss": 0.4569, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.10429379784755353, |
|
"grad_norm": 1.4140625, |
|
"learning_rate": 1.5640599001663892e-05, |
|
"loss": 0.4751, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.1065128148230334, |
|
"grad_norm": 2.1875, |
|
"learning_rate": 1.5973377703826956e-05, |
|
"loss": 0.4719, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.10873183179851326, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 1.6306156405990016e-05, |
|
"loss": 0.4455, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.11095084877399312, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 1.6638935108153077e-05, |
|
"loss": 0.5068, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.11316986574947298, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 1.697171381031614e-05, |
|
"loss": 0.5076, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.11538888272495285, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 1.7304492512479204e-05, |
|
"loss": 0.449, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.11760789970043271, |
|
"grad_norm": 2.859375, |
|
"learning_rate": 1.7637271214642264e-05, |
|
"loss": 0.4298, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.11982691667591257, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 1.7970049916805324e-05, |
|
"loss": 0.5298, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.12204593365139244, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 1.8302828618968388e-05, |
|
"loss": 0.5398, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.12426495062687229, |
|
"grad_norm": 1.40625, |
|
"learning_rate": 1.8635607321131448e-05, |
|
"loss": 0.5533, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.12648396760235217, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 1.896838602329451e-05, |
|
"loss": 0.422, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.12870298457783202, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 1.9301164725457572e-05, |
|
"loss": 0.4807, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.13092200155331188, |
|
"grad_norm": 1.484375, |
|
"learning_rate": 1.9633943427620632e-05, |
|
"loss": 0.5209, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.13314101852879173, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 1.9966722129783693e-05, |
|
"loss": 0.4522, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.13536003550427161, |
|
"grad_norm": 1.3828125, |
|
"learning_rate": 2.0299500831946756e-05, |
|
"loss": 0.4684, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.13757905247975147, |
|
"grad_norm": 1.59375, |
|
"learning_rate": 2.063227953410982e-05, |
|
"loss": 0.3982, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.13979806945523132, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 2.096505823627288e-05, |
|
"loss": 0.5518, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.1420170864307112, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 2.129783693843594e-05, |
|
"loss": 0.5107, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.14423610340619106, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 2.1630615640599004e-05, |
|
"loss": 0.4422, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.1464551203816709, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 2.1963394342762064e-05, |
|
"loss": 0.5578, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.1486741373571508, |
|
"grad_norm": 1.4921875, |
|
"learning_rate": 2.2296173044925124e-05, |
|
"loss": 0.4516, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.15089315433263065, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.2628951747088188e-05, |
|
"loss": 0.3979, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.1531121713081105, |
|
"grad_norm": 1.4140625, |
|
"learning_rate": 2.296173044925125e-05, |
|
"loss": 0.4895, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.15533118828359038, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 2.329450915141431e-05, |
|
"loss": 0.4505, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.15755020525907024, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 2.3627287853577372e-05, |
|
"loss": 0.5008, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.1597692222345501, |
|
"grad_norm": 1.5390625, |
|
"learning_rate": 2.3960066555740432e-05, |
|
"loss": 0.523, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.16198823921002994, |
|
"grad_norm": 1.5546875, |
|
"learning_rate": 2.4292845257903493e-05, |
|
"loss": 0.4813, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.16420725618550983, |
|
"grad_norm": 1.46875, |
|
"learning_rate": 2.4625623960066556e-05, |
|
"loss": 0.4995, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.16642627316098968, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 2.495840266222962e-05, |
|
"loss": 0.474, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.16864529013646953, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 2.529118136439268e-05, |
|
"loss": 0.4201, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.17086430711194942, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.562396006655574e-05, |
|
"loss": 0.5332, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.17308332408742927, |
|
"grad_norm": 2.53125, |
|
"learning_rate": 2.5956738768718804e-05, |
|
"loss": 0.5105, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.17530234106290912, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 2.6289517470881864e-05, |
|
"loss": 0.4679, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.177521358038389, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 2.6622296173044925e-05, |
|
"loss": 0.4276, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.17974037501386886, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 2.6955074875207988e-05, |
|
"loss": 0.4462, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.1819593919893487, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 2.728785357737105e-05, |
|
"loss": 0.4534, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.1841784089648286, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 2.762063227953411e-05, |
|
"loss": 0.4699, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.18639742594030845, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.795341098169717e-05, |
|
"loss": 0.483, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.1886164429157883, |
|
"grad_norm": 2.765625, |
|
"learning_rate": 2.8286189683860236e-05, |
|
"loss": 0.4776, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.19083545989126816, |
|
"grad_norm": 2.0, |
|
"learning_rate": 2.8618968386023296e-05, |
|
"loss": 0.4361, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.19305447686674804, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 2.8951747088186356e-05, |
|
"loss": 0.4566, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.1952734938422279, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 2.928452579034942e-05, |
|
"loss": 0.4609, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.19749251081770774, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 2.961730449251248e-05, |
|
"loss": 0.5019, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.19971152779318763, |
|
"grad_norm": 2.125, |
|
"learning_rate": 2.995008319467554e-05, |
|
"loss": 0.4844, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.20193054476866748, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 2.999958848436878e-05, |
|
"loss": 0.4974, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.20414956174414733, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.9998050673796383e-05, |
|
"loss": 0.4591, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.20636857871962722, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 2.999537386964595e-05, |
|
"loss": 0.4319, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.20858759569510707, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 2.9991558275201416e-05, |
|
"loss": 0.4425, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.21080661267058692, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 2.998660418022961e-05, |
|
"loss": 0.4729, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.2130256296460668, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 2.9980511960958247e-05, |
|
"loss": 0.4049, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.21524464662154666, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 2.9973282080047365e-05, |
|
"loss": 0.4646, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.2174636635970265, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 2.996491508655417e-05, |
|
"loss": 0.4725, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.21968268057250637, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 2.995541161589137e-05, |
|
"loss": 0.4448, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.22190169754798625, |
|
"grad_norm": 1.3046875, |
|
"learning_rate": 2.994477238977889e-05, |
|
"loss": 0.4673, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.2241207145234661, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 2.9932998216189077e-05, |
|
"loss": 0.423, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.22633973149894596, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 2.992008998928534e-05, |
|
"loss": 0.4803, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.22855874847442584, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 2.990604868935424e-05, |
|
"loss": 0.4604, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.2307777654499057, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 2.989087538273105e-05, |
|
"loss": 0.5155, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.23299678242538555, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 2.9874571221718776e-05, |
|
"loss": 0.5122, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.23521579940086543, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 2.985713744450063e-05, |
|
"loss": 0.4475, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.23743481637634528, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.9838575375046034e-05, |
|
"loss": 0.4281, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.23965383335182514, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 2.9818886423010024e-05, |
|
"loss": 0.4383, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.24187285032730502, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 2.979807208362625e-05, |
|
"loss": 0.449, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.24409186730278487, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 2.9776133937593395e-05, |
|
"loss": 0.5002, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.24631088427826472, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 2.9753073650955128e-05, |
|
"loss": 0.4667, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.24852990125374458, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 2.9728892974973592e-05, |
|
"loss": 0.4431, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.25074891822922446, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 2.970359374599641e-05, |
|
"loss": 0.4684, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.25296793520470434, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.967717788531722e-05, |
|
"loss": 0.4599, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.25518695218018417, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 2.9649647399029764e-05, |
|
"loss": 0.5047, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.25740596915566405, |
|
"grad_norm": 2.265625, |
|
"learning_rate": 2.9621004377875558e-05, |
|
"loss": 0.4824, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.2596249861311439, |
|
"grad_norm": 1.3515625, |
|
"learning_rate": 2.959125099708509e-05, |
|
"loss": 0.4928, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.26184400310662376, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 2.9560389516212638e-05, |
|
"loss": 0.5386, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.26406302008210364, |
|
"grad_norm": 0.96875, |
|
"learning_rate": 2.9528422278964687e-05, |
|
"loss": 0.4092, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.26628203705758346, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 2.949535171302192e-05, |
|
"loss": 0.4929, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.26850105403306335, |
|
"grad_norm": 1.5859375, |
|
"learning_rate": 2.9461180329854867e-05, |
|
"loss": 0.4893, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.27072007100854323, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 2.9425910724533165e-05, |
|
"loss": 0.4554, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.27293908798402305, |
|
"grad_norm": 1.65625, |
|
"learning_rate": 2.9389545575528496e-05, |
|
"loss": 0.4702, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.27515810495950294, |
|
"grad_norm": 1.375, |
|
"learning_rate": 2.9352087644511162e-05, |
|
"loss": 0.5056, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.2773771219349828, |
|
"grad_norm": 2.859375, |
|
"learning_rate": 2.9313539776140362e-05, |
|
"loss": 0.428, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.27959613891046264, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 2.9273904897848174e-05, |
|
"loss": 0.4827, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.2818151558859425, |
|
"grad_norm": 1.359375, |
|
"learning_rate": 2.9233186019617214e-05, |
|
"loss": 0.5082, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.2840341728614224, |
|
"grad_norm": 1.421875, |
|
"learning_rate": 2.9191386233752062e-05, |
|
"loss": 0.4803, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.28625318983690223, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 2.9148508714644424e-05, |
|
"loss": 0.5205, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.2884722068123821, |
|
"grad_norm": 1.546875, |
|
"learning_rate": 2.9104556718532054e-05, |
|
"loss": 0.4965, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.290691223787862, |
|
"grad_norm": 1.5078125, |
|
"learning_rate": 2.9059533583251487e-05, |
|
"loss": 0.4376, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.2929102407633418, |
|
"grad_norm": 1.7578125, |
|
"learning_rate": 2.9013442727984517e-05, |
|
"loss": 0.4618, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.2951292577388217, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 2.8966287652998562e-05, |
|
"loss": 0.4943, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.2973482747143016, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 2.891807193938085e-05, |
|
"loss": 0.4582, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.2995672916897814, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 2.8868799248766436e-05, |
|
"loss": 0.5133, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.3017863086652613, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 2.8818473323060143e-05, |
|
"loss": 0.4628, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.3040053256407412, |
|
"grad_norm": 1.609375, |
|
"learning_rate": 2.87670979841524e-05, |
|
"loss": 0.4408, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.306224342616221, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 2.8714677133628963e-05, |
|
"loss": 0.5004, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.3084433595917009, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 2.866121475247467e-05, |
|
"loss": 0.4031, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.31066237656718076, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 2.8606714900771055e-05, |
|
"loss": 0.5192, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.3128813935426606, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 2.8551181717388066e-05, |
|
"loss": 0.5397, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.3151004105181405, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 2.849461941966972e-05, |
|
"loss": 0.4446, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.3173194274936203, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 2.8437032303113823e-05, |
|
"loss": 0.4464, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.3195384444691002, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 2.8378424741045773e-05, |
|
"loss": 0.526, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.32175746144458006, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 2.831880118428644e-05, |
|
"loss": 0.4087, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.3239764784200599, |
|
"grad_norm": 1.125, |
|
"learning_rate": 2.8258166160814135e-05, |
|
"loss": 0.4833, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.32619549539553977, |
|
"grad_norm": 1.34375, |
|
"learning_rate": 2.8196524275420758e-05, |
|
"loss": 0.4403, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.32841451237101965, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 2.813388020936211e-05, |
|
"loss": 0.4728, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.3306335293464995, |
|
"grad_norm": 1.0390625, |
|
"learning_rate": 2.8070238720002364e-05, |
|
"loss": 0.4389, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.33285254632197936, |
|
"grad_norm": 0.90625, |
|
"learning_rate": 2.800560464045278e-05, |
|
"loss": 0.482, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.33507156329745924, |
|
"grad_norm": 1.6796875, |
|
"learning_rate": 2.7939982879204686e-05, |
|
"loss": 0.4157, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.33729058027293907, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 2.787337841975668e-05, |
|
"loss": 0.4593, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.33950959724841895, |
|
"grad_norm": 1.2421875, |
|
"learning_rate": 2.78057963202362e-05, |
|
"loss": 0.4071, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.34172861422389883, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 2.773724171301538e-05, |
|
"loss": 0.4307, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.34394763119937866, |
|
"grad_norm": 2.75, |
|
"learning_rate": 2.7667719804321285e-05, |
|
"loss": 0.5196, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.34616664817485854, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 2.7597235873840544e-05, |
|
"loss": 0.4529, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.3483856651503384, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 2.7525795274318386e-05, |
|
"loss": 0.5671, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.35060468212581825, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 2.745340343115213e-05, |
|
"loss": 0.506, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.3528236991012981, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.7380065841979196e-05, |
|
"loss": 0.4834, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.355042716076778, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 2.7305788076259565e-05, |
|
"loss": 0.4635, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.35726173305225783, |
|
"grad_norm": 1.7109375, |
|
"learning_rate": 2.7230575774852843e-05, |
|
"loss": 0.4873, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.3594807500277377, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 2.715443464958986e-05, |
|
"loss": 0.4133, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.3616997670032176, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 2.707737048283891e-05, |
|
"loss": 0.4954, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.3639187839786974, |
|
"grad_norm": 1.3671875, |
|
"learning_rate": 2.699938912706663e-05, |
|
"loss": 0.4585, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.3661378009541773, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 2.6920496504393507e-05, |
|
"loss": 0.5479, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.3683568179296572, |
|
"grad_norm": 2.515625, |
|
"learning_rate": 2.6840698606144197e-05, |
|
"loss": 0.4712, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.370575834905137, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 2.6760001492392474e-05, |
|
"loss": 0.5046, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.3727948518806169, |
|
"grad_norm": 1.53125, |
|
"learning_rate": 2.6678411291501038e-05, |
|
"loss": 0.522, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.3750138688560967, |
|
"grad_norm": 1.6796875, |
|
"learning_rate": 2.6595934199656108e-05, |
|
"loss": 0.4852, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.3772328858315766, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 2.6512576480396862e-05, |
|
"loss": 0.4745, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.3794519028070565, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 2.6428344464139756e-05, |
|
"loss": 0.4903, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.3816709197825363, |
|
"grad_norm": 1.625, |
|
"learning_rate": 2.6343244547697798e-05, |
|
"loss": 0.3906, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.3838899367580162, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 2.6257283193794742e-05, |
|
"loss": 0.471, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.3861089537334961, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 2.617046693057429e-05, |
|
"loss": 0.4878, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.3883279707089759, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 2.6082802351104317e-05, |
|
"loss": 0.4718, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.3905469876844558, |
|
"grad_norm": 1.8828125, |
|
"learning_rate": 2.5994296112876222e-05, |
|
"loss": 0.4649, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.39276600465993566, |
|
"grad_norm": 1.2109375, |
|
"learning_rate": 2.5904954937299267e-05, |
|
"loss": 0.3973, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.3949850216354155, |
|
"grad_norm": 1.4609375, |
|
"learning_rate": 2.5814785609190197e-05, |
|
"loss": 0.4785, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.39720403861089537, |
|
"grad_norm": 2.171875, |
|
"learning_rate": 2.5723794976257947e-05, |
|
"loss": 0.4512, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.39942305558637525, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 2.5631989948583623e-05, |
|
"loss": 0.4931, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.4016420725618551, |
|
"grad_norm": 1.1796875, |
|
"learning_rate": 2.553937749809572e-05, |
|
"loss": 0.4183, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.40386108953733496, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 2.544596465804068e-05, |
|
"loss": 0.4321, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.40608010651281484, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 2.5351758522448724e-05, |
|
"loss": 0.489, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.40829912348829467, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 2.5256766245595166e-05, |
|
"loss": 0.4398, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.41051814046377455, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 2.516099504145703e-05, |
|
"loss": 0.4831, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.41273715743925443, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 2.5064452183165283e-05, |
|
"loss": 0.4365, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.41495617441473426, |
|
"grad_norm": 1.359375, |
|
"learning_rate": 2.496714500245241e-05, |
|
"loss": 0.4309, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.41717519139021414, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 2.4869080889095693e-05, |
|
"loss": 0.5378, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.419394208365694, |
|
"grad_norm": 1.75, |
|
"learning_rate": 2.477026729035595e-05, |
|
"loss": 0.4905, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.42161322534117385, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 2.4670711710412026e-05, |
|
"loss": 0.4187, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.42383224231665373, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 2.457042170979086e-05, |
|
"loss": 0.4817, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.4260512592921336, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 2.4469404904793338e-05, |
|
"loss": 0.5108, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.42827027626761344, |
|
"grad_norm": 1.46875, |
|
"learning_rate": 2.4367668966915885e-05, |
|
"loss": 0.5112, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.4304892932430933, |
|
"grad_norm": 3.046875, |
|
"learning_rate": 2.4265221622267876e-05, |
|
"loss": 0.5353, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.43270831021857314, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 2.4162070650984893e-05, |
|
"loss": 0.4684, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.434927327194053, |
|
"grad_norm": 1.5078125, |
|
"learning_rate": 2.4058223886637872e-05, |
|
"loss": 0.4374, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.4371463441695329, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 2.3953689215638194e-05, |
|
"loss": 0.3753, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.43936536114501273, |
|
"grad_norm": 1.4921875, |
|
"learning_rate": 2.3848474576638807e-05, |
|
"loss": 0.4427, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.4415843781204926, |
|
"grad_norm": 1.25, |
|
"learning_rate": 2.3742587959931285e-05, |
|
"loss": 0.5074, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.4438033950959725, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 2.3636037406839076e-05, |
|
"loss": 0.4841, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.4460224120714523, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 2.3528831009106786e-05, |
|
"loss": 0.4643, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.4482414290469322, |
|
"grad_norm": 1.25, |
|
"learning_rate": 2.3420976908285687e-05, |
|
"loss": 0.4976, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.4504604460224121, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 2.3312483295115424e-05, |
|
"loss": 0.5453, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.4526794629978919, |
|
"grad_norm": 2.921875, |
|
"learning_rate": 2.320335840890198e-05, |
|
"loss": 0.46, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.4548984799733718, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 2.3093610536891965e-05, |
|
"loss": 0.5129, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.4571174969488517, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 2.2983248013643253e-05, |
|
"loss": 0.4429, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.4593365139243315, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 2.2872279220392054e-05, |
|
"loss": 0.4817, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.4615555308998114, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 2.2760712584416386e-05, |
|
"loss": 0.51, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.46377454787529127, |
|
"grad_norm": 1.2734375, |
|
"learning_rate": 2.2648556578396107e-05, |
|
"loss": 0.5001, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.4659935648507711, |
|
"grad_norm": 2.34375, |
|
"learning_rate": 2.2535819719769487e-05, |
|
"loss": 0.4739, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.468212581826251, |
|
"grad_norm": 1.484375, |
|
"learning_rate": 2.242251057008633e-05, |
|
"loss": 0.474, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.47043159880173085, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 2.2308637734357826e-05, |
|
"loss": 0.4622, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.4726506157772107, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 2.219420986040305e-05, |
|
"loss": 0.5075, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.47486963275269056, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 2.2079235638192203e-05, |
|
"loss": 0.4817, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.47708864972817044, |
|
"grad_norm": 1.3671875, |
|
"learning_rate": 2.1963723799186706e-05, |
|
"loss": 0.445, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.47930766670365027, |
|
"grad_norm": 1.6796875, |
|
"learning_rate": 2.184768311567608e-05, |
|
"loss": 0.475, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.48152668367913015, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 2.1731122400111764e-05, |
|
"loss": 0.4418, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.48374570065461003, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 2.161405050443789e-05, |
|
"loss": 0.4722, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.48596471763008986, |
|
"grad_norm": 1.75, |
|
"learning_rate": 2.1496476319419002e-05, |
|
"loss": 0.4814, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.48818373460556974, |
|
"grad_norm": 1.59375, |
|
"learning_rate": 2.137840877396491e-05, |
|
"loss": 0.5435, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.49040275158104957, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 2.125985683445258e-05, |
|
"loss": 0.4521, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.49262176855652945, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 2.114082950404519e-05, |
|
"loss": 0.452, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.49484078553200933, |
|
"grad_norm": 1.2734375, |
|
"learning_rate": 2.1021335822008447e-05, |
|
"loss": 0.5176, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.49705980250748916, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 2.0901384863024078e-05, |
|
"loss": 0.4307, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.49927881948296904, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 2.0780985736500696e-05, |
|
"loss": 0.4856, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.5014978364584489, |
|
"grad_norm": 1.9609375, |
|
"learning_rate": 2.0660147585881994e-05, |
|
"loss": 0.4339, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.5037168534339288, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 2.0538879587952382e-05, |
|
"loss": 0.4902, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.5059358704094087, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 2.0417190952140064e-05, |
|
"loss": 0.4343, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.5081548873848885, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 2.029509091981765e-05, |
|
"loss": 0.4416, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.5103739043603683, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 2.0172588763600335e-05, |
|
"loss": 0.4802, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.5125929213358482, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 2.0049693786641734e-05, |
|
"loss": 0.4794, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.5148119383113281, |
|
"grad_norm": 2.8125, |
|
"learning_rate": 1.9926415321927347e-05, |
|
"loss": 0.5092, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.517030955286808, |
|
"grad_norm": 1.3984375, |
|
"learning_rate": 1.980276273156581e-05, |
|
"loss": 0.4595, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.5192499722622878, |
|
"grad_norm": 1.484375, |
|
"learning_rate": 1.9678745406077886e-05, |
|
"loss": 0.452, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.5214689892377676, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 1.9554372763683337e-05, |
|
"loss": 0.4528, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.5236880062132475, |
|
"grad_norm": 1.5, |
|
"learning_rate": 1.9429654249585684e-05, |
|
"loss": 0.4743, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.5259070231887274, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 1.9304599335254894e-05, |
|
"loss": 0.4476, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.5281260401642073, |
|
"grad_norm": 1.546875, |
|
"learning_rate": 1.9179217517708117e-05, |
|
"loss": 0.447, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.5303450571396872, |
|
"grad_norm": 1.625, |
|
"learning_rate": 1.9053518318788428e-05, |
|
"loss": 0.4606, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.5325640741151669, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 1.8927511284441722e-05, |
|
"loss": 0.4561, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.5347830910906468, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 1.880120598399178e-05, |
|
"loss": 0.4522, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.5370021080661267, |
|
"grad_norm": 1.9921875, |
|
"learning_rate": 1.8674612009413536e-05, |
|
"loss": 0.4943, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.5392211250416066, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 1.8547738974604623e-05, |
|
"loss": 0.5822, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.5414401420170865, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 1.842059651465531e-05, |
|
"loss": 0.4273, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.5436591589925663, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 1.829319428511673e-05, |
|
"loss": 0.4704, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.5458781759680461, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 1.816554196126767e-05, |
|
"loss": 0.5452, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.548097192943526, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 1.803764923737974e-05, |
|
"loss": 0.436, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.5503162099190059, |
|
"grad_norm": 1.7578125, |
|
"learning_rate": 1.7909525825981214e-05, |
|
"loss": 0.4518, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.5525352268944858, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 1.778118145711942e-05, |
|
"loss": 0.4138, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.5547542438699656, |
|
"grad_norm": 2.46875, |
|
"learning_rate": 1.7652625877621793e-05, |
|
"loss": 0.4332, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.5569732608454455, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 1.7523868850355704e-05, |
|
"loss": 0.4842, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.5591922778209253, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 1.7394920153487022e-05, |
|
"loss": 0.4935, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.5614112947964052, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 1.7265789579737528e-05, |
|
"loss": 0.5129, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.563630311771885, |
|
"grad_norm": 1.65625, |
|
"learning_rate": 1.7136486935641256e-05, |
|
"loss": 0.4281, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.5658493287473649, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 1.7007022040799726e-05, |
|
"loss": 0.4634, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.5680683457228448, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 1.687740472713623e-05, |
|
"loss": 0.5225, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.5702873626983247, |
|
"grad_norm": 1.9921875, |
|
"learning_rate": 1.674764483814918e-05, |
|
"loss": 0.4838, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.5725063796738045, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 1.661775222816453e-05, |
|
"loss": 0.4291, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.5747253966492843, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 1.648773676158747e-05, |
|
"loss": 0.3925, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.5769444136247642, |
|
"grad_norm": 1.734375, |
|
"learning_rate": 1.6357608312153223e-05, |
|
"loss": 0.4385, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.5791634306002441, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 1.6227376762177272e-05, |
|
"loss": 0.4144, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.581382447575724, |
|
"grad_norm": 2.71875, |
|
"learning_rate": 1.6097052001804825e-05, |
|
"loss": 0.5116, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.5836014645512038, |
|
"grad_norm": 1.953125, |
|
"learning_rate": 1.5966643928259753e-05, |
|
"loss": 0.4869, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.5858204815266836, |
|
"grad_norm": 1.9140625, |
|
"learning_rate": 1.5836162445092963e-05, |
|
"loss": 0.4644, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.5880394985021635, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 1.5705617461430282e-05, |
|
"loss": 0.3844, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.5902585154776434, |
|
"grad_norm": 1.75, |
|
"learning_rate": 1.5575018891219944e-05, |
|
"loss": 0.4233, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.5924775324531233, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 1.5444376652479706e-05, |
|
"loss": 0.5273, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.5946965494286032, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 1.531370066654362e-05, |
|
"loss": 0.4783, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.5969155664040829, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 1.5183000857308604e-05, |
|
"loss": 0.4747, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.5991345833795628, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 1.5052287150480774e-05, |
|
"loss": 0.4335, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.6013536003550427, |
|
"grad_norm": 1.75, |
|
"learning_rate": 1.4921569472821673e-05, |
|
"loss": 0.4489, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.6035726173305226, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 1.4790857751394398e-05, |
|
"loss": 0.4495, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.6057916343060025, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 1.4660161912809718e-05, |
|
"loss": 0.4516, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.6080106512814824, |
|
"grad_norm": 1.4765625, |
|
"learning_rate": 1.4529491882472209e-05, |
|
"loss": 0.4418, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.6102296682569621, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 1.4398857583826501e-05, |
|
"loss": 0.4701, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.612448685232442, |
|
"grad_norm": 2.390625, |
|
"learning_rate": 1.4268268937603659e-05, |
|
"loss": 0.4957, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.6146677022079219, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 1.413773586106777e-05, |
|
"loss": 0.5176, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.6168867191834018, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 1.400726826726282e-05, |
|
"loss": 0.4252, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.6191057361588816, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 1.3876876064259836e-05, |
|
"loss": 0.473, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.6213247531343615, |
|
"grad_norm": 1.6796875, |
|
"learning_rate": 1.3746569154404477e-05, |
|
"loss": 0.4589, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.6235437701098413, |
|
"grad_norm": 2.734375, |
|
"learning_rate": 1.3616357433564993e-05, |
|
"loss": 0.4083, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.6257627870853212, |
|
"grad_norm": 1.75, |
|
"learning_rate": 1.348625079038071e-05, |
|
"loss": 0.4638, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.6279818040608011, |
|
"grad_norm": 2.0625, |
|
"learning_rate": 1.335625910551108e-05, |
|
"loss": 0.4703, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.630200821036281, |
|
"grad_norm": 1.7109375, |
|
"learning_rate": 1.3226392250885288e-05, |
|
"loss": 0.4683, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.6324198380117608, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 1.3096660088952581e-05, |
|
"loss": 0.4641, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.6346388549872406, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 1.2967072471933255e-05, |
|
"loss": 0.4324, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.6368578719627205, |
|
"grad_norm": 1.6953125, |
|
"learning_rate": 1.283763924107046e-05, |
|
"loss": 0.5183, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.6390768889382004, |
|
"grad_norm": 1.4921875, |
|
"learning_rate": 1.2708370225882848e-05, |
|
"loss": 0.4178, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.6412959059136802, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 1.2579275243418074e-05, |
|
"loss": 0.4503, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.6435149228891601, |
|
"grad_norm": 1.21875, |
|
"learning_rate": 1.245036409750725e-05, |
|
"loss": 0.4796, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.64573393986464, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 1.2321646578020452e-05, |
|
"loss": 0.4437, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.6479529568401198, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 1.219313246012321e-05, |
|
"loss": 0.4504, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.6501719738155997, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 1.2064831503534185e-05, |
|
"loss": 0.5122, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.6523909907910795, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 1.1936753451783973e-05, |
|
"loss": 0.4294, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.6546100077665594, |
|
"grad_norm": 2.296875, |
|
"learning_rate": 1.1808908031475151e-05, |
|
"loss": 0.4895, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.6568290247420393, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 1.1681304951543635e-05, |
|
"loss": 0.4824, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.6590480417175192, |
|
"grad_norm": 1.875, |
|
"learning_rate": 1.1553953902521321e-05, |
|
"loss": 0.525, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.661267058692999, |
|
"grad_norm": 1.8359375, |
|
"learning_rate": 1.1426864555800195e-05, |
|
"loss": 0.4289, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.6634860756684788, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 1.1300046562897837e-05, |
|
"loss": 0.4695, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.6657050926439587, |
|
"grad_norm": 1.7109375, |
|
"learning_rate": 1.1173509554724461e-05, |
|
"loss": 0.465, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.6679241096194386, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 1.104726314085153e-05, |
|
"loss": 0.4641, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.6701431265949185, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 1.0921316908781965e-05, |
|
"loss": 0.4559, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.6723621435703984, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 1.079568042322205e-05, |
|
"loss": 0.4938, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.6745811605458781, |
|
"grad_norm": 1.2734375, |
|
"learning_rate": 1.0670363225355054e-05, |
|
"loss": 0.4247, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.676800177521358, |
|
"grad_norm": 1.5390625, |
|
"learning_rate": 1.0545374832116658e-05, |
|
"loss": 0.4739, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.6790191944968379, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 1.042072473547221e-05, |
|
"loss": 0.4923, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.6812382114723178, |
|
"grad_norm": 1.5, |
|
"learning_rate": 1.0296422401695867e-05, |
|
"loss": 0.5248, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.6834572284477977, |
|
"grad_norm": 1.5390625, |
|
"learning_rate": 1.017247727065172e-05, |
|
"loss": 0.4336, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.6856762454232775, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 1.0048898755076885e-05, |
|
"loss": 0.4915, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.6878952623987573, |
|
"grad_norm": 1.515625, |
|
"learning_rate": 9.925696239866679e-06, |
|
"loss": 0.4908, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.6901142793742372, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 9.802879081361927e-06, |
|
"loss": 0.4512, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.6923332963497171, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 9.680456606638376e-06, |
|
"loss": 0.4356, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.694552313325197, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 9.558438112798397e-06, |
|
"loss": 0.4321, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.6967713303006768, |
|
"grad_norm": 1.421875, |
|
"learning_rate": 9.436832866264942e-06, |
|
"loss": 0.4288, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.6989903472761566, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 9.3156501020778e-06, |
|
"loss": 0.4119, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.7012093642516365, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 9.194899023192295e-06, |
|
"loss": 0.4729, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.7034283812271164, |
|
"grad_norm": 2.03125, |
|
"learning_rate": 9.074588799780359e-06, |
|
"loss": 0.4438, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.7056473982025963, |
|
"grad_norm": 1.703125, |
|
"learning_rate": 8.95472856853414e-06, |
|
"loss": 0.4509, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.7078664151780761, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 8.835327431972136e-06, |
|
"loss": 0.4812, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.710085432153556, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 8.716394457747915e-06, |
|
"loss": 0.4796, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.7123044491290358, |
|
"grad_norm": 0.9375, |
|
"learning_rate": 8.597938677961505e-06, |
|
"loss": 0.4138, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.7145234661045157, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 8.479969088473462e-06, |
|
"loss": 0.4161, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.7167424830799956, |
|
"grad_norm": 2.203125, |
|
"learning_rate": 8.362494648221697e-06, |
|
"loss": 0.4685, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.7189615000554754, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 8.245524278541116e-06, |
|
"loss": 0.4476, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.7211805170309553, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 8.129066862486115e-06, |
|
"loss": 0.5104, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.7233995340064352, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 8.013131244155964e-06, |
|
"loss": 0.4467, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.725618550981915, |
|
"grad_norm": 2.234375, |
|
"learning_rate": 7.89772622802316e-06, |
|
"loss": 0.4317, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.7278375679573948, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 7.782860578264806e-06, |
|
"loss": 0.4398, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.7300565849328747, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 7.668543018097014e-06, |
|
"loss": 0.5054, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.7322756019083546, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 7.5547822291124715e-06, |
|
"loss": 0.4968, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.7344946188838345, |
|
"grad_norm": 1.5, |
|
"learning_rate": 7.441586850621102e-06, |
|
"loss": 0.4202, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.7367136358593144, |
|
"grad_norm": 1.765625, |
|
"learning_rate": 7.328965478993994e-06, |
|
"loss": 0.4447, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.7389326528347941, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 7.2169266670105555e-06, |
|
"loss": 0.4869, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.741151669810274, |
|
"grad_norm": 1.3203125, |
|
"learning_rate": 7.105478923209001e-06, |
|
"loss": 0.4744, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.7433706867857539, |
|
"grad_norm": 0.76171875, |
|
"learning_rate": 6.994630711240201e-06, |
|
"loss": 0.4054, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.7455897037612338, |
|
"grad_norm": 1.9765625, |
|
"learning_rate": 6.884390449224898e-06, |
|
"loss": 0.4307, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.7478087207367137, |
|
"grad_norm": 1.921875, |
|
"learning_rate": 6.774766509114435e-06, |
|
"loss": 0.4728, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.7500277377121934, |
|
"grad_norm": 1.53125, |
|
"learning_rate": 6.66576721605496e-06, |
|
"loss": 0.4254, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.7522467546876733, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 6.557400847755183e-06, |
|
"loss": 0.4508, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.7544657716631532, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 6.449675633857772e-06, |
|
"loss": 0.4814, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.7566847886386331, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 6.3425997553143315e-06, |
|
"loss": 0.459, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.758903805614113, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 6.236181343764144e-06, |
|
"loss": 0.464, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.7611228225895929, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 6.130428480916626e-06, |
|
"loss": 0.4897, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.7633418395650726, |
|
"grad_norm": 1.5625, |
|
"learning_rate": 6.025349197937577e-06, |
|
"loss": 0.4614, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.7655608565405525, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 5.920951474839266e-06, |
|
"loss": 0.444, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.7677798735160324, |
|
"grad_norm": 1.1640625, |
|
"learning_rate": 5.817243239874434e-06, |
|
"loss": 0.4649, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.7699988904915123, |
|
"grad_norm": 1.8046875, |
|
"learning_rate": 5.714232368934163e-06, |
|
"loss": 0.4758, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.7722179074669921, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 5.611926684949779e-06, |
|
"loss": 0.4519, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.774436924442472, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 5.510333957298756e-06, |
|
"loss": 0.4729, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.7766559414179518, |
|
"grad_norm": 1.28125, |
|
"learning_rate": 5.409461901214679e-06, |
|
"loss": 0.4557, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.7788749583934317, |
|
"grad_norm": 1.2265625, |
|
"learning_rate": 5.3093181772013545e-06, |
|
"loss": 0.4433, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.7810939753689116, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 5.209910390451007e-06, |
|
"loss": 0.4767, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.7833129923443914, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 5.111246090266763e-06, |
|
"loss": 0.4422, |
|
"step": 7060 |
|
}, |
|
{ |
|
"epoch": 0.7855320093198713, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 5.0133327694893035e-06, |
|
"loss": 0.4276, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 0.7877510262953512, |
|
"grad_norm": 1.5859375, |
|
"learning_rate": 4.916177863927856e-06, |
|
"loss": 0.482, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.789970043270831, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 4.819788751795485e-06, |
|
"loss": 0.4817, |
|
"step": 7120 |
|
}, |
|
{ |
|
"epoch": 0.7921890602463109, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 4.7241727531487925e-06, |
|
"loss": 0.4702, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 0.7944080772217907, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 4.629337129331983e-06, |
|
"loss": 0.4494, |
|
"step": 7160 |
|
}, |
|
{ |
|
"epoch": 0.7966270941972706, |
|
"grad_norm": 1.9453125, |
|
"learning_rate": 4.535289082425438e-06, |
|
"loss": 0.4411, |
|
"step": 7180 |
|
}, |
|
{ |
|
"epoch": 0.7988461111727505, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 4.442035754698759e-06, |
|
"loss": 0.4671, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.8010651281482304, |
|
"grad_norm": 2.359375, |
|
"learning_rate": 4.349584228068369e-06, |
|
"loss": 0.4077, |
|
"step": 7220 |
|
}, |
|
{ |
|
"epoch": 0.8032841451237102, |
|
"grad_norm": 1.3984375, |
|
"learning_rate": 4.257941523559703e-06, |
|
"loss": 0.5378, |
|
"step": 7240 |
|
}, |
|
{ |
|
"epoch": 0.80550316209919, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 4.167114600773983e-06, |
|
"loss": 0.4719, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 0.8077221790746699, |
|
"grad_norm": 1.140625, |
|
"learning_rate": 4.0771103573597125e-06, |
|
"loss": 0.4596, |
|
"step": 7280 |
|
}, |
|
{ |
|
"epoch": 0.8099411960501498, |
|
"grad_norm": 1.5703125, |
|
"learning_rate": 3.987935628488841e-06, |
|
"loss": 0.4812, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.8121602130256297, |
|
"grad_norm": 1.65625, |
|
"learning_rate": 3.899597186337676e-06, |
|
"loss": 0.4636, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 0.8143792300011095, |
|
"grad_norm": 1.609375, |
|
"learning_rate": 3.812101739572605e-06, |
|
"loss": 0.4605, |
|
"step": 7340 |
|
}, |
|
{ |
|
"epoch": 0.8165982469765893, |
|
"grad_norm": 2.09375, |
|
"learning_rate": 3.725455932840593e-06, |
|
"loss": 0.4368, |
|
"step": 7360 |
|
}, |
|
{ |
|
"epoch": 0.8188172639520692, |
|
"grad_norm": 1.8125, |
|
"learning_rate": 3.6396663462645917e-06, |
|
"loss": 0.4686, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 0.8210362809275491, |
|
"grad_norm": 1.875, |
|
"learning_rate": 3.554739494943813e-06, |
|
"loss": 0.4701, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.823255297903029, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 3.470681828458962e-06, |
|
"loss": 0.4595, |
|
"step": 7420 |
|
}, |
|
{ |
|
"epoch": 0.8254743148785089, |
|
"grad_norm": 1.328125, |
|
"learning_rate": 3.3874997303824416e-06, |
|
"loss": 0.4265, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 0.8276933318539886, |
|
"grad_norm": 1.5, |
|
"learning_rate": 3.305199517793557e-06, |
|
"loss": 0.4927, |
|
"step": 7460 |
|
}, |
|
{ |
|
"epoch": 0.8299123488294685, |
|
"grad_norm": 1.7109375, |
|
"learning_rate": 3.2237874407987776e-06, |
|
"loss": 0.4211, |
|
"step": 7480 |
|
}, |
|
{ |
|
"epoch": 0.8321313658049484, |
|
"grad_norm": 2.28125, |
|
"learning_rate": 3.1432696820570993e-06, |
|
"loss": 0.4771, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.8343503827804283, |
|
"grad_norm": 1.7109375, |
|
"learning_rate": 3.0636523563104985e-06, |
|
"loss": 0.4934, |
|
"step": 7520 |
|
}, |
|
{ |
|
"epoch": 0.8365693997559082, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 2.9849415099195886e-06, |
|
"loss": 0.5184, |
|
"step": 7540 |
|
}, |
|
{ |
|
"epoch": 0.838788416731388, |
|
"grad_norm": 1.234375, |
|
"learning_rate": 2.9071431204044123e-06, |
|
"loss": 0.4622, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 0.8410074337068678, |
|
"grad_norm": 1.6171875, |
|
"learning_rate": 2.8302630959905084e-06, |
|
"loss": 0.4491, |
|
"step": 7580 |
|
}, |
|
{ |
|
"epoch": 0.8432264506823477, |
|
"grad_norm": 2.46875, |
|
"learning_rate": 2.7543072751602246e-06, |
|
"loss": 0.4093, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.8454454676578276, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 2.6792814262093214e-06, |
|
"loss": 0.4312, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 0.8476644846333075, |
|
"grad_norm": 1.71875, |
|
"learning_rate": 2.605191246808912e-06, |
|
"loss": 0.455, |
|
"step": 7640 |
|
}, |
|
{ |
|
"epoch": 0.8498835016087873, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 2.5320423635727824e-06, |
|
"loss": 0.4839, |
|
"step": 7660 |
|
}, |
|
{ |
|
"epoch": 0.8521025185842672, |
|
"grad_norm": 2.15625, |
|
"learning_rate": 2.45984033163006e-06, |
|
"loss": 0.5647, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 0.854321535559747, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 2.388590634203366e-06, |
|
"loss": 0.4108, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.8565405525352269, |
|
"grad_norm": 1.90625, |
|
"learning_rate": 2.3182986821923934e-06, |
|
"loss": 0.3933, |
|
"step": 7720 |
|
}, |
|
{ |
|
"epoch": 0.8587595695107068, |
|
"grad_norm": 2.015625, |
|
"learning_rate": 2.2489698137629904e-06, |
|
"loss": 0.413, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 0.8609785864861866, |
|
"grad_norm": 2.140625, |
|
"learning_rate": 2.1806092939417732e-06, |
|
"loss": 0.516, |
|
"step": 7760 |
|
}, |
|
{ |
|
"epoch": 0.8631976034616665, |
|
"grad_norm": 1.40625, |
|
"learning_rate": 2.1132223142162714e-06, |
|
"loss": 0.4403, |
|
"step": 7780 |
|
}, |
|
{ |
|
"epoch": 0.8654166204371463, |
|
"grad_norm": 1.859375, |
|
"learning_rate": 2.046813992140679e-06, |
|
"loss": 0.4355, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.8676356374126262, |
|
"grad_norm": 1.4375, |
|
"learning_rate": 1.981389370947218e-06, |
|
"loss": 0.4744, |
|
"step": 7820 |
|
}, |
|
{ |
|
"epoch": 0.869854654388106, |
|
"grad_norm": 2.359375, |
|
"learning_rate": 1.9169534191631243e-06, |
|
"loss": 0.4106, |
|
"step": 7840 |
|
}, |
|
{ |
|
"epoch": 0.8720736713635859, |
|
"grad_norm": 2.046875, |
|
"learning_rate": 1.853511030233354e-06, |
|
"loss": 0.4519, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 0.8742926883390658, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 1.79106702214893e-06, |
|
"loss": 0.4568, |
|
"step": 7880 |
|
}, |
|
{ |
|
"epoch": 0.8765117053145457, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 1.7296261370810695e-06, |
|
"loss": 0.4734, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.8787307222900255, |
|
"grad_norm": 1.828125, |
|
"learning_rate": 1.669193041021041e-06, |
|
"loss": 0.547, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 0.8809497392655053, |
|
"grad_norm": 1.4453125, |
|
"learning_rate": 1.6097723234258188e-06, |
|
"loss": 0.5001, |
|
"step": 7940 |
|
}, |
|
{ |
|
"epoch": 0.8831687562409852, |
|
"grad_norm": 1.6875, |
|
"learning_rate": 1.5513684968695574e-06, |
|
"loss": 0.447, |
|
"step": 7960 |
|
}, |
|
{ |
|
"epoch": 0.8853877732164651, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 1.4939859967008768e-06, |
|
"loss": 0.4161, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 0.887606790191945, |
|
"grad_norm": 1.78125, |
|
"learning_rate": 1.437629180706037e-06, |
|
"loss": 0.4606, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.8898258071674249, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 1.382302328778e-06, |
|
"loss": 0.4126, |
|
"step": 8020 |
|
}, |
|
{ |
|
"epoch": 0.8920448241429046, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 1.328009642591394e-06, |
|
"loss": 0.4313, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 0.8942638411183845, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 1.2747552452834388e-06, |
|
"loss": 0.4436, |
|
"step": 8060 |
|
}, |
|
{ |
|
"epoch": 0.8964828580938644, |
|
"grad_norm": 1.6015625, |
|
"learning_rate": 1.2225431811408133e-06, |
|
"loss": 0.4591, |
|
"step": 8080 |
|
}, |
|
{ |
|
"epoch": 0.8987018750693443, |
|
"grad_norm": 1.484375, |
|
"learning_rate": 1.1713774152925195e-06, |
|
"loss": 0.4103, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.9009208920448242, |
|
"grad_norm": 1.890625, |
|
"learning_rate": 1.1212618334087693e-06, |
|
"loss": 0.4196, |
|
"step": 8120 |
|
}, |
|
{ |
|
"epoch": 0.903139909020304, |
|
"grad_norm": 2.125, |
|
"learning_rate": 1.0722002414058868e-06, |
|
"loss": 0.4701, |
|
"step": 8140 |
|
}, |
|
{ |
|
"epoch": 0.9053589259957838, |
|
"grad_norm": 1.8671875, |
|
"learning_rate": 1.0241963651572867e-06, |
|
"loss": 0.4452, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 0.9075779429712637, |
|
"grad_norm": 2.21875, |
|
"learning_rate": 9.772538502105093e-07, |
|
"loss": 0.4122, |
|
"step": 8180 |
|
}, |
|
{ |
|
"epoch": 0.9097969599467436, |
|
"grad_norm": 1.625, |
|
"learning_rate": 9.313762615103761e-07, |
|
"loss": 0.4265, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.9120159769222235, |
|
"grad_norm": 1.9296875, |
|
"learning_rate": 8.865670831282513e-07, |
|
"loss": 0.4678, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 0.9142349938977034, |
|
"grad_norm": 1.671875, |
|
"learning_rate": 8.42829717997457e-07, |
|
"loss": 0.4137, |
|
"step": 8240 |
|
}, |
|
{ |
|
"epoch": 0.9164540108731832, |
|
"grad_norm": 1.6484375, |
|
"learning_rate": 8.001674876548471e-07, |
|
"loss": 0.4939, |
|
"step": 8260 |
|
}, |
|
{ |
|
"epoch": 0.918673027848663, |
|
"grad_norm": 1.640625, |
|
"learning_rate": 7.585836319885525e-07, |
|
"loss": 0.4888, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 0.9208920448241429, |
|
"grad_norm": 1.578125, |
|
"learning_rate": 7.180813089919403e-07, |
|
"loss": 0.4352, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.9231110617996228, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 6.78663594523788e-07, |
|
"loss": 0.5086, |
|
"step": 8320 |
|
}, |
|
{ |
|
"epoch": 0.9253300787751026, |
|
"grad_norm": 1.96875, |
|
"learning_rate": 6.403334820746876e-07, |
|
"loss": 0.4695, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 0.9275490957505825, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 6.030938825397225e-07, |
|
"loss": 0.3946, |
|
"step": 8360 |
|
}, |
|
{ |
|
"epoch": 0.9297681127260623, |
|
"grad_norm": 1.609375, |
|
"learning_rate": 5.669476239973975e-07, |
|
"loss": 0.4797, |
|
"step": 8380 |
|
}, |
|
{ |
|
"epoch": 0.9319871297015422, |
|
"grad_norm": 1.796875, |
|
"learning_rate": 5.318974514948672e-07, |
|
"loss": 0.4274, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.9342061466770221, |
|
"grad_norm": 1.453125, |
|
"learning_rate": 4.979460268394726e-07, |
|
"loss": 0.4369, |
|
"step": 8420 |
|
}, |
|
{ |
|
"epoch": 0.936425163652502, |
|
"grad_norm": 1.6640625, |
|
"learning_rate": 4.6509592839659666e-07, |
|
"loss": 0.4452, |
|
"step": 8440 |
|
}, |
|
{ |
|
"epoch": 0.9386441806279818, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 4.333496508938506e-07, |
|
"loss": 0.4949, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 0.9408631976034617, |
|
"grad_norm": 2.078125, |
|
"learning_rate": 4.02709605231627e-07, |
|
"loss": 0.4815, |
|
"step": 8480 |
|
}, |
|
{ |
|
"epoch": 0.9430822145789415, |
|
"grad_norm": 2.109375, |
|
"learning_rate": 3.731781182999983e-07, |
|
"loss": 0.5533, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.9453012315544214, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 3.447574328020109e-07, |
|
"loss": 0.4741, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 0.9475202485299012, |
|
"grad_norm": 1.5546875, |
|
"learning_rate": 3.1744970708337205e-07, |
|
"loss": 0.5182, |
|
"step": 8540 |
|
}, |
|
{ |
|
"epoch": 0.9497392655053811, |
|
"grad_norm": 1.4296875, |
|
"learning_rate": 2.912570149685323e-07, |
|
"loss": 0.4612, |
|
"step": 8560 |
|
}, |
|
{ |
|
"epoch": 0.951958282480861, |
|
"grad_norm": 1.3984375, |
|
"learning_rate": 2.661813456032014e-07, |
|
"loss": 0.4253, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 0.9541772994563409, |
|
"grad_norm": 2.125, |
|
"learning_rate": 2.4222460330327933e-07, |
|
"loss": 0.5029, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.9563963164318207, |
|
"grad_norm": 1.9921875, |
|
"learning_rate": 2.1938860741023858e-07, |
|
"loss": 0.4444, |
|
"step": 8620 |
|
}, |
|
{ |
|
"epoch": 0.9586153334073005, |
|
"grad_norm": 1.390625, |
|
"learning_rate": 1.9767509215296297e-07, |
|
"loss": 0.4607, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 0.9608343503827804, |
|
"grad_norm": 1.7890625, |
|
"learning_rate": 1.7708570651604306e-07, |
|
"loss": 0.5045, |
|
"step": 8660 |
|
}, |
|
{ |
|
"epoch": 0.9630533673582603, |
|
"grad_norm": 1.484375, |
|
"learning_rate": 1.5762201411454626e-07, |
|
"loss": 0.4525, |
|
"step": 8680 |
|
}, |
|
{ |
|
"epoch": 0.9652723843337402, |
|
"grad_norm": 1.28125, |
|
"learning_rate": 1.3928549307527183e-07, |
|
"loss": 0.4484, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.9674914013092201, |
|
"grad_norm": 1.7265625, |
|
"learning_rate": 1.2207753592450078e-07, |
|
"loss": 0.4194, |
|
"step": 8720 |
|
}, |
|
{ |
|
"epoch": 0.9697104182846998, |
|
"grad_norm": 1.7734375, |
|
"learning_rate": 1.059994494822386e-07, |
|
"loss": 0.4822, |
|
"step": 8740 |
|
}, |
|
{ |
|
"epoch": 0.9719294352601797, |
|
"grad_norm": 1.7421875, |
|
"learning_rate": 9.105245476297653e-08, |
|
"loss": 0.4584, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 0.9741484522356596, |
|
"grad_norm": 1.5546875, |
|
"learning_rate": 7.723768688296217e-08, |
|
"loss": 0.4508, |
|
"step": 8780 |
|
}, |
|
{ |
|
"epoch": 0.9763674692111395, |
|
"grad_norm": 1.4921875, |
|
"learning_rate": 6.455619497399534e-08, |
|
"loss": 0.4953, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.9785864861866194, |
|
"grad_norm": 1.9375, |
|
"learning_rate": 5.300894210375329e-08, |
|
"loss": 0.4791, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 0.9808055031620991, |
|
"grad_norm": 1.4765625, |
|
"learning_rate": 4.259680520265596e-08, |
|
"loss": 0.4083, |
|
"step": 8840 |
|
}, |
|
{ |
|
"epoch": 0.983024520137579, |
|
"grad_norm": 1.6328125, |
|
"learning_rate": 3.3320574997267595e-08, |
|
"loss": 0.449, |
|
"step": 8860 |
|
}, |
|
{ |
|
"epoch": 0.9852435371130589, |
|
"grad_norm": 1.84375, |
|
"learning_rate": 2.5180955950243056e-08, |
|
"loss": 0.5022, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 0.9874625540885388, |
|
"grad_norm": 1.203125, |
|
"learning_rate": 1.8178566206837334e-08, |
|
"loss": 0.54, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.9896815710640187, |
|
"grad_norm": 1.8203125, |
|
"learning_rate": 1.231393754795307e-08, |
|
"loss": 0.4788, |
|
"step": 8920 |
|
}, |
|
{ |
|
"epoch": 0.9919005880394985, |
|
"grad_norm": 1.5546875, |
|
"learning_rate": 7.587515349762874e-09, |
|
"loss": 0.5039, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 0.9941196050149783, |
|
"grad_norm": 1.8984375, |
|
"learning_rate": 3.9996585498797145e-09, |
|
"loss": 0.5317, |
|
"step": 8960 |
|
}, |
|
{ |
|
"epoch": 0.9963386219904582, |
|
"grad_norm": 2.0, |
|
"learning_rate": 1.550639620103711e-09, |
|
"loss": 0.4578, |
|
"step": 8980 |
|
}, |
|
{ |
|
"epoch": 0.9985576389659381, |
|
"grad_norm": 1.5234375, |
|
"learning_rate": 2.406445457253659e-10, |
|
"loss": 0.474, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"step": 9013, |
|
"total_flos": 5.485972481640161e+17, |
|
"train_loss": 0.4824531834622566, |
|
"train_runtime": 26326.043, |
|
"train_samples_per_second": 0.685, |
|
"train_steps_per_second": 0.342 |
|
} |
|
], |
|
"logging_steps": 20, |
|
"max_steps": 9013, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 100, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.485972481640161e+17, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|