{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.9516041326808048, "eval_steps": 50, "global_step": 3500, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0027188689505165853, "grad_norm": 0.484375, "learning_rate": 5e-06, "loss": 0.7094, "step": 10 }, { "epoch": 0.005437737901033171, "grad_norm": 0.609375, "learning_rate": 1e-05, "loss": 0.7117, "step": 20 }, { "epoch": 0.008156606851549755, "grad_norm": 0.515625, "learning_rate": 1.5e-05, "loss": 0.7047, "step": 30 }, { "epoch": 0.010875475802066341, "grad_norm": 0.353515625, "learning_rate": 2e-05, "loss": 0.7059, "step": 40 }, { "epoch": 0.013594344752582926, "grad_norm": 0.43359375, "learning_rate": 2.5e-05, "loss": 0.7016, "step": 50 }, { "epoch": 0.013594344752582926, "eval_loss": 0.6993125081062317, "eval_runtime": 144.0611, "eval_samples_per_second": 3.471, "eval_steps_per_second": 0.111, "step": 50 }, { "epoch": 0.01631321370309951, "grad_norm": 0.31640625, "learning_rate": 3e-05, "loss": 0.7023, "step": 60 }, { "epoch": 0.019032082653616094, "grad_norm": 0.3125, "learning_rate": 3.5e-05, "loss": 0.6961, "step": 70 }, { "epoch": 0.021750951604132682, "grad_norm": 0.357421875, "learning_rate": 4e-05, "loss": 0.6957, "step": 80 }, { "epoch": 0.024469820554649267, "grad_norm": 0.3515625, "learning_rate": 4.5e-05, "loss": 0.6902, "step": 90 }, { "epoch": 0.027188689505165852, "grad_norm": 0.5390625, "learning_rate": 5e-05, "loss": 0.6828, "step": 100 }, { "epoch": 0.027188689505165852, "eval_loss": 0.6763437390327454, "eval_runtime": 120.6338, "eval_samples_per_second": 4.145, "eval_steps_per_second": 0.133, "step": 100 }, { "epoch": 0.029907558455682437, "grad_norm": 0.87890625, "learning_rate": 4.9860257126886535e-05, "loss": 0.6664, "step": 110 }, { "epoch": 0.03262642740619902, "grad_norm": 1.125, "learning_rate": 4.972051425377306e-05, "loss": 0.6402, "step": 120 }, { "epoch": 0.03534529635671561, "grad_norm": 1.7578125, "learning_rate": 4.9580771380659594e-05, "loss": 0.591, "step": 130 }, { "epoch": 0.03806416530723219, "grad_norm": 1.4609375, "learning_rate": 4.944102850754612e-05, "loss": 0.5555, "step": 140 }, { "epoch": 0.040783034257748776, "grad_norm": 2.71875, "learning_rate": 4.9301285634432645e-05, "loss": 0.5365, "step": 150 }, { "epoch": 0.040783034257748776, "eval_loss": 0.5141562223434448, "eval_runtime": 107.3892, "eval_samples_per_second": 4.656, "eval_steps_per_second": 0.149, "step": 150 }, { "epoch": 0.043501903208265365, "grad_norm": 2.40625, "learning_rate": 4.916154276131917e-05, "loss": 0.4957, "step": 160 }, { "epoch": 0.046220772158781946, "grad_norm": 2.5625, "learning_rate": 4.9021799888205704e-05, "loss": 0.4451, "step": 170 }, { "epoch": 0.048939641109298535, "grad_norm": 2.28125, "learning_rate": 4.888205701509223e-05, "loss": 0.4641, "step": 180 }, { "epoch": 0.051658510059815116, "grad_norm": 1.6875, "learning_rate": 4.874231414197876e-05, "loss": 0.434, "step": 190 }, { "epoch": 0.054377379010331704, "grad_norm": 2.671875, "learning_rate": 4.860257126886529e-05, "loss": 0.3877, "step": 200 }, { "epoch": 0.054377379010331704, "eval_loss": 0.41612499952316284, "eval_runtime": 106.701, "eval_samples_per_second": 4.686, "eval_steps_per_second": 0.15, "step": 200 }, { "epoch": 0.057096247960848286, "grad_norm": 2.015625, "learning_rate": 4.846282839575182e-05, "loss": 0.3903, "step": 210 }, { "epoch": 0.059815116911364874, "grad_norm": 3.34375, "learning_rate": 4.8323085522638347e-05, "loss": 0.3504, "step": 220 }, { "epoch": 0.06253398586188146, "grad_norm": 3.28125, "learning_rate": 4.818334264952488e-05, "loss": 0.3375, "step": 230 }, { "epoch": 0.06525285481239804, "grad_norm": 2.203125, "learning_rate": 4.8043599776411405e-05, "loss": 0.3176, "step": 240 }, { "epoch": 0.06797172376291463, "grad_norm": 2.78125, "learning_rate": 4.790385690329793e-05, "loss": 0.3045, "step": 250 }, { "epoch": 0.06797172376291463, "eval_loss": 0.32106250524520874, "eval_runtime": 105.1814, "eval_samples_per_second": 4.754, "eval_steps_per_second": 0.152, "step": 250 }, { "epoch": 0.07069059271343121, "grad_norm": 1.5703125, "learning_rate": 4.7764114030184464e-05, "loss": 0.3239, "step": 260 }, { "epoch": 0.0734094616639478, "grad_norm": 1.84375, "learning_rate": 4.762437115707099e-05, "loss": 0.2965, "step": 270 }, { "epoch": 0.07612833061446438, "grad_norm": 3.203125, "learning_rate": 4.748462828395752e-05, "loss": 0.3126, "step": 280 }, { "epoch": 0.07884719956498097, "grad_norm": 2.5625, "learning_rate": 4.734488541084405e-05, "loss": 0.2654, "step": 290 }, { "epoch": 0.08156606851549755, "grad_norm": 2.03125, "learning_rate": 4.720514253773058e-05, "loss": 0.3025, "step": 300 }, { "epoch": 0.08156606851549755, "eval_loss": 0.2882343828678131, "eval_runtime": 101.5553, "eval_samples_per_second": 4.923, "eval_steps_per_second": 0.158, "step": 300 }, { "epoch": 0.08428493746601413, "grad_norm": 2.296875, "learning_rate": 4.7065399664617106e-05, "loss": 0.2987, "step": 310 }, { "epoch": 0.08700380641653073, "grad_norm": 1.296875, "learning_rate": 4.692565679150364e-05, "loss": 0.2868, "step": 320 }, { "epoch": 0.08972267536704731, "grad_norm": 2.40625, "learning_rate": 4.6785913918390165e-05, "loss": 0.3149, "step": 330 }, { "epoch": 0.09244154431756389, "grad_norm": 2.015625, "learning_rate": 4.664617104527669e-05, "loss": 0.231, "step": 340 }, { "epoch": 0.09516041326808047, "grad_norm": 1.9140625, "learning_rate": 4.650642817216322e-05, "loss": 0.2977, "step": 350 }, { "epoch": 0.09516041326808047, "eval_loss": 0.2687968611717224, "eval_runtime": 135.1857, "eval_samples_per_second": 3.699, "eval_steps_per_second": 0.118, "step": 350 }, { "epoch": 0.09787928221859707, "grad_norm": 1.640625, "learning_rate": 4.636668529904975e-05, "loss": 0.2553, "step": 360 }, { "epoch": 0.10059815116911365, "grad_norm": 2.171875, "learning_rate": 4.622694242593628e-05, "loss": 0.2471, "step": 370 }, { "epoch": 0.10331702011963023, "grad_norm": 1.8203125, "learning_rate": 4.608719955282281e-05, "loss": 0.2753, "step": 380 }, { "epoch": 0.10603588907014681, "grad_norm": 2.09375, "learning_rate": 4.594745667970934e-05, "loss": 0.2913, "step": 390 }, { "epoch": 0.10875475802066341, "grad_norm": 1.8515625, "learning_rate": 4.5807713806595866e-05, "loss": 0.2674, "step": 400 }, { "epoch": 0.10875475802066341, "eval_loss": 0.25398436188697815, "eval_runtime": 111.6096, "eval_samples_per_second": 4.48, "eval_steps_per_second": 0.143, "step": 400 }, { "epoch": 0.11147362697117999, "grad_norm": 1.8828125, "learning_rate": 4.56679709334824e-05, "loss": 0.2763, "step": 410 }, { "epoch": 0.11419249592169657, "grad_norm": 3.125, "learning_rate": 4.5528228060368925e-05, "loss": 0.2797, "step": 420 }, { "epoch": 0.11691136487221315, "grad_norm": 1.5703125, "learning_rate": 4.538848518725545e-05, "loss": 0.252, "step": 430 }, { "epoch": 0.11963023382272975, "grad_norm": 2.734375, "learning_rate": 4.5248742314141976e-05, "loss": 0.3126, "step": 440 }, { "epoch": 0.12234910277324633, "grad_norm": 2.234375, "learning_rate": 4.510899944102851e-05, "loss": 0.2595, "step": 450 }, { "epoch": 0.12234910277324633, "eval_loss": 0.2485624998807907, "eval_runtime": 117.7724, "eval_samples_per_second": 4.245, "eval_steps_per_second": 0.136, "step": 450 }, { "epoch": 0.1250679717237629, "grad_norm": 1.7734375, "learning_rate": 4.4969256567915035e-05, "loss": 0.2341, "step": 460 }, { "epoch": 0.1277868406742795, "grad_norm": 2.15625, "learning_rate": 4.482951369480157e-05, "loss": 0.2337, "step": 470 }, { "epoch": 0.13050570962479607, "grad_norm": 1.1875, "learning_rate": 4.46897708216881e-05, "loss": 0.2757, "step": 480 }, { "epoch": 0.13322457857531267, "grad_norm": 3.390625, "learning_rate": 4.4550027948574626e-05, "loss": 0.2219, "step": 490 }, { "epoch": 0.13594344752582926, "grad_norm": 2.34375, "learning_rate": 4.441028507546116e-05, "loss": 0.2371, "step": 500 }, { "epoch": 0.13594344752582926, "eval_loss": 0.23534375429153442, "eval_runtime": 104.6381, "eval_samples_per_second": 4.778, "eval_steps_per_second": 0.153, "step": 500 }, { "epoch": 0.13866231647634583, "grad_norm": 3.984375, "learning_rate": 4.4270542202347684e-05, "loss": 0.2367, "step": 510 }, { "epoch": 0.14138118542686243, "grad_norm": 1.8515625, "learning_rate": 4.413079932923421e-05, "loss": 0.2249, "step": 520 }, { "epoch": 0.14410005437737902, "grad_norm": 1.6484375, "learning_rate": 4.3991056456120736e-05, "loss": 0.2861, "step": 530 }, { "epoch": 0.1468189233278956, "grad_norm": 1.1953125, "learning_rate": 4.385131358300727e-05, "loss": 0.2504, "step": 540 }, { "epoch": 0.14953779227841218, "grad_norm": 2.03125, "learning_rate": 4.3711570709893795e-05, "loss": 0.2458, "step": 550 }, { "epoch": 0.14953779227841218, "eval_loss": 0.2247031182050705, "eval_runtime": 110.8821, "eval_samples_per_second": 4.509, "eval_steps_per_second": 0.144, "step": 550 }, { "epoch": 0.15225666122892875, "grad_norm": 4.875, "learning_rate": 4.357182783678033e-05, "loss": 0.2205, "step": 560 }, { "epoch": 0.15497553017944535, "grad_norm": 1.8828125, "learning_rate": 4.343208496366685e-05, "loss": 0.2121, "step": 570 }, { "epoch": 0.15769439912996194, "grad_norm": 2.671875, "learning_rate": 4.3292342090553386e-05, "loss": 0.2286, "step": 580 }, { "epoch": 0.1604132680804785, "grad_norm": 1.8828125, "learning_rate": 4.315259921743992e-05, "loss": 0.2234, "step": 590 }, { "epoch": 0.1631321370309951, "grad_norm": 2.609375, "learning_rate": 4.3012856344326444e-05, "loss": 0.2261, "step": 600 }, { "epoch": 0.1631321370309951, "eval_loss": 0.22798436880111694, "eval_runtime": 102.5588, "eval_samples_per_second": 4.875, "eval_steps_per_second": 0.156, "step": 600 }, { "epoch": 0.1658510059815117, "grad_norm": 2.75, "learning_rate": 4.287311347121297e-05, "loss": 0.1937, "step": 610 }, { "epoch": 0.16856987493202827, "grad_norm": 2.921875, "learning_rate": 4.2733370598099496e-05, "loss": 0.2187, "step": 620 }, { "epoch": 0.17128874388254486, "grad_norm": 2.453125, "learning_rate": 4.259362772498603e-05, "loss": 0.2353, "step": 630 }, { "epoch": 0.17400761283306146, "grad_norm": 2.671875, "learning_rate": 4.2453884851872554e-05, "loss": 0.2376, "step": 640 }, { "epoch": 0.17672648178357803, "grad_norm": 1.8828125, "learning_rate": 4.231414197875909e-05, "loss": 0.2688, "step": 650 }, { "epoch": 0.17672648178357803, "eval_loss": 0.21031250059604645, "eval_runtime": 102.6274, "eval_samples_per_second": 4.872, "eval_steps_per_second": 0.156, "step": 650 }, { "epoch": 0.17944535073409462, "grad_norm": 1.34375, "learning_rate": 4.217439910564561e-05, "loss": 0.2137, "step": 660 }, { "epoch": 0.1821642196846112, "grad_norm": 1.59375, "learning_rate": 4.2034656232532146e-05, "loss": 0.245, "step": 670 }, { "epoch": 0.18488308863512778, "grad_norm": 2.140625, "learning_rate": 4.189491335941867e-05, "loss": 0.2625, "step": 680 }, { "epoch": 0.18760195758564438, "grad_norm": 1.8203125, "learning_rate": 4.17551704863052e-05, "loss": 0.2325, "step": 690 }, { "epoch": 0.19032082653616095, "grad_norm": 2.9375, "learning_rate": 4.161542761319172e-05, "loss": 0.2058, "step": 700 }, { "epoch": 0.19032082653616095, "eval_loss": 0.20640625059604645, "eval_runtime": 103.0629, "eval_samples_per_second": 4.851, "eval_steps_per_second": 0.155, "step": 700 }, { "epoch": 0.19303969548667754, "grad_norm": 1.328125, "learning_rate": 4.1475684740078256e-05, "loss": 0.222, "step": 710 }, { "epoch": 0.19575856443719414, "grad_norm": 1.359375, "learning_rate": 4.133594186696479e-05, "loss": 0.2537, "step": 720 }, { "epoch": 0.1984774333877107, "grad_norm": 1.1875, "learning_rate": 4.1196198993851314e-05, "loss": 0.2147, "step": 730 }, { "epoch": 0.2011963023382273, "grad_norm": 2.09375, "learning_rate": 4.105645612073785e-05, "loss": 0.1797, "step": 740 }, { "epoch": 0.2039151712887439, "grad_norm": 1.2109375, "learning_rate": 4.091671324762437e-05, "loss": 0.202, "step": 750 }, { "epoch": 0.2039151712887439, "eval_loss": 0.1913750022649765, "eval_runtime": 103.5861, "eval_samples_per_second": 4.827, "eval_steps_per_second": 0.154, "step": 750 }, { "epoch": 0.20663404023926046, "grad_norm": 2.328125, "learning_rate": 4.0776970374510905e-05, "loss": 0.2025, "step": 760 }, { "epoch": 0.20935290918977706, "grad_norm": 0.828125, "learning_rate": 4.063722750139743e-05, "loss": 0.2231, "step": 770 }, { "epoch": 0.21207177814029363, "grad_norm": 2.078125, "learning_rate": 4.049748462828396e-05, "loss": 0.2158, "step": 780 }, { "epoch": 0.21479064709081022, "grad_norm": 1.890625, "learning_rate": 4.035774175517048e-05, "loss": 0.2059, "step": 790 }, { "epoch": 0.21750951604132682, "grad_norm": 0.984375, "learning_rate": 4.0217998882057016e-05, "loss": 0.1792, "step": 800 }, { "epoch": 0.21750951604132682, "eval_loss": 0.20859375596046448, "eval_runtime": 99.6459, "eval_samples_per_second": 5.018, "eval_steps_per_second": 0.161, "step": 800 }, { "epoch": 0.22022838499184338, "grad_norm": 1.765625, "learning_rate": 4.007825600894354e-05, "loss": 0.2092, "step": 810 }, { "epoch": 0.22294725394235998, "grad_norm": 1.7265625, "learning_rate": 3.9938513135830074e-05, "loss": 0.184, "step": 820 }, { "epoch": 0.22566612289287658, "grad_norm": 1.578125, "learning_rate": 3.979877026271661e-05, "loss": 0.1796, "step": 830 }, { "epoch": 0.22838499184339314, "grad_norm": 2.015625, "learning_rate": 3.965902738960313e-05, "loss": 0.219, "step": 840 }, { "epoch": 0.23110386079390974, "grad_norm": 1.078125, "learning_rate": 3.9519284516489665e-05, "loss": 0.21, "step": 850 }, { "epoch": 0.23110386079390974, "eval_loss": 0.18590624630451202, "eval_runtime": 100.5192, "eval_samples_per_second": 4.974, "eval_steps_per_second": 0.159, "step": 850 }, { "epoch": 0.2338227297444263, "grad_norm": 1.1484375, "learning_rate": 3.937954164337619e-05, "loss": 0.1995, "step": 860 }, { "epoch": 0.2365415986949429, "grad_norm": 2.375, "learning_rate": 3.923979877026272e-05, "loss": 0.2072, "step": 870 }, { "epoch": 0.2392604676454595, "grad_norm": 2.0, "learning_rate": 3.910005589714924e-05, "loss": 0.1809, "step": 880 }, { "epoch": 0.24197933659597606, "grad_norm": 3.203125, "learning_rate": 3.8960313024035775e-05, "loss": 0.2188, "step": 890 }, { "epoch": 0.24469820554649266, "grad_norm": 1.8828125, "learning_rate": 3.88205701509223e-05, "loss": 0.2067, "step": 900 }, { "epoch": 0.24469820554649266, "eval_loss": 0.18979687988758087, "eval_runtime": 109.9141, "eval_samples_per_second": 4.549, "eval_steps_per_second": 0.146, "step": 900 }, { "epoch": 0.24741707449700925, "grad_norm": 2.421875, "learning_rate": 3.8680827277808834e-05, "loss": 0.2491, "step": 910 }, { "epoch": 0.2501359434475258, "grad_norm": 3.015625, "learning_rate": 3.854108440469536e-05, "loss": 0.1773, "step": 920 }, { "epoch": 0.2528548123980424, "grad_norm": 1.6953125, "learning_rate": 3.840134153158189e-05, "loss": 0.1618, "step": 930 }, { "epoch": 0.255573681348559, "grad_norm": 1.9296875, "learning_rate": 3.826159865846842e-05, "loss": 0.2098, "step": 940 }, { "epoch": 0.2582925502990756, "grad_norm": 3.625, "learning_rate": 3.812185578535495e-05, "loss": 0.2236, "step": 950 }, { "epoch": 0.2582925502990756, "eval_loss": 0.18760937452316284, "eval_runtime": 105.2992, "eval_samples_per_second": 4.748, "eval_steps_per_second": 0.152, "step": 950 }, { "epoch": 0.26101141924959215, "grad_norm": 1.6328125, "learning_rate": 3.798211291224148e-05, "loss": 0.228, "step": 960 }, { "epoch": 0.26373028820010874, "grad_norm": 1.8515625, "learning_rate": 3.7842370039128e-05, "loss": 0.1924, "step": 970 }, { "epoch": 0.26644915715062534, "grad_norm": 1.2421875, "learning_rate": 3.7702627166014535e-05, "loss": 0.1952, "step": 980 }, { "epoch": 0.26916802610114193, "grad_norm": 2.140625, "learning_rate": 3.756288429290106e-05, "loss": 0.226, "step": 990 }, { "epoch": 0.27188689505165853, "grad_norm": 1.3671875, "learning_rate": 3.7423141419787594e-05, "loss": 0.2095, "step": 1000 }, { "epoch": 0.27188689505165853, "eval_loss": 0.19670312106609344, "eval_runtime": 100.0411, "eval_samples_per_second": 4.998, "eval_steps_per_second": 0.16, "step": 1000 }, { "epoch": 0.27460576400217507, "grad_norm": 1.4765625, "learning_rate": 3.728339854667412e-05, "loss": 0.197, "step": 1010 }, { "epoch": 0.27732463295269166, "grad_norm": 1.703125, "learning_rate": 3.714365567356065e-05, "loss": 0.1979, "step": 1020 }, { "epoch": 0.28004350190320826, "grad_norm": 1.9375, "learning_rate": 3.700391280044718e-05, "loss": 0.2003, "step": 1030 }, { "epoch": 0.28276237085372485, "grad_norm": 1.6953125, "learning_rate": 3.686416992733371e-05, "loss": 0.1868, "step": 1040 }, { "epoch": 0.28548123980424145, "grad_norm": 1.53125, "learning_rate": 3.6724427054220237e-05, "loss": 0.198, "step": 1050 }, { "epoch": 0.28548123980424145, "eval_loss": 0.19362500309944153, "eval_runtime": 100.7072, "eval_samples_per_second": 4.965, "eval_steps_per_second": 0.159, "step": 1050 }, { "epoch": 0.28820010875475804, "grad_norm": 1.640625, "learning_rate": 3.658468418110676e-05, "loss": 0.2059, "step": 1060 }, { "epoch": 0.2909189777052746, "grad_norm": 1.1640625, "learning_rate": 3.6444941307993295e-05, "loss": 0.1568, "step": 1070 }, { "epoch": 0.2936378466557912, "grad_norm": 1.1328125, "learning_rate": 3.630519843487982e-05, "loss": 0.1812, "step": 1080 }, { "epoch": 0.2963567156063078, "grad_norm": 2.1875, "learning_rate": 3.6165455561766354e-05, "loss": 0.1819, "step": 1090 }, { "epoch": 0.29907558455682437, "grad_norm": 1.546875, "learning_rate": 3.602571268865288e-05, "loss": 0.1739, "step": 1100 }, { "epoch": 0.29907558455682437, "eval_loss": 0.18107812106609344, "eval_runtime": 101.4517, "eval_samples_per_second": 4.928, "eval_steps_per_second": 0.158, "step": 1100 }, { "epoch": 0.30179445350734097, "grad_norm": 1.15625, "learning_rate": 3.588596981553941e-05, "loss": 0.2062, "step": 1110 }, { "epoch": 0.3045133224578575, "grad_norm": 2.0, "learning_rate": 3.574622694242594e-05, "loss": 0.2333, "step": 1120 }, { "epoch": 0.3072321914083741, "grad_norm": 1.7734375, "learning_rate": 3.560648406931247e-05, "loss": 0.1878, "step": 1130 }, { "epoch": 0.3099510603588907, "grad_norm": 2.515625, "learning_rate": 3.5466741196198996e-05, "loss": 0.2058, "step": 1140 }, { "epoch": 0.3126699293094073, "grad_norm": 2.09375, "learning_rate": 3.532699832308552e-05, "loss": 0.206, "step": 1150 }, { "epoch": 0.3126699293094073, "eval_loss": 0.182679682970047, "eval_runtime": 99.428, "eval_samples_per_second": 5.029, "eval_steps_per_second": 0.161, "step": 1150 }, { "epoch": 0.3153887982599239, "grad_norm": 1.3125, "learning_rate": 3.518725544997205e-05, "loss": 0.2013, "step": 1160 }, { "epoch": 0.3181076672104405, "grad_norm": 1.75, "learning_rate": 3.504751257685858e-05, "loss": 0.2018, "step": 1170 }, { "epoch": 0.320826536160957, "grad_norm": 1.265625, "learning_rate": 3.4907769703745107e-05, "loss": 0.2129, "step": 1180 }, { "epoch": 0.3235454051114736, "grad_norm": 1.984375, "learning_rate": 3.476802683063164e-05, "loss": 0.1912, "step": 1190 }, { "epoch": 0.3262642740619902, "grad_norm": 2.171875, "learning_rate": 3.462828395751817e-05, "loss": 0.1822, "step": 1200 }, { "epoch": 0.3262642740619902, "eval_loss": 0.18774999678134918, "eval_runtime": 99.1074, "eval_samples_per_second": 5.045, "eval_steps_per_second": 0.161, "step": 1200 }, { "epoch": 0.3289831430125068, "grad_norm": 4.5625, "learning_rate": 3.44885410844047e-05, "loss": 0.1758, "step": 1210 }, { "epoch": 0.3317020119630234, "grad_norm": 1.9921875, "learning_rate": 3.434879821129123e-05, "loss": 0.2345, "step": 1220 }, { "epoch": 0.33442088091353994, "grad_norm": 1.3359375, "learning_rate": 3.4209055338177756e-05, "loss": 0.2056, "step": 1230 }, { "epoch": 0.33713974986405654, "grad_norm": 1.8515625, "learning_rate": 3.406931246506428e-05, "loss": 0.2017, "step": 1240 }, { "epoch": 0.33985861881457313, "grad_norm": 1.9375, "learning_rate": 3.392956959195081e-05, "loss": 0.1789, "step": 1250 }, { "epoch": 0.33985861881457313, "eval_loss": 0.18862499296665192, "eval_runtime": 100.4085, "eval_samples_per_second": 4.98, "eval_steps_per_second": 0.159, "step": 1250 }, { "epoch": 0.3425774877650897, "grad_norm": 1.3046875, "learning_rate": 3.378982671883734e-05, "loss": 0.1966, "step": 1260 }, { "epoch": 0.3452963567156063, "grad_norm": 2.234375, "learning_rate": 3.3650083845723866e-05, "loss": 0.201, "step": 1270 }, { "epoch": 0.3480152256661229, "grad_norm": 1.375, "learning_rate": 3.35103409726104e-05, "loss": 0.186, "step": 1280 }, { "epoch": 0.35073409461663946, "grad_norm": 2.3125, "learning_rate": 3.3370598099496925e-05, "loss": 0.1931, "step": 1290 }, { "epoch": 0.35345296356715605, "grad_norm": 1.6015625, "learning_rate": 3.323085522638346e-05, "loss": 0.1633, "step": 1300 }, { "epoch": 0.35345296356715605, "eval_loss": 0.17839062213897705, "eval_runtime": 97.1074, "eval_samples_per_second": 5.149, "eval_steps_per_second": 0.165, "step": 1300 }, { "epoch": 0.35617183251767265, "grad_norm": 1.4453125, "learning_rate": 3.309111235326999e-05, "loss": 0.1765, "step": 1310 }, { "epoch": 0.35889070146818924, "grad_norm": 3.640625, "learning_rate": 3.2951369480156516e-05, "loss": 0.1793, "step": 1320 }, { "epoch": 0.36160957041870584, "grad_norm": 1.34375, "learning_rate": 3.281162660704304e-05, "loss": 0.1976, "step": 1330 }, { "epoch": 0.3643284393692224, "grad_norm": 1.5, "learning_rate": 3.267188373392957e-05, "loss": 0.2001, "step": 1340 }, { "epoch": 0.367047308319739, "grad_norm": 0.9140625, "learning_rate": 3.25321408608161e-05, "loss": 0.1723, "step": 1350 }, { "epoch": 0.367047308319739, "eval_loss": 0.17289061844348907, "eval_runtime": 97.8393, "eval_samples_per_second": 5.11, "eval_steps_per_second": 0.164, "step": 1350 }, { "epoch": 0.36976617727025557, "grad_norm": 1.734375, "learning_rate": 3.2392397987702626e-05, "loss": 0.2236, "step": 1360 }, { "epoch": 0.37248504622077216, "grad_norm": 1.2109375, "learning_rate": 3.225265511458916e-05, "loss": 0.1914, "step": 1370 }, { "epoch": 0.37520391517128876, "grad_norm": 2.828125, "learning_rate": 3.2112912241475685e-05, "loss": 0.1715, "step": 1380 }, { "epoch": 0.37792278412180536, "grad_norm": 2.265625, "learning_rate": 3.197316936836222e-05, "loss": 0.1694, "step": 1390 }, { "epoch": 0.3806416530723219, "grad_norm": 1.671875, "learning_rate": 3.183342649524874e-05, "loss": 0.1828, "step": 1400 }, { "epoch": 0.3806416530723219, "eval_loss": 0.1787421852350235, "eval_runtime": 105.4565, "eval_samples_per_second": 4.741, "eval_steps_per_second": 0.152, "step": 1400 }, { "epoch": 0.3833605220228385, "grad_norm": 1.4296875, "learning_rate": 3.1693683622135276e-05, "loss": 0.224, "step": 1410 }, { "epoch": 0.3860793909733551, "grad_norm": 2.484375, "learning_rate": 3.15539407490218e-05, "loss": 0.1992, "step": 1420 }, { "epoch": 0.3887982599238717, "grad_norm": 3.0625, "learning_rate": 3.141419787590833e-05, "loss": 0.2106, "step": 1430 }, { "epoch": 0.3915171288743883, "grad_norm": 1.1875, "learning_rate": 3.127445500279486e-05, "loss": 0.1732, "step": 1440 }, { "epoch": 0.3942359978249048, "grad_norm": 1.5625, "learning_rate": 3.1134712129681386e-05, "loss": 0.1865, "step": 1450 }, { "epoch": 0.3942359978249048, "eval_loss": 0.1751093715429306, "eval_runtime": 97.1038, "eval_samples_per_second": 5.149, "eval_steps_per_second": 0.165, "step": 1450 }, { "epoch": 0.3969548667754214, "grad_norm": 1.8515625, "learning_rate": 3.099496925656792e-05, "loss": 0.2113, "step": 1460 }, { "epoch": 0.399673735725938, "grad_norm": 1.453125, "learning_rate": 3.0855226383454444e-05, "loss": 0.2173, "step": 1470 }, { "epoch": 0.4023926046764546, "grad_norm": 2.046875, "learning_rate": 3.071548351034098e-05, "loss": 0.1893, "step": 1480 }, { "epoch": 0.4051114736269712, "grad_norm": 2.0, "learning_rate": 3.05757406372275e-05, "loss": 0.1885, "step": 1490 }, { "epoch": 0.4078303425774878, "grad_norm": 1.4140625, "learning_rate": 3.0435997764114032e-05, "loss": 0.1642, "step": 1500 }, { "epoch": 0.4078303425774878, "eval_loss": 0.1752343773841858, "eval_runtime": 97.2789, "eval_samples_per_second": 5.14, "eval_steps_per_second": 0.164, "step": 1500 }, { "epoch": 0.41054921152800433, "grad_norm": 2.359375, "learning_rate": 3.0296254891000558e-05, "loss": 0.189, "step": 1510 }, { "epoch": 0.4132680804785209, "grad_norm": 1.5546875, "learning_rate": 3.015651201788709e-05, "loss": 0.1399, "step": 1520 }, { "epoch": 0.4159869494290375, "grad_norm": 0.9921875, "learning_rate": 3.0016769144773617e-05, "loss": 0.2088, "step": 1530 }, { "epoch": 0.4187058183795541, "grad_norm": 1.7421875, "learning_rate": 2.9877026271660146e-05, "loss": 0.2048, "step": 1540 }, { "epoch": 0.4214246873300707, "grad_norm": 1.1640625, "learning_rate": 2.973728339854668e-05, "loss": 0.1898, "step": 1550 }, { "epoch": 0.4214246873300707, "eval_loss": 0.17342187464237213, "eval_runtime": 95.7506, "eval_samples_per_second": 5.222, "eval_steps_per_second": 0.167, "step": 1550 }, { "epoch": 0.42414355628058725, "grad_norm": 2.71875, "learning_rate": 2.9597540525433204e-05, "loss": 0.2081, "step": 1560 }, { "epoch": 0.42686242523110385, "grad_norm": 2.34375, "learning_rate": 2.9457797652319734e-05, "loss": 0.1575, "step": 1570 }, { "epoch": 0.42958129418162044, "grad_norm": 3.21875, "learning_rate": 2.931805477920626e-05, "loss": 0.1849, "step": 1580 }, { "epoch": 0.43230016313213704, "grad_norm": 1.7421875, "learning_rate": 2.9178311906092792e-05, "loss": 0.2059, "step": 1590 }, { "epoch": 0.43501903208265363, "grad_norm": 1.2421875, "learning_rate": 2.9038569032979318e-05, "loss": 0.1641, "step": 1600 }, { "epoch": 0.43501903208265363, "eval_loss": 0.1657109409570694, "eval_runtime": 96.6429, "eval_samples_per_second": 5.174, "eval_steps_per_second": 0.166, "step": 1600 }, { "epoch": 0.4377379010331702, "grad_norm": 1.953125, "learning_rate": 2.889882615986585e-05, "loss": 0.1906, "step": 1610 }, { "epoch": 0.44045676998368677, "grad_norm": 1.6328125, "learning_rate": 2.8759083286752376e-05, "loss": 0.1907, "step": 1620 }, { "epoch": 0.44317563893420336, "grad_norm": 4.625, "learning_rate": 2.8619340413638906e-05, "loss": 0.2193, "step": 1630 }, { "epoch": 0.44589450788471996, "grad_norm": 1.515625, "learning_rate": 2.847959754052543e-05, "loss": 0.1714, "step": 1640 }, { "epoch": 0.44861337683523655, "grad_norm": 1.8125, "learning_rate": 2.8339854667411964e-05, "loss": 0.1771, "step": 1650 }, { "epoch": 0.44861337683523655, "eval_loss": 0.18095313012599945, "eval_runtime": 95.4318, "eval_samples_per_second": 5.239, "eval_steps_per_second": 0.168, "step": 1650 }, { "epoch": 0.45133224578575315, "grad_norm": 1.9296875, "learning_rate": 2.820011179429849e-05, "loss": 0.1917, "step": 1660 }, { "epoch": 0.4540511147362697, "grad_norm": 1.8046875, "learning_rate": 2.806036892118502e-05, "loss": 0.1915, "step": 1670 }, { "epoch": 0.4567699836867863, "grad_norm": 2.53125, "learning_rate": 2.7920626048071552e-05, "loss": 0.1436, "step": 1680 }, { "epoch": 0.4594888526373029, "grad_norm": 0.921875, "learning_rate": 2.7780883174958078e-05, "loss": 0.1683, "step": 1690 }, { "epoch": 0.4622077215878195, "grad_norm": 1.640625, "learning_rate": 2.764114030184461e-05, "loss": 0.1762, "step": 1700 }, { "epoch": 0.4622077215878195, "eval_loss": 0.1666562557220459, "eval_runtime": 96.3207, "eval_samples_per_second": 5.191, "eval_steps_per_second": 0.166, "step": 1700 }, { "epoch": 0.46492659053833607, "grad_norm": 2.484375, "learning_rate": 2.7501397428731136e-05, "loss": 0.1977, "step": 1710 }, { "epoch": 0.4676454594888526, "grad_norm": 2.171875, "learning_rate": 2.7361654555617665e-05, "loss": 0.2066, "step": 1720 }, { "epoch": 0.4703643284393692, "grad_norm": 2.5, "learning_rate": 2.722191168250419e-05, "loss": 0.1756, "step": 1730 }, { "epoch": 0.4730831973898858, "grad_norm": 1.5625, "learning_rate": 2.7082168809390724e-05, "loss": 0.1871, "step": 1740 }, { "epoch": 0.4758020663404024, "grad_norm": 1.8046875, "learning_rate": 2.694242593627725e-05, "loss": 0.2212, "step": 1750 }, { "epoch": 0.4758020663404024, "eval_loss": 0.15598437190055847, "eval_runtime": 94.5918, "eval_samples_per_second": 5.286, "eval_steps_per_second": 0.169, "step": 1750 }, { "epoch": 0.478520935290919, "grad_norm": 2.1875, "learning_rate": 2.680268306316378e-05, "loss": 0.1998, "step": 1760 }, { "epoch": 0.4812398042414356, "grad_norm": 2.046875, "learning_rate": 2.6662940190050305e-05, "loss": 0.1886, "step": 1770 }, { "epoch": 0.4839586731919521, "grad_norm": 1.5390625, "learning_rate": 2.6523197316936838e-05, "loss": 0.1857, "step": 1780 }, { "epoch": 0.4866775421424687, "grad_norm": 1.2421875, "learning_rate": 2.638345444382337e-05, "loss": 0.1601, "step": 1790 }, { "epoch": 0.4893964110929853, "grad_norm": 1.3984375, "learning_rate": 2.6243711570709896e-05, "loss": 0.1989, "step": 1800 }, { "epoch": 0.4893964110929853, "eval_loss": 0.17681249976158142, "eval_runtime": 97.019, "eval_samples_per_second": 5.154, "eval_steps_per_second": 0.165, "step": 1800 }, { "epoch": 0.4921152800435019, "grad_norm": 0.94140625, "learning_rate": 2.6103968697596425e-05, "loss": 0.1391, "step": 1810 }, { "epoch": 0.4948341489940185, "grad_norm": 2.765625, "learning_rate": 2.596422582448295e-05, "loss": 0.211, "step": 1820 }, { "epoch": 0.49755301794453505, "grad_norm": 2.859375, "learning_rate": 2.5824482951369484e-05, "loss": 0.1558, "step": 1830 }, { "epoch": 0.5002718868950516, "grad_norm": 2.796875, "learning_rate": 2.568474007825601e-05, "loss": 0.1764, "step": 1840 }, { "epoch": 0.5029907558455683, "grad_norm": 1.609375, "learning_rate": 2.554499720514254e-05, "loss": 0.1547, "step": 1850 }, { "epoch": 0.5029907558455683, "eval_loss": 0.1746249943971634, "eval_runtime": 95.3118, "eval_samples_per_second": 5.246, "eval_steps_per_second": 0.168, "step": 1850 }, { "epoch": 0.5057096247960848, "grad_norm": 2.1875, "learning_rate": 2.5405254332029065e-05, "loss": 0.1657, "step": 1860 }, { "epoch": 0.5084284937466014, "grad_norm": 1.828125, "learning_rate": 2.5265511458915597e-05, "loss": 0.1985, "step": 1870 }, { "epoch": 0.511147362697118, "grad_norm": 1.1875, "learning_rate": 2.5125768585802123e-05, "loss": 0.1828, "step": 1880 }, { "epoch": 0.5138662316476346, "grad_norm": 1.1953125, "learning_rate": 2.4986025712688656e-05, "loss": 0.1603, "step": 1890 }, { "epoch": 0.5165851005981512, "grad_norm": 1.8203125, "learning_rate": 2.484628283957518e-05, "loss": 0.1718, "step": 1900 }, { "epoch": 0.5165851005981512, "eval_loss": 0.1659843772649765, "eval_runtime": 117.6794, "eval_samples_per_second": 4.249, "eval_steps_per_second": 0.136, "step": 1900 }, { "epoch": 0.5193039695486678, "grad_norm": 0.7421875, "learning_rate": 2.470653996646171e-05, "loss": 0.1812, "step": 1910 }, { "epoch": 0.5220228384991843, "grad_norm": 1.4375, "learning_rate": 2.456679709334824e-05, "loss": 0.1624, "step": 1920 }, { "epoch": 0.524741707449701, "grad_norm": 1.90625, "learning_rate": 2.442705422023477e-05, "loss": 0.1344, "step": 1930 }, { "epoch": 0.5274605764002175, "grad_norm": 1.6015625, "learning_rate": 2.42873113471213e-05, "loss": 0.1823, "step": 1940 }, { "epoch": 0.5301794453507341, "grad_norm": 2.515625, "learning_rate": 2.4147568474007825e-05, "loss": 0.1742, "step": 1950 }, { "epoch": 0.5301794453507341, "eval_loss": 0.1723593771457672, "eval_runtime": 104.5426, "eval_samples_per_second": 4.783, "eval_steps_per_second": 0.153, "step": 1950 }, { "epoch": 0.5328983143012507, "grad_norm": 4.375, "learning_rate": 2.4007825600894354e-05, "loss": 0.2232, "step": 1960 }, { "epoch": 0.5356171832517672, "grad_norm": 1.6640625, "learning_rate": 2.3868082727780886e-05, "loss": 0.1532, "step": 1970 }, { "epoch": 0.5383360522022839, "grad_norm": 2.265625, "learning_rate": 2.3728339854667416e-05, "loss": 0.171, "step": 1980 }, { "epoch": 0.5410549211528004, "grad_norm": 1.4375, "learning_rate": 2.358859698155394e-05, "loss": 0.2081, "step": 1990 }, { "epoch": 0.5437737901033171, "grad_norm": 1.6640625, "learning_rate": 2.344885410844047e-05, "loss": 0.182, "step": 2000 }, { "epoch": 0.5437737901033171, "eval_loss": 0.1640625, "eval_runtime": 99.8495, "eval_samples_per_second": 5.008, "eval_steps_per_second": 0.16, "step": 2000 }, { "epoch": 0.5464926590538336, "grad_norm": 1.203125, "learning_rate": 2.3309111235327e-05, "loss": 0.1363, "step": 2010 }, { "epoch": 0.5492115280043501, "grad_norm": 1.8984375, "learning_rate": 2.316936836221353e-05, "loss": 0.1666, "step": 2020 }, { "epoch": 0.5519303969548668, "grad_norm": 1.6796875, "learning_rate": 2.302962548910006e-05, "loss": 0.1405, "step": 2030 }, { "epoch": 0.5546492659053833, "grad_norm": 1.3046875, "learning_rate": 2.2889882615986584e-05, "loss": 0.1768, "step": 2040 }, { "epoch": 0.5573681348559, "grad_norm": 3.140625, "learning_rate": 2.2750139742873114e-05, "loss": 0.2092, "step": 2050 }, { "epoch": 0.5573681348559, "eval_loss": 0.1575937569141388, "eval_runtime": 99.6448, "eval_samples_per_second": 5.018, "eval_steps_per_second": 0.161, "step": 2050 }, { "epoch": 0.5600870038064165, "grad_norm": 2.09375, "learning_rate": 2.2610396869759643e-05, "loss": 0.1888, "step": 2060 }, { "epoch": 0.5628058727569332, "grad_norm": 1.921875, "learning_rate": 2.2470653996646172e-05, "loss": 0.1454, "step": 2070 }, { "epoch": 0.5655247417074497, "grad_norm": 1.0703125, "learning_rate": 2.2330911123532698e-05, "loss": 0.1697, "step": 2080 }, { "epoch": 0.5682436106579662, "grad_norm": 3.171875, "learning_rate": 2.219116825041923e-05, "loss": 0.19, "step": 2090 }, { "epoch": 0.5709624796084829, "grad_norm": 1.546875, "learning_rate": 2.205142537730576e-05, "loss": 0.1925, "step": 2100 }, { "epoch": 0.5709624796084829, "eval_loss": 0.16167187690734863, "eval_runtime": 99.0392, "eval_samples_per_second": 5.049, "eval_steps_per_second": 0.162, "step": 2100 }, { "epoch": 0.5736813485589994, "grad_norm": 1.8828125, "learning_rate": 2.191168250419229e-05, "loss": 0.1822, "step": 2110 }, { "epoch": 0.5764002175095161, "grad_norm": 1.4609375, "learning_rate": 2.1771939631078815e-05, "loss": 0.1685, "step": 2120 }, { "epoch": 0.5791190864600326, "grad_norm": 1.0546875, "learning_rate": 2.1632196757965344e-05, "loss": 0.2201, "step": 2130 }, { "epoch": 0.5818379554105492, "grad_norm": 1.0625, "learning_rate": 2.1492453884851873e-05, "loss": 0.1572, "step": 2140 }, { "epoch": 0.5845568243610658, "grad_norm": 0.9140625, "learning_rate": 2.1352711011738403e-05, "loss": 0.1507, "step": 2150 }, { "epoch": 0.5845568243610658, "eval_loss": 0.15681250393390656, "eval_runtime": 96.4445, "eval_samples_per_second": 5.184, "eval_steps_per_second": 0.166, "step": 2150 }, { "epoch": 0.5872756933115824, "grad_norm": 2.9375, "learning_rate": 2.1212968138624932e-05, "loss": 0.1653, "step": 2160 }, { "epoch": 0.589994562262099, "grad_norm": 1.5625, "learning_rate": 2.1073225265511458e-05, "loss": 0.1815, "step": 2170 }, { "epoch": 0.5927134312126155, "grad_norm": 1.15625, "learning_rate": 2.0933482392397987e-05, "loss": 0.1157, "step": 2180 }, { "epoch": 0.5954323001631321, "grad_norm": 1.890625, "learning_rate": 2.0793739519284516e-05, "loss": 0.1535, "step": 2190 }, { "epoch": 0.5981511691136487, "grad_norm": 1.21875, "learning_rate": 2.0653996646171045e-05, "loss": 0.1547, "step": 2200 }, { "epoch": 0.5981511691136487, "eval_loss": 0.16277343034744263, "eval_runtime": 96.1155, "eval_samples_per_second": 5.202, "eval_steps_per_second": 0.166, "step": 2200 }, { "epoch": 0.6008700380641653, "grad_norm": 2.09375, "learning_rate": 2.0514253773057575e-05, "loss": 0.1582, "step": 2210 }, { "epoch": 0.6035889070146819, "grad_norm": 1.3203125, "learning_rate": 2.0374510899944104e-05, "loss": 0.1592, "step": 2220 }, { "epoch": 0.6063077759651985, "grad_norm": 2.6875, "learning_rate": 2.0234768026830633e-05, "loss": 0.1771, "step": 2230 }, { "epoch": 0.609026644915715, "grad_norm": 1.9921875, "learning_rate": 2.0095025153717162e-05, "loss": 0.169, "step": 2240 }, { "epoch": 0.6117455138662317, "grad_norm": 2.140625, "learning_rate": 1.995528228060369e-05, "loss": 0.1685, "step": 2250 }, { "epoch": 0.6117455138662317, "eval_loss": 0.16175781190395355, "eval_runtime": 99.1628, "eval_samples_per_second": 5.042, "eval_steps_per_second": 0.161, "step": 2250 }, { "epoch": 0.6144643828167482, "grad_norm": 1.234375, "learning_rate": 1.9815539407490218e-05, "loss": 0.165, "step": 2260 }, { "epoch": 0.6171832517672649, "grad_norm": 2.140625, "learning_rate": 1.9675796534376747e-05, "loss": 0.1928, "step": 2270 }, { "epoch": 0.6199021207177814, "grad_norm": 1.6484375, "learning_rate": 1.9536053661263276e-05, "loss": 0.2038, "step": 2280 }, { "epoch": 0.622620989668298, "grad_norm": 1.828125, "learning_rate": 1.9396310788149805e-05, "loss": 0.1871, "step": 2290 }, { "epoch": 0.6253398586188146, "grad_norm": 1.078125, "learning_rate": 1.9256567915036335e-05, "loss": 0.1733, "step": 2300 }, { "epoch": 0.6253398586188146, "eval_loss": 0.15731249749660492, "eval_runtime": 95.4546, "eval_samples_per_second": 5.238, "eval_steps_per_second": 0.168, "step": 2300 }, { "epoch": 0.6280587275693311, "grad_norm": 1.40625, "learning_rate": 1.911682504192286e-05, "loss": 0.1706, "step": 2310 }, { "epoch": 0.6307775965198478, "grad_norm": 3.078125, "learning_rate": 1.897708216880939e-05, "loss": 0.1856, "step": 2320 }, { "epoch": 0.6334964654703643, "grad_norm": 1.375, "learning_rate": 1.8837339295695922e-05, "loss": 0.1852, "step": 2330 }, { "epoch": 0.636215334420881, "grad_norm": 2.3125, "learning_rate": 1.869759642258245e-05, "loss": 0.1748, "step": 2340 }, { "epoch": 0.6389342033713975, "grad_norm": 1.2890625, "learning_rate": 1.8557853549468977e-05, "loss": 0.1576, "step": 2350 }, { "epoch": 0.6389342033713975, "eval_loss": 0.15459375083446503, "eval_runtime": 94.7107, "eval_samples_per_second": 5.279, "eval_steps_per_second": 0.169, "step": 2350 }, { "epoch": 0.641653072321914, "grad_norm": 1.4921875, "learning_rate": 1.8418110676355507e-05, "loss": 0.1657, "step": 2360 }, { "epoch": 0.6443719412724307, "grad_norm": 1.8359375, "learning_rate": 1.8278367803242036e-05, "loss": 0.1725, "step": 2370 }, { "epoch": 0.6470908102229472, "grad_norm": 1.90625, "learning_rate": 1.8138624930128565e-05, "loss": 0.1667, "step": 2380 }, { "epoch": 0.6498096791734639, "grad_norm": 2.203125, "learning_rate": 1.7998882057015094e-05, "loss": 0.1468, "step": 2390 }, { "epoch": 0.6525285481239804, "grad_norm": 2.3125, "learning_rate": 1.785913918390162e-05, "loss": 0.1852, "step": 2400 }, { "epoch": 0.6525285481239804, "eval_loss": 0.14464062452316284, "eval_runtime": 109.621, "eval_samples_per_second": 4.561, "eval_steps_per_second": 0.146, "step": 2400 }, { "epoch": 0.655247417074497, "grad_norm": 2.015625, "learning_rate": 1.771939631078815e-05, "loss": 0.1942, "step": 2410 }, { "epoch": 0.6579662860250136, "grad_norm": 1.9453125, "learning_rate": 1.757965343767468e-05, "loss": 0.1513, "step": 2420 }, { "epoch": 0.6606851549755302, "grad_norm": 1.1875, "learning_rate": 1.7439910564561208e-05, "loss": 0.1445, "step": 2430 }, { "epoch": 0.6634040239260468, "grad_norm": 3.171875, "learning_rate": 1.7300167691447737e-05, "loss": 0.1874, "step": 2440 }, { "epoch": 0.6661228928765633, "grad_norm": 2.5625, "learning_rate": 1.7160424818334266e-05, "loss": 0.1697, "step": 2450 }, { "epoch": 0.6661228928765633, "eval_loss": 0.14278124272823334, "eval_runtime": 99.7567, "eval_samples_per_second": 5.012, "eval_steps_per_second": 0.16, "step": 2450 }, { "epoch": 0.6688417618270799, "grad_norm": 1.6171875, "learning_rate": 1.7020681945220796e-05, "loss": 0.1505, "step": 2460 }, { "epoch": 0.6715606307775965, "grad_norm": 1.1875, "learning_rate": 1.6880939072107325e-05, "loss": 0.2015, "step": 2470 }, { "epoch": 0.6742794997281131, "grad_norm": 1.5625, "learning_rate": 1.6741196198993854e-05, "loss": 0.1782, "step": 2480 }, { "epoch": 0.6769983686786297, "grad_norm": 3.140625, "learning_rate": 1.660145332588038e-05, "loss": 0.159, "step": 2490 }, { "epoch": 0.6797172376291463, "grad_norm": 1.265625, "learning_rate": 1.646171045276691e-05, "loss": 0.2079, "step": 2500 }, { "epoch": 0.6797172376291463, "eval_loss": 0.15110155940055847, "eval_runtime": 101.2462, "eval_samples_per_second": 4.938, "eval_steps_per_second": 0.158, "step": 2500 }, { "epoch": 0.6824361065796629, "grad_norm": 2.046875, "learning_rate": 1.632196757965344e-05, "loss": 0.1347, "step": 2510 }, { "epoch": 0.6851549755301795, "grad_norm": 1.140625, "learning_rate": 1.6182224706539968e-05, "loss": 0.1575, "step": 2520 }, { "epoch": 0.687873844480696, "grad_norm": 1.8046875, "learning_rate": 1.6042481833426497e-05, "loss": 0.1613, "step": 2530 }, { "epoch": 0.6905927134312126, "grad_norm": 2.015625, "learning_rate": 1.5902738960313023e-05, "loss": 0.1625, "step": 2540 }, { "epoch": 0.6933115823817292, "grad_norm": 3.421875, "learning_rate": 1.5762996087199552e-05, "loss": 0.1748, "step": 2550 }, { "epoch": 0.6933115823817292, "eval_loss": 0.15357030928134918, "eval_runtime": 102.5541, "eval_samples_per_second": 4.875, "eval_steps_per_second": 0.156, "step": 2550 }, { "epoch": 0.6960304513322458, "grad_norm": 2.3125, "learning_rate": 1.562325321408608e-05, "loss": 0.1767, "step": 2560 }, { "epoch": 0.6987493202827624, "grad_norm": 1.734375, "learning_rate": 1.5483510340972614e-05, "loss": 0.1507, "step": 2570 }, { "epoch": 0.7014681892332789, "grad_norm": 2.0625, "learning_rate": 1.534376746785914e-05, "loss": 0.13, "step": 2580 }, { "epoch": 0.7041870581837956, "grad_norm": 1.3671875, "learning_rate": 1.5204024594745669e-05, "loss": 0.1576, "step": 2590 }, { "epoch": 0.7069059271343121, "grad_norm": 1.0546875, "learning_rate": 1.5064281721632198e-05, "loss": 0.173, "step": 2600 }, { "epoch": 0.7069059271343121, "eval_loss": 0.15370312333106995, "eval_runtime": 102.0222, "eval_samples_per_second": 4.901, "eval_steps_per_second": 0.157, "step": 2600 }, { "epoch": 0.7096247960848288, "grad_norm": 0.62890625, "learning_rate": 1.4924538848518726e-05, "loss": 0.1421, "step": 2610 }, { "epoch": 0.7123436650353453, "grad_norm": 1.8125, "learning_rate": 1.4784795975405255e-05, "loss": 0.1548, "step": 2620 }, { "epoch": 0.7150625339858618, "grad_norm": 0.96484375, "learning_rate": 1.4645053102291784e-05, "loss": 0.1422, "step": 2630 }, { "epoch": 0.7177814029363785, "grad_norm": 1.8125, "learning_rate": 1.4505310229178312e-05, "loss": 0.1577, "step": 2640 }, { "epoch": 0.720500271886895, "grad_norm": 1.046875, "learning_rate": 1.4365567356064841e-05, "loss": 0.1542, "step": 2650 }, { "epoch": 0.720500271886895, "eval_loss": 0.16090625524520874, "eval_runtime": 105.0015, "eval_samples_per_second": 4.762, "eval_steps_per_second": 0.152, "step": 2650 }, { "epoch": 0.7232191408374117, "grad_norm": 3.53125, "learning_rate": 1.4225824482951369e-05, "loss": 0.1655, "step": 2660 }, { "epoch": 0.7259380097879282, "grad_norm": 0.96484375, "learning_rate": 1.4086081609837898e-05, "loss": 0.1508, "step": 2670 }, { "epoch": 0.7286568787384448, "grad_norm": 1.9609375, "learning_rate": 1.3946338736724427e-05, "loss": 0.1455, "step": 2680 }, { "epoch": 0.7313757476889614, "grad_norm": 2.171875, "learning_rate": 1.3806595863610958e-05, "loss": 0.1408, "step": 2690 }, { "epoch": 0.734094616639478, "grad_norm": 1.125, "learning_rate": 1.3666852990497486e-05, "loss": 0.1562, "step": 2700 }, { "epoch": 0.734094616639478, "eval_loss": 0.15767186880111694, "eval_runtime": 98.7176, "eval_samples_per_second": 5.065, "eval_steps_per_second": 0.162, "step": 2700 }, { "epoch": 0.7368134855899946, "grad_norm": 0.875, "learning_rate": 1.3527110117384015e-05, "loss": 0.1621, "step": 2710 }, { "epoch": 0.7395323545405111, "grad_norm": 1.578125, "learning_rate": 1.3387367244270544e-05, "loss": 0.1673, "step": 2720 }, { "epoch": 0.7422512234910277, "grad_norm": 1.1640625, "learning_rate": 1.3247624371157072e-05, "loss": 0.1708, "step": 2730 }, { "epoch": 0.7449700924415443, "grad_norm": 1.4765625, "learning_rate": 1.3107881498043601e-05, "loss": 0.1295, "step": 2740 }, { "epoch": 0.7476889613920609, "grad_norm": 1.9765625, "learning_rate": 1.2968138624930128e-05, "loss": 0.1518, "step": 2750 }, { "epoch": 0.7476889613920609, "eval_loss": 0.1566796898841858, "eval_runtime": 99.834, "eval_samples_per_second": 5.008, "eval_steps_per_second": 0.16, "step": 2750 }, { "epoch": 0.7504078303425775, "grad_norm": 3.34375, "learning_rate": 1.2828395751816658e-05, "loss": 0.2034, "step": 2760 }, { "epoch": 0.7531266992930941, "grad_norm": 1.9765625, "learning_rate": 1.2688652878703185e-05, "loss": 0.1786, "step": 2770 }, { "epoch": 0.7558455682436107, "grad_norm": 1.5, "learning_rate": 1.2548910005589715e-05, "loss": 0.1941, "step": 2780 }, { "epoch": 0.7585644371941273, "grad_norm": 2.109375, "learning_rate": 1.2409167132476244e-05, "loss": 0.1586, "step": 2790 }, { "epoch": 0.7612833061446438, "grad_norm": 2.59375, "learning_rate": 1.2269424259362773e-05, "loss": 0.1914, "step": 2800 }, { "epoch": 0.7612833061446438, "eval_loss": 0.14214062690734863, "eval_runtime": 99.6815, "eval_samples_per_second": 5.016, "eval_steps_per_second": 0.161, "step": 2800 }, { "epoch": 0.7640021750951604, "grad_norm": 2.6875, "learning_rate": 1.2129681386249302e-05, "loss": 0.1537, "step": 2810 }, { "epoch": 0.766721044045677, "grad_norm": 1.5234375, "learning_rate": 1.198993851313583e-05, "loss": 0.1528, "step": 2820 }, { "epoch": 0.7694399129961936, "grad_norm": 1.5234375, "learning_rate": 1.1850195640022359e-05, "loss": 0.1504, "step": 2830 }, { "epoch": 0.7721587819467102, "grad_norm": 1.875, "learning_rate": 1.1710452766908888e-05, "loss": 0.1599, "step": 2840 }, { "epoch": 0.7748776508972267, "grad_norm": 2.21875, "learning_rate": 1.1570709893795418e-05, "loss": 0.1833, "step": 2850 }, { "epoch": 0.7748776508972267, "eval_loss": 0.15465624630451202, "eval_runtime": 98.484, "eval_samples_per_second": 5.077, "eval_steps_per_second": 0.162, "step": 2850 }, { "epoch": 0.7775965198477434, "grad_norm": 2.796875, "learning_rate": 1.1430967020681945e-05, "loss": 0.1537, "step": 2860 }, { "epoch": 0.7803153887982599, "grad_norm": 2.796875, "learning_rate": 1.1291224147568474e-05, "loss": 0.1916, "step": 2870 }, { "epoch": 0.7830342577487766, "grad_norm": 1.875, "learning_rate": 1.1151481274455004e-05, "loss": 0.1445, "step": 2880 }, { "epoch": 0.7857531266992931, "grad_norm": 3.0625, "learning_rate": 1.1011738401341531e-05, "loss": 0.187, "step": 2890 }, { "epoch": 0.7884719956498096, "grad_norm": 1.34375, "learning_rate": 1.0871995528228062e-05, "loss": 0.1492, "step": 2900 }, { "epoch": 0.7884719956498096, "eval_loss": 0.14948438107967377, "eval_runtime": 117.0767, "eval_samples_per_second": 4.271, "eval_steps_per_second": 0.137, "step": 2900 }, { "epoch": 0.7911908646003263, "grad_norm": 1.15625, "learning_rate": 1.073225265511459e-05, "loss": 0.211, "step": 2910 }, { "epoch": 0.7939097335508428, "grad_norm": 1.2109375, "learning_rate": 1.0592509782001119e-05, "loss": 0.1555, "step": 2920 }, { "epoch": 0.7966286025013595, "grad_norm": 1.578125, "learning_rate": 1.0452766908887646e-05, "loss": 0.1666, "step": 2930 }, { "epoch": 0.799347471451876, "grad_norm": 1.8046875, "learning_rate": 1.0313024035774176e-05, "loss": 0.1936, "step": 2940 }, { "epoch": 0.8020663404023926, "grad_norm": 2.421875, "learning_rate": 1.0173281162660705e-05, "loss": 0.1484, "step": 2950 }, { "epoch": 0.8020663404023926, "eval_loss": 0.1609531193971634, "eval_runtime": 104.0405, "eval_samples_per_second": 4.806, "eval_steps_per_second": 0.154, "step": 2950 }, { "epoch": 0.8047852093529092, "grad_norm": 1.34375, "learning_rate": 1.0033538289547234e-05, "loss": 0.1312, "step": 2960 }, { "epoch": 0.8075040783034257, "grad_norm": 1.265625, "learning_rate": 9.893795416433763e-06, "loss": 0.1239, "step": 2970 }, { "epoch": 0.8102229472539424, "grad_norm": 2.265625, "learning_rate": 9.754052543320291e-06, "loss": 0.1733, "step": 2980 }, { "epoch": 0.8129418162044589, "grad_norm": 1.359375, "learning_rate": 9.61430967020682e-06, "loss": 0.1964, "step": 2990 }, { "epoch": 0.8156606851549756, "grad_norm": 1.46875, "learning_rate": 9.474566797093348e-06, "loss": 0.1463, "step": 3000 }, { "epoch": 0.8156606851549756, "eval_loss": 0.15464062988758087, "eval_runtime": 103.2751, "eval_samples_per_second": 4.841, "eval_steps_per_second": 0.155, "step": 3000 }, { "epoch": 0.8183795541054921, "grad_norm": 1.84375, "learning_rate": 9.334823923979877e-06, "loss": 0.1547, "step": 3010 }, { "epoch": 0.8210984230560087, "grad_norm": 1.7265625, "learning_rate": 9.195081050866406e-06, "loss": 0.1783, "step": 3020 }, { "epoch": 0.8238172920065253, "grad_norm": 1.609375, "learning_rate": 9.055338177752935e-06, "loss": 0.1595, "step": 3030 }, { "epoch": 0.8265361609570419, "grad_norm": 1.296875, "learning_rate": 8.915595304639463e-06, "loss": 0.1625, "step": 3040 }, { "epoch": 0.8292550299075585, "grad_norm": 1.671875, "learning_rate": 8.775852431525992e-06, "loss": 0.1821, "step": 3050 }, { "epoch": 0.8292550299075585, "eval_loss": 0.15335936844348907, "eval_runtime": 100.4729, "eval_samples_per_second": 4.976, "eval_steps_per_second": 0.159, "step": 3050 }, { "epoch": 0.831973898858075, "grad_norm": 3.0625, "learning_rate": 8.636109558412521e-06, "loss": 0.1855, "step": 3060 }, { "epoch": 0.8346927678085916, "grad_norm": 1.328125, "learning_rate": 8.496366685299049e-06, "loss": 0.1513, "step": 3070 }, { "epoch": 0.8374116367591082, "grad_norm": 1.0859375, "learning_rate": 8.35662381218558e-06, "loss": 0.1566, "step": 3080 }, { "epoch": 0.8401305057096248, "grad_norm": 1.90625, "learning_rate": 8.216880939072108e-06, "loss": 0.1611, "step": 3090 }, { "epoch": 0.8428493746601414, "grad_norm": 1.9453125, "learning_rate": 8.077138065958637e-06, "loss": 0.1432, "step": 3100 }, { "epoch": 0.8428493746601414, "eval_loss": 0.15146875381469727, "eval_runtime": 100.1544, "eval_samples_per_second": 4.992, "eval_steps_per_second": 0.16, "step": 3100 }, { "epoch": 0.845568243610658, "grad_norm": 1.8359375, "learning_rate": 7.937395192845164e-06, "loss": 0.1737, "step": 3110 }, { "epoch": 0.8482871125611745, "grad_norm": 1.0703125, "learning_rate": 7.797652319731694e-06, "loss": 0.1423, "step": 3120 }, { "epoch": 0.8510059815116912, "grad_norm": 2.046875, "learning_rate": 7.657909446618223e-06, "loss": 0.1262, "step": 3130 }, { "epoch": 0.8537248504622077, "grad_norm": 0.8515625, "learning_rate": 7.518166573504752e-06, "loss": 0.1402, "step": 3140 }, { "epoch": 0.8564437194127243, "grad_norm": 1.3125, "learning_rate": 7.3784237003912805e-06, "loss": 0.1803, "step": 3150 }, { "epoch": 0.8564437194127243, "eval_loss": 0.13735155761241913, "eval_runtime": 98.6763, "eval_samples_per_second": 5.067, "eval_steps_per_second": 0.162, "step": 3150 }, { "epoch": 0.8591625883632409, "grad_norm": 1.2265625, "learning_rate": 7.238680827277809e-06, "loss": 0.1801, "step": 3160 }, { "epoch": 0.8618814573137574, "grad_norm": 3.9375, "learning_rate": 7.098937954164338e-06, "loss": 0.1819, "step": 3170 }, { "epoch": 0.8646003262642741, "grad_norm": 1.875, "learning_rate": 6.9591950810508665e-06, "loss": 0.1458, "step": 3180 }, { "epoch": 0.8673191952147906, "grad_norm": 2.9375, "learning_rate": 6.819452207937395e-06, "loss": 0.168, "step": 3190 }, { "epoch": 0.8700380641653073, "grad_norm": 1.9609375, "learning_rate": 6.679709334823925e-06, "loss": 0.1505, "step": 3200 }, { "epoch": 0.8700380641653073, "eval_loss": 0.14777344465255737, "eval_runtime": 99.0119, "eval_samples_per_second": 5.05, "eval_steps_per_second": 0.162, "step": 3200 }, { "epoch": 0.8727569331158238, "grad_norm": 9.0, "learning_rate": 6.539966461710453e-06, "loss": 0.1582, "step": 3210 }, { "epoch": 0.8754758020663403, "grad_norm": 1.890625, "learning_rate": 6.400223588596982e-06, "loss": 0.176, "step": 3220 }, { "epoch": 0.878194671016857, "grad_norm": 1.25, "learning_rate": 6.26048071548351e-06, "loss": 0.1346, "step": 3230 }, { "epoch": 0.8809135399673735, "grad_norm": 1.953125, "learning_rate": 6.1207378423700394e-06, "loss": 0.1424, "step": 3240 }, { "epoch": 0.8836324089178902, "grad_norm": 1.8671875, "learning_rate": 5.980994969256568e-06, "loss": 0.1413, "step": 3250 }, { "epoch": 0.8836324089178902, "eval_loss": 0.14879687130451202, "eval_runtime": 95.0133, "eval_samples_per_second": 5.262, "eval_steps_per_second": 0.168, "step": 3250 }, { "epoch": 0.8863512778684067, "grad_norm": 1.8359375, "learning_rate": 5.841252096143097e-06, "loss": 0.1432, "step": 3260 }, { "epoch": 0.8890701468189234, "grad_norm": 2.203125, "learning_rate": 5.701509223029626e-06, "loss": 0.1352, "step": 3270 }, { "epoch": 0.8917890157694399, "grad_norm": 1.578125, "learning_rate": 5.561766349916155e-06, "loss": 0.1572, "step": 3280 }, { "epoch": 0.8945078847199565, "grad_norm": 2.078125, "learning_rate": 5.422023476802683e-06, "loss": 0.1746, "step": 3290 }, { "epoch": 0.8972267536704731, "grad_norm": 1.5703125, "learning_rate": 5.282280603689212e-06, "loss": 0.1812, "step": 3300 }, { "epoch": 0.8972267536704731, "eval_loss": 0.1419062465429306, "eval_runtime": 95.5103, "eval_samples_per_second": 5.235, "eval_steps_per_second": 0.168, "step": 3300 }, { "epoch": 0.8999456226209896, "grad_norm": 1.2265625, "learning_rate": 5.142537730575741e-06, "loss": 0.1398, "step": 3310 }, { "epoch": 0.9026644915715063, "grad_norm": 2.28125, "learning_rate": 5.002794857462269e-06, "loss": 0.143, "step": 3320 }, { "epoch": 0.9053833605220228, "grad_norm": 2.34375, "learning_rate": 4.863051984348798e-06, "loss": 0.1623, "step": 3330 }, { "epoch": 0.9081022294725394, "grad_norm": 3.21875, "learning_rate": 4.723309111235328e-06, "loss": 0.1893, "step": 3340 }, { "epoch": 0.910821098423056, "grad_norm": 1.5703125, "learning_rate": 4.583566238121856e-06, "loss": 0.1617, "step": 3350 }, { "epoch": 0.910821098423056, "eval_loss": 0.15714062750339508, "eval_runtime": 94.6634, "eval_samples_per_second": 5.282, "eval_steps_per_second": 0.169, "step": 3350 }, { "epoch": 0.9135399673735726, "grad_norm": 1.7734375, "learning_rate": 4.443823365008385e-06, "loss": 0.1508, "step": 3360 }, { "epoch": 0.9162588363240892, "grad_norm": 5.09375, "learning_rate": 4.304080491894914e-06, "loss": 0.1355, "step": 3370 }, { "epoch": 0.9189777052746058, "grad_norm": 2.09375, "learning_rate": 4.164337618781442e-06, "loss": 0.1457, "step": 3380 }, { "epoch": 0.9216965742251223, "grad_norm": 1.296875, "learning_rate": 4.024594745667971e-06, "loss": 0.1391, "step": 3390 }, { "epoch": 0.924415443175639, "grad_norm": 2.046875, "learning_rate": 3.8848518725545e-06, "loss": 0.1853, "step": 3400 }, { "epoch": 0.924415443175639, "eval_loss": 0.14689843356609344, "eval_runtime": 109.943, "eval_samples_per_second": 4.548, "eval_steps_per_second": 0.146, "step": 3400 }, { "epoch": 0.9271343121261555, "grad_norm": 2.265625, "learning_rate": 3.7451089994410285e-06, "loss": 0.1904, "step": 3410 }, { "epoch": 0.9298531810766721, "grad_norm": 1.34375, "learning_rate": 3.6053661263275578e-06, "loss": 0.1342, "step": 3420 }, { "epoch": 0.9325720500271887, "grad_norm": 2.15625, "learning_rate": 3.465623253214086e-06, "loss": 0.1583, "step": 3430 }, { "epoch": 0.9352909189777052, "grad_norm": 1.859375, "learning_rate": 3.325880380100615e-06, "loss": 0.1763, "step": 3440 }, { "epoch": 0.9380097879282219, "grad_norm": 1.703125, "learning_rate": 3.1861375069871442e-06, "loss": 0.1367, "step": 3450 }, { "epoch": 0.9380097879282219, "eval_loss": 0.1392187476158142, "eval_runtime": 96.262, "eval_samples_per_second": 5.194, "eval_steps_per_second": 0.166, "step": 3450 }, { "epoch": 0.9407286568787384, "grad_norm": 1.15625, "learning_rate": 3.0463946338736726e-06, "loss": 0.184, "step": 3460 }, { "epoch": 0.9434475258292551, "grad_norm": 1.2734375, "learning_rate": 2.9066517607602015e-06, "loss": 0.1794, "step": 3470 }, { "epoch": 0.9461663947797716, "grad_norm": 2.9375, "learning_rate": 2.7669088876467303e-06, "loss": 0.1705, "step": 3480 }, { "epoch": 0.9488852637302883, "grad_norm": 1.0078125, "learning_rate": 2.627166014533259e-06, "loss": 0.1846, "step": 3490 }, { "epoch": 0.9516041326808048, "grad_norm": 0.97265625, "learning_rate": 2.487423141419788e-06, "loss": 0.1917, "step": 3500 }, { "epoch": 0.9516041326808048, "eval_loss": 0.14869531989097595, "eval_runtime": 96.4372, "eval_samples_per_second": 5.185, "eval_steps_per_second": 0.166, "step": 3500 } ], "logging_steps": 10, "max_steps": 3678, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.05068095973888e+18, "train_batch_size": 32, "trial_name": null, "trial_params": null }