{ "best_metric": null, "best_model_checkpoint": null, "epoch": 500.0, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 2.5, "grad_norm": 8.514540942565027, "learning_rate": 5e-06, "loss": 0.585, "step": 5 }, { "epoch": 5.0, "grad_norm": 3.9267725944427414, "learning_rate": 1e-05, "loss": 0.4444, "step": 10 }, { "epoch": 7.5, "grad_norm": 1.359630389285586, "learning_rate": 1.5e-05, "loss": 0.1918, "step": 15 }, { "epoch": 10.0, "grad_norm": 0.8675984069699776, "learning_rate": 2e-05, "loss": 0.0255, "step": 20 }, { "epoch": 12.5, "grad_norm": 1.3166063968433408, "learning_rate": 2.5e-05, "loss": 0.0094, "step": 25 }, { "epoch": 15.0, "grad_norm": 1.4298321029000574, "learning_rate": 3e-05, "loss": 0.0066, "step": 30 }, { "epoch": 17.5, "grad_norm": 0.39016346214571596, "learning_rate": 3.5e-05, "loss": 0.0054, "step": 35 }, { "epoch": 20.0, "grad_norm": 0.2790201303316679, "learning_rate": 4e-05, "loss": 0.0062, "step": 40 }, { "epoch": 22.5, "grad_norm": 0.08501619599523778, "learning_rate": 4.5e-05, "loss": 0.004, "step": 45 }, { "epoch": 25.0, "grad_norm": 0.062067934755517584, "learning_rate": 5e-05, "loss": 0.0029, "step": 50 }, { "epoch": 27.5, "grad_norm": 0.052666305930186126, "learning_rate": 4.9996924362330555e-05, "loss": 0.0028, "step": 55 }, { "epoch": 30.0, "grad_norm": 0.03938136111188392, "learning_rate": 4.998769829017084e-05, "loss": 0.0028, "step": 60 }, { "epoch": 32.5, "grad_norm": 0.03394672644933696, "learning_rate": 4.997232430583686e-05, "loss": 0.0028, "step": 65 }, { "epoch": 35.0, "grad_norm": 0.037661243334742, "learning_rate": 4.995080661242243e-05, "loss": 0.0028, "step": 70 }, { "epoch": 37.5, "grad_norm": 0.039184906370568255, "learning_rate": 4.992315109265007e-05, "loss": 0.0027, "step": 75 }, { "epoch": 40.0, "grad_norm": 0.04435589843124744, "learning_rate": 4.988936530726276e-05, "loss": 0.0028, "step": 80 }, { "epoch": 42.5, "grad_norm": 0.02657151726999846, "learning_rate": 4.984945849295686e-05, "loss": 0.0026, "step": 85 }, { "epoch": 45.0, "grad_norm": 0.03472385835097059, "learning_rate": 4.980344155985695e-05, "loss": 0.0027, "step": 90 }, { "epoch": 47.5, "grad_norm": 0.031070975559645563, "learning_rate": 4.975132708853304e-05, "loss": 0.0026, "step": 95 }, { "epoch": 50.0, "grad_norm": 0.03730805984645713, "learning_rate": 4.9693129326561254e-05, "loss": 0.0027, "step": 100 }, { "epoch": 52.5, "grad_norm": 0.030401417138752773, "learning_rate": 4.96288641846286e-05, "loss": 0.0027, "step": 105 }, { "epoch": 55.0, "grad_norm": 0.030056720425659458, "learning_rate": 4.955854923218321e-05, "loss": 0.0027, "step": 110 }, { "epoch": 57.5, "grad_norm": 0.02714242610959012, "learning_rate": 4.948220369263099e-05, "loss": 0.0025, "step": 115 }, { "epoch": 60.0, "grad_norm": 0.021468304046160087, "learning_rate": 4.939984843808013e-05, "loss": 0.0026, "step": 120 }, { "epoch": 62.5, "grad_norm": 0.027579476441235425, "learning_rate": 4.931150598363494e-05, "loss": 0.0027, "step": 125 }, { "epoch": 65.0, "grad_norm": 0.041471384473431336, "learning_rate": 4.921720048124035e-05, "loss": 0.0028, "step": 130 }, { "epoch": 67.5, "grad_norm": 0.030743428074535287, "learning_rate": 4.9116957713079084e-05, "loss": 0.0027, "step": 135 }, { "epoch": 70.0, "grad_norm": 0.03200739761669791, "learning_rate": 4.901080508452305e-05, "loss": 0.0027, "step": 140 }, { "epoch": 72.5, "grad_norm": 0.02266656589663142, "learning_rate": 4.889877161664096e-05, "loss": 0.0027, "step": 145 }, { "epoch": 75.0, "grad_norm": 0.03858968870626217, "learning_rate": 4.878088793826428e-05, "loss": 0.0026, "step": 150 }, { "epoch": 77.5, "grad_norm": 0.0436229252420695, "learning_rate": 4.865718627761363e-05, "loss": 0.0027, "step": 155 }, { "epoch": 80.0, "grad_norm": 0.021505547895787954, "learning_rate": 4.8527700453487873e-05, "loss": 0.0026, "step": 160 }, { "epoch": 82.5, "grad_norm": 0.020985544413640905, "learning_rate": 4.839246586601841e-05, "loss": 0.0027, "step": 165 }, { "epoch": 85.0, "grad_norm": 0.023894609006279273, "learning_rate": 4.825151948699116e-05, "loss": 0.0028, "step": 170 }, { "epoch": 87.5, "grad_norm": 0.02603189523521318, "learning_rate": 4.8104899849738795e-05, "loss": 0.0026, "step": 175 }, { "epoch": 90.0, "grad_norm": 0.033072587670050196, "learning_rate": 4.795264703860616e-05, "loss": 0.0027, "step": 180 }, { "epoch": 92.5, "grad_norm": 0.029455707311945873, "learning_rate": 4.779480267799158e-05, "loss": 0.0027, "step": 185 }, { "epoch": 95.0, "grad_norm": 0.031066744462249135, "learning_rate": 4.763140992096718e-05, "loss": 0.0027, "step": 190 }, { "epoch": 97.5, "grad_norm": 0.03179467575788789, "learning_rate": 4.7462513437481266e-05, "loss": 0.0027, "step": 195 }, { "epoch": 100.0, "grad_norm": 0.04391757622955937, "learning_rate": 4.7288159402146e-05, "loss": 0.0027, "step": 200 }, { "epoch": 102.5, "grad_norm": 0.029394885252518814, "learning_rate": 4.7108395481613736e-05, "loss": 0.0028, "step": 205 }, { "epoch": 105.0, "grad_norm": 0.027320800240786685, "learning_rate": 4.692327082154542e-05, "loss": 0.0027, "step": 210 }, { "epoch": 107.5, "grad_norm": 0.030873484414188097, "learning_rate": 4.6732836033174634e-05, "loss": 0.0027, "step": 215 }, { "epoch": 110.0, "grad_norm": 0.048697733723929194, "learning_rate": 4.653714317947105e-05, "loss": 0.0027, "step": 220 }, { "epoch": 112.5, "grad_norm": 0.025217762723189756, "learning_rate": 4.6336245760906896e-05, "loss": 0.0026, "step": 225 }, { "epoch": 115.0, "grad_norm": 0.034325242650264584, "learning_rate": 4.6130198700830455e-05, "loss": 0.0027, "step": 230 }, { "epoch": 117.5, "grad_norm": 0.0392844047749773, "learning_rate": 4.591905833045059e-05, "loss": 0.0027, "step": 235 }, { "epoch": 120.0, "grad_norm": 0.02965200107252582, "learning_rate": 4.570288237343632e-05, "loss": 0.0026, "step": 240 }, { "epoch": 122.5, "grad_norm": 0.020210067730727373, "learning_rate": 4.5481729930135805e-05, "loss": 0.0026, "step": 245 }, { "epoch": 125.0, "grad_norm": 0.027623657423726022, "learning_rate": 4.5255661461418854e-05, "loss": 0.0026, "step": 250 }, { "epoch": 127.5, "grad_norm": 0.023255879069235798, "learning_rate": 4.502473877214754e-05, "loss": 0.0027, "step": 255 }, { "epoch": 130.0, "grad_norm": 0.014768241964023569, "learning_rate": 4.478902499427933e-05, "loss": 0.0027, "step": 260 }, { "epoch": 132.5, "grad_norm": 0.03668790355624439, "learning_rate": 4.454858456960754e-05, "loss": 0.0027, "step": 265 }, { "epoch": 135.0, "grad_norm": 0.028968111779672595, "learning_rate": 4.430348323214347e-05, "loss": 0.0026, "step": 270 }, { "epoch": 137.5, "grad_norm": 0.02287695653834384, "learning_rate": 4.4053787990145465e-05, "loss": 0.0027, "step": 275 }, { "epoch": 140.0, "grad_norm": 0.014260556347040123, "learning_rate": 4.379956710779951e-05, "loss": 0.0027, "step": 280 }, { "epoch": 142.5, "grad_norm": 0.029771289364298407, "learning_rate": 4.3540890086556435e-05, "loss": 0.0026, "step": 285 }, { "epoch": 145.0, "grad_norm": 0.03788961751665883, "learning_rate": 4.327782764613099e-05, "loss": 0.0027, "step": 290 }, { "epoch": 147.5, "grad_norm": 0.03971119187487509, "learning_rate": 4.301045170516773e-05, "loss": 0.0026, "step": 295 }, { "epoch": 150.0, "grad_norm": 0.03370808514177945, "learning_rate": 4.2738835361579175e-05, "loss": 0.0026, "step": 300 }, { "epoch": 152.5, "grad_norm": 0.04042652697228835, "learning_rate": 4.2463052872561584e-05, "loss": 0.0027, "step": 305 }, { "epoch": 155.0, "grad_norm": 0.03521656241677526, "learning_rate": 4.2183179634293794e-05, "loss": 0.0028, "step": 310 }, { "epoch": 157.5, "grad_norm": 0.03092781528226213, "learning_rate": 4.1899292161324627e-05, "loss": 0.0026, "step": 315 }, { "epoch": 160.0, "grad_norm": 0.020310997173564212, "learning_rate": 4.1611468065654586e-05, "loss": 0.0026, "step": 320 }, { "epoch": 162.5, "grad_norm": 0.015790567180341695, "learning_rate": 4.1319786035517534e-05, "loss": 0.0027, "step": 325 }, { "epoch": 165.0, "grad_norm": 0.0194147951616345, "learning_rate": 4.1024325813868065e-05, "loss": 0.0027, "step": 330 }, { "epoch": 167.5, "grad_norm": 0.030730482667734805, "learning_rate": 4.072516817658065e-05, "loss": 0.0027, "step": 335 }, { "epoch": 170.0, "grad_norm": 0.029837731953442564, "learning_rate": 4.0422394910366236e-05, "loss": 0.0026, "step": 340 }, { "epoch": 172.5, "grad_norm": 0.02492166177013941, "learning_rate": 4.0116088790412645e-05, "loss": 0.0026, "step": 345 }, { "epoch": 175.0, "grad_norm": 0.022790129612325603, "learning_rate": 3.980633355775461e-05, "loss": 0.0027, "step": 350 }, { "epoch": 177.5, "grad_norm": 0.0367286949474086, "learning_rate": 3.949321389637986e-05, "loss": 0.0027, "step": 355 }, { "epoch": 180.0, "grad_norm": 0.026449210914021956, "learning_rate": 3.917681541007734e-05, "loss": 0.0025, "step": 360 }, { "epoch": 182.5, "grad_norm": 0.025906355217603764, "learning_rate": 3.885722459903399e-05, "loss": 0.0027, "step": 365 }, { "epoch": 185.0, "grad_norm": 0.013510506972149507, "learning_rate": 3.853452883618644e-05, "loss": 0.0027, "step": 370 }, { "epoch": 187.5, "grad_norm": 0.02126782273762181, "learning_rate": 3.8208816343334156e-05, "loss": 0.0027, "step": 375 }, { "epoch": 190.0, "grad_norm": 0.02778025887186039, "learning_rate": 3.788017616702048e-05, "loss": 0.0027, "step": 380 }, { "epoch": 192.5, "grad_norm": 0.02023158670094667, "learning_rate": 3.754869815418815e-05, "loss": 0.0026, "step": 385 }, { "epoch": 195.0, "grad_norm": 0.031240526748051056, "learning_rate": 3.721447292761609e-05, "loss": 0.0027, "step": 390 }, { "epoch": 197.5, "grad_norm": 0.02268049791605829, "learning_rate": 3.687759186114403e-05, "loss": 0.0026, "step": 395 }, { "epoch": 200.0, "grad_norm": 0.02998474683024178, "learning_rate": 3.6538147054691817e-05, "loss": 0.0027, "step": 400 }, { "epoch": 202.5, "grad_norm": 0.026430986104193692, "learning_rate": 3.619623130908018e-05, "loss": 0.0027, "step": 405 }, { "epoch": 205.0, "grad_norm": 0.010605276850508178, "learning_rate": 3.5851938100659964e-05, "loss": 0.0027, "step": 410 }, { "epoch": 207.5, "grad_norm": 0.02068209169464807, "learning_rate": 3.550536155575662e-05, "loss": 0.0027, "step": 415 }, { "epoch": 210.0, "grad_norm": 0.03300339570363238, "learning_rate": 3.515659642493697e-05, "loss": 0.0026, "step": 420 }, { "epoch": 212.5, "grad_norm": 0.026016657903437497, "learning_rate": 3.480573805710538e-05, "loss": 0.0027, "step": 425 }, { "epoch": 215.0, "grad_norm": 0.024770792616580162, "learning_rate": 3.4452882373436316e-05, "loss": 0.0026, "step": 430 }, { "epoch": 217.5, "grad_norm": 0.0265444448484458, "learning_rate": 3.4098125841150466e-05, "loss": 0.0026, "step": 435 }, { "epoch": 220.0, "grad_norm": 0.01924921498743014, "learning_rate": 3.37415654471415e-05, "loss": 0.0026, "step": 440 }, { "epoch": 222.5, "grad_norm": 0.027603952267821173, "learning_rate": 3.3383298671460944e-05, "loss": 0.0026, "step": 445 }, { "epoch": 225.0, "grad_norm": 0.0254709928199966, "learning_rate": 3.3023423460667985e-05, "loss": 0.0026, "step": 450 }, { "epoch": 227.5, "grad_norm": 0.031183288661866662, "learning_rate": 3.2662038201051914e-05, "loss": 0.0027, "step": 455 }, { "epoch": 230.0, "grad_norm": 0.03425285898227886, "learning_rate": 3.2299241691734304e-05, "loss": 0.0026, "step": 460 }, { "epoch": 232.5, "grad_norm": 0.028752151160391863, "learning_rate": 3.1935133117658306e-05, "loss": 0.0027, "step": 465 }, { "epoch": 235.0, "grad_norm": 0.028976330470022123, "learning_rate": 3.156981202247248e-05, "loss": 0.0026, "step": 470 }, { "epoch": 237.5, "grad_norm": 0.026920539664030775, "learning_rate": 3.1203378281316515e-05, "loss": 0.0026, "step": 475 }, { "epoch": 240.0, "grad_norm": 0.020989002709489214, "learning_rate": 3.0835932073516444e-05, "loss": 0.0027, "step": 480 }, { "epoch": 242.5, "grad_norm": 0.024523095658861207, "learning_rate": 3.0467573855196558e-05, "loss": 0.0027, "step": 485 }, { "epoch": 245.0, "grad_norm": 0.015580401724056997, "learning_rate": 3.0098404331815695e-05, "loss": 0.0027, "step": 490 }, { "epoch": 247.5, "grad_norm": 0.022996840296875175, "learning_rate": 2.9728524430635417e-05, "loss": 0.0027, "step": 495 }, { "epoch": 250.0, "grad_norm": 0.021854877708908236, "learning_rate": 2.9358035273127483e-05, "loss": 0.0026, "step": 500 }, { "epoch": 252.5, "grad_norm": 0.025003129400125675, "learning_rate": 2.8987038147328238e-05, "loss": 0.0026, "step": 505 }, { "epoch": 255.0, "grad_norm": 0.02759119149124416, "learning_rate": 2.86156344801475e-05, "loss": 0.0026, "step": 510 }, { "epoch": 257.5, "grad_norm": 0.024349103530368846, "learning_rate": 2.824392580963944e-05, "loss": 0.0027, "step": 515 }, { "epoch": 260.0, "grad_norm": 0.02918200596645593, "learning_rate": 2.787201375724307e-05, "loss": 0.0027, "step": 520 }, { "epoch": 262.5, "grad_norm": 0.017661985537795705, "learning_rate": 2.7500000000000004e-05, "loss": 0.0026, "step": 525 }, { "epoch": 265.0, "grad_norm": 0.02880881596677379, "learning_rate": 2.7127986242756936e-05, "loss": 0.0026, "step": 530 }, { "epoch": 267.5, "grad_norm": 0.04090166133876468, "learning_rate": 2.6756074190360563e-05, "loss": 0.0028, "step": 535 }, { "epoch": 270.0, "grad_norm": 0.022838636369005745, "learning_rate": 2.63843655198525e-05, "loss": 0.0027, "step": 540 }, { "epoch": 272.5, "grad_norm": 0.03417808915995595, "learning_rate": 2.6012961852671767e-05, "loss": 0.0026, "step": 545 }, { "epoch": 275.0, "grad_norm": 0.02391514801170413, "learning_rate": 2.5641964726872526e-05, "loss": 0.0026, "step": 550 }, { "epoch": 277.5, "grad_norm": 0.027050966559574396, "learning_rate": 2.527147556936459e-05, "loss": 0.0027, "step": 555 }, { "epoch": 280.0, "grad_norm": 0.029392338368449532, "learning_rate": 2.4901595668184314e-05, "loss": 0.0027, "step": 560 }, { "epoch": 282.5, "grad_norm": 0.03213320609449678, "learning_rate": 2.453242614480345e-05, "loss": 0.0027, "step": 565 }, { "epoch": 285.0, "grad_norm": 0.017580210273689653, "learning_rate": 2.416406792648355e-05, "loss": 0.0027, "step": 570 }, { "epoch": 287.5, "grad_norm": 0.025395175371267904, "learning_rate": 2.3796621718683487e-05, "loss": 0.0027, "step": 575 }, { "epoch": 290.0, "grad_norm": 0.02830946439085919, "learning_rate": 2.3430187977527533e-05, "loss": 0.0027, "step": 580 }, { "epoch": 292.5, "grad_norm": 0.04065072947108163, "learning_rate": 2.3064866882341696e-05, "loss": 0.0027, "step": 585 }, { "epoch": 295.0, "grad_norm": 0.028524952343665373, "learning_rate": 2.27007583082657e-05, "loss": 0.0026, "step": 590 }, { "epoch": 297.5, "grad_norm": 0.0284550999234347, "learning_rate": 2.233796179894809e-05, "loss": 0.0026, "step": 595 }, { "epoch": 300.0, "grad_norm": 0.03018543474602915, "learning_rate": 2.1976576539332024e-05, "loss": 0.0026, "step": 600 }, { "epoch": 302.5, "grad_norm": 0.017475284356752212, "learning_rate": 2.1616701328539057e-05, "loss": 0.0026, "step": 605 }, { "epoch": 305.0, "grad_norm": 0.02381743267091071, "learning_rate": 2.1258434552858502e-05, "loss": 0.0026, "step": 610 }, { "epoch": 307.5, "grad_norm": 0.030282727334382083, "learning_rate": 2.090187415884955e-05, "loss": 0.0027, "step": 615 }, { "epoch": 310.0, "grad_norm": 0.02239088576419377, "learning_rate": 2.054711762656369e-05, "loss": 0.0026, "step": 620 }, { "epoch": 312.5, "grad_norm": 0.020772310557568244, "learning_rate": 2.0194261942894628e-05, "loss": 0.0026, "step": 625 }, { "epoch": 315.0, "grad_norm": 0.03490126977201953, "learning_rate": 1.984340357506303e-05, "loss": 0.0026, "step": 630 }, { "epoch": 317.5, "grad_norm": 0.015968447976387346, "learning_rate": 1.949463844424338e-05, "loss": 0.0026, "step": 635 }, { "epoch": 320.0, "grad_norm": 0.023986554939275363, "learning_rate": 1.914806189934003e-05, "loss": 0.0028, "step": 640 }, { "epoch": 322.5, "grad_norm": 0.03420061182554338, "learning_rate": 1.8803768690919832e-05, "loss": 0.0027, "step": 645 }, { "epoch": 325.0, "grad_norm": 0.01792951575408546, "learning_rate": 1.8461852945308196e-05, "loss": 0.0027, "step": 650 }, { "epoch": 327.5, "grad_norm": 0.023802622532695907, "learning_rate": 1.8122408138855974e-05, "loss": 0.0026, "step": 655 }, { "epoch": 330.0, "grad_norm": 0.023316524889500846, "learning_rate": 1.778552707238391e-05, "loss": 0.0026, "step": 660 }, { "epoch": 332.5, "grad_norm": 0.013376585581553524, "learning_rate": 1.7451301845811857e-05, "loss": 0.0026, "step": 665 }, { "epoch": 335.0, "grad_norm": 0.019080334175566694, "learning_rate": 1.711982383297953e-05, "loss": 0.0026, "step": 670 }, { "epoch": 337.5, "grad_norm": 0.01654436855407368, "learning_rate": 1.6791183656665846e-05, "loss": 0.0026, "step": 675 }, { "epoch": 340.0, "grad_norm": 0.022278657387787754, "learning_rate": 1.6465471163813574e-05, "loss": 0.0027, "step": 680 }, { "epoch": 342.5, "grad_norm": 0.015375948509190046, "learning_rate": 1.6142775400966023e-05, "loss": 0.0027, "step": 685 }, { "epoch": 345.0, "grad_norm": 0.01684675867751184, "learning_rate": 1.582318458992267e-05, "loss": 0.0027, "step": 690 }, { "epoch": 347.5, "grad_norm": 0.032002543536054376, "learning_rate": 1.5506786103620146e-05, "loss": 0.0027, "step": 695 }, { "epoch": 350.0, "grad_norm": 0.021043035856906086, "learning_rate": 1.5193666442245402e-05, "loss": 0.0027, "step": 700 }, { "epoch": 352.5, "grad_norm": 0.01601931955071107, "learning_rate": 1.4883911209587368e-05, "loss": 0.0026, "step": 705 }, { "epoch": 355.0, "grad_norm": 0.02601502051924795, "learning_rate": 1.4577605089633773e-05, "loss": 0.0026, "step": 710 }, { "epoch": 357.5, "grad_norm": 0.02710347701415514, "learning_rate": 1.427483182341936e-05, "loss": 0.0026, "step": 715 }, { "epoch": 360.0, "grad_norm": 0.03320644743977735, "learning_rate": 1.3975674186131937e-05, "loss": 0.0027, "step": 720 }, { "epoch": 362.5, "grad_norm": 0.01591474795789328, "learning_rate": 1.3680213964482475e-05, "loss": 0.0027, "step": 725 }, { "epoch": 365.0, "grad_norm": 0.026155373607272086, "learning_rate": 1.3388531934345416e-05, "loss": 0.0027, "step": 730 }, { "epoch": 367.5, "grad_norm": 0.024271143128982743, "learning_rate": 1.3100707838675377e-05, "loss": 0.0026, "step": 735 }, { "epoch": 370.0, "grad_norm": 0.016369678310535946, "learning_rate": 1.2816820365706206e-05, "loss": 0.0027, "step": 740 }, { "epoch": 372.5, "grad_norm": 0.016955558044549856, "learning_rate": 1.2536947127438415e-05, "loss": 0.0026, "step": 745 }, { "epoch": 375.0, "grad_norm": 0.02239783849742695, "learning_rate": 1.2261164638420832e-05, "loss": 0.0026, "step": 750 }, { "epoch": 377.5, "grad_norm": 0.029924950679270815, "learning_rate": 1.198954829483227e-05, "loss": 0.0027, "step": 755 }, { "epoch": 380.0, "grad_norm": 0.028621427234193805, "learning_rate": 1.1722172353869008e-05, "loss": 0.0025, "step": 760 }, { "epoch": 382.5, "grad_norm": 0.02561136953891063, "learning_rate": 1.1459109913443567e-05, "loss": 0.0026, "step": 765 }, { "epoch": 385.0, "grad_norm": 0.028256041486332584, "learning_rate": 1.12004328922005e-05, "loss": 0.0027, "step": 770 }, { "epoch": 387.5, "grad_norm": 0.029729889846126924, "learning_rate": 1.094621200985454e-05, "loss": 0.0027, "step": 775 }, { "epoch": 390.0, "grad_norm": 0.014383243995550028, "learning_rate": 1.0696516767856546e-05, "loss": 0.0026, "step": 780 }, { "epoch": 392.5, "grad_norm": 0.015879736540587025, "learning_rate": 1.0451415430392474e-05, "loss": 0.0026, "step": 785 }, { "epoch": 395.0, "grad_norm": 0.024629301532455877, "learning_rate": 1.0210975005720677e-05, "loss": 0.0026, "step": 790 }, { "epoch": 397.5, "grad_norm": 0.029707059150893382, "learning_rate": 9.975261227852472e-06, "loss": 0.0027, "step": 795 }, { "epoch": 400.0, "grad_norm": 0.0318537391592188, "learning_rate": 9.744338538581147e-06, "loss": 0.0027, "step": 800 }, { "epoch": 402.5, "grad_norm": 0.017706015342844352, "learning_rate": 9.518270069864195e-06, "loss": 0.0027, "step": 805 }, { "epoch": 405.0, "grad_norm": 0.01724610650030409, "learning_rate": 9.297117626563687e-06, "loss": 0.0026, "step": 810 }, { "epoch": 407.5, "grad_norm": 0.018957851028085164, "learning_rate": 9.080941669549423e-06, "loss": 0.0026, "step": 815 }, { "epoch": 410.0, "grad_norm": 0.025418915769378568, "learning_rate": 8.86980129916955e-06, "loss": 0.0026, "step": 820 }, { "epoch": 412.5, "grad_norm": 0.01583223922193541, "learning_rate": 8.663754239093109e-06, "loss": 0.0027, "step": 825 }, { "epoch": 415.0, "grad_norm": 0.027164074218547888, "learning_rate": 8.462856820528952e-06, "loss": 0.0026, "step": 830 }, { "epoch": 417.5, "grad_norm": 0.028647280916262126, "learning_rate": 8.26716396682537e-06, "loss": 0.0027, "step": 835 }, { "epoch": 420.0, "grad_norm": 0.019553123738546497, "learning_rate": 8.076729178454588e-06, "loss": 0.0026, "step": 840 }, { "epoch": 422.5, "grad_norm": 0.025379025955769637, "learning_rate": 7.89160451838626e-06, "loss": 0.0027, "step": 845 }, { "epoch": 425.0, "grad_norm": 0.01750943138582632, "learning_rate": 7.711840597853998e-06, "loss": 0.0026, "step": 850 }, { "epoch": 427.5, "grad_norm": 0.02520822601658372, "learning_rate": 7.537486562518735e-06, "loss": 0.0026, "step": 855 }, { "epoch": 430.0, "grad_norm": 0.01993484997782361, "learning_rate": 7.368590079032822e-06, "loss": 0.0026, "step": 860 }, { "epoch": 432.5, "grad_norm": 0.019158523070010834, "learning_rate": 7.205197322008425e-06, "loss": 0.0027, "step": 865 }, { "epoch": 435.0, "grad_norm": 0.015239850907125138, "learning_rate": 7.047352961393844e-06, "loss": 0.0027, "step": 870 }, { "epoch": 437.5, "grad_norm": 0.020559758942860218, "learning_rate": 6.8951001502612065e-06, "loss": 0.0027, "step": 875 }, { "epoch": 440.0, "grad_norm": 0.02381279446911383, "learning_rate": 6.748480513008844e-06, "loss": 0.0025, "step": 880 }, { "epoch": 442.5, "grad_norm": 0.014589626300082677, "learning_rate": 6.607534133981594e-06, "loss": 0.0026, "step": 885 }, { "epoch": 445.0, "grad_norm": 0.026727121558605703, "learning_rate": 6.472299546512134e-06, "loss": 0.0026, "step": 890 }, { "epoch": 447.5, "grad_norm": 0.01580348324753083, "learning_rate": 6.342813722386374e-06, "loss": 0.0027, "step": 895 }, { "epoch": 450.0, "grad_norm": 0.02334311857794958, "learning_rate": 6.219112061735721e-06, "loss": 0.0026, "step": 900 }, { "epoch": 452.5, "grad_norm": 0.022995449553396018, "learning_rate": 6.1012283833590465e-06, "loss": 0.0026, "step": 905 }, { "epoch": 455.0, "grad_norm": 0.03391053587500504, "learning_rate": 5.989194915476954e-06, "loss": 0.0027, "step": 910 }, { "epoch": 457.5, "grad_norm": 0.0281194121062883, "learning_rate": 5.883042286920918e-06, "loss": 0.0025, "step": 915 }, { "epoch": 460.0, "grad_norm": 0.018983581090637, "learning_rate": 5.782799518759658e-06, "loss": 0.0025, "step": 920 }, { "epoch": 462.5, "grad_norm": 0.024476355758836905, "learning_rate": 5.688494016365067e-06, "loss": 0.0026, "step": 925 }, { "epoch": 465.0, "grad_norm": 0.023061048781595993, "learning_rate": 5.600151561919871e-06, "loss": 0.0027, "step": 930 }, { "epoch": 467.5, "grad_norm": 0.012496893287463113, "learning_rate": 5.517796307369017e-06, "loss": 0.0026, "step": 935 }, { "epoch": 470.0, "grad_norm": 0.01991683594417727, "learning_rate": 5.44145076781679e-06, "loss": 0.0026, "step": 940 }, { "epoch": 472.5, "grad_norm": 0.028037001206555136, "learning_rate": 5.371135815371398e-06, "loss": 0.0027, "step": 945 }, { "epoch": 475.0, "grad_norm": 0.026283232329898983, "learning_rate": 5.3068706734387484e-06, "loss": 0.0026, "step": 950 }, { "epoch": 477.5, "grad_norm": 0.026815406305102027, "learning_rate": 5.248672911466959e-06, "loss": 0.0026, "step": 955 }, { "epoch": 480.0, "grad_norm": 0.01847886314594556, "learning_rate": 5.196558440143059e-06, "loss": 0.0025, "step": 960 }, { "epoch": 482.5, "grad_norm": 0.019370493214584474, "learning_rate": 5.150541507043143e-06, "loss": 0.0027, "step": 965 }, { "epoch": 485.0, "grad_norm": 0.021779585772489417, "learning_rate": 5.110634692737244e-06, "loss": 0.0026, "step": 970 }, { "epoch": 487.5, "grad_norm": 0.02541037015533656, "learning_rate": 5.0768489073499295e-06, "loss": 0.0027, "step": 975 }, { "epoch": 490.0, "grad_norm": 0.02945361459676501, "learning_rate": 5.049193387577574e-06, "loss": 0.0026, "step": 980 }, { "epoch": 492.5, "grad_norm": 0.017922772556181753, "learning_rate": 5.027675694163144e-06, "loss": 0.0028, "step": 985 }, { "epoch": 495.0, "grad_norm": 0.023023048365518207, "learning_rate": 5.012301709829164e-06, "loss": 0.0026, "step": 990 }, { "epoch": 497.5, "grad_norm": 0.020646776127788924, "learning_rate": 5.003075637669448e-06, "loss": 0.0027, "step": 995 }, { "epoch": 500.0, "grad_norm": 0.0148034159233786, "learning_rate": 5e-06, "loss": 0.0027, "step": 1000 }, { "epoch": 500.0, "step": 1000, "total_flos": 28457722773504.0, "train_loss": 0.008928275344893336, "train_runtime": 1701.2794, "train_samples_per_second": 18.809, "train_steps_per_second": 0.588 } ], "logging_steps": 5, "max_steps": 1000, "num_input_tokens_seen": 0, "num_train_epochs": 500, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 28457722773504.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }