{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.146177459435755, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00146177459435755, "grad_norm": 1.1480557918548584, "learning_rate": 2.4752475247524754e-05, "loss": 2.0355, "step": 10 }, { "epoch": 0.0029235491887151, "grad_norm": 1.2355608940124512, "learning_rate": 2.4502450245024505e-05, "loss": 0.9343, "step": 20 }, { "epoch": 0.00438532378307265, "grad_norm": 1.6818450689315796, "learning_rate": 2.4252425242524256e-05, "loss": 0.294, "step": 30 }, { "epoch": 0.0058470983774302, "grad_norm": 0.5871298909187317, "learning_rate": 2.4002400240024003e-05, "loss": 0.2372, "step": 40 }, { "epoch": 0.00730887297178775, "grad_norm": 0.7146134972572327, "learning_rate": 2.3752375237523754e-05, "loss": 0.2836, "step": 50 }, { "epoch": 0.0087706475661453, "grad_norm": 0.7070990800857544, "learning_rate": 2.3502350235023505e-05, "loss": 0.2088, "step": 60 }, { "epoch": 0.01023242216050285, "grad_norm": 0.4255610704421997, "learning_rate": 2.3252325232523252e-05, "loss": 0.1978, "step": 70 }, { "epoch": 0.0116941967548604, "grad_norm": 0.49717941880226135, "learning_rate": 2.3002300230023006e-05, "loss": 0.1968, "step": 80 }, { "epoch": 0.01315597134921795, "grad_norm": 0.40119174122810364, "learning_rate": 2.2752275227522754e-05, "loss": 0.1913, "step": 90 }, { "epoch": 0.0146177459435755, "grad_norm": 0.7866990566253662, "learning_rate": 2.2502250225022505e-05, "loss": 0.2131, "step": 100 }, { "epoch": 0.016079520537933052, "grad_norm": 0.8951889872550964, "learning_rate": 2.2252225222522255e-05, "loss": 0.1765, "step": 110 }, { "epoch": 0.0175412951322906, "grad_norm": 0.29221057891845703, "learning_rate": 2.2002200220022003e-05, "loss": 0.1407, "step": 120 }, { "epoch": 0.019003069726648152, "grad_norm": 0.49533528089523315, "learning_rate": 2.1752175217521754e-05, "loss": 0.1644, "step": 130 }, { "epoch": 0.0204648443210057, "grad_norm": 0.3729006350040436, "learning_rate": 2.1502150215021505e-05, "loss": 0.099, "step": 140 }, { "epoch": 0.021926618915363252, "grad_norm": 0.8305023312568665, "learning_rate": 2.1252125212521255e-05, "loss": 0.1278, "step": 150 }, { "epoch": 0.0233883935097208, "grad_norm": 0.4704670011997223, "learning_rate": 2.1002100210021003e-05, "loss": 0.1521, "step": 160 }, { "epoch": 0.024850168104078352, "grad_norm": 0.787459135055542, "learning_rate": 2.0752075207520754e-05, "loss": 0.1326, "step": 170 }, { "epoch": 0.0263119426984359, "grad_norm": 0.5290590524673462, "learning_rate": 2.0502050205020504e-05, "loss": 0.1146, "step": 180 }, { "epoch": 0.027773717292793453, "grad_norm": 0.46197742223739624, "learning_rate": 2.0252025202520252e-05, "loss": 0.1278, "step": 190 }, { "epoch": 0.029235491887151, "grad_norm": 0.6425529718399048, "learning_rate": 2.0002000200020006e-05, "loss": 0.1063, "step": 200 }, { "epoch": 0.030697266481508553, "grad_norm": 0.7542579770088196, "learning_rate": 1.9751975197519753e-05, "loss": 0.117, "step": 210 }, { "epoch": 0.032159041075866104, "grad_norm": 0.38565704226493835, "learning_rate": 1.9501950195019504e-05, "loss": 0.1298, "step": 220 }, { "epoch": 0.03362081567022365, "grad_norm": 0.31743547320365906, "learning_rate": 1.9251925192519255e-05, "loss": 0.1127, "step": 230 }, { "epoch": 0.0350825902645812, "grad_norm": 0.9804937839508057, "learning_rate": 1.9001900190019003e-05, "loss": 0.1271, "step": 240 }, { "epoch": 0.03654436485893875, "grad_norm": 0.5180727243423462, "learning_rate": 1.8751875187518753e-05, "loss": 0.1222, "step": 250 }, { "epoch": 0.038006139453296305, "grad_norm": 0.5256487131118774, "learning_rate": 1.8501850185018504e-05, "loss": 0.0975, "step": 260 }, { "epoch": 0.03946791404765385, "grad_norm": 0.3315429091453552, "learning_rate": 1.8251825182518255e-05, "loss": 0.0949, "step": 270 }, { "epoch": 0.0409296886420114, "grad_norm": 0.4036615788936615, "learning_rate": 1.8001800180018002e-05, "loss": 0.1127, "step": 280 }, { "epoch": 0.04239146323636895, "grad_norm": 0.35830047726631165, "learning_rate": 1.7751775177517753e-05, "loss": 0.1042, "step": 290 }, { "epoch": 0.043853237830726505, "grad_norm": 0.7400040626525879, "learning_rate": 1.7501750175017504e-05, "loss": 0.1095, "step": 300 }, { "epoch": 0.04531501242508405, "grad_norm": 0.4160325229167938, "learning_rate": 1.725172517251725e-05, "loss": 0.0921, "step": 310 }, { "epoch": 0.0467767870194416, "grad_norm": 0.6937474012374878, "learning_rate": 1.7001700170017006e-05, "loss": 0.1096, "step": 320 }, { "epoch": 0.04823856161379915, "grad_norm": 0.3352717161178589, "learning_rate": 1.6751675167516753e-05, "loss": 0.0958, "step": 330 }, { "epoch": 0.049700336208156705, "grad_norm": 0.5310590267181396, "learning_rate": 1.6501650165016504e-05, "loss": 0.0866, "step": 340 }, { "epoch": 0.05116211080251425, "grad_norm": 0.44820308685302734, "learning_rate": 1.6251625162516255e-05, "loss": 0.0768, "step": 350 }, { "epoch": 0.0526238853968718, "grad_norm": 0.3603011965751648, "learning_rate": 1.6001600160016002e-05, "loss": 0.1096, "step": 360 }, { "epoch": 0.05408565999122935, "grad_norm": 0.4311069846153259, "learning_rate": 1.5751575157515753e-05, "loss": 0.0957, "step": 370 }, { "epoch": 0.055547434585586905, "grad_norm": 0.36236947774887085, "learning_rate": 1.5501550155015504e-05, "loss": 0.1151, "step": 380 }, { "epoch": 0.05700920917994445, "grad_norm": 0.405987948179245, "learning_rate": 1.5251525152515253e-05, "loss": 0.1184, "step": 390 }, { "epoch": 0.058470983774302, "grad_norm": 0.7587026357650757, "learning_rate": 1.5001500150015002e-05, "loss": 0.0836, "step": 400 }, { "epoch": 0.05993275836865955, "grad_norm": 0.7230392694473267, "learning_rate": 1.4751475147514754e-05, "loss": 0.0796, "step": 410 }, { "epoch": 0.061394532963017105, "grad_norm": 0.3412224352359772, "learning_rate": 1.4501450145014504e-05, "loss": 0.0883, "step": 420 }, { "epoch": 0.06285630755737465, "grad_norm": 0.4818565845489502, "learning_rate": 1.4251425142514253e-05, "loss": 0.0683, "step": 430 }, { "epoch": 0.06431808215173221, "grad_norm": 0.6868996620178223, "learning_rate": 1.4001400140014002e-05, "loss": 0.0923, "step": 440 }, { "epoch": 0.06577985674608976, "grad_norm": 0.3734653890132904, "learning_rate": 1.3751375137513753e-05, "loss": 0.0841, "step": 450 }, { "epoch": 0.0672416313404473, "grad_norm": 0.4107135534286499, "learning_rate": 1.3501350135013502e-05, "loss": 0.0797, "step": 460 }, { "epoch": 0.06870340593480485, "grad_norm": 0.43387386202812195, "learning_rate": 1.3251325132513251e-05, "loss": 0.0721, "step": 470 }, { "epoch": 0.0701651805291624, "grad_norm": 0.5933758020401001, "learning_rate": 1.3001300130013003e-05, "loss": 0.1408, "step": 480 }, { "epoch": 0.07162695512351995, "grad_norm": 0.42749252915382385, "learning_rate": 1.2751275127512753e-05, "loss": 0.0926, "step": 490 }, { "epoch": 0.0730887297178775, "grad_norm": 0.5452368855476379, "learning_rate": 1.2501250125012502e-05, "loss": 0.0871, "step": 500 }, { "epoch": 0.07455050431223505, "grad_norm": 0.42915862798690796, "learning_rate": 1.2251225122512252e-05, "loss": 0.0779, "step": 510 }, { "epoch": 0.07601227890659261, "grad_norm": 0.5489547252655029, "learning_rate": 1.2001200120012002e-05, "loss": 0.0649, "step": 520 }, { "epoch": 0.07747405350095016, "grad_norm": 0.3804190158843994, "learning_rate": 1.1751175117511752e-05, "loss": 0.0692, "step": 530 }, { "epoch": 0.0789358280953077, "grad_norm": 0.4363003075122833, "learning_rate": 1.1501150115011503e-05, "loss": 0.1051, "step": 540 }, { "epoch": 0.08039760268966525, "grad_norm": 0.352827787399292, "learning_rate": 1.1251125112511252e-05, "loss": 0.0924, "step": 550 }, { "epoch": 0.0818593772840228, "grad_norm": 0.3126923143863678, "learning_rate": 1.1001100110011001e-05, "loss": 0.08, "step": 560 }, { "epoch": 0.08332115187838035, "grad_norm": 0.8426673412322998, "learning_rate": 1.0751075107510752e-05, "loss": 0.0856, "step": 570 }, { "epoch": 0.0847829264727379, "grad_norm": 0.862802267074585, "learning_rate": 1.0501050105010501e-05, "loss": 0.0839, "step": 580 }, { "epoch": 0.08624470106709545, "grad_norm": 0.3358764946460724, "learning_rate": 1.0251025102510252e-05, "loss": 0.0661, "step": 590 }, { "epoch": 0.08770647566145301, "grad_norm": 0.27090683579444885, "learning_rate": 1.0001000100010003e-05, "loss": 0.0756, "step": 600 }, { "epoch": 0.08916825025581056, "grad_norm": 0.2274910807609558, "learning_rate": 9.750975097509752e-06, "loss": 0.0803, "step": 610 }, { "epoch": 0.0906300248501681, "grad_norm": 0.38698479533195496, "learning_rate": 9.500950095009501e-06, "loss": 0.0792, "step": 620 }, { "epoch": 0.09209179944452565, "grad_norm": 0.4514479637145996, "learning_rate": 9.250925092509252e-06, "loss": 0.0732, "step": 630 }, { "epoch": 0.0935535740388832, "grad_norm": 0.5197333693504333, "learning_rate": 9.000900090009001e-06, "loss": 0.0669, "step": 640 }, { "epoch": 0.09501534863324075, "grad_norm": 0.4484429657459259, "learning_rate": 8.750875087508752e-06, "loss": 0.0842, "step": 650 }, { "epoch": 0.0964771232275983, "grad_norm": 0.2479228973388672, "learning_rate": 8.500850085008503e-06, "loss": 0.0599, "step": 660 }, { "epoch": 0.09793889782195585, "grad_norm": 0.46314725279808044, "learning_rate": 8.250825082508252e-06, "loss": 0.0707, "step": 670 }, { "epoch": 0.09940067241631341, "grad_norm": 0.33118072152137756, "learning_rate": 8.000800080008001e-06, "loss": 0.0742, "step": 680 }, { "epoch": 0.10086244701067096, "grad_norm": 0.4030800759792328, "learning_rate": 7.750775077507752e-06, "loss": 0.0759, "step": 690 }, { "epoch": 0.1023242216050285, "grad_norm": 0.31615331768989563, "learning_rate": 7.500750075007501e-06, "loss": 0.0912, "step": 700 }, { "epoch": 0.10378599619938605, "grad_norm": 0.32550713419914246, "learning_rate": 7.250725072507252e-06, "loss": 0.0864, "step": 710 }, { "epoch": 0.1052477707937436, "grad_norm": 0.47825929522514343, "learning_rate": 7.000700070007001e-06, "loss": 0.0735, "step": 720 }, { "epoch": 0.10670954538810115, "grad_norm": 0.534595787525177, "learning_rate": 6.750675067506751e-06, "loss": 0.0742, "step": 730 }, { "epoch": 0.1081713199824587, "grad_norm": 0.5879881978034973, "learning_rate": 6.500650065006502e-06, "loss": 0.0816, "step": 740 }, { "epoch": 0.10963309457681626, "grad_norm": 0.7548266053199768, "learning_rate": 6.250625062506251e-06, "loss": 0.0612, "step": 750 }, { "epoch": 0.11109486917117381, "grad_norm": 1.1165082454681396, "learning_rate": 6.000600060006001e-06, "loss": 0.0766, "step": 760 }, { "epoch": 0.11255664376553136, "grad_norm": 0.5112439393997192, "learning_rate": 5.750575057505752e-06, "loss": 0.0628, "step": 770 }, { "epoch": 0.1140184183598889, "grad_norm": 0.416564017534256, "learning_rate": 5.500550055005501e-06, "loss": 0.0751, "step": 780 }, { "epoch": 0.11548019295424646, "grad_norm": 0.9259045124053955, "learning_rate": 5.250525052505251e-06, "loss": 0.1037, "step": 790 }, { "epoch": 0.116941967548604, "grad_norm": 0.3240310847759247, "learning_rate": 5.0005000500050015e-06, "loss": 0.0826, "step": 800 }, { "epoch": 0.11840374214296155, "grad_norm": 0.32784464955329895, "learning_rate": 4.750475047504751e-06, "loss": 0.0849, "step": 810 }, { "epoch": 0.1198655167373191, "grad_norm": 0.198726087808609, "learning_rate": 4.500450045004501e-06, "loss": 0.0554, "step": 820 }, { "epoch": 0.12132729133167666, "grad_norm": 0.47690388560295105, "learning_rate": 4.250425042504251e-06, "loss": 0.07, "step": 830 }, { "epoch": 0.12278906592603421, "grad_norm": 0.6009281873703003, "learning_rate": 4.0004000400040005e-06, "loss": 0.1019, "step": 840 }, { "epoch": 0.12425084052039176, "grad_norm": 0.37418490648269653, "learning_rate": 3.7503750375037505e-06, "loss": 0.0935, "step": 850 }, { "epoch": 0.1257126151147493, "grad_norm": 0.36421412229537964, "learning_rate": 3.5003500350035005e-06, "loss": 0.0883, "step": 860 }, { "epoch": 0.12717438970910686, "grad_norm": 0.44606733322143555, "learning_rate": 3.250325032503251e-06, "loss": 0.1029, "step": 870 }, { "epoch": 0.12863616430346442, "grad_norm": 0.4448641836643219, "learning_rate": 3.0003000300030004e-06, "loss": 0.0802, "step": 880 }, { "epoch": 0.13009793889782195, "grad_norm": 0.42409053444862366, "learning_rate": 2.7502750275027504e-06, "loss": 0.0688, "step": 890 }, { "epoch": 0.13155971349217951, "grad_norm": 0.43361061811447144, "learning_rate": 2.5002500250025008e-06, "loss": 0.0575, "step": 900 }, { "epoch": 0.13302148808653705, "grad_norm": 0.3293386995792389, "learning_rate": 2.2502250225022503e-06, "loss": 0.0547, "step": 910 }, { "epoch": 0.1344832626808946, "grad_norm": 1.1192256212234497, "learning_rate": 2.0002000200020003e-06, "loss": 0.0809, "step": 920 }, { "epoch": 0.13594503727525215, "grad_norm": 0.45779263973236084, "learning_rate": 1.7501750175017502e-06, "loss": 0.0993, "step": 930 }, { "epoch": 0.1374068118696097, "grad_norm": 0.3892732560634613, "learning_rate": 1.5001500150015002e-06, "loss": 0.0467, "step": 940 }, { "epoch": 0.13886858646396727, "grad_norm": 0.621871829032898, "learning_rate": 1.2501250125012504e-06, "loss": 0.0756, "step": 950 }, { "epoch": 0.1403303610583248, "grad_norm": 0.37552908062934875, "learning_rate": 1.0001000100010001e-06, "loss": 0.0774, "step": 960 }, { "epoch": 0.14179213565268237, "grad_norm": 0.42678821086883545, "learning_rate": 7.500750075007501e-07, "loss": 0.0657, "step": 970 }, { "epoch": 0.1432539102470399, "grad_norm": 0.552952766418457, "learning_rate": 5.000500050005001e-07, "loss": 0.0517, "step": 980 }, { "epoch": 0.14471568484139746, "grad_norm": 0.2972164750099182, "learning_rate": 2.5002500250025003e-07, "loss": 0.0597, "step": 990 }, { "epoch": 0.146177459435755, "grad_norm": 0.501569926738739, "learning_rate": 0.0, "loss": 0.0774, "step": 1000 } ], "logging_steps": 10, "max_steps": 1000, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2645444001792000.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }