{ "best_metric": null, "best_model_checkpoint": null, "epoch": 0.05660991522665195, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0005660991522665195, "grad_norm": 0.4538244605064392, "learning_rate": 2e-05, "loss": 3.4057, "step": 10 }, { "epoch": 0.001132198304533039, "grad_norm": 0.5858215689659119, "learning_rate": 4e-05, "loss": 3.0604, "step": 20 }, { "epoch": 0.0016982974567995583, "grad_norm": 0.612471342086792, "learning_rate": 6e-05, "loss": 3.3496, "step": 30 }, { "epoch": 0.002264396609066078, "grad_norm": 0.9295674562454224, "learning_rate": 8e-05, "loss": 2.9204, "step": 40 }, { "epoch": 0.0028304957613325974, "grad_norm": 0.7428744435310364, "learning_rate": 0.0001, "loss": 2.7989, "step": 50 }, { "epoch": 0.0033965949135991167, "grad_norm": 39.52121353149414, "learning_rate": 0.00012, "loss": 2.3753, "step": 60 }, { "epoch": 0.003962694065865636, "grad_norm": 1.2908339500427246, "learning_rate": 0.00014, "loss": 2.3282, "step": 70 }, { "epoch": 0.004528793218132156, "grad_norm": 1.0259243249893188, "learning_rate": 0.00016, "loss": 2.0235, "step": 80 }, { "epoch": 0.0050948923703986754, "grad_norm": 5.856877326965332, "learning_rate": 0.00018, "loss": 2.3282, "step": 90 }, { "epoch": 0.005660991522665195, "grad_norm": 1.6842005252838135, "learning_rate": 0.0002, "loss": 2.1914, "step": 100 }, { "epoch": 0.006227090674931714, "grad_norm": 1.1379739046096802, "learning_rate": 0.0001999390827019096, "loss": 2.1828, "step": 110 }, { "epoch": 0.006793189827198233, "grad_norm": 1.5961283445358276, "learning_rate": 0.00019975640502598244, "loss": 2.1712, "step": 120 }, { "epoch": 0.0073592889794647535, "grad_norm": 1.2035201787948608, "learning_rate": 0.00019945218953682734, "loss": 1.8216, "step": 130 }, { "epoch": 0.007925388131731272, "grad_norm": 3.0688140392303467, "learning_rate": 0.00019902680687415705, "loss": 2.0315, "step": 140 }, { "epoch": 0.008491487283997793, "grad_norm": 1.050741195678711, "learning_rate": 0.00019848077530122083, "loss": 1.9974, "step": 150 }, { "epoch": 0.009057586436264312, "grad_norm": 1.5697944164276123, "learning_rate": 0.00019781476007338058, "loss": 1.8635, "step": 160 }, { "epoch": 0.009623685588530832, "grad_norm": 1.8822702169418335, "learning_rate": 0.00019702957262759965, "loss": 1.7645, "step": 170 }, { "epoch": 0.010189784740797351, "grad_norm": 6.6875224113464355, "learning_rate": 0.0001961261695938319, "loss": 2.0548, "step": 180 }, { "epoch": 0.01075588389306387, "grad_norm": 1.6249600648880005, "learning_rate": 0.00019510565162951537, "loss": 1.9403, "step": 190 }, { "epoch": 0.01132198304533039, "grad_norm": 1.3810372352600098, "learning_rate": 0.00019396926207859084, "loss": 1.8465, "step": 200 }, { "epoch": 0.011888082197596909, "grad_norm": 1.6512261629104614, "learning_rate": 0.00019271838545667876, "loss": 1.764, "step": 210 }, { "epoch": 0.012454181349863428, "grad_norm": 2.0129904747009277, "learning_rate": 0.0001913545457642601, "loss": 1.6833, "step": 220 }, { "epoch": 0.013020280502129947, "grad_norm": 2.3854947090148926, "learning_rate": 0.0001898794046299167, "loss": 1.9643, "step": 230 }, { "epoch": 0.013586379654396467, "grad_norm": 11.102593421936035, "learning_rate": 0.00018829475928589271, "loss": 1.8319, "step": 240 }, { "epoch": 0.014152478806662988, "grad_norm": 1.0218788385391235, "learning_rate": 0.00018660254037844388, "loss": 1.7456, "step": 250 }, { "epoch": 0.014718577958929507, "grad_norm": 2.1926181316375732, "learning_rate": 0.0001848048096156426, "loss": 1.7705, "step": 260 }, { "epoch": 0.015284677111196026, "grad_norm": 1.2894445657730103, "learning_rate": 0.00018290375725550417, "loss": 1.759, "step": 270 }, { "epoch": 0.015850776263462544, "grad_norm": 1.2821546792984009, "learning_rate": 0.00018090169943749476, "loss": 1.7809, "step": 280 }, { "epoch": 0.016416875415729067, "grad_norm": 1.6037951707839966, "learning_rate": 0.00017880107536067218, "loss": 1.6338, "step": 290 }, { "epoch": 0.016982974567995586, "grad_norm": 1.6165717840194702, "learning_rate": 0.0001766044443118978, "loss": 1.59, "step": 300 }, { "epoch": 0.017549073720262105, "grad_norm": 1.3858145475387573, "learning_rate": 0.00017431448254773944, "loss": 1.7459, "step": 310 }, { "epoch": 0.018115172872528625, "grad_norm": 1.5735255479812622, "learning_rate": 0.0001719339800338651, "loss": 1.7295, "step": 320 }, { "epoch": 0.018681272024795144, "grad_norm": 1.284432053565979, "learning_rate": 0.00016946583704589973, "loss": 1.769, "step": 330 }, { "epoch": 0.019247371177061663, "grad_norm": 0.984366238117218, "learning_rate": 0.00016691306063588583, "loss": 1.6516, "step": 340 }, { "epoch": 0.019813470329328182, "grad_norm": 1.6160484552383423, "learning_rate": 0.00016427876096865394, "loss": 1.529, "step": 350 }, { "epoch": 0.020379569481594702, "grad_norm": 1.8237900733947754, "learning_rate": 0.0001615661475325658, "loss": 1.7505, "step": 360 }, { "epoch": 0.02094566863386122, "grad_norm": 1.297549843788147, "learning_rate": 0.00015877852522924732, "loss": 1.7133, "step": 370 }, { "epoch": 0.02151176778612774, "grad_norm": 2.0918915271759033, "learning_rate": 0.0001559192903470747, "loss": 1.6225, "step": 380 }, { "epoch": 0.02207786693839426, "grad_norm": 1.0555062294006348, "learning_rate": 0.0001529919264233205, "loss": 1.6404, "step": 390 }, { "epoch": 0.02264396609066078, "grad_norm": 1.5048798322677612, "learning_rate": 0.00015000000000000001, "loss": 1.6544, "step": 400 }, { "epoch": 0.0232100652429273, "grad_norm": 1.7201550006866455, "learning_rate": 0.00014694715627858908, "loss": 1.5341, "step": 410 }, { "epoch": 0.023776164395193818, "grad_norm": 1.2342567443847656, "learning_rate": 0.00014383711467890774, "loss": 1.5762, "step": 420 }, { "epoch": 0.024342263547460337, "grad_norm": 0.8844823241233826, "learning_rate": 0.00014067366430758004, "loss": 1.4964, "step": 430 }, { "epoch": 0.024908362699726856, "grad_norm": 2.017977714538574, "learning_rate": 0.00013746065934159123, "loss": 1.5507, "step": 440 }, { "epoch": 0.025474461851993375, "grad_norm": 1.0709903240203857, "learning_rate": 0.00013420201433256689, "loss": 1.51, "step": 450 }, { "epoch": 0.026040561004259895, "grad_norm": 0.863139271736145, "learning_rate": 0.00013090169943749476, "loss": 1.5191, "step": 460 }, { "epoch": 0.026606660156526414, "grad_norm": 1.096670389175415, "learning_rate": 0.0001275637355816999, "loss": 1.6605, "step": 470 }, { "epoch": 0.027172759308792933, "grad_norm": 0.9680797457695007, "learning_rate": 0.00012419218955996676, "loss": 1.6433, "step": 480 }, { "epoch": 0.027738858461059456, "grad_norm": 1.6639995574951172, "learning_rate": 0.00012079116908177593, "loss": 1.5499, "step": 490 }, { "epoch": 0.028304957613325975, "grad_norm": 1.3906327486038208, "learning_rate": 0.00011736481776669306, "loss": 1.5884, "step": 500 }, { "epoch": 0.028871056765592495, "grad_norm": 1.403045654296875, "learning_rate": 0.00011391731009600654, "loss": 1.4895, "step": 510 }, { "epoch": 0.029437155917859014, "grad_norm": 1.4162797927856445, "learning_rate": 0.00011045284632676536, "loss": 1.7302, "step": 520 }, { "epoch": 0.030003255070125533, "grad_norm": 1.0728657245635986, "learning_rate": 0.00010697564737441252, "loss": 1.6658, "step": 530 }, { "epoch": 0.030569354222392053, "grad_norm": 0.9048101902008057, "learning_rate": 0.00010348994967025012, "loss": 1.6423, "step": 540 }, { "epoch": 0.031135453374658572, "grad_norm": 4.7274250984191895, "learning_rate": 0.0001, "loss": 1.6433, "step": 550 }, { "epoch": 0.03170155252692509, "grad_norm": 1.6377885341644287, "learning_rate": 9.651005032974994e-05, "loss": 1.6475, "step": 560 }, { "epoch": 0.03226765167919161, "grad_norm": 2.1424057483673096, "learning_rate": 9.302435262558747e-05, "loss": 1.5888, "step": 570 }, { "epoch": 0.03283375083145813, "grad_norm": 0.9035511612892151, "learning_rate": 8.954715367323468e-05, "loss": 1.6005, "step": 580 }, { "epoch": 0.03339984998372465, "grad_norm": 1.209153413772583, "learning_rate": 8.608268990399349e-05, "loss": 1.6187, "step": 590 }, { "epoch": 0.03396594913599117, "grad_norm": 1.4936399459838867, "learning_rate": 8.263518223330697e-05, "loss": 1.6622, "step": 600 }, { "epoch": 0.03453204828825769, "grad_norm": 0.9744408130645752, "learning_rate": 7.920883091822408e-05, "loss": 1.5553, "step": 610 }, { "epoch": 0.03509814744052421, "grad_norm": 1.1214041709899902, "learning_rate": 7.580781044003324e-05, "loss": 1.4749, "step": 620 }, { "epoch": 0.03566424659279073, "grad_norm": 1.5141066312789917, "learning_rate": 7.243626441830009e-05, "loss": 1.6162, "step": 630 }, { "epoch": 0.03623034574505725, "grad_norm": 1.1192114353179932, "learning_rate": 6.909830056250527e-05, "loss": 1.6655, "step": 640 }, { "epoch": 0.03679644489732377, "grad_norm": 0.8808345794677734, "learning_rate": 6.579798566743314e-05, "loss": 1.5585, "step": 650 }, { "epoch": 0.03736254404959029, "grad_norm": 1.2207506895065308, "learning_rate": 6.25393406584088e-05, "loss": 1.5282, "step": 660 }, { "epoch": 0.03792864320185681, "grad_norm": 0.8557892441749573, "learning_rate": 5.9326335692419995e-05, "loss": 1.6304, "step": 670 }, { "epoch": 0.038494742354123326, "grad_norm": 1.4455292224884033, "learning_rate": 5.616288532109225e-05, "loss": 1.7623, "step": 680 }, { "epoch": 0.039060841506389846, "grad_norm": 1.7669495344161987, "learning_rate": 5.305284372141095e-05, "loss": 1.5934, "step": 690 }, { "epoch": 0.039626940658656365, "grad_norm": 2.2413482666015625, "learning_rate": 5.000000000000002e-05, "loss": 1.692, "step": 700 }, { "epoch": 0.040193039810922884, "grad_norm": 0.9742591977119446, "learning_rate": 4.700807357667952e-05, "loss": 1.5471, "step": 710 }, { "epoch": 0.040759138963189404, "grad_norm": 1.437410593032837, "learning_rate": 4.4080709652925336e-05, "loss": 1.5466, "step": 720 }, { "epoch": 0.04132523811545592, "grad_norm": 1.625700831413269, "learning_rate": 4.12214747707527e-05, "loss": 1.5798, "step": 730 }, { "epoch": 0.04189133726772244, "grad_norm": 0.8843604922294617, "learning_rate": 3.843385246743417e-05, "loss": 1.5987, "step": 740 }, { "epoch": 0.04245743641998896, "grad_norm": 1.7022398710250854, "learning_rate": 3.5721239031346066e-05, "loss": 1.7433, "step": 750 }, { "epoch": 0.04302353557225548, "grad_norm": 0.8799572587013245, "learning_rate": 3.308693936411421e-05, "loss": 1.4654, "step": 760 }, { "epoch": 0.043589634724522, "grad_norm": 0.8340188264846802, "learning_rate": 3.053416295410026e-05, "loss": 1.5339, "step": 770 }, { "epoch": 0.04415573387678852, "grad_norm": 1.684546947479248, "learning_rate": 2.8066019966134904e-05, "loss": 1.5587, "step": 780 }, { "epoch": 0.04472183302905504, "grad_norm": 1.205370306968689, "learning_rate": 2.5685517452260567e-05, "loss": 1.3387, "step": 790 }, { "epoch": 0.04528793218132156, "grad_norm": 1.310747504234314, "learning_rate": 2.339555568810221e-05, "loss": 1.538, "step": 800 }, { "epoch": 0.04585403133358808, "grad_norm": 2.451975107192993, "learning_rate": 2.119892463932781e-05, "loss": 1.6725, "step": 810 }, { "epoch": 0.0464201304858546, "grad_norm": 2.1125898361206055, "learning_rate": 1.9098300562505266e-05, "loss": 1.4746, "step": 820 }, { "epoch": 0.046986229638121116, "grad_norm": 0.9616561532020569, "learning_rate": 1.7096242744495837e-05, "loss": 1.4724, "step": 830 }, { "epoch": 0.047552328790387635, "grad_norm": 1.1507611274719238, "learning_rate": 1.5195190384357404e-05, "loss": 1.4674, "step": 840 }, { "epoch": 0.048118427942654154, "grad_norm": 1.4164016246795654, "learning_rate": 1.339745962155613e-05, "loss": 1.5415, "step": 850 }, { "epoch": 0.048684527094920674, "grad_norm": 1.2756904363632202, "learning_rate": 1.1705240714107302e-05, "loss": 1.4425, "step": 860 }, { "epoch": 0.04925062624718719, "grad_norm": 1.3417842388153076, "learning_rate": 1.0120595370083318e-05, "loss": 1.4356, "step": 870 }, { "epoch": 0.04981672539945371, "grad_norm": 2.374289035797119, "learning_rate": 8.645454235739903e-06, "loss": 1.4808, "step": 880 }, { "epoch": 0.05038282455172023, "grad_norm": 3.114461660385132, "learning_rate": 7.281614543321269e-06, "loss": 1.6683, "step": 890 }, { "epoch": 0.05094892370398675, "grad_norm": 1.7116568088531494, "learning_rate": 6.030737921409169e-06, "loss": 1.6627, "step": 900 }, { "epoch": 0.05151502285625327, "grad_norm": 1.497970461845398, "learning_rate": 4.8943483704846475e-06, "loss": 1.4333, "step": 910 }, { "epoch": 0.05208112200851979, "grad_norm": 1.462307333946228, "learning_rate": 3.873830406168111e-06, "loss": 1.3954, "step": 920 }, { "epoch": 0.05264722116078631, "grad_norm": 0.9947201609611511, "learning_rate": 2.970427372400353e-06, "loss": 1.5113, "step": 930 }, { "epoch": 0.05321332031305283, "grad_norm": 1.0234606266021729, "learning_rate": 2.1852399266194314e-06, "loss": 1.5614, "step": 940 }, { "epoch": 0.05377941946531935, "grad_norm": 1.0096086263656616, "learning_rate": 1.5192246987791981e-06, "loss": 1.4893, "step": 950 }, { "epoch": 0.05434551861758587, "grad_norm": 1.2131421566009521, "learning_rate": 9.731931258429638e-07, "loss": 1.5647, "step": 960 }, { "epoch": 0.05491161776985239, "grad_norm": 1.0574455261230469, "learning_rate": 5.478104631726711e-07, "loss": 1.3718, "step": 970 }, { "epoch": 0.05547771692211891, "grad_norm": 1.1654020547866821, "learning_rate": 2.4359497401758024e-07, "loss": 1.6251, "step": 980 }, { "epoch": 0.05604381607438543, "grad_norm": 1.3310041427612305, "learning_rate": 6.09172980904238e-08, "loss": 1.6333, "step": 990 }, { "epoch": 0.05660991522665195, "grad_norm": 1.0006204843521118, "learning_rate": 0.0, "loss": 1.534, "step": 1000 } ], "logging_steps": 10, "max_steps": 1000, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.297589686332416e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }