|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 10.0, |
|
"eval_steps": 500, |
|
"global_step": 1210, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08264462809917356, |
|
"grad_norm": 9.173369407653809, |
|
"learning_rate": 1.6393442622950818e-05, |
|
"loss": 1.2749, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.1652892561983471, |
|
"grad_norm": 3.0318551063537598, |
|
"learning_rate": 3.2786885245901635e-05, |
|
"loss": 0.7164, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.24793388429752067, |
|
"grad_norm": 1.8877290487289429, |
|
"learning_rate": 4.918032786885246e-05, |
|
"loss": 0.4175, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.3305785123966942, |
|
"grad_norm": 2.459096670150757, |
|
"learning_rate": 6.557377049180327e-05, |
|
"loss": 0.3476, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.4132231404958678, |
|
"grad_norm": 1.8838213682174683, |
|
"learning_rate": 8.19672131147541e-05, |
|
"loss": 0.2797, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.49586776859504134, |
|
"grad_norm": 1.053584337234497, |
|
"learning_rate": 9.836065573770493e-05, |
|
"loss": 0.2348, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.5785123966942148, |
|
"grad_norm": 2.7973430156707764, |
|
"learning_rate": 9.998486219870769e-05, |
|
"loss": 0.224, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.6611570247933884, |
|
"grad_norm": 1.6049909591674805, |
|
"learning_rate": 9.993254576451651e-05, |
|
"loss": 0.2017, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.743801652892562, |
|
"grad_norm": 0.7833113074302673, |
|
"learning_rate": 9.984290291014105e-05, |
|
"loss": 0.2074, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.8264462809917356, |
|
"grad_norm": 3.117323160171509, |
|
"learning_rate": 9.971600064692222e-05, |
|
"loss": 0.1826, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.9090909090909091, |
|
"grad_norm": 0.858580470085144, |
|
"learning_rate": 9.955193383898376e-05, |
|
"loss": 0.1674, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.9917355371900827, |
|
"grad_norm": 0.7289248108863831, |
|
"learning_rate": 9.935082513231775e-05, |
|
"loss": 0.1584, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 1.0743801652892562, |
|
"grad_norm": 2.072802782058716, |
|
"learning_rate": 9.911282486310213e-05, |
|
"loss": 0.154, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 1.1570247933884297, |
|
"grad_norm": 1.2961859703063965, |
|
"learning_rate": 9.883811094531906e-05, |
|
"loss": 0.1567, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 1.2396694214876034, |
|
"grad_norm": 0.9249287843704224, |
|
"learning_rate": 9.85268887377574e-05, |
|
"loss": 0.1537, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 1.322314049586777, |
|
"grad_norm": 1.7591155767440796, |
|
"learning_rate": 9.817939089049964e-05, |
|
"loss": 0.1435, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 1.4049586776859504, |
|
"grad_norm": 3.4378185272216797, |
|
"learning_rate": 9.779587717100729e-05, |
|
"loss": 0.1329, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 1.487603305785124, |
|
"grad_norm": 0.6799660921096802, |
|
"learning_rate": 9.737663426993513e-05, |
|
"loss": 0.1388, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 1.5702479338842976, |
|
"grad_norm": 0.6710700988769531, |
|
"learning_rate": 9.69219755868194e-05, |
|
"loss": 0.147, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 1.6528925619834711, |
|
"grad_norm": 1.6649482250213623, |
|
"learning_rate": 9.643224099579998e-05, |
|
"loss": 0.1219, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 1.7355371900826446, |
|
"grad_norm": 1.4101303815841675, |
|
"learning_rate": 9.59077965915521e-05, |
|
"loss": 0.1132, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 1.8181818181818183, |
|
"grad_norm": 0.7230204939842224, |
|
"learning_rate": 9.534903441561693e-05, |
|
"loss": 0.1196, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 1.9008264462809916, |
|
"grad_norm": 2.3801050186157227, |
|
"learning_rate": 9.47563721633361e-05, |
|
"loss": 0.106, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 1.9834710743801653, |
|
"grad_norm": 1.3875033855438232, |
|
"learning_rate": 9.413025287160905e-05, |
|
"loss": 0.1181, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 2.0661157024793386, |
|
"grad_norm": 0.49982619285583496, |
|
"learning_rate": 9.347114458770657e-05, |
|
"loss": 0.0937, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 2.1487603305785123, |
|
"grad_norm": 1.5207865238189697, |
|
"learning_rate": 9.277954001938818e-05, |
|
"loss": 0.1005, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 2.231404958677686, |
|
"grad_norm": 0.5808604955673218, |
|
"learning_rate": 9.205595616658495e-05, |
|
"loss": 0.1046, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 2.3140495867768593, |
|
"grad_norm": 0.5397279262542725, |
|
"learning_rate": 9.1300933934923e-05, |
|
"loss": 0.095, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 2.396694214876033, |
|
"grad_norm": 2.023026704788208, |
|
"learning_rate": 9.051503773137646e-05, |
|
"loss": 0.1054, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 2.479338842975207, |
|
"grad_norm": 0.8514553904533386, |
|
"learning_rate": 8.969885504235256e-05, |
|
"loss": 0.1058, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 2.56198347107438, |
|
"grad_norm": 0.7182636260986328, |
|
"learning_rate": 8.885299599452382e-05, |
|
"loss": 0.1155, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 2.644628099173554, |
|
"grad_norm": 1.0667051076889038, |
|
"learning_rate": 8.797809289873587e-05, |
|
"loss": 0.1009, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 2.7272727272727275, |
|
"grad_norm": 0.7455232739448547, |
|
"learning_rate": 8.70747997773317e-05, |
|
"loss": 0.0969, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 2.809917355371901, |
|
"grad_norm": 0.5841661691665649, |
|
"learning_rate": 8.614379187524592e-05, |
|
"loss": 0.0885, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 2.8925619834710745, |
|
"grad_norm": 0.5531947016716003, |
|
"learning_rate": 8.518576515523424e-05, |
|
"loss": 0.0909, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 2.975206611570248, |
|
"grad_norm": 0.7314962148666382, |
|
"learning_rate": 8.42014357776155e-05, |
|
"loss": 0.0821, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 3.0578512396694215, |
|
"grad_norm": 1.354279637336731, |
|
"learning_rate": 8.319153956491568e-05, |
|
"loss": 0.0908, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 3.1404958677685952, |
|
"grad_norm": 0.621191143989563, |
|
"learning_rate": 8.215683145181312e-05, |
|
"loss": 0.0822, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 3.2231404958677685, |
|
"grad_norm": 0.6993005871772766, |
|
"learning_rate": 8.109808492079718e-05, |
|
"loss": 0.0791, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 3.3057851239669422, |
|
"grad_norm": 1.1259503364562988, |
|
"learning_rate": 8.001609142396149e-05, |
|
"loss": 0.089, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 3.3884297520661155, |
|
"grad_norm": 1.8521549701690674, |
|
"learning_rate": 7.891165979136429e-05, |
|
"loss": 0.0741, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 3.4710743801652892, |
|
"grad_norm": 0.6957982778549194, |
|
"learning_rate": 7.778561562639818e-05, |
|
"loss": 0.076, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 3.553719008264463, |
|
"grad_norm": 0.4278518259525299, |
|
"learning_rate": 7.663880068862106e-05, |
|
"loss": 0.0752, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 3.6363636363636362, |
|
"grad_norm": 0.9472731351852417, |
|
"learning_rate": 7.547207226450979e-05, |
|
"loss": 0.0899, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 3.71900826446281, |
|
"grad_norm": 0.472150593996048, |
|
"learning_rate": 7.428630252660704e-05, |
|
"loss": 0.0746, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 3.8016528925619832, |
|
"grad_norm": 0.614167332649231, |
|
"learning_rate": 7.308237788154003e-05, |
|
"loss": 0.0785, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 3.884297520661157, |
|
"grad_norm": 0.5245425701141357, |
|
"learning_rate": 7.186119830739883e-05, |
|
"loss": 0.0756, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 3.9669421487603307, |
|
"grad_norm": 0.4369754195213318, |
|
"learning_rate": 7.062367668096967e-05, |
|
"loss": 0.0673, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 4.049586776859504, |
|
"grad_norm": 0.7638271450996399, |
|
"learning_rate": 6.93707380953258e-05, |
|
"loss": 0.0582, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 4.132231404958677, |
|
"grad_norm": 0.7099161744117737, |
|
"learning_rate": 6.810331916828622e-05, |
|
"loss": 0.065, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 4.214876033057851, |
|
"grad_norm": 0.9365837574005127, |
|
"learning_rate": 6.682236734225944e-05, |
|
"loss": 0.0736, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 4.297520661157025, |
|
"grad_norm": 0.5708351135253906, |
|
"learning_rate": 6.552884017599517e-05, |
|
"loss": 0.0582, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 4.380165289256198, |
|
"grad_norm": 2.0133728981018066, |
|
"learning_rate": 6.422370462877396e-05, |
|
"loss": 0.0737, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 4.462809917355372, |
|
"grad_norm": 0.615628719329834, |
|
"learning_rate": 6.29079363375694e-05, |
|
"loss": 0.0619, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 4.545454545454545, |
|
"grad_norm": 1.595961570739746, |
|
"learning_rate": 6.15825188877235e-05, |
|
"loss": 0.0646, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 4.628099173553719, |
|
"grad_norm": 0.5787972807884216, |
|
"learning_rate": 6.0248443077680316e-05, |
|
"loss": 0.0727, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 4.710743801652892, |
|
"grad_norm": 0.37428799271583557, |
|
"learning_rate": 5.890670617832764e-05, |
|
"loss": 0.0694, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 4.793388429752066, |
|
"grad_norm": 0.35877710580825806, |
|
"learning_rate": 5.755831118750016e-05, |
|
"loss": 0.0615, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 4.87603305785124, |
|
"grad_norm": 0.7337440252304077, |
|
"learning_rate": 5.620426608020156e-05, |
|
"loss": 0.0693, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 4.958677685950414, |
|
"grad_norm": 0.4097314476966858, |
|
"learning_rate": 5.484558305510609e-05, |
|
"loss": 0.0584, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 5.041322314049586, |
|
"grad_norm": 0.9892732501029968, |
|
"learning_rate": 5.348327777790262e-05, |
|
"loss": 0.0654, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 5.12396694214876, |
|
"grad_norm": 0.9509891867637634, |
|
"learning_rate": 5.211836862204715e-05, |
|
"loss": 0.0594, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 5.206611570247934, |
|
"grad_norm": 0.6350495219230652, |
|
"learning_rate": 5.075187590749101e-05, |
|
"loss": 0.0568, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 5.289256198347108, |
|
"grad_norm": 0.4029739201068878, |
|
"learning_rate": 4.93848211379541e-05, |
|
"loss": 0.0572, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 5.371900826446281, |
|
"grad_norm": 0.8811182379722595, |
|
"learning_rate": 4.8018226237313165e-05, |
|
"loss": 0.0602, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 5.454545454545454, |
|
"grad_norm": 0.9591042995452881, |
|
"learning_rate": 4.665311278567593e-05, |
|
"loss": 0.0571, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 5.537190082644628, |
|
"grad_norm": 0.4754871726036072, |
|
"learning_rate": 4.5290501255712415e-05, |
|
"loss": 0.0489, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 5.619834710743802, |
|
"grad_norm": 1.2373918294906616, |
|
"learning_rate": 4.3931410249813806e-05, |
|
"loss": 0.0559, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 5.702479338842975, |
|
"grad_norm": 0.8961721062660217, |
|
"learning_rate": 4.2576855738649714e-05, |
|
"loss": 0.0532, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 5.785123966942149, |
|
"grad_norm": 0.4459110498428345, |
|
"learning_rate": 4.122785030169256e-05, |
|
"loss": 0.0655, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 5.867768595041323, |
|
"grad_norm": 0.38759222626686096, |
|
"learning_rate": 3.988540237027702e-05, |
|
"loss": 0.0542, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 5.950413223140496, |
|
"grad_norm": 1.9169152975082397, |
|
"learning_rate": 3.8550515473760514e-05, |
|
"loss": 0.0507, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 6.033057851239669, |
|
"grad_norm": 2.071638345718384, |
|
"learning_rate": 3.722418748934785e-05, |
|
"loss": 0.0559, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 6.115702479338843, |
|
"grad_norm": 0.7945961952209473, |
|
"learning_rate": 3.590740989614131e-05, |
|
"loss": 0.0542, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 6.198347107438017, |
|
"grad_norm": 0.488382488489151, |
|
"learning_rate": 3.460116703397336e-05, |
|
"loss": 0.0599, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 6.2809917355371905, |
|
"grad_norm": 0.5963483452796936, |
|
"learning_rate": 3.330643536757638e-05, |
|
"loss": 0.051, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 6.363636363636363, |
|
"grad_norm": 0.6972360610961914, |
|
"learning_rate": 3.2024182756639185e-05, |
|
"loss": 0.0574, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 6.446280991735537, |
|
"grad_norm": 0.6945924758911133, |
|
"learning_rate": 3.075536773229624e-05, |
|
"loss": 0.0533, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 6.528925619834711, |
|
"grad_norm": 0.3093964755535126, |
|
"learning_rate": 2.9500938780590275e-05, |
|
"loss": 0.0567, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 6.6115702479338845, |
|
"grad_norm": 1.643970251083374, |
|
"learning_rate": 2.826183363344391e-05, |
|
"loss": 0.0527, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 6.694214876033058, |
|
"grad_norm": 0.8777986168861389, |
|
"learning_rate": 2.7038978567670558e-05, |
|
"loss": 0.0543, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 6.776859504132231, |
|
"grad_norm": 0.5457648038864136, |
|
"learning_rate": 2.5833287712548198e-05, |
|
"loss": 0.0533, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 6.859504132231405, |
|
"grad_norm": 0.49185365438461304, |
|
"learning_rate": 2.4645662366474188e-05, |
|
"loss": 0.0551, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 6.9421487603305785, |
|
"grad_norm": 0.3884984254837036, |
|
"learning_rate": 2.3476990323211267e-05, |
|
"loss": 0.0513, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 7.024793388429752, |
|
"grad_norm": 0.5958797335624695, |
|
"learning_rate": 2.2328145208229095e-05, |
|
"loss": 0.0631, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 7.107438016528926, |
|
"grad_norm": 1.0943728685379028, |
|
"learning_rate": 2.119998582563692e-05, |
|
"loss": 0.0531, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 7.190082644628099, |
|
"grad_norm": 0.691371500492096, |
|
"learning_rate": 2.0093355516195888e-05, |
|
"loss": 0.0554, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 7.2727272727272725, |
|
"grad_norm": 5.228278636932373, |
|
"learning_rate": 1.900908152689062e-05, |
|
"loss": 0.0509, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 7.355371900826446, |
|
"grad_norm": 0.3314078748226166, |
|
"learning_rate": 1.7947974392531612e-05, |
|
"loss": 0.0419, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 7.43801652892562, |
|
"grad_norm": 0.6459057927131653, |
|
"learning_rate": 1.6910827329850616e-05, |
|
"loss": 0.0443, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 7.520661157024794, |
|
"grad_norm": 0.29878008365631104, |
|
"learning_rate": 1.589841564454176e-05, |
|
"loss": 0.0424, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 7.6033057851239665, |
|
"grad_norm": 0.4006846249103546, |
|
"learning_rate": 1.4911496151692012e-05, |
|
"loss": 0.0479, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 7.68595041322314, |
|
"grad_norm": 0.5304248332977295, |
|
"learning_rate": 1.3950806610033957e-05, |
|
"loss": 0.0436, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 7.768595041322314, |
|
"grad_norm": 0.7876132726669312, |
|
"learning_rate": 1.3017065170443948e-05, |
|
"loss": 0.0448, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 7.851239669421488, |
|
"grad_norm": 0.9319537281990051, |
|
"learning_rate": 1.2110969839097797e-05, |
|
"loss": 0.045, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 7.933884297520661, |
|
"grad_norm": 1.810137391090393, |
|
"learning_rate": 1.1233197955685409e-05, |
|
"loss": 0.0432, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 8.016528925619834, |
|
"grad_norm": 1.6881096363067627, |
|
"learning_rate": 1.03844056870744e-05, |
|
"loss": 0.0492, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 8.099173553719009, |
|
"grad_norm": 1.5425840616226196, |
|
"learning_rate": 9.565227536801135e-06, |
|
"loss": 0.0454, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 8.181818181818182, |
|
"grad_norm": 1.0203477144241333, |
|
"learning_rate": 8.776275870755924e-06, |
|
"loss": 0.0551, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 8.264462809917354, |
|
"grad_norm": 1.097283959388733, |
|
"learning_rate": 8.018140459416962e-06, |
|
"loss": 0.039, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 8.347107438016529, |
|
"grad_norm": 0.6623409390449524, |
|
"learning_rate": 7.291388036975072e-06, |
|
"loss": 0.0438, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 8.429752066115702, |
|
"grad_norm": 0.38722532987594604, |
|
"learning_rate": 6.596561877679036e-06, |
|
"loss": 0.0541, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 8.512396694214877, |
|
"grad_norm": 0.1958349198102951, |
|
"learning_rate": 5.93418138971803e-06, |
|
"loss": 0.042, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 8.59504132231405, |
|
"grad_norm": 0.44341278076171875, |
|
"learning_rate": 5.304741726944873e-06, |
|
"loss": 0.0434, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 8.677685950413224, |
|
"grad_norm": 2.145524024963379, |
|
"learning_rate": 4.70871341873021e-06, |
|
"loss": 0.0357, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 8.760330578512397, |
|
"grad_norm": 1.4434798955917358, |
|
"learning_rate": 4.146542018224447e-06, |
|
"loss": 0.0405, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 8.84297520661157, |
|
"grad_norm": 0.6765860915184021, |
|
"learning_rate": 3.6186477692903954e-06, |
|
"loss": 0.0426, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 8.925619834710744, |
|
"grad_norm": 0.3515833020210266, |
|
"learning_rate": 3.1254252923553994e-06, |
|
"loss": 0.0464, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 9.008264462809917, |
|
"grad_norm": 0.4808438718318939, |
|
"learning_rate": 2.667243289418059e-06, |
|
"loss": 0.045, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 9.090909090909092, |
|
"grad_norm": 0.4166981875896454, |
|
"learning_rate": 2.244444268429857e-06, |
|
"loss": 0.0477, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 9.173553719008265, |
|
"grad_norm": 0.5427719950675964, |
|
"learning_rate": 1.8573442872578616e-06, |
|
"loss": 0.0494, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 9.256198347107437, |
|
"grad_norm": 0.3531731963157654, |
|
"learning_rate": 1.5062327174197644e-06, |
|
"loss": 0.0431, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 9.338842975206612, |
|
"grad_norm": 1.0387060642242432, |
|
"learning_rate": 1.191372027768034e-06, |
|
"loss": 0.0478, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 9.421487603305785, |
|
"grad_norm": 0.3243672251701355, |
|
"learning_rate": 9.129975882847364e-07, |
|
"loss": 0.0393, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 9.50413223140496, |
|
"grad_norm": 0.33455690741539, |
|
"learning_rate": 6.713174941338162e-07, |
|
"loss": 0.0419, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 9.586776859504132, |
|
"grad_norm": 0.4129684567451477, |
|
"learning_rate": 4.6651241010226e-07, |
|
"loss": 0.0463, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 9.669421487603305, |
|
"grad_norm": 2.450713872909546, |
|
"learning_rate": 2.9873543554652106e-07, |
|
"loss": 0.0448, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 9.75206611570248, |
|
"grad_norm": 0.22442331910133362, |
|
"learning_rate": 1.681119899450856e-07, |
|
"loss": 0.0462, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 9.834710743801653, |
|
"grad_norm": 0.9303910732269287, |
|
"learning_rate": 7.473971914280787e-08, |
|
"loss": 0.0449, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 9.917355371900827, |
|
"grad_norm": 1.7050765752792358, |
|
"learning_rate": 1.8688422357004966e-08, |
|
"loss": 0.0365, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"grad_norm": 0.60956209897995, |
|
"learning_rate": 0.0, |
|
"loss": 0.0377, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"step": 1210, |
|
"total_flos": 1.3059661832211168e+17, |
|
"train_loss": 0.09883986462738888, |
|
"train_runtime": 1198.7826, |
|
"train_samples_per_second": 49.142, |
|
"train_steps_per_second": 1.009 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1210, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 10, |
|
"save_steps": 10000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.3059661832211168e+17, |
|
"train_batch_size": 49, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|