|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 0.50003543837267, |
|
"eval_steps": 500, |
|
"global_step": 7055, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 7.0876745339854e-05, |
|
"eval_accuracy": 0.6605764392507908, |
|
"eval_loss": 1.4750796556472778, |
|
"eval_runtime": 7.2429, |
|
"eval_samples_per_second": 47.909, |
|
"eval_steps_per_second": 0.414, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.00070876745339854, |
|
"grad_norm": 1.8515625, |
|
"learning_rate": 1e-05, |
|
"loss": 1.538, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.00141753490679708, |
|
"grad_norm": 1.0, |
|
"learning_rate": 2e-05, |
|
"loss": 1.5311, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.00212630236019562, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 3e-05, |
|
"loss": 1.5247, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.00283506981359416, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 4e-05, |
|
"loss": 1.531, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.0035438372669927, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 5e-05, |
|
"loss": 1.5146, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.00425260472039124, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 6e-05, |
|
"loss": 1.5055, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.00496137217378978, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 7e-05, |
|
"loss": 1.5345, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.00567013962718832, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 8e-05, |
|
"loss": 1.5133, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.00637890708058686, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 9e-05, |
|
"loss": 1.5043, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.0070876745339854, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.513, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.00779644198738394, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5083, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.00850520944078248, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5098, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.00921397689418102, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4936, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.00992274434757956, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5091, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.0106315118009781, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5073, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.01134027925437664, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5161, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.01204904670777518, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5067, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.01275781416117372, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4984, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.01346658161457226, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5053, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.0141753490679708, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4937, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.01488411652136934, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4946, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.01559288397476788, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5031, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.01630165142816642, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4857, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.01701041888156496, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5134, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.0177191863349635, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5069, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.01842795378836204, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5037, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.01913672124176058, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4999, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.01984548869515912, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5038, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.02055425614855766, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4985, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.0212630236019562, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5112, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.02197179105535474, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5083, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.02268055850875328, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5104, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.02338932596215182, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5132, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.02409809341555036, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4956, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.0248068608689489, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4949, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.02551562832234744, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5018, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.02622439577574598, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4821, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.02693316322914452, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5092, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.02764193068254306, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4998, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.0283506981359416, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4958, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.02905946558934014, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5031, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.02976823304273868, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4929, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.03047700049613722, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4987, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.03118576794953576, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4853, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.0318945354029343, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4968, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.03260330285633284, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4872, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.03331207030973138, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5061, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.03402083776312992, |
|
"grad_norm": 0.6484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5078, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.03472960521652846, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.504, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.035438372669927, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5071, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.035438372669927, |
|
"eval_accuracy": 0.6646616911746595, |
|
"eval_loss": 1.4113208055496216, |
|
"eval_runtime": 7.1977, |
|
"eval_samples_per_second": 48.21, |
|
"eval_steps_per_second": 0.417, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.03614714012332554, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5028, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.03685590757672408, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5042, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.03756467503012262, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4969, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.03827344248352116, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5016, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.0389822099369197, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5008, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.03969097739031824, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5126, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.04039974484371678, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4949, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.04110851229711532, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4912, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.04181727975051386, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4912, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.0425260472039124, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4945, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.04323481465731094, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4977, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.04394358211070948, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4908, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.04465234956410802, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4937, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.04536111701750656, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4966, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.0460698844709051, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4982, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.04677865192430364, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4889, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.04748741937770218, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4948, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.04819618683110072, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5004, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.04890495428449926, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4888, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.0496137217378978, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4981, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.05032248919129634, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.504, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.05103125664469488, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4909, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.05174002409809342, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4971, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.05244879155149196, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4994, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.0531575590048905, |
|
"grad_norm": 0.60546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4926, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.05386632645828904, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4945, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.05457509391168758, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5003, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.05528386136508612, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4998, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.05599262881848466, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5005, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.0567013962718832, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.509, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.05741016372528174, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4885, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.05811893117868028, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4957, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.05882769863207882, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5006, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.05953646608547736, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.493, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.0602452335388759, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5008, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.06095400099227444, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4966, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.06166276844567298, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5013, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.06237153589907152, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4937, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.06308030335247006, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4815, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.0637890708058686, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4878, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.06449783825926714, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4948, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.06520660571266568, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5051, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.06591537316606422, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4993, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.06662414061946276, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4984, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.0673329080728613, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5001, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.06804167552625984, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4899, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.06875044297965838, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4881, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.06945921043305692, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5096, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.07016797788645546, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4902, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.070876745339854, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5003, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.070876745339854, |
|
"eval_accuracy": 0.6648911127297294, |
|
"eval_loss": 1.4080075025558472, |
|
"eval_runtime": 7.1527, |
|
"eval_samples_per_second": 48.513, |
|
"eval_steps_per_second": 0.419, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.07158551279325254, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4882, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.07229428024665108, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5019, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.07300304770004962, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.488, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.07371181515344816, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5065, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.0744205826068467, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4906, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.07512935006024524, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5036, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.07583811751364378, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4968, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.07654688496704232, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5076, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.07725565242044086, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5008, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.0779644198738394, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4951, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.07867318732723794, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4844, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.07938195478063648, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.499, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.08009072223403502, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4794, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.08079948968743356, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.508, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.0815082571408321, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4997, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.08221702459423064, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4987, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.08292579204762918, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5028, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.08363455950102772, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4991, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.08434332695442626, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5075, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.0850520944078248, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4845, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.08576086186122334, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4993, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.08646962931462188, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.502, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.08717839676802042, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4859, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.08788716422141896, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5029, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.0885959316748175, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5035, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.08930469912821604, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5057, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.09001346658161458, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4862, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.09072223403501312, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4941, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.09143100148841166, |
|
"grad_norm": 0.427734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4995, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.0921397689418102, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4891, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.09284853639520874, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5021, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.09355730384860728, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4896, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.09426607130200582, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4889, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.09497483875540436, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4978, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.0956836062088029, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.505, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.09639237366220144, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4934, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.09710114111559998, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4956, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.09780990856899852, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4973, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.09851867602239706, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4906, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.0992274434757956, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4942, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.09993621092919414, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4881, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.10064497838259268, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4837, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.10135374583599122, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4954, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.10206251328938976, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4838, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.1027712807427883, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4998, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.10348004819618684, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4998, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.10418881564958538, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4876, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.10489758310298392, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4965, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.10560635055638246, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4875, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.106315118009781, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4959, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.106315118009781, |
|
"eval_accuracy": 0.6653851431949415, |
|
"eval_loss": 1.4062947034835815, |
|
"eval_runtime": 7.3356, |
|
"eval_samples_per_second": 47.304, |
|
"eval_steps_per_second": 0.409, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.10702388546317954, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4954, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.10773265291657808, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4988, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.10844142036997662, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4895, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.10915018782337516, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5043, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.1098589552767737, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4746, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.11056772273017224, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4898, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.11127649018357078, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4876, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.11198525763696932, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4966, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.11269402509036786, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5048, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.1134027925437664, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.482, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.11411155999716494, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4951, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.11482032745056348, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.485, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.11552909490396202, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.491, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.11623786235736056, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4785, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.1169466298107591, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4971, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.11765539726415764, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5013, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.11836416471755618, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5077, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.11907293217095472, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4939, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.11978169962435326, |
|
"grad_norm": 0.63671875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4829, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.1204904670777518, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4964, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.12119923453115033, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4969, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.12190800198454887, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4955, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.12261676943794741, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4913, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.12332553689134595, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4905, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.1240343043447445, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5034, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.12474307179814303, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4991, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.12545183925154157, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5019, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.12616060670494011, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4943, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.12686937415833865, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4971, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.1275781416117372, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4961, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.12828690906513573, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4888, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.12899567651853427, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4974, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.12970444397193281, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4866, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.13041321142533135, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4878, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.1311219788787299, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4855, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.13183074633212843, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.503, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.13253951378552697, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4955, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.13324828123892551, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4956, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.13395704869232405, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4922, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.1346658161457226, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5161, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.13537458359912113, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4909, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.13608335105251967, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4936, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.13679211850591821, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4882, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.13750088595931675, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4745, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.1382096534127153, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4964, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.13891842086611383, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5041, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.13962718831951237, |
|
"grad_norm": 0.62890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4916, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.14033595577291091, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4992, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.14104472322630945, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5001, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.141753490679708, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5019, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.141753490679708, |
|
"eval_accuracy": 0.6655364488217514, |
|
"eval_loss": 1.4053725004196167, |
|
"eval_runtime": 7.1455, |
|
"eval_samples_per_second": 48.562, |
|
"eval_steps_per_second": 0.42, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.14246225813310653, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5015, |
|
"step": 2010 |
|
}, |
|
{ |
|
"epoch": 0.14317102558650507, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4842, |
|
"step": 2020 |
|
}, |
|
{ |
|
"epoch": 0.14387979303990361, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4868, |
|
"step": 2030 |
|
}, |
|
{ |
|
"epoch": 0.14458856049330215, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4947, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.1452973279467007, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4964, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 0.14600609540009923, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5029, |
|
"step": 2060 |
|
}, |
|
{ |
|
"epoch": 0.14671486285349777, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4974, |
|
"step": 2070 |
|
}, |
|
{ |
|
"epoch": 0.1474236303068963, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.488, |
|
"step": 2080 |
|
}, |
|
{ |
|
"epoch": 0.14813239776029485, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5008, |
|
"step": 2090 |
|
}, |
|
{ |
|
"epoch": 0.1488411652136934, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5018, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.14954993266709193, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5048, |
|
"step": 2110 |
|
}, |
|
{ |
|
"epoch": 0.15025870012049047, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5077, |
|
"step": 2120 |
|
}, |
|
{ |
|
"epoch": 0.150967467573889, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4962, |
|
"step": 2130 |
|
}, |
|
{ |
|
"epoch": 0.15167623502728755, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4966, |
|
"step": 2140 |
|
}, |
|
{ |
|
"epoch": 0.1523850024806861, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4926, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 0.15309376993408463, |
|
"grad_norm": 0.640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4895, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.15380253738748317, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4877, |
|
"step": 2170 |
|
}, |
|
{ |
|
"epoch": 0.1545113048408817, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4981, |
|
"step": 2180 |
|
}, |
|
{ |
|
"epoch": 0.15522007229428025, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4947, |
|
"step": 2190 |
|
}, |
|
{ |
|
"epoch": 0.1559288397476788, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4891, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.15663760720107733, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5004, |
|
"step": 2210 |
|
}, |
|
{ |
|
"epoch": 0.15734637465447587, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4877, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.1580551421078744, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4966, |
|
"step": 2230 |
|
}, |
|
{ |
|
"epoch": 0.15876390956127295, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4934, |
|
"step": 2240 |
|
}, |
|
{ |
|
"epoch": 0.1594726770146715, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4981, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 0.16018144446807003, |
|
"grad_norm": 0.6328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5087, |
|
"step": 2260 |
|
}, |
|
{ |
|
"epoch": 0.16089021192146857, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4989, |
|
"step": 2270 |
|
}, |
|
{ |
|
"epoch": 0.1615989793748671, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4785, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.16230774682826565, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4978, |
|
"step": 2290 |
|
}, |
|
{ |
|
"epoch": 0.1630165142816642, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4931, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.16372528173506273, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5119, |
|
"step": 2310 |
|
}, |
|
{ |
|
"epoch": 0.16443404918846127, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5097, |
|
"step": 2320 |
|
}, |
|
{ |
|
"epoch": 0.1651428166418598, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.481, |
|
"step": 2330 |
|
}, |
|
{ |
|
"epoch": 0.16585158409525835, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4884, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.1665603515486569, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4934, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 0.16726911900205543, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4789, |
|
"step": 2360 |
|
}, |
|
{ |
|
"epoch": 0.16797788645545397, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5063, |
|
"step": 2370 |
|
}, |
|
{ |
|
"epoch": 0.1686866539088525, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5095, |
|
"step": 2380 |
|
}, |
|
{ |
|
"epoch": 0.16939542136225105, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4949, |
|
"step": 2390 |
|
}, |
|
{ |
|
"epoch": 0.1701041888156496, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4918, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.17081295626904813, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5029, |
|
"step": 2410 |
|
}, |
|
{ |
|
"epoch": 0.17152172372244667, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4945, |
|
"step": 2420 |
|
}, |
|
{ |
|
"epoch": 0.1722304911758452, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4933, |
|
"step": 2430 |
|
}, |
|
{ |
|
"epoch": 0.17293925862924375, |
|
"grad_norm": 0.61328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4949, |
|
"step": 2440 |
|
}, |
|
{ |
|
"epoch": 0.1736480260826423, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5043, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 0.17435679353604083, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4971, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.17506556098943937, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.49, |
|
"step": 2470 |
|
}, |
|
{ |
|
"epoch": 0.1757743284428379, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4892, |
|
"step": 2480 |
|
}, |
|
{ |
|
"epoch": 0.17648309589623645, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5006, |
|
"step": 2490 |
|
}, |
|
{ |
|
"epoch": 0.177191863349635, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4891, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.177191863349635, |
|
"eval_accuracy": 0.6656455296224749, |
|
"eval_loss": 1.4047455787658691, |
|
"eval_runtime": 7.1711, |
|
"eval_samples_per_second": 48.388, |
|
"eval_steps_per_second": 0.418, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.17790063080303353, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4883, |
|
"step": 2510 |
|
}, |
|
{ |
|
"epoch": 0.17860939825643207, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4906, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.1793181657098306, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4946, |
|
"step": 2530 |
|
}, |
|
{ |
|
"epoch": 0.18002693316322915, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4933, |
|
"step": 2540 |
|
}, |
|
{ |
|
"epoch": 0.1807357006166277, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4794, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 0.18144446807002623, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5019, |
|
"step": 2560 |
|
}, |
|
{ |
|
"epoch": 0.18215323552342477, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4956, |
|
"step": 2570 |
|
}, |
|
{ |
|
"epoch": 0.1828620029768233, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5043, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.18357077043022185, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4928, |
|
"step": 2590 |
|
}, |
|
{ |
|
"epoch": 0.1842795378836204, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5021, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.18498830533701893, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5033, |
|
"step": 2610 |
|
}, |
|
{ |
|
"epoch": 0.18569707279041747, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4875, |
|
"step": 2620 |
|
}, |
|
{ |
|
"epoch": 0.186405840243816, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4913, |
|
"step": 2630 |
|
}, |
|
{ |
|
"epoch": 0.18711460769721455, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4929, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.1878233751506131, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4948, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 0.18853214260401163, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5015, |
|
"step": 2660 |
|
}, |
|
{ |
|
"epoch": 0.18924091005741017, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4996, |
|
"step": 2670 |
|
}, |
|
{ |
|
"epoch": 0.1899496775108087, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.487, |
|
"step": 2680 |
|
}, |
|
{ |
|
"epoch": 0.19065844496420725, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4944, |
|
"step": 2690 |
|
}, |
|
{ |
|
"epoch": 0.1913672124176058, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4889, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.19207597987100433, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4996, |
|
"step": 2710 |
|
}, |
|
{ |
|
"epoch": 0.19278474732440287, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4908, |
|
"step": 2720 |
|
}, |
|
{ |
|
"epoch": 0.1934935147778014, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4877, |
|
"step": 2730 |
|
}, |
|
{ |
|
"epoch": 0.19420228223119995, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4824, |
|
"step": 2740 |
|
}, |
|
{ |
|
"epoch": 0.1949110496845985, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4925, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 0.19561981713799703, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4895, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.19632858459139557, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.489, |
|
"step": 2770 |
|
}, |
|
{ |
|
"epoch": 0.1970373520447941, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4968, |
|
"step": 2780 |
|
}, |
|
{ |
|
"epoch": 0.19774611949819265, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4985, |
|
"step": 2790 |
|
}, |
|
{ |
|
"epoch": 0.1984548869515912, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.482, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.19916365440498973, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5019, |
|
"step": 2810 |
|
}, |
|
{ |
|
"epoch": 0.19987242185838827, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4914, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.2005811893117868, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4989, |
|
"step": 2830 |
|
}, |
|
{ |
|
"epoch": 0.20128995676518535, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4927, |
|
"step": 2840 |
|
}, |
|
{ |
|
"epoch": 0.2019987242185839, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.496, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 0.20270749167198243, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4828, |
|
"step": 2860 |
|
}, |
|
{ |
|
"epoch": 0.20341625912538097, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4945, |
|
"step": 2870 |
|
}, |
|
{ |
|
"epoch": 0.2041250265787795, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4906, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.20483379403217805, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5005, |
|
"step": 2890 |
|
}, |
|
{ |
|
"epoch": 0.2055425614855766, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5009, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.20625132893897513, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4862, |
|
"step": 2910 |
|
}, |
|
{ |
|
"epoch": 0.20696009639237367, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4833, |
|
"step": 2920 |
|
}, |
|
{ |
|
"epoch": 0.2076688638457722, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4975, |
|
"step": 2930 |
|
}, |
|
{ |
|
"epoch": 0.20837763129917075, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4992, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.2090863987525693, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4878, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 0.20979516620596783, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4978, |
|
"step": 2960 |
|
}, |
|
{ |
|
"epoch": 0.21050393365936637, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.484, |
|
"step": 2970 |
|
}, |
|
{ |
|
"epoch": 0.2112127011127649, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5065, |
|
"step": 2980 |
|
}, |
|
{ |
|
"epoch": 0.21192146856616345, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4829, |
|
"step": 2990 |
|
}, |
|
{ |
|
"epoch": 0.212630236019562, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4916, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.212630236019562, |
|
"eval_accuracy": 0.6657292755275465, |
|
"eval_loss": 1.4040066003799438, |
|
"eval_runtime": 7.1205, |
|
"eval_samples_per_second": 48.733, |
|
"eval_steps_per_second": 0.421, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.21333900347296053, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4859, |
|
"step": 3010 |
|
}, |
|
{ |
|
"epoch": 0.21404777092635907, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4907, |
|
"step": 3020 |
|
}, |
|
{ |
|
"epoch": 0.2147565383797576, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4771, |
|
"step": 3030 |
|
}, |
|
{ |
|
"epoch": 0.21546530583315615, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4868, |
|
"step": 3040 |
|
}, |
|
{ |
|
"epoch": 0.2161740732865547, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4941, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 0.21688284073995323, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4951, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.21759160819335177, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4915, |
|
"step": 3070 |
|
}, |
|
{ |
|
"epoch": 0.2183003756467503, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4869, |
|
"step": 3080 |
|
}, |
|
{ |
|
"epoch": 0.21900914310014885, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5057, |
|
"step": 3090 |
|
}, |
|
{ |
|
"epoch": 0.2197179105535474, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4941, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.22042667800694593, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5017, |
|
"step": 3110 |
|
}, |
|
{ |
|
"epoch": 0.22113544546034447, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.511, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.221844212913743, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4903, |
|
"step": 3130 |
|
}, |
|
{ |
|
"epoch": 0.22255298036714155, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5098, |
|
"step": 3140 |
|
}, |
|
{ |
|
"epoch": 0.2232617478205401, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4989, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 0.22397051527393863, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4935, |
|
"step": 3160 |
|
}, |
|
{ |
|
"epoch": 0.22467928272733717, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4907, |
|
"step": 3170 |
|
}, |
|
{ |
|
"epoch": 0.2253880501807357, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5033, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 0.22609681763413425, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4982, |
|
"step": 3190 |
|
}, |
|
{ |
|
"epoch": 0.2268055850875328, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4853, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.22751435254093133, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.498, |
|
"step": 3210 |
|
}, |
|
{ |
|
"epoch": 0.22822311999432987, |
|
"grad_norm": 0.6171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.489, |
|
"step": 3220 |
|
}, |
|
{ |
|
"epoch": 0.2289318874477284, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4758, |
|
"step": 3230 |
|
}, |
|
{ |
|
"epoch": 0.22964065490112695, |
|
"grad_norm": 0.69921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4896, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 0.2303494223545255, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.479, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 0.23105818980792403, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4968, |
|
"step": 3260 |
|
}, |
|
{ |
|
"epoch": 0.23176695726132257, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4913, |
|
"step": 3270 |
|
}, |
|
{ |
|
"epoch": 0.2324757247147211, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4856, |
|
"step": 3280 |
|
}, |
|
{ |
|
"epoch": 0.23318449216811965, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4969, |
|
"step": 3290 |
|
}, |
|
{ |
|
"epoch": 0.2338932596215182, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4968, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.23460202707491673, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4904, |
|
"step": 3310 |
|
}, |
|
{ |
|
"epoch": 0.23531079452831527, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.484, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.2360195619817138, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4928, |
|
"step": 3330 |
|
}, |
|
{ |
|
"epoch": 0.23672832943511235, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4964, |
|
"step": 3340 |
|
}, |
|
{ |
|
"epoch": 0.2374370968885109, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4974, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 0.23814586434190943, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.503, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 0.23885463179530797, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5029, |
|
"step": 3370 |
|
}, |
|
{ |
|
"epoch": 0.2395633992487065, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.501, |
|
"step": 3380 |
|
}, |
|
{ |
|
"epoch": 0.24027216670210505, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4885, |
|
"step": 3390 |
|
}, |
|
{ |
|
"epoch": 0.2409809341555036, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4904, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.24168970160890213, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4851, |
|
"step": 3410 |
|
}, |
|
{ |
|
"epoch": 0.24239846906230067, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4783, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 0.2431072365156992, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4948, |
|
"step": 3430 |
|
}, |
|
{ |
|
"epoch": 0.24381600396909775, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4754, |
|
"step": 3440 |
|
}, |
|
{ |
|
"epoch": 0.2445247714224963, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4936, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 0.24523353887589483, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4814, |
|
"step": 3460 |
|
}, |
|
{ |
|
"epoch": 0.24594230632929337, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.505, |
|
"step": 3470 |
|
}, |
|
{ |
|
"epoch": 0.2466510737826919, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4888, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 0.24735984123609045, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4974, |
|
"step": 3490 |
|
}, |
|
{ |
|
"epoch": 0.248068608689489, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.496, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.248068608689489, |
|
"eval_accuracy": 0.665728571780445, |
|
"eval_loss": 1.403409481048584, |
|
"eval_runtime": 7.1308, |
|
"eval_samples_per_second": 48.662, |
|
"eval_steps_per_second": 0.421, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.24877737614288753, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4988, |
|
"step": 3510 |
|
}, |
|
{ |
|
"epoch": 0.24948614359628607, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5036, |
|
"step": 3520 |
|
}, |
|
{ |
|
"epoch": 0.2501949110496846, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5038, |
|
"step": 3530 |
|
}, |
|
{ |
|
"epoch": 0.25090367850308315, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4947, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 0.2516124459564817, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4914, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 0.25232121340988023, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4855, |
|
"step": 3560 |
|
}, |
|
{ |
|
"epoch": 0.25302998086327877, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4845, |
|
"step": 3570 |
|
}, |
|
{ |
|
"epoch": 0.2537387483166773, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5041, |
|
"step": 3580 |
|
}, |
|
{ |
|
"epoch": 0.25444751577007585, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4803, |
|
"step": 3590 |
|
}, |
|
{ |
|
"epoch": 0.2551562832234744, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4938, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.25586505067687293, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4938, |
|
"step": 3610 |
|
}, |
|
{ |
|
"epoch": 0.25657381813027147, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4972, |
|
"step": 3620 |
|
}, |
|
{ |
|
"epoch": 0.25728258558367, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4941, |
|
"step": 3630 |
|
}, |
|
{ |
|
"epoch": 0.25799135303706855, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4881, |
|
"step": 3640 |
|
}, |
|
{ |
|
"epoch": 0.2587001204904671, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4906, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 0.25940888794386563, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4937, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 0.26011765539726417, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4904, |
|
"step": 3670 |
|
}, |
|
{ |
|
"epoch": 0.2608264228506627, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5034, |
|
"step": 3680 |
|
}, |
|
{ |
|
"epoch": 0.26153519030406125, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4727, |
|
"step": 3690 |
|
}, |
|
{ |
|
"epoch": 0.2622439577574598, |
|
"grad_norm": 0.421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4919, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.26295272521085833, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4842, |
|
"step": 3710 |
|
}, |
|
{ |
|
"epoch": 0.26366149266425687, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4953, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 0.2643702601176554, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4739, |
|
"step": 3730 |
|
}, |
|
{ |
|
"epoch": 0.26507902757105395, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4913, |
|
"step": 3740 |
|
}, |
|
{ |
|
"epoch": 0.2657877950244525, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5003, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 0.26649656247785103, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4919, |
|
"step": 3760 |
|
}, |
|
{ |
|
"epoch": 0.26720532993124957, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4818, |
|
"step": 3770 |
|
}, |
|
{ |
|
"epoch": 0.2679140973846481, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5003, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 0.26862286483804665, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4934, |
|
"step": 3790 |
|
}, |
|
{ |
|
"epoch": 0.2693316322914452, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4869, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.27004039974484373, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4896, |
|
"step": 3810 |
|
}, |
|
{ |
|
"epoch": 0.27074916719824227, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5021, |
|
"step": 3820 |
|
}, |
|
{ |
|
"epoch": 0.2714579346516408, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5048, |
|
"step": 3830 |
|
}, |
|
{ |
|
"epoch": 0.27216670210503935, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4963, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 0.2728754695584379, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4967, |
|
"step": 3850 |
|
}, |
|
{ |
|
"epoch": 0.27358423701183643, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5036, |
|
"step": 3860 |
|
}, |
|
{ |
|
"epoch": 0.27429300446523497, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4944, |
|
"step": 3870 |
|
}, |
|
{ |
|
"epoch": 0.2750017719186335, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4951, |
|
"step": 3880 |
|
}, |
|
{ |
|
"epoch": 0.27571053937203205, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5006, |
|
"step": 3890 |
|
}, |
|
{ |
|
"epoch": 0.2764193068254306, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4886, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.27712807427882913, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4931, |
|
"step": 3910 |
|
}, |
|
{ |
|
"epoch": 0.27783684173222767, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4926, |
|
"step": 3920 |
|
}, |
|
{ |
|
"epoch": 0.2785456091856262, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4962, |
|
"step": 3930 |
|
}, |
|
{ |
|
"epoch": 0.27925437663902475, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4933, |
|
"step": 3940 |
|
}, |
|
{ |
|
"epoch": 0.2799631440924233, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4959, |
|
"step": 3950 |
|
}, |
|
{ |
|
"epoch": 0.28067191154582183, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.49, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 0.28138067899922037, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5087, |
|
"step": 3970 |
|
}, |
|
{ |
|
"epoch": 0.2820894464526189, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4939, |
|
"step": 3980 |
|
}, |
|
{ |
|
"epoch": 0.28279821390601745, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4913, |
|
"step": 3990 |
|
}, |
|
{ |
|
"epoch": 0.283506981359416, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.495, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.283506981359416, |
|
"eval_accuracy": 0.6657313867688508, |
|
"eval_loss": 1.403212308883667, |
|
"eval_runtime": 7.176, |
|
"eval_samples_per_second": 48.355, |
|
"eval_steps_per_second": 0.418, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.28421574881281453, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4793, |
|
"step": 4010 |
|
}, |
|
{ |
|
"epoch": 0.28492451626621307, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.484, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 0.2856332837196116, |
|
"grad_norm": 0.578125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4967, |
|
"step": 4030 |
|
}, |
|
{ |
|
"epoch": 0.28634205117301015, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4952, |
|
"step": 4040 |
|
}, |
|
{ |
|
"epoch": 0.2870508186264087, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.494, |
|
"step": 4050 |
|
}, |
|
{ |
|
"epoch": 0.28775958607980723, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4888, |
|
"step": 4060 |
|
}, |
|
{ |
|
"epoch": 0.28846835353320577, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4784, |
|
"step": 4070 |
|
}, |
|
{ |
|
"epoch": 0.2891771209866043, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4924, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 0.28988588844000285, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5047, |
|
"step": 4090 |
|
}, |
|
{ |
|
"epoch": 0.2905946558934014, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4895, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.29130342334679993, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4852, |
|
"step": 4110 |
|
}, |
|
{ |
|
"epoch": 0.29201219080019847, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5075, |
|
"step": 4120 |
|
}, |
|
{ |
|
"epoch": 0.292720958253597, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5061, |
|
"step": 4130 |
|
}, |
|
{ |
|
"epoch": 0.29342972570699555, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4953, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 0.2941384931603941, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4956, |
|
"step": 4150 |
|
}, |
|
{ |
|
"epoch": 0.2948472606137926, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4978, |
|
"step": 4160 |
|
}, |
|
{ |
|
"epoch": 0.29555602806719117, |
|
"grad_norm": 0.7421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4759, |
|
"step": 4170 |
|
}, |
|
{ |
|
"epoch": 0.2962647955205897, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4955, |
|
"step": 4180 |
|
}, |
|
{ |
|
"epoch": 0.29697356297398825, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4955, |
|
"step": 4190 |
|
}, |
|
{ |
|
"epoch": 0.2976823304273868, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.482, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.2983910978807853, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4774, |
|
"step": 4210 |
|
}, |
|
{ |
|
"epoch": 0.29909986533418387, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.484, |
|
"step": 4220 |
|
}, |
|
{ |
|
"epoch": 0.2998086327875824, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5073, |
|
"step": 4230 |
|
}, |
|
{ |
|
"epoch": 0.30051740024098095, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5047, |
|
"step": 4240 |
|
}, |
|
{ |
|
"epoch": 0.3012261676943795, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4898, |
|
"step": 4250 |
|
}, |
|
{ |
|
"epoch": 0.301934935147778, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4914, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 0.30264370260117657, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4838, |
|
"step": 4270 |
|
}, |
|
{ |
|
"epoch": 0.3033524700545751, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4908, |
|
"step": 4280 |
|
}, |
|
{ |
|
"epoch": 0.30406123750797365, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4838, |
|
"step": 4290 |
|
}, |
|
{ |
|
"epoch": 0.3047700049613722, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4871, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.3054787724147707, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5091, |
|
"step": 4310 |
|
}, |
|
{ |
|
"epoch": 0.30618753986816927, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4968, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 0.3068963073215678, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4926, |
|
"step": 4330 |
|
}, |
|
{ |
|
"epoch": 0.30760507477496635, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4982, |
|
"step": 4340 |
|
}, |
|
{ |
|
"epoch": 0.3083138422283649, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5001, |
|
"step": 4350 |
|
}, |
|
{ |
|
"epoch": 0.3090226096817634, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4939, |
|
"step": 4360 |
|
}, |
|
{ |
|
"epoch": 0.30973137713516197, |
|
"grad_norm": 0.49609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4979, |
|
"step": 4370 |
|
}, |
|
{ |
|
"epoch": 0.3104401445885605, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4799, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 0.31114891204195905, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4945, |
|
"step": 4390 |
|
}, |
|
{ |
|
"epoch": 0.3118576794953576, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4948, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.3125664469487561, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4942, |
|
"step": 4410 |
|
}, |
|
{ |
|
"epoch": 0.31327521440215467, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5039, |
|
"step": 4420 |
|
}, |
|
{ |
|
"epoch": 0.3139839818555532, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4882, |
|
"step": 4430 |
|
}, |
|
{ |
|
"epoch": 0.31469274930895175, |
|
"grad_norm": 0.4296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4925, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 0.3154015167623503, |
|
"grad_norm": 0.65625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5131, |
|
"step": 4450 |
|
}, |
|
{ |
|
"epoch": 0.3161102842157488, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5026, |
|
"step": 4460 |
|
}, |
|
{ |
|
"epoch": 0.31681905166914737, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4959, |
|
"step": 4470 |
|
}, |
|
{ |
|
"epoch": 0.3175278191225459, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4972, |
|
"step": 4480 |
|
}, |
|
{ |
|
"epoch": 0.31823658657594445, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4922, |
|
"step": 4490 |
|
}, |
|
{ |
|
"epoch": 0.318945354029343, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4934, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.318945354029343, |
|
"eval_accuracy": 0.6658446900521828, |
|
"eval_loss": 1.4029836654663086, |
|
"eval_runtime": 7.1758, |
|
"eval_samples_per_second": 48.357, |
|
"eval_steps_per_second": 0.418, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.3196541214827415, |
|
"grad_norm": 0.57421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4693, |
|
"step": 4510 |
|
}, |
|
{ |
|
"epoch": 0.32036288893614007, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.498, |
|
"step": 4520 |
|
}, |
|
{ |
|
"epoch": 0.3210716563895386, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4872, |
|
"step": 4530 |
|
}, |
|
{ |
|
"epoch": 0.32178042384293715, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4866, |
|
"step": 4540 |
|
}, |
|
{ |
|
"epoch": 0.3224891912963357, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4919, |
|
"step": 4550 |
|
}, |
|
{ |
|
"epoch": 0.3231979587497342, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4765, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 0.32390672620313277, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4919, |
|
"step": 4570 |
|
}, |
|
{ |
|
"epoch": 0.3246154936565313, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5069, |
|
"step": 4580 |
|
}, |
|
{ |
|
"epoch": 0.32532426110992985, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5014, |
|
"step": 4590 |
|
}, |
|
{ |
|
"epoch": 0.3260330285633284, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5081, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.3267417960167269, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5007, |
|
"step": 4610 |
|
}, |
|
{ |
|
"epoch": 0.32745056347012547, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4755, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 0.328159330923524, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4875, |
|
"step": 4630 |
|
}, |
|
{ |
|
"epoch": 0.32886809837692255, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4871, |
|
"step": 4640 |
|
}, |
|
{ |
|
"epoch": 0.3295768658303211, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5034, |
|
"step": 4650 |
|
}, |
|
{ |
|
"epoch": 0.3302856332837196, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4828, |
|
"step": 4660 |
|
}, |
|
{ |
|
"epoch": 0.33099440073711817, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.489, |
|
"step": 4670 |
|
}, |
|
{ |
|
"epoch": 0.3317031681905167, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5142, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 0.33241193564391525, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4918, |
|
"step": 4690 |
|
}, |
|
{ |
|
"epoch": 0.3331207030973138, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4862, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.3338294705507123, |
|
"grad_norm": 0.61328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4979, |
|
"step": 4710 |
|
}, |
|
{ |
|
"epoch": 0.33453823800411087, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4968, |
|
"step": 4720 |
|
}, |
|
{ |
|
"epoch": 0.3352470054575094, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4929, |
|
"step": 4730 |
|
}, |
|
{ |
|
"epoch": 0.33595577291090795, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4713, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 0.3366645403643065, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4866, |
|
"step": 4750 |
|
}, |
|
{ |
|
"epoch": 0.337373307817705, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4993, |
|
"step": 4760 |
|
}, |
|
{ |
|
"epoch": 0.33808207527110357, |
|
"grad_norm": 0.58203125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4966, |
|
"step": 4770 |
|
}, |
|
{ |
|
"epoch": 0.3387908427245021, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5022, |
|
"step": 4780 |
|
}, |
|
{ |
|
"epoch": 0.33949961017790065, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4908, |
|
"step": 4790 |
|
}, |
|
{ |
|
"epoch": 0.3402083776312992, |
|
"grad_norm": 0.66015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4884, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.3409171450846977, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5059, |
|
"step": 4810 |
|
}, |
|
{ |
|
"epoch": 0.34162591253809627, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4907, |
|
"step": 4820 |
|
}, |
|
{ |
|
"epoch": 0.3423346799914948, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4842, |
|
"step": 4830 |
|
}, |
|
{ |
|
"epoch": 0.34304344744489335, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4984, |
|
"step": 4840 |
|
}, |
|
{ |
|
"epoch": 0.3437522148982919, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4828, |
|
"step": 4850 |
|
}, |
|
{ |
|
"epoch": 0.3444609823516904, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4865, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 0.34516974980508897, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5001, |
|
"step": 4870 |
|
}, |
|
{ |
|
"epoch": 0.3458785172584875, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5026, |
|
"step": 4880 |
|
}, |
|
{ |
|
"epoch": 0.34658728471188605, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.488, |
|
"step": 4890 |
|
}, |
|
{ |
|
"epoch": 0.3472960521652846, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4846, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.3480048196186831, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4972, |
|
"step": 4910 |
|
}, |
|
{ |
|
"epoch": 0.34871358707208167, |
|
"grad_norm": 0.46484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.487, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 0.3494223545254802, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4945, |
|
"step": 4930 |
|
}, |
|
{ |
|
"epoch": 0.35013112197887875, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4942, |
|
"step": 4940 |
|
}, |
|
{ |
|
"epoch": 0.3508398894322773, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4944, |
|
"step": 4950 |
|
}, |
|
{ |
|
"epoch": 0.3515486568856758, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5099, |
|
"step": 4960 |
|
}, |
|
{ |
|
"epoch": 0.35225742433907437, |
|
"grad_norm": 0.494140625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4951, |
|
"step": 4970 |
|
}, |
|
{ |
|
"epoch": 0.3529661917924729, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5049, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.35367495924587145, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4764, |
|
"step": 4990 |
|
}, |
|
{ |
|
"epoch": 0.35438372669927, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4849, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.35438372669927, |
|
"eval_accuracy": 0.6660431467347894, |
|
"eval_loss": 1.4028522968292236, |
|
"eval_runtime": 7.1426, |
|
"eval_samples_per_second": 48.581, |
|
"eval_steps_per_second": 0.42, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.3550924941526685, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4884, |
|
"step": 5010 |
|
}, |
|
{ |
|
"epoch": 0.35580126160606707, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5002, |
|
"step": 5020 |
|
}, |
|
{ |
|
"epoch": 0.3565100290594656, |
|
"grad_norm": 0.498046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5019, |
|
"step": 5030 |
|
}, |
|
{ |
|
"epoch": 0.35721879651286415, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4813, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 0.3579275639662627, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4826, |
|
"step": 5050 |
|
}, |
|
{ |
|
"epoch": 0.3586363314196612, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5028, |
|
"step": 5060 |
|
}, |
|
{ |
|
"epoch": 0.35934509887305977, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4953, |
|
"step": 5070 |
|
}, |
|
{ |
|
"epoch": 0.3600538663264583, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4826, |
|
"step": 5080 |
|
}, |
|
{ |
|
"epoch": 0.36076263377985684, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4877, |
|
"step": 5090 |
|
}, |
|
{ |
|
"epoch": 0.3614714012332554, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4823, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.3621801686866539, |
|
"grad_norm": 0.55859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4957, |
|
"step": 5110 |
|
}, |
|
{ |
|
"epoch": 0.36288893614005246, |
|
"grad_norm": 0.546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4935, |
|
"step": 5120 |
|
}, |
|
{ |
|
"epoch": 0.363597703593451, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5076, |
|
"step": 5130 |
|
}, |
|
{ |
|
"epoch": 0.36430647104684954, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4848, |
|
"step": 5140 |
|
}, |
|
{ |
|
"epoch": 0.3650152385002481, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4892, |
|
"step": 5150 |
|
}, |
|
{ |
|
"epoch": 0.3657240059536466, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4944, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 0.36643277340704516, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4946, |
|
"step": 5170 |
|
}, |
|
{ |
|
"epoch": 0.3671415408604437, |
|
"grad_norm": 0.474609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4826, |
|
"step": 5180 |
|
}, |
|
{ |
|
"epoch": 0.36785030831384224, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4874, |
|
"step": 5190 |
|
}, |
|
{ |
|
"epoch": 0.3685590757672408, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5055, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.3692678432206393, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5005, |
|
"step": 5210 |
|
}, |
|
{ |
|
"epoch": 0.36997661067403786, |
|
"grad_norm": 0.46875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4842, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 0.3706853781274364, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.501, |
|
"step": 5230 |
|
}, |
|
{ |
|
"epoch": 0.37139414558083494, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5031, |
|
"step": 5240 |
|
}, |
|
{ |
|
"epoch": 0.3721029130342335, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4887, |
|
"step": 5250 |
|
}, |
|
{ |
|
"epoch": 0.372811680487632, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4779, |
|
"step": 5260 |
|
}, |
|
{ |
|
"epoch": 0.37352044794103056, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4779, |
|
"step": 5270 |
|
}, |
|
{ |
|
"epoch": 0.3742292153944291, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5006, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 0.37493798284782764, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5109, |
|
"step": 5290 |
|
}, |
|
{ |
|
"epoch": 0.3756467503012262, |
|
"grad_norm": 0.451171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4779, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.3763555177546247, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4861, |
|
"step": 5310 |
|
}, |
|
{ |
|
"epoch": 0.37706428520802326, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5001, |
|
"step": 5320 |
|
}, |
|
{ |
|
"epoch": 0.3777730526614218, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4876, |
|
"step": 5330 |
|
}, |
|
{ |
|
"epoch": 0.37848182011482034, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.486, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 0.3791905875682189, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4956, |
|
"step": 5350 |
|
}, |
|
{ |
|
"epoch": 0.3798993550216174, |
|
"grad_norm": 0.61328125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4972, |
|
"step": 5360 |
|
}, |
|
{ |
|
"epoch": 0.38060812247501596, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4926, |
|
"step": 5370 |
|
}, |
|
{ |
|
"epoch": 0.3813168899284145, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4982, |
|
"step": 5380 |
|
}, |
|
{ |
|
"epoch": 0.38202565738181304, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4949, |
|
"step": 5390 |
|
}, |
|
{ |
|
"epoch": 0.3827344248352116, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4928, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.3834431922886101, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4841, |
|
"step": 5410 |
|
}, |
|
{ |
|
"epoch": 0.38415195974200866, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4975, |
|
"step": 5420 |
|
}, |
|
{ |
|
"epoch": 0.3848607271954072, |
|
"grad_norm": 0.63671875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4907, |
|
"step": 5430 |
|
}, |
|
{ |
|
"epoch": 0.38556949464880574, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4933, |
|
"step": 5440 |
|
}, |
|
{ |
|
"epoch": 0.3862782621022043, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5083, |
|
"step": 5450 |
|
}, |
|
{ |
|
"epoch": 0.3869870295556028, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4766, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 0.38769579700900136, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.485, |
|
"step": 5470 |
|
}, |
|
{ |
|
"epoch": 0.3884045644623999, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4891, |
|
"step": 5480 |
|
}, |
|
{ |
|
"epoch": 0.38911333191579844, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.495, |
|
"step": 5490 |
|
}, |
|
{ |
|
"epoch": 0.389822099369197, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4833, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.389822099369197, |
|
"eval_accuracy": 0.6661451900644985, |
|
"eval_loss": 1.4023902416229248, |
|
"eval_runtime": 7.2106, |
|
"eval_samples_per_second": 48.124, |
|
"eval_steps_per_second": 0.416, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.3905308668225955, |
|
"grad_norm": 0.466796875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5, |
|
"step": 5510 |
|
}, |
|
{ |
|
"epoch": 0.39123963427599406, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4789, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 0.3919484017293926, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.498, |
|
"step": 5530 |
|
}, |
|
{ |
|
"epoch": 0.39265716918279114, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5051, |
|
"step": 5540 |
|
}, |
|
{ |
|
"epoch": 0.3933659366361897, |
|
"grad_norm": 0.66015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5041, |
|
"step": 5550 |
|
}, |
|
{ |
|
"epoch": 0.3940747040895882, |
|
"grad_norm": 0.458984375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4736, |
|
"step": 5560 |
|
}, |
|
{ |
|
"epoch": 0.39478347154298676, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4863, |
|
"step": 5570 |
|
}, |
|
{ |
|
"epoch": 0.3954922389963853, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4962, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 0.39620100644978384, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4904, |
|
"step": 5590 |
|
}, |
|
{ |
|
"epoch": 0.3969097739031824, |
|
"grad_norm": 0.54296875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4955, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.3976185413565809, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4925, |
|
"step": 5610 |
|
}, |
|
{ |
|
"epoch": 0.39832730880997946, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4945, |
|
"step": 5620 |
|
}, |
|
{ |
|
"epoch": 0.399036076263378, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5, |
|
"step": 5630 |
|
}, |
|
{ |
|
"epoch": 0.39974484371677654, |
|
"grad_norm": 0.6015625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4946, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 0.4004536111701751, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4868, |
|
"step": 5650 |
|
}, |
|
{ |
|
"epoch": 0.4011623786235736, |
|
"grad_norm": 0.52734375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4841, |
|
"step": 5660 |
|
}, |
|
{ |
|
"epoch": 0.40187114607697216, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.494, |
|
"step": 5670 |
|
}, |
|
{ |
|
"epoch": 0.4025799135303707, |
|
"grad_norm": 0.51171875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4844, |
|
"step": 5680 |
|
}, |
|
{ |
|
"epoch": 0.40328868098376924, |
|
"grad_norm": 0.55078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4932, |
|
"step": 5690 |
|
}, |
|
{ |
|
"epoch": 0.4039974484371678, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5066, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.4047062158905663, |
|
"grad_norm": 0.5390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.498, |
|
"step": 5710 |
|
}, |
|
{ |
|
"epoch": 0.40541498334396486, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5048, |
|
"step": 5720 |
|
}, |
|
{ |
|
"epoch": 0.4061237507973634, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4863, |
|
"step": 5730 |
|
}, |
|
{ |
|
"epoch": 0.40683251825076194, |
|
"grad_norm": 0.66796875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4899, |
|
"step": 5740 |
|
}, |
|
{ |
|
"epoch": 0.4075412857041605, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4792, |
|
"step": 5750 |
|
}, |
|
{ |
|
"epoch": 0.408250053157559, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4915, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 0.40895882061095756, |
|
"grad_norm": 0.515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4923, |
|
"step": 5770 |
|
}, |
|
{ |
|
"epoch": 0.4096675880643561, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4757, |
|
"step": 5780 |
|
}, |
|
{ |
|
"epoch": 0.41037635551775464, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4972, |
|
"step": 5790 |
|
}, |
|
{ |
|
"epoch": 0.4110851229711532, |
|
"grad_norm": 0.5625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4846, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.4117938904245517, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4888, |
|
"step": 5810 |
|
}, |
|
{ |
|
"epoch": 0.41250265787795026, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4996, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 0.4132114253313488, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4927, |
|
"step": 5830 |
|
}, |
|
{ |
|
"epoch": 0.41392019278474734, |
|
"grad_norm": 0.4765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4861, |
|
"step": 5840 |
|
}, |
|
{ |
|
"epoch": 0.4146289602381459, |
|
"grad_norm": 0.53125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4917, |
|
"step": 5850 |
|
}, |
|
{ |
|
"epoch": 0.4153377276915444, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5034, |
|
"step": 5860 |
|
}, |
|
{ |
|
"epoch": 0.41604649514494296, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5052, |
|
"step": 5870 |
|
}, |
|
{ |
|
"epoch": 0.4167552625983415, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5002, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 0.41746403005174004, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5054, |
|
"step": 5890 |
|
}, |
|
{ |
|
"epoch": 0.4181727975051386, |
|
"grad_norm": 0.5546875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.498, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.4188815649585371, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4894, |
|
"step": 5910 |
|
}, |
|
{ |
|
"epoch": 0.41959033241193566, |
|
"grad_norm": 0.59765625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5121, |
|
"step": 5920 |
|
}, |
|
{ |
|
"epoch": 0.4202990998653342, |
|
"grad_norm": 0.48828125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4975, |
|
"step": 5930 |
|
}, |
|
{ |
|
"epoch": 0.42100786731873274, |
|
"grad_norm": 0.6484375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4825, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 0.4217166347721313, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4985, |
|
"step": 5950 |
|
}, |
|
{ |
|
"epoch": 0.4224254022255298, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4922, |
|
"step": 5960 |
|
}, |
|
{ |
|
"epoch": 0.42313416967892836, |
|
"grad_norm": 0.50390625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4964, |
|
"step": 5970 |
|
}, |
|
{ |
|
"epoch": 0.4238429371323269, |
|
"grad_norm": 0.53515625, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4982, |
|
"step": 5980 |
|
}, |
|
{ |
|
"epoch": 0.42455170458572544, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4778, |
|
"step": 5990 |
|
}, |
|
{ |
|
"epoch": 0.425260472039124, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4909, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.425260472039124, |
|
"eval_accuracy": 0.6661219664101509, |
|
"eval_loss": 1.4022555351257324, |
|
"eval_runtime": 7.2496, |
|
"eval_samples_per_second": 47.865, |
|
"eval_steps_per_second": 0.414, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.4259692394925225, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5043, |
|
"step": 6010 |
|
}, |
|
{ |
|
"epoch": 0.42667800694592106, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.507, |
|
"step": 6020 |
|
}, |
|
{ |
|
"epoch": 0.4273867743993196, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 0.0001, |
|
"loss": 1.4907, |
|
"step": 6030 |
|
}, |
|
{ |
|
"epoch": 0.42809554185271814, |
|
"grad_norm": 0.5, |
|
"learning_rate": 0.0001, |
|
"loss": 1.491, |
|
"step": 6040 |
|
}, |
|
{ |
|
"epoch": 0.4288043093061167, |
|
"grad_norm": 0.5859375, |
|
"learning_rate": 0.0001, |
|
"loss": 1.5023, |
|
"step": 6050 |
|
}, |
|
{ |
|
"epoch": 0.4295130767595152, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.997960964140947e-05, |
|
"loss": 1.4967, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 0.43022184421291376, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.991845519630678e-05, |
|
"loss": 1.4897, |
|
"step": 6070 |
|
}, |
|
{ |
|
"epoch": 0.4309306116663123, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 9.981658654313457e-05, |
|
"loss": 1.4917, |
|
"step": 6080 |
|
}, |
|
{ |
|
"epoch": 0.43163937911971084, |
|
"grad_norm": 0.5, |
|
"learning_rate": 9.967408676742751e-05, |
|
"loss": 1.5015, |
|
"step": 6090 |
|
}, |
|
{ |
|
"epoch": 0.4323481465731094, |
|
"grad_norm": 0.486328125, |
|
"learning_rate": 9.949107209404665e-05, |
|
"loss": 1.4798, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.4330569140265079, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 9.926769179238466e-05, |
|
"loss": 1.4874, |
|
"step": 6110 |
|
}, |
|
{ |
|
"epoch": 0.43376568147990646, |
|
"grad_norm": 0.62109375, |
|
"learning_rate": 9.900412805461967e-05, |
|
"loss": 1.4951, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 0.434474448933305, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 9.870059584711668e-05, |
|
"loss": 1.4917, |
|
"step": 6130 |
|
}, |
|
{ |
|
"epoch": 0.43518321638670354, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 9.835734273509786e-05, |
|
"loss": 1.5102, |
|
"step": 6140 |
|
}, |
|
{ |
|
"epoch": 0.4358919838401021, |
|
"grad_norm": 0.482421875, |
|
"learning_rate": 9.797464868072488e-05, |
|
"loss": 1.485, |
|
"step": 6150 |
|
}, |
|
{ |
|
"epoch": 0.4366007512935006, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 9.755282581475769e-05, |
|
"loss": 1.4967, |
|
"step": 6160 |
|
}, |
|
{ |
|
"epoch": 0.43730951874689916, |
|
"grad_norm": 0.4609375, |
|
"learning_rate": 9.709221818197624e-05, |
|
"loss": 1.4827, |
|
"step": 6170 |
|
}, |
|
{ |
|
"epoch": 0.4380182862002977, |
|
"grad_norm": 0.56640625, |
|
"learning_rate": 9.659320146057262e-05, |
|
"loss": 1.4919, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 0.43872705365369624, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 9.60561826557425e-05, |
|
"loss": 1.5128, |
|
"step": 6190 |
|
}, |
|
{ |
|
"epoch": 0.4394358211070948, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 9.548159976772592e-05, |
|
"loss": 1.496, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.4401445885604933, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 9.486992143456792e-05, |
|
"loss": 1.4835, |
|
"step": 6210 |
|
}, |
|
{ |
|
"epoch": 0.44085335601389186, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 9.422164654989072e-05, |
|
"loss": 1.4907, |
|
"step": 6220 |
|
}, |
|
{ |
|
"epoch": 0.4415621234672904, |
|
"grad_norm": 0.5703125, |
|
"learning_rate": 9.353730385598887e-05, |
|
"loss": 1.4982, |
|
"step": 6230 |
|
}, |
|
{ |
|
"epoch": 0.44227089092068894, |
|
"grad_norm": 0.447265625, |
|
"learning_rate": 9.281745151257946e-05, |
|
"loss": 1.493, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 0.4429796583740875, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 9.206267664155907e-05, |
|
"loss": 1.4849, |
|
"step": 6250 |
|
}, |
|
{ |
|
"epoch": 0.443688425827486, |
|
"grad_norm": 0.484375, |
|
"learning_rate": 9.12735948481387e-05, |
|
"loss": 1.495, |
|
"step": 6260 |
|
}, |
|
{ |
|
"epoch": 0.44439719328088456, |
|
"grad_norm": 0.47265625, |
|
"learning_rate": 9.045084971874738e-05, |
|
"loss": 1.4908, |
|
"step": 6270 |
|
}, |
|
{ |
|
"epoch": 0.4451059607342831, |
|
"grad_norm": 0.462890625, |
|
"learning_rate": 8.959511229611376e-05, |
|
"loss": 1.4864, |
|
"step": 6280 |
|
}, |
|
{ |
|
"epoch": 0.44581472818768164, |
|
"grad_norm": 0.48046875, |
|
"learning_rate": 8.870708053195413e-05, |
|
"loss": 1.4839, |
|
"step": 6290 |
|
}, |
|
{ |
|
"epoch": 0.4465234956410802, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 8.778747871771292e-05, |
|
"loss": 1.4995, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.4472322630944787, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 8.683705689382024e-05, |
|
"loss": 1.4856, |
|
"step": 6310 |
|
}, |
|
{ |
|
"epoch": 0.44794103054787726, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 8.585659023794818e-05, |
|
"loss": 1.4933, |
|
"step": 6320 |
|
}, |
|
{ |
|
"epoch": 0.4486497980012758, |
|
"grad_norm": 0.490234375, |
|
"learning_rate": 8.484687843276469e-05, |
|
"loss": 1.4798, |
|
"step": 6330 |
|
}, |
|
{ |
|
"epoch": 0.44935856545467434, |
|
"grad_norm": 0.439453125, |
|
"learning_rate": 8.380874501370097e-05, |
|
"loss": 1.4972, |
|
"step": 6340 |
|
}, |
|
{ |
|
"epoch": 0.4500673329080729, |
|
"grad_norm": 0.45703125, |
|
"learning_rate": 8.274303669726426e-05, |
|
"loss": 1.482, |
|
"step": 6350 |
|
}, |
|
{ |
|
"epoch": 0.4507761003614714, |
|
"grad_norm": 0.44140625, |
|
"learning_rate": 8.165062269044353e-05, |
|
"loss": 1.4946, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 0.45148486781486996, |
|
"grad_norm": 0.443359375, |
|
"learning_rate": 8.053239398177191e-05, |
|
"loss": 1.487, |
|
"step": 6370 |
|
}, |
|
{ |
|
"epoch": 0.4521936352682685, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 7.938926261462366e-05, |
|
"loss": 1.5043, |
|
"step": 6380 |
|
}, |
|
{ |
|
"epoch": 0.45290240272166704, |
|
"grad_norm": 0.4921875, |
|
"learning_rate": 7.822216094333847e-05, |
|
"loss": 1.4912, |
|
"step": 6390 |
|
}, |
|
{ |
|
"epoch": 0.4536111701750656, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 7.703204087277988e-05, |
|
"loss": 1.4895, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.4543199376284641, |
|
"grad_norm": 0.470703125, |
|
"learning_rate": 7.58198730819481e-05, |
|
"loss": 1.5003, |
|
"step": 6410 |
|
}, |
|
{ |
|
"epoch": 0.45502870508186266, |
|
"grad_norm": 0.5078125, |
|
"learning_rate": 7.45866462322802e-05, |
|
"loss": 1.4872, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 0.4557374725352612, |
|
"grad_norm": 0.435546875, |
|
"learning_rate": 7.333336616128369e-05, |
|
"loss": 1.4993, |
|
"step": 6430 |
|
}, |
|
{ |
|
"epoch": 0.45644623998865974, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 7.206105506216106e-05, |
|
"loss": 1.5037, |
|
"step": 6440 |
|
}, |
|
{ |
|
"epoch": 0.4571550074420583, |
|
"grad_norm": 0.43359375, |
|
"learning_rate": 7.077075065009433e-05, |
|
"loss": 1.5001, |
|
"step": 6450 |
|
}, |
|
{ |
|
"epoch": 0.4578637748954568, |
|
"grad_norm": 0.478515625, |
|
"learning_rate": 6.946350531586959e-05, |
|
"loss": 1.5007, |
|
"step": 6460 |
|
}, |
|
{ |
|
"epoch": 0.45857254234885536, |
|
"grad_norm": 0.40625, |
|
"learning_rate": 6.814038526753205e-05, |
|
"loss": 1.4763, |
|
"step": 6470 |
|
}, |
|
{ |
|
"epoch": 0.4592813098022539, |
|
"grad_norm": 0.4453125, |
|
"learning_rate": 6.680246966077151e-05, |
|
"loss": 1.4987, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 0.45999007725565244, |
|
"grad_norm": 0.3984375, |
|
"learning_rate": 6.545084971874738e-05, |
|
"loss": 1.4925, |
|
"step": 6490 |
|
}, |
|
{ |
|
"epoch": 0.460698844709051, |
|
"grad_norm": 0.40234375, |
|
"learning_rate": 6.408662784207149e-05, |
|
"loss": 1.4923, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.460698844709051, |
|
"eval_accuracy": 0.666518879775364, |
|
"eval_loss": 1.39996337890625, |
|
"eval_runtime": 7.2215, |
|
"eval_samples_per_second": 48.051, |
|
"eval_steps_per_second": 0.415, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.4614076121624495, |
|
"grad_norm": 0.42578125, |
|
"learning_rate": 6.271091670967436e-05, |
|
"loss": 1.4809, |
|
"step": 6510 |
|
}, |
|
{ |
|
"epoch": 0.46211637961584806, |
|
"grad_norm": 0.41015625, |
|
"learning_rate": 6.132483837128823e-05, |
|
"loss": 1.4809, |
|
"step": 6520 |
|
}, |
|
{ |
|
"epoch": 0.4628251470692466, |
|
"grad_norm": 0.41796875, |
|
"learning_rate": 5.992952333228728e-05, |
|
"loss": 1.5065, |
|
"step": 6530 |
|
}, |
|
{ |
|
"epoch": 0.46353391452264514, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 5.85261096316312e-05, |
|
"loss": 1.4931, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 0.4642426819760437, |
|
"grad_norm": 0.431640625, |
|
"learning_rate": 5.7115741913664264e-05, |
|
"loss": 1.4766, |
|
"step": 6550 |
|
}, |
|
{ |
|
"epoch": 0.4649514494294422, |
|
"grad_norm": 0.412109375, |
|
"learning_rate": 5.569957049452703e-05, |
|
"loss": 1.49, |
|
"step": 6560 |
|
}, |
|
{ |
|
"epoch": 0.46566021688284076, |
|
"grad_norm": 0.423828125, |
|
"learning_rate": 5.427875042394199e-05, |
|
"loss": 1.4955, |
|
"step": 6570 |
|
}, |
|
{ |
|
"epoch": 0.4663689843362393, |
|
"grad_norm": 0.4375, |
|
"learning_rate": 5.2854440543138406e-05, |
|
"loss": 1.4972, |
|
"step": 6580 |
|
}, |
|
{ |
|
"epoch": 0.46707775178963784, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 5.142780253968481e-05, |
|
"loss": 1.4879, |
|
"step": 6590 |
|
}, |
|
{ |
|
"epoch": 0.4677865192430364, |
|
"grad_norm": 0.380859375, |
|
"learning_rate": 5e-05, |
|
"loss": 1.493, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.4684952866964349, |
|
"grad_norm": 0.3671875, |
|
"learning_rate": 4.85721974603152e-05, |
|
"loss": 1.4995, |
|
"step": 6610 |
|
}, |
|
{ |
|
"epoch": 0.46920405414983346, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 4.71455594568616e-05, |
|
"loss": 1.4771, |
|
"step": 6620 |
|
}, |
|
{ |
|
"epoch": 0.469912821603232, |
|
"grad_norm": 0.419921875, |
|
"learning_rate": 4.5721249576058027e-05, |
|
"loss": 1.4815, |
|
"step": 6630 |
|
}, |
|
{ |
|
"epoch": 0.47062158905663054, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 4.4300429505472976e-05, |
|
"loss": 1.4733, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.4713303565100291, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 4.288425808633575e-05, |
|
"loss": 1.4976, |
|
"step": 6650 |
|
}, |
|
{ |
|
"epoch": 0.4720391239634276, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 4.147389036836881e-05, |
|
"loss": 1.4884, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 0.47274789141682616, |
|
"grad_norm": 0.375, |
|
"learning_rate": 4.007047666771274e-05, |
|
"loss": 1.4969, |
|
"step": 6670 |
|
}, |
|
{ |
|
"epoch": 0.4734566588702247, |
|
"grad_norm": 0.375, |
|
"learning_rate": 3.8675161628711776e-05, |
|
"loss": 1.5137, |
|
"step": 6680 |
|
}, |
|
{ |
|
"epoch": 0.47416542632362324, |
|
"grad_norm": 0.365234375, |
|
"learning_rate": 3.728908329032567e-05, |
|
"loss": 1.4889, |
|
"step": 6690 |
|
}, |
|
{ |
|
"epoch": 0.4748741937770218, |
|
"grad_norm": 0.408203125, |
|
"learning_rate": 3.591337215792852e-05, |
|
"loss": 1.4875, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.4755829612304203, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 3.4549150281252636e-05, |
|
"loss": 1.5153, |
|
"step": 6710 |
|
}, |
|
{ |
|
"epoch": 0.47629172868381886, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 3.3197530339228487e-05, |
|
"loss": 1.4904, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 0.4770004961372174, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 3.1859614732467954e-05, |
|
"loss": 1.4851, |
|
"step": 6730 |
|
}, |
|
{ |
|
"epoch": 0.47770926359061594, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 3.053649468413043e-05, |
|
"loss": 1.4979, |
|
"step": 6740 |
|
}, |
|
{ |
|
"epoch": 0.4784180310440145, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 2.9229249349905684e-05, |
|
"loss": 1.492, |
|
"step": 6750 |
|
}, |
|
{ |
|
"epoch": 0.479126798497413, |
|
"grad_norm": 0.345703125, |
|
"learning_rate": 2.7938944937838923e-05, |
|
"loss": 1.4843, |
|
"step": 6760 |
|
}, |
|
{ |
|
"epoch": 0.47983556595081156, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 2.6666633838716314e-05, |
|
"loss": 1.4781, |
|
"step": 6770 |
|
}, |
|
{ |
|
"epoch": 0.4805443334042101, |
|
"grad_norm": 0.349609375, |
|
"learning_rate": 2.5413353767719805e-05, |
|
"loss": 1.4949, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 0.48125310085760864, |
|
"grad_norm": 0.35546875, |
|
"learning_rate": 2.418012691805191e-05, |
|
"loss": 1.4961, |
|
"step": 6790 |
|
}, |
|
{ |
|
"epoch": 0.4819618683110072, |
|
"grad_norm": 0.38671875, |
|
"learning_rate": 2.296795912722014e-05, |
|
"loss": 1.4942, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.4826706357644057, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 2.1777839056661554e-05, |
|
"loss": 1.49, |
|
"step": 6810 |
|
}, |
|
{ |
|
"epoch": 0.48337940321780426, |
|
"grad_norm": 0.361328125, |
|
"learning_rate": 2.061073738537635e-05, |
|
"loss": 1.5085, |
|
"step": 6820 |
|
}, |
|
{ |
|
"epoch": 0.4840881706712028, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 1.946760601822809e-05, |
|
"loss": 1.4926, |
|
"step": 6830 |
|
}, |
|
{ |
|
"epoch": 0.48479693812460134, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 1.8349377309556486e-05, |
|
"loss": 1.4815, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 0.4855057055779999, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 1.725696330273575e-05, |
|
"loss": 1.5001, |
|
"step": 6850 |
|
}, |
|
{ |
|
"epoch": 0.4862144730313984, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 1.619125498629904e-05, |
|
"loss": 1.481, |
|
"step": 6860 |
|
}, |
|
{ |
|
"epoch": 0.48692324048479696, |
|
"grad_norm": 0.353515625, |
|
"learning_rate": 1.5153121567235335e-05, |
|
"loss": 1.4781, |
|
"step": 6870 |
|
}, |
|
{ |
|
"epoch": 0.4876320079381955, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 1.414340976205183e-05, |
|
"loss": 1.4906, |
|
"step": 6880 |
|
}, |
|
{ |
|
"epoch": 0.48834077539159404, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 1.3162943106179749e-05, |
|
"loss": 1.5031, |
|
"step": 6890 |
|
}, |
|
{ |
|
"epoch": 0.4890495428449926, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 1.2212521282287092e-05, |
|
"loss": 1.4878, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.4897583102983911, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 1.1292919468045877e-05, |
|
"loss": 1.4929, |
|
"step": 6910 |
|
}, |
|
{ |
|
"epoch": 0.49046707775178966, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 1.0404887703886251e-05, |
|
"loss": 1.4889, |
|
"step": 6920 |
|
}, |
|
{ |
|
"epoch": 0.4911758452051882, |
|
"grad_norm": 0.359375, |
|
"learning_rate": 9.549150281252633e-06, |
|
"loss": 1.4783, |
|
"step": 6930 |
|
}, |
|
{ |
|
"epoch": 0.49188461265858674, |
|
"grad_norm": 0.322265625, |
|
"learning_rate": 8.7264051518613e-06, |
|
"loss": 1.4928, |
|
"step": 6940 |
|
}, |
|
{ |
|
"epoch": 0.4925933801119853, |
|
"grad_norm": 0.3359375, |
|
"learning_rate": 7.937323358440935e-06, |
|
"loss": 1.489, |
|
"step": 6950 |
|
}, |
|
{ |
|
"epoch": 0.4933021475653838, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 7.182548487420554e-06, |
|
"loss": 1.4899, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 0.49401091501878236, |
|
"grad_norm": 0.33984375, |
|
"learning_rate": 6.462696144011149e-06, |
|
"loss": 1.4859, |
|
"step": 6970 |
|
}, |
|
{ |
|
"epoch": 0.4947196824721809, |
|
"grad_norm": 0.34375, |
|
"learning_rate": 5.778353450109286e-06, |
|
"loss": 1.4944, |
|
"step": 6980 |
|
}, |
|
{ |
|
"epoch": 0.49542844992557944, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 5.13007856543209e-06, |
|
"loss": 1.481, |
|
"step": 6990 |
|
}, |
|
{ |
|
"epoch": 0.496137217378978, |
|
"grad_norm": 0.333984375, |
|
"learning_rate": 4.5184002322740785e-06, |
|
"loss": 1.4965, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.496137217378978, |
|
"eval_accuracy": 0.6668890507507222, |
|
"eval_loss": 1.397897720336914, |
|
"eval_runtime": 7.477, |
|
"eval_samples_per_second": 46.409, |
|
"eval_steps_per_second": 0.401, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.4968459848323765, |
|
"grad_norm": 0.373046875, |
|
"learning_rate": 3.9438173442575e-06, |
|
"loss": 1.492, |
|
"step": 7010 |
|
}, |
|
{ |
|
"epoch": 0.49755475228577506, |
|
"grad_norm": 0.310546875, |
|
"learning_rate": 3.406798539427386e-06, |
|
"loss": 1.4714, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 0.4982635197391736, |
|
"grad_norm": 0.31640625, |
|
"learning_rate": 2.9077818180237693e-06, |
|
"loss": 1.4839, |
|
"step": 7030 |
|
}, |
|
{ |
|
"epoch": 0.49897228719257214, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 2.4471741852423237e-06, |
|
"loss": 1.4976, |
|
"step": 7040 |
|
}, |
|
{ |
|
"epoch": 0.4996810546459707, |
|
"grad_norm": 0.328125, |
|
"learning_rate": 2.0253513192751373e-06, |
|
"loss": 1.4857, |
|
"step": 7050 |
|
}, |
|
{ |
|
"epoch": 0.50003543837267, |
|
"step": 7055, |
|
"total_flos": 1.568873006179696e+20, |
|
"train_loss": 1.4944760516549913, |
|
"train_runtime": 7219.1988, |
|
"train_samples_per_second": 125.076, |
|
"train_steps_per_second": 0.977 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 7055, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 1000, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.568873006179696e+20, |
|
"train_batch_size": 1, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|