DeepSeek-Llama-8-GGUF / trainer_state.json
bmas10's picture
Upload 8 files
52f5fba verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9936305732484074,
"eval_steps": 500,
"global_step": 1059,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.028308563340410473,
"grad_norm": 10.472460746765137,
"learning_rate": 4.971671388101983e-05,
"loss": 5.6874,
"step": 10
},
{
"epoch": 0.056617126680820945,
"grad_norm": 1.0690819025039673,
"learning_rate": 4.9244570349386214e-05,
"loss": 1.622,
"step": 20
},
{
"epoch": 0.08492569002123142,
"grad_norm": 0.6105159521102905,
"learning_rate": 4.8772426817752596e-05,
"loss": 2.0585,
"step": 30
},
{
"epoch": 0.11323425336164189,
"grad_norm": 0.9013139009475708,
"learning_rate": 4.830028328611898e-05,
"loss": 1.3675,
"step": 40
},
{
"epoch": 0.14154281670205238,
"grad_norm": 1.0703097581863403,
"learning_rate": 4.7828139754485365e-05,
"loss": 1.7004,
"step": 50
},
{
"epoch": 0.16985138004246284,
"grad_norm": 0.715630829334259,
"learning_rate": 4.7355996222851754e-05,
"loss": 1.2608,
"step": 60
},
{
"epoch": 0.19815994338287332,
"grad_norm": 0.6118766069412231,
"learning_rate": 4.6883852691218135e-05,
"loss": 1.3334,
"step": 70
},
{
"epoch": 0.22646850672328378,
"grad_norm": 1.0044665336608887,
"learning_rate": 4.641170915958452e-05,
"loss": 1.3244,
"step": 80
},
{
"epoch": 0.25477707006369427,
"grad_norm": 1.927817940711975,
"learning_rate": 4.59395656279509e-05,
"loss": 1.2348,
"step": 90
},
{
"epoch": 0.28308563340410475,
"grad_norm": 1.5741430521011353,
"learning_rate": 4.546742209631728e-05,
"loss": 1.22,
"step": 100
},
{
"epoch": 0.31139419674451524,
"grad_norm": 0.6977642774581909,
"learning_rate": 4.499527856468367e-05,
"loss": 1.0786,
"step": 110
},
{
"epoch": 0.33970276008492567,
"grad_norm": 2.7766706943511963,
"learning_rate": 4.452313503305005e-05,
"loss": 1.1531,
"step": 120
},
{
"epoch": 0.36801132342533616,
"grad_norm": 2.0790321826934814,
"learning_rate": 4.405099150141643e-05,
"loss": 1.0915,
"step": 130
},
{
"epoch": 0.39631988676574664,
"grad_norm": 3.8764405250549316,
"learning_rate": 4.357884796978281e-05,
"loss": 1.1757,
"step": 140
},
{
"epoch": 0.42462845010615713,
"grad_norm": 0.8142156600952148,
"learning_rate": 4.31067044381492e-05,
"loss": 1.0972,
"step": 150
},
{
"epoch": 0.45293701344656756,
"grad_norm": 2.6679811477661133,
"learning_rate": 4.263456090651558e-05,
"loss": 1.1941,
"step": 160
},
{
"epoch": 0.48124557678697805,
"grad_norm": 3.724242925643921,
"learning_rate": 4.216241737488197e-05,
"loss": 1.0199,
"step": 170
},
{
"epoch": 0.5095541401273885,
"grad_norm": 0.9096172451972961,
"learning_rate": 4.169027384324835e-05,
"loss": 0.9365,
"step": 180
},
{
"epoch": 0.537862703467799,
"grad_norm": 0.8280671834945679,
"learning_rate": 4.1218130311614734e-05,
"loss": 0.7999,
"step": 190
},
{
"epoch": 0.5661712668082095,
"grad_norm": 0.7605137825012207,
"learning_rate": 4.0745986779981115e-05,
"loss": 0.9142,
"step": 200
},
{
"epoch": 0.5944798301486199,
"grad_norm": 1.1094071865081787,
"learning_rate": 4.02738432483475e-05,
"loss": 0.9304,
"step": 210
},
{
"epoch": 0.6227883934890305,
"grad_norm": 1.2188092470169067,
"learning_rate": 3.980169971671388e-05,
"loss": 1.0773,
"step": 220
},
{
"epoch": 0.6510969568294409,
"grad_norm": 1.421103835105896,
"learning_rate": 3.932955618508027e-05,
"loss": 0.9702,
"step": 230
},
{
"epoch": 0.6794055201698513,
"grad_norm": 0.6401333212852478,
"learning_rate": 3.885741265344665e-05,
"loss": 0.8283,
"step": 240
},
{
"epoch": 0.7077140835102619,
"grad_norm": 1.2169103622436523,
"learning_rate": 3.8385269121813036e-05,
"loss": 0.7306,
"step": 250
},
{
"epoch": 0.7360226468506723,
"grad_norm": 4.206311225891113,
"learning_rate": 3.791312559017942e-05,
"loss": 0.7564,
"step": 260
},
{
"epoch": 0.7643312101910829,
"grad_norm": 0.7803528308868408,
"learning_rate": 3.74409820585458e-05,
"loss": 0.9409,
"step": 270
},
{
"epoch": 0.7926397735314933,
"grad_norm": 1.2564524412155151,
"learning_rate": 3.696883852691218e-05,
"loss": 0.8487,
"step": 280
},
{
"epoch": 0.8209483368719037,
"grad_norm": 1.0345911979675293,
"learning_rate": 3.649669499527857e-05,
"loss": 0.9284,
"step": 290
},
{
"epoch": 0.8492569002123143,
"grad_norm": 2.6895930767059326,
"learning_rate": 3.602455146364495e-05,
"loss": 1.1654,
"step": 300
},
{
"epoch": 0.8775654635527247,
"grad_norm": 1.2171903848648071,
"learning_rate": 3.555240793201133e-05,
"loss": 0.8589,
"step": 310
},
{
"epoch": 0.9058740268931351,
"grad_norm": 2.6197869777679443,
"learning_rate": 3.5080264400377714e-05,
"loss": 0.8405,
"step": 320
},
{
"epoch": 0.9341825902335457,
"grad_norm": 0.9666792750358582,
"learning_rate": 3.4608120868744095e-05,
"loss": 1.1513,
"step": 330
},
{
"epoch": 0.9624911535739561,
"grad_norm": 0.9216092824935913,
"learning_rate": 3.4135977337110484e-05,
"loss": 0.8712,
"step": 340
},
{
"epoch": 0.9907997169143666,
"grad_norm": 2.9435744285583496,
"learning_rate": 3.366383380547687e-05,
"loss": 0.9633,
"step": 350
},
{
"epoch": 1.0,
"eval_loss": 0.920014500617981,
"eval_runtime": 45.4818,
"eval_samples_per_second": 15.545,
"eval_steps_per_second": 1.957,
"step": 354
},
{
"epoch": 1.0169851380042463,
"grad_norm": 0.7391465306282043,
"learning_rate": 3.3191690273843253e-05,
"loss": 0.6779,
"step": 360
},
{
"epoch": 1.0452937013446568,
"grad_norm": 1.267843246459961,
"learning_rate": 3.2719546742209635e-05,
"loss": 0.9072,
"step": 370
},
{
"epoch": 1.0736022646850671,
"grad_norm": 0.7193703651428223,
"learning_rate": 3.2247403210576016e-05,
"loss": 1.0308,
"step": 380
},
{
"epoch": 1.1019108280254777,
"grad_norm": 3.1680991649627686,
"learning_rate": 3.17752596789424e-05,
"loss": 0.8272,
"step": 390
},
{
"epoch": 1.1302193913658882,
"grad_norm": 1.0723551511764526,
"learning_rate": 3.130311614730878e-05,
"loss": 1.02,
"step": 400
},
{
"epoch": 1.1585279547062988,
"grad_norm": 0.7101927995681763,
"learning_rate": 3.083097261567517e-05,
"loss": 0.942,
"step": 410
},
{
"epoch": 1.186836518046709,
"grad_norm": 1.9213780164718628,
"learning_rate": 3.0358829084041553e-05,
"loss": 0.8884,
"step": 420
},
{
"epoch": 1.2151450813871196,
"grad_norm": 4.313545227050781,
"learning_rate": 2.9886685552407934e-05,
"loss": 1.055,
"step": 430
},
{
"epoch": 1.2434536447275302,
"grad_norm": 1.1699409484863281,
"learning_rate": 2.9414542020774316e-05,
"loss": 0.7015,
"step": 440
},
{
"epoch": 1.2717622080679405,
"grad_norm": 0.7084372043609619,
"learning_rate": 2.8942398489140697e-05,
"loss": 0.8342,
"step": 450
},
{
"epoch": 1.300070771408351,
"grad_norm": 1.1760873794555664,
"learning_rate": 2.8470254957507082e-05,
"loss": 0.7567,
"step": 460
},
{
"epoch": 1.3283793347487616,
"grad_norm": 6.9120192527771,
"learning_rate": 2.799811142587347e-05,
"loss": 0.8156,
"step": 470
},
{
"epoch": 1.356687898089172,
"grad_norm": 0.7138371467590332,
"learning_rate": 2.7525967894239852e-05,
"loss": 0.8024,
"step": 480
},
{
"epoch": 1.3849964614295824,
"grad_norm": 0.6835730075836182,
"learning_rate": 2.7053824362606233e-05,
"loss": 0.7977,
"step": 490
},
{
"epoch": 1.413305024769993,
"grad_norm": 6.436087608337402,
"learning_rate": 2.6581680830972615e-05,
"loss": 0.8102,
"step": 500
},
{
"epoch": 1.4416135881104033,
"grad_norm": 3.4291305541992188,
"learning_rate": 2.6109537299339e-05,
"loss": 0.7643,
"step": 510
},
{
"epoch": 1.4699221514508138,
"grad_norm": 0.9250390529632568,
"learning_rate": 2.563739376770538e-05,
"loss": 0.8417,
"step": 520
},
{
"epoch": 1.4982307147912244,
"grad_norm": 3.264242649078369,
"learning_rate": 2.516525023607177e-05,
"loss": 0.9988,
"step": 530
},
{
"epoch": 1.5265392781316347,
"grad_norm": 1.8651660680770874,
"learning_rate": 2.469310670443815e-05,
"loss": 0.8652,
"step": 540
},
{
"epoch": 1.5548478414720452,
"grad_norm": 5.576003074645996,
"learning_rate": 2.4220963172804533e-05,
"loss": 0.7481,
"step": 550
},
{
"epoch": 1.5831564048124558,
"grad_norm": 4.8853302001953125,
"learning_rate": 2.3748819641170918e-05,
"loss": 0.7857,
"step": 560
},
{
"epoch": 1.611464968152866,
"grad_norm": 2.0183866024017334,
"learning_rate": 2.3276676109537303e-05,
"loss": 0.607,
"step": 570
},
{
"epoch": 1.6397735314932769,
"grad_norm": 3.133368730545044,
"learning_rate": 2.2804532577903684e-05,
"loss": 0.8706,
"step": 580
},
{
"epoch": 1.6680820948336872,
"grad_norm": 2.440894365310669,
"learning_rate": 2.2332389046270066e-05,
"loss": 0.8035,
"step": 590
},
{
"epoch": 1.6963906581740975,
"grad_norm": 3.0257763862609863,
"learning_rate": 2.186024551463645e-05,
"loss": 1.1186,
"step": 600
},
{
"epoch": 1.7246992215145083,
"grad_norm": 2.741161823272705,
"learning_rate": 2.1388101983002835e-05,
"loss": 1.0302,
"step": 610
},
{
"epoch": 1.7530077848549186,
"grad_norm": 3.0165324211120605,
"learning_rate": 2.0915958451369217e-05,
"loss": 1.0235,
"step": 620
},
{
"epoch": 1.7813163481953291,
"grad_norm": 1.8402531147003174,
"learning_rate": 2.0443814919735602e-05,
"loss": 0.8455,
"step": 630
},
{
"epoch": 1.8096249115357397,
"grad_norm": 1.0427038669586182,
"learning_rate": 1.9971671388101983e-05,
"loss": 0.7868,
"step": 640
},
{
"epoch": 1.83793347487615,
"grad_norm": 4.955197334289551,
"learning_rate": 1.9499527856468365e-05,
"loss": 0.8728,
"step": 650
},
{
"epoch": 1.8662420382165605,
"grad_norm": 1.1331288814544678,
"learning_rate": 1.9027384324834753e-05,
"loss": 0.8719,
"step": 660
},
{
"epoch": 1.894550601556971,
"grad_norm": 1.395735263824463,
"learning_rate": 1.8555240793201135e-05,
"loss": 0.8224,
"step": 670
},
{
"epoch": 1.9228591648973814,
"grad_norm": 1.110823154449463,
"learning_rate": 1.8083097261567516e-05,
"loss": 0.6859,
"step": 680
},
{
"epoch": 1.951167728237792,
"grad_norm": 1.224063754081726,
"learning_rate": 1.76109537299339e-05,
"loss": 0.79,
"step": 690
},
{
"epoch": 1.9794762915782025,
"grad_norm": 1.4636934995651245,
"learning_rate": 1.7138810198300283e-05,
"loss": 0.9543,
"step": 700
},
{
"epoch": 2.0,
"eval_loss": 0.8783606290817261,
"eval_runtime": 45.4943,
"eval_samples_per_second": 15.54,
"eval_steps_per_second": 1.956,
"step": 708
},
{
"epoch": 2.0056617126680822,
"grad_norm": 3.3927221298217773,
"learning_rate": 1.6666666666666667e-05,
"loss": 1.0075,
"step": 710
},
{
"epoch": 2.0339702760084926,
"grad_norm": 2.4237897396087646,
"learning_rate": 1.6194523135033052e-05,
"loss": 0.7245,
"step": 720
},
{
"epoch": 2.062278839348903,
"grad_norm": 0.9672299027442932,
"learning_rate": 1.5722379603399434e-05,
"loss": 0.7631,
"step": 730
},
{
"epoch": 2.0905874026893136,
"grad_norm": 2.908951997756958,
"learning_rate": 1.5250236071765817e-05,
"loss": 1.0795,
"step": 740
},
{
"epoch": 2.118895966029724,
"grad_norm": 3.8468856811523438,
"learning_rate": 1.4778092540132202e-05,
"loss": 0.6312,
"step": 750
},
{
"epoch": 2.1472045293701343,
"grad_norm": 1.352995753288269,
"learning_rate": 1.4305949008498584e-05,
"loss": 0.8062,
"step": 760
},
{
"epoch": 2.175513092710545,
"grad_norm": 1.1038804054260254,
"learning_rate": 1.3833805476864967e-05,
"loss": 0.8566,
"step": 770
},
{
"epoch": 2.2038216560509554,
"grad_norm": 1.2990654706954956,
"learning_rate": 1.3361661945231352e-05,
"loss": 0.7693,
"step": 780
},
{
"epoch": 2.2321302193913657,
"grad_norm": 2.6625759601593018,
"learning_rate": 1.2889518413597735e-05,
"loss": 0.6126,
"step": 790
},
{
"epoch": 2.2604387827317765,
"grad_norm": 5.676569938659668,
"learning_rate": 1.2417374881964118e-05,
"loss": 0.7988,
"step": 800
},
{
"epoch": 2.2887473460721868,
"grad_norm": 2.953200340270996,
"learning_rate": 1.1945231350330501e-05,
"loss": 0.6122,
"step": 810
},
{
"epoch": 2.3170559094125975,
"grad_norm": 2.253474473953247,
"learning_rate": 1.1473087818696884e-05,
"loss": 1.1378,
"step": 820
},
{
"epoch": 2.345364472753008,
"grad_norm": 1.5058188438415527,
"learning_rate": 1.1000944287063268e-05,
"loss": 0.8095,
"step": 830
},
{
"epoch": 2.373673036093418,
"grad_norm": 0.8082104325294495,
"learning_rate": 1.0528800755429651e-05,
"loss": 0.5857,
"step": 840
},
{
"epoch": 2.401981599433829,
"grad_norm": 2.0527071952819824,
"learning_rate": 1.0056657223796034e-05,
"loss": 0.9159,
"step": 850
},
{
"epoch": 2.4302901627742393,
"grad_norm": 4.757397651672363,
"learning_rate": 9.584513692162419e-06,
"loss": 0.7801,
"step": 860
},
{
"epoch": 2.4585987261146496,
"grad_norm": 2.750612735748291,
"learning_rate": 9.1123701605288e-06,
"loss": 0.7435,
"step": 870
},
{
"epoch": 2.4869072894550603,
"grad_norm": 0.7405831813812256,
"learning_rate": 8.640226628895184e-06,
"loss": 0.7491,
"step": 880
},
{
"epoch": 2.5152158527954707,
"grad_norm": 3.5916926860809326,
"learning_rate": 8.168083097261569e-06,
"loss": 0.896,
"step": 890
},
{
"epoch": 2.543524416135881,
"grad_norm": 2.436352252960205,
"learning_rate": 7.69593956562795e-06,
"loss": 0.7018,
"step": 900
},
{
"epoch": 2.5718329794762917,
"grad_norm": 0.8077714443206787,
"learning_rate": 7.223796033994334e-06,
"loss": 1.1473,
"step": 910
},
{
"epoch": 2.600141542816702,
"grad_norm": 2.01489520072937,
"learning_rate": 6.751652502360718e-06,
"loss": 0.8547,
"step": 920
},
{
"epoch": 2.6284501061571124,
"grad_norm": 0.9167873859405518,
"learning_rate": 6.279508970727101e-06,
"loss": 0.9682,
"step": 930
},
{
"epoch": 2.656758669497523,
"grad_norm": 0.8447218537330627,
"learning_rate": 5.807365439093485e-06,
"loss": 0.5817,
"step": 940
},
{
"epoch": 2.6850672328379335,
"grad_norm": 1.5092198848724365,
"learning_rate": 5.335221907459869e-06,
"loss": 1.2028,
"step": 950
},
{
"epoch": 2.713375796178344,
"grad_norm": 2.282589912414551,
"learning_rate": 4.863078375826251e-06,
"loss": 0.6571,
"step": 960
},
{
"epoch": 2.7416843595187546,
"grad_norm": 1.5408430099487305,
"learning_rate": 4.390934844192634e-06,
"loss": 0.6459,
"step": 970
},
{
"epoch": 2.769992922859165,
"grad_norm": 5.208029270172119,
"learning_rate": 3.918791312559018e-06,
"loss": 0.8532,
"step": 980
},
{
"epoch": 2.798301486199575,
"grad_norm": 1.4328051805496216,
"learning_rate": 3.4466477809254016e-06,
"loss": 0.8735,
"step": 990
},
{
"epoch": 2.826610049539986,
"grad_norm": 2.1999919414520264,
"learning_rate": 2.9745042492917848e-06,
"loss": 1.0141,
"step": 1000
},
{
"epoch": 2.8549186128803963,
"grad_norm": 1.0183438062667847,
"learning_rate": 2.502360717658168e-06,
"loss": 0.7974,
"step": 1010
},
{
"epoch": 2.8832271762208066,
"grad_norm": 3.3981127738952637,
"learning_rate": 2.0302171860245516e-06,
"loss": 0.8471,
"step": 1020
},
{
"epoch": 2.9115357395612174,
"grad_norm": 2.758704662322998,
"learning_rate": 1.5580736543909348e-06,
"loss": 0.8021,
"step": 1030
},
{
"epoch": 2.9398443029016277,
"grad_norm": 0.8233171105384827,
"learning_rate": 1.0859301227573182e-06,
"loss": 0.7704,
"step": 1040
},
{
"epoch": 2.968152866242038,
"grad_norm": 3.0792815685272217,
"learning_rate": 6.137865911237016e-07,
"loss": 0.6247,
"step": 1050
},
{
"epoch": 2.9936305732484074,
"eval_loss": 0.861165463924408,
"eval_runtime": 45.4972,
"eval_samples_per_second": 15.539,
"eval_steps_per_second": 1.956,
"step": 1059
}
],
"logging_steps": 10,
"max_steps": 1059,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 1.95177370947158e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}