bge-m3-finetuned-200k / trainer_state.json
meandyou200175's picture
Upload fine-tuned BGE-M3 model
6d15ced verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 500,
"global_step": 25872,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01932591218305504,
"grad_norm": 135.7185821533203,
"learning_rate": 1.9088098918083462e-06,
"loss": 2.7419,
"step": 500
},
{
"epoch": 0.03865182436611008,
"grad_norm": 71.08179473876953,
"learning_rate": 3.840803709428131e-06,
"loss": 1.6608,
"step": 1000
},
{
"epoch": 0.05797773654916512,
"grad_norm": 102.95220184326172,
"learning_rate": 5.772797527047914e-06,
"loss": 1.5107,
"step": 1500
},
{
"epoch": 0.07730364873222016,
"grad_norm": 72.49132537841797,
"learning_rate": 7.704791344667698e-06,
"loss": 1.3103,
"step": 2000
},
{
"epoch": 0.0966295609152752,
"grad_norm": 78.74112701416016,
"learning_rate": 9.636785162287482e-06,
"loss": 1.3606,
"step": 2500
},
{
"epoch": 0.11595547309833024,
"grad_norm": 132.5709991455078,
"learning_rate": 9.825631334822196e-06,
"loss": 1.2151,
"step": 3000
},
{
"epoch": 0.13528138528138528,
"grad_norm": 89.91647338867188,
"learning_rate": 9.61089159938155e-06,
"loss": 1.1923,
"step": 3500
},
{
"epoch": 0.15460729746444032,
"grad_norm": 31.426151275634766,
"learning_rate": 9.396581343411786e-06,
"loss": 1.1765,
"step": 4000
},
{
"epoch": 0.17393320964749537,
"grad_norm": 93.264404296875,
"learning_rate": 9.18184160797114e-06,
"loss": 1.1631,
"step": 4500
},
{
"epoch": 0.1932591218305504,
"grad_norm": 94.13616943359375,
"learning_rate": 8.967101872530494e-06,
"loss": 1.1176,
"step": 5000
},
{
"epoch": 0.21258503401360543,
"grad_norm": 67.35560607910156,
"learning_rate": 8.752362137089849e-06,
"loss": 1.1277,
"step": 5500
},
{
"epoch": 0.23191094619666047,
"grad_norm": 34.71173858642578,
"learning_rate": 8.537622401649201e-06,
"loss": 1.0943,
"step": 6000
},
{
"epoch": 0.25123685837971554,
"grad_norm": 134.21011352539062,
"learning_rate": 8.322882666208556e-06,
"loss": 1.052,
"step": 6500
},
{
"epoch": 0.27056277056277056,
"grad_norm": 100.96661376953125,
"learning_rate": 8.10857241023879e-06,
"loss": 1.0651,
"step": 7000
},
{
"epoch": 0.2898886827458256,
"grad_norm": 95.49282836914062,
"learning_rate": 7.893832674798145e-06,
"loss": 1.0536,
"step": 7500
},
{
"epoch": 0.30921459492888065,
"grad_norm": 94.79146575927734,
"learning_rate": 7.6790929393575e-06,
"loss": 1.0546,
"step": 8000
},
{
"epoch": 0.32854050711193566,
"grad_norm": 67.1232681274414,
"learning_rate": 7.464353203916853e-06,
"loss": 1.0366,
"step": 8500
},
{
"epoch": 0.34786641929499074,
"grad_norm": 48.50230026245117,
"learning_rate": 7.250042947947089e-06,
"loss": 1.0106,
"step": 9000
},
{
"epoch": 0.36719233147804575,
"grad_norm": 28.64300537109375,
"learning_rate": 7.035303212506442e-06,
"loss": 1.0843,
"step": 9500
},
{
"epoch": 0.3865182436611008,
"grad_norm": 48.05903625488281,
"learning_rate": 6.820563477065797e-06,
"loss": 1.0454,
"step": 10000
},
{
"epoch": 0.40584415584415584,
"grad_norm": 53.515464782714844,
"learning_rate": 6.605823741625152e-06,
"loss": 0.9767,
"step": 10500
},
{
"epoch": 0.42517006802721086,
"grad_norm": 2.856135606765747,
"learning_rate": 6.391084006184504e-06,
"loss": 0.9433,
"step": 11000
},
{
"epoch": 0.44449598021026593,
"grad_norm": 125.483154296875,
"learning_rate": 6.17677375021474e-06,
"loss": 0.9314,
"step": 11500
},
{
"epoch": 0.46382189239332094,
"grad_norm": 45.01868438720703,
"learning_rate": 5.962034014774094e-06,
"loss": 0.9568,
"step": 12000
},
{
"epoch": 0.483147804576376,
"grad_norm": 15.953567504882812,
"learning_rate": 5.747294279333448e-06,
"loss": 0.8972,
"step": 12500
},
{
"epoch": 0.5024737167594311,
"grad_norm": 85.09109497070312,
"learning_rate": 5.532554543892802e-06,
"loss": 0.9156,
"step": 13000
},
{
"epoch": 0.5217996289424861,
"grad_norm": 53.42656707763672,
"learning_rate": 5.317814808452156e-06,
"loss": 0.9581,
"step": 13500
},
{
"epoch": 0.5411255411255411,
"grad_norm": 72.82543182373047,
"learning_rate": 5.103504552482392e-06,
"loss": 0.9065,
"step": 14000
},
{
"epoch": 0.5604514533085961,
"grad_norm": 88.66947174072266,
"learning_rate": 4.8887648170417454e-06,
"loss": 0.9134,
"step": 14500
},
{
"epoch": 0.5797773654916512,
"grad_norm": 86.12175750732422,
"learning_rate": 4.6740250816011e-06,
"loss": 0.8819,
"step": 15000
},
{
"epoch": 0.5991032776747063,
"grad_norm": 61.66287612915039,
"learning_rate": 4.459285346160454e-06,
"loss": 0.8798,
"step": 15500
},
{
"epoch": 0.6184291898577613,
"grad_norm": 51.16645431518555,
"learning_rate": 4.244975090190689e-06,
"loss": 0.8437,
"step": 16000
},
{
"epoch": 0.6377551020408163,
"grad_norm": 135.60870361328125,
"learning_rate": 4.0302353547500435e-06,
"loss": 0.8831,
"step": 16500
},
{
"epoch": 0.6570810142238713,
"grad_norm": 89.30845642089844,
"learning_rate": 3.815495619309397e-06,
"loss": 0.8872,
"step": 17000
},
{
"epoch": 0.6764069264069265,
"grad_norm": 54.0753288269043,
"learning_rate": 3.6007558838687514e-06,
"loss": 0.8717,
"step": 17500
},
{
"epoch": 0.6957328385899815,
"grad_norm": 28.510011672973633,
"learning_rate": 3.3860161484281057e-06,
"loss": 0.8599,
"step": 18000
},
{
"epoch": 0.7150587507730365,
"grad_norm": 42.08302307128906,
"learning_rate": 3.1712764129874596e-06,
"loss": 0.7934,
"step": 18500
},
{
"epoch": 0.7343846629560915,
"grad_norm": 54.24616622924805,
"learning_rate": 2.956966157017695e-06,
"loss": 0.7987,
"step": 19000
},
{
"epoch": 0.7537105751391465,
"grad_norm": 66.43406677246094,
"learning_rate": 2.742226421577049e-06,
"loss": 0.8707,
"step": 19500
},
{
"epoch": 0.7730364873222016,
"grad_norm": 47.240234375,
"learning_rate": 2.527486686136403e-06,
"loss": 0.8444,
"step": 20000
},
{
"epoch": 0.7923623995052567,
"grad_norm": 50.62797164916992,
"learning_rate": 2.312746950695757e-06,
"loss": 0.8591,
"step": 20500
},
{
"epoch": 0.8116883116883117,
"grad_norm": 41.326271057128906,
"learning_rate": 2.0980072152551107e-06,
"loss": 0.7583,
"step": 21000
},
{
"epoch": 0.8310142238713667,
"grad_norm": 101.81986236572266,
"learning_rate": 1.883267479814465e-06,
"loss": 0.9015,
"step": 21500
},
{
"epoch": 0.8503401360544217,
"grad_norm": 109.75096893310547,
"learning_rate": 1.6689572238447004e-06,
"loss": 0.798,
"step": 22000
},
{
"epoch": 0.8696660482374768,
"grad_norm": 74.9537582397461,
"learning_rate": 1.4542174884040545e-06,
"loss": 0.8254,
"step": 22500
},
{
"epoch": 0.8889919604205319,
"grad_norm": 53.3721923828125,
"learning_rate": 1.2394777529634084e-06,
"loss": 0.8392,
"step": 23000
},
{
"epoch": 0.9083178726035869,
"grad_norm": 74.43936157226562,
"learning_rate": 1.0247380175227625e-06,
"loss": 0.7987,
"step": 23500
},
{
"epoch": 0.9276437847866419,
"grad_norm": 22.659399032592773,
"learning_rate": 8.099982820821166e-07,
"loss": 0.8334,
"step": 24000
},
{
"epoch": 0.946969696969697,
"grad_norm": 57.379234313964844,
"learning_rate": 5.952585466414706e-07,
"loss": 0.7609,
"step": 24500
},
{
"epoch": 0.966295609152752,
"grad_norm": 48.525848388671875,
"learning_rate": 3.8051881120082463e-07,
"loss": 0.8165,
"step": 25000
},
{
"epoch": 0.985621521335807,
"grad_norm": 29.62494468688965,
"learning_rate": 1.6577907576017866e-07,
"loss": 0.8133,
"step": 25500
}
],
"logging_steps": 500,
"max_steps": 25872,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}