|
{
|
|
"best_metric": 1.9407862424850464,
|
|
"best_model_checkpoint": "pvt-small-224-ConcreteClassifier-PVT\\checkpoint-9635",
|
|
"epoch": 11.0,
|
|
"eval_steps": 500,
|
|
"global_step": 21197,
|
|
"is_hyper_param_search": false,
|
|
"is_local_process_zero": true,
|
|
"is_world_process_zero": true,
|
|
"log_history": [
|
|
{
|
|
"epoch": 0.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.7631,
|
|
"step": 10
|
|
},
|
|
{
|
|
"epoch": 0.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0945,
|
|
"step": 20
|
|
},
|
|
{
|
|
"epoch": 0.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9938,
|
|
"step": 30
|
|
},
|
|
{
|
|
"epoch": 0.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9829,
|
|
"step": 40
|
|
},
|
|
{
|
|
"epoch": 0.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0147,
|
|
"step": 50
|
|
},
|
|
{
|
|
"epoch": 0.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.096,
|
|
"step": 60
|
|
},
|
|
{
|
|
"epoch": 0.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.006,
|
|
"step": 70
|
|
},
|
|
{
|
|
"epoch": 0.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9784,
|
|
"step": 80
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9937,
|
|
"step": 90
|
|
},
|
|
{
|
|
"epoch": 0.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9689,
|
|
"step": 100
|
|
},
|
|
{
|
|
"epoch": 0.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0182,
|
|
"step": 110
|
|
},
|
|
{
|
|
"epoch": 0.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9828,
|
|
"step": 120
|
|
},
|
|
{
|
|
"epoch": 0.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9499,
|
|
"step": 130
|
|
},
|
|
{
|
|
"epoch": 0.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9696,
|
|
"step": 140
|
|
},
|
|
{
|
|
"epoch": 0.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9334,
|
|
"step": 150
|
|
},
|
|
{
|
|
"epoch": 0.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9834,
|
|
"step": 160
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9474,
|
|
"step": 170
|
|
},
|
|
{
|
|
"epoch": 0.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9558,
|
|
"step": 180
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 190
|
|
},
|
|
{
|
|
"epoch": 0.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9384,
|
|
"step": 200
|
|
},
|
|
{
|
|
"epoch": 0.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9978,
|
|
"step": 210
|
|
},
|
|
{
|
|
"epoch": 0.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9604,
|
|
"step": 220
|
|
},
|
|
{
|
|
"epoch": 0.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.971,
|
|
"step": 230
|
|
},
|
|
{
|
|
"epoch": 0.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9705,
|
|
"step": 240
|
|
},
|
|
{
|
|
"epoch": 0.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9544,
|
|
"step": 250
|
|
},
|
|
{
|
|
"epoch": 0.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9343,
|
|
"step": 260
|
|
},
|
|
{
|
|
"epoch": 0.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9263,
|
|
"step": 270
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9695,
|
|
"step": 280
|
|
},
|
|
{
|
|
"epoch": 0.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9348,
|
|
"step": 290
|
|
},
|
|
{
|
|
"epoch": 0.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9506,
|
|
"step": 300
|
|
},
|
|
{
|
|
"epoch": 0.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9976,
|
|
"step": 310
|
|
},
|
|
{
|
|
"epoch": 0.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9527,
|
|
"step": 320
|
|
},
|
|
{
|
|
"epoch": 0.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9526,
|
|
"step": 330
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9623,
|
|
"step": 340
|
|
},
|
|
{
|
|
"epoch": 0.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9455,
|
|
"step": 350
|
|
},
|
|
{
|
|
"epoch": 0.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9286,
|
|
"step": 360
|
|
},
|
|
{
|
|
"epoch": 0.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9476,
|
|
"step": 370
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9484,
|
|
"step": 380
|
|
},
|
|
{
|
|
"epoch": 0.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 390
|
|
},
|
|
{
|
|
"epoch": 0.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9471,
|
|
"step": 400
|
|
},
|
|
{
|
|
"epoch": 0.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9434,
|
|
"step": 410
|
|
},
|
|
{
|
|
"epoch": 0.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.928,
|
|
"step": 420
|
|
},
|
|
{
|
|
"epoch": 0.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 430
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.8994,
|
|
"step": 440
|
|
},
|
|
{
|
|
"epoch": 0.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.029,
|
|
"step": 450
|
|
},
|
|
{
|
|
"epoch": 0.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9449,
|
|
"step": 460
|
|
},
|
|
{
|
|
"epoch": 0.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9539,
|
|
"step": 470
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9541,
|
|
"step": 480
|
|
},
|
|
{
|
|
"epoch": 0.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.956,
|
|
"step": 490
|
|
},
|
|
{
|
|
"epoch": 0.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9707,
|
|
"step": 500
|
|
},
|
|
{
|
|
"epoch": 0.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 510
|
|
},
|
|
{
|
|
"epoch": 0.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9656,
|
|
"step": 520
|
|
},
|
|
{
|
|
"epoch": 0.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 530
|
|
},
|
|
{
|
|
"epoch": 0.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.955,
|
|
"step": 540
|
|
},
|
|
{
|
|
"epoch": 0.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9288,
|
|
"step": 550
|
|
},
|
|
{
|
|
"epoch": 0.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9574,
|
|
"step": 560
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.956,
|
|
"step": 570
|
|
},
|
|
{
|
|
"epoch": 0.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9587,
|
|
"step": 580
|
|
},
|
|
{
|
|
"epoch": 0.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9473,
|
|
"step": 590
|
|
},
|
|
{
|
|
"epoch": 0.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 600
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9648,
|
|
"step": 610
|
|
},
|
|
{
|
|
"epoch": 0.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9314,
|
|
"step": 620
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.961,
|
|
"step": 630
|
|
},
|
|
{
|
|
"epoch": 0.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9386,
|
|
"step": 640
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 650
|
|
},
|
|
{
|
|
"epoch": 0.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.967,
|
|
"step": 660
|
|
},
|
|
{
|
|
"epoch": 0.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9451,
|
|
"step": 670
|
|
},
|
|
{
|
|
"epoch": 0.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 680
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 690
|
|
},
|
|
{
|
|
"epoch": 0.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 700
|
|
},
|
|
{
|
|
"epoch": 0.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 710
|
|
},
|
|
{
|
|
"epoch": 0.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 720
|
|
},
|
|
{
|
|
"epoch": 0.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 730
|
|
},
|
|
{
|
|
"epoch": 0.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9188,
|
|
"step": 740
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0077,
|
|
"step": 750
|
|
},
|
|
{
|
|
"epoch": 0.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 760
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9348,
|
|
"step": 770
|
|
},
|
|
{
|
|
"epoch": 0.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9512,
|
|
"step": 780
|
|
},
|
|
{
|
|
"epoch": 0.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9409,
|
|
"step": 790
|
|
},
|
|
{
|
|
"epoch": 0.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 800
|
|
},
|
|
{
|
|
"epoch": 0.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9656,
|
|
"step": 810
|
|
},
|
|
{
|
|
"epoch": 0.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 820
|
|
},
|
|
{
|
|
"epoch": 0.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9471,
|
|
"step": 830
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 840
|
|
},
|
|
{
|
|
"epoch": 0.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9386,
|
|
"step": 850
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9568,
|
|
"step": 860
|
|
},
|
|
{
|
|
"epoch": 0.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9498,
|
|
"step": 870
|
|
},
|
|
{
|
|
"epoch": 0.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9621,
|
|
"step": 880
|
|
},
|
|
{
|
|
"epoch": 0.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 890
|
|
},
|
|
{
|
|
"epoch": 0.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 900
|
|
},
|
|
{
|
|
"epoch": 0.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 910
|
|
},
|
|
{
|
|
"epoch": 0.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 920
|
|
},
|
|
{
|
|
"epoch": 0.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9427,
|
|
"step": 930
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9586,
|
|
"step": 940
|
|
},
|
|
{
|
|
"epoch": 0.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9588,
|
|
"step": 950
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9511,
|
|
"step": 960
|
|
},
|
|
{
|
|
"epoch": 0.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 970
|
|
},
|
|
{
|
|
"epoch": 0.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9424,
|
|
"step": 980
|
|
},
|
|
{
|
|
"epoch": 0.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 990
|
|
},
|
|
{
|
|
"epoch": 0.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 1000
|
|
},
|
|
{
|
|
"epoch": 0.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 1010
|
|
},
|
|
{
|
|
"epoch": 0.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9338,
|
|
"step": 1020
|
|
},
|
|
{
|
|
"epoch": 0.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 1030
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 1040
|
|
},
|
|
{
|
|
"epoch": 0.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9347,
|
|
"step": 1050
|
|
},
|
|
{
|
|
"epoch": 0.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9458,
|
|
"step": 1060
|
|
},
|
|
{
|
|
"epoch": 0.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9389,
|
|
"step": 1070
|
|
},
|
|
{
|
|
"epoch": 0.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9445,
|
|
"step": 1080
|
|
},
|
|
{
|
|
"epoch": 0.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9594,
|
|
"step": 1090
|
|
},
|
|
{
|
|
"epoch": 0.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9666,
|
|
"step": 1100
|
|
},
|
|
{
|
|
"epoch": 0.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9567,
|
|
"step": 1110
|
|
},
|
|
{
|
|
"epoch": 0.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9527,
|
|
"step": 1120
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9434,
|
|
"step": 1130
|
|
},
|
|
{
|
|
"epoch": 0.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9252,
|
|
"step": 1140
|
|
},
|
|
{
|
|
"epoch": 0.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9541,
|
|
"step": 1150
|
|
},
|
|
{
|
|
"epoch": 0.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9619,
|
|
"step": 1160
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9531,
|
|
"step": 1170
|
|
},
|
|
{
|
|
"epoch": 0.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 1180
|
|
},
|
|
{
|
|
"epoch": 0.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 1190
|
|
},
|
|
{
|
|
"epoch": 0.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 1200
|
|
},
|
|
{
|
|
"epoch": 0.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9166,
|
|
"step": 1210
|
|
},
|
|
{
|
|
"epoch": 0.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0,
|
|
"step": 1220
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9459,
|
|
"step": 1230
|
|
},
|
|
{
|
|
"epoch": 0.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9397,
|
|
"step": 1240
|
|
},
|
|
{
|
|
"epoch": 0.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9722,
|
|
"step": 1250
|
|
},
|
|
{
|
|
"epoch": 0.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9704,
|
|
"step": 1260
|
|
},
|
|
{
|
|
"epoch": 0.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.957,
|
|
"step": 1270
|
|
},
|
|
{
|
|
"epoch": 0.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9551,
|
|
"step": 1280
|
|
},
|
|
{
|
|
"epoch": 0.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9534,
|
|
"step": 1290
|
|
},
|
|
{
|
|
"epoch": 0.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9718,
|
|
"step": 1300
|
|
},
|
|
{
|
|
"epoch": 0.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9597,
|
|
"step": 1310
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9527,
|
|
"step": 1320
|
|
},
|
|
{
|
|
"epoch": 0.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 1330
|
|
},
|
|
{
|
|
"epoch": 0.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9636,
|
|
"step": 1340
|
|
},
|
|
{
|
|
"epoch": 0.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 1350
|
|
},
|
|
{
|
|
"epoch": 0.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9551,
|
|
"step": 1360
|
|
},
|
|
{
|
|
"epoch": 0.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0172,
|
|
"step": 1370
|
|
},
|
|
{
|
|
"epoch": 0.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 1380
|
|
},
|
|
{
|
|
"epoch": 0.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 1390
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9877,
|
|
"step": 1400
|
|
},
|
|
{
|
|
"epoch": 0.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.976,
|
|
"step": 1410
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9371,
|
|
"step": 1420
|
|
},
|
|
{
|
|
"epoch": 0.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9797,
|
|
"step": 1430
|
|
},
|
|
{
|
|
"epoch": 0.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9719,
|
|
"step": 1440
|
|
},
|
|
{
|
|
"epoch": 0.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.925,
|
|
"step": 1450
|
|
},
|
|
{
|
|
"epoch": 0.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9756,
|
|
"step": 1460
|
|
},
|
|
{
|
|
"epoch": 0.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 1470
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 1480
|
|
},
|
|
{
|
|
"epoch": 0.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 1490
|
|
},
|
|
{
|
|
"epoch": 0.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 1500
|
|
},
|
|
{
|
|
"epoch": 0.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9347,
|
|
"step": 1510
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9342,
|
|
"step": 1520
|
|
},
|
|
{
|
|
"epoch": 0.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9762,
|
|
"step": 1530
|
|
},
|
|
{
|
|
"epoch": 0.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 1540
|
|
},
|
|
{
|
|
"epoch": 0.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9493,
|
|
"step": 1550
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9309,
|
|
"step": 1560
|
|
},
|
|
{
|
|
"epoch": 0.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9875,
|
|
"step": 1570
|
|
},
|
|
{
|
|
"epoch": 0.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 1580
|
|
},
|
|
{
|
|
"epoch": 0.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9231,
|
|
"step": 1590
|
|
},
|
|
{
|
|
"epoch": 0.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9163,
|
|
"step": 1600
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9894,
|
|
"step": 1610
|
|
},
|
|
{
|
|
"epoch": 0.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 1620
|
|
},
|
|
{
|
|
"epoch": 0.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9237,
|
|
"step": 1630
|
|
},
|
|
{
|
|
"epoch": 0.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.1552,
|
|
"step": 1640
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0399,
|
|
"step": 1650
|
|
},
|
|
{
|
|
"epoch": 0.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.971,
|
|
"step": 1660
|
|
},
|
|
{
|
|
"epoch": 0.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0006,
|
|
"step": 1670
|
|
},
|
|
{
|
|
"epoch": 0.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9451,
|
|
"step": 1680
|
|
},
|
|
{
|
|
"epoch": 0.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.993,
|
|
"step": 1690
|
|
},
|
|
{
|
|
"epoch": 0.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9586,
|
|
"step": 1700
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.963,
|
|
"step": 1710
|
|
},
|
|
{
|
|
"epoch": 0.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9641,
|
|
"step": 1720
|
|
},
|
|
{
|
|
"epoch": 0.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9555,
|
|
"step": 1730
|
|
},
|
|
{
|
|
"epoch": 0.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 1740
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.934,
|
|
"step": 1750
|
|
},
|
|
{
|
|
"epoch": 0.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 2.0062,
|
|
"step": 1760
|
|
},
|
|
{
|
|
"epoch": 0.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9315,
|
|
"step": 1770
|
|
},
|
|
{
|
|
"epoch": 0.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9626,
|
|
"step": 1780
|
|
},
|
|
{
|
|
"epoch": 0.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9135,
|
|
"step": 1790
|
|
},
|
|
{
|
|
"epoch": 0.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9756,
|
|
"step": 1800
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9628,
|
|
"step": 1810
|
|
},
|
|
{
|
|
"epoch": 0.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.952,
|
|
"step": 1820
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9326,
|
|
"step": 1830
|
|
},
|
|
{
|
|
"epoch": 0.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9552,
|
|
"step": 1840
|
|
},
|
|
{
|
|
"epoch": 0.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9527,
|
|
"step": 1850
|
|
},
|
|
{
|
|
"epoch": 0.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 1860
|
|
},
|
|
{
|
|
"epoch": 0.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 1870
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 1880
|
|
},
|
|
{
|
|
"epoch": 0.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9465,
|
|
"step": 1890
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 1900
|
|
},
|
|
{
|
|
"epoch": 0.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9271,
|
|
"step": 1910
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.981,
|
|
"step": 1920
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.1556420233463035
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03848003848003848
|
|
},
|
|
"eval_loss": 1.9584157466888428,
|
|
"eval_precision": {
|
|
"precision": 0.022234574763757644
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 201.5846,
|
|
"eval_samples_per_second": 19.123,
|
|
"eval_steps_per_second": 9.564,
|
|
"step": 1927
|
|
},
|
|
{
|
|
"epoch": 1.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9356,
|
|
"step": 1930
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9557,
|
|
"step": 1940
|
|
},
|
|
{
|
|
"epoch": 1.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 1950
|
|
},
|
|
{
|
|
"epoch": 1.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9192,
|
|
"step": 1960
|
|
},
|
|
{
|
|
"epoch": 1.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9903,
|
|
"step": 1970
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9343,
|
|
"step": 1980
|
|
},
|
|
{
|
|
"epoch": 1.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9753,
|
|
"step": 1990
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9178,
|
|
"step": 2000
|
|
},
|
|
{
|
|
"epoch": 1.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.925,
|
|
"step": 2010
|
|
},
|
|
{
|
|
"epoch": 1.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.955,
|
|
"step": 2020
|
|
},
|
|
{
|
|
"epoch": 1.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9774,
|
|
"step": 2030
|
|
},
|
|
{
|
|
"epoch": 1.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9351,
|
|
"step": 2040
|
|
},
|
|
{
|
|
"epoch": 1.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9794,
|
|
"step": 2050
|
|
},
|
|
{
|
|
"epoch": 1.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 2060
|
|
},
|
|
{
|
|
"epoch": 1.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9512,
|
|
"step": 2070
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9483,
|
|
"step": 2080
|
|
},
|
|
{
|
|
"epoch": 1.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9376,
|
|
"step": 2090
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9282,
|
|
"step": 2100
|
|
},
|
|
{
|
|
"epoch": 1.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.936,
|
|
"step": 2110
|
|
},
|
|
{
|
|
"epoch": 1.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9811,
|
|
"step": 2120
|
|
},
|
|
{
|
|
"epoch": 1.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 2130
|
|
},
|
|
{
|
|
"epoch": 1.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9484,
|
|
"step": 2140
|
|
},
|
|
{
|
|
"epoch": 1.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9356,
|
|
"step": 2150
|
|
},
|
|
{
|
|
"epoch": 1.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 2160
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.944,
|
|
"step": 2170
|
|
},
|
|
{
|
|
"epoch": 1.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9451,
|
|
"step": 2180
|
|
},
|
|
{
|
|
"epoch": 1.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 2190
|
|
},
|
|
{
|
|
"epoch": 1.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 2200
|
|
},
|
|
{
|
|
"epoch": 1.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9374,
|
|
"step": 2210
|
|
},
|
|
{
|
|
"epoch": 1.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 2220
|
|
},
|
|
{
|
|
"epoch": 1.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 2230
|
|
},
|
|
{
|
|
"epoch": 1.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9339,
|
|
"step": 2240
|
|
},
|
|
{
|
|
"epoch": 1.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9285,
|
|
"step": 2250
|
|
},
|
|
{
|
|
"epoch": 1.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 2260
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9719,
|
|
"step": 2270
|
|
},
|
|
{
|
|
"epoch": 1.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9526,
|
|
"step": 2280
|
|
},
|
|
{
|
|
"epoch": 1.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9551,
|
|
"step": 2290
|
|
},
|
|
{
|
|
"epoch": 1.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 2300
|
|
},
|
|
{
|
|
"epoch": 1.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 2310
|
|
},
|
|
{
|
|
"epoch": 1.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 2320
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9403,
|
|
"step": 2330
|
|
},
|
|
{
|
|
"epoch": 1.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9462,
|
|
"step": 2340
|
|
},
|
|
{
|
|
"epoch": 1.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9344,
|
|
"step": 2350
|
|
},
|
|
{
|
|
"epoch": 1.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9686,
|
|
"step": 2360
|
|
},
|
|
{
|
|
"epoch": 1.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9476,
|
|
"step": 2370
|
|
},
|
|
{
|
|
"epoch": 1.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.955,
|
|
"step": 2380
|
|
},
|
|
{
|
|
"epoch": 1.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 2390
|
|
},
|
|
{
|
|
"epoch": 1.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.947,
|
|
"step": 2400
|
|
},
|
|
{
|
|
"epoch": 1.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9498,
|
|
"step": 2410
|
|
},
|
|
{
|
|
"epoch": 1.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 2420
|
|
},
|
|
{
|
|
"epoch": 1.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 2430
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 2440
|
|
},
|
|
{
|
|
"epoch": 1.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 2450
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 2460
|
|
},
|
|
{
|
|
"epoch": 1.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9401,
|
|
"step": 2470
|
|
},
|
|
{
|
|
"epoch": 1.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 2480
|
|
},
|
|
{
|
|
"epoch": 1.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9474,
|
|
"step": 2490
|
|
},
|
|
{
|
|
"epoch": 1.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 2500
|
|
},
|
|
{
|
|
"epoch": 1.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 2510
|
|
},
|
|
{
|
|
"epoch": 1.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9523,
|
|
"step": 2520
|
|
},
|
|
{
|
|
"epoch": 1.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9497,
|
|
"step": 2530
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9332,
|
|
"step": 2540
|
|
},
|
|
{
|
|
"epoch": 1.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 2550
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9374,
|
|
"step": 2560
|
|
},
|
|
{
|
|
"epoch": 1.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9349,
|
|
"step": 2570
|
|
},
|
|
{
|
|
"epoch": 1.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9223,
|
|
"step": 2580
|
|
},
|
|
{
|
|
"epoch": 1.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9679,
|
|
"step": 2590
|
|
},
|
|
{
|
|
"epoch": 1.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9427,
|
|
"step": 2600
|
|
},
|
|
{
|
|
"epoch": 1.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 2610
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9311,
|
|
"step": 2620
|
|
},
|
|
{
|
|
"epoch": 1.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.951,
|
|
"step": 2630
|
|
},
|
|
{
|
|
"epoch": 1.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9154,
|
|
"step": 2640
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.965,
|
|
"step": 2650
|
|
},
|
|
{
|
|
"epoch": 1.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9611,
|
|
"step": 2660
|
|
},
|
|
{
|
|
"epoch": 1.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9356,
|
|
"step": 2670
|
|
},
|
|
{
|
|
"epoch": 1.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9417,
|
|
"step": 2680
|
|
},
|
|
{
|
|
"epoch": 1.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9597,
|
|
"step": 2690
|
|
},
|
|
{
|
|
"epoch": 1.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9337,
|
|
"step": 2700
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9329,
|
|
"step": 2710
|
|
},
|
|
{
|
|
"epoch": 1.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.934,
|
|
"step": 2720
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.921,
|
|
"step": 2730
|
|
},
|
|
{
|
|
"epoch": 1.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 2740
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 2750
|
|
},
|
|
{
|
|
"epoch": 1.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 2760
|
|
},
|
|
{
|
|
"epoch": 1.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9527,
|
|
"step": 2770
|
|
},
|
|
{
|
|
"epoch": 1.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 2780
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9506,
|
|
"step": 2790
|
|
},
|
|
{
|
|
"epoch": 1.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9293,
|
|
"step": 2800
|
|
},
|
|
{
|
|
"epoch": 1.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9577,
|
|
"step": 2810
|
|
},
|
|
{
|
|
"epoch": 1.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9549,
|
|
"step": 2820
|
|
},
|
|
{
|
|
"epoch": 1.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9555,
|
|
"step": 2830
|
|
},
|
|
{
|
|
"epoch": 1.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9427,
|
|
"step": 2840
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9417,
|
|
"step": 2850
|
|
},
|
|
{
|
|
"epoch": 1.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9328,
|
|
"step": 2860
|
|
},
|
|
{
|
|
"epoch": 1.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 2870
|
|
},
|
|
{
|
|
"epoch": 1.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9619,
|
|
"step": 2880
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 2890
|
|
},
|
|
{
|
|
"epoch": 1.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9323,
|
|
"step": 2900
|
|
},
|
|
{
|
|
"epoch": 1.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9226,
|
|
"step": 2910
|
|
},
|
|
{
|
|
"epoch": 1.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9836,
|
|
"step": 2920
|
|
},
|
|
{
|
|
"epoch": 1.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 2930
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9512,
|
|
"step": 2940
|
|
},
|
|
{
|
|
"epoch": 1.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9445,
|
|
"step": 2950
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 2960
|
|
},
|
|
{
|
|
"epoch": 1.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 2970
|
|
},
|
|
{
|
|
"epoch": 1.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.942,
|
|
"step": 2980
|
|
},
|
|
{
|
|
"epoch": 1.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9563,
|
|
"step": 2990
|
|
},
|
|
{
|
|
"epoch": 1.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 3000
|
|
},
|
|
{
|
|
"epoch": 1.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9591,
|
|
"step": 3010
|
|
},
|
|
{
|
|
"epoch": 1.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9421,
|
|
"step": 3020
|
|
},
|
|
{
|
|
"epoch": 1.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9447,
|
|
"step": 3030
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9345,
|
|
"step": 3040
|
|
},
|
|
{
|
|
"epoch": 1.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9399,
|
|
"step": 3050
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9321,
|
|
"step": 3060
|
|
},
|
|
{
|
|
"epoch": 1.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9367,
|
|
"step": 3070
|
|
},
|
|
{
|
|
"epoch": 1.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 3080
|
|
},
|
|
{
|
|
"epoch": 1.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9531,
|
|
"step": 3090
|
|
},
|
|
{
|
|
"epoch": 1.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9573,
|
|
"step": 3100
|
|
},
|
|
{
|
|
"epoch": 1.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 3110
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9541,
|
|
"step": 3120
|
|
},
|
|
{
|
|
"epoch": 1.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9457,
|
|
"step": 3130
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9443,
|
|
"step": 3140
|
|
},
|
|
{
|
|
"epoch": 1.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9431,
|
|
"step": 3150
|
|
},
|
|
{
|
|
"epoch": 1.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9507,
|
|
"step": 3160
|
|
},
|
|
{
|
|
"epoch": 1.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9282,
|
|
"step": 3170
|
|
},
|
|
{
|
|
"epoch": 1.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9731,
|
|
"step": 3180
|
|
},
|
|
{
|
|
"epoch": 1.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 3190
|
|
},
|
|
{
|
|
"epoch": 1.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 3200
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9439,
|
|
"step": 3210
|
|
},
|
|
{
|
|
"epoch": 1.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9339,
|
|
"step": 3220
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 3230
|
|
},
|
|
{
|
|
"epoch": 1.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9429,
|
|
"step": 3240
|
|
},
|
|
{
|
|
"epoch": 1.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9539,
|
|
"step": 3250
|
|
},
|
|
{
|
|
"epoch": 1.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 3260
|
|
},
|
|
{
|
|
"epoch": 1.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 3270
|
|
},
|
|
{
|
|
"epoch": 1.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.953,
|
|
"step": 3280
|
|
},
|
|
{
|
|
"epoch": 1.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9504,
|
|
"step": 3290
|
|
},
|
|
{
|
|
"epoch": 1.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9493,
|
|
"step": 3300
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9371,
|
|
"step": 3310
|
|
},
|
|
{
|
|
"epoch": 1.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 3320
|
|
},
|
|
{
|
|
"epoch": 1.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9654,
|
|
"step": 3330
|
|
},
|
|
{
|
|
"epoch": 1.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9607,
|
|
"step": 3340
|
|
},
|
|
{
|
|
"epoch": 1.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.956,
|
|
"step": 3350
|
|
},
|
|
{
|
|
"epoch": 1.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 3360
|
|
},
|
|
{
|
|
"epoch": 1.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9138,
|
|
"step": 3370
|
|
},
|
|
{
|
|
"epoch": 1.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.928,
|
|
"step": 3380
|
|
},
|
|
{
|
|
"epoch": 1.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9845,
|
|
"step": 3390
|
|
},
|
|
{
|
|
"epoch": 1.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9536,
|
|
"step": 3400
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9458,
|
|
"step": 3410
|
|
},
|
|
{
|
|
"epoch": 1.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9565,
|
|
"step": 3420
|
|
},
|
|
{
|
|
"epoch": 1.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.935,
|
|
"step": 3430
|
|
},
|
|
{
|
|
"epoch": 1.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9366,
|
|
"step": 3440
|
|
},
|
|
{
|
|
"epoch": 1.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.931,
|
|
"step": 3450
|
|
},
|
|
{
|
|
"epoch": 1.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9534,
|
|
"step": 3460
|
|
},
|
|
{
|
|
"epoch": 1.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 3470
|
|
},
|
|
{
|
|
"epoch": 1.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9455,
|
|
"step": 3480
|
|
},
|
|
{
|
|
"epoch": 1.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 3490
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9459,
|
|
"step": 3500
|
|
},
|
|
{
|
|
"epoch": 1.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9354,
|
|
"step": 3510
|
|
},
|
|
{
|
|
"epoch": 1.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9056,
|
|
"step": 3520
|
|
},
|
|
{
|
|
"epoch": 1.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9871,
|
|
"step": 3530
|
|
},
|
|
{
|
|
"epoch": 1.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9884,
|
|
"step": 3540
|
|
},
|
|
{
|
|
"epoch": 1.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.944,
|
|
"step": 3550
|
|
},
|
|
{
|
|
"epoch": 1.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 3560
|
|
},
|
|
{
|
|
"epoch": 1.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9536,
|
|
"step": 3570
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9269,
|
|
"step": 3580
|
|
},
|
|
{
|
|
"epoch": 1.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9259,
|
|
"step": 3590
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9648,
|
|
"step": 3600
|
|
},
|
|
{
|
|
"epoch": 1.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9642,
|
|
"step": 3610
|
|
},
|
|
{
|
|
"epoch": 1.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9677,
|
|
"step": 3620
|
|
},
|
|
{
|
|
"epoch": 1.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 3630
|
|
},
|
|
{
|
|
"epoch": 1.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9517,
|
|
"step": 3640
|
|
},
|
|
{
|
|
"epoch": 1.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9523,
|
|
"step": 3650
|
|
},
|
|
{
|
|
"epoch": 1.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 3660
|
|
},
|
|
{
|
|
"epoch": 1.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 3670
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 3680
|
|
},
|
|
{
|
|
"epoch": 1.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 3690
|
|
},
|
|
{
|
|
"epoch": 1.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 3700
|
|
},
|
|
{
|
|
"epoch": 1.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9575,
|
|
"step": 3710
|
|
},
|
|
{
|
|
"epoch": 1.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 3720
|
|
},
|
|
{
|
|
"epoch": 1.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9534,
|
|
"step": 3730
|
|
},
|
|
{
|
|
"epoch": 1.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9484,
|
|
"step": 3740
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9277,
|
|
"step": 3750
|
|
},
|
|
{
|
|
"epoch": 1.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9326,
|
|
"step": 3760
|
|
},
|
|
{
|
|
"epoch": 1.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 3770
|
|
},
|
|
{
|
|
"epoch": 1.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9568,
|
|
"step": 3780
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9535,
|
|
"step": 3790
|
|
},
|
|
{
|
|
"epoch": 1.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9523,
|
|
"step": 3800
|
|
},
|
|
{
|
|
"epoch": 1.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 3810
|
|
},
|
|
{
|
|
"epoch": 1.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9607,
|
|
"step": 3820
|
|
},
|
|
{
|
|
"epoch": 1.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9499,
|
|
"step": 3830
|
|
},
|
|
{
|
|
"epoch": 1.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 3840
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.951,
|
|
"step": 3850
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.1556420233463035
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03848003848003848
|
|
},
|
|
"eval_loss": 1.9446961879730225,
|
|
"eval_precision": {
|
|
"precision": 0.022234574763757644
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 78.0133,
|
|
"eval_samples_per_second": 49.415,
|
|
"eval_steps_per_second": 24.714,
|
|
"step": 3854
|
|
},
|
|
{
|
|
"epoch": 2.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 3860
|
|
},
|
|
{
|
|
"epoch": 2.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9355,
|
|
"step": 3870
|
|
},
|
|
{
|
|
"epoch": 2.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9457,
|
|
"step": 3880
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 3890
|
|
},
|
|
{
|
|
"epoch": 2.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9332,
|
|
"step": 3900
|
|
},
|
|
{
|
|
"epoch": 2.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9269,
|
|
"step": 3910
|
|
},
|
|
{
|
|
"epoch": 2.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9283,
|
|
"step": 3920
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9559,
|
|
"step": 3930
|
|
},
|
|
{
|
|
"epoch": 2.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9414,
|
|
"step": 3940
|
|
},
|
|
{
|
|
"epoch": 2.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 3950
|
|
},
|
|
{
|
|
"epoch": 2.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9506,
|
|
"step": 3960
|
|
},
|
|
{
|
|
"epoch": 2.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 3970
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 3980
|
|
},
|
|
{
|
|
"epoch": 2.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 3990
|
|
},
|
|
{
|
|
"epoch": 2.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 4000
|
|
},
|
|
{
|
|
"epoch": 2.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9581,
|
|
"step": 4010
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9388,
|
|
"step": 4020
|
|
},
|
|
{
|
|
"epoch": 2.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.935,
|
|
"step": 4030
|
|
},
|
|
{
|
|
"epoch": 2.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9466,
|
|
"step": 4040
|
|
},
|
|
{
|
|
"epoch": 2.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9417,
|
|
"step": 4050
|
|
},
|
|
{
|
|
"epoch": 2.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9401,
|
|
"step": 4060
|
|
},
|
|
{
|
|
"epoch": 2.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9596,
|
|
"step": 4070
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9571,
|
|
"step": 4080
|
|
},
|
|
{
|
|
"epoch": 2.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 4090
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9431,
|
|
"step": 4100
|
|
},
|
|
{
|
|
"epoch": 2.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.942,
|
|
"step": 4110
|
|
},
|
|
{
|
|
"epoch": 2.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9499,
|
|
"step": 4120
|
|
},
|
|
{
|
|
"epoch": 2.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 4130
|
|
},
|
|
{
|
|
"epoch": 2.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9363,
|
|
"step": 4140
|
|
},
|
|
{
|
|
"epoch": 2.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 4150
|
|
},
|
|
{
|
|
"epoch": 2.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 4160
|
|
},
|
|
{
|
|
"epoch": 2.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9266,
|
|
"step": 4170
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 4180
|
|
},
|
|
{
|
|
"epoch": 2.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 4190
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9429,
|
|
"step": 4200
|
|
},
|
|
{
|
|
"epoch": 2.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9465,
|
|
"step": 4210
|
|
},
|
|
{
|
|
"epoch": 2.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.952,
|
|
"step": 4220
|
|
},
|
|
{
|
|
"epoch": 2.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9493,
|
|
"step": 4230
|
|
},
|
|
{
|
|
"epoch": 2.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9393,
|
|
"step": 4240
|
|
},
|
|
{
|
|
"epoch": 2.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9531,
|
|
"step": 4250
|
|
},
|
|
{
|
|
"epoch": 2.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 4260
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9493,
|
|
"step": 4270
|
|
},
|
|
{
|
|
"epoch": 2.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.944,
|
|
"step": 4280
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9572,
|
|
"step": 4290
|
|
},
|
|
{
|
|
"epoch": 2.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9311,
|
|
"step": 4300
|
|
},
|
|
{
|
|
"epoch": 2.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9446,
|
|
"step": 4310
|
|
},
|
|
{
|
|
"epoch": 2.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9557,
|
|
"step": 4320
|
|
},
|
|
{
|
|
"epoch": 2.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9396,
|
|
"step": 4330
|
|
},
|
|
{
|
|
"epoch": 2.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9253,
|
|
"step": 4340
|
|
},
|
|
{
|
|
"epoch": 2.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9697,
|
|
"step": 4350
|
|
},
|
|
{
|
|
"epoch": 2.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9489,
|
|
"step": 4360
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9489,
|
|
"step": 4370
|
|
},
|
|
{
|
|
"epoch": 2.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9589,
|
|
"step": 4380
|
|
},
|
|
{
|
|
"epoch": 2.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 4390
|
|
},
|
|
{
|
|
"epoch": 2.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 4400
|
|
},
|
|
{
|
|
"epoch": 2.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.958,
|
|
"step": 4410
|
|
},
|
|
{
|
|
"epoch": 2.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9288,
|
|
"step": 4420
|
|
},
|
|
{
|
|
"epoch": 2.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9225,
|
|
"step": 4430
|
|
},
|
|
{
|
|
"epoch": 2.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9194,
|
|
"step": 4440
|
|
},
|
|
{
|
|
"epoch": 2.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9816,
|
|
"step": 4450
|
|
},
|
|
{
|
|
"epoch": 2.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 4460
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 4470
|
|
},
|
|
{
|
|
"epoch": 2.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9567,
|
|
"step": 4480
|
|
},
|
|
{
|
|
"epoch": 2.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.955,
|
|
"step": 4490
|
|
},
|
|
{
|
|
"epoch": 2.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9545,
|
|
"step": 4500
|
|
},
|
|
{
|
|
"epoch": 2.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9386,
|
|
"step": 4510
|
|
},
|
|
{
|
|
"epoch": 2.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9513,
|
|
"step": 4520
|
|
},
|
|
{
|
|
"epoch": 2.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9269,
|
|
"step": 4530
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9262,
|
|
"step": 4540
|
|
},
|
|
{
|
|
"epoch": 2.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9299,
|
|
"step": 4550
|
|
},
|
|
{
|
|
"epoch": 2.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 4560
|
|
},
|
|
{
|
|
"epoch": 2.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9701,
|
|
"step": 4570
|
|
},
|
|
{
|
|
"epoch": 2.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9508,
|
|
"step": 4580
|
|
},
|
|
{
|
|
"epoch": 2.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 4590
|
|
},
|
|
{
|
|
"epoch": 2.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9614,
|
|
"step": 4600
|
|
},
|
|
{
|
|
"epoch": 2.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 4610
|
|
},
|
|
{
|
|
"epoch": 2.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9401,
|
|
"step": 4620
|
|
},
|
|
{
|
|
"epoch": 2.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9388,
|
|
"step": 4630
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9659,
|
|
"step": 4640
|
|
},
|
|
{
|
|
"epoch": 2.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9382,
|
|
"step": 4650
|
|
},
|
|
{
|
|
"epoch": 2.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 4660
|
|
},
|
|
{
|
|
"epoch": 2.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9567,
|
|
"step": 4670
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9468,
|
|
"step": 4680
|
|
},
|
|
{
|
|
"epoch": 2.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9426,
|
|
"step": 4690
|
|
},
|
|
{
|
|
"epoch": 2.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9659,
|
|
"step": 4700
|
|
},
|
|
{
|
|
"epoch": 2.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9491,
|
|
"step": 4710
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 4720
|
|
},
|
|
{
|
|
"epoch": 2.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 4730
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 4740
|
|
},
|
|
{
|
|
"epoch": 2.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 4750
|
|
},
|
|
{
|
|
"epoch": 2.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 4760
|
|
},
|
|
{
|
|
"epoch": 2.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9248,
|
|
"step": 4770
|
|
},
|
|
{
|
|
"epoch": 2.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9324,
|
|
"step": 4780
|
|
},
|
|
{
|
|
"epoch": 2.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9833,
|
|
"step": 4790
|
|
},
|
|
{
|
|
"epoch": 2.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9301,
|
|
"step": 4800
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 4810
|
|
},
|
|
{
|
|
"epoch": 2.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9267,
|
|
"step": 4820
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9304,
|
|
"step": 4830
|
|
},
|
|
{
|
|
"epoch": 2.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9502,
|
|
"step": 4840
|
|
},
|
|
{
|
|
"epoch": 2.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 4850
|
|
},
|
|
{
|
|
"epoch": 2.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9399,
|
|
"step": 4860
|
|
},
|
|
{
|
|
"epoch": 2.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 4870
|
|
},
|
|
{
|
|
"epoch": 2.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.92,
|
|
"step": 4880
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 4890
|
|
},
|
|
{
|
|
"epoch": 2.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 4900
|
|
},
|
|
{
|
|
"epoch": 2.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9496,
|
|
"step": 4910
|
|
},
|
|
{
|
|
"epoch": 2.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 4920
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 4930
|
|
},
|
|
{
|
|
"epoch": 2.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 4940
|
|
},
|
|
{
|
|
"epoch": 2.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 4950
|
|
},
|
|
{
|
|
"epoch": 2.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.951,
|
|
"step": 4960
|
|
},
|
|
{
|
|
"epoch": 2.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.8918,
|
|
"step": 4970
|
|
},
|
|
{
|
|
"epoch": 2.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 4980
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9264,
|
|
"step": 4990
|
|
},
|
|
{
|
|
"epoch": 2.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9679,
|
|
"step": 5000
|
|
},
|
|
{
|
|
"epoch": 2.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9513,
|
|
"step": 5010
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9212,
|
|
"step": 5020
|
|
},
|
|
{
|
|
"epoch": 2.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.929,
|
|
"step": 5030
|
|
},
|
|
{
|
|
"epoch": 2.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 5040
|
|
},
|
|
{
|
|
"epoch": 2.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 5050
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9432,
|
|
"step": 5060
|
|
},
|
|
{
|
|
"epoch": 2.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9235,
|
|
"step": 5070
|
|
},
|
|
{
|
|
"epoch": 2.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9589,
|
|
"step": 5080
|
|
},
|
|
{
|
|
"epoch": 2.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9806,
|
|
"step": 5090
|
|
},
|
|
{
|
|
"epoch": 2.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9347,
|
|
"step": 5100
|
|
},
|
|
{
|
|
"epoch": 2.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9781,
|
|
"step": 5110
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9489,
|
|
"step": 5120
|
|
},
|
|
{
|
|
"epoch": 2.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 5130
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9579,
|
|
"step": 5140
|
|
},
|
|
{
|
|
"epoch": 2.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.967,
|
|
"step": 5150
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 5160
|
|
},
|
|
{
|
|
"epoch": 2.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9492,
|
|
"step": 5170
|
|
},
|
|
{
|
|
"epoch": 2.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9345,
|
|
"step": 5180
|
|
},
|
|
{
|
|
"epoch": 2.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 5190
|
|
},
|
|
{
|
|
"epoch": 2.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 5200
|
|
},
|
|
{
|
|
"epoch": 2.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.96,
|
|
"step": 5210
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9477,
|
|
"step": 5220
|
|
},
|
|
{
|
|
"epoch": 2.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9319,
|
|
"step": 5230
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 5240
|
|
},
|
|
{
|
|
"epoch": 2.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9389,
|
|
"step": 5250
|
|
},
|
|
{
|
|
"epoch": 2.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9474,
|
|
"step": 5260
|
|
},
|
|
{
|
|
"epoch": 2.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 5270
|
|
},
|
|
{
|
|
"epoch": 2.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9403,
|
|
"step": 5280
|
|
},
|
|
{
|
|
"epoch": 2.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9624,
|
|
"step": 5290
|
|
},
|
|
{
|
|
"epoch": 2.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9514,
|
|
"step": 5300
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9217,
|
|
"step": 5310
|
|
},
|
|
{
|
|
"epoch": 2.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9347,
|
|
"step": 5320
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9502,
|
|
"step": 5330
|
|
},
|
|
{
|
|
"epoch": 2.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.99,
|
|
"step": 5340
|
|
},
|
|
{
|
|
"epoch": 2.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9212,
|
|
"step": 5350
|
|
},
|
|
{
|
|
"epoch": 2.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 5360
|
|
},
|
|
{
|
|
"epoch": 2.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 5370
|
|
},
|
|
{
|
|
"epoch": 2.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 5380
|
|
},
|
|
{
|
|
"epoch": 2.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 5390
|
|
},
|
|
{
|
|
"epoch": 2.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9239,
|
|
"step": 5400
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9821,
|
|
"step": 5410
|
|
},
|
|
{
|
|
"epoch": 2.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.934,
|
|
"step": 5420
|
|
},
|
|
{
|
|
"epoch": 2.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 5430
|
|
},
|
|
{
|
|
"epoch": 2.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9296,
|
|
"step": 5440
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9525,
|
|
"step": 5450
|
|
},
|
|
{
|
|
"epoch": 2.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9197,
|
|
"step": 5460
|
|
},
|
|
{
|
|
"epoch": 2.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9298,
|
|
"step": 5470
|
|
},
|
|
{
|
|
"epoch": 2.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9653,
|
|
"step": 5480
|
|
},
|
|
{
|
|
"epoch": 2.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.947,
|
|
"step": 5490
|
|
},
|
|
{
|
|
"epoch": 2.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9438,
|
|
"step": 5500
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9581,
|
|
"step": 5510
|
|
},
|
|
{
|
|
"epoch": 2.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9582,
|
|
"step": 5520
|
|
},
|
|
{
|
|
"epoch": 2.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9202,
|
|
"step": 5530
|
|
},
|
|
{
|
|
"epoch": 2.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9561,
|
|
"step": 5540
|
|
},
|
|
{
|
|
"epoch": 2.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 5550
|
|
},
|
|
{
|
|
"epoch": 2.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 5560
|
|
},
|
|
{
|
|
"epoch": 2.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9328,
|
|
"step": 5570
|
|
},
|
|
{
|
|
"epoch": 2.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 5580
|
|
},
|
|
{
|
|
"epoch": 2.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9488,
|
|
"step": 5590
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 5600
|
|
},
|
|
{
|
|
"epoch": 2.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.959,
|
|
"step": 5610
|
|
},
|
|
{
|
|
"epoch": 2.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9649,
|
|
"step": 5620
|
|
},
|
|
{
|
|
"epoch": 2.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9498,
|
|
"step": 5630
|
|
},
|
|
{
|
|
"epoch": 2.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9353,
|
|
"step": 5640
|
|
},
|
|
{
|
|
"epoch": 2.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 5650
|
|
},
|
|
{
|
|
"epoch": 2.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 5660
|
|
},
|
|
{
|
|
"epoch": 2.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 5670
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 5680
|
|
},
|
|
{
|
|
"epoch": 2.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 5690
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9389,
|
|
"step": 5700
|
|
},
|
|
{
|
|
"epoch": 2.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9575,
|
|
"step": 5710
|
|
},
|
|
{
|
|
"epoch": 2.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9573,
|
|
"step": 5720
|
|
},
|
|
{
|
|
"epoch": 2.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 5730
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 5740
|
|
},
|
|
{
|
|
"epoch": 2.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 5750
|
|
},
|
|
{
|
|
"epoch": 2.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9192,
|
|
"step": 5760
|
|
},
|
|
{
|
|
"epoch": 2.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9526,
|
|
"step": 5770
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9799,
|
|
"step": 5780
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.13618677042801555
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03424657534246575
|
|
},
|
|
"eval_loss": 1.9497852325439453,
|
|
"eval_precision": {
|
|
"precision": 0.019455252918287935
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 77.5559,
|
|
"eval_samples_per_second": 49.706,
|
|
"eval_steps_per_second": 24.859,
|
|
"step": 5781
|
|
},
|
|
{
|
|
"epoch": 3.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9658,
|
|
"step": 5790
|
|
},
|
|
{
|
|
"epoch": 3.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 5800
|
|
},
|
|
{
|
|
"epoch": 3.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 5810
|
|
},
|
|
{
|
|
"epoch": 3.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.957,
|
|
"step": 5820
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 5830
|
|
},
|
|
{
|
|
"epoch": 3.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 5840
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 5850
|
|
},
|
|
{
|
|
"epoch": 3.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 5860
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9464,
|
|
"step": 5870
|
|
},
|
|
{
|
|
"epoch": 3.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9605,
|
|
"step": 5880
|
|
},
|
|
{
|
|
"epoch": 3.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9525,
|
|
"step": 5890
|
|
},
|
|
{
|
|
"epoch": 3.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9427,
|
|
"step": 5900
|
|
},
|
|
{
|
|
"epoch": 3.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9498,
|
|
"step": 5910
|
|
},
|
|
{
|
|
"epoch": 3.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.944,
|
|
"step": 5920
|
|
},
|
|
{
|
|
"epoch": 3.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 5930
|
|
},
|
|
{
|
|
"epoch": 3.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9408,
|
|
"step": 5940
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9169,
|
|
"step": 5950
|
|
},
|
|
{
|
|
"epoch": 3.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9782,
|
|
"step": 5960
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9376,
|
|
"step": 5970
|
|
},
|
|
{
|
|
"epoch": 3.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 5980
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9446,
|
|
"step": 5990
|
|
},
|
|
{
|
|
"epoch": 3.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9434,
|
|
"step": 6000
|
|
},
|
|
{
|
|
"epoch": 3.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9421,
|
|
"step": 6010
|
|
},
|
|
{
|
|
"epoch": 3.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9364,
|
|
"step": 6020
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9587,
|
|
"step": 6030
|
|
},
|
|
{
|
|
"epoch": 3.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9465,
|
|
"step": 6040
|
|
},
|
|
{
|
|
"epoch": 3.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9405,
|
|
"step": 6050
|
|
},
|
|
{
|
|
"epoch": 3.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 6060
|
|
},
|
|
{
|
|
"epoch": 3.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9451,
|
|
"step": 6070
|
|
},
|
|
{
|
|
"epoch": 3.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9602,
|
|
"step": 6080
|
|
},
|
|
{
|
|
"epoch": 3.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9448,
|
|
"step": 6090
|
|
},
|
|
{
|
|
"epoch": 3.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9438,
|
|
"step": 6100
|
|
},
|
|
{
|
|
"epoch": 3.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.935,
|
|
"step": 6110
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9541,
|
|
"step": 6120
|
|
},
|
|
{
|
|
"epoch": 3.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 6130
|
|
},
|
|
{
|
|
"epoch": 3.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9538,
|
|
"step": 6140
|
|
},
|
|
{
|
|
"epoch": 3.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 6150
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9523,
|
|
"step": 6160
|
|
},
|
|
{
|
|
"epoch": 3.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 6170
|
|
},
|
|
{
|
|
"epoch": 3.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9503,
|
|
"step": 6180
|
|
},
|
|
{
|
|
"epoch": 3.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.946,
|
|
"step": 6190
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9468,
|
|
"step": 6200
|
|
},
|
|
{
|
|
"epoch": 3.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9536,
|
|
"step": 6210
|
|
},
|
|
{
|
|
"epoch": 3.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 6220
|
|
},
|
|
{
|
|
"epoch": 3.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 6230
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 6240
|
|
},
|
|
{
|
|
"epoch": 3.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 6250
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9534,
|
|
"step": 6260
|
|
},
|
|
{
|
|
"epoch": 3.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9348,
|
|
"step": 6270
|
|
},
|
|
{
|
|
"epoch": 3.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9276,
|
|
"step": 6280
|
|
},
|
|
{
|
|
"epoch": 3.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 6290
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9499,
|
|
"step": 6300
|
|
},
|
|
{
|
|
"epoch": 3.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9618,
|
|
"step": 6310
|
|
},
|
|
{
|
|
"epoch": 3.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 6320
|
|
},
|
|
{
|
|
"epoch": 3.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 6330
|
|
},
|
|
{
|
|
"epoch": 3.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 6340
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9393,
|
|
"step": 6350
|
|
},
|
|
{
|
|
"epoch": 3.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 6360
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9318,
|
|
"step": 6370
|
|
},
|
|
{
|
|
"epoch": 3.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9617,
|
|
"step": 6380
|
|
},
|
|
{
|
|
"epoch": 3.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9354,
|
|
"step": 6390
|
|
},
|
|
{
|
|
"epoch": 3.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 6400
|
|
},
|
|
{
|
|
"epoch": 3.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9657,
|
|
"step": 6410
|
|
},
|
|
{
|
|
"epoch": 3.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9473,
|
|
"step": 6420
|
|
},
|
|
{
|
|
"epoch": 3.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9572,
|
|
"step": 6430
|
|
},
|
|
{
|
|
"epoch": 3.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.941,
|
|
"step": 6440
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9316,
|
|
"step": 6450
|
|
},
|
|
{
|
|
"epoch": 3.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9525,
|
|
"step": 6460
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 6470
|
|
},
|
|
{
|
|
"epoch": 3.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9664,
|
|
"step": 6480
|
|
},
|
|
{
|
|
"epoch": 3.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9279,
|
|
"step": 6490
|
|
},
|
|
{
|
|
"epoch": 3.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9534,
|
|
"step": 6500
|
|
},
|
|
{
|
|
"epoch": 3.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9162,
|
|
"step": 6510
|
|
},
|
|
{
|
|
"epoch": 3.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9218,
|
|
"step": 6520
|
|
},
|
|
{
|
|
"epoch": 3.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9517,
|
|
"step": 6530
|
|
},
|
|
{
|
|
"epoch": 3.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9523,
|
|
"step": 6540
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9532,
|
|
"step": 6550
|
|
},
|
|
{
|
|
"epoch": 3.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 6560
|
|
},
|
|
{
|
|
"epoch": 3.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9576,
|
|
"step": 6570
|
|
},
|
|
{
|
|
"epoch": 3.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 6580
|
|
},
|
|
{
|
|
"epoch": 3.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.944,
|
|
"step": 6590
|
|
},
|
|
{
|
|
"epoch": 3.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9472,
|
|
"step": 6600
|
|
},
|
|
{
|
|
"epoch": 3.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 6610
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9513,
|
|
"step": 6620
|
|
},
|
|
{
|
|
"epoch": 3.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9519,
|
|
"step": 6630
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9256,
|
|
"step": 6640
|
|
},
|
|
{
|
|
"epoch": 3.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9366,
|
|
"step": 6650
|
|
},
|
|
{
|
|
"epoch": 3.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9464,
|
|
"step": 6660
|
|
},
|
|
{
|
|
"epoch": 3.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9364,
|
|
"step": 6670
|
|
},
|
|
{
|
|
"epoch": 3.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9324,
|
|
"step": 6680
|
|
},
|
|
{
|
|
"epoch": 3.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9474,
|
|
"step": 6690
|
|
},
|
|
{
|
|
"epoch": 3.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9316,
|
|
"step": 6700
|
|
},
|
|
{
|
|
"epoch": 3.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 6710
|
|
},
|
|
{
|
|
"epoch": 3.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9277,
|
|
"step": 6720
|
|
},
|
|
{
|
|
"epoch": 3.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9321,
|
|
"step": 6730
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9355,
|
|
"step": 6740
|
|
},
|
|
{
|
|
"epoch": 3.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9642,
|
|
"step": 6750
|
|
},
|
|
{
|
|
"epoch": 3.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 6760
|
|
},
|
|
{
|
|
"epoch": 3.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 6770
|
|
},
|
|
{
|
|
"epoch": 3.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9563,
|
|
"step": 6780
|
|
},
|
|
{
|
|
"epoch": 3.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 6790
|
|
},
|
|
{
|
|
"epoch": 3.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 6800
|
|
},
|
|
{
|
|
"epoch": 3.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9438,
|
|
"step": 6810
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 6820
|
|
},
|
|
{
|
|
"epoch": 3.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 6830
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 6840
|
|
},
|
|
{
|
|
"epoch": 3.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9433,
|
|
"step": 6850
|
|
},
|
|
{
|
|
"epoch": 3.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9526,
|
|
"step": 6860
|
|
},
|
|
{
|
|
"epoch": 3.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9679,
|
|
"step": 6870
|
|
},
|
|
{
|
|
"epoch": 3.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 6880
|
|
},
|
|
{
|
|
"epoch": 3.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9332,
|
|
"step": 6890
|
|
},
|
|
{
|
|
"epoch": 3.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9424,
|
|
"step": 6900
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9502,
|
|
"step": 6910
|
|
},
|
|
{
|
|
"epoch": 3.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9504,
|
|
"step": 6920
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9308,
|
|
"step": 6930
|
|
},
|
|
{
|
|
"epoch": 3.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 6940
|
|
},
|
|
{
|
|
"epoch": 3.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9349,
|
|
"step": 6950
|
|
},
|
|
{
|
|
"epoch": 3.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 6960
|
|
},
|
|
{
|
|
"epoch": 3.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9487,
|
|
"step": 6970
|
|
},
|
|
{
|
|
"epoch": 3.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9396,
|
|
"step": 6980
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9337,
|
|
"step": 6990
|
|
},
|
|
{
|
|
"epoch": 3.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9433,
|
|
"step": 7000
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.933,
|
|
"step": 7010
|
|
},
|
|
{
|
|
"epoch": 3.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9064,
|
|
"step": 7020
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9386,
|
|
"step": 7030
|
|
},
|
|
{
|
|
"epoch": 3.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9323,
|
|
"step": 7040
|
|
},
|
|
{
|
|
"epoch": 3.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.946,
|
|
"step": 7050
|
|
},
|
|
{
|
|
"epoch": 3.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9297,
|
|
"step": 7060
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.8934,
|
|
"step": 7070
|
|
},
|
|
{
|
|
"epoch": 3.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9523,
|
|
"step": 7080
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9222,
|
|
"step": 7090
|
|
},
|
|
{
|
|
"epoch": 3.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9464,
|
|
"step": 7100
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 7110
|
|
},
|
|
{
|
|
"epoch": 3.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9425,
|
|
"step": 7120
|
|
},
|
|
{
|
|
"epoch": 3.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9329,
|
|
"step": 7130
|
|
},
|
|
{
|
|
"epoch": 3.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9393,
|
|
"step": 7140
|
|
},
|
|
{
|
|
"epoch": 3.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 7150
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9641,
|
|
"step": 7160
|
|
},
|
|
{
|
|
"epoch": 3.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9414,
|
|
"step": 7170
|
|
},
|
|
{
|
|
"epoch": 3.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9574,
|
|
"step": 7180
|
|
},
|
|
{
|
|
"epoch": 3.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9491,
|
|
"step": 7190
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9342,
|
|
"step": 7200
|
|
},
|
|
{
|
|
"epoch": 3.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9548,
|
|
"step": 7210
|
|
},
|
|
{
|
|
"epoch": 3.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9333,
|
|
"step": 7220
|
|
},
|
|
{
|
|
"epoch": 3.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9731,
|
|
"step": 7230
|
|
},
|
|
{
|
|
"epoch": 3.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9384,
|
|
"step": 7240
|
|
},
|
|
{
|
|
"epoch": 3.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 7250
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 7260
|
|
},
|
|
{
|
|
"epoch": 3.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 7270
|
|
},
|
|
{
|
|
"epoch": 3.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9445,
|
|
"step": 7280
|
|
},
|
|
{
|
|
"epoch": 3.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9498,
|
|
"step": 7290
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9376,
|
|
"step": 7300
|
|
},
|
|
{
|
|
"epoch": 3.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9477,
|
|
"step": 7310
|
|
},
|
|
{
|
|
"epoch": 3.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.931,
|
|
"step": 7320
|
|
},
|
|
{
|
|
"epoch": 3.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9003,
|
|
"step": 7330
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 7340
|
|
},
|
|
{
|
|
"epoch": 3.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9369,
|
|
"step": 7350
|
|
},
|
|
{
|
|
"epoch": 3.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9529,
|
|
"step": 7360
|
|
},
|
|
{
|
|
"epoch": 3.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.946,
|
|
"step": 7370
|
|
},
|
|
{
|
|
"epoch": 3.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9339,
|
|
"step": 7380
|
|
},
|
|
{
|
|
"epoch": 3.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 7390
|
|
},
|
|
{
|
|
"epoch": 3.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9303,
|
|
"step": 7400
|
|
},
|
|
{
|
|
"epoch": 3.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.967,
|
|
"step": 7410
|
|
},
|
|
{
|
|
"epoch": 3.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9406,
|
|
"step": 7420
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9433,
|
|
"step": 7430
|
|
},
|
|
{
|
|
"epoch": 3.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 7440
|
|
},
|
|
{
|
|
"epoch": 3.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9309,
|
|
"step": 7450
|
|
},
|
|
{
|
|
"epoch": 3.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 7460
|
|
},
|
|
{
|
|
"epoch": 3.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 7470
|
|
},
|
|
{
|
|
"epoch": 3.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 7480
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9622,
|
|
"step": 7490
|
|
},
|
|
{
|
|
"epoch": 3.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9342,
|
|
"step": 7500
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9044,
|
|
"step": 7510
|
|
},
|
|
{
|
|
"epoch": 3.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 7520
|
|
},
|
|
{
|
|
"epoch": 3.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9608,
|
|
"step": 7530
|
|
},
|
|
{
|
|
"epoch": 3.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9298,
|
|
"step": 7540
|
|
},
|
|
{
|
|
"epoch": 3.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9511,
|
|
"step": 7550
|
|
},
|
|
{
|
|
"epoch": 3.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9408,
|
|
"step": 7560
|
|
},
|
|
{
|
|
"epoch": 3.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9408,
|
|
"step": 7570
|
|
},
|
|
{
|
|
"epoch": 3.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9561,
|
|
"step": 7580
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 7590
|
|
},
|
|
{
|
|
"epoch": 3.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9429,
|
|
"step": 7600
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9495,
|
|
"step": 7610
|
|
},
|
|
{
|
|
"epoch": 3.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 7620
|
|
},
|
|
{
|
|
"epoch": 3.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 7630
|
|
},
|
|
{
|
|
"epoch": 3.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 7640
|
|
},
|
|
{
|
|
"epoch": 3.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 7650
|
|
},
|
|
{
|
|
"epoch": 3.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9408,
|
|
"step": 7660
|
|
},
|
|
{
|
|
"epoch": 3.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9428,
|
|
"step": 7670
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9222,
|
|
"step": 7680
|
|
},
|
|
{
|
|
"epoch": 3.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9615,
|
|
"step": 7690
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9458,
|
|
"step": 7700
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.17665369649805449
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.04289493575207861
|
|
},
|
|
"eval_loss": 1.9412175416946411,
|
|
"eval_precision": {
|
|
"precision": 0.025236242356864926
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 77.9759,
|
|
"eval_samples_per_second": 49.438,
|
|
"eval_steps_per_second": 24.726,
|
|
"step": 7708
|
|
},
|
|
{
|
|
"epoch": 4.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 7710
|
|
},
|
|
{
|
|
"epoch": 4.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9409,
|
|
"step": 7720
|
|
},
|
|
{
|
|
"epoch": 4.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9472,
|
|
"step": 7730
|
|
},
|
|
{
|
|
"epoch": 4.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 7740
|
|
},
|
|
{
|
|
"epoch": 4.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9466,
|
|
"step": 7750
|
|
},
|
|
{
|
|
"epoch": 4.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9387,
|
|
"step": 7760
|
|
},
|
|
{
|
|
"epoch": 4.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 7770
|
|
},
|
|
{
|
|
"epoch": 4.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9473,
|
|
"step": 7780
|
|
},
|
|
{
|
|
"epoch": 4.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9302,
|
|
"step": 7790
|
|
},
|
|
{
|
|
"epoch": 4.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9263,
|
|
"step": 7800
|
|
},
|
|
{
|
|
"epoch": 4.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9636,
|
|
"step": 7810
|
|
},
|
|
{
|
|
"epoch": 4.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9634,
|
|
"step": 7820
|
|
},
|
|
{
|
|
"epoch": 4.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9548,
|
|
"step": 7830
|
|
},
|
|
{
|
|
"epoch": 4.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9345,
|
|
"step": 7840
|
|
},
|
|
{
|
|
"epoch": 4.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 7850
|
|
},
|
|
{
|
|
"epoch": 4.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9491,
|
|
"step": 7860
|
|
},
|
|
{
|
|
"epoch": 4.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9305,
|
|
"step": 7870
|
|
},
|
|
{
|
|
"epoch": 4.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 7880
|
|
},
|
|
{
|
|
"epoch": 4.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9428,
|
|
"step": 7890
|
|
},
|
|
{
|
|
"epoch": 4.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.929,
|
|
"step": 7900
|
|
},
|
|
{
|
|
"epoch": 4.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9336,
|
|
"step": 7910
|
|
},
|
|
{
|
|
"epoch": 4.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 7920
|
|
},
|
|
{
|
|
"epoch": 4.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9622,
|
|
"step": 7930
|
|
},
|
|
{
|
|
"epoch": 4.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 7940
|
|
},
|
|
{
|
|
"epoch": 4.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9243,
|
|
"step": 7950
|
|
},
|
|
{
|
|
"epoch": 4.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 7960
|
|
},
|
|
{
|
|
"epoch": 4.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9542,
|
|
"step": 7970
|
|
},
|
|
{
|
|
"epoch": 4.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9191,
|
|
"step": 7980
|
|
},
|
|
{
|
|
"epoch": 4.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9585,
|
|
"step": 7990
|
|
},
|
|
{
|
|
"epoch": 4.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9433,
|
|
"step": 8000
|
|
},
|
|
{
|
|
"epoch": 4.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9774,
|
|
"step": 8010
|
|
},
|
|
{
|
|
"epoch": 4.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9324,
|
|
"step": 8020
|
|
},
|
|
{
|
|
"epoch": 4.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9455,
|
|
"step": 8030
|
|
},
|
|
{
|
|
"epoch": 4.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 8040
|
|
},
|
|
{
|
|
"epoch": 4.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9473,
|
|
"step": 8050
|
|
},
|
|
{
|
|
"epoch": 4.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9287,
|
|
"step": 8060
|
|
},
|
|
{
|
|
"epoch": 4.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9451,
|
|
"step": 8070
|
|
},
|
|
{
|
|
"epoch": 4.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9786,
|
|
"step": 8080
|
|
},
|
|
{
|
|
"epoch": 4.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9257,
|
|
"step": 8090
|
|
},
|
|
{
|
|
"epoch": 4.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9269,
|
|
"step": 8100
|
|
},
|
|
{
|
|
"epoch": 4.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9279,
|
|
"step": 8110
|
|
},
|
|
{
|
|
"epoch": 4.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9575,
|
|
"step": 8120
|
|
},
|
|
{
|
|
"epoch": 4.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9488,
|
|
"step": 8130
|
|
},
|
|
{
|
|
"epoch": 4.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 8140
|
|
},
|
|
{
|
|
"epoch": 4.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 8150
|
|
},
|
|
{
|
|
"epoch": 4.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9317,
|
|
"step": 8160
|
|
},
|
|
{
|
|
"epoch": 4.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9328,
|
|
"step": 8170
|
|
},
|
|
{
|
|
"epoch": 4.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9304,
|
|
"step": 8180
|
|
},
|
|
{
|
|
"epoch": 4.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9382,
|
|
"step": 8190
|
|
},
|
|
{
|
|
"epoch": 4.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9339,
|
|
"step": 8200
|
|
},
|
|
{
|
|
"epoch": 4.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9357,
|
|
"step": 8210
|
|
},
|
|
{
|
|
"epoch": 4.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9401,
|
|
"step": 8220
|
|
},
|
|
{
|
|
"epoch": 4.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9633,
|
|
"step": 8230
|
|
},
|
|
{
|
|
"epoch": 4.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9443,
|
|
"step": 8240
|
|
},
|
|
{
|
|
"epoch": 4.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9585,
|
|
"step": 8250
|
|
},
|
|
{
|
|
"epoch": 4.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.954,
|
|
"step": 8260
|
|
},
|
|
{
|
|
"epoch": 4.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 8270
|
|
},
|
|
{
|
|
"epoch": 4.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 8280
|
|
},
|
|
{
|
|
"epoch": 4.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 8290
|
|
},
|
|
{
|
|
"epoch": 4.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9369,
|
|
"step": 8300
|
|
},
|
|
{
|
|
"epoch": 4.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9428,
|
|
"step": 8310
|
|
},
|
|
{
|
|
"epoch": 4.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9492,
|
|
"step": 8320
|
|
},
|
|
{
|
|
"epoch": 4.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 8330
|
|
},
|
|
{
|
|
"epoch": 4.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9559,
|
|
"step": 8340
|
|
},
|
|
{
|
|
"epoch": 4.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9503,
|
|
"step": 8350
|
|
},
|
|
{
|
|
"epoch": 4.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9615,
|
|
"step": 8360
|
|
},
|
|
{
|
|
"epoch": 4.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9387,
|
|
"step": 8370
|
|
},
|
|
{
|
|
"epoch": 4.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 8380
|
|
},
|
|
{
|
|
"epoch": 4.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.947,
|
|
"step": 8390
|
|
},
|
|
{
|
|
"epoch": 4.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9563,
|
|
"step": 8400
|
|
},
|
|
{
|
|
"epoch": 4.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9457,
|
|
"step": 8410
|
|
},
|
|
{
|
|
"epoch": 4.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9399,
|
|
"step": 8420
|
|
},
|
|
{
|
|
"epoch": 4.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9534,
|
|
"step": 8430
|
|
},
|
|
{
|
|
"epoch": 4.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.93,
|
|
"step": 8440
|
|
},
|
|
{
|
|
"epoch": 4.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9513,
|
|
"step": 8450
|
|
},
|
|
{
|
|
"epoch": 4.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 8460
|
|
},
|
|
{
|
|
"epoch": 4.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9232,
|
|
"step": 8470
|
|
},
|
|
{
|
|
"epoch": 4.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 8480
|
|
},
|
|
{
|
|
"epoch": 4.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9544,
|
|
"step": 8490
|
|
},
|
|
{
|
|
"epoch": 4.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 8500
|
|
},
|
|
{
|
|
"epoch": 4.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 8510
|
|
},
|
|
{
|
|
"epoch": 4.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9432,
|
|
"step": 8520
|
|
},
|
|
{
|
|
"epoch": 4.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9374,
|
|
"step": 8530
|
|
},
|
|
{
|
|
"epoch": 4.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 8540
|
|
},
|
|
{
|
|
"epoch": 4.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 8550
|
|
},
|
|
{
|
|
"epoch": 4.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9077,
|
|
"step": 8560
|
|
},
|
|
{
|
|
"epoch": 4.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9779,
|
|
"step": 8570
|
|
},
|
|
{
|
|
"epoch": 4.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9489,
|
|
"step": 8580
|
|
},
|
|
{
|
|
"epoch": 4.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9406,
|
|
"step": 8590
|
|
},
|
|
{
|
|
"epoch": 4.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9254,
|
|
"step": 8600
|
|
},
|
|
{
|
|
"epoch": 4.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 8610
|
|
},
|
|
{
|
|
"epoch": 4.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.946,
|
|
"step": 8620
|
|
},
|
|
{
|
|
"epoch": 4.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 8630
|
|
},
|
|
{
|
|
"epoch": 4.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9333,
|
|
"step": 8640
|
|
},
|
|
{
|
|
"epoch": 4.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9384,
|
|
"step": 8650
|
|
},
|
|
{
|
|
"epoch": 4.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9443,
|
|
"step": 8660
|
|
},
|
|
{
|
|
"epoch": 4.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.962,
|
|
"step": 8670
|
|
},
|
|
{
|
|
"epoch": 4.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.958,
|
|
"step": 8680
|
|
},
|
|
{
|
|
"epoch": 4.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9449,
|
|
"step": 8690
|
|
},
|
|
{
|
|
"epoch": 4.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 8700
|
|
},
|
|
{
|
|
"epoch": 4.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9468,
|
|
"step": 8710
|
|
},
|
|
{
|
|
"epoch": 4.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9566,
|
|
"step": 8720
|
|
},
|
|
{
|
|
"epoch": 4.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 8730
|
|
},
|
|
{
|
|
"epoch": 4.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 8740
|
|
},
|
|
{
|
|
"epoch": 4.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9565,
|
|
"step": 8750
|
|
},
|
|
{
|
|
"epoch": 4.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9292,
|
|
"step": 8760
|
|
},
|
|
{
|
|
"epoch": 4.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9345,
|
|
"step": 8770
|
|
},
|
|
{
|
|
"epoch": 4.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9686,
|
|
"step": 8780
|
|
},
|
|
{
|
|
"epoch": 4.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9181,
|
|
"step": 8790
|
|
},
|
|
{
|
|
"epoch": 4.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9606,
|
|
"step": 8800
|
|
},
|
|
{
|
|
"epoch": 4.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.937,
|
|
"step": 8810
|
|
},
|
|
{
|
|
"epoch": 4.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9302,
|
|
"step": 8820
|
|
},
|
|
{
|
|
"epoch": 4.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9756,
|
|
"step": 8830
|
|
},
|
|
{
|
|
"epoch": 4.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9396,
|
|
"step": 8840
|
|
},
|
|
{
|
|
"epoch": 4.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 8850
|
|
},
|
|
{
|
|
"epoch": 4.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.962,
|
|
"step": 8860
|
|
},
|
|
{
|
|
"epoch": 4.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9464,
|
|
"step": 8870
|
|
},
|
|
{
|
|
"epoch": 4.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 8880
|
|
},
|
|
{
|
|
"epoch": 4.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9495,
|
|
"step": 8890
|
|
},
|
|
{
|
|
"epoch": 4.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 8900
|
|
},
|
|
{
|
|
"epoch": 4.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9363,
|
|
"step": 8910
|
|
},
|
|
{
|
|
"epoch": 4.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9567,
|
|
"step": 8920
|
|
},
|
|
{
|
|
"epoch": 4.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9389,
|
|
"step": 8930
|
|
},
|
|
{
|
|
"epoch": 4.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9319,
|
|
"step": 8940
|
|
},
|
|
{
|
|
"epoch": 4.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 8950
|
|
},
|
|
{
|
|
"epoch": 4.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.908,
|
|
"step": 8960
|
|
},
|
|
{
|
|
"epoch": 4.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9273,
|
|
"step": 8970
|
|
},
|
|
{
|
|
"epoch": 4.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 8980
|
|
},
|
|
{
|
|
"epoch": 4.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9577,
|
|
"step": 8990
|
|
},
|
|
{
|
|
"epoch": 4.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 9000
|
|
},
|
|
{
|
|
"epoch": 4.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9585,
|
|
"step": 9010
|
|
},
|
|
{
|
|
"epoch": 4.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 9020
|
|
},
|
|
{
|
|
"epoch": 4.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9484,
|
|
"step": 9030
|
|
},
|
|
{
|
|
"epoch": 4.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9468,
|
|
"step": 9040
|
|
},
|
|
{
|
|
"epoch": 4.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9492,
|
|
"step": 9050
|
|
},
|
|
{
|
|
"epoch": 4.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9426,
|
|
"step": 9060
|
|
},
|
|
{
|
|
"epoch": 4.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 9070
|
|
},
|
|
{
|
|
"epoch": 4.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9491,
|
|
"step": 9080
|
|
},
|
|
{
|
|
"epoch": 4.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9539,
|
|
"step": 9090
|
|
},
|
|
{
|
|
"epoch": 4.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 9100
|
|
},
|
|
{
|
|
"epoch": 4.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 9110
|
|
},
|
|
{
|
|
"epoch": 4.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 9120
|
|
},
|
|
{
|
|
"epoch": 4.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9438,
|
|
"step": 9130
|
|
},
|
|
{
|
|
"epoch": 4.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 9140
|
|
},
|
|
{
|
|
"epoch": 4.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 9150
|
|
},
|
|
{
|
|
"epoch": 4.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9393,
|
|
"step": 9160
|
|
},
|
|
{
|
|
"epoch": 4.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 9170
|
|
},
|
|
{
|
|
"epoch": 4.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 9180
|
|
},
|
|
{
|
|
"epoch": 4.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 9190
|
|
},
|
|
{
|
|
"epoch": 4.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9312,
|
|
"step": 9200
|
|
},
|
|
{
|
|
"epoch": 4.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9432,
|
|
"step": 9210
|
|
},
|
|
{
|
|
"epoch": 4.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 9220
|
|
},
|
|
{
|
|
"epoch": 4.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 9230
|
|
},
|
|
{
|
|
"epoch": 4.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 9240
|
|
},
|
|
{
|
|
"epoch": 4.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9497,
|
|
"step": 9250
|
|
},
|
|
{
|
|
"epoch": 4.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9563,
|
|
"step": 9260
|
|
},
|
|
{
|
|
"epoch": 4.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 9270
|
|
},
|
|
{
|
|
"epoch": 4.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9485,
|
|
"step": 9280
|
|
},
|
|
{
|
|
"epoch": 4.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 9290
|
|
},
|
|
{
|
|
"epoch": 4.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9426,
|
|
"step": 9300
|
|
},
|
|
{
|
|
"epoch": 4.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 9310
|
|
},
|
|
{
|
|
"epoch": 4.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9337,
|
|
"step": 9320
|
|
},
|
|
{
|
|
"epoch": 4.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 9330
|
|
},
|
|
{
|
|
"epoch": 4.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9385,
|
|
"step": 9340
|
|
},
|
|
{
|
|
"epoch": 4.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9565,
|
|
"step": 9350
|
|
},
|
|
{
|
|
"epoch": 4.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9384,
|
|
"step": 9360
|
|
},
|
|
{
|
|
"epoch": 4.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 9370
|
|
},
|
|
{
|
|
"epoch": 4.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 9380
|
|
},
|
|
{
|
|
"epoch": 4.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9125,
|
|
"step": 9390
|
|
},
|
|
{
|
|
"epoch": 4.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9585,
|
|
"step": 9400
|
|
},
|
|
{
|
|
"epoch": 4.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 9410
|
|
},
|
|
{
|
|
"epoch": 4.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9282,
|
|
"step": 9420
|
|
},
|
|
{
|
|
"epoch": 4.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 9430
|
|
},
|
|
{
|
|
"epoch": 4.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9634,
|
|
"step": 9440
|
|
},
|
|
{
|
|
"epoch": 4.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9432,
|
|
"step": 9450
|
|
},
|
|
{
|
|
"epoch": 4.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 9460
|
|
},
|
|
{
|
|
"epoch": 4.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9443,
|
|
"step": 9470
|
|
},
|
|
{
|
|
"epoch": 4.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 9480
|
|
},
|
|
{
|
|
"epoch": 4.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 9490
|
|
},
|
|
{
|
|
"epoch": 4.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9465,
|
|
"step": 9500
|
|
},
|
|
{
|
|
"epoch": 4.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 9510
|
|
},
|
|
{
|
|
"epoch": 4.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9322,
|
|
"step": 9520
|
|
},
|
|
{
|
|
"epoch": 4.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9429,
|
|
"step": 9530
|
|
},
|
|
{
|
|
"epoch": 4.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9294,
|
|
"step": 9540
|
|
},
|
|
{
|
|
"epoch": 4.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.966,
|
|
"step": 9550
|
|
},
|
|
{
|
|
"epoch": 4.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 9560
|
|
},
|
|
{
|
|
"epoch": 4.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9477,
|
|
"step": 9570
|
|
},
|
|
{
|
|
"epoch": 4.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 9580
|
|
},
|
|
{
|
|
"epoch": 4.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9563,
|
|
"step": 9590
|
|
},
|
|
{
|
|
"epoch": 4.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9393,
|
|
"step": 9600
|
|
},
|
|
{
|
|
"epoch": 4.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 9610
|
|
},
|
|
{
|
|
"epoch": 4.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9558,
|
|
"step": 9620
|
|
},
|
|
{
|
|
"epoch": 5.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 9630
|
|
},
|
|
{
|
|
"epoch": 5.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.17665369649805449
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.04289493575207861
|
|
},
|
|
"eval_loss": 1.9407862424850464,
|
|
"eval_precision": {
|
|
"precision": 0.025236242356864926
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 76.9258,
|
|
"eval_samples_per_second": 50.113,
|
|
"eval_steps_per_second": 25.063,
|
|
"step": 9635
|
|
},
|
|
{
|
|
"epoch": 5.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 9640
|
|
},
|
|
{
|
|
"epoch": 5.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9466,
|
|
"step": 9650
|
|
},
|
|
{
|
|
"epoch": 5.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9411,
|
|
"step": 9660
|
|
},
|
|
{
|
|
"epoch": 5.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9391,
|
|
"step": 9670
|
|
},
|
|
{
|
|
"epoch": 5.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9364,
|
|
"step": 9680
|
|
},
|
|
{
|
|
"epoch": 5.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9595,
|
|
"step": 9690
|
|
},
|
|
{
|
|
"epoch": 5.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 9700
|
|
},
|
|
{
|
|
"epoch": 5.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 9710
|
|
},
|
|
{
|
|
"epoch": 5.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9405,
|
|
"step": 9720
|
|
},
|
|
{
|
|
"epoch": 5.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9598,
|
|
"step": 9730
|
|
},
|
|
{
|
|
"epoch": 5.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9447,
|
|
"step": 9740
|
|
},
|
|
{
|
|
"epoch": 5.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9469,
|
|
"step": 9750
|
|
},
|
|
{
|
|
"epoch": 5.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.936,
|
|
"step": 9760
|
|
},
|
|
{
|
|
"epoch": 5.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9361,
|
|
"step": 9770
|
|
},
|
|
{
|
|
"epoch": 5.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9465,
|
|
"step": 9780
|
|
},
|
|
{
|
|
"epoch": 5.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9325,
|
|
"step": 9790
|
|
},
|
|
{
|
|
"epoch": 5.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 9800
|
|
},
|
|
{
|
|
"epoch": 5.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 9810
|
|
},
|
|
{
|
|
"epoch": 5.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9248,
|
|
"step": 9820
|
|
},
|
|
{
|
|
"epoch": 5.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9434,
|
|
"step": 9830
|
|
},
|
|
{
|
|
"epoch": 5.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9291,
|
|
"step": 9840
|
|
},
|
|
{
|
|
"epoch": 5.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 9850
|
|
},
|
|
{
|
|
"epoch": 5.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9049,
|
|
"step": 9860
|
|
},
|
|
{
|
|
"epoch": 5.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 9870
|
|
},
|
|
{
|
|
"epoch": 5.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 9880
|
|
},
|
|
{
|
|
"epoch": 5.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9415,
|
|
"step": 9890
|
|
},
|
|
{
|
|
"epoch": 5.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9756,
|
|
"step": 9900
|
|
},
|
|
{
|
|
"epoch": 5.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9386,
|
|
"step": 9910
|
|
},
|
|
{
|
|
"epoch": 5.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9301,
|
|
"step": 9920
|
|
},
|
|
{
|
|
"epoch": 5.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 9930
|
|
},
|
|
{
|
|
"epoch": 5.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9548,
|
|
"step": 9940
|
|
},
|
|
{
|
|
"epoch": 5.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9657,
|
|
"step": 9950
|
|
},
|
|
{
|
|
"epoch": 5.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.945,
|
|
"step": 9960
|
|
},
|
|
{
|
|
"epoch": 5.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9526,
|
|
"step": 9970
|
|
},
|
|
{
|
|
"epoch": 5.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9477,
|
|
"step": 9980
|
|
},
|
|
{
|
|
"epoch": 5.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.952,
|
|
"step": 9990
|
|
},
|
|
{
|
|
"epoch": 5.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 10000
|
|
},
|
|
{
|
|
"epoch": 5.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 10010
|
|
},
|
|
{
|
|
"epoch": 5.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9458,
|
|
"step": 10020
|
|
},
|
|
{
|
|
"epoch": 5.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 10030
|
|
},
|
|
{
|
|
"epoch": 5.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 10040
|
|
},
|
|
{
|
|
"epoch": 5.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 10050
|
|
},
|
|
{
|
|
"epoch": 5.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 10060
|
|
},
|
|
{
|
|
"epoch": 5.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 10070
|
|
},
|
|
{
|
|
"epoch": 5.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 10080
|
|
},
|
|
{
|
|
"epoch": 5.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 10090
|
|
},
|
|
{
|
|
"epoch": 5.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9483,
|
|
"step": 10100
|
|
},
|
|
{
|
|
"epoch": 5.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9278,
|
|
"step": 10110
|
|
},
|
|
{
|
|
"epoch": 5.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9271,
|
|
"step": 10120
|
|
},
|
|
{
|
|
"epoch": 5.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 10130
|
|
},
|
|
{
|
|
"epoch": 5.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9341,
|
|
"step": 10140
|
|
},
|
|
{
|
|
"epoch": 5.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 10150
|
|
},
|
|
{
|
|
"epoch": 5.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.918,
|
|
"step": 10160
|
|
},
|
|
{
|
|
"epoch": 5.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9338,
|
|
"step": 10170
|
|
},
|
|
{
|
|
"epoch": 5.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 10180
|
|
},
|
|
{
|
|
"epoch": 5.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9636,
|
|
"step": 10190
|
|
},
|
|
{
|
|
"epoch": 5.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9425,
|
|
"step": 10200
|
|
},
|
|
{
|
|
"epoch": 5.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9239,
|
|
"step": 10210
|
|
},
|
|
{
|
|
"epoch": 5.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 10220
|
|
},
|
|
{
|
|
"epoch": 5.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9801,
|
|
"step": 10230
|
|
},
|
|
{
|
|
"epoch": 5.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9297,
|
|
"step": 10240
|
|
},
|
|
{
|
|
"epoch": 5.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9544,
|
|
"step": 10250
|
|
},
|
|
{
|
|
"epoch": 5.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 10260
|
|
},
|
|
{
|
|
"epoch": 5.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9295,
|
|
"step": 10270
|
|
},
|
|
{
|
|
"epoch": 5.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 10280
|
|
},
|
|
{
|
|
"epoch": 5.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9489,
|
|
"step": 10290
|
|
},
|
|
{
|
|
"epoch": 5.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9641,
|
|
"step": 10300
|
|
},
|
|
{
|
|
"epoch": 5.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 10310
|
|
},
|
|
{
|
|
"epoch": 5.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.925,
|
|
"step": 10320
|
|
},
|
|
{
|
|
"epoch": 5.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9391,
|
|
"step": 10330
|
|
},
|
|
{
|
|
"epoch": 5.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9426,
|
|
"step": 10340
|
|
},
|
|
{
|
|
"epoch": 5.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.951,
|
|
"step": 10350
|
|
},
|
|
{
|
|
"epoch": 5.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.955,
|
|
"step": 10360
|
|
},
|
|
{
|
|
"epoch": 5.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9401,
|
|
"step": 10370
|
|
},
|
|
{
|
|
"epoch": 5.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 10380
|
|
},
|
|
{
|
|
"epoch": 5.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9264,
|
|
"step": 10390
|
|
},
|
|
{
|
|
"epoch": 5.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9397,
|
|
"step": 10400
|
|
},
|
|
{
|
|
"epoch": 5.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9305,
|
|
"step": 10410
|
|
},
|
|
{
|
|
"epoch": 5.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9285,
|
|
"step": 10420
|
|
},
|
|
{
|
|
"epoch": 5.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9221,
|
|
"step": 10430
|
|
},
|
|
{
|
|
"epoch": 5.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 10440
|
|
},
|
|
{
|
|
"epoch": 5.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9425,
|
|
"step": 10450
|
|
},
|
|
{
|
|
"epoch": 5.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9596,
|
|
"step": 10460
|
|
},
|
|
{
|
|
"epoch": 5.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 10470
|
|
},
|
|
{
|
|
"epoch": 5.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9427,
|
|
"step": 10480
|
|
},
|
|
{
|
|
"epoch": 5.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 10490
|
|
},
|
|
{
|
|
"epoch": 5.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9427,
|
|
"step": 10500
|
|
},
|
|
{
|
|
"epoch": 5.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9405,
|
|
"step": 10510
|
|
},
|
|
{
|
|
"epoch": 5.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9321,
|
|
"step": 10520
|
|
},
|
|
{
|
|
"epoch": 5.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 10530
|
|
},
|
|
{
|
|
"epoch": 5.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 10540
|
|
},
|
|
{
|
|
"epoch": 5.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9516,
|
|
"step": 10550
|
|
},
|
|
{
|
|
"epoch": 5.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9371,
|
|
"step": 10560
|
|
},
|
|
{
|
|
"epoch": 5.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9446,
|
|
"step": 10570
|
|
},
|
|
{
|
|
"epoch": 5.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9374,
|
|
"step": 10580
|
|
},
|
|
{
|
|
"epoch": 5.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9682,
|
|
"step": 10590
|
|
},
|
|
{
|
|
"epoch": 5.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 10600
|
|
},
|
|
{
|
|
"epoch": 5.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9424,
|
|
"step": 10610
|
|
},
|
|
{
|
|
"epoch": 5.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9497,
|
|
"step": 10620
|
|
},
|
|
{
|
|
"epoch": 5.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 10630
|
|
},
|
|
{
|
|
"epoch": 5.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9715,
|
|
"step": 10640
|
|
},
|
|
{
|
|
"epoch": 5.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9503,
|
|
"step": 10650
|
|
},
|
|
{
|
|
"epoch": 5.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9623,
|
|
"step": 10660
|
|
},
|
|
{
|
|
"epoch": 5.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9432,
|
|
"step": 10670
|
|
},
|
|
{
|
|
"epoch": 5.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 10680
|
|
},
|
|
{
|
|
"epoch": 5.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9645,
|
|
"step": 10690
|
|
},
|
|
{
|
|
"epoch": 5.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 10700
|
|
},
|
|
{
|
|
"epoch": 5.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 10710
|
|
},
|
|
{
|
|
"epoch": 5.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9347,
|
|
"step": 10720
|
|
},
|
|
{
|
|
"epoch": 5.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9674,
|
|
"step": 10730
|
|
},
|
|
{
|
|
"epoch": 5.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 10740
|
|
},
|
|
{
|
|
"epoch": 5.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9487,
|
|
"step": 10750
|
|
},
|
|
{
|
|
"epoch": 5.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9223,
|
|
"step": 10760
|
|
},
|
|
{
|
|
"epoch": 5.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9382,
|
|
"step": 10770
|
|
},
|
|
{
|
|
"epoch": 5.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9707,
|
|
"step": 10780
|
|
},
|
|
{
|
|
"epoch": 5.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9595,
|
|
"step": 10790
|
|
},
|
|
{
|
|
"epoch": 5.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9353,
|
|
"step": 10800
|
|
},
|
|
{
|
|
"epoch": 5.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9548,
|
|
"step": 10810
|
|
},
|
|
{
|
|
"epoch": 5.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 10820
|
|
},
|
|
{
|
|
"epoch": 5.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9371,
|
|
"step": 10830
|
|
},
|
|
{
|
|
"epoch": 5.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9316,
|
|
"step": 10840
|
|
},
|
|
{
|
|
"epoch": 5.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9244,
|
|
"step": 10850
|
|
},
|
|
{
|
|
"epoch": 5.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9322,
|
|
"step": 10860
|
|
},
|
|
{
|
|
"epoch": 5.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9484,
|
|
"step": 10870
|
|
},
|
|
{
|
|
"epoch": 5.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9485,
|
|
"step": 10880
|
|
},
|
|
{
|
|
"epoch": 5.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9679,
|
|
"step": 10890
|
|
},
|
|
{
|
|
"epoch": 5.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9401,
|
|
"step": 10900
|
|
},
|
|
{
|
|
"epoch": 5.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9439,
|
|
"step": 10910
|
|
},
|
|
{
|
|
"epoch": 5.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 10920
|
|
},
|
|
{
|
|
"epoch": 5.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 10930
|
|
},
|
|
{
|
|
"epoch": 5.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9587,
|
|
"step": 10940
|
|
},
|
|
{
|
|
"epoch": 5.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9387,
|
|
"step": 10950
|
|
},
|
|
{
|
|
"epoch": 5.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.959,
|
|
"step": 10960
|
|
},
|
|
{
|
|
"epoch": 5.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9449,
|
|
"step": 10970
|
|
},
|
|
{
|
|
"epoch": 5.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 10980
|
|
},
|
|
{
|
|
"epoch": 5.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 10990
|
|
},
|
|
{
|
|
"epoch": 5.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 11000
|
|
},
|
|
{
|
|
"epoch": 5.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9562,
|
|
"step": 11010
|
|
},
|
|
{
|
|
"epoch": 5.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9549,
|
|
"step": 11020
|
|
},
|
|
{
|
|
"epoch": 5.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9322,
|
|
"step": 11030
|
|
},
|
|
{
|
|
"epoch": 5.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9551,
|
|
"step": 11040
|
|
},
|
|
{
|
|
"epoch": 5.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 11050
|
|
},
|
|
{
|
|
"epoch": 5.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.951,
|
|
"step": 11060
|
|
},
|
|
{
|
|
"epoch": 5.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9303,
|
|
"step": 11070
|
|
},
|
|
{
|
|
"epoch": 5.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9343,
|
|
"step": 11080
|
|
},
|
|
{
|
|
"epoch": 5.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9496,
|
|
"step": 11090
|
|
},
|
|
{
|
|
"epoch": 5.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 11100
|
|
},
|
|
{
|
|
"epoch": 5.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9563,
|
|
"step": 11110
|
|
},
|
|
{
|
|
"epoch": 5.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9291,
|
|
"step": 11120
|
|
},
|
|
{
|
|
"epoch": 5.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9516,
|
|
"step": 11130
|
|
},
|
|
{
|
|
"epoch": 5.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9541,
|
|
"step": 11140
|
|
},
|
|
{
|
|
"epoch": 5.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9448,
|
|
"step": 11150
|
|
},
|
|
{
|
|
"epoch": 5.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 11160
|
|
},
|
|
{
|
|
"epoch": 5.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9448,
|
|
"step": 11170
|
|
},
|
|
{
|
|
"epoch": 5.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 11180
|
|
},
|
|
{
|
|
"epoch": 5.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 11190
|
|
},
|
|
{
|
|
"epoch": 5.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9425,
|
|
"step": 11200
|
|
},
|
|
{
|
|
"epoch": 5.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9495,
|
|
"step": 11210
|
|
},
|
|
{
|
|
"epoch": 5.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9299,
|
|
"step": 11220
|
|
},
|
|
{
|
|
"epoch": 5.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 11230
|
|
},
|
|
{
|
|
"epoch": 5.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9585,
|
|
"step": 11240
|
|
},
|
|
{
|
|
"epoch": 5.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9438,
|
|
"step": 11250
|
|
},
|
|
{
|
|
"epoch": 5.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9355,
|
|
"step": 11260
|
|
},
|
|
{
|
|
"epoch": 5.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9471,
|
|
"step": 11270
|
|
},
|
|
{
|
|
"epoch": 5.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9193,
|
|
"step": 11280
|
|
},
|
|
{
|
|
"epoch": 5.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9572,
|
|
"step": 11290
|
|
},
|
|
{
|
|
"epoch": 5.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9644,
|
|
"step": 11300
|
|
},
|
|
{
|
|
"epoch": 5.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 11310
|
|
},
|
|
{
|
|
"epoch": 5.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9584,
|
|
"step": 11320
|
|
},
|
|
{
|
|
"epoch": 5.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 11330
|
|
},
|
|
{
|
|
"epoch": 5.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9213,
|
|
"step": 11340
|
|
},
|
|
{
|
|
"epoch": 5.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 11350
|
|
},
|
|
{
|
|
"epoch": 5.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9472,
|
|
"step": 11360
|
|
},
|
|
{
|
|
"epoch": 5.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 11370
|
|
},
|
|
{
|
|
"epoch": 5.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9538,
|
|
"step": 11380
|
|
},
|
|
{
|
|
"epoch": 5.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9502,
|
|
"step": 11390
|
|
},
|
|
{
|
|
"epoch": 5.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9348,
|
|
"step": 11400
|
|
},
|
|
{
|
|
"epoch": 5.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9488,
|
|
"step": 11410
|
|
},
|
|
{
|
|
"epoch": 5.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9585,
|
|
"step": 11420
|
|
},
|
|
{
|
|
"epoch": 5.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.945,
|
|
"step": 11430
|
|
},
|
|
{
|
|
"epoch": 5.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 11440
|
|
},
|
|
{
|
|
"epoch": 5.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9599,
|
|
"step": 11450
|
|
},
|
|
{
|
|
"epoch": 5.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9445,
|
|
"step": 11460
|
|
},
|
|
{
|
|
"epoch": 5.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 11470
|
|
},
|
|
{
|
|
"epoch": 5.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 11480
|
|
},
|
|
{
|
|
"epoch": 5.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9242,
|
|
"step": 11490
|
|
},
|
|
{
|
|
"epoch": 5.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9447,
|
|
"step": 11500
|
|
},
|
|
{
|
|
"epoch": 5.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9472,
|
|
"step": 11510
|
|
},
|
|
{
|
|
"epoch": 5.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9703,
|
|
"step": 11520
|
|
},
|
|
{
|
|
"epoch": 5.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9566,
|
|
"step": 11530
|
|
},
|
|
{
|
|
"epoch": 5.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 11540
|
|
},
|
|
{
|
|
"epoch": 5.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9345,
|
|
"step": 11550
|
|
},
|
|
{
|
|
"epoch": 6.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 11560
|
|
},
|
|
{
|
|
"epoch": 6.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.17665369649805449
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.04289493575207861
|
|
},
|
|
"eval_loss": 1.9427300691604614,
|
|
"eval_precision": {
|
|
"precision": 0.025236242356864926
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 77.3874,
|
|
"eval_samples_per_second": 49.814,
|
|
"eval_steps_per_second": 24.914,
|
|
"step": 11562
|
|
},
|
|
{
|
|
"epoch": 6.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9544,
|
|
"step": 11570
|
|
},
|
|
{
|
|
"epoch": 6.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 11580
|
|
},
|
|
{
|
|
"epoch": 6.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9266,
|
|
"step": 11590
|
|
},
|
|
{
|
|
"epoch": 6.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.937,
|
|
"step": 11600
|
|
},
|
|
{
|
|
"epoch": 6.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 11610
|
|
},
|
|
{
|
|
"epoch": 6.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 11620
|
|
},
|
|
{
|
|
"epoch": 6.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9284,
|
|
"step": 11630
|
|
},
|
|
{
|
|
"epoch": 6.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9544,
|
|
"step": 11640
|
|
},
|
|
{
|
|
"epoch": 6.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 11650
|
|
},
|
|
{
|
|
"epoch": 6.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9539,
|
|
"step": 11660
|
|
},
|
|
{
|
|
"epoch": 6.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 11670
|
|
},
|
|
{
|
|
"epoch": 6.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9234,
|
|
"step": 11680
|
|
},
|
|
{
|
|
"epoch": 6.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9336,
|
|
"step": 11690
|
|
},
|
|
{
|
|
"epoch": 6.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9397,
|
|
"step": 11700
|
|
},
|
|
{
|
|
"epoch": 6.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9411,
|
|
"step": 11710
|
|
},
|
|
{
|
|
"epoch": 6.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9233,
|
|
"step": 11720
|
|
},
|
|
{
|
|
"epoch": 6.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9312,
|
|
"step": 11730
|
|
},
|
|
{
|
|
"epoch": 6.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9187,
|
|
"step": 11740
|
|
},
|
|
{
|
|
"epoch": 6.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9728,
|
|
"step": 11750
|
|
},
|
|
{
|
|
"epoch": 6.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 11760
|
|
},
|
|
{
|
|
"epoch": 6.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9603,
|
|
"step": 11770
|
|
},
|
|
{
|
|
"epoch": 6.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 11780
|
|
},
|
|
{
|
|
"epoch": 6.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9679,
|
|
"step": 11790
|
|
},
|
|
{
|
|
"epoch": 6.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 11800
|
|
},
|
|
{
|
|
"epoch": 6.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 11810
|
|
},
|
|
{
|
|
"epoch": 6.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9476,
|
|
"step": 11820
|
|
},
|
|
{
|
|
"epoch": 6.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9263,
|
|
"step": 11830
|
|
},
|
|
{
|
|
"epoch": 6.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 11840
|
|
},
|
|
{
|
|
"epoch": 6.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9329,
|
|
"step": 11850
|
|
},
|
|
{
|
|
"epoch": 6.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9514,
|
|
"step": 11860
|
|
},
|
|
{
|
|
"epoch": 6.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9343,
|
|
"step": 11870
|
|
},
|
|
{
|
|
"epoch": 6.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9511,
|
|
"step": 11880
|
|
},
|
|
{
|
|
"epoch": 6.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9594,
|
|
"step": 11890
|
|
},
|
|
{
|
|
"epoch": 6.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9575,
|
|
"step": 11900
|
|
},
|
|
{
|
|
"epoch": 6.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9399,
|
|
"step": 11910
|
|
},
|
|
{
|
|
"epoch": 6.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9474,
|
|
"step": 11920
|
|
},
|
|
{
|
|
"epoch": 6.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9595,
|
|
"step": 11930
|
|
},
|
|
{
|
|
"epoch": 6.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9476,
|
|
"step": 11940
|
|
},
|
|
{
|
|
"epoch": 6.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 11950
|
|
},
|
|
{
|
|
"epoch": 6.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9429,
|
|
"step": 11960
|
|
},
|
|
{
|
|
"epoch": 6.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 11970
|
|
},
|
|
{
|
|
"epoch": 6.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9447,
|
|
"step": 11980
|
|
},
|
|
{
|
|
"epoch": 6.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9408,
|
|
"step": 11990
|
|
},
|
|
{
|
|
"epoch": 6.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9468,
|
|
"step": 12000
|
|
},
|
|
{
|
|
"epoch": 6.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 12010
|
|
},
|
|
{
|
|
"epoch": 6.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9608,
|
|
"step": 12020
|
|
},
|
|
{
|
|
"epoch": 6.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9256,
|
|
"step": 12030
|
|
},
|
|
{
|
|
"epoch": 6.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 12040
|
|
},
|
|
{
|
|
"epoch": 6.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 12050
|
|
},
|
|
{
|
|
"epoch": 6.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9548,
|
|
"step": 12060
|
|
},
|
|
{
|
|
"epoch": 6.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9526,
|
|
"step": 12070
|
|
},
|
|
{
|
|
"epoch": 6.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9508,
|
|
"step": 12080
|
|
},
|
|
{
|
|
"epoch": 6.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 12090
|
|
},
|
|
{
|
|
"epoch": 6.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9319,
|
|
"step": 12100
|
|
},
|
|
{
|
|
"epoch": 6.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9488,
|
|
"step": 12110
|
|
},
|
|
{
|
|
"epoch": 6.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.919,
|
|
"step": 12120
|
|
},
|
|
{
|
|
"epoch": 6.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 12130
|
|
},
|
|
{
|
|
"epoch": 6.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9304,
|
|
"step": 12140
|
|
},
|
|
{
|
|
"epoch": 6.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9628,
|
|
"step": 12150
|
|
},
|
|
{
|
|
"epoch": 6.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.933,
|
|
"step": 12160
|
|
},
|
|
{
|
|
"epoch": 6.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9667,
|
|
"step": 12170
|
|
},
|
|
{
|
|
"epoch": 6.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.919,
|
|
"step": 12180
|
|
},
|
|
{
|
|
"epoch": 6.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9281,
|
|
"step": 12190
|
|
},
|
|
{
|
|
"epoch": 6.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 12200
|
|
},
|
|
{
|
|
"epoch": 6.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9363,
|
|
"step": 12210
|
|
},
|
|
{
|
|
"epoch": 6.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9386,
|
|
"step": 12220
|
|
},
|
|
{
|
|
"epoch": 6.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9556,
|
|
"step": 12230
|
|
},
|
|
{
|
|
"epoch": 6.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9363,
|
|
"step": 12240
|
|
},
|
|
{
|
|
"epoch": 6.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9536,
|
|
"step": 12250
|
|
},
|
|
{
|
|
"epoch": 6.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 12260
|
|
},
|
|
{
|
|
"epoch": 6.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9727,
|
|
"step": 12270
|
|
},
|
|
{
|
|
"epoch": 6.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 12280
|
|
},
|
|
{
|
|
"epoch": 6.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9399,
|
|
"step": 12290
|
|
},
|
|
{
|
|
"epoch": 6.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9403,
|
|
"step": 12300
|
|
},
|
|
{
|
|
"epoch": 6.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9325,
|
|
"step": 12310
|
|
},
|
|
{
|
|
"epoch": 6.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9502,
|
|
"step": 12320
|
|
},
|
|
{
|
|
"epoch": 6.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9302,
|
|
"step": 12330
|
|
},
|
|
{
|
|
"epoch": 6.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9351,
|
|
"step": 12340
|
|
},
|
|
{
|
|
"epoch": 6.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.946,
|
|
"step": 12350
|
|
},
|
|
{
|
|
"epoch": 6.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9326,
|
|
"step": 12360
|
|
},
|
|
{
|
|
"epoch": 6.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 12370
|
|
},
|
|
{
|
|
"epoch": 6.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9569,
|
|
"step": 12380
|
|
},
|
|
{
|
|
"epoch": 6.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 12390
|
|
},
|
|
{
|
|
"epoch": 6.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 12400
|
|
},
|
|
{
|
|
"epoch": 6.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9382,
|
|
"step": 12410
|
|
},
|
|
{
|
|
"epoch": 6.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9519,
|
|
"step": 12420
|
|
},
|
|
{
|
|
"epoch": 6.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9212,
|
|
"step": 12430
|
|
},
|
|
{
|
|
"epoch": 6.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.929,
|
|
"step": 12440
|
|
},
|
|
{
|
|
"epoch": 6.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9354,
|
|
"step": 12450
|
|
},
|
|
{
|
|
"epoch": 6.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 12460
|
|
},
|
|
{
|
|
"epoch": 6.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9466,
|
|
"step": 12470
|
|
},
|
|
{
|
|
"epoch": 6.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9391,
|
|
"step": 12480
|
|
},
|
|
{
|
|
"epoch": 6.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9259,
|
|
"step": 12490
|
|
},
|
|
{
|
|
"epoch": 6.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 12500
|
|
},
|
|
{
|
|
"epoch": 6.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9424,
|
|
"step": 12510
|
|
},
|
|
{
|
|
"epoch": 6.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9551,
|
|
"step": 12520
|
|
},
|
|
{
|
|
"epoch": 6.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 12530
|
|
},
|
|
{
|
|
"epoch": 6.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9261,
|
|
"step": 12540
|
|
},
|
|
{
|
|
"epoch": 6.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9577,
|
|
"step": 12550
|
|
},
|
|
{
|
|
"epoch": 6.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9313,
|
|
"step": 12560
|
|
},
|
|
{
|
|
"epoch": 6.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9199,
|
|
"step": 12570
|
|
},
|
|
{
|
|
"epoch": 6.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 12580
|
|
},
|
|
{
|
|
"epoch": 6.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.953,
|
|
"step": 12590
|
|
},
|
|
{
|
|
"epoch": 6.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 12600
|
|
},
|
|
{
|
|
"epoch": 6.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9471,
|
|
"step": 12610
|
|
},
|
|
{
|
|
"epoch": 6.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 12620
|
|
},
|
|
{
|
|
"epoch": 6.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 12630
|
|
},
|
|
{
|
|
"epoch": 6.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 12640
|
|
},
|
|
{
|
|
"epoch": 6.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 12650
|
|
},
|
|
{
|
|
"epoch": 6.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9493,
|
|
"step": 12660
|
|
},
|
|
{
|
|
"epoch": 6.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 12670
|
|
},
|
|
{
|
|
"epoch": 6.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9177,
|
|
"step": 12680
|
|
},
|
|
{
|
|
"epoch": 6.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9316,
|
|
"step": 12690
|
|
},
|
|
{
|
|
"epoch": 6.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 12700
|
|
},
|
|
{
|
|
"epoch": 6.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.964,
|
|
"step": 12710
|
|
},
|
|
{
|
|
"epoch": 6.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9324,
|
|
"step": 12720
|
|
},
|
|
{
|
|
"epoch": 6.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 12730
|
|
},
|
|
{
|
|
"epoch": 6.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9659,
|
|
"step": 12740
|
|
},
|
|
{
|
|
"epoch": 6.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 12750
|
|
},
|
|
{
|
|
"epoch": 6.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9326,
|
|
"step": 12760
|
|
},
|
|
{
|
|
"epoch": 6.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9511,
|
|
"step": 12770
|
|
},
|
|
{
|
|
"epoch": 6.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9314,
|
|
"step": 12780
|
|
},
|
|
{
|
|
"epoch": 6.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9627,
|
|
"step": 12790
|
|
},
|
|
{
|
|
"epoch": 6.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.92,
|
|
"step": 12800
|
|
},
|
|
{
|
|
"epoch": 6.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9569,
|
|
"step": 12810
|
|
},
|
|
{
|
|
"epoch": 6.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 12820
|
|
},
|
|
{
|
|
"epoch": 6.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9256,
|
|
"step": 12830
|
|
},
|
|
{
|
|
"epoch": 6.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9342,
|
|
"step": 12840
|
|
},
|
|
{
|
|
"epoch": 6.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 12850
|
|
},
|
|
{
|
|
"epoch": 6.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9354,
|
|
"step": 12860
|
|
},
|
|
{
|
|
"epoch": 6.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.956,
|
|
"step": 12870
|
|
},
|
|
{
|
|
"epoch": 6.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9549,
|
|
"step": 12880
|
|
},
|
|
{
|
|
"epoch": 6.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9196,
|
|
"step": 12890
|
|
},
|
|
{
|
|
"epoch": 6.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 12900
|
|
},
|
|
{
|
|
"epoch": 6.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 12910
|
|
},
|
|
{
|
|
"epoch": 6.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9608,
|
|
"step": 12920
|
|
},
|
|
{
|
|
"epoch": 6.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 12930
|
|
},
|
|
{
|
|
"epoch": 6.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9511,
|
|
"step": 12940
|
|
},
|
|
{
|
|
"epoch": 6.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9277,
|
|
"step": 12950
|
|
},
|
|
{
|
|
"epoch": 6.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 12960
|
|
},
|
|
{
|
|
"epoch": 6.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9499,
|
|
"step": 12970
|
|
},
|
|
{
|
|
"epoch": 6.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9334,
|
|
"step": 12980
|
|
},
|
|
{
|
|
"epoch": 6.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.923,
|
|
"step": 12990
|
|
},
|
|
{
|
|
"epoch": 6.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 13000
|
|
},
|
|
{
|
|
"epoch": 6.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 13010
|
|
},
|
|
{
|
|
"epoch": 6.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.957,
|
|
"step": 13020
|
|
},
|
|
{
|
|
"epoch": 6.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 13030
|
|
},
|
|
{
|
|
"epoch": 6.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9517,
|
|
"step": 13040
|
|
},
|
|
{
|
|
"epoch": 6.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9448,
|
|
"step": 13050
|
|
},
|
|
{
|
|
"epoch": 6.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9297,
|
|
"step": 13060
|
|
},
|
|
{
|
|
"epoch": 6.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9317,
|
|
"step": 13070
|
|
},
|
|
{
|
|
"epoch": 6.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9439,
|
|
"step": 13080
|
|
},
|
|
{
|
|
"epoch": 6.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9527,
|
|
"step": 13090
|
|
},
|
|
{
|
|
"epoch": 6.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 13100
|
|
},
|
|
{
|
|
"epoch": 6.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 13110
|
|
},
|
|
{
|
|
"epoch": 6.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9254,
|
|
"step": 13120
|
|
},
|
|
{
|
|
"epoch": 6.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9722,
|
|
"step": 13130
|
|
},
|
|
{
|
|
"epoch": 6.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9611,
|
|
"step": 13140
|
|
},
|
|
{
|
|
"epoch": 6.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 13150
|
|
},
|
|
{
|
|
"epoch": 6.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9319,
|
|
"step": 13160
|
|
},
|
|
{
|
|
"epoch": 6.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9545,
|
|
"step": 13170
|
|
},
|
|
{
|
|
"epoch": 6.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 13180
|
|
},
|
|
{
|
|
"epoch": 6.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9457,
|
|
"step": 13190
|
|
},
|
|
{
|
|
"epoch": 6.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 13200
|
|
},
|
|
{
|
|
"epoch": 6.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 13210
|
|
},
|
|
{
|
|
"epoch": 6.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.947,
|
|
"step": 13220
|
|
},
|
|
{
|
|
"epoch": 6.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9168,
|
|
"step": 13230
|
|
},
|
|
{
|
|
"epoch": 6.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9118,
|
|
"step": 13240
|
|
},
|
|
{
|
|
"epoch": 6.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9839,
|
|
"step": 13250
|
|
},
|
|
{
|
|
"epoch": 6.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9411,
|
|
"step": 13260
|
|
},
|
|
{
|
|
"epoch": 6.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9575,
|
|
"step": 13270
|
|
},
|
|
{
|
|
"epoch": 6.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9477,
|
|
"step": 13280
|
|
},
|
|
{
|
|
"epoch": 6.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9482,
|
|
"step": 13290
|
|
},
|
|
{
|
|
"epoch": 6.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9485,
|
|
"step": 13300
|
|
},
|
|
{
|
|
"epoch": 6.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9662,
|
|
"step": 13310
|
|
},
|
|
{
|
|
"epoch": 6.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.937,
|
|
"step": 13320
|
|
},
|
|
{
|
|
"epoch": 6.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 13330
|
|
},
|
|
{
|
|
"epoch": 6.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9661,
|
|
"step": 13340
|
|
},
|
|
{
|
|
"epoch": 6.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9593,
|
|
"step": 13350
|
|
},
|
|
{
|
|
"epoch": 6.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9306,
|
|
"step": 13360
|
|
},
|
|
{
|
|
"epoch": 6.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9415,
|
|
"step": 13370
|
|
},
|
|
{
|
|
"epoch": 6.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9525,
|
|
"step": 13380
|
|
},
|
|
{
|
|
"epoch": 6.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9245,
|
|
"step": 13390
|
|
},
|
|
{
|
|
"epoch": 6.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9474,
|
|
"step": 13400
|
|
},
|
|
{
|
|
"epoch": 6.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 13410
|
|
},
|
|
{
|
|
"epoch": 6.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 13420
|
|
},
|
|
{
|
|
"epoch": 6.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 13430
|
|
},
|
|
{
|
|
"epoch": 6.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.963,
|
|
"step": 13440
|
|
},
|
|
{
|
|
"epoch": 6.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.957,
|
|
"step": 13450
|
|
},
|
|
{
|
|
"epoch": 6.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 13460
|
|
},
|
|
{
|
|
"epoch": 6.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9837,
|
|
"step": 13470
|
|
},
|
|
{
|
|
"epoch": 7.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 13480
|
|
},
|
|
{
|
|
"epoch": 7.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.1556420233463035
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03848003848003848
|
|
},
|
|
"eval_loss": 1.9433287382125854,
|
|
"eval_precision": {
|
|
"precision": 0.022234574763757644
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 77.8252,
|
|
"eval_samples_per_second": 49.534,
|
|
"eval_steps_per_second": 24.773,
|
|
"step": 13489
|
|
},
|
|
{
|
|
"epoch": 7.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 13490
|
|
},
|
|
{
|
|
"epoch": 7.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9573,
|
|
"step": 13500
|
|
},
|
|
{
|
|
"epoch": 7.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9405,
|
|
"step": 13510
|
|
},
|
|
{
|
|
"epoch": 7.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9472,
|
|
"step": 13520
|
|
},
|
|
{
|
|
"epoch": 7.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 13530
|
|
},
|
|
{
|
|
"epoch": 7.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9592,
|
|
"step": 13540
|
|
},
|
|
{
|
|
"epoch": 7.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9641,
|
|
"step": 13550
|
|
},
|
|
{
|
|
"epoch": 7.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.947,
|
|
"step": 13560
|
|
},
|
|
{
|
|
"epoch": 7.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 13570
|
|
},
|
|
{
|
|
"epoch": 7.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.942,
|
|
"step": 13580
|
|
},
|
|
{
|
|
"epoch": 7.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 13590
|
|
},
|
|
{
|
|
"epoch": 7.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9406,
|
|
"step": 13600
|
|
},
|
|
{
|
|
"epoch": 7.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9525,
|
|
"step": 13610
|
|
},
|
|
{
|
|
"epoch": 7.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 13620
|
|
},
|
|
{
|
|
"epoch": 7.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9318,
|
|
"step": 13630
|
|
},
|
|
{
|
|
"epoch": 7.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9361,
|
|
"step": 13640
|
|
},
|
|
{
|
|
"epoch": 7.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9303,
|
|
"step": 13650
|
|
},
|
|
{
|
|
"epoch": 7.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9401,
|
|
"step": 13660
|
|
},
|
|
{
|
|
"epoch": 7.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9638,
|
|
"step": 13670
|
|
},
|
|
{
|
|
"epoch": 7.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 13680
|
|
},
|
|
{
|
|
"epoch": 7.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 13690
|
|
},
|
|
{
|
|
"epoch": 7.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9485,
|
|
"step": 13700
|
|
},
|
|
{
|
|
"epoch": 7.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 13710
|
|
},
|
|
{
|
|
"epoch": 7.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9668,
|
|
"step": 13720
|
|
},
|
|
{
|
|
"epoch": 7.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9477,
|
|
"step": 13730
|
|
},
|
|
{
|
|
"epoch": 7.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9496,
|
|
"step": 13740
|
|
},
|
|
{
|
|
"epoch": 7.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 13750
|
|
},
|
|
{
|
|
"epoch": 7.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9664,
|
|
"step": 13760
|
|
},
|
|
{
|
|
"epoch": 7.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9347,
|
|
"step": 13770
|
|
},
|
|
{
|
|
"epoch": 7.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 13780
|
|
},
|
|
{
|
|
"epoch": 7.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9385,
|
|
"step": 13790
|
|
},
|
|
{
|
|
"epoch": 7.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 13800
|
|
},
|
|
{
|
|
"epoch": 7.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9555,
|
|
"step": 13810
|
|
},
|
|
{
|
|
"epoch": 7.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 13820
|
|
},
|
|
{
|
|
"epoch": 7.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 13830
|
|
},
|
|
{
|
|
"epoch": 7.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9545,
|
|
"step": 13840
|
|
},
|
|
{
|
|
"epoch": 7.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9361,
|
|
"step": 13850
|
|
},
|
|
{
|
|
"epoch": 7.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9405,
|
|
"step": 13860
|
|
},
|
|
{
|
|
"epoch": 7.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 13870
|
|
},
|
|
{
|
|
"epoch": 7.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9733,
|
|
"step": 13880
|
|
},
|
|
{
|
|
"epoch": 7.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 13890
|
|
},
|
|
{
|
|
"epoch": 7.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 13900
|
|
},
|
|
{
|
|
"epoch": 7.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 13910
|
|
},
|
|
{
|
|
"epoch": 7.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9457,
|
|
"step": 13920
|
|
},
|
|
{
|
|
"epoch": 7.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 13930
|
|
},
|
|
{
|
|
"epoch": 7.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9256,
|
|
"step": 13940
|
|
},
|
|
{
|
|
"epoch": 7.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9255,
|
|
"step": 13950
|
|
},
|
|
{
|
|
"epoch": 7.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 13960
|
|
},
|
|
{
|
|
"epoch": 7.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 13970
|
|
},
|
|
{
|
|
"epoch": 7.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 13980
|
|
},
|
|
{
|
|
"epoch": 7.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.931,
|
|
"step": 13990
|
|
},
|
|
{
|
|
"epoch": 7.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9608,
|
|
"step": 14000
|
|
},
|
|
{
|
|
"epoch": 7.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 14010
|
|
},
|
|
{
|
|
"epoch": 7.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.951,
|
|
"step": 14020
|
|
},
|
|
{
|
|
"epoch": 7.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9425,
|
|
"step": 14030
|
|
},
|
|
{
|
|
"epoch": 7.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 14040
|
|
},
|
|
{
|
|
"epoch": 7.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 14050
|
|
},
|
|
{
|
|
"epoch": 7.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9313,
|
|
"step": 14060
|
|
},
|
|
{
|
|
"epoch": 7.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 14070
|
|
},
|
|
{
|
|
"epoch": 7.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9215,
|
|
"step": 14080
|
|
},
|
|
{
|
|
"epoch": 7.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9291,
|
|
"step": 14090
|
|
},
|
|
{
|
|
"epoch": 7.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.935,
|
|
"step": 14100
|
|
},
|
|
{
|
|
"epoch": 7.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9188,
|
|
"step": 14110
|
|
},
|
|
{
|
|
"epoch": 7.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9577,
|
|
"step": 14120
|
|
},
|
|
{
|
|
"epoch": 7.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9561,
|
|
"step": 14130
|
|
},
|
|
{
|
|
"epoch": 7.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 14140
|
|
},
|
|
{
|
|
"epoch": 7.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9346,
|
|
"step": 14150
|
|
},
|
|
{
|
|
"epoch": 7.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9458,
|
|
"step": 14160
|
|
},
|
|
{
|
|
"epoch": 7.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9331,
|
|
"step": 14170
|
|
},
|
|
{
|
|
"epoch": 7.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9351,
|
|
"step": 14180
|
|
},
|
|
{
|
|
"epoch": 7.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 14190
|
|
},
|
|
{
|
|
"epoch": 7.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 14200
|
|
},
|
|
{
|
|
"epoch": 7.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.933,
|
|
"step": 14210
|
|
},
|
|
{
|
|
"epoch": 7.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9536,
|
|
"step": 14220
|
|
},
|
|
{
|
|
"epoch": 7.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.941,
|
|
"step": 14230
|
|
},
|
|
{
|
|
"epoch": 7.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 14240
|
|
},
|
|
{
|
|
"epoch": 7.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9356,
|
|
"step": 14250
|
|
},
|
|
{
|
|
"epoch": 7.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 14260
|
|
},
|
|
{
|
|
"epoch": 7.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 14270
|
|
},
|
|
{
|
|
"epoch": 7.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9554,
|
|
"step": 14280
|
|
},
|
|
{
|
|
"epoch": 7.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9483,
|
|
"step": 14290
|
|
},
|
|
{
|
|
"epoch": 7.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 14300
|
|
},
|
|
{
|
|
"epoch": 7.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9244,
|
|
"step": 14310
|
|
},
|
|
{
|
|
"epoch": 7.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 14320
|
|
},
|
|
{
|
|
"epoch": 7.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 14330
|
|
},
|
|
{
|
|
"epoch": 7.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9513,
|
|
"step": 14340
|
|
},
|
|
{
|
|
"epoch": 7.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9426,
|
|
"step": 14350
|
|
},
|
|
{
|
|
"epoch": 7.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9289,
|
|
"step": 14360
|
|
},
|
|
{
|
|
"epoch": 7.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9271,
|
|
"step": 14370
|
|
},
|
|
{
|
|
"epoch": 7.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9607,
|
|
"step": 14380
|
|
},
|
|
{
|
|
"epoch": 7.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9377,
|
|
"step": 14390
|
|
},
|
|
{
|
|
"epoch": 7.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 14400
|
|
},
|
|
{
|
|
"epoch": 7.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9432,
|
|
"step": 14410
|
|
},
|
|
{
|
|
"epoch": 7.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9618,
|
|
"step": 14420
|
|
},
|
|
{
|
|
"epoch": 7.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9425,
|
|
"step": 14430
|
|
},
|
|
{
|
|
"epoch": 7.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 14440
|
|
},
|
|
{
|
|
"epoch": 7.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9268,
|
|
"step": 14450
|
|
},
|
|
{
|
|
"epoch": 7.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.936,
|
|
"step": 14460
|
|
},
|
|
{
|
|
"epoch": 7.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 14470
|
|
},
|
|
{
|
|
"epoch": 7.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 14480
|
|
},
|
|
{
|
|
"epoch": 7.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 14490
|
|
},
|
|
{
|
|
"epoch": 7.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9488,
|
|
"step": 14500
|
|
},
|
|
{
|
|
"epoch": 7.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9383,
|
|
"step": 14510
|
|
},
|
|
{
|
|
"epoch": 7.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9415,
|
|
"step": 14520
|
|
},
|
|
{
|
|
"epoch": 7.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9586,
|
|
"step": 14530
|
|
},
|
|
{
|
|
"epoch": 7.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9568,
|
|
"step": 14540
|
|
},
|
|
{
|
|
"epoch": 7.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 14550
|
|
},
|
|
{
|
|
"epoch": 7.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 14560
|
|
},
|
|
{
|
|
"epoch": 7.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9606,
|
|
"step": 14570
|
|
},
|
|
{
|
|
"epoch": 7.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9455,
|
|
"step": 14580
|
|
},
|
|
{
|
|
"epoch": 7.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 14590
|
|
},
|
|
{
|
|
"epoch": 7.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 14600
|
|
},
|
|
{
|
|
"epoch": 7.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9476,
|
|
"step": 14610
|
|
},
|
|
{
|
|
"epoch": 7.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9464,
|
|
"step": 14620
|
|
},
|
|
{
|
|
"epoch": 7.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9599,
|
|
"step": 14630
|
|
},
|
|
{
|
|
"epoch": 7.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.933,
|
|
"step": 14640
|
|
},
|
|
{
|
|
"epoch": 7.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9622,
|
|
"step": 14650
|
|
},
|
|
{
|
|
"epoch": 7.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9584,
|
|
"step": 14660
|
|
},
|
|
{
|
|
"epoch": 7.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9429,
|
|
"step": 14670
|
|
},
|
|
{
|
|
"epoch": 7.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 14680
|
|
},
|
|
{
|
|
"epoch": 7.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9403,
|
|
"step": 14690
|
|
},
|
|
{
|
|
"epoch": 7.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.936,
|
|
"step": 14700
|
|
},
|
|
{
|
|
"epoch": 7.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 14710
|
|
},
|
|
{
|
|
"epoch": 7.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9331,
|
|
"step": 14720
|
|
},
|
|
{
|
|
"epoch": 7.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9529,
|
|
"step": 14730
|
|
},
|
|
{
|
|
"epoch": 7.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9255,
|
|
"step": 14740
|
|
},
|
|
{
|
|
"epoch": 7.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9385,
|
|
"step": 14750
|
|
},
|
|
{
|
|
"epoch": 7.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 14760
|
|
},
|
|
{
|
|
"epoch": 7.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 14770
|
|
},
|
|
{
|
|
"epoch": 7.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 14780
|
|
},
|
|
{
|
|
"epoch": 7.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.93,
|
|
"step": 14790
|
|
},
|
|
{
|
|
"epoch": 7.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 14800
|
|
},
|
|
{
|
|
"epoch": 7.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9415,
|
|
"step": 14810
|
|
},
|
|
{
|
|
"epoch": 7.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9328,
|
|
"step": 14820
|
|
},
|
|
{
|
|
"epoch": 7.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9529,
|
|
"step": 14830
|
|
},
|
|
{
|
|
"epoch": 7.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9532,
|
|
"step": 14840
|
|
},
|
|
{
|
|
"epoch": 7.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9255,
|
|
"step": 14850
|
|
},
|
|
{
|
|
"epoch": 7.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9364,
|
|
"step": 14860
|
|
},
|
|
{
|
|
"epoch": 7.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.931,
|
|
"step": 14870
|
|
},
|
|
{
|
|
"epoch": 7.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9492,
|
|
"step": 14880
|
|
},
|
|
{
|
|
"epoch": 7.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9556,
|
|
"step": 14890
|
|
},
|
|
{
|
|
"epoch": 7.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 14900
|
|
},
|
|
{
|
|
"epoch": 7.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9519,
|
|
"step": 14910
|
|
},
|
|
{
|
|
"epoch": 7.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9406,
|
|
"step": 14920
|
|
},
|
|
{
|
|
"epoch": 7.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 14930
|
|
},
|
|
{
|
|
"epoch": 7.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9389,
|
|
"step": 14940
|
|
},
|
|
{
|
|
"epoch": 7.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9449,
|
|
"step": 14950
|
|
},
|
|
{
|
|
"epoch": 7.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9286,
|
|
"step": 14960
|
|
},
|
|
{
|
|
"epoch": 7.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 14970
|
|
},
|
|
{
|
|
"epoch": 7.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.956,
|
|
"step": 14980
|
|
},
|
|
{
|
|
"epoch": 7.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9369,
|
|
"step": 14990
|
|
},
|
|
{
|
|
"epoch": 7.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9321,
|
|
"step": 15000
|
|
},
|
|
{
|
|
"epoch": 7.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 15010
|
|
},
|
|
{
|
|
"epoch": 7.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9542,
|
|
"step": 15020
|
|
},
|
|
{
|
|
"epoch": 7.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9503,
|
|
"step": 15030
|
|
},
|
|
{
|
|
"epoch": 7.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9493,
|
|
"step": 15040
|
|
},
|
|
{
|
|
"epoch": 7.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 15050
|
|
},
|
|
{
|
|
"epoch": 7.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 15060
|
|
},
|
|
{
|
|
"epoch": 7.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9295,
|
|
"step": 15070
|
|
},
|
|
{
|
|
"epoch": 7.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9256,
|
|
"step": 15080
|
|
},
|
|
{
|
|
"epoch": 7.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9374,
|
|
"step": 15090
|
|
},
|
|
{
|
|
"epoch": 7.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 15100
|
|
},
|
|
{
|
|
"epoch": 7.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9103,
|
|
"step": 15110
|
|
},
|
|
{
|
|
"epoch": 7.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9173,
|
|
"step": 15120
|
|
},
|
|
{
|
|
"epoch": 7.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9138,
|
|
"step": 15130
|
|
},
|
|
{
|
|
"epoch": 7.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9622,
|
|
"step": 15140
|
|
},
|
|
{
|
|
"epoch": 7.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9637,
|
|
"step": 15150
|
|
},
|
|
{
|
|
"epoch": 7.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 15160
|
|
},
|
|
{
|
|
"epoch": 7.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9356,
|
|
"step": 15170
|
|
},
|
|
{
|
|
"epoch": 7.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 15180
|
|
},
|
|
{
|
|
"epoch": 7.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9209,
|
|
"step": 15190
|
|
},
|
|
{
|
|
"epoch": 7.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9247,
|
|
"step": 15200
|
|
},
|
|
{
|
|
"epoch": 7.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9585,
|
|
"step": 15210
|
|
},
|
|
{
|
|
"epoch": 7.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 15220
|
|
},
|
|
{
|
|
"epoch": 7.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 15230
|
|
},
|
|
{
|
|
"epoch": 7.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9538,
|
|
"step": 15240
|
|
},
|
|
{
|
|
"epoch": 7.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9465,
|
|
"step": 15250
|
|
},
|
|
{
|
|
"epoch": 7.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9424,
|
|
"step": 15260
|
|
},
|
|
{
|
|
"epoch": 7.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9611,
|
|
"step": 15270
|
|
},
|
|
{
|
|
"epoch": 7.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 15280
|
|
},
|
|
{
|
|
"epoch": 7.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9411,
|
|
"step": 15290
|
|
},
|
|
{
|
|
"epoch": 7.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 15300
|
|
},
|
|
{
|
|
"epoch": 7.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 15310
|
|
},
|
|
{
|
|
"epoch": 7.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 15320
|
|
},
|
|
{
|
|
"epoch": 7.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9541,
|
|
"step": 15330
|
|
},
|
|
{
|
|
"epoch": 7.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 15340
|
|
},
|
|
{
|
|
"epoch": 7.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 15350
|
|
},
|
|
{
|
|
"epoch": 7.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 15360
|
|
},
|
|
{
|
|
"epoch": 7.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 15370
|
|
},
|
|
{
|
|
"epoch": 7.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9279,
|
|
"step": 15380
|
|
},
|
|
{
|
|
"epoch": 7.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 15390
|
|
},
|
|
{
|
|
"epoch": 7.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 15400
|
|
},
|
|
{
|
|
"epoch": 8.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9529,
|
|
"step": 15410
|
|
},
|
|
{
|
|
"epoch": 8.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.1556420233463035
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03848003848003848
|
|
},
|
|
"eval_loss": 1.9431941509246826,
|
|
"eval_precision": {
|
|
"precision": 0.022234574763757644
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 76.9544,
|
|
"eval_samples_per_second": 50.095,
|
|
"eval_steps_per_second": 25.054,
|
|
"step": 15416
|
|
},
|
|
{
|
|
"epoch": 8.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 15420
|
|
},
|
|
{
|
|
"epoch": 8.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 15430
|
|
},
|
|
{
|
|
"epoch": 8.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 15440
|
|
},
|
|
{
|
|
"epoch": 8.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9514,
|
|
"step": 15450
|
|
},
|
|
{
|
|
"epoch": 8.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9308,
|
|
"step": 15460
|
|
},
|
|
{
|
|
"epoch": 8.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9416,
|
|
"step": 15470
|
|
},
|
|
{
|
|
"epoch": 8.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9699,
|
|
"step": 15480
|
|
},
|
|
{
|
|
"epoch": 8.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9319,
|
|
"step": 15490
|
|
},
|
|
{
|
|
"epoch": 8.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 15500
|
|
},
|
|
{
|
|
"epoch": 8.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9338,
|
|
"step": 15510
|
|
},
|
|
{
|
|
"epoch": 8.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 15520
|
|
},
|
|
{
|
|
"epoch": 8.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.923,
|
|
"step": 15530
|
|
},
|
|
{
|
|
"epoch": 8.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9537,
|
|
"step": 15540
|
|
},
|
|
{
|
|
"epoch": 8.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9206,
|
|
"step": 15550
|
|
},
|
|
{
|
|
"epoch": 8.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9462,
|
|
"step": 15560
|
|
},
|
|
{
|
|
"epoch": 8.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.957,
|
|
"step": 15570
|
|
},
|
|
{
|
|
"epoch": 8.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9448,
|
|
"step": 15580
|
|
},
|
|
{
|
|
"epoch": 8.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9495,
|
|
"step": 15590
|
|
},
|
|
{
|
|
"epoch": 8.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9297,
|
|
"step": 15600
|
|
},
|
|
{
|
|
"epoch": 8.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9284,
|
|
"step": 15610
|
|
},
|
|
{
|
|
"epoch": 8.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9255,
|
|
"step": 15620
|
|
},
|
|
{
|
|
"epoch": 8.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 15630
|
|
},
|
|
{
|
|
"epoch": 8.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9625,
|
|
"step": 15640
|
|
},
|
|
{
|
|
"epoch": 8.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9514,
|
|
"step": 15650
|
|
},
|
|
{
|
|
"epoch": 8.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 15660
|
|
},
|
|
{
|
|
"epoch": 8.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.926,
|
|
"step": 15670
|
|
},
|
|
{
|
|
"epoch": 8.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 15680
|
|
},
|
|
{
|
|
"epoch": 8.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 15690
|
|
},
|
|
{
|
|
"epoch": 8.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9332,
|
|
"step": 15700
|
|
},
|
|
{
|
|
"epoch": 8.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9233,
|
|
"step": 15710
|
|
},
|
|
{
|
|
"epoch": 8.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 15720
|
|
},
|
|
{
|
|
"epoch": 8.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9503,
|
|
"step": 15730
|
|
},
|
|
{
|
|
"epoch": 8.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9172,
|
|
"step": 15740
|
|
},
|
|
{
|
|
"epoch": 8.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9519,
|
|
"step": 15750
|
|
},
|
|
{
|
|
"epoch": 8.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 15760
|
|
},
|
|
{
|
|
"epoch": 8.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9267,
|
|
"step": 15770
|
|
},
|
|
{
|
|
"epoch": 8.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9593,
|
|
"step": 15780
|
|
},
|
|
{
|
|
"epoch": 8.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9413,
|
|
"step": 15790
|
|
},
|
|
{
|
|
"epoch": 8.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 15800
|
|
},
|
|
{
|
|
"epoch": 8.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9363,
|
|
"step": 15810
|
|
},
|
|
{
|
|
"epoch": 8.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9275,
|
|
"step": 15820
|
|
},
|
|
{
|
|
"epoch": 8.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9316,
|
|
"step": 15830
|
|
},
|
|
{
|
|
"epoch": 8.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9306,
|
|
"step": 15840
|
|
},
|
|
{
|
|
"epoch": 8.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.979,
|
|
"step": 15850
|
|
},
|
|
{
|
|
"epoch": 8.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9507,
|
|
"step": 15860
|
|
},
|
|
{
|
|
"epoch": 8.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 15870
|
|
},
|
|
{
|
|
"epoch": 8.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9491,
|
|
"step": 15880
|
|
},
|
|
{
|
|
"epoch": 8.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 15890
|
|
},
|
|
{
|
|
"epoch": 8.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9284,
|
|
"step": 15900
|
|
},
|
|
{
|
|
"epoch": 8.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9561,
|
|
"step": 15910
|
|
},
|
|
{
|
|
"epoch": 8.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9269,
|
|
"step": 15920
|
|
},
|
|
{
|
|
"epoch": 8.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9215,
|
|
"step": 15930
|
|
},
|
|
{
|
|
"epoch": 8.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9626,
|
|
"step": 15940
|
|
},
|
|
{
|
|
"epoch": 8.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9607,
|
|
"step": 15950
|
|
},
|
|
{
|
|
"epoch": 8.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 15960
|
|
},
|
|
{
|
|
"epoch": 8.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9474,
|
|
"step": 15970
|
|
},
|
|
{
|
|
"epoch": 8.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 15980
|
|
},
|
|
{
|
|
"epoch": 8.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 15990
|
|
},
|
|
{
|
|
"epoch": 8.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9212,
|
|
"step": 16000
|
|
},
|
|
{
|
|
"epoch": 8.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 16010
|
|
},
|
|
{
|
|
"epoch": 8.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9468,
|
|
"step": 16020
|
|
},
|
|
{
|
|
"epoch": 8.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9513,
|
|
"step": 16030
|
|
},
|
|
{
|
|
"epoch": 8.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9386,
|
|
"step": 16040
|
|
},
|
|
{
|
|
"epoch": 8.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9806,
|
|
"step": 16050
|
|
},
|
|
{
|
|
"epoch": 8.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 16060
|
|
},
|
|
{
|
|
"epoch": 8.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9548,
|
|
"step": 16070
|
|
},
|
|
{
|
|
"epoch": 8.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.945,
|
|
"step": 16080
|
|
},
|
|
{
|
|
"epoch": 8.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9354,
|
|
"step": 16090
|
|
},
|
|
{
|
|
"epoch": 8.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 16100
|
|
},
|
|
{
|
|
"epoch": 8.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9369,
|
|
"step": 16110
|
|
},
|
|
{
|
|
"epoch": 8.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 16120
|
|
},
|
|
{
|
|
"epoch": 8.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9477,
|
|
"step": 16130
|
|
},
|
|
{
|
|
"epoch": 8.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 16140
|
|
},
|
|
{
|
|
"epoch": 8.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 16150
|
|
},
|
|
{
|
|
"epoch": 8.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9362,
|
|
"step": 16160
|
|
},
|
|
{
|
|
"epoch": 8.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9234,
|
|
"step": 16170
|
|
},
|
|
{
|
|
"epoch": 8.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9282,
|
|
"step": 16180
|
|
},
|
|
{
|
|
"epoch": 8.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9281,
|
|
"step": 16190
|
|
},
|
|
{
|
|
"epoch": 8.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9154,
|
|
"step": 16200
|
|
},
|
|
{
|
|
"epoch": 8.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9439,
|
|
"step": 16210
|
|
},
|
|
{
|
|
"epoch": 8.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9602,
|
|
"step": 16220
|
|
},
|
|
{
|
|
"epoch": 8.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 16230
|
|
},
|
|
{
|
|
"epoch": 8.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9438,
|
|
"step": 16240
|
|
},
|
|
{
|
|
"epoch": 8.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 16250
|
|
},
|
|
{
|
|
"epoch": 8.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 16260
|
|
},
|
|
{
|
|
"epoch": 8.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9459,
|
|
"step": 16270
|
|
},
|
|
{
|
|
"epoch": 8.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9497,
|
|
"step": 16280
|
|
},
|
|
{
|
|
"epoch": 8.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 16290
|
|
},
|
|
{
|
|
"epoch": 8.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9716,
|
|
"step": 16300
|
|
},
|
|
{
|
|
"epoch": 8.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9466,
|
|
"step": 16310
|
|
},
|
|
{
|
|
"epoch": 8.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 16320
|
|
},
|
|
{
|
|
"epoch": 8.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9538,
|
|
"step": 16330
|
|
},
|
|
{
|
|
"epoch": 8.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.959,
|
|
"step": 16340
|
|
},
|
|
{
|
|
"epoch": 8.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9448,
|
|
"step": 16350
|
|
},
|
|
{
|
|
"epoch": 8.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9334,
|
|
"step": 16360
|
|
},
|
|
{
|
|
"epoch": 8.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 16370
|
|
},
|
|
{
|
|
"epoch": 8.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9239,
|
|
"step": 16380
|
|
},
|
|
{
|
|
"epoch": 8.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9506,
|
|
"step": 16390
|
|
},
|
|
{
|
|
"epoch": 8.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9519,
|
|
"step": 16400
|
|
},
|
|
{
|
|
"epoch": 8.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.935,
|
|
"step": 16410
|
|
},
|
|
{
|
|
"epoch": 8.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9338,
|
|
"step": 16420
|
|
},
|
|
{
|
|
"epoch": 8.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 16430
|
|
},
|
|
{
|
|
"epoch": 8.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 16440
|
|
},
|
|
{
|
|
"epoch": 8.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9623,
|
|
"step": 16450
|
|
},
|
|
{
|
|
"epoch": 8.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.954,
|
|
"step": 16460
|
|
},
|
|
{
|
|
"epoch": 8.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9421,
|
|
"step": 16470
|
|
},
|
|
{
|
|
"epoch": 8.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9433,
|
|
"step": 16480
|
|
},
|
|
{
|
|
"epoch": 8.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 16490
|
|
},
|
|
{
|
|
"epoch": 8.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9421,
|
|
"step": 16500
|
|
},
|
|
{
|
|
"epoch": 8.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.936,
|
|
"step": 16510
|
|
},
|
|
{
|
|
"epoch": 8.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 16520
|
|
},
|
|
{
|
|
"epoch": 8.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9247,
|
|
"step": 16530
|
|
},
|
|
{
|
|
"epoch": 8.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9245,
|
|
"step": 16540
|
|
},
|
|
{
|
|
"epoch": 8.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 16550
|
|
},
|
|
{
|
|
"epoch": 8.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 16560
|
|
},
|
|
{
|
|
"epoch": 8.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9315,
|
|
"step": 16570
|
|
},
|
|
{
|
|
"epoch": 8.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9605,
|
|
"step": 16580
|
|
},
|
|
{
|
|
"epoch": 8.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9502,
|
|
"step": 16590
|
|
},
|
|
{
|
|
"epoch": 8.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9406,
|
|
"step": 16600
|
|
},
|
|
{
|
|
"epoch": 8.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9485,
|
|
"step": 16610
|
|
},
|
|
{
|
|
"epoch": 8.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 16620
|
|
},
|
|
{
|
|
"epoch": 8.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9439,
|
|
"step": 16630
|
|
},
|
|
{
|
|
"epoch": 8.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 16640
|
|
},
|
|
{
|
|
"epoch": 8.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9459,
|
|
"step": 16650
|
|
},
|
|
{
|
|
"epoch": 8.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9555,
|
|
"step": 16660
|
|
},
|
|
{
|
|
"epoch": 8.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 16670
|
|
},
|
|
{
|
|
"epoch": 8.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 16680
|
|
},
|
|
{
|
|
"epoch": 8.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9519,
|
|
"step": 16690
|
|
},
|
|
{
|
|
"epoch": 8.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 16700
|
|
},
|
|
{
|
|
"epoch": 8.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9276,
|
|
"step": 16710
|
|
},
|
|
{
|
|
"epoch": 8.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.941,
|
|
"step": 16720
|
|
},
|
|
{
|
|
"epoch": 8.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9385,
|
|
"step": 16730
|
|
},
|
|
{
|
|
"epoch": 8.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9507,
|
|
"step": 16740
|
|
},
|
|
{
|
|
"epoch": 8.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9542,
|
|
"step": 16750
|
|
},
|
|
{
|
|
"epoch": 8.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9517,
|
|
"step": 16760
|
|
},
|
|
{
|
|
"epoch": 8.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9499,
|
|
"step": 16770
|
|
},
|
|
{
|
|
"epoch": 8.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9389,
|
|
"step": 16780
|
|
},
|
|
{
|
|
"epoch": 8.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9517,
|
|
"step": 16790
|
|
},
|
|
{
|
|
"epoch": 8.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 16800
|
|
},
|
|
{
|
|
"epoch": 8.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 16810
|
|
},
|
|
{
|
|
"epoch": 8.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9353,
|
|
"step": 16820
|
|
},
|
|
{
|
|
"epoch": 8.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9385,
|
|
"step": 16830
|
|
},
|
|
{
|
|
"epoch": 8.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9775,
|
|
"step": 16840
|
|
},
|
|
{
|
|
"epoch": 8.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9428,
|
|
"step": 16850
|
|
},
|
|
{
|
|
"epoch": 8.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9621,
|
|
"step": 16860
|
|
},
|
|
{
|
|
"epoch": 8.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 16870
|
|
},
|
|
{
|
|
"epoch": 8.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 16880
|
|
},
|
|
{
|
|
"epoch": 8.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.945,
|
|
"step": 16890
|
|
},
|
|
{
|
|
"epoch": 8.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9495,
|
|
"step": 16900
|
|
},
|
|
{
|
|
"epoch": 8.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9343,
|
|
"step": 16910
|
|
},
|
|
{
|
|
"epoch": 8.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9488,
|
|
"step": 16920
|
|
},
|
|
{
|
|
"epoch": 8.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9557,
|
|
"step": 16930
|
|
},
|
|
{
|
|
"epoch": 8.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 16940
|
|
},
|
|
{
|
|
"epoch": 8.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9264,
|
|
"step": 16950
|
|
},
|
|
{
|
|
"epoch": 8.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9445,
|
|
"step": 16960
|
|
},
|
|
{
|
|
"epoch": 8.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9405,
|
|
"step": 16970
|
|
},
|
|
{
|
|
"epoch": 8.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9288,
|
|
"step": 16980
|
|
},
|
|
{
|
|
"epoch": 8.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9303,
|
|
"step": 16990
|
|
},
|
|
{
|
|
"epoch": 8.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9343,
|
|
"step": 17000
|
|
},
|
|
{
|
|
"epoch": 8.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.96,
|
|
"step": 17010
|
|
},
|
|
{
|
|
"epoch": 8.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9827,
|
|
"step": 17020
|
|
},
|
|
{
|
|
"epoch": 8.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 17030
|
|
},
|
|
{
|
|
"epoch": 8.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9549,
|
|
"step": 17040
|
|
},
|
|
{
|
|
"epoch": 8.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9556,
|
|
"step": 17050
|
|
},
|
|
{
|
|
"epoch": 8.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 17060
|
|
},
|
|
{
|
|
"epoch": 8.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9451,
|
|
"step": 17070
|
|
},
|
|
{
|
|
"epoch": 8.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9343,
|
|
"step": 17080
|
|
},
|
|
{
|
|
"epoch": 8.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9334,
|
|
"step": 17090
|
|
},
|
|
{
|
|
"epoch": 8.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9506,
|
|
"step": 17100
|
|
},
|
|
{
|
|
"epoch": 8.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9242,
|
|
"step": 17110
|
|
},
|
|
{
|
|
"epoch": 8.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9516,
|
|
"step": 17120
|
|
},
|
|
{
|
|
"epoch": 8.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9516,
|
|
"step": 17130
|
|
},
|
|
{
|
|
"epoch": 8.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9232,
|
|
"step": 17140
|
|
},
|
|
{
|
|
"epoch": 8.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9175,
|
|
"step": 17150
|
|
},
|
|
{
|
|
"epoch": 8.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9588,
|
|
"step": 17160
|
|
},
|
|
{
|
|
"epoch": 8.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9634,
|
|
"step": 17170
|
|
},
|
|
{
|
|
"epoch": 8.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.941,
|
|
"step": 17180
|
|
},
|
|
{
|
|
"epoch": 8.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 17190
|
|
},
|
|
{
|
|
"epoch": 8.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.956,
|
|
"step": 17200
|
|
},
|
|
{
|
|
"epoch": 8.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9446,
|
|
"step": 17210
|
|
},
|
|
{
|
|
"epoch": 8.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9285,
|
|
"step": 17220
|
|
},
|
|
{
|
|
"epoch": 8.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9628,
|
|
"step": 17230
|
|
},
|
|
{
|
|
"epoch": 8.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 17240
|
|
},
|
|
{
|
|
"epoch": 8.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9459,
|
|
"step": 17250
|
|
},
|
|
{
|
|
"epoch": 8.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 17260
|
|
},
|
|
{
|
|
"epoch": 8.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9357,
|
|
"step": 17270
|
|
},
|
|
{
|
|
"epoch": 8.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9496,
|
|
"step": 17280
|
|
},
|
|
{
|
|
"epoch": 8.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9393,
|
|
"step": 17290
|
|
},
|
|
{
|
|
"epoch": 8.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 17300
|
|
},
|
|
{
|
|
"epoch": 8.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9135,
|
|
"step": 17310
|
|
},
|
|
{
|
|
"epoch": 8.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.975,
|
|
"step": 17320
|
|
},
|
|
{
|
|
"epoch": 8.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9387,
|
|
"step": 17330
|
|
},
|
|
{
|
|
"epoch": 9.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9305,
|
|
"step": 17340
|
|
},
|
|
{
|
|
"epoch": 9.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.1556420233463035
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03848003848003848
|
|
},
|
|
"eval_loss": 1.9463131427764893,
|
|
"eval_precision": {
|
|
"precision": 0.022234574763757644
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 76.7556,
|
|
"eval_samples_per_second": 50.224,
|
|
"eval_steps_per_second": 25.119,
|
|
"step": 17343
|
|
},
|
|
{
|
|
"epoch": 9.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9564,
|
|
"step": 17350
|
|
},
|
|
{
|
|
"epoch": 9.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 17360
|
|
},
|
|
{
|
|
"epoch": 9.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 17370
|
|
},
|
|
{
|
|
"epoch": 9.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9357,
|
|
"step": 17380
|
|
},
|
|
{
|
|
"epoch": 9.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9358,
|
|
"step": 17390
|
|
},
|
|
{
|
|
"epoch": 9.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9372,
|
|
"step": 17400
|
|
},
|
|
{
|
|
"epoch": 9.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9376,
|
|
"step": 17410
|
|
},
|
|
{
|
|
"epoch": 9.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9475,
|
|
"step": 17420
|
|
},
|
|
{
|
|
"epoch": 9.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9539,
|
|
"step": 17430
|
|
},
|
|
{
|
|
"epoch": 9.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9399,
|
|
"step": 17440
|
|
},
|
|
{
|
|
"epoch": 9.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 17450
|
|
},
|
|
{
|
|
"epoch": 9.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9518,
|
|
"step": 17460
|
|
},
|
|
{
|
|
"epoch": 9.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9369,
|
|
"step": 17470
|
|
},
|
|
{
|
|
"epoch": 9.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9352,
|
|
"step": 17480
|
|
},
|
|
{
|
|
"epoch": 9.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 17490
|
|
},
|
|
{
|
|
"epoch": 9.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9335,
|
|
"step": 17500
|
|
},
|
|
{
|
|
"epoch": 9.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 17510
|
|
},
|
|
{
|
|
"epoch": 9.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9403,
|
|
"step": 17520
|
|
},
|
|
{
|
|
"epoch": 9.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 17530
|
|
},
|
|
{
|
|
"epoch": 9.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9421,
|
|
"step": 17540
|
|
},
|
|
{
|
|
"epoch": 9.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.924,
|
|
"step": 17550
|
|
},
|
|
{
|
|
"epoch": 9.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 17560
|
|
},
|
|
{
|
|
"epoch": 9.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9226,
|
|
"step": 17570
|
|
},
|
|
{
|
|
"epoch": 9.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9669,
|
|
"step": 17580
|
|
},
|
|
{
|
|
"epoch": 9.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9318,
|
|
"step": 17590
|
|
},
|
|
{
|
|
"epoch": 9.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 17600
|
|
},
|
|
{
|
|
"epoch": 9.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9491,
|
|
"step": 17610
|
|
},
|
|
{
|
|
"epoch": 9.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 17620
|
|
},
|
|
{
|
|
"epoch": 9.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9367,
|
|
"step": 17630
|
|
},
|
|
{
|
|
"epoch": 9.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9294,
|
|
"step": 17640
|
|
},
|
|
{
|
|
"epoch": 9.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9431,
|
|
"step": 17650
|
|
},
|
|
{
|
|
"epoch": 9.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 17660
|
|
},
|
|
{
|
|
"epoch": 9.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9592,
|
|
"step": 17670
|
|
},
|
|
{
|
|
"epoch": 9.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.941,
|
|
"step": 17680
|
|
},
|
|
{
|
|
"epoch": 9.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 17690
|
|
},
|
|
{
|
|
"epoch": 9.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9339,
|
|
"step": 17700
|
|
},
|
|
{
|
|
"epoch": 9.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9589,
|
|
"step": 17710
|
|
},
|
|
{
|
|
"epoch": 9.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9428,
|
|
"step": 17720
|
|
},
|
|
{
|
|
"epoch": 9.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9511,
|
|
"step": 17730
|
|
},
|
|
{
|
|
"epoch": 9.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9589,
|
|
"step": 17740
|
|
},
|
|
{
|
|
"epoch": 9.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9659,
|
|
"step": 17750
|
|
},
|
|
{
|
|
"epoch": 9.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9361,
|
|
"step": 17760
|
|
},
|
|
{
|
|
"epoch": 9.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 17770
|
|
},
|
|
{
|
|
"epoch": 9.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9396,
|
|
"step": 17780
|
|
},
|
|
{
|
|
"epoch": 9.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9517,
|
|
"step": 17790
|
|
},
|
|
{
|
|
"epoch": 9.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9323,
|
|
"step": 17800
|
|
},
|
|
{
|
|
"epoch": 9.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9315,
|
|
"step": 17810
|
|
},
|
|
{
|
|
"epoch": 9.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9702,
|
|
"step": 17820
|
|
},
|
|
{
|
|
"epoch": 9.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9314,
|
|
"step": 17830
|
|
},
|
|
{
|
|
"epoch": 9.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.95,
|
|
"step": 17840
|
|
},
|
|
{
|
|
"epoch": 9.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9556,
|
|
"step": 17850
|
|
},
|
|
{
|
|
"epoch": 9.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9196,
|
|
"step": 17860
|
|
},
|
|
{
|
|
"epoch": 9.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.942,
|
|
"step": 17870
|
|
},
|
|
{
|
|
"epoch": 9.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 17880
|
|
},
|
|
{
|
|
"epoch": 9.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9203,
|
|
"step": 17890
|
|
},
|
|
{
|
|
"epoch": 9.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9366,
|
|
"step": 17900
|
|
},
|
|
{
|
|
"epoch": 9.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9553,
|
|
"step": 17910
|
|
},
|
|
{
|
|
"epoch": 9.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9312,
|
|
"step": 17920
|
|
},
|
|
{
|
|
"epoch": 9.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.961,
|
|
"step": 17930
|
|
},
|
|
{
|
|
"epoch": 9.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9179,
|
|
"step": 17940
|
|
},
|
|
{
|
|
"epoch": 9.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9552,
|
|
"step": 17950
|
|
},
|
|
{
|
|
"epoch": 9.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.954,
|
|
"step": 17960
|
|
},
|
|
{
|
|
"epoch": 9.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 17970
|
|
},
|
|
{
|
|
"epoch": 9.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 17980
|
|
},
|
|
{
|
|
"epoch": 9.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 17990
|
|
},
|
|
{
|
|
"epoch": 9.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 18000
|
|
},
|
|
{
|
|
"epoch": 9.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9236,
|
|
"step": 18010
|
|
},
|
|
{
|
|
"epoch": 9.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9534,
|
|
"step": 18020
|
|
},
|
|
{
|
|
"epoch": 9.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9289,
|
|
"step": 18030
|
|
},
|
|
{
|
|
"epoch": 9.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 18040
|
|
},
|
|
{
|
|
"epoch": 9.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9638,
|
|
"step": 18050
|
|
},
|
|
{
|
|
"epoch": 9.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9324,
|
|
"step": 18060
|
|
},
|
|
{
|
|
"epoch": 9.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9397,
|
|
"step": 18070
|
|
},
|
|
{
|
|
"epoch": 9.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9335,
|
|
"step": 18080
|
|
},
|
|
{
|
|
"epoch": 9.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9607,
|
|
"step": 18090
|
|
},
|
|
{
|
|
"epoch": 9.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 18100
|
|
},
|
|
{
|
|
"epoch": 9.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9571,
|
|
"step": 18110
|
|
},
|
|
{
|
|
"epoch": 9.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9499,
|
|
"step": 18120
|
|
},
|
|
{
|
|
"epoch": 9.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 18130
|
|
},
|
|
{
|
|
"epoch": 9.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9346,
|
|
"step": 18140
|
|
},
|
|
{
|
|
"epoch": 9.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9305,
|
|
"step": 18150
|
|
},
|
|
{
|
|
"epoch": 9.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9567,
|
|
"step": 18160
|
|
},
|
|
{
|
|
"epoch": 9.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9457,
|
|
"step": 18170
|
|
},
|
|
{
|
|
"epoch": 9.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9571,
|
|
"step": 18180
|
|
},
|
|
{
|
|
"epoch": 9.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.945,
|
|
"step": 18190
|
|
},
|
|
{
|
|
"epoch": 9.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 18200
|
|
},
|
|
{
|
|
"epoch": 9.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.928,
|
|
"step": 18210
|
|
},
|
|
{
|
|
"epoch": 9.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.931,
|
|
"step": 18220
|
|
},
|
|
{
|
|
"epoch": 9.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9481,
|
|
"step": 18230
|
|
},
|
|
{
|
|
"epoch": 9.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9337,
|
|
"step": 18240
|
|
},
|
|
{
|
|
"epoch": 9.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9231,
|
|
"step": 18250
|
|
},
|
|
{
|
|
"epoch": 9.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9565,
|
|
"step": 18260
|
|
},
|
|
{
|
|
"epoch": 9.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9578,
|
|
"step": 18270
|
|
},
|
|
{
|
|
"epoch": 9.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9476,
|
|
"step": 18280
|
|
},
|
|
{
|
|
"epoch": 9.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9484,
|
|
"step": 18290
|
|
},
|
|
{
|
|
"epoch": 9.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9468,
|
|
"step": 18300
|
|
},
|
|
{
|
|
"epoch": 9.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9559,
|
|
"step": 18310
|
|
},
|
|
{
|
|
"epoch": 9.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9575,
|
|
"step": 18320
|
|
},
|
|
{
|
|
"epoch": 9.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9355,
|
|
"step": 18330
|
|
},
|
|
{
|
|
"epoch": 9.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9259,
|
|
"step": 18340
|
|
},
|
|
{
|
|
"epoch": 9.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9469,
|
|
"step": 18350
|
|
},
|
|
{
|
|
"epoch": 9.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9438,
|
|
"step": 18360
|
|
},
|
|
{
|
|
"epoch": 9.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9351,
|
|
"step": 18370
|
|
},
|
|
{
|
|
"epoch": 9.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9368,
|
|
"step": 18380
|
|
},
|
|
{
|
|
"epoch": 9.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 18390
|
|
},
|
|
{
|
|
"epoch": 9.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9395,
|
|
"step": 18400
|
|
},
|
|
{
|
|
"epoch": 9.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9375,
|
|
"step": 18410
|
|
},
|
|
{
|
|
"epoch": 9.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 18420
|
|
},
|
|
{
|
|
"epoch": 9.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9414,
|
|
"step": 18430
|
|
},
|
|
{
|
|
"epoch": 9.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.945,
|
|
"step": 18440
|
|
},
|
|
{
|
|
"epoch": 9.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9527,
|
|
"step": 18450
|
|
},
|
|
{
|
|
"epoch": 9.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9531,
|
|
"step": 18460
|
|
},
|
|
{
|
|
"epoch": 9.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9273,
|
|
"step": 18470
|
|
},
|
|
{
|
|
"epoch": 9.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9357,
|
|
"step": 18480
|
|
},
|
|
{
|
|
"epoch": 9.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9472,
|
|
"step": 18490
|
|
},
|
|
{
|
|
"epoch": 9.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9382,
|
|
"step": 18500
|
|
},
|
|
{
|
|
"epoch": 9.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9709,
|
|
"step": 18510
|
|
},
|
|
{
|
|
"epoch": 9.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9406,
|
|
"step": 18520
|
|
},
|
|
{
|
|
"epoch": 9.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 18530
|
|
},
|
|
{
|
|
"epoch": 9.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9665,
|
|
"step": 18540
|
|
},
|
|
{
|
|
"epoch": 9.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9424,
|
|
"step": 18550
|
|
},
|
|
{
|
|
"epoch": 9.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9356,
|
|
"step": 18560
|
|
},
|
|
{
|
|
"epoch": 9.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9364,
|
|
"step": 18570
|
|
},
|
|
{
|
|
"epoch": 9.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9446,
|
|
"step": 18580
|
|
},
|
|
{
|
|
"epoch": 9.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9441,
|
|
"step": 18590
|
|
},
|
|
{
|
|
"epoch": 9.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 18600
|
|
},
|
|
{
|
|
"epoch": 9.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 18610
|
|
},
|
|
{
|
|
"epoch": 9.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.943,
|
|
"step": 18620
|
|
},
|
|
{
|
|
"epoch": 9.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9546,
|
|
"step": 18630
|
|
},
|
|
{
|
|
"epoch": 9.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9331,
|
|
"step": 18640
|
|
},
|
|
{
|
|
"epoch": 9.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9589,
|
|
"step": 18650
|
|
},
|
|
{
|
|
"epoch": 9.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.944,
|
|
"step": 18660
|
|
},
|
|
{
|
|
"epoch": 9.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9336,
|
|
"step": 18670
|
|
},
|
|
{
|
|
"epoch": 9.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 18680
|
|
},
|
|
{
|
|
"epoch": 9.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9391,
|
|
"step": 18690
|
|
},
|
|
{
|
|
"epoch": 9.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9399,
|
|
"step": 18700
|
|
},
|
|
{
|
|
"epoch": 9.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 18710
|
|
},
|
|
{
|
|
"epoch": 9.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 18720
|
|
},
|
|
{
|
|
"epoch": 9.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9354,
|
|
"step": 18730
|
|
},
|
|
{
|
|
"epoch": 9.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9504,
|
|
"step": 18740
|
|
},
|
|
{
|
|
"epoch": 9.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9397,
|
|
"step": 18750
|
|
},
|
|
{
|
|
"epoch": 9.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9357,
|
|
"step": 18760
|
|
},
|
|
{
|
|
"epoch": 9.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9312,
|
|
"step": 18770
|
|
},
|
|
{
|
|
"epoch": 9.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 18780
|
|
},
|
|
{
|
|
"epoch": 9.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9472,
|
|
"step": 18790
|
|
},
|
|
{
|
|
"epoch": 9.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.938,
|
|
"step": 18800
|
|
},
|
|
{
|
|
"epoch": 9.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 18810
|
|
},
|
|
{
|
|
"epoch": 9.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.955,
|
|
"step": 18820
|
|
},
|
|
{
|
|
"epoch": 9.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9407,
|
|
"step": 18830
|
|
},
|
|
{
|
|
"epoch": 9.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9521,
|
|
"step": 18840
|
|
},
|
|
{
|
|
"epoch": 9.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 18850
|
|
},
|
|
{
|
|
"epoch": 9.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 18860
|
|
},
|
|
{
|
|
"epoch": 9.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9232,
|
|
"step": 18870
|
|
},
|
|
{
|
|
"epoch": 9.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9007,
|
|
"step": 18880
|
|
},
|
|
{
|
|
"epoch": 9.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9883,
|
|
"step": 18890
|
|
},
|
|
{
|
|
"epoch": 9.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9389,
|
|
"step": 18900
|
|
},
|
|
{
|
|
"epoch": 9.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9447,
|
|
"step": 18910
|
|
},
|
|
{
|
|
"epoch": 9.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9359,
|
|
"step": 18920
|
|
},
|
|
{
|
|
"epoch": 9.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9531,
|
|
"step": 18930
|
|
},
|
|
{
|
|
"epoch": 9.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9393,
|
|
"step": 18940
|
|
},
|
|
{
|
|
"epoch": 9.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9566,
|
|
"step": 18950
|
|
},
|
|
{
|
|
"epoch": 9.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9486,
|
|
"step": 18960
|
|
},
|
|
{
|
|
"epoch": 9.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9516,
|
|
"step": 18970
|
|
},
|
|
{
|
|
"epoch": 9.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9436,
|
|
"step": 18980
|
|
},
|
|
{
|
|
"epoch": 9.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9397,
|
|
"step": 18990
|
|
},
|
|
{
|
|
"epoch": 9.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9485,
|
|
"step": 19000
|
|
},
|
|
{
|
|
"epoch": 9.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9448,
|
|
"step": 19010
|
|
},
|
|
{
|
|
"epoch": 9.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 19020
|
|
},
|
|
{
|
|
"epoch": 9.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 19030
|
|
},
|
|
{
|
|
"epoch": 9.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9423,
|
|
"step": 19040
|
|
},
|
|
{
|
|
"epoch": 9.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9369,
|
|
"step": 19050
|
|
},
|
|
{
|
|
"epoch": 9.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 19060
|
|
},
|
|
{
|
|
"epoch": 9.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9458,
|
|
"step": 19070
|
|
},
|
|
{
|
|
"epoch": 9.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9652,
|
|
"step": 19080
|
|
},
|
|
{
|
|
"epoch": 9.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9461,
|
|
"step": 19090
|
|
},
|
|
{
|
|
"epoch": 9.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9458,
|
|
"step": 19100
|
|
},
|
|
{
|
|
"epoch": 9.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9449,
|
|
"step": 19110
|
|
},
|
|
{
|
|
"epoch": 9.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9425,
|
|
"step": 19120
|
|
},
|
|
{
|
|
"epoch": 9.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9383,
|
|
"step": 19130
|
|
},
|
|
{
|
|
"epoch": 9.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9331,
|
|
"step": 19140
|
|
},
|
|
{
|
|
"epoch": 9.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9462,
|
|
"step": 19150
|
|
},
|
|
{
|
|
"epoch": 9.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9444,
|
|
"step": 19160
|
|
},
|
|
{
|
|
"epoch": 9.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9565,
|
|
"step": 19170
|
|
},
|
|
{
|
|
"epoch": 9.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9569,
|
|
"step": 19180
|
|
},
|
|
{
|
|
"epoch": 9.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9501,
|
|
"step": 19190
|
|
},
|
|
{
|
|
"epoch": 9.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9469,
|
|
"step": 19200
|
|
},
|
|
{
|
|
"epoch": 9.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9326,
|
|
"step": 19210
|
|
},
|
|
{
|
|
"epoch": 9.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9363,
|
|
"step": 19220
|
|
},
|
|
{
|
|
"epoch": 9.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9545,
|
|
"step": 19230
|
|
},
|
|
{
|
|
"epoch": 9.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9427,
|
|
"step": 19240
|
|
},
|
|
{
|
|
"epoch": 9.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9404,
|
|
"step": 19250
|
|
},
|
|
{
|
|
"epoch": 9.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9498,
|
|
"step": 19260
|
|
},
|
|
{
|
|
"epoch": 10.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 19270
|
|
},
|
|
{
|
|
"epoch": 10.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.1556420233463035
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03848003848003848
|
|
},
|
|
"eval_loss": 1.9412217140197754,
|
|
"eval_precision": {
|
|
"precision": 0.022234574763757644
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 75.9999,
|
|
"eval_samples_per_second": 50.724,
|
|
"eval_steps_per_second": 25.368,
|
|
"step": 19270
|
|
},
|
|
{
|
|
"epoch": 10.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9594,
|
|
"step": 19280
|
|
},
|
|
{
|
|
"epoch": 10.01,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9263,
|
|
"step": 19290
|
|
},
|
|
{
|
|
"epoch": 10.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9506,
|
|
"step": 19300
|
|
},
|
|
{
|
|
"epoch": 10.02,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9357,
|
|
"step": 19310
|
|
},
|
|
{
|
|
"epoch": 10.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9277,
|
|
"step": 19320
|
|
},
|
|
{
|
|
"epoch": 10.03,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9428,
|
|
"step": 19330
|
|
},
|
|
{
|
|
"epoch": 10.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9209,
|
|
"step": 19340
|
|
},
|
|
{
|
|
"epoch": 10.04,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9331,
|
|
"step": 19350
|
|
},
|
|
{
|
|
"epoch": 10.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9599,
|
|
"step": 19360
|
|
},
|
|
{
|
|
"epoch": 10.05,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 19370
|
|
},
|
|
{
|
|
"epoch": 10.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9291,
|
|
"step": 19380
|
|
},
|
|
{
|
|
"epoch": 10.06,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9421,
|
|
"step": 19390
|
|
},
|
|
{
|
|
"epoch": 10.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.942,
|
|
"step": 19400
|
|
},
|
|
{
|
|
"epoch": 10.07,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9483,
|
|
"step": 19410
|
|
},
|
|
{
|
|
"epoch": 10.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9547,
|
|
"step": 19420
|
|
},
|
|
{
|
|
"epoch": 10.08,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9336,
|
|
"step": 19430
|
|
},
|
|
{
|
|
"epoch": 10.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 19440
|
|
},
|
|
{
|
|
"epoch": 10.09,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9599,
|
|
"step": 19450
|
|
},
|
|
{
|
|
"epoch": 10.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.934,
|
|
"step": 19460
|
|
},
|
|
{
|
|
"epoch": 10.1,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9107,
|
|
"step": 19470
|
|
},
|
|
{
|
|
"epoch": 10.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9662,
|
|
"step": 19480
|
|
},
|
|
{
|
|
"epoch": 10.11,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 19490
|
|
},
|
|
{
|
|
"epoch": 10.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9378,
|
|
"step": 19500
|
|
},
|
|
{
|
|
"epoch": 10.12,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9485,
|
|
"step": 19510
|
|
},
|
|
{
|
|
"epoch": 10.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9339,
|
|
"step": 19520
|
|
},
|
|
{
|
|
"epoch": 10.13,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9335,
|
|
"step": 19530
|
|
},
|
|
{
|
|
"epoch": 10.14,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9132,
|
|
"step": 19540
|
|
},
|
|
{
|
|
"epoch": 10.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9616,
|
|
"step": 19550
|
|
},
|
|
{
|
|
"epoch": 10.15,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9411,
|
|
"step": 19560
|
|
},
|
|
{
|
|
"epoch": 10.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9383,
|
|
"step": 19570
|
|
},
|
|
{
|
|
"epoch": 10.16,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9402,
|
|
"step": 19580
|
|
},
|
|
{
|
|
"epoch": 10.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.957,
|
|
"step": 19590
|
|
},
|
|
{
|
|
"epoch": 10.17,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9637,
|
|
"step": 19600
|
|
},
|
|
{
|
|
"epoch": 10.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9462,
|
|
"step": 19610
|
|
},
|
|
{
|
|
"epoch": 10.18,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9471,
|
|
"step": 19620
|
|
},
|
|
{
|
|
"epoch": 10.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9508,
|
|
"step": 19630
|
|
},
|
|
{
|
|
"epoch": 10.19,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9522,
|
|
"step": 19640
|
|
},
|
|
{
|
|
"epoch": 10.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9327,
|
|
"step": 19650
|
|
},
|
|
{
|
|
"epoch": 10.2,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9171,
|
|
"step": 19660
|
|
},
|
|
{
|
|
"epoch": 10.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9613,
|
|
"step": 19670
|
|
},
|
|
{
|
|
"epoch": 10.21,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.948,
|
|
"step": 19680
|
|
},
|
|
{
|
|
"epoch": 10.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9323,
|
|
"step": 19690
|
|
},
|
|
{
|
|
"epoch": 10.22,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9516,
|
|
"step": 19700
|
|
},
|
|
{
|
|
"epoch": 10.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9543,
|
|
"step": 19710
|
|
},
|
|
{
|
|
"epoch": 10.23,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9331,
|
|
"step": 19720
|
|
},
|
|
{
|
|
"epoch": 10.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9398,
|
|
"step": 19730
|
|
},
|
|
{
|
|
"epoch": 10.24,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9495,
|
|
"step": 19740
|
|
},
|
|
{
|
|
"epoch": 10.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9566,
|
|
"step": 19750
|
|
},
|
|
{
|
|
"epoch": 10.25,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9257,
|
|
"step": 19760
|
|
},
|
|
{
|
|
"epoch": 10.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9337,
|
|
"step": 19770
|
|
},
|
|
{
|
|
"epoch": 10.26,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9321,
|
|
"step": 19780
|
|
},
|
|
{
|
|
"epoch": 10.27,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9406,
|
|
"step": 19790
|
|
},
|
|
{
|
|
"epoch": 10.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9379,
|
|
"step": 19800
|
|
},
|
|
{
|
|
"epoch": 10.28,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9395,
|
|
"step": 19810
|
|
},
|
|
{
|
|
"epoch": 10.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9452,
|
|
"step": 19820
|
|
},
|
|
{
|
|
"epoch": 10.29,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 19830
|
|
},
|
|
{
|
|
"epoch": 10.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9381,
|
|
"step": 19840
|
|
},
|
|
{
|
|
"epoch": 10.3,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 19850
|
|
},
|
|
{
|
|
"epoch": 10.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9374,
|
|
"step": 19860
|
|
},
|
|
{
|
|
"epoch": 10.31,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 19870
|
|
},
|
|
{
|
|
"epoch": 10.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 19880
|
|
},
|
|
{
|
|
"epoch": 10.32,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9609,
|
|
"step": 19890
|
|
},
|
|
{
|
|
"epoch": 10.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9392,
|
|
"step": 19900
|
|
},
|
|
{
|
|
"epoch": 10.33,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.941,
|
|
"step": 19910
|
|
},
|
|
{
|
|
"epoch": 10.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9467,
|
|
"step": 19920
|
|
},
|
|
{
|
|
"epoch": 10.34,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 19930
|
|
},
|
|
{
|
|
"epoch": 10.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9298,
|
|
"step": 19940
|
|
},
|
|
{
|
|
"epoch": 10.35,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9541,
|
|
"step": 19950
|
|
},
|
|
{
|
|
"epoch": 10.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9394,
|
|
"step": 19960
|
|
},
|
|
{
|
|
"epoch": 10.36,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9281,
|
|
"step": 19970
|
|
},
|
|
{
|
|
"epoch": 10.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9645,
|
|
"step": 19980
|
|
},
|
|
{
|
|
"epoch": 10.37,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9263,
|
|
"step": 19990
|
|
},
|
|
{
|
|
"epoch": 10.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9417,
|
|
"step": 20000
|
|
},
|
|
{
|
|
"epoch": 10.38,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9408,
|
|
"step": 20010
|
|
},
|
|
{
|
|
"epoch": 10.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9463,
|
|
"step": 20020
|
|
},
|
|
{
|
|
"epoch": 10.39,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9558,
|
|
"step": 20030
|
|
},
|
|
{
|
|
"epoch": 10.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9387,
|
|
"step": 20040
|
|
},
|
|
{
|
|
"epoch": 10.4,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9412,
|
|
"step": 20050
|
|
},
|
|
{
|
|
"epoch": 10.41,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9443,
|
|
"step": 20060
|
|
},
|
|
{
|
|
"epoch": 10.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9509,
|
|
"step": 20070
|
|
},
|
|
{
|
|
"epoch": 10.42,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 20080
|
|
},
|
|
{
|
|
"epoch": 10.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9493,
|
|
"step": 20090
|
|
},
|
|
{
|
|
"epoch": 10.43,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.962,
|
|
"step": 20100
|
|
},
|
|
{
|
|
"epoch": 10.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9306,
|
|
"step": 20110
|
|
},
|
|
{
|
|
"epoch": 10.44,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9484,
|
|
"step": 20120
|
|
},
|
|
{
|
|
"epoch": 10.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9376,
|
|
"step": 20130
|
|
},
|
|
{
|
|
"epoch": 10.45,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9303,
|
|
"step": 20140
|
|
},
|
|
{
|
|
"epoch": 10.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9446,
|
|
"step": 20150
|
|
},
|
|
{
|
|
"epoch": 10.46,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9453,
|
|
"step": 20160
|
|
},
|
|
{
|
|
"epoch": 10.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9417,
|
|
"step": 20170
|
|
},
|
|
{
|
|
"epoch": 10.47,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9415,
|
|
"step": 20180
|
|
},
|
|
{
|
|
"epoch": 10.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9432,
|
|
"step": 20190
|
|
},
|
|
{
|
|
"epoch": 10.48,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9318,
|
|
"step": 20200
|
|
},
|
|
{
|
|
"epoch": 10.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9515,
|
|
"step": 20210
|
|
},
|
|
{
|
|
"epoch": 10.49,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9455,
|
|
"step": 20220
|
|
},
|
|
{
|
|
"epoch": 10.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9506,
|
|
"step": 20230
|
|
},
|
|
{
|
|
"epoch": 10.5,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9278,
|
|
"step": 20240
|
|
},
|
|
{
|
|
"epoch": 10.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9516,
|
|
"step": 20250
|
|
},
|
|
{
|
|
"epoch": 10.51,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9507,
|
|
"step": 20260
|
|
},
|
|
{
|
|
"epoch": 10.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 20270
|
|
},
|
|
{
|
|
"epoch": 10.52,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9435,
|
|
"step": 20280
|
|
},
|
|
{
|
|
"epoch": 10.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9247,
|
|
"step": 20290
|
|
},
|
|
{
|
|
"epoch": 10.53,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9336,
|
|
"step": 20300
|
|
},
|
|
{
|
|
"epoch": 10.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9583,
|
|
"step": 20310
|
|
},
|
|
{
|
|
"epoch": 10.54,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9645,
|
|
"step": 20320
|
|
},
|
|
{
|
|
"epoch": 10.55,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9364,
|
|
"step": 20330
|
|
},
|
|
{
|
|
"epoch": 10.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9544,
|
|
"step": 20340
|
|
},
|
|
{
|
|
"epoch": 10.56,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9616,
|
|
"step": 20350
|
|
},
|
|
{
|
|
"epoch": 10.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9309,
|
|
"step": 20360
|
|
},
|
|
{
|
|
"epoch": 10.57,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9422,
|
|
"step": 20370
|
|
},
|
|
{
|
|
"epoch": 10.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9317,
|
|
"step": 20380
|
|
},
|
|
{
|
|
"epoch": 10.58,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 20390
|
|
},
|
|
{
|
|
"epoch": 10.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9578,
|
|
"step": 20400
|
|
},
|
|
{
|
|
"epoch": 10.59,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9524,
|
|
"step": 20410
|
|
},
|
|
{
|
|
"epoch": 10.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9596,
|
|
"step": 20420
|
|
},
|
|
{
|
|
"epoch": 10.6,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9371,
|
|
"step": 20430
|
|
},
|
|
{
|
|
"epoch": 10.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9494,
|
|
"step": 20440
|
|
},
|
|
{
|
|
"epoch": 10.61,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9418,
|
|
"step": 20450
|
|
},
|
|
{
|
|
"epoch": 10.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9476,
|
|
"step": 20460
|
|
},
|
|
{
|
|
"epoch": 10.62,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9437,
|
|
"step": 20470
|
|
},
|
|
{
|
|
"epoch": 10.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9313,
|
|
"step": 20480
|
|
},
|
|
{
|
|
"epoch": 10.63,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9641,
|
|
"step": 20490
|
|
},
|
|
{
|
|
"epoch": 10.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9414,
|
|
"step": 20500
|
|
},
|
|
{
|
|
"epoch": 10.64,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9533,
|
|
"step": 20510
|
|
},
|
|
{
|
|
"epoch": 10.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9382,
|
|
"step": 20520
|
|
},
|
|
{
|
|
"epoch": 10.65,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9248,
|
|
"step": 20530
|
|
},
|
|
{
|
|
"epoch": 10.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9628,
|
|
"step": 20540
|
|
},
|
|
{
|
|
"epoch": 10.66,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.936,
|
|
"step": 20550
|
|
},
|
|
{
|
|
"epoch": 10.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9479,
|
|
"step": 20560
|
|
},
|
|
{
|
|
"epoch": 10.67,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9361,
|
|
"step": 20570
|
|
},
|
|
{
|
|
"epoch": 10.68,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.939,
|
|
"step": 20580
|
|
},
|
|
{
|
|
"epoch": 10.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9456,
|
|
"step": 20590
|
|
},
|
|
{
|
|
"epoch": 10.69,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9257,
|
|
"step": 20600
|
|
},
|
|
{
|
|
"epoch": 10.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9487,
|
|
"step": 20610
|
|
},
|
|
{
|
|
"epoch": 10.7,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9443,
|
|
"step": 20620
|
|
},
|
|
{
|
|
"epoch": 10.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9505,
|
|
"step": 20630
|
|
},
|
|
{
|
|
"epoch": 10.71,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 20640
|
|
},
|
|
{
|
|
"epoch": 10.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9517,
|
|
"step": 20650
|
|
},
|
|
{
|
|
"epoch": 10.72,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9434,
|
|
"step": 20660
|
|
},
|
|
{
|
|
"epoch": 10.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.929,
|
|
"step": 20670
|
|
},
|
|
{
|
|
"epoch": 10.73,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9498,
|
|
"step": 20680
|
|
},
|
|
{
|
|
"epoch": 10.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9563,
|
|
"step": 20690
|
|
},
|
|
{
|
|
"epoch": 10.74,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9561,
|
|
"step": 20700
|
|
},
|
|
{
|
|
"epoch": 10.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9395,
|
|
"step": 20710
|
|
},
|
|
{
|
|
"epoch": 10.75,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 20720
|
|
},
|
|
{
|
|
"epoch": 10.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9428,
|
|
"step": 20730
|
|
},
|
|
{
|
|
"epoch": 10.76,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9478,
|
|
"step": 20740
|
|
},
|
|
{
|
|
"epoch": 10.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9351,
|
|
"step": 20750
|
|
},
|
|
{
|
|
"epoch": 10.77,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9532,
|
|
"step": 20760
|
|
},
|
|
{
|
|
"epoch": 10.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.949,
|
|
"step": 20770
|
|
},
|
|
{
|
|
"epoch": 10.78,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9433,
|
|
"step": 20780
|
|
},
|
|
{
|
|
"epoch": 10.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9598,
|
|
"step": 20790
|
|
},
|
|
{
|
|
"epoch": 10.79,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9395,
|
|
"step": 20800
|
|
},
|
|
{
|
|
"epoch": 10.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9433,
|
|
"step": 20810
|
|
},
|
|
{
|
|
"epoch": 10.8,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9621,
|
|
"step": 20820
|
|
},
|
|
{
|
|
"epoch": 10.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9455,
|
|
"step": 20830
|
|
},
|
|
{
|
|
"epoch": 10.81,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9298,
|
|
"step": 20840
|
|
},
|
|
{
|
|
"epoch": 10.82,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.951,
|
|
"step": 20850
|
|
},
|
|
{
|
|
"epoch": 10.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9578,
|
|
"step": 20860
|
|
},
|
|
{
|
|
"epoch": 10.83,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.947,
|
|
"step": 20870
|
|
},
|
|
{
|
|
"epoch": 10.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9293,
|
|
"step": 20880
|
|
},
|
|
{
|
|
"epoch": 10.84,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9277,
|
|
"step": 20890
|
|
},
|
|
{
|
|
"epoch": 10.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9464,
|
|
"step": 20900
|
|
},
|
|
{
|
|
"epoch": 10.85,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9503,
|
|
"step": 20910
|
|
},
|
|
{
|
|
"epoch": 10.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9323,
|
|
"step": 20920
|
|
},
|
|
{
|
|
"epoch": 10.86,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9365,
|
|
"step": 20930
|
|
},
|
|
{
|
|
"epoch": 10.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9419,
|
|
"step": 20940
|
|
},
|
|
{
|
|
"epoch": 10.87,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.942,
|
|
"step": 20950
|
|
},
|
|
{
|
|
"epoch": 10.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9552,
|
|
"step": 20960
|
|
},
|
|
{
|
|
"epoch": 10.88,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9278,
|
|
"step": 20970
|
|
},
|
|
{
|
|
"epoch": 10.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9429,
|
|
"step": 20980
|
|
},
|
|
{
|
|
"epoch": 10.89,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9473,
|
|
"step": 20990
|
|
},
|
|
{
|
|
"epoch": 10.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9576,
|
|
"step": 21000
|
|
},
|
|
{
|
|
"epoch": 10.9,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.94,
|
|
"step": 21010
|
|
},
|
|
{
|
|
"epoch": 10.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9302,
|
|
"step": 21020
|
|
},
|
|
{
|
|
"epoch": 10.91,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9369,
|
|
"step": 21030
|
|
},
|
|
{
|
|
"epoch": 10.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9396,
|
|
"step": 21040
|
|
},
|
|
{
|
|
"epoch": 10.92,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9296,
|
|
"step": 21050
|
|
},
|
|
{
|
|
"epoch": 10.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9538,
|
|
"step": 21060
|
|
},
|
|
{
|
|
"epoch": 10.93,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9471,
|
|
"step": 21070
|
|
},
|
|
{
|
|
"epoch": 10.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9442,
|
|
"step": 21080
|
|
},
|
|
{
|
|
"epoch": 10.94,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9454,
|
|
"step": 21090
|
|
},
|
|
{
|
|
"epoch": 10.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9281,
|
|
"step": 21100
|
|
},
|
|
{
|
|
"epoch": 10.95,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9646,
|
|
"step": 21110
|
|
},
|
|
{
|
|
"epoch": 10.96,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9283,
|
|
"step": 21120
|
|
},
|
|
{
|
|
"epoch": 10.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9508,
|
|
"step": 21130
|
|
},
|
|
{
|
|
"epoch": 10.97,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9648,
|
|
"step": 21140
|
|
},
|
|
{
|
|
"epoch": 10.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9391,
|
|
"step": 21150
|
|
},
|
|
{
|
|
"epoch": 10.98,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9556,
|
|
"step": 21160
|
|
},
|
|
{
|
|
"epoch": 10.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.9528,
|
|
"step": 21170
|
|
},
|
|
{
|
|
"epoch": 10.99,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.934,
|
|
"step": 21180
|
|
},
|
|
{
|
|
"epoch": 11.0,
|
|
"learning_rate": 0.001,
|
|
"loss": 1.945,
|
|
"step": 21190
|
|
},
|
|
{
|
|
"epoch": 11.0,
|
|
"eval_accuracy": {
|
|
"accuracy": 0.1556420233463035
|
|
},
|
|
"eval_f1": {
|
|
"f1": 0.03848003848003848
|
|
},
|
|
"eval_loss": 1.943160057067871,
|
|
"eval_precision": {
|
|
"precision": 0.022234574763757644
|
|
},
|
|
"eval_recall": {
|
|
"recall": 0.14285714285714285
|
|
},
|
|
"eval_runtime": 75.9934,
|
|
"eval_samples_per_second": 50.728,
|
|
"eval_steps_per_second": 25.371,
|
|
"step": 21197
|
|
}
|
|
],
|
|
"logging_steps": 10,
|
|
"max_steps": 57810,
|
|
"num_input_tokens_seen": 0,
|
|
"num_train_epochs": 30,
|
|
"save_steps": 500,
|
|
"total_flos": 3.6721287534351974e+18,
|
|
"train_batch_size": 2,
|
|
"trial_name": null,
|
|
"trial_params": null
|
|
}
|
|
|