organamnist-beit-base-finetuned / trainer_state.json
selmamalak's picture
End of training
08a6f39 verified
raw
history blame contribute delete
No virus
90.8 kB
{
"best_metric": 0.986134647974118,
"best_model_checkpoint": "beit-base-patch16-224-pt22k-ft22k-finetuned-lora-medmnistv2/checkpoint-4864",
"epoch": 9.990749306197966,
"eval_steps": 500,
"global_step": 5400,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02,
"grad_norm": 3.66642165184021,
"learning_rate": 0.004990740740740741,
"loss": 1.7617,
"step": 10
},
{
"epoch": 0.04,
"grad_norm": 5.545452117919922,
"learning_rate": 0.004982407407407408,
"loss": 1.1989,
"step": 20
},
{
"epoch": 0.06,
"grad_norm": 5.292081356048584,
"learning_rate": 0.004973148148148148,
"loss": 1.0735,
"step": 30
},
{
"epoch": 0.07,
"grad_norm": 3.0206360816955566,
"learning_rate": 0.004963888888888889,
"loss": 1.1936,
"step": 40
},
{
"epoch": 0.09,
"grad_norm": 3.947781562805176,
"learning_rate": 0.004954629629629629,
"loss": 1.0373,
"step": 50
},
{
"epoch": 0.11,
"grad_norm": 1.839845895767212,
"learning_rate": 0.004945370370370371,
"loss": 0.9221,
"step": 60
},
{
"epoch": 0.13,
"grad_norm": 2.843251943588257,
"learning_rate": 0.004936111111111112,
"loss": 0.8724,
"step": 70
},
{
"epoch": 0.15,
"grad_norm": 3.0122690200805664,
"learning_rate": 0.0049268518518518515,
"loss": 0.9201,
"step": 80
},
{
"epoch": 0.17,
"grad_norm": 3.403461456298828,
"learning_rate": 0.004917592592592593,
"loss": 0.9221,
"step": 90
},
{
"epoch": 0.19,
"grad_norm": 2.4073173999786377,
"learning_rate": 0.004908333333333333,
"loss": 0.8766,
"step": 100
},
{
"epoch": 0.2,
"grad_norm": 1.793009877204895,
"learning_rate": 0.004899074074074074,
"loss": 0.8742,
"step": 110
},
{
"epoch": 0.22,
"grad_norm": 1.8483041524887085,
"learning_rate": 0.004889814814814815,
"loss": 0.8097,
"step": 120
},
{
"epoch": 0.24,
"grad_norm": 1.997031807899475,
"learning_rate": 0.0048805555555555555,
"loss": 0.888,
"step": 130
},
{
"epoch": 0.26,
"grad_norm": 3.958829641342163,
"learning_rate": 0.004871296296296296,
"loss": 0.8276,
"step": 140
},
{
"epoch": 0.28,
"grad_norm": 2.8835649490356445,
"learning_rate": 0.004862037037037037,
"loss": 0.8469,
"step": 150
},
{
"epoch": 0.3,
"grad_norm": 2.2343573570251465,
"learning_rate": 0.004852777777777778,
"loss": 0.8354,
"step": 160
},
{
"epoch": 0.31,
"grad_norm": 1.8973902463912964,
"learning_rate": 0.004843518518518519,
"loss": 0.7974,
"step": 170
},
{
"epoch": 0.33,
"grad_norm": 2.303403854370117,
"learning_rate": 0.0048342592592592595,
"loss": 0.7531,
"step": 180
},
{
"epoch": 0.35,
"grad_norm": 2.833897352218628,
"learning_rate": 0.004825,
"loss": 0.7917,
"step": 190
},
{
"epoch": 0.37,
"grad_norm": 2.833991527557373,
"learning_rate": 0.004815740740740741,
"loss": 0.9165,
"step": 200
},
{
"epoch": 0.39,
"grad_norm": 2.109811305999756,
"learning_rate": 0.004806481481481482,
"loss": 0.7347,
"step": 210
},
{
"epoch": 0.41,
"grad_norm": 4.833127498626709,
"learning_rate": 0.004797222222222223,
"loss": 0.8852,
"step": 220
},
{
"epoch": 0.43,
"grad_norm": 6.393899440765381,
"learning_rate": 0.004787962962962963,
"loss": 0.7416,
"step": 230
},
{
"epoch": 0.44,
"grad_norm": 2.243924856185913,
"learning_rate": 0.004778703703703704,
"loss": 0.7616,
"step": 240
},
{
"epoch": 0.46,
"grad_norm": 2.55934739112854,
"learning_rate": 0.004769444444444444,
"loss": 0.6881,
"step": 250
},
{
"epoch": 0.48,
"grad_norm": 2.707200765609741,
"learning_rate": 0.004760185185185185,
"loss": 0.7597,
"step": 260
},
{
"epoch": 0.5,
"grad_norm": 3.238736867904663,
"learning_rate": 0.004750925925925926,
"loss": 0.6598,
"step": 270
},
{
"epoch": 0.52,
"grad_norm": 1.6335750818252563,
"learning_rate": 0.004741666666666667,
"loss": 0.6988,
"step": 280
},
{
"epoch": 0.54,
"grad_norm": 1.6415876150131226,
"learning_rate": 0.004732407407407407,
"loss": 0.7615,
"step": 290
},
{
"epoch": 0.56,
"grad_norm": 2.09512996673584,
"learning_rate": 0.004723148148148148,
"loss": 0.7345,
"step": 300
},
{
"epoch": 0.57,
"grad_norm": 2.1141433715820312,
"learning_rate": 0.004713888888888889,
"loss": 0.7503,
"step": 310
},
{
"epoch": 0.59,
"grad_norm": 1.1782424449920654,
"learning_rate": 0.00470462962962963,
"loss": 0.6997,
"step": 320
},
{
"epoch": 0.61,
"grad_norm": 2.7585694789886475,
"learning_rate": 0.004695370370370371,
"loss": 0.7703,
"step": 330
},
{
"epoch": 0.63,
"grad_norm": 1.3906846046447754,
"learning_rate": 0.004686111111111111,
"loss": 0.6595,
"step": 340
},
{
"epoch": 0.65,
"grad_norm": 1.9824657440185547,
"learning_rate": 0.004676851851851852,
"loss": 0.612,
"step": 350
},
{
"epoch": 0.67,
"grad_norm": 3.896270990371704,
"learning_rate": 0.004667592592592593,
"loss": 0.8571,
"step": 360
},
{
"epoch": 0.68,
"grad_norm": 1.860967993736267,
"learning_rate": 0.004658333333333333,
"loss": 0.774,
"step": 370
},
{
"epoch": 0.7,
"grad_norm": 1.4278061389923096,
"learning_rate": 0.004649074074074074,
"loss": 0.7693,
"step": 380
},
{
"epoch": 0.72,
"grad_norm": 4.371230125427246,
"learning_rate": 0.004639814814814815,
"loss": 0.6989,
"step": 390
},
{
"epoch": 0.74,
"grad_norm": 1.6411677598953247,
"learning_rate": 0.004630555555555555,
"loss": 0.8566,
"step": 400
},
{
"epoch": 0.76,
"grad_norm": 3.412655830383301,
"learning_rate": 0.004621296296296296,
"loss": 0.6902,
"step": 410
},
{
"epoch": 0.78,
"grad_norm": 1.9510608911514282,
"learning_rate": 0.004612037037037038,
"loss": 0.8284,
"step": 420
},
{
"epoch": 0.8,
"grad_norm": 1.8800582885742188,
"learning_rate": 0.004602777777777778,
"loss": 0.6096,
"step": 430
},
{
"epoch": 0.81,
"grad_norm": 2.7876181602478027,
"learning_rate": 0.0045935185185185185,
"loss": 0.7293,
"step": 440
},
{
"epoch": 0.83,
"grad_norm": 2.0592384338378906,
"learning_rate": 0.004584259259259259,
"loss": 0.756,
"step": 450
},
{
"epoch": 0.85,
"grad_norm": 3.5372824668884277,
"learning_rate": 0.004575,
"loss": 0.6865,
"step": 460
},
{
"epoch": 0.87,
"grad_norm": 1.7686892747879028,
"learning_rate": 0.004565740740740741,
"loss": 0.7477,
"step": 470
},
{
"epoch": 0.89,
"grad_norm": 2.8619141578674316,
"learning_rate": 0.004556481481481482,
"loss": 0.6861,
"step": 480
},
{
"epoch": 0.91,
"grad_norm": 2.4887492656707764,
"learning_rate": 0.0045472222222222225,
"loss": 0.7623,
"step": 490
},
{
"epoch": 0.93,
"grad_norm": 1.9150561094284058,
"learning_rate": 0.004537962962962963,
"loss": 0.7462,
"step": 500
},
{
"epoch": 0.94,
"grad_norm": 2.1251327991485596,
"learning_rate": 0.004528703703703704,
"loss": 0.673,
"step": 510
},
{
"epoch": 0.96,
"grad_norm": 2.338467597961426,
"learning_rate": 0.004519444444444444,
"loss": 0.6898,
"step": 520
},
{
"epoch": 0.98,
"grad_norm": 2.199749708175659,
"learning_rate": 0.004510185185185186,
"loss": 0.6776,
"step": 530
},
{
"epoch": 1.0,
"grad_norm": 1.5489593744277954,
"learning_rate": 0.0045009259259259264,
"loss": 0.6786,
"step": 540
},
{
"epoch": 1.0,
"eval_accuracy": 0.9339084886766292,
"eval_f1": 0.938537564429046,
"eval_loss": 0.17764028906822205,
"eval_precision": 0.9507286953627705,
"eval_recall": 0.934120062770795,
"eval_runtime": 54.7942,
"eval_samples_per_second": 118.462,
"eval_steps_per_second": 7.41,
"step": 540
},
{
"epoch": 1.02,
"grad_norm": 2.148123264312744,
"learning_rate": 0.004491666666666666,
"loss": 0.592,
"step": 550
},
{
"epoch": 1.04,
"grad_norm": 3.1778488159179688,
"learning_rate": 0.004482407407407407,
"loss": 0.93,
"step": 560
},
{
"epoch": 1.05,
"grad_norm": 1.9821240901947021,
"learning_rate": 0.004473148148148149,
"loss": 0.8283,
"step": 570
},
{
"epoch": 1.07,
"grad_norm": 1.9005764722824097,
"learning_rate": 0.004463888888888889,
"loss": 0.7101,
"step": 580
},
{
"epoch": 1.09,
"grad_norm": 2.126354217529297,
"learning_rate": 0.0044546296296296296,
"loss": 0.6671,
"step": 590
},
{
"epoch": 1.11,
"grad_norm": 2.0805599689483643,
"learning_rate": 0.00444537037037037,
"loss": 0.6665,
"step": 600
},
{
"epoch": 1.13,
"grad_norm": 1.5277152061462402,
"learning_rate": 0.004436111111111111,
"loss": 0.7092,
"step": 610
},
{
"epoch": 1.15,
"grad_norm": 4.088949680328369,
"learning_rate": 0.004426851851851852,
"loss": 0.8679,
"step": 620
},
{
"epoch": 1.17,
"grad_norm": 4.5228376388549805,
"learning_rate": 0.004417592592592593,
"loss": 0.7648,
"step": 630
},
{
"epoch": 1.18,
"grad_norm": 4.287468910217285,
"learning_rate": 0.0044083333333333335,
"loss": 0.725,
"step": 640
},
{
"epoch": 1.2,
"grad_norm": 3.0292155742645264,
"learning_rate": 0.004399074074074074,
"loss": 0.7416,
"step": 650
},
{
"epoch": 1.22,
"grad_norm": 2.364569664001465,
"learning_rate": 0.004389814814814815,
"loss": 0.7576,
"step": 660
},
{
"epoch": 1.24,
"grad_norm": 1.8046070337295532,
"learning_rate": 0.004380555555555555,
"loss": 0.6917,
"step": 670
},
{
"epoch": 1.26,
"grad_norm": 3.8957250118255615,
"learning_rate": 0.004371296296296297,
"loss": 0.6671,
"step": 680
},
{
"epoch": 1.28,
"grad_norm": 2.4471426010131836,
"learning_rate": 0.0043620370370370375,
"loss": 0.6749,
"step": 690
},
{
"epoch": 1.3,
"grad_norm": 1.9041227102279663,
"learning_rate": 0.0043527777777777775,
"loss": 0.7423,
"step": 700
},
{
"epoch": 1.31,
"grad_norm": 3.001101016998291,
"learning_rate": 0.004343518518518519,
"loss": 0.7775,
"step": 710
},
{
"epoch": 1.33,
"grad_norm": 2.1323304176330566,
"learning_rate": 0.004334259259259259,
"loss": 0.6961,
"step": 720
},
{
"epoch": 1.35,
"grad_norm": 3.0839684009552,
"learning_rate": 0.004325,
"loss": 0.6791,
"step": 730
},
{
"epoch": 1.37,
"grad_norm": 1.924004316329956,
"learning_rate": 0.004315740740740741,
"loss": 0.663,
"step": 740
},
{
"epoch": 1.39,
"grad_norm": 1.854284644126892,
"learning_rate": 0.0043064814814814814,
"loss": 0.706,
"step": 750
},
{
"epoch": 1.41,
"grad_norm": 2.4816770553588867,
"learning_rate": 0.004297222222222222,
"loss": 0.6623,
"step": 760
},
{
"epoch": 1.42,
"grad_norm": 1.8469676971435547,
"learning_rate": 0.004287962962962963,
"loss": 0.6025,
"step": 770
},
{
"epoch": 1.44,
"grad_norm": 1.6703932285308838,
"learning_rate": 0.004278703703703704,
"loss": 0.7209,
"step": 780
},
{
"epoch": 1.46,
"grad_norm": 1.3802216053009033,
"learning_rate": 0.004269444444444445,
"loss": 0.6823,
"step": 790
},
{
"epoch": 1.48,
"grad_norm": 2.1210741996765137,
"learning_rate": 0.004260185185185185,
"loss": 0.6577,
"step": 800
},
{
"epoch": 1.5,
"grad_norm": 1.841590404510498,
"learning_rate": 0.004250925925925926,
"loss": 0.6049,
"step": 810
},
{
"epoch": 1.52,
"grad_norm": 2.4515421390533447,
"learning_rate": 0.004241666666666667,
"loss": 0.6555,
"step": 820
},
{
"epoch": 1.54,
"grad_norm": 2.086291551589966,
"learning_rate": 0.004232407407407408,
"loss": 0.7836,
"step": 830
},
{
"epoch": 1.55,
"grad_norm": 1.714977741241455,
"learning_rate": 0.004223148148148149,
"loss": 0.6868,
"step": 840
},
{
"epoch": 1.57,
"grad_norm": 1.9800652265548706,
"learning_rate": 0.0042138888888888885,
"loss": 0.7313,
"step": 850
},
{
"epoch": 1.59,
"grad_norm": 2.726518154144287,
"learning_rate": 0.00420462962962963,
"loss": 0.6946,
"step": 860
},
{
"epoch": 1.61,
"grad_norm": 2.3700504302978516,
"learning_rate": 0.00419537037037037,
"loss": 0.7765,
"step": 870
},
{
"epoch": 1.63,
"grad_norm": 3.294623851776123,
"learning_rate": 0.004186111111111111,
"loss": 0.7001,
"step": 880
},
{
"epoch": 1.65,
"grad_norm": 1.3391900062561035,
"learning_rate": 0.004176851851851852,
"loss": 0.7045,
"step": 890
},
{
"epoch": 1.67,
"grad_norm": 3.3717267513275146,
"learning_rate": 0.0041675925925925925,
"loss": 0.7396,
"step": 900
},
{
"epoch": 1.68,
"grad_norm": 2.7791171073913574,
"learning_rate": 0.004158333333333333,
"loss": 0.6946,
"step": 910
},
{
"epoch": 1.7,
"grad_norm": 2.9420201778411865,
"learning_rate": 0.004149074074074074,
"loss": 0.7215,
"step": 920
},
{
"epoch": 1.72,
"grad_norm": 2.337606191635132,
"learning_rate": 0.004139814814814815,
"loss": 0.6967,
"step": 930
},
{
"epoch": 1.74,
"grad_norm": 2.43172025680542,
"learning_rate": 0.004130555555555556,
"loss": 0.7989,
"step": 940
},
{
"epoch": 1.76,
"grad_norm": 1.683398723602295,
"learning_rate": 0.0041212962962962965,
"loss": 0.6185,
"step": 950
},
{
"epoch": 1.78,
"grad_norm": 2.1114211082458496,
"learning_rate": 0.004112037037037037,
"loss": 0.7013,
"step": 960
},
{
"epoch": 1.79,
"grad_norm": 2.566915512084961,
"learning_rate": 0.004102777777777778,
"loss": 0.6091,
"step": 970
},
{
"epoch": 1.81,
"grad_norm": 4.286575794219971,
"learning_rate": 0.004093518518518519,
"loss": 0.7474,
"step": 980
},
{
"epoch": 1.83,
"grad_norm": 1.9286409616470337,
"learning_rate": 0.004084259259259259,
"loss": 0.7089,
"step": 990
},
{
"epoch": 1.85,
"grad_norm": 2.5214929580688477,
"learning_rate": 0.004075,
"loss": 0.7412,
"step": 1000
},
{
"epoch": 1.87,
"grad_norm": 2.4169070720672607,
"learning_rate": 0.004065740740740741,
"loss": 0.826,
"step": 1010
},
{
"epoch": 1.89,
"grad_norm": 2.7352206707000732,
"learning_rate": 0.004056481481481481,
"loss": 0.7868,
"step": 1020
},
{
"epoch": 1.91,
"grad_norm": 1.969298243522644,
"learning_rate": 0.004047222222222222,
"loss": 0.6795,
"step": 1030
},
{
"epoch": 1.92,
"grad_norm": 3.015155792236328,
"learning_rate": 0.004037962962962964,
"loss": 0.7852,
"step": 1040
},
{
"epoch": 1.94,
"grad_norm": 2.2428641319274902,
"learning_rate": 0.004028703703703704,
"loss": 0.7271,
"step": 1050
},
{
"epoch": 1.96,
"grad_norm": 2.1830852031707764,
"learning_rate": 0.004019444444444444,
"loss": 0.6171,
"step": 1060
},
{
"epoch": 1.98,
"grad_norm": 4.2572455406188965,
"learning_rate": 0.004010185185185185,
"loss": 0.715,
"step": 1070
},
{
"epoch": 2.0,
"grad_norm": 1.7257713079452515,
"learning_rate": 0.004000925925925926,
"loss": 0.7397,
"step": 1080
},
{
"epoch": 2.0,
"eval_accuracy": 0.9406871052226159,
"eval_f1": 0.9414537311136683,
"eval_loss": 0.17834338545799255,
"eval_precision": 0.9539065276276052,
"eval_recall": 0.9346301030411137,
"eval_runtime": 55.3829,
"eval_samples_per_second": 117.202,
"eval_steps_per_second": 7.331,
"step": 1081
},
{
"epoch": 2.02,
"grad_norm": 1.4031952619552612,
"learning_rate": 0.003991666666666667,
"loss": 0.652,
"step": 1090
},
{
"epoch": 2.04,
"grad_norm": 2.210618257522583,
"learning_rate": 0.003982407407407408,
"loss": 0.6434,
"step": 1100
},
{
"epoch": 2.05,
"grad_norm": 2.780142068862915,
"learning_rate": 0.0039731481481481475,
"loss": 0.6638,
"step": 1110
},
{
"epoch": 2.07,
"grad_norm": 2.0362918376922607,
"learning_rate": 0.003963888888888889,
"loss": 0.6194,
"step": 1120
},
{
"epoch": 2.09,
"grad_norm": 2.7346765995025635,
"learning_rate": 0.00395462962962963,
"loss": 0.6721,
"step": 1130
},
{
"epoch": 2.11,
"grad_norm": 1.8441007137298584,
"learning_rate": 0.00394537037037037,
"loss": 0.7129,
"step": 1140
},
{
"epoch": 2.13,
"grad_norm": 2.510021448135376,
"learning_rate": 0.003936111111111112,
"loss": 0.669,
"step": 1150
},
{
"epoch": 2.15,
"grad_norm": 2.9301140308380127,
"learning_rate": 0.003926851851851852,
"loss": 0.6342,
"step": 1160
},
{
"epoch": 2.16,
"grad_norm": 1.851173996925354,
"learning_rate": 0.003917592592592592,
"loss": 0.6602,
"step": 1170
},
{
"epoch": 2.18,
"grad_norm": 2.490886926651001,
"learning_rate": 0.003908333333333333,
"loss": 0.6195,
"step": 1180
},
{
"epoch": 2.2,
"grad_norm": 1.5079678297042847,
"learning_rate": 0.0038990740740740743,
"loss": 0.8126,
"step": 1190
},
{
"epoch": 2.22,
"grad_norm": 1.9084036350250244,
"learning_rate": 0.003889814814814815,
"loss": 0.6769,
"step": 1200
},
{
"epoch": 2.24,
"grad_norm": 1.8284987211227417,
"learning_rate": 0.0038814814814814814,
"loss": 0.7317,
"step": 1210
},
{
"epoch": 2.26,
"grad_norm": 3.351156234741211,
"learning_rate": 0.0038722222222222226,
"loss": 0.7049,
"step": 1220
},
{
"epoch": 2.28,
"grad_norm": 2.314993143081665,
"learning_rate": 0.003862962962962963,
"loss": 0.6291,
"step": 1230
},
{
"epoch": 2.29,
"grad_norm": 2.27850341796875,
"learning_rate": 0.003853703703703704,
"loss": 0.6749,
"step": 1240
},
{
"epoch": 2.31,
"grad_norm": 2.4541099071502686,
"learning_rate": 0.003844444444444444,
"loss": 0.7365,
"step": 1250
},
{
"epoch": 2.33,
"grad_norm": 1.6301252841949463,
"learning_rate": 0.0038351851851851854,
"loss": 0.6236,
"step": 1260
},
{
"epoch": 2.35,
"grad_norm": 2.035531520843506,
"learning_rate": 0.003825925925925926,
"loss": 0.7176,
"step": 1270
},
{
"epoch": 2.37,
"grad_norm": 2.2524566650390625,
"learning_rate": 0.0038166666666666666,
"loss": 0.7119,
"step": 1280
},
{
"epoch": 2.39,
"grad_norm": 1.778144121170044,
"learning_rate": 0.003807407407407408,
"loss": 0.7021,
"step": 1290
},
{
"epoch": 2.41,
"grad_norm": 1.8808122873306274,
"learning_rate": 0.003798148148148148,
"loss": 0.6876,
"step": 1300
},
{
"epoch": 2.42,
"grad_norm": 2.720100164413452,
"learning_rate": 0.003788888888888889,
"loss": 0.7559,
"step": 1310
},
{
"epoch": 2.44,
"grad_norm": 1.9608194828033447,
"learning_rate": 0.0037796296296296297,
"loss": 0.6799,
"step": 1320
},
{
"epoch": 2.46,
"grad_norm": 1.8149750232696533,
"learning_rate": 0.0037703703703703705,
"loss": 0.6701,
"step": 1330
},
{
"epoch": 2.48,
"grad_norm": 1.8840572834014893,
"learning_rate": 0.0037611111111111113,
"loss": 0.7146,
"step": 1340
},
{
"epoch": 2.5,
"grad_norm": 1.7513312101364136,
"learning_rate": 0.0037518518518518517,
"loss": 0.6712,
"step": 1350
},
{
"epoch": 2.52,
"grad_norm": 1.748304009437561,
"learning_rate": 0.0037425925925925925,
"loss": 0.7049,
"step": 1360
},
{
"epoch": 2.53,
"grad_norm": 1.2645372152328491,
"learning_rate": 0.0037333333333333337,
"loss": 0.6088,
"step": 1370
},
{
"epoch": 2.55,
"grad_norm": 3.14270281791687,
"learning_rate": 0.003724074074074074,
"loss": 0.7163,
"step": 1380
},
{
"epoch": 2.57,
"grad_norm": 2.0332484245300293,
"learning_rate": 0.003714814814814815,
"loss": 0.7047,
"step": 1390
},
{
"epoch": 2.59,
"grad_norm": 1.6158325672149658,
"learning_rate": 0.0037055555555555557,
"loss": 0.6448,
"step": 1400
},
{
"epoch": 2.61,
"grad_norm": 1.9182847738265991,
"learning_rate": 0.0036962962962962965,
"loss": 0.711,
"step": 1410
},
{
"epoch": 2.63,
"grad_norm": 2.140184164047241,
"learning_rate": 0.0036870370370370373,
"loss": 0.6617,
"step": 1420
},
{
"epoch": 2.65,
"grad_norm": 1.746243953704834,
"learning_rate": 0.0036777777777777776,
"loss": 0.7103,
"step": 1430
},
{
"epoch": 2.66,
"grad_norm": 2.6956822872161865,
"learning_rate": 0.003668518518518519,
"loss": 0.7163,
"step": 1440
},
{
"epoch": 2.68,
"grad_norm": 2.7663381099700928,
"learning_rate": 0.0036592592592592592,
"loss": 0.6509,
"step": 1450
},
{
"epoch": 2.7,
"grad_norm": 2.410922050476074,
"learning_rate": 0.00365,
"loss": 0.6046,
"step": 1460
},
{
"epoch": 2.72,
"grad_norm": 1.8373881578445435,
"learning_rate": 0.0036407407407407404,
"loss": 0.701,
"step": 1470
},
{
"epoch": 2.74,
"grad_norm": 1.8466787338256836,
"learning_rate": 0.0036314814814814816,
"loss": 0.6391,
"step": 1480
},
{
"epoch": 2.76,
"grad_norm": 2.0422921180725098,
"learning_rate": 0.0036222222222222224,
"loss": 0.7001,
"step": 1490
},
{
"epoch": 2.78,
"grad_norm": 1.739327311515808,
"learning_rate": 0.003612962962962963,
"loss": 0.6194,
"step": 1500
},
{
"epoch": 2.79,
"grad_norm": 2.094775915145874,
"learning_rate": 0.003603703703703704,
"loss": 0.5824,
"step": 1510
},
{
"epoch": 2.81,
"grad_norm": 2.5036680698394775,
"learning_rate": 0.003594444444444445,
"loss": 0.6477,
"step": 1520
},
{
"epoch": 2.83,
"grad_norm": 1.7037166357040405,
"learning_rate": 0.003585185185185185,
"loss": 0.6822,
"step": 1530
},
{
"epoch": 2.85,
"grad_norm": 2.193714141845703,
"learning_rate": 0.003575925925925926,
"loss": 0.7301,
"step": 1540
},
{
"epoch": 2.87,
"grad_norm": 4.449968338012695,
"learning_rate": 0.0035666666666666668,
"loss": 0.7114,
"step": 1550
},
{
"epoch": 2.89,
"grad_norm": 1.640833854675293,
"learning_rate": 0.0035574074074074076,
"loss": 0.5631,
"step": 1560
},
{
"epoch": 2.9,
"grad_norm": 2.4068825244903564,
"learning_rate": 0.003548148148148148,
"loss": 0.5896,
"step": 1570
},
{
"epoch": 2.92,
"grad_norm": 2.7234010696411133,
"learning_rate": 0.0035388888888888887,
"loss": 0.7563,
"step": 1580
},
{
"epoch": 2.94,
"grad_norm": 1.8094590902328491,
"learning_rate": 0.00352962962962963,
"loss": 0.7116,
"step": 1590
},
{
"epoch": 2.96,
"grad_norm": 2.4650044441223145,
"learning_rate": 0.0035203703703703703,
"loss": 0.6583,
"step": 1600
},
{
"epoch": 2.98,
"grad_norm": 1.8076032400131226,
"learning_rate": 0.003511111111111111,
"loss": 0.5859,
"step": 1610
},
{
"epoch": 3.0,
"grad_norm": 2.924779176712036,
"learning_rate": 0.0035018518518518523,
"loss": 0.7151,
"step": 1620
},
{
"epoch": 3.0,
"eval_accuracy": 0.9551686951163149,
"eval_f1": 0.957151302239187,
"eval_loss": 0.12974682450294495,
"eval_precision": 0.961140704317014,
"eval_recall": 0.9555174635857528,
"eval_runtime": 55.4288,
"eval_samples_per_second": 117.105,
"eval_steps_per_second": 7.325,
"step": 1621
},
{
"epoch": 3.02,
"grad_norm": 2.4308605194091797,
"learning_rate": 0.0034925925925925927,
"loss": 0.6032,
"step": 1630
},
{
"epoch": 3.03,
"grad_norm": 1.6040972471237183,
"learning_rate": 0.0034833333333333335,
"loss": 0.6163,
"step": 1640
},
{
"epoch": 3.05,
"grad_norm": 1.7093302011489868,
"learning_rate": 0.003474074074074074,
"loss": 0.6484,
"step": 1650
},
{
"epoch": 3.07,
"grad_norm": 2.0212676525115967,
"learning_rate": 0.003464814814814815,
"loss": 0.5747,
"step": 1660
},
{
"epoch": 3.09,
"grad_norm": 3.725684642791748,
"learning_rate": 0.0034555555555555555,
"loss": 0.6488,
"step": 1670
},
{
"epoch": 3.11,
"grad_norm": 2.5696334838867188,
"learning_rate": 0.0034462962962962963,
"loss": 0.7223,
"step": 1680
},
{
"epoch": 3.13,
"grad_norm": 1.7049779891967773,
"learning_rate": 0.003437037037037037,
"loss": 0.641,
"step": 1690
},
{
"epoch": 3.15,
"grad_norm": 2.035477638244629,
"learning_rate": 0.003427777777777778,
"loss": 0.6237,
"step": 1700
},
{
"epoch": 3.16,
"grad_norm": 1.8789253234863281,
"learning_rate": 0.0034185185185185187,
"loss": 0.6312,
"step": 1710
},
{
"epoch": 3.18,
"grad_norm": 2.2245302200317383,
"learning_rate": 0.003409259259259259,
"loss": 0.5985,
"step": 1720
},
{
"epoch": 3.2,
"grad_norm": 2.150331735610962,
"learning_rate": 0.0034000000000000002,
"loss": 0.612,
"step": 1730
},
{
"epoch": 3.22,
"grad_norm": 1.986059546470642,
"learning_rate": 0.003390740740740741,
"loss": 0.6001,
"step": 1740
},
{
"epoch": 3.24,
"grad_norm": 1.4887031316757202,
"learning_rate": 0.0033814814814814814,
"loss": 0.6574,
"step": 1750
},
{
"epoch": 3.26,
"grad_norm": 1.1965304613113403,
"learning_rate": 0.003372222222222222,
"loss": 0.6766,
"step": 1760
},
{
"epoch": 3.27,
"grad_norm": 1.564257025718689,
"learning_rate": 0.0033629629629629634,
"loss": 0.6955,
"step": 1770
},
{
"epoch": 3.29,
"grad_norm": 5.508277416229248,
"learning_rate": 0.003353703703703704,
"loss": 0.6097,
"step": 1780
},
{
"epoch": 3.31,
"grad_norm": 1.9245549440383911,
"learning_rate": 0.0033444444444444446,
"loss": 0.6515,
"step": 1790
},
{
"epoch": 3.33,
"grad_norm": 2.3552732467651367,
"learning_rate": 0.003335185185185185,
"loss": 0.6401,
"step": 1800
},
{
"epoch": 3.35,
"grad_norm": 1.622826099395752,
"learning_rate": 0.003325925925925926,
"loss": 0.7,
"step": 1810
},
{
"epoch": 3.37,
"grad_norm": 1.6205705404281616,
"learning_rate": 0.0033166666666666665,
"loss": 0.5968,
"step": 1820
},
{
"epoch": 3.39,
"grad_norm": 1.9269992113113403,
"learning_rate": 0.0033074074074074073,
"loss": 0.6,
"step": 1830
},
{
"epoch": 3.4,
"grad_norm": 2.3607101440429688,
"learning_rate": 0.0032981481481481486,
"loss": 0.5655,
"step": 1840
},
{
"epoch": 3.42,
"grad_norm": 1.4156101942062378,
"learning_rate": 0.003288888888888889,
"loss": 0.6071,
"step": 1850
},
{
"epoch": 3.44,
"grad_norm": 2.652838706970215,
"learning_rate": 0.0032796296296296297,
"loss": 0.5959,
"step": 1860
},
{
"epoch": 3.46,
"grad_norm": 1.7326959371566772,
"learning_rate": 0.00327037037037037,
"loss": 0.7051,
"step": 1870
},
{
"epoch": 3.48,
"grad_norm": 1.4709230661392212,
"learning_rate": 0.0032611111111111113,
"loss": 0.68,
"step": 1880
},
{
"epoch": 3.5,
"grad_norm": 2.1479110717773438,
"learning_rate": 0.003251851851851852,
"loss": 0.6491,
"step": 1890
},
{
"epoch": 3.52,
"grad_norm": 2.462580442428589,
"learning_rate": 0.0032425925925925925,
"loss": 0.6129,
"step": 1900
},
{
"epoch": 3.53,
"grad_norm": 2.8849940299987793,
"learning_rate": 0.0032333333333333333,
"loss": 0.6553,
"step": 1910
},
{
"epoch": 3.55,
"grad_norm": 1.6244118213653564,
"learning_rate": 0.003224074074074074,
"loss": 0.5845,
"step": 1920
},
{
"epoch": 3.57,
"grad_norm": 2.803816795349121,
"learning_rate": 0.003214814814814815,
"loss": 0.6695,
"step": 1930
},
{
"epoch": 3.59,
"grad_norm": 2.4138996601104736,
"learning_rate": 0.0032055555555555552,
"loss": 0.707,
"step": 1940
},
{
"epoch": 3.61,
"grad_norm": 1.1193430423736572,
"learning_rate": 0.0031962962962962965,
"loss": 0.6003,
"step": 1950
},
{
"epoch": 3.63,
"grad_norm": 1.8593242168426514,
"learning_rate": 0.0031870370370370373,
"loss": 0.5724,
"step": 1960
},
{
"epoch": 3.64,
"grad_norm": 2.084005117416382,
"learning_rate": 0.0031777777777777776,
"loss": 0.6246,
"step": 1970
},
{
"epoch": 3.66,
"grad_norm": 1.32781982421875,
"learning_rate": 0.0031685185185185184,
"loss": 0.6075,
"step": 1980
},
{
"epoch": 3.68,
"grad_norm": 2.452892303466797,
"learning_rate": 0.0031592592592592597,
"loss": 0.6104,
"step": 1990
},
{
"epoch": 3.7,
"grad_norm": 1.6961791515350342,
"learning_rate": 0.00315,
"loss": 0.5483,
"step": 2000
},
{
"epoch": 3.72,
"grad_norm": 2.94992733001709,
"learning_rate": 0.003140740740740741,
"loss": 0.5403,
"step": 2010
},
{
"epoch": 3.74,
"grad_norm": 2.3624625205993652,
"learning_rate": 0.0031314814814814816,
"loss": 0.6748,
"step": 2020
},
{
"epoch": 3.76,
"grad_norm": 2.0412802696228027,
"learning_rate": 0.0031222222222222224,
"loss": 0.4809,
"step": 2030
},
{
"epoch": 3.77,
"grad_norm": 1.5728198289871216,
"learning_rate": 0.003112962962962963,
"loss": 0.5755,
"step": 2040
},
{
"epoch": 3.79,
"grad_norm": 1.3974460363388062,
"learning_rate": 0.0031037037037037036,
"loss": 0.5436,
"step": 2050
},
{
"epoch": 3.81,
"grad_norm": 1.5604791641235352,
"learning_rate": 0.003094444444444445,
"loss": 0.6197,
"step": 2060
},
{
"epoch": 3.83,
"grad_norm": 2.040276288986206,
"learning_rate": 0.003085185185185185,
"loss": 0.6062,
"step": 2070
},
{
"epoch": 3.85,
"grad_norm": 3.173293113708496,
"learning_rate": 0.003075925925925926,
"loss": 0.717,
"step": 2080
},
{
"epoch": 3.87,
"grad_norm": 2.018458843231201,
"learning_rate": 0.0030666666666666663,
"loss": 0.6418,
"step": 2090
},
{
"epoch": 3.89,
"grad_norm": 1.7057709693908691,
"learning_rate": 0.0030574074074074076,
"loss": 0.5948,
"step": 2100
},
{
"epoch": 3.9,
"grad_norm": 2.133119821548462,
"learning_rate": 0.0030481481481481484,
"loss": 0.6099,
"step": 2110
},
{
"epoch": 3.92,
"grad_norm": 2.217524290084839,
"learning_rate": 0.0030388888888888887,
"loss": 0.6263,
"step": 2120
},
{
"epoch": 3.94,
"grad_norm": 2.3708674907684326,
"learning_rate": 0.00302962962962963,
"loss": 0.5851,
"step": 2130
},
{
"epoch": 3.96,
"grad_norm": 1.4289004802703857,
"learning_rate": 0.0030203703703703707,
"loss": 0.5971,
"step": 2140
},
{
"epoch": 3.98,
"grad_norm": 2.463360071182251,
"learning_rate": 0.003011111111111111,
"loss": 0.5889,
"step": 2150
},
{
"epoch": 4.0,
"grad_norm": 1.5189619064331055,
"learning_rate": 0.003001851851851852,
"loss": 0.4964,
"step": 2160
},
{
"epoch": 4.0,
"eval_accuracy": 0.97350177168387,
"eval_f1": 0.9729772233764705,
"eval_loss": 0.07412911206483841,
"eval_precision": 0.9765336155795084,
"eval_recall": 0.9701628526053132,
"eval_runtime": 55.303,
"eval_samples_per_second": 117.372,
"eval_steps_per_second": 7.341,
"step": 2162
},
{
"epoch": 4.01,
"grad_norm": 2.313246965408325,
"learning_rate": 0.0029925925925925927,
"loss": 0.5979,
"step": 2170
},
{
"epoch": 4.03,
"grad_norm": 1.554916501045227,
"learning_rate": 0.0029833333333333335,
"loss": 0.621,
"step": 2180
},
{
"epoch": 4.05,
"grad_norm": 2.7611353397369385,
"learning_rate": 0.002974074074074074,
"loss": 0.6662,
"step": 2190
},
{
"epoch": 4.07,
"grad_norm": 1.206809639930725,
"learning_rate": 0.0029648148148148147,
"loss": 0.6061,
"step": 2200
},
{
"epoch": 4.09,
"grad_norm": 1.876120686531067,
"learning_rate": 0.002955555555555556,
"loss": 0.5656,
"step": 2210
},
{
"epoch": 4.11,
"grad_norm": 2.117579936981201,
"learning_rate": 0.0029462962962962963,
"loss": 0.611,
"step": 2220
},
{
"epoch": 4.13,
"grad_norm": 2.014042377471924,
"learning_rate": 0.002937037037037037,
"loss": 0.5893,
"step": 2230
},
{
"epoch": 4.14,
"grad_norm": 1.6792641878128052,
"learning_rate": 0.0029277777777777783,
"loss": 0.5691,
"step": 2240
},
{
"epoch": 4.16,
"grad_norm": 1.2664446830749512,
"learning_rate": 0.0029185185185185186,
"loss": 0.5729,
"step": 2250
},
{
"epoch": 4.18,
"grad_norm": 3.1098408699035645,
"learning_rate": 0.0029092592592592594,
"loss": 0.6115,
"step": 2260
},
{
"epoch": 4.2,
"grad_norm": 2.1186132431030273,
"learning_rate": 0.0029,
"loss": 0.596,
"step": 2270
},
{
"epoch": 4.22,
"grad_norm": 1.8666993379592896,
"learning_rate": 0.002890740740740741,
"loss": 0.6021,
"step": 2280
},
{
"epoch": 4.24,
"grad_norm": 1.655548334121704,
"learning_rate": 0.0028814814814814814,
"loss": 0.5928,
"step": 2290
},
{
"epoch": 4.26,
"grad_norm": 2.1081762313842773,
"learning_rate": 0.002872222222222222,
"loss": 0.5683,
"step": 2300
},
{
"epoch": 4.27,
"grad_norm": 1.4986019134521484,
"learning_rate": 0.0028629629629629626,
"loss": 0.562,
"step": 2310
},
{
"epoch": 4.29,
"grad_norm": 2.6111056804656982,
"learning_rate": 0.002853703703703704,
"loss": 0.5804,
"step": 2320
},
{
"epoch": 4.31,
"grad_norm": 1.8590542078018188,
"learning_rate": 0.0028444444444444446,
"loss": 0.5409,
"step": 2330
},
{
"epoch": 4.33,
"grad_norm": 1.897940754890442,
"learning_rate": 0.002835185185185185,
"loss": 0.6299,
"step": 2340
},
{
"epoch": 4.35,
"grad_norm": 1.9114952087402344,
"learning_rate": 0.002825925925925926,
"loss": 0.5333,
"step": 2350
},
{
"epoch": 4.37,
"grad_norm": 2.015868663787842,
"learning_rate": 0.002816666666666667,
"loss": 0.5611,
"step": 2360
},
{
"epoch": 4.38,
"grad_norm": 1.5808693170547485,
"learning_rate": 0.0028074074074074073,
"loss": 0.6212,
"step": 2370
},
{
"epoch": 4.4,
"grad_norm": 1.8288767337799072,
"learning_rate": 0.002798148148148148,
"loss": 0.604,
"step": 2380
},
{
"epoch": 4.42,
"grad_norm": 1.389092206954956,
"learning_rate": 0.002788888888888889,
"loss": 0.4657,
"step": 2390
},
{
"epoch": 4.44,
"grad_norm": 2.551008701324463,
"learning_rate": 0.0027796296296296297,
"loss": 0.6214,
"step": 2400
},
{
"epoch": 4.46,
"grad_norm": 2.26882266998291,
"learning_rate": 0.0027703703703703705,
"loss": 0.5511,
"step": 2410
},
{
"epoch": 4.48,
"grad_norm": 2.3850460052490234,
"learning_rate": 0.002761111111111111,
"loss": 0.5768,
"step": 2420
},
{
"epoch": 4.5,
"grad_norm": 1.9451379776000977,
"learning_rate": 0.002751851851851852,
"loss": 0.5732,
"step": 2430
},
{
"epoch": 4.51,
"grad_norm": 1.6541566848754883,
"learning_rate": 0.0027425925925925925,
"loss": 0.549,
"step": 2440
},
{
"epoch": 4.53,
"grad_norm": 1.3802433013916016,
"learning_rate": 0.0027333333333333333,
"loss": 0.5486,
"step": 2450
},
{
"epoch": 4.55,
"grad_norm": 1.941901683807373,
"learning_rate": 0.0027240740740740745,
"loss": 0.6119,
"step": 2460
},
{
"epoch": 4.57,
"grad_norm": 1.9203604459762573,
"learning_rate": 0.002714814814814815,
"loss": 0.632,
"step": 2470
},
{
"epoch": 4.59,
"grad_norm": 2.170319080352783,
"learning_rate": 0.0027055555555555557,
"loss": 0.5819,
"step": 2480
},
{
"epoch": 4.61,
"grad_norm": 1.8074395656585693,
"learning_rate": 0.002696296296296296,
"loss": 0.6101,
"step": 2490
},
{
"epoch": 4.63,
"grad_norm": 2.010608196258545,
"learning_rate": 0.0026870370370370373,
"loss": 0.492,
"step": 2500
},
{
"epoch": 4.64,
"grad_norm": 1.3907346725463867,
"learning_rate": 0.002677777777777778,
"loss": 0.5658,
"step": 2510
},
{
"epoch": 4.66,
"grad_norm": 1.8057767152786255,
"learning_rate": 0.0026685185185185184,
"loss": 0.509,
"step": 2520
},
{
"epoch": 4.68,
"grad_norm": 2.036869764328003,
"learning_rate": 0.002659259259259259,
"loss": 0.5711,
"step": 2530
},
{
"epoch": 4.7,
"grad_norm": 1.673647165298462,
"learning_rate": 0.00265,
"loss": 0.5692,
"step": 2540
},
{
"epoch": 4.72,
"grad_norm": 2.044961929321289,
"learning_rate": 0.002640740740740741,
"loss": 0.6153,
"step": 2550
},
{
"epoch": 4.74,
"grad_norm": 1.3360215425491333,
"learning_rate": 0.002631481481481481,
"loss": 0.5785,
"step": 2560
},
{
"epoch": 4.75,
"grad_norm": 2.4942688941955566,
"learning_rate": 0.0026222222222222224,
"loss": 0.5423,
"step": 2570
},
{
"epoch": 4.77,
"grad_norm": 1.556859016418457,
"learning_rate": 0.002612962962962963,
"loss": 0.5783,
"step": 2580
},
{
"epoch": 4.79,
"grad_norm": 3.010946750640869,
"learning_rate": 0.0026037037037037036,
"loss": 0.5786,
"step": 2590
},
{
"epoch": 4.81,
"grad_norm": 1.5896031856536865,
"learning_rate": 0.0025944444444444444,
"loss": 0.6075,
"step": 2600
},
{
"epoch": 4.83,
"grad_norm": 1.9418363571166992,
"learning_rate": 0.0025851851851851856,
"loss": 0.5482,
"step": 2610
},
{
"epoch": 4.85,
"grad_norm": 1.9787338972091675,
"learning_rate": 0.002576851851851852,
"loss": 0.6026,
"step": 2620
},
{
"epoch": 4.87,
"grad_norm": 2.7021665573120117,
"learning_rate": 0.0025675925925925927,
"loss": 0.4977,
"step": 2630
},
{
"epoch": 4.88,
"grad_norm": 2.2835893630981445,
"learning_rate": 0.0025583333333333335,
"loss": 0.5375,
"step": 2640
},
{
"epoch": 4.9,
"grad_norm": 1.9748303890228271,
"learning_rate": 0.0025490740740740743,
"loss": 0.5263,
"step": 2650
},
{
"epoch": 4.92,
"grad_norm": 1.6271171569824219,
"learning_rate": 0.0025398148148148146,
"loss": 0.5275,
"step": 2660
},
{
"epoch": 4.94,
"grad_norm": 2.3140103816986084,
"learning_rate": 0.0025305555555555554,
"loss": 0.5493,
"step": 2670
},
{
"epoch": 4.96,
"grad_norm": 2.2506508827209473,
"learning_rate": 0.0025212962962962967,
"loss": 0.5756,
"step": 2680
},
{
"epoch": 4.98,
"grad_norm": 2.5812437534332275,
"learning_rate": 0.002512037037037037,
"loss": 0.511,
"step": 2690
},
{
"epoch": 5.0,
"grad_norm": 1.886813998222351,
"learning_rate": 0.002502777777777778,
"loss": 0.5509,
"step": 2700
},
{
"epoch": 5.0,
"eval_accuracy": 0.9770451394238175,
"eval_f1": 0.9783131597050811,
"eval_loss": 0.06713523715734482,
"eval_precision": 0.9775503610869588,
"eval_recall": 0.9795962655538212,
"eval_runtime": 55.3992,
"eval_samples_per_second": 117.168,
"eval_steps_per_second": 7.329,
"step": 2702
},
{
"epoch": 5.01,
"grad_norm": 2.990206241607666,
"learning_rate": 0.0024935185185185186,
"loss": 0.5212,
"step": 2710
},
{
"epoch": 5.03,
"grad_norm": 1.2570441961288452,
"learning_rate": 0.0024842592592592594,
"loss": 0.4703,
"step": 2720
},
{
"epoch": 5.05,
"grad_norm": 1.6648197174072266,
"learning_rate": 0.002475,
"loss": 0.6173,
"step": 2730
},
{
"epoch": 5.07,
"grad_norm": 1.9291369915008545,
"learning_rate": 0.002465740740740741,
"loss": 0.5929,
"step": 2740
},
{
"epoch": 5.09,
"grad_norm": 2.2894320487976074,
"learning_rate": 0.0024564814814814814,
"loss": 0.5282,
"step": 2750
},
{
"epoch": 5.11,
"grad_norm": NaN,
"learning_rate": 0.002448148148148148,
"loss": 0.5085,
"step": 2760
},
{
"epoch": 5.12,
"grad_norm": 2.4065709114074707,
"learning_rate": 0.002438888888888889,
"loss": 0.5983,
"step": 2770
},
{
"epoch": 5.14,
"grad_norm": 1.4784891605377197,
"learning_rate": 0.0024296296296296297,
"loss": 0.5586,
"step": 2780
},
{
"epoch": 5.16,
"grad_norm": 3.6442089080810547,
"learning_rate": 0.0024203703703703705,
"loss": 0.5824,
"step": 2790
},
{
"epoch": 5.18,
"grad_norm": 1.4887707233428955,
"learning_rate": 0.0024111111111111113,
"loss": 0.6023,
"step": 2800
},
{
"epoch": 5.2,
"grad_norm": 1.4839690923690796,
"learning_rate": 0.002401851851851852,
"loss": 0.5486,
"step": 2810
},
{
"epoch": 5.22,
"grad_norm": 1.522079586982727,
"learning_rate": 0.0023925925925925924,
"loss": 0.5803,
"step": 2820
},
{
"epoch": 5.24,
"grad_norm": 1.6155500411987305,
"learning_rate": 0.0023833333333333337,
"loss": 0.5765,
"step": 2830
},
{
"epoch": 5.25,
"grad_norm": 1.4608234167099,
"learning_rate": 0.002374074074074074,
"loss": 0.5163,
"step": 2840
},
{
"epoch": 5.27,
"grad_norm": 1.8083865642547607,
"learning_rate": 0.002364814814814815,
"loss": 0.5533,
"step": 2850
},
{
"epoch": 5.29,
"grad_norm": 1.7674206495285034,
"learning_rate": 0.0023555555555555556,
"loss": 0.5558,
"step": 2860
},
{
"epoch": 5.31,
"grad_norm": 1.7677381038665771,
"learning_rate": 0.0023462962962962964,
"loss": 0.551,
"step": 2870
},
{
"epoch": 5.33,
"grad_norm": 2.006486415863037,
"learning_rate": 0.0023370370370370368,
"loss": 0.5311,
"step": 2880
},
{
"epoch": 5.35,
"grad_norm": 1.3190593719482422,
"learning_rate": 0.002327777777777778,
"loss": 0.5462,
"step": 2890
},
{
"epoch": 5.37,
"grad_norm": 2.1701364517211914,
"learning_rate": 0.002318518518518519,
"loss": 0.5056,
"step": 2900
},
{
"epoch": 5.38,
"grad_norm": 3.071043014526367,
"learning_rate": 0.002309259259259259,
"loss": 0.5611,
"step": 2910
},
{
"epoch": 5.4,
"grad_norm": 1.7673370838165283,
"learning_rate": 0.0023,
"loss": 0.5369,
"step": 2920
},
{
"epoch": 5.42,
"grad_norm": 1.6603857278823853,
"learning_rate": 0.0022907407407407408,
"loss": 0.6108,
"step": 2930
},
{
"epoch": 5.44,
"grad_norm": 3.754216194152832,
"learning_rate": 0.0022814814814814816,
"loss": 0.5355,
"step": 2940
},
{
"epoch": 5.46,
"grad_norm": 2.2037136554718018,
"learning_rate": 0.0022722222222222224,
"loss": 0.4998,
"step": 2950
},
{
"epoch": 5.48,
"grad_norm": 2.2701072692871094,
"learning_rate": 0.002262962962962963,
"loss": 0.5672,
"step": 2960
},
{
"epoch": 5.49,
"grad_norm": 2.246368408203125,
"learning_rate": 0.0022537037037037035,
"loss": 0.5309,
"step": 2970
},
{
"epoch": 5.51,
"grad_norm": 1.4481878280639648,
"learning_rate": 0.0022444444444444443,
"loss": 0.4786,
"step": 2980
},
{
"epoch": 5.53,
"grad_norm": 1.7569130659103394,
"learning_rate": 0.002235185185185185,
"loss": 0.579,
"step": 2990
},
{
"epoch": 5.55,
"grad_norm": 2.219710350036621,
"learning_rate": 0.002225925925925926,
"loss": 0.4768,
"step": 3000
},
{
"epoch": 5.57,
"grad_norm": 1.3611927032470703,
"learning_rate": 0.0022166666666666667,
"loss": 0.4504,
"step": 3010
},
{
"epoch": 5.59,
"grad_norm": 1.528106451034546,
"learning_rate": 0.0022074074074074075,
"loss": 0.491,
"step": 3020
},
{
"epoch": 5.61,
"grad_norm": 1.934448003768921,
"learning_rate": 0.0021981481481481483,
"loss": 0.5553,
"step": 3030
},
{
"epoch": 5.62,
"grad_norm": 1.9539951086044312,
"learning_rate": 0.002188888888888889,
"loss": 0.514,
"step": 3040
},
{
"epoch": 5.64,
"grad_norm": 3.347055196762085,
"learning_rate": 0.00217962962962963,
"loss": 0.565,
"step": 3050
},
{
"epoch": 5.66,
"grad_norm": 2.0984296798706055,
"learning_rate": 0.0021703703703703702,
"loss": 0.4917,
"step": 3060
},
{
"epoch": 5.68,
"grad_norm": 4.030628681182861,
"learning_rate": 0.002161111111111111,
"loss": 0.5485,
"step": 3070
},
{
"epoch": 5.7,
"grad_norm": 1.62152898311615,
"learning_rate": 0.002151851851851852,
"loss": 0.5468,
"step": 3080
},
{
"epoch": 5.72,
"grad_norm": 2.245067596435547,
"learning_rate": 0.0021425925925925926,
"loss": 0.4992,
"step": 3090
},
{
"epoch": 5.74,
"grad_norm": 2.4511606693267822,
"learning_rate": 0.0021333333333333334,
"loss": 0.5478,
"step": 3100
},
{
"epoch": 5.75,
"grad_norm": 3.2524254322052,
"learning_rate": 0.0021240740740740742,
"loss": 0.5006,
"step": 3110
},
{
"epoch": 5.77,
"grad_norm": 1.8704936504364014,
"learning_rate": 0.002114814814814815,
"loss": 0.5177,
"step": 3120
},
{
"epoch": 5.79,
"grad_norm": 1.6047933101654053,
"learning_rate": 0.0021064814814814813,
"loss": 0.5491,
"step": 3130
},
{
"epoch": 5.81,
"grad_norm": 2.038342237472534,
"learning_rate": 0.002097222222222222,
"loss": 0.5096,
"step": 3140
},
{
"epoch": 5.83,
"grad_norm": 1.1834994554519653,
"learning_rate": 0.002087962962962963,
"loss": 0.4611,
"step": 3150
},
{
"epoch": 5.85,
"grad_norm": 3.6137051582336426,
"learning_rate": 0.0020787037037037037,
"loss": 0.5056,
"step": 3160
},
{
"epoch": 5.86,
"grad_norm": 1.7401528358459473,
"learning_rate": 0.0020694444444444445,
"loss": 0.5285,
"step": 3170
},
{
"epoch": 5.88,
"grad_norm": 1.1825543642044067,
"learning_rate": 0.0020601851851851853,
"loss": 0.5166,
"step": 3180
},
{
"epoch": 5.9,
"grad_norm": 1.469771385192871,
"learning_rate": 0.002050925925925926,
"loss": 0.5246,
"step": 3190
},
{
"epoch": 5.92,
"grad_norm": 1.9187328815460205,
"learning_rate": 0.0020416666666666665,
"loss": 0.5687,
"step": 3200
},
{
"epoch": 5.94,
"grad_norm": 1.27570641040802,
"learning_rate": 0.0020324074074074077,
"loss": 0.5367,
"step": 3210
},
{
"epoch": 5.96,
"grad_norm": 2.436992883682251,
"learning_rate": 0.002023148148148148,
"loss": 0.5108,
"step": 3220
},
{
"epoch": 5.98,
"grad_norm": 2.1716506481170654,
"learning_rate": 0.002013888888888889,
"loss": 0.5081,
"step": 3230
},
{
"epoch": 5.99,
"grad_norm": 2.3047661781311035,
"learning_rate": 0.0020046296296296296,
"loss": 0.5746,
"step": 3240
},
{
"epoch": 6.0,
"eval_accuracy": 0.9753504852873209,
"eval_f1": 0.9794812735446677,
"eval_loss": 0.06419458985328674,
"eval_precision": 0.9810236528071776,
"eval_recall": 0.9788061689635984,
"eval_runtime": 55.1158,
"eval_samples_per_second": 117.77,
"eval_steps_per_second": 7.366,
"step": 3243
},
{
"epoch": 6.01,
"grad_norm": 2.193244457244873,
"learning_rate": 0.0019953703703703704,
"loss": 0.4844,
"step": 3250
},
{
"epoch": 6.03,
"grad_norm": 1.7774107456207275,
"learning_rate": 0.0019861111111111112,
"loss": 0.5153,
"step": 3260
},
{
"epoch": 6.05,
"grad_norm": 1.8430052995681763,
"learning_rate": 0.001976851851851852,
"loss": 0.5332,
"step": 3270
},
{
"epoch": 6.07,
"grad_norm": 2.268313407897949,
"learning_rate": 0.001967592592592593,
"loss": 0.4762,
"step": 3280
},
{
"epoch": 6.09,
"grad_norm": 2.3177573680877686,
"learning_rate": 0.001958333333333333,
"loss": 0.5045,
"step": 3290
},
{
"epoch": 6.11,
"grad_norm": 2.091614007949829,
"learning_rate": 0.0019490740740740742,
"loss": 0.426,
"step": 3300
},
{
"epoch": 6.12,
"grad_norm": 2.227275848388672,
"learning_rate": 0.0019398148148148148,
"loss": 0.5102,
"step": 3310
},
{
"epoch": 6.14,
"grad_norm": 1.47825026512146,
"learning_rate": 0.0019305555555555556,
"loss": 0.4559,
"step": 3320
},
{
"epoch": 6.16,
"grad_norm": 2.3970351219177246,
"learning_rate": 0.0019212962962962962,
"loss": 0.5202,
"step": 3330
},
{
"epoch": 6.18,
"grad_norm": 1.6059942245483398,
"learning_rate": 0.0019120370370370372,
"loss": 0.4684,
"step": 3340
},
{
"epoch": 6.2,
"grad_norm": 1.4332858324050903,
"learning_rate": 0.0019027777777777778,
"loss": 0.5029,
"step": 3350
},
{
"epoch": 6.22,
"grad_norm": 2.158965587615967,
"learning_rate": 0.0018935185185185186,
"loss": 0.5185,
"step": 3360
},
{
"epoch": 6.23,
"grad_norm": 2.077367067337036,
"learning_rate": 0.0018842592592592591,
"loss": 0.472,
"step": 3370
},
{
"epoch": 6.25,
"grad_norm": 2.0703649520874023,
"learning_rate": 0.001875,
"loss": 0.4818,
"step": 3380
},
{
"epoch": 6.27,
"grad_norm": 2.000805139541626,
"learning_rate": 0.001865740740740741,
"loss": 0.4643,
"step": 3390
},
{
"epoch": 6.29,
"grad_norm": 1.556027889251709,
"learning_rate": 0.0018564814814814815,
"loss": 0.468,
"step": 3400
},
{
"epoch": 6.31,
"grad_norm": 1.2694203853607178,
"learning_rate": 0.0018472222222222223,
"loss": 0.4437,
"step": 3410
},
{
"epoch": 6.33,
"grad_norm": 2.0669643878936768,
"learning_rate": 0.001837962962962963,
"loss": 0.4565,
"step": 3420
},
{
"epoch": 6.35,
"grad_norm": 1.6692304611206055,
"learning_rate": 0.001828703703703704,
"loss": 0.4573,
"step": 3430
},
{
"epoch": 6.36,
"grad_norm": 1.6164546012878418,
"learning_rate": 0.0018194444444444445,
"loss": 0.4282,
"step": 3440
},
{
"epoch": 6.38,
"grad_norm": 1.9918891191482544,
"learning_rate": 0.0018101851851851853,
"loss": 0.4522,
"step": 3450
},
{
"epoch": 6.4,
"grad_norm": 1.4127854108810425,
"learning_rate": 0.0018009259259259259,
"loss": 0.5317,
"step": 3460
},
{
"epoch": 6.42,
"grad_norm": 1.8190535306930542,
"learning_rate": 0.0017916666666666667,
"loss": 0.5193,
"step": 3470
},
{
"epoch": 6.44,
"grad_norm": 1.8793113231658936,
"learning_rate": 0.0017824074074074072,
"loss": 0.5322,
"step": 3480
},
{
"epoch": 6.46,
"grad_norm": 1.1124266386032104,
"learning_rate": 0.0017731481481481483,
"loss": 0.4849,
"step": 3490
},
{
"epoch": 6.48,
"grad_norm": 1.5575320720672607,
"learning_rate": 0.001763888888888889,
"loss": 0.5254,
"step": 3500
},
{
"epoch": 6.49,
"grad_norm": 1.3481545448303223,
"learning_rate": 0.0017546296296296296,
"loss": 0.4869,
"step": 3510
},
{
"epoch": 6.51,
"grad_norm": 1.0486959218978882,
"learning_rate": 0.0017453703703703704,
"loss": 0.5048,
"step": 3520
},
{
"epoch": 6.53,
"grad_norm": 1.848264217376709,
"learning_rate": 0.001736111111111111,
"loss": 0.5064,
"step": 3530
},
{
"epoch": 6.55,
"grad_norm": 1.6750601530075073,
"learning_rate": 0.001726851851851852,
"loss": 0.4459,
"step": 3540
},
{
"epoch": 6.57,
"grad_norm": 1.5514721870422363,
"learning_rate": 0.0017175925925925926,
"loss": 0.4382,
"step": 3550
},
{
"epoch": 6.59,
"grad_norm": 1.3411380052566528,
"learning_rate": 0.0017083333333333334,
"loss": 0.4007,
"step": 3560
},
{
"epoch": 6.6,
"grad_norm": 1.8607478141784668,
"learning_rate": 0.001699074074074074,
"loss": 0.4776,
"step": 3570
},
{
"epoch": 6.62,
"grad_norm": 1.6788519620895386,
"learning_rate": 0.0016898148148148148,
"loss": 0.4499,
"step": 3580
},
{
"epoch": 6.64,
"grad_norm": 1.4055275917053223,
"learning_rate": 0.0016805555555555558,
"loss": 0.5017,
"step": 3590
},
{
"epoch": 6.66,
"grad_norm": 1.9145076274871826,
"learning_rate": 0.0016712962962962964,
"loss": 0.4196,
"step": 3600
},
{
"epoch": 6.68,
"grad_norm": 1.7690577507019043,
"learning_rate": 0.0016620370370370372,
"loss": 0.4792,
"step": 3610
},
{
"epoch": 6.7,
"grad_norm": 1.713212490081787,
"learning_rate": 0.0016527777777777778,
"loss": 0.4418,
"step": 3620
},
{
"epoch": 6.72,
"grad_norm": 2.4528722763061523,
"learning_rate": 0.0016435185185185185,
"loss": 0.4519,
"step": 3630
},
{
"epoch": 6.73,
"grad_norm": 1.8348699808120728,
"learning_rate": 0.0016342592592592591,
"loss": 0.3959,
"step": 3640
},
{
"epoch": 6.75,
"grad_norm": 1.38786780834198,
"learning_rate": 0.0016250000000000001,
"loss": 0.4433,
"step": 3650
},
{
"epoch": 6.77,
"grad_norm": 0.8413279056549072,
"learning_rate": 0.0016157407407407407,
"loss": 0.4869,
"step": 3660
},
{
"epoch": 6.79,
"grad_norm": 1.3703649044036865,
"learning_rate": 0.0016064814814814815,
"loss": 0.5021,
"step": 3670
},
{
"epoch": 6.81,
"grad_norm": 1.6931146383285522,
"learning_rate": 0.001597222222222222,
"loss": 0.416,
"step": 3680
},
{
"epoch": 6.83,
"grad_norm": 1.314029574394226,
"learning_rate": 0.001587962962962963,
"loss": 0.4489,
"step": 3690
},
{
"epoch": 6.85,
"grad_norm": 1.6631927490234375,
"learning_rate": 0.001578703703703704,
"loss": 0.4334,
"step": 3700
},
{
"epoch": 6.86,
"grad_norm": 1.821455478668213,
"learning_rate": 0.0015694444444444445,
"loss": 0.4307,
"step": 3710
},
{
"epoch": 6.88,
"grad_norm": 1.4613134860992432,
"learning_rate": 0.0015601851851851853,
"loss": 0.414,
"step": 3720
},
{
"epoch": 6.9,
"grad_norm": 1.9156243801116943,
"learning_rate": 0.0015509259259259259,
"loss": 0.5022,
"step": 3730
},
{
"epoch": 6.92,
"grad_norm": 1.9036259651184082,
"learning_rate": 0.0015416666666666669,
"loss": 0.4562,
"step": 3740
},
{
"epoch": 6.94,
"grad_norm": 1.6884894371032715,
"learning_rate": 0.0015324074074074075,
"loss": 0.4418,
"step": 3750
},
{
"epoch": 6.96,
"grad_norm": 1.4793485403060913,
"learning_rate": 0.0015231481481481483,
"loss": 0.4091,
"step": 3760
},
{
"epoch": 6.98,
"grad_norm": 1.9029099941253662,
"learning_rate": 0.0015138888888888888,
"loss": 0.4827,
"step": 3770
},
{
"epoch": 6.99,
"grad_norm": 1.2703421115875244,
"learning_rate": 0.0015046296296296296,
"loss": 0.4066,
"step": 3780
},
{
"epoch": 7.0,
"eval_accuracy": 0.9565552303189031,
"eval_f1": 0.96141521763694,
"eval_loss": 0.11959208548069,
"eval_precision": 0.9693468985322017,
"eval_recall": 0.9563045431610278,
"eval_runtime": 55.0215,
"eval_samples_per_second": 117.972,
"eval_steps_per_second": 7.379,
"step": 3783
},
{
"epoch": 7.01,
"grad_norm": 1.4109677076339722,
"learning_rate": 0.0014953703703703702,
"loss": 0.434,
"step": 3790
},
{
"epoch": 7.03,
"grad_norm": 1.6477872133255005,
"learning_rate": 0.0014861111111111112,
"loss": 0.466,
"step": 3800
},
{
"epoch": 7.05,
"grad_norm": 0.9931581020355225,
"learning_rate": 0.001476851851851852,
"loss": 0.4282,
"step": 3810
},
{
"epoch": 7.07,
"grad_norm": 2.313886880874634,
"learning_rate": 0.0014675925925925926,
"loss": 0.4249,
"step": 3820
},
{
"epoch": 7.09,
"grad_norm": 1.5501474142074585,
"learning_rate": 0.0014583333333333334,
"loss": 0.4062,
"step": 3830
},
{
"epoch": 7.1,
"grad_norm": 2.267270088195801,
"learning_rate": 0.001449074074074074,
"loss": 0.4546,
"step": 3840
},
{
"epoch": 7.12,
"grad_norm": 2.5071542263031006,
"learning_rate": 0.001439814814814815,
"loss": 0.4828,
"step": 3850
},
{
"epoch": 7.14,
"grad_norm": 1.464120864868164,
"learning_rate": 0.0014305555555555556,
"loss": 0.416,
"step": 3860
},
{
"epoch": 7.16,
"grad_norm": 1.3284567594528198,
"learning_rate": 0.0014212962962962964,
"loss": 0.3704,
"step": 3870
},
{
"epoch": 7.18,
"grad_norm": 1.5376476049423218,
"learning_rate": 0.001412037037037037,
"loss": 0.454,
"step": 3880
},
{
"epoch": 7.2,
"grad_norm": 1.0760722160339355,
"learning_rate": 0.0014027777777777777,
"loss": 0.4538,
"step": 3890
},
{
"epoch": 7.22,
"grad_norm": 2.0362236499786377,
"learning_rate": 0.0013935185185185188,
"loss": 0.4178,
"step": 3900
},
{
"epoch": 7.23,
"grad_norm": 1.949922800064087,
"learning_rate": 0.0013842592592592593,
"loss": 0.3457,
"step": 3910
},
{
"epoch": 7.25,
"grad_norm": 1.173089623451233,
"learning_rate": 0.0013750000000000001,
"loss": 0.389,
"step": 3920
},
{
"epoch": 7.27,
"grad_norm": 1.4733856916427612,
"learning_rate": 0.0013657407407407407,
"loss": 0.456,
"step": 3930
},
{
"epoch": 7.29,
"grad_norm": 1.5104173421859741,
"learning_rate": 0.0013564814814814815,
"loss": 0.406,
"step": 3940
},
{
"epoch": 7.31,
"grad_norm": 2.4948642253875732,
"learning_rate": 0.001347222222222222,
"loss": 0.4195,
"step": 3950
},
{
"epoch": 7.33,
"grad_norm": 2.137315511703491,
"learning_rate": 0.001337962962962963,
"loss": 0.4857,
"step": 3960
},
{
"epoch": 7.35,
"grad_norm": 1.20357084274292,
"learning_rate": 0.0013287037037037037,
"loss": 0.4287,
"step": 3970
},
{
"epoch": 7.36,
"grad_norm": 1.5241217613220215,
"learning_rate": 0.0013194444444444445,
"loss": 0.4535,
"step": 3980
},
{
"epoch": 7.38,
"grad_norm": 1.4119207859039307,
"learning_rate": 0.001310185185185185,
"loss": 0.486,
"step": 3990
},
{
"epoch": 7.4,
"grad_norm": 1.3262003660202026,
"learning_rate": 0.0013009259259259259,
"loss": 0.4833,
"step": 4000
},
{
"epoch": 7.42,
"grad_norm": 1.0021650791168213,
"learning_rate": 0.0012916666666666669,
"loss": 0.3996,
"step": 4010
},
{
"epoch": 7.44,
"grad_norm": 1.8423185348510742,
"learning_rate": 0.0012824074074074075,
"loss": 0.4569,
"step": 4020
},
{
"epoch": 7.46,
"grad_norm": 2.010634183883667,
"learning_rate": 0.0012731481481481483,
"loss": 0.4673,
"step": 4030
},
{
"epoch": 7.47,
"grad_norm": 2.974384307861328,
"learning_rate": 0.0012638888888888888,
"loss": 0.3992,
"step": 4040
},
{
"epoch": 7.49,
"grad_norm": 1.022455096244812,
"learning_rate": 0.0012546296296296296,
"loss": 0.3598,
"step": 4050
},
{
"epoch": 7.51,
"grad_norm": 1.5852398872375488,
"learning_rate": 0.0012453703703703704,
"loss": 0.4115,
"step": 4060
},
{
"epoch": 7.53,
"grad_norm": 1.3596739768981934,
"learning_rate": 0.0012361111111111112,
"loss": 0.3534,
"step": 4070
},
{
"epoch": 7.55,
"grad_norm": 1.9260822534561157,
"learning_rate": 0.001226851851851852,
"loss": 0.4297,
"step": 4080
},
{
"epoch": 7.57,
"grad_norm": 1.5783950090408325,
"learning_rate": 0.0012175925925925926,
"loss": 0.368,
"step": 4090
},
{
"epoch": 7.59,
"grad_norm": 1.7927485704421997,
"learning_rate": 0.0012083333333333334,
"loss": 0.3701,
"step": 4100
},
{
"epoch": 7.6,
"grad_norm": 1.1680549383163452,
"learning_rate": 0.0011990740740740742,
"loss": 0.3883,
"step": 4110
},
{
"epoch": 7.62,
"grad_norm": 1.4916154146194458,
"learning_rate": 0.0011898148148148148,
"loss": 0.373,
"step": 4120
},
{
"epoch": 7.64,
"grad_norm": 1.3623180389404297,
"learning_rate": 0.0011805555555555556,
"loss": 0.3697,
"step": 4130
},
{
"epoch": 7.66,
"grad_norm": 1.2618119716644287,
"learning_rate": 0.0011712962962962964,
"loss": 0.4114,
"step": 4140
},
{
"epoch": 7.68,
"grad_norm": 1.7506533861160278,
"learning_rate": 0.001162037037037037,
"loss": 0.4003,
"step": 4150
},
{
"epoch": 7.7,
"grad_norm": 1.0284491777420044,
"learning_rate": 0.001152777777777778,
"loss": 0.3749,
"step": 4160
},
{
"epoch": 7.72,
"grad_norm": 1.344994306564331,
"learning_rate": 0.0011435185185185185,
"loss": 0.4025,
"step": 4170
},
{
"epoch": 7.73,
"grad_norm": 3.066920042037964,
"learning_rate": 0.0011342592592592593,
"loss": 0.4203,
"step": 4180
},
{
"epoch": 7.75,
"grad_norm": 1.4180749654769897,
"learning_rate": 0.0011250000000000001,
"loss": 0.4364,
"step": 4190
},
{
"epoch": 7.77,
"grad_norm": 1.5211139917373657,
"learning_rate": 0.0011157407407407407,
"loss": 0.4133,
"step": 4200
},
{
"epoch": 7.79,
"grad_norm": 1.9516364336013794,
"learning_rate": 0.0011064814814814815,
"loss": 0.4232,
"step": 4210
},
{
"epoch": 7.81,
"grad_norm": 2.458101511001587,
"learning_rate": 0.0010972222222222223,
"loss": 0.434,
"step": 4220
},
{
"epoch": 7.83,
"grad_norm": 1.2099213600158691,
"learning_rate": 0.0010879629629629629,
"loss": 0.3169,
"step": 4230
},
{
"epoch": 7.84,
"grad_norm": 1.0573714971542358,
"learning_rate": 0.0010787037037037037,
"loss": 0.3777,
"step": 4240
},
{
"epoch": 7.86,
"grad_norm": 1.4449975490570068,
"learning_rate": 0.0010694444444444445,
"loss": 0.4426,
"step": 4250
},
{
"epoch": 7.88,
"grad_norm": 1.335119366645813,
"learning_rate": 0.001060185185185185,
"loss": 0.4643,
"step": 4260
},
{
"epoch": 7.9,
"grad_norm": 2.1732048988342285,
"learning_rate": 0.001050925925925926,
"loss": 0.3903,
"step": 4270
},
{
"epoch": 7.92,
"grad_norm": 1.193305492401123,
"learning_rate": 0.0010416666666666667,
"loss": 0.3744,
"step": 4280
},
{
"epoch": 7.94,
"grad_norm": 0.9779842495918274,
"learning_rate": 0.0010324074074074074,
"loss": 0.4124,
"step": 4290
},
{
"epoch": 7.96,
"grad_norm": 2.7513339519500732,
"learning_rate": 0.0010231481481481482,
"loss": 0.3917,
"step": 4300
},
{
"epoch": 7.97,
"grad_norm": 1.756395936012268,
"learning_rate": 0.0010138888888888888,
"loss": 0.4301,
"step": 4310
},
{
"epoch": 7.99,
"grad_norm": 1.5804221630096436,
"learning_rate": 0.0010046296296296296,
"loss": 0.4046,
"step": 4320
},
{
"epoch": 8.0,
"eval_accuracy": 0.979818209828994,
"eval_f1": 0.9833721625743178,
"eval_loss": 0.04691172018647194,
"eval_precision": 0.9853003976747072,
"eval_recall": 0.9820557579494299,
"eval_runtime": 54.8438,
"eval_samples_per_second": 118.354,
"eval_steps_per_second": 7.403,
"step": 4324
},
{
"epoch": 8.01,
"grad_norm": 1.8008298873901367,
"learning_rate": 0.0009953703703703704,
"loss": 0.4452,
"step": 4330
},
{
"epoch": 8.03,
"grad_norm": 3.602733850479126,
"learning_rate": 0.000986111111111111,
"loss": 0.3281,
"step": 4340
},
{
"epoch": 8.05,
"grad_norm": 1.5571168661117554,
"learning_rate": 0.0009768518518518518,
"loss": 0.432,
"step": 4350
},
{
"epoch": 8.07,
"grad_norm": 1.5973519086837769,
"learning_rate": 0.0009675925925925926,
"loss": 0.3682,
"step": 4360
},
{
"epoch": 8.09,
"grad_norm": 1.1785231828689575,
"learning_rate": 0.0009583333333333334,
"loss": 0.3233,
"step": 4370
},
{
"epoch": 8.1,
"grad_norm": 1.9037755727767944,
"learning_rate": 0.0009490740740740742,
"loss": 0.3733,
"step": 4380
},
{
"epoch": 8.12,
"grad_norm": 1.3710215091705322,
"learning_rate": 0.0009398148148148149,
"loss": 0.3562,
"step": 4390
},
{
"epoch": 8.14,
"grad_norm": 1.4512821435928345,
"learning_rate": 0.0009305555555555556,
"loss": 0.3391,
"step": 4400
},
{
"epoch": 8.16,
"grad_norm": 1.5068678855895996,
"learning_rate": 0.0009212962962962964,
"loss": 0.3703,
"step": 4410
},
{
"epoch": 8.18,
"grad_norm": 1.2810806035995483,
"learning_rate": 0.000912037037037037,
"loss": 0.3729,
"step": 4420
},
{
"epoch": 8.2,
"grad_norm": 1.9714285135269165,
"learning_rate": 0.0009027777777777777,
"loss": 0.3988,
"step": 4430
},
{
"epoch": 8.21,
"grad_norm": 1.3327577114105225,
"learning_rate": 0.0008935185185185185,
"loss": 0.4024,
"step": 4440
},
{
"epoch": 8.23,
"grad_norm": 1.0256530046463013,
"learning_rate": 0.0008842592592592592,
"loss": 0.3818,
"step": 4450
},
{
"epoch": 8.25,
"grad_norm": 2.233092784881592,
"learning_rate": 0.0008749999999999999,
"loss": 0.3668,
"step": 4460
},
{
"epoch": 8.27,
"grad_norm": 1.3951568603515625,
"learning_rate": 0.0008657407407407407,
"loss": 0.44,
"step": 4470
},
{
"epoch": 8.29,
"grad_norm": 1.827919840812683,
"learning_rate": 0.0008564814814814815,
"loss": 0.3801,
"step": 4480
},
{
"epoch": 8.31,
"grad_norm": 2.934967041015625,
"learning_rate": 0.0008472222222222223,
"loss": 0.422,
"step": 4490
},
{
"epoch": 8.33,
"grad_norm": 1.2205818891525269,
"learning_rate": 0.000837962962962963,
"loss": 0.3991,
"step": 4500
},
{
"epoch": 8.34,
"grad_norm": 1.424593448638916,
"learning_rate": 0.0008287037037037038,
"loss": 0.418,
"step": 4510
},
{
"epoch": 8.36,
"grad_norm": 1.0254913568496704,
"learning_rate": 0.0008194444444444445,
"loss": 0.3876,
"step": 4520
},
{
"epoch": 8.38,
"grad_norm": 0.9939424991607666,
"learning_rate": 0.0008101851851851852,
"loss": 0.4267,
"step": 4530
},
{
"epoch": 8.4,
"grad_norm": 1.7678611278533936,
"learning_rate": 0.000800925925925926,
"loss": 0.3916,
"step": 4540
},
{
"epoch": 8.42,
"grad_norm": 1.0343542098999023,
"learning_rate": 0.0007916666666666666,
"loss": 0.3939,
"step": 4550
},
{
"epoch": 8.44,
"grad_norm": 0.9549526572227478,
"learning_rate": 0.0007824074074074073,
"loss": 0.3785,
"step": 4560
},
{
"epoch": 8.46,
"grad_norm": 1.0010316371917725,
"learning_rate": 0.0007731481481481481,
"loss": 0.3508,
"step": 4570
},
{
"epoch": 8.47,
"grad_norm": 2.0805888175964355,
"learning_rate": 0.0007638888888888889,
"loss": 0.3614,
"step": 4580
},
{
"epoch": 8.49,
"grad_norm": 1.181626558303833,
"learning_rate": 0.0007546296296296297,
"loss": 0.359,
"step": 4590
},
{
"epoch": 8.51,
"grad_norm": 1.138376235961914,
"learning_rate": 0.0007453703703703704,
"loss": 0.3513,
"step": 4600
},
{
"epoch": 8.53,
"grad_norm": 1.0916022062301636,
"learning_rate": 0.0007361111111111111,
"loss": 0.3199,
"step": 4610
},
{
"epoch": 8.55,
"grad_norm": 2.3587114810943604,
"learning_rate": 0.0007268518518518519,
"loss": 0.3658,
"step": 4620
},
{
"epoch": 8.57,
"grad_norm": 1.2504198551177979,
"learning_rate": 0.0007175925925925926,
"loss": 0.3952,
"step": 4630
},
{
"epoch": 8.58,
"grad_norm": 1.3484505414962769,
"learning_rate": 0.0007083333333333333,
"loss": 0.3542,
"step": 4640
},
{
"epoch": 8.6,
"grad_norm": 2.4245710372924805,
"learning_rate": 0.0006990740740740741,
"loss": 0.3421,
"step": 4650
},
{
"epoch": 8.62,
"grad_norm": 1.2606480121612549,
"learning_rate": 0.0006898148148148148,
"loss": 0.3404,
"step": 4660
},
{
"epoch": 8.64,
"grad_norm": 1.913200855255127,
"learning_rate": 0.0006805555555555556,
"loss": 0.4015,
"step": 4670
},
{
"epoch": 8.66,
"grad_norm": 1.5260008573532104,
"learning_rate": 0.0006712962962962964,
"loss": 0.3425,
"step": 4680
},
{
"epoch": 8.68,
"grad_norm": 1.221198320388794,
"learning_rate": 0.0006620370370370372,
"loss": 0.4078,
"step": 4690
},
{
"epoch": 8.7,
"grad_norm": 1.630136489868164,
"learning_rate": 0.0006527777777777778,
"loss": 0.3508,
"step": 4700
},
{
"epoch": 8.71,
"grad_norm": 1.3053523302078247,
"learning_rate": 0.0006435185185185185,
"loss": 0.2958,
"step": 4710
},
{
"epoch": 8.73,
"grad_norm": 1.5121403932571411,
"learning_rate": 0.0006342592592592593,
"loss": 0.3552,
"step": 4720
},
{
"epoch": 8.75,
"grad_norm": 1.1451386213302612,
"learning_rate": 0.000625,
"loss": 0.3688,
"step": 4730
},
{
"epoch": 8.77,
"grad_norm": 2.4750773906707764,
"learning_rate": 0.0006157407407407407,
"loss": 0.3435,
"step": 4740
},
{
"epoch": 8.79,
"grad_norm": 0.6995977163314819,
"learning_rate": 0.0006064814814814815,
"loss": 0.3152,
"step": 4750
},
{
"epoch": 8.81,
"grad_norm": 1.6774075031280518,
"learning_rate": 0.0005972222222222223,
"loss": 0.2898,
"step": 4760
},
{
"epoch": 8.83,
"grad_norm": 1.1041797399520874,
"learning_rate": 0.000587962962962963,
"loss": 0.3086,
"step": 4770
},
{
"epoch": 8.84,
"grad_norm": 1.2594033479690552,
"learning_rate": 0.0005787037037037037,
"loss": 0.3981,
"step": 4780
},
{
"epoch": 8.86,
"grad_norm": 0.9759043455123901,
"learning_rate": 0.0005694444444444445,
"loss": 0.356,
"step": 4790
},
{
"epoch": 8.88,
"grad_norm": 1.4041826725006104,
"learning_rate": 0.0005601851851851852,
"loss": 0.3514,
"step": 4800
},
{
"epoch": 8.9,
"grad_norm": 0.9928153157234192,
"learning_rate": 0.000550925925925926,
"loss": 0.388,
"step": 4810
},
{
"epoch": 8.92,
"grad_norm": 1.0990333557128906,
"learning_rate": 0.0005416666666666668,
"loss": 0.3013,
"step": 4820
},
{
"epoch": 8.94,
"grad_norm": 1.2321481704711914,
"learning_rate": 0.0005324074074074074,
"loss": 0.4016,
"step": 4830
},
{
"epoch": 8.95,
"grad_norm": 1.1745514869689941,
"learning_rate": 0.0005231481481481481,
"loss": 0.3601,
"step": 4840
},
{
"epoch": 8.97,
"grad_norm": 1.633880615234375,
"learning_rate": 0.0005138888888888889,
"loss": 0.3291,
"step": 4850
},
{
"epoch": 8.99,
"grad_norm": 1.1216607093811035,
"learning_rate": 0.0005046296296296297,
"loss": 0.3314,
"step": 4860
},
{
"epoch": 9.0,
"eval_accuracy": 0.986134647974118,
"eval_f1": 0.987440563210657,
"eval_loss": 0.038755957037210464,
"eval_precision": 0.9891828577753293,
"eval_recall": 0.98596866524898,
"eval_runtime": 55.0297,
"eval_samples_per_second": 117.955,
"eval_steps_per_second": 7.378,
"step": 4864
},
{
"epoch": 9.01,
"grad_norm": 1.2497656345367432,
"learning_rate": 0.0004953703703703704,
"loss": 0.3183,
"step": 4870
},
{
"epoch": 9.03,
"grad_norm": 1.5892603397369385,
"learning_rate": 0.00048611111111111115,
"loss": 0.3222,
"step": 4880
},
{
"epoch": 9.05,
"grad_norm": 1.4494744539260864,
"learning_rate": 0.00047685185185185184,
"loss": 0.3573,
"step": 4890
},
{
"epoch": 9.07,
"grad_norm": 1.7034820318222046,
"learning_rate": 0.0004675925925925926,
"loss": 0.3194,
"step": 4900
},
{
"epoch": 9.08,
"grad_norm": 0.8668962121009827,
"learning_rate": 0.0004583333333333333,
"loss": 0.3483,
"step": 4910
},
{
"epoch": 9.1,
"grad_norm": 1.081874132156372,
"learning_rate": 0.0004490740740740741,
"loss": 0.3626,
"step": 4920
},
{
"epoch": 9.12,
"grad_norm": 1.78743314743042,
"learning_rate": 0.0004398148148148148,
"loss": 0.3651,
"step": 4930
},
{
"epoch": 9.14,
"grad_norm": 1.1324867010116577,
"learning_rate": 0.00043055555555555555,
"loss": 0.2714,
"step": 4940
},
{
"epoch": 9.16,
"grad_norm": 1.9258543252944946,
"learning_rate": 0.0004212962962962963,
"loss": 0.3635,
"step": 4950
},
{
"epoch": 9.18,
"grad_norm": 1.2044012546539307,
"learning_rate": 0.00041203703703703704,
"loss": 0.3126,
"step": 4960
},
{
"epoch": 9.2,
"grad_norm": 1.3639435768127441,
"learning_rate": 0.00040277777777777783,
"loss": 0.3203,
"step": 4970
},
{
"epoch": 9.21,
"grad_norm": 1.7811472415924072,
"learning_rate": 0.0003935185185185185,
"loss": 0.3208,
"step": 4980
},
{
"epoch": 9.23,
"grad_norm": 1.4253028631210327,
"learning_rate": 0.00038425925925925927,
"loss": 0.3326,
"step": 4990
},
{
"epoch": 9.25,
"grad_norm": 1.4700595140457153,
"learning_rate": 0.000375,
"loss": 0.3015,
"step": 5000
},
{
"epoch": 9.27,
"grad_norm": 2.1158525943756104,
"learning_rate": 0.0003657407407407407,
"loss": 0.3472,
"step": 5010
},
{
"epoch": 9.29,
"grad_norm": 1.325528621673584,
"learning_rate": 0.0003564814814814815,
"loss": 0.3365,
"step": 5020
},
{
"epoch": 9.31,
"grad_norm": 1.173003077507019,
"learning_rate": 0.00034722222222222224,
"loss": 0.3213,
"step": 5030
},
{
"epoch": 9.32,
"grad_norm": 2.355008840560913,
"learning_rate": 0.000337962962962963,
"loss": 0.3386,
"step": 5040
},
{
"epoch": 9.34,
"grad_norm": 1.198581576347351,
"learning_rate": 0.0003287037037037037,
"loss": 0.2985,
"step": 5050
},
{
"epoch": 9.36,
"grad_norm": 1.7199565172195435,
"learning_rate": 0.0003194444444444444,
"loss": 0.3666,
"step": 5060
},
{
"epoch": 9.38,
"grad_norm": 1.359098196029663,
"learning_rate": 0.0003101851851851852,
"loss": 0.327,
"step": 5070
},
{
"epoch": 9.4,
"grad_norm": 1.2990363836288452,
"learning_rate": 0.0003009259259259259,
"loss": 0.3673,
"step": 5080
},
{
"epoch": 9.42,
"grad_norm": 1.5110586881637573,
"learning_rate": 0.0002916666666666667,
"loss": 0.3124,
"step": 5090
},
{
"epoch": 9.44,
"grad_norm": 1.6193897724151611,
"learning_rate": 0.0002824074074074074,
"loss": 0.3056,
"step": 5100
},
{
"epoch": 9.45,
"grad_norm": 1.2994916439056396,
"learning_rate": 0.0002731481481481482,
"loss": 0.3041,
"step": 5110
},
{
"epoch": 9.47,
"grad_norm": 1.1582506895065308,
"learning_rate": 0.0002638888888888889,
"loss": 0.325,
"step": 5120
},
{
"epoch": 9.49,
"grad_norm": 1.2252171039581299,
"learning_rate": 0.0002546296296296296,
"loss": 0.3541,
"step": 5130
},
{
"epoch": 9.51,
"grad_norm": 1.0199075937271118,
"learning_rate": 0.0002453703703703704,
"loss": 0.3763,
"step": 5140
},
{
"epoch": 9.53,
"grad_norm": 1.113032341003418,
"learning_rate": 0.00023611111111111112,
"loss": 0.3459,
"step": 5150
},
{
"epoch": 9.55,
"grad_norm": 1.1221752166748047,
"learning_rate": 0.00022685185185185186,
"loss": 0.284,
"step": 5160
},
{
"epoch": 9.57,
"grad_norm": 1.0831656455993652,
"learning_rate": 0.0002175925925925926,
"loss": 0.3117,
"step": 5170
},
{
"epoch": 9.58,
"grad_norm": 0.723981499671936,
"learning_rate": 0.00020833333333333332,
"loss": 0.3231,
"step": 5180
},
{
"epoch": 9.6,
"grad_norm": 1.0349414348602295,
"learning_rate": 0.0001990740740740741,
"loss": 0.3344,
"step": 5190
},
{
"epoch": 9.62,
"grad_norm": 0.9438874125480652,
"learning_rate": 0.0001898148148148148,
"loss": 0.3272,
"step": 5200
},
{
"epoch": 9.64,
"grad_norm": 1.0570435523986816,
"learning_rate": 0.00018055555555555555,
"loss": 0.3309,
"step": 5210
},
{
"epoch": 9.66,
"grad_norm": 0.9774361848831177,
"learning_rate": 0.00017129629629629632,
"loss": 0.3129,
"step": 5220
},
{
"epoch": 9.68,
"grad_norm": 1.1588292121887207,
"learning_rate": 0.00016203703703703703,
"loss": 0.3287,
"step": 5230
},
{
"epoch": 9.69,
"grad_norm": 1.1384470462799072,
"learning_rate": 0.00015277777777777777,
"loss": 0.2838,
"step": 5240
},
{
"epoch": 9.71,
"grad_norm": 1.3432296514511108,
"learning_rate": 0.00014351851851851852,
"loss": 0.3424,
"step": 5250
},
{
"epoch": 9.73,
"grad_norm": 0.6609634160995483,
"learning_rate": 0.00013425925925925926,
"loss": 0.2843,
"step": 5260
},
{
"epoch": 9.75,
"grad_norm": 1.2155038118362427,
"learning_rate": 0.000125,
"loss": 0.3159,
"step": 5270
},
{
"epoch": 9.77,
"grad_norm": 1.5623424053192139,
"learning_rate": 0.00011574074074074073,
"loss": 0.2891,
"step": 5280
},
{
"epoch": 9.79,
"grad_norm": 1.9647743701934814,
"learning_rate": 0.00010648148148148149,
"loss": 0.3394,
"step": 5290
},
{
"epoch": 9.81,
"grad_norm": 1.1939747333526611,
"learning_rate": 9.722222222222223e-05,
"loss": 0.3545,
"step": 5300
},
{
"epoch": 9.82,
"grad_norm": 1.5705734491348267,
"learning_rate": 8.796296296296297e-05,
"loss": 0.372,
"step": 5310
},
{
"epoch": 9.84,
"grad_norm": 1.9765841960906982,
"learning_rate": 7.87037037037037e-05,
"loss": 0.3555,
"step": 5320
},
{
"epoch": 9.86,
"grad_norm": 1.168407917022705,
"learning_rate": 6.944444444444444e-05,
"loss": 0.2545,
"step": 5330
},
{
"epoch": 9.88,
"grad_norm": 0.8777741193771362,
"learning_rate": 6.018518518518519e-05,
"loss": 0.3408,
"step": 5340
},
{
"epoch": 9.9,
"grad_norm": 1.2033889293670654,
"learning_rate": 5.092592592592593e-05,
"loss": 0.3518,
"step": 5350
},
{
"epoch": 9.92,
"grad_norm": 1.9429413080215454,
"learning_rate": 4.1666666666666665e-05,
"loss": 0.2779,
"step": 5360
},
{
"epoch": 9.94,
"grad_norm": 1.5541284084320068,
"learning_rate": 3.240740740740741e-05,
"loss": 0.3169,
"step": 5370
},
{
"epoch": 9.95,
"grad_norm": 0.8645866513252258,
"learning_rate": 2.3148148148148147e-05,
"loss": 0.2976,
"step": 5380
},
{
"epoch": 9.97,
"grad_norm": 1.3550152778625488,
"learning_rate": 1.388888888888889e-05,
"loss": 0.2891,
"step": 5390
},
{
"epoch": 9.99,
"grad_norm": 1.666207194328308,
"learning_rate": 4.6296296296296296e-06,
"loss": 0.2865,
"step": 5400
},
{
"epoch": 9.99,
"eval_accuracy": 0.9830534586350331,
"eval_f1": 0.9868535728073075,
"eval_loss": 0.0450444296002388,
"eval_precision": 0.9880250018061109,
"eval_recall": 0.986225426158784,
"eval_runtime": 55.3663,
"eval_samples_per_second": 117.237,
"eval_steps_per_second": 7.333,
"step": 5400
},
{
"epoch": 9.99,
"step": 5400,
"total_flos": 2.695096772600047e+19,
"train_loss": 0.5508582998205114,
"train_runtime": 6179.724,
"train_samples_per_second": 55.959,
"train_steps_per_second": 0.874
},
{
"epoch": 9.99,
"eval_accuracy": 0.9328945888176398,
"eval_f1": 0.9340102214756545,
"eval_loss": 0.23724627494812012,
"eval_precision": 0.9416197933012,
"eval_recall": 0.9295532071921557,
"eval_runtime": 150.877,
"eval_samples_per_second": 117.831,
"eval_steps_per_second": 7.37,
"step": 5400
},
{
"epoch": 9.99,
"eval_accuracy": 0.9328945888176398,
"eval_f1": 0.9340102214756545,
"eval_loss": 0.23724627494812012,
"eval_precision": 0.9416197933012,
"eval_recall": 0.9295532071921557,
"eval_runtime": 150.8426,
"eval_samples_per_second": 117.858,
"eval_steps_per_second": 7.372,
"step": 5400
}
],
"logging_steps": 10,
"max_steps": 5400,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 500,
"total_flos": 2.695096772600047e+19,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}