Training in progress, step 600, checkpoint
Browse files
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 349243752
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:170b9839a6f9cb4849a79d372fdd8978c15d9f959a62f6fc4739c2fab1a62cb9
|
3 |
size 349243752
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 177909253
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fde78758d4bfbd2f73a3973efb9b2f5d957019b5c9fc4ad6115838e371515762
|
3 |
size 177909253
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14645
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:996933921c4deed1d34ddceb89eb427578e0472ba88129c76719f873c95cb898
|
3 |
size 14645
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1465
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6557e2f348eac233d3d621679f2b4c7e01f4a6a61c82ffe28816c1229ea4afae
|
3 |
size 1465
|
last-checkpoint/trainer_state.json
CHANGED
@@ -2,9 +2,9 @@
|
|
2 |
"best_global_step": null,
|
3 |
"best_metric": null,
|
4 |
"best_model_checkpoint": null,
|
5 |
-
"epoch": 0.
|
6 |
"eval_steps": 500,
|
7 |
-
"global_step":
|
8 |
"is_hyper_param_search": false,
|
9 |
"is_local_process_zero": true,
|
10 |
"is_world_process_zero": true,
|
@@ -883,6 +883,181 @@
|
|
883 |
"learning_rate": 6.1274380295443624e-06,
|
884 |
"loss": 1.345,
|
885 |
"step": 500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
886 |
}
|
887 |
],
|
888 |
"logging_steps": 4,
|
@@ -902,7 +1077,7 @@
|
|
902 |
"attributes": {}
|
903 |
}
|
904 |
},
|
905 |
-
"total_flos":
|
906 |
"train_batch_size": 24,
|
907 |
"trial_name": null,
|
908 |
"trial_params": null
|
|
|
2 |
"best_global_step": null,
|
3 |
"best_metric": null,
|
4 |
"best_model_checkpoint": null,
|
5 |
+
"epoch": 0.5970149253731343,
|
6 |
"eval_steps": 500,
|
7 |
+
"global_step": 600,
|
8 |
"is_hyper_param_search": false,
|
9 |
"is_local_process_zero": true,
|
10 |
"is_world_process_zero": true,
|
|
|
883 |
"learning_rate": 6.1274380295443624e-06,
|
884 |
"loss": 1.345,
|
885 |
"step": 500
|
886 |
+
},
|
887 |
+
{
|
888 |
+
"epoch": 0.5014925373134328,
|
889 |
+
"grad_norm": 0.16558654606342316,
|
890 |
+
"learning_rate": 6.057762090680162e-06,
|
891 |
+
"loss": 1.3696,
|
892 |
+
"step": 504
|
893 |
+
},
|
894 |
+
{
|
895 |
+
"epoch": 0.5054726368159204,
|
896 |
+
"grad_norm": 0.17274197936058044,
|
897 |
+
"learning_rate": 5.9876866595667085e-06,
|
898 |
+
"loss": 1.4371,
|
899 |
+
"step": 508
|
900 |
+
},
|
901 |
+
{
|
902 |
+
"epoch": 0.5094527363184079,
|
903 |
+
"grad_norm": 0.18550808727741241,
|
904 |
+
"learning_rate": 5.917230303186e-06,
|
905 |
+
"loss": 1.491,
|
906 |
+
"step": 512
|
907 |
+
},
|
908 |
+
{
|
909 |
+
"epoch": 0.5134328358208955,
|
910 |
+
"grad_norm": 0.16486109793186188,
|
911 |
+
"learning_rate": 5.84641168944888e-06,
|
912 |
+
"loss": 1.3596,
|
913 |
+
"step": 516
|
914 |
+
},
|
915 |
+
{
|
916 |
+
"epoch": 0.5174129353233831,
|
917 |
+
"grad_norm": 0.15757621824741364,
|
918 |
+
"learning_rate": 5.775249582248848e-06,
|
919 |
+
"loss": 1.3464,
|
920 |
+
"step": 520
|
921 |
+
},
|
922 |
+
{
|
923 |
+
"epoch": 0.5213930348258706,
|
924 |
+
"grad_norm": 0.1419944018125534,
|
925 |
+
"learning_rate": 5.703762836490429e-06,
|
926 |
+
"loss": 1.3415,
|
927 |
+
"step": 524
|
928 |
+
},
|
929 |
+
{
|
930 |
+
"epoch": 0.5253731343283582,
|
931 |
+
"grad_norm": 0.24455907940864563,
|
932 |
+
"learning_rate": 5.631970393093435e-06,
|
933 |
+
"loss": 1.4322,
|
934 |
+
"step": 528
|
935 |
+
},
|
936 |
+
{
|
937 |
+
"epoch": 0.5293532338308458,
|
938 |
+
"grad_norm": 0.15628467500209808,
|
939 |
+
"learning_rate": 5.559891273974433e-06,
|
940 |
+
"loss": 1.4357,
|
941 |
+
"step": 532
|
942 |
+
},
|
943 |
+
{
|
944 |
+
"epoch": 0.5333333333333333,
|
945 |
+
"grad_norm": 0.15226367115974426,
|
946 |
+
"learning_rate": 5.487544577006756e-06,
|
947 |
+
"loss": 1.3864,
|
948 |
+
"step": 536
|
949 |
+
},
|
950 |
+
{
|
951 |
+
"epoch": 0.5373134328358209,
|
952 |
+
"grad_norm": 0.15544599294662476,
|
953 |
+
"learning_rate": 5.414949470960377e-06,
|
954 |
+
"loss": 1.3954,
|
955 |
+
"step": 540
|
956 |
+
},
|
957 |
+
{
|
958 |
+
"epoch": 0.5412935323383085,
|
959 |
+
"grad_norm": 0.14596706628799438,
|
960 |
+
"learning_rate": 5.3421251904230245e-06,
|
961 |
+
"loss": 1.2908,
|
962 |
+
"step": 544
|
963 |
+
},
|
964 |
+
{
|
965 |
+
"epoch": 0.545273631840796,
|
966 |
+
"grad_norm": 0.14396768808364868,
|
967 |
+
"learning_rate": 5.269091030703825e-06,
|
968 |
+
"loss": 1.3339,
|
969 |
+
"step": 548
|
970 |
+
},
|
971 |
+
{
|
972 |
+
"epoch": 0.5492537313432836,
|
973 |
+
"grad_norm": 0.14690382778644562,
|
974 |
+
"learning_rate": 5.195866342720883e-06,
|
975 |
+
"loss": 1.4412,
|
976 |
+
"step": 552
|
977 |
+
},
|
978 |
+
{
|
979 |
+
"epoch": 0.5532338308457712,
|
980 |
+
"grad_norm": 0.164686918258667,
|
981 |
+
"learning_rate": 5.122470527874117e-06,
|
982 |
+
"loss": 1.3025,
|
983 |
+
"step": 556
|
984 |
+
},
|
985 |
+
{
|
986 |
+
"epoch": 0.5572139303482587,
|
987 |
+
"grad_norm": 0.15687449276447296,
|
988 |
+
"learning_rate": 5.0489230329047216e-06,
|
989 |
+
"loss": 1.4286,
|
990 |
+
"step": 560
|
991 |
+
},
|
992 |
+
{
|
993 |
+
"epoch": 0.5611940298507463,
|
994 |
+
"grad_norm": 0.16417358815670013,
|
995 |
+
"learning_rate": 4.975243344742612e-06,
|
996 |
+
"loss": 1.3747,
|
997 |
+
"step": 564
|
998 |
+
},
|
999 |
+
{
|
1000 |
+
"epoch": 0.5651741293532339,
|
1001 |
+
"grad_norm": 0.140806183218956,
|
1002 |
+
"learning_rate": 4.90145098534323e-06,
|
1003 |
+
"loss": 1.3345,
|
1004 |
+
"step": 568
|
1005 |
+
},
|
1006 |
+
{
|
1007 |
+
"epoch": 0.5691542288557214,
|
1008 |
+
"grad_norm": 0.14276030659675598,
|
1009 |
+
"learning_rate": 4.827565506515056e-06,
|
1010 |
+
"loss": 1.3701,
|
1011 |
+
"step": 572
|
1012 |
+
},
|
1013 |
+
{
|
1014 |
+
"epoch": 0.573134328358209,
|
1015 |
+
"grad_norm": 0.18743367493152618,
|
1016 |
+
"learning_rate": 4.753606484739227e-06,
|
1017 |
+
"loss": 1.323,
|
1018 |
+
"step": 576
|
1019 |
+
},
|
1020 |
+
{
|
1021 |
+
"epoch": 0.5771144278606966,
|
1022 |
+
"grad_norm": 0.13849185407161713,
|
1023 |
+
"learning_rate": 4.679593515982602e-06,
|
1024 |
+
"loss": 1.3634,
|
1025 |
+
"step": 580
|
1026 |
+
},
|
1027 |
+
{
|
1028 |
+
"epoch": 0.5810945273631841,
|
1029 |
+
"grad_norm": 0.16938801109790802,
|
1030 |
+
"learning_rate": 4.605546210505675e-06,
|
1031 |
+
"loss": 1.4166,
|
1032 |
+
"step": 584
|
1033 |
+
},
|
1034 |
+
{
|
1035 |
+
"epoch": 0.5850746268656717,
|
1036 |
+
"grad_norm": 0.1600942611694336,
|
1037 |
+
"learning_rate": 4.531484187666699e-06,
|
1038 |
+
"loss": 1.3556,
|
1039 |
+
"step": 588
|
1040 |
+
},
|
1041 |
+
{
|
1042 |
+
"epoch": 0.5890547263681593,
|
1043 |
+
"grad_norm": 0.18938744068145752,
|
1044 |
+
"learning_rate": 4.457427070723396e-06,
|
1045 |
+
"loss": 1.3305,
|
1046 |
+
"step": 592
|
1047 |
+
},
|
1048 |
+
{
|
1049 |
+
"epoch": 0.5930348258706468,
|
1050 |
+
"grad_norm": 0.1494676023721695,
|
1051 |
+
"learning_rate": 4.383394481633633e-06,
|
1052 |
+
"loss": 1.3438,
|
1053 |
+
"step": 596
|
1054 |
+
},
|
1055 |
+
{
|
1056 |
+
"epoch": 0.5970149253731343,
|
1057 |
+
"grad_norm": 0.17030707001686096,
|
1058 |
+
"learning_rate": 4.309406035856457e-06,
|
1059 |
+
"loss": 1.3414,
|
1060 |
+
"step": 600
|
1061 |
}
|
1062 |
],
|
1063 |
"logging_steps": 4,
|
|
|
1077 |
"attributes": {}
|
1078 |
}
|
1079 |
},
|
1080 |
+
"total_flos": 4.505167220095058e+17,
|
1081 |
"train_batch_size": 24,
|
1082 |
"trial_name": null,
|
1083 |
"trial_params": null
|