sandernotenbaert commited on
Commit
5baf346
·
verified ·
1 Parent(s): 569262c

Training in progress, step 7000, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66f0bf969e7c4f10819387c514099288befb4d6a92b18fedb97919efb7838146
3
  size 30214176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc781d4fe42e6432654e1870cb89122dc1779571f97b3368cac763d149c179c6
3
  size 30214176
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50326021a12350d0fd340fb39b69dcb99c4d9d7625607f1ea78848e8533735ef
3
  size 291962
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c480a79986a9324e56009cbfffe3ce2c159495030d19916cb8eceb85e6eb6c47
3
  size 291962
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2aba6f3b60d25e9d80bfa48704513608d3c51338a99fc68164f2cda5539a2e88
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f88d979933c32dc472f41ac1057650933f8e3f97e8756c84857a83fcd3723c20
3
  size 14244
last-checkpoint/scaler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b18b3eda07668461b2815107443aa6a6c68056a4e1c0be8b74117cc02ff9391
3
  size 988
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50284aeeef15b76b3e429d017933740783393c66d3932640c0600cbe6c9ede27
3
  size 988
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2167d86470a0af7e2c7e124cb3b9e089d70c4424875cd5dabb7a3f9dabe21d16
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32ea2c91d36474b100bc2bfe7dbfa12a58f0438100ede285e747c4f76e5d7201
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 6500,
3
- "best_metric": 1.5628445148468018,
4
- "best_model_checkpoint": "./results/hierarchical_music_t5_small_finetune/checkpoint-6500",
5
- "epoch": 2.894058717128148,
6
  "eval_steps": 500,
7
- "global_step": 6500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -1022,6 +1022,84 @@
1022
  "eval_samples_per_second": 389.105,
1023
  "eval_steps_per_second": 48.653,
1024
  "step": 6500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1025
  }
1026
  ],
1027
  "logging_steps": 50,
@@ -1050,7 +1128,7 @@
1050
  "attributes": {}
1051
  }
1052
  },
1053
- "total_flos": 2.8160444868950016e+16,
1054
  "train_batch_size": 4,
1055
  "trial_name": null,
1056
  "trial_params": null
 
1
  {
2
+ "best_global_step": 7000,
3
+ "best_metric": 1.5605802536010742,
4
+ "best_model_checkpoint": "./results/hierarchical_music_t5_small_finetune/checkpoint-7000",
5
+ "epoch": 3.1171003200222622,
6
  "eval_steps": 500,
7
+ "global_step": 7000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
1022
  "eval_samples_per_second": 389.105,
1023
  "eval_steps_per_second": 48.653,
1024
  "step": 6500
1025
+ },
1026
+ {
1027
+ "epoch": 2.916321135383331,
1028
+ "grad_norm": 1.0452390909194946,
1029
+ "learning_rate": 5e-05,
1030
+ "loss": 1.6794,
1031
+ "step": 6550
1032
+ },
1033
+ {
1034
+ "epoch": 2.938583553638514,
1035
+ "grad_norm": 1.338881254196167,
1036
+ "learning_rate": 5e-05,
1037
+ "loss": 1.6678,
1038
+ "step": 6600
1039
+ },
1040
+ {
1041
+ "epoch": 2.960845971893697,
1042
+ "grad_norm": 0.989860475063324,
1043
+ "learning_rate": 5e-05,
1044
+ "loss": 1.6753,
1045
+ "step": 6650
1046
+ },
1047
+ {
1048
+ "epoch": 2.98310839014888,
1049
+ "grad_norm": 1.1380687952041626,
1050
+ "learning_rate": 5e-05,
1051
+ "loss": 1.6639,
1052
+ "step": 6700
1053
+ },
1054
+ {
1055
+ "epoch": 3.0057882287463475,
1056
+ "grad_norm": 1.2292852401733398,
1057
+ "learning_rate": 5e-05,
1058
+ "loss": 1.697,
1059
+ "step": 6750
1060
+ },
1061
+ {
1062
+ "epoch": 3.0280506470015305,
1063
+ "grad_norm": 1.1919242143630981,
1064
+ "learning_rate": 5e-05,
1065
+ "loss": 1.6714,
1066
+ "step": 6800
1067
+ },
1068
+ {
1069
+ "epoch": 3.0503130652567134,
1070
+ "grad_norm": 1.1312869787216187,
1071
+ "learning_rate": 5e-05,
1072
+ "loss": 1.6641,
1073
+ "step": 6850
1074
+ },
1075
+ {
1076
+ "epoch": 3.0725754835118964,
1077
+ "grad_norm": 1.3589369058609009,
1078
+ "learning_rate": 5e-05,
1079
+ "loss": 1.655,
1080
+ "step": 6900
1081
+ },
1082
+ {
1083
+ "epoch": 3.0948379017670793,
1084
+ "grad_norm": 1.257063627243042,
1085
+ "learning_rate": 5e-05,
1086
+ "loss": 1.6661,
1087
+ "step": 6950
1088
+ },
1089
+ {
1090
+ "epoch": 3.1171003200222622,
1091
+ "grad_norm": 1.3228737115859985,
1092
+ "learning_rate": 5e-05,
1093
+ "loss": 1.653,
1094
+ "step": 7000
1095
+ },
1096
+ {
1097
+ "epoch": 3.1171003200222622,
1098
+ "eval_loss": 1.5605802536010742,
1099
+ "eval_runtime": 42.9947,
1100
+ "eval_samples_per_second": 371.464,
1101
+ "eval_steps_per_second": 46.448,
1102
+ "step": 7000
1103
  }
1104
  ],
1105
  "logging_steps": 50,
 
1128
  "attributes": {}
1129
  }
1130
  },
1131
+ "total_flos": 2.864099321051136e+16,
1132
  "train_batch_size": 4,
1133
  "trial_name": null,
1134
  "trial_params": null