fats-fme commited on
Commit
22eba06
·
verified ·
1 Parent(s): 2184ef9

Training in progress, step 200, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:98c209b7bf72e9a0283cbadfa15442c82a5e4eb9519b9d7364acd840d847689f
3
  size 67240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab8c8dd89fe6080fa7fbcec9ea73e9231a0a82c17ca756a794c22c4555da03ca
3
  size 67240
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b69036cb9e8b134c2b1a58619820e6464ce30832ba3994ff33678ebb4607d3d7
3
  size 138758
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb1f79bf8cd7bfde47c4b8900225e45de1a79da8b283da85951074d8b1045e0e
3
  size 138758
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34b25434515ff1ffa5ef57e151ef89f031c4dd81c881229710479bc4c6a54e6d
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edbcf2d15803e2421fecb3f78827c12e7e0cfd18cfa9f829993ebbbd3c374be1
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:691263e55c7bb34b6e090bd89cf079f362100d7bd758be475bd3e7091c1a4ed6
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2d754412c61116546142914503e7369d0cc35d3c380a07e5218f595d76b6d96
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 11.5,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-100",
4
- "epoch": 0.004233789877008404,
5
  "eval_steps": 100,
6
- "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -93,6 +93,84 @@
93
  "eval_samples_per_second": 98.748,
94
  "eval_steps_per_second": 49.379,
95
  "step": 100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
96
  }
97
  ],
98
  "logging_steps": 10,
@@ -107,7 +185,7 @@
107
  "early_stopping_threshold": 0.0
108
  },
109
  "attributes": {
110
- "early_stopping_patience_counter": 0
111
  }
112
  },
113
  "TrainerControl": {
@@ -116,12 +194,12 @@
116
  "should_evaluate": false,
117
  "should_log": false,
118
  "should_save": true,
119
- "should_training_stop": false
120
  },
121
  "attributes": {}
122
  }
123
  },
124
- "total_flos": 8201347792896.0,
125
  "train_batch_size": 2,
126
  "trial_name": null,
127
  "trial_params": null
 
1
  {
2
  "best_metric": 11.5,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-100",
4
+ "epoch": 0.008467579754016808,
5
  "eval_steps": 100,
6
+ "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
93
  "eval_samples_per_second": 98.748,
94
  "eval_steps_per_second": 49.379,
95
  "step": 100
96
+ },
97
+ {
98
+ "epoch": 0.004657168864709245,
99
+ "grad_norm": 0.0001554292975924909,
100
+ "learning_rate": 9.755282581475769e-05,
101
+ "loss": 46.0,
102
+ "step": 110
103
+ },
104
+ {
105
+ "epoch": 0.005080547852410085,
106
+ "grad_norm": 0.00017063747509382665,
107
+ "learning_rate": 9.045084971874738e-05,
108
+ "loss": 46.0,
109
+ "step": 120
110
+ },
111
+ {
112
+ "epoch": 0.005503926840110925,
113
+ "grad_norm": 0.0002251509140478447,
114
+ "learning_rate": 7.938926261462366e-05,
115
+ "loss": 46.0,
116
+ "step": 130
117
+ },
118
+ {
119
+ "epoch": 0.005927305827811766,
120
+ "grad_norm": 0.00025318050757050514,
121
+ "learning_rate": 6.545084971874738e-05,
122
+ "loss": 46.0,
123
+ "step": 140
124
+ },
125
+ {
126
+ "epoch": 0.006350684815512606,
127
+ "grad_norm": 0.00024476656108163297,
128
+ "learning_rate": 5e-05,
129
+ "loss": 46.0,
130
+ "step": 150
131
+ },
132
+ {
133
+ "epoch": 0.006774063803213447,
134
+ "grad_norm": 0.0004327484348323196,
135
+ "learning_rate": 3.4549150281252636e-05,
136
+ "loss": 46.0,
137
+ "step": 160
138
+ },
139
+ {
140
+ "epoch": 0.007197442790914287,
141
+ "grad_norm": 0.00031539108022116125,
142
+ "learning_rate": 2.061073738537635e-05,
143
+ "loss": 46.0,
144
+ "step": 170
145
+ },
146
+ {
147
+ "epoch": 0.007620821778615127,
148
+ "grad_norm": 0.00022764148889109492,
149
+ "learning_rate": 9.549150281252633e-06,
150
+ "loss": 46.0,
151
+ "step": 180
152
+ },
153
+ {
154
+ "epoch": 0.008044200766315967,
155
+ "grad_norm": 0.00025987360277213156,
156
+ "learning_rate": 2.4471741852423237e-06,
157
+ "loss": 46.0,
158
+ "step": 190
159
+ },
160
+ {
161
+ "epoch": 0.008467579754016808,
162
+ "grad_norm": 0.00024760179803706706,
163
+ "learning_rate": 0.0,
164
+ "loss": 46.0,
165
+ "step": 200
166
+ },
167
+ {
168
+ "epoch": 0.008467579754016808,
169
+ "eval_loss": 11.5,
170
+ "eval_runtime": 101.2299,
171
+ "eval_samples_per_second": 98.242,
172
+ "eval_steps_per_second": 49.126,
173
+ "step": 200
174
  }
175
  ],
176
  "logging_steps": 10,
 
185
  "early_stopping_threshold": 0.0
186
  },
187
  "attributes": {
188
+ "early_stopping_patience_counter": 1
189
  }
190
  },
191
  "TrainerControl": {
 
194
  "should_evaluate": false,
195
  "should_log": false,
196
  "should_save": true,
197
+ "should_training_stop": true
198
  },
199
  "attributes": {}
200
  }
201
  },
202
+ "total_flos": 16362094854144.0,
203
  "train_batch_size": 2,
204
  "trial_name": null,
205
  "trial_params": null