sara-nabhani commited on
Commit
f2fcb6e
·
1 Parent(s): 9ab49ae

Training in progress, step 1400

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5bbc7079009dcb015a118d176ee96ef3fa9f81f5e32596bd48e471250f7cd019
3
  size 2843370360
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51a5c4269980c54f82bc3da566ba2c9f23039e5604c94f38ff35d314135c4a3a
3
  size 2843370360
last-checkpoint/pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27d9a7aedcce194e7334bb3ff1bbd48d5ae2a679ca306c2690cc2f3503f52f78
3
  size 1421660981
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f99a4feff5614541872efc0ede3d0a150a715ce72bba1d6deecbce49223acefa
3
  size 1421660981
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:37cb7810664b13600938b962e95e12c94aaeb9d2b88cf011246a5ed3f6f5583d
3
  size 14639
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b0b77b5900549b5144ba01f867f512f9757676e061b6356487054256b9dbf9a5
3
  size 14639
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b31560d2c60ddf21d021e78a5d14bba7dc363334c5f64fc9eb49d0b82d96fc2a
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4735c8454ebaaf0f7e5072d8f53d7dcffa20fd292c672caa239ebbcbcd191bb
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
- "best_metric": 0.35946421270983664,
3
- "best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-0/checkpoint-800",
4
- "epoch": 4.733727810650888,
5
- "global_step": 800,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
@@ -154,11 +154,122 @@
154
  "eval_samples_per_second": 609.632,
155
  "eval_steps_per_second": 19.292,
156
  "step": 800
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
157
  }
158
  ],
159
  "max_steps": 1690,
160
  "num_train_epochs": 10,
161
- "total_flos": 4392119246594136.0,
162
  "trial_name": null,
163
  "trial_params": null
164
  }
 
1
  {
2
+ "best_metric": 0.40153906255014676,
3
+ "best_model_checkpoint": "/home2/s5432073/language-tech-project/results/ltp-roberta-large-defaultltp-roberta-large-default-0/checkpoint-1200",
4
+ "epoch": 8.284023668639053,
5
+ "global_step": 1400,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
 
154
  "eval_samples_per_second": 609.632,
155
  "eval_steps_per_second": 19.292,
156
  "step": 800
157
+ },
158
+ {
159
+ "epoch": 5.92,
160
+ "learning_rate": 4.299065420560748e-06,
161
+ "loss": 0.2582,
162
+ "step": 1000
163
+ },
164
+ {
165
+ "epoch": 5.92,
166
+ "eval_f1": 0.3779236416443889,
167
+ "eval_f1_all": [
168
+ 0.5119617224880383,
169
+ 0.5549872122762148,
170
+ 0.09523809523809525,
171
+ 0.24793388429752064,
172
+ 0.6457739791073124,
173
+ 0.10112359550561797,
174
+ 0.31351351351351353,
175
+ 0.0,
176
+ 0.7588785046728972,
177
+ 0.6312056737588652,
178
+ 0.44999999999999996,
179
+ 0.5300546448087431,
180
+ 0.0,
181
+ 0.015384615384615385,
182
+ 0.5868995633187772,
183
+ 0.17857142857142858,
184
+ 0.6494401378122309,
185
+ 0.7022222222222223,
186
+ 0.12851405622489961,
187
+ 0.45676998368678634
188
+ ],
189
+ "eval_loss": 0.30811235308647156,
190
+ "eval_runtime": 3.5499,
191
+ "eval_samples_per_second": 534.102,
192
+ "eval_steps_per_second": 16.902,
193
+ "step": 1000
194
+ },
195
+ {
196
+ "epoch": 7.1,
197
+ "learning_rate": 3.0529595015576325e-06,
198
+ "loss": 0.2446,
199
+ "step": 1200
200
+ },
201
+ {
202
+ "epoch": 7.1,
203
+ "eval_f1": 0.40153906255014676,
204
+ "eval_f1_all": [
205
+ 0.5172413793103449,
206
+ 0.5871121718377088,
207
+ 0.12,
208
+ 0.3759398496240602,
209
+ 0.630952380952381,
210
+ 0.17258883248730966,
211
+ 0.3768115942028986,
212
+ 0.0,
213
+ 0.7563683866753755,
214
+ 0.6462513199577613,
215
+ 0.4758064516129032,
216
+ 0.5527369826435248,
217
+ 0.0,
218
+ 0.015037593984962405,
219
+ 0.5758683729433272,
220
+ 0.21203438395415472,
221
+ 0.6746794871794872,
222
+ 0.7136563876651982,
223
+ 0.17777777777777778,
224
+ 0.44991789819376027
225
+ ],
226
+ "eval_loss": 0.30632713437080383,
227
+ "eval_runtime": 3.2144,
228
+ "eval_samples_per_second": 589.85,
229
+ "eval_steps_per_second": 18.666,
230
+ "step": 1200
231
+ },
232
+ {
233
+ "epoch": 8.28,
234
+ "learning_rate": 1.8068535825545173e-06,
235
+ "loss": 0.2344,
236
+ "step": 1400
237
+ },
238
+ {
239
+ "epoch": 8.28,
240
+ "eval_f1": 0.39818355623683893,
241
+ "eval_f1_all": [
242
+ 0.5059665871121718,
243
+ 0.6030267753201397,
244
+ 0.13071895424836602,
245
+ 0.3609022556390977,
246
+ 0.6425073457394711,
247
+ 0.17616580310880828,
248
+ 0.3827751196172249,
249
+ 0.0,
250
+ 0.7609942638623327,
251
+ 0.6363636363636364,
252
+ 0.4672131147540983,
253
+ 0.5517241379310345,
254
+ 0.0,
255
+ 0.015151515151515152,
256
+ 0.5907099035933392,
257
+ 0.14012738853503187,
258
+ 0.7006960556844547,
259
+ 0.7053571428571428,
260
+ 0.1484375,
261
+ 0.4448336252189142
262
+ ],
263
+ "eval_loss": 0.30558517575263977,
264
+ "eval_runtime": 3.0733,
265
+ "eval_samples_per_second": 616.921,
266
+ "eval_steps_per_second": 19.523,
267
+ "step": 1400
268
  }
269
  ],
270
  "max_steps": 1690,
271
  "num_train_epochs": 10,
272
+ "total_flos": 7696217086325736.0,
273
  "trial_name": null,
274
  "trial_params": null
275
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:27d9a7aedcce194e7334bb3ff1bbd48d5ae2a679ca306c2690cc2f3503f52f78
3
  size 1421660981
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f99a4feff5614541872efc0ede3d0a150a715ce72bba1d6deecbce49223acefa
3
  size 1421660981