besimray commited on
Commit
b622b07
·
verified ·
1 Parent(s): d6bf7a4

Training in progress, step 40, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:061c34a8ed6809cd0904d70b19b43324b3658413d60d98a7f6d48e2c3bffaef4
3
  size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7c64749f73b3f3a5d89b895d9a7cda164f3d9c079c61a743d42dbfdad4b0209
3
  size 335604696
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1b94daf7dfa353715d4f279c393bf368a95cc77df273199877d1783020bbc946
3
  size 170920084
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:690b1653700e58de40b3ff6b61c00bbbac153cf7c5c51151346d224efe7c87fe
3
  size 170920084
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0c68f05fddafe7d9625cc40fb5e03db579f2837d87b6225adf6c39031a56218a
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d9c10501b47aba97f7694378ee8a126f4d92e6dc5888f3a053ba82ed4ba7880d
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:60663a94a33586da5717f6f80de424ce9fe5b18a8c8d13d4ca09aa40f102443b
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f924e37bc06756f5535d9fa2079568e2b7869291abee642e58937943662c2f6f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.3060170114040375,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-30",
4
- "epoch": 0.05545286506469501,
5
  "eval_steps": 5,
6
- "global_step": 30,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -273,6 +273,92 @@
273
  "eval_samples_per_second": 1.311,
274
  "eval_steps_per_second": 0.655,
275
  "step": 30
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
276
  }
277
  ],
278
  "logging_steps": 1,
@@ -287,7 +373,7 @@
287
  "early_stopping_threshold": 0.0
288
  },
289
  "attributes": {
290
- "early_stopping_patience_counter": 0
291
  }
292
  },
293
  "TrainerControl": {
@@ -296,12 +382,12 @@
296
  "should_evaluate": false,
297
  "should_log": false,
298
  "should_save": true,
299
- "should_training_stop": false
300
  },
301
  "attributes": {}
302
  }
303
  },
304
- "total_flos": 4.243520835551232e+16,
305
  "train_batch_size": 2,
306
  "trial_name": null,
307
  "trial_params": null
 
1
  {
2
  "best_metric": 0.3060170114040375,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-30",
4
+ "epoch": 0.07393715341959335,
5
  "eval_steps": 5,
6
+ "global_step": 40,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
273
  "eval_samples_per_second": 1.311,
274
  "eval_steps_per_second": 0.655,
275
  "step": 30
276
+ },
277
+ {
278
+ "epoch": 0.05730129390018484,
279
+ "grad_norm": 6.309504508972168,
280
+ "learning_rate": 0.00019909497617679348,
281
+ "loss": 1.9929,
282
+ "step": 31
283
+ },
284
+ {
285
+ "epoch": 0.059149722735674676,
286
+ "grad_norm": 3.596278190612793,
287
+ "learning_rate": 0.0001990068775649202,
288
+ "loss": 1.3422,
289
+ "step": 32
290
+ },
291
+ {
292
+ "epoch": 0.06099815157116451,
293
+ "grad_norm": 3.6826868057250977,
294
+ "learning_rate": 0.00019891470916809362,
295
+ "loss": 1.3938,
296
+ "step": 33
297
+ },
298
+ {
299
+ "epoch": 0.06284658040665435,
300
+ "grad_norm": 2.7460691928863525,
301
+ "learning_rate": 0.00019881847477499557,
302
+ "loss": 0.8931,
303
+ "step": 34
304
+ },
305
+ {
306
+ "epoch": 0.06469500924214418,
307
+ "grad_norm": 4.063459873199463,
308
+ "learning_rate": 0.00019871817834144504,
309
+ "loss": 1.4139,
310
+ "step": 35
311
+ },
312
+ {
313
+ "epoch": 0.06469500924214418,
314
+ "eval_loss": 0.3002864420413971,
315
+ "eval_runtime": 173.4727,
316
+ "eval_samples_per_second": 1.314,
317
+ "eval_steps_per_second": 0.657,
318
+ "step": 35
319
+ },
320
+ {
321
+ "epoch": 0.066543438077634,
322
+ "grad_norm": 3.00803804397583,
323
+ "learning_rate": 0.0001986138239902355,
324
+ "loss": 0.9237,
325
+ "step": 36
326
+ },
327
+ {
328
+ "epoch": 0.06839186691312385,
329
+ "grad_norm": 3.4707858562469482,
330
+ "learning_rate": 0.0001985054160109657,
331
+ "loss": 1.6732,
332
+ "step": 37
333
+ },
334
+ {
335
+ "epoch": 0.07024029574861368,
336
+ "grad_norm": 4.129047393798828,
337
+ "learning_rate": 0.00019839295885986296,
338
+ "loss": 1.4183,
339
+ "step": 38
340
+ },
341
+ {
342
+ "epoch": 0.07208872458410351,
343
+ "grad_norm": 3.468799352645874,
344
+ "learning_rate": 0.0001982764571596004,
345
+ "loss": 1.1402,
346
+ "step": 39
347
+ },
348
+ {
349
+ "epoch": 0.07393715341959335,
350
+ "grad_norm": 2.87471079826355,
351
+ "learning_rate": 0.00019815591569910654,
352
+ "loss": 1.4329,
353
+ "step": 40
354
+ },
355
+ {
356
+ "epoch": 0.07393715341959335,
357
+ "eval_loss": 0.3075679838657379,
358
+ "eval_runtime": 173.655,
359
+ "eval_samples_per_second": 1.313,
360
+ "eval_steps_per_second": 0.656,
361
+ "step": 40
362
  }
363
  ],
364
  "logging_steps": 1,
 
373
  "early_stopping_threshold": 0.0
374
  },
375
  "attributes": {
376
+ "early_stopping_patience_counter": 1
377
  }
378
  },
379
  "TrainerControl": {
 
382
  "should_evaluate": false,
383
  "should_log": false,
384
  "should_save": true,
385
+ "should_training_stop": true
386
  },
387
  "attributes": {}
388
  }
389
  },
390
+ "total_flos": 5.658027780734976e+16,
391
  "train_batch_size": 2,
392
  "trial_name": null,
393
  "trial_params": null