Nexspear commited on
Commit
c1e2880
·
verified ·
1 Parent(s): f8a3a3d

Training in progress, step 238, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:410b91c48e370417f87515a2eaae1a7f535d115bdfa83c50e64334080b1f0b62
3
  size 161533192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5f938f590c16a2da7e7003b5726decdca766113cb8c6bf1132392eec1032131
3
  size 161533192
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:912ad228eb3b6d149e12e4ade24a7080c1aa79a5940fd0cf0ff1e0684af11de3
3
  size 82460660
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:242a26a061ea5d980a98bce60edfbf0494218ef54aba0d781b0dca84f1a1c629
3
  size 82460660
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1bbb06d4550f69a6f5bcf553617c3e68e7fa013d474d5f9b6a819f651cff51fa
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:771fda137b95b0a53343c99caba02f0ddc4fde33ba4274f7584011ab1e463479
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5f9839d107756d9c8815de9164f2ebf92c05b3536704a349ca5892084df7663e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbcef9424696e41c7961bd91f0570d39d59ef33af28ed19a0eb9e4f50ed1b09a
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.06877950101146325,
5
  "eval_steps": 34,
6
- "global_step": 204,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -539,6 +539,91 @@
539
  "eval_samples_per_second": 14.615,
540
  "eval_steps_per_second": 1.829,
541
  "step": 204
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
542
  }
543
  ],
544
  "logging_steps": 3,
@@ -558,7 +643,7 @@
558
  "attributes": {}
559
  }
560
  },
561
- "total_flos": 2.85207786877354e+17,
562
  "train_batch_size": 8,
563
  "trial_name": null,
564
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.08024275118004046,
5
  "eval_steps": 34,
6
+ "global_step": 238,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
539
  "eval_samples_per_second": 14.615,
540
  "eval_steps_per_second": 1.829,
541
  "step": 204
542
+ },
543
+ {
544
+ "epoch": 0.06979096426163182,
545
+ "grad_norm": 1.039337396621704,
546
+ "learning_rate": 2.459724913431772e-05,
547
+ "loss": 0.439,
548
+ "step": 207
549
+ },
550
+ {
551
+ "epoch": 0.07080242751180041,
552
+ "grad_norm": 1.222737431526184,
553
+ "learning_rate": 2.399335149726463e-05,
554
+ "loss": 0.4838,
555
+ "step": 210
556
+ },
557
+ {
558
+ "epoch": 0.07181389076196898,
559
+ "grad_norm": 1.278668999671936,
560
+ "learning_rate": 2.3390041714589514e-05,
561
+ "loss": 0.4612,
562
+ "step": 213
563
+ },
564
+ {
565
+ "epoch": 0.07282535401213756,
566
+ "grad_norm": 1.1525593996047974,
567
+ "learning_rate": 2.2787672102216042e-05,
568
+ "loss": 0.4372,
569
+ "step": 216
570
+ },
571
+ {
572
+ "epoch": 0.07383681726230613,
573
+ "grad_norm": 1.3022117614746094,
574
+ "learning_rate": 2.2186594427034864e-05,
575
+ "loss": 0.4593,
576
+ "step": 219
577
+ },
578
+ {
579
+ "epoch": 0.07484828051247472,
580
+ "grad_norm": 1.4199026823043823,
581
+ "learning_rate": 2.1587159701481716e-05,
582
+ "loss": 0.455,
583
+ "step": 222
584
+ },
585
+ {
586
+ "epoch": 0.07585974376264329,
587
+ "grad_norm": 1.3410009145736694,
588
+ "learning_rate": 2.098971797855599e-05,
589
+ "loss": 0.6084,
590
+ "step": 225
591
+ },
592
+ {
593
+ "epoch": 0.07687120701281187,
594
+ "grad_norm": 1.2653465270996094,
595
+ "learning_rate": 2.0394618147399713e-05,
596
+ "loss": 0.497,
597
+ "step": 228
598
+ },
599
+ {
600
+ "epoch": 0.07788267026298044,
601
+ "grad_norm": 1.2599753141403198,
602
+ "learning_rate": 1.980220772955602e-05,
603
+ "loss": 0.4794,
604
+ "step": 231
605
+ },
606
+ {
607
+ "epoch": 0.07889413351314903,
608
+ "grad_norm": 1.176132321357727,
609
+ "learning_rate": 1.921283267602643e-05,
610
+ "loss": 0.4134,
611
+ "step": 234
612
+ },
613
+ {
614
+ "epoch": 0.0799055967633176,
615
+ "grad_norm": 1.2982177734375,
616
+ "learning_rate": 1.8626837165245165e-05,
617
+ "loss": 0.4404,
618
+ "step": 237
619
+ },
620
+ {
621
+ "epoch": 0.08024275118004046,
622
+ "eval_loss": 0.4182414412498474,
623
+ "eval_runtime": 341.5144,
624
+ "eval_samples_per_second": 14.626,
625
+ "eval_steps_per_second": 1.83,
626
+ "step": 238
627
  }
628
  ],
629
  "logging_steps": 3,
 
643
  "attributes": {}
644
  }
645
  },
646
+ "total_flos": 3.3274241802357965e+17,
647
  "train_batch_size": 8,
648
  "trial_name": null,
649
  "trial_params": null