joe611 commited on
Commit
6447e7c
1 Parent(s): f12d8a5

Training in progress, epoch 111, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0ad2b988d8f0404432c148548cf4bef2c3e459fc5cb3d4b25049082154418178
3
  size 166496880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:292c73693098875692abcb4fe84b6b84cc28407036bb27a962300192d5d9f048
3
  size 166496880
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bfe570c8a4fa1b73bfe1c715f7f40db03cc6b86d7cb2e76ca103416b45d58494
3
  size 330495866
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c70174351bbc61e80037e803568d9a6f4338016d31c52d7c41f2228630dfa92a
3
  size 330495866
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c0f7c3a56d7769670ed1811eae3d35cdb7046cabac5d4616921ff952a0230d6c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c387bfd9a993f978958650fcc5a25ab904bd56418ed7491f4791553eb1382ae
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3ce54cfb1e1954af7d5a410661038938f3b5da2d5bc8c7db9c67e10ccc637ab7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91165b4467d36cadf91977e383e5a88c4e25b217f011b6ddd68eb9dee41f87cb
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.2962397336959839,
3
  "best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-45000",
4
- "epoch": 110.0,
5
  "eval_steps": 500,
6
- "global_step": 55000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -15698,6 +15698,151 @@
15698
  "eval_samples_per_second": 7.28,
15699
  "eval_steps_per_second": 0.946,
15700
  "step": 55000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15701
  }
15702
  ],
15703
  "logging_steps": 30,
@@ -15717,7 +15862,7 @@
15717
  "attributes": {}
15718
  }
15719
  },
15720
- "total_flos": 1.8921271209984e+19,
15721
  "train_batch_size": 2,
15722
  "trial_name": null,
15723
  "trial_params": null
 
1
  {
2
  "best_metric": 0.2962397336959839,
3
  "best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-45000",
4
+ "epoch": 111.0,
5
  "eval_steps": 500,
6
+ "global_step": 55500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
15698
  "eval_samples_per_second": 7.28,
15699
  "eval_steps_per_second": 0.946,
15700
  "step": 55000
15701
+ },
15702
+ {
15703
+ "epoch": 110.04,
15704
+ "grad_norm": 35.87388610839844,
15705
+ "learning_rate": 1.651235264858177e-06,
15706
+ "loss": 0.2443,
15707
+ "step": 55020
15708
+ },
15709
+ {
15710
+ "epoch": 110.1,
15711
+ "grad_norm": 47.754180908203125,
15712
+ "learning_rate": 1.6465721173164e-06,
15713
+ "loss": 0.2129,
15714
+ "step": 55050
15715
+ },
15716
+ {
15717
+ "epoch": 110.16,
15718
+ "grad_norm": 26.22478675842285,
15719
+ "learning_rate": 1.6419142652949793e-06,
15720
+ "loss": 0.1971,
15721
+ "step": 55080
15722
+ },
15723
+ {
15724
+ "epoch": 110.22,
15725
+ "grad_norm": 51.385337829589844,
15726
+ "learning_rate": 1.6372617161493014e-06,
15727
+ "loss": 0.2324,
15728
+ "step": 55110
15729
+ },
15730
+ {
15731
+ "epoch": 110.28,
15732
+ "grad_norm": 54.496002197265625,
15733
+ "learning_rate": 1.6326144772263752e-06,
15734
+ "loss": 0.2315,
15735
+ "step": 55140
15736
+ },
15737
+ {
15738
+ "epoch": 110.34,
15739
+ "grad_norm": 73.95658111572266,
15740
+ "learning_rate": 1.627972555864824e-06,
15741
+ "loss": 0.2268,
15742
+ "step": 55170
15743
+ },
15744
+ {
15745
+ "epoch": 110.4,
15746
+ "grad_norm": 56.15229415893555,
15747
+ "learning_rate": 1.6233359593948777e-06,
15748
+ "loss": 0.1911,
15749
+ "step": 55200
15750
+ },
15751
+ {
15752
+ "epoch": 110.46,
15753
+ "grad_norm": 26.1507511138916,
15754
+ "learning_rate": 1.618704695138353e-06,
15755
+ "loss": 0.2455,
15756
+ "step": 55230
15757
+ },
15758
+ {
15759
+ "epoch": 110.52,
15760
+ "grad_norm": 73.8873291015625,
15761
+ "learning_rate": 1.6140787704086502e-06,
15762
+ "loss": 0.1911,
15763
+ "step": 55260
15764
+ },
15765
+ {
15766
+ "epoch": 110.58,
15767
+ "grad_norm": 101.59705352783203,
15768
+ "learning_rate": 1.6094581925107356e-06,
15769
+ "loss": 0.2273,
15770
+ "step": 55290
15771
+ },
15772
+ {
15773
+ "epoch": 110.64,
15774
+ "grad_norm": 31.966201782226562,
15775
+ "learning_rate": 1.6048429687411294e-06,
15776
+ "loss": 0.2295,
15777
+ "step": 55320
15778
+ },
15779
+ {
15780
+ "epoch": 110.7,
15781
+ "grad_norm": 60.9480094909668,
15782
+ "learning_rate": 1.600233106387904e-06,
15783
+ "loss": 0.2284,
15784
+ "step": 55350
15785
+ },
15786
+ {
15787
+ "epoch": 110.76,
15788
+ "grad_norm": 65.5320816040039,
15789
+ "learning_rate": 1.5956286127306591e-06,
15790
+ "loss": 0.2274,
15791
+ "step": 55380
15792
+ },
15793
+ {
15794
+ "epoch": 110.82,
15795
+ "grad_norm": 67.60034942626953,
15796
+ "learning_rate": 1.591029495040518e-06,
15797
+ "loss": 0.2207,
15798
+ "step": 55410
15799
+ },
15800
+ {
15801
+ "epoch": 110.88,
15802
+ "grad_norm": 134.33175659179688,
15803
+ "learning_rate": 1.586435760580118e-06,
15804
+ "loss": 0.2187,
15805
+ "step": 55440
15806
+ },
15807
+ {
15808
+ "epoch": 110.94,
15809
+ "grad_norm": 66.95919799804688,
15810
+ "learning_rate": 1.5818474166035907e-06,
15811
+ "loss": 0.2397,
15812
+ "step": 55470
15813
+ },
15814
+ {
15815
+ "epoch": 111.0,
15816
+ "grad_norm": 59.430965423583984,
15817
+ "learning_rate": 1.5772644703565564e-06,
15818
+ "loss": 0.2212,
15819
+ "step": 55500
15820
+ },
15821
+ {
15822
+ "epoch": 111.0,
15823
+ "eval_loss": 0.31890785694122314,
15824
+ "eval_map": 0.8013,
15825
+ "eval_map_50": 0.9453,
15826
+ "eval_map_75": 0.8992,
15827
+ "eval_map_chicken": 0.8086,
15828
+ "eval_map_duck": 0.7247,
15829
+ "eval_map_large": 0.7895,
15830
+ "eval_map_medium": 0.8062,
15831
+ "eval_map_plant": 0.8707,
15832
+ "eval_map_small": 0.3336,
15833
+ "eval_mar_1": 0.3229,
15834
+ "eval_mar_10": 0.8357,
15835
+ "eval_mar_100": 0.8399,
15836
+ "eval_mar_100_chicken": 0.8488,
15837
+ "eval_mar_100_duck": 0.7742,
15838
+ "eval_mar_100_plant": 0.8967,
15839
+ "eval_mar_large": 0.8245,
15840
+ "eval_mar_medium": 0.8465,
15841
+ "eval_mar_small": 0.3881,
15842
+ "eval_runtime": 11.8276,
15843
+ "eval_samples_per_second": 8.455,
15844
+ "eval_steps_per_second": 1.099,
15845
+ "step": 55500
15846
  }
15847
  ],
15848
  "logging_steps": 30,
 
15862
  "attributes": {}
15863
  }
15864
  },
15865
+ "total_flos": 1.90932827664384e+19,
15866
  "train_batch_size": 2,
15867
  "trial_name": null,
15868
  "trial_params": null