besimray commited on
Commit
fa5849f
·
verified ·
1 Parent(s): 201787d

Training in progress, step 210, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7ce726842d250f20c86c371d9e87a8ce9846060e663f40a17aa5220f549de934
3
  size 125048
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6e8b163ef36cc7e57a8ee15a6ebe44a66dd7b467f2cc54f8de9f11808fbf4a74
3
  size 125048
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:89d5bd83a26023b846c84f45639c04652bc00657132d2cf830dd829784d28815
3
  size 162868
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e11899b82d3568219c123451efb8f9c3baa4ecaaf187d3d8094ff509232f7bf
3
  size 162868
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8ebd8401c143f16f72b022f70333a3cfcf5c180eb2884806e8d3b16f8611859c
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:578c17b2105d3c6bea4883770e5007551a3f2302d9968187a7d2ba2aece88a9c
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:53ada7ec389029bea855f553cfcb8ba6729038e26afcd278435a9c0b241e2783
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99f3078781d07b22b8aceece06e9b1a7abc294ae09750b1339d4cf73fa58b867
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 11.022467613220215,
3
- "best_model_checkpoint": "miner_id_24/checkpoint-200",
4
- "epoch": 0.009039343743644212,
5
  "eval_steps": 5,
6
- "global_step": 200,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1735,6 +1735,92 @@
1735
  "eval_samples_per_second": 52.886,
1736
  "eval_steps_per_second": 26.446,
1737
  "step": 200
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1738
  }
1739
  ],
1740
  "logging_steps": 1,
@@ -1763,7 +1849,7 @@
1763
  "attributes": {}
1764
  }
1765
  },
1766
- "total_flos": 2101346304000.0,
1767
  "train_batch_size": 2,
1768
  "trial_name": null,
1769
  "trial_params": null
 
1
  {
2
+ "best_metric": 11.02186393737793,
3
+ "best_model_checkpoint": "miner_id_24/checkpoint-210",
4
+ "epoch": 0.009491310930826422,
5
  "eval_steps": 5,
6
+ "global_step": 210,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1735
  "eval_samples_per_second": 52.886,
1736
  "eval_steps_per_second": 26.446,
1737
  "step": 200
1738
+ },
1739
+ {
1740
+ "epoch": 0.009084540462362432,
1741
+ "grad_norm": 0.6612194776535034,
1742
+ "learning_rate": 0.00013393410881565876,
1743
+ "loss": 44.1568,
1744
+ "step": 201
1745
+ },
1746
+ {
1747
+ "epoch": 0.009129737181080653,
1748
+ "grad_norm": 0.5365848541259766,
1749
+ "learning_rate": 0.00013333031729088419,
1750
+ "loss": 44.0318,
1751
+ "step": 202
1752
+ },
1753
+ {
1754
+ "epoch": 0.009174933899798874,
1755
+ "grad_norm": 0.43999558687210083,
1756
+ "learning_rate": 0.0001327251556873117,
1757
+ "loss": 44.0544,
1758
+ "step": 203
1759
+ },
1760
+ {
1761
+ "epoch": 0.009220130618517095,
1762
+ "grad_norm": 0.5535528659820557,
1763
+ "learning_rate": 0.00013211864888076457,
1764
+ "loss": 44.0657,
1765
+ "step": 204
1766
+ },
1767
+ {
1768
+ "epoch": 0.009265327337235316,
1769
+ "grad_norm": 0.5289484262466431,
1770
+ "learning_rate": 0.0001315108218023621,
1771
+ "loss": 44.0946,
1772
+ "step": 205
1773
+ },
1774
+ {
1775
+ "epoch": 0.009265327337235316,
1776
+ "eval_loss": 11.022246360778809,
1777
+ "eval_runtime": 175.5631,
1778
+ "eval_samples_per_second": 53.069,
1779
+ "eval_steps_per_second": 26.537,
1780
+ "step": 205
1781
+ },
1782
+ {
1783
+ "epoch": 0.009310524055953538,
1784
+ "grad_norm": 0.515040934085846,
1785
+ "learning_rate": 0.00013090169943749476,
1786
+ "loss": 44.1026,
1787
+ "step": 206
1788
+ },
1789
+ {
1790
+ "epoch": 0.009355720774671759,
1791
+ "grad_norm": 0.43807700276374817,
1792
+ "learning_rate": 0.00013029130682479722,
1793
+ "loss": 44.0529,
1794
+ "step": 207
1795
+ },
1796
+ {
1797
+ "epoch": 0.00940091749338998,
1798
+ "grad_norm": 0.40383437275886536,
1799
+ "learning_rate": 0.00012967966905511906,
1800
+ "loss": 44.0854,
1801
+ "step": 208
1802
+ },
1803
+ {
1804
+ "epoch": 0.009446114212108201,
1805
+ "grad_norm": 0.42450079321861267,
1806
+ "learning_rate": 0.00012906681127049338,
1807
+ "loss": 44.0488,
1808
+ "step": 209
1809
+ },
1810
+ {
1811
+ "epoch": 0.009491310930826422,
1812
+ "grad_norm": 0.5043962597846985,
1813
+ "learning_rate": 0.00012845275866310324,
1814
+ "loss": 44.047,
1815
+ "step": 210
1816
+ },
1817
+ {
1818
+ "epoch": 0.009491310930826422,
1819
+ "eval_loss": 11.02186393737793,
1820
+ "eval_runtime": 176.4502,
1821
+ "eval_samples_per_second": 52.802,
1822
+ "eval_steps_per_second": 26.404,
1823
+ "step": 210
1824
  }
1825
  ],
1826
  "logging_steps": 1,
 
1849
  "attributes": {}
1850
  }
1851
  },
1852
+ "total_flos": 2206413619200.0,
1853
  "train_batch_size": 2,
1854
  "trial_name": null,
1855
  "trial_params": null