rizkyjun commited on
Commit
2e117bc
1 Parent(s): 8ae8fec

Training in progress, step 2100, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6c59c9a73081018f9935c7827972a4c96c091e134bec87a67b615ed60a5f241c
3
  size 31466288
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4663053c60e5f267803a23dfc4bf6b7799e1bea84539190c45eeb0994148fd0
3
  size 31466288
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21103e6050d033638c5a313b15a95db6f7f0ee96b52eaab7515f511c5bdf2c1a
3
  size 62950917
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db2037ccc8f01df1eb9dd3cb62165d668a40294b816104d67ba12a3f0bcb38fa
3
  size 62950917
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da5eddeaa757c6d016160549ec4b158dd730ee5eb61636139ede6dfade684bcc
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f733f68e9af3a413f7f5c8bf128e0ef057938927e43869c070a32516f9bd1d62
3
  size 14575
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b2d4b4cadcabd3e17adcc39c43d8312038a1f5c7f53125f288ba4de5ed25396b
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1238b3bfa75e49a19396161e9e7b72ab89cdd1a3f63b51c0ab4d6e8d216c5a5
3
  size 627
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 1.5527715682983398,
3
- "best_model_checkpoint": "./outputs/checkpoint-1900",
4
- "epoch": 1.384335154826958,
5
  "eval_steps": 100,
6
- "global_step": 1900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -273,13 +273,41 @@
273
  "eval_samples_per_second": 11.278,
274
  "eval_steps_per_second": 1.411,
275
  "step": 1900
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
276
  }
277
  ],
278
  "logging_steps": 100,
279
  "max_steps": 4116,
280
  "num_train_epochs": 3,
281
  "save_steps": 100,
282
- "total_flos": 4.9544045126846054e+17,
283
  "trial_name": null,
284
  "trial_params": null
285
  }
 
1
  {
2
+ "best_metric": 1.5026806592941284,
3
+ "best_model_checkpoint": "./outputs/checkpoint-2100",
4
+ "epoch": 1.530054644808743,
5
  "eval_steps": 100,
6
+ "global_step": 2100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
273
  "eval_samples_per_second": 11.278,
274
  "eval_steps_per_second": 1.411,
275
  "step": 1900
276
+ },
277
+ {
278
+ "epoch": 1.46,
279
+ "learning_rate": 0.0002,
280
+ "loss": 1.5057,
281
+ "step": 2000
282
+ },
283
+ {
284
+ "epoch": 1.46,
285
+ "eval_loss": 1.5266777276992798,
286
+ "eval_runtime": 556.4442,
287
+ "eval_samples_per_second": 11.275,
288
+ "eval_steps_per_second": 1.411,
289
+ "step": 2000
290
+ },
291
+ {
292
+ "epoch": 1.53,
293
+ "learning_rate": 0.0002,
294
+ "loss": 1.4851,
295
+ "step": 2100
296
+ },
297
+ {
298
+ "epoch": 1.53,
299
+ "eval_loss": 1.5026806592941284,
300
+ "eval_runtime": 556.1829,
301
+ "eval_samples_per_second": 11.28,
302
+ "eval_steps_per_second": 1.411,
303
+ "step": 2100
304
  }
305
  ],
306
  "logging_steps": 100,
307
  "max_steps": 4116,
308
  "num_train_epochs": 3,
309
  "save_steps": 100,
310
+ "total_flos": 5.47414668374655e+17,
311
  "trial_name": null,
312
  "trial_params": null
313
  }