Hanzalwi commited on
Commit
990f638
1 Parent(s): caade6b

Training in progress, step 100, checkpoint

Browse files
last-checkpoint/README.md CHANGED
@@ -236,4 +236,23 @@ The following `bitsandbytes` quantization config was used during training:
236
  ### Framework versions
237
 
238
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
239
  - PEFT 0.6.3.dev0
 
236
  ### Framework versions
237
 
238
 
239
+ - PEFT 0.6.3.dev0
240
+ ## Training procedure
241
+
242
+
243
+ The following `bitsandbytes` quantization config was used during training:
244
+ - quant_method: bitsandbytes
245
+ - load_in_8bit: True
246
+ - load_in_4bit: False
247
+ - llm_int8_threshold: 6.0
248
+ - llm_int8_skip_modules: None
249
+ - llm_int8_enable_fp32_cpu_offload: False
250
+ - llm_int8_has_fp16_weight: False
251
+ - bnb_4bit_quant_type: fp4
252
+ - bnb_4bit_use_double_quant: False
253
+ - bnb_4bit_compute_dtype: float32
254
+
255
+ ### Framework versions
256
+
257
+
258
  - PEFT 0.6.3.dev0
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef2c8347f1be9aebfc1271f5f427065d90f0e50efb3a58333837af0c92cfabd6
3
  size 9444296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a48e0e5bde33bb82c52da7473d5f0b81aa3ad4d16f1b412eac2591e9ab1a27e1
3
  size 9444296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2323219fe41e5b63b7d472b1ca12c211f60c05cc1bc4517c5ee44eb63d4d78ba
3
  size 18902665
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85ad78154fee72f2aca6fca0e93077aac0cc3d7717cd9c9ea57c8c724ce6383b
3
  size 18902665
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bbebb324d09ed3b79588ce704d8e867cae0d7b652313decf15a03bfb1ba6792e
3
  size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43b3279df66a37918fc86e90e9c6a8fa486ff8c675a0a5adbdb77a496e973bf1
3
  size 14575
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5efc416a6883409dd7ab6f5c779e107c7c2baa7af6e12ed9fbd9dd73b8b20784
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:216f76b8039f833c337db298c81f13b12082d5fd4f9d866cecd34b2ca7550b37
3
  size 627
last-checkpoint/tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:490b3432398e9c45dda0581bf34f6daa9411bda59cd0cd831670d7c86f90d2dd
3
- size 14500570
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17a208233d2ee8d8c83b23bc214df737c44806a1919f444e89b31e586cd956ba
3
+ size 14500471
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 2.005772113800049,
3
- "best_model_checkpoint": "./outputs/checkpoint-500",
4
- "epoch": 0.7102272727272727,
5
  "eval_steps": 100,
6
- "global_step": 500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -11,72 +11,16 @@
11
  {
12
  "epoch": 0.14,
13
  "learning_rate": 0.0002,
14
- "loss": 2.4473,
15
  "step": 100
16
  },
17
  {
18
  "epoch": 0.14,
19
- "eval_loss": 2.1156835556030273,
20
- "eval_runtime": 56.5007,
21
- "eval_samples_per_second": 26.407,
22
- "eval_steps_per_second": 3.31,
23
  "step": 100
24
- },
25
- {
26
- "epoch": 0.28,
27
- "learning_rate": 0.0002,
28
- "loss": 2.253,
29
- "step": 200
30
- },
31
- {
32
- "epoch": 0.28,
33
- "eval_loss": 2.073469877243042,
34
- "eval_runtime": 54.097,
35
- "eval_samples_per_second": 27.58,
36
- "eval_steps_per_second": 3.457,
37
- "step": 200
38
- },
39
- {
40
- "epoch": 0.43,
41
- "learning_rate": 0.0002,
42
- "loss": 2.2236,
43
- "step": 300
44
- },
45
- {
46
- "epoch": 0.43,
47
- "eval_loss": 2.046644687652588,
48
- "eval_runtime": 53.9984,
49
- "eval_samples_per_second": 27.63,
50
- "eval_steps_per_second": 3.463,
51
- "step": 300
52
- },
53
- {
54
- "epoch": 0.57,
55
- "learning_rate": 0.0002,
56
- "loss": 2.1911,
57
- "step": 400
58
- },
59
- {
60
- "epoch": 0.57,
61
- "eval_loss": 2.0298001766204834,
62
- "eval_runtime": 54.0695,
63
- "eval_samples_per_second": 27.594,
64
- "eval_steps_per_second": 3.459,
65
- "step": 400
66
- },
67
- {
68
- "epoch": 0.71,
69
- "learning_rate": 0.0002,
70
- "loss": 2.1704,
71
- "step": 500
72
- },
73
- {
74
- "epoch": 0.71,
75
- "eval_loss": 2.005772113800049,
76
- "eval_runtime": 54.1197,
77
- "eval_samples_per_second": 27.569,
78
- "eval_steps_per_second": 3.455,
79
- "step": 500
80
  }
81
  ],
82
  "logging_steps": 100,
@@ -84,7 +28,7 @@
84
  "num_input_tokens_seen": 0,
85
  "num_train_epochs": 3,
86
  "save_steps": 100,
87
- "total_flos": 1.52069905096704e+16,
88
  "trial_name": null,
89
  "trial_params": null
90
  }
 
1
  {
2
+ "best_metric": 2.11466121673584,
3
+ "best_model_checkpoint": "./outputs/checkpoint-100",
4
+ "epoch": 0.14204545454545456,
5
  "eval_steps": 100,
6
+ "global_step": 100,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
11
  {
12
  "epoch": 0.14,
13
  "learning_rate": 0.0002,
14
+ "loss": 2.4493,
15
  "step": 100
16
  },
17
  {
18
  "epoch": 0.14,
19
+ "eval_loss": 2.11466121673584,
20
+ "eval_runtime": 56.3914,
21
+ "eval_samples_per_second": 26.458,
22
+ "eval_steps_per_second": 3.316,
23
  "step": 100
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  }
25
  ],
26
  "logging_steps": 100,
 
28
  "num_input_tokens_seen": 0,
29
  "num_train_epochs": 3,
30
  "save_steps": 100,
31
+ "total_flos": 3017571420487680.0,
32
  "trial_name": null,
33
  "trial_params": null
34
  }
last-checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f41f795a4d38cf0b20c4ab7c8b00a3bbdbb028ecd4a46fa726fb31c2bc81f568
3
  size 4283
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:127332f99b100bbb3a9659e58e9fcfa7f341ada72cf3863477eab4b88b505759
3
  size 4283