emdemor commited on
Commit
f98de7e
1 Parent(s): e8aa5f0

Training in progress, step 50

Browse files
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "o_proj",
25
- "v_proj",
26
- "down_proj",
27
- "k_proj",
28
  "gate_proj",
29
- "up_proj"
 
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
 
 
23
  "gate_proj",
24
+ "k_proj",
25
+ "up_proj",
26
+ "down_proj",
27
+ "v_proj",
28
+ "q_proj",
29
+ "o_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f7cecf1438a0318c70037968c17dd00367784b98780908cb83a5baf1c4377e7d
3
  size 35668592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68ea8ad8be0b23821d0ba7b1377dc7d1787101493604a4b60e1bf54b56e4d915
3
  size 35668592
metrics.json CHANGED
@@ -1,3 +1 @@
1
- {"Step":50,"eval_loss":0.9085000753,"eval_runtime":68.9775,"eval_samples_per_second":1.45,"eval_steps_per_second":0.188,"epoch":0.08}
2
- {"Step":100,"eval_loss":0.7916119099,"eval_runtime":68.7621,"eval_samples_per_second":1.454,"eval_steps_per_second":0.189,"epoch":0.16}
3
- {"Step":150,"eval_loss":0.7562382221,"eval_runtime":68.6488,"eval_samples_per_second":1.457,"eval_steps_per_second":0.189,"epoch":0.24}
 
1
+ {"Step":50,"eval_loss":0.9260319471,"eval_runtime":60.9012,"eval_samples_per_second":1.642,"eval_steps_per_second":0.213,"epoch":0.0967117988}
 
 
state.json CHANGED
@@ -1,46 +1,15 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.24,
5
  "eval_steps": 50,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
- "log_history": [
11
- {
12
- "epoch": 0.08,
13
- "grad_norm": 0.03769562020897865,
14
- "learning_rate": 0.00013031914893617022,
15
- "loss": 1.0215,
16
- "step": 50
17
- },
18
- {
19
- "epoch": 0.08,
20
- "eval_loss": 0.908500075340271,
21
- "eval_runtime": 68.9775,
22
- "eval_samples_per_second": 1.45,
23
- "eval_steps_per_second": 0.188,
24
- "step": 50
25
- },
26
- {
27
- "epoch": 0.16,
28
- "grad_norm": 0.028151430189609528,
29
- "learning_rate": 0.0002632978723404255,
30
- "loss": 0.8497,
31
- "step": 100
32
- },
33
- {
34
- "epoch": 0.16,
35
- "eval_loss": 0.791611909866333,
36
- "eval_runtime": 68.7621,
37
- "eval_samples_per_second": 1.454,
38
- "eval_steps_per_second": 0.189,
39
- "step": 100
40
- }
41
- ],
42
  "logging_steps": 50,
43
- "max_steps": 1875,
44
  "num_input_tokens_seen": 0,
45
  "num_train_epochs": 3,
46
  "save_steps": 50,
@@ -50,13 +19,13 @@
50
  "should_epoch_stop": false,
51
  "should_evaluate": false,
52
  "should_log": false,
53
- "should_save": true,
54
  "should_training_stop": false
55
  },
56
  "attributes": {}
57
  }
58
  },
59
- "total_flos": 1.474428848529408e+16,
60
  "train_batch_size": 8,
61
  "trial_name": null,
62
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.09671179883945841,
5
  "eval_steps": 50,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
+ "log_history": [],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  "logging_steps": 50,
12
+ "max_steps": 1551,
13
  "num_input_tokens_seen": 0,
14
  "num_train_epochs": 3,
15
  "save_steps": 50,
 
19
  "should_epoch_stop": false,
20
  "should_evaluate": false,
21
  "should_log": false,
22
+ "should_save": false,
23
  "should_training_stop": false
24
  },
25
  "attributes": {}
26
  }
27
  },
28
+ "total_flos": 0,
29
  "train_batch_size": 8,
30
  "trial_name": null,
31
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:00989b9f531236a6c251afdcbb9c72bf447e3c288ce640ca158662681219037a
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb123c014e107e989a46cae0ffd2cd1b676fa12aa284894769198e87fcfa9e71
3
  size 5432