nblinh commited on
Commit
29751ba
·
verified ·
1 Parent(s): 8cd6bd5

End of training

Browse files
README.md CHANGED
@@ -78,7 +78,7 @@ resume_from_checkpoint: null
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
- sequence_len: 1024
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
@@ -103,7 +103,7 @@ xformers_attention: true
103
 
104
  This model is a fine-tuned version of [Qwen/Qwen1.5-0.5B-Chat](https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
- - Loss: 0.6344
107
 
108
  ## Model description
109
 
@@ -135,7 +135,7 @@ The following hyperparameters were used during training:
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
- | 0.5982 | 0.0001 | 10 | 0.6344 |
139
 
140
 
141
  ### Framework versions
 
78
  s2_attention: null
79
  sample_packing: false
80
  saves_per_epoch: 1
81
+ sequence_len: 2048
82
  strict: false
83
  tf32: false
84
  tokenizer_type: AutoTokenizer
 
103
 
104
  This model is a fine-tuned version of [Qwen/Qwen1.5-0.5B-Chat](https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat) on the None dataset.
105
  It achieves the following results on the evaluation set:
106
+ - Loss: 0.3129
107
 
108
  ## Model description
109
 
 
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
+ | 0.6412 | 0.0001 | 10 | 0.3129 |
139
 
140
 
141
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
 
24
  "o_proj",
25
- "up_proj",
26
  "v_proj",
27
- "k_proj",
28
  "down_proj",
29
- "gate_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "gate_proj",
24
+ "k_proj",
25
  "o_proj",
 
26
  "v_proj",
27
+ "q_proj",
28
  "down_proj",
29
+ "up_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fbcb485708cf73bd3e8c2b73acc10f131747be9e1a8af3c42a342619eb9f871
3
  size 30398410
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90c22ccae38b8f2676cbe082cb2ff6b0ed7cf5001e4a4a9320a7698f41b3b763
3
  size 30398410
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cc4ec68431f55df9b6dc0e1694da8d6d7390eca202d7e0dba6ef380aa6f2589a
3
  size 30322120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3fb949e5a5abcac622e882d40aef4a8fedf6e36fa56178b25f714c34872949ee
3
  size 30322120
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1de78312fc7b8bf2bd311bd2109386cd6196800e3c13b07c9dce2a163690857b
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6a31c5aa7aace24fede26d1803f3475f502732524a502a70e1277a113c8ffc3
3
  size 6776