givyboy commited on
Commit
fad1d68
1 Parent(s): b2f12f6

Training in progress, step 10

Browse files
adapter_config.json CHANGED
@@ -14,12 +14,10 @@
14
  "r": 32,
15
  "revision": null,
16
  "target_modules": [
17
- "q_proj",
18
- "k_proj",
19
- "v_proj",
20
  "dense",
21
- "fc1",
22
- "fc2"
23
  ],
24
  "task_type": "CAUSAL_LM"
25
  }
 
14
  "r": 32,
15
  "revision": null,
16
  "target_modules": [
17
+ "query_key_value",
 
 
18
  "dense",
19
+ "dense_h_to_4h",
20
+ "dense_4h_to_h"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f844dc64bdd37df1b779194f38d4e2278e8bcd6348eecdc74cfa9d4ea504b96
3
+ size 188881930
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:21459078538af91abf09b23675132ba1528303b58cbebf946982d36a309e9df3
3
- size 188794608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a02a3082928136c078cea9ce852282c03ac48dea0165b72e7240486e55509e3e
3
+ size 20980048
tokenizer.json CHANGED
@@ -1,11 +1,6 @@
1
  {
2
  "version": "1.0",
3
- "truncation": {
4
- "direction": "Right",
5
- "max_length": 1024,
6
- "strategy": "LongestFirst",
7
- "stride": 0
8
- },
9
  "padding": null,
10
  "added_tokens": [
11
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23ffcf3cfab6918b4e6ab67033e93f0e179391475b1936cf9a03ecee26044b0f
3
  size 4728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a13cce3d91a11e42c90de80e0b18754955a0ff72cf8be1834c3bd87f78bea489
3
  size 4728