BlitherBoom commited on
Commit
94f4d49
·
1 Parent(s): 86be6b0

Training in progress, step 10000

Browse files
adapter_config.json CHANGED
@@ -16,8 +16,8 @@
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
- "v_proj",
20
- "q_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
 
16
  "rank_pattern": {},
17
  "revision": null,
18
  "target_modules": [
19
+ "q_proj",
20
+ "v_proj"
21
  ],
22
  "task_type": "CAUSAL_LM"
23
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:804c1ff50ea7d0653709320c185a2e53ef78b825228ef8dff265fb16ccf4e0c0
3
  size 27280152
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a4ccab09717884655be6350e064e226ca1090458744546937a5c3331144f11a
3
  size 27280152
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cf54746626c99fede497f222957e68a6cf384031c1367bf6aaa200987f31fcbc
3
  size 4664
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae57feb87edd5ff405c3f99d817f57b96390629bef0ab11a291c6eaabcf37f9e
3
  size 4664