huseinzol05 commited on
Commit
9c07e27
·
verified ·
1 Parent(s): cd49ec2

Upload model

Browse files
Files changed (3) hide show
  1. README.md +2 -8
  2. adapter_config.json +5 -5
  3. adapter_model.safetensors +1 -1
README.md CHANGED
@@ -1,12 +1,6 @@
1
  ---
2
- base_model: unsloth/Llama-3.2-3B-Instruct
3
  library_name: transformers
4
- language:
5
- - en
6
- - ms
7
- - id
8
- - ta
9
- - zh
10
  ---
11
 
12
  # Llama-3.2-3B-Malaysian-Reasoning LoRA
@@ -17,6 +11,6 @@ Full README at [mesolitica/Llama-3.2-3B-Malaysian-Reasoning](https://huggingface
17
 
18
  ## Merging
19
 
20
- Because Llama 3.2 3B is using tied weight embedding, so merging required to clone embedding into lm head after that `addmm` as usual, script at https://github.com/mesolitica/malaya/blob/master/session/small-malaysian-reasoning/merge-3b.ipynb
21
 
22
  If the model is not a tied weight, you can use default `merge_and_unload` function from PEFT or unsloth.
 
1
  ---
 
2
  library_name: transformers
3
+ tags: []
 
 
 
 
 
4
  ---
5
 
6
  # Llama-3.2-3B-Malaysian-Reasoning LoRA
 
11
 
12
  ## Merging
13
 
14
+ Because Llama 3.2 3B is using tied weight embedding, so merging required to clone embedding into lm head, script at https://github.com/mesolitica/malaya/blob/master/session/small-malaysian-reasoning/merge-3b.ipynb
15
 
16
  If the model is not a tied weight, you can use default `merge_and_unload` function from PEFT or unsloth.
adapter_config.json CHANGED
@@ -23,14 +23,14 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "gate_proj",
27
- "o_proj",
28
  "down_proj",
29
- "embed_tokens",
30
- "lm_head",
31
- "v_proj",
32
  "k_proj",
 
 
33
  "up_proj",
 
 
 
34
  "q_proj"
35
  ],
36
  "task_type": "CAUSAL_LM",
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
 
 
26
  "down_proj",
 
 
 
27
  "k_proj",
28
+ "gate_proj",
29
+ "embed_tokens",
30
  "up_proj",
31
+ "v_proj",
32
+ "lm_head",
33
+ "o_proj",
34
  "q_proj"
35
  ],
36
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:116678e3ee5899352cfcb86b2a8ef828a19bf36eaa601e2149555f2d43abc906
3
  size 3401110784
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b08bdc1217aed1d16721c796e997a4336aa2ce5c10b90ccd15bb6e1f88d117aa
3
  size 3401110784