Upload 8 files
Browse files- .gitattributes +6 -0
- adapter.gguf +3 -0
- adapter_config.json +32 -0
- adapter_model.safetensors +3 -0
- cra-v1-guided-7b-F16.gguf +3 -0
- cra-v1-guided-7b-Q4_K_M.gguf +3 -0
- cra-v1-guided-7b-Q5_K.gguf +3 -0
- cra-v1-guided-7b-Q6_K.gguf +3 -0
- cra-v1-guided-7b-Q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
adapter.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
cra-v1-guided-7b-F16.gguf filter=lfs diff=lfs merge=lfs -text
|
38 |
+
cra-v1-guided-7b-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
39 |
+
cra-v1-guided-7b-Q5_K.gguf filter=lfs diff=lfs merge=lfs -text
|
40 |
+
cra-v1-guided-7b-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
41 |
+
cra-v1-guided-7b-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
adapter.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:991a5e4efe59cb19f7567daa57330f4bffe8e3bb5cf60a6c10b6ac1ff70d435a
|
3 |
+
size 80767488
|
adapter_config.json
ADDED
@@ -0,0 +1,32 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"alpha_pattern": {},
|
3 |
+
"auto_mapping": null,
|
4 |
+
"base_model_name_or_path": "/llm-downloader-destination/base/fireworks/qwen2p5-7b-instruct/hf",
|
5 |
+
"bias": "none",
|
6 |
+
"fan_in_fan_out": false,
|
7 |
+
"inference_mode": true,
|
8 |
+
"init_lora_weights": true,
|
9 |
+
"layers_pattern": null,
|
10 |
+
"layers_to_transform": null,
|
11 |
+
"loftq_config": {},
|
12 |
+
"lora_alpha": 64,
|
13 |
+
"lora_dropout": 0.05,
|
14 |
+
"megatron_config": null,
|
15 |
+
"megatron_core": "megatron.core",
|
16 |
+
"modules_to_save": null,
|
17 |
+
"peft_type": "LORA",
|
18 |
+
"r": 8,
|
19 |
+
"rank_pattern": {},
|
20 |
+
"revision": null,
|
21 |
+
"target_modules": [
|
22 |
+
"down_proj",
|
23 |
+
"up_proj",
|
24 |
+
"q_proj",
|
25 |
+
"gate_proj",
|
26 |
+
"v_proj",
|
27 |
+
"o_proj",
|
28 |
+
"k_proj"
|
29 |
+
],
|
30 |
+
"task_type": "CAUSAL_LM",
|
31 |
+
"use_rslora": false
|
32 |
+
}
|
adapter_model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:90f8666df804ba05e6ec7d33243ff60dd87bf45d9809fde1ac614d2ef08c8ec4
|
3 |
+
size 40422208
|
cra-v1-guided-7b-F16.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:03f91868f8a86fd9c01b3fa27bb03a089cb25f2fc644fc4a1481cfb005006ecd
|
3 |
+
size 15237853600
|
cra-v1-guided-7b-Q4_K_M.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29b4ca0b2d1873b397b4d68cf19df72da57c177c7491a4bd04949d9315d5b73b
|
3 |
+
size 4683073952
|
cra-v1-guided-7b-Q5_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d8806e64197651a24f209df598846f344581f7806ffaa35c83386c61f403f4fc
|
3 |
+
size 5444831648
|
cra-v1-guided-7b-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:364e722c77ab865ca1a56006cffcc7f3c9db03d3159c841ac6d38803c78a54f8
|
3 |
+
size 6254199200
|
cra-v1-guided-7b-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:12119349ee64bbbf702e64171d1980f9814617898071d931f368171189c7ac30
|
3 |
+
size 8098525600
|