Upload folder using huggingface_hub
Browse files- .gitattributes +2 -0
- README.md +53 -3
- taide-it-llama3-16b-Q6_K.gguf +3 -0
- taide-it-llama3-16b-Q8_0.gguf +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
taide-it-llama3-16b-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
37 |
+
taide-it-llama3-16b-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,5 +1,55 @@
|
|
1 |
---
|
2 |
-
|
3 |
-
|
4 |
-
|
|
|
|
|
|
|
5 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
+
base_model: []
|
3 |
+
library_name: transformers
|
4 |
+
tags:
|
5 |
+
- mergekit
|
6 |
+
- merge
|
7 |
+
|
8 |
---
|
9 |
+
# llama3-15b-v02
|
10 |
+
|
11 |
+
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
|
12 |
+
|
13 |
+
## Merge Details
|
14 |
+
### Merge Method
|
15 |
+
|
16 |
+
This model was merged using the [linear](https://arxiv.org/abs/2203.05482) merge method.
|
17 |
+
|
18 |
+
### Models Merged
|
19 |
+
|
20 |
+
The following models were included in the merge:
|
21 |
+
* D:/text-generation-webui/models/meta-llama_Meta-Llama-3-8B-Instruct
|
22 |
+
* D:/text-generation-webui/models/taide_Llama3-TAIDE-LX-8B-Chat-Alpha1
|
23 |
+
|
24 |
+
### Configuration
|
25 |
+
|
26 |
+
The following YAML configuration was used to produce this model:
|
27 |
+
|
28 |
+
```yaml
|
29 |
+
dtype: bfloat16
|
30 |
+
merge_method: linear # use linear so we can include multiple models, albeit at a zero weight
|
31 |
+
parameters:
|
32 |
+
weight: 1.0 # weight everything as 1 unless specified otherwise - linear with one model weighted at 1 is a no-op like passthrough
|
33 |
+
slices:
|
34 |
+
- sources:
|
35 |
+
- layer_range: [0, 1]
|
36 |
+
model: D:/text-generation-webui/models/meta-llama_Meta-Llama-3-8B-Instruct
|
37 |
+
- layer_range: [0, 1]
|
38 |
+
model: D:/text-generation-webui/models/taide_Llama3-TAIDE-LX-8B-Chat-Alpha1
|
39 |
+
parameters:
|
40 |
+
weight: 0
|
41 |
+
- sources:
|
42 |
+
- layer_range: [1, 24]
|
43 |
+
model: D:/text-generation-webui/models/meta-llama_Meta-Llama-3-8B-Instruct
|
44 |
+
- layer_range: [1, 24]
|
45 |
+
model: D:/text-generation-webui/models/taide_Llama3-TAIDE-LX-8B-Chat-Alpha1
|
46 |
+
- sources:
|
47 |
+
- layer_range: [24, 32]
|
48 |
+
model: D:/text-generation-webui/models/taide_Llama3-TAIDE-LX-8B-Chat-Alpha1
|
49 |
+
parameters:
|
50 |
+
weight: 0
|
51 |
+
- layer_range: [24, 32]
|
52 |
+
model: D:/text-generation-webui/models/meta-llama_Meta-Llama-3-8B-Instruct
|
53 |
+
|
54 |
+
|
55 |
+
```
|
taide-it-llama3-16b-Q6_K.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bc4c5338fb881a3bc5f26a9a29a43ef9c929e27104ec8c9703ebc2df9a9a1f64
|
3 |
+
size 6596006176
|
taide-it-llama3-16b-Q8_0.gguf
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6acb625a6fb595d27e18df7e43330b3f621a532070d52c8f9a1bf3e8e86d96e8
|
3 |
+
size 8540770592
|