Update README.md
Browse files
README.md
CHANGED
@@ -4,3 +4,11 @@ license: apache-2.0
|
|
4 |
# hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF
|
5 |
This model was converted to GGUF format from [`VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct`](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) using llama.cpp.
|
6 |
Refer to the [original model card](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) for more details on the model.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4 |
# hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF
|
5 |
This model was converted to GGUF format from [`VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct`](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) using llama.cpp.
|
6 |
Refer to the [original model card](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) for more details on the model.
|
7 |
+
|
8 |
+
Since HuggingFace supports files up to 50GB, the Q6_K quant is split into two files instead.
|
9 |
+
You probably need to merge them again before you can use them. You can use llama.cpp for that.
|
10 |
+
Use
|
11 |
+
```shell
|
12 |
+
./llama-gguf-split -h
|
13 |
+
```
|
14 |
+
to find out more about the function, after you installed llama.cpp.
|