hwarnecke commited on
Commit
c4cf124
1 Parent(s): 2dd11ce

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -0
README.md CHANGED
@@ -4,3 +4,11 @@ license: apache-2.0
4
  # hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF
5
  This model was converted to GGUF format from [`VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct`](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) using llama.cpp.
6
  Refer to the [original model card](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) for more details on the model.
 
 
 
 
 
 
 
 
 
4
  # hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF
5
  This model was converted to GGUF format from [`VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct`](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) using llama.cpp.
6
  Refer to the [original model card](https://huggingface.co/VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct) for more details on the model.
7
+
8
+ Since HuggingFace supports files up to 50GB, the Q6_K quant is split into two files instead.
9
+ You probably need to merge them again before you can use them. You can use llama.cpp for that.
10
+ Use
11
+ ```shell
12
+ ./llama-gguf-split -h
13
+ ```
14
+ to find out more about the function, after you installed llama.cpp.