hwarnecke's picture
Update README.md
c4cf124 verified
|
raw
history blame
707 Bytes
metadata
license: apache-2.0

hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF

This model was converted to GGUF format from VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct using llama.cpp. Refer to the original model card for more details on the model.

Since HuggingFace supports files up to 50GB, the Q6_K quant is split into two files instead. You probably need to merge them again before you can use them. You can use llama.cpp for that. Use

./llama-gguf-split -h

to find out more about the function, after you installed llama.cpp.