metadata
license: apache-2.0
hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF
This model was converted to GGUF format from VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct
using llama.cpp.
Refer to the original model card for more details on the model.
Since HuggingFace supports files up to 50GB, the Q6_K quant is split into two files instead. You probably need to merge them again before you can use them. You can use llama.cpp for that. Use
./llama-gguf-split -h
to find out more about the function, after you installed llama.cpp.