Edit model card

hwarnecke/SauerkrautLM-Nemo-12b-Instruct-Q6_K-GGUF

This model was converted to GGUF format from VAGOsolutions/Llama-3.1-SauerkrautLM-70b-Instruct using llama.cpp. Refer to the original model card for more details on the model.

Since HuggingFace supports files up to 50GB, the Q6_K quant is split into two files instead. You probably need to merge them again before you can use them. You can use llama.cpp for that. Use

./llama-gguf-split -h

to find out more about the function, after you installed llama.cpp.

Downloads last month
3
GGUF
Model size
70.6B params
Architecture
llama

4-bit

6-bit

Inference API
Unable to determine this model's library. Check the docs .