--- license: llama2 --- # Model Card for Model ID **dragon-llama-qa-tool** is a 4_K_M quantized GGUF version of DRAGON Llama, providing a fast, small inference implementation. Load in your favorite GGUF inference engine, or try with llmware as follows: from llmware.models import ModelCatalog qa_tool = ModelCatalog().load_model("llmware/dragon-llama-qa-tool") response = qa_tool.inference(query, text_sample) ### Model Description - **Developed by:** llmware - **Model type:** GGUF - **Language(s) (NLP):** English - **License:** Apache 2.0 - **Quantized from model:** llmware/dragon-llama (finetuned llama) ## Uses All of the DRAGON models use the following prompt wrapper template: " " + context + "\n" + question + "\n: " ## Model Card Contact Darren Oberst & llmware team