--- license: apache-2.0 datasets: - timdettmers/openassistant-guanaco - databricks/databricks-dolly-15k --- # Mastermax Llama 7B This is a a Llama2 7B base model that was fined tuned on additional datasets, in attempts improve performance. ## How to use with HugginFace pipeline ``` from transformers import AutoModelForCausalLM, AutoTokenizer, pineline model = AutoModelForCausalLM.from_pretrained( "lifeofcoding/mastermax-llama-7b", load_in_4bit=True) tokenizer = AutoTokenizer.from_pretrained("lifeofcoding/mastermax-llama-7b", trust_remote_code=True) # Generate text using the pipeline pipe = pipeline(task="text-generation", model=model, tokenizer=tokenizer, max_length=200) result = pipe(f"[INST] {prompt} [/INST]") generated_text = result[0]['generated_text'] ``` - **Developed by:** [LifeOfCoding](https://www.livingthecode.life/) - **Finetuned from model:** [GuardRail Llama2 7B Sharded](https://huggingface.co/guardrail/llama-2-7b-guanaco-instruct-sharded)