Edit model card

Nous-Hermes-2-Mistral-7B-DPO

I converted NousResearch/Nous-Hermes-2-Mistral-7B-DPO to GGUF and quantized it to my favorite quantizations. Se their original model card for all the details.

I quickly quantized this model using a modified version of AutoGGUF from Maxime Labonne

Here is my Ollama modelfile. According to llama.cpp the model was trained on 32k tokens, but I set the ollama modelfile to 16k so that 16 GB Macs can still run it.

FROM ./nous-hermes-2-mistral-7b-dpo.Q5_K_M.gguf
PARAMETER num_ctx 16384
TEMPLATE """<|im_start|>system
{{ .System }}<|im_end|>
<|im_start|>user
{{ .Prompt }}<|im_end|>
<|im_start|>assistant
"""
PARAMETER stop "<|im_start|>"
PARAMETER stop "<|im_end|>"
Downloads last month
377
GGUF
Model size
7.24B params
Architecture
llama

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .

Model tree for olafgeibig/Nous-Hermes-2-Mistral-7B-DPO-GGUF

Quantized
(163)
this model

Dataset used to train olafgeibig/Nous-Hermes-2-Mistral-7B-DPO-GGUF