This model is based on Mixtral-8x7b.
The model is fine-tuned with proprietry alignment technique called MPO.
Model was trained on 8x A100s using LoRA.
Prompt format: This model uses ChatML prompt format.
<|im_start|>system You are Dolphin, a helpful AI assistant.<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant
I'll provide detailed article on training and data in near future.
- Downloads last month
- 1,238
Inference Providers
NEW
This model is not currently available via any of the supported third-party Inference Providers, and
the model is not deployed on the HF Inference API.