LM-Cocktail phi-2 v1.1
This is a 0.5-0.5 merge of two models based on phi-2. Here are the models used to create this merge:
I named this model "LMCocktail phi-2 v1.1" because I see it as a continuation of the v1.
I used Yhyu13/phi-2-sft-dpo-gpt4_en-ep1 and it "outputs significantly longer result" than the one used in v1 by Yhyu13.
I also used venkycs/phi-2-instruct "a fine-tuned version of microsoft/phi-2 on the filtered ultrachat200k dataset using the SFT technique".
The main reason I created this model was to merge it with cognitivecomputations/dolphin-2_6-phi-2, and I will create a repo for it when I do it.
Code
The LM-cocktail is novel technique for merging multiple models: https://arxiv.org/abs/2311.13534
Code is backed up by this repo: https://github.com/FlagOpen/FlagEmbedding.git
Merging script is available under the ./scripts folder.
- Downloads last month
- 13