Edit model card

LM-Cocktail phi-2 v1.1

This is a 0.5-0.5 merge of two models based on phi-2. Here are the models used to create this merge:

  1. venkycs/phi-2-instruct
  2. Yhyu13/phi-2-sft-dpo-gpt4_en-ep1

I named this model "LMCocktail phi-2 v1.1" because I see it as a continuation of the v1.

I used Yhyu13/phi-2-sft-dpo-gpt4_en-ep1 and it "outputs significantly longer result" than the one used in v1 by Yhyu13.

I also used venkycs/phi-2-instruct "a fine-tuned version of microsoft/phi-2 on the filtered ultrachat200k dataset using the SFT technique".

The main reason I created this model was to merge it with cognitivecomputations/dolphin-2_6-phi-2, and I will create a repo for it when I do it.

Code

The LM-cocktail is novel technique for merging multiple models: https://arxiv.org/abs/2311.13534

Code is backed up by this repo: https://github.com/FlagOpen/FlagEmbedding.git

Merging script is available under the ./scripts folder.

Downloads last month
13
Safetensors
Model size
2.78B params
Tensor type
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.