Edit model card

LM-Cocktail phi-2 v1.1

This is a 0.5-0.5 merge of two models based on phi-2. Here are the models used to create this merge:

  1. venkycs/phi-2-instruct
  2. Yhyu13/phi-2-sft-dpo-gpt4_en-ep1

I named this model "LMCocktail phi-2 v1.1" because I see it as a continuation of the v1.

I used Yhyu13/phi-2-sft-dpo-gpt4_en-ep1 and it "outputs significantly longer result" than the one used in v1 by Yhyu13.

I also used venkycs/phi-2-instruct "a fine-tuned version of microsoft/phi-2 on the filtered ultrachat200k dataset using the SFT technique".

The main reason I created this model was to merge it with cognitivecomputations/dolphin-2_6-phi-2, and I will create a repo for it when I do it.

Code

The LM-cocktail is novel technique for merging multiple models: https://arxiv.org/abs/2311.13534

Code is backed up by this repo: https://github.com/FlagOpen/FlagEmbedding.git

Merging script is available under the ./scripts folder.

Downloads last month
1
Safetensors
Model size
2.78B params
Tensor type
F32
·