Which models are actually merged?

#3
by kurnevsky - opened

The readme says it's cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser and teknium/OpenHermes-2.5-Mistral-7B but mergekit_moe_config.yml contains cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser and cognitivecomputations/dolphin-2.1-mistral-7b.

I am merging a new configuration because there seems to be some pollution of the repository. Here is the new config

base_model: cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser 
gate_mode: hidden
dtype: bfloat16
experts:
  - source_model: cognitivecomputations/dolphin-2.6-mistral-7b-dpo-laser 
    positive_prompts:
 
      
  - source_model: mlabonne/NeuralBeagle14-7B
    positive_prompts:
kurnevsky changed discussion status to closed

Sign up or log in to comment