asiansoul's picture
Update README.md
2388789 verified
|
raw
history blame
3.19 kB
metadata
base_model:
  - nayohan/llama3-8b-it-translation-general-en-ko-1sent
  - MaziyarPanahi/Llama-3-8B-Instruct-DPO-v0.3
  - cognitivecomputations/dolphin-2.9-llama3-8b
  - NousResearch/Hermes-2-Pro-Llama-3-8B
  - winglian/llama-3-8b-1m-PoSE
  - asiansoul/Llama-3-Open-Ko-Linear-8B
  - NousResearch/Meta-Llama-3-8B
  - Danielbrdz/Barcenas-Llama3-8b-ORPO
  - NousResearch/Meta-Llama-3-8B-Instruct
library_name: transformers
tags:
  - mergekit
  - merge

Versatile-Llama-3-8B

I'm not going to say that this Merge model is the best model ever made. I'm not going to tell you that you'll enjoy chatting with my merge model.

All I want to say is thank you for taking time out of your day to visit. Without users like you, my work would be meaningless.

Not test this model, just based on brainstorming, release this.

Merge Details

Merge Method

This model was merged using the DARE TIES merge method using NousResearch/Meta-Llama-3-8B as a base.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

models:
  - model: NousResearch/Meta-Llama-3-8B
    # Base model providing a general foundation without specific parameters

  - model: NousResearch/Meta-Llama-3-8B-Instruct
    parameters:
      density: 0.60  
      weight: 0.25  

  - model: winglian/llama-3-8b-1m-PoSE
    parameters:
      density: 0.55  
      weight: 0.15  

  - model: MaziyarPanahi/Llama-3-8B-Instruct-DPO-v0.3
    parameters:
      density: 0.55  
      weight: 0.15  

  - model: asiansoul/Llama-3-Open-Ko-Linear-8B
    parameters:
      density: 0.55  
      weight: 0.2  

  - model: nayohan/llama3-8b-it-translation-general-en-ko-1sent
    parameters:
      density: 0.55  
      weight: 0.1  

  - model: cognitivecomputations/dolphin-2.9-llama3-8b
    parameters:
      density: 0.55  
      weight: 0.1  

  - model: Danielbrdz/Barcenas-Llama3-8b-ORPO
    parameters:
      density: 0.55  
      weight: 0.05 

  - model: NousResearch/Hermes-2-Pro-Llama-3-8B
    parameters:
      density: 0.55  
      weight: 0.1  

merge_method: dare_ties
base_model: NousResearch/Meta-Llama-3-8B
parameters:
  int8_mask: true
dtype: bfloat16