Edit model card

Merge:

layer_slices:
  - model: Gryphe/MythoMax-L2-13b
    start: 0
    end: 16
  - model: Undi95/MM-ReMM-L2-20B-Part1
    start: 8
    end: 20
  - model: Gryphe/MythoMax-L2-13b
    start: 17
    end: 32
  - model: Undi95/MM-ReMM-L2-20B-Part1
    start: 21
    end: 40

Models used

  • Gryphe/MythoMax-L2-13b
  • Undi95/ReMM-v2.1-L2-13B

Part1 = ReMM v2.1 merged /w MythoMax low weight to keep consistency. I call this "dilution" and result show consistency and coherency without repeat/loop beside the small amount of duplicated datas.

Prompt template: Alpaca

Below is an instruction that describes a task. Write a response that completes the request.

### Instruction:
{prompt}

### Response:

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 51.14
ARC (25-shot) 60.84
HellaSwag (10-shot) 85.18
MMLU (5-shot) 56.45
TruthfulQA (0-shot) 53.33
Winogrande (5-shot) 75.77
GSM8K (5-shot) 7.73
DROP (3-shot) 18.66
Downloads last month
2,492
Safetensors
Model size
20B params
Tensor type
FP16
·
BF16
·
F32
·

Collection including Undi95/MM-ReMM-L2-20B