Edit model card
YAML Metadata Warning: empty or missing yaml metadata in repo card (https://huggingface.co/docs/hub/model-cards#model-card-metadata)

Quantization made by Richard Erkhov.

Github

Discord

Request more models

SeverusWestLake-7B-DPO - GGUF

Original model description:

license: mit library_name: transformers tags: - mergekit - merge base_model: - FelixChao/Sectumsempra-7B-DPO - cognitivecomputations/WestLake-7B-v2-laser pipeline_tag: text-generation model-index: - name: SeverusWestLake-7B-DPO results: - task: type: text-generation name: Text Generation dataset: name: AI2 Reasoning Challenge (25-Shot) type: ai2_arc config: ARC-Challenge split: test args: num_few_shot: 25 metrics: - type: acc_norm value: 72.18 name: normalized accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=s3nh/SeverusWestLake-7B-DPO name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: HellaSwag (10-Shot) type: hellaswag split: validation args: num_few_shot: 10 metrics: - type: acc_norm value: 88.94 name: normalized accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=s3nh/SeverusWestLake-7B-DPO name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU (5-Shot) type: cais/mmlu config: all split: test args: num_few_shot: 5 metrics: - type: acc value: 64.65 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=s3nh/SeverusWestLake-7B-DPO name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: TruthfulQA (0-shot) type: truthful_qa config: multiple_choice split: validation args: num_few_shot: 0 metrics: - type: mc2 value: 71.49 source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=s3nh/SeverusWestLake-7B-DPO name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: Winogrande (5-shot) type: winogrande config: winogrande_xl split: validation args: num_few_shot: 5 metrics: - type: acc value: 86.11 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=s3nh/SeverusWestLake-7B-DPO name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GSM8k (5-shot) type: gsm8k config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 69.14 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=s3nh/SeverusWestLake-7B-DPO name: Open LLM Leaderboard

SeverusWestLake-7B-DPO

image

This is a merge of pre-trained language models created using mergekit.

Merge Details

Merge Method

This model was merged using the SLERP merge method.

Models Merged

The following models were included in the merge:

Configuration

The following YAML configuration was used to produce this model:

base_model: FelixChao/Sectumsempra-7B-DPO
dtype: bfloat16
merge_method: slerp
parameters:
  t:
  - filter: self_attn
    value: [0.0, 0.5, 0.3, 0.7, 1.0]
  - filter: mlp
    value: [1.0, 0.5, 0.7, 0.3, 0.0]
  - value: 0.5
slices:
- sources:
  - layer_range: [0, 32]
    model: FelixChao/Sectumsempra-7B-DPO
  - layer_range: [0, 32]
    model: cognitivecomputations/WestLake-7B-v2-laser

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 75.42
AI2 Reasoning Challenge (25-Shot) 72.18
HellaSwag (10-Shot) 88.94
MMLU (5-Shot) 64.65
TruthfulQA (0-shot) 71.49
Winogrande (5-shot) 86.11
GSM8k (5-shot) 69.14
Downloads last month
2
GGUF
Model size
7.24B params
Architecture
llama

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

Inference API
Unable to determine this model's library. Check the docs .