File size: 3,860 Bytes
a5db143 97a6908 a5db143 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 27518fe 97a6908 a5db143 97a6908 a5db143 27518fe a5db143 27518fe |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 |
---
base_model: jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9
inference: false
library_name: transformers
license: apache-2.0
merged_models:
- jsfs11/MixtureofMerges-MoE-2x7b-v7
- jsfs11/MixtureofMerges-MoE-2x7bRP-v8
model-index:
- name: MixtureofMerges-MoE-2x7b-SLERPv0.9
results:
- dataset:
args:
num_few_shot: 25
config: ARC-Challenge
name: AI2 Reasoning Challenge (25-Shot)
split: test
type: ai2_arc
metrics:
- name: normalized accuracy
type: acc_norm
value: 73.12
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 10
name: HellaSwag (10-Shot)
split: validation
type: hellaswag
metrics:
- name: normalized accuracy
type: acc_norm
value: 88.76
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 5
config: all
name: MMLU (5-Shot)
split: test
type: cais/mmlu
metrics:
- name: accuracy
type: acc
value: 65.0
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 0
config: multiple_choice
name: TruthfulQA (0-shot)
split: validation
type: truthful_qa
metrics:
- type: mc2
value: 74.83
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 5
config: winogrande_xl
name: Winogrande (5-shot)
split: validation
type: winogrande
metrics:
- name: accuracy
type: acc
value: 83.58
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9
task:
name: Text Generation
type: text-generation
- dataset:
args:
num_few_shot: 5
config: main
name: GSM8k (5-shot)
split: test
type: gsm8k
metrics:
- name: accuracy
type: acc
value: 69.22
source:
name: Open LLM Leaderboard
url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9
task:
name: Text Generation
type: text-generation
pipeline_tag: text-generation
quantized_by: Suparious
tags:
- 4-bit
- AWQ
- text-generation
- autotrain_compatible
- endpoints_compatible
- merge
- mergekit
- lazymergekit
- jsfs11/MixtureofMerges-MoE-2x7b-v7
- jsfs11/MixtureofMerges-MoE-2x7bRP-v8
---
# jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9 AWQ
- Model creator: [jsfs11](https://huggingface.co/jsfs11)
- Original model: [MixtureofMerges-MoE-2x7b-SLERPv0.9](https://huggingface.co/jsfs11/MixtureofMerges-MoE-2x7b-SLERPv0.9)
## Model Summary
MixtureofMerges-MoE-2x7b-SLERPv0.9 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
* [jsfs11/MixtureofMerges-MoE-2x7b-v7](https://huggingface.co/jsfs11/MixtureofMerges-MoE-2x7b-v7)
* [jsfs11/MixtureofMerges-MoE-2x7bRP-v8](https://huggingface.co/jsfs11/MixtureofMerges-MoE-2x7bRP-v8)
|