maldv's picture
Update README.md
960d73a verified
---
license: cc-by-nc-4.0
tags:
- merge
- conversational
- multi-task
pipeline_tag: text-generation
base_model:
- paulml/OmniBeagleSquaredMBX-v3-7B
- ZySec-AI/ZySec-7B-v1
- liminerity/Omningotex-7b-slerp
- localfultonextractor/Erosumika-7B
- KatyTheCutie/LemonadeRP-4.5.3
- cgato/Thespis-Krangled-7b
- CorticalStack/pastiche-crown-clown-7b-dare
- snorkelai/Snorkel-Mistral-PairRM-DPO
- MTSAIR/multi_verse_model
model-index:
- name: winter-garden-7b-alpha - "Smart Assistant"
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: AI2 Reasoning Challenge (25-Shot)
type: ai2_arc
config: ARC-Challenge
split: test
args:
num_few_shot: 25
metrics:
- type: acc_norm
value: 65.19
name: normalized accuracy
source:
url: >-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=maldv/winter-garden-7b-alpha
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: HellaSwag (10-Shot)
type: hellaswag
split: validation
args:
num_few_shot: 10
metrics:
- type: acc_norm
value: 85.36
name: normalized accuracy
source:
url: >-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=maldv/winter-garden-7b-alpha
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: MMLU (5-Shot)
type: cais/mmlu
config: all
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 65.2
name: accuracy
source:
url: >-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=maldv/winter-garden-7b-alpha
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: TruthfulQA (0-shot)
type: truthful_qa
config: multiple_choice
split: validation
args:
num_few_shot: 0
metrics:
- type: mc2
value: 50.94
source:
url: >-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=maldv/winter-garden-7b-alpha
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: Winogrande (5-shot)
type: winogrande
config: winogrande_xl
split: validation
args:
num_few_shot: 5
metrics:
- type: acc
value: 80.35
name: accuracy
source:
url: >-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=maldv/winter-garden-7b-alpha
name: Open LLM Leaderboard
- task:
type: text-generation
name: Text Generation
dataset:
name: GSM8k (5-shot)
type: gsm8k
config: main
split: test
args:
num_few_shot: 5
metrics:
- type: acc
value: 54.44
name: accuracy
source:
url: >-
https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=maldv/winter-garden-7b-alpha
name: Open LLM Leaderboard
---
# Winter Garden 7B - α - "Smart Assistant"
It was mentioned that we are in the open ai dark winter; so I thought I would make myself a nice winter garden.
## An experiment
I've merged four partitions successfully in the past, so lets go for 9! I started with:
* Mistral-7B-v0.1
and merged in
* OmniBeagleSquaredMBX-v3-7B
* ZySec-7B-v1
* Omningotex-7b-slerp
* Erosumika-7B
* LemonadeRP-4.5.3
* Thespis-Krangled-7b
* pastiche-crown-clown-7b-dare
* Snorkel-Mistral-PairRM-DPO
* multi_verse_model
### 9-partition merge
All of the layers were partitioned in to 9 random bins. Alternating models were slerped at [0...1], and [1...0] gradients; except attention, which was slerped at 0.03.
This means that the model is still predominantly ordered around base mistral - including half of the input and output layers, and 28% of attention.
### Other
Includes fast tokenizer.
## Chat Template
I put a conversational chat template, which takes "name", "to" (optional), and "content" as the turns. It is designed to follow a transcript style chat which is used by some of the models. This type of use-case is best done by outlining a scene and creating a character card.
```
### {% title %}
{% metadata %}
USER: Hello
ASSISTANT: Hi, how are you?
```
It leans to being a coder when given an `### Instruction`, follows `<s>[INST][/INST]`, and likes `<|user|>`, `<|assistant|>` as well.
A quite cheery and intelligent model. Very good with science and math, but still capable of a decent amount of creativity for a 7b model.
## Scores
Metric | Score
---|---
Average | 66.91
ARC | 65.19
HellaSwag | 85.36
MMLU | 65.2
TruthfulQA | 50.94
Winogrande | 80.35
GSM8K | 54.44
[Details](https://huggingface.co/datasets/open-llm-leaderboard/details_maldv__winter-garden-7b-alpha)