SnowStorm-v1.15-4x8B-B
Collection
v1.15B, vision, based on Stheno-v3.1
•
3 items
•
Updated
•
1
Experimental RP-oriented MoE, the idea was to get a model that would be equal to or better than Mixtral 8x7B and it's finetunes in RP/ERP tasks.
There's:
base_model: Sao10K_L3-8B-Stheno-v3.1
gate_mode: random
dtype: bfloat16
experts_per_token: 2
experts:
- source_model: Nitral-AI_Poppy_Porpoise-1.0-L3-8B
- source_model: NeverSleep_Llama-3-Lumimaid-8B-v0.1-OAS
- source_model: openlynn_Llama-3-Soliloquy-8B-v2
- source_model: Sao10K_L3-8B-Stheno-v3.1
Detailed results can be found here
Metric | Value |
---|---|
Avg. | 68.01 |
AI2 Reasoning Challenge (25-Shot) | 60.67 |
HellaSwag (10-Shot) | 81.60 |
MMLU (5-Shot) | 68.12 |
TruthfulQA (0-shot) | 51.69 |
Winogrande (5-shot) | 76.56 |
GSM8k (5-shot) | 69.45 |