Edit model card

GGUF

Experimental RP-oriented MoE, the idea was to get a model that would be equal to or better than Mixtral 8x7B and it's finetunes in RP/ERP tasks.

There's:

Llama 3 SnowStorm v1.15B 4x8B

base_model: Sao10K_L3-8B-Stheno-v3.1
gate_mode: random
dtype: bfloat16
experts_per_token: 2
experts:
  - source_model: Nitral-AI_Poppy_Porpoise-1.0-L3-8B
  - source_model: NeverSleep_Llama-3-Lumimaid-8B-v0.1-OAS
  - source_model: openlynn_Llama-3-Soliloquy-8B-v2
  - source_model: Sao10K_L3-8B-Stheno-v3.1

Models used

Difference(from SnowStorm v1.0)

Vision

llama3_mmproj

image/png

Prompt format: Llama 3

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

Metric Value
Avg. 68.01
AI2 Reasoning Challenge (25-Shot) 60.67
HellaSwag (10-Shot) 81.60
MMLU (5-Shot) 68.12
TruthfulQA (0-shot) 51.69
Winogrande (5-shot) 76.56
GSM8k (5-shot) 69.45
Downloads last month
857
Safetensors
Model size
24.9B params
Tensor type
BF16
·

Collection including xxx777xxxASD/L3-SnowStorm-v1.15-4x8B-B

Evaluation results