Erosumika-7B
This is an attempt to create a model that combines multiple "established" 7Bs and a very small WIP private dataset with Eros' raw creative power. In terms of instruction formats, ChatML and Alpaca work best. The merge isn't purely ChatML, and as such, my previous attempts to integrate it with ChatML strings out of the box were Sisyphean and uninformed.
imatrix GGUF - many thanks to Lewdiculous!
Merge config.yml:
- I was asked to upload the merge configuration I used, sadly the one for the 'sumitest02' model is lost to time, like tears in rain:
- sources:
- model: localfultonextractor/sumitest02
layer_range: [0, 32]
- model: tavtav/eros-7b-test
layer_range: [0, 32]
merge_method: slerp
base_model: localfultonextractor/sumitest02
parameters:
t:
- filter: self_attn
value: [0, 0.2, 0.4, 0.55, 0.8]
- filter: mlp
value: [0.7, 0.3, 0.4, 0.3, 0]
- value: 0.37 # fallback for rest of tensors
dtype: float16
Open LLM Leaderboard Evaluation Results
Detailed results can be found here
| Metric |Value|
|---------------------------------|----:|
|Avg. |64.92|
|AI2 Reasoning Challenge (25-Shot)|62.88|
|HellaSwag (10-Shot) |85.90|
|MMLU (5-Shot) |60.64|
|TruthfulQA (0-shot) |67.59|
|Winogrande (5-shot) |75.30|
|GSM8k (5-shot) |37.23|
- Downloads last month
- 0