Edit model card

image/png

Erosumika-7B

This is an attempt to create a model that combines multiple "established" 7Bs and a very small WIP private dataset with Eros' raw creative power. In terms of instruction formats, ChatML and Alpaca work best. The merge isn't purely ChatML, and as such, my previous attempts to integrate it with ChatML strings out of the box were Sisyphean and uninformed.

GGUF

exl2, 4bpw

exl2, 6bpw

imatrix GGUF - many thanks to Lewdiculous!

Merge config.yml:

  • I was asked to upload the merge configuration I used, sadly the one for the 'sumitest02' model is lost to time, like tears in rain:
  - sources:
      - model: localfultonextractor/sumitest02
        layer_range: [0, 32]
      - model: tavtav/eros-7b-test
        layer_range: [0, 32]
merge_method: slerp
base_model: localfultonextractor/sumitest02
parameters:
  t:
    - filter: self_attn
      value: [0, 0.2, 0.4, 0.55, 0.8]
    - filter: mlp
      value: [0.7, 0.3, 0.4, 0.3, 0]
    - value: 0.37 # fallback for rest of tensors
dtype: float16

Open LLM Leaderboard Evaluation Results

Detailed results can be found here

|             Metric              |Value|
|---------------------------------|----:|
|Avg.                             |64.92|
|AI2 Reasoning Challenge (25-Shot)|62.88|
|HellaSwag (10-Shot)              |85.90|
|MMLU (5-Shot)                    |60.64|
|TruthfulQA (0-shot)              |67.59|
|Winogrande (5-shot)              |75.30|
|GSM8k (5-shot)                   |37.23|
Downloads last month
0
Safetensors
Model size
7.24B params
Tensor type
FP16
·
Inference API
Input a message to start chatting with localfultonextractor/Erosumika-7B.
Inference API (serverless) does not yet support adapter-transformers models for this pipeline type.