Erosumika GGUF quants
Collection
6 items
•
Updated
•
1
This repo contains GGUF format model files for Erosumika 7B.
This is an attempt to create a model that combines multiple "established" 7Bs and a very small WIP private dataset with Eros' raw creative power. In terms of instruction formats, ChatML and Alpaca work best. The merge isn't purely ChatML, and as such, my previous attempts to integrate it with ChatML strings out of the box were Sisyphean and uninformed.
- sources:
- model: localfultonextractor/sumitest02
layer_range: [0, 32]
- model: tavtav/eros-7b-test
layer_range: [0, 32]
merge_method: slerp
base_model: localfultonextractor/sumitest02
parameters:
t:
- filter: self_attn
value: [0, 0.2, 0.4, 0.55, 0.8]
- filter: mlp
value: [0.7, 0.3, 0.4, 0.3, 0]
- value: 0.37 # fallback for rest of tensors
dtype: float16