Fimbul-Airo-18B / README.md
Burn Oil
Update README.md
50531dc verified
|
raw
history blame
1.56 kB
---
base_model:
- Sao10K/Fimbulvetr-11B-v2
- Undi95/Mistral-11B-CC-Air-RP
library_name: transformers
tags:
- mergekit
- merge
- πŸ‘
---
# Fimbul-Airo-18B πŸ‘
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). πŸ‘
I tested it for thirtneen.second πŸ‘
Works pretty good, seems uncensored. I'll update with more results/observations as I continue to test.
## Merge Details
### Merge Method
This model was merged using the passthrough merge method. Taking models and smashing em all together πŸ‘
### Models Merged
The following models were included in the merge:
* [Sao10K/Fimbulvetr-11B-v2](https://huggingface.co/Sao10K/Fimbulvetr-11B-v2) πŸ‘
* [Undi95/Mistral-11B-CC-Air-RP](https://huggingface.co/Undi95/Mistral-11B-CC-Air-RP) πŸ‘
* [CollectiveCognition-v1.1-Mistral-7B](https://huggingface.co/teknium/CollectiveCognition-v1.1-Mistral-7B)
* [airoboros-mistral2.2-7b](https://huggingface.co/teknium/airoboros-mistral2.2-7b/)
* PIPPA dataset 11B qlora
* LimaRPv3 dataset 11B qlora
### The Sauce
The following YAML configuration was used to produce this model:
```yaml
slices:
- sources:
- model: Sao10K/Fimbulvetr-11B-v2
layer_range: [0, 40]
- sources:
- model: Undi95/Mistral-11B-CC-Air-RP
layer_range: [8, 48]
merge_method: passthrough
dtype: bfloat16
πŸ‘
```
### Prompt Format: Alpaca πŸ‘
```
### Instruction:
<Prompt>
### Input:
<Insert Context Here>
### Response:
```
πŸ‘
Don't forget to take care of yourself and have a wonderful day!