Update README.md
Browse files
README.md
CHANGED
@@ -5,8 +5,8 @@ language:
|
|
5 |
---
|
6 |
An experiment with gradient merges using [the following script](https://github.com/TehVenomm/LM_Transformers_BlockMerge), with [Chronos](https://huggingface.co/elinas/chronos-13b) as its primary model, augmented by [Hermes](https://huggingface.co/NousResearch/Nous-Hermes-13b) and [Wizard-Vicuna Uncensored](https://huggingface.co/TheBloke/Wizard-Vicuna-13B-Uncensored-HF).
|
7 |
|
8 |
-
Chronos is a wonderful model, though doesn't feel very smart. Hermes and WizardLM have been merged gradually, primarily in the higher layers (10+) in an attempt to rectify some of this behaviour
|
9 |
-
I'd say the end product is about 60% Chronos, with 20% Hermes and 20% Wizard added in gradually increasing amounts.
|
10 |
|
11 |
This model primarily uses Alpaca formatting, so for optimal model performance, use:
|
12 |
```
|
|
|
5 |
---
|
6 |
An experiment with gradient merges using [the following script](https://github.com/TehVenomm/LM_Transformers_BlockMerge), with [Chronos](https://huggingface.co/elinas/chronos-13b) as its primary model, augmented by [Hermes](https://huggingface.co/NousResearch/Nous-Hermes-13b) and [Wizard-Vicuna Uncensored](https://huggingface.co/TheBloke/Wizard-Vicuna-13B-Uncensored-HF).
|
7 |
|
8 |
+
Chronos is a wonderful model, though doesn't feel very smart. Hermes and WizardLM have been merged gradually, primarily in the higher layers (10+) in an attempt to rectify some of this behaviour.
|
9 |
+
I'd say the end product is about 60% Chronos, with 20% Hermes and 20% Wizard added in gradually increasing amounts. The result feels surprisingly robust, though I'll let you be the final judge of that!
|
10 |
|
11 |
This model primarily uses Alpaca formatting, so for optimal model performance, use:
|
12 |
```
|