Gryphe commited on
Commit
27b830f
·
1 Parent(s): 36d249d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -5,8 +5,8 @@ language:
5
  ---
6
  An experiment with gradient merges using [the following script](https://github.com/TehVenomm/LM_Transformers_BlockMerge), with [Chronos](https://huggingface.co/elinas/chronos-13b) as its primary model, augmented by [Hermes](https://huggingface.co/NousResearch/Nous-Hermes-13b) and [Wizard-Vicuna Uncensored](https://huggingface.co/TheBloke/Wizard-Vicuna-13B-Uncensored-HF).
7
 
8
- Chronos is a wonderful model, though doesn't feel very smart. Hermes and WizardLM have been merged gradually, primarily in the higher layers (10+) in an attempt to rectify some of this behaviour without affecting Chronos' lengthy replies.
9
- I'd say the end product is about 60% Chronos, with 20% Hermes and 20% Wizard added in gradually increasing amounts.
10
 
11
  This model primarily uses Alpaca formatting, so for optimal model performance, use:
12
  ```
 
5
  ---
6
  An experiment with gradient merges using [the following script](https://github.com/TehVenomm/LM_Transformers_BlockMerge), with [Chronos](https://huggingface.co/elinas/chronos-13b) as its primary model, augmented by [Hermes](https://huggingface.co/NousResearch/Nous-Hermes-13b) and [Wizard-Vicuna Uncensored](https://huggingface.co/TheBloke/Wizard-Vicuna-13B-Uncensored-HF).
7
 
8
+ Chronos is a wonderful model, though doesn't feel very smart. Hermes and WizardLM have been merged gradually, primarily in the higher layers (10+) in an attempt to rectify some of this behaviour.
9
+ I'd say the end product is about 60% Chronos, with 20% Hermes and 20% Wizard added in gradually increasing amounts. The result feels surprisingly robust, though I'll let you be the final judge of that!
10
 
11
  This model primarily uses Alpaca formatting, so for optimal model performance, use:
12
  ```