grimjim commited on
Commit
1a7d2d5
1 Parent(s): a38702b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -54
README.md CHANGED
@@ -1,54 +1,54 @@
1
- ---
2
- base_model:
3
- - grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B
4
- - grimjim/cuckoo-starling-7B
5
- library_name: transformers
6
- tags:
7
- - mergekit
8
- - merge
9
- license: cc-by-nc-4.0
10
- pipeline_tag: text-generation
11
- ---
12
- # rogue-enchantress-32k-7B-GGUF
13
-
14
- This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
-
16
- An ambition of this merge was to augment text generation with the potential creative richness of the WizardLM-2 7B and Zephyr-7B-Beta models, the reasoning of the Starling-LM-7B-beta model, and extended context length of Mistral v0.2.
17
-
18
- The resulting model is very attentive to character card descriptions and capable of applying reasoning. This model is in the smarter side, attentive to context and formatting. The model is creative and "wants" to write, incorporating details cooperatively with occasional runaway narration if it finds that the prompt leans that way.
19
-
20
- Tested with ChatML Instruct prompts, temperature 1.0, and minP 0.02.
21
-
22
- - Full weights: [grimjim/rogue-enchantress-32k-7B](https://huggingface.co/grimjim/rogue-enchantress-32k-7B)
23
- - GGUF quants: [grimjim/rogue-enchantress-32k-7B-GGUF](https://huggingface.co/grimjim/rogue-enchantress-32k-7B-GGUF)
24
-
25
- ## Merge Details
26
- ### Merge Method
27
-
28
- This model was merged using the SLERP merge method.
29
-
30
- ### Models Merged
31
-
32
- The following models were included in the merge:
33
- * [grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B](https://huggingface.co/grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B)
34
- * [grimjim/cuckoo-starling-7B](https://huggingface.co/grimjim/cuckoo-starling-7B)
35
-
36
- ### Configuration
37
-
38
- The following YAML configuration was used to produce this model:
39
-
40
- ```yaml
41
- slices:
42
- - sources:
43
- - model: grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B
44
- layer_range: [0,32]
45
- - model: grimjim/cuckoo-starling-7B
46
- layer_range: [0,32]
47
- merge_method: slerp
48
- base_model: grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B
49
- parameters:
50
- t:
51
- - value: 0.5
52
- dtype: bfloat16
53
-
54
- ```
 
1
+ ---
2
+ base_model:
3
+ - grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B
4
+ - grimjim/cuckoo-starling-7B
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ license: cc-by-nc-4.0
10
+ pipeline_tag: text-generation
11
+ ---
12
+ # rogue-enchantress-32k-7B-GGUF
13
+
14
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
15
+
16
+ An ambition of this merge was to augment text generation with the potential creative richness of the WizardLM-2 7B and Zephyr-7B-Beta models, the reasoning of the Starling-LM-7B-beta model, and extended context length of Mistral v0.2.
17
+
18
+ The resulting model is very attentive to character card descriptions and capable of applying reasoning. This model is in the smarter side, following context and formatting. The model is creative and "wants" to write, incorporating details cooperatively with occasional runaway narration if it finds that the prompt leans that way.
19
+
20
+ Tested with ChatML Instruct prompts, temperature 1.0, and minP 0.02.
21
+
22
+ - Full weights: [grimjim/rogue-enchantress-32k-7B](https://huggingface.co/grimjim/rogue-enchantress-32k-7B)
23
+ - GGUF quants: [grimjim/rogue-enchantress-32k-7B-GGUF](https://huggingface.co/grimjim/rogue-enchantress-32k-7B-GGUF)
24
+
25
+ ## Merge Details
26
+ ### Merge Method
27
+
28
+ This model was merged using the SLERP merge method.
29
+
30
+ ### Models Merged
31
+
32
+ The following models were included in the merge:
33
+ * [grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B](https://huggingface.co/grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B)
34
+ * [grimjim/cuckoo-starling-7B](https://huggingface.co/grimjim/cuckoo-starling-7B)
35
+
36
+ ### Configuration
37
+
38
+ The following YAML configuration was used to produce this model:
39
+
40
+ ```yaml
41
+ slices:
42
+ - sources:
43
+ - model: grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B
44
+ layer_range: [0,32]
45
+ - model: grimjim/cuckoo-starling-7B
46
+ layer_range: [0,32]
47
+ merge_method: slerp
48
+ base_model: grimjim/zephyr-wizard-kuno-royale-BF16-merge-7B
49
+ parameters:
50
+ t:
51
+ - value: 0.5
52
+ dtype: bfloat16
53
+
54
+ ```