grimjim commited on
Commit
32ffbc4
1 Parent(s): 4b3603d

Initial release

Browse files
.gitattributes CHANGED
@@ -4,6 +4,7 @@
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
 
7
  *.gz filter=lfs diff=lfs merge=lfs -text
8
  *.h5 filter=lfs diff=lfs merge=lfs -text
9
  *.joblib filter=lfs diff=lfs merge=lfs -text
 
4
  *.bz2 filter=lfs diff=lfs merge=lfs -text
5
  *.ckpt filter=lfs diff=lfs merge=lfs -text
6
  *.ftz filter=lfs diff=lfs merge=lfs -text
7
+ *.gguf filter=lfs diff=lfs merge=lfs -text
8
  *.gz filter=lfs diff=lfs merge=lfs -text
9
  *.h5 filter=lfs diff=lfs merge=lfs -text
10
  *.joblib filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,52 @@
1
- ---
2
- license: cc-by-nc-4.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - grimjim/rogue-enchantress-32k-7B
4
+ - grimjim/kunoichi-lemon-royale-v2-32K-7B
5
+ library_name: transformers
6
+ tags:
7
+ - mergekit
8
+ - merge
9
+ license: cc-by-nc-4.0
10
+ pipeline_tag: text-generation
11
+ ---
12
+ # kunoichi-lemon-royale-v3-32K-7B
13
+
14
+ This is a merge of pre-trained Mistral 7B language models created using [mergekit](https://github.com/cg123/mergekit).
15
+
16
+ With this merger, we explore *merge densification*, a merge approach that attempts to transfer and adapt some benefits of denser models. A highly creative model, which itself was merged from multiple dense models, was merged in at very low weight in order to lightly modify the base model. The result was expected to improve variability in output without significantly impacting the coherence in the base model.
17
+
18
+ Tested with ChatML instruct templates, temperature 1.0, and minP 0.02. Practical context length should be at least 16K.
19
+
20
+ The additional model merge weight of 0.02 was deliberately chosen to be on par with the minP setting.
21
+
22
+ - Full weights: [grimjim/kunoichi-lemon-royale-v3-32K-7B](https://huggingface.co/grimjim/kunoichi-lemon-royale-v3-32K-7B)
23
+ - GGUF quants: [grimjim/kunoichi-lemon-royale-v3-32K-7B-GGUF](https://huggingface.co/grimjim/kunoichi-lemon-royale-v3-32K-7B-GGUF)
24
+
25
+ ## Merge Details
26
+ ### Merge Method
27
+
28
+ This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [grimjim/kunoichi-lemon-royale-v2-32K-7B](https://huggingface.co/grimjim/kunoichi-lemon-royale-v2-32K-7B) as a base.
29
+
30
+ ### Models Merged
31
+
32
+ The following model was also included in the merge:
33
+ * [grimjim/rogue-enchantress-32k-7B](https://huggingface.co/grimjim/rogue-enchantress-32k-7B)
34
+
35
+ ### Configuration
36
+
37
+ The following YAML configuration was used to produce this model:
38
+
39
+ ```yaml
40
+ base_model: grimjim/kunoichi-lemon-royale-v2-32K-7B
41
+ dtype: bfloat16
42
+ merge_method: task_arithmetic
43
+ slices:
44
+ - sources:
45
+ - layer_range: [0, 32]
46
+ model: grimjim/kunoichi-lemon-royale-v2-32K-7B
47
+ - layer_range: [0, 32]
48
+ model: grimjim/rogue-enchantress-32k-7B
49
+ parameters:
50
+ weight: 0.02
51
+
52
+ ```
kunoichi-lemon-royale-v3-32K-7B.Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6cddc2118f034bbef68554daff3bd385a30853138bb3524f895e9a8e47359ee6
3
+ size 4368439616
kunoichi-lemon-royale-v3-32K-7B.Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0ab07da7dce842aba247802dada8f625b7d008ca0eabbcbc499fb35f081587c2
3
+ size 5131409728
kunoichi-lemon-royale-v3-32K-7B.Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bbf3065df79ad54b215dca474a055ed1ae79f0399e26812ce641de8ace7eb484
3
+ size 5942065472
kunoichi-lemon-royale-v3-32K-7B.Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:265b436a8bfd5af113895d67a80bfc1dd6d0244afe90b4a940252a8558571199
3
+ size 7695857984