koesn commited on
Commit
ac24db1
1 Parent(s): 9d86f8c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +74 -0
README.md CHANGED
@@ -1,3 +1,77 @@
1
  ---
2
  license: apache-2.0
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: apache-2.0
3
  ---
4
+ # Pastiche-Crown-Clown-7B-dare
5
+
6
+ ## Description
7
+ This repo contains GGUF format model files for Pastiche-Crown-Clown-7B-dare.
8
+
9
+ ## Files Provided
10
+ | Name | Quant | Bits | File Size | Remark |
11
+ | ----------------------------------------- | ------- | ---- | --------- | -------------------------------- |
12
+ | pastiche-crown-clown-7b-dare.IQ3_XXS.gguf | IQ3_XXS | 3 | 3.02 GB | 3.06 bpw quantization |
13
+ | pastiche-crown-clown-7b-dare.IQ3_S.gguf | IQ3_S | 3 | 3.18 GB | 3.44 bpw quantization |
14
+ | pastiche-crown-clown-7b-dare.IQ3_M.gguf | IQ3_M | 3 | 3.28 GB | 3.66 bpw quantization mix |
15
+ | pastiche-crown-clown-7b-dare.Q4_0.gguf | Q4_0 | 4 | 4.11 GB | 3.56G, +0.2166 ppl |
16
+ | pastiche-crown-clown-7b-dare.IQ4_NL.gguf | IQ4_NL | 4 | 4.16 GB | 4.25 bpw non-linear quantization |
17
+ | pastiche-crown-clown-7b-dare.Q4_K_M.gguf | Q4_K_M | 4 | 4.37 GB | 3.80G, +0.0532 ppl |
18
+ | pastiche-crown-clown-7b-dare.Q5_K_M.gguf | Q5_K_M | 5 | 5.13 GB | 4.45G, +0.0122 ppl |
19
+ | pastiche-crown-clown-7b-dare.Q6_K.gguf | Q6_K | 6 | 5.94 GB | 5.15G, +0.0008 ppl |
20
+ | pastiche-crown-clown-7b-dare.Q8_0.gguf | Q8_0 | 8 | 7.70 GB | 6.70G, +0.0004 ppl |
21
+
22
+ ## Parameters
23
+ | path | type | architecture | rope_theta | sliding_win | max_pos_embed |
24
+ | ------------------------------------------ | ------- | ------------------ | ---------- | ----------- | ------------- |
25
+ | CorticalStack/pastiche-crown-clown-7b-dare | mistral | MistralForCausalLM | 10000.0 | 4096 | 32768 |
26
+
27
+ ## Benchmarks
28
+ ![](https://i.ibb.co/Srwfpj3/pastiche-crown-clown-7b-dare.png)
29
+
30
+ # Original Model Card
31
+
32
+ ---
33
+ license: apache-2.0
34
+ tags:
35
+ - merge
36
+ - mergekit
37
+ - bardsai/jaskier-7b-dpo-v5.6
38
+ - mlabonne/AlphaMonarch-7B
39
+ - mlabonne/NeuralMonarch-7B
40
+ - macadeliccc/MBX-7B-v3-DPO
41
+ ---
42
+
43
+ <img src="pastiche-crown-clown.png" alt="Pastiche crown clown logo" width="800" style="margin-left:'auto' margin-right:'auto' display:'block'"/>
44
+
45
+ # pastiche-crown-clown-7B-dare
46
+
47
+ pastiche-crown-clown-7B-dare is a DARE merge of the following models using [mergekit](https://github.com/cg123/mergekit):
48
+ * [bardsai/jaskier-7b-dpo-v5.6](https://huggingface.co/bardsai/jaskier-7b-dpo-v5.6)
49
+ * [mlabonne/AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B)
50
+ * [mlabonne/NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B)
51
+ * [macadeliccc/MBX-7B-v3-DPO](https://huggingface.co/macadeliccc/MBX-7B-v3-DPO)
52
+
53
+ See the paper [Language Models are Super Mario: Absorbing Abilities from Homologous Models as a Free Lunch](https://arxiv.org/abs/2311.03099) for more on the method.
54
+
55
+ ## 🧩 Configuration
56
+
57
+ ```yaml
58
+ models:
59
+ - model: bardsai/jaskier-7b-dpo-v5.6
60
+ - model: mlabonne/AlphaMonarch-7B
61
+ parameters:
62
+ density: 0.53
63
+ weight: 0.2
64
+ - model: mlabonne/NeuralMonarch-7B
65
+ parameters:
66
+ density: 0.53
67
+ weight: 0.4
68
+ - model: macadeliccc/MBX-7B-v3-DPO
69
+ parameters:
70
+ density: 0.53
71
+ weight: 0.4
72
+ merge_method: dare_ties
73
+ base_model: bardsai/jaskier-7b-dpo-v5.6
74
+ parameters:
75
+ int8_mask: true
76
+ dtype: bfloat16
77
+ ```