aashish1904 commited on
Commit
f64c67a
·
verified ·
1 Parent(s): 030a8b4

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ ---
3
+
4
+ base_model:
5
+ - Qwen/Qwen2.5-14B-Instruct
6
+ - Lambent/qwen2.5-lumen-rebased-14B
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+
12
+
13
+ ---
14
+
15
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
16
+
17
+
18
+ # QuantFactory/qwen2.5-reinstruct-alternate-lumen-14B-GGUF
19
+ This is quantized version of [Lambent/qwen2.5-reinstruct-alternate-lumen-14B](https://huggingface.co/Lambent/qwen2.5-reinstruct-alternate-lumen-14B) created using llama.cpp
20
+
21
+ # Original Model Card
22
+
23
+ # qwenreinstruct
24
+
25
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
26
+
27
+ ## Merge Details
28
+
29
+ Extracted an approximate LoRA of v000000/Qwen2.5-Lumen-14B, rank 128 difference between that and Instruct,
30
+ and first applied this to Lambent/qwen2.5-14B-alternate-instruct-slerp which had no issues with EQ-Bench.
31
+
32
+ Then, here, re-applied a density and weight of original Instruct which in previous merges gave me no issues with EQ-Bench.
33
+
34
+ This one has EQ-Bench of 77.6713 and no "emotions don't match reference error" (if possibly still one not parsed).
35
+ This is similar to Lumen and original Instruct and slightly exceeds both (within margin of error).
36
+ My hope is that it has healed Instruct somewhat and regained its intelligence.
37
+
38
+ ### Merge Method
39
+
40
+ This model was merged using the della merge method using [Lambent/qwen2.5-lumen-rebased-14B](https://huggingface.co/Lambent/qwen2.5-lumen-rebased-14B) as a base.
41
+
42
+ ### Models Merged
43
+
44
+ The following models were included in the merge:
45
+ * [Qwen/Qwen2.5-14B-Instruct](https://huggingface.co/Qwen/Qwen2.5-14B-Instruct)
46
+
47
+ ### Configuration
48
+
49
+ The following YAML configuration was used to produce this model:
50
+
51
+ ```yaml
52
+ models:
53
+ - model: Qwen/Qwen2.5-14B-Instruct
54
+ parameters:
55
+ weight: 0.3
56
+ density: 0.4
57
+ merge_method: della
58
+ base_model: Lambent/qwen2.5-lumen-rebased-14B
59
+ parameters:
60
+ epsilon: 0.05
61
+ lambda: 1
62
+ dtype: bfloat16
63
+ tokenizer_source: base
64
+
65
+
66
+ ```
67
+