wolfram commited on
Commit
da7a544
1 Parent(s): bb4b39a

Upload folder using huggingface_hub

Browse files
Files changed (3) hide show
  1. .gitattributes +1 -0
  2. README.md +93 -0
  3. miqu-1-120b.IQ3_XXS.gguf +3 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ miqu-1-120b.IQ3_XXS.gguf filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - 152334H/miqu-1-70b-sf
4
+ language:
5
+ - en
6
+ - de
7
+ - fr
8
+ - es
9
+ - it
10
+ library_name: transformers
11
+ tags:
12
+ - mergekit
13
+ - merge
14
+ ---
15
+ # miqu-1-120b
16
+
17
+ ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6303ca537373aacccd85d8a7/LxO9j7OykuabKLYQHIodG.jpeg)
18
+
19
+ This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
20
+
21
+ Inspired by [Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2), [MegaDolphin-120b](https://huggingface.co/cognitivecomputations/MegaDolphin-120b), and [goliath-120b](https://huggingface.co/alpindale/goliath-120b).
22
+
23
+ Thanks for the support, [CopilotKit](https://github.com/CopilotKit/CopilotKit) - the open-source platform for building in-app AI Copilots into any product, with any LLM model. Check out their GitHub.
24
+
25
+ ## Prompt template: Mistral
26
+
27
+ ```
28
+ <s>[INST] {prompt} [/INST]
29
+ ```
30
+
31
+ See also: [🐺🐦‍⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instruct with **17** different instruct templates : LocalLLaMA](https://www.reddit.com/r/LocalLLaMA/comments/18ljvxb/llm_prompt_format_comparisontest_mixtral_8x7b/)
32
+
33
+ ## Model Details
34
+
35
+ * Max Context: 32764 tokens (kept the weird number from the original/base model)
36
+ * Layers: 140
37
+
38
+ ## Merge Details
39
+
40
+ ### Merge Method
41
+
42
+ This model was merged using the passthrough merge method.
43
+
44
+ ### Models Merged
45
+
46
+ The following models were included in the merge:
47
+ * [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
48
+
49
+ ### Configuration
50
+
51
+ The following YAML configuration was used to produce this model:
52
+
53
+ ```yaml
54
+ dtype: float16
55
+ merge_method: passthrough
56
+ slices:
57
+ - sources:
58
+ - layer_range: [0, 20]
59
+ model: 152334H/miqu-1-70b-sf
60
+ - sources:
61
+ - layer_range: [10, 30]
62
+ model: 152334H/miqu-1-70b-sf
63
+ - sources:
64
+ - layer_range: [20, 40]
65
+ model: 152334H/miqu-1-70b-sf
66
+ - sources:
67
+ - layer_range: [30, 50]
68
+ model: 152334H/miqu-1-70b-sf
69
+ - sources:
70
+ - layer_range: [40, 60]
71
+ model: 152334H/miqu-1-70b-sf
72
+ - sources:
73
+ - layer_range: [50, 70]
74
+ model: 152334H/miqu-1-70b-sf
75
+ - sources:
76
+ - layer_range: [60, 80]
77
+ model: 152334H/miqu-1-70b-sf
78
+ ```
79
+
80
+ ## Credits & Special Thanks
81
+
82
+ * original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
83
+ * leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
84
+ * f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
85
+ * mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
86
+ * mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
87
+ - gguf quantization: [ggerganov/llama.cpp: Port of Facebook's LLaMA model in C/C++](https://github.com/ggerganov/llama.cpp)
88
+
89
+ ### Support
90
+
91
+ * [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
92
+
93
+ #### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
miqu-1-120b.IQ3_XXS.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:084229d79fe5d836738c3666c5734f937b0bb699cec34f1bc7cf75435944e395
3
+ size 46880781760