wolfram commited on
Commit
6d89d59
1 Parent(s): f1538db

Update README.md

Browse files

Finalized README.

Files changed (1) hide show
  1. README.md +12 -13
README.md CHANGED
@@ -12,13 +12,13 @@ tags:
12
  - mergekit
13
  - merge
14
  ---
15
- # miqu-1-120b-GGUF
16
 
17
  ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6303ca537373aacccd85d8a7/LxO9j7OykuabKLYQHIodG.jpeg)
18
 
19
- - EXL2: [2.4bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2) | [2.65bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2) | [3.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2) | [4.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2) | [5.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2)
20
- - **More** GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/)
21
- - HF FP16: [wolfram/miqu-1-120b](https://huggingface.co/wolfram/miqu-1-120b)
22
 
23
  This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
24
 
@@ -38,8 +38,8 @@ See also: [🐺🐦‍⬛ LLM Prompt Format Comparison/Test: Mixtral 8x7B Instru
38
 
39
  ## Model Details
40
 
41
- - Max Context: 32764 tokens (kept the weird number from the original/base model)
42
- - Layers: 140
43
 
44
  ## Merge Details
45
 
@@ -85,15 +85,14 @@ slices:
85
 
86
  ## Credits & Special Thanks
87
 
88
- - original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
89
- - leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
90
- - f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
91
- - mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
92
- - mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
93
- - gguf quantization: [ggerganov/llama.cpp: Port of Facebook's LLaMA model in C/C++](https://github.com/ggerganov/llama.cpp)
94
 
95
  ### Support
96
 
97
- - [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
98
 
99
  #### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.
 
12
  - mergekit
13
  - merge
14
  ---
15
+ # miqu-1-120b
16
 
17
  ![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/6303ca537373aacccd85d8a7/LxO9j7OykuabKLYQHIodG.jpeg)
18
 
19
+ * EXL2: [2.4bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.4bpw-h6-exl2) | [2.65bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-2.65bpw-h6-exl2) | [3.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-3.0bpw-h6-exl2) | [4.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-4.0bpw-h6-exl2) | [5.0bpw](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-5.0bpw-h6-exl2)
20
+ * **More** GGUF: [Q2_K-Q5_K_M](https://huggingface.co/LoneStriker/wolfram_miqu-1-120b-GGUF/)
21
+ * HF FP16: [wolfram/miqu-1-120b](https://huggingface.co/wolfram/miqu-1-120b)
22
 
23
  This is a 120b frankenmerge of [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b) created by interleaving layers of [miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf) with itself using [mergekit](https://github.com/cg123/mergekit).
24
 
 
38
 
39
  ## Model Details
40
 
41
+ * Max Context: 32764 tokens (kept the weird number from the original/base model)
42
+ * Layers: 140
43
 
44
  ## Merge Details
45
 
 
85
 
86
  ## Credits & Special Thanks
87
 
88
+ * original (unreleased) model: [mistralai (Mistral AI_)](https://huggingface.co/mistralai)
89
+ * leaked model: [miqudev/miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b)
90
+ * f16 model: [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
91
+ * mergekit: [arcee-ai/mergekit: Tools for merging pretrained large language models.](https://github.com/arcee-ai/mergekit)
92
+ * mergekit_config.yml: [nsfwthrowitaway69/Venus-120b-v1.2](https://huggingface.co/nsfwthrowitaway69/Venus-120b-v1.2)
 
93
 
94
  ### Support
95
 
96
+ * [My Ko-fi page](https://ko-fi.com/wolframravenwolf) if you'd like to tip me to say thanks or request specific models to be tested or merged with priority. Also consider supporting your favorite model creators, quantizers, or frontend/backend devs if you can afford to do so. They deserve it!
97
 
98
  #### DISCLAIMER: THIS IS [BASED ON A LEAKED ASSET](https://huggingface.co/miqudev/miqu-1-70b/discussions/10) AND HAS NO LICENSE ASSOCIATED WITH IT. USE AT YOUR OWN RISK.