LoneStriker commited on
Commit
4f52793
1 Parent(s): c723973

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -1,35 +1,9 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ Mixtral_7Bx5_MoE_30B-Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
2
+ Mixtral_7Bx5_MoE_30B-Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
3
+ Mixtral_7Bx5_MoE_30B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
4
+ Mixtral_7Bx5_MoE_30B-Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
5
+ Mixtral_7Bx5_MoE_30B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
6
+ Mixtral_7Bx5_MoE_30B-Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
7
+ Mixtral_7Bx5_MoE_30B-Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
8
+ Mixtral_7Bx5_MoE_30B-Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
9
+ Mixtral_7Bx5_MoE_30B-Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
Mixtral_7Bx5_MoE_30B-Q3_K_L.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85fcfae8644f132cb3abd4963b3a11f09ab3ab6d8eda12baf24188aa6adef4dd
3
+ size 15449972384
Mixtral_7Bx5_MoE_30B-Q3_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c33a0231525e53f7c29943bed49d8cd509f47ec8e0e36ec54b9b4476eea7672
3
+ size 14266130080
Mixtral_7Bx5_MoE_30B-Q3_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5c47923f52d41bc39de7659609e4f97827fbe09615d3a3a0e13d61369d17ea0
3
+ size 12854746784
Mixtral_7Bx5_MoE_30B-Q4_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b4a9e512db5d76606dc493a19d371a0b829e698e481fee2a3eb635732eb7eaf
3
+ size 18022235808
Mixtral_7Bx5_MoE_30B-Q4_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16aa649e3b369b6b8d1320a246a5bbbc6ccea4712d513ddf758192e2d2650802
3
+ size 16942726816
Mixtral_7Bx5_MoE_30B-Q5_K_M.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1558080dc19db67c8e6cef6fa0063438f45419c4d01fbfdc4cff5bccb0359b4
3
+ size 21134016160
Mixtral_7Bx5_MoE_30B-Q5_K_S.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e04577c0ba55eee1d1ab9a9d5ad8b25d12099f74e438461b5c0523cfd97b544
3
+ size 20501200544
Mixtral_7Bx5_MoE_30B-Q6_K.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d35896fa06d17244ec0e6a16685bfafe3739d2e76b59b1a15b20ada3224569d5
3
+ size 24440282784
Mixtral_7Bx5_MoE_30B-Q8_0.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c698e5578ae5d00af934d7196b5ada5beab7dbdc168ffb6fd671a8114a0f9811
3
+ size 31655059104
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: cc-by-nc-4.0
3
+ ---
4
+
5
+ # Mixtral MOE 5x7B
6
+
7
+ MoE of the following models :
8
+
9
+ * [Toten5/Marcoroni-neural-chat-7B-v1](https://huggingface.co/Toten5/Marcoroni-neural-chat-7B-v1)
10
+ * [NurtureAI/neural-chat-7b-v3-16k](https://huggingface.co/NurtureAI/neural-chat-7b-v3-16k)
11
+ * [mncai/mistral-7b-dpo-v6](https://huggingface.co/mncai/mistral-7b-dpo-v6)
12
+ * [cookinai/CatMacaroni-Slerp](https://huggingface.co/cookinai/CatMacaroni-Slerp)
13
+ * [ignos/Mistral-T5-7B-v1](https://huggingface.co/ignos/Mistral-T5-7B-v1)
14
+
15
+
16
+ gpu code example
17
+
18
+ ```
19
+ import torch
20
+ from transformers import AutoTokenizer, AutoModelForCausalLM
21
+ import math
22
+
23
+ ## v2 models
24
+ model_path = "cloudyu/Mixtral_7Bx5_MoE_30B"
25
+
26
+ tokenizer = AutoTokenizer.from_pretrained(model_path, use_default_system_prompt=False)
27
+ model = AutoModelForCausalLM.from_pretrained(
28
+ model_path, torch_dtype=torch.float32, device_map='auto',local_files_only=False, load_in_4bit=True
29
+ )
30
+ print(model)
31
+ prompt = input("please input prompt:")
32
+ while len(prompt) > 0:
33
+ input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to("cuda")
34
+
35
+ generation_output = model.generate(
36
+ input_ids=input_ids, max_new_tokens=500,repetition_penalty=1.2
37
+ )
38
+ print(tokenizer.decode(generation_output[0]))
39
+ prompt = input("please input prompt:")
40
+ ```
41
+
42
+ CPU example
43
+
44
+ ```
45
+ import torch
46
+ from transformers import AutoTokenizer, AutoModelForCausalLM
47
+ import math
48
+
49
+ ## v2 models
50
+ model_path = "cloudyu/Mixtral_7Bx5_MoE_30B"
51
+
52
+ tokenizer = AutoTokenizer.from_pretrained(model_path, use_default_system_prompt=False)
53
+ model = AutoModelForCausalLM.from_pretrained(
54
+ model_path, torch_dtype=torch.float32, device_map='cpu',local_files_only=False
55
+ )
56
+ print(model)
57
+ prompt = input("please input prompt:")
58
+ while len(prompt) > 0:
59
+ input_ids = tokenizer(prompt, return_tensors="pt").input_ids
60
+
61
+ generation_output = model.generate(
62
+ input_ids=input_ids, max_new_tokens=500,repetition_penalty=1.2
63
+ )
64
+ print(tokenizer.decode(generation_output[0]))
65
+ prompt = input("please input prompt:")
66
+
67
+ ```