LoneStriker commited on
Commit
d4752bf
β€’
1 Parent(s): 01a157f

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +82 -0
README.md ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - ko
5
+ - en
6
+ tags:
7
+ - moe
8
+ ---
9
+
10
+ # **Synatra-Mixtral-8x7B**
11
+ <img src="./Synatra-Mixtral.png" alt="Synatra-Mixtral-8x7B" width="512"/>
12
+
13
+
14
+ **Synatra-Mixtral-8x7B** is a fine-tuned version of the Mixtral-8x7B-Instruct-v0.1 model using **Korean** datasets.
15
+
16
+ This model features overwhelmingly superior comprehension and inference capabilities and is licensed under apache-2.0.
17
+
18
+ # **Join Our Discord**
19
+
20
+ [Server Link](https://discord.gg/MrBt3PXdXc)
21
+
22
+ # **License**
23
+
24
+ **OPEN**, Apache-2.0.
25
+
26
+ # **Model Details**
27
+
28
+ **Base Model**
29
+ [mistralai/Mixtral-8x7B-Instruct-v0.1](https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1)
30
+
31
+ **Trained On**
32
+ A100 80GB * 6
33
+
34
+ **Instruction format**
35
+
36
+ It follows **Alpaca** format.
37
+ ```
38
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
39
+
40
+ ### Instruction:
41
+ {input}
42
+
43
+ ### Response:
44
+ {output}
45
+ ```
46
+
47
+ # **Model Benchmark**
48
+ TBD
49
+
50
+ # **Implementation Code**
51
+
52
+ ```python
53
+ from transformers import AutoModelForCausalLM, AutoTokenizer
54
+
55
+ device = "cuda" # the device to load the model onto
56
+
57
+ model = AutoModelForCausalLM.from_pretrained("maywell/Synatra-Mixtral-8x7B")
58
+ tokenizer = AutoTokenizer.from_pretrained("maywell/Synatra-Mixtral-8x7B")
59
+
60
+ messages = [
61
+ {"role": "user", "content": "μ•„μΈμŠˆνƒ€μΈμ˜ μƒλŒ€μ„±μ΄λ‘ μ— λŒ€ν•΄μ„œ μžμ„Ένžˆ μ„€λͺ…ν•΄μ€˜."},
62
+ ]
63
+
64
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
65
+
66
+ model_inputs = encodeds.to(device)
67
+ model.to(device)
68
+
69
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
70
+ decoded = tokenizer.batch_decode(generated_ids)
71
+ print(decoded[0])
72
+ ```
73
+
74
+ # **Author's Message**
75
+
76
+ This model's training got sponsered by no one but support from people around Earth.
77
+
78
+ [Support Me](https://www.buymeacoffee.com/mwell)
79
+
80
+ Contact Me on Discord - **is.maywell**
81
+
82
+ Follow me on twitter: https://twitter.com/stablefluffy