RichardErkhov commited on
Commit
70718a2
β€’
1 Parent(s): 8516d1b

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +157 -0
README.md ADDED
@@ -0,0 +1,157 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ DataVortexS-10.7B-dpo-v1.2 - GGUF
11
+ - Model creator: https://huggingface.co/Edentns/
12
+ - Original model: https://huggingface.co/Edentns/DataVortexS-10.7B-dpo-v1.2/
13
+
14
+
15
+ | Name | Quant method | Size |
16
+ | ---- | ---- | ---- |
17
+ | [DataVortexS-10.7B-dpo-v1.2.Q2_K.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q2_K.gguf) | Q2_K | 3.73GB |
18
+ | [DataVortexS-10.7B-dpo-v1.2.IQ3_XS.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.IQ3_XS.gguf) | IQ3_XS | 4.14GB |
19
+ | [DataVortexS-10.7B-dpo-v1.2.IQ3_S.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.IQ3_S.gguf) | IQ3_S | 4.37GB |
20
+ | [DataVortexS-10.7B-dpo-v1.2.Q3_K_S.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q3_K_S.gguf) | Q3_K_S | 4.34GB |
21
+ | [DataVortexS-10.7B-dpo-v1.2.IQ3_M.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.IQ3_M.gguf) | IQ3_M | 4.51GB |
22
+ | [DataVortexS-10.7B-dpo-v1.2.Q3_K.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q3_K.gguf) | Q3_K | 4.84GB |
23
+ | [DataVortexS-10.7B-dpo-v1.2.Q3_K_M.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q3_K_M.gguf) | Q3_K_M | 4.84GB |
24
+ | [DataVortexS-10.7B-dpo-v1.2.Q3_K_L.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q3_K_L.gguf) | Q3_K_L | 5.26GB |
25
+ | [DataVortexS-10.7B-dpo-v1.2.IQ4_XS.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.IQ4_XS.gguf) | IQ4_XS | 5.43GB |
26
+ | [DataVortexS-10.7B-dpo-v1.2.Q4_0.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q4_0.gguf) | Q4_0 | 5.66GB |
27
+ | [DataVortexS-10.7B-dpo-v1.2.IQ4_NL.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.IQ4_NL.gguf) | IQ4_NL | 5.72GB |
28
+ | [DataVortexS-10.7B-dpo-v1.2.Q4_K_S.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q4_K_S.gguf) | Q4_K_S | 5.7GB |
29
+ | [DataVortexS-10.7B-dpo-v1.2.Q4_K.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q4_K.gguf) | Q4_K | 6.02GB |
30
+ | [DataVortexS-10.7B-dpo-v1.2.Q4_K_M.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q4_K_M.gguf) | Q4_K_M | 6.02GB |
31
+ | [DataVortexS-10.7B-dpo-v1.2.Q4_1.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q4_1.gguf) | Q4_1 | 6.27GB |
32
+ | [DataVortexS-10.7B-dpo-v1.2.Q5_0.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q5_0.gguf) | Q5_0 | 6.89GB |
33
+ | [DataVortexS-10.7B-dpo-v1.2.Q5_K_S.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q5_K_S.gguf) | Q5_K_S | 6.89GB |
34
+ | [DataVortexS-10.7B-dpo-v1.2.Q5_K.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q5_K.gguf) | Q5_K | 7.08GB |
35
+ | [DataVortexS-10.7B-dpo-v1.2.Q5_K_M.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q5_K_M.gguf) | Q5_K_M | 7.08GB |
36
+ | [DataVortexS-10.7B-dpo-v1.2.Q5_1.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q5_1.gguf) | Q5_1 | 7.51GB |
37
+ | [DataVortexS-10.7B-dpo-v1.2.Q6_K.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q6_K.gguf) | Q6_K | 8.2GB |
38
+ | [DataVortexS-10.7B-dpo-v1.2.Q8_0.gguf](https://huggingface.co/RichardErkhov/Edentns_-_DataVortexS-10.7B-dpo-v1.2-gguf/blob/main/DataVortexS-10.7B-dpo-v1.2.Q8_0.gguf) | Q8_0 | 10.62GB |
39
+
40
+
41
+
42
+
43
+ Original model description:
44
+ ---
45
+ tags:
46
+ - text-generation
47
+ license: cc-by-nc-sa-4.0
48
+ language:
49
+ - ko
50
+ base_model: megastudy/M-SOLAR-10.7B-v1.3
51
+ pipeline_tag: text-generation
52
+ ---
53
+
54
+ # **DataVortexS-10.7B-dpo-v1.2**
55
+
56
+ <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;">
57
+
58
+ ## Our Team
59
+
60
+ | Research & Engineering | Product Management |
61
+ | :--------------------: | :----------------: |
62
+ | Kwangseok Yang | Seunghyun Choi |
63
+ | Jeongwon Choi | Hyoseok Choi |
64
+
65
+ ## **Model Details**
66
+
67
+ ### **Base Model**
68
+
69
+ [megastudy/M-SOLAR-10.7B-v1.3](https://huggingface.co/megastudy/M-SOLAR-10.7B-v1.3)
70
+
71
+ ### **Trained On**
72
+
73
+ - **OS**: Ubuntu 22.04
74
+ - **GPU**: H100 80GB 4ea
75
+ - **transformers**: v4.36.2
76
+
77
+ ### **Instruction format**
78
+
79
+ It follows **Alpaca (Chat)** format.
80
+
81
+ E.g.
82
+
83
+ ```python
84
+ text = """\
85
+ ### System:
86
+ 당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€.
87
+
88
+ ### User:
89
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?
90
+
91
+ ### Assistant:
92
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€.
93
+
94
+ ### User:
95
+ μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?
96
+ """
97
+ ```
98
+
99
+ ## **Model Benchmark**
100
+
101
+ ### **[Ko LM Eval Harness](https://github.com/Beomi/ko-lm-evaluation-harness)**
102
+
103
+ | Task | 0-shot | 5-shot | 10-shot | 50-shot |
104
+ | :--------------- | -----------: | -----------: | -----------: | ----------: |
105
+ | kobest_boolq | 0.86665 | 0.932254 | 0.940132 | 0.941561 |
106
+ | kobest_copa | 0.723415 | 0.780594 | 0.778814 | 0.79982 |
107
+ | kobest_hellaswag | 0.471639 | 0.466883 | 0.472548 | 0.488648 |
108
+ | kobest_sentineg | 0.78514 | 0.964734 | 0.972281 | 0.972289 |
109
+ | **Average** | **0.711711** | **0.786116** | **0.790944** | **0.80058** |
110
+
111
+ ### **[Ko-LLM-Leaderboard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard)**
112
+
113
+ | Average | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
114
+ | ------: | -----: | -----------: | ------: | ------------: | --------------: |
115
+ | 56.53 | 52.73 | 64.83 | 52.99 | 58.36 | 53.72 |
116
+
117
+ ## **Implementation Code**
118
+
119
+ This model contains the chat_template instruction format.
120
+ You can use the code below.
121
+
122
+ ```python
123
+ from transformers import AutoModelForCausalLM, AutoTokenizer
124
+
125
+ device = "cuda" # the device to load the model onto
126
+
127
+ model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-dpo-v1.2")
128
+ tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-dpo-v1.2")
129
+
130
+ messages = [
131
+ {"role": "system", "content": "당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€."},
132
+ {"role": "user", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?"},
133
+ {"role": "assistant", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€."},
134
+ {"role": "user", "content": "μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?"}
135
+ ]
136
+
137
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
138
+
139
+ model_inputs = encodeds.to(device)
140
+ model.to(device)
141
+
142
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
143
+ decoded = tokenizer.batch_decode(generated_ids)
144
+ print(decoded[0])
145
+ ```
146
+
147
+ ## **License**
148
+
149
+ The model is licensed under the [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) license, which allows others to copy, modify, and share the work non-commercially, as long as they give appropriate credit and distribute any derivative works under the same license.
150
+
151
+ <div align="center">
152
+ <a href="https://edentns.com/">
153
+ <img src="./Logo.png" alt="Logo" style="height: 3em;">
154
+ </a>
155
+ </div>
156
+
157
+