JeongwonChoi commited on
Commit
6e2b3c7
Β·
verified Β·
1 Parent(s): cce246a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -36
README.md CHANGED
@@ -1,77 +1,107 @@
1
  ---
2
  tags:
3
- - text-generation
4
  license: cc-by-nc-sa-4.0
5
  language:
6
- - ko
7
  base_model: yanolja/KoSOLAR-10.7B-v0.1
8
  pipeline_tag: text-generation
 
 
9
  ---
10
 
11
  # **DataVortexS-10.7B-v0.3**
12
- <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;">
13
-
14
- ## **License**
15
 
16
- [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/)
17
 
18
  ## **Model Details**
19
 
20
  ### **Base Model**
21
- [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1)
 
22
 
23
  ### **Trained On**
24
- H100 80GB 1ea
 
 
 
 
 
 
 
25
 
26
  ### **Instruction format**
27
 
28
- It follows **(No Input) Alpaca** format.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29
 
30
  ## **Model Benchmark**
31
 
32
- ### **Ko-LLM-Leaderboard**
 
 
33
 
34
- On Benchmarking...
 
 
 
 
 
 
 
 
 
35
 
36
- # **Implementation Code**
37
 
38
- Since, chat_template already contains insturction format above.
39
  You can use the code below.
40
 
41
  ```python
42
  from transformers import AutoModelForCausalLM, AutoTokenizer
43
 
44
- device = "cuda"
45
 
46
- model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-v0.3", device_map=device)
47
  tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-v0.3")
48
 
49
  messages = [
50
- { "role": "user", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?" }
 
 
 
51
  ]
52
 
53
- encoded = tokenizer.apply_chat_template(
54
- messages,
55
- add_generation_prompt=True,
56
- return_tensors="pt",
57
- return_token_type_ids=False
58
- ).to(device)
59
-
60
- decoded = model.generate(
61
- input_ids=encoded,
62
- temperature=0.2,
63
- top_p=0.9,
64
- repetition_penalty=1.2,
65
- do_sample=True,
66
- max_length=4096,
67
- eos_token_id=tokenizer.eos_token_id,
68
- pad_token_id=tokenizer.eos_token_id
69
- )
70
- decoded = decoded[0][encoded.shape[1]:decoded[0].shape[-1]]
71
- decoded_text = tokenizer.decode(decoded, skip_special_tokens=True)
72
- print(decoded_text)
73
  ```
74
 
 
 
 
 
75
  <div align="center">
76
  <a href="https://edentns.com/">
77
  <img src="./Logo.png" alt="Logo" style="height: 3em;">
 
1
  ---
2
  tags:
3
+ - text-generation
4
  license: cc-by-nc-sa-4.0
5
  language:
6
+ - ko
7
  base_model: yanolja/KoSOLAR-10.7B-v0.1
8
  pipeline_tag: text-generation
9
+ datasets:
10
+ - jojo0217/korean_rlhf_dataset
11
  ---
12
 
13
  # **DataVortexS-10.7B-v0.3**
 
 
 
14
 
15
+ <img src="./DataVortex.png" alt="DataVortex" style="height: 8em;">
16
 
17
  ## **Model Details**
18
 
19
  ### **Base Model**
20
+
21
+ [yanolja/KoSOLAR-10.7B-v0.1](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.1)
22
 
23
  ### **Trained On**
24
+
25
+ - **OS**: Ubuntu 20.04
26
+ - **GPU**: H100 80GB 4ea
27
+ - **transformers**: v4.36.2
28
+
29
+ ### **Dataset**
30
+
31
+ - [jojo0217/korean_rlhf_dataset](https://huggingface.co/datasets/jojo0217/korean_rlhf_dataset)
32
 
33
  ### **Instruction format**
34
 
35
+ It follows **Alpaca** format.
36
+
37
+ E.g.
38
+
39
+ ```python
40
+ text = """\
41
+ 당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€.
42
+
43
+ ### Instruction:
44
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?
45
+
46
+ ### Response:
47
+ λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€.
48
+
49
+ ### Instruction:
50
+ μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?
51
+ """
52
+ ```
53
 
54
  ## **Model Benchmark**
55
 
56
+ ### **[Ko-LLM-Leaderboard](https://huggingface.co/spaces/upstage/open-ko-llm-leaderboard)**
57
+
58
+ On Benchmarking ...
59
 
60
+ | Model | Average | Ko-ARC | Ko-HellaSwag | Ko-MMLU | Ko-TruthfulQA | Ko-CommonGen V2 |
61
+ | ---------------------------- | ------- | ------ | ------------ | ------- | ------------- | --------------- |
62
+ | DataVortexM-7B-Instruct-v0.1 | 39.81 | 34.13 | 42.35 | 38.73 | 45.46 | 38.37 |
63
+ | DataVortexS-10.7B-v0.1 | 0 | 0 | 0 | 0 | 0 | 0 |
64
+ | DataVortexS-10.7B-v0.2 | 0 | 0 | 0 | 0 | 0 | 0 |
65
+ | **DataVortexS-10.7B-v0.3** | **0** | **0** | **0** | **0** | **0** | **0** |
66
+ | DataVortexS-10.7B-v0.4 | 0 | 0 | 0 | 0 | 0 | 0 |
67
+ | DataVortexS-10.7B-v0.5 | 0 | 0 | 0 | 0 | 0 | 0 |
68
+ | DataVortexTL-1.1B-v0.1 | 0 | 0 | 0 | 0 | 0 | 0 |
69
+ | DataVortexS-10.7B-dpo-v0.1 | 0 | 0 | 0 | 0 | 0 | 0 |
70
 
71
+ ## **Implementation Code**
72
 
73
+ This model contains the chat_template instruction format.
74
  You can use the code below.
75
 
76
  ```python
77
  from transformers import AutoModelForCausalLM, AutoTokenizer
78
 
79
+ device = "cuda" # the device to load the model onto
80
 
81
+ model = AutoModelForCausalLM.from_pretrained("Edentns/DataVortexS-10.7B-v0.3")
82
  tokenizer = AutoTokenizer.from_pretrained("Edentns/DataVortexS-10.7B-v0.3")
83
 
84
  messages = [
85
+ {"role": "system", "content": "당신은 μ‚¬λžŒλ“€μ΄ 정보λ₯Ό 찾을 수 μžˆλ„λ‘ λ„μ™€μ£ΌλŠ” 인곡지λŠ₯ λΉ„μ„œμž…λ‹ˆλ‹€."},
86
+ {"role": "user", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ–΄λ””μ•Ό?"},
87
+ {"role": "assistant", "content": "λŒ€ν•œλ―Όκ΅­μ˜ μˆ˜λ„λŠ” μ„œμšΈμž…λ‹ˆλ‹€."},
88
+ {"role": "user", "content": "μ„œμšΈ μΈκ΅¬λŠ” 총 λͺ‡ λͺ…이야?"}
89
  ]
90
 
91
+ encodeds = tokenizer.apply_chat_template(messages, return_tensors="pt")
92
+
93
+ model_inputs = encodeds.to(device)
94
+ model.to(device)
95
+
96
+ generated_ids = model.generate(model_inputs, max_new_tokens=1000, do_sample=True)
97
+ decoded = tokenizer.batch_decode(generated_ids)
98
+ print(decoded[0])
 
 
 
 
 
 
 
 
 
 
 
 
99
  ```
100
 
101
+ ## **License**
102
+
103
+ The model is licensed under the [cc-by-nc-sa-4.0](https://creativecommons.org/licenses/by-nc-sa/4.0/) license, which allows others to copy, modify, and share the work non-commercially, as long as they give appropriate credit and distribute any derivative works under the same license.
104
+
105
  <div align="center">
106
  <a href="https://edentns.com/">
107
  <img src="./Logo.png" alt="Logo" style="height: 3em;">