kyujinpy commited on
Commit
7f45a1e
1 Parent(s): a347647

Upload README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -2,12 +2,12 @@
2
  language:
3
  - en
4
  datasets:
5
- - Intel/orca_dpo_pairs
6
  pipeline_tag: text-generation
7
  license: cc-by-nc-sa-4.0
8
  ---
9
 
10
- # **Sakura-SOLRCA-Instruct-DPO**
11
  <img src='./sakura.png' width=512>
12
 
13
  **(주)미디어그룹사람과숲과 (주)마커의 LLM 연구 컨소시엄에서 개발된 모델입니다**
@@ -18,7 +18,7 @@ license: cc-by-nc-sa-4.0
18
 
19
  **Method**
20
  Using DPO method.
21
- With [Intel/orca_dpo_pairs](https://huggingface.co/datasets/Intel/orca_dpo_pairs).
22
 
23
  I shared the information about my model. (training and code)
24
  Please see: ⭐[Sakura-SOLAR](https://github.com/KyujinHan/Sakura-SOLAR-DPO).
@@ -30,7 +30,6 @@ Please see: ⭐[Sakura-SOLAR](https://github.com/KyujinHan/Sakura-SOLAR-DPO).
30
 
31
  | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
32
  | --- | --- | --- | --- | --- | --- | --- | --- |
33
- | Sakura-SOLRCA-Instruct-DPO | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
34
  | Sakura-SOLAR-Instruct-DPO-v2 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
35
  | Sakura-SOLAR-Instruct-DPO-v1 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
36
  | [kyujinpy/Sakura-SOLAR-Instruct](https://huggingface.co/kyujinpy/Sakura-SOLAR-Instruct) | 74.40 | 70.99 | 88.42 | 66.33 | 71.79 | 83.66 | 65.20
@@ -42,7 +41,7 @@ Please see: ⭐[Sakura-SOLAR](https://github.com/KyujinHan/Sakura-SOLAR-DPO).
42
  from transformers import AutoModelForCausalLM, AutoTokenizer
43
  import torch
44
 
45
- repo = "kyujinpy/Sakura-SOLRCA-Instruct-DPO"
46
  OpenOrca = AutoModelForCausalLM.from_pretrained(
47
  repo,
48
  return_dict=True,
 
2
  language:
3
  - en
4
  datasets:
5
+ - argilla/distilabel-math-preference-dpo
6
  pipeline_tag: text-generation
7
  license: cc-by-nc-sa-4.0
8
  ---
9
 
10
+ # **Sakura-SOLAR-Instruct-DPO-v2**
11
  <img src='./sakura.png' width=512>
12
 
13
  **(주)미디어그룹사람과숲과 (주)마커의 LLM 연구 컨소시엄에서 개발된 모델입니다**
 
18
 
19
  **Method**
20
  Using DPO method.
21
+ With [argilla/distilabel-math-preference-dpo](https://huggingface.co/datasets/argilla/distilabel-math-preference-dpo).
22
 
23
  I shared the information about my model. (training and code)
24
  Please see: ⭐[Sakura-SOLAR](https://github.com/KyujinHan/Sakura-SOLAR-DPO).
 
30
 
31
  | Model | Average | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
32
  | --- | --- | --- | --- | --- | --- | --- | --- |
 
33
  | Sakura-SOLAR-Instruct-DPO-v2 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
34
  | Sakura-SOLAR-Instruct-DPO-v1 | NaN | NaN | NaN | NaN | NaN | NaN | NaN |
35
  | [kyujinpy/Sakura-SOLAR-Instruct](https://huggingface.co/kyujinpy/Sakura-SOLAR-Instruct) | 74.40 | 70.99 | 88.42 | 66.33 | 71.79 | 83.66 | 65.20
 
41
  from transformers import AutoModelForCausalLM, AutoTokenizer
42
  import torch
43
 
44
+ repo = "kyujinpy/Sakura-SOLAR-Instruct-DPO-v2"
45
  OpenOrca = AutoModelForCausalLM.from_pretrained(
46
  repo,
47
  return_dict=True,