seungduk commited on
Commit
6c9d44b
1 Parent(s): f56998c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +39 -7
README.md CHANGED
@@ -13,6 +13,14 @@ model-index:
13
 
14
  If you're passionate about the field of Large Language Models and wish to exchange knowledge and insights, we warmly invite you to join our Discord server. It's worth noting that Korean is the primary language used in this server. The landscape of LLM is evolving rapidly, and without active sharing, our collective knowledge risks becoming outdated swiftly. Let's collaborate and drive greater impact together! Join us here: [Discord Link](https://discord.gg/b27bAHg95m).
15
 
 
 
 
 
 
 
 
 
16
  ## About the Model
17
 
18
  This model is a fine-tuned version of [yanolja/KoSOLAR-10.7B-v0.2](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.2), which is a Korean vocabulary-extended version of [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0). Specifically, we employed Direct Preference Optimization (DPO) based on [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory).
@@ -22,10 +30,34 @@ This model is a fine-tuned version of [yanolja/KoSOLAR-10.7B-v0.2](https://huggi
22
  - Korean-translated version of [argilla/ultrafeedback-binarized-preferences-cleaned](https://huggingface.co/datasets/argilla/ultrafeedback-binarized-preferences-cleaned)
23
  - No other dataset was used
24
 
25
- ## Our Dedicated Team
26
- | Research | Engineering | Product Management |
27
- |-----------------|-----------------|--------------------|
28
- | Myeongho Jeong | Sanghoon Han | Bokyung Huh |
29
- | Seungtaek Choi | Suhyun Kang | |
30
- | Seungduk Kim | Rifqi Alfi | |
31
- | | Geon Kim | |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
  If you're passionate about the field of Large Language Models and wish to exchange knowledge and insights, we warmly invite you to join our Discord server. It's worth noting that Korean is the primary language used in this server. The landscape of LLM is evolving rapidly, and without active sharing, our collective knowledge risks becoming outdated swiftly. Let's collaborate and drive greater impact together! Join us here: [Discord Link](https://discord.gg/b27bAHg95m).
15
 
16
+ ## Our Dedicated Team (Alphabetical Order)
17
+ | Research | Engineering | Product Management | UX Design |
18
+ |-----------------|-----------------|--------------------|--------------
19
+ | Myeongho Jeong | Geon Kim | Bokyung Huh | Eunsue Choi |
20
+ | Seungduk Kim | Rifqi Alfi | | |
21
+ | Seungtaek Choi | Sanghoon Han | | |
22
+ | | Suhyun Kang | | |
23
+
24
  ## About the Model
25
 
26
  This model is a fine-tuned version of [yanolja/KoSOLAR-10.7B-v0.2](https://huggingface.co/yanolja/KoSOLAR-10.7B-v0.2), which is a Korean vocabulary-extended version of [upstage/SOLAR-10.7B-v1.0](https://huggingface.co/upstage/SOLAR-10.7B-v1.0). Specifically, we employed Direct Preference Optimization (DPO) based on [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory).
 
30
  - Korean-translated version of [argilla/ultrafeedback-binarized-preferences-cleaned](https://huggingface.co/datasets/argilla/ultrafeedback-binarized-preferences-cleaned)
31
  - No other dataset was used
32
 
33
+ ## Citation
34
+
35
+ ```
36
+ @misc{cui2023ultrafeedback,
37
+ title={UltraFeedback: Boosting Language Models with High-quality Feedback},
38
+ author={Ganqu Cui and Lifan Yuan and Ning Ding and Guanming Yao and Wei Zhu and Yuan Ni and Guotong Xie and Zhiyuan Liu and Maosong Sun},
39
+ year={2023},
40
+ eprint={2310.01377},
41
+ archivePrefix={arXiv},
42
+ primaryClass={cs.CL}
43
+ }
44
+ ```
45
+ ```
46
+ @misc{SlimOrcaDedup,
47
+ title = {SlimOrca Dedup: A Deduplicated Subset of SlimOrca},
48
+ author = {Wing Lian and Guan Wang and Bleys Goodson and Eugene Pentland and Austin Cook and Chanvichet Vong and "Teknium" and Nathan Hoos},
49
+ year = {2023},
50
+ publisher = {HuggingFace},
51
+ url = {https://huggingface.co/datasets/Open-Orca/SlimOrca-Dedup/}
52
+ }
53
+ ```
54
+ ```
55
+ @misc{mukherjee2023orca,
56
+ title={Orca: Progressive Learning from Complex Explanation Traces of GPT-4},
57
+ author={Subhabrata Mukherjee and Arindam Mitra and Ganesh Jawahar and Sahaj Agarwal and Hamid Palangi and Ahmed Awadallah},
58
+ year={2023},
59
+ eprint={2306.02707},
60
+ archivePrefix={arXiv},
61
+ primaryClass={cs.CL}
62
+ }
63
+ ```