File size: 3,469 Bytes
808ffb8 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 3b8ec83 2ea066e 808ffb8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 |
---
language: ko
---
# Pretrained BART in Korean
This is pretrained BART model with multiple Korean Datasets.
I used multiple datasets for generalizing the model for both colloquial and written texts.
The training is supported by [TPU Research Cloud](https://sites.research.google/trc/) program.
The script which is used to pre-train model is [here](https://github.com/cosmoquester/transformers-bart-pretrain).
When you use the reference API, you must wrap the sentence with `[BOS]` and `[EOS]` like below example.
```
[BOS] ์๋
ํ์ธ์? ๋ฐ๊ฐ์์~~ [EOS]
```
You can also test mask filling performance using `[MASK]` token like this.
```
[BOS] [MASK] ๋จน์์ด? [EOS]
```
## Benchmark
<table>
<tr>
<th style="text-align:center">Dataset</th>
<td style="text-align:center">KLUE NLI dev</th>
<td style="text-align:center">NSMC test</td>
<td style="text-align:center">QuestionPair test</td>
<td colspan="2" style="text-align:center">KLUE TC dev</td>
<td colspan="3" style="text-align:center">KLUE STS dev</td>
<td colspan="3" style="text-align:center">KorSTS dev</td>
<td colspan="2" style="text-align:center">HateSpeech dev</td>
</tr>
<tr>
<th style="text-align:center">Metric</th>
<!-- KLUE NLI -->
<td style="text-align:center">Acc</th>
<!-- NSMC -->
<td style="text-align:center">Acc</td>
<!-- QuestionPair -->
<td style="text-align:center">Acc</td>
<!-- KLUE TC -->
<td style="text-align:center">Acc</td>
<td style="text-align:center">F1</td>
<!-- KLUE STS -->
<td style="text-align:center">F1</td>
<td style="text-align:center">Pearson</td>
<td style="text-align:center">Spearman</td>
<!-- KorSTS -->
<td style="text-align:center">F1</td>
<td style="text-align:center">Pearson</td>
<td style="text-align:center">Spearman</td>
<!-- HateSpeech -->
<td style="text-align:center">Bias Acc</td>
<td style="text-align:center">Hate Acc</td>
</tr>
<tr>
<th style="text-align:center">Score</th>
<!-- KLUE NLI -->
<td style="text-align:center">0.5253</th>
<!-- NSMC -->
<td style="text-align:center">0.8425</td>
<!-- QuestionPair -->
<td style="text-align:center">0.8945</td>
<!-- KLUE TC -->
<td style="text-align:center">0.8047</td>
<td style="text-align:center">0.7988</td>
<!-- KLUE STS -->
<td style="text-align:center">0.7411</td>
<td style="text-align:center">0.7471</td>
<td style="text-align:center">0.7399</td>
<!-- KorSTS -->
<td style="text-align:center">0.7725</td>
<td style="text-align:center">0.6503</td>
<td style="text-align:center">0.6191</td>
<!-- HateSpeech -->
<td style="text-align:center">0.7537</td>
<td style="text-align:center">0.5605</td>
</tr>
</table>
- The performance was measured using [the notebooks here](https://github.com/cosmoquester/transformers-bart-finetune) with colab.
## Used Datasets
### [๋ชจ๋์ ๋ง๋ญ์น](https://corpus.korean.go.kr/)
- ์ผ์ ๋ํ ๋ง๋ญ์น 2020
- ๊ตฌ์ด ๋ง๋ญ์น
- ๋ฌธ์ด ๋ง๋ญ์น
- ์ ๋ฌธ ๋ง๋ญ์น
### AIhub
- [๊ฐ๋ฐฉ๋ฐ์ดํฐ ์ ๋ฌธ๋ถ์ผ๋ง๋ญ์น](https://aihub.or.kr/aidata/30717)
- [๊ฐ๋ฐฉ๋ฐ์ดํฐ ํ๊ตญ์ด๋ํ์์ฝ](https://aihub.or.kr/aidata/30714)
- [๊ฐ๋ฐฉ๋ฐ์ดํฐ ๊ฐ์ฑ ๋ํ ๋ง๋ญ์น](https://aihub.or.kr/aidata/7978)
- [๊ฐ๋ฐฉ๋ฐ์ดํฐ ํ๊ตญ์ด ์์ฑ](https://aihub.or.kr/aidata/105)
- [๊ฐ๋ฐฉ๋ฐ์ดํฐ ํ๊ตญ์ด SNS](https://aihub.or.kr/aidata/30718)
### [์ธ์ข
๋ง๋ญ์น](https://ithub.korean.go.kr/)
|