Datasets:
Languages:
Korean
Multilinguality:
monolingual
Size Categories:
10M<n<100M
Language Creators:
crowdsourced
Annotations Creators:
no-annotation
Source Datasets:
original
Tags:
License:
annotations_creators: | |
- no-annotation | |
language_creators: | |
- crowdsourced | |
language: | |
- ko | |
license: | |
- cc-by-sa-4.0 | |
multilinguality: | |
- monolingual | |
pretty_name: KcBERT Pre-Training Corpus (Korean News Comments) | |
size_categories: | |
- 10M<n<100M | |
source_datasets: | |
- original | |
task_categories: | |
- fill-mask | |
- text-generation | |
task_ids: | |
- masked-language-modeling | |
- language-modeling | |
# KcBERT Pre-Training Corpus (Korean News Comments) | |
## Dataset Description | |
- **Homepage:** [KcBERT Pre-Training Corpus](https://www.kaggle.com/datasets/junbumlee/kcbert-pretraining-corpus-korean-news-comments) | |
- **Repository:** [Beomi/KcBERT](https://github.com/Beomi/KcBERT) | |
- **Paper:** [Needs More Information] | |
- **Leaderboard:** [Needs More Information] | |
- **Point of Contact:** [Needs More Information] | |
## KcBERT | |
[beomi/kcbert-base](https://huggingface.co/beomi/kcbert-base) | |
Github KcBERT Repo: [https://github.com/Beomi/KcBERT](https://github.com/Beomi/KcBERT) | |
KcBERT is Korean Comments BERT pretrained on this Corpus set. | |
(You can use it via Huggingface's Transformers library!) | |
This Kaggle Dataset contains **CLEANED** dataset preprocessed with the code below. | |
```python | |
import re | |
import emoji | |
from soynlp.normalizer import repeat_normalize | |
emojis = ''.join(emoji.UNICODE_EMOJI.keys()) | |
pattern = re.compile(f'[^ .,?!/@$%~%·∼()\x00-\x7Fㄱ-힣{emojis}]+') | |
url_pattern = re.compile( | |
r'https?:\/\/(www\.)?[-a-zA-Z0-9@:%._\+~#=]{1,256}\.[a-zA-Z0-9()]{1,6}\b([-a-zA-Z0-9()@:%_\+.~#?&//=]*)') | |
def clean(x): | |
x = pattern.sub(' ', x) | |
x = url_pattern.sub('', x) | |
x = x.strip() | |
x = repeat_normalize(x, num_repeats=2) | |
return x | |
``` | |
### License | |
[CC BY-SA 4.0](https://creativecommons.org/licenses/by-sa/4.0/) | |
## Dataset Structure | |
### Data Instance | |
```pycon | |
>>> from datasets import load_dataset | |
>>> dataset = load_dataset("Bingsu/KcBERT_Pre-Training_Corpus") | |
>>> dataset | |
DatasetDict({ | |
train: Dataset({ | |
features: ['text'], | |
num_rows: 86246285 | |
}) | |
}) | |
``` | |
### Data Size | |
download: 7.90 GiB<br> | |
generated: 11.86 GiB<br> | |
total: 19.76 GiB | |
※ You can download this dataset from [kaggle](https://www.kaggle.com/datasets/junbumlee/kcbert-pretraining-corpus-korean-news-comments), and it's 5 GiB. (12.48 GiB when uncompressed) | |
### Data Fields | |
- text: `string` | |
### Data Splits | |
| | train | | |
| ---------- | -------- | | |
| # of texts | 86246285 | | |