Datasets:
e9t
/

Languages:
Korean
Multilinguality:
monolingual
Size Categories:
100K<n<1M
Language Creators:
found
Annotations Creators:
crowdsourced
Source Datasets:
original
License:
albertvillanova HF staff commited on
Commit
e8ac4f5
1 Parent(s): 5c17b49

Delete legacy JSON metadata (#5)

Browse files

- Delete legacy JSON metadata (f43539fc1d17086f7aa07475438473c9bda42631)

Files changed (1) hide show
  1. dataset_infos.json +0 -1
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"default": {"description": "This is a movie review dataset in the Korean language. Reviews were scraped from Naver movies. The dataset construction is based on the method noted in Large movie review dataset from Maas et al., 2011.\n", "citation": "@InProceedings{Park:2016,\n title = \"Naver Sentiment Movie Corpus\",\n author = \"Lucy Park\",\n year = \"2016\",\n howpublished = {\\url{https://github.com/e9t/nsmc}}\n}\n", "homepage": "https://github.com/e9t/nsmc/", "license": "CC0 1.0 Universal (CC0 1.0)", "features": {"id": {"dtype": "string", "id": null, "_type": "Value"}, "document": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 2, "names": ["negative", "positive"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "builder_name": "nsmc", "config_name": "default", "version": {"version_str": "1.1.0", "description": null, "major": 1, "minor": 1, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 16423803, "num_examples": 150000, "dataset_name": "nsmc"}, "test": {"name": "test", "num_bytes": 5491417, "num_examples": 50000, "dataset_name": "nsmc"}}, "download_checksums": {"https://raw.githubusercontent.com/e9t/nsmc/master/ratings_train.txt": {"num_bytes": 14628807, "checksum": "e03b7d14e9e41be8d464a28057cd25d7396c53e67aa7fd5f7e552c59b0ee2940"}, "https://raw.githubusercontent.com/e9t/nsmc/master/ratings_test.txt": {"num_bytes": 4893335, "checksum": "8ac9f64052f11dbf6ae0acb5e038f03d90a76f0eda7820cfb3a92d02edfcebda"}}, "download_size": 19522142, "post_processing_size": null, "dataset_size": 21915220, "size_in_bytes": 41437362}}