civil_comments / dataset_infos.json
system's picture
system HF staff
Update files from the datasets library (from 1.0.0)
5539069
raw history blame
No virus
3.25 kB
{"default": {"description": "\nThe comments in this dataset come from an archive of the Civil Comments\nplatform, a commenting plugin for independent news sites. These public comments\nwere created from 2015 - 2017 and appeared on approximately 50 English-language\nnews sites across the world. When Civil Comments shut down in 2017, they chose\nto make the public comments available in a lasting open archive to enable future\nresearch. The original data, published on figshare, includes the public comment\ntext, some associated metadata such as article IDs, timestamps and\ncommenter-generated \"civility\" labels, but does not include user ids. Jigsaw\nextended this dataset by adding additional labels for toxicity and identity\nmentions. This data set is an exact replica of the data released for the\nJigsaw Unintended Bias in Toxicity Classification Kaggle challenge. This\ndataset is released under CC0, as is the underlying comment text.\n", "citation": "\n@article{DBLP:journals/corr/abs-1903-04561,\n author = {Daniel Borkan and\n Lucas Dixon and\n Jeffrey Sorensen and\n Nithum Thain and\n Lucy Vasserman},\n title = {Nuanced Metrics for Measuring Unintended Bias with Real Data for Text\n Classification},\n journal = {CoRR},\n volume = {abs/1903.04561},\n year = {2019},\n url = {http://arxiv.org/abs/1903.04561},\n archivePrefix = {arXiv},\n eprint = {1903.04561},\n timestamp = {Sun, 31 Mar 2019 19:01:24 +0200},\n biburl = {https://dblp.org/rec/bib/journals/corr/abs-1903-04561},\n bibsource = {dblp computer science bibliography, https://dblp.org}\n}\n", "homepage": "https://www.kaggle.com/c/jigsaw-unintended-bias-in-toxicity-classification/data", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "toxicity": {"dtype": "float32", "id": null, "_type": "Value"}, "severe_toxicity": {"dtype": "float32", "id": null, "_type": "Value"}, "obscene": {"dtype": "float32", "id": null, "_type": "Value"}, "threat": {"dtype": "float32", "id": null, "_type": "Value"}, "insult": {"dtype": "float32", "id": null, "_type": "Value"}, "identity_attack": {"dtype": "float32", "id": null, "_type": "Value"}, "sexual_explicit": {"dtype": "float32", "id": null, "_type": "Value"}}, "supervised_keys": {"input": "text", "output": "toxicity"}, "builder_name": "civil_comments", "config_name": "default", "version": {"version_str": "0.9.0", "description": null, "datasets_version_to_prepare": null, "major": 0, "minor": 9, "patch": 0}, "splits": {"test": {"name": "test", "num_bytes": 32073013, "num_examples": 97320, "dataset_name": "civil_comments"}, "train": {"name": "train", "num_bytes": 596835730, "num_examples": 1804874, "dataset_name": "civil_comments"}, "validation": {"name": "validation", "num_bytes": 32326369, "num_examples": 97320, "dataset_name": "civil_comments"}}, "download_checksums": {"https://storage.googleapis.com/jigsaw-unintended-bias-in-toxicity-classification/civil_comments.zip": {"num_bytes": 414947977, "checksum": "767b71a3d9dc7a2eceb234d0c3e7e38604e11f59c12ba1cbb888ffd4ce6b6271"}}, "download_size": 414947977, "dataset_size": 661235112, "size_in_bytes": 1076183089}}