albertvillanova HF staff commited on
Commit
99c1098
·
verified ·
1 Parent(s): 3c8cc81

Convert dataset to Parquet

Browse files

Convert dataset to Parquet.

This dataset uses `tasks`, which are deprecated and will raise an error after the next release of `datasets`. See: https://github.com/huggingface/datasets/pull/6999

README.md ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ dataset_info:
3
+ features:
4
+ - name: text
5
+ dtype: string
6
+ - name: label
7
+ dtype:
8
+ class_label:
9
+ names:
10
+ '0': World
11
+ '1': Sports
12
+ '2': Business
13
+ '3': Sci/Tech
14
+ splits:
15
+ - name: train
16
+ num_bytes: 29817303
17
+ num_examples: 120000
18
+ - name: test
19
+ num_bytes: 1879474
20
+ num_examples: 7600
21
+ download_size: 19820267
22
+ dataset_size: 31696777
23
+ configs:
24
+ - config_name: default
25
+ data_files:
26
+ - split: train
27
+ path: data/train-*
28
+ - split: test
29
+ path: data/test-*
30
+ ---
data/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f43f50be12dbc50a87a88f7cf78d32ed16651d402bb63b590b105cd2ec54f346
3
+ size 1234829
data/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b0df9c3bc6b51f5ccb303f8e622b2a7a932b3d21c6f0d38a1e403e7b947d780
3
+ size 18585438
dataset_infos.json CHANGED
@@ -1 +1,55 @@
1
- {"default": {"description": "AG is a collection of more than 1 million news articles. News articles have been\ngathered from more than 2000 news sources by ComeToMyHead in more than 1 year of\nactivity. ComeToMyHead is an academic news search engine which has been running\nsince July, 2004. The dataset is provided by the academic comunity for research\npurposes in data mining (clustering, classification, etc), information retrieval\n(ranking, search, etc), xml, data compression, data streaming, and any other\nnon-commercial activity. For more information, please refer to the link\nhttp://www.di.unipi.it/~gulli/AG_corpus_of_news_articles.html .\n\nThe AG's news topic classification dataset is constructed by Xiang Zhang\n(xiang.zhang@nyu.edu) from the dataset above. It is used as a text\nclassification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann\nLeCun. Character-level Convolutional Networks for Text Classification. Advances\nin Neural Information Processing Systems 28 (NIPS 2015).\n", "citation": "@inproceedings{Zhang2015CharacterlevelCN,\n title={Character-level Convolutional Networks for Text Classification},\n author={Xiang Zhang and Junbo Jake Zhao and Yann LeCun},\n booktitle={NIPS},\n year={2015}\n}\n", "homepage": "http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html", "license": "", "features": {"text": {"dtype": "string", "id": null, "_type": "Value"}, "label": {"num_classes": 4, "names": ["World", "Sports", "Business", "Sci/Tech"], "names_file": null, "id": null, "_type": "ClassLabel"}}, "post_processed": null, "supervised_keys": null, "task_templates": [{"task": "text-classification", "text_column": "text", "label_column": "label", "labels": ["Business", "Sci/Tech", "Sports", "World"]}], "builder_name": "ag_news", "config_name": "default", "version": {"version_str": "0.0.0", "description": null, "major": 0, "minor": 0, "patch": 0}, "splits": {"train": {"name": "train", "num_bytes": 29817351, "num_examples": 120000, "dataset_name": "ag_news"}, "test": {"name": "test", "num_bytes": 1879478, "num_examples": 7600, "dataset_name": "ag_news"}}, "download_checksums": {"https://raw.githubusercontent.com/mhjabreel/CharCnn_Keras/master/data/ag_news_csv/train.csv": {"num_bytes": 29470338, "checksum": "76a0a2d2f92b286371fe4d4044640910a04a803fdd2538e0f3f29a5c6f6b672e"}, "https://raw.githubusercontent.com/mhjabreel/CharCnn_Keras/master/data/ag_news_csv/test.csv": {"num_bytes": 1857427, "checksum": "521465c2428ed7f02f8d6db6ffdd4b5447c1c701962353eb2c40d548c3c85699"}}, "download_size": 31327765, "post_processing_size": null, "dataset_size": 31696829, "size_in_bytes": 63024594}}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "default": {
3
+ "description": "AG is a collection of more than 1 million news articles. News articles have been\ngathered from more than 2000 news sources by ComeToMyHead in more than 1 year of\nactivity. ComeToMyHead is an academic news search engine which has been running\nsince July, 2004. The dataset is provided by the academic comunity for research\npurposes in data mining (clustering, classification, etc), information retrieval\n(ranking, search, etc), xml, data compression, data streaming, and any other\nnon-commercial activity. For more information, please refer to the link\nhttp://www.di.unipi.it/~gulli/AG_corpus_of_news_articles.html .\nThe AG's news topic classification dataset is constructed by Xiang Zhang\n(xiang.zhang@nyu.edu) from the dataset above. It is used as a text\nclassification benchmark in the following paper: Xiang Zhang, Junbo Zhao, Yann\nLeCun. Character-level Convolutional Networks for Text Classification. Advances\nin Neural Information Processing Systems 28 (NIPS 2015).\n",
4
+ "citation": "@inproceedings{Zhang2015CharacterlevelCN,\n title={Character-level Convolutional Networks for Text Classification},\n author={Xiang Zhang and Junbo Jake Zhao and Yann LeCun},\n booktitle={NIPS},\n year={2015}\n}\n",
5
+ "homepage": "http://groups.di.unipi.it/~gulli/AG_corpus_of_news_articles.html",
6
+ "license": "",
7
+ "features": {
8
+ "text": {
9
+ "dtype": "string",
10
+ "_type": "Value"
11
+ },
12
+ "label": {
13
+ "names": [
14
+ "World",
15
+ "Sports",
16
+ "Business",
17
+ "Sci/Tech"
18
+ ],
19
+ "_type": "ClassLabel"
20
+ }
21
+ },
22
+ "task_templates": [
23
+ {
24
+ "task": "text-classification",
25
+ "label_column": "label"
26
+ }
27
+ ],
28
+ "builder_name": "parquet",
29
+ "dataset_name": "labels",
30
+ "config_name": "default",
31
+ "version": {
32
+ "version_str": "0.0.0",
33
+ "major": 0,
34
+ "minor": 0,
35
+ "patch": 0
36
+ },
37
+ "splits": {
38
+ "train": {
39
+ "name": "train",
40
+ "num_bytes": 29817303,
41
+ "num_examples": 120000,
42
+ "dataset_name": null
43
+ },
44
+ "test": {
45
+ "name": "test",
46
+ "num_bytes": 1879474,
47
+ "num_examples": 7600,
48
+ "dataset_name": null
49
+ }
50
+ },
51
+ "download_size": 19820267,
52
+ "dataset_size": 31696777,
53
+ "size_in_bytes": 51517044
54
+ }
55
+ }