Alvant commited on
Commit
a62f0eb
1 Parent(s): 9fb65f3

add reuters

Browse files
Files changed (3) hide show
  1. README.md +26 -0
  2. data/Reuters_BOW.csv.gz +3 -0
  3. data/Reuters_NOOW.csv.gz +3 -0
README.md CHANGED
@@ -3,4 +3,30 @@ license: other
3
  license_name: topicnet
4
  license_link: >-
5
  https://github.com/machine-intelligence-laboratory/TopicNet/blob/master/LICENSE.txt
 
 
 
 
 
 
 
 
 
 
6
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  license_name: topicnet
4
  license_link: >-
5
  https://github.com/machine-intelligence-laboratory/TopicNet/blob/master/LICENSE.txt
6
+ configs:
7
+ - config_name: "bag-of-words"
8
+ default: true
9
+ data_files:
10
+ - split: train
11
+ path: "data/Reuters_BOW.csv.gz"
12
+ - config_name: "natural-order-of-words"
13
+ data_files:
14
+ - split: train
15
+ path: "data/Reuters_NOOW.csv.gz"
16
  ---
17
+
18
+
19
+ # Reuters
20
+
21
+ The Reuters Corpus contains 10,788 news documents totaling 1.3 million words. The documents have been classified into 90 topics, and grouped into two sets, called "training" and "test"; thus, the text with fileid 'test/14826' is a document drawn from the test set. This split is for training and testing algorithms that automatically detect the topic of a document, as we will see in chap-data-intensive.
22
+
23
+ * Language: English
24
+ * Number of topics: 90
25
+ * Number of articles: ~10.000
26
+ * Year: 2000
27
+
28
+
29
+ ## References
30
+
31
+ * NLTK datasets: https://www.nltk.org/book/ch02.html.
32
+ * Dataset site: https://trec.nist.gov/data/reuters/reuters.html.
data/Reuters_BOW.csv.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b09d902e738cbe694267af9fc2b99b47ec8659c4df65d1cb1dee4a2082514030
3
+ size 5306187
data/Reuters_NOOW.csv.gz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c6df18d47ddd64b70d1b4c96b0f557a918ac4d4076fb52a42293e9df9a21e204
3
+ size 5145624