autotrain-data-processor commited on
Commit
ca47807
1 Parent(s): 2665e96

Processed data from AutoTrain data processor ([2022-06-09 12:28 ]

Browse files
README.md ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ languages:
3
+ - en
4
+ task_categories:
5
+ - text-classification
6
+
7
+ ---
8
+ # AutoTrain Dataset for project: car-review-project
9
+
10
+ ## Dataset Descritpion
11
+
12
+ This dataset has been automatically processed by AutoTrain for project car-review-project.
13
+
14
+ ### Languages
15
+
16
+ The BCP-47 code for the dataset's language is en.
17
+
18
+ ## Dataset Structure
19
+
20
+ ### Data Instances
21
+
22
+ A sample from this dataset looks as follows:
23
+
24
+ ```json
25
+ [
26
+ {
27
+ "text": " ",
28
+ "target": 1
29
+ },
30
+ {
31
+ "text": " Mazda truck costs less than the sister look-a-like Ford; Mazda is a \"A\" series of the Ford Ranger, [...]",
32
+ "target": 2
33
+ }
34
+ ]
35
+ ```
36
+
37
+ ### Dataset Fields
38
+
39
+ The dataset has the following fields (also called "features"):
40
+
41
+ ```json
42
+ {
43
+ "text": "Value(dtype='string', id=None)",
44
+ "target": "ClassLabel(num_classes=3, names=['great', 'ok', 'poor'], id=None)"
45
+ }
46
+ ```
47
+
48
+ ### Dataset Splits
49
+
50
+ This dataset is split into a train and validation split. The split sizes are as follow:
51
+
52
+ | Split name | Num samples |
53
+ | ------------ | ------------------- |
54
+ | train | 19731 |
55
+ | valid | 4935 |
processed/dataset_dict.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"splits": ["train", "valid"]}
processed/train/dataset.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a9299249ee2eb347e92a478ba7ba97e91e0b62fef18c8b100e183ddf90d9a15
3
+ size 11805360
processed/train/dataset_info.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": null,
3
+ "citation": "",
4
+ "config_name": null,
5
+ "dataset_size": null,
6
+ "description": "AutoTrain generated dataset",
7
+ "download_checksums": null,
8
+ "download_size": null,
9
+ "features": {
10
+ "text": {
11
+ "dtype": "string",
12
+ "id": null,
13
+ "_type": "Value"
14
+ },
15
+ "target": {
16
+ "num_classes": 3,
17
+ "names": [
18
+ "great",
19
+ "ok",
20
+ "poor"
21
+ ],
22
+ "id": null,
23
+ "_type": "ClassLabel"
24
+ }
25
+ },
26
+ "homepage": "",
27
+ "license": "",
28
+ "post_processed": null,
29
+ "post_processing_size": null,
30
+ "size_in_bytes": null,
31
+ "splits": {
32
+ "train": {
33
+ "name": "train",
34
+ "num_bytes": 11800540,
35
+ "num_examples": 19731,
36
+ "dataset_name": null
37
+ }
38
+ },
39
+ "supervised_keys": null,
40
+ "task_templates": null,
41
+ "version": null
42
+ }
processed/train/state.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "dataset.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "833c82d9af851800",
8
+ "_format_columns": [
9
+ "target",
10
+ "text"
11
+ ],
12
+ "_format_kwargs": {},
13
+ "_format_type": null,
14
+ "_indexes": {},
15
+ "_output_all_columns": false,
16
+ "_split": null
17
+ }
processed/valid/dataset.arrow ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8822e2fd0dafcec70f386cf0f7e81bdd55b49072105f38b43c1e2e373480afa
3
+ size 2966488
processed/valid/dataset_info.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "builder_name": null,
3
+ "citation": "",
4
+ "config_name": null,
5
+ "dataset_size": null,
6
+ "description": "AutoTrain generated dataset",
7
+ "download_checksums": null,
8
+ "download_size": null,
9
+ "features": {
10
+ "text": {
11
+ "dtype": "string",
12
+ "id": null,
13
+ "_type": "Value"
14
+ },
15
+ "target": {
16
+ "num_classes": 3,
17
+ "names": [
18
+ "great",
19
+ "ok",
20
+ "poor"
21
+ ],
22
+ "id": null,
23
+ "_type": "ClassLabel"
24
+ }
25
+ },
26
+ "homepage": "",
27
+ "license": "",
28
+ "post_processed": null,
29
+ "post_processing_size": null,
30
+ "size_in_bytes": null,
31
+ "splits": {
32
+ "valid": {
33
+ "name": "valid",
34
+ "num_bytes": 2964968,
35
+ "num_examples": 4935,
36
+ "dataset_name": null
37
+ }
38
+ },
39
+ "supervised_keys": null,
40
+ "task_templates": null,
41
+ "version": null
42
+ }
processed/valid/state.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_data_files": [
3
+ {
4
+ "filename": "dataset.arrow"
5
+ }
6
+ ],
7
+ "_fingerprint": "606441e98f6797b9",
8
+ "_format_columns": [
9
+ "target",
10
+ "text"
11
+ ],
12
+ "_format_kwargs": {},
13
+ "_format_type": null,
14
+ "_indexes": {},
15
+ "_output_all_columns": false,
16
+ "_split": null
17
+ }