File size: 2,807 Bytes
f918a75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16ea747
f918a75
3047cec
f918a75
16ea747
f918a75
455a656
 
 
 
f918a75
16ea747
 
 
 
 
 
 
 
 
 
f918a75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
{
  "default": {
    "description": "The RVL-CDIP (Ryerson Vision Lab Complex Document Information Processing) dataset consists of 400,000 grayscale images in 16 classes, with 25,000 images per class. There are 320,000 training images, 40,000 validation images, and 40,000 test images.\n",
    "citation": "@inproceedings{harley2015icdar,\n    title = {Evaluation of Deep Convolutional Nets for Document Image Classification and Retrieval},\n    author = {Adam W Harley and Alex Ufkes and Konstantinos G Derpanis},\n    booktitle = {International Conference on Document Analysis and Recognition ({ICDAR})}},\n    year = {2015}\n}\n",
    "homepage": "https://www.cs.cmu.edu/~aharley/rvl-cdip/",
    "license": "https://www.industrydocuments.ucsf.edu/help/copyright/",
    "features": {
      "image": {
        "decode": true,
        "id": null,
        "_type": "Image"
      },
      "label": {
        "num_classes": 16,
        "names": [
          "letter",
          "form",
          "email",
          "handwritten",
          "advertisement",
          "scientific report",
          "scientific publication",
          "specification",
          "file folder",
          "news article",
          "budget",
          "invoice",
          "presentation",
          "questionnaire",
          "resume",
          "memo"
        ],
        "id": null,
        "_type": "ClassLabel"
      },
      "id": {
        "_type": "Value",
        "dtype": "string"
      },
      "words": {
        "_type": "Sequence",
        "feature": {
          "dtype": "string",
          "_type": "Value"
        }
      },
      "boxes": {
        "_type": "Sequence",
        "feature": {
          "_type": "Sequence",
          "feature": {
            "dtype": "int32",
            "_type": "Value"
          }
        }
      }
    },
    "post_processed": null,
    "supervised_keys": {
      "input": "image",
      "output": "label"
    },
    "task_templates": [
      {
        "task": "image-classification",
        "image_column": "image",
        "label_column": "label"
      }
    ],
    "builder_name": "rvl_cdip_easyOCR",
    "config_name": "default",
    "version": {
      "version_str": "1.0.0",
      "description": null,
      "major": 1,
      "minor": 0,
      "patch": 0
    },
    "splits": {
      "train": {
        "name": "train",
        "num_bytes": 38816373360,
        "num_examples": 320000,
        "dataset_name": "rvl_cdip_easyOCR"
      },
      "test": {
        "name": "test",
        "num_bytes": 4863300853,
        "num_examples": 40000,
        "dataset_name": "rvl_cdip_easyOCR"
      },
      "validation": {
        "name": "validation",
        "num_bytes": 4868685208,
        "num_examples": 40000,
        "dataset_name": "rvl_cdip_easyOCR"
      }
    }
  }
}