Datasets:

Languages:
English
ArXiv:
License:
parquet-converter commited on
Commit
0bfcbb3
1 Parent(s): a014c2c

Update parquet files

Browse files
.gitattributes DELETED
@@ -1,37 +0,0 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ftz filter=lfs diff=lfs merge=lfs -text
6
- *.gz filter=lfs diff=lfs merge=lfs -text
7
- *.h5 filter=lfs diff=lfs merge=lfs -text
8
- *.joblib filter=lfs diff=lfs merge=lfs -text
9
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
10
- *.model filter=lfs diff=lfs merge=lfs -text
11
- *.msgpack filter=lfs diff=lfs merge=lfs -text
12
- *.onnx filter=lfs diff=lfs merge=lfs -text
13
- *.ot filter=lfs diff=lfs merge=lfs -text
14
- *.parquet filter=lfs diff=lfs merge=lfs -text
15
- *.pb filter=lfs diff=lfs merge=lfs -text
16
- *.pt filter=lfs diff=lfs merge=lfs -text
17
- *.pth filter=lfs diff=lfs merge=lfs -text
18
- *.rar filter=lfs diff=lfs merge=lfs -text
19
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
20
- *.tar.* filter=lfs diff=lfs merge=lfs -text
21
- *.tflite filter=lfs diff=lfs merge=lfs -text
22
- *.tgz filter=lfs diff=lfs merge=lfs -text
23
- *.wasm filter=lfs diff=lfs merge=lfs -text
24
- *.xz filter=lfs diff=lfs merge=lfs -text
25
- *.zip filter=lfs diff=lfs merge=lfs -text
26
- *.zstandard filter=lfs diff=lfs merge=lfs -text
27
- *tfevents* filter=lfs diff=lfs merge=lfs -text
28
- # Audio files - uncompressed
29
- *.pcm filter=lfs diff=lfs merge=lfs -text
30
- *.sam filter=lfs diff=lfs merge=lfs -text
31
- *.raw filter=lfs diff=lfs merge=lfs -text
32
- # Audio files - compressed
33
- *.aac filter=lfs diff=lfs merge=lfs -text
34
- *.flac filter=lfs diff=lfs merge=lfs -text
35
- *.mp3 filter=lfs diff=lfs merge=lfs -text
36
- *.ogg filter=lfs diff=lfs merge=lfs -text
37
- *.wav filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
README.md DELETED
@@ -1,356 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - crowdsourced
4
- language_creators:
5
- - crowdsourced
6
- language:
7
- - en
8
- license:
9
- - cc-by-4.0
10
- multilinguality:
11
- - monolingual
12
- pretty_name: TextVQA
13
- size_categories:
14
- - 10K<n<100K
15
- source_datasets:
16
- - original
17
- task_categories:
18
- - visual-question-answering
19
- task_ids:
20
- - visual-question-answering
21
- dataset_info:
22
- - config_name: train
23
- features:
24
- - name: image_id
25
- dtype: string
26
- - name: question_id
27
- dtype: int32
28
- - name: question
29
- dtype: string
30
- - name: question_tokens
31
- sequence: string
32
- - name: image
33
- dtype: image
34
- - name: image_width
35
- dtype: int32
36
- - name: image_height
37
- dtype: int32
38
- - name: flickr_original_url
39
- dtype: string
40
- - name: flickr_300k_url
41
- dtype: string
42
- - name: answers
43
- sequence: string
44
- - name: image_classes
45
- sequence: string
46
- - name: set_name
47
- dtype: string
48
- splits:
49
- - name: train
50
- num_bytes: 21381310
51
- num_examples: 34602
52
- - name: validation
53
- num_bytes: 3077854
54
- num_examples: 5000
55
- - name: test
56
- num_bytes: 3025046
57
- num_examples: 5734
58
- download_size: 8070116310
59
- dataset_size: 27484210
60
- - config_name: val
61
- features:
62
- - name: image_id
63
- dtype: string
64
- - name: question_id
65
- dtype: int32
66
- - name: question
67
- dtype: string
68
- - name: question_tokens
69
- sequence: string
70
- - name: image
71
- dtype: image
72
- - name: image_width
73
- dtype: int32
74
- - name: image_height
75
- dtype: int32
76
- - name: flickr_original_url
77
- dtype: string
78
- - name: flickr_300k_url
79
- dtype: string
80
- - name: answers
81
- sequence: string
82
- - name: image_classes
83
- sequence: string
84
- - name: set_name
85
- dtype: string
86
- splits:
87
- - name: train
88
- num_bytes: 21381310
89
- num_examples: 34602
90
- - name: validation
91
- num_bytes: 3077854
92
- num_examples: 5000
93
- - name: test
94
- num_bytes: 3025046
95
- num_examples: 5734
96
- download_size: 8070116310
97
- dataset_size: 27484210
98
- - config_name: test
99
- features:
100
- - name: image_id
101
- dtype: string
102
- - name: question_id
103
- dtype: int32
104
- - name: question
105
- dtype: string
106
- - name: question_tokens
107
- sequence: string
108
- - name: image
109
- dtype: image
110
- - name: image_width
111
- dtype: int32
112
- - name: image_height
113
- dtype: int32
114
- - name: flickr_original_url
115
- dtype: string
116
- - name: flickr_300k_url
117
- dtype: string
118
- - name: answers
119
- sequence: string
120
- - name: image_classes
121
- sequence: string
122
- - name: set_name
123
- dtype: string
124
- splits:
125
- - name: train
126
- num_bytes: 21381310
127
- num_examples: 34602
128
- - name: validation
129
- num_bytes: 3077854
130
- num_examples: 5000
131
- - name: test
132
- num_bytes: 3025046
133
- num_examples: 5734
134
- download_size: 8070116310
135
- dataset_size: 27484210
136
- - config_name: textvqa
137
- features:
138
- - name: image_id
139
- dtype: string
140
- - name: question_id
141
- dtype: int32
142
- - name: question
143
- dtype: string
144
- - name: question_tokens
145
- sequence: string
146
- - name: image
147
- dtype: image
148
- - name: image_width
149
- dtype: int32
150
- - name: image_height
151
- dtype: int32
152
- - name: flickr_original_url
153
- dtype: string
154
- - name: flickr_300k_url
155
- dtype: string
156
- - name: answers
157
- sequence: string
158
- - name: image_classes
159
- sequence: string
160
- - name: set_name
161
- dtype: string
162
- splits:
163
- - name: train
164
- num_bytes: 22073350
165
- num_examples: 34602
166
- - name: validation
167
- num_bytes: 3177854
168
- num_examples: 5000
169
- - name: test
170
- num_bytes: 3139726
171
- num_examples: 5734
172
- download_size: 8070116310
173
- dataset_size: 28390930
174
- ---
175
-
176
- # Dataset Card for TextVQA
177
-
178
- ## Table of Contents
179
- - [Dataset Description](#dataset-description)
180
- - [Dataset Summary](#dataset-summary)
181
- - [Supported Tasks](#supported-tasks-and-leaderboards)
182
- - [Languages](#languages)
183
- - [Dataset Structure](#dataset-structure)
184
- - [Data Instances](#data-instances)
185
- - [Data Fields](#data-instances)
186
- - [Data Splits](#data-instances)
187
- - [Dataset Creation](#dataset-creation)
188
- - [Curation Rationale](#curation-rationale)
189
- - [Source Data](#source-data)
190
- - [Annotations](#annotations)
191
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
192
- - [Considerations for Using the Data](#considerations-for-using-the-data)
193
- - [Social Impact of Dataset](#social-impact-of-dataset)
194
- - [Discussion of Biases](#discussion-of-biases)
195
- - [Other Known Limitations](#other-known-limitations)
196
- - [Additional Information](#additional-information)
197
- - [Dataset Curators](#dataset-curators)
198
- - [Licensing Information](#licensing-information)
199
- - [Citation Information](#citation-information)
200
- - [Contributions](#contributions)
201
-
202
- ## Dataset Description
203
-
204
- - **Homepage:** https://textvqa.org
205
- - **Repository:** https://github.com/facebookresearch/mmf
206
- - **Paper:** https://arxiv.org/abs/1904.08920
207
- - **Leaderboard:** https://eval.ai/web/challenges/challenge-page/874/overview
208
- - **Point of Contact:** mailto:amanpreet@nyu.edu
209
-
210
- ### Dataset Summary
211
-
212
- TextVQA requires models to read and reason about text in images to answer questions about them.
213
- Specifically, models need to incorporate a new modality of text present in the images and reason
214
- over it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images
215
- from the OpenImages dataset. The dataset uses [VQA accuracy](https://visualqa.org/evaluation.html) metric for evaluation.
216
-
217
- ### Supported Tasks and Leaderboards
218
-
219
- - `visual-question-answering`: The dataset can be used for Visual Question Answering tasks where given an image, you have to answer a question based on the image. For the TextVQA dataset specifically, the questions require reading and reasoning about the scene text in the given image.
220
-
221
- ### Languages
222
-
223
- The questions in the dataset are in English.
224
-
225
- ## Dataset Structure
226
-
227
- ### Data Instances
228
-
229
- A typical sample mainly contains the question in `question` field, an image object in `image` field, OpenImage image id in `image_id` and lot of other useful metadata. 10 answers per questions are contained in the `answers` attribute. For test set, 10 empty strings are contained in the `answers` field as the answers are not available for it.
230
-
231
- An example look like below:
232
-
233
- ```
234
- {'question': 'who is this copyrighted by?',
235
- 'image_id': '00685bc495504d61',
236
- 'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=384x512 at 0x276021C5EB8>,
237
- 'image_classes': ['Vehicle', 'Tower', 'Airplane', 'Aircraft'],
238
- 'flickr_original_url': 'https://farm2.staticflickr.com/5067/5620759429_4ea686e643_o.jpg',
239
- 'flickr_300k_url': 'https://c5.staticflickr.com/6/5067/5620759429_f43a649fb5_z.jpg',
240
- 'image_width': 786,
241
- 'image_height': 1024,
242
- 'answers': ['simon clancy',
243
- 'simon ciancy',
244
- 'simon clancy',
245
- 'simon clancy',
246
- 'the brand is bayard',
247
- 'simon clancy',
248
- 'simon clancy',
249
- 'simon clancy',
250
- 'simon clancy',
251
- 'simon clancy'],
252
- 'question_tokens': ['who', 'is', 'this', 'copyrighted', 'by'],
253
- 'question_id': 3,
254
- 'set_name': 'train'
255
- },
256
- ```
257
-
258
- ### Data Fields
259
-
260
- - `question`: string, the question that is being asked about the image
261
- - `image_id`: string, id of the image which is same as the OpenImages id
262
- - `image`: A `PIL.Image.Image` object containing the image about which the question is being asked. Note that when accessing the image column: `dataset[0]["image"]` the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the `"image"` column, *i.e.* `dataset[0]["image"]` should **always** be preferred over `dataset["image"][0]`.
263
- - `image_classes`: List[str], The OpenImages classes to which the image belongs to.
264
- - `flickr_original_url`: string, URL to original image on Flickr
265
- - `flickr_300k_url`: string, URL to resized and low-resolution image on Flickr.
266
- - `image_width`: int, Width of the original image.
267
- - `image_height`: int, Height of the original image.
268
- - `question_tokens`: List[str], A pre-tokenized list of question.
269
- - `answers`: List[str], List of 10 human-annotated answers for the question. These 10 answers are collected from 10 different users. The list will contain empty strings for test set for which we don't have the answers.
270
- - `question_id`: int, Unique id of the question.
271
- - `set_name`: string, the set to which this question belongs.
272
-
273
- ### Data Splits
274
-
275
- There are three splits. `train`, `validation` and `test`. The `train` and `validation` sets share images with OpenImages `train` set and have their answers available. For test set answers, we return a list of ten empty strings. To get inference results and numbers on `test` set, you need to go to the [EvalAI leaderboard](https://eval.ai/web/challenges/challenge-page/874/overview) and upload your predictions there. Please see instructions at [https://textvqa.org/challenge/](https://textvqa.org/challenge/).
276
-
277
- ## Dataset Creation
278
-
279
- ### Curation Rationale
280
-
281
- From the paper:
282
-
283
- > Studies have shown that a dominant class of questions asked by visually impaired users on images of their surroundings involves reading text in the image. But today’s VQA models can not read! Our paper takes a first step towards addressing this problem. First, we introduce a new “TextVQA” dataset to facilitate progress on this important problem. Existing datasets either have a small proportion of questions about text (e.g., the VQA dataset) or are too small (e.g., the VizWiz dataset). TextVQA contains 45,336 questions on 28,408 images that require reasoning about text to answer.
284
-
285
- ### Source Data
286
-
287
- #### Initial Data Collection and Normalization
288
-
289
- The initial images were sourced from [OpenImages](https://storage.googleapis.com/openimages/web/factsfigures_v4.html) v4 dataset. These were first filtered based on automatic heuristics using an OCR system where we only took images which had at least some text detected in them. See [annotation process](#annotation-process) section to understand the next stages.
290
-
291
- #### Who are the source language producers?
292
-
293
- English Crowdsource Annotators
294
-
295
- ### Annotations
296
-
297
- #### Annotation process
298
-
299
- After the automatic process of filter the images that contain text, the images were manually verified using human annotators making sure that they had text. In next stage, the annotators were asked to write questions involving scene text for the image. For some images, in this stage, two questions were collected whenever possible. Finally, in the last stage, ten different human annotators answer the questions asked in last stage.
300
-
301
- #### Who are the annotators?
302
-
303
- Annotators are from one of the major data collection platforms such as AMT. Exact details are not mentioned in the paper.
304
-
305
- ### Personal and Sensitive Information
306
-
307
- The dataset does have similar PII issues as OpenImages and can at some times contain human faces, license plates, and documents. Using provided `image_classes` data field is one option to try to filter out some of this information.
308
-
309
- ## Considerations for Using the Data
310
-
311
- ### Social Impact of Dataset
312
-
313
- The paper helped realize the importance of scene text recognition and reasoning in general purpose machine learning applications and has led to many follow-up works including [TextCaps](https://textvqa.org/textcaps) and [TextOCR](https://textvqa.org/textocr). Similar datasets were introduced over the time which specifically focus on sight-disabled users such as [VizWiz](https://vizwiz.org) or focusing specifically on the same problem as TextVQA like [STVQA](https://paperswithcode.com/dataset/st-vqa), [DocVQA](https://arxiv.org/abs/2007.00398v3) and [OCRVQA](https://ocr-vqa.github.io/). Currently, most methods train on combined dataset from TextVQA and STVQA to achieve state-of-the-art performance on both datasets.
314
-
315
- ### Discussion of Biases
316
-
317
- Question-only bias where a model is able to answer the question without even looking at the image is discussed in the [paper](https://arxiv.org/abs/1904.08920) which was a major issue with original VQA dataset. The outlier bias in answers is prevented by collecting 10 different answers which are also taken in consideration by the evaluation metric.
318
-
319
- ### Other Known Limitations
320
-
321
- - The dataset is english only but does involve images with non-English latin characters so can involve some multi-lingual understanding.
322
- - The performance on the dataset is also dependent on the quality of OCR used as the OCR errors can directly lead to wrong answers.
323
- - The metric used for calculating accuracy is same as [VQA accuracy](https://visualqa.org/evaluation.html). This involves one-to-one matching with the given answers and thus doesn't allow analyzing one-off errors through OCR.
324
-
325
- ## Additional Information
326
-
327
- ### Dataset Curators
328
-
329
- - [Amanpreet Singh](https://github.com/apsdehal)
330
- - Vivek Natarjan
331
- - Meet Shah
332
- - Yu Jiang
333
- - Xinlei Chen
334
- - Dhruv Batra
335
- - Devi Parikh
336
- - Marcus Rohrbach
337
-
338
- ### Licensing Information
339
-
340
- CC by 4.0
341
-
342
- ### Citation Information
343
-
344
- ```bibtex
345
- @inproceedings{singh2019towards,
346
- title={Towards VQA Models That Can Read},
347
- author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Batra, Dhruv and Parikh, Devi and Rohrbach, Marcus},
348
- booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
349
- pages={8317-8326},
350
- year={2019}
351
- }
352
- ```
353
-
354
- ### Contributions
355
-
356
- Thanks to [@apsdehal](https://github.com/apsdehal) for adding this dataset.
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
dataset_infos.json DELETED
@@ -1 +0,0 @@
1
- {"train": {"description": "TextVQA requires models to read and reason about text in images to answer questions about them. \nSpecifically, models need to incorporate a new modality of text present in the images and reason \nover it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images\nfrom the OpenImages dataset. \n", "citation": "\n@inproceedings{singh2019towards,\n title={Towards VQA Models That Can Read},\n author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Parikh, Devi and Rohrbach, Marcus},\n booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},\n pages={8317-8326},\n year={2019}\n}\n", "homepage": "https://textvqa.org", "license": "CC BY 4.0", "features": {"image_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "int32", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "question_tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image": {"decode": true, "id": null, "_type": "Image"}, "image_width": {"dtype": "int32", "id": null, "_type": "Value"}, "image_height": {"dtype": "int32", "id": null, "_type": "Value"}, "flickr_original_url": {"dtype": "string", "id": null, "_type": "Value"}, "flickr_300k_url": {"dtype": "string", "id": null, "_type": "Value"}, "answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image_classes": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "set_name": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "textvqa", "config_name": "train", "version": {"version_str": "0.5.1", "description": "", "major": 0, "minor": 5, "patch": 1}, "splits": {"train": {"name": "train", "num_bytes": 21381310, "num_examples": 34602, "dataset_name": "textvqa"}, "validation": {"name": "validation", "num_bytes": 3077854, "num_examples": 5000, "dataset_name": "textvqa"}, "test": {"name": "test", "num_bytes": 3025046, "num_examples": 5734, "dataset_name": "textvqa"}}, "download_checksums": {"https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_train.json": {"num_bytes": 21634937, "checksum": "95f5c407db56cba56a177799dcd685a7cc0ec7c0d851b59910acf7786d31b68a"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_val.json": {"num_bytes": 3116162, "checksum": "4ceb5aadc1a41719d0a3e4dfdf06838bcfee1db569a9a65ee67d31c99893081d"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_test.json": {"num_bytes": 2770520, "checksum": "d8d4b738101087bac5a6182d22d9aef3772e08e77827e6cf6116808910b75db2"}, "https://dl.fbaipublicfiles.com/textvqa/images/train_val_images.zip": {"num_bytes": 7072297970, "checksum": "ecf35005640d0708eae185aab1c0a10f89b2db7420b29185a1ed92a8f4290498"}, "https://dl.fbaipublicfiles.com/textvqa/images/test_images.zip": {"num_bytes": 970296721, "checksum": "1276b908994c444c46484fb21e9e15fcda1be9c675f6ad727489e52eea68cbcd"}}, "download_size": 8070116310, "post_processing_size": null, "dataset_size": 27484210, "size_in_bytes": 8097600520}, "val": {"description": "TextVQA requires models to read and reason about text in images to answer questions about them. \nSpecifically, models need to incorporate a new modality of text present in the images and reason \nover it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images\nfrom the OpenImages dataset. \n", "citation": "\n@inproceedings{singh2019towards,\n title={Towards VQA Models That Can Read},\n author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Parikh, Devi and Rohrbach, Marcus},\n booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},\n pages={8317-8326},\n year={2019}\n}\n", "homepage": "https://textvqa.org", "license": "CC BY 4.0", "features": {"image_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "int32", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "question_tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image": {"decode": true, "id": null, "_type": "Image"}, "image_width": {"dtype": "int32", "id": null, "_type": "Value"}, "image_height": {"dtype": "int32", "id": null, "_type": "Value"}, "flickr_original_url": {"dtype": "string", "id": null, "_type": "Value"}, "flickr_300k_url": {"dtype": "string", "id": null, "_type": "Value"}, "answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image_classes": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "set_name": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "textvqa", "config_name": "val", "version": {"version_str": "0.5.1", "description": "", "major": 0, "minor": 5, "patch": 1}, "splits": {"train": {"name": "train", "num_bytes": 21381310, "num_examples": 34602, "dataset_name": "textvqa"}, "validation": {"name": "validation", "num_bytes": 3077854, "num_examples": 5000, "dataset_name": "textvqa"}, "test": {"name": "test", "num_bytes": 3025046, "num_examples": 5734, "dataset_name": "textvqa"}}, "download_checksums": {"https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_train.json": {"num_bytes": 21634937, "checksum": "95f5c407db56cba56a177799dcd685a7cc0ec7c0d851b59910acf7786d31b68a"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_val.json": {"num_bytes": 3116162, "checksum": "4ceb5aadc1a41719d0a3e4dfdf06838bcfee1db569a9a65ee67d31c99893081d"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_test.json": {"num_bytes": 2770520, "checksum": "d8d4b738101087bac5a6182d22d9aef3772e08e77827e6cf6116808910b75db2"}, "https://dl.fbaipublicfiles.com/textvqa/images/train_val_images.zip": {"num_bytes": 7072297970, "checksum": "ecf35005640d0708eae185aab1c0a10f89b2db7420b29185a1ed92a8f4290498"}, "https://dl.fbaipublicfiles.com/textvqa/images/test_images.zip": {"num_bytes": 970296721, "checksum": "1276b908994c444c46484fb21e9e15fcda1be9c675f6ad727489e52eea68cbcd"}}, "download_size": 8070116310, "post_processing_size": null, "dataset_size": 27484210, "size_in_bytes": 8097600520}, "test": {"description": "TextVQA requires models to read and reason about text in images to answer questions about them. \nSpecifically, models need to incorporate a new modality of text present in the images and reason \nover it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images\nfrom the OpenImages dataset. \n", "citation": "\n@inproceedings{singh2019towards,\n title={Towards VQA Models That Can Read},\n author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Parikh, Devi and Rohrbach, Marcus},\n booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},\n pages={8317-8326},\n year={2019}\n}\n", "homepage": "https://textvqa.org", "license": "CC BY 4.0", "features": {"image_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "int32", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "question_tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image": {"decode": true, "id": null, "_type": "Image"}, "image_width": {"dtype": "int32", "id": null, "_type": "Value"}, "image_height": {"dtype": "int32", "id": null, "_type": "Value"}, "flickr_original_url": {"dtype": "string", "id": null, "_type": "Value"}, "flickr_300k_url": {"dtype": "string", "id": null, "_type": "Value"}, "answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image_classes": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "set_name": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "textvqa", "config_name": "test", "version": {"version_str": "0.5.1", "description": "", "major": 0, "minor": 5, "patch": 1}, "splits": {"train": {"name": "train", "num_bytes": 21381310, "num_examples": 34602, "dataset_name": "textvqa"}, "validation": {"name": "validation", "num_bytes": 3077854, "num_examples": 5000, "dataset_name": "textvqa"}, "test": {"name": "test", "num_bytes": 3025046, "num_examples": 5734, "dataset_name": "textvqa"}}, "download_checksums": {"https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_train.json": {"num_bytes": 21634937, "checksum": "95f5c407db56cba56a177799dcd685a7cc0ec7c0d851b59910acf7786d31b68a"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_val.json": {"num_bytes": 3116162, "checksum": "4ceb5aadc1a41719d0a3e4dfdf06838bcfee1db569a9a65ee67d31c99893081d"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_test.json": {"num_bytes": 2770520, "checksum": "d8d4b738101087bac5a6182d22d9aef3772e08e77827e6cf6116808910b75db2"}, "https://dl.fbaipublicfiles.com/textvqa/images/train_val_images.zip": {"num_bytes": 7072297970, "checksum": "ecf35005640d0708eae185aab1c0a10f89b2db7420b29185a1ed92a8f4290498"}, "https://dl.fbaipublicfiles.com/textvqa/images/test_images.zip": {"num_bytes": 970296721, "checksum": "1276b908994c444c46484fb21e9e15fcda1be9c675f6ad727489e52eea68cbcd"}}, "download_size": 8070116310, "post_processing_size": null, "dataset_size": 27484210, "size_in_bytes": 8097600520}, "textvqa": {"description": "TextVQA requires models to read and reason about text in images to answer questions about them.\nSpecifically, models need to incorporate a new modality of text present in the images and reason\nover it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images\nfrom the OpenImages dataset.\n", "citation": "\n@inproceedings{singh2019towards,\n title={Towards VQA Models That Can Read},\n author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Batra, Dhruv and Parikh, Devi and Rohrbach, Marcus},\n booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},\n pages={8317-8326},\n year={2019}\n}\n", "homepage": "https://textvqa.org", "license": "CC BY 4.0", "features": {"image_id": {"dtype": "string", "id": null, "_type": "Value"}, "question_id": {"dtype": "int32", "id": null, "_type": "Value"}, "question": {"dtype": "string", "id": null, "_type": "Value"}, "question_tokens": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image": {"decode": true, "id": null, "_type": "Image"}, "image_width": {"dtype": "int32", "id": null, "_type": "Value"}, "image_height": {"dtype": "int32", "id": null, "_type": "Value"}, "flickr_original_url": {"dtype": "string", "id": null, "_type": "Value"}, "flickr_300k_url": {"dtype": "string", "id": null, "_type": "Value"}, "answers": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "image_classes": {"feature": {"dtype": "string", "id": null, "_type": "Value"}, "length": -1, "id": null, "_type": "Sequence"}, "set_name": {"dtype": "string", "id": null, "_type": "Value"}}, "post_processed": null, "supervised_keys": null, "task_templates": null, "builder_name": "textvqa", "config_name": "textvqa", "version": {"version_str": "0.5.1", "description": null, "major": 0, "minor": 5, "patch": 1}, "splits": {"train": {"name": "train", "num_bytes": 22073350, "num_examples": 34602, "dataset_name": "textvqa"}, "validation": {"name": "validation", "num_bytes": 3177854, "num_examples": 5000, "dataset_name": "textvqa"}, "test": {"name": "test", "num_bytes": 3139726, "num_examples": 5734, "dataset_name": "textvqa"}}, "download_checksums": {"https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_train.json": {"num_bytes": 21634937, "checksum": "95f5c407db56cba56a177799dcd685a7cc0ec7c0d851b59910acf7786d31b68a"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_val.json": {"num_bytes": 3116162, "checksum": "4ceb5aadc1a41719d0a3e4dfdf06838bcfee1db569a9a65ee67d31c99893081d"}, "https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_test.json": {"num_bytes": 2770520, "checksum": "d8d4b738101087bac5a6182d22d9aef3772e08e77827e6cf6116808910b75db2"}, "https://dl.fbaipublicfiles.com/textvqa/images/train_val_images.zip": {"num_bytes": 7072297970, "checksum": "ecf35005640d0708eae185aab1c0a10f89b2db7420b29185a1ed92a8f4290498"}, "https://dl.fbaipublicfiles.com/textvqa/images/test_images.zip": {"num_bytes": 970296721, "checksum": "1276b908994c444c46484fb21e9e15fcda1be9c675f6ad727489e52eea68cbcd"}}, "download_size": 8070116310, "post_processing_size": null, "dataset_size": 28390930, "size_in_bytes": 8098507240}}
 
 
textvqa.py DELETED
@@ -1,133 +0,0 @@
1
- # coding=utf-8
2
- # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor.
3
- #
4
- # Licensed under the Apache License, Version 2.0 (the "License");
5
- # you may not use this file except in compliance with the License.
6
- # You may obtain a copy of the License at
7
- #
8
- # http://www.apache.org/licenses/LICENSE-2.0
9
- #
10
- # Unless required by applicable law or agreed to in writing, software
11
- # distributed under the License is distributed on an "AS IS" BASIS,
12
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
- # See the License for the specific language governing permissions and
14
- # limitations under the License.
15
- """TextVQA dataset"""
16
-
17
- import copy
18
- import json
19
- import os
20
-
21
- import datasets
22
-
23
-
24
- _CITATION = """
25
- @inproceedings{singh2019towards,
26
- title={Towards VQA Models That Can Read},
27
- author={Singh, Amanpreet and Natarjan, Vivek and Shah, Meet and Jiang, Yu and Chen, Xinlei and Batra, Dhruv and Parikh, Devi and Rohrbach, Marcus},
28
- booktitle={Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition},
29
- pages={8317-8326},
30
- year={2019}
31
- }
32
- """
33
-
34
- _DESCRIPTION = """\
35
- TextVQA requires models to read and reason about text in images to answer questions about them.
36
- Specifically, models need to incorporate a new modality of text present in the images and reason
37
- over it to answer TextVQA questions. TextVQA dataset contains 45,336 questions over 28,408 images
38
- from the OpenImages dataset.
39
- """
40
-
41
- _HOMEPAGE = "https://textvqa.org"
42
-
43
- _LICENSE = "CC BY 4.0"
44
-
45
- _SPLITS = ["train", "val", "test"]
46
- _URLS = {
47
- f"{split}_annotations": f"https://dl.fbaipublicfiles.com/textvqa/data/TextVQA_0.5.1_{split}.json"
48
- for split in _SPLITS
49
- }
50
- # TextVQA val and train images are packed together
51
- _URLS["train_val_images"] = "https://dl.fbaipublicfiles.com/textvqa/images/train_val_images.zip"
52
- _URLS["test_images"] = "https://dl.fbaipublicfiles.com/textvqa/images/test_images.zip"
53
- _NUM_ANSWERS_PER_QUESTION = 10
54
-
55
-
56
- class Textvqa(datasets.GeneratorBasedBuilder):
57
- """TextVQA dataset."""
58
-
59
- BUILDER_CONFIGS = [
60
- datasets.BuilderConfig(
61
- name="textvqa",
62
- version=datasets.Version("0.5.1"),
63
- description=_DESCRIPTION,
64
- )
65
- ]
66
-
67
- DEFAULT_CONFIG_NAME = "textvqa"
68
-
69
- def _info(self):
70
- features = datasets.Features(
71
- {
72
- "image_id": datasets.Value("string"),
73
- "question_id": datasets.Value("int32"),
74
- "question": datasets.Value("string"),
75
- "question_tokens": datasets.Sequence(datasets.Value("string")),
76
- "image": datasets.Image(),
77
- "image_width": datasets.Value("int32"),
78
- "image_height": datasets.Value("int32"),
79
- "flickr_original_url": datasets.Value("string"),
80
- "flickr_300k_url": datasets.Value("string"),
81
- "answers": datasets.Sequence(datasets.Value("string")),
82
- "image_classes": datasets.Sequence(datasets.Value("string")),
83
- "set_name": datasets.Value("string"),
84
- }
85
- )
86
-
87
- return datasets.DatasetInfo(
88
- description=_DESCRIPTION,
89
- features=features,
90
- supervised_keys=None,
91
- homepage=_HOMEPAGE,
92
- license=_LICENSE,
93
- citation=_CITATION,
94
- )
95
-
96
- def _split_generators(self, dl_manager):
97
- downloaded_files = dl_manager.download_and_extract(_URLS)
98
- return [
99
- datasets.SplitGenerator(
100
- name=datasets.Split.TRAIN,
101
- gen_kwargs={
102
- "annotations_path": downloaded_files["train_annotations"],
103
- "images_path": downloaded_files["train_val_images"],
104
- },
105
- ),
106
- datasets.SplitGenerator(
107
- name=datasets.Split.VALIDATION,
108
- gen_kwargs={
109
- "annotations_path": downloaded_files["val_annotations"],
110
- "images_path": downloaded_files["train_val_images"],
111
- },
112
- ),
113
- datasets.SplitGenerator(
114
- name=datasets.Split.TEST,
115
- gen_kwargs={
116
- "annotations_path": downloaded_files["test_annotations"],
117
- "images_path": downloaded_files["test_images"],
118
- },
119
- ),
120
- ]
121
-
122
- def _generate_examples(self, annotations_path: str, images_path: str):
123
- with open(annotations_path, "r", encoding="utf-8") as f:
124
- data = json.load(f)["data"]
125
- idx = 0
126
- for item in data:
127
- item = copy.deepcopy(item)
128
- item["answers"] = item.get("answers", ["" for _ in range(_NUM_ANSWERS_PER_QUESTION)])
129
- image_id = item["image_id"]
130
- image_subfolder = "train_images" if item["set_name"] != "test" else "test_images"
131
- item["image"] = os.path.join(images_path, image_subfolder, f"{image_id}.jpg")
132
- yield idx, item
133
- idx += 1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
textvqa/textvqa-test-00000-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f7f6a1f2827ede03ad8a24100becd03db24f891232b3d34760440529ddf70ff
3
+ size 337437133
textvqa/textvqa-test-00001-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f61f9e3c36d27d033ab9eb91a84525658dda9b21f0a137af96ebd0c01eaf00c5
3
+ size 332110120
textvqa/textvqa-test-00002-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7580290d14081e8b9e7113f6a3673d1fee3f3bfaa60acbb8b9e520c8f66b12dc
3
+ size 288497292
textvqa/textvqa-train-00000-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3d3cec2dbb976d1e3447b740ce3f16950108e1fe7cacc0c0c36160afb98cd99
3
+ size 347722145
textvqa/textvqa-train-00001-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0585602aab3e2aac0b05cd4974ca5f4d9c7534cf409098645e95ff131d2bca14
3
+ size 342910247
textvqa/textvqa-train-00002-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c10b5e59a5f9c8e3e26823e2aa34e23e912854f1ae6b9fc3d2defb3889bddf96
3
+ size 331112410
textvqa/textvqa-train-00003-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d662ac601be2f006854abd0e623829dba85d48392134688d791cf48466a1b9e2
3
+ size 349212749
textvqa/textvqa-train-00004-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e959e71d45ef3fb5524a2057def9a023165da68e3822f63d699f7ee2720a13f1
3
+ size 338870268
textvqa/textvqa-train-00005-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bd5ae5588e2dccd5b98534dd51973730e19e21bbf1c5a1a098b25da309b1e15
3
+ size 341369415
textvqa/textvqa-train-00006-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e73425fb7eb04b22ecbc325e200986646a3e1ae229104f2d565691266fbad81
3
+ size 436223327
textvqa/textvqa-train-00007-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88e9b55fcf488f318b33aac7bbf047d0d3e546529af52e0355cc889b3076effe
3
+ size 341766128
textvqa/textvqa-train-00008-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c3ec3c9ad10f196b8b7114d96ee5057ed7c8774dde76cd22d5e2c2ea876f86e
3
+ size 331363152
textvqa/textvqa-train-00009-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:00c0b199f409086c199fabcbb7319e72df8461284e733a0fa82f2569de23326c
3
+ size 439368965
textvqa/textvqa-train-00010-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ee12116dfb007b812def0413855ac472c00dcc75d1ae9805279041e5615240e
3
+ size 427573056
textvqa/textvqa-train-00011-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f33649dd9f615228c336b89ce2815ed422a626a7eb02dcd2a4d66f35e2d81e0f
3
+ size 419412129
textvqa/textvqa-train-00012-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b89ca2f50ea7964cafb3eef01600972168d39747406cf23301012346db5bab25
3
+ size 340488455
textvqa/textvqa-train-00013-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d27a699968f859c8714145272f44db98691315543f88e7f2362290a73bd5a3f1
3
+ size 374117145
textvqa/textvqa-train-00014-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d89a8fba441bcf9b23f9d2fb68de94bdff7f0e0e415194d1fd7e13352fc156c1
3
+ size 319256370
textvqa/textvqa-train-00015-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0534e51129b50c558a6121eb22514968519bd69388300dde2f3848a73c791386
3
+ size 558107055
textvqa/textvqa-train-00016-of-00017.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7891a524edb437c919c17c35e1cbb8fc200303c08c359edf3b5898fc91e5d70
3
+ size 143526574
textvqa/textvqa-validation-00000-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ccde2690b7501f869b8f40bf80ec3108c0a864345d9795cc5eceb489357c31c
3
+ size 372708039
textvqa/textvqa-validation-00001-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:290f90392be1b9c3627dea3469657533187dc7c2b5063256d9ea83425dfefbf6
3
+ size 360380383
textvqa/textvqa-validation-00002-of-00003.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bd9a8264169eb876dc1265984e3e7a277c1fb820a83937d83777c6f794c7119
3
+ size 183351948