actdan2016 commited on
Commit
f851e93
1 Parent(s): 2651894

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -476
README.md CHANGED
@@ -1,476 +0,0 @@
1
- ---
2
- annotations_creators:
3
- - found
4
- language_creators:
5
- - found
6
- language:
7
- - en
8
- license:
9
- - cc-by-4.0
10
- multilinguality:
11
- - monolingual
12
- size_categories:
13
- - 10M<n<100M
14
- source_datasets:
15
- - original
16
- task_categories:
17
- - image-to-text
18
- task_ids:
19
- - image-captioning
20
- paperswithcode_id: redcaps
21
- pretty_name: RedCaps
22
- ---
23
-
24
- # Dataset Card for RedCaps
25
-
26
- ## Table of Contents
27
- - [dan Table of Contents](#table-of-contents)
28
- - [dan Dataset Description](#dataset-description)
29
- - [Dataset dan Summary](#dataset-summary)
30
- - [Dataset dan Preprocessing](#dataset-preprocessing)
31
- - [Supported Tasks and Leaderboards](#supported-tasks-and-leaderboards)
32
- - [Languages](#languages)
33
- - [Dataset Structure](#dataset-structure)
34
- - [Data Instances](#data-instances)
35
- - [Data Fields](#data-fields)
36
- - [Data Splits](#data-splits)
37
- - [Dataset Creation](#dataset-creation)
38
- - [Curation Rationale](#curation-rationale)
39
- - [Source Data](#source-data)
40
- - [Annotations](#annotations)
41
- - [Personal and Sensitive Information](#personal-and-sensitive-information)
42
- - [Considerations for Using the Data](#considerations-for-using-the-data)
43
- - [Social Impact of Dataset](#social-impact-of-dataset)
44
- - [Discussion of Biases](#discussion-of-biases)
45
- - [Other Known Limitations](#other-known-limitations)
46
- - [Additional Information](#additional-information)
47
- - [Dataset Curators](#dataset-curators)
48
- - [Licensing Information](#licensing-information)
49
- - [Citation Information](#citation-information)
50
- - [Contributions](#contributions)
51
-
52
- ## Dataset Description
53
-
54
- - **Homepage:** [RedCaps homepage](https://redcaps.xyz/)
55
- - **Repository:** [RedCaps repository](https://github.com/redcaps-dataset/redcaps-downloader)
56
- - **Paper:** [RedCaps: web-curated image-text data created by the people, for the people](https://arxiv.org/abs/2111.11431)
57
- - **Leaderboard:**
58
- - **Point of Contact:** [Karan Desai](mailto:kdexd@umich.edu)
59
-
60
- ### Dataset Summary
61
-
62
- RedCaps is a large-scale dataset of 12M image-text pairs collected from Reddit.
63
- Images and captions from Reddit depict and describe a wide variety of objects and scenes.
64
- The data is collected from a manually curated set of subreddits (350 total),
65
- which give coarse image labels and allow steering of the dataset composition
66
- without labeling individual instances. RedCaps data is created *by the people, for the people* – it contains everyday things that users like to share on social media, for example hobbies (r/crafts) and pets (r/shiba). Captions often contain specific and
67
- fine-grained descriptions (northern cardinal, taj mahal). Subreddit names provide relevant image
68
- labels (r/shiba) even when captions may not (mlem!), and sometimes may group many visually
69
- unrelated images through a common semantic meaning (r/perfectfit).
70
-
71
- ### Dataset Preprocessing
72
-
73
- This dataset doesn't download the images locally by default. Instead, it exposes URLs to the images. To fetch the images, use the following code:
74
-
75
- ```python
76
- from concurrent.futures import ThreadPoolExecutor
77
- from functools import partial
78
- import io
79
- import urllib
80
-
81
- import PIL.Image
82
-
83
- from datasets import load_dataset
84
- from datasets.utils.file_utils import get_datasets_user_agent
85
-
86
-
87
- USER_AGENT = get_datasets_user_agent()
88
-
89
-
90
- def fetch_single_image(image_url, timeout=None, retries=0):
91
- for _ in range(retries + 1):
92
- try:
93
- request = urllib.request.Request(
94
- image_url,
95
- data=None,
96
- headers={"user-agent": USER_AGENT},
97
- )
98
- with urllib.request.urlopen(request, timeout=timeout) as req:
99
- image = PIL.Image.open(io.BytesIO(req.read()))
100
- break
101
- except Exception:
102
- image = None
103
- return image
104
-
105
-
106
- def fetch_images(batch, num_threads, timeout=None, retries=0):
107
- fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries)
108
- with ThreadPoolExecutor(max_workers=num_threads) as executor:
109
- batch["image"] = list(executor.map(fetch_single_image_with_args, batch["image_url"]))
110
- return batch
111
-
112
-
113
- num_threads = 20
114
- dset = load_dataset("red_caps", "rabbits_2017")
115
- dset = dset.map(fetch_images, batched=True, batch_size=100, fn_kwargs={"num_threads": num_threads})
116
- ```
117
-
118
- Some image links point to more than one image. You can process and downloaded those as follows:
119
-
120
- ```python
121
- from concurrent.futures import ThreadPoolExecutor
122
- from functools import partial
123
- import io
124
- import os
125
- import re
126
- import urllib
127
-
128
- import PIL.Image
129
-
130
- import datasets
131
- from datasets import load_dataset
132
- from datasets.utils.file_utils import get_datasets_user_agent
133
-
134
-
135
- USER_AGENT = get_datasets_user_agent()
136
-
137
-
138
- def fetch_single_image(image_url, timeout=None, retries=0):
139
- for _ in range(retries + 1):
140
- try:
141
- request = urllib.request.Request(
142
- image_url,
143
- data=None,
144
- headers={"user-agent": USER_AGENT},
145
- )
146
- with urllib.request.urlopen(request, timeout=timeout) as req:
147
- image = PIL.Image.open(io.BytesIO(req.read()))
148
- break
149
- except Exception:
150
- image = None
151
- return image
152
-
153
-
154
- def fetch_images(batch, num_threads, timeout=None, retries=0):
155
- fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries)
156
- with ThreadPoolExecutor(max_workers=num_threads) as executor:
157
- batch["image"] = list(executor.map(lambda image_urls: [fetch_single_image_with_args(image_url) for image_url in image_urls], batch["image_url"]))
158
- return batch
159
-
160
-
161
- def process_image_urls(batch):
162
- processed_batch_image_urls = []
163
- for image_url in batch["image_url"]:
164
- processed_example_image_urls = []
165
- image_url_splits = re.findall(r"http\S+", image_url)
166
- for image_url_split in image_url_splits:
167
- if "imgur" in image_url_split and "," in image_url_split:
168
- for image_url_part in image_url_split.split(","):
169
- if not image_url_part:
170
- continue
171
- image_url_part = image_url_part.strip()
172
- root, ext = os.path.splitext(image_url_part)
173
- if not root.startswith("http"):
174
- root = "http://i.imgur.com/" + root
175
- root = root.split("#")[0]
176
- if not ext:
177
- ext = ".jpg"
178
- ext = re.split(r"[?%]", ext)[0]
179
- image_url_part = root + ext
180
- processed_example_image_urls.append(image_url_part)
181
- else:
182
- processed_example_image_urls.append(image_url_split)
183
- processed_batch_image_urls.append(processed_example_image_urls)
184
- batch["image_url"] = processed_batch_image_urls
185
- return batch
186
-
187
-
188
- dset = load_dataset("red_caps", "rabbits_2017")
189
- dset = dset.map(process_image_urls, batched=True, num_proc=4)
190
- features = dset["train"].features.copy()
191
- features["image"] = datasets.Sequence(datasets.Image())
192
- num_threads = 20
193
- dset = dset.map(fetch_images, batched=True, batch_size=100, features=features, fn_kwargs={"num_threads": num_threads})
194
- ```
195
-
196
- Note that in the above code, we use the `datasets.Sequence` feature to represent a list of images for the multi-image links.
197
-
198
- ### Supported Tasks and Leaderboards
199
-
200
- From the paper:
201
- > We have used our dataset to train deep neural networks that perform image captioning, and
202
- that learn transferable visual representations for a variety of downstream visual recognition tasks
203
- (image classification, object detection, instance segmentation).
204
-
205
- > We anticipate that the dataset could be used for a variety of vision-and-language (V&L) tasks,
206
- such as image or text retrieval or text-to-image synthesis.
207
-
208
- ### Languages
209
-
210
- All of the subreddits in RedCaps use English as their primary language.
211
-
212
- ## Dataset Structure
213
-
214
- ### Data Instances
215
-
216
- Each instance in RedCaps represents a single Reddit image post:
217
-
218
- ```
219
- {
220
- 'image_id': 'bpzj7r',
221
- 'author': 'djasz1',
222
- 'image_url': 'https://i.redd.it/ho0wntksivy21.jpg',
223
- 'raw_caption': 'Found on a friend’s property in the Keys FL. She is now happily living in my house.',
224
- 'caption': 'found on a friend's property in the keys fl. she is now happily living in my house.', 'subreddit': 3,
225
- 'score': 72,
226
- 'created_utc': datetime.datetime(2019, 5, 18, 1, 36, 41),
227
- 'permalink': '/r/airplants/comments/bpzj7r/found_on_a_friends_property_in_the_keys_fl_she_is/', 'crosspost_parents': None
228
- }
229
- ```
230
-
231
- ### Data Fields
232
-
233
- - `image_id`: Unique alphanumeric ID of the image post (assigned by Reddit).
234
- - `author`: Reddit username of the image post author.
235
- - `image_url`: Static URL for downloading the image associated with the post.
236
- - `raw_caption`: Textual description of the image, written by the post author.
237
- - `caption`: Cleaned version of "raw_caption" by us (see Q35).
238
- - `subreddit`: Name of subreddit where the post was submitted.
239
- - `score`: Net upvotes (discounting downvotes) received by the image post. This field is equal to `None` if the image post is a crosspost.
240
- - `created_utc`: Integer time epoch (in UTC) when the post was submitted to Reddit.
241
- - `permalink`: Partial URL of the Reddit post (https://reddit.com/<permalink>).
242
- - `crosspost_parents`: List of parent posts. This field is optional.
243
-
244
-
245
- ### Data Splits
246
-
247
- All the data is contained in training set. The training set has nearly 12M (12,011,111) instances.
248
-
249
- From the paper:
250
- > We intend our dataset to be primarily used for pre-training with one or more specific downstream task(s) in mind. Hence, all instances in our dataset would be used for training while
251
- the validation split is derived from downstream task(s). If users require a validation split, we
252
- recommend sampling it such that it follows the same subreddit distribution as entire dataset.
253
-
254
- ## Dataset Creation
255
-
256
- ### Curation Rationale
257
-
258
- From the paper:
259
- > Large datasets of image-text pairs are widely used for pre-training generic representations
260
- that transfer to a variety of downstream vision and vision-and-language tasks. Existing public
261
- datasets of this kind were curated from search engine results (SBU Captions [1]) or HTML
262
- alt-text from arbitrary web pages (Conceptual Captions [2, 31]). They performed complex
263
- data filtering to deal with noisy web data. Due to aggressive filtering, their data collection is
264
- inefficient and diversity is artificially supressed. We argue that the quality of data depends on
265
- its source, and the human intent behind its creation. In this work, we explore Reddit – a social
266
- media platform, for curating high quality data. We introduce RedCaps – a large dataset of
267
- 12M image-text pairs from Reddit. While we expect the use-cases of RedCaps to be similar to
268
- existing datasets, we discuss how Reddit as a data source leads to fast and lightweight collection,
269
- better data quality, lets us easily steer the data distribution, and facilitates ethically responsible data curation.
270
-
271
- ### Source Data
272
-
273
- #### Initial Data Collection and Normalization
274
-
275
- From the paper:
276
- > **Data Collection Pipeline**
277
- Reddit’s uniform structure allows us to parallelize data collection as independent tasks – each task
278
- involves collecting posts submitted to a single subreddit in one year. Our collection pipeline has three steps: (1) subreddit selection, (2) image post filtering, and (3) caption cleaning.
279
- **Step 1**. Subreddit selection: We collect data from a manually curated set of subreddits. Subreddits
280
- have their own rules, community norms, and moderators so curating subreddits allows us to steer the
281
- dataset’s composition without annotating individual instances. We select subreddits with a high volume of images posts, where images tend to be photographs (rather than memes, drawings, screenshots,
282
- etc) and post titles tend to describe image content (rather than making jokes, political commentary,
283
- etc). We do not select any NSFW, banned, or quarantined subreddits. We want to minimize the
284
- number of people that appear in RedCaps, so we omit subreddits whose primary purpose is to share or
285
- comment on images of people (such as celebrity pics or user selfies). We choose subreddits focused on
286
- general photography (r/pics, r/itookapicture), animals (r/axolotls, r/birdsofprey, r/dachshund),
287
- plants (r/roses, r/succulents), objects (r/classiccars, r/trains, r/mechanicalkeyboards), food
288
- (r/steak, r/macarons), scenery (r/cityporn1
289
- , r/desertporn), or activities (r/carpentry, r/kayaking).
290
- In total we collect data from 350 subreddits; the full list can be found in Appendix A.
291
- **Step 2**. Image post filtering: We use Pushshift [41] and Reddit [42, 43] APIs to download all image
292
- posts submitted to our selected subreddits from 2008–2020. Posts are collected at least six months
293
- after their creation to let upvotes stabilize. We only collect posts with images hosted on three domains:
294
- Reddit (i.redd.it), Imgur (i.imgur.com), and Flickr (staticflickr.com). Some image posts contain
295
- multiple images (gallery posts) – in this case we only collect the first image and associate it with
296
- the caption. We discard posts with < 2 upvotes to avoid unappealing content, and we discard posts
297
- marked NSFW (by their authors or subreddit moderators) to avoid pornographic or disturbing content.
298
- **Step 3**. Caption cleaning: We expect Reddit post titles to be less noisy than other large-scale
299
- sources of image captions such as alt-text [2, 31], so we apply minimal text cleaning. We lowercase
300
- captions and use ftfy [44] to remove character accents, emojis, and non-latin characters, following
301
- [29, 35, 36]. Then we apply simple pattern matching to discard all sub-strings enclosed in brackets
302
- ((.*), [.*]). These sub-strings usually give non-semantic information: original content tags [oc],
303
- image resolutions (800x600 px), camera specs (shot with iPhone), self-promotion [Instagram:
304
- @user], and other references (link in comments). Finally, like [31] we replace social media
305
- handles (words starting with ‘@’) with a [USR] token to protect user privacy and reduce redundancy.
306
- Due to such filtering, ≈12K (0.1%) captions in our dataset are empty strings. We do not discard them,
307
- as subreddit names alone provide meaningful supervision. Unlike CC-3M or CC-12M that discard
308
- captions without nouns or that don’t overlap image tags, we do not discard any instances in this step.
309
- Through this pipeline, we collect 13.4M instances from 350 subreddits. Our collection pipeline is
310
- less resource-intensive than existing datasets – we do not require webpage crawlers, search engines,
311
- or large databases of indexed webpages. RedCaps is easily extensible in the future by selecting more
312
- subreddits and collecting posts from future years. Next, we perform additional filtering to mitigate
313
- user privacy risks and harmful stereotypes in RedCaps, resulting in final size of 12M instances.
314
-
315
- #### Who are the source language producers?
316
-
317
- Reddit is the singular data source for RedCaps.
318
-
319
- ### Annotations
320
-
321
- #### Annotation process
322
-
323
- The dataset is built using fully automatic data collection pipeline which doesn't require any human annotators.
324
-
325
- #### Who are the annotators?
326
-
327
- The annotation process doesn't require any human annotators.
328
-
329
- ### Personal and Sensitive Information
330
-
331
- From the paper:
332
- > **Does the dataset relate to people?**
333
- The dataset pertains to people in that people wrote the captions and posted images to Reddit
334
- that we curate in RedCaps. We made specific design choices while curating RedCaps to avoid
335
- large quantities of images containing people:
336
- (a) We collect data from manually curated subreddits in which most contain primarily pertains
337
- to animals, objects, places, or activities. We exclude all subreddits whose primary purpose
338
- is to share and describe images of people (such as celebrity photos or user selfies).
339
- (b) We use an off-the-shelf face detector to find and remove images with potential presence of
340
- human faces. We manually checked 50K random images in RedCaps (Q16) and found 79
341
- images with identifiable human faces – the entire dataset may have ≈19K (0.15%) images
342
- with identifiable people. Refer Section 2.2 in the main paper.
343
-
344
- > **Is it possible to identify one or more natural persons, either directly or indirectly (i.e., in
345
- combination with other data) from the dataset?**
346
- Yes, all instances in RedCaps include Reddit usernames of their post authors. This could be
347
- used to look up the Reddit user profile, and some Reddit users may have identifying information
348
- in their profiles. Some images may contain human faces which could be identified by
349
- appearance. However, note that all this information is already public on Reddit, and searching it
350
- in RedCaps is no easier than searching directly on Reddit.
351
-
352
- > **Were the individuals in question notified about the data collection?**
353
- No. Reddit users are anonymous by default, and are not required to share their personal contact
354
- information (email, phone numbers, etc.). Hence, the only way to notify the authors of RedCaps
355
- image posts is by sending them private messages on Reddit. This is practically difficult to do
356
- manually, and will be classified as spam and blocked by Reddit if attempted to programmatically
357
- send a templated message to millions of users.
358
-
359
- > **Did the individuals in question consent to the collection and use of their data?**
360
- Users did not explicitly consent to the use of their data in our dataset. However, by uploading
361
- their data on Reddit, they consent that it would appear on the Reddit plaform and will be
362
- accessible via the official Reddit API (which we use to collect RedCaps).
363
-
364
- > **If consent was obtained, were the consenting individuals provided with a mechanism to
365
- revoke their consent in the future or for certain uses?**
366
- Users have full control over the presence of their data in our dataset. If users wish to revoke
367
- their consent, they can delete the underlying Reddit post – it will be automatically removed
368
- dfrom RedCaps since we distributed images as URLs. Moreover, we provide an opt-out request
369
- form on our dataset website for anybody to request removal of an individual instance if it is
370
- potentially harmful (e.g. NSFW, violates privacy, harmful stereotypes, etc.).
371
-
372
- ## Considerations for Using the Data
373
-
374
- ### Social Impact of Dataset
375
-
376
- From the paper:
377
- > **Has an analysis of the potential impact of the dataset and its use on data subjects (e.g.,
378
- a data protection impact analysis) been conducted?**
379
- No.
380
-
381
- ### Discussion of Biases
382
-
383
- From the paper:
384
- > **Harmful Stereotypes**: Another concern with
385
- Reddit data is that images or language may represent harmful stereotypes about gender, race, or other
386
- characteristics of people [48, 49, 51]. We select only non-NSFW subreddits with active moderation
387
- for collecting data. This stands in contrast to less curated uses of Reddit data, such as GPT-2 [35]
388
- whose training data includes at least 63K documents from banned or quarantined subreddits which
389
- may contain toxic language [53]. We attempt to further reduce harmful stereotypes in two ways:
390
- > * **NSFW images**: We use the InceptionV3 [54] model from [55] to filter images detected as porn or hentai with confidence ≥ 0.9. Similar to face filtering, we estimated precision of our filtering and estimated amount of missed detections, shown in Table 1. The model detects 87K images with low
391
- precision (∼1%) – most detections are non-NSFW images with pink and beige hues.
392
- > * **Potentially derogatory language**: We filter instances whose captions contain words or phrases from a common blocklist [56]. It is important to note that such coarse filtering might suppress language from marginalized groups reclaiming slurs [51]; however, as RedCaps is not intended to describe people, we believe this is a pragmatic tradeoff to avoid propagating harmful labels.
393
-
394
- > **Reddit demographics**: Reddit’s user demographics are not representative of the population at large.
395
- Compared to US adults, Reddit users skew male (69% vs 49%), young (58% 18-29 years old vs
396
- 22%), college educated (36% vs 28%), and politically liberal (41% vs 25%) [57]. Reddit users
397
- are predominantly white (63%) [57], and 49% of desktop traffic to Reddit comes from the United
398
- States [58]. All of the subreddits in RedCaps use English as their primary language. Taken together,
399
- these demographic biases likely also bias the types of objects and places that appear in images on
400
- Reddit, and the language used to describe these images. We do not offer explicit countermeasures to
401
- these biases, but users of RedCaps should keep in mind that size doesn’t guarantee diversity [51].
402
- Subtler issues may also exist, such as imbalanced representation of demographic groups [59] or
403
- gender bias in object co-occurrence [60] or language [61]. These are hard to control in internet
404
- data, so we release RedCaps with explicit instructions on suitable use-cases; specifically requesting models not be trained to identify people, or make decisions that impact people. We document these instructions and other terms-of-use in a datasheet [45], provided in Appendix G.
405
-
406
- > **Does the dataset contain data that, if viewed directly, might be offensive, insulting, threatening, or might otherwise cause anxiety?**
407
- The scale of RedCaps means that we are unable to verify the contents of all images and
408
- captions. However we have tried to minimize the possibility that RedCaps contains data that
409
- might be offensive, insulting, threatening, or might cause anxiety via the following mitigations:
410
- (a) We manually curate the set of subreddits from which to collect data; we only chose
411
- subreddits that are not marked NSFW and which generally contain non-offensive content.
412
- (b) Within our curated subreddits, we did not include any posts marked NSFW.
413
- (c) We removed all instances whose captions contained any of the 400 potentially offensive
414
- words or phrases. Refer Section 2.2 in the main paper.
415
- (d) We remove all instances whose images were flagged NSFW by an off-the-shelf detector.
416
- We manually checked 50K random images in RedCaps and found one image containing
417
- nudity (exposed buttocks; no identifiable face). Refer Section 2.2 in the main paper
418
-
419
- > **Does the dataset identify any subpopulations (e.g., by age, gender)?**
420
- RedCaps does not explicitly identify any subpopulations. Since some images contain people
421
- and captions are free-form natural language written by Reddit users, it is possible that some
422
- captions may identify people appearing in individual images as part of a subpopulation.
423
-
424
- > **Were any ethical review processes conducted (e.g., by an institutional review board)?**
425
- We did not conduct a formal ethical review process via institutional review boards. However,
426
- as described in Section 2.2 of the main paper and Q16 we employed several filtering mechanisms
427
- to try and remove instances that could be problematic.
428
-
429
- ### Other Known Limitations
430
-
431
- From the paper:
432
- > **Are there any errors, sources of noise, or redundancies in the dataset?**
433
- RedCaps is noisy by design since image-text pairs on the internet are noisy and unstructured.
434
- Some instances may also have duplicate images and captions – Reddit users may have shared
435
- the same image post in multiple subreddits. Such redundancies constitute a very small fraction
436
- of the dataset, and should have almost no effect in training large-scale models.
437
-
438
- > **Does the dataset contain data that might be considered confidential (e.g., data that is
439
- protected by legal privilege or by doctor-patient confidentiality, data that includes the
440
- content of individuals non-public communications)?**
441
- No, the subreddits included in RedCaps do not cover topics that may be considered confidential. All posts were publicly shared on Reddit prior to inclusion in RedCaps.
442
-
443
- ## Additional Information
444
-
445
- ### Dataset Curators
446
-
447
- From the paper:
448
- > Four researchers at the University of Michigan (affiliated as of 2021) have created RedCaps:
449
- Karan Desai, Gaurav Kaul, Zubin Aysola, and Justin Johnson.
450
-
451
- ### Licensing Information
452
-
453
- The image metadata is licensed under CC-BY 4.0 license. Additionally, uses of this dataset are subject to Reddit API terms (https://www.reddit.com/wiki/
454
- api-terms) and users must comply with Reddit User Agreeement, Content Policy,
455
- and Privacy Policy – all accessible at https://www.redditinc.com/policies.
456
-
457
- From the paper:
458
- > RedCaps should only be used for non-commercial research. RedCaps should not be used for any tasks that involve identifying features related to people (facial recognition, gender, age, ethnicity identification, etc.) or make decisions that impact people (mortgages, job applications, criminal sentences; or moderation decisions about user-uploaded data that could result in bans from a website). Any commercial and for-profit uses of RedCaps are restricted – it should not be used to train models that will be deployed in production systems as part of a product offered by businesses or government agencies.
459
-
460
-
461
- ### Citation Information
462
-
463
- ```bibtex
464
- @misc{desai2021redcaps,
465
- title={RedCaps: web-curated image-text data created by the people, for the people},
466
- author={Karan Desai and Gaurav Kaul and Zubin Aysola and Justin Johnson},
467
- year={2021},
468
- eprint={2111.11431},
469
- archivePrefix={arXiv},
470
- primaryClass={cs.CV}
471
- }
472
- ```
473
-
474
- ### Contributions
475
-
476
- Thanks to [@mariosasko](https://github.com/mariosasko) for adding this dataset.