|
--- |
|
task_categories: |
|
- visual-question-answering |
|
language: |
|
- en |
|
pretty_name: YTTB-VQA |
|
size_categories: |
|
- n<1K |
|
license: cc-by-nc-4.0 |
|
--- |
|
# Dataset Card for Dataset Name |
|
|
|
## Dataset Description |
|
|
|
- **Homepage:** https://gordonhu608.github.io/bliva/ |
|
- **Repository:** https://github.com/mlpc-ucsd/BLIVA.git |
|
- **Paper:** |
|
- **Point of Contact:** w1hu@ucsd.edu |
|
|
|
### Dataset Summary |
|
|
|
The YTTB-VQA Dataset is a collection of 400 Youtube thumbnail question-answer pairs to evaluate the visual perception abilities of in-text images. It covers 11 |
|
categories, including technology, sports, entertainment, food, news, history, music, nature, cars, and education. |
|
|
|
### Supported Tasks and Leaderboards |
|
|
|
This dataset supports many tasks, including visual question answering, image captioning, etc. |
|
|
|
### License |
|
CC-By-NC-4.0 |
|
|
|
### Languages |
|
|
|
The language of the data is primarily English. |
|
|
|
## Getting Started |
|
|
|
### Creating the dataset |
|
|
|
Run the following command to download the images and create the dataset: |
|
|
|
```python3 create_dataset.py``` |
|
|
|
You will find the images in `images_new` and the dataset in `youtube_new.json`. |
|
|
|
## Dataset Structure |
|
|
|
### Data Instances |
|
|
|
A data instance in this dataset represents entries from a collection augmented by human-generated questions submitted to BLIVA. The answer is then entered into the answer field. |
|
|
|
### Data Fields |
|
|
|
**video_id:** a unique string representing a specific YouTube thumbnail image.<br> |
|
**question:** representing a human-generated question.<br> |
|
**video_classes:** representing a specific category for the YouTube thumbnail image.<br> |
|
**answers:** This represents a ground truth answer for the question made about the YouTube thumbnail image.<br> |
|
**video link** Representing the URL link for each YouTube video. |
|
|
|
### Data Splits |
|
|
|
The data are unsplit. |
|
|
|
## Dataset Creation |
|
|
|
### Source Data |
|
|
|
#### Initial Data Collection and Normalization |
|
|
|
We randomly selected YouTube videos with text-rich thumbnails from different categories during the data collection. |
|
We recorded the unique video ID for each YouTube video and obtained the high-resolution thumbnail from the |
|
URL ”http://img.youtube.com/vi/YouTube-Video-ID/maxresdefault.jpg”. |
|
|
|
### Annotations |
|
|
|
#### Annotation process |
|
|
|
We created the annotation file with the following fields: ”video id,” question,” video classes,” answers,” and ”video link" in JSON format. |
|
|
|
## Considerations for Using the Data |
|
|
|
### Discussion of Biases |
|
|
|
Although our dataset spans 11 categories, the ratio within each category varies. For example, 18% of the dataset pertains to education, while only 2% is dedicated to news. |
|
|
|
### Acknowledgments |
|
|
|
The youtube thumbnails dataset is purely for academic research and not for any monetary uses. For any of the authors who saw our dataset and found their thumbnail images used inappropriately, please get in touch with us directly by this email at w1hu@ucsd.edu and we will remove the image immediately. |